[tip:x86/vdso] x86/vdso: Use unsigned int consistently for vsyscall_gtod_data:: Seq

From: tip-bot for Thomas Gleixner
Date: Thu Oct 04 2018 - 17:06:26 EST


Commit-ID: 77e9c678c54f2d9214796c1c5bd0c7c7ccedd932
Gitweb: https://git.kernel.org/tip/77e9c678c54f2d9214796c1c5bd0c7c7ccedd932
Author: Thomas Gleixner <tglx@xxxxxxxxxxxxx>
AuthorDate: Mon, 17 Sep 2018 14:45:37 +0200
Committer: Thomas Gleixner <tglx@xxxxxxxxxxxxx>
CommitDate: Thu, 4 Oct 2018 23:00:25 +0200

x86/vdso: Use unsigned int consistently for vsyscall_gtod_data:: Seq

The sequence count in vgtod_data is unsigned int, but the call sites use
unsigned long, which is a pointless exercise. Fix the call sites and
replace 'unsigned' with unsinged 'int' while at it.

Signed-off-by: Thomas Gleixner <tglx@xxxxxxxxxxxxx>
Acked-by: Andy Lutomirski <luto@xxxxxxxxxx>
Cc: Peter Zijlstra <peterz@xxxxxxxxxxxxx>
Cc: Matt Rickard <matt@xxxxxxxxxxxxxxx>
Cc: Stephen Boyd <sboyd@xxxxxxxxxx>
Cc: John Stultz <john.stultz@xxxxxxxxxx>
Cc: Florian Weimer <fweimer@xxxxxxxxxx>
Cc: "K. Y. Srinivasan" <kys@xxxxxxxxxxxxx>
Cc: Vitaly Kuznetsov <vkuznets@xxxxxxxxxx>
Cc: devel@xxxxxxxxxxxxxxxxxxxxxx
Cc: virtualization@xxxxxxxxxxxxxxxxxxxxxxxxxx
Cc: Paolo Bonzini <pbonzini@xxxxxxxxxx>
Cc: Arnd Bergmann <arnd@xxxxxxxx>
Cc: Juergen Gross <jgross@xxxxxxxx>
Link: https://lkml.kernel.org/r/20180917130707.236250416@xxxxxxxxxxxxx

---
arch/x86/entry/vdso/vclock_gettime.c | 8 ++++----
arch/x86/include/asm/vgtod.h | 10 +++++-----
2 files changed, 9 insertions(+), 9 deletions(-)

diff --git a/arch/x86/entry/vdso/vclock_gettime.c b/arch/x86/entry/vdso/vclock_gettime.c
index 6a950854034f..b50ee064beff 100644
--- a/arch/x86/entry/vdso/vclock_gettime.c
+++ b/arch/x86/entry/vdso/vclock_gettime.c
@@ -208,7 +208,7 @@ notrace static inline u64 vgetsns(int *mode)
/* Code size doesn't matter (vdso is 4k anyway) and this is faster. */
notrace static int __always_inline do_realtime(struct timespec *ts)
{
- unsigned long seq;
+ unsigned int seq;
u64 ns;
int mode;

@@ -229,7 +229,7 @@ notrace static int __always_inline do_realtime(struct timespec *ts)

notrace static int __always_inline do_monotonic(struct timespec *ts)
{
- unsigned long seq;
+ unsigned int seq;
u64 ns;
int mode;

@@ -250,7 +250,7 @@ notrace static int __always_inline do_monotonic(struct timespec *ts)

notrace static void do_realtime_coarse(struct timespec *ts)
{
- unsigned long seq;
+ unsigned int seq;
do {
seq = gtod_read_begin(gtod);
ts->tv_sec = gtod->wall_time_coarse_sec;
@@ -260,7 +260,7 @@ notrace static void do_realtime_coarse(struct timespec *ts)

notrace static void do_monotonic_coarse(struct timespec *ts)
{
- unsigned long seq;
+ unsigned int seq;
do {
seq = gtod_read_begin(gtod);
ts->tv_sec = gtod->monotonic_time_coarse_sec;
diff --git a/arch/x86/include/asm/vgtod.h b/arch/x86/include/asm/vgtod.h
index 53748541c487..91cad1f01027 100644
--- a/arch/x86/include/asm/vgtod.h
+++ b/arch/x86/include/asm/vgtod.h
@@ -15,9 +15,9 @@ typedef unsigned long gtod_long_t;
* so be carefull by modifying this structure.
*/
struct vsyscall_gtod_data {
- unsigned seq;
+ unsigned int seq;

- int vclock_mode;
+ int vclock_mode;
u64 cycle_last;
u64 mask;
u32 mult;
@@ -44,9 +44,9 @@ static inline bool vclock_was_used(int vclock)
return READ_ONCE(vclocks_used) & (1 << vclock);
}

-static inline unsigned gtod_read_begin(const struct vsyscall_gtod_data *s)
+static inline unsigned int gtod_read_begin(const struct vsyscall_gtod_data *s)
{
- unsigned ret;
+ unsigned int ret;

repeat:
ret = READ_ONCE(s->seq);
@@ -59,7 +59,7 @@ repeat:
}

static inline int gtod_read_retry(const struct vsyscall_gtod_data *s,
- unsigned start)
+ unsigned int start)
{
smp_rmb();
return unlikely(s->seq != start);