Loading arch/tile/include/asm/vdso.h +3 −2 Original line number Original line Diff line number Diff line Loading @@ -15,6 +15,7 @@ #ifndef __TILE_VDSO_H__ #ifndef __TILE_VDSO_H__ #define __TILE_VDSO_H__ #define __TILE_VDSO_H__ #include <linux/seqlock.h> #include <linux/types.h> #include <linux/types.h> /* /* Loading @@ -26,8 +27,8 @@ */ */ struct vdso_data { struct vdso_data { __u64 tz_update_count; /* Timezone atomicity ctr */ seqcount_t tz_seq; /* Timezone seqlock */ __u64 tb_update_count; /* Timebase atomicity ctr */ seqcount_t tb_seq; /* Timebase seqlock */ __u64 xtime_tod_stamp; /* TOD clock for xtime */ __u64 xtime_tod_stamp; /* TOD clock for xtime */ __u64 xtime_clock_sec; /* Kernel time second */ __u64 xtime_clock_sec; /* Kernel time second */ __u64 xtime_clock_nsec; /* Kernel time nanosecond */ __u64 xtime_clock_nsec; /* Kernel time nanosecond */ Loading arch/tile/kernel/time.c +6 −10 Original line number Original line Diff line number Diff line Loading @@ -249,13 +249,10 @@ cycles_t ns2cycles(unsigned long nsecs) void update_vsyscall_tz(void) void update_vsyscall_tz(void) { { /* Userspace gettimeofday will spin while this value is odd. */ write_seqcount_begin(&vdso_data->tz_seq); ++vdso_data->tz_update_count; smp_wmb(); vdso_data->tz_minuteswest = sys_tz.tz_minuteswest; vdso_data->tz_minuteswest = sys_tz.tz_minuteswest; vdso_data->tz_dsttime = sys_tz.tz_dsttime; vdso_data->tz_dsttime = sys_tz.tz_dsttime; smp_wmb(); write_seqcount_end(&vdso_data->tz_seq); ++vdso_data->tz_update_count; } } void update_vsyscall(struct timekeeper *tk) void update_vsyscall(struct timekeeper *tk) Loading @@ -266,9 +263,8 @@ void update_vsyscall(struct timekeeper *tk) if (clock != &cycle_counter_cs) if (clock != &cycle_counter_cs) return; return; /* Userspace gettimeofday will spin while this value is odd. */ write_seqcount_begin(&vdso_data->tb_seq); ++vdso_data->tb_update_count; smp_wmb(); vdso_data->xtime_tod_stamp = tk->tkr.cycle_last; vdso_data->xtime_tod_stamp = tk->tkr.cycle_last; vdso_data->xtime_clock_sec = tk->xtime_sec; vdso_data->xtime_clock_sec = tk->xtime_sec; vdso_data->xtime_clock_nsec = tk->tkr.xtime_nsec; vdso_data->xtime_clock_nsec = tk->tkr.xtime_nsec; Loading @@ -276,6 +272,6 @@ void update_vsyscall(struct timekeeper *tk) vdso_data->wtom_clock_nsec = wtm->tv_nsec; vdso_data->wtom_clock_nsec = wtm->tv_nsec; vdso_data->mult = tk->tkr.mult; vdso_data->mult = tk->tkr.mult; vdso_data->shift = tk->tkr.shift; vdso_data->shift = tk->tkr.shift; smp_wmb(); ++vdso_data->tb_update_count; write_seqcount_end(&vdso_data->tb_seq); } } arch/tile/kernel/vdso/vgettimeofday.c +15 −32 Original line number Original line Diff line number Diff line Loading @@ -53,50 +53,33 @@ inline unsigned long get_datapage(void) int __vdso_gettimeofday(struct timeval *tv, struct timezone *tz) int __vdso_gettimeofday(struct timeval *tv, struct timezone *tz) { { cycles_t cycles; cycles_t cycles; unsigned long count, sec, ns; unsigned count; volatile struct vdso_data *vdso_data; unsigned long sec, ns; struct vdso_data *vdso = (struct vdso_data *)get_datapage(); vdso_data = (struct vdso_data *)get_datapage(); /* The use of the timezone is obsolete, normally tz is NULL. */ /* The use of the timezone is obsolete, normally tz is NULL. */ if (unlikely(tz != NULL)) { if (unlikely(tz != NULL)) { while (1) { do { /* Spin until the update finish. */ count = read_seqcount_begin(&vdso->tz_seq); count = vdso_data->tz_update_count; tz->tz_minuteswest = vdso->tz_minuteswest; if (count & 1) tz->tz_dsttime = vdso->tz_dsttime; continue; } while (unlikely(read_seqcount_retry(&vdso->tz_seq, count))); tz->tz_minuteswest = vdso_data->tz_minuteswest; tz->tz_dsttime = vdso_data->tz_dsttime; /* Check whether updated, read again if so. */ if (count == vdso_data->tz_update_count) break; } } } if (unlikely(tv == NULL)) if (unlikely(tv == NULL)) return 0; return 0; while (1) { do { /* Spin until the update finish. */ count = read_seqcount_begin(&vdso->tb_seq); count = vdso_data->tb_update_count; sec = vdso->xtime_clock_sec; if (count & 1) cycles = get_cycles() - vdso->xtime_tod_stamp; continue; ns = (cycles * vdso->mult) + vdso->xtime_clock_nsec; ns >>= vdso->shift; sec = vdso_data->xtime_clock_sec; cycles = get_cycles() - vdso_data->xtime_tod_stamp; ns = (cycles * vdso_data->mult) + vdso_data->xtime_clock_nsec; ns >>= vdso_data->shift; if (ns >= NSEC_PER_SEC) { if (ns >= NSEC_PER_SEC) { ns -= NSEC_PER_SEC; ns -= NSEC_PER_SEC; sec += 1; sec += 1; } } } while (unlikely(read_seqcount_retry(&vdso->tb_seq, count))); /* Check whether updated, read again if so. */ if (count == vdso_data->tb_update_count) break; } tv->tv_sec = sec; tv->tv_sec = sec; tv->tv_usec = ns / 1000; tv->tv_usec = ns / 1000; Loading Loading
arch/tile/include/asm/vdso.h +3 −2 Original line number Original line Diff line number Diff line Loading @@ -15,6 +15,7 @@ #ifndef __TILE_VDSO_H__ #ifndef __TILE_VDSO_H__ #define __TILE_VDSO_H__ #define __TILE_VDSO_H__ #include <linux/seqlock.h> #include <linux/types.h> #include <linux/types.h> /* /* Loading @@ -26,8 +27,8 @@ */ */ struct vdso_data { struct vdso_data { __u64 tz_update_count; /* Timezone atomicity ctr */ seqcount_t tz_seq; /* Timezone seqlock */ __u64 tb_update_count; /* Timebase atomicity ctr */ seqcount_t tb_seq; /* Timebase seqlock */ __u64 xtime_tod_stamp; /* TOD clock for xtime */ __u64 xtime_tod_stamp; /* TOD clock for xtime */ __u64 xtime_clock_sec; /* Kernel time second */ __u64 xtime_clock_sec; /* Kernel time second */ __u64 xtime_clock_nsec; /* Kernel time nanosecond */ __u64 xtime_clock_nsec; /* Kernel time nanosecond */ Loading
arch/tile/kernel/time.c +6 −10 Original line number Original line Diff line number Diff line Loading @@ -249,13 +249,10 @@ cycles_t ns2cycles(unsigned long nsecs) void update_vsyscall_tz(void) void update_vsyscall_tz(void) { { /* Userspace gettimeofday will spin while this value is odd. */ write_seqcount_begin(&vdso_data->tz_seq); ++vdso_data->tz_update_count; smp_wmb(); vdso_data->tz_minuteswest = sys_tz.tz_minuteswest; vdso_data->tz_minuteswest = sys_tz.tz_minuteswest; vdso_data->tz_dsttime = sys_tz.tz_dsttime; vdso_data->tz_dsttime = sys_tz.tz_dsttime; smp_wmb(); write_seqcount_end(&vdso_data->tz_seq); ++vdso_data->tz_update_count; } } void update_vsyscall(struct timekeeper *tk) void update_vsyscall(struct timekeeper *tk) Loading @@ -266,9 +263,8 @@ void update_vsyscall(struct timekeeper *tk) if (clock != &cycle_counter_cs) if (clock != &cycle_counter_cs) return; return; /* Userspace gettimeofday will spin while this value is odd. */ write_seqcount_begin(&vdso_data->tb_seq); ++vdso_data->tb_update_count; smp_wmb(); vdso_data->xtime_tod_stamp = tk->tkr.cycle_last; vdso_data->xtime_tod_stamp = tk->tkr.cycle_last; vdso_data->xtime_clock_sec = tk->xtime_sec; vdso_data->xtime_clock_sec = tk->xtime_sec; vdso_data->xtime_clock_nsec = tk->tkr.xtime_nsec; vdso_data->xtime_clock_nsec = tk->tkr.xtime_nsec; Loading @@ -276,6 +272,6 @@ void update_vsyscall(struct timekeeper *tk) vdso_data->wtom_clock_nsec = wtm->tv_nsec; vdso_data->wtom_clock_nsec = wtm->tv_nsec; vdso_data->mult = tk->tkr.mult; vdso_data->mult = tk->tkr.mult; vdso_data->shift = tk->tkr.shift; vdso_data->shift = tk->tkr.shift; smp_wmb(); ++vdso_data->tb_update_count; write_seqcount_end(&vdso_data->tb_seq); } }
arch/tile/kernel/vdso/vgettimeofday.c +15 −32 Original line number Original line Diff line number Diff line Loading @@ -53,50 +53,33 @@ inline unsigned long get_datapage(void) int __vdso_gettimeofday(struct timeval *tv, struct timezone *tz) int __vdso_gettimeofday(struct timeval *tv, struct timezone *tz) { { cycles_t cycles; cycles_t cycles; unsigned long count, sec, ns; unsigned count; volatile struct vdso_data *vdso_data; unsigned long sec, ns; struct vdso_data *vdso = (struct vdso_data *)get_datapage(); vdso_data = (struct vdso_data *)get_datapage(); /* The use of the timezone is obsolete, normally tz is NULL. */ /* The use of the timezone is obsolete, normally tz is NULL. */ if (unlikely(tz != NULL)) { if (unlikely(tz != NULL)) { while (1) { do { /* Spin until the update finish. */ count = read_seqcount_begin(&vdso->tz_seq); count = vdso_data->tz_update_count; tz->tz_minuteswest = vdso->tz_minuteswest; if (count & 1) tz->tz_dsttime = vdso->tz_dsttime; continue; } while (unlikely(read_seqcount_retry(&vdso->tz_seq, count))); tz->tz_minuteswest = vdso_data->tz_minuteswest; tz->tz_dsttime = vdso_data->tz_dsttime; /* Check whether updated, read again if so. */ if (count == vdso_data->tz_update_count) break; } } } if (unlikely(tv == NULL)) if (unlikely(tv == NULL)) return 0; return 0; while (1) { do { /* Spin until the update finish. */ count = read_seqcount_begin(&vdso->tb_seq); count = vdso_data->tb_update_count; sec = vdso->xtime_clock_sec; if (count & 1) cycles = get_cycles() - vdso->xtime_tod_stamp; continue; ns = (cycles * vdso->mult) + vdso->xtime_clock_nsec; ns >>= vdso->shift; sec = vdso_data->xtime_clock_sec; cycles = get_cycles() - vdso_data->xtime_tod_stamp; ns = (cycles * vdso_data->mult) + vdso_data->xtime_clock_nsec; ns >>= vdso_data->shift; if (ns >= NSEC_PER_SEC) { if (ns >= NSEC_PER_SEC) { ns -= NSEC_PER_SEC; ns -= NSEC_PER_SEC; sec += 1; sec += 1; } } } while (unlikely(read_seqcount_retry(&vdso->tb_seq, count))); /* Check whether updated, read again if so. */ if (count == vdso_data->tb_update_count) break; } tv->tv_sec = sec; tv->tv_sec = sec; tv->tv_usec = ns / 1000; tv->tv_usec = ns / 1000; Loading