time: Rename timekeeper::tkr to timekeeper::tkr_mono

In preparation of adding another tkr field, rename this one to
tkr_mono. Also rename tk_read_base::base_mono to tk_read_base::base,
since the structure is not specific to CLOCK_MONOTONIC and the mono
name got added to the tk_read_base instance.

Lots of trivial churn.

Signed-off-by: Peter Zijlstra (Intel) <peterz@infradead.org>
Acked-by: John Stultz <john.stultz@linaro.org>
Cc: Thomas Gleixner <tglx@linutronix.de>
Link: http://lkml.kernel.org/r/20150319093400.344679419@infradead.org
Signed-off-by: Ingo Molnar <mingo@kernel.org>
This commit is contained in:
Peter Zijlstra 2015-03-19 10:09:06 +01:00 committed by Ingo Molnar
parent 32fea568ae
commit 876e78818d
7 changed files with 126 additions and 126 deletions

View File

@ -200,7 +200,7 @@ up_fail:
void update_vsyscall(struct timekeeper *tk) void update_vsyscall(struct timekeeper *tk)
{ {
struct timespec xtime_coarse; struct timespec xtime_coarse;
u32 use_syscall = strcmp(tk->tkr.clock->name, "arch_sys_counter"); u32 use_syscall = strcmp(tk->tkr_mono.clock->name, "arch_sys_counter");
++vdso_data->tb_seq_count; ++vdso_data->tb_seq_count;
smp_wmb(); smp_wmb();
@ -213,11 +213,11 @@ void update_vsyscall(struct timekeeper *tk)
vdso_data->wtm_clock_nsec = tk->wall_to_monotonic.tv_nsec; vdso_data->wtm_clock_nsec = tk->wall_to_monotonic.tv_nsec;
if (!use_syscall) { if (!use_syscall) {
vdso_data->cs_cycle_last = tk->tkr.cycle_last; vdso_data->cs_cycle_last = tk->tkr_mono.cycle_last;
vdso_data->xtime_clock_sec = tk->xtime_sec; vdso_data->xtime_clock_sec = tk->xtime_sec;
vdso_data->xtime_clock_nsec = tk->tkr.xtime_nsec; vdso_data->xtime_clock_nsec = tk->tkr_mono.xtime_nsec;
vdso_data->cs_mult = tk->tkr.mult; vdso_data->cs_mult = tk->tkr_mono.mult;
vdso_data->cs_shift = tk->tkr.shift; vdso_data->cs_shift = tk->tkr_mono.shift;
} }
smp_wmb(); smp_wmb();

View File

@ -215,20 +215,20 @@ void update_vsyscall(struct timekeeper *tk)
{ {
u64 nsecps; u64 nsecps;
if (tk->tkr.clock != &clocksource_tod) if (tk->tkr_mono.clock != &clocksource_tod)
return; return;
/* Make userspace gettimeofday spin until we're done. */ /* Make userspace gettimeofday spin until we're done. */
++vdso_data->tb_update_count; ++vdso_data->tb_update_count;
smp_wmb(); smp_wmb();
vdso_data->xtime_tod_stamp = tk->tkr.cycle_last; vdso_data->xtime_tod_stamp = tk->tkr_mono.cycle_last;
vdso_data->xtime_clock_sec = tk->xtime_sec; vdso_data->xtime_clock_sec = tk->xtime_sec;
vdso_data->xtime_clock_nsec = tk->tkr.xtime_nsec; vdso_data->xtime_clock_nsec = tk->tkr_mono.xtime_nsec;
vdso_data->wtom_clock_sec = vdso_data->wtom_clock_sec =
tk->xtime_sec + tk->wall_to_monotonic.tv_sec; tk->xtime_sec + tk->wall_to_monotonic.tv_sec;
vdso_data->wtom_clock_nsec = tk->tkr.xtime_nsec + vdso_data->wtom_clock_nsec = tk->tkr_mono.xtime_nsec +
+ ((u64) tk->wall_to_monotonic.tv_nsec << tk->tkr.shift); + ((u64) tk->wall_to_monotonic.tv_nsec << tk->tkr_mono.shift);
nsecps = (u64) NSEC_PER_SEC << tk->tkr.shift; nsecps = (u64) NSEC_PER_SEC << tk->tkr_mono.shift;
while (vdso_data->wtom_clock_nsec >= nsecps) { while (vdso_data->wtom_clock_nsec >= nsecps) {
vdso_data->wtom_clock_nsec -= nsecps; vdso_data->wtom_clock_nsec -= nsecps;
vdso_data->wtom_clock_sec++; vdso_data->wtom_clock_sec++;
@ -236,7 +236,7 @@ void update_vsyscall(struct timekeeper *tk)
vdso_data->xtime_coarse_sec = tk->xtime_sec; vdso_data->xtime_coarse_sec = tk->xtime_sec;
vdso_data->xtime_coarse_nsec = vdso_data->xtime_coarse_nsec =
(long)(tk->tkr.xtime_nsec >> tk->tkr.shift); (long)(tk->tkr_mono.xtime_nsec >> tk->tkr_mono.shift);
vdso_data->wtom_coarse_sec = vdso_data->wtom_coarse_sec =
vdso_data->xtime_coarse_sec + tk->wall_to_monotonic.tv_sec; vdso_data->xtime_coarse_sec + tk->wall_to_monotonic.tv_sec;
vdso_data->wtom_coarse_nsec = vdso_data->wtom_coarse_nsec =
@ -246,8 +246,8 @@ void update_vsyscall(struct timekeeper *tk)
vdso_data->wtom_coarse_sec++; vdso_data->wtom_coarse_sec++;
} }
vdso_data->tk_mult = tk->tkr.mult; vdso_data->tk_mult = tk->tkr_mono.mult;
vdso_data->tk_shift = tk->tkr.shift; vdso_data->tk_shift = tk->tkr_mono.shift;
smp_wmb(); smp_wmb();
++vdso_data->tb_update_count; ++vdso_data->tb_update_count;
} }

View File

@ -257,34 +257,34 @@ void update_vsyscall_tz(void)
void update_vsyscall(struct timekeeper *tk) void update_vsyscall(struct timekeeper *tk)
{ {
if (tk->tkr.clock != &cycle_counter_cs) if (tk->tkr_mono.clock != &cycle_counter_cs)
return; return;
write_seqcount_begin(&vdso_data->tb_seq); write_seqcount_begin(&vdso_data->tb_seq);
vdso_data->cycle_last = tk->tkr.cycle_last; vdso_data->cycle_last = tk->tkr_mono.cycle_last;
vdso_data->mask = tk->tkr.mask; vdso_data->mask = tk->tkr_mono.mask;
vdso_data->mult = tk->tkr.mult; vdso_data->mult = tk->tkr_mono.mult;
vdso_data->shift = tk->tkr.shift; vdso_data->shift = tk->tkr_mono.shift;
vdso_data->wall_time_sec = tk->xtime_sec; vdso_data->wall_time_sec = tk->xtime_sec;
vdso_data->wall_time_snsec = tk->tkr.xtime_nsec; vdso_data->wall_time_snsec = tk->tkr_mono.xtime_nsec;
vdso_data->monotonic_time_sec = tk->xtime_sec vdso_data->monotonic_time_sec = tk->xtime_sec
+ tk->wall_to_monotonic.tv_sec; + tk->wall_to_monotonic.tv_sec;
vdso_data->monotonic_time_snsec = tk->tkr.xtime_nsec vdso_data->monotonic_time_snsec = tk->tkr_mono.xtime_nsec
+ ((u64)tk->wall_to_monotonic.tv_nsec + ((u64)tk->wall_to_monotonic.tv_nsec
<< tk->tkr.shift); << tk->tkr_mono.shift);
while (vdso_data->monotonic_time_snsec >= while (vdso_data->monotonic_time_snsec >=
(((u64)NSEC_PER_SEC) << tk->tkr.shift)) { (((u64)NSEC_PER_SEC) << tk->tkr_mono.shift)) {
vdso_data->monotonic_time_snsec -= vdso_data->monotonic_time_snsec -=
((u64)NSEC_PER_SEC) << tk->tkr.shift; ((u64)NSEC_PER_SEC) << tk->tkr_mono.shift;
vdso_data->monotonic_time_sec++; vdso_data->monotonic_time_sec++;
} }
vdso_data->wall_time_coarse_sec = tk->xtime_sec; vdso_data->wall_time_coarse_sec = tk->xtime_sec;
vdso_data->wall_time_coarse_nsec = (long)(tk->tkr.xtime_nsec >> vdso_data->wall_time_coarse_nsec = (long)(tk->tkr_mono.xtime_nsec >>
tk->tkr.shift); tk->tkr_mono.shift);
vdso_data->monotonic_time_coarse_sec = vdso_data->monotonic_time_coarse_sec =
vdso_data->wall_time_coarse_sec + tk->wall_to_monotonic.tv_sec; vdso_data->wall_time_coarse_sec + tk->wall_to_monotonic.tv_sec;

View File

@ -31,30 +31,30 @@ void update_vsyscall(struct timekeeper *tk)
gtod_write_begin(vdata); gtod_write_begin(vdata);
/* copy vsyscall data */ /* copy vsyscall data */
vdata->vclock_mode = tk->tkr.clock->archdata.vclock_mode; vdata->vclock_mode = tk->tkr_mono.clock->archdata.vclock_mode;
vdata->cycle_last = tk->tkr.cycle_last; vdata->cycle_last = tk->tkr_mono.cycle_last;
vdata->mask = tk->tkr.mask; vdata->mask = tk->tkr_mono.mask;
vdata->mult = tk->tkr.mult; vdata->mult = tk->tkr_mono.mult;
vdata->shift = tk->tkr.shift; vdata->shift = tk->tkr_mono.shift;
vdata->wall_time_sec = tk->xtime_sec; vdata->wall_time_sec = tk->xtime_sec;
vdata->wall_time_snsec = tk->tkr.xtime_nsec; vdata->wall_time_snsec = tk->tkr_mono.xtime_nsec;
vdata->monotonic_time_sec = tk->xtime_sec vdata->monotonic_time_sec = tk->xtime_sec
+ tk->wall_to_monotonic.tv_sec; + tk->wall_to_monotonic.tv_sec;
vdata->monotonic_time_snsec = tk->tkr.xtime_nsec vdata->monotonic_time_snsec = tk->tkr_mono.xtime_nsec
+ ((u64)tk->wall_to_monotonic.tv_nsec + ((u64)tk->wall_to_monotonic.tv_nsec
<< tk->tkr.shift); << tk->tkr_mono.shift);
while (vdata->monotonic_time_snsec >= while (vdata->monotonic_time_snsec >=
(((u64)NSEC_PER_SEC) << tk->tkr.shift)) { (((u64)NSEC_PER_SEC) << tk->tkr_mono.shift)) {
vdata->monotonic_time_snsec -= vdata->monotonic_time_snsec -=
((u64)NSEC_PER_SEC) << tk->tkr.shift; ((u64)NSEC_PER_SEC) << tk->tkr_mono.shift;
vdata->monotonic_time_sec++; vdata->monotonic_time_sec++;
} }
vdata->wall_time_coarse_sec = tk->xtime_sec; vdata->wall_time_coarse_sec = tk->xtime_sec;
vdata->wall_time_coarse_nsec = (long)(tk->tkr.xtime_nsec >> vdata->wall_time_coarse_nsec = (long)(tk->tkr_mono.xtime_nsec >>
tk->tkr.shift); tk->tkr_mono.shift);
vdata->monotonic_time_coarse_sec = vdata->monotonic_time_coarse_sec =
vdata->wall_time_coarse_sec + tk->wall_to_monotonic.tv_sec; vdata->wall_time_coarse_sec + tk->wall_to_monotonic.tv_sec;

View File

@ -1070,19 +1070,19 @@ static void update_pvclock_gtod(struct timekeeper *tk)
struct pvclock_gtod_data *vdata = &pvclock_gtod_data; struct pvclock_gtod_data *vdata = &pvclock_gtod_data;
u64 boot_ns; u64 boot_ns;
boot_ns = ktime_to_ns(ktime_add(tk->tkr.base_mono, tk->offs_boot)); boot_ns = ktime_to_ns(ktime_add(tk->tkr_mono.base, tk->offs_boot));
write_seqcount_begin(&vdata->seq); write_seqcount_begin(&vdata->seq);
/* copy pvclock gtod data */ /* copy pvclock gtod data */
vdata->clock.vclock_mode = tk->tkr.clock->archdata.vclock_mode; vdata->clock.vclock_mode = tk->tkr_mono.clock->archdata.vclock_mode;
vdata->clock.cycle_last = tk->tkr.cycle_last; vdata->clock.cycle_last = tk->tkr_mono.cycle_last;
vdata->clock.mask = tk->tkr.mask; vdata->clock.mask = tk->tkr_mono.mask;
vdata->clock.mult = tk->tkr.mult; vdata->clock.mult = tk->tkr_mono.mult;
vdata->clock.shift = tk->tkr.shift; vdata->clock.shift = tk->tkr_mono.shift;
vdata->boot_ns = boot_ns; vdata->boot_ns = boot_ns;
vdata->nsec_base = tk->tkr.xtime_nsec; vdata->nsec_base = tk->tkr_mono.xtime_nsec;
write_seqcount_end(&vdata->seq); write_seqcount_end(&vdata->seq);
} }

View File

@ -16,16 +16,16 @@
* @read: Read function of @clock * @read: Read function of @clock
* @mask: Bitmask for two's complement subtraction of non 64bit clocks * @mask: Bitmask for two's complement subtraction of non 64bit clocks
* @cycle_last: @clock cycle value at last update * @cycle_last: @clock cycle value at last update
* @mult: NTP adjusted multiplier for scaled math conversion * @mult: (NTP adjusted) multiplier for scaled math conversion
* @shift: Shift value for scaled math conversion * @shift: Shift value for scaled math conversion
* @xtime_nsec: Shifted (fractional) nano seconds offset for readout * @xtime_nsec: Shifted (fractional) nano seconds offset for readout
* @base_mono: ktime_t (nanoseconds) base time for readout * @base: ktime_t (nanoseconds) base time for readout
* *
* This struct has size 56 byte on 64 bit. Together with a seqcount it * This struct has size 56 byte on 64 bit. Together with a seqcount it
* occupies a single 64byte cache line. * occupies a single 64byte cache line.
* *
* The struct is separate from struct timekeeper as it is also used * The struct is separate from struct timekeeper as it is also used
* for a fast NMI safe accessor to clock monotonic. * for a fast NMI safe accessors.
*/ */
struct tk_read_base { struct tk_read_base {
struct clocksource *clock; struct clocksource *clock;
@ -35,12 +35,12 @@ struct tk_read_base {
u32 mult; u32 mult;
u32 shift; u32 shift;
u64 xtime_nsec; u64 xtime_nsec;
ktime_t base_mono; ktime_t base;
}; };
/** /**
* struct timekeeper - Structure holding internal timekeeping values. * struct timekeeper - Structure holding internal timekeeping values.
* @tkr: The readout base structure * @tkr_mono: The readout base structure for CLOCK_MONOTONIC
* @xtime_sec: Current CLOCK_REALTIME time in seconds * @xtime_sec: Current CLOCK_REALTIME time in seconds
* @ktime_sec: Current CLOCK_MONOTONIC time in seconds * @ktime_sec: Current CLOCK_MONOTONIC time in seconds
* @wall_to_monotonic: CLOCK_REALTIME to CLOCK_MONOTONIC offset * @wall_to_monotonic: CLOCK_REALTIME to CLOCK_MONOTONIC offset
@ -76,7 +76,7 @@ struct tk_read_base {
* used instead. * used instead.
*/ */
struct timekeeper { struct timekeeper {
struct tk_read_base tkr; struct tk_read_base tkr_mono;
u64 xtime_sec; u64 xtime_sec;
unsigned long ktime_sec; unsigned long ktime_sec;
struct timespec64 wall_to_monotonic; struct timespec64 wall_to_monotonic;

View File

@ -68,8 +68,8 @@ bool __read_mostly persistent_clock_exist = false;
static inline void tk_normalize_xtime(struct timekeeper *tk) static inline void tk_normalize_xtime(struct timekeeper *tk)
{ {
while (tk->tkr.xtime_nsec >= ((u64)NSEC_PER_SEC << tk->tkr.shift)) { while (tk->tkr_mono.xtime_nsec >= ((u64)NSEC_PER_SEC << tk->tkr_mono.shift)) {
tk->tkr.xtime_nsec -= (u64)NSEC_PER_SEC << tk->tkr.shift; tk->tkr_mono.xtime_nsec -= (u64)NSEC_PER_SEC << tk->tkr_mono.shift;
tk->xtime_sec++; tk->xtime_sec++;
} }
} }
@ -79,20 +79,20 @@ static inline struct timespec64 tk_xtime(struct timekeeper *tk)
struct timespec64 ts; struct timespec64 ts;
ts.tv_sec = tk->xtime_sec; ts.tv_sec = tk->xtime_sec;
ts.tv_nsec = (long)(tk->tkr.xtime_nsec >> tk->tkr.shift); ts.tv_nsec = (long)(tk->tkr_mono.xtime_nsec >> tk->tkr_mono.shift);
return ts; return ts;
} }
static void tk_set_xtime(struct timekeeper *tk, const struct timespec64 *ts) static void tk_set_xtime(struct timekeeper *tk, const struct timespec64 *ts)
{ {
tk->xtime_sec = ts->tv_sec; tk->xtime_sec = ts->tv_sec;
tk->tkr.xtime_nsec = (u64)ts->tv_nsec << tk->tkr.shift; tk->tkr_mono.xtime_nsec = (u64)ts->tv_nsec << tk->tkr_mono.shift;
} }
static void tk_xtime_add(struct timekeeper *tk, const struct timespec64 *ts) static void tk_xtime_add(struct timekeeper *tk, const struct timespec64 *ts)
{ {
tk->xtime_sec += ts->tv_sec; tk->xtime_sec += ts->tv_sec;
tk->tkr.xtime_nsec += (u64)ts->tv_nsec << tk->tkr.shift; tk->tkr_mono.xtime_nsec += (u64)ts->tv_nsec << tk->tkr_mono.shift;
tk_normalize_xtime(tk); tk_normalize_xtime(tk);
} }
@ -136,8 +136,8 @@ static long timekeeping_last_warning;
static void timekeeping_check_update(struct timekeeper *tk, cycle_t offset) static void timekeeping_check_update(struct timekeeper *tk, cycle_t offset)
{ {
cycle_t max_cycles = tk->tkr.clock->max_cycles; cycle_t max_cycles = tk->tkr_mono.clock->max_cycles;
const char *name = tk->tkr.clock->name; const char *name = tk->tkr_mono.clock->name;
if (offset > max_cycles) { if (offset > max_cycles) {
printk_deferred("WARNING: timekeeping: Cycle offset (%lld) is larger than allowed by the '%s' clock's max_cycles value (%lld): time overflow danger\n", printk_deferred("WARNING: timekeeping: Cycle offset (%lld) is larger than allowed by the '%s' clock's max_cycles value (%lld): time overflow danger\n",
@ -246,11 +246,11 @@ static void tk_setup_internals(struct timekeeper *tk, struct clocksource *clock)
u64 tmp, ntpinterval; u64 tmp, ntpinterval;
struct clocksource *old_clock; struct clocksource *old_clock;
old_clock = tk->tkr.clock; old_clock = tk->tkr_mono.clock;
tk->tkr.clock = clock; tk->tkr_mono.clock = clock;
tk->tkr.read = clock->read; tk->tkr_mono.read = clock->read;
tk->tkr.mask = clock->mask; tk->tkr_mono.mask = clock->mask;
tk->tkr.cycle_last = tk->tkr.read(clock); tk->tkr_mono.cycle_last = tk->tkr_mono.read(clock);
/* Do the ns -> cycle conversion first, using original mult */ /* Do the ns -> cycle conversion first, using original mult */
tmp = NTP_INTERVAL_LENGTH; tmp = NTP_INTERVAL_LENGTH;
@ -274,11 +274,11 @@ static void tk_setup_internals(struct timekeeper *tk, struct clocksource *clock)
if (old_clock) { if (old_clock) {
int shift_change = clock->shift - old_clock->shift; int shift_change = clock->shift - old_clock->shift;
if (shift_change < 0) if (shift_change < 0)
tk->tkr.xtime_nsec >>= -shift_change; tk->tkr_mono.xtime_nsec >>= -shift_change;
else else
tk->tkr.xtime_nsec <<= shift_change; tk->tkr_mono.xtime_nsec <<= shift_change;
} }
tk->tkr.shift = clock->shift; tk->tkr_mono.shift = clock->shift;
tk->ntp_error = 0; tk->ntp_error = 0;
tk->ntp_error_shift = NTP_SCALE_SHIFT - clock->shift; tk->ntp_error_shift = NTP_SCALE_SHIFT - clock->shift;
@ -289,7 +289,7 @@ static void tk_setup_internals(struct timekeeper *tk, struct clocksource *clock)
* active clocksource. These value will be adjusted via NTP * active clocksource. These value will be adjusted via NTP
* to counteract clock drifting. * to counteract clock drifting.
*/ */
tk->tkr.mult = clock->mult; tk->tkr_mono.mult = clock->mult;
tk->ntp_err_mult = 0; tk->ntp_err_mult = 0;
} }
@ -318,11 +318,11 @@ static inline s64 timekeeping_get_ns(struct tk_read_base *tkr)
static inline s64 timekeeping_get_ns_raw(struct timekeeper *tk) static inline s64 timekeeping_get_ns_raw(struct timekeeper *tk)
{ {
struct clocksource *clock = tk->tkr.clock; struct clocksource *clock = tk->tkr_mono.clock;
cycle_t delta; cycle_t delta;
s64 nsec; s64 nsec;
delta = timekeeping_get_delta(&tk->tkr); delta = timekeeping_get_delta(&tk->tkr_mono);
/* convert delta to nanoseconds. */ /* convert delta to nanoseconds. */
nsec = clocksource_cyc2ns(delta, clock->mult, clock->shift); nsec = clocksource_cyc2ns(delta, clock->mult, clock->shift);
@ -428,7 +428,7 @@ u64 notrace ktime_get_mono_fast_ns(void)
do { do {
seq = raw_read_seqcount(&tk_fast_mono.seq); seq = raw_read_seqcount(&tk_fast_mono.seq);
tkr = tk_fast_mono.base + (seq & 0x01); tkr = tk_fast_mono.base + (seq & 0x01);
now = ktime_to_ns(tkr->base_mono) + timekeeping_get_ns(tkr); now = ktime_to_ns(tkr->base) + timekeeping_get_ns(tkr);
} while (read_seqcount_retry(&tk_fast_mono.seq, seq)); } while (read_seqcount_retry(&tk_fast_mono.seq, seq));
return now; return now;
@ -456,7 +456,7 @@ static cycle_t dummy_clock_read(struct clocksource *cs)
static void halt_fast_timekeeper(struct timekeeper *tk) static void halt_fast_timekeeper(struct timekeeper *tk)
{ {
static struct tk_read_base tkr_dummy; static struct tk_read_base tkr_dummy;
struct tk_read_base *tkr = &tk->tkr; struct tk_read_base *tkr = &tk->tkr_mono;
memcpy(&tkr_dummy, tkr, sizeof(tkr_dummy)); memcpy(&tkr_dummy, tkr, sizeof(tkr_dummy));
cycles_at_suspend = tkr->read(tkr->clock); cycles_at_suspend = tkr->read(tkr->clock);
@ -472,8 +472,8 @@ static inline void update_vsyscall(struct timekeeper *tk)
xt = timespec64_to_timespec(tk_xtime(tk)); xt = timespec64_to_timespec(tk_xtime(tk));
wm = timespec64_to_timespec(tk->wall_to_monotonic); wm = timespec64_to_timespec(tk->wall_to_monotonic);
update_vsyscall_old(&xt, &wm, tk->tkr.clock, tk->tkr.mult, update_vsyscall_old(&xt, &wm, tk->tkr_mono.clock, tk->tkr_mono.mult,
tk->tkr.cycle_last); tk->tkr_mono.cycle_last);
} }
static inline void old_vsyscall_fixup(struct timekeeper *tk) static inline void old_vsyscall_fixup(struct timekeeper *tk)
@ -490,11 +490,11 @@ static inline void old_vsyscall_fixup(struct timekeeper *tk)
* (shifted nanoseconds), and CONFIG_GENERIC_TIME_VSYSCALL_OLD * (shifted nanoseconds), and CONFIG_GENERIC_TIME_VSYSCALL_OLD
* users are removed, this can be killed. * users are removed, this can be killed.
*/ */
remainder = tk->tkr.xtime_nsec & ((1ULL << tk->tkr.shift) - 1); remainder = tk->tkr_mono.xtime_nsec & ((1ULL << tk->tkr_mono.shift) - 1);
tk->tkr.xtime_nsec -= remainder; tk->tkr_mono.xtime_nsec -= remainder;
tk->tkr.xtime_nsec += 1ULL << tk->tkr.shift; tk->tkr_mono.xtime_nsec += 1ULL << tk->tkr_mono.shift;
tk->ntp_error += remainder << tk->ntp_error_shift; tk->ntp_error += remainder << tk->ntp_error_shift;
tk->ntp_error -= (1ULL << tk->tkr.shift) << tk->ntp_error_shift; tk->ntp_error -= (1ULL << tk->tkr_mono.shift) << tk->ntp_error_shift;
} }
#else #else
#define old_vsyscall_fixup(tk) #define old_vsyscall_fixup(tk)
@ -559,7 +559,7 @@ static inline void tk_update_ktime_data(struct timekeeper *tk)
*/ */
seconds = (u64)(tk->xtime_sec + tk->wall_to_monotonic.tv_sec); seconds = (u64)(tk->xtime_sec + tk->wall_to_monotonic.tv_sec);
nsec = (u32) tk->wall_to_monotonic.tv_nsec; nsec = (u32) tk->wall_to_monotonic.tv_nsec;
tk->tkr.base_mono = ns_to_ktime(seconds * NSEC_PER_SEC + nsec); tk->tkr_mono.base = ns_to_ktime(seconds * NSEC_PER_SEC + nsec);
/* Update the monotonic raw base */ /* Update the monotonic raw base */
tk->base_raw = timespec64_to_ktime(tk->raw_time); tk->base_raw = timespec64_to_ktime(tk->raw_time);
@ -569,7 +569,7 @@ static inline void tk_update_ktime_data(struct timekeeper *tk)
* wall_to_monotonic can be greater/equal one second. Take * wall_to_monotonic can be greater/equal one second. Take
* this into account before updating tk->ktime_sec. * this into account before updating tk->ktime_sec.
*/ */
nsec += (u32)(tk->tkr.xtime_nsec >> tk->tkr.shift); nsec += (u32)(tk->tkr_mono.xtime_nsec >> tk->tkr_mono.shift);
if (nsec >= NSEC_PER_SEC) if (nsec >= NSEC_PER_SEC)
seconds++; seconds++;
tk->ktime_sec = seconds; tk->ktime_sec = seconds;
@ -592,7 +592,7 @@ static void timekeeping_update(struct timekeeper *tk, unsigned int action)
memcpy(&shadow_timekeeper, &tk_core.timekeeper, memcpy(&shadow_timekeeper, &tk_core.timekeeper,
sizeof(tk_core.timekeeper)); sizeof(tk_core.timekeeper));
update_fast_timekeeper(&tk->tkr); update_fast_timekeeper(&tk->tkr_mono);
} }
/** /**
@ -604,18 +604,18 @@ static void timekeeping_update(struct timekeeper *tk, unsigned int action)
*/ */
static void timekeeping_forward_now(struct timekeeper *tk) static void timekeeping_forward_now(struct timekeeper *tk)
{ {
struct clocksource *clock = tk->tkr.clock; struct clocksource *clock = tk->tkr_mono.clock;
cycle_t cycle_now, delta; cycle_t cycle_now, delta;
s64 nsec; s64 nsec;
cycle_now = tk->tkr.read(clock); cycle_now = tk->tkr_mono.read(clock);
delta = clocksource_delta(cycle_now, tk->tkr.cycle_last, tk->tkr.mask); delta = clocksource_delta(cycle_now, tk->tkr_mono.cycle_last, tk->tkr_mono.mask);
tk->tkr.cycle_last = cycle_now; tk->tkr_mono.cycle_last = cycle_now;
tk->tkr.xtime_nsec += delta * tk->tkr.mult; tk->tkr_mono.xtime_nsec += delta * tk->tkr_mono.mult;
/* If arch requires, add in get_arch_timeoffset() */ /* If arch requires, add in get_arch_timeoffset() */
tk->tkr.xtime_nsec += (u64)arch_gettimeoffset() << tk->tkr.shift; tk->tkr_mono.xtime_nsec += (u64)arch_gettimeoffset() << tk->tkr_mono.shift;
tk_normalize_xtime(tk); tk_normalize_xtime(tk);
@ -640,7 +640,7 @@ int __getnstimeofday64(struct timespec64 *ts)
seq = read_seqcount_begin(&tk_core.seq); seq = read_seqcount_begin(&tk_core.seq);
ts->tv_sec = tk->xtime_sec; ts->tv_sec = tk->xtime_sec;
nsecs = timekeeping_get_ns(&tk->tkr); nsecs = timekeeping_get_ns(&tk->tkr_mono);
} while (read_seqcount_retry(&tk_core.seq, seq)); } while (read_seqcount_retry(&tk_core.seq, seq));
@ -680,8 +680,8 @@ ktime_t ktime_get(void)
do { do {
seq = read_seqcount_begin(&tk_core.seq); seq = read_seqcount_begin(&tk_core.seq);
base = tk->tkr.base_mono; base = tk->tkr_mono.base;
nsecs = timekeeping_get_ns(&tk->tkr); nsecs = timekeeping_get_ns(&tk->tkr_mono);
} while (read_seqcount_retry(&tk_core.seq, seq)); } while (read_seqcount_retry(&tk_core.seq, seq));
@ -706,8 +706,8 @@ ktime_t ktime_get_with_offset(enum tk_offsets offs)
do { do {
seq = read_seqcount_begin(&tk_core.seq); seq = read_seqcount_begin(&tk_core.seq);
base = ktime_add(tk->tkr.base_mono, *offset); base = ktime_add(tk->tkr_mono.base, *offset);
nsecs = timekeeping_get_ns(&tk->tkr); nsecs = timekeeping_get_ns(&tk->tkr_mono);
} while (read_seqcount_retry(&tk_core.seq, seq)); } while (read_seqcount_retry(&tk_core.seq, seq));
@ -777,7 +777,7 @@ void ktime_get_ts64(struct timespec64 *ts)
do { do {
seq = read_seqcount_begin(&tk_core.seq); seq = read_seqcount_begin(&tk_core.seq);
ts->tv_sec = tk->xtime_sec; ts->tv_sec = tk->xtime_sec;
nsec = timekeeping_get_ns(&tk->tkr); nsec = timekeeping_get_ns(&tk->tkr_mono);
tomono = tk->wall_to_monotonic; tomono = tk->wall_to_monotonic;
} while (read_seqcount_retry(&tk_core.seq, seq)); } while (read_seqcount_retry(&tk_core.seq, seq));
@ -863,7 +863,7 @@ void getnstime_raw_and_real(struct timespec *ts_raw, struct timespec *ts_real)
ts_real->tv_nsec = 0; ts_real->tv_nsec = 0;
nsecs_raw = timekeeping_get_ns_raw(tk); nsecs_raw = timekeeping_get_ns_raw(tk);
nsecs_real = timekeeping_get_ns(&tk->tkr); nsecs_real = timekeeping_get_ns(&tk->tkr_mono);
} while (read_seqcount_retry(&tk_core.seq, seq)); } while (read_seqcount_retry(&tk_core.seq, seq));
@ -1046,7 +1046,7 @@ static int change_clocksource(void *data)
*/ */
if (try_module_get(new->owner)) { if (try_module_get(new->owner)) {
if (!new->enable || new->enable(new) == 0) { if (!new->enable || new->enable(new) == 0) {
old = tk->tkr.clock; old = tk->tkr_mono.clock;
tk_setup_internals(tk, new); tk_setup_internals(tk, new);
if (old->disable) if (old->disable)
old->disable(old); old->disable(old);
@ -1074,11 +1074,11 @@ int timekeeping_notify(struct clocksource *clock)
{ {
struct timekeeper *tk = &tk_core.timekeeper; struct timekeeper *tk = &tk_core.timekeeper;
if (tk->tkr.clock == clock) if (tk->tkr_mono.clock == clock)
return 0; return 0;
stop_machine(change_clocksource, clock, NULL); stop_machine(change_clocksource, clock, NULL);
tick_clock_notify(); tick_clock_notify();
return tk->tkr.clock == clock ? 0 : -1; return tk->tkr_mono.clock == clock ? 0 : -1;
} }
/** /**
@ -1119,7 +1119,7 @@ int timekeeping_valid_for_hres(void)
do { do {
seq = read_seqcount_begin(&tk_core.seq); seq = read_seqcount_begin(&tk_core.seq);
ret = tk->tkr.clock->flags & CLOCK_SOURCE_VALID_FOR_HRES; ret = tk->tkr_mono.clock->flags & CLOCK_SOURCE_VALID_FOR_HRES;
} while (read_seqcount_retry(&tk_core.seq, seq)); } while (read_seqcount_retry(&tk_core.seq, seq));
@ -1138,7 +1138,7 @@ u64 timekeeping_max_deferment(void)
do { do {
seq = read_seqcount_begin(&tk_core.seq); seq = read_seqcount_begin(&tk_core.seq);
ret = tk->tkr.clock->max_idle_ns; ret = tk->tkr_mono.clock->max_idle_ns;
} while (read_seqcount_retry(&tk_core.seq, seq)); } while (read_seqcount_retry(&tk_core.seq, seq));
@ -1303,7 +1303,7 @@ void timekeeping_inject_sleeptime64(struct timespec64 *delta)
void timekeeping_resume(void) void timekeeping_resume(void)
{ {
struct timekeeper *tk = &tk_core.timekeeper; struct timekeeper *tk = &tk_core.timekeeper;
struct clocksource *clock = tk->tkr.clock; struct clocksource *clock = tk->tkr_mono.clock;
unsigned long flags; unsigned long flags;
struct timespec64 ts_new, ts_delta; struct timespec64 ts_new, ts_delta;
struct timespec tmp; struct timespec tmp;
@ -1331,16 +1331,16 @@ void timekeeping_resume(void)
* The less preferred source will only be tried if there is no better * The less preferred source will only be tried if there is no better
* usable source. The rtc part is handled separately in rtc core code. * usable source. The rtc part is handled separately in rtc core code.
*/ */
cycle_now = tk->tkr.read(clock); cycle_now = tk->tkr_mono.read(clock);
if ((clock->flags & CLOCK_SOURCE_SUSPEND_NONSTOP) && if ((clock->flags & CLOCK_SOURCE_SUSPEND_NONSTOP) &&
cycle_now > tk->tkr.cycle_last) { cycle_now > tk->tkr_mono.cycle_last) {
u64 num, max = ULLONG_MAX; u64 num, max = ULLONG_MAX;
u32 mult = clock->mult; u32 mult = clock->mult;
u32 shift = clock->shift; u32 shift = clock->shift;
s64 nsec = 0; s64 nsec = 0;
cycle_delta = clocksource_delta(cycle_now, tk->tkr.cycle_last, cycle_delta = clocksource_delta(cycle_now, tk->tkr_mono.cycle_last,
tk->tkr.mask); tk->tkr_mono.mask);
/* /*
* "cycle_delta * mutl" may cause 64 bits overflow, if the * "cycle_delta * mutl" may cause 64 bits overflow, if the
@ -1366,7 +1366,7 @@ void timekeeping_resume(void)
__timekeeping_inject_sleeptime(tk, &ts_delta); __timekeeping_inject_sleeptime(tk, &ts_delta);
/* Re-base the last cycle value */ /* Re-base the last cycle value */
tk->tkr.cycle_last = cycle_now; tk->tkr_mono.cycle_last = cycle_now;
tk->ntp_error = 0; tk->ntp_error = 0;
timekeeping_suspended = 0; timekeeping_suspended = 0;
timekeeping_update(tk, TK_MIRROR | TK_CLOCK_WAS_SET); timekeeping_update(tk, TK_MIRROR | TK_CLOCK_WAS_SET);
@ -1519,15 +1519,15 @@ static __always_inline void timekeeping_apply_adjustment(struct timekeeper *tk,
* *
* XXX - TODO: Doc ntp_error calculation. * XXX - TODO: Doc ntp_error calculation.
*/ */
if ((mult_adj > 0) && (tk->tkr.mult + mult_adj < mult_adj)) { if ((mult_adj > 0) && (tk->tkr_mono.mult + mult_adj < mult_adj)) {
/* NTP adjustment caused clocksource mult overflow */ /* NTP adjustment caused clocksource mult overflow */
WARN_ON_ONCE(1); WARN_ON_ONCE(1);
return; return;
} }
tk->tkr.mult += mult_adj; tk->tkr_mono.mult += mult_adj;
tk->xtime_interval += interval; tk->xtime_interval += interval;
tk->tkr.xtime_nsec -= offset; tk->tkr_mono.xtime_nsec -= offset;
tk->ntp_error -= (interval - offset) << tk->ntp_error_shift; tk->ntp_error -= (interval - offset) << tk->ntp_error_shift;
} }
@ -1589,13 +1589,13 @@ static void timekeeping_adjust(struct timekeeper *tk, s64 offset)
tk->ntp_err_mult = 0; tk->ntp_err_mult = 0;
} }
if (unlikely(tk->tkr.clock->maxadj && if (unlikely(tk->tkr_mono.clock->maxadj &&
(abs(tk->tkr.mult - tk->tkr.clock->mult) (abs(tk->tkr_mono.mult - tk->tkr_mono.clock->mult)
> tk->tkr.clock->maxadj))) { > tk->tkr_mono.clock->maxadj))) {
printk_once(KERN_WARNING printk_once(KERN_WARNING
"Adjusting %s more than 11%% (%ld vs %ld)\n", "Adjusting %s more than 11%% (%ld vs %ld)\n",
tk->tkr.clock->name, (long)tk->tkr.mult, tk->tkr_mono.clock->name, (long)tk->tkr_mono.mult,
(long)tk->tkr.clock->mult + tk->tkr.clock->maxadj); (long)tk->tkr_mono.clock->mult + tk->tkr_mono.clock->maxadj);
} }
/* /*
@ -1612,9 +1612,9 @@ static void timekeeping_adjust(struct timekeeper *tk, s64 offset)
* We'll correct this error next time through this function, when * We'll correct this error next time through this function, when
* xtime_nsec is not as small. * xtime_nsec is not as small.
*/ */
if (unlikely((s64)tk->tkr.xtime_nsec < 0)) { if (unlikely((s64)tk->tkr_mono.xtime_nsec < 0)) {
s64 neg = -(s64)tk->tkr.xtime_nsec; s64 neg = -(s64)tk->tkr_mono.xtime_nsec;
tk->tkr.xtime_nsec = 0; tk->tkr_mono.xtime_nsec = 0;
tk->ntp_error += neg << tk->ntp_error_shift; tk->ntp_error += neg << tk->ntp_error_shift;
} }
} }
@ -1629,13 +1629,13 @@ static void timekeeping_adjust(struct timekeeper *tk, s64 offset)
*/ */
static inline unsigned int accumulate_nsecs_to_secs(struct timekeeper *tk) static inline unsigned int accumulate_nsecs_to_secs(struct timekeeper *tk)
{ {
u64 nsecps = (u64)NSEC_PER_SEC << tk->tkr.shift; u64 nsecps = (u64)NSEC_PER_SEC << tk->tkr_mono.shift;
unsigned int clock_set = 0; unsigned int clock_set = 0;
while (tk->tkr.xtime_nsec >= nsecps) { while (tk->tkr_mono.xtime_nsec >= nsecps) {
int leap; int leap;
tk->tkr.xtime_nsec -= nsecps; tk->tkr_mono.xtime_nsec -= nsecps;
tk->xtime_sec++; tk->xtime_sec++;
/* Figure out if its a leap sec and apply if needed */ /* Figure out if its a leap sec and apply if needed */
@ -1680,9 +1680,9 @@ static cycle_t logarithmic_accumulation(struct timekeeper *tk, cycle_t offset,
/* Accumulate one shifted interval */ /* Accumulate one shifted interval */
offset -= interval; offset -= interval;
tk->tkr.cycle_last += interval; tk->tkr_mono.cycle_last += interval;
tk->tkr.xtime_nsec += tk->xtime_interval << shift; tk->tkr_mono.xtime_nsec += tk->xtime_interval << shift;
*clock_set |= accumulate_nsecs_to_secs(tk); *clock_set |= accumulate_nsecs_to_secs(tk);
/* Accumulate raw time */ /* Accumulate raw time */
@ -1725,8 +1725,8 @@ void update_wall_time(void)
#ifdef CONFIG_ARCH_USES_GETTIMEOFFSET #ifdef CONFIG_ARCH_USES_GETTIMEOFFSET
offset = real_tk->cycle_interval; offset = real_tk->cycle_interval;
#else #else
offset = clocksource_delta(tk->tkr.read(tk->tkr.clock), offset = clocksource_delta(tk->tkr_mono.read(tk->tkr_mono.clock),
tk->tkr.cycle_last, tk->tkr.mask); tk->tkr_mono.cycle_last, tk->tkr_mono.mask);
#endif #endif
/* Check if there's really nothing to do */ /* Check if there's really nothing to do */
@ -1890,8 +1890,8 @@ ktime_t ktime_get_update_offsets_tick(ktime_t *offs_real, ktime_t *offs_boot,
do { do {
seq = read_seqcount_begin(&tk_core.seq); seq = read_seqcount_begin(&tk_core.seq);
base = tk->tkr.base_mono; base = tk->tkr_mono.base;
nsecs = tk->tkr.xtime_nsec >> tk->tkr.shift; nsecs = tk->tkr_mono.xtime_nsec >> tk->tkr_mono.shift;
*offs_real = tk->offs_real; *offs_real = tk->offs_real;
*offs_boot = tk->offs_boot; *offs_boot = tk->offs_boot;
@ -1922,8 +1922,8 @@ ktime_t ktime_get_update_offsets_now(ktime_t *offs_real, ktime_t *offs_boot,
do { do {
seq = read_seqcount_begin(&tk_core.seq); seq = read_seqcount_begin(&tk_core.seq);
base = tk->tkr.base_mono; base = tk->tkr_mono.base;
nsecs = timekeeping_get_ns(&tk->tkr); nsecs = timekeeping_get_ns(&tk->tkr_mono);
*offs_real = tk->offs_real; *offs_real = tk->offs_real;
*offs_boot = tk->offs_boot; *offs_boot = tk->offs_boot;