TICK_NSEC
TICK_NSEC, lapic_clockevent.shift);
loops_per_usec = loops_per_jiffy / (TICK_NSEC / 1000);
#define SSIF_MSG_JIFFIES ((SSIF_MSG_USEC * 1000) / TICK_NSEC)
#define SSIF_REQ_RETRY_JIFFIES ((SSIF_REQ_RETRY_USEC * 1000) / TICK_NSEC)
#define SSIF_MSG_PART_JIFFIES ((SSIF_MSG_PART_USEC * 1000) / TICK_NSEC)
#define CPUFREQ_DBS_MIN_SAMPLING_INTERVAL (2 * TICK_NSEC / NSEC_PER_USEC)
if (time_before64(time, cpu->last_io_update + 2 * TICK_NSEC))
if (delta_ns > TICK_NSEC) {
if (delta_ns > TICK_NSEC)
#define CPUIDLE_POLL_MAX (TICK_NSEC / 16)
if (tick_nohz_tick_stopped() && predicted_ns < TICK_NSEC)
delta_tick = TICK_NSEC / 2;
if (predicted_ns < TICK_NSEC)
if (drv->states[idx].target_residency_ns < TICK_NSEC &&
predicted_ns < TICK_NSEC) && !tick_nohz_tick_stopped()) {
if (data->tick_wakeup && data->next_timer_ns > TICK_NSEC) {
if (measured_ns <= TICK_NSEC)
ktime_t delta_tick = TICK_NSEC / 2;
drv->states[min_idx].target_residency_ns < TICK_NSEC)
if (drv->states[idx].target_residency_ns < TICK_NSEC &&
duration_ns = TICK_NSEC / 2;
duration_ns >= TICK_NSEC) || tick_nohz_tick_stopped())
seq_printf(m, "usec_in_jiffy=%lu, statistics=%d\n", TICK_NSEC / 1000,
#if (TICK_NSEC % (NSEC_PER_SEC / AHZ)) == 0
u64 tmp = (u64)x * TICK_NSEC;
#define USLEEP_RANGE_UPPER_BOUND ((TICK_NSEC << max_slack_shift) / NSEC_PER_USEC)
#if !((((NSEC_PER_SEC << 2) / TICK_NSEC) << (SEC_JIFFIE_SC - 2)) & 0x80000000)
TICK_NSEC -1) / (u64)TICK_NSEC))
TICK_NSEC -1) / (u64)TICK_NSEC))
(long)((u64)((u64)MAX_JIFFY_OFFSET * TICK_NSEC) / NSEC_PER_SEC)
(SH_DIV((MAX_JIFFY_OFFSET >> SEC_JIFFIE_SC) * TICK_NSEC, NSEC_PER_SEC, 1) - 1)
return ktime_add(ktime_get(), TICK_NSEC);
*delta_next = TICK_NSEC;
#define LOW_RES_NSEC TICK_NSEC
if (delta > 0 && delta < 2*TICK_NSEC)
u64 now, period = TICK_NSEC;
max = (TICK_NSEC / 100) * sysctl_perf_cpu_time_max_percent;
max_clock = wrap_max(old_clock, gtod + TICK_NSEC);
if (delta_ns <= TICK_NSEC)
u64 other, cputime = TICK_NSEC * ticks;
cputime = TICK_NSEC;
cputime = ticks * TICK_NSEC;
if (vtime->stime >= TICK_NSEC) {
if (vtime->gtime >= TICK_NSEC) {
if (vtime->utime >= TICK_NSEC) {
p->node_stamp += 2 * TICK_NSEC;
delay += 2 * TICK_NSEC;
u64 max_slice = cfs_rq_max_slice(cfs_rq) + TICK_NSEC;
WARN_ON_ONCE(clocksource_wdtest_jiffies.uncertainty_margin != TICK_NSEC);
.uncertainty_margin = TICK_NSEC,
.mult = TICK_NSEC << JIFFIES_SHIFT, /* details above */
nval += TICK_NSEC;
val = TICK_NSEC;
.mult = TICK_NSEC << JIFFIES_SHIFT, /* details above */
const unsigned long TIME_SET_NSEC_FUZZ = TICK_NSEC * 5;
*oldval = TICK_NSEC;
ktime_t next = ktime_add_ns(dev->next_event, TICK_NSEC);
next = ktime_add_ns(next, TICK_NSEC);
next = ktime_add_ns(next, TICK_NSEC);
tick_next_period = ktime_add_ns(tick_next_period, TICK_NSEC);
if (unlikely(delta >= TICK_NSEC)) {
s64 incr = TICK_NSEC;
TICK_NSEC);
nextp = ktime_add_ns(last_jiffies_update, TICK_NSEC);
u64 offset = TICK_NSEC >> 1;
hrtimer_forward_now(&ts->sched_timer, TICK_NSEC);
div_u64_rem(tick_next_period, TICK_NSEC, &rem);
tick_next_period += TICK_NSEC - rem;
hrtimer_forward(timer, now, TICK_NSEC);
hrtimer_forward(&ts->sched_timer, now, TICK_NSEC);
next_tick = basemono + TICK_NSEC;
if (delta <= (u64)TICK_NSEC) {
long nsec = value->tv_nsec + TICK_NSEC - 1;
value->tv_sec = div_u64_rem((u64)jiffies * TICK_NSEC,
#if (TICK_NSEC % (NSEC_PER_SEC / USER_HZ)) == 0
return div_u64((u64)x * TICK_NSEC, NSEC_PER_SEC / USER_HZ);
#if (TICK_NSEC % (NSEC_PER_SEC / USER_HZ)) == 0
x = div_u64(x * TICK_NSEC, (NSEC_PER_SEC / USER_HZ));
return DIV_ROUND_UP_ULL(nextevt, TICK_NSEC) * TICK_NSEC;
tevt->local = basem + (u64)(nextevt - basej) * TICK_NSEC;
tevt->global = basem + (u64)(nextevt_global - basej) * TICK_NSEC;
tevt->local = basem + (u64)(nextevt_local - basej) * TICK_NSEC;
tmp = div_u64(next_tmigr - basem, TICK_NSEC);
if (delta < TICK_NSEC)