next_event
struct event *event, *next_event;
list_for_each_entry_safe(event, next_event, &client->event_list, link)
struct mlxsw_sp_port_mapping_event *event, *next_event;
list_for_each_entry_safe(event, next_event, &event_queue, list) {
struct mlxsw_sp_port_mapping_event *event, *next_event;
list_for_each_entry_safe(event, next_event, &events->queue, list) {
u8 next_event;
ctrl->next_event = 0;
taskInfo = &(ctrl->event_queue[ctrl->next_event]);
ctrl->next_event = (ctrl->next_event + 1) % 10;
taskInfo = &(ctrl->event_queue[ctrl->next_event]);
ctrl->next_event = (ctrl->next_event + 1) % 10;
taskInfo = &(ctrl->event_queue[ctrl->next_event]);
ctrl->next_event = (ctrl->next_event + 1) % 10;
event = next_event (dev, GADGETFS_CONNECT);
event = next_event (dev, GADGETFS_SETUP);
next_event (dev, GADGETFS_DISCONNECT);
next_event (dev, GADGETFS_SUSPEND);
ktime_t next_event;
struct perf_event *next_event)
value = local64_read(&next_event->count);
local64_set(&next_event->count, value);
swap(event->total_time_enabled, next_event->total_time_enabled);
swap(event->total_time_running, next_event->total_time_running);
perf_event_update_userpage(next_event);
struct perf_event *event, *next_event;
next_event = list_first_entry(&next_ctx->event_list,
&next_event->event_entry != &next_ctx->event_list) {
__perf_event_sync_stat(event, next_event);
next_event = list_next_entry(next_event, event_entry);
dev->next_event = KTIME_MAX;
dev->next_event = KTIME_MAX;
dev->next_event = ktime_add_ns(ktime_get(), delta);
dev->next_event = ktime_add_ns(ktime_get(), delta);
dev->next_event = expires;
return clockevents_program_event(dev, dev->next_event, false);
dev->next_event = KTIME_MAX;
td->evtdev->next_event = expires;
bc->next_event = KTIME_MAX;
tick_program_event(td->evtdev->next_event, 1);
clockevents_program_event(bc, bc->next_event, 1);
ktime_t next = ktime_add_ns(dev->next_event, TICK_NSEC);
ktime_t now, next_event;
dev->next_event = KTIME_MAX;
next_event = KTIME_MAX;
if (td->evtdev->next_event <= now) {
} else if (td->evtdev->next_event < next_event) {
next_event = td->evtdev->next_event;
if (next_event != KTIME_MAX)
tick_broadcast_set_event(dev, next_cpu, next_event);
if (bc->next_event == KTIME_MAX)
if (dev->next_event < bc->next_event)
} else if (dev->next_event < bc->next_event) {
tick_broadcast_set_event(bc, cpu, dev->next_event);
if (dev->next_event == KTIME_MAX)
if (dev->next_event <= now) {
tick_program_event(dev->next_event, 1);
clockevents_program_event(wd, dev->next_event, 1);
ktime_t next = dev->next_event;
ktime_t next_event = 0;
next_event = td->evtdev->next_event;
tick_setup_oneshot(newdev, handler, next_event);
dev->next_event = KTIME_MAX;
ktime_t next_event)
clockevents_program_event(newdev, next_event, true);
dev->next_event, hrtimer_active(&ts->sched_timer),
return __this_cpu_read(tick_cpu_device.evtdev)->next_event;
ktime_t next_event;
*delta_next = ktime_sub(dev->next_event, now);
next_event = tick_nohz_next_event(ts, cpu);
if (!next_event)
next_event = min_t(u64, next_event,
return ktime_sub(next_event, now);
dev->next_event = KTIME_MAX;
(unsigned long long) ktime_to_ns(dev->next_event));
if (iter->head != iter->next_event)
iter->next_event = iter->head + length;
iter->next_event = 0;
iter->next_event = 0;
iter->next_event = iter->head;
if (iter->head == iter->next_event) {
iter->head = iter->next_event;
if (iter->next_event >= rb_page_size(iter->head_page)) {
unsigned long next_event;
s64 next_event;
next_event = q->now + 5LLU * NSEC_PER_SEC;
if (next_event > event)
next_event = event;
if (likely(next_event > q->now))
qdisc_watchdog_schedule_ns(&q->watchdog, next_event);
next_event:
goto next_event;
goto next_event;
goto next_event;
goto next_event;
goto next_event;
goto next_event;
goto next_event;
goto next_event;
next_event:
goto next_event;
next_event:
goto next_event;
next_event: