timers
verify_heap (EV_A_ timers, timercnt);
if (timercnt && ANHE_at (timers [HEAP0]) < mn_now)
ev_timer *w = (ev_timer *)ANHE_w (timers [HEAP0]);
ANHE_at_cache (timers [HEAP0]);
downheap (timers, timercnt, HEAP0);
while (timercnt && ANHE_at (timers [HEAP0]) < mn_now);
ANHE *he = timers + i + HEAP0;
ev_tstamp to = ANHE_at (timers [HEAP0]) - mn_now;
array_needsize (ANHE, timers, timermax, ev_active (w) + 1, EMPTY2);
ANHE_w (timers [ev_active (w)]) = (WT)w;
ANHE_at_cache (timers [ev_active (w)]);
upheap (timers, ev_active (w));
assert (("libev: internal timer heap corruption", ANHE_w (timers [active]) == (WT)w));
timers [active] = timers [timercnt + HEAP0];
adjustheap (timers, timercnt, active);
ANHE_at_cache (timers [ev_active (w)]);
adjustheap (timers, timercnt, ev_active (w));
if (ev_cb ((ev_timer *)ANHE_w (timers [i])) == stat_timer_cb)
cb (EV_A_ EV_STAT, ((char *)ANHE_w (timers [i])) - offsetof (struct ev_stat, timer));
cb (EV_A_ EV_TIMER, ANHE_w (timers [i]));
VARx(ANHE *, timers)
#define timers ((loop)->timers)
if (heap_insert(ctx->timers, id) < 0)
(void) heap_for_each(ctx->timers, print_timer, (void *)ctx);
if (heap_element(ctx->timers, del->index) != del)
if (heap_delete(ctx->timers, del->index) < 0)
(void) heap_for_each(ctx->timers, print_timer, (void *)ctx);
if (heap_element(ctx->timers, timer->index) != timer)
if (heap_element(ctx->timers, timer->index) != timer)
result = heap_increased(ctx->timers, timer->index);
result = heap_decreased(ctx->timers, timer->index);
(void) heap_for_each(ctx->timers, print_timer, (void *)ctx);
(void) heap_for_each(ctx->timers, free_timer, NULL);
(void) heap_free(ctx->timers);
heap_context timers;
unsigned int timers = 0, timers_shift = 22;
timers = 1;
timers = 2;
timers = 3;
t3_write_reg(adap, A_TP_CMM_TIMER_BASE, V_CMTIMERMAXNUM(timers) | m);
u16 timers;
u16 timers;
P1B_SYSCTL(CTL_P1003_1B_TIMERS, timers);
bbr_log_to_processing(struct tcp_bbr *bbr, uint32_t cts, int32_t ret, int32_t timers, uint8_t hpts_calling)
log.u_bbr.flex1 = timers;
int32_t timers = (bbr->r_ctl.rc_hpts_flags & PACE_TMR_MASK);
if (timers == 0) {
if (timers & PACE_TMR_DELACK) {
} else if (timers & PACE_TMR_PERSIT) {
} else if (timers & PACE_TMR_RACK) {
} else if (timers & PACE_TMR_TLP) {
} else if (timers & PACE_TMR_RXT) {
} else if (timers & PACE_TMR_KEEP) {
bbr_log_to_processing(bbr, cts, ret, timers, hpts_calling);
rack_log_to_processing(struct tcp_rack *rack, uint32_t cts, int32_t ret, int32_t timers)
log.u_bbr.flex1 = timers;
int32_t timers = (rack->r_ctl.rc_hpts_flags & PACE_TMR_MASK);
if (timers == 0) {
if ((timers & PACE_TMR_RACK) &&
if (timers & PACE_TMR_DELACK) {
} else if (timers & PACE_TMR_RACK) {
} else if (timers & PACE_TMR_TLP) {
} else if (timers & PACE_TMR_RXT) {
} else if (timers & PACE_TMR_PERSIT) {
} else if (timers & PACE_TMR_KEEP) {
rack_log_to_processing(rack, cts, ret, timers);
timers = regs[TP_CMM_TIMER_BASE / 4] & 0xfffffff,
timers - 1, timers - cq_cntxt);
printf(" Timers: 0x%08x - 0x%08x [%u]\n", timers,
pstructs - 1, pstructs - timers);
tmr_add(&timers, 1, 5, tmr_status, ctx);
TAILQ_INIT(&timers.head);
timers.res = 1;
tmr_add(&timers, 1, 5, &tmr_watchdog, &ctx);
static struct timers timers;
tmr_run(struct timers *tmrs)
tmr_add(struct timers *tmrs, int id, int timeout, tmr_cb func, void *arg)
tmr_add(&timers, 1, 5, tmr_watchdog, ctx);
tmr_run(&timers);
tmr_add(&timers, 1, 1, tmr_creg, ctx);
tmr_add(&timers, 1, 30, tmr_creg, ctx);
tmr_add(&timers, 1, 1, tmr_cgreg, ctx);
tmr_add(&timers, 1, 30, tmr_cgreg, ctx);
tmr_add(&timers, 1, 15, tmr_status, ctx);