Symbol: idle
arch/alpha/kernel/smp.c
290
secondary_cpu_start(int cpuid, struct task_struct *idle)
arch/alpha/kernel/smp.c
301
ipcb = &task_thread_info(idle)->pcb;
arch/alpha/kernel/smp.c
321
cpuid, idle->state, ipcb->flags));
arch/alpha/kernel/smp.c
361
smp_boot_one_cpu(int cpuid, struct task_struct *idle)
arch/alpha/kernel/smp.c
369
if (secondary_cpu_start(cpuid, idle))
arch/arc/kernel/smp.c
199
int __cpu_up(unsigned int cpu, struct task_struct *idle)
arch/arc/kernel/smp.c
203
secondary_idle_tsk = idle;
arch/arc/kernel/smp.c
205
pr_info("Idle Task [%d] %p", cpu, idle);
arch/arm/common/mcpm_platsmp.c
29
static int mcpm_boot_secondary(unsigned int cpu, struct task_struct *idle)
arch/arm/include/asm/smp.h
94
int (*smp_boot_secondary)(unsigned int cpu, struct task_struct *idle);
arch/arm/kernel/psci_smp.c
45
static int psci_boot_secondary(unsigned int cpu, struct task_struct *idle)
arch/arm/kernel/smp.c
131
int __cpu_up(unsigned int cpu, struct task_struct *idle)
arch/arm/kernel/smp.c
146
secondary_data.stack = task_stack_page(idle) + THREAD_START_SP;
arch/arm/kernel/smp.c
155
secondary_data.task = idle;
arch/arm/kernel/smp.c
161
ret = smp_ops.smp_boot_secondary(cpu, idle);
arch/arm/mach-actions/platsmp.c
79
static int s500_smp_boot_secondary(unsigned int cpu, struct task_struct *idle)
arch/arm/mach-alpine/platsmp.c
17
static int alpine_boot_secondary(unsigned int cpu, struct task_struct *idle)
arch/arm/mach-aspeed/platsmp.c
15
static int aspeed_g6_boot_secondary(unsigned int cpu, struct task_struct *idle)
arch/arm/mach-at91/pm.c
872
void (*idle)(void);
arch/arm/mach-at91/pm.c
877
{ .idle = at91rm9200_standby, .memctrl = AT91_MEMCTRL_MC},
arch/arm/mach-at91/pm.c
878
{ .idle = at91sam9_sdram_standby, .memctrl = AT91_MEMCTRL_SDRAMC},
arch/arm/mach-at91/pm.c
879
{ .idle = at91_ddr_standby, .memctrl = AT91_MEMCTRL_DDRSDR},
arch/arm/mach-at91/pm.c
880
{ .idle = sama5d3_ddr_standby, .memctrl = AT91_MEMCTRL_DDRSDR},
arch/arm/mach-at91/pm.c
881
{ .idle = sama7g5_standby, },
arch/arm/mach-at91/pm.c
919
standby = ramc->idle;
arch/arm/mach-axxia/platsmp.c
31
static int axxia_boot_secondary(unsigned int cpu, struct task_struct *idle)
arch/arm/mach-bcm/bcm63xx_smp.c
107
struct task_struct *idle)
arch/arm/mach-bcm/platsmp-brcmstb.c
339
static int brcmstb_boot_secondary(unsigned int cpu, struct task_struct *idle)
arch/arm/mach-bcm/platsmp.c
158
static int kona_boot_secondary(unsigned int cpu, struct task_struct *idle)
arch/arm/mach-bcm/platsmp.c
223
static int bcm23550_boot_secondary(unsigned int cpu, struct task_struct *idle)
arch/arm/mach-bcm/platsmp.c
249
ret = kona_boot_secondary(cpu, idle);
arch/arm/mach-bcm/platsmp.c
264
static int nsp_boot_secondary(unsigned int cpu, struct task_struct *idle)
arch/arm/mach-bcm/platsmp.c
285
static int bcm2836_boot_secondary(unsigned int cpu, struct task_struct *idle)
arch/arm/mach-berlin/platsmp.c
44
static int berlin_boot_secondary(unsigned int cpu, struct task_struct *idle)
arch/arm/mach-exynos/platsmp.c
317
static int exynos_boot_secondary(unsigned int cpu, struct task_struct *idle)
arch/arm/mach-hisi/platmcpm.c
98
static int hip04_boot_secondary(unsigned int l_cpu, struct task_struct *idle)
arch/arm/mach-hisi/platsmp.c
117
static int hix5hd2_boot_secondary(unsigned int cpu, struct task_struct *idle)
arch/arm/mach-hisi/platsmp.c
152
static int hip01_boot_secondary(unsigned int cpu, struct task_struct *idle)
arch/arm/mach-hisi/platsmp.c
84
static int hi3xxx_boot_secondary(unsigned int cpu, struct task_struct *idle)
arch/arm/mach-imx/platsmp.c
123
static int ls1021a_boot_secondary(unsigned int cpu, struct task_struct *idle)
arch/arm/mach-imx/platsmp.c
43
static int imx_boot_secondary(unsigned int cpu, struct task_struct *idle)
arch/arm/mach-mediatek/platsmp.c
69
static int mtk_boot_secondary(unsigned int cpu, struct task_struct *idle)
arch/arm/mach-meson/platsmp.c
156
struct task_struct *idle)
arch/arm/mach-meson/platsmp.c
212
struct task_struct *idle)
arch/arm/mach-milbeaut/platsmp.c
23
static int m10v_boot_secondary(unsigned int l_cpu, struct task_struct *idle)
arch/arm/mach-mmp/platsmp.c
12
static int mmp3_boot_secondary(unsigned int cpu, struct task_struct *idle)
arch/arm/mach-mstar/mstarv7.c
74
static int mstarv7_boot_secondary(unsigned int cpu, struct task_struct *idle)
arch/arm/mach-mvebu/platsmp-a9.c
25
struct task_struct *idle)
arch/arm/mach-mvebu/platsmp.c
216
static int mv98dx3236_boot_secondary(unsigned int cpu, struct task_struct *idle)
arch/arm/mach-mvebu/platsmp.c
50
static int armada_xp_boot_secondary(unsigned int cpu, struct task_struct *idle)
arch/arm/mach-npcm/platsmp.c
22
struct task_struct *idle)
arch/arm/mach-omap2/omap-smp.c
173
static int omap4_boot_secondary(unsigned int cpu, struct task_struct *idle)
arch/arm/mach-omap2/pm33xx-core.c
100
idle_fn = idle;
arch/arm/mach-omap2/pm33xx-core.c
105
static int am33xx_suspend_init(int (*idle)(u32 wfi_flags))
arch/arm/mach-omap2/pm33xx-core.c
114
return amx3_common_init(idle);
arch/arm/mach-omap2/pm33xx-core.c
117
static int am43xx_suspend_init(int (*idle)(u32 wfi_flags))
arch/arm/mach-omap2/pm33xx-core.c
127
ret = amx3_common_init(idle);
arch/arm/mach-omap2/pm33xx-core.c
80
static int amx3_common_init(int (*idle)(u32 wfi_flags))
arch/arm/mach-omap2/vc.c
373
struct omap3_vc_timings *c, u32 idle)
arch/arm/mach-omap2/vc.c
377
val = (voltdm->vc_param->on - idle) / voltdm->pmic->slew_rate;
arch/arm/mach-qcom/platsmp.c
336
static int msm8660_boot_secondary(unsigned int cpu, struct task_struct *idle)
arch/arm/mach-qcom/platsmp.c
341
static int cortex_a7_boot_secondary(unsigned int cpu, struct task_struct *idle)
arch/arm/mach-qcom/platsmp.c
346
static int kpssv1_boot_secondary(unsigned int cpu, struct task_struct *idle)
arch/arm/mach-qcom/platsmp.c
351
static int kpssv2_boot_secondary(unsigned int cpu, struct task_struct *idle)
arch/arm/mach-rockchip/platsmp.c
116
static int rockchip_boot_secondary(unsigned int cpu, struct task_struct *idle)
arch/arm/mach-shmobile/platsmp-apmu.c
256
struct task_struct *idle)
arch/arm/mach-shmobile/smp-emev2.c
24
static int emev2_boot_secondary(unsigned int cpu, struct task_struct *idle)
arch/arm/mach-shmobile/smp-r8a7779.c
28
static int r8a7779_boot_secondary(unsigned int cpu, struct task_struct *idle)
arch/arm/mach-shmobile/smp-sh73a0.c
32
static int sh73a0_boot_secondary(unsigned int cpu, struct task_struct *idle)
arch/arm/mach-socfpga/platsmp.c
21
static int socfpga_boot_secondary(unsigned int cpu, struct task_struct *idle)
arch/arm/mach-socfpga/platsmp.c
46
static int socfpga_a10_boot_secondary(unsigned int cpu, struct task_struct *idle)
arch/arm/mach-spear/platsmp.c
56
static int spear13xx_boot_secondary(unsigned int cpu, struct task_struct *idle)
arch/arm/mach-sti/platsmp.c
30
static int sti_boot_secondary(unsigned int cpu, struct task_struct *idle)
arch/arm/mach-sunxi/mc_smp.c
391
static int sunxi_mc_smp_boot_secondary(unsigned int l_cpu, struct task_struct *idle)
arch/arm/mach-sunxi/platsmp.c
157
struct task_struct *idle)
arch/arm/mach-sunxi/platsmp.c
72
struct task_struct *idle)
arch/arm/mach-tegra/platsmp.c
129
static int tegra114_boot_secondary(unsigned int cpu, struct task_struct *idle)
arch/arm/mach-tegra/platsmp.c
159
struct task_struct *idle)
arch/arm/mach-tegra/platsmp.c
162
return tegra20_boot_secondary(cpu, idle);
arch/arm/mach-tegra/platsmp.c
164
return tegra30_boot_secondary(cpu, idle);
arch/arm/mach-tegra/platsmp.c
166
return tegra114_boot_secondary(cpu, idle);
arch/arm/mach-tegra/platsmp.c
168
return tegra114_boot_secondary(cpu, idle);
arch/arm/mach-tegra/platsmp.c
42
static int tegra20_boot_secondary(unsigned int cpu, struct task_struct *idle)
arch/arm/mach-tegra/platsmp.c
70
static int tegra30_boot_secondary(unsigned int cpu, struct task_struct *idle)
arch/arm/mach-ux500/platsmp.c
67
static int ux500_boot_secondary(unsigned int cpu, struct task_struct *idle)
arch/arm/mach-versatile/platsmp.c
66
int versatile_boot_secondary(unsigned int cpu, struct task_struct *idle)
arch/arm/mach-versatile/platsmp.h
10
extern int versatile_boot_secondary(unsigned int cpu, struct task_struct *idle);
arch/arm/mach-zynq/platsmp.c
84
static int zynq_boot_secondary(unsigned int cpu, struct task_struct *idle)
arch/arm64/kernel/smp.c
111
int __cpu_up(unsigned int cpu, struct task_struct *idle)
arch/arm64/kernel/smp.c
120
secondary_data.task = idle;
arch/arm64/kernel/smp.c
124
ret = boot_secondary(cpu, idle);
arch/arm64/kernel/smp.c
99
static int boot_secondary(unsigned int cpu, struct task_struct *idle)
arch/hexagon/kernel/smp.c
174
int __cpu_up(unsigned int cpu, struct task_struct *idle)
arch/hexagon/kernel/smp.c
176
struct thread_info *thread = (struct thread_info *)idle->stack;
arch/loongarch/include/asm/smp.h
34
void loongson_boot_secondary(int cpu, struct task_struct *idle);
arch/loongarch/kernel/smp.c
396
void loongson_boot_secondary(int cpu, struct task_struct *idle)
arch/loongarch/kernel/smp.c
403
cpuboot_data.stack = (unsigned long)__KSTK_TOS(idle);
arch/loongarch/kernel/smp.c
404
cpuboot_data.thread_info = (unsigned long)task_thread_info(idle);
arch/m68k/fpsp040/fpsp.h
335
.set IDLE_SIZE,4 | size of idle frame
arch/m68k/fpsp040/fpsp.h
336
.set IDLE_FRAME,LV-IDLE_SIZE | start of idle frame
arch/m68k/fpsp040/fpsp.h
60
| if the FPU state after the exception is idle.
arch/mips/cavium-octeon/smp.c
209
static int octeon_boot_secondary(int cpu, struct task_struct *idle)
arch/mips/cavium-octeon/smp.c
216
octeon_processor_sp = __KSTK_TOS(idle);
arch/mips/cavium-octeon/smp.c
217
octeon_processor_gp = (unsigned long)(task_thread_info(idle));
arch/mips/include/asm/octeon/cvmx-dpi-defs.h
184
uint64_t idle:1;
arch/mips/include/asm/octeon/cvmx-dpi-defs.h
190
uint64_t idle:1;
arch/mips/include/asm/octeon/cvmx-dpi-defs.h
201
uint64_t idle:1;
arch/mips/include/asm/octeon/cvmx-dpi-defs.h
209
uint64_t idle:1;
arch/mips/include/asm/octeon/cvmx-npei-defs.h
966
uint64_t idle:1;
arch/mips/include/asm/octeon/cvmx-npei-defs.h
972
uint64_t idle:1;
arch/mips/include/asm/smp-ops.h
27
int (*boot_secondary)(int cpu, struct task_struct *idle);
arch/mips/kernel/smp-bmips.c
191
static int bmips_boot_secondary(int cpu, struct task_struct *idle)
arch/mips/kernel/smp-bmips.c
193
bmips_smp_boot_sp = __KSTK_TOS(idle);
arch/mips/kernel/smp-bmips.c
194
bmips_smp_boot_gp = (unsigned long)task_thread_info(idle);
arch/mips/kernel/smp-cps.c
617
static int cps_boot_secondary(int cpu, struct task_struct *idle)
arch/mips/kernel/smp-cps.c
630
vpe_cfg->sp = __KSTK_TOS(idle);
arch/mips/kernel/smp-cps.c
631
vpe_cfg->gp = (unsigned long)task_thread_info(idle);
arch/mips/kernel/smp-mt.c
143
static int vsmp_boot_secondary(int cpu, struct task_struct *idle)
arch/mips/kernel/smp-mt.c
145
struct thread_info *gp = task_thread_info(idle);
arch/mips/kernel/smp-mt.c
163
write_tc_gpr_sp( __KSTK_TOS(idle));
arch/mips/kernel/smp-up.c
42
static int up_boot_secondary(int cpu, struct task_struct *idle)
arch/mips/loongson64/smp.c
104
startargs[1] = (unsigned long)__KSTK_TOS(idle);
arch/mips/loongson64/smp.c
105
startargs[2] = (unsigned long)task_thread_info(idle);
arch/mips/loongson64/smp.c
146
static void legacy_ipi_write_buf(int cpu, struct task_struct *idle)
arch/mips/loongson64/smp.c
152
startargs[1] = (unsigned long)__KSTK_TOS(idle);
arch/mips/loongson64/smp.c
153
startargs[2] = (unsigned long)task_thread_info(idle);
arch/mips/loongson64/smp.c
41
static void (*ipi_write_buf)(int cpu, struct task_struct *idle);
arch/mips/loongson64/smp.c
503
static int loongson3_boot_secondary(int cpu, struct task_struct *idle)
arch/mips/loongson64/smp.c
507
ipi_write_buf(cpu, idle);
arch/mips/loongson64/smp.c
98
static void csr_ipi_write_buf(int cpu, struct task_struct *idle)
arch/mips/sgi-ip27/ip27-smp.c
147
static int ip27_boot_secondary(int cpu, struct task_struct *idle)
arch/mips/sgi-ip27/ip27-smp.c
149
unsigned long gp = (unsigned long)task_thread_info(idle);
arch/mips/sgi-ip27/ip27-smp.c
150
unsigned long sp = __KSTK_TOS(idle);
arch/mips/sgi-ip30/ip30-smp.c
110
static int __init ip30_smp_boot_secondary(int cpu, struct task_struct *idle)
arch/mips/sgi-ip30/ip30-smp.c
115
mpc->stackaddr = (void *)__KSTK_TOS(idle);
arch/mips/sgi-ip30/ip30-smp.c
118
mpc->lnch_parm = task_thread_info(idle);
arch/mips/sibyte/bcm1480/smp.c
107
static int bcm1480_boot_secondary(int cpu, struct task_struct *idle)
arch/mips/sibyte/bcm1480/smp.c
112
__KSTK_TOS(idle),
arch/mips/sibyte/bcm1480/smp.c
113
(unsigned long)task_thread_info(idle), 0);
arch/mips/sibyte/sb1250/smp.c
101
__KSTK_TOS(idle),
arch/mips/sibyte/sb1250/smp.c
102
(unsigned long)task_thread_info(idle), 0);
arch/mips/sibyte/sb1250/smp.c
96
static int sb1250_boot_secondary(int cpu, struct task_struct *idle)
arch/openrisc/kernel/smp.c
102
int __cpu_up(unsigned int cpu, struct task_struct *idle)
arch/openrisc/kernel/smp.c
110
secondary_thread_info = task_thread_info(idle);
arch/openrisc/kernel/smp.c
113
boot_secondary(cpu, idle);
arch/openrisc/kernel/smp.c
52
static void boot_secondary(unsigned int cpu, struct task_struct *idle)
arch/parisc/kernel/process.c
222
if (args->idle) /* idle thread */
arch/parisc/kernel/smp.c
333
static int smp_boot_one_cpu(int cpuid, struct task_struct *idle)
arch/parisc/kernel/smp.c
364
smp_init_current_idle_task = idle ;
arch/powerpc/include/asm/idle.h
49
get_lppaca()->idle = 1;
arch/powerpc/include/asm/idle.h
56
get_lppaca()->idle = 0;
arch/powerpc/include/asm/idle.h
68
if (unlikely(get_lppaca()->idle == 1)) {
arch/powerpc/include/asm/idle.h
84
if (get_lppaca()->idle == 1) {
arch/powerpc/include/asm/lppaca.h
70
u8 idle; /* Indicate OS is idle */
arch/powerpc/include/asm/paravirt.h
74
return lppaca_of(vcpu).idle;
arch/powerpc/kernel/smp.c
1280
static void cpu_idle_thread_init(unsigned int cpu, struct task_struct *idle)
arch/powerpc/kernel/smp.c
1283
paca_ptrs[cpu]->__current = idle;
arch/powerpc/kernel/smp.c
1284
paca_ptrs[cpu]->kstack = (unsigned long)task_stack_page(idle) +
arch/powerpc/kernel/smp.c
1287
task_thread_info(idle)->cpu = cpu;
arch/powerpc/kernel/smp.c
1288
secondary_current = current_set[cpu] = idle;
arch/powerpc/platforms/44x/cpm.c
164
__ATTR(idle, 0644, cpm_idle_show, cpm_idle_store);
arch/powerpc/sysdev/mpic_timer.c
120
map = casc_priv->cascade_map & priv->idle;
arch/powerpc/sysdev/mpic_timer.c
126
priv->idle &= ~casc_priv->cascade_map;
arch/powerpc/sysdev/mpic_timer.c
221
if (priv->idle & (1 << i)) {
arch/powerpc/sysdev/mpic_timer.c
223
priv->idle &= ~(1 << i);
arch/powerpc/sysdev/mpic_timer.c
339
priv->idle |= casc_priv->cascade_map;
arch/powerpc/sysdev/mpic_timer.c
342
priv->idle |= TIMER_OFFSET(handle->num);
arch/powerpc/sysdev/mpic_timer.c
451
priv->idle |= TIMER_OFFSET((offset + j));
arch/powerpc/sysdev/mpic_timer.c
67
unsigned int idle;
arch/s390/include/asm/vtime.h
31
struct s390_idle_data *idle = this_cpu_ptr(&s390_idle);
arch/s390/include/asm/vtime.h
40
__this_cpu_add(mt_cycles[i], cycles_new[i] - idle->mt_cycles_enter[i]);
arch/s390/include/asm/vtime.h
51
lc->steal_timer += idle->clock_idle_enter - lc->last_update_clock;
arch/s390/include/asm/vtime.h
53
lc->system_timer += lc->last_update_timer - idle->timer_idle_enter;
arch/s390/kernel/idle.c
26
struct s390_idle_data *idle = this_cpu_ptr(&s390_idle);
arch/s390/kernel/idle.c
29
idle_time = get_lowcore()->int_clock - idle->clock_idle_enter;
arch/s390/kernel/idle.c
32
__atomic64_add(idle_time, &idle->idle_time);
arch/s390/kernel/idle.c
33
__atomic64_add_const(1, &idle->idle_count);
arch/s390/kernel/idle.c
39
struct s390_idle_data *idle = this_cpu_ptr(&s390_idle);
arch/s390/kernel/idle.c
48
stcctm(MT_DIAG, smp_cpu_mtid, (u64 *)&idle->mt_cycles_enter);
arch/s390/kernel/idle.c
49
idle->clock_idle_enter = get_tod_clock_fast();
arch/s390/kernel/idle.c
50
idle->timer_idle_enter = get_cpu_timer();
arch/s390/kernel/idle.c
58
struct s390_idle_data *idle = &per_cpu(s390_idle, dev->id);
arch/s390/kernel/idle.c
60
return sysfs_emit(buf, "%lu\n", READ_ONCE(idle->idle_count));
arch/s390/kernel/idle.c
67
struct s390_idle_data *idle = &per_cpu(s390_idle, dev->id);
arch/s390/kernel/idle.c
69
return sysfs_emit(buf, "%lu\n", READ_ONCE(idle->idle_time) >> 12);
arch/s390/kvm/sigp.c
73
bool idle;
arch/s390/kvm/sigp.c
75
idle = is_vcpu_idle(vcpu);
arch/s390/kvm/sigp.c
83
|| (idle && psw->addr != 0)
arch/s390/kvm/sigp.c
84
|| (!idle && (asn == p_asn || asn == s_asn))) {
arch/sparc/kernel/kernel.h
109
int smp4m_boot_one_cpu(int i, struct task_struct *idle);
arch/sparc/kernel/kernel.h
130
int smp4d_boot_one_cpu(int i, struct task_struct *idle);
arch/sparc/kernel/leon_smp.c
181
int leon_boot_one_cpu(int i, struct task_struct *idle)
arch/sparc/kernel/leon_smp.c
185
current_set[i] = task_thread_info(idle);
arch/sparc/kernel/smp_64.c
347
static int smp_boot_one_cpu(unsigned int cpu, struct task_struct *idle)
arch/sparc/kernel/smp_64.c
357
cpu_new_thread = task_thread_info(idle);
arch/sparc/kernel/sun4d_smp.c
122
int smp4d_boot_one_cpu(int i, struct task_struct *idle)
arch/sparc/kernel/sun4d_smp.c
129
current_set[i] = task_thread_info(idle);
arch/sparc/kernel/sun4m_smp.c
79
int smp4m_boot_one_cpu(int i, struct task_struct *idle)
arch/sparc/kernel/sun4m_smp.c
86
current_set[i] = task_thread_info(idle);
arch/um/kernel/smp.c
141
struct task_struct *idle;
arch/um/kernel/smp.c
151
idle = cpu_tasks[cpu];
arch/um/kernel/smp.c
152
idle->thread_info.cpu = cpu;
arch/um/kernel/smp.c
155
idle->active_mm = mm;
arch/um/kernel/smp.c
157
idle->thread.request.thread.proc = start_secondary;
arch/um/kernel/smp.c
158
idle->thread.request.thread.arg = NULL;
arch/um/kernel/smp.c
160
new_thread(task_stack_page(idle), &idle->thread.switch_buf,
arch/um/kernel/smp.c
162
os_start_secondary(opaque, &idle->thread.switch_buf);
arch/um/kernel/time.c
346
static void time_travel_ext_wait(bool idle)
arch/um/kernel/time.c
367
time_travel_handle_message(&msg, idle ? TTMH_IDLE : TTMH_POLL);
arch/um/kernel/time.c
384
static void __time_travel_update_time(unsigned long long ns, bool idle)
arch/um/kernel/time.c
387
time_travel_ext_wait(idle);
arch/um/kernel/time.c
531
static void time_travel_update_time(unsigned long long next, bool idle)
arch/um/kernel/time.c
537
bool finished = idle;
arch/um/kernel/time.c
546
__time_travel_update_time(e->time, idle);
arch/um/kernel/time.c
732
static inline void time_travel_update_time(unsigned long long ns, bool idle)
arch/x86/hyperv/hv_vtl.c
125
struct task_struct *idle = idle_thread_get(cpu);
arch/x86/hyperv/hv_vtl.c
126
if (IS_ERR(idle))
arch/x86/hyperv/hv_vtl.c
127
return PTR_ERR(idle);
arch/x86/hyperv/hv_vtl.c
129
rsp = (unsigned long)idle->thread.sp;
arch/x86/include/asm/stackprotector.h
44
static inline void cpu_init_stack_canary(int cpu, struct task_struct *idle)
arch/x86/include/asm/stackprotector.h
46
per_cpu(__stack_chk_guard, cpu) = idle->stack_canary;
arch/x86/include/asm/stackprotector.h
53
static inline void cpu_init_stack_canary(int cpu, struct task_struct *idle)
arch/x86/kernel/smpboot.c
1008
idle->thread.sp = (unsigned long)task_pt_regs(idle);
arch/x86/kernel/smpboot.c
1013
initial_stack = idle->thread.sp;
arch/x86/kernel/smpboot.c
970
int common_cpu_up(unsigned int cpu, struct task_struct *idle)
arch/x86/kernel/smpboot.c
977
per_cpu(current_task, cpu) = idle;
arch/x86/kernel/smpboot.c
978
cpu_init_stack_canary(cpu, idle);
arch/x86/kernel/smpboot.c
987
per_cpu(cpu_current_top_of_stack, cpu) = task_top_of_stack(idle);
arch/x86/kernel/smpboot.c
998
static int do_boot_cpu(u32 apicid, unsigned int cpu, struct task_struct *idle)
arch/x86/xen/smp_pv.c
224
cpu_initialize_context(unsigned int cpu, struct task_struct *idle)
arch/x86/xen/smp_pv.c
253
ctxt->user_regs.esp = (unsigned long)task_pt_regs(idle);
arch/x86/xen/smp_pv.c
272
ctxt->kernel_sp = task_top_of_stack(idle);
arch/x86/xen/smp_pv.c
289
static int xen_pv_kick_ap(unsigned int cpu, struct task_struct *idle)
arch/x86/xen/smp_pv.c
293
rc = common_cpu_up(cpu, idle);
arch/x86/xen/smp_pv.c
302
rc = cpu_initialize_context(cpu, idle);
arch/xtensa/kernel/smp.c
234
int __cpu_up(unsigned int cpu, struct task_struct *idle)
arch/xtensa/kernel/smp.c
241
start_info.stack = (unsigned long)task_pt_regs(idle);
arch/xtensa/kernel/smp.c
245
__func__, cpu, idle, start_info.stack);
arch/xtensa/kernel/smp.c
248
ret = boot_secondary(cpu, idle);
block/bfq-iosched.h
57
struct rb_root idle;
block/bfq-wf2q.c
1143
else if (!is_in_service && entity->tree == &st->idle)
block/bfq-wf2q.c
365
bfq_extract(&st->idle, entity);
block/bfq-wf2q.c
604
bfq_insert(&st->idle, entity);
block/bfq-wf2q.c
941
if (entity->tree == &st->idle) {
block/mq-deadline.c
882
u32 rt, be, idle;
block/mq-deadline.c
887
idle = dd_queued(dd, DD_IDLE_PRIO);
block/mq-deadline.c
890
seq_printf(m, "%u %u %u\n", rt, be, idle);
block/mq-deadline.c
910
u32 rt, be, idle;
block/mq-deadline.c
915
idle = dd_owned_by_driver(dd, DD_IDLE_PRIO);
block/mq-deadline.c
918
seq_printf(m, "%u %u %u\n", rt, be, idle);
drivers/ata/libata-sff.c
1467
unsigned int handled, idle, polling;
drivers/ata/libata-sff.c
1474
handled = idle = polling = 0;
drivers/ata/libata-sff.c
1486
idle |= 1 << i;
drivers/ata/libata-sff.c
1507
if (idle & (1 << i)) {
drivers/atm/idt77252.c
1537
unsigned long base, idle, jump;
drivers/atm/idt77252.c
1545
idle = card->tst[card->tst_index ^ 1];
drivers/atm/idt77252.c
1551
if ((pc ^ idle) & ~(card->tst_size - 1)) {
drivers/atm/idt77252.c
1562
idle = card->tst[card->tst_index ^ 1];
drivers/atm/idt77252.c
1566
write_sram(card, idle + e,
drivers/atm/idt77252.c
1577
write_sram(card, idle + e,
drivers/atm/idt77252.c
1601
unsigned long idle;
drivers/atm/idt77252.c
1624
idle = card->tst[card->tst_index ^ 1];
drivers/atm/idt77252.c
1640
write_sram(card, idle + e, data);
drivers/atm/idt77252.c
1677
unsigned long idle;
drivers/atm/idt77252.c
1680
idle = card->tst[card->tst_index ^ 1];
drivers/atm/idt77252.c
1690
write_sram(card, idle + e, TSTE_OPC_VAR);
drivers/block/swim3.c
1134
fs->state = idle;
drivers/block/swim3.c
315
if (fs->cur_req || fs->state != idle) {
drivers/block/swim3.c
494
case idle:
drivers/block/swim3.c
536
fs->state = idle;
drivers/block/swim3.c
572
fs->state = idle;
drivers/block/swim3.c
595
fs->state = idle;
drivers/block/swim3.c
623
fs->state = idle;
drivers/block/swim3.c
651
fs->state = idle;
drivers/block/swim3.c
687
fs->state = idle;
drivers/block/swim3.c
769
fs->state = idle;
drivers/block/swim3.c
778
fs->state = idle;
drivers/block/swim3.c
793
fs->state = idle;
drivers/block/swim3.c
818
if (fs->state != idle && fs->state != available) {
drivers/block/swim3.c
849
fs->state = idle;
drivers/block/ublk_drv.c
2831
bool *idle = data;
drivers/block/ublk_drv.c
2834
*idle = false;
drivers/block/ublk_drv.c
2842
bool idle;
drivers/block/ublk_drv.c
2846
idle = true;
drivers/block/ublk_drv.c
2848
ublk_check_inflight_rq, &idle);
drivers/block/ublk_drv.c
2849
if (idle)
drivers/block/ublk_drv.c
5036
struct count_busy *idle = data;
drivers/block/ublk_drv.c
5038
if (!blk_mq_request_started(rq) && rq->mq_hctx->driver_data == idle->ubq)
drivers/block/ublk_drv.c
5039
idle->nr_busy += 1;
drivers/block/zram/zram_drv.c
2990
static DEVICE_ATTR_WO(idle);
drivers/dma/altera-msgdma.c
179
bool idle;
drivers/dma/altera-msgdma.c
491
mdev->idle = true;
drivers/dma/altera-msgdma.c
520
mdev->idle = false;
drivers/dma/altera-msgdma.c
551
if (!mdev->idle)
drivers/dma/altera-msgdma.c
664
mdev->idle = true;
drivers/dma/altera-msgdma.c
740
mdev->idle = true;
drivers/dma/fsldma.c
1179
chan->idle = true;
drivers/dma/fsldma.c
1344
if (unlikely(!chan->idle))
drivers/dma/fsldma.c
585
if (!chan->idle) {
drivers/dma/fsldma.c
624
chan->idle = false;
drivers/dma/fsldma.c
842
chan->idle = true;
drivers/dma/fsldma.c
989
chan->idle = true;
drivers/dma/fsldma.h
178
bool idle; /* DMA controller is idle */
drivers/dma/mmp_pdma.c
1088
chan->idle = list_empty(&chan->chain_running);
drivers/dma/mmp_pdma.c
125
bool idle; /* channel statue machine */
drivers/dma/mmp_pdma.c
459
if (!chan->idle) {
drivers/dma/mmp_pdma.c
493
chan->idle = false;
drivers/dma/mmp_pdma.c
567
chan->idle = true;
drivers/dma/mmp_pdma.c
595
chan->idle = true;
drivers/dma/mmp_pdma.c
921
chan->idle = true;
drivers/dma/xilinx/xilinx_dma.c
1386
if (!chan->idle)
drivers/dma/xilinx/xilinx_dma.c
1471
chan->idle = false;
drivers/dma/xilinx/xilinx_dma.c
1487
if (!chan->idle)
drivers/dma/xilinx/xilinx_dma.c
1543
chan->idle = false;
drivers/dma/xilinx/xilinx_dma.c
1578
chan->idle = false;
drivers/dma/xilinx/xilinx_dma.c
1583
if (!chan->idle)
drivers/dma/xilinx/xilinx_dma.c
1642
chan->idle = false;
drivers/dma/xilinx/xilinx_dma.c
1663
if (!chan->idle)
drivers/dma/xilinx/xilinx_dma.c
1712
chan->idle = false;
drivers/dma/xilinx/xilinx_dma.c
1808
chan->idle = true;
drivers/dma/xilinx/xilinx_dma.c
1896
chan->idle = true;
drivers/dma/xilinx/xilinx_dma.c
1953
chan->idle = true;
drivers/dma/xilinx/xilinx_dma.c
2637
chan->idle = true;
drivers/dma/xilinx/xilinx_dma.c
2963
chan->idle = true;
drivers/dma/xilinx/xilinx_dma.c
442
bool idle;
drivers/dma/xilinx/zynqmp_dma.c
236
bool idle;
drivers/dma/xilinx/zynqmp_dma.c
373
chan->idle = true;
drivers/dma/xilinx/zynqmp_dma.c
489
chan->idle = true;
drivers/dma/xilinx/zynqmp_dma.c
529
chan->idle = false;
drivers/dma/xilinx/zynqmp_dma.c
594
if (!chan->idle)
drivers/dma/xilinx/zynqmp_dma.c
743
chan->idle = true;
drivers/dma/xilinx/zynqmp_dma.c
787
if (chan->idle) {
drivers/dma/xilinx/zynqmp_dma.c
961
chan->idle = true;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.c
705
void amdgpu_amdkfd_set_compute_idle(struct amdgpu_device *adev, bool idle)
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.c
707
enum amd_powergating_state state = idle ? AMD_PG_STATE_GATE : AMD_PG_STATE_UNGATE;
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.c
711
pr_debug("GFXOFF is %s\n", idle ? "enabled" : "disabled");
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.c
712
amdgpu_gfx_off_ctrl(adev, idle);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.c
722
(void)amdgpu_dpm_switch_power_profile(adev, PP_SMC_POWER_PROFILE_COMPUTE, !idle);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd.h
178
void amdgpu_amdkfd_set_compute_idle(struct amdgpu_device *adev, bool idle);
drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c
208
struct amdgpu_vmid **idle,
drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c
222
list_for_each_entry_reverse((*idle), &id_mgr->ids_lru, list) {
drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c
227
*fence = amdgpu_sync_peek_fence(&(*idle)->active, r);
drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c
236
*idle = NULL;
drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c
390
struct amdgpu_vmid *idle = NULL;
drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c
395
r = amdgpu_vmid_grab_idle(ring, &idle, fence);
drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c
396
if (r || !idle)
drivers/gpu/drm/amd/amdgpu/amdgpu_ids.c
410
id = idle;
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
191
list_move(&vm_bo->vm_status, &vm_bo->vm->idle);
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
2594
INIT_LIST_HEAD(&vm->idle);
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
281
list_for_each_entry_safe(vm_bo, tmp, &vm->idle, vm_status) {
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
3072
list_for_each_entry_safe(bo_va, tmp, &vm->idle, base.vm_status) {
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.h
371
struct list_head idle;
drivers/gpu/drm/amd/display/dc/dcn10/dcn10_hw_sequencer_debug.c
404
s.idle);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/dml21_translation_helper.c
811
context->bw_ctx.bw.dcn.clk.idle_dramclk_khz = in_ctx->v21.mode_programming.programming->min_clocks.dcn4x.idle.uclk_khz;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/dml21_translation_helper.c
812
context->bw_ctx.bw.dcn.clk.idle_fclk_khz = in_ctx->v21.mode_programming.programming->min_clocks.dcn4x.idle.fclk_khz;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/inc/dml_top_types.h
390
} idle;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_dpmm/dml2_dpmm_dcn4.c
200
in_out->programming->min_clocks.dcn4x.idle.uclk_khz = dml_round_up(min_uclk_avg > min_uclk_latency ? min_uclk_avg : min_uclk_latency);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_dpmm/dml2_dpmm_dcn4.c
201
in_out->programming->min_clocks.dcn4x.idle.fclk_khz = dml_round_up(min_fclk_avg > min_fclk_latency ? min_fclk_avg : min_fclk_latency);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_dpmm/dml2_dpmm_dcn4.c
202
in_out->programming->min_clocks.dcn4x.idle.dcfclk_khz = dml_round_up(min_dcfclk_avg > min_dcfclk_latency ? min_dcfclk_avg : min_dcfclk_latency);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_dpmm/dml2_dpmm_dcn4.c
320
result = round_up_to_next_dpm(&display_cfg->min_clocks.dcn4x.idle.dcfclk_khz, &state_table->dcfclk);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_dpmm/dml2_dpmm_dcn4.c
322
result = round_up_to_next_dpm(&display_cfg->min_clocks.dcn4x.idle.fclk_khz, &state_table->fclk);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_dpmm/dml2_dpmm_dcn4.c
324
result = round_up_to_next_dpm(&display_cfg->min_clocks.dcn4x.idle.uclk_khz, &state_table->uclk);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_dpmm/dml2_dpmm_dcn4.c
361
if (display_cfg->min_clocks.dcn4x.idle.dcfclk_khz <= state_table->dcfclk.clk_values_khz[index] &&
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_dpmm/dml2_dpmm_dcn4.c
362
display_cfg->min_clocks.dcn4x.idle.fclk_khz <= state_table->fclk.clk_values_khz[index] &&
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_dpmm/dml2_dpmm_dcn4.c
363
display_cfg->min_clocks.dcn4x.idle.uclk_khz <= state_table->uclk.clk_values_khz[index]) {
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_dpmm/dml2_dpmm_dcn4.c
364
display_cfg->min_clocks.dcn4x.idle.dcfclk_khz = state_table->dcfclk.clk_values_khz[index];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_dpmm/dml2_dpmm_dcn4.c
365
display_cfg->min_clocks.dcn4x.idle.fclk_khz = state_table->fclk.clk_values_khz[index];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_dpmm/dml2_dpmm_dcn4.c
366
display_cfg->min_clocks.dcn4x.idle.uclk_khz = state_table->uclk.clk_values_khz[index];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_dpmm/dml2_dpmm_dcn4.c
599
in_out->programming->min_clocks.dcn4x.idle.uclk_khz = in_out->soc_bb->clk_table.uclk.clk_values_khz[in_out->soc_bb->clk_table.uclk.num_clk_values - 1];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_dpmm/dml2_dpmm_dcn4.c
605
in_out->programming->min_clocks.dcn4x.idle.fclk_khz = in_out->soc_bb->clk_table.fclk.clk_values_khz[in_out->soc_bb->clk_table.fclk.num_clk_values - 1];
drivers/gpu/drm/amd/display/dc/hwss/dcn10/dcn10_hwseq.c
547
s.idle);
drivers/gpu/drm/amd/display/dc/hwss/dcn20/dcn20_hwseq.c
177
s.idle,
drivers/gpu/drm/amd/display/dc/hwss/dcn30/dcn30_hwseq.c
196
s.idle,
drivers/gpu/drm/amd/display/dc/inc/hw/mpc.h
341
uint32_t idle;
drivers/gpu/drm/amd/display/dc/mpc/dcn10/dcn10_mpc.c
456
REG_GET_2(MPCC_STATUS[mpcc_inst], MPCC_IDLE, &s->idle,
drivers/gpu/drm/amd/display/dc/mpc/dcn20/dcn20_mpc.c
557
REG_GET_2(MPCC_STATUS[mpcc_inst], MPCC_IDLE, &s->idle,
drivers/gpu/drm/amd/display/dc/mpc/dcn30/dcn30_mpc.c
1486
REG_GET_2(MPCC_STATUS[mpcc_inst], MPCC_IDLE, &s->idle,
drivers/gpu/drm/etnaviv/etnaviv_gpu.c
1051
seq_printf(m, "\tidle: 0x%08x\n", idle);
drivers/gpu/drm/etnaviv/etnaviv_gpu.c
1052
idle |= ~gpu->idle_mask & ~VIVS_HI_IDLE_STATE_AXI_LP;
drivers/gpu/drm/etnaviv/etnaviv_gpu.c
1053
if ((idle & VIVS_HI_IDLE_STATE_FE) == 0)
drivers/gpu/drm/etnaviv/etnaviv_gpu.c
1055
if ((idle & VIVS_HI_IDLE_STATE_DE) == 0)
drivers/gpu/drm/etnaviv/etnaviv_gpu.c
1057
if ((idle & VIVS_HI_IDLE_STATE_PE) == 0)
drivers/gpu/drm/etnaviv/etnaviv_gpu.c
1059
if ((idle & VIVS_HI_IDLE_STATE_SH) == 0)
drivers/gpu/drm/etnaviv/etnaviv_gpu.c
1061
if ((idle & VIVS_HI_IDLE_STATE_PA) == 0)
drivers/gpu/drm/etnaviv/etnaviv_gpu.c
1063
if ((idle & VIVS_HI_IDLE_STATE_SE) == 0)
drivers/gpu/drm/etnaviv/etnaviv_gpu.c
1065
if ((idle & VIVS_HI_IDLE_STATE_RA) == 0)
drivers/gpu/drm/etnaviv/etnaviv_gpu.c
1067
if ((idle & VIVS_HI_IDLE_STATE_TX) == 0)
drivers/gpu/drm/etnaviv/etnaviv_gpu.c
1069
if ((idle & VIVS_HI_IDLE_STATE_VG) == 0)
drivers/gpu/drm/etnaviv/etnaviv_gpu.c
1071
if ((idle & VIVS_HI_IDLE_STATE_IM) == 0)
drivers/gpu/drm/etnaviv/etnaviv_gpu.c
1073
if ((idle & VIVS_HI_IDLE_STATE_FP) == 0)
drivers/gpu/drm/etnaviv/etnaviv_gpu.c
1075
if ((idle & VIVS_HI_IDLE_STATE_TS) == 0)
drivers/gpu/drm/etnaviv/etnaviv_gpu.c
1077
if ((idle & VIVS_HI_IDLE_STATE_BL) == 0)
drivers/gpu/drm/etnaviv/etnaviv_gpu.c
1079
if ((idle & VIVS_HI_IDLE_STATE_ASYNCFE) == 0)
drivers/gpu/drm/etnaviv/etnaviv_gpu.c
1081
if ((idle & VIVS_HI_IDLE_STATE_MC) == 0)
drivers/gpu/drm/etnaviv/etnaviv_gpu.c
1083
if ((idle & VIVS_HI_IDLE_STATE_PPA) == 0)
drivers/gpu/drm/etnaviv/etnaviv_gpu.c
1085
if ((idle & VIVS_HI_IDLE_STATE_WD) == 0)
drivers/gpu/drm/etnaviv/etnaviv_gpu.c
1087
if ((idle & VIVS_HI_IDLE_STATE_NN) == 0)
drivers/gpu/drm/etnaviv/etnaviv_gpu.c
1089
if ((idle & VIVS_HI_IDLE_STATE_TP) == 0)
drivers/gpu/drm/etnaviv/etnaviv_gpu.c
1091
if (idle & VIVS_HI_IDLE_STATE_AXI_LP)
drivers/gpu/drm/etnaviv/etnaviv_gpu.c
1691
u32 idle = gpu_read(gpu, VIVS_HI_IDLE_STATE);
drivers/gpu/drm/etnaviv/etnaviv_gpu.c
1693
if ((idle & gpu->idle_mask) == gpu->idle_mask)
drivers/gpu/drm/etnaviv/etnaviv_gpu.c
1699
idle);
drivers/gpu/drm/etnaviv/etnaviv_gpu.c
1987
u32 idle, mask;
drivers/gpu/drm/etnaviv/etnaviv_gpu.c
1996
idle = gpu_read(gpu, VIVS_HI_IDLE_STATE) & mask;
drivers/gpu/drm/etnaviv/etnaviv_gpu.c
1997
if (idle != mask) {
drivers/gpu/drm/etnaviv/etnaviv_gpu.c
1999
idle);
drivers/gpu/drm/etnaviv/etnaviv_gpu.c
537
u32 control, idle;
drivers/gpu/drm/etnaviv/etnaviv_gpu.c
586
idle = gpu_read(gpu, VIVS_HI_IDLE_STATE);
drivers/gpu/drm/etnaviv/etnaviv_gpu.c
589
if ((idle & VIVS_HI_IDLE_STATE_FE) == 0) {
drivers/gpu/drm/etnaviv/etnaviv_gpu.c
613
idle = gpu_read(gpu, VIVS_HI_IDLE_STATE);
drivers/gpu/drm/etnaviv/etnaviv_gpu.c
617
idle & VIVS_HI_IDLE_STATE_FE ? "" : "not ",
drivers/gpu/drm/etnaviv/etnaviv_gpu.c
973
u32 dma_lo, dma_hi, axi, idle;
drivers/gpu/drm/etnaviv/etnaviv_gpu.c
985
idle = gpu_read(gpu, VIVS_HI_IDLE_STATE);
drivers/gpu/drm/i915/gt/intel_engine_cs.c
1841
bool idle = true;
drivers/gpu/drm/i915/gt/intel_engine_cs.c
1852
idle = false;
drivers/gpu/drm/i915/gt/intel_engine_cs.c
1857
idle = false;
drivers/gpu/drm/i915/gt/intel_engine_cs.c
1861
return idle;
drivers/gpu/drm/imagination/pvr_ccb.c
427
bool idle;
drivers/gpu/drm/imagination/pvr_ccb.c
431
idle = (READ_ONCE(ctrl->write_offset) == READ_ONCE(ctrl->read_offset));
drivers/gpu/drm/imagination/pvr_ccb.c
435
return idle;
drivers/gpu/drm/imagination/pvr_device.h
236
struct list_head idle;
drivers/gpu/drm/imagination/pvr_queue.c
1322
list_add_tail(&queue->node, &pvr_dev->queues.idle);
drivers/gpu/drm/imagination/pvr_queue.c
1351
list_for_each_entry(queue, &pvr_dev->queues.idle, node)
drivers/gpu/drm/imagination/pvr_queue.c
1365
list_for_each_entry(queue, &pvr_dev->queues.idle, node)
drivers/gpu/drm/imagination/pvr_queue.c
1432
INIT_LIST_HEAD(&pvr_dev->queues.idle);
drivers/gpu/drm/imagination/pvr_queue.c
584
list_move_tail(&queue->node, &pvr_dev->queues.idle);
drivers/gpu/drm/imagination/pvr_queue.c
847
list_move_tail(&queue->node, &pvr_dev->queues.idle);
drivers/gpu/drm/nouveau/nvkm/engine/fifo/ga100.c
284
.idle = ga100_runq_idle,
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk104.c
400
.idle = gk104_runq_idle,
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gk208.c
42
.idle = gk104_runq_idle,
drivers/gpu/drm/nouveau/nvkm/engine/fifo/gv100.c
172
.idle = gk104_runq_idle,
drivers/gpu/drm/nouveau/nvkm/engine/fifo/runl.c
93
if (runq->func->idle(runq))
drivers/gpu/drm/nouveau/nvkm/engine/fifo/runq.h
13
bool (*idle)(struct nvkm_runq *);
drivers/gpu/drm/nouveau/nvkm/engine/gr/g84.c
121
bool idle, timeout = false;
drivers/gpu/drm/nouveau/nvkm/engine/gr/g84.c
132
idle = true;
drivers/gpu/drm/nouveau/nvkm/engine/gr/g84.c
134
for (tmp = nvkm_rd32(device, 0x400380); tmp && idle; tmp >>= 3) {
drivers/gpu/drm/nouveau/nvkm/engine/gr/g84.c
136
idle = false;
drivers/gpu/drm/nouveau/nvkm/engine/gr/g84.c
139
for (tmp = nvkm_rd32(device, 0x400384); tmp && idle; tmp >>= 3) {
drivers/gpu/drm/nouveau/nvkm/engine/gr/g84.c
141
idle = false;
drivers/gpu/drm/nouveau/nvkm/engine/gr/g84.c
144
for (tmp = nvkm_rd32(device, 0x400388); tmp && idle; tmp >>= 3) {
drivers/gpu/drm/nouveau/nvkm/engine/gr/g84.c
146
idle = false;
drivers/gpu/drm/nouveau/nvkm/engine/gr/g84.c
148
} while (!idle &&
drivers/gpu/drm/panthor/panthor_sched.c
2240
&sched->groups.idle[group->priority] :
drivers/gpu/drm/panthor/panthor_sched.c
2260
&sched->groups.idle[group->priority] :
drivers/gpu/drm/panthor/panthor_sched.c
2413
list_move_tail(&group->run_node, &sched->groups.idle[prio]);
drivers/gpu/drm/panthor/panthor_sched.c
242
struct list_head idle[PANTHOR_CSG_PRIORITY_COUNT];
drivers/gpu/drm/panthor/panthor_sched.c
2535
tick_ctx_pick_groups_from_list(sched, &ctx, &sched->groups.idle[prio],
drivers/gpu/drm/panthor/panthor_sched.c
2769
&sched->groups.idle[group->priority] :
drivers/gpu/drm/panthor/panthor_sched.c
2929
&sched->groups.idle[group->priority]);
drivers/gpu/drm/panthor/panthor_sched.c
2967
for (i = 0; i < ARRAY_SIZE(sched->groups.idle); i++) {
drivers/gpu/drm/panthor/panthor_sched.c
2968
list_for_each_entry_safe(group, group_tmp, &sched->groups.idle[i], run_node)
drivers/gpu/drm/panthor/panthor_sched.c
3753
&sched->groups.idle[group->priority]);
drivers/gpu/drm/panthor/panthor_sched.c
4082
drm_WARN_ON(ddev, !list_empty(&sched->groups.idle[prio]));
drivers/gpu/drm/panthor/panthor_sched.c
4148
INIT_LIST_HEAD(&sched->groups.idle[prio]);
drivers/gpu/drm/rockchip/rockchip_drm_vop.c
1281
unsigned int idle;
drivers/gpu/drm/rockchip/rockchip_drm_vop.c
1303
idle, !idle, 5, 30 * 1000);
drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c
635
bool idle = false;
drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c
647
idle = list_empty(&man->error);
drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c
652
return idle;
drivers/gpu/drm/xe/xe_bo.c
673
bool idle = false;
drivers/gpu/drm/xe/xe_bo.c
695
if (!idle) {
drivers/gpu/drm/xe/xe_bo.c
712
idle = true;
drivers/gpu/host1x/cdma.c
664
bool idle = list_empty(&cdma->sync_queue);
drivers/gpu/host1x/cdma.c
674
if (job->timeout && idle)
drivers/hid/usbhid/hid-core.c
1320
static int usbhid_idle(struct hid_device *hid, int report, int idle,
drivers/hid/usbhid/hid-core.c
1331
return hid_set_idle(dev, ifnum, report, idle);
drivers/hid/usbhid/hid-core.c
1352
.idle = usbhid_idle,
drivers/hid/usbhid/hid-core.c
657
static int hid_set_idle(struct usb_device *dev, int ifnum, int report, int idle)
drivers/hid/usbhid/hid-core.c
660
HID_REQ_SET_IDLE, USB_TYPE_CLASS | USB_RECIP_INTERFACE, (idle << 8) | report,
drivers/i2c/busses/i2c-i801.c
1325
gpio_data.idle = I2C_MUX_GPIO_NO_IDLE;
drivers/i2c/muxes/i2c-mux-gpio.c
120
if (device_property_read_u32(dev, "idle-state", &mux->data.idle))
drivers/i2c/muxes/i2c-mux-gpio.c
121
mux->data.idle = I2C_MUX_GPIO_NO_IDLE;
drivers/i2c/muxes/i2c-mux-gpio.c
177
if (mux->data.idle != I2C_MUX_GPIO_NO_IDLE) {
drivers/i2c/muxes/i2c-mux-gpio.c
178
initial_state = mux->data.idle;
drivers/i2c/muxes/i2c-mux-gpio.c
51
i2c_mux_gpio_set(mux, mux->data.idle);
drivers/i2c/muxes/i2c-mux-reg.c
135
if (!of_property_read_u32(np, "idle-state", &mux->data.idle))
drivers/i2c/muxes/i2c-mux-reg.c
73
return i2c_mux_reg_set(mux, mux->data.idle);
drivers/iio/adc/ti_am335x_adc.c
705
unsigned int idle;
drivers/iio/adc/ti_am335x_adc.c
707
idle = tiadc_readl(adc_dev, REG_CTRL);
drivers/iio/adc/ti_am335x_adc.c
708
idle &= ~(CNTRLREG_SSENB);
drivers/iio/adc/ti_am335x_adc.c
709
tiadc_writel(adc_dev, REG_CTRL, idle | CNTRLREG_POWERDOWN);
drivers/input/touchscreen/atmel_mxt_ts.c
100
u8 idle;
drivers/input/touchscreen/atmel_mxt_ts.c
2314
struct t7_config deepsleep = { .active = 0, .idle = 0 };
drivers/input/touchscreen/atmel_mxt_ts.c
2327
new_config->active, new_config->idle);
drivers/input/touchscreen/atmel_mxt_ts.c
2344
if (data->t7_cfg.active == 0 || data->t7_cfg.idle == 0) {
drivers/input/touchscreen/atmel_mxt_ts.c
2353
data->t7_cfg.idle = 100;
drivers/input/touchscreen/atmel_mxt_ts.c
2359
data->t7_cfg.active, data->t7_cfg.idle);
drivers/input/touchscreen/ti_am335x_tsc.c
519
unsigned int idle;
drivers/input/touchscreen/ti_am335x_tsc.c
523
idle = titsc_readl(ts_dev, REG_IRQENABLE);
drivers/input/touchscreen/ti_am335x_tsc.c
525
(idle | IRQENB_HW_PEN));
drivers/input/touchscreen/zforce_ts.c
208
static int zforce_scan_frequency(struct zforce_ts *ts, u16 idle, u16 finger,
drivers/input/touchscreen/zforce_ts.c
213
(idle & 0xff), ((idle >> 8) & 0xff),
drivers/input/touchscreen/zforce_ts.c
219
idle, finger, stylus);
drivers/irqchip/irq-omap-intc.c
109
intc_writel(INTC_IDLE, intc_context.idle);
drivers/irqchip/irq-omap-intc.c
60
u32 idle;
drivers/irqchip/irq-omap-intc.c
90
intc_context.idle =
drivers/macintosh/adb-iop.c
241
if (adb_iop_state == idle)
drivers/macintosh/adb-iop.c
65
adb_iop_state = idle;
drivers/macintosh/adb-iop.c
72
if (adb_iop_state == idle)
drivers/macintosh/via-cuda.c
217
cuda_state = idle;
drivers/macintosh/via-cuda.c
258
cuda_state = idle;
drivers/macintosh/via-cuda.c
523
if (cuda_state == idle)
drivers/macintosh/via-cuda.c
594
case idle:
drivers/macintosh/via-cuda.c
619
cuda_state = idle;
drivers/macintosh/via-cuda.c
645
cuda_state = idle;
drivers/macintosh/via-cuda.c
711
cuda_state = idle;
drivers/macintosh/via-cuda.c
713
if (cuda_state == idle && TREQ_asserted(in_8(&via[B]))) {
drivers/macintosh/via-macii.c
154
macii_state = idle;
drivers/macintosh/via-macii.c
270
if (macii_state == idle)
drivers/macintosh/via-macii.c
291
if (current_req && macii_state == idle)
drivers/macintosh/via-macii.c
386
case idle:
drivers/macintosh/via-macii.c
475
macii_state = idle;
drivers/macintosh/via-macii.c
507
macii_state = idle;
drivers/macintosh/via-macii.c
543
if (macii_state == idle) {
drivers/macintosh/via-macii.c
550
if (macii_state == idle) {
drivers/macintosh/via-pmu.c
1172
if (pmu_state == idle)
drivers/macintosh/via-pmu.c
1237
if (!req || pmu_state != idle
drivers/macintosh/via-pmu.c
1274
} while (pmu_suspended && (adb_int_pending || pmu_state != idle
drivers/macintosh/via-pmu.c
1283
while((pmu_state != idle && pmu_state != locked) || !req->complete)
drivers/macintosh/via-pmu.c
1313
if (!adb_int_pending && pmu_state == idle && !req_awaiting_reply) {
drivers/macintosh/via-pmu.c
1503
pmu_state = idle;
drivers/macintosh/via-pmu.c
1544
pmu_state = idle;
drivers/macintosh/via-pmu.c
1559
pmu_state = idle;
drivers/macintosh/via-pmu.c
1629
if (pmu_state == idle) {
drivers/macintosh/via-pmu.c
1686
pmu_state = idle;
drivers/macintosh/via-pmu.c
2539
while (pmu_state != idle)
drivers/macintosh/via-pmu.c
357
pmu_state = idle;
drivers/macintosh/via-pmu.c
390
pmu_state = idle;
drivers/macintosh/via-pmu.c
492
} while (pmu_state != idle);
drivers/macintosh/via-pmu.c
593
while (interrupt_data[0][0] || pmu_state != idle) {
drivers/macintosh/via-pmu.c
598
if (pmu_state == idle)
drivers/md/dm-bufio.c
2291
cache_remove_range(&c->cache, block, block + n_blocks, idle, __free_buffer_wake);
drivers/md/dm-cache-policy-internal.h
36
bool idle, struct policy_work **result)
drivers/md/dm-cache-policy-internal.h
38
return p->get_background_work(p, idle, result);
drivers/md/dm-cache-policy-smq.c
1141
static bool clean_target_met(struct smq_policy *mq, bool idle)
drivers/md/dm-cache-policy-smq.c
1147
if (idle || mq->cleaner) {
drivers/md/dm-cache-policy-smq.c
1185
static void queue_writeback(struct smq_policy *mq, bool idle)
drivers/md/dm-cache-policy-smq.c
1191
e = q_peek(&mq->dirty, mq->dirty.nr_levels, idle);
drivers/md/dm-cache-policy-smq.c
1445
static int smq_get_background_work(struct dm_cache_policy *p, bool idle,
drivers/md/dm-cache-policy-smq.c
1455
if (!clean_target_met(mq, idle)) {
drivers/md/dm-cache-policy-smq.c
1456
queue_writeback(mq, idle);
drivers/md/dm-cache-policy.h
78
int (*get_background_work)(struct dm_cache_policy *p, bool idle,
drivers/md/dm-cache-target.c
1594
bool idle = dm_iot_idle_for(&cache->tracker, HZ);
drivers/md/dm-cache-target.c
1598
if (idle && current_volume <= cache->migration_threshold)
drivers/md/dm-vdo/funnel-workqueue.c
146
if ((atomic_read(&queue->idle) != 1) || (atomic_cmpxchg(&queue->idle, 1, 0) != 1))
drivers/md/dm-vdo/funnel-workqueue.c
190
atomic_set(&queue->idle, 1);
drivers/md/dm-vdo/funnel-workqueue.c
217
atomic_set(&queue->idle, 0);
drivers/md/dm-vdo/funnel-workqueue.c
491
thread_status = atomic_read(&queue->idle) ? "idle" : "running";
drivers/md/dm-vdo/funnel-workqueue.c
59
atomic_t idle;
drivers/md/dm-vdo/physical-zone.c
216
idle_pbn_lock *idle;
drivers/md/dm-vdo/physical-zone.c
221
idle = container_of(lock, idle_pbn_lock, lock);
drivers/md/dm-vdo/physical-zone.c
222
INIT_LIST_HEAD(&idle->entry);
drivers/md/dm-vdo/physical-zone.c
223
list_add_tail(&idle->entry, &pool->idle_list);
drivers/md/dm-vdo/physical-zone.c
295
idle_pbn_lock *idle;
drivers/md/dm-vdo/physical-zone.c
311
idle = list_entry(idle_entry, idle_pbn_lock, entry);
drivers/md/dm-vdo/physical-zone.c
312
idle->lock.holder_count = 0;
drivers/md/dm-vdo/physical-zone.c
313
set_pbn_lock_type(&idle->lock, type);
drivers/md/dm-vdo/physical-zone.c
315
*lock_ptr = &idle->lock;
drivers/md/dm-writecache.c
1958
unsigned long idle;
drivers/md/dm-writecache.c
1963
idle = dm_iot_idle_time(&wc->iot);
drivers/md/dm-writecache.c
1964
if (idle >= wc->pause)
drivers/md/dm-writecache.c
1966
idle = wc->pause - idle;
drivers/md/dm-writecache.c
1967
if (idle > HZ)
drivers/md/dm-writecache.c
1968
idle = HZ;
drivers/md/dm-writecache.c
1969
schedule_timeout_idle(idle);
drivers/md/dm-zoned-metadata.c
1936
unsigned int idx, bool idle)
drivers/md/dm-zoned-metadata.c
1946
if (idle && list_empty(zone_list))
drivers/md/dm-zoned-metadata.c
2013
unsigned int dev_idx, bool idle)
drivers/md/dm-zoned-metadata.c
2029
zone = dmz_get_rnd_zone_for_reclaim(zmd, dev_idx, idle);
drivers/md/dm-zoned.h
246
unsigned int dev_idx, bool idle);
drivers/md/md.c
9056
bool idle = true;
drivers/md/md.c
9065
idle = false;
drivers/md/md.c
9070
idle = false;
drivers/md/md.c
9073
return idle;
drivers/media/dvb-frontends/stv090x.c
3785
u32 reg, idle = 0, fifo_full = 1;
drivers/media/dvb-frontends/stv090x.c
3820
while ((!idle) && (i < 10)) {
drivers/media/dvb-frontends/stv090x.c
3822
idle = STV090x_GETFIELD_Px(reg, TX_IDLE_FIELD);
drivers/media/dvb-frontends/stv090x.c
3837
u32 reg, idle = 0, fifo_full = 1;
drivers/media/dvb-frontends/stv090x.c
3878
while ((!idle) && (i < 10)) {
drivers/media/dvb-frontends/stv090x.c
3880
idle = STV090x_GETFIELD_Px(reg, TX_IDLE_FIELD);
drivers/media/pci/bt8xx/bttv-driver.c
3511
struct bttv_buffer_set idle;
drivers/media/pci/bt8xx/bttv-driver.c
3518
memset(&idle, 0, sizeof(idle));
drivers/media/pci/bt8xx/bttv-driver.c
3522
btv->curr = idle;
drivers/media/pci/bt8xx/bttv-driver.c
3524
bttv_buffer_activate_video(btv, &idle);
drivers/media/pci/cx88/cx88-input.c
540
if (samples == 0xff && ir->dev->idle)
drivers/media/platform/mediatek/vpu/mtk_vpu.c
656
unsigned int running, pc, vpu_to_host, host_to_vpu, wdt, idle, ra, sp;
drivers/media/platform/mediatek/vpu/mtk_vpu.c
675
idle = vpu_cfg_readl(vpu, VPU_IDLE_REG);
drivers/media/platform/mediatek/vpu/mtk_vpu.c
690
host_to_vpu, vpu_to_host, sp, ra, idle);
drivers/media/platform/ti/omap3isp/isp.c
399
static void isp_core_init(struct isp_device *isp, int idle)
drivers/media/platform/ti/omap3isp/isp.c
402
((idle ? ISP_SYSCONFIG_MIDLEMODE_SMARTSTANDBY :
drivers/media/rc/ene_ir.c
942
static void ene_set_idle(struct rc_dev *rdev, bool idle)
drivers/media/rc/ene_ir.c
946
if (idle) {
drivers/media/rc/imon_raw.c
74
if (packet_no == 0x0a && !imon->rcdev->idle) {
drivers/media/rc/rc-ir-raw.c
173
if (dev->idle && !ev->pulse)
drivers/media/rc/rc-ir-raw.c
175
else if (dev->idle)
drivers/media/rc/rc-ir-raw.c
201
void ir_raw_event_set_idle(struct rc_dev *dev, bool idle)
drivers/media/rc/rc-ir-raw.c
206
dev_dbg(&dev->dev, "%s idle mode\n", idle ? "enter" : "leave");
drivers/media/rc/rc-ir-raw.c
208
if (idle) {
drivers/media/rc/rc-ir-raw.c
215
dev->s_idle(dev, idle);
drivers/media/rc/rc-ir-raw.c
217
dev->idle = idle;
drivers/media/rc/rc-ir-raw.c
624
dev->idle = true;
drivers/media/rc/rc-loopback.c
145
if (lodev->idle != enable) {
drivers/media/rc/rc-loopback.c
147
lodev->idle = enable;
drivers/media/rc/rc-loopback.c
248
loopdev.idle = true;
drivers/media/rc/rc-loopback.c
27
bool idle;
drivers/media/rc/winbond-cir.c
334
wbcir_idle_rx(struct rc_dev *dev, bool idle)
drivers/media/rc/winbond-cir.c
338
if (!idle && data->rxstate == WBCIR_RXSTATE_INACTIVE)
drivers/media/rc/winbond-cir.c
341
if (idle && data->rxstate != WBCIR_RXSTATE_INACTIVE) {
drivers/media/rc/winbond-cir.c
514
if (enable && data->dev->idle)
drivers/mfd/db8500-prcmu.c
1137
struct prcmu_auto_pm_config *idle)
drivers/mfd/db8500-prcmu.c
1143
BUG_ON((sleep == NULL) || (idle == NULL));
drivers/mfd/db8500-prcmu.c
1152
idle_cfg = (idle->sva_auto_pm_enable & 0xF);
drivers/mfd/db8500-prcmu.c
1153
idle_cfg = ((idle_cfg << 4) | (idle->sia_auto_pm_enable & 0xF));
drivers/mfd/db8500-prcmu.c
1154
idle_cfg = ((idle_cfg << 8) | (idle->sva_power_on & 0xFF));
drivers/mfd/db8500-prcmu.c
1155
idle_cfg = ((idle_cfg << 8) | (idle->sia_power_on & 0xFF));
drivers/mfd/db8500-prcmu.c
1156
idle_cfg = ((idle_cfg << 4) | (idle->sva_policy & 0xF));
drivers/mfd/db8500-prcmu.c
1157
idle_cfg = ((idle_cfg << 4) | (idle->sia_policy & 0xF));
drivers/mfd/db8500-prcmu.c
1172
(idle->sva_auto_pm_enable == PRCMU_AUTO_PM_ON) ||
drivers/mfd/db8500-prcmu.c
1173
(idle->sia_auto_pm_enable == PRCMU_AUTO_PM_ON));
drivers/misc/mei/init.c
346
bool idle = (dev->dev_state == MEI_DEV_ENABLED &&
drivers/misc/mei/init.c
352
idle,
drivers/misc/mei/init.c
358
return idle;
drivers/misc/sgi-xp/xpc_main.c
601
int idle = atomic_read(&ch->kthreads_idle);
drivers/misc/sgi-xp/xpc_main.c
607
if (idle > 0) {
drivers/misc/sgi-xp/xpc_main.c
608
wakeup = (needed > idle) ? idle : needed;
drivers/net/ethernet/airoha/airoha_ppe.c
1261
u32 idle;
drivers/net/ethernet/airoha/airoha_ppe.c
1268
idle = airoha_ppe_entry_idle_time(eth->ppe, e);
drivers/net/ethernet/airoha/airoha_ppe.c
1269
f->stats.lastused = jiffies - idle * HZ;
drivers/net/ethernet/airoha/airoha_ppe.c
914
int idle;
drivers/net/ethernet/airoha/airoha_ppe.c
924
idle = now - ts;
drivers/net/ethernet/airoha/airoha_ppe.c
926
return idle < 0 ? idle + ts_mask + 1 : idle;
drivers/net/ethernet/airoha/airoha_ppe.c
942
int idle;
drivers/net/ethernet/airoha/airoha_ppe.c
956
idle = airoha_ppe_get_entry_idle_time(ppe, ib1);
drivers/net/ethernet/airoha/airoha_ppe.c
957
if (idle >= min_idle)
drivers/net/ethernet/airoha/airoha_ppe.c
960
min_idle = idle;
drivers/net/ethernet/mediatek/mtk_ppe.c
561
int idle;
drivers/net/ethernet/mediatek/mtk_ppe.c
563
idle = __mtk_foe_entry_idle_time(ppe, entry->data.ib1);
drivers/net/ethernet/mediatek/mtk_ppe.c
578
if (cur_idle >= idle)
drivers/net/ethernet/mediatek/mtk_ppe.c
581
idle = cur_idle;
drivers/net/ethernet/mediatek/mtk_ppe_offload.c
545
u32 idle;
drivers/net/ethernet/mediatek/mtk_ppe_offload.c
552
idle = mtk_foe_entry_idle_time(eth->ppe[entry->ppe_index], entry);
drivers/net/ethernet/mediatek/mtk_ppe_offload.c
553
f->stats.lastused = jiffies - idle * HZ;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
2344
bool idle;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
2347
err = read_poll_timeout_atomic(fbnic_all_idle, idle, idle, 2, 500000,
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
2351
err = read_poll_timeout_atomic(fbnic_all_idle, idle, idle,
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
2362
err = read_poll_timeout_atomic(fbnic_all_idle, idle, idle, 2, 500000,
drivers/net/ethernet/natsemi/ns83820.c
1415
dev->rx_info.idle = 1;
drivers/net/ethernet/natsemi/ns83820.c
399
unsigned long idle;
drivers/net/ethernet/natsemi/ns83820.c
461
if (test_and_clear_bit(0, &dev->rx_info.idle)) {
drivers/net/ethernet/natsemi/ns83820.c
715
dev->rx_info.idle = 1;
drivers/net/ethernet/natsemi/ns83820.c
811
if (dev->rx_info.idle)
drivers/net/ethernet/realtek/rtase/rtase_main.c
1681
u32 idle = tp->tx_qos[queue].idleslope * RTASE_1T_CLOCK;
drivers/net/ethernet/realtek/rtase/rtase_main.c
1684
val = u32_encode_bits(idle / RTASE_1T_POWER, RTASE_IDLESLOPE_INT_MASK);
drivers/net/ethernet/realtek/rtase/rtase_main.c
1685
idle %= RTASE_1T_POWER;
drivers/net/ethernet/realtek/rtase/rtase_main.c
1688
idle *= 2;
drivers/net/ethernet/realtek/rtase/rtase_main.c
1689
if ((idle / RTASE_1T_POWER) == 1)
drivers/net/ethernet/realtek/rtase/rtase_main.c
1692
idle %= RTASE_1T_POWER;
drivers/net/wireless/ath/wil6210/debugfs.c
155
uint64_t idle = txdata->idle * 100;
drivers/net/wireless/ath/wil6210/debugfs.c
159
do_div(idle, total);
drivers/net/wireless/ath/wil6210/debugfs.c
161
(int)idle);
drivers/net/wireless/ath/wil6210/debugfs.c
166
txdata->idle = 0ULL;
drivers/net/wireless/ath/wil6210/txrx.c
1123
txdata->idle = 0;
drivers/net/wireless/ath/wil6210/txrx.c
1979
txdata->idle += get_cycles() - txdata->last_idle;
drivers/net/wireless/ath/wil6210/txrx.c
2132
txdata->idle += get_cycles() - txdata->last_idle;
drivers/net/wireless/ath/wil6210/txrx_edma.c
1502
txdata->idle += get_cycles() - txdata->last_idle;
drivers/net/wireless/ath/wil6210/wil6210.h
635
cycles_t idle, last_idle, begin;
drivers/net/wireless/broadcom/b43/main.c
4743
static void b43_set_synth_pu_delay(struct b43_wldev *dev, bool idle)
drivers/net/wireless/broadcom/b43/main.c
4749
if (b43_is_mode(dev->wl, NL80211_IFTYPE_ADHOC) || idle)
drivers/net/wireless/broadcom/b43/phy_ht.c
626
u8 *idle = phy_ht->idle_tssi;
drivers/net/wireless/broadcom/b43/phy_ht.c
687
idle[0] << B43_PHY_HT_TXPCTL_IDLE_TSSI_C1_SHIFT);
drivers/net/wireless/broadcom/b43/phy_ht.c
690
idle[1] << B43_PHY_HT_TXPCTL_IDLE_TSSI_C2_SHIFT);
drivers/net/wireless/broadcom/b43/phy_ht.c
693
idle[2] << B43_PHY_HT_TXPCTL_IDLE_TSSI2_C3_SHIFT);
drivers/net/wireless/broadcom/b43/phy_n.c
4047
u8 idle[2];
drivers/net/wireless/broadcom/b43/phy_n.c
4079
idle[0] = nphy->pwr_ctl_info[0].idle_tssi_2g;
drivers/net/wireless/broadcom/b43/phy_n.c
4080
idle[1] = nphy->pwr_ctl_info[1].idle_tssi_2g;
drivers/net/wireless/broadcom/b43/phy_n.c
4088
idle[c] = nphy->pwr_ctl_info[c].idle_tssi_2g;
drivers/net/wireless/broadcom/b43/phy_n.c
4096
idle[c] = nphy->pwr_ctl_info[c].idle_tssi_5g;
drivers/net/wireless/broadcom/b43/phy_n.c
4104
idle[c] = nphy->pwr_ctl_info[c].idle_tssi_5g;
drivers/net/wireless/broadcom/b43/phy_n.c
4112
idle[c] = nphy->pwr_ctl_info[c].idle_tssi_5g;
drivers/net/wireless/broadcom/b43/phy_n.c
4119
idle[0] = nphy->pwr_ctl_info[0].idle_tssi_5g;
drivers/net/wireless/broadcom/b43/phy_n.c
4120
idle[1] = nphy->pwr_ctl_info[1].idle_tssi_5g;
drivers/net/wireless/broadcom/b43/phy_n.c
4187
idle[0] << B43_NPHY_TXPCTL_ITSSI_0_SHIFT |
drivers/net/wireless/broadcom/b43/phy_n.c
4188
idle[1] << B43_NPHY_TXPCTL_ITSSI_1_SHIFT |
drivers/net/wireless/broadcom/b43/phy_n.c
4199
if (dev->phy.rev < 3 && (i <= (31 - idle[c] + 1)))
drivers/net/wireless/broadcom/b43legacy/main.c
3145
bool idle) {
drivers/net/wireless/broadcom/b43legacy/main.c
3148
if (b43legacy_is_mode(dev->wl, NL80211_IFTYPE_ADHOC) || idle)
drivers/net/wireless/broadcom/brcm80211/brcmfmac/btcoex.c
316
goto idle;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/btcoex.c
336
goto idle;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/btcoex.c
340
goto idle;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/btcoex.c
345
idle:
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
113
u32 idle;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
3176
sinfo->inactive_time = le32_to_cpu(sta_info_le.idle) * 1000;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
8076
survey->idle = val;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
8217
info->time_busy = ACS_MSRMNT_DELAY - survey.idle;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwil_types.h
619
__le32 idle; /* time since data pkt rx'd from sta */
drivers/net/wireless/intel/iwlwifi/dvm/rxon.c
1395
if (changes & BSS_CHANGED_IDLE && vif->cfg.idle) {
drivers/net/wireless/intel/iwlwifi/dvm/rxon.c
565
(!ctx_bss->vif->cfg.idle &&
drivers/net/wireless/intel/iwlwifi/dvm/rxon.c
569
} else if (!ctx_pan->vif->cfg.idle &&
drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c
3166
if (changes & BSS_CHANGED_IDLE && !vif->cfg.idle)
drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c
4180
if (!vif->cfg.idle)
drivers/net/wireless/intel/iwlwifi/mvm/mld-mac80211.c
784
if (changes & BSS_CHANGED_IDLE && !vif->cfg.idle)
drivers/net/wireless/realtek/rtw89/efuse.c
140
int rtw89_cnv_efuse_state_ax(struct rtw89_dev *rtwdev, bool idle)
drivers/net/wireless/realtek/rtw89/efuse.h
26
int rtw89_cnv_efuse_state_ax(struct rtw89_dev *rtwdev, bool idle);
drivers/net/wireless/realtek/rtw89/efuse.h
29
int rtw89_cnv_efuse_state_be(struct rtw89_dev *rtwdev, bool idle);
drivers/net/wireless/realtek/rtw89/efuse_be.c
143
int rtw89_cnv_efuse_state_be(struct rtw89_dev *rtwdev, bool idle)
drivers/net/wireless/realtek/rtw89/efuse_be.c
148
if (idle) {
drivers/net/wireless/realtek/rtw89/mac.h
1094
int (*cnv_efuse_state)(struct rtw89_dev *rtwdev, bool idle);
drivers/net/wireless/ti/wlcore/main.c
3127
bool idle)
drivers/net/wireless/ti/wlcore/main.c
3131
if (idle == cur_idle)
drivers/net/wireless/ti/wlcore/main.c
3134
if (idle) {
drivers/net/wireless/ti/wlcore/main.c
4452
wl1271_sta_handle_idle(wl, wlvif, vif->cfg.idle);
drivers/net/wireless/virtual/mac80211_hwsim.c
1853
if (!data2->started || (data2->idle && !data2->tmp_chan) ||
drivers/net/wireless/virtual/mac80211_hwsim.c
2068
if (data->idle && !data->tmp_chan) {
drivers/net/wireless/virtual/mac80211_hwsim.c
2458
data->idle = !!(conf->flags & IEEE80211_CONF_IDLE);
drivers/net/wireless/virtual/mac80211_hwsim.c
6171
if ((data2->idle && !data2->tmp_chan) || !data2->started)
drivers/net/wireless/virtual/mac80211_hwsim.c
711
bool started, idle, scanning;
drivers/perf/xgene_pmu.c
548
XGENE_PMU_EVENT_ATTR(rank-idle, 0x1f),
drivers/phy/motorola/phy-mapphone-mdm6600.c
607
goto idle;
drivers/phy/motorola/phy-mapphone-mdm6600.c
618
idle:
drivers/phy/tegra/xusb-tegra124.c
386
unsigned int index, bool idle)
drivers/phy/tegra/xusb-tegra124.c
392
if (idle)
drivers/phy/tegra/xusb-tegra210.c
1034
unsigned int index, bool idle)
drivers/phy/tegra/xusb-tegra210.c
1044
if (idle)
drivers/phy/tegra/xusb.c
1385
unsigned int port, bool idle)
drivers/phy/tegra/xusb.c
1388
return padctl->soc->ops->hsic_set_idle(padctl, port, idle);
drivers/phy/tegra/xusb.h
408
unsigned int index, bool idle);
drivers/pmdomain/rockchip/pm-domains.c
117
#define DOMAIN(_name, pwr, status, req, idle, ack, wakeup) \
drivers/pmdomain/rockchip/pm-domains.c
123
.idle_mask = (idle), \
drivers/pmdomain/rockchip/pm-domains.c
128
#define DOMAIN_M(_name, pwr, status, req, idle, ack, wakeup) \
drivers/pmdomain/rockchip/pm-domains.c
136
.idle_mask = (idle), \
drivers/pmdomain/rockchip/pm-domains.c
141
#define DOMAIN_M_G(_name, pwr, status, req, idle, ack, g_mask, wakeup, keepon) \
drivers/pmdomain/rockchip/pm-domains.c
149
.idle_mask = (idle), \
drivers/pmdomain/rockchip/pm-domains.c
155
#define DOMAIN_M_G_SD(_name, pwr, status, req, idle, ack, g_mask, mem, wakeup, keepon) \
drivers/pmdomain/rockchip/pm-domains.c
163
.idle_mask = (idle), \
drivers/pmdomain/rockchip/pm-domains.c
169
#define DOMAIN_M_O_R(_name, p_offset, pwr, status, m_offset, m_status, r_status, r_offset, req, idle, ack, wakeup, regulator) \
drivers/pmdomain/rockchip/pm-domains.c
182
.idle_mask = (idle), \
drivers/pmdomain/rockchip/pm-domains.c
188
#define DOMAIN_M_O_R_G(_name, p_offset, pwr, status, m_offset, m_status, r_status, r_offset, req, idle, ack, g_mask, wakeup) \
drivers/pmdomain/rockchip/pm-domains.c
201
.idle_mask = (idle), \
drivers/pmdomain/rockchip/pm-domains.c
207
#define DOMAIN_RK3036(_name, req, ack, idle, wakeup) \
drivers/pmdomain/rockchip/pm-domains.c
213
.idle_mask = (idle), \
drivers/pmdomain/rockchip/pm-domains.c
220
#define DOMAIN_RV1126(name, pwr, req, idle, wakeup) \
drivers/pmdomain/rockchip/pm-domains.c
221
DOMAIN_M(name, pwr, pwr, req, idle, idle, wakeup)
drivers/pmdomain/rockchip/pm-domains.c
247
#define DOMAIN_RK3576(name, p_offset, pwr, status, r_status, r_offset, req, idle, g_mask, wakeup) \
drivers/pmdomain/rockchip/pm-domains.c
248
DOMAIN_M_O_R_G(name, p_offset, pwr, status, 0, r_status, r_status, r_offset, req, idle, idle, g_mask, wakeup)
drivers/pmdomain/rockchip/pm-domains.c
353
#define DOMAIN_RK3588(name, p_offset, pwr, status, m_offset, m_status, r_status, r_offset, req, idle, wakeup, regulator) \
drivers/pmdomain/rockchip/pm-domains.c
354
DOMAIN_M_O_R(name, p_offset, pwr, status, m_offset, m_status, r_status, r_offset, req, idle, idle, wakeup, regulator)
drivers/pmdomain/rockchip/pm-domains.c
395
bool idle)
drivers/pmdomain/rockchip/pm-domains.c
410
idle ? (pd_info->req_mask | pd_info->req_w_mask) :
drivers/pmdomain/rockchip/pm-domains.c
414
pd_info->req_mask, idle ? -1U : 0);
drivers/pmdomain/rockchip/pm-domains.c
419
target_ack = idle ? pd_info->ack_mask : 0;
drivers/pmdomain/rockchip/pm-domains.c
431
is_idle, is_idle == idle, 0, 10000);
drivers/scsi/aha152x.c
2840
for (i = idle; i<maxstate; i++) {
drivers/scsi/ibmvscsi_tgt/ibmvscsi_tgt.c
976
static long ibmvscsis_ready_for_suspend(struct scsi_info *vscsi, bool idle)
drivers/scsi/ibmvscsi_tgt/ibmvscsi_tgt.c
999
if (idle) {
drivers/scsi/mac53c94.c
150
if (state->phase != idle || state->current_req != NULL)
drivers/scsi/mac53c94.c
351
state->phase = idle;
drivers/scsi/mac53c94.c
95
if (state->phase == idle)
drivers/scsi/mesh.c
1003
ms->phase = idle;
drivers/scsi/mesh.c
1538
case idle:
drivers/scsi/mesh.c
1611
ms->phase = idle;
drivers/scsi/mesh.c
1642
if (ms->phase == idle)
drivers/scsi/mesh.c
1775
while(ms->phase != idle) {
drivers/scsi/mesh.c
391
ms->phase = idle;
drivers/scsi/mesh.c
467
ms->phase = idle;
drivers/scsi/mesh.c
553
if (ms->phase != idle || ms->current_req != NULL) {
drivers/scsi/mesh.c
559
while (ms->phase == idle) {
drivers/scsi/mesh.c
615
ms->phase = idle;
drivers/scsi/mesh.c
858
case idle:
drivers/scsi/ncr53c8xx.c
1826
ncrcmd idle [ 2];
drivers/scsi/ncr53c8xx.c
3413
*p++ =PADDR (idle);
drivers/scsi/ncr53c8xx.c
4419
np->scripth->tryloop [qidx] = cpu_to_scr(NCB_SCRIPT_PHYS (np, idle));
drivers/scsi/ncr53c8xx.c
5084
cpu_to_scr(NCB_SCRIPT_PHYS (np, idle));
drivers/scsi/ncr53c8xx.c
7171
cp->start.schedule.l_paddr = cpu_to_scr(NCB_SCRIPT_PHYS (np, idle));
drivers/scsi/ncr53c8xx.c
8215
np->script0->idle[0] =
drivers/scsi/pcmcia/sym53c500_cs.c
501
scp->phase = idle;
drivers/scsi/sym53c8xx_2/sym_fw.c
107
scripta0->idle[0] = cpu_to_scr(SCR_NO_OP);
drivers/scsi/sym53c8xx_2/sym_fw.c
151
scripta0->idle[0] = cpu_to_scr(SCR_NO_OP);
drivers/scsi/sym53c8xx_2/sym_fw.h
40
SYM_GEN_A(s, idle) SYM_GEN_A(s, ungetjob) \
drivers/scsi/sym53c8xx_2/sym_fw1.h
93
u32 idle [ 4];
drivers/scsi/sym53c8xx_2/sym_fw1.h
95
u32 idle [ 2];
drivers/scsi/sym53c8xx_2/sym_fw2.h
89
u32 idle [ 4];
drivers/scsi/sym53c8xx_2/sym_fw2.h
91
u32 idle [ 2];
drivers/scsi/sym53c8xx_2/sym_hipd.c
4902
cp->phys.head.go.start = cpu_to_scr(SCRIPTA_BA(np, idle));
drivers/scsi/sym53c8xx_2/sym_hipd.c
5738
np->idletask.start = cpu_to_scr(SCRIPTA_BA(np, idle));
drivers/scsi/sym53c8xx_2/sym_hipd.c
5742
np->notask.start = cpu_to_scr(SCRIPTA_BA(np, idle));
drivers/scsi/sym53c8xx_2/sym_hipd.c
5746
np->bad_itl.start = cpu_to_scr(SCRIPTA_BA(np, idle));
drivers/scsi/sym53c8xx_2/sym_hipd.c
5750
np->bad_itlq.start = cpu_to_scr(SCRIPTA_BA(np, idle));
drivers/soc/hisilicon/kunpeng_hccs.c
1249
struct hccs_port_info *port, u8 *idle)
drivers/soc/hisilicon/kunpeng_hccs.c
1269
*idle = *((u8 *)desc.rsp.data);
drivers/soc/hisilicon/kunpeng_hccs.c
1281
u8 idle;
drivers/soc/hisilicon/kunpeng_hccs.c
1293
&idle);
drivers/soc/hisilicon/kunpeng_hccs.c
1299
} else if (idle == 0) {
drivers/soc/renesas/r9a06g032-smp.c
36
struct task_struct *idle)
drivers/staging/media/atomisp/pci/sh_css_hrt.c
30
bool not_idle = false, idle;
drivers/staging/media/atomisp/pci/sh_css_hrt.c
33
idle = sp_ctrl_getbit(SP0_ID, SP_SC_REG, SP_IDLE_BIT);
drivers/staging/media/atomisp/pci/sh_css_hrt.c
34
not_idle |= !idle;
drivers/staging/media/atomisp/pci/sh_css_hrt.c
35
if (!idle)
drivers/staging/media/atomisp/pci/sh_css_hrt.c
38
idle = isp_ctrl_getbit(ISP0_ID, ISP_SC_REG, ISP_IDLE_BIT);
drivers/staging/media/atomisp/pci/sh_css_hrt.c
39
not_idle |= !idle;
drivers/staging/media/atomisp/pci/sh_css_hrt.c
40
if (!idle)
drivers/thermal/intel/intel_powerclamp.c
146
static bool check_invalid(cpumask_var_t mask, u8 idle)
drivers/thermal/intel/intel_powerclamp.c
148
if (cpumask_equal(cpu_present_mask, mask) && idle > MAX_ALL_CPU_IDLE)
drivers/thermal/renesas/rcar_thermal.c
403
int idle = IDLE_INTERVAL;
drivers/thermal/renesas/rcar_thermal.c
444
idle = 0; /* polling delay is not needed */
drivers/thermal/renesas/rcar_thermal.c
494
idle);
drivers/usb/gadget/function/f_hid.c
1198
hidg->idle = 1;
drivers/usb/gadget/function/f_hid.c
61
unsigned char idle;
drivers/usb/gadget/function/f_hid.c
886
((u8 *) req->buf)[0] = hidg->idle;
drivers/usb/gadget/function/f_hid.c
921
hidg->idle = value >> 8;
drivers/usb/host/xhci-tegra.c
623
bool idle, enable;
drivers/usb/host/xhci-tegra.c
680
idle = false;
drivers/usb/host/xhci-tegra.c
682
idle = true;
drivers/usb/host/xhci-tegra.c
689
idle);
drivers/usb/host/xhci-tegra.c
696
idle ? "idle" : "busy", err);
drivers/usb/musb/musb_host.c
1998
int idle = 0;
drivers/usb/musb/musb_host.c
2096
idle = 1;
drivers/usb/musb/musb_host.c
2102
idle = list_empty(head);
drivers/usb/musb/musb_host.c
2108
if (idle)
fs/proc/stat.c
108
idle += get_idle_time(&kcpustat, i);
fs/proc/stat.c
130
seq_put_decimal_ull(p, " ", nsec_to_clock_t(idle));
fs/proc/stat.c
149
idle = get_idle_time(&kcpustat, i);
fs/proc/stat.c
160
seq_put_decimal_ull(p, " ", nsec_to_clock_t(idle));
fs/proc/stat.c
27
u64 idle, idle_usecs = -1ULL;
fs/proc/stat.c
34
idle = kcs->cpustat[CPUTIME_IDLE];
fs/proc/stat.c
36
idle = idle_usecs * NSEC_PER_USEC;
fs/proc/stat.c
38
return idle;
fs/proc/stat.c
85
u64 user, nice, system, idle, iowait, irq, softirq, steal;
fs/proc/stat.c
92
user = nice = system = idle = iowait =
fs/proc/uptime.c
15
struct timespec64 idle;
fs/proc/uptime.c
31
idle.tv_sec = div_u64_rem(idle_nsec, NSEC_PER_SEC, &rem);
fs/proc/uptime.c
32
idle.tv_nsec = rem;
fs/proc/uptime.c
36
(unsigned long) idle.tv_sec,
fs/proc/uptime.c
37
(idle.tv_nsec / (NSEC_PER_SEC / 100)));
fs/smb/client/smbdirect.c
1196
if (sc->idle.keepalive == SMBDIRECT_KEEPALIVE_PENDING) {
fs/smb/client/smbdirect.c
1197
sc->idle.keepalive = SMBDIRECT_KEEPALIVE_SENT;
fs/smb/client/smbdirect.c
1202
mod_delayed_work(sc->workqueue, &sc->idle.timer_work,
fs/smb/client/smbdirect.c
1841
container_of(work, struct smbdirect_socket, idle.immediate_work);
fs/smb/client/smbdirect.c
1854
container_of(work, struct smbdirect_socket, idle.timer_work.work);
fs/smb/client/smbdirect.c
1857
if (sc->idle.keepalive != SMBDIRECT_KEEPALIVE_NONE) {
fs/smb/client/smbdirect.c
1860
sc->idle.keepalive);
fs/smb/client/smbdirect.c
1872
sc->idle.keepalive = SMBDIRECT_KEEPALIVE_PENDING;
fs/smb/client/smbdirect.c
1873
mod_delayed_work(sc->workqueue, &sc->idle.timer_work,
fs/smb/client/smbdirect.c
1876
queue_work(sc->workqueue, &sc->idle.immediate_work);
fs/smb/client/smbdirect.c
1927
disable_delayed_work_sync(&sc->idle.timer_work);
fs/smb/client/smbdirect.c
1929
disable_work_sync(&sc->idle.immediate_work);
fs/smb/client/smbdirect.c
220
disable_work(&sc->idle.immediate_work);
fs/smb/client/smbdirect.c
221
disable_delayed_work(&sc->idle.timer_work);
fs/smb/client/smbdirect.c
2275
INIT_WORK(&sc->idle.immediate_work, send_immediate_empty_message);
fs/smb/client/smbdirect.c
2276
INIT_DELAYED_WORK(&sc->idle.timer_work, idle_connection_timer);
fs/smb/client/smbdirect.c
2281
sc->idle.keepalive = SMBDIRECT_KEEPALIVE_PENDING;
fs/smb/client/smbdirect.c
2282
mod_delayed_work(sc->workqueue, &sc->idle.timer_work,
fs/smb/client/smbdirect.c
2308
disable_delayed_work_sync(&sc->idle.timer_work);
fs/smb/client/smbdirect.c
275
disable_work(&sc->idle.immediate_work);
fs/smb/client/smbdirect.c
276
disable_delayed_work(&sc->idle.timer_work);
fs/smb/client/smbdirect.c
759
queue_work(sc->workqueue, &sc->idle.immediate_work);
fs/smb/client/smbdirect.c
801
sc->idle.keepalive = SMBDIRECT_KEEPALIVE_NONE;
fs/smb/client/smbdirect.c
802
mod_delayed_work(sc->workqueue, &sc->idle.timer_work,
fs/smb/client/smbdirect.c
885
queue_work(sc->workqueue, &sc->idle.immediate_work);
fs/smb/common/smbdirect/smbdirect_socket.h
150
} idle;
fs/smb/common/smbdirect/smbdirect_socket.h
380
INIT_WORK(&sc->idle.immediate_work, __smbdirect_socket_disabled_work);
fs/smb/common/smbdirect/smbdirect_socket.h
381
disable_work_sync(&sc->idle.immediate_work);
fs/smb/common/smbdirect/smbdirect_socket.h
382
INIT_DELAYED_WORK(&sc->idle.timer_work, __smbdirect_socket_disabled_work);
fs/smb/common/smbdirect/smbdirect_socket.h
383
disable_delayed_work_sync(&sc->idle.timer_work);
fs/smb/server/transport_rdma.c
1059
queue_work(sc->workqueue, &sc->idle.immediate_work);
fs/smb/server/transport_rdma.c
1169
if (sc->idle.keepalive == SMBDIRECT_KEEPALIVE_PENDING) {
fs/smb/server/transport_rdma.c
1170
sc->idle.keepalive = SMBDIRECT_KEEPALIVE_SENT;
fs/smb/server/transport_rdma.c
1175
mod_delayed_work(sc->workqueue, &sc->idle.timer_work,
fs/smb/server/transport_rdma.c
2145
sc->idle.keepalive = SMBDIRECT_KEEPALIVE_PENDING;
fs/smb/server/transport_rdma.c
2146
mod_delayed_work(sc->workqueue, &sc->idle.timer_work,
fs/smb/server/transport_rdma.c
2605
INIT_WORK(&sc->idle.immediate_work, smb_direct_send_immediate_work);
fs/smb/server/transport_rdma.c
261
disable_delayed_work(&sc->idle.timer_work);
fs/smb/server/transport_rdma.c
262
disable_work(&sc->idle.immediate_work);
fs/smb/server/transport_rdma.c
317
disable_work(&sc->idle.immediate_work);
fs/smb/server/transport_rdma.c
318
disable_delayed_work(&sc->idle.timer_work);
fs/smb/server/transport_rdma.c
375
container_of(work, struct smbdirect_socket, idle.immediate_work);
fs/smb/server/transport_rdma.c
386
container_of(work, struct smbdirect_socket, idle.timer_work.work);
fs/smb/server/transport_rdma.c
389
if (sc->idle.keepalive != SMBDIRECT_KEEPALIVE_NONE) {
fs/smb/server/transport_rdma.c
401
sc->idle.keepalive = SMBDIRECT_KEEPALIVE_PENDING;
fs/smb/server/transport_rdma.c
402
mod_delayed_work(sc->workqueue, &sc->idle.timer_work,
fs/smb/server/transport_rdma.c
404
queue_work(sc->workqueue, &sc->idle.immediate_work);
fs/smb/server/transport_rdma.c
442
INIT_DELAYED_WORK(&sc->idle.timer_work, smb_direct_idle_connection_timer);
fs/smb/server/transport_rdma.c
488
disable_delayed_work_sync(&sc->idle.timer_work);
fs/smb/server/transport_rdma.c
489
disable_work_sync(&sc->idle.immediate_work);
fs/smb/server/transport_rdma.c
649
sc->idle.keepalive = SMBDIRECT_KEEPALIVE_NONE;
fs/smb/server/transport_rdma.c
650
mod_delayed_work(sc->workqueue, &sc->idle.timer_work,
fs/smb/server/transport_rdma.c
714
queue_work(sc->workqueue, &sc->idle.immediate_work);
fs/smb/server/transport_rdma.c
842
sc->idle.keepalive = SMBDIRECT_KEEPALIVE_NONE;
fs/smb/server/transport_rdma.c
843
mod_delayed_work(sc->workqueue, &sc->idle.timer_work,
include/linux/hid.h
1219
static inline int hid_hw_idle(struct hid_device *hdev, int report, int idle,
include/linux/hid.h
1222
if (hdev->ll_driver->idle)
include/linux/hid.h
1223
return hdev->ll_driver->idle(hdev, report, idle, reqtype);
include/linux/hid.h
942
int (*idle)(struct hid_device *hdev, int report, int idle, int reqtype);
include/linux/irqflags.h
31
struct task_struct *idle);
include/linux/irqflags.h
39
struct task_struct *idle) {}
include/linux/mfd/db8500-prcmu.h
500
struct prcmu_auto_pm_config *idle);
include/linux/mfd/db8500-prcmu.h
599
struct prcmu_auto_pm_config *idle)
include/linux/mfd/wm8994/pdata.h
120
bool idle;
include/linux/page-flags.h
673
FOLIO_FLAG(idle, FOLIO_HEAD_PAGE)
include/linux/page-flags.h
679
FOLIO_FLAG_FALSE(idle)
include/linux/phy/tegra/xusb.h
19
unsigned int port, bool idle);
include/linux/platform_data/i2c-mux-gpio.h
29
unsigned idle;
include/linux/platform_data/i2c-mux-reg.h
32
u32 idle;
include/linux/platform_data/pm33xx.h
49
int (*init)(int (*idle)(u32 wfi_flags));
include/linux/sched/ext.h
253
bool idle;
include/linux/sched/task.h
41
int idle;
include/linux/sched/task.h
64
extern void init_idle(struct task_struct *idle, int cpu);
include/media/rc-core.h
161
bool idle;
include/media/rc-core.h
317
void ir_raw_event_set_idle(struct rc_dev *dev, bool idle);
include/media/rc-core.h
325
dev->idle = true;
include/net/ax25.h
243
unsigned long t1, t2, t3, idle, rtt;
include/net/mac80211.h
1989
bool idle;
include/net/netrom.h
76
unsigned long t1, t2, t4, idle;
include/net/rose.h
142
unsigned long t1, t2, t3, hb, idle;
include/trace/events/mmflags.h
165
IF_HAVE_PG_IDLE(idle) \
include/uapi/drm/radeon_drm.h
595
int idle;
include/uapi/linux/ax25.h
104
unsigned int idle, idletimer;
include/uapi/linux/ax25.h
94
unsigned int idle, idletimer;
kernel/cpu.c
1619
struct task_struct *idle;
kernel/cpu.c
1638
idle = idle_thread_get(cpu);
kernel/cpu.c
1639
if (IS_ERR(idle)) {
kernel/cpu.c
1640
ret = PTR_ERR(idle);
kernel/cpu.c
1647
scs_task_reset(idle);
kernel/cpu.c
1648
kasan_unpoison_task_stack(idle);
kernel/cpu.c
866
struct task_struct *idle = idle_thread_get(cpu);
kernel/cpu.c
883
ret = __cpu_up(cpu, idle);
kernel/cpu.c
910
struct task_struct *idle = idle_thread_get(cpu);
kernel/cpu.c
911
struct mm_struct *mm = idle->active_mm;
kernel/cpu.c
918
idle->active_mm = NULL;
kernel/debug/kdb/kdb_main.c
2165
int idle = 0, daemon = 0;
kernel/debug/kdb/kdb_main.c
2171
++idle;
kernel/debug/kdb/kdb_main.c
2177
if (idle || daemon) {
kernel/debug/kdb/kdb_main.c
2178
if (idle)
kernel/debug/kdb/kdb_main.c
2180
idle, idle == 1 ? "" : "es",
kernel/fork.c
2547
static inline void init_idle_pids(struct task_struct *idle)
kernel/fork.c
2552
INIT_HLIST_NODE(&idle->pid_links[type]); /* not really needed */
kernel/fork.c
2553
init_task_pid(idle, type, &init_struct_pid);
kernel/fork.c
2571
.idle = 1,
kernel/locking/lockdep.c
4603
void lockdep_cleanup_dead_cpu(unsigned int cpu, struct task_struct *idle)
kernel/locking/lockdep.c
4610
if (idle)
kernel/locking/lockdep.c
4611
print_irqtrace_events(idle);
kernel/sched/core.c
1276
if (set_nr_and_not_polling(task_thread_info(rq->idle), TIF_NEED_RESCHED))
kernel/sched/core.c
333
rq->core->core_forceidle_count && rq->curr == rq->idle)
kernel/sched/core.c
3779
if (set_nr_if_polling(cpu_rq(cpu)->idle)) {
kernel/sched/core.c
5969
return (task_rq(t)->idle == t);
kernel/sched/core.c
6166
if (p == rq_i->idle) {
kernel/sched/core.c
6243
if (rq->core->core_forceidle_count && next == rq->idle)
kernel/sched/core.c
6263
if (dst->curr != dst->idle)
kernel/sched/core.c
6277
if (p->core_occupation > dst->idle->core_occupation)
kernel/sched/core.c
6540
put_prev_set_next_task(rq, rq->donor, rq->idle);
kernel/sched/core.c
6541
rq_set_donor(rq, rq->idle);
kernel/sched/core.c
6542
set_tsk_need_resched(rq->idle);
kernel/sched/core.c
6543
return rq->idle;
kernel/sched/core.c
6859
if (next == rq->idle)
kernel/sched/core.c
7358
if (unlikely(p == rq->idle)) {
kernel/sched/core.c
7932
void __init init_idle(struct task_struct *idle, int cpu)
kernel/sched/core.c
7941
raw_spin_lock_irqsave(&idle->pi_lock, flags);
kernel/sched/core.c
7944
idle->__state = TASK_RUNNING;
kernel/sched/core.c
7945
idle->se.exec_start = sched_clock();
kernel/sched/core.c
7950
idle->flags |= PF_KTHREAD | PF_NO_SETAFFINITY;
kernel/sched/core.c
7951
kthread_set_per_cpu(idle, cpu);
kernel/sched/core.c
7957
set_cpus_allowed_common(idle, &ac);
kernel/sched/core.c
7969
__set_task_cpu(idle, cpu);
kernel/sched/core.c
7972
rq->idle = idle;
kernel/sched/core.c
7973
rq_set_donor(rq, idle);
kernel/sched/core.c
7974
rcu_assign_pointer(rq->curr, idle);
kernel/sched/core.c
7975
idle->on_rq = TASK_ON_RQ_QUEUED;
kernel/sched/core.c
7976
idle->on_cpu = 1;
kernel/sched/core.c
7978
raw_spin_unlock_irqrestore(&idle->pi_lock, flags);
kernel/sched/core.c
7981
init_idle_preempt_count(idle, cpu);
kernel/sched/core.c
7986
idle->sched_class = &idle_sched_class;
kernel/sched/core.c
7987
ftrace_graph_init_idle_task(idle, cpu);
kernel/sched/core.c
7988
vtime_init_idle(idle, cpu);
kernel/sched/core.c
7989
sprintf(idle->comm, "%s/%d", INIT_TASK_COMM, cpu);
kernel/sched/core.c
9907
return css_tg(css)->idle;
kernel/sched/core.c
9911
struct cftype *cft, s64 idle)
kernel/sched/core.c
9915
ret = sched_group_set_idle(css_tg(css), idle);
kernel/sched/core.c
9917
scx_group_set_idle(css_tg(css), idle);
kernel/sched/core_sched.c
280
if (p == rq_i->idle)
kernel/sched/cputime.c
408
} else if (p == this_rq()->idle) {
kernel/sched/cputime.c
501
else if ((p != this_rq()->idle) || (irq_count() != HARDIRQ_OFFSET))
kernel/sched/deadline.c
1427
bool idle = idle_rq(rq);
kernel/sched/deadline.c
1448
if (dl_se->dl_defer_idle && !idle)
kernel/sched/deadline.c
1470
if (dl_se->dl_defer_idle && idle) {
kernel/sched/deadline.c
1495
if (idle)
kernel/sched/ext.c
3300
tg->scx.idle = false;
kernel/sched/ext.c
3447
void scx_group_set_idle(struct task_group *tg, bool idle)
kernel/sched/ext.c
3455
tg_cgrp(tg), idle);
kernel/sched/ext.c
3458
tg->scx.idle = idle;
kernel/sched/ext.c
4751
bool idle;
kernel/sched/ext.c
4755
idle = list_empty(&rq->scx.runnable_list) &&
kernel/sched/ext.c
4758
if (idle && !SCX_HAS_OP(sch, dump_cpu))
kernel/sched/ext.c
4798
&dctx, cpu, idle);
kernel/sched/ext.c
4806
if (idle && used == seq_buf_used(&ns))
kernel/sched/ext.c
5535
static void sched_ext_ops__update_idle(s32 cpu, bool idle) {}
kernel/sched/ext.c
5550
static void sched_ext_ops__cgroup_set_idle(struct cgroup *cgrp, bool idle) {}
kernel/sched/ext.c
5557
static void sched_ext_ops__dump_cpu(struct scx_dump_ctx *ctx, s32 cpu, bool idle) {}
kernel/sched/ext.h
62
void __scx_update_idle(struct rq *rq, bool idle, bool do_notify);
kernel/sched/ext.h
64
static inline void scx_update_idle(struct rq *rq, bool idle, bool do_notify)
kernel/sched/ext.h
67
__scx_update_idle(rq, idle, do_notify);
kernel/sched/ext.h
70
static inline void scx_update_idle(struct rq *rq, bool idle, bool do_notify) {}
kernel/sched/ext.h
82
void scx_group_set_idle(struct task_group *tg, bool idle);
kernel/sched/ext.h
92
static inline void scx_group_set_idle(struct task_group *tg, bool idle) {}
kernel/sched/ext_idle.c
690
static void update_builtin_idle(int cpu, bool idle)
kernel/sched/ext_idle.c
695
assign_cpu(cpu, idle_cpus, idle);
kernel/sched/ext_idle.c
702
if (idle) {
kernel/sched/ext_idle.c
733
void __scx_update_idle(struct rq *rq, bool idle, bool do_notify)
kernel/sched/ext_idle.c
756
update_builtin_idle(cpu, idle);
kernel/sched/ext_idle.c
771
SCX_CALL_OP(sch, SCX_KF_REST, update_idle, rq, cpu_of(rq), idle);
kernel/sched/ext_internal.h
520
void (*update_idle)(s32 cpu, bool idle);
kernel/sched/ext_internal.h
606
void (*dump_cpu)(struct scx_dump_ctx *ctx, s32 cpu, bool idle);
kernel/sched/ext_internal.h
720
void (*cgroup_set_idle)(struct cgroup *cgrp, bool idle);
kernel/sched/fair.c
10390
if (!env->idle)
kernel/sched/fair.c
10414
if (!env->idle || !busiest->sum_nr_running)
kernel/sched/fair.c
10521
} else if (env->idle && sched_reduced_capacity(rq, env->sd)) {
kernel/sched/fair.c
10533
if (!local_group && env->idle && sgs->sum_h_nr_running &&
kernel/sched/fair.c
10761
if (rq->curr != rq->idle && rq->curr != p)
kernel/sched/fair.c
11076
if (!sched_feat(SIS_UTIL) || env->idle == CPU_NEWLY_IDLE)
kernel/sched/fair.c
11156
if (env->idle != CPU_NEWLY_IDLE ||
kernel/sched/fair.c
11285
if (env->idle && env->imbalance == 0) {
kernel/sched/fair.c
11497
if (!env->idle) {
kernel/sched/fair.c
11705
return env->idle && sched_use_asym_prio(env->sd, env->dst_cpu) &&
kernel/sched/fair.c
11743
if (env->idle &&
kernel/sched/fair.c
11778
if (env->idle == CPU_NEWLY_IDLE) {
kernel/sched/fair.c
11825
enum cpu_idle_type idle)
kernel/sched/fair.c
11832
__schedstat_add(sd->lb_imbalance_load[idle], env->imbalance);
kernel/sched/fair.c
11835
__schedstat_add(sd->lb_imbalance_util[idle], env->imbalance);
kernel/sched/fair.c
11838
__schedstat_add(sd->lb_imbalance_task[idle], env->imbalance);
kernel/sched/fair.c
11841
__schedstat_add(sd->lb_imbalance_misfit[idle], env->imbalance);
kernel/sched/fair.c
11866
struct sched_domain *sd, enum cpu_idle_type idle,
kernel/sched/fair.c
11880
.idle = idle,
kernel/sched/fair.c
11890
schedstat_inc(sd->lb_count[idle]);
kernel/sched/fair.c
11908
schedstat_inc(sd->lb_nobusyg[idle]);
kernel/sched/fair.c
11914
schedstat_inc(sd->lb_nobusyq[idle]);
kernel/sched/fair.c
11920
update_lb_imbalance_stat(&env, sd, idle);
kernel/sched/fair.c
12037
schedstat_inc(sd->lb_failed[idle]);
kernel/sched/fair.c
12047
if (idle != CPU_NEWLY_IDLE &&
kernel/sched/fair.c
12119
schedstat_inc(sd->lb_balanced[idle]);
kernel/sched/fair.c
12136
if (env.idle == CPU_NEWLY_IDLE ||
kernel/sched/fair.c
12244
.idle = CPU_IDLE,
kernel/sched/fair.c
12331
static void sched_balance_domains(struct rq *rq, enum cpu_idle_type idle)
kernel/sched/fair.c
12335
int busy = idle != CPU_IDLE && !sched_idle_cpu(cpu);
kernel/sched/fair.c
12366
if (sched_balance_rq(cpu, rq, sd, idle, &continue_balancing)) {
kernel/sched/fair.c
12372
idle = idle_cpu(cpu);
kernel/sched/fair.c
12373
busy = !idle && !sched_idle_cpu(cpu);
kernel/sched/fair.c
12836
static bool nohz_idle_balance(struct rq *this_rq, enum cpu_idle_type idle)
kernel/sched/fair.c
12845
if (idle != CPU_IDLE)
kernel/sched/fair.c
12905
static inline bool nohz_idle_balance(struct rq *this_rq, enum cpu_idle_type idle)
kernel/sched/fair.c
13077
enum cpu_idle_type idle = this_rq->idle_balance;
kernel/sched/fair.c
13086
if (nohz_idle_balance(this_rq, idle))
kernel/sched/fair.c
13091
sched_balance_domains(this_rq, idle);
kernel/sched/fair.c
13865
int sched_group_set_idle(struct task_group *tg, long idle)
kernel/sched/fair.c
13872
if (idle < 0 || idle > 1)
kernel/sched/fair.c
13877
if (tg->idle == idle) {
kernel/sched/fair.c
13882
tg->idle = idle;
kernel/sched/fair.c
13894
grp_cfs_rq->idle = idle;
kernel/sched/fair.c
457
return tg->idle > 0;
kernel/sched/fair.c
462
return cfs_rq->idle > 0;
kernel/sched/fair.c
5697
cfs_b->idle = 0;
kernel/sched/fair.c
6091
if (rq->curr == rq->idle && rq->cfs.nr_queued)
kernel/sched/fair.c
6272
if (cfs_b->idle && !throttled)
kernel/sched/fair.c
6277
cfs_b->idle = 1;
kernel/sched/fair.c
6300
cfs_b->idle = 0;
kernel/sched/fair.c
6507
int idle = 0;
kernel/sched/fair.c
6516
idle = do_sched_cfs_period_timer(cfs_b, overrun, flags);
kernel/sched/fair.c
6549
if (idle)
kernel/sched/fair.c
6553
return idle ? HRTIMER_NORESTART : HRTIMER_RESTART;
kernel/sched/fair.c
7470
struct cpuidle_state *idle = idle_get_state(rq);
kernel/sched/fair.c
7471
if (idle && idle->exit_latency < min_exit_latency) {
kernel/sched/fair.c
7477
min_exit_latency = idle->exit_latency;
kernel/sched/fair.c
7480
} else if ((!idle || idle->exit_latency == min_exit_latency) &&
kernel/sched/fair.c
7624
bool idle = true;
kernel/sched/fair.c
7629
idle = false;
kernel/sched/fair.c
7643
if (idle)
kernel/sched/fair.c
8988
goto idle;
kernel/sched/fair.c
9039
idle:
kernel/sched/fair.c
9336
enum cpu_idle_type idle;
kernel/sched/fair.c
9433
if (env->idle == CPU_IDLE)
kernel/sched/fair.c
9541
if (env->idle == CPU_NEWLY_IDLE ||
kernel/sched/fair.c
9600
schedstat_inc(env->sd->lb_hot_gained[env->idle]);
kernel/sched/fair.c
9636
schedstat_inc(env->sd->lb_gained[env->idle]);
kernel/sched/fair.c
9674
if (env->idle && env->src_rq->nr_running <= 1)
kernel/sched/fair.c
9754
if (env->idle == CPU_NEWLY_IDLE)
kernel/sched/fair.c
9778
schedstat_add(env->sd->lb_gained[env->idle], detached);
kernel/sched/idle.c
504
return rq->idle;
kernel/sched/idle.c
550
struct sched_entity *se = &rq->idle->se;
kernel/sched/idle.c
569
DEFINE_SCHED_CLASS(idle) = {
kernel/sched/rt.c
105
int idle = 0;
kernel/sched/rt.c
115
idle = do_sched_rt_period_timer(rt_b, overrun);
kernel/sched/rt.c
118
if (idle)
kernel/sched/rt.c
122
return idle ? HRTIMER_NORESTART : HRTIMER_RESTART;
kernel/sched/rt.c
780
int i, idle = 1, throttled = 0;
kernel/sched/rt.c
838
if (rt_rq->rt_nr_running && rq->curr == rq->idle)
kernel/sched/rt.c
842
idle = 0;
kernel/sched/rt.c
845
idle = 0;
kernel/sched/rt.c
860
return idle;
kernel/sched/sched.h
1149
struct task_struct *idle;
kernel/sched/sched.h
1398
return rq->curr == rq->idle && !rq->nr_running && !rq->ttwu_pending;
kernel/sched/sched.h
457
u8 idle;
kernel/sched/sched.h
479
int idle;
kernel/sched/sched.h
612
extern int sched_group_set_idle(struct task_group *tg, long idle);
kernel/sched/sched.h
618
static inline int sched_group_set_idle(struct task_group *tg, long idle) { return 0; }
kernel/sched/sched.h
749
int idle;
kernel/sched/stats.h
335
if (prev != rq->idle)
kernel/sched/stats.h
338
if (next != rq->idle)
kernel/sched/syscalls.c
194
return cpu_rq(cpu)->idle;
kernel/sched/syscalls.c
202
if (sched_core_enabled(rq) && rq->curr == rq->idle)
kernel/time/tick-internal.h
178
u64 timer_base_try_to_set_idle(unsigned long basej, u64 basem, bool *idle);
kernel/time/tick-sched.c
789
ktime_t now, idle;
kernel/time/tick-sched.c
805
idle = ktime_add(*sleeptime, delta);
kernel/time/tick-sched.c
807
idle = *sleeptime;
kernel/time/tick-sched.c
811
return ktime_to_us(idle);
kernel/time/timer.c
2184
bool *idle)
kernel/time/timer.c
2196
if (idle)
kernel/time/timer.c
2197
*idle = true;
kernel/time/timer.c
2223
timer_use_tmigr(basej, basem, &nextevt, idle,
kernel/time/timer.c
2236
if (idle) {
kernel/time/timer.c
2260
*idle = base_local->is_idle;
kernel/time/timer.c
2307
u64 timer_base_try_to_set_idle(unsigned long basej, u64 basem, bool *idle)
kernel/time/timer.c
2309
if (*idle)
kernel/time/timer.c
2312
return __get_next_timer_interrupt(basej, basem, idle);
kernel/time/timer_migration.c
1014
if (!tmc->available || !tmc->idle) {
kernel/time/timer_migration.c
1207
if (!tmc->idle) {
kernel/time/timer_migration.c
1390
tmc->idle = true;
kernel/time/timer_migration.c
1429
if (WARN_ON_ONCE(tmc->idle))
kernel/time/timer_migration.c
1463
WARN_ON_ONCE(!tmc->available || tmc->idle);
kernel/time/timer_migration.c
1514
tmc->idle = timer_base_is_idle();
kernel/time/timer_migration.c
1515
if (!tmc->idle)
kernel/time/timer_migration.c
737
if (WARN_ON_ONCE(!tmc->idle))
kernel/time/timer_migration.c
741
tmc->idle = false;
kernel/time/timer_migration.h
101
bool idle;
kernel/workqueue.c
6414
bool idle = true;
kernel/workqueue.c
6419
idle = false;
kernel/workqueue.c
6423
if (idle) /* Nothing to print for idle workqueue */
net/ax25/af_ax25.c
1853
ax25_info.idle = ax25->idle / (60 * HZ);
net/ax25/af_ax25.c
1982
ax25->idle / (60 * HZ),
net/ax25/af_ax25.c
447
ax25->idle = ax25_ctl.arg * 60 * HZ;
net/ax25/af_ax25.c
478
ax25->idle = msecs_to_jiffies(ax25_dev->values[AX25_VALUES_IDLE]);
net/ax25/af_ax25.c
512
ax25->idle = msecs_to_jiffies(AX25_DEF_IDLE);
net/ax25/af_ax25.c
631
ax25->idle = opt * 60 * HZ;
net/ax25/af_ax25.c
765
val = ax25->idle / (60 * HZ);
net/ax25/af_ax25.c
971
ax25->idle = oax25->idle;
net/ax25/ax25_timer.c
73
if (ax25->idle > 0)
net/ax25/ax25_timer.c
74
mod_timer(&ax25->idletimer, jiffies + ax25->idle);
net/core/pktgen.c
3347
ktime_t idle = ns_to_ktime(pkt_dev->idle_acc);
net/core/pktgen.c
3351
(unsigned long long)ktime_to_us(ktime_sub(elapsed, idle)),
net/core/pktgen.c
3352
(unsigned long long)ktime_to_us(idle),
net/core/pktgen.c
571
u64 idle;
net/core/pktgen.c
696
idle = pkt_dev->idle_acc;
net/core/pktgen.c
697
do_div(idle, NSEC_PER_USEC);
net/core/pktgen.c
720
(unsigned long long) idle);
net/mac80211/chan.c
1078
if (new_idle != sdata->vif.cfg.idle) {
net/mac80211/chan.c
1079
sdata->vif.cfg.idle = new_idle;
net/mac80211/iface.c
1876
sdata->vif.cfg.idle = true;
net/mac80211/trace.h
450
__field(bool, idle)
net/mac80211/trace.h
474
__entry->idle = sdata->vif.cfg.idle;
net/netrom/af_netrom.c
1326
nr->idle / (60 * HZ),
net/netrom/af_netrom.c
340
nr->idle = opt * 60 * HZ;
net/netrom/af_netrom.c
383
val = nr->idle / (60 * HZ);
net/netrom/af_netrom.c
463
nr->idle =
net/netrom/af_netrom.c
509
nr->idle = onr->idle;
net/netrom/nr_timer.c
73
if (nr->idle > 0)
net/netrom/nr_timer.c
74
sk_reset_timer(sk, &nr->idletimer, jiffies + nr->idle);
net/rose/af_rose.c
1503
rose->idle / (60 * HZ),
net/rose/af_rose.c
443
rose->idle = opt * 60 * HZ;
net/rose/af_rose.c
494
val = rose->idle / (60 * HZ);
net/rose/af_rose.c
582
rose->idle = msecs_to_jiffies(sysctl_rose_no_activity_timeout);
net/rose/af_rose.c
627
rose->idle = orose->idle;
net/rose/rose_timer.c
96
if (rose->idle > 0) {
net/rose/rose_timer.c
98
rose->idletimer.expires = jiffies + rose->idle;
net/rxrpc/ar-internal.h
895
u32 idle; /* Max time since last data packet (msec) */
net/rxrpc/call_object.c
223
if (p->timeouts.idle)
net/rxrpc/call_object.c
224
call->next_req_timo = umin(p->timeouts.idle, 1);
net/rxrpc/sendmsg.c
583
if (p->call.nr_timeouts >= 2 && p->call.timeouts.idle > 60 * 60 * 1000)
net/rxrpc/sendmsg.c
742
WRITE_ONCE(call->next_req_timo, p.call.timeouts.idle);
sound/soc/codecs/wm8994.c
116
bool idle;
sound/soc/codecs/wm8994.c
120
idle = !wm8994->jack_mic;
sound/soc/codecs/wm8994.c
141
if (rates[i].idle != idle)
sound/soc/codecs/wm8994.c
146
else if (rates[best].idle != idle)
sound/soc/codecs/wm8994.c
155
idle ? "idle" : "active");
sound/soc/rockchip/rockchip_sai.c
133
unsigned int reg, idle, val;
sound/soc/rockchip/rockchip_sai.c
138
idle = SAI_STATUS_FS_IDLE;
sound/soc/rockchip/rockchip_sai.c
139
idle = sai->version >= SAI_VER_2311 ? idle >> 1 : idle;
sound/soc/rockchip/rockchip_sai.c
142
idle = SAI_XFER_FS_IDLE;
sound/soc/rockchip/rockchip_sai.c
146
(val & idle), 10, TIMEOUT_US);
sound/soc/rockchip/rockchip_sai.c
156
unsigned int idle = 0;
sound/soc/rockchip/rockchip_sai.c
162
idle |= SAI_STATUS_TX_IDLE;
sound/soc/rockchip/rockchip_sai.c
164
idle |= SAI_STATUS_RX_IDLE;
sound/soc/rockchip/rockchip_sai.c
165
idle = sai->version >= SAI_VER_2311 ? idle >> 1 : idle;
sound/soc/rockchip/rockchip_sai.c
169
idle |= SAI_XFER_TX_IDLE;
sound/soc/rockchip/rockchip_sai.c
171
idle |= SAI_XFER_RX_IDLE;
sound/soc/rockchip/rockchip_sai.c
175
(val & idle), 10, TIMEOUT_US);
sound/soc/ti/omap-mcbsp.c
477
int idle;
sound/soc/ti/omap-mcbsp.c
500
idle = !((MCBSP_READ_CACHE(mcbsp, SPCR2) |
sound/soc/ti/omap-mcbsp.c
503
if (idle) {
tools/perf/builtin-kwork.c
1595
u64 idle_time = stat->cpus_runtime[MAX_NR_CPUS].idle;
tools/perf/builtin-kwork.c
2024
stat->cpus_runtime[work->cpu].idle += work->total_runtime;
tools/perf/builtin-kwork.c
2025
stat->cpus_runtime[MAX_NR_CPUS].idle += work->total_runtime;
tools/perf/builtin-sched.c
2436
struct thread *idle = idle_threads[i];
tools/perf/builtin-sched.c
2438
if (idle) {
tools/perf/builtin-sched.c
2441
itr = thread__priv(idle);
tools/perf/builtin-sched.c
2445
thread__delete(idle);
tools/perf/builtin-sched.c
2525
struct thread *idle;
tools/perf/builtin-sched.c
2528
idle = get_idle_thread(sample->cpu);
tools/perf/builtin-sched.c
2529
if (idle == NULL) {
tools/perf/builtin-sched.c
2534
itr = thread__priv(idle);
tools/perf/builtin-top.c
832
if (al.sym == NULL || !al.sym->idle) {
tools/perf/util/bpf_lock_contention.c
743
struct thread *idle = machine__findnew_thread(machine,
tools/perf/util/bpf_lock_contention.c
746
thread__set_comm(idle, "swapper", /*timestamp=*/0);
tools/perf/util/kwork.h
188
u64 idle;
tools/perf/util/symbol.c
377
sym->idle = symbol__is_idle(name);
tools/perf/util/symbol.h
60
u8 idle:1;
tools/sched_ext/scx_flatcg.c
58
__u64 sum = 0, idle = 0;
tools/sched_ext/scx_flatcg.c
92
idle = v;
tools/sched_ext/scx_flatcg.c
96
delta_idle = idle - *last_idle;
tools/sched_ext/scx_flatcg.c
98
*last_idle = idle;
tools/sched_ext/scx_qmap.bpf.c
621
void BPF_STRUCT_OPS(qmap_dump_cpu, struct scx_dump_ctx *dctx, s32 cpu, bool idle)
tools/sched_ext/scx_qmap.bpf.c
626
if (suppress_dump || idle)
tools/sched_ext/scx_userland.bpf.c
258
void BPF_STRUCT_OPS(userland_update_idle, s32 cpu, bool idle)
tools/sched_ext/scx_userland.bpf.c
264
if (!idle)
tools/testing/selftests/landlock/tsync_test.c
102
ASSERT_EQ(0, pthread_create(&t1, NULL, idle, NULL));
tools/testing/selftests/landlock/tsync_test.c
103
ASSERT_EQ(0, pthread_create(&t2, NULL, idle, NULL));
tools/testing/selftests/landlock/tsync_test.c
221
ASSERT_EQ(0, pthread_create(&threads[i], NULL, idle, NULL));
tools/testing/selftests/landlock/tsync_test.c
73
ASSERT_EQ(0, pthread_create(&t1, NULL, idle, &no_new_privs1));
tools/testing/selftests/landlock/tsync_test.c
74
ASSERT_EQ(0, pthread_create(&t2, NULL, idle, &no_new_privs2));
tools/testing/selftests/sched_ext/maximal.bpf.c
67
void BPF_STRUCT_OPS(maximal_update_idle, s32 cpu, bool idle)
tools/testing/selftests/sched_ext/numa.bpf.c
25
bool idle;
tools/testing/selftests/sched_ext/numa.bpf.c
28
idle = bpf_cpumask_test_cpu(cpu, idle_cpumask);
tools/testing/selftests/sched_ext/numa.bpf.c
31
return idle;