arch/arm64/kernel/proton-pack.c
659
bool ssbs = false, kthread = tsk->flags & PF_KTHREAD;
arch/arm64/kernel/proton-pack.c
663
else if (spectre_v4_mitigations_dynamic() && !kthread)
drivers/block/aoe/aoecmd.c
1270
task = kthread_run(kthread, k, "%s", k->name);
drivers/crypto/ccp/ccp-dev-v3.c
466
struct task_struct *kthread;
drivers/crypto/ccp/ccp-dev-v3.c
470
kthread = kthread_run(ccp_cmd_queue_thread, cmd_q,
drivers/crypto/ccp/ccp-dev-v3.c
472
if (IS_ERR(kthread)) {
drivers/crypto/ccp/ccp-dev-v3.c
474
PTR_ERR(kthread));
drivers/crypto/ccp/ccp-dev-v3.c
475
ret = PTR_ERR(kthread);
drivers/crypto/ccp/ccp-dev-v3.c
479
cmd_q->kthread = kthread;
drivers/crypto/ccp/ccp-dev-v3.c
505
if (ccp->cmd_q[i].kthread)
drivers/crypto/ccp/ccp-dev-v3.c
506
kthread_stop(ccp->cmd_q[i].kthread);
drivers/crypto/ccp/ccp-dev-v3.c
544
if (ccp->cmd_q[i].kthread)
drivers/crypto/ccp/ccp-dev-v3.c
545
kthread_stop(ccp->cmd_q[i].kthread);
drivers/crypto/ccp/ccp-dev-v5.c
1045
if (ccp->cmd_q[i].kthread)
drivers/crypto/ccp/ccp-dev-v5.c
1046
kthread_stop(ccp->cmd_q[i].kthread);
drivers/crypto/ccp/ccp-dev-v5.c
949
struct task_struct *kthread;
drivers/crypto/ccp/ccp-dev-v5.c
953
kthread = kthread_run(ccp_cmd_queue_thread, cmd_q,
drivers/crypto/ccp/ccp-dev-v5.c
955
if (IS_ERR(kthread)) {
drivers/crypto/ccp/ccp-dev-v5.c
957
PTR_ERR(kthread));
drivers/crypto/ccp/ccp-dev-v5.c
958
ret = PTR_ERR(kthread);
drivers/crypto/ccp/ccp-dev-v5.c
962
cmd_q->kthread = kthread;
drivers/crypto/ccp/ccp-dev-v5.c
993
if (ccp->cmd_q[i].kthread)
drivers/crypto/ccp/ccp-dev-v5.c
994
kthread_stop(ccp->cmd_q[i].kthread);
drivers/crypto/ccp/ccp-dev.c
339
wake_up_process(ccp->cmd_q[i].kthread);
drivers/crypto/ccp/ccp-dev.c
371
wake_up_process(ccp->cmd_q[i].kthread);
drivers/crypto/ccp/ccp-dev.c
567
wake_up_process(ccp->cmd_q[i].kthread);
drivers/crypto/ccp/ccp-dev.c
594
wake_up_process(ccp->cmd_q[i].kthread);
drivers/crypto/ccp/ccp-dev.h
294
struct task_struct *kthread;
drivers/media/cec/core/cec-core.c
266
adap->kthread = kthread_run(cec_thread_func, adap, "cec-%s", name);
drivers/media/cec/core/cec-core.c
267
if (IS_ERR(adap->kthread)) {
drivers/media/cec/core/cec-core.c
269
res = PTR_ERR(adap->kthread);
drivers/media/cec/core/cec-core.c
283
kthread_stop(adap->kthread);
drivers/media/cec/core/cec-core.c
392
kthread_stop(adap->kthread);
drivers/media/cec/core/cec-pin-priv.h
181
struct task_struct *kthread;
drivers/media/cec/core/cec-pin.c
1168
if (!pin->kthread) {
drivers/media/cec/core/cec-pin.c
1169
pin->kthread = kthread_run(cec_pin_thread_func, adap,
drivers/media/cec/core/cec-pin.c
1171
if (IS_ERR(pin->kthread)) {
drivers/media/cec/core/cec-pin.c
1172
int err = PTR_ERR(pin->kthread);
drivers/media/cec/core/cec-pin.c
1175
pin->kthread = NULL;
drivers/media/cec/core/cec-pin.c
1181
} else if (pin->kthread) {
drivers/media/cec/core/cec-pin.c
1323
if (pin->kthread)
drivers/media/cec/core/cec-pin.c
1324
kthread_stop(pin->kthread);
drivers/media/cec/core/cec-pin.c
1325
pin->kthread = NULL;
drivers/media/i2c/msp3400-driver.c
305
if (NULL == state->kthread)
drivers/media/i2c/msp3400-driver.c
846
state->kthread = kthread_run(thread_func, client, "msp34xx");
drivers/media/i2c/msp3400-driver.c
848
if (IS_ERR(state->kthread))
drivers/media/i2c/msp3400-driver.c
861
if (state->kthread) {
drivers/media/i2c/msp3400-driver.c
863
kthread_stop(state->kthread);
drivers/media/i2c/msp3400-driver.h
109
struct task_struct *kthread;
drivers/media/pci/cx88/cx88-video.c
1496
core->kthread = kthread_run(cx88_audio_thread,
drivers/media/pci/cx88/cx88-video.c
1498
if (IS_ERR(core->kthread)) {
drivers/media/pci/cx88/cx88-video.c
1499
err = PTR_ERR(core->kthread);
drivers/media/pci/cx88/cx88-video.c
1529
if (core->kthread) {
drivers/media/pci/cx88/cx88-video.c
1530
kthread_stop(core->kthread);
drivers/media/pci/cx88/cx88-video.c
1531
core->kthread = NULL;
drivers/media/pci/cx88/cx88.h
376
struct task_struct *kthread;
drivers/media/pci/pt1/pt1.c
1316
if (pt1->kthread)
drivers/media/pci/pt1/pt1.c
1317
kthread_stop(pt1->kthread);
drivers/media/pci/pt1/pt1.c
681
if (!pt1->kthread) {
drivers/media/pci/pt1/pt1.c
682
pt1->kthread = kthread_run(pt1_thread, pt1, "earth-pt1");
drivers/media/pci/pt1/pt1.c
683
if (IS_ERR(pt1->kthread)) {
drivers/media/pci/pt1/pt1.c
684
ret = PTR_ERR(pt1->kthread);
drivers/media/pci/pt1/pt1.c
685
pt1->kthread = NULL;
drivers/media/pci/pt1/pt1.c
715
if (count == 0 && pt1->kthread) {
drivers/media/pci/pt1/pt1.c
716
kthread_stop(pt1->kthread);
drivers/media/pci/pt1/pt1.c
717
pt1->kthread = NULL;
drivers/media/pci/pt1/pt1.c
79
struct task_struct *kthread;
drivers/media/pci/saa7164/saa7164-core.c
1406
dev->kthread = kthread_run(saa7164_thread_function, dev,
drivers/media/pci/saa7164/saa7164-core.c
1408
if (IS_ERR(dev->kthread)) {
drivers/media/pci/saa7164/saa7164-core.c
1409
dev->kthread = NULL;
drivers/media/pci/saa7164/saa7164-core.c
1446
if (fw_debug && dev->kthread) {
drivers/media/pci/saa7164/saa7164-core.c
1447
kthread_stop(dev->kthread);
drivers/media/pci/saa7164/saa7164-core.c
1448
dev->kthread = NULL;
drivers/media/pci/saa7164/saa7164.h
482
struct task_struct *kthread;
drivers/media/pci/solo6x10/solo6x10-v4l2.c
285
solo_dev->kthread = kthread_run(solo_thread, solo_dev, SOLO6X10_NAME "_disp");
drivers/media/pci/solo6x10/solo6x10-v4l2.c
287
if (IS_ERR(solo_dev->kthread)) {
drivers/media/pci/solo6x10/solo6x10-v4l2.c
288
ret = PTR_ERR(solo_dev->kthread);
drivers/media/pci/solo6x10/solo6x10-v4l2.c
289
solo_dev->kthread = NULL;
drivers/media/pci/solo6x10/solo6x10-v4l2.c
299
if (!solo_dev->kthread)
drivers/media/pci/solo6x10/solo6x10-v4l2.c
303
kthread_stop(solo_dev->kthread);
drivers/media/pci/solo6x10/solo6x10-v4l2.c
304
solo_dev->kthread = NULL;
drivers/media/pci/solo6x10/solo6x10.h
268
struct task_struct *kthread;
drivers/media/radio/radio-miropcm20.c
384
IS_ERR_OR_NULL(dev->kthread)) {
drivers/media/radio/radio-miropcm20.c
385
dev->kthread = kthread_run(pcm20_thread, dev, "%s",
drivers/media/radio/radio-miropcm20.c
387
if (IS_ERR(dev->kthread)) {
drivers/media/radio/radio-miropcm20.c
390
return PTR_ERR(dev->kthread);
drivers/media/radio/radio-miropcm20.c
400
if (v4l2_fh_is_singular_file(file) && !IS_ERR_OR_NULL(dev->kthread)) {
drivers/media/radio/radio-miropcm20.c
401
kthread_stop(dev->kthread);
drivers/media/radio/radio-miropcm20.c
402
dev->kthread = NULL;
drivers/media/radio/radio-miropcm20.c
73
struct task_struct *kthread;
drivers/media/test-drivers/vimc/vimc-streamer.c
206
if (stream->kthread)
drivers/media/test-drivers/vimc/vimc-streamer.c
213
stream->kthread = kthread_run(vimc_streamer_thread, stream,
drivers/media/test-drivers/vimc/vimc-streamer.c
216
if (IS_ERR(stream->kthread)) {
drivers/media/test-drivers/vimc/vimc-streamer.c
217
ret = PTR_ERR(stream->kthread);
drivers/media/test-drivers/vimc/vimc-streamer.c
220
stream->kthread = NULL;
drivers/media/test-drivers/vimc/vimc-streamer.c
225
if (!stream->kthread)
drivers/media/test-drivers/vimc/vimc-streamer.c
228
ret = kthread_stop(stream->kthread);
drivers/media/test-drivers/vimc/vimc-streamer.c
238
stream->kthread = NULL;
drivers/media/test-drivers/vimc/vimc-streamer.h
38
struct task_struct *kthread;
drivers/misc/sgi-xp/xpc_main.c
1182
struct task_struct *kthread;
drivers/misc/sgi-xp/xpc_main.c
1231
kthread = kthread_run(xpc_hb_checker, NULL, XPC_HB_CHECK_THREAD_NAME);
drivers/misc/sgi-xp/xpc_main.c
1232
if (IS_ERR(kthread)) {
drivers/misc/sgi-xp/xpc_main.c
1243
kthread = kthread_run(xpc_initiate_discovery, NULL,
drivers/misc/sgi-xp/xpc_main.c
1245
if (IS_ERR(kthread)) {
drivers/misc/sgi-xp/xpc_main.c
577
struct task_struct *kthread;
drivers/misc/sgi-xp/xpc_main.c
588
kthread = kthread_run(xpc_activating, (void *)((u64)partid), "xpc%02d",
drivers/misc/sgi-xp/xpc_main.c
590
if (IS_ERR(kthread)) {
drivers/misc/sgi-xp/xpc_main.c
767
struct task_struct *kthread;
drivers/misc/sgi-xp/xpc_main.c
796
kthread = kthread_run(xpc_kthread_start, (void *)args,
drivers/misc/sgi-xp/xpc_main.c
798
if (IS_ERR(kthread)) {
drivers/scsi/bnx2fc/bnx2fc_fcoe.c
2713
bg->kthread = l2_thread;
drivers/scsi/bnx2fc/bnx2fc_fcoe.c
2780
l2_thread = bg->kthread;
drivers/scsi/bnx2fc/bnx2fc_fcoe.c
2781
bg->kthread = NULL;
drivers/scsi/bnx2fc/bnx2fc_fcoe.c
470
wake_up_process(bg->kthread);
drivers/usb/atm/ueagle-atm.c
120
struct task_struct *kthread;
drivers/usb/atm/ueagle-atm.c
2184
sc->kthread = kthread_create(uea_kthread, sc, "ueagle-atm");
drivers/usb/atm/ueagle-atm.c
2185
if (IS_ERR(sc->kthread)) {
drivers/usb/atm/ueagle-atm.c
2187
ret = PTR_ERR(sc->kthread);
drivers/usb/atm/ueagle-atm.c
2212
ret = kthread_stop(sc->kthread);
drivers/usb/atm/ueagle-atm.c
2613
wake_up_process(sc->kthread);
drivers/vfio/vfio_iommu_type1.c
3147
bool kthread = current->mm == NULL;
drivers/vfio/vfio_iommu_type1.c
3164
if (kthread)
drivers/vfio/vfio_iommu_type1.c
3192
if (kthread)
fs/smb/server/transport_tcp.c
314
struct task_struct *kthread;
fs/smb/server/transport_tcp.c
316
kthread = kthread_run(ksmbd_kthread_fn, (void *)iface, "ksmbd-%s",
fs/smb/server/transport_tcp.c
318
if (IS_ERR(kthread)) {
fs/smb/server/transport_tcp.c
319
rc = PTR_ERR(kthread);
fs/smb/server/transport_tcp.c
322
iface->ksmbd_kthread = kthread;
fs/smb/server/transport_tcp.c
41
static void tcp_stop_kthread(struct task_struct *kthread);
fs/smb/server/transport_tcp.c
607
static void tcp_stop_kthread(struct task_struct *kthread)
fs/smb/server/transport_tcp.c
611
if (!kthread)
fs/smb/server/transport_tcp.c
614
ret = kthread_stop(kthread);
include/linux/console.h
470
struct task_struct *kthread;
include/linux/kthread.h
11
struct kthread;
include/linux/kthread.h
122
void kthread_do_exit(struct kthread *, long);
include/linux/kthread.h
21
static inline struct kthread *tsk_is_kthread(struct task_struct *p)
include/linux/sched/task.h
30
u32 kthread:1;
include/media/cec.h
252
struct task_struct *kthread;
include/scsi/libfcoe.h
332
struct task_struct *kthread;
kernel/bpf/cpumap.c
482
rcpu->kthread = kthread_create_on_node(cpu_map_kthread_run, rcpu, numa,
kernel/bpf/cpumap.c
485
if (IS_ERR(rcpu->kthread)) {
kernel/bpf/cpumap.c
486
err = PTR_ERR(rcpu->kthread);
kernel/bpf/cpumap.c
491
kthread_bind(rcpu->kthread, cpu);
kernel/bpf/cpumap.c
492
wake_up_process(rcpu->kthread);
kernel/bpf/cpumap.c
532
kthread_stop(rcpu->kthread);
kernel/bpf/cpumap.c
69
struct task_struct *kthread;
kernel/bpf/cpumap.c
813
wake_up_process(rcpu->kthread);
kernel/bpf/cpumap.c
829
wake_up_process(bq->obj->kthread);
kernel/exit.c
899
struct kthread *kthread;
kernel/exit.c
905
kthread = tsk_is_kthread(tsk);
kernel/exit.c
906
if (unlikely(kthread))
kernel/exit.c
907
kthread_do_exit(kthread, code);
kernel/fork.c
2054
if (args->kthread)
kernel/fork.c
2151
if (args->kthread) {
kernel/fork.c
2570
.kthread = 1,
kernel/fork.c
2711
.kthread = 1,
kernel/kthread.c
102
struct kthread *kthread;
kernel/kthread.c
107
kthread = kzalloc_obj(*kthread);
kernel/kthread.c
108
if (!kthread)
kernel/kthread.c
111
init_completion(&kthread->exited);
kernel/kthread.c
112
init_completion(&kthread->parked);
kernel/kthread.c
113
INIT_LIST_HEAD(&kthread->affinity_node);
kernel/kthread.c
114
p->vfork_done = &kthread->exited;
kernel/kthread.c
116
kthread->task = p;
kernel/kthread.c
117
kthread->node = tsk_fork_get_node(current);
kernel/kthread.c
118
p->worker_private = kthread;
kernel/kthread.c
124
struct kthread *kthread;
kernel/kthread.c
129
kthread = to_kthread(k);
kernel/kthread.c
130
if (!kthread)
kernel/kthread.c
134
WARN_ON_ONCE(kthread->blkcg_css);
kernel/kthread.c
137
kfree(kthread->full_name);
kernel/kthread.c
138
kfree(kthread);
kernel/kthread.c
1704
struct kthread *kthread;
kernel/kthread.c
1708
kthread = to_kthread(current);
kernel/kthread.c
1709
if (!kthread)
kernel/kthread.c
1712
if (kthread->blkcg_css) {
kernel/kthread.c
1713
css_put(kthread->blkcg_css);
kernel/kthread.c
1714
kthread->blkcg_css = NULL;
kernel/kthread.c
1718
kthread->blkcg_css = css;
kernel/kthread.c
1730
struct kthread *kthread;
kernel/kthread.c
1733
kthread = to_kthread(current);
kernel/kthread.c
1734
if (kthread)
kernel/kthread.c
1735
return kthread->blkcg_css;
kernel/kthread.c
178
struct kthread *kthread = tsk_is_kthread(current);
kernel/kthread.c
180
if (!kthread)
kernel/kthread.c
183
return kthread->flags & (BIT(KTHREAD_SHOULD_STOP) | BIT(KTHREAD_SHOULD_PARK));
kernel/kthread.c
219
struct kthread *kthread = tsk_is_kthread(task);
kernel/kthread.c
220
if (kthread)
kernel/kthread.c
221
return kthread->threadfn;
kernel/kthread.c
251
struct kthread *kthread = tsk_is_kthread(task);
kernel/kthread.c
254
if (kthread)
kernel/kthread.c
255
copy_from_kernel_nofault(&data, &kthread->data, sizeof(data));
kernel/kthread.c
259
static void __kthread_parkme(struct kthread *self)
kernel/kthread.c
294
void kthread_do_exit(struct kthread *kthread, long result)
kernel/kthread.c
296
kthread->result = result;
kernel/kthread.c
297
if (!list_empty(&kthread->affinity_node)) {
kernel/kthread.c
299
list_del(&kthread->affinity_node);
kernel/kthread.c
302
if (kthread->preferred_affinity) {
kernel/kthread.c
303
kfree(kthread->preferred_affinity);
kernel/kthread.c
304
kthread->preferred_affinity = NULL;
kernel/kthread.c
330
static void kthread_fetch_affinity(struct kthread *kthread, struct cpumask *cpumask)
kernel/kthread.c
336
if (kthread->preferred_affinity) {
kernel/kthread.c
337
pref = kthread->preferred_affinity;
kernel/kthread.c
339
if (kthread->node == NUMA_NO_NODE)
kernel/kthread.c
342
pref = cpumask_of_node(kthread->node);
kernel/kthread.c
352
struct kthread *kthread = to_kthread(current);
kernel/kthread.c
364
WARN_ON_ONCE(!list_empty(&kthread->affinity_node));
kernel/kthread.c
365
list_add_tail(&kthread->affinity_node, &kthread_affinity_list);
kernel/kthread.c
373
kthread_fetch_affinity(kthread, affinity);
kernel/kthread.c
388
struct kthread *self;
kernel/kthread.c
459
pid = kernel_thread(kthread, create, create->full_name,
kernel/kthread.c
587
struct kthread *kthread = to_kthread(p);
kernel/kthread.c
589
WARN_ON_ONCE(kthread->started);
kernel/kthread.c
603
struct kthread *kthread = to_kthread(p);
kernel/kthread.c
605
WARN_ON_ONCE(kthread->started);
kernel/kthread.c
638
struct kthread *kthread = to_kthread(k);
kernel/kthread.c
639
if (!kthread)
kernel/kthread.c
645
clear_bit(KTHREAD_IS_PER_CPU, &kthread->flags);
kernel/kthread.c
649
kthread->cpu = cpu;
kernel/kthread.c
650
set_bit(KTHREAD_IS_PER_CPU, &kthread->flags);
kernel/kthread.c
655
struct kthread *kthread = tsk_is_kthread(p);
kernel/kthread.c
656
if (!kthread)
kernel/kthread.c
659
return test_bit(KTHREAD_IS_PER_CPU, &kthread->flags);
kernel/kthread.c
672
struct kthread *kthread = to_kthread(k);
kernel/kthread.c
674
if (!test_bit(KTHREAD_SHOULD_PARK, &kthread->flags))
kernel/kthread.c
680
if (test_bit(KTHREAD_IS_PER_CPU, &kthread->flags))
kernel/kthread.c
681
__kthread_bind(k, kthread->cpu, TASK_PARKED);
kernel/kthread.c
683
clear_bit(KTHREAD_SHOULD_PARK, &kthread->flags);
kernel/kthread.c
705
struct kthread *kthread = to_kthread(k);
kernel/kthread.c
710
if (WARN_ON_ONCE(test_bit(KTHREAD_SHOULD_PARK, &kthread->flags)))
kernel/kthread.c
713
set_bit(KTHREAD_SHOULD_PARK, &kthread->flags);
kernel/kthread.c
720
wait_for_completion(&kthread->parked);
kernel/kthread.c
749
struct kthread *kthread;
kernel/kthread.c
755
kthread = to_kthread(k);
kernel/kthread.c
756
set_bit(KTHREAD_SHOULD_STOP, &kthread->flags);
kernel/kthread.c
760
wait_for_completion(&kthread->exited);
kernel/kthread.c
761
ret = kthread->result;
kernel/kthread.c
82
static inline struct kthread *to_kthread(struct task_struct *k)
kernel/kthread.c
841
struct kthread *kthread = to_kthread(p);
kernel/kthread.c
845
if (!wait_task_inactive(p, TASK_UNINTERRUPTIBLE) || kthread->started) {
kernel/kthread.c
850
WARN_ON_ONCE(kthread->preferred_affinity);
kernel/kthread.c
855
kthread->preferred_affinity = kzalloc(sizeof(struct cpumask), GFP_KERNEL);
kernel/kthread.c
856
if (!kthread->preferred_affinity) {
kernel/kthread.c
862
cpumask_copy(kthread->preferred_affinity, mask);
kernel/kthread.c
863
WARN_ON_ONCE(!list_empty(&kthread->affinity_node));
kernel/kthread.c
864
list_add_tail(&kthread->affinity_node, &kthread_affinity_list);
kernel/kthread.c
865
kthread_fetch_affinity(kthread, affinity);
kernel/kthread.c
881
struct kthread *k;
kernel/kthread.c
90
struct kthread *kthread = to_kthread(tsk);
kernel/kthread.c
92
if (!kthread || !kthread->full_name) {
kernel/kthread.c
97
strscpy_pad(buf, kthread->full_name, buf_size);
kernel/printk/nbcon.c
1352
if (!con->kthread)
kernel/printk/nbcon.c
1355
kthread_stop(con->kthread);
kernel/printk/nbcon.c
1356
con->kthread = NULL;
kernel/printk/nbcon.c
1381
if (con->kthread)
kernel/printk/nbcon.c
1390
con->kthread = kt;
kernel/printk/nbcon.c
1396
sched_set_normal(con->kthread, -20);
kernel/printk/printk_ringbuffer_kunit_test.c
237
static void prbtest_add_kthread_cleanup(struct kunit *test, struct task_struct *kthread)
kernel/printk/printk_ringbuffer_kunit_test.c
241
err = kunit_add_action_or_reset(test, prbtest_kthread_cleanup, kthread);
kernel/sched/core.c
3022
bool kthread = p->flags & PF_KTHREAD;
kernel/sched/core.c
3026
if (kthread || is_migration_disabled(p)) {
kernel/sched/core.c
3040
if (!kthread && !cpumask_subset(ctx->new_mask, cpu_allowed_mask)) {
kernel/trace/trace_events_trigger.c
32
struct task_struct *kthread;
kernel/trace/trace_events_trigger.c
34
kthread = kthread_create(trigger_kthread_fn, NULL,
kernel/trace/trace_events_trigger.c
36
if (!IS_ERR(kthread))
kernel/trace/trace_events_trigger.c
37
WRITE_ONCE(trigger_kthread, kthread);
kernel/trace/trace_hwlat.c
165
if (!kdata->kthread)
kernel/trace/trace_hwlat.c
393
struct task_struct *kthread;
kernel/trace/trace_hwlat.c
396
kthread = kdata->kthread;
kernel/trace/trace_hwlat.c
398
if (!kthread)
kernel/trace/trace_hwlat.c
401
kthread_stop(kthread);
kernel/trace/trace_hwlat.c
402
kdata->kthread = NULL;
kernel/trace/trace_hwlat.c
419
struct task_struct *kthread;
kernel/trace/trace_hwlat.c
423
if (kdata->kthread)
kernel/trace/trace_hwlat.c
426
kthread = kthread_create(kthread_fn, NULL, "hwlatd");
kernel/trace/trace_hwlat.c
427
if (IS_ERR(kthread)) {
kernel/trace/trace_hwlat.c
443
set_cpus_allowed_ptr(kthread, current_mask);
kernel/trace/trace_hwlat.c
445
kdata->kthread = kthread;
kernel/trace/trace_hwlat.c
446
wake_up_process(kthread);
kernel/trace/trace_hwlat.c
458
struct task_struct *kthread;
kernel/trace/trace_hwlat.c
460
kthread = per_cpu(hwlat_per_cpu_data, cpu).kthread;
kernel/trace/trace_hwlat.c
461
if (kthread)
kernel/trace/trace_hwlat.c
462
kthread_stop(kthread);
kernel/trace/trace_hwlat.c
463
per_cpu(hwlat_per_cpu_data, cpu).kthread = NULL;
kernel/trace/trace_hwlat.c
487
struct task_struct *kthread;
kernel/trace/trace_hwlat.c
490
if (per_cpu(hwlat_per_cpu_data, cpu).kthread)
kernel/trace/trace_hwlat.c
493
kthread = kthread_run_on_cpu(kthread_fn, NULL, cpu, "hwlatd/%u");
kernel/trace/trace_hwlat.c
494
if (IS_ERR(kthread)) {
kernel/trace/trace_hwlat.c
499
per_cpu(hwlat_per_cpu_data, cpu).kthread = kthread;
kernel/trace/trace_hwlat.c
74
struct task_struct *kthread;
kernel/trace/trace_osnoise.c
1665
this_cpu_osn_var()->kthread = NULL;
kernel/trace/trace_osnoise.c
1802
wake_up_process(tlat->kthread);
kernel/trace/trace_osnoise.c
1808
wake_up_process(tlat->kthread);
kernel/trace/trace_osnoise.c
1881
tlat->kthread = current;
kernel/trace/trace_osnoise.c
1934
struct task_struct *kthread;
kernel/trace/trace_osnoise.c
1936
kthread = xchg_relaxed(&(per_cpu(per_cpu_osnoise_var, cpu).kthread), NULL);
kernel/trace/trace_osnoise.c
1937
if (kthread) {
kernel/trace/trace_osnoise.c
1940
kthread_stop(kthread);
kernel/trace/trace_osnoise.c
1947
kill_pid(kthread->thread_pid, SIGKILL, 1);
kernel/trace/trace_osnoise.c
1948
put_task_struct(kthread);
kernel/trace/trace_osnoise.c
1985
struct task_struct *kthread;
kernel/trace/trace_osnoise.c
1990
if (per_cpu(per_cpu_osnoise_var, cpu).kthread)
kernel/trace/trace_osnoise.c
2006
kthread = kthread_run_on_cpu(main, NULL, cpu, comm);
kernel/trace/trace_osnoise.c
2008
if (IS_ERR(kthread)) {
kernel/trace/trace_osnoise.c
2013
per_cpu(per_cpu_osnoise_var, cpu).kthread = kthread;
kernel/trace/trace_osnoise.c
2044
struct task_struct *kthread;
kernel/trace/trace_osnoise.c
2046
kthread = xchg_relaxed(&(per_cpu(per_cpu_osnoise_var, cpu).kthread), NULL);
kernel/trace/trace_osnoise.c
2047
if (!WARN_ON(!kthread))
kernel/trace/trace_osnoise.c
2048
kthread_stop(kthread);
kernel/trace/trace_osnoise.c
209
struct task_struct *kthread;
kernel/trace/trace_osnoise.c
242
struct task_struct *kthread;
kernel/trace/trace_osnoise.c
2423
osn_var->kthread = current;
kernel/trace/trace_osnoise.c
2516
tlat->kthread = current;
kernel/trace/trace_osnoise.c
2568
if (tlat_var->kthread)
kernel/trace/trace_osnoise.c
2578
if (osn_var->kthread) {
kernel/trace/trace_osnoise.c
2579
put_task_struct(osn_var->kthread);
kernel/trace/trace_osnoise.c
2580
osn_var->kthread = NULL;
kernel/trace/trace_osnoise.c
277
if (tlat_var->kthread)
net/batman-adv/tp_meter.c
903
struct task_struct *kthread;
net/batman-adv/tp_meter.c
908
kthread = kthread_create(batadv_tp_send, tp_vars, "kbatadv_tp_meter");
net/batman-adv/tp_meter.c
909
if (IS_ERR(kthread)) {
net/batman-adv/tp_meter.c
925
wake_up_process(kthread);
samples/bpf/xdp_sample_user.c
105
struct record kthread;
samples/bpf/xdp_sample_user.c
1409
&rec->kthread);
samples/bpf/xdp_sample_user.c
482
rec->kthread.cpu = alloc_record_per_cpu();
samples/bpf/xdp_sample_user.c
483
if (!rec->kthread.cpu) {
samples/bpf/xdp_sample_user.c
536
free(rec->kthread.cpu);
samples/bpf/xdp_sample_user.c
563
free(r->kthread.cpu);
samples/bpf/xdp_sample_user.c
754
rec = &stats_rec->kthread;
samples/bpf/xdp_sample_user.c
755
prev = &stats_prev->kthread;
samples/bpf/xdp_sample_user.c
792
rec = &stats_rec->kthread;
samples/bpf/xdp_sample_user.c
793
prev = &stats_prev->kthread;