arch/x86/events/intel/lbr.c
364
struct x86_perf_task_context *task_ctx = ctx;
arch/x86/events/intel/lbr.c
366
u64 tos = task_ctx->tos;
arch/x86/events/intel/lbr.c
371
for (i = 0; i < task_ctx->valid_lbrs; i++) {
arch/x86/events/intel/lbr.c
373
wrlbr_all(&task_ctx->lbr[i], lbr_idx, need_info);
arch/x86/events/intel/lbr.c
387
wrmsrq(MSR_LBR_SELECT, task_ctx->lbr_sel);
arch/x86/events/intel/lbr.c
392
struct x86_perf_task_context_arch_lbr *task_ctx = ctx;
arch/x86/events/intel/lbr.c
393
struct lbr_entry *entries = task_ctx->entries;
arch/x86/events/intel/lbr.c
413
struct x86_perf_task_context_arch_lbr_xsave *task_ctx = ctx;
arch/x86/events/intel/lbr.c
415
xrstors(&task_ctx->xsave, XFEATURE_MASK_LBR);
arch/x86/events/intel/lbr.c
462
struct x86_perf_task_context *task_ctx = ctx;
arch/x86/events/intel/lbr.c
472
if (!rdlbr_all(&task_ctx->lbr[i], lbr_idx, need_info))
arch/x86/events/intel/lbr.c
475
task_ctx->valid_lbrs = i;
arch/x86/events/intel/lbr.c
476
task_ctx->tos = tos;
arch/x86/events/intel/lbr.c
479
rdmsrq(MSR_LBR_SELECT, task_ctx->lbr_sel);
arch/x86/events/intel/lbr.c
484
struct x86_perf_task_context_arch_lbr *task_ctx = ctx;
arch/x86/events/intel/lbr.c
485
struct lbr_entry *entries = task_ctx->entries;
arch/x86/events/intel/lbr.c
504
struct x86_perf_task_context_arch_lbr_xsave *task_ctx = ctx;
arch/x86/events/intel/lbr.c
506
xsaves(&task_ctx->xsave, XFEATURE_MASK_LBR);
arch/x86/events/intel/lbr.c
531
void *task_ctx;
arch/x86/events/intel/lbr.c
543
task_ctx = ctx_data ? ctx_data->data : NULL;
arch/x86/events/intel/lbr.c
544
if (task_ctx) {
arch/x86/events/intel/lbr.c
546
__intel_pmu_lbr_restore(task_ctx);
arch/x86/events/intel/lbr.c
548
__intel_pmu_lbr_save(task_ctx);
drivers/net/ethernet/qlogic/qed/qed_cxt.h
230
u32 tid, u8 ctx_type, void **task_ctx);
drivers/scsi/bnx2fc/bnx2fc.h
210
struct fcoe_task_ctx_entry **task_ctx;
drivers/scsi/bnx2fc/bnx2fc_els.c
775
interface->hba->task_ctx[task_idx];
drivers/scsi/bnx2fc/bnx2fc_hwi.c
1016
task_page = (struct fcoe_task_ctx_entry *)hba->task_ctx[task_idx];
drivers/scsi/bnx2fc/bnx2fc_hwi.c
1896
hba->task_ctx = kzalloc((task_ctx_arr_sz * sizeof(void *)),
drivers/scsi/bnx2fc/bnx2fc_hwi.c
1898
if (!hba->task_ctx) {
drivers/scsi/bnx2fc/bnx2fc_hwi.c
1918
hba->task_ctx[i] = dma_alloc_coherent(&hba->pcidev->dev,
drivers/scsi/bnx2fc/bnx2fc_hwi.c
1922
if (!hba->task_ctx[i]) {
drivers/scsi/bnx2fc/bnx2fc_hwi.c
1936
if (hba->task_ctx[i]) {
drivers/scsi/bnx2fc/bnx2fc_hwi.c
1939
hba->task_ctx[i], hba->task_ctx_dma[i]);
drivers/scsi/bnx2fc/bnx2fc_hwi.c
1940
hba->task_ctx[i] = NULL;
drivers/scsi/bnx2fc/bnx2fc_hwi.c
1947
kfree(hba->task_ctx);
drivers/scsi/bnx2fc/bnx2fc_hwi.c
1948
hba->task_ctx = NULL;
drivers/scsi/bnx2fc/bnx2fc_hwi.c
1970
if (hba->task_ctx) {
drivers/scsi/bnx2fc/bnx2fc_hwi.c
1972
if (hba->task_ctx[i]) {
drivers/scsi/bnx2fc/bnx2fc_hwi.c
1974
hba->task_ctx[i],
drivers/scsi/bnx2fc/bnx2fc_hwi.c
1976
hba->task_ctx[i] = NULL;
drivers/scsi/bnx2fc/bnx2fc_hwi.c
1979
kfree(hba->task_ctx);
drivers/scsi/bnx2fc/bnx2fc_hwi.c
1980
hba->task_ctx = NULL;
drivers/scsi/bnx2fc/bnx2fc_io.c
1028
interface->hba->task_ctx[task_idx];
drivers/scsi/bnx2fc/bnx2fc_io.c
2069
task_page = (struct fcoe_task_ctx_entry *) hba->task_ctx[task_idx];
drivers/scsi/bnx2fc/bnx2fc_io.c
758
interface->hba->task_ctx[task_idx];
drivers/scsi/bnx2fc/bnx2fc_io.c
894
interface->hba->task_ctx[task_idx];
drivers/scsi/bnx2fc/bnx2fc_io.c
970
interface->hba->task_ctx[task_idx];
drivers/scsi/qedf/qedf.h
516
struct fcoe_task_context *task_ctx, struct fcoe_wqe *sqe);
drivers/scsi/qedf/qedf_io.c
587
struct qedf_ioreq *io_req, struct fcoe_task_context *task_ctx,
drivers/scsi/qedf/qedf_io.c
604
io_req->task = task_ctx;
drivers/scsi/qedf/qedf_io.c
605
memset(task_ctx, 0, sizeof(struct fcoe_task_context));
drivers/scsi/qedf/qedf_io.c
623
io_req->task_params->context = task_ctx;
drivers/scsi/qedf/qedf_io.c
677
struct fcoe_task_context *task_ctx, struct fcoe_wqe *sqe)
drivers/scsi/qedf/qedf_io.c
695
memset(task_ctx, 0, sizeof(struct fcoe_task_context));
drivers/scsi/qedf/qedf_io.c
699
io_req->task = task_ctx;
drivers/scsi/qedf/qedf_io.c
702
io_req->task_params->context = task_ctx;
drivers/scsi/qedf/qedf_io.c
852
struct fcoe_task_context *task_ctx;
drivers/scsi/qedf/qedf_io.c
905
task_ctx = qedf_get_task_mem(&qedf->tasks, xid);
drivers/scsi/qedf/qedf_io.c
906
if (!task_ctx) {
drivers/scsi/qedf/qedf_io.c
915
qedf_init_task(fcport, lport, io_req, task_ctx, sqe);
drivers/scsi/qedi/qedi_fw.c
120
memset(task_ctx, '\0', sizeof(*task_ctx));
drivers/scsi/qedi/qedi_fw.c
264
struct iscsi_task_context *task_ctx;
drivers/scsi/qedi/qedi_fw.c
273
task_ctx = qedi_get_task_mem(&qedi->tasks, cmd->task_id);
drivers/scsi/qedi/qedi_fw.c
305
memset(task_ctx, '\0', sizeof(*task_ctx));
drivers/scsi/qedi/qedi_fw.c
88
struct iscsi_task_context *task_ctx;
drivers/scsi/qedi/qedi_fw.c
95
task_ctx = qedi_get_task_mem(&qedi->tasks, cmd->task_id);
include/linux/perf_event.h
1135
struct perf_event_context *task_ctx;
kernel/events/core.c
1059
guard(perf_ctx_lock)(cpuctx, cpuctx->task_ctx);
kernel/events/core.c
249
return this_cpu_ptr(&perf_cpu_context)->task_ctx;
kernel/events/core.c
2624
WARN_ON_ONCE(cpuctx->task_ctx != ctx);
kernel/events/core.c
2625
cpuctx->task_ctx = NULL;
kernel/events/core.c
286
struct perf_event_context *task_ctx = cpuctx->task_ctx;
kernel/events/core.c
291
perf_ctx_lock(cpuctx, task_ctx);
kernel/events/core.c
2949
if (!cpuctx->task_ctx)
kernel/events/core.c
2952
if (WARN_ON_ONCE(ctx != cpuctx->task_ctx))
kernel/events/core.c
2987
struct perf_event_context *task_ctx,
kernel/events/core.c
3005
if (task_ctx) {
kernel/events/core.c
3006
for_each_epc(epc, task_ctx, pmu, 0)
kernel/events/core.c
3009
task_ctx_sched_out(task_ctx, pmu, event_type);
kernel/events/core.c
3024
perf_event_sched_in(cpuctx, task_ctx, pmu, 0);
kernel/events/core.c
3029
if (task_ctx) {
kernel/events/core.c
3030
for_each_epc(epc, task_ctx, pmu, 0)
kernel/events/core.c
3038
struct perf_event_context *task_ctx = cpuctx->task_ctx;
kernel/events/core.c
3040
perf_ctx_lock(cpuctx, task_ctx);
kernel/events/core.c
3041
ctx_resched(cpuctx, task_ctx, pmu, EVENT_ALL|EVENT_CPU);
kernel/events/core.c
3042
perf_ctx_unlock(cpuctx, task_ctx);
kernel/events/core.c
3056
struct perf_event_context *task_ctx = cpuctx->task_ctx;
kernel/events/core.c
3063
task_ctx = ctx;
kernel/events/core.c
3079
WARN_ON_ONCE(reprogram && cpuctx->task_ctx && cpuctx->task_ctx != ctx);
kernel/events/core.c
3080
} else if (task_ctx) {
kernel/events/core.c
3081
raw_spin_lock(&task_ctx->lock);
kernel/events/core.c
3099
ctx_resched(cpuctx, task_ctx, event->pmu_ctx->pmu,
kernel/events/core.c
3106
perf_ctx_unlock(cpuctx, task_ctx);
kernel/events/core.c
314
WARN_ON_ONCE(task_ctx != ctx);
kernel/events/core.c
321
perf_ctx_unlock(cpuctx, task_ctx);
kernel/events/core.c
3237
struct perf_event_context *task_ctx;
kernel/events/core.c
3261
task_ctx = cpuctx->task_ctx;
kernel/events/core.c
3263
WARN_ON_ONCE(task_ctx != ctx);
kernel/events/core.c
3265
ctx_resched(cpuctx, task_ctx, event->pmu_ctx->pmu, get_event_type(event));
kernel/events/core.c
3592
WARN_ON_ONCE(cpuctx->task_ctx);
kernel/events/core.c
3627
WARN_ON_ONCE(cpuctx->task_ctx != ctx);
kernel/events/core.c
3629
cpuctx->task_ctx = NULL;
kernel/events/core.c
388
struct perf_event_context *task_ctx = NULL;
kernel/events/core.c
3906
perf_ctx_lock(cpuctx, cpuctx->task_ctx);
kernel/events/core.c
3912
perf_ctx_unlock(cpuctx, cpuctx->task_ctx);
kernel/events/core.c
3923
if (prev == next || cpuctx->task_ctx)
kernel/events/core.c
396
task_ctx = ctx;
kernel/events/core.c
399
perf_ctx_lock(cpuctx, task_ctx);
kernel/events/core.c
415
if (WARN_ON_ONCE(cpuctx->task_ctx != ctx))
kernel/events/core.c
4208
cpuctx->task_ctx = ctx;
kernel/events/core.c
4210
WARN_ON_ONCE(cpuctx->task_ctx != ctx);
kernel/events/core.c
424
perf_ctx_unlock(cpuctx, task_ctx);
kernel/events/core.c
4257
if (cpuctx->task_ctx == ctx) {
kernel/events/core.c
4292
perf_ctx_sched_task_cb(cpuctx->task_ctx, task, true);
kernel/events/core.c
4610
perf_ctx_lock(cpuctx, cpuctx->task_ctx);
kernel/events/core.c
4641
perf_ctx_unlock(cpuctx, cpuctx->task_ctx);
kernel/events/core.c
4825
if (ctx->task && cpuctx->task_ctx != ctx)
kernel/events/core.c
6421
guard(perf_ctx_lock)(cpuctx, cpuctx->task_ctx);
kernel/events/core.c
6428
if (cpuctx->task_ctx) {
kernel/events/core.c
6429
perf_ctx_disable(cpuctx->task_ctx, EVENT_GUEST);
kernel/events/core.c
6430
task_ctx_sched_out(cpuctx->task_ctx, NULL, EVENT_GUEST);
kernel/events/core.c
6434
if (cpuctx->task_ctx)
kernel/events/core.c
6435
perf_ctx_enable(cpuctx->task_ctx, EVENT_GUEST);
kernel/events/core.c
6447
guard(perf_ctx_lock)(cpuctx, cpuctx->task_ctx);
kernel/events/core.c
6453
if (cpuctx->task_ctx)
kernel/events/core.c
6454
perf_ctx_disable(cpuctx->task_ctx, EVENT_GUEST);
kernel/events/core.c
6456
perf_event_sched_in(cpuctx, cpuctx->task_ctx, NULL, EVENT_GUEST);
kernel/events/core.c
6458
if (cpuctx->task_ctx)
kernel/events/core.c
6459
perf_ctx_enable(cpuctx->task_ctx, EVENT_GUEST);
kernel/events/core.c
8934
struct perf_event_context *task_ctx)
kernel/events/core.c
8946
if (task_ctx) {
kernel/events/core.c
8947
perf_iterate_ctx(task_ctx, output, data, false);
kernel/events/core.c
9055
if (cpuctx->task_ctx)
kernel/events/core.c
9056
perf_iterate_ctx(cpuctx->task_ctx, __perf_event_output_stop,
kernel/events/core.c
9101
struct perf_event_context *task_ctx;
kernel/events/core.c
9165
struct perf_event_context *task_ctx,
kernel/events/core.c
9177
.task_ctx = task_ctx,
kernel/events/core.c
9194
task_ctx);
security/apparmor/domain.c
1226
struct aa_task_ctx *ctx = task_ctx(current);
security/apparmor/domain.c
1398
struct aa_task_ctx *ctx = task_ctx(current);
security/apparmor/domain.c
937
ctx = task_ctx(current);
security/apparmor/lsm.c
112
aa_free_task_ctx(task_ctx(task));
security/apparmor/lsm.c
118
struct aa_task_ctx *new = task_ctx(task);
security/apparmor/lsm.c
120
aa_dup_task_ctx(new, task_ctx(current));
security/apparmor/lsm.c
785
struct aa_task_ctx *ctx = task_ctx(current);
security/apparmor/lsm.c
827
struct aa_task_ctx *ctx = task_ctx(current);
security/apparmor/lsm.c
979
aa_clear_task_ctx_trans(task_ctx(current));
security/apparmor/task.c
100
struct aa_task_ctx *ctx = task_ctx(current);
security/apparmor/task.c
120
struct aa_task_ctx *ctx = task_ctx(current);
security/apparmor/task.c
160
struct aa_task_ctx *ctx = task_ctx(current);
security/apparmor/task.c
50
struct aa_task_ctx *ctx = task_ctx(current);
security/apparmor/task.c
76
aa_clear_task_ctx_trans(task_ctx(current));
tools/sched_ext/scx_flatcg.bpf.c
130
} task_ctx SEC(".maps");
tools/sched_ext/scx_flatcg.bpf.c
325
taskc = bpf_task_storage_get(&task_ctx, p, 0, 0);
tools/sched_ext/scx_flatcg.bpf.c
351
taskc = bpf_task_storage_get(&task_ctx, p, 0, 0);
tools/sched_ext/scx_flatcg.bpf.c
558
taskc = bpf_task_storage_get(&task_ctx, p, 0, 0);
tools/sched_ext/scx_flatcg.bpf.c
815
taskc = bpf_task_storage_get(&task_ctx, p, 0,
tools/sched_ext/scx_qmap.bpf.c
112
__type(value, struct task_ctx);
tools/sched_ext/scx_qmap.bpf.c
151
static struct task_ctx *lookup_task_ctx(struct task_struct *p)
tools/sched_ext/scx_qmap.bpf.c
153
struct task_ctx *tctx;
tools/sched_ext/scx_qmap.bpf.c
165
struct task_ctx *tctx;
tools/sched_ext/scx_qmap.bpf.c
199
struct task_ctx *tctx;
tools/sched_ext/scx_qmap.bpf.c
295
struct task_ctx *tctx;
tools/sched_ext/scx_qmap.bpf.c
319
struct task_ctx *tctx;
tools/sched_ext/scx_qmap.bpf.c
370
struct task_ctx *tctx;
tools/sched_ext/scx_qmap.bpf.c
415
struct task_ctx *tctx;
tools/sched_ext/scx_qmap.bpf.c
499
struct task_ctx *tctx;
tools/sched_ext/scx_qmap.bpf.c
638
struct task_ctx *taskc;
tools/sched_ext/scx_userland.bpf.c
148
struct task_ctx *tctx;
tools/sched_ext/scx_userland.bpf.c
208
struct task_ctx *tctx;
tools/sched_ext/scx_userland.bpf.c
92
__type(value, struct task_ctx);
tools/testing/selftests/sched_ext/select_cpu_dfl_nodispatch.bpf.c
27
__type(value, struct task_ctx);
tools/testing/selftests/sched_ext/select_cpu_dfl_nodispatch.bpf.c
37
struct task_ctx *tctx;
tools/testing/selftests/sched_ext/select_cpu_dfl_nodispatch.bpf.c
56
struct task_ctx *tctx;