Symbol: affinity
arch/alpha/include/asm/gct.h
22
gct6_handle affinity;
arch/alpha/kernel/sys_dp264.c
136
cpu_set_irq_affinity(unsigned int irq, cpumask_t affinity)
arch/alpha/kernel/sys_dp264.c
142
if (cpumask_test_cpu(cpu, &affinity))
arch/alpha/kernel/sys_dp264.c
151
dp264_set_affinity(struct irq_data *d, const struct cpumask *affinity,
arch/alpha/kernel/sys_dp264.c
155
cpu_set_irq_affinity(d->irq, *affinity);
arch/alpha/kernel/sys_dp264.c
163
clipper_set_affinity(struct irq_data *d, const struct cpumask *affinity,
arch/alpha/kernel/sys_dp264.c
167
cpu_set_irq_affinity(d->irq - 16, *affinity);
arch/alpha/kernel/sys_titan.c
135
titan_cpu_set_irq_affinity(unsigned int irq, cpumask_t affinity)
arch/alpha/kernel/sys_titan.c
140
if (cpumask_test_cpu(cpu, &affinity))
arch/alpha/kernel/sys_titan.c
149
titan_set_irq_affinity(struct irq_data *d, const struct cpumask *affinity,
arch/alpha/kernel/sys_titan.c
154
titan_cpu_set_irq_affinity(irq - 16, *affinity);
arch/arm/kernel/setup.c
626
u32 i, affinity;
arch/arm/kernel/setup.c
640
affinity = MPIDR_AFFINITY_LEVEL(mask, i);
arch/arm/kernel/setup.c
646
ls = fls(affinity);
arch/arm/kernel/setup.c
647
fs[i] = affinity ? ffs(affinity) - 1 : 0;
arch/arm64/kernel/setup.c
113
u32 i, affinity, fs[4], bits[4], ls;
arch/arm64/kernel/setup.c
127
affinity = MPIDR_AFFINITY_LEVEL(mask, i);
arch/arm64/kernel/setup.c
133
ls = fls(affinity);
arch/arm64/kernel/setup.c
134
fs[i] = affinity ? ffs(affinity) - 1 : 0;
arch/arm64/kvm/psci.c
55
unsigned long affinity)
arch/arm64/kvm/psci.c
57
return !(affinity & ~MPIDR_HWID_BITMASK);
arch/mips/sgi-ip27/ip27-irq.c
154
cpumask_copy(desc->irq_common_data.affinity, &hub->h_cpus);
arch/powerpc/platforms/cell/spufs/inode.c
383
int affinity;
arch/powerpc/platforms/cell/spufs/inode.c
406
affinity = flags & (SPU_CREATE_AFFINITY_MEM | SPU_CREATE_AFFINITY_SPU);
arch/powerpc/platforms/cell/spufs/inode.c
407
if (affinity) {
arch/powerpc/platforms/cell/spufs/inode.c
425
if (affinity) {
arch/powerpc/platforms/cell/spufs/inode.c
437
if (affinity)
arch/powerpc/sysdev/xive/common.c
595
const struct cpumask *affinity)
arch/powerpc/sysdev/xive/common.c
609
for_each_cpu_and(cpu, affinity, cpu_online_mask) {
arch/powerpc/sysdev/xive/common.c
626
return xive_find_target_in_mask(affinity, fuzz++);
arch/sparc/kernel/irq_64.c
349
static int irq_choose_cpu(unsigned int irq, const struct cpumask *affinity)
arch/sparc/kernel/irq_64.c
353
if (cpumask_equal(affinity, cpu_online_mask)) {
arch/sparc/kernel/irq_64.c
356
cpuid = cpumask_first_and(affinity, cpu_online_mask);
arch/sparc/kernel/irq_64.c
363
#define irq_choose_cpu(irq, affinity) \
arch/sparc/kernel/leon_kernel.c
107
static int irq_choose_cpu(const struct cpumask *affinity)
arch/sparc/kernel/leon_kernel.c
109
unsigned int cpu = cpumask_first_and(affinity, cpu_online_mask);
arch/sparc/kernel/leon_kernel.c
111
if (cpumask_subset(cpu_online_mask, affinity) || cpu >= nr_cpu_ids)
arch/sparc/kernel/leon_kernel.c
117
#define irq_choose_cpu(affinity) boot_cpu_id
drivers/acpi/irq.c
319
return info.affinity;
drivers/base/platform.c
183
const struct cpumask **affinity)
drivers/base/platform.c
252
if (ret > 0 && affinity)
drivers/base/platform.c
253
*affinity = get_irq_affinity(dev, num);
drivers/firmware/arm_sdei.c
548
void *arg, u64 flags, u64 affinity)
drivers/firmware/arm_sdei.c
552
flags, affinity, NULL);
drivers/gpu/drm/amd/amdgpu/amdgpu_umsch_mm.h
83
uint32_t affinity;
drivers/gpu/drm/amd/amdgpu/umsch_mm_v4_0.c
351
add_queue.affinity.u32All = input_ptr->affinity;
drivers/gpu/drm/amd/include/umsch_mm_4_0_api_def.h
292
union UMSCH_AFFINITY affinity;
drivers/gpu/drm/amd/include/umsch_mm_4_0_api_def.h
411
union UMSCH_AFFINITY affinity;
drivers/gpu/drm/panfrost/panfrost_job.c
137
u64 affinity;
drivers/gpu/drm/panfrost/panfrost_job.c
144
affinity = pfdev->features.shader_present;
drivers/gpu/drm/panfrost/panfrost_job.c
146
job_write(pfdev, JS_AFFINITY_NEXT_LO(js), lower_32_bits(affinity));
drivers/gpu/drm/panfrost/panfrost_job.c
147
job_write(pfdev, JS_AFFINITY_NEXT_HI(js), upper_32_bits(affinity));
drivers/hwtracing/coresight/coresight-trbe.c
1477
const struct cpumask *affinity;
drivers/hwtracing/coresight/coresight-trbe.c
1480
drvdata->irq = platform_get_irq_affinity(pdev, 0, &affinity);
drivers/hwtracing/coresight/coresight-trbe.c
1491
cpumask_copy(&drvdata->supported_cpus, affinity);
drivers/hwtracing/coresight/coresight-trbe.c
1498
affinity, drvdata->handle);
drivers/infiniband/hw/hfi1/affinity.c
1042
mutex_lock(&affinity->lock);
drivers/infiniband/hw/hfi1/affinity.c
1072
for (i = 0; i < affinity->num_core_siblings; i++) {
drivers/infiniband/hw/hfi1/affinity.c
1073
find_hw_thread_mask(i, hw_thread_mask, affinity);
drivers/infiniband/hw/hfi1/affinity.c
1143
mutex_unlock(&affinity->lock);
drivers/infiniband/hw/hfi1/affinity.c
1159
struct hfi1_affinity_node_list *affinity = &node_affinity;
drivers/infiniband/hw/hfi1/affinity.c
1160
struct cpu_mask_set *set = &affinity->proc;
drivers/infiniband/hw/hfi1/affinity.c
1165
mutex_lock(&affinity->lock);
drivers/infiniband/hw/hfi1/affinity.c
1168
mutex_unlock(&affinity->lock);
drivers/infiniband/hw/hfi1/affinity.c
955
struct hfi1_affinity_node_list *affinity)
drivers/infiniband/hw/hfi1/affinity.c
960
cpumask_copy(hw_thread_mask, &affinity->proc.mask);
drivers/infiniband/hw/hfi1/affinity.c
962
if (affinity->num_core_siblings == 0)
drivers/infiniband/hw/hfi1/affinity.c
965
num_cores = rounddown(node_affinity.num_online_cpus / affinity->num_core_siblings,
drivers/infiniband/hw/hfi1/affinity.c
983
struct hfi1_affinity_node_list *affinity = &node_affinity;
drivers/infiniband/hw/hfi1/affinity.c
984
struct cpu_mask_set *set = &affinity->proc;
drivers/iommu/hyperv-iommu.c
198
const struct cpumask *affinity;
drivers/iommu/hyperv-iommu.c
205
affinity = irq_data_get_effective_affinity_mask(irq_data);
drivers/iommu/hyperv-iommu.c
206
cpu = cpumask_first_and(affinity, cpu_online_mask);
drivers/irqchip/irq-apple-aic.c
645
info->affinity = NULL;
drivers/irqchip/irq-apple-aic.c
660
info->affinity = mask;
drivers/irqchip/irq-bcm7038-l1.c
173
__bcm7038_l1_unmask(d, intc->affinity[d->hwirq]);
drivers/irqchip/irq-bcm7038-l1.c
183
__bcm7038_l1_mask(d, intc->affinity[d->hwirq]);
drivers/irqchip/irq-bcm7038-l1.c
202
was_disabled = !!(intc->cpus[intc->affinity[hw]]->mask_cache[word] &
drivers/irqchip/irq-bcm7038-l1.c
204
__bcm7038_l1_mask(d, intc->affinity[hw]);
drivers/irqchip/irq-bcm7038-l1.c
205
intc->affinity[hw] = first_cpu;
drivers/irqchip/irq-bcm7038-l1.c
47
u8 affinity[MAX_WORDS * IRQS_PER_WORD];
drivers/irqchip/irq-gic-v3.c
1718
info->affinity = NULL;
drivers/irqchip/irq-gic-v3.c
1753
info->affinity = mask;
drivers/irqchip/irq-gic-v3.c
926
u64 affinity;
drivers/irqchip/irq-gic-v3.c
976
affinity = gic_cpu_to_affinity(smp_processor_id());
drivers/irqchip/irq-gic-v3.c
978
gic_write_irouter(affinity, base + GICD_IROUTER + i * 8);
drivers/irqchip/irq-gic-v3.c
981
gic_write_irouter(affinity, base + GICD_IROUTERnE + i * 8);
drivers/irqchip/irq-loongson-eiointc.c
145
static int eiointc_set_irq_affinity(struct irq_data *d, const struct cpumask *affinity, bool force)
drivers/irqchip/irq-loongson-eiointc.c
154
cpu = cpumask_first_and_and(&priv->cpuspan_map, affinity, cpu_online_mask);
drivers/net/ethernet/mellanox/mlx5/core/lag/port_sel.c
82
u8 affinity;
drivers/net/ethernet/mellanox/mlx5/core/lag/port_sel.c
85
affinity = ports[idx];
drivers/net/ethernet/mellanox/mlx5/core/lag/port_sel.c
87
dest.vport.vhca_id = MLX5_CAP_GEN(ldev->pf[affinity - 1].dev,
drivers/of/irq.c
569
return info.affinity;
drivers/pci/controller/pci-hyperv.c
1803
static int hv_compose_msi_req_get_cpu(const struct cpumask *affinity)
drivers/pci/controller/pci-hyperv.c
1805
return cpumask_first_and(affinity, cpu_online_mask);
drivers/pci/msi/api.c
354
if (!desc->affinity)
drivers/pci/msi/api.c
362
return &desc->affinity[idx].mask;
drivers/pci/msi/msi.c
307
desc.affinity = masks;
drivers/pci/msi/msi.c
635
desc.affinity = masks ? curmsk : NULL;
drivers/perf/apple_m1_cpu_pmu.c
500
unsigned long affinity = m1_pmu_event_affinity[evtype];
drivers/perf/apple_m1_cpu_pmu.c
511
for_each_set_bit(idx, &affinity, M1_PMU_NR_COUNTERS) {
drivers/perf/arm_pmu.c
29
static int armpmu_count_irq_users(const struct cpumask *affinity,
drivers/perf/arm_pmu.c
587
static int armpmu_count_irq_users(const struct cpumask *affinity, const int irq)
drivers/perf/arm_pmu.c
591
for_each_cpu(cpu, affinity) {
drivers/perf/arm_pmu.c
600
armpmu_find_irq_ops(const struct cpumask *affinity, int irq)
drivers/perf/arm_pmu.c
605
for_each_cpu(cpu, affinity) {
drivers/perf/arm_pmu.c
634
const struct cpumask *affinity = *armpmu ? &(*armpmu)->supported_cpus :
drivers/perf/arm_pmu.c
668
} else if (armpmu_count_irq_users(affinity, irq) == 0) {
drivers/perf/arm_pmu.c
669
err = request_percpu_nmi(irq, handler, "arm-pmu", affinity, pcpu_armpmu);
drivers/perf/arm_pmu.c
674
affinity, pcpu_armpmu);
drivers/perf/arm_pmu.c
682
irq_ops = armpmu_find_irq_ops(affinity, irq);
drivers/perf/arm_pmu_platform.c
117
const struct cpumask *affinity;
drivers/perf/arm_pmu_platform.c
120
irq = platform_get_irq_affinity(pdev, 0, &affinity);
drivers/perf/arm_pmu_platform.c
122
return pmu_parse_percpu_irq(pmu, irq, affinity);
drivers/perf/arm_pmu_platform.c
46
const struct cpumask *affinity)
drivers/perf/arm_pmu_platform.c
51
cpumask_copy(&pmu->supported_cpus, affinity);
drivers/perf/arm_spe_pmu.c
1333
const struct cpumask *affinity;
drivers/perf/arm_spe_pmu.c
1336
irq = platform_get_irq_affinity(pdev, 0, &affinity);
drivers/perf/arm_spe_pmu.c
1345
cpumask_copy(&spe_pmu->supported_cpus, affinity);
drivers/resctrl/mpam_devices.c
449
cpumask_t *affinity)
drivers/resctrl/mpam_devices.c
451
return acpi_pptt_get_cpumask_from_cache_id(cache_id, affinity);
drivers/resctrl/mpam_devices.c
458
static void get_cpumask_from_node_id(u32 node_id, cpumask_t *affinity)
drivers/resctrl/mpam_devices.c
464
cpumask_set_cpu(cpu, affinity);
drivers/resctrl/mpam_devices.c
468
static int mpam_ris_get_affinity(struct mpam_msc *msc, cpumask_t *affinity,
drivers/resctrl/mpam_devices.c
478
affinity);
drivers/resctrl/mpam_devices.c
485
if (cpumask_empty(affinity))
drivers/resctrl/mpam_devices.c
490
get_cpumask_from_node_id(comp->comp_id, affinity);
drivers/resctrl/mpam_devices.c
497
cpumask_and(affinity, affinity, &msc->accessibility);
drivers/resctrl/mpam_devices.c
545
err = mpam_ris_get_affinity(msc, &ris->affinity, type, class, comp);
drivers/resctrl/mpam_devices.c
557
cpumask_or(&comp->affinity, &comp->affinity, &ris->affinity);
drivers/resctrl/mpam_devices.c
558
cpumask_or(&class->affinity, &class->affinity, &ris->affinity);
drivers/resctrl/mpam_devices.c
578
cpumask_andnot(&class->affinity, &class->affinity, &ris->affinity);
drivers/resctrl/mpam_devices.c
579
cpumask_andnot(&comp->affinity, &comp->affinity, &ris->affinity);
drivers/resctrl/mpam_internal.h
245
cpumask_t affinity;
drivers/resctrl/mpam_internal.h
282
cpumask_t affinity;
drivers/resctrl/mpam_internal.h
323
cpumask_t affinity;
drivers/resctrl/mpam_internal.h
392
cpumask_t *affinity);
include/linux/cpu_rmap.h
40
const struct cpumask *affinity);
include/linux/interrupt.h
129
const struct cpumask *affinity;
include/linux/interrupt.h
185
const cpumask_t *affinity, void __percpu *percpu_dev_id);
include/linux/interrupt.h
201
const struct cpumask *affinity, void __percpu *dev_id);
include/linux/interrupt.h
365
struct irq_affinity_desc *affinity);
include/linux/interrupt.h
414
struct irq_affinity_desc *affinity)
include/linux/irq.h
153
cpumask_var_t affinity;
include/linux/irq.h
884
return d->common->affinity;
include/linux/irq.h
894
cpumask_copy(d->common->affinity, m);
include/linux/irq.h
940
const struct irq_affinity_desc *affinity);
include/linux/irq.h
944
const struct irq_affinity_desc *affinity);
include/linux/irqdomain.h
375
const struct irq_affinity_desc *affinity);
include/linux/irqdomain.h
482
const struct irq_affinity_desc *affinity);
include/linux/irqdomain.h
59
const struct cpumask *affinity;
include/linux/irqdomain.h
610
const struct irq_affinity_desc *affinity);
include/linux/msi.h
189
struct irq_affinity_desc *affinity;
include/ras/ras_event.h
195
__field(u8, affinity)
include/ras/ras_event.h
208
__entry->affinity = proc->affinity_level;
include/ras/ras_event.h
210
__entry->affinity = ~0;
include/ras/ras_event.h
238
__entry->affinity, __entry->mpidr, __entry->midr,
kernel/irq/chip.c
915
if (cpumask_test_cpu(cpu, action->affinity))
kernel/irq/cpuhotplug.c
105
affinity = irq_desc_get_pending_mask(desc);
kernel/irq/cpuhotplug.c
107
affinity = irq_data_get_affinity_mask(d);
kernel/irq/cpuhotplug.c
113
if (!cpumask_intersects(affinity, cpu_online_mask)) {
kernel/irq/cpuhotplug.c
123
affinity = cpu_online_mask;
kernel/irq/cpuhotplug.c
132
err = irq_do_set_affinity(d, affinity, false);
kernel/irq/cpuhotplug.c
139
if (err == -ENOSPC && !irqd_affinity_is_managed(d) && affinity != cpu_online_mask) {
kernel/irq/cpuhotplug.c
141
d->irq, cpumask_pr_args(affinity));
kernel/irq/cpuhotplug.c
143
affinity = cpu_online_mask;
kernel/irq/cpuhotplug.c
146
err = irq_do_set_affinity(d, affinity, false);
kernel/irq/cpuhotplug.c
209
const struct cpumask *affinity = irq_data_get_affinity_mask(data);
kernel/irq/cpuhotplug.c
212
!irq_data_get_irq_chip(data) || !cpumask_test_cpu(cpu, affinity))
kernel/irq/cpuhotplug.c
226
irq_set_affinity_locked(data, affinity, false);
kernel/irq/cpuhotplug.c
58
const struct cpumask *affinity;
kernel/irq/devres.c
227
const struct irq_affinity_desc *affinity)
kernel/irq/devres.c
236
base = __irq_alloc_descs(irq, from, cnt, node, owner, affinity);
kernel/irq/ipi.c
94
cpumask_copy(data->common->affinity, dest);
kernel/irq/irq_test.c
120
struct irq_affinity_desc affinity = {
kernel/irq/irq_test.c
128
virq = irq_test_setup_fake_irq(test, &affinity);
kernel/irq/irq_test.c
172
struct irq_affinity_desc affinity = {
kernel/irq/irq_test.c
185
cpumask_copy(&affinity.mask, cpumask_of(1));
kernel/irq/irq_test.c
187
virq = irq_test_setup_fake_irq(test, &affinity);
kernel/irq/irqdesc.c
106
free_cpumask_var(desc->irq_common_data.affinity);
kernel/irq/irqdesc.c
116
desc_smp_init(struct irq_desc *desc, int node, const struct cpumask *affinity) { }
kernel/irq/irqdesc.c
121
const struct cpumask *affinity, struct module *owner)
kernel/irq/irqdesc.c
140
desc_smp_init(desc, node, affinity);
kernel/irq/irqdesc.c
213
const struct cpumask *affinity,
kernel/irq/irqdesc.c
229
desc_set_defaults(irq, desc, node, affinity, owner);
kernel/irq/irqdesc.c
434
const struct cpumask *affinity,
kernel/irq/irqdesc.c
444
ret = init_desc(desc, irq, node, flags, affinity, owner);
kernel/irq/irqdesc.c
498
const struct irq_affinity_desc *affinity,
kernel/irq/irqdesc.c
505
if (affinity) {
kernel/irq/irqdesc.c
507
if (cpumask_empty(&affinity[i].mask))
kernel/irq/irqdesc.c
516
if (affinity) {
kernel/irq/irqdesc.c
517
if (affinity->is_managed) {
kernel/irq/irqdesc.c
522
mask = &affinity->mask;
kernel/irq/irqdesc.c
524
affinity++;
kernel/irq/irqdesc.c
57
if (!zalloc_cpumask_var_node(&desc->irq_common_data.affinity,
kernel/irq/irqdesc.c
637
const struct irq_affinity_desc *affinity,
kernel/irq/irqdesc.c
64
free_cpumask_var(desc->irq_common_data.affinity);
kernel/irq/irqdesc.c
74
free_cpumask_var(desc->irq_common_data.affinity);
kernel/irq/irqdesc.c
86
static void desc_smp_init(struct irq_desc *desc, int node, const struct cpumask *affinity)
kernel/irq/irqdesc.c
88
if (!affinity)
kernel/irq/irqdesc.c
888
struct module *owner, const struct irq_affinity_desc *affinity)
kernel/irq/irqdesc.c
89
affinity = irq_default_affinity;
kernel/irq/irqdesc.c
90
cpumask_copy(desc->irq_common_data.affinity, affinity);
kernel/irq/irqdesc.c
918
return alloc_descs(start, cnt, node, affinity, owner);
kernel/irq/irqdomain.c
1284
int node, const struct irq_affinity_desc *affinity)
kernel/irq/irqdomain.c
1290
affinity);
kernel/irq/irqdomain.c
1296
affinity);
kernel/irq/irqdomain.c
1299
affinity);
kernel/irq/irqdomain.c
1629
bool realloc, const struct irq_affinity_desc *affinity)
kernel/irq/irqdomain.c
1637
affinity);
kernel/irq/irqdomain.c
1697
bool realloc, const struct irq_affinity_desc *affinity)
kernel/irq/irqdomain.c
1709
realloc, affinity);
kernel/irq/irqdomain.c
2073
bool realloc, const struct irq_affinity_desc *affinity)
kernel/irq/irqdomain.c
30
bool realloc, const struct irq_affinity_desc *affinity);
kernel/irq/irqdomain.c
783
const struct irq_affinity_desc *affinity)
kernel/irq/irqdomain.c
792
affinity);
kernel/irq/irqdomain.c
822
const struct irq_affinity_desc *affinity)
kernel/irq/irqdomain.c
843
virq = irq_create_mapping_affinity_locked(domain, hwirq, affinity);
kernel/irq/manage.c
1446
if (cpumask_intersects(old->affinity, new->affinity) ||
kernel/irq/manage.c
2421
if (cpumask_intersects(desc->percpu_enabled, action->affinity)) {
kernel/irq/manage.c
2423
cpumask_first_and(desc->percpu_enabled, action->affinity));
kernel/irq/manage.c
2481
const char *devname, const cpumask_t *affinity,
kernel/irq/manage.c
2486
if (!affinity)
kernel/irq/manage.c
2487
affinity = cpu_possible_mask;
kernel/irq/manage.c
2497
action->affinity = affinity;
kernel/irq/manage.c
2504
if (!cpumask_equal(affinity, cpu_possible_mask))
kernel/irq/manage.c
2527
const cpumask_t *affinity, void __percpu *dev_id)
kernel/irq/manage.c
2541
action = create_percpu_irqaction(handler, 0, devname, affinity, dev_id);
kernel/irq/manage.c
2585
const struct cpumask *affinity, void __percpu *dev_id)
kernel/irq/manage.c
2604
(!affinity || cpumask_equal(affinity, cpu_possible_mask)))
kernel/irq/manage.c
2608
name, affinity, dev_id);
kernel/irq/manage.c
293
cpumask_copy(desc->irq_common_data.affinity, mask);
kernel/irq/manage.c
355
cpumask_copy(desc->irq_common_data.affinity, mask);
kernel/irq/manage.c
419
int irq_update_affinity_desc(unsigned int irq, struct irq_affinity_desc *affinity)
kernel/irq/manage.c
447
if (affinity->is_managed) {
kernel/irq/manage.c
452
cpumask_copy(desc->irq_common_data.affinity, &affinity->mask);
kernel/irq/manage.c
532
cpumask_copy(cpumask, desc->irq_common_data.affinity);
kernel/irq/manage.c
611
if (cpumask_intersects(desc->irq_common_data.affinity,
kernel/irq/manage.c
613
set = desc->irq_common_data.affinity;
kernel/irq/msi.c
1338
desc->affinity);
kernel/irq/msi.c
157
desc = msi_alloc_desc(dev, init_desc->nvec_used, init_desc->affinity);
kernel/irq/msi.c
77
const struct irq_affinity_desc *affinity)
kernel/irq/msi.c
86
if (affinity) {
kernel/irq/msi.c
87
desc->affinity = kmemdup_array(affinity, nvec, sizeof(*desc->affinity), GFP_KERNEL);
kernel/irq/msi.c
88
if (!desc->affinity) {
kernel/irq/msi.c
98
kfree(desc->affinity);
kernel/irq/proc.c
57
mask = desc->irq_common_data.affinity;
kernel/kthread.c
353
cpumask_var_t affinity;
kernel/kthread.c
358
if (!zalloc_cpumask_var(&affinity, GFP_KERNEL)) {
kernel/kthread.c
373
kthread_fetch_affinity(kthread, affinity);
kernel/kthread.c
374
set_cpus_allowed_ptr(current, affinity);
kernel/kthread.c
377
free_cpumask_var(affinity);
kernel/kthread.c
842
cpumask_var_t affinity;
kernel/kthread.c
852
if (!zalloc_cpumask_var(&affinity, GFP_KERNEL))
kernel/kthread.c
865
kthread_fetch_affinity(kthread, affinity);
kernel/kthread.c
868
set_cpus_allowed_force(p, affinity);
kernel/kthread.c
872
free_cpumask_var(affinity);
kernel/kthread.c
880
cpumask_var_t affinity;
kernel/kthread.c
889
if (!zalloc_cpumask_var(&affinity, GFP_KERNEL))
kernel/kthread.c
914
kthread_fetch_affinity(k, affinity);
kernel/kthread.c
915
set_cpus_allowed_ptr(k->task, affinity);
kernel/kthread.c
919
free_cpumask_var(affinity);
kernel/rcu/tree.c
4163
cpumask_var_t affinity;
kernel/rcu/tree.c
4166
if (!zalloc_cpumask_var(&affinity, GFP_KERNEL))
kernel/rcu/tree.c
4170
cpumask_set_cpu(cpu, affinity);
kernel/rcu/tree.c
4172
kthread_affine_preferred(t, affinity);
kernel/rcu/tree.c
4174
free_cpumask_var(affinity);
lib/cpu_rmap.c
168
const struct cpumask *affinity)
lib/cpu_rmap.c
191
for_each_cpu(cpu, affinity) {
net/ceph/osdmap.c
1230
u32 *affinity;
net/ceph/osdmap.c
1232
affinity = kvmalloc(array_size(max, sizeof(*affinity)),
net/ceph/osdmap.c
1234
if (!affinity)
net/ceph/osdmap.c
1237
memcpy(affinity, map->osd_primary_affinity,
net/ceph/osdmap.c
1238
to_copy * sizeof(*affinity));
net/ceph/osdmap.c
1241
map->osd_primary_affinity = affinity;
tools/perf/builtin-record.c
100
struct mmap_cpu_mask affinity;
tools/perf/builtin-record.c
1290
if (opts->affinity != PERF_AFFINITY_SYS)
tools/perf/builtin-record.c
1296
opts->nr_cblocks, opts->affinity,
tools/perf/builtin-record.c
1578
if (rec->opts.affinity != PERF_AFFINITY_SYS &&
tools/perf/builtin-record.c
1579
!bitmap_equal(thread->mask->affinity.bits, map->affinity_mask.bits,
tools/perf/builtin-record.c
1580
thread->mask->affinity.nbits)) {
tools/perf/builtin-record.c
1581
bitmap_zero(thread->mask->affinity.bits, thread->mask->affinity.nbits);
tools/perf/builtin-record.c
1582
bitmap_or(thread->mask->affinity.bits, thread->mask->affinity.bits,
tools/perf/builtin-record.c
1583
map->affinity_mask.bits, thread->mask->affinity.nbits);
tools/perf/builtin-record.c
1584
sched_setaffinity(0, MMAP_CPU_MASK_BYTES(&thread->mask->affinity),
tools/perf/builtin-record.c
1585
(cpu_set_t *)thread->mask->affinity.bits);
tools/perf/builtin-record.c
1588
mmap_cpu_mask__scnprintf(&thread->mask->affinity, "affinity");
tools/perf/builtin-record.c
2351
MMAP_CPU_MASK_BYTES(&(thread_data[t].mask->affinity)),
tools/perf/builtin-record.c
2352
(cpu_set_t *)(thread_data[t].mask->affinity.bits));
tools/perf/builtin-record.c
2371
sched_setaffinity(0, MMAP_CPU_MASK_BYTES(&thread->mask->affinity),
tools/perf/builtin-record.c
2372
(cpu_set_t *)thread->mask->affinity.bits);
tools/perf/builtin-record.c
3093
opts->affinity = PERF_AFFINITY_NODE;
tools/perf/builtin-record.c
3095
opts->affinity = PERF_AFFINITY_CPU;
tools/perf/builtin-record.c
3122
mask->affinity.bits = NULL;
tools/perf/builtin-record.c
3126
ret = record__mmap_cpu_mask_alloc(&mask->affinity, nr_bits);
tools/perf/builtin-record.c
3138
record__mmap_cpu_mask_free(&mask->affinity);
tools/perf/builtin-record.c
3781
__set_bit(perf_cpu_map__cpu(cpus, t).cpu, rec->thread_masks[t].affinity.bits);
tools/perf/builtin-record.c
3786
mmap_cpu_mask__scnprintf(&rec->thread_masks[t].affinity, "affinity");
tools/perf/builtin-record.c
3832
ret = record__mmap_cpu_mask_init_spec(&thread_mask.affinity, affinity_spec[s]);
tools/perf/builtin-record.c
3845
if (!bitmap_and(thread_mask.affinity.bits, thread_mask.affinity.bits,
tools/perf/builtin-record.c
3846
cpus_mask.bits, thread_mask.affinity.nbits)) {
tools/perf/builtin-record.c
3859
if (bitmap_intersects(thread_mask.affinity.bits, full_mask.affinity.bits,
tools/perf/builtin-record.c
3860
thread_mask.affinity.nbits)) {
tools/perf/builtin-record.c
3868
bitmap_or(full_mask.affinity.bits, full_mask.affinity.bits,
tools/perf/builtin-record.c
3869
thread_mask.affinity.bits, full_mask.maps.nbits);
tools/perf/builtin-record.c
3883
mmap_cpu_mask__scnprintf(&rec->thread_masks[t].affinity, "affinity");
tools/perf/builtin-record.c
4113
rec->opts.affinity = PERF_AFFINITY_SYS;
tools/perf/builtin-record.c
4194
if (rec->opts.affinity != PERF_AFFINITY_SYS) {
tools/perf/builtin-record.c
4415
pr_debug("affinity: %s\n", affinity_tags[rec->opts.affinity]);
tools/perf/builtin-stat.c
2429
bool affinity = true, affinity_set = false;
tools/perf/builtin-stat.c
2558
OPT_BOOLEAN_SET(0, "affinity", &affinity, &affinity_set,
tools/perf/builtin-stat.c
2618
evsel_list->no_affinity = !affinity;
tools/perf/util/affinity.c
25
int affinity__setup(struct affinity *a)
tools/perf/util/affinity.c
49
void affinity__set(struct affinity *a, int cpu)
tools/perf/util/affinity.c
72
static void __affinity__cleanup(struct affinity *a)
tools/perf/util/affinity.c
82
void affinity__cleanup(struct affinity *a)
tools/perf/util/affinity.h
14
void affinity__cleanup(struct affinity *a);
tools/perf/util/affinity.h
15
void affinity__set(struct affinity *a, int cpu);
tools/perf/util/affinity.h
16
int affinity__setup(struct affinity *a);
tools/perf/util/evlist.c
1013
bool auxtrace_overwrite, int nr_cblocks, int affinity, int flush,
tools/perf/util/evlist.c
1023
.affinity = affinity,
tools/perf/util/evlist.c
430
.affinity = NULL,
tools/perf/util/evlist.c
441
itr->affinity = &itr->saved_affinity;
tools/perf/util/evlist.c
445
if (itr->affinity)
tools/perf/util/evlist.c
446
affinity__set(itr->affinity, itr->cpu.cpu);
tools/perf/util/evlist.c
458
if (!itr->affinity)
tools/perf/util/evlist.c
461
affinity__cleanup(itr->affinity);
tools/perf/util/evlist.c
462
itr->affinity = NULL;
tools/perf/util/evlist.c
481
if (evlist_cpu_itr->affinity)
tools/perf/util/evlist.c
482
affinity__set(evlist_cpu_itr->affinity, evlist_cpu_itr->cpu.cpu);
tools/perf/util/evlist.h
200
int affinity, int flush, int comp_level);
tools/perf/util/evlist.h
367
struct affinity *affinity;
tools/perf/util/evlist.h
369
struct affinity saved_affinity;
tools/perf/util/mmap.c
106
if (affinity != PERF_AFFINITY_SYS && cpu__max_node() > 1) {
tools/perf/util/mmap.c
142
struct perf_cpu cpu __maybe_unused, int affinity __maybe_unused)
tools/perf/util/mmap.c
176
ret = perf_mmap__aio_bind(map, i, map->core.cpu, mp->affinity);
tools/perf/util/mmap.c
269
if (mp->affinity == PERF_AFFINITY_NODE && cpu__max_node() > 1)
tools/perf/util/mmap.c
271
else if (mp->affinity == PERF_AFFINITY_CPU)
tools/perf/util/mmap.c
285
if (mp->affinity != PERF_AFFINITY_SYS &&
tools/perf/util/mmap.c
98
static int perf_mmap__aio_bind(struct mmap *map, int idx, struct perf_cpu cpu, int affinity)
tools/perf/util/mmap.h
48
int nr_cblocks, affinity, flush, comp_level;
tools/perf/util/record.h
75
int affinity;
tools/testing/selftests/bpf/bench.c
19
.affinity = false,
tools/testing/selftests/bpf/bench.c
349
env.affinity = true;
tools/testing/selftests/bpf/bench.c
358
env.affinity = true;
tools/testing/selftests/bpf/bench.c
366
env.affinity = true;
tools/testing/selftests/bpf/bench.c
705
if (env.affinity)
tools/testing/selftests/bpf/bench.c
728
if (env.affinity)
tools/testing/selftests/bpf/bench.h
27
bool affinity;
tools/testing/selftests/bpf/benchs/bench_bpf_hashmap_lookup.c
255
if (env.affinity)
tools/testing/selftests/bpf/benchs/bench_sockmap.c
239
!env.affinity)
tools/testing/selftests/rseq/basic_test.c
18
cpu_set_t affinity, test_affinity;
tools/testing/selftests/rseq/basic_test.c
21
sched_getaffinity(0, sizeof(affinity), &affinity);
tools/testing/selftests/rseq/basic_test.c
24
if (CPU_ISSET(i, &affinity)) {
tools/testing/selftests/rseq/basic_test.c
39
sched_setaffinity(0, sizeof(affinity), &affinity);
tools/testing/selftests/rseq/slice_test.c
125
cpu_set_t affinity;
tools/testing/selftests/rseq/slice_test.c
127
ASSERT_EQ(sched_getaffinity(0, sizeof(affinity), &affinity), 0);
tools/testing/selftests/rseq/slice_test.c
131
if (!CPU_ISSET(i, &affinity))
tools/testing/selftests/rseq/slice_test.c
134
CPU_ZERO(&affinity);
tools/testing/selftests/rseq/slice_test.c
135
CPU_SET(i, &affinity);
tools/testing/selftests/rseq/slice_test.c
136
ASSERT_EQ(sched_setaffinity(0, sizeof(affinity), &affinity), 0);
tools/testing/selftests/seccomp/seccomp_benchmark.c
207
affinity();
tools/testing/selftests/ublk/kublk.c
1044
if (info->affinity)
tools/testing/selftests/ublk/kublk.c
1158
tinfo[i].affinity = &affinity_buf[i];
tools/testing/selftests/ublk/kublk.c
333
cpu_set_t *affinity;
tools/testing/selftests/ublk/kublk.c
342
ret = ublk_ctrl_get_affinity(dev, &affinity);
tools/testing/selftests/ublk/kublk.c
355
if (affinity) {
tools/testing/selftests/ublk/kublk.c
360
ublk_print_cpu_set(&affinity[i], buf, sizeof(buf));
tools/testing/selftests/ublk/kublk.c
364
free(affinity);
tools/testing/selftests/ublk/kublk.c
959
cpu_set_t *affinity;
tools/testing/selftests/ublk/kublk.c
966
if (pthread_setaffinity_np(pthread_self(), sizeof(*info->affinity), info->affinity) < 0)