arch/arm64/include/asm/cpufeature.h
356
bool (*matches)(const struct arm64_cpu_capabilities *caps, int scope);
arch/arm64/include/asm/cpufeature.h
424
int scope)
arch/arm64/include/asm/cpufeature.h
429
if (caps->matches(caps, scope))
arch/arm64/include/asm/cpufeature.h
645
static inline bool supports_csv2p3(int scope)
arch/arm64/include/asm/cpufeature.h
650
if (scope == SCOPE_LOCAL_CPU)
arch/arm64/include/asm/cpufeature.h
660
static inline bool supports_clearbhb(int scope)
arch/arm64/include/asm/cpufeature.h
664
if (scope == SCOPE_LOCAL_CPU)
arch/arm64/include/asm/spectre.h
85
bool has_spectre_v2(const struct arm64_cpu_capabilities *cap, int scope);
arch/arm64/include/asm/spectre.h
88
bool has_spectre_v3a(const struct arm64_cpu_capabilities *cap, int scope);
arch/arm64/include/asm/spectre.h
92
bool has_spectre_v4(const struct arm64_cpu_capabilities *cap, int scope);
arch/arm64/include/asm/spectre.h
99
bool is_spectre_bhb_affected(const struct arm64_cpu_capabilities *entry, int scope);
arch/arm64/kernel/cpu_errata.c
103
WARN_ON(scope != SCOPE_LOCAL_CPU || preemptible());
arch/arm64/kernel/cpu_errata.c
114
int scope)
arch/arm64/kernel/cpu_errata.c
120
WARN_ON(scope != SCOPE_LOCAL_CPU || preemptible());
arch/arm64/kernel/cpu_errata.c
162
need_arm_si_l1_workaround_4311569(const struct arm64_cpu_capabilities *entry, int scope)
arch/arm64/kernel/cpu_errata.c
190
int scope)
arch/arm64/kernel/cpu_errata.c
192
return is_affected_midr_range_list(entry, scope) && is_kernel_in_hyp_mode();
arch/arm64/kernel/cpu_errata.c
247
int scope)
arch/arm64/kernel/cpu_errata.c
251
if (!is_affected_midr_range_list(entry, scope) ||
arch/arm64/kernel/cpu_errata.c
265
int scope)
arch/arm64/kernel/cpu_errata.c
270
WARN_ON(scope != SCOPE_LOCAL_CPU || preemptible());
arch/arm64/kernel/cpu_errata.c
290
static bool has_impdef_pmuv3(const struct arm64_cpu_capabilities *entry, int scope)
arch/arm64/kernel/cpu_errata.c
72
is_affected_midr_range(const struct arm64_cpu_capabilities *entry, int scope)
arch/arm64/kernel/cpu_errata.c
77
WARN_ON(scope != SCOPE_LOCAL_CPU || preemptible());
arch/arm64/kernel/cpu_errata.c
92
int scope)
arch/arm64/kernel/cpu_errata.c
94
WARN_ON(scope != SCOPE_LOCAL_CPU || preemptible());
arch/arm64/kernel/cpu_errata.c
99
is_kryo_midr(const struct arm64_cpu_capabilities *entry, int scope)
arch/arm64/kernel/cpufeature.c
1591
has_always(const struct arm64_cpu_capabilities *entry, int scope)
arch/arm64/kernel/cpufeature.c
1624
read_scoped_sysreg(const struct arm64_cpu_capabilities *entry, int scope)
arch/arm64/kernel/cpufeature.c
1626
WARN_ON(scope == SCOPE_LOCAL_CPU && preemptible());
arch/arm64/kernel/cpufeature.c
1627
if (scope == SCOPE_SYSTEM)
arch/arm64/kernel/cpufeature.c
1634
has_user_cpuid_feature(const struct arm64_cpu_capabilities *entry, int scope)
arch/arm64/kernel/cpufeature.c
1638
u64 val = read_scoped_sysreg(entry, scope);
arch/arm64/kernel/cpufeature.c
1654
has_cpuid_feature(const struct arm64_cpu_capabilities *entry, int scope)
arch/arm64/kernel/cpufeature.c
1656
u64 val = read_scoped_sysreg(entry, scope);
arch/arm64/kernel/cpufeature.c
1709
static bool has_32bit_el0(const struct arm64_cpu_capabilities *entry, int scope)
arch/arm64/kernel/cpufeature.c
1711
if (!has_cpuid_feature(entry, scope))
arch/arm64/kernel/cpufeature.c
1714
if (scope == SCOPE_SYSTEM)
arch/arm64/kernel/cpufeature.c
1720
static bool has_useable_gicv3_cpuif(const struct arm64_cpu_capabilities *entry, int scope)
arch/arm64/kernel/cpufeature.c
1724
if (!has_cpuid_feature(entry, scope))
arch/arm64/kernel/cpufeature.c
1736
int scope)
arch/arm64/kernel/cpufeature.c
1740
if (scope == SCOPE_SYSTEM)
arch/arm64/kernel/cpufeature.c
1761
int scope)
arch/arm64/kernel/cpufeature.c
1765
if (scope == SCOPE_SYSTEM)
arch/arm64/kernel/cpufeature.c
1774
has_useable_cnp(const struct arm64_cpu_capabilities *entry, int scope)
arch/arm64/kernel/cpufeature.c
1787
return has_cpuid_feature(entry, scope);
arch/arm64/kernel/cpufeature.c
1794
int scope)
arch/arm64/kernel/cpufeature.c
1821
if (has_cpuid_feature(entry, scope))
arch/arm64/kernel/cpufeature.c
1869
static bool has_nv1(const struct arm64_cpu_capabilities *entry, int scope)
arch/arm64/kernel/cpufeature.c
1887
!(has_cpuid_feature(entry, scope) ||
arch/arm64/kernel/cpufeature.c
1910
static bool has_lpa2(const struct arm64_cpu_capabilities *entry, int scope)
arch/arm64/kernel/cpufeature.c
1918
static bool has_lpa2(const struct arm64_cpu_capabilities *entry, int scope)
arch/arm64/kernel/cpufeature.c
1925
static bool has_pmuv3(const struct arm64_cpu_capabilities *entry, int scope)
arch/arm64/kernel/cpufeature.c
2108
int scope)
arch/arm64/kernel/cpufeature.c
2113
if (!cpucap_multi_entry_cap_matches(cap, scope)) {
arch/arm64/kernel/cpufeature.c
2163
static bool has_bbml2_noabort(const struct arm64_cpu_capabilities *caps, int scope)
arch/arm64/kernel/cpufeature.c
2186
static bool has_rasv1p1(const struct arm64_cpu_capabilities *__unused, int scope)
arch/arm64/kernel/cpufeature.c
2200
return (has_cpuid_feature(&rasv1p1_caps[0], scope) ||
arch/arm64/kernel/cpufeature.c
2201
(has_cpuid_feature(&rasv1p1_caps[1], scope) &&
arch/arm64/kernel/cpufeature.c
2202
has_cpuid_feature(&rasv1p1_caps[2], scope)));
arch/arm64/kernel/cpufeature.c
2207
static bool has_address_auth_cpucap(const struct arm64_cpu_capabilities *entry, int scope)
arch/arm64/kernel/cpufeature.c
2212
WARN_ON(scope == SCOPE_SYSTEM);
arch/arm64/kernel/cpufeature.c
2225
if (scope & SCOPE_BOOT_CPU)
arch/arm64/kernel/cpufeature.c
2234
int scope)
arch/arm64/kernel/cpufeature.c
2236
bool api = has_address_auth_cpucap(cpucap_ptrs[ARM64_HAS_ADDRESS_AUTH_IMP_DEF], scope);
arch/arm64/kernel/cpufeature.c
2237
bool apa = has_address_auth_cpucap(cpucap_ptrs[ARM64_HAS_ADDRESS_AUTH_ARCH_QARMA5], scope);
arch/arm64/kernel/cpufeature.c
2238
bool apa3 = has_address_auth_cpucap(cpucap_ptrs[ARM64_HAS_ADDRESS_AUTH_ARCH_QARMA3], scope);
arch/arm64/kernel/cpufeature.c
2274
int scope)
arch/arm64/kernel/cpufeature.c
2288
int scope)
arch/arm64/kernel/cpufeature.c
2317
int scope)
arch/arm64/kernel/cpufeature.c
2492
test_has_mpam(const struct arm64_cpu_capabilities *entry, int scope)
arch/arm64/kernel/cpufeature.c
2494
if (!has_cpuid_feature(entry, scope))
arch/arm64/kernel/cpufeature.c
2514
test_has_mpam_hcr(const struct arm64_cpu_capabilities *entry, int scope)
arch/arm64/kernel/cpufeature.c
2522
test_has_gicv5_legacy(const struct arm64_cpu_capabilities *entry, int scope)
arch/arm64/kernel/cpufeature.c
3249
static bool has_sve_feature(const struct arm64_cpu_capabilities *cap, int scope)
arch/arm64/kernel/cpufeature.c
3251
return system_supports_sve() && has_user_cpuid_feature(cap, scope);
arch/arm64/kernel/cpufeature.c
3256
static bool has_sme_feature(const struct arm64_cpu_capabilities *cap, int scope)
arch/arm64/kernel/cpufeature.c
3258
return system_supports_sme() && has_user_cpuid_feature(cap, scope);
arch/arm64/kernel/cpufeature.c
3395
static bool compat_has_neon(const struct arm64_cpu_capabilities *cap, int scope)
arch/arm64/kernel/cpufeature.c
3404
WARN_ON(scope == SCOPE_LOCAL_CPU && preemptible());
arch/arm64/kernel/cpufeature.c
3405
if (scope == SCOPE_SYSTEM)
arch/arm64/kernel/proton-pack.c
197
bool has_spectre_v2(const struct arm64_cpu_capabilities *entry, int scope)
arch/arm64/kernel/proton-pack.c
199
WARN_ON(scope != SCOPE_LOCAL_CPU || preemptible());
arch/arm64/kernel/proton-pack.c
320
bool has_spectre_v3a(const struct arm64_cpu_capabilities *entry, int scope)
arch/arm64/kernel/proton-pack.c
328
WARN_ON(scope != SCOPE_LOCAL_CPU || preemptible());
arch/arm64/kernel/proton-pack.c
501
bool has_spectre_v4(const struct arm64_cpu_capabilities *cap, int scope)
arch/arm64/kernel/proton-pack.c
505
WARN_ON(scope != SCOPE_LOCAL_CPU || preemptible());
arch/arm64/kernel/proton-pack.c
838
static bool is_spectre_bhb_safe(int scope)
arch/arm64/kernel/proton-pack.c
854
if (scope != SCOPE_LOCAL_CPU)
arch/arm64/kernel/proton-pack.c
958
static bool supports_ecbhb(int scope)
arch/arm64/kernel/proton-pack.c
962
if (scope == SCOPE_LOCAL_CPU)
arch/arm64/kernel/proton-pack.c
974
int scope)
arch/arm64/kernel/proton-pack.c
976
WARN_ON(scope != SCOPE_LOCAL_CPU || preemptible());
arch/arm64/kernel/proton-pack.c
978
if (supports_csv2p3(scope))
arch/arm64/kernel/proton-pack.c
981
if (is_spectre_bhb_safe(scope))
arch/arm64/kernel/proton-pack.c
989
if (scope == SCOPE_LOCAL_CPU && !supports_clearbhb(SCOPE_LOCAL_CPU))
arch/arm64/kvm/nested.c
1000
struct s1e2_tlbi_scope *scope)
arch/arm64/kvm/nested.c
1015
scope->type = TLBI_ALL;
arch/arm64/kvm/nested.c
1041
scope->type = TLBI_VA;
arch/arm64/kvm/nested.c
1042
scope->size = ttl_to_size(FIELD_GET(TLBI_TTL_MASK, val));
arch/arm64/kvm/nested.c
1043
if (!scope->size)
arch/arm64/kvm/nested.c
1044
scope->size = SZ_1G;
arch/arm64/kvm/nested.c
1045
scope->va = tlbi_va_s1_to_va(val) & ~(scope->size - 1);
arch/arm64/kvm/nested.c
1046
scope->asid = FIELD_GET(TLBIR_ASID_MASK, val);
arch/arm64/kvm/nested.c
1054
scope->type = TLBI_ASID;
arch/arm64/kvm/nested.c
1055
scope->asid = FIELD_GET(TLBIR_ASID_MASK, val);
arch/arm64/kvm/nested.c
1069
scope->type = TLBI_VAA;
arch/arm64/kvm/nested.c
1070
scope->size = ttl_to_size(FIELD_GET(TLBI_TTL_MASK, val));
arch/arm64/kvm/nested.c
1071
if (!scope->size)
arch/arm64/kvm/nested.c
1072
scope->size = SZ_1G;
arch/arm64/kvm/nested.c
1073
scope->va = tlbi_va_s1_to_va(val) & ~(scope->size - 1);
arch/arm64/kvm/nested.c
1099
scope->type = TLBI_VA;
arch/arm64/kvm/nested.c
1100
scope->va = decode_range_tlbi(val, &scope->size, &scope->asid);
arch/arm64/kvm/nested.c
1114
scope->type = TLBI_VAA;
arch/arm64/kvm/nested.c
1115
scope->va = decode_range_tlbi(val, &scope->size, NULL);
arch/arm64/kvm/nested.c
1122
struct s1e2_tlbi_scope scope = {};
arch/arm64/kvm/nested.c
1124
compute_s1_tlbi_range(vcpu, inst, val, &scope);
arch/arm64/kvm/nested.c
1127
invalidate_vncr_va(vcpu->kvm, &scope);
arch/arm64/kvm/nested.c
950
struct s1e2_tlbi_scope *scope)
arch/arm64/kvm/nested.c
969
switch (scope->type) {
arch/arm64/kvm/nested.c
974
if (va_end <= scope->va ||
arch/arm64/kvm/nested.c
975
va_start >= (scope->va + scope->size))
arch/arm64/kvm/nested.c
977
if (vt->wr.nG && vt->wr.asid != scope->asid)
arch/arm64/kvm/nested.c
982
if (va_end <= scope->va ||
arch/arm64/kvm/nested.c
983
va_start >= (scope->va + scope->size))
arch/arm64/kvm/nested.c
988
if (!vt->wr.nG || vt->wr.asid != scope->asid)
arch/m68k/include/asm/syscalls.h
8
asmlinkage int sys_cacheflush(unsigned long addr, int scope, int cache,
arch/m68k/kernel/sys_m68k.c
228
cache_flush_060 (unsigned long addr, int scope, int cache, unsigned long len)
arch/m68k/kernel/sys_m68k.c
238
switch (scope)
arch/m68k/kernel/sys_m68k.c
378
sys_cacheflush (unsigned long addr, int scope, int cache, unsigned long len)
arch/m68k/kernel/sys_m68k.c
382
if (scope < FLUSH_SCOPE_LINE || scope > FLUSH_SCOPE_ALL ||
arch/m68k/kernel/sys_m68k.c
386
if (scope == FLUSH_SCOPE_ALL) {
arch/m68k/kernel/sys_m68k.c
411
if (scope == FLUSH_SCOPE_LINE && len < 256) {
arch/m68k/kernel/sys_m68k.c
444
if (len>=3*PAGE_SIZE && scope<FLUSH_SCOPE_PAGE)
arch/m68k/kernel/sys_m68k.c
445
scope=FLUSH_SCOPE_PAGE;
arch/m68k/kernel/sys_m68k.c
446
if (len>=10*PAGE_SIZE && scope<FLUSH_SCOPE_ALL)
arch/m68k/kernel/sys_m68k.c
447
scope=FLUSH_SCOPE_ALL;
arch/m68k/kernel/sys_m68k.c
449
ret = cache_flush_040 (addr, scope, cache, len);
arch/m68k/kernel/sys_m68k.c
451
ret = cache_flush_060 (addr, scope, cache, len);
arch/m68k/kernel/sys_m68k.c
537
sys_cacheflush (unsigned long addr, int scope, int cache, unsigned long len)
arch/m68k/kernel/sys_m68k.c
68
cache_flush_040 (unsigned long addr, int scope, int cache, unsigned long len)
arch/m68k/kernel/sys_m68k.c
72
switch (scope)
arch/powerpc/include/asm/io_event_irq.h
40
uint8_t scope; /* 0x02 Error/Event Scope */
arch/powerpc/perf/kvm-hv-pmu.c
393
.scope = PERF_PMU_SCOPE_SYS_WIDE,
arch/powerpc/platforms/powernv/eeh-powernv.c
858
uint8_t scope;
arch/powerpc/platforms/powernv/eeh-powernv.c
871
scope = OPAL_RESET_PCI_FUNDAMENTAL;
arch/powerpc/platforms/powernv/eeh-powernv.c
874
scope = OPAL_RESET_PCI_HOT;
arch/powerpc/platforms/powernv/eeh-powernv.c
884
rc = opal_pci_reset(id, scope, OPAL_ASSERT_RESET);
arch/powerpc/platforms/pseries/mobility.c
180
static int update_dt_node(struct device_node *dn, s32 scope)
arch/powerpc/platforms/pseries/mobility.c
204
scope);
arch/powerpc/platforms/pseries/mobility.c
295
static int pseries_devicetree_update(s32 scope)
arch/powerpc/platforms/pseries/mobility.c
311
rc = mobility_rtas_call(update_nodes_token, rtas_buf, scope);
arch/powerpc/platforms/pseries/mobility.c
340
update_dt_node(np, scope);
arch/powerpc/platforms/pseries/mobility.c
77
static int mobility_rtas_call(int token, char *buf, s32 scope)
arch/powerpc/platforms/pseries/mobility.c
84
rc = rtas_call(token, 2, 1, NULL, rtas_data_buf, scope);
arch/riscv/mm/cacheflush.c
246
int riscv_set_icache_flush_ctx(unsigned long ctx, unsigned long scope)
arch/riscv/mm/cacheflush.c
251
switch (scope) {
arch/riscv/mm/cacheflush.c
263
switch (scope) {
arch/s390/kernel/cache.c
160
pvt = (ct.ci[level].scope == CACHE_SCOPE_PRIVATE) ? 1 : 0;
arch/s390/kernel/cache.c
42
unsigned char scope : 2;
arch/s390/kernel/cache.c
93
if (ci->scope != CACHE_SCOPE_SHARED && ci->scope != CACHE_SCOPE_PRIVATE)
arch/x86/events/intel/cstate.c
404
.scope = PERF_PMU_SCOPE_CORE,
arch/x86/events/intel/cstate.c
420
.scope = PERF_PMU_SCOPE_PKG,
arch/x86/events/intel/cstate.c
436
.scope = PERF_PMU_SCOPE_CLUSTER,
arch/x86/events/intel/cstate.c
736
cstate_pkg_pmu.scope = PERF_PMU_SCOPE_DIE;
arch/x86/events/rapl.c
174
static inline unsigned int get_rapl_pmu_idx(int cpu, int scope)
arch/x86/events/rapl.c
181
switch (scope) {
arch/x86/events/rapl.c
204
if (event->pmu->scope == PERF_PMU_SCOPE_CORE)
arch/x86/events/rapl.c
391
rapl_pmus_scope = rapl_pmus->pmu.scope;
arch/x86/events/rapl.c
761
rapl_pmus->pmu.scope = rapl_pmu_scope;
arch/x86/kernel/cpu/resctrl/core.c
454
static int get_domain_id_from_scope(int cpu, enum resctrl_scope scope)
arch/x86/kernel/cpu/resctrl/core.c
456
switch (scope) {
arch/x86/kernel/cpu/resctrl/core.c
459
return get_cpu_cacheinfo_id(cpu, scope);
drivers/acpi/acpica/aclocal.h
366
scope,
drivers/acpi/acpica/aclocal.h
683
struct acpi_scope_state scope;
drivers/acpi/acpica/aclocal.h
917
union acpi_generic_state *scope; /* Current scope */
drivers/acpi/acpica/acnamesp.h
162
acpi_ns_convert_to_unicode(struct acpi_namespace_node *scope,
drivers/acpi/acpica/acnamesp.h
167
acpi_ns_convert_to_resource(struct acpi_namespace_node *scope,
drivers/acpi/acpica/acnamesp.h
172
acpi_ns_convert_to_reference(struct acpi_namespace_node *scope,
drivers/acpi/acpica/acparser.h
166
union acpi_parse_object *acpi_ps_find(union acpi_parse_object *scope,
drivers/acpi/acpica/acparser.h
79
union acpi_parse_object *acpi_ps_find_name(union acpi_parse_object *scope,
drivers/acpi/acpica/dsobject.c
124
walk_state->scope_info->scope.node;
drivers/acpi/acpica/dspkginit.c
406
scope_info.scope.node = element->reference.node; /* Prefix node */
drivers/acpi/acpica/dspkginit.c
457
scope_info.scope.node->name.ascii));
drivers/acpi/acpica/dswscope.c
100
scope_info->scope.node = node;
drivers/acpi/acpica/dswscope.c
114
scope.node),
drivers/acpi/acpica/dswscope.c
123
acpi_ut_get_node_name(scope_info->scope.node),
drivers/acpi/acpica/dswscope.c
164
acpi_ut_get_node_name(scope_info->scope.node),
drivers/acpi/acpica/dswscope.c
171
scope.node),
drivers/acpi/acpica/dswscope.c
40
walk_state->scope_info = scope_info->scope.next;
drivers/acpi/acpica/dswstate.c
693
if (walk_state->parser_state.scope) {
drivers/acpi/acpica/exconfig.c
124
start_node = walk_state->scope_info->scope.node;
drivers/acpi/acpica/excreate.c
272
walk_state->scope_info->scope.node;
drivers/acpi/acpica/exoparg1.c
881
scope_info->scope.
drivers/acpi/acpica/nsaccess.c
321
if ((!scope_info) || (!scope_info->scope.node)) {
drivers/acpi/acpica/nsaccess.c
328
prefix_node = scope_info->scope.node;
drivers/acpi/acpica/nsconvert.c
286
acpi_ns_convert_to_unicode(struct acpi_namespace_node *scope,
drivers/acpi/acpica/nsconvert.c
353
acpi_ns_convert_to_resource(struct acpi_namespace_node *scope,
drivers/acpi/acpica/nsconvert.c
432
acpi_ns_convert_to_reference(struct acpi_namespace_node *scope,
drivers/acpi/acpica/nsconvert.c
454
scope_info.scope.node =
drivers/acpi/acpica/nsconvert.c
455
ACPI_CAST_PTR(struct acpi_namespace_node, scope);
drivers/acpi/acpica/nsnames.c
349
if (prefix_scope && prefix_scope->scope.node) {
drivers/acpi/acpica/nsnames.c
351
acpi_ns_get_normalized_pathname(prefix_scope->scope.node,
drivers/acpi/acpica/nsutils.c
673
scope_info.scope.node = prefix_node;
drivers/acpi/acpica/psloop.c
249
if ((parser_state->scope->parse_scope.op) &&
drivers/acpi/acpica/psloop.c
250
((parser_state->scope->parse_scope.op->common.
drivers/acpi/acpica/psloop.c
252
|| (parser_state->scope->parse_scope.op->common.
drivers/acpi/acpica/psobject.c
447
walk_state->parser_state.scope->parse_scope.arg_count--;
drivers/acpi/acpica/psparse.c
337
parser_state->aml = parser_state->scope->parse_scope.pkg_end;
drivers/acpi/acpica/psparse.c
359
parser_state->aml = parser_state->scope->parse_scope.pkg_end;
drivers/acpi/acpica/psscope.c
115
union acpi_generic_state *scope;
drivers/acpi/acpica/psscope.c
119
scope = acpi_ut_create_generic_state();
drivers/acpi/acpica/psscope.c
120
if (!scope) {
drivers/acpi/acpica/psscope.c
124
scope->common.descriptor_type = ACPI_DESC_TYPE_STATE_PSCOPE;
drivers/acpi/acpica/psscope.c
125
scope->parse_scope.op = op;
drivers/acpi/acpica/psscope.c
126
scope->parse_scope.arg_list = remaining_args;
drivers/acpi/acpica/psscope.c
127
scope->parse_scope.arg_count = arg_count;
drivers/acpi/acpica/psscope.c
128
scope->parse_scope.pkg_end = parser_state->pkg_end;
drivers/acpi/acpica/psscope.c
132
acpi_ut_push_generic_state(&parser_state->scope, scope);
drivers/acpi/acpica/psscope.c
138
scope->parse_scope.arg_end = parser_state->pkg_end;
drivers/acpi/acpica/psscope.c
142
scope->parse_scope.arg_end = ACPI_TO_POINTER(ACPI_MAX_PTR);
drivers/acpi/acpica/psscope.c
168
union acpi_generic_state *scope = parser_state->scope;
drivers/acpi/acpica/psscope.c
174
if (scope->common.next) {
drivers/acpi/acpica/psscope.c
175
scope = acpi_ut_pop_generic_state(&parser_state->scope);
drivers/acpi/acpica/psscope.c
179
*op = scope->parse_scope.op;
drivers/acpi/acpica/psscope.c
180
*arg_list = scope->parse_scope.arg_list;
drivers/acpi/acpica/psscope.c
181
*arg_count = scope->parse_scope.arg_count;
drivers/acpi/acpica/psscope.c
182
parser_state->pkg_end = scope->parse_scope.pkg_end;
drivers/acpi/acpica/psscope.c
186
acpi_ut_delete_generic_state(scope);
drivers/acpi/acpica/psscope.c
215
union acpi_generic_state *scope;
drivers/acpi/acpica/psscope.c
225
while (parser_state->scope) {
drivers/acpi/acpica/psscope.c
226
scope = acpi_ut_pop_generic_state(&parser_state->scope);
drivers/acpi/acpica/psscope.c
227
acpi_ut_delete_generic_state(scope);
drivers/acpi/acpica/psscope.c
32
return (parser_state->scope->parse_scope.op);
drivers/acpi/acpica/psscope.c
53
((parser_state->aml >= parser_state->scope->parse_scope.arg_end
drivers/acpi/acpica/psscope.c
54
|| !parser_state->scope->parse_scope.arg_count)));
drivers/acpi/acpica/psscope.c
74
union acpi_generic_state *scope;
drivers/acpi/acpica/psscope.c
78
scope = acpi_ut_create_generic_state();
drivers/acpi/acpica/psscope.c
79
if (!scope) {
drivers/acpi/acpica/psscope.c
83
scope->common.descriptor_type = ACPI_DESC_TYPE_STATE_RPSCOPE;
drivers/acpi/acpica/psscope.c
84
scope->parse_scope.op = root_op;
drivers/acpi/acpica/psscope.c
85
scope->parse_scope.arg_count = ACPI_VAR_ARGS;
drivers/acpi/acpica/psscope.c
86
scope->parse_scope.arg_end = parser_state->aml_end;
drivers/acpi/acpica/psscope.c
87
scope->parse_scope.pkg_end = parser_state->aml_end;
drivers/acpi/acpica/psscope.c
89
parser_state->scope = scope;
drivers/acpi/property.c
109
acpi_get_parent(handle, &scope);
drivers/acpi/property.c
118
result = acpi_extract_properties(scope, desc, &dn->data);
drivers/acpi/property.c
123
if (acpi_enumerate_nondev_subnodes(scope, desc, &dn->data, &dn->fwnode))
drivers/acpi/property.c
143
static bool acpi_nondev_subnode_ok(acpi_handle scope,
drivers/acpi/property.c
158
if (!scope)
drivers/acpi/property.c
161
status = acpi_get_handle(scope, link->package.elements[1].string.pointer,
drivers/acpi/property.c
179
static bool acpi_add_nondev_subnodes(acpi_handle scope,
drivers/acpi/property.c
213
result = acpi_nondev_subnode_ok(scope, link, list,
drivers/acpi/property.c
232
acpi_handle_debug(scope, "subnode %s: Unknown scope\n",
drivers/acpi/property.c
244
acpi_handle_info(scope, "subnode %s: Unexpected reference\n",
drivers/acpi/property.c
257
static bool acpi_enumerate_nondev_subnodes(acpi_handle scope,
drivers/acpi/property.c
284
return acpi_add_nondev_subnodes(scope, links, &data->subnodes,
drivers/acpi/property.c
537
static bool acpi_extract_properties(acpi_handle scope, union acpi_object *desc,
drivers/acpi/property.c
564
acpi_data_add_buffer_props(scope, data, properties);
drivers/acpi/property.c
71
static bool acpi_enumerate_nondev_subnodes(acpi_handle scope,
drivers/acpi/property.c
86
acpi_handle scope = NULL;
drivers/acpi/property.c
912
acpi_handle scope, handle;
drivers/acpi/property.c
918
scope = to_acpi_device_node(fwnode)->handle;
drivers/acpi/property.c
920
scope = to_acpi_data_node(fwnode)->handle;
drivers/acpi/property.c
926
status = acpi_get_handle(scope, refstring, &handle);
drivers/acpi/property.c
928
acpi_handle_debug(scope, "Unable to get an ACPI handle for %s\n",
drivers/counter/counter-chrdev.c
285
switch (watch.component.scope) {
drivers/counter/counter-chrdev.c
319
if (watch.component.scope != COUNTER_SCOPE_SIGNAL)
drivers/counter/counter-chrdev.c
326
if (watch.component.scope != COUNTER_SCOPE_COUNT)
drivers/counter/counter-chrdev.c
333
if (watch.component.scope != COUNTER_SCOPE_COUNT)
drivers/counter/counter-chrdev.c
340
if (watch.component.scope != COUNTER_SCOPE_COUNT)
drivers/counter/counter-chrdev.c
488
const enum counter_scope scope,
drivers/counter/counter-chrdev.c
499
if (scope != COUNTER_SCOPE_SIGNAL)
drivers/counter/counter-chrdev.c
506
switch (scope) {
drivers/counter/counter-chrdev.c
528
const enum counter_scope scope = comp_node->component.scope;
drivers/counter/counter-chrdev.c
545
switch (scope) {
drivers/counter/counter-chrdev.c
566
switch (scope) {
drivers/counter/counter-chrdev.c
583
switch (scope) {
drivers/counter/counter-chrdev.c
598
switch (scope) {
drivers/counter/counter-chrdev.c
618
return counter_get_array_data(counter, scope, comp_node->parent,
drivers/counter/counter-sysfs.c
1062
const enum counter_scope scope = COUNTER_SCOPE_DEVICE;
drivers/counter/counter-sysfs.c
1085
scope, NULL);
drivers/counter/counter-sysfs.c
1091
scope, NULL);
drivers/counter/counter-sysfs.c
1097
&counter_events_queue_size_comp, scope, NULL);
drivers/counter/counter-sysfs.c
1103
counter->num_ext, scope, NULL);
drivers/counter/counter-sysfs.c
113
switch (a->scope) {
drivers/counter/counter-sysfs.c
154
switch (a->scope) {
drivers/counter/counter-sysfs.c
182
switch (a->scope) {
drivers/counter/counter-sysfs.c
279
switch (a->scope) {
drivers/counter/counter-sysfs.c
310
switch (a->scope) {
drivers/counter/counter-sysfs.c
342
switch (a->scope) {
drivers/counter/counter-sysfs.c
371
if (a->scope != COUNTER_SCOPE_SIGNAL ||
drivers/counter/counter-sysfs.c
394
a->scope != COUNTER_SCOPE_SIGNAL)
drivers/counter/counter-sysfs.c
40
enum counter_scope scope;
drivers/counter/counter-sysfs.c
421
switch (a->scope) {
drivers/counter/counter-sysfs.c
457
switch (a->scope) {
drivers/counter/counter-sysfs.c
570
const enum counter_scope scope,
drivers/counter/counter-sysfs.c
583
counter_attr->scope = scope;
drivers/counter/counter-sysfs.c
749
const enum counter_scope scope,
drivers/counter/counter-sysfs.c
755
err = counter_attr_create(dev, group, ext, scope, parent);
drivers/counter/counter-sysfs.c
766
const enum counter_scope scope,
drivers/counter/counter-sysfs.c
793
err = counter_ext_attrs_create(dev, group, &ext, scope, parent,
drivers/counter/counter-sysfs.c
806
const enum counter_scope scope,
drivers/counter/counter-sysfs.c
819
err = counter_array_attrs_create(dev, group, ext, scope,
drivers/counter/counter-sysfs.c
824
err = counter_ext_attrs_create(dev, group, ext, scope,
drivers/counter/counter-sysfs.c
844
const enum counter_scope scope = COUNTER_SCOPE_SIGNAL;
drivers/counter/counter-sysfs.c
852
err = counter_attr_create(dev, cattr_group, &comp, scope, signal);
drivers/counter/counter-sysfs.c
863
signal->num_ext, scope, signal);
drivers/counter/counter-sysfs.c
944
const enum counter_scope scope = COUNTER_SCOPE_COUNT;
drivers/counter/counter-sysfs.c
953
err = counter_attr_create(dev, cattr_group, &comp, scope, count);
drivers/counter/counter-sysfs.c
966
err = counter_attr_create(dev, cattr_group, &comp, scope, count);
drivers/counter/counter-sysfs.c
972
count->num_ext, scope, count);
drivers/crypto/intel/qat/qat_common/icp_qat_uclo.h
294
char scope;
drivers/crypto/intel/qat/qat_common/qat_uclo.c
224
if (init_mem->scope != ICP_QAT_UOF_LOCAL_SCOPE) {
drivers/cxl/core/features.c
543
enum fwctl_rpc_scope scope)
drivers/cxl/core/features.c
595
if (effects & imm_mask && scope >= FWCTL_RPC_DEBUG_WRITE_FULL)
drivers/cxl/core/features.c
603
if (!(effects & imm_mask) && scope >= FWCTL_RPC_DEBUG_WRITE)
drivers/cxl/core/features.c
611
enum fwctl_rpc_scope scope,
drivers/cxl/core/features.c
623
return cxlctl_validate_set_features(cxlfs, rpc_in, scope);
drivers/cxl/core/features.c
645
static void *cxlctl_fw_rpc(struct fwctl_uctx *uctx, enum fwctl_rpc_scope scope,
drivers/cxl/core/features.c
654
if (!cxlctl_validate_hw_command(cxlfs, rpc_in, scope, opcode))
drivers/dma/idxd/perfmon.c
460
idxd_pmu->pmu.scope = PERF_PMU_SCOPE_SYS_WIDE;
drivers/firmware/xilinx/zynqmp.c
1758
struct zynqmp_pm_shutdown_scope *scope;
drivers/firmware/xilinx/zynqmp.c
1760
scope = zynqmp_pm_is_shutdown_scope_valid(buf);
drivers/firmware/xilinx/zynqmp.c
1761
if (!scope)
drivers/firmware/xilinx/zynqmp.c
1765
scope->subtype);
drivers/firmware/xilinx/zynqmp.c
1771
selected_scope = scope;
drivers/fwctl/main.c
125
ucmd->uctx, cmd->scope, inbuf, cmd->in_len, &out_len);
drivers/fwctl/main.c
95
switch (cmd->scope) {
drivers/fwctl/mlx5/main.c
180
static bool mlx5ctl_validate_rpc(const void *in, enum fwctl_rpc_scope scope)
drivers/fwctl/mlx5/main.c
291
return scope >= FWCTL_RPC_DEBUG_READ_ONLY;
drivers/fwctl/mlx5/main.c
294
return scope >= FWCTL_RPC_DEBUG_WRITE;
drivers/fwctl/mlx5/main.c
300
return scope >= FWCTL_RPC_DEBUG_READ_ONLY;
drivers/fwctl/mlx5/main.c
306
static void *mlx5ctl_fw_rpc(struct fwctl_uctx *uctx, enum fwctl_rpc_scope scope,
drivers/fwctl/mlx5/main.c
324
if (!mlx5ctl_validate_rpc(rpc_in, scope))
drivers/fwctl/pds/main.c
258
switch (entries[i].scope) {
drivers/fwctl/pds/main.c
262
entries[i].scope = FWCTL_RPC_CONFIGURATION;
drivers/fwctl/pds/main.c
265
entries[i].scope = FWCTL_RPC_DEBUG_READ_ONLY;
drivers/fwctl/pds/main.c
268
entries[i].scope = FWCTL_RPC_DEBUG_WRITE;
drivers/fwctl/pds/main.c
271
entries[i].scope = FWCTL_RPC_DEBUG_WRITE_FULL;
drivers/fwctl/pds/main.c
275
ep, le32_to_cpu(entries[i].id), entries[i].scope);
drivers/fwctl/pds/main.c
283
enum fwctl_rpc_scope scope)
drivers/fwctl/pds/main.c
339
if (scope < op_entry[i].scope)
drivers/fwctl/pds/main.c
350
static void *pdsfc_fw_rpc(struct fwctl_uctx *uctx, enum fwctl_rpc_scope scope,
drivers/fwctl/pds/main.c
365
err = pdsfc_validate_rpc(pdsfc, rpc, scope);
drivers/gpu/drm/i915/i915_pmu.c
1189
pmu->base.scope = PERF_PMU_SCOPE_SYS_WIDE;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/fifo.c
370
msg->nv2080EngineType, msg->chid, msg->exceptType, msg->scope,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/nvrm/fifo.h
347
NvU32 scope;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/fifo.c
110
msg->scope, msg->partitionAttributionId,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/nvrm/fifo.h
182
NvU32 scope;
drivers/gpu/drm/xe/xe_pmu.c
574
pmu->base.scope = PERF_PMU_SCOPE_SYS_WIDE;
drivers/i2c/busses/i2c-rtl9300.c
34
enum rtl9300_i2c_reg_scope scope;
drivers/i2c/busses/i2c-rtl9300.c
408
if (drv_data->field_desc[i].scope == REG_SCOPE_MASTER)
drivers/i2c/busses/i2c-rtl9300.c
470
{ .field = REG_FIELD(reg, msb, lsb), .scope = REG_SCOPE_GLOBAL }
drivers/i2c/busses/i2c-rtl9300.c
472
{ .field = REG_FIELD(reg, msb, lsb), .scope = REG_SCOPE_MASTER }
drivers/infiniband/core/multicast.c
322
if (comp_mask & IB_SA_MCMEMBER_REC_SCOPE && src->scope != dst->scope)
drivers/infiniband/core/sa_query.c
492
{ MCMEMBER_REC_FIELD(scope),
drivers/infiniband/core/uverbs_uapi.c
230
if (def->scope == UAPI_SCOPE_OBJECT) {
drivers/infiniband/core/uverbs_uapi.c
241
if (def->scope == UAPI_SCOPE_METHOD &&
drivers/infiniband/core/uverbs_uapi.c
253
if (def->scope == UAPI_SCOPE_METHOD &&
drivers/infiniband/hw/cxgb4/cm.c
3274
if (ifp->scope == IFA_LINK &&
drivers/infiniband/hw/efa/efa_admin_cmds_defs.h
606
u8 scope;
drivers/infiniband/hw/efa/efa_com_cmd.c
781
cmd.scope = params->scope;
drivers/infiniband/hw/efa/efa_com_cmd.c
793
cmd.type, cmd.scope, cmd.scope_modifier, err);
drivers/infiniband/hw/efa/efa_com_cmd.h
253
u8 scope;
drivers/infiniband/hw/efa/efa_verbs.c
2237
params.scope = EFA_ADMIN_GET_STATS_SCOPE_ALL;
drivers/iommu/intel/dmar.c
222
struct acpi_dmar_device_scope *scope;
drivers/iommu/intel/dmar.c
228
for (; start < end; start += scope->length) {
drivers/iommu/intel/dmar.c
229
scope = start;
drivers/iommu/intel/dmar.c
230
if (scope->entry_type != ACPI_DMAR_SCOPE_TYPE_ENDPOINT &&
drivers/iommu/intel/dmar.c
231
scope->entry_type != ACPI_DMAR_SCOPE_TYPE_BRIDGE)
drivers/iommu/intel/dmar.c
234
path = (struct acpi_dmar_pci_path *)(scope + 1);
drivers/iommu/intel/dmar.c
235
level = (scope->length - sizeof(*scope)) / sizeof(*path);
drivers/iommu/intel/dmar.c
236
if (!dmar_match_pci_path(info, scope->bus, path, level))
drivers/iommu/intel/dmar.c
248
if ((scope->entry_type == ACPI_DMAR_SCOPE_TYPE_ENDPOINT &&
drivers/iommu/intel/dmar.c
250
(scope->entry_type == ACPI_DMAR_SCOPE_TYPE_BRIDGE &&
drivers/iommu/intel/dmar.c
733
struct acpi_dmar_device_scope *scope;
drivers/iommu/intel/dmar.c
743
for (scope = (void *)(drhd + 1);
drivers/iommu/intel/dmar.c
744
(unsigned long)scope < ((unsigned long)drhd) + drhd->header.length;
drivers/iommu/intel/dmar.c
745
scope = ((void *)scope) + scope->length) {
drivers/iommu/intel/dmar.c
746
if (scope->entry_type != ACPI_DMAR_SCOPE_TYPE_NAMESPACE)
drivers/iommu/intel/dmar.c
748
if (scope->enumeration_id != device_number)
drivers/iommu/intel/dmar.c
751
path = (void *)(scope + 1);
drivers/iommu/intel/dmar.c
754
scope->bus, path->device, path->function);
drivers/iommu/intel/dmar.c
757
dmaru->devices[i].bus = scope->bus;
drivers/iommu/intel/dmar.c
84
struct acpi_dmar_device_scope *scope;
drivers/iommu/intel/dmar.c
88
scope = start;
drivers/iommu/intel/dmar.c
89
if (scope->entry_type == ACPI_DMAR_SCOPE_TYPE_NAMESPACE ||
drivers/iommu/intel/dmar.c
90
scope->entry_type == ACPI_DMAR_SCOPE_TYPE_ENDPOINT ||
drivers/iommu/intel/dmar.c
91
scope->entry_type == ACPI_DMAR_SCOPE_TYPE_BRIDGE)
drivers/iommu/intel/dmar.c
93
else if (scope->entry_type != ACPI_DMAR_SCOPE_TYPE_IOAPIC &&
drivers/iommu/intel/dmar.c
94
scope->entry_type != ACPI_DMAR_SCOPE_TYPE_HPET) {
drivers/iommu/intel/dmar.c
97
start += scope->length;
drivers/iommu/intel/irq_remapping.c
828
static int ir_parse_one_hpet_scope(struct acpi_dmar_device_scope *scope,
drivers/iommu/intel/irq_remapping.c
836
bus = scope->bus;
drivers/iommu/intel/irq_remapping.c
837
path = (struct acpi_dmar_pci_path *)(scope + 1);
drivers/iommu/intel/irq_remapping.c
838
count = (scope->length - sizeof(struct acpi_dmar_device_scope))
drivers/iommu/intel/irq_remapping.c
853
ir_hpet[count].id == scope->enumeration_id)
drivers/iommu/intel/irq_remapping.c
864
ir_hpet[free].id = scope->enumeration_id;
drivers/iommu/intel/irq_remapping.c
868
scope->enumeration_id, drhd->address);
drivers/iommu/intel/irq_remapping.c
873
static int ir_parse_one_ioapic_scope(struct acpi_dmar_device_scope *scope,
drivers/iommu/intel/irq_remapping.c
881
bus = scope->bus;
drivers/iommu/intel/irq_remapping.c
882
path = (struct acpi_dmar_pci_path *)(scope + 1);
drivers/iommu/intel/irq_remapping.c
883
count = (scope->length - sizeof(struct acpi_dmar_device_scope))
drivers/iommu/intel/irq_remapping.c
898
ir_ioapic[count].id == scope->enumeration_id)
drivers/iommu/intel/irq_remapping.c
911
ir_ioapic[free].id = scope->enumeration_id;
drivers/iommu/intel/irq_remapping.c
913
scope->enumeration_id, drhd->address, iommu->seq_id);
drivers/iommu/intel/irq_remapping.c
923
struct acpi_dmar_device_scope *scope;
drivers/iommu/intel/irq_remapping.c
931
scope = start;
drivers/iommu/intel/irq_remapping.c
932
if (scope->entry_type == ACPI_DMAR_SCOPE_TYPE_IOAPIC)
drivers/iommu/intel/irq_remapping.c
933
ret = ir_parse_one_ioapic_scope(scope, iommu, drhd);
drivers/iommu/intel/irq_remapping.c
934
else if (scope->entry_type == ACPI_DMAR_SCOPE_TYPE_HPET)
drivers/iommu/intel/irq_remapping.c
935
ret = ir_parse_one_hpet_scope(scope, iommu, drhd);
drivers/iommu/intel/irq_remapping.c
936
start += scope->length;
drivers/iommu/intel/perfmon.c
549
iommu_pmu->pmu.scope = PERF_PMU_SCOPE_SYS_WIDE;
drivers/net/ethernet/amazon/ena/ena_admin_defs.h
395
u8 scope;
drivers/net/ethernet/intel/iavf/iavf_main.c
5070
if (shaper->handle.scope == NET_SHAPER_SCOPE_QUEUE) {
drivers/net/ethernet/intel/iavf/iavf_main.c
5135
enum net_shaper_scope scope,
drivers/net/ethernet/intel/iavf/iavf_main.c
5138
if (scope != NET_SHAPER_SCOPE_QUEUE)
drivers/net/ethernet/microsoft/mana/mana_en.c
903
if (shaper->handle.scope != NET_SHAPER_SCOPE_NETDEV) {
drivers/net/ethernet/microsoft/mana/mana_en.c
946
apc->handle.scope = NET_SHAPER_SCOPE_UNSPEC;
drivers/net/ethernet/microsoft/mana/mana_en.c
954
enum net_shaper_scope scope,
drivers/net/netdevsim/netdev.c
632
enum net_shaper_scope scope,
drivers/net/wireless/ath/ath11k/mac.c
9507
u32 count, scope;
drivers/net/wireless/ath/ath11k/mac.c
9533
scope = ipv6_addr_src_scope(&ifa6->addr);
drivers/net/wireless/ath/ath11k/mac.c
9534
if (scope == IPV6_ADDR_SCOPE_LINKLOCAL ||
drivers/net/wireless/ath/ath11k/mac.c
9535
scope == IPV6_ADDR_SCOPE_GLOBAL) {
drivers/net/wireless/ath/ath11k/mac.c
9541
scope);
drivers/net/wireless/ath/ath11k/mac.c
9544
ath11k_warn(ar->ab, "Unsupported ipv6 scope: %d\n", scope);
drivers/net/wireless/ath/ath11k/mac.c
9554
scope = ipv6_addr_src_scope(&ifaca6->aca_addr);
drivers/net/wireless/ath/ath11k/mac.c
9555
if (scope == IPV6_ADDR_SCOPE_LINKLOCAL ||
drivers/net/wireless/ath/ath11k/mac.c
9556
scope == IPV6_ADDR_SCOPE_GLOBAL) {
drivers/net/wireless/ath/ath11k/mac.c
9562
scope);
drivers/net/wireless/ath/ath11k/mac.c
9565
ath11k_warn(ar->ab, "Unsupported ipv scope: %d\n", scope);
drivers/net/wireless/ath/ath12k/wow.c
647
u32 count = 0, scope;
drivers/net/wireless/ath/ath12k/wow.c
668
scope = ipv6_addr_src_scope(&ifa6->addr);
drivers/net/wireless/ath/ath12k/wow.c
669
if (scope != IPV6_ADDR_SCOPE_LINKLOCAL &&
drivers/net/wireless/ath/ath12k/wow.c
670
scope != IPV6_ADDR_SCOPE_GLOBAL) {
drivers/net/wireless/ath/ath12k/wow.c
672
"Unsupported ipv6 scope: %d\n", scope);
drivers/net/wireless/ath/ath12k/wow.c
681
scope);
drivers/net/wireless/ath/ath12k/wow.c
695
scope = ipv6_addr_src_scope(&ifaca6->aca_addr);
drivers/net/wireless/ath/ath12k/wow.c
696
if (scope != IPV6_ADDR_SCOPE_LINKLOCAL &&
drivers/net/wireless/ath/ath12k/wow.c
697
scope != IPV6_ADDR_SCOPE_GLOBAL) {
drivers/net/wireless/ath/ath12k/wow.c
699
"Unsupported ipv scope: %d\n", scope);
drivers/net/wireless/ath/ath12k/wow.c
708
scope);
drivers/pci/tsm.c
403
ssize_t pci_tsm_guest_req(struct pci_dev *pdev, enum pci_tsm_req_scope scope,
drivers/pci/tsm.c
412
if (scope > PCI_TSM_REQ_STATE_CHANGE)
drivers/pci/tsm.c
433
return to_pci_tsm_ops(pdev->tsm)->guest_req(tdi, scope, req_in, in_len,
drivers/scsi/aic7xxx/aicasm/aicasm.c
460
dump_scope(scope_t *scope)
drivers/scsi/aic7xxx/aicasm/aicasm.c
467
emit_patch(scope, 0);
drivers/scsi/aic7xxx/aicasm/aicasm.c
472
cur_scope = TAILQ_FIRST(&scope->inner_scope);
drivers/scsi/aic7xxx/aicasm/aicasm.c
484
emit_patch(scope, 1);
drivers/scsi/aic7xxx/aicasm/aicasm.c
488
emit_patch(scope_t *scope, int patch)
drivers/scsi/aic7xxx/aicasm/aicasm.c
493
pinfo = &scope->patches[patch];
drivers/scsi/aic7xxx/aicasm/aicasm.c
507
new_patch->patch_func = scope->func_num;
drivers/scsi/aic7xxx/aicasm/aicasm.c
508
new_patch->begin = scope->begin_addr;
drivers/scsi/aic7xxx/aicasm/aicasm.c
511
new_patch->begin = scope->end_addr;
drivers/scsi/aic7xxx/aicasm/aicasm.c
780
process_scope(scope_t *scope)
drivers/scsi/aic7xxx/aicasm/aicasm.c
791
cur_scope = TAILQ_LAST(&scope->inner_scope, scope_tailq);
drivers/scsi/aic7xxx/aicasm/aicasm.c
80
static void dump_scope(scope_t *scope);
drivers/scsi/aic7xxx/aicasm/aicasm.c
81
static void emit_patch(scope_t *scope, int patch);
drivers/scsi/aic7xxx/aicasm/aicasm.c
824
scope->inner_scope_patches += skip_patch_count;
drivers/scsi/aic7xxx/aicasm/aicasm.h
90
struct scope *scope_alloc(void);
drivers/scsi/aic7xxx/aicasm/aicasm.h
91
void process_scope(struct scope *);
drivers/scsi/aic7xxx/aicasm/aicasm_symbol.h
169
SLIST_ENTRY(scope) scope_stack_links;
drivers/scsi/aic7xxx/aicasm/aicasm_symbol.h
170
TAILQ_ENTRY(scope) scope_links;
drivers/scsi/aic7xxx/aicasm/aicasm_symbol.h
171
TAILQ_HEAD(, scope) inner_scope;
drivers/scsi/aic7xxx/aicasm/aicasm_symbol.h
181
SLIST_HEAD(scope_list, scope);
drivers/scsi/aic7xxx/aicasm/aicasm_symbol.h
182
TAILQ_HEAD(scope_tailq, scope);
drivers/scsi/bnx2fc/bnx2fc_io.c
1909
u16 scope = 0, qualifier = 0;
drivers/scsi/bnx2fc/bnx2fc_io.c
1991
scope = fcp_rsp->retry_delay_timer
drivers/scsi/bnx2fc/bnx2fc_io.c
1997
if (scope > 0 && qualifier > 0 &&
drivers/scsi/qedf/qedf_io.c
1124
u16 scope, qualifier = 0;
drivers/scsi/qedf/qedf_io.c
1249
scope = fcp_rsp->retry_delay_timer & 0xC000;
drivers/scsi/qedf/qedf_io.c
1267
if ((scope == 1 || scope == 2) &&
drivers/scsi/qedf/qedf_io.c
1279
scope, qualifier);
drivers/scsi/qedf/qedf_io.c
1292
scope, qualifier);
drivers/scsi/qla2xxx/qla_inline.h
278
u8 scope;
drivers/scsi/qla2xxx/qla_inline.h
290
scope = (sts_qual & SQ_SCOPE_MASK) >> SQ_SCOPE_SHIFT;
drivers/scsi/qla2xxx/qla_inline.h
292
if (scope != 1 && scope != 2)
drivers/target/iscsi/iscsi_target_parameters.c
531
if (!leading && (param->scope & SCOPE_SESSION_WIDE)) {
drivers/target/iscsi/iscsi_target_parameters.c
556
new_param->scope = param->scope;
drivers/target/iscsi/iscsi_target_parameters.c
71
char *name, char *value, u8 phase, u8 scope, u8 sender,
drivers/target/iscsi/iscsi_target_parameters.c
96
param->scope = scope;
drivers/target/iscsi/iscsi_target_parameters.h
19
u8 scope;
drivers/target/iscsi/iscsi_target_parameters.h
195
#define IS_SCOPE_CONNECTION_ONLY(p) ((p)->scope & SCOPE_CONNECTION_ONLY)
drivers/target/iscsi/iscsi_target_parameters.h
196
#define IS_SCOPE_SESSION_WIDE(p) ((p)->scope & SCOPE_SESSION_WIDE)
drivers/target/target_core_pr.c
2263
core_scsi3_pro_reserve(struct se_cmd *cmd, int type, int scope, u64 res_key)
drivers/target/target_core_pr.c
2313
if (scope != PR_SCOPE_LU_SCOPE) {
drivers/target/target_core_pr.c
2314
pr_err("SPC-3 PR: Illegal SCOPE: 0x%02x\n", scope);
drivers/target/target_core_pr.c
2358
(pr_res_holder->pr_res_scope != scope)) {
drivers/target/target_core_pr.c
2391
pr_reg->pr_res_scope = scope;
drivers/target/target_core_pr.c
2417
core_scsi3_emulate_pro_reserve(struct se_cmd *cmd, int type, int scope,
drivers/target/target_core_pr.c
2427
return core_scsi3_pro_reserve(cmd, type, scope, res_key);
drivers/target/target_core_pr.c
2503
core_scsi3_emulate_pro_release(struct se_cmd *cmd, int type, int scope,
drivers/target/target_core_pr.c
2586
(pr_res_holder->pr_res_scope != scope)) {
drivers/target/target_core_pr.c
2748
int scope,
drivers/target/target_core_pr.c
2768
pr_reg->pr_res_scope = scope;
drivers/target/target_core_pr.c
2812
core_scsi3_pro_preempt(struct se_cmd *cmd, int type, int scope, u64 res_key,
drivers/target/target_core_pr.c
2841
if (scope != PR_SCOPE_LU_SCOPE) {
drivers/target/target_core_pr.c
2842
pr_err("SPC-3 PR: Illegal SCOPE: 0x%02x\n", scope);
drivers/target/target_core_pr.c
2961
type, scope, preempt_type);
drivers/target/target_core_pr.c
3067
type, scope, preempt_type);
drivers/target/target_core_pr.c
3081
if ((prh_type != type) || (prh_scope != scope)) {
drivers/target/target_core_pr.c
3122
core_scsi3_emulate_pro_preempt(struct se_cmd *cmd, int type, int scope,
drivers/target/target_core_pr.c
3132
return core_scsi3_pro_preempt(cmd, type, scope, res_key,
drivers/target/target_core_pr.c
3159
int new_reg = 0, type, scope, matching_iname;
drivers/target/target_core_pr.c
3433
scope = pr_res_holder->pr_res_type;
drivers/target/target_core_pr.c
3489
pr_reg->pr_res_scope = scope;
drivers/target/target_core_pr.c
3586
int sa, scope, type, aptpl;
drivers/target/target_core_pr.c
3623
scope = (cdb[2] & 0xf0);
drivers/target/target_core_pr.c
3696
ret = core_scsi3_emulate_pro_reserve(cmd, type, scope, res_key);
drivers/target/target_core_pr.c
3699
ret = core_scsi3_emulate_pro_release(cmd, type, scope, res_key);
drivers/target/target_core_pr.c
3705
ret = core_scsi3_emulate_pro_preempt(cmd, type, scope,
drivers/target/target_core_pr.c
3709
ret = core_scsi3_emulate_pro_preempt(cmd, type, scope,
drivers/usb/typec/ucsi/psy.c
38
u8 scope = POWER_SUPPLY_SCOPE_UNKNOWN;
drivers/usb/typec/ucsi/psy.c
41
device_property_read_u8(dev, "scope", &scope);
drivers/usb/typec/ucsi/psy.c
42
if (scope == POWER_SUPPLY_SCOPE_UNKNOWN) {
drivers/usb/typec/ucsi/psy.c
47
scope = POWER_SUPPLY_SCOPE_SYSTEM;
drivers/usb/typec/ucsi/psy.c
49
scope = POWER_SUPPLY_SCOPE_DEVICE;
drivers/usb/typec/ucsi/psy.c
51
val->intval = scope;
fs/nfsd/nfsctl.c
1597
const char *scope = NULL;
fs/nfsd/nfsctl.c
1653
scope = nla_data(attr);
fs/nfsd/nfsctl.c
1660
ret = nfsd_svc(nrpools, nthreads, net, current_cred(), scope);
fs/nfsd/nfsd.h
99
const struct cred *cred, const char *scope);
fs/nfsd/nfssvc.c
751
nfsd_svc(int n, int *nthreads, struct net *net, const struct cred *cred, const char *scope)
fs/nfsd/nfssvc.c
761
strscpy(nn->nfsd_name, scope ? scope : utsname()->nodename,
fs/resctrl/pseudo_lock.c
227
enum resctrl_scope scope = plr->s->res->ctrl_scope;
fs/resctrl/pseudo_lock.c
231
if (WARN_ON_ONCE(scope != RESCTRL_L2_CACHE && scope != RESCTRL_L3_CACHE))
fs/resctrl/pseudo_lock.c
244
ci = get_cpu_cacheinfo_level(plr->cpu, scope);
include/linux/btf_ids.h
148
#define __BTF_SET_START(name, scope) \
include/linux/btf_ids.h
151
"." #scope " __BTF_ID__set__" #name "; \n" \
include/linux/btf_ids.h
192
#define __BTF_SET8_START(name, scope, flags) \
include/linux/btf_ids.h
196
"." #scope " __BTF_ID__set8__" #name "; \n" \
include/linux/btf_ids.h
89
#define __BTF_ID_LIST(name, scope) \
include/linux/btf_ids.h
92
"." #scope " " #name "; \n" \
include/linux/cleanup.h
438
for (CLASS(_name, scope)(args); \
include/linux/cleanup.h
439
__guard_ptr(_name)(&scope) || !__is_cond_ptr(_name); \
include/linux/cleanup.h
450
for (CLASS(_name, scope)(args); true; ({ goto _label; })) \
include/linux/cleanup.h
451
if (!__guard_ptr(_name)(&scope)) { \
include/linux/fwctl.h
56
void *(*fw_rpc)(struct fwctl_uctx *uctx, enum fwctl_rpc_scope scope,
include/linux/inetdevice.h
199
__be32 inet_select_addr(const struct net_device *dev, __be32 dst, int scope);
include/linux/inetdevice.h
201
__be32 local, int scope);
include/linux/pci-tsm.h
216
ssize_t pci_tsm_guest_req(struct pci_dev *pdev, enum pci_tsm_req_scope scope,
include/linux/pci-tsm.h
235
enum pci_tsm_req_scope scope,
include/linux/pci-tsm.h
57
enum pci_tsm_req_scope scope,
include/linux/pds/pds_adminq.h
1351
u8 scope;
include/linux/perf_event.h
352
unsigned int scope;
include/linux/pid_namespace.h
71
int scope = MEMFD_NOEXEC_SCOPE_EXEC;
include/linux/pid_namespace.h
74
scope = max(scope, READ_ONCE(ns->memfd_noexec_scope));
include/linux/pid_namespace.h
76
return scope;
include/linux/tty_port.h
284
#define scoped_tty() ((struct tty_struct *)(__guard_ptr(tty_port_tty)(&scope)))
include/net/addrconf.h
89
u16 scope;
include/net/flow.h
108
fl4->flowi4_scope = scope;
include/net/flow.h
97
__u32 mark, __u8 tos, __u8 scope,
include/net/if_inet6.h
246
unsigned char scope = broadcast[5] & 0xF;
include/net/if_inet6.h
253
buf[5] = 0x10 | scope; /* scope from broadcast address */
include/net/if_inet6.h
50
__u16 scope;
include/net/ip.h
640
unsigned char scope = broadcast[5] & 0xF;
include/net/ip.h
648
buf[5] = 0x10 | scope; /* scope from broadcast address */
include/net/ip_fib.h
178
unsigned char scope;
include/net/ip_fib.h
198
unsigned char scope;
include/net/ip_fib.h
209
unsigned char scope);
include/net/ip_fib.h
575
int fib_check_nh(struct net *net, struct fib_nh *nh, u32 table, u8 scope,
include/net/net_shaper.h
117
enum net_shaper_scope scope, unsigned long *cap);
include/net/net_shaper.h
28
enum net_shaper_scope scope;
include/net/nexthop.h
486
int fib_check_nexthop(struct nexthop *nh, u8 scope,
include/net/route.h
188
int oif, __u8 scope)
include/net/route.h
193
.flowi4_scope = scope,
include/net/sctp/structs.h
1177
enum sctp_scope scope, gfp_t gfp,
include/net/sctp/structs.h
1205
const enum sctp_scope scope);
include/net/sctp/structs.h
2111
enum sctp_scope scope, gfp_t gfp);
include/net/sctp/structs.h
2150
enum sctp_scope scope, gfp_t gfp);
include/net/sctp/structs.h
464
enum sctp_scope (*scope)(union sctp_addr *);
include/rdma/ib_sa.h
338
u8 scope;
include/rdma/uverbs_ioctl.h
333
u8 scope;
include/rdma/uverbs_ioctl.h
369
.scope = UAPI_SCOPE_OBJECT, \
include/rdma/uverbs_ioctl.h
380
.scope = UAPI_SCOPE_OBJECT, \
include/rdma/uverbs_ioctl.h
394
.scope = UAPI_SCOPE_OBJECT, \
include/rdma/uverbs_ioctl.h
409
.scope = UAPI_SCOPE_METHOD, \
include/rdma/uverbs_ioctl.h
421
.scope = UAPI_SCOPE_OBJECT, .func_is_supported = _func, \
include/trace/events/fib.h
29
__field( __u8, scope )
include/trace/events/fib.h
50
__entry->scope = flp->flowi4_scope;
include/trace/events/fib.h
98
__entry->tos, __entry->scope, __entry->flags,
include/trace/events/fib6.h
27
__field( __u8, scope )
include/trace/events/fib6.h
48
__entry->scope = flp->flowi6_scope;
include/trace/events/fib6.h
84
__entry->flowlabel, __entry->tos, __entry->scope,
include/uapi/fwctl/fwctl.h
133
__u32 scope;
include/uapi/linux/counter.h
47
__u8 scope;
include/uapi/linux/tipc.h
160
signed char scope;
include/uapi/linux/tipc.h
205
__u32 scope; /* cluster/node */
kernel/events/core.c
12578
static inline const struct cpumask *perf_scope_cpu_topology_cpumask(unsigned int scope, int cpu)
kernel/events/core.c
12580
switch (scope) {
kernel/events/core.c
12596
static inline struct cpumask *perf_scope_cpumask(unsigned int scope)
kernel/events/core.c
12598
switch (scope) {
kernel/events/core.c
12618
struct cpumask *mask = perf_scope_cpumask(pmu->scope);
kernel/events/core.c
12644
if (n == 3 && pmu->scope == PERF_PMU_SCOPE_NONE)
kernel/events/core.c
12772
if (WARN_ONCE(pmu->scope >= PERF_PMU_MAX_SCOPE,
kernel/events/core.c
13049
if (pmu->scope != PERF_PMU_SCOPE_NONE && event->cpu >= 0) {
kernel/events/core.c
13054
cpumask = perf_scope_cpu_topology_cpumask(pmu->scope, event->cpu);
kernel/events/core.c
13055
pmu_cpumask = perf_scope_cpumask(pmu->scope);
kernel/events/core.c
15091
unsigned int scope;
kernel/events/core.c
15096
for (scope = PERF_PMU_SCOPE_NONE + 1; scope < PERF_PMU_MAX_SCOPE; scope++) {
kernel/events/core.c
15097
const struct cpumask *cpumask = perf_scope_cpu_topology_cpumask(scope, cpu);
kernel/events/core.c
15098
struct cpumask *pmu_cpumask = perf_scope_cpumask(scope);
kernel/events/core.c
15100
target[scope] = -1;
kernel/events/core.c
15106
target[scope] = cpumask_any_but(cpumask, cpu);
kernel/events/core.c
15107
if (target[scope] < nr_cpu_ids)
kernel/events/core.c
15108
cpumask_set_cpu(target[scope], pmu_cpumask);
kernel/events/core.c
15113
if (pmu->scope == PERF_PMU_SCOPE_NONE ||
kernel/events/core.c
15114
WARN_ON_ONCE(pmu->scope >= PERF_PMU_MAX_SCOPE))
kernel/events/core.c
15117
if (target[pmu->scope] >= 0 && target[pmu->scope] < nr_cpu_ids)
kernel/events/core.c
15118
perf_pmu_migrate_context(pmu, cpu, target[pmu->scope]);
kernel/events/core.c
15153
unsigned int scope;
kernel/events/core.c
15161
for (scope = PERF_PMU_SCOPE_NONE + 1; scope < PERF_PMU_MAX_SCOPE; scope++) {
kernel/events/core.c
15162
pmu_cpumask = perf_scope_cpumask(scope);
kernel/events/core.c
15170
for (scope = PERF_PMU_SCOPE_NONE + 1; scope < PERF_PMU_MAX_SCOPE; scope++) {
kernel/events/core.c
15171
const struct cpumask *cpumask = perf_scope_cpu_topology_cpumask(scope, cpu);
kernel/events/core.c
15173
pmu_cpumask = perf_scope_cpumask(scope);
kernel/events/core.c
4779
static inline const struct cpumask *perf_scope_cpu_topology_cpumask(unsigned int scope, int cpu);
kernel/events/core.c
4790
const struct cpumask *cpumask = perf_scope_cpu_topology_cpumask(event->pmu->scope, event_cpu);
kernel/irq/internals.h
193
#define scoped_irqdesc ((struct irq_desc *)(__guard_ptr(irqdesc_lock)(&scope)))
kernel/pid_sysctl.h
13
int err, scope, parent_scope;
kernel/pid_sysctl.h
23
scope = max(READ_ONCE(ns->memfd_noexec_scope), parent_scope);
kernel/pid_sysctl.h
25
table_copy.data = &scope;
kernel/pid_sysctl.h
30
WRITE_ONCE(ns->memfd_noexec_scope, scope);
kernel/sched/core.c
7391
scope->flags |= ENQUEUE_REPLENISH;
kernel/sched/core.c
7399
scope->flags |= ENQUEUE_HEAD;
kernel/sched/core.c
9183
if (scope->running)
kernel/sched/core.c
9185
queued = scope->queued;
kernel/sched/deadline.c
1156
struct rq_flags *rf = &scope.rf;
kernel/sched/fair.c
5804
rq = scope.rq;
kernel/sched/syscalls.c
1528
struct rq *rq = scope.rq;
kernel/sched/syscalls.c
290
struct sched_change_ctx *scope)
kernel/sched/syscalls.c
304
scope->flags |= ENQUEUE_REPLENISH;
kernel/sched/syscalls.c
311
struct sched_change_ctx *scope)
kernel/sched/syscalls.c
687
__setscheduler_dl_pi(newprio, policy, p, scope);
kernel/sched/syscalls.c
691
if (scope->queued) {
kernel/sched/syscalls.c
697
scope->flags |= ENQUEUE_HEAD;
kernel/time/posix-timers.c
79
#define scoped_timer (scope)
kernel/workqueue.c
4811
enum wq_affn_scope scope;
kernel/workqueue.c
4818
scope = wq_affn_dfl;
kernel/workqueue.c
4820
scope = attrs->affn_scope;
kernel/workqueue.c
4822
pt = &wq_pod_types[scope];
net/core/netdev-genl.c
807
netdev_nl_qstats_get_dump_one(struct net_device *netdev, unsigned int scope,
net/core/netdev-genl.c
814
switch (scope) {
net/core/netdev-genl.c
832
unsigned int scope;
net/core/netdev-genl.c
835
scope = 0;
net/core/netdev-genl.c
837
scope = nla_get_uint(info->attrs[NETDEV_A_QSTATS_SCOPE]);
net/core/netdev-genl.c
851
err = netdev_nl_qstats_get_dump_one(netdev, scope, skb,
net/core/netdev-genl.c
863
err = netdev_nl_qstats_get_dump_one(netdev, scope, skb,
net/core/pktgen.c
2262
if ((ifp->scope & IFA_LINK) &&
net/ipv4/arp.c
400
int scope;
net/ipv4/arp.c
407
scope = RT_SCOPE_HOST;
net/ipv4/arp.c
413
scope = RT_SCOPE_HOST;
net/ipv4/arp.c
417
scope = RT_SCOPE_LINK;
net/ipv4/arp.c
430
return !inet_confirm_addr(net, in_dev, sip, tip, scope);
net/ipv4/arp.c
436
int scope = RT_SCOPE_LINK;
net/ipv4/arp.c
447
return !!inet_confirm_addr(net, in_dev, sip, 0, scope);
net/ipv4/devinet.c
1344
int scope)
net/ipv4/devinet.c
1352
ifa->ifa_scope <= scope)
net/ipv4/devinet.c
1359
__be32 inet_select_addr(const struct net_device *dev, __be32 dst, int scope)
net/ipv4/devinet.c
1380
if (min(ifa->ifa_scope, localnet_scope) > scope)
net/ipv4/devinet.c
1403
addr = in_dev_select_addr(in_dev, scope);
net/ipv4/devinet.c
1420
addr = in_dev_select_addr(in_dev, scope);
net/ipv4/devinet.c
1431
__be32 local, int scope)
net/ipv4/devinet.c
1446
min_scope <= scope) {
net/ipv4/devinet.c
1461
if (min_scope <= scope) {
net/ipv4/devinet.c
1483
__be32 dst, __be32 local, int scope)
net/ipv4/devinet.c
1489
return confirm_addr_indev(in_dev, dst, local, scope);
net/ipv4/devinet.c
1495
addr = confirm_addr_indev(in_dev, dst, local, scope);
net/ipv4/fib_frontend.c
1396
frn->scope = res.scope;
net/ipv4/fib_frontend.c
279
int scope;
net/ipv4/fib_frontend.c
290
scope = RT_SCOPE_UNIVERSE;
net/ipv4/fib_frontend.c
298
.flowi4_scope = scope,
net/ipv4/fib_frontend.c
304
scope = RT_SCOPE_LINK;
net/ipv4/fib_frontend.c
307
return inet_select_addr(dev, ip_hdr(skb)->saddr, scope);
net/ipv4/fib_lookup.h
60
u8 scope;
net/ipv4/fib_semantics.c
103
.scope = RT_SCOPE_LINK,
net/ipv4/fib_semantics.c
107
.scope = RT_SCOPE_LINK,
net/ipv4/fib_semantics.c
111
.scope = RT_SCOPE_UNIVERSE,
net/ipv4/fib_semantics.c
1144
u8 scope, struct netlink_ext_ack *extack)
net/ipv4/fib_semantics.c
115
.scope = RT_SCOPE_UNIVERSE,
net/ipv4/fib_semantics.c
1153
if (scope >= RT_SCOPE_LINK) {
net/ipv4/fib_semantics.c
1183
.flowi4_scope = scope + 1,
net/ipv4/fib_semantics.c
119
.scope = RT_SCOPE_UNIVERSE,
net/ipv4/fib_semantics.c
1220
nh->fib_nh_scope = res.scope;
net/ipv4/fib_semantics.c
123
.scope = RT_SCOPE_UNIVERSE,
net/ipv4/fib_semantics.c
127
.scope = RT_SCOPE_UNIVERSE,
net/ipv4/fib_semantics.c
1272
int fib_check_nh(struct net *net, struct fib_nh *nh, u32 table, u8 scope,
net/ipv4/fib_semantics.c
1278
err = fib_check_nh_v4_gw(net, nh, table, scope, extack);
net/ipv4/fib_semantics.c
1288
unsigned char scope)
net/ipv4/fib_semantics.c
1294
return inet_select_addr(nhc->nhc_dev, 0, scope);
net/ipv4/fib_semantics.c
1297
saddr = inet_select_addr(nh->fib_nh_dev, nh->fib_nh_gw4, scope);
net/ipv4/fib_semantics.c
131
.scope = RT_SCOPE_NOWHERE,
net/ipv4/fib_semantics.c
135
.scope = RT_SCOPE_NOWHERE,
net/ipv4/fib_semantics.c
1363
if (fib_props[cfg->fc_type].scope > cfg->fc_scope) {
net/ipv4/fib_semantics.c
1933
int scope = RT_SCOPE_NOWHERE;
net/ipv4/fib_semantics.c
1938
scope = -1;
net/ipv4/fib_semantics.c
1954
nexthop_nh->fib_nh_scope != scope) {
net/ipv4/fib_semantics.c
2028
if (next_fi->fib_scope != res->scope ||
net/ipv4/fib_semantics.c
316
static unsigned int fib_info_hashfn_1(int init_val, u8 protocol, u8 scope,
net/ipv4/fib_semantics.c
321
val ^= (protocol << 8) | scope;
net/ipv4/fib_semantics.c
91
.scope = RT_SCOPE_NOWHERE,
net/ipv4/fib_semantics.c
95
.scope = RT_SCOPE_UNIVERSE,
net/ipv4/fib_semantics.c
99
.scope = RT_SCOPE_HOST,
net/ipv4/fib_trie.c
1605
res->scope = fi->fib_scope;
net/ipv4/nexthop.c
1612
static int nexthop_check_scope(struct nh_info *nhi, u8 scope,
net/ipv4/nexthop.c
1615
if (scope == RT_SCOPE_HOST && nhi->fib_nhc.nhc_gw_family) {
net/ipv4/nexthop.c
1621
if (nhi->fib_nhc.nhc_flags & RTNH_F_ONLINK && scope >= RT_SCOPE_LINK) {
net/ipv4/nexthop.c
1633
int fib_check_nexthop(struct nexthop *nh, u8 scope,
net/ipv4/nexthop.c
1649
if (scope == RT_SCOPE_HOST) {
net/ipv4/nexthop.c
1657
err = nexthop_check_scope(nhi, scope, extack);
net/ipv4/nexthop.c
1665
err = nexthop_check_scope(nhi, scope, extack);
net/ipv4/ping.c
710
u8 scope;
net/ipv4/ping.c
773
scope = ip_sendmsg_scope(inet, &ipc, msg);
net/ipv4/ping.c
784
ipc.tos & INET_DSCP_MASK, scope,
net/ipv4/raw.c
491
u8 scope;
net/ipv4/raw.c
585
scope = ip_sendmsg_scope(inet, &ipc, msg);
net/ipv4/raw.c
610
ipc.tos & INET_DSCP_MASK, scope,
net/ipv4/route.c
515
__u8 scope = RT_SCOPE_UNIVERSE;
net/ipv4/route.c
521
scope = ip_sock_rt_scope(sk);
net/ipv4/route.c
526
flowi4_init_output(fl4, oif, mark, tos & INET_DSCP_MASK, scope,
net/ipv4/udp.c
1285
u8 scope;
net/ipv4/udp.c
1408
scope = ip_sendmsg_scope(inet, &ipc, msg);
net/ipv4/udp.c
1409
if (scope == RT_SCOPE_LINK)
net/ipv4/udp.c
1445
ipc.tos & INET_DSCP_MASK, scope,
net/ipv6/addrconf.c
1141
ifa->scope = cfg->scope;
net/ipv6/addrconf.c
1479
cfg.scope = ipv6_addr_scope(cfg.pfx);
net/ipv6/addrconf.c
1537
int scope;
net/ipv6/addrconf.c
1637
if (ret >= dst->scope)
net/ipv6/addrconf.c
1828
dst.scope = __ipv6_addr_src_scope(dst_type);
net/ipv6/addrconf.c
1856
dst.scope <= IPV6_ADDR_SCOPE_LINKLOCAL ||
net/ipv6/addrconf.c
1918
if (ifp->scope > IFA_LINK)
net/ipv6/addrconf.c
1920
if (ifp->scope == IFA_LINK &&
net/ipv6/addrconf.c
2007
!(ifp->scope&(IFA_LINK|IFA_HOST) || strict))) {
net/ipv6/addrconf.c
2100
!(ifp->scope&(IFA_LINK|IFA_HOST) || strict)) {
net/ipv6/addrconf.c
2188
.scope = ifp->scope,
net/ipv6/addrconf.c
2409
if (ifp->scope > IFA_LINK)
net/ipv6/addrconf.c
2411
if (ifp->scope == IFA_LINK && !(ifp->flags&IFA_F_TENTATIVE)) {
net/ipv6/addrconf.c
2689
.scope = addr_type & IPV6_ADDR_SCOPE_MASK,
net/ipv6/addrconf.c
3047
cfg->scope = ipv6_addr_scope(cfg->pfx);
net/ipv6/addrconf.c
3187
int plen, int scope, u8 proto)
net/ipv6/addrconf.c
3196
.scope = scope,
net/ipv6/addrconf.c
3217
int scope, plen;
net/ipv6/addrconf.c
3226
scope = IPV6_ADDR_COMPATv4;
net/ipv6/addrconf.c
3234
scope = IFA_LINK;
net/ipv6/addrconf.c
3239
add_addr(idev, &addr, plen, scope, IFAPROT_UNSPEC);
net/ipv6/addrconf.c
3249
int flag = scope;
net/ipv6/addrconf.c
3298
.scope = IFA_LINK,
net/ipv6/addrconf.c
4300
if (ifpiter->scope > IFA_LINK)
net/ipv6/addrconf.c
4302
if (ifp != ifpiter && ifpiter->scope == IFA_LINK &&
net/ipv6/addrconf.c
4331
send_mld = ifp->scope == IFA_LINK && ipv6_lonely_lladdr(ifp);
net/ipv6/addrconf.c
4494
ifp->scope,
net/ipv6/addrconf.c
5085
u8 scope, int ifindex)
net/ipv6/addrconf.c
5093
ifm->ifa_scope = scope;
net/ipv6/addrconf.c
5148
put_ifaddrmsg(nlh, ifa->prefix_len, ifa->flags, rt_scope(ifa->scope),
net/ipv6/addrconf.c
5216
u8 scope = RT_SCOPE_UNIVERSE;
net/ipv6/addrconf.c
5221
scope = RT_SCOPE_SITE;
net/ipv6/addrconf.c
5234
put_ifaddrmsg(nlh, 128, IFA_F_PERMANENT, scope, ifindex);
net/ipv6/addrconf.c
5252
u8 scope = RT_SCOPE_UNIVERSE;
net/ipv6/addrconf.c
5256
scope = RT_SCOPE_SITE;
net/ipv6/addrconf.c
5269
put_ifaddrmsg(nlh, 128, IFA_F_PERMANENT, scope, ifindex);
net/ipv6/addrconf_core.c
20
#define IPV6_ADDR_SCOPE_TYPE(scope) ((scope) << 16)
net/ipv6/addrconf_core.c
22
static inline unsigned int ipv6_addr_scope2type(unsigned int scope)
net/ipv6/addrconf_core.c
24
switch (scope) {
net/ipv6/addrconf_core.c
35
return IPV6_ADDR_SCOPE_TYPE(scope);
net/netfilter/nf_nat_redirect.c
112
unsigned int scope = ipv6_addr_scope(&ipv6_hdr(skb)->daddr);
net/netfilter/nf_nat_redirect.c
122
if (!nf_nat_redirect_ipv6_usable(ifa, scope))
net/netfilter/nf_nat_redirect.c
83
static bool nf_nat_redirect_ipv6_usable(const struct inet6_ifaddr *ifa, unsigned int scope)
net/netfilter/nf_nat_redirect.c
93
if (scope) {
net/netfilter/nf_nat_redirect.c
96
if (!(scope & ifa_scope))
net/sctp/associola.c
1555
enum sctp_scope scope, gfp_t gfp)
net/sctp/associola.c
1574
scope, gfp, flags);
net/sctp/associola.c
288
enum sctp_scope scope, gfp_t gfp)
net/sctp/associola.c
296
if (!sctp_association_init(asoc, ep, sk, scope, gfp))
net/sctp/associola.c
55
enum sctp_scope scope, gfp_t gfp)
net/sctp/bind_addr.c
34
union sctp_addr *addr, enum sctp_scope scope,
net/sctp/bind_addr.c
45
enum sctp_scope scope, gfp_t gfp,
net/sctp/bind_addr.c
460
union sctp_addr *addr, enum sctp_scope scope,
net/sctp/bind_addr.c
466
error = sctp_copy_local_addr_list(net, dest, scope, gfp, flags);
net/sctp/bind_addr.c
467
} else if (sctp_in_scope(net, addr, scope)) {
net/sctp/bind_addr.c
506
enum sctp_scope scope)
net/sctp/bind_addr.c
528
if (addr_scope <= scope)
net/sctp/bind_addr.c
532
if (addr_scope <= scope || SCTP_SCOPE_PRIVATE == addr_scope)
net/sctp/bind_addr.c
536
if (addr_scope <= scope || SCTP_SCOPE_LINK == addr_scope)
net/sctp/bind_addr.c
56
error = sctp_copy_one_addr(net, dest, &addr->a, scope,
net/sctp/bind_addr.c
574
return af->scope((union sctp_addr *)addr);
net/sctp/bind_addr.c
66
if (list_empty(&dest->address_list) && (SCTP_SCOPE_GLOBAL == scope)) {
net/sctp/ipv6.c
1105
.scope = sctp_v6_scope,
net/sctp/ipv6.c
286
enum sctp_scope scope;
net/sctp/ipv6.c
335
scope = sctp_scope(daddr);
net/sctp/ipv6.c
379
scope > sctp_scope(&laddr->a))
net/sctp/protocol.c
1172
.scope = sctp_v4_scope,
net/sctp/protocol.c
136
enum sctp_scope scope, gfp_t gfp, int copy_flags)
net/sctp/protocol.c
146
if (!sctp_in_scope(net, &addr->a, scope))
net/sctp/sm_make_chunk.c
1625
enum sctp_scope scope;
net/sctp/sm_make_chunk.c
1629
scope = sctp_scope(sctp_source(chunk));
net/sctp/sm_make_chunk.c
1630
asoc = sctp_association_new(ep, ep->base.sk, scope, gfp);
net/sctp/sm_make_chunk.c
1733
enum sctp_scope scope;
net/sctp/sm_make_chunk.c
1830
scope = sctp_scope(sctp_source(chunk));
net/sctp/sm_make_chunk.c
1831
retval = sctp_association_new(ep, ep->base.sk, scope, gfp);
net/sctp/sm_make_chunk.c
2519
enum sctp_scope scope;
net/sctp/sm_make_chunk.c
2544
scope = sctp_scope(peer_addr);
net/sctp/sm_make_chunk.c
2545
if (sctp_in_scope(net, &addr, scope))
net/sctp/socket.c
1076
enum sctp_scope scope;
net/sctp/socket.c
1091
scope = sctp_scope(daddr);
net/sctp/socket.c
1092
asoc = sctp_association_new(ep, sk, scope, GFP_KERNEL);
net/sctp/socket.c
1096
err = sctp_assoc_set_bind_addr_from_ep(asoc, scope, GFP_KERNEL);
net/shaper/shaper.c
1074
if (WARN_ON_ONCE(node.handle.scope != NET_SHAPER_SCOPE_NETDEV))
net/shaper/shaper.c
1099
update_node = node.handle.scope != NET_SHAPER_SCOPE_NETDEV;
net/shaper/shaper.c
113
if (handle->scope == NET_SHAPER_SCOPE_UNSPEC)
net/shaper/shaper.c
1134
if (handle.scope == NET_SHAPER_SCOPE_NODE) {
net/shaper/shaper.c
120
if (nla_put_u32(msg, NET_SHAPER_A_HANDLE_SCOPE, handle->scope) ||
net/shaper/shaper.c
121
(handle->scope >= NET_SHAPER_SCOPE_QUEUE &&
net/shaper/shaper.c
1237
if (leaves[i].parent.scope == NET_SHAPER_SCOPE_NODE &&
net/shaper/shaper.c
1286
enum net_shaper_scope scope, unsigned long flags,
net/shaper/shaper.c
1297
nla_put_u32(msg, NET_SHAPER_A_CAPS_SCOPE, scope))
net/shaper/shaper.c
1319
enum net_shaper_scope scope;
net/shaper/shaper.c
1328
scope = nla_get_u32(info->attrs[NET_SHAPER_A_CAPS_SCOPE]);
net/shaper/shaper.c
1330
ops->capabilities(binding, scope, &flags);
net/shaper/shaper.c
1338
ret = net_shaper_cap_fill_one(msg, binding, scope, flags, info);
net/shaper/shaper.c
1355
enum net_shaper_scope scope;
net/shaper/shaper.c
1360
for (scope = 0; scope <= NET_SHAPER_SCOPE_MAX; ++scope) {
net/shaper/shaper.c
1363
ops->capabilities(binding, scope, &flags);
net/shaper/shaper.c
1367
ret = net_shaper_cap_fill_one(skb, binding, scope, flags,
net/shaper/shaper.c
1429
handle.scope = NET_SHAPER_SCOPE_QUEUE;
net/shaper/shaper.c
1444
if (parent_handle.scope != NET_SHAPER_SCOPE_NODE)
net/shaper/shaper.c
249
return FIELD_PREP(NET_SHAPER_SCOPE_MASK, handle->scope) |
net/shaper/shaper.c
256
handle->scope = FIELD_GET(NET_SHAPER_SCOPE_MASK, index);
net/shaper/shaper.c
263
switch (handle->scope) {
net/shaper/shaper.c
267
parent->scope = NET_SHAPER_SCOPE_UNSPEC;
net/shaper/shaper.c
272
parent->scope = NET_SHAPER_SCOPE_NETDEV;
net/shaper/shaper.c
347
if (handle->scope == NET_SHAPER_SCOPE_NODE &&
net/shaper/shaper.c
460
handle->scope = nla_get_u32(tb[NET_SHAPER_A_HANDLE_SCOPE]);
net/shaper/shaper.c
470
else if (handle->scope == NET_SHAPER_SCOPE_NODE)
net/shaper/shaper.c
486
ops->capabilities(binding, shaper->handle.scope, &caps);
net/shaper/shaper.c
512
if (shaper->handle.scope == NET_SHAPER_SCOPE_QUEUE &&
net/shaper/shaper.c
563
if (shaper->handle.scope == NET_SHAPER_SCOPE_UNSPEC) {
net/shaper/shaper.c
608
ops->capabilities(binding, shaper->handle.scope, &caps);
net/shaper/shaper.c
612
shaper->handle.scope);
net/shaper/shaper.c
640
if (shaper->handle.scope != NET_SHAPER_SCOPE_QUEUE) {
net/shaper/shaper.c
645
if (node->handle.scope == NET_SHAPER_SCOPE_NODE) {
net/shaper/shaper.c
672
if (shaper->handle.scope != NET_SHAPER_SCOPE_NODE &&
net/shaper/shaper.c
673
shaper->handle.scope != NET_SHAPER_SCOPE_NETDEV) {
net/shaper/shaper.c
684
if (shaper->parent.scope != NET_SHAPER_SCOPE_NODE &&
net/shaper/shaper.c
685
shaper->parent.scope != NET_SHAPER_SCOPE_NETDEV) {
net/shaper/shaper.c
876
if (handle.scope == NET_SHAPER_SCOPE_NODE &&
net/shaper/shaper.c
916
if (parent_handle.scope == NET_SHAPER_SCOPE_NODE) {
net/shaper/shaper.c
966
if (node->handle.scope == NET_SHAPER_SCOPE_NODE) {
net/shaper/shaper.c
974
node->handle.scope, node->handle.id);
net/shaper/shaper.c
981
if (node->parent.scope == NET_SHAPER_SCOPE_UNSPEC) {
net/shaper/shaper.c
993
if (node->parent.scope == NET_SHAPER_SCOPE_NODE) {
net/shaper/shaper.c
997
node->parent.scope, node->parent.id);
net/tipc/addr.h
53
signed char scope;
net/tipc/addr.h
64
static inline void tipc_uaddr(struct tipc_uaddr *ua, u32 atype, u32 scope,
net/tipc/addr.h
69
ua->scope = scope;
net/tipc/group.c
168
bool global = mreq->scope != TIPC_NODE_SCOPE;
net/tipc/group.c
184
grp->scope = mreq->scope;
net/tipc/group.c
364
int *scope)
net/tipc/group.c
369
*scope = grp->scope;
net/tipc/group.c
754
if (grp->scope == TIPC_NODE_SCOPE && node != tipc_own_addr(grp->net))
net/tipc/group.c
90
u32 scope;
net/tipc/group.c
945
if (grp->scope == TIPC_NODE_SCOPE)
net/tipc/group.c
949
if (grp->scope == TIPC_CLUSTER_SCOPE)
net/tipc/group.h
55
int *scope);
net/tipc/msg.c
704
u32 scope = msg_lookup_scope(msg);
net/tipc/msg.c
722
tipc_uaddr(&ua, TIPC_SERVICE_RANGE, scope,
net/tipc/msg.c
724
sk.node = tipc_scope2node(net, scope);
net/tipc/name_distr.c
127
if (p->scope == TIPC_NODE_SCOPE)
net/tipc/name_distr.c
241
tipc_uaddr(&ua, TIPC_SERVICE_RANGE, p->scope, p->sr.type,
net/tipc/name_distr.c
94
if (p->scope == TIPC_NODE_SCOPE) {
net/tipc/name_table.c
1003
if (nla_put_u32(msg->skb, TIPC_NLA_PUBL_SCOPE, p->scope))
net/tipc/name_table.c
240
p->scope = ua->scope;
net/tipc/name_table.c
654
if (p->scope != ua->scope)
net/tipc/name_table.c
683
u8 scope = ua->scope;
net/tipc/name_table.c
693
if (scope == p->scope || scope == TIPC_ANY_SCOPE)
net/tipc/name_table.c
749
if (p->scope != ua->scope)
net/tipc/name_table.h
80
u16 scope;
net/tipc/socket.c
1018
ua->scope = msg_lookup_scope(hdr);
net/tipc/socket.c
1167
ua->scope = msg_lookup_scope(hdr);
net/tipc/socket.c
1218
ua.scope = TIPC_ANY_SCOPE;
net/tipc/socket.c
1220
ua.scope = TIPC_CLUSTER_SCOPE;
net/tipc/socket.c
1237
ua.scope = msg_lookup_scope(hdr);
net/tipc/socket.c
1468
ua->scope = tipc_node2scope(skaddr.node);
net/tipc/socket.c
1491
msg_set_lookup_scope(hdr, ua->scope);
net/tipc/socket.c
1703
srcaddr->sock.scope = 0;
net/tipc/socket.c
1715
srcaddr->member.scope = 0;
net/tipc/socket.c
2936
tipc_uaddr(&_ua, TIPC_SERVICE_RANGE, p->scope,
net/tipc/socket.c
2942
if (p->scope != ua->scope)
net/tipc/socket.c
3077
if (mreq->scope > TIPC_NODE_SCOPE)
net/tipc/socket.c
3079
if (mreq->scope != TIPC_NODE_SCOPE)
net/tipc/socket.c
3080
mreq->scope = TIPC_CLUSTER_SCOPE;
net/tipc/socket.c
3087
msg_set_lookup_scope(hdr, mreq->scope);
net/tipc/socket.c
3090
tipc_uaddr(&ua, TIPC_SERVICE_RANGE, mreq->scope,
net/tipc/socket.c
3111
int scope;
net/tipc/socket.c
3116
tipc_group_self(grp, &ua.sr, &scope);
net/tipc/socket.c
3117
ua.scope = scope;
net/tipc/socket.c
3237
int len, scope;
net/tipc/socket.c
3277
tipc_group_self(tsk->group, &seq, &scope);
net/tipc/socket.c
3338
tsk1->peer.scope = TIPC_NODE_SCOPE;
net/tipc/socket.c
3343
tsk2->peer.scope = TIPC_NODE_SCOPE;
net/tipc/socket.c
687
if (ua->scope < 0) {
net/tipc/socket.c
689
ua->scope = -ua->scope;
net/tipc/socket.c
692
if (ua->scope != TIPC_NODE_SCOPE)
net/tipc/socket.c
693
ua->scope = TIPC_CLUSTER_SCOPE;
net/tipc/socket.c
765
addr->scope = 0;
net/tipc/subscr.c
97
if (filter & TIPC_SUB_CLUSTER_SCOPE && p->scope == TIPC_NODE_SCOPE)
net/tipc/subscr.c
99
if (filter & TIPC_SUB_NODE_SCOPE && p->scope != TIPC_NODE_SCOPE)
net/tipc/topsrv.c
540
saddr.scope = TIPC_NODE_SCOPE;
security/landlock/access.h
52
access_mask_t scope : LANDLOCK_NUM_SCOPE;
security/landlock/fs.c
1797
.scope = LANDLOCK_SCOPE_SIGNAL,
security/landlock/ruleset.h
278
ruleset->access_masks[layer_level].scope |= mask;
security/landlock/ruleset.h
301
return ruleset->access_masks[layer_level].scope;
security/landlock/task.c
181
access_mask_t scope)
security/landlock/task.c
207
if (landlock_get_scope_mask(client, client_layer) & scope)
security/landlock/task.c
220
if (landlock_get_scope_mask(client, client_layer) & scope) {
security/landlock/task.c
262
.scope = LANDLOCK_SCOPE_ABSTRACT_UNIX_SOCKET,
security/landlock/task.c
335
.scope = LANDLOCK_SCOPE_SIGNAL,
security/landlock/task.c
373
signal_scope.scope);
security/landlock/task.c
414
signal_scope.scope);
tools/counter/counter_example.c
23
.component.scope = COUNTER_SCOPE_COUNT,
tools/counter/counter_example.c
33
.component.scope = COUNTER_SCOPE_COUNT,
tools/counter/counter_watch_events.c
24
.component.scope = COUNTER_SCOPE_COUNT,
tools/counter/counter_watch_events.c
266
watches[i].component.scope = ret;
tools/counter/counter_watch_events.c
76
counter_scope_name[watch[i].component.scope],
tools/include/linux/btf_ids.h
135
#define __BTF_SET_START(name, scope) \
tools/include/linux/btf_ids.h
138
"." #scope " __BTF_ID__set__" #name "; \n" \
tools/include/linux/btf_ids.h
76
#define __BTF_ID_LIST(name, scope) \
tools/include/linux/btf_ids.h
79
"." #scope " " #name "; \n" \
tools/net/ynl/ynltool/qstats.c
136
if (scope == NETDEV_QSTATS_SCOPE_QUEUE &&
tools/net/ynl/ynltool/qstats.c
15
static enum netdev_qstats_scope scope; /* default - device */
tools/net/ynl/ynltool/qstats.c
241
qstats_dump(enum netdev_qstats_scope scope)
tools/net/ynl/ynltool/qstats.c
260
if (scope)
tools/net/ynl/ynltool/qstats.c
261
netdev_qstats_get_req_set_scope(req, scope);
tools/net/ynl/ynltool/qstats.c
291
scope = NETDEV_QSTATS_SCOPE_QUEUE;
tools/net/ynl/ynltool/qstats.c
293
scope = 0;
tools/net/ynl/ynltool/qstats.c
305
qstats = qstats_dump(scope);
tools/power/x86/turbostat/turbostat.c
10249
struct pmt_counter **pmt_get_scope_root(enum counter_scope scope)
tools/power/x86/turbostat/turbostat.c
10251
switch (scope) {
tools/power/x86/turbostat/turbostat.c
10276
unsigned int lsb, unsigned int msb, unsigned int offset, enum counter_scope scope,
tools/power/x86/turbostat/turbostat.c
10281
struct pmt_counter **const pmt_root = pmt_get_scope_root(scope);
tools/power/x86/turbostat/turbostat.c
10323
pcounter->scope = scope;
tools/power/x86/turbostat/turbostat.c
10329
conflict += pcounter->scope != scope;
tools/power/x86/turbostat/turbostat.c
10575
unsigned int width, enum counter_scope scope, enum counter_type type, enum counter_format format, int flags, int id)
tools/power/x86/turbostat/turbostat.c
10584
__func__, msr_num, path, name, width, scope, type, format, flags, id);
tools/power/x86/turbostat/turbostat.c
10586
switch (scope) {
tools/power/x86/turbostat/turbostat.c
10641
switch (scope) {
tools/power/x86/turbostat/turbostat.c
10684
unsigned int width, enum counter_scope scope, enum counter_type type, enum counter_format format)
tools/power/x86/turbostat/turbostat.c
10697
pinfo->scope = scope;
tools/power/x86/turbostat/turbostat.c
10705
enum counter_scope scope, enum counter_type type, enum counter_format format)
tools/power/x86/turbostat/turbostat.c
10709
switch (scope) {
tools/power/x86/turbostat/turbostat.c
10732
pinfo = make_perf_counter_info(perf_device, perf_event, name_buffer, width, scope, type, format);
tools/power/x86/turbostat/turbostat.c
10737
switch (scope) {
tools/power/x86/turbostat/turbostat.c
10759
fprintf(stderr, "%s: %s/%s, name: %s, scope%d\n", __func__, pinfo->device, pinfo->event, pinfo->name, pinfo->scope);
tools/power/x86/turbostat/turbostat.c
10773
enum counter_scope scope = SCOPE_CPU;
tools/power/x86/turbostat/turbostat.c
10801
scope = SCOPE_CPU;
tools/power/x86/turbostat/turbostat.c
10805
scope = SCOPE_CORE;
tools/power/x86/turbostat/turbostat.c
10809
scope = SCOPE_PACKAGE;
tools/power/x86/turbostat/turbostat.c
10880
if (add_perf_counter(perf_device, perf_event, name_buffer, width, scope, type, format))
tools/power/x86/turbostat/turbostat.c
10883
if (add_counter(msr_num, path, name_buffer, width, scope, type, format, 0, 0))
tools/power/x86/turbostat/turbostat.c
10991
enum counter_scope scope = 0;
tools/power/x86/turbostat/turbostat.c
11025
scope = SCOPE_CPU;
tools/power/x86/turbostat/turbostat.c
11028
scope = SCOPE_CORE;
tools/power/x86/turbostat/turbostat.c
11031
scope = SCOPE_PACKAGE;
tools/power/x86/turbostat/turbostat.c
11186
pmt_add_counter(guid, seq, name, type, lsb, msb, offset, scope, format, domain_id, PMT_OPEN_REQUIRED);
tools/power/x86/turbostat/turbostat.c
116
enum counter_scope scope;
tools/power/x86/turbostat/turbostat.c
1969
enum counter_scope scope;
tools/power/x86/turbostat/turbostat.c
2656
switch (pc->scope) {
tools/power/x86/turbostat/turbostat.c
572
unsigned int width, enum counter_scope scope, enum counter_type type, enum counter_format format, int flags, int package_num);
tools/power/x86/turbostat/turbostat.c
9928
switch (pinfo->scope) {
tools/testing/selftests/landlock/scoped_common.h
13
const __u16 scope)
tools/testing/selftests/landlock/scoped_common.h
17
.scoped = scope,