Symbol: id4
arch/x86/kernel/cpu/cacheinfo.c
223
static int cpuid4_info_fill_done(struct _cpuid4_info *id4, union _cpuid4_leaf_eax eax,
arch/x86/kernel/cpu/cacheinfo.c
229
id4->eax = eax;
arch/x86/kernel/cpu/cacheinfo.c
230
id4->ebx = ebx;
arch/x86/kernel/cpu/cacheinfo.c
231
id4->ecx = ecx;
arch/x86/kernel/cpu/cacheinfo.c
232
id4->size = (ecx.split.number_of_sets + 1) *
arch/x86/kernel/cpu/cacheinfo.c
240
static int amd_fill_cpuid4_info(int index, struct _cpuid4_info *id4)
arch/x86/kernel/cpu/cacheinfo.c
252
return cpuid4_info_fill_done(id4, eax, ebx, ecx);
arch/x86/kernel/cpu/cacheinfo.c
255
static int intel_fill_cpuid4_info(int index, struct _cpuid4_info *id4)
arch/x86/kernel/cpu/cacheinfo.c
264
return cpuid4_info_fill_done(id4, eax, ebx, ecx);
arch/x86/kernel/cpu/cacheinfo.c
267
static int fill_cpuid4_info(int index, struct _cpuid4_info *id4)
arch/x86/kernel/cpu/cacheinfo.c
272
amd_fill_cpuid4_info(index, id4) :
arch/x86/kernel/cpu/cacheinfo.c
273
intel_fill_cpuid4_info(index, id4);
arch/x86/kernel/cpu/cacheinfo.c
297
static unsigned int get_cache_id(u32 apicid, const struct _cpuid4_info *id4)
arch/x86/kernel/cpu/cacheinfo.c
302
num_threads_sharing = 1 + id4->eax.split.num_threads_sharing;
arch/x86/kernel/cpu/cacheinfo.c
332
struct _cpuid4_info id4 = {};
arch/x86/kernel/cpu/cacheinfo.c
334
if (!amd_fill_cpuid4_info(llc_index, &id4))
arch/x86/kernel/cpu/cacheinfo.c
335
c->topo.llc_id = get_cache_id(c->topo.apicid, &id4);
arch/x86/kernel/cpu/cacheinfo.c
413
static unsigned int calc_cache_topo_id(struct cpuinfo_x86 *c, const struct _cpuid4_info *id4)
arch/x86/kernel/cpu/cacheinfo.c
418
num_threads_sharing = 1 + id4->eax.split.num_threads_sharing;
arch/x86/kernel/cpu/cacheinfo.c
443
struct _cpuid4_info id4 = {};
arch/x86/kernel/cpu/cacheinfo.c
446
ret = intel_fill_cpuid4_info(i, &id4);
arch/x86/kernel/cpu/cacheinfo.c
450
switch (id4.eax.split.level) {
arch/x86/kernel/cpu/cacheinfo.c
452
if (id4.eax.split.type == CTYPE_DATA)
arch/x86/kernel/cpu/cacheinfo.c
453
l1d = id4.size / 1024;
arch/x86/kernel/cpu/cacheinfo.c
454
else if (id4.eax.split.type == CTYPE_INST)
arch/x86/kernel/cpu/cacheinfo.c
455
l1i = id4.size / 1024;
arch/x86/kernel/cpu/cacheinfo.c
458
l2 = id4.size / 1024;
arch/x86/kernel/cpu/cacheinfo.c
459
l2_id = calc_cache_topo_id(c, &id4);
arch/x86/kernel/cpu/cacheinfo.c
462
l3 = id4.size / 1024;
arch/x86/kernel/cpu/cacheinfo.c
463
l3_id = calc_cache_topo_id(c, &id4);
arch/x86/kernel/cpu/cacheinfo.c
489
const struct _cpuid4_info *id4)
arch/x86/kernel/cpu/cacheinfo.c
515
nshared = id4->eax.split.num_threads_sharing + 1;
arch/x86/kernel/cpu/cacheinfo.c
548
const struct _cpuid4_info *id4)
arch/x86/kernel/cpu/cacheinfo.c
557
if (__cache_amd_cpumap_setup(cpu, index, id4))
arch/x86/kernel/cpu/cacheinfo.c
562
num_threads_sharing = 1 + id4->eax.split.num_threads_sharing;
arch/x86/kernel/cpu/cacheinfo.c
584
static void ci_info_init(struct cacheinfo *ci, const struct _cpuid4_info *id4,
arch/x86/kernel/cpu/cacheinfo.c
587
ci->id = id4->id;
arch/x86/kernel/cpu/cacheinfo.c
589
ci->level = id4->eax.split.level;
arch/x86/kernel/cpu/cacheinfo.c
590
ci->type = cache_type_map[id4->eax.split.type];
arch/x86/kernel/cpu/cacheinfo.c
591
ci->coherency_line_size = id4->ebx.split.coherency_line_size + 1;
arch/x86/kernel/cpu/cacheinfo.c
592
ci->ways_of_associativity = id4->ebx.split.ways_of_associativity + 1;
arch/x86/kernel/cpu/cacheinfo.c
593
ci->size = id4->size;
arch/x86/kernel/cpu/cacheinfo.c
594
ci->number_of_sets = id4->ecx.split.number_of_sets + 1;
arch/x86/kernel/cpu/cacheinfo.c
595
ci->physical_line_partition = id4->ebx.split.physical_line_partition + 1;
arch/x86/kernel/cpu/cacheinfo.c
617
struct _cpuid4_info id4 = {};
arch/x86/kernel/cpu/cacheinfo.c
621
ret = fill_cpuid4_info(idx, &id4);
arch/x86/kernel/cpu/cacheinfo.c
625
id4.id = get_cache_id(apicid, &id4);
arch/x86/kernel/cpu/cacheinfo.c
630
ci_info_init(ci++, &id4, nb);
arch/x86/kernel/cpu/cacheinfo.c
631
__cache_cpumap_setup(cpu, idx, &id4);
drivers/gpu/drm/nouveau/dispnv04/tvmodesnv17.c
276
0.051285 * id4, 26.168746 * id4, -4.361449 * id4, -0.000001 * id4,
drivers/gpu/drm/nouveau/dispnv04/tvmodesnv17.c
280
29.832207 * id4, 50.047322 * id4, -25.380017 * id4, 2.546422 * id4,
drivers/gpu/drm/nouveau/dispnv04/tvmodesnv17.c
284
33.541131 * id4, -34.149302 * id4, 5.691537 * id4, 0.000002 * id4,
drivers/gpu/drm/nouveau/dispnv04/tvmodesnv17.c
288
-41.470726 * id4, -17.775823 * id4, 13.057821 * id4, -1.15823 * id4,
drivers/gpu/drm/nouveau/dispnv04/tvmodesnv17.c
296
-3.402339 * id4, 0.000209 * id4, -0.000092 * id4, 0.000010 * id4,
drivers/gpu/drm/nouveau/dispnv04/tvmodesnv17.c
300
8.293120 * id4, -1.192888 * id4, -0.947652 * id4, 0.094507 * id4,
drivers/gpu/drm/nouveau/dispnv04/tvmodesnv17.c
304
39.541978 * id4, 5.680053 * id4, -0.946676 * id4, 0.000000 * id4,
drivers/gpu/drm/nouveau/dispnv04/tvmodesnv17.c
308
20.687300 * id4, 3.014003 * id4, -0.557786 * id4, -0.01311 * id4,
drivers/net/dsa/microchip/ksz_common.c
3995
u8 id1, id2, id4;
drivers/net/dsa/microchip/ksz_common.c
4039
ret = ksz_read8(dev, REG_KSZ8864_CHIP_ID, &id4);
drivers/net/dsa/microchip/ksz_common.c
4042
if (id4 & SW_KSZ8864)
drivers/net/dsa/microchip/ksz_common.c
4071
&id4);
drivers/net/dsa/microchip/ksz_common.c
4075
if (id4 == SKU_ID_KSZ8563)
drivers/net/dsa/microchip/ksz_common.c
4077
else if (id4 == SKU_ID_KSZ9563)
net/mptcp/options.c
1698
u8 id1, id2, id3, id4;
net/mptcp/options.c
1703
id4 = i + 3 < opts->rm_list.nr ? opts->rm_list.ids[i + 3] : TCPOPT_NOP;
net/mptcp/options.c
1704
put_unaligned_be32(id1 << 24 | id2 << 16 | id3 << 8 | id4, ptr);
tools/testing/selftests/bpf/prog_tests/cgroup_mprog_opts.c
114
ASSERT_EQ(optq.prog_ids[3], id4, "prog_ids[3]");
tools/testing/selftests/bpf/prog_tests/cgroup_mprog_opts.c
156
__u32 fd1, fd2, fd3, fd4, id1, id2, id3, id4;
tools/testing/selftests/bpf/prog_tests/cgroup_mprog_opts.c
177
id4 = id_from_prog_fd(fd4);
tools/testing/selftests/bpf/prog_tests/cgroup_mprog_opts.c
23
__u32 fd1, fd2, fd3, fd4, id1, id2, id3, id4;
tools/testing/selftests/bpf/prog_tests/cgroup_mprog_opts.c
246
ASSERT_EQ(optq.prog_ids[3], id4, "prog_ids[3]");
tools/testing/selftests/bpf/prog_tests/cgroup_mprog_opts.c
44
id4 = id_from_prog_fd(fd4);
tools/testing/selftests/bpf/prog_tests/tc_opts.c
1031
__u32 fd1, fd2, fd3, fd4, id1, id2, id3, id4;
tools/testing/selftests/bpf/prog_tests/tc_opts.c
1048
id4 = id_from_prog_fd(fd4);
tools/testing/selftests/bpf/prog_tests/tc_opts.c
1051
ASSERT_NEQ(id3, id4, "prog_ids_3_4");
tools/testing/selftests/bpf/prog_tests/tc_opts.c
1122
ASSERT_EQ(optq.prog_ids[0], id4, "prog_ids[0]");
tools/testing/selftests/bpf/prog_tests/tc_opts.c
116
__u32 fd1, fd2, fd3, fd4, id1, id2, id3, id4;
tools/testing/selftests/bpf/prog_tests/tc_opts.c
1171
__u32 fd1, fd2, fd3, fd4, id1, id2, id3, id4;
tools/testing/selftests/bpf/prog_tests/tc_opts.c
1188
id4 = id_from_prog_fd(fd4);
tools/testing/selftests/bpf/prog_tests/tc_opts.c
1191
ASSERT_NEQ(id3, id4, "prog_ids_3_4");
tools/testing/selftests/bpf/prog_tests/tc_opts.c
1265
ASSERT_EQ(optq.prog_ids[3], id4, "prog_ids[3]");
tools/testing/selftests/bpf/prog_tests/tc_opts.c
1311
__u32 fd1, fd2, fd3, fd4, id1, id2, id3, id4;
tools/testing/selftests/bpf/prog_tests/tc_opts.c
133
id4 = id_from_prog_fd(fd4);
tools/testing/selftests/bpf/prog_tests/tc_opts.c
1331
id4 = id_from_prog_fd(fd4);
tools/testing/selftests/bpf/prog_tests/tc_opts.c
1334
ASSERT_NEQ(id3, id4, "prog_ids_3_4");
tools/testing/selftests/bpf/prog_tests/tc_opts.c
136
ASSERT_NEQ(id3, id4, "prog_ids_3_4");
tools/testing/selftests/bpf/prog_tests/tc_opts.c
1654
__u32 fd1, fd2, fd3, fd4, id1, id2, id3, id4;
tools/testing/selftests/bpf/prog_tests/tc_opts.c
1671
id4 = id_from_prog_fd(fd4);
tools/testing/selftests/bpf/prog_tests/tc_opts.c
1674
ASSERT_NEQ(id3, id4, "prog_ids_3_4");
tools/testing/selftests/bpf/prog_tests/tc_opts.c
1717
ASSERT_EQ(optq.prog_ids[3], id4, "prog_ids[3]");
tools/testing/selftests/bpf/prog_tests/tc_opts.c
1740
ASSERT_EQ(optq.prog_ids[2], id4, "prog_ids[2]");
tools/testing/selftests/bpf/prog_tests/tc_opts.c
1825
__u32 fd1, fd2, fd3, fd4, id1, id2, id3, id4;
tools/testing/selftests/bpf/prog_tests/tc_opts.c
1842
id4 = id_from_prog_fd(fd4);
tools/testing/selftests/bpf/prog_tests/tc_opts.c
1845
ASSERT_NEQ(id3, id4, "prog_ids_3_4");
tools/testing/selftests/bpf/prog_tests/tc_opts.c
1888
ASSERT_EQ(optq.prog_ids[3], id4, "prog_ids[3]");
tools/testing/selftests/bpf/prog_tests/tc_opts.c
1912
ASSERT_EQ(optq.prog_ids[2], id4, "prog_ids[2]");
tools/testing/selftests/bpf/prog_tests/tc_opts.c
1962
ASSERT_EQ(optq.prog_ids[1], id4, "prog_ids[1]");
tools/testing/selftests/bpf/prog_tests/tc_opts.c
1984
ASSERT_EQ(optq.prog_ids[0], id4, "prog_ids[0]");
tools/testing/selftests/bpf/prog_tests/tc_opts.c
2032
__u32 fd1, fd2, fd3, fd4, id1, id2, id3, id4;
tools/testing/selftests/bpf/prog_tests/tc_opts.c
2049
id4 = id_from_prog_fd(fd4);
tools/testing/selftests/bpf/prog_tests/tc_opts.c
2052
ASSERT_NEQ(id3, id4, "prog_ids_3_4");
tools/testing/selftests/bpf/prog_tests/tc_opts.c
2095
ASSERT_EQ(optq.prog_ids[3], id4, "prog_ids[3]");
tools/testing/selftests/bpf/prog_tests/tc_opts.c
2119
ASSERT_EQ(optq.prog_ids[2], id4, "prog_ids[2]");
tools/testing/selftests/bpf/prog_tests/tc_opts.c
2178
ASSERT_EQ(optq.prog_ids[1], id4, "prog_ids[1]");
tools/testing/selftests/bpf/prog_tests/tc_opts.c
219
ASSERT_EQ(optq.prog_ids[0], id4, "prog_ids[0]");
tools/testing/selftests/bpf/prog_tests/tc_opts.c
2465
__u32 fd1, fd2, fd3, fd4, id1, id2, id3, id4;
tools/testing/selftests/bpf/prog_tests/tc_opts.c
2483
id4 = id_from_prog_fd(fd4);
tools/testing/selftests/bpf/prog_tests/tc_opts.c
2549
ASSERT_EQ(optq.prog_ids[3], id4, "prog_ids[3]");
tools/testing/selftests/bpf/prog_tests/tc_opts.c
2590
ASSERT_EQ(prog_ids[3], id4, "prog_ids[3]");
tools/testing/selftests/bpf/prog_tests/tc_opts.c
2648
ASSERT_EQ(prog_ids[3], id4, "prog_ids[3]");
tools/testing/selftests/bpf/prog_tests/tc_opts.c
268
__u32 fd1, fd2, fd3, fd4, id1, id2, id3, id4;
tools/testing/selftests/bpf/prog_tests/tc_opts.c
285
id4 = id_from_prog_fd(fd4);
tools/testing/selftests/bpf/prog_tests/tc_opts.c
288
ASSERT_NEQ(id3, id4, "prog_ids_3_4");
tools/testing/selftests/bpf/prog_tests/tc_opts.c
374
ASSERT_EQ(optq.prog_ids[3], id4, "prog_ids[3]");