arch/arc/include/asm/arcregs.h
224
unsigned int pad:7, c:1, num_entries:8, num_cores:8, ver:8;
arch/arc/include/asm/arcregs.h
226
unsigned int ver:8, num_cores:8, num_entries:8, c:1, pad:7;
arch/arm/probes/kprobes/test-core.c
699
unsigned num_entries;
arch/arm/probes/kprobes/test-core.c
738
struct coverage_entry *entry = coverage->base + coverage->num_entries;
arch/arm/probes/kprobes/test-core.c
740
if (coverage->num_entries == MAX_COVERAGE_ENTRIES - 1) {
arch/arm/probes/kprobes/test-core.c
745
++coverage->num_entries;
arch/arm/probes/kprobes/test-core.c
768
coverage.num_entries = 0;
arch/arm/probes/kprobes/test-core.c
844
struct coverage_entry *end = coverage.base + coverage.num_entries;
arch/arm/probes/kprobes/test-core.c
891
struct coverage_entry *end = coverage.base + coverage.num_entries;
arch/loongarch/include/asm/module.h
120
for (i = 0; i < sec->num_entries; i++)
arch/loongarch/include/asm/module.h
16
int num_entries;
arch/loongarch/include/asm/module.h
91
for (i = 0; i < sec->num_entries; i++) {
arch/loongarch/kernel/module-sections.c
159
mod->arch.got.num_entries = 0;
arch/loongarch/kernel/module-sections.c
16
int i = got_sec->num_entries;
arch/loongarch/kernel/module-sections.c
167
mod->arch.plt.num_entries = 0;
arch/loongarch/kernel/module-sections.c
175
mod->arch.plt_idx.num_entries = 0;
arch/loongarch/kernel/module-sections.c
26
got_sec->num_entries++;
arch/loongarch/kernel/module-sections.c
27
if (got_sec->num_entries > got_sec->max_entries) {
arch/loongarch/kernel/module-sections.c
50
nr = plt_sec->num_entries;
arch/loongarch/kernel/module-sections.c
58
plt_sec->num_entries++;
arch/loongarch/kernel/module-sections.c
59
plt_idx_sec->num_entries++;
arch/loongarch/kernel/module-sections.c
60
BUG_ON(plt_sec->num_entries > plt_sec->max_entries);
arch/loongarch/kernel/unwind_orc.c
248
unsigned int num_entries = orc_ip_size / sizeof(int);
arch/loongarch/kernel/unwind_orc.c
252
num_entries != orc_size / sizeof(*orc));
arch/loongarch/kernel/unwind_orc.c
262
sort(orc_ip, num_entries, sizeof(int), orc_sort_cmp, orc_sort_swap);
arch/loongarch/kernel/unwind_orc.c
267
mod->arch.num_orcs = num_entries;
arch/loongarch/kernel/unwind_orc.c
276
size_t num_entries = orc_ip_size / sizeof(int);
arch/loongarch/kernel/unwind_orc.c
279
if (!num_entries || orc_ip_size % sizeof(int) != 0 ||
arch/loongarch/kernel/unwind_orc.c
281
num_entries != orc_size / sizeof(struct orc_entry)) {
arch/loongarch/kernel/unwind_orc.c
296
num_entries, LOOKUP_START_IP + (LOOKUP_BLOCK_SIZE * i));
arch/loongarch/kernel/unwind_orc.c
306
orc = __orc_find(__start_orc_unwind_ip, __start_orc_unwind, num_entries, LOOKUP_STOP_IP);
arch/loongarch/kernel/unwind_orc.c
60
unsigned int num_entries, unsigned long ip)
arch/loongarch/kernel/unwind_orc.c
64
int *last = ip_table + num_entries - 1;
arch/loongarch/kernel/unwind_orc.c
66
if (!num_entries)
arch/parisc/include/asm/pdc.h
62
int pdc_pci_irt_size(unsigned long *num_entries, unsigned long hpa);
arch/parisc/include/asm/pdc.h
63
int pdc_pci_irt(unsigned long num_entries, unsigned long hpa, void *tbl);
arch/parisc/include/asm/pdcpat.h
219
extern int pdc_pat_get_irt_size(unsigned long *num_entries, unsigned long cell_num);
arch/parisc/include/asm/pdcpat.h
224
#define pdc_pat_get_irt_size(num_entries, cell_numn) PDC_BAD_PROC
arch/parisc/kernel/firmware.c
1001
*num_entries = pdc_result[0];
arch/parisc/kernel/firmware.c
1016
int pdc_pci_irt(unsigned long num_entries, unsigned long hpa, void *tbl)
arch/parisc/kernel/firmware.c
1024
pdc_result[0] = num_entries;
arch/parisc/kernel/firmware.c
1562
int pdc_pat_get_irt_size(unsigned long *num_entries, unsigned long cell_num)
arch/parisc/kernel/firmware.c
1570
*num_entries = pdc_result[0];
arch/parisc/kernel/firmware.c
992
int pdc_pci_irt_size(unsigned long *num_entries, unsigned long hpa)
arch/powerpc/platforms/pseries/htmdump.c
231
u64 *num_entries;
arch/powerpc/platforms/pseries/htmdump.c
258
num_entries = htm_status_buf + 0x10;
arch/powerpc/platforms/pseries/htmdump.c
263
to_copy = 32 + (be64_to_cpu(*num_entries) * htmstatus_flag);
arch/powerpc/platforms/pseries/htmdump.c
278
u64 *num_entries;
arch/powerpc/platforms/pseries/htmdump.c
304
num_entries = htm_info_buf + 0x10;
arch/powerpc/platforms/pseries/htmdump.c
305
to_copy = 32 + (be64_to_cpu(*num_entries) * 16);
arch/riscv/include/asm/module.h
17
int num_entries;
arch/riscv/include/asm/module.h
41
for (i = 0; i < sec->num_entries; i++) {
arch/riscv/include/asm/module.h
94
for (i = 0; i < sec->num_entries; i++) {
arch/riscv/kernel/module-sections.c
17
int i = got_sec->num_entries;
arch/riscv/kernel/module-sections.c
196
mod->arch.plt.num_entries = 0;
arch/riscv/kernel/module-sections.c
203
mod->arch.got.num_entries = 0;
arch/riscv/kernel/module-sections.c
210
mod->arch.got_plt.num_entries = 0;
arch/riscv/kernel/module-sections.c
27
got_sec->num_entries++;
arch/riscv/kernel/module-sections.c
28
BUG_ON(got_sec->num_entries > got_sec->max_entries);
arch/riscv/kernel/module-sections.c
39
int i = plt_sec->num_entries;
arch/riscv/kernel/module-sections.c
52
plt_sec->num_entries++;
arch/riscv/kernel/module-sections.c
53
got_plt_sec->num_entries++;
arch/riscv/kernel/module-sections.c
54
BUG_ON(plt_sec->num_entries > plt_sec->max_entries);
arch/sparc/include/asm/hypervisor.h
3136
unsigned long num_entries);
arch/sparc/include/asm/hypervisor.h
3139
unsigned long *num_entries);
arch/sparc/include/asm/hypervisor.h
3148
unsigned long num_entries);
arch/sparc/include/asm/hypervisor.h
3151
unsigned long *num_entries);
arch/sparc/include/asm/hypervisor.h
3160
unsigned long num_entries);
arch/sparc/include/asm/hypervisor.h
3163
unsigned long *num_entries);
arch/sparc/include/asm/iommu-common.h
36
unsigned long num_entries,
arch/sparc/include/asm/vio.h
279
u32 num_entries;
arch/sparc/include/asm/vio.h
315
if (++index == dr->num_entries)
arch/sparc/include/asm/vio.h
323
return dr->num_entries - 1;
arch/sparc/kernel/iommu-common.c
53
unsigned long num_entries,
arch/sparc/kernel/iommu-common.c
78
iommu->poolsize = num_entries/iommu->nr_pools;
arch/sparc/kernel/iommu-common.c
80
iommu->poolsize = (num_entries * 3 / 4)/iommu->nr_pools;
arch/sparc/kernel/iommu-common.c
94
p->end = num_entries;
arch/sparc/kernel/irq_64.c
987
unsigned long num_entries = (qmask + 1) / 64;
arch/sparc/kernel/irq_64.c
990
status = sun4v_cpu_qconf(type, paddr, num_entries);
arch/sparc/kernel/irq_64.c
993
"err %lu\n", type, paddr, num_entries, status);
arch/sparc/kernel/ldc.c
1006
size = num_entries * LDC_PACKET_SIZE;
arch/sparc/kernel/ldc.c
201
static unsigned long __advance(unsigned long off, unsigned long num_entries)
arch/sparc/kernel/ldc.c
204
if (off == (num_entries * LDC_PACKET_SIZE))
arch/sparc/kernel/ldc.c
975
static int alloc_queue(const char *name, unsigned long num_entries,
arch/sparc/kernel/ldc.c
981
size = num_entries * LDC_PACKET_SIZE;
arch/sparc/kernel/ldc.c
999
static void free_queue(unsigned long num_entries, struct ldc_packet *q)
arch/sparc/kernel/pci_sun4v.h
35
unsigned long num_entries);
arch/sparc/kernel/pci_sun4v.h
39
unsigned long *num_entries);
arch/sparc/kernel/viohs.c
196
u.pkt.num_descr = dr->num_entries;
arch/sparc/kernel/viohs.c
462
dr->num_entries = pkt->num_descr;
arch/x86/boot/compressed/acpi.c
203
u32 num_entries, size, len;
arch/x86/boot/compressed/acpi.c
240
num_entries = (len - sizeof(struct acpi_table_header)) / size;
arch/x86/boot/compressed/acpi.c
243
while (num_entries--) {
arch/x86/boot/startup/sev-shared.c
615
pc->num_entries = 1;
arch/x86/coco/sev/svsm.c
103
pc->num_entries = 0;
arch/x86/coco/sev/svsm.c
118
pc->num_entries++;
arch/x86/coco/sev/svsm.c
119
if (pc->num_entries == SVSM_PVALIDATE_MAX_COUNT)
arch/x86/coco/sev/svsm.c
133
pc->num_entries = 0;
arch/x86/coco/sev/svsm.c
150
pc->num_entries++;
arch/x86/coco/sev/svsm.c
151
if (pc->num_entries == SVSM_PVALIDATE_MAX_COUNT)
arch/x86/coco/sev/svsm.c
225
if (pc->cur_index < pc->num_entries)
arch/x86/include/asm/efi.h
395
extern int __init efi_memmap_alloc(unsigned int num_entries,
arch/x86/include/asm/sev.h
358
u16 num_entries;
arch/x86/kernel/cpu/microcode/amd.c
1048
equiv_table.num_entries = equiv_tbl_len / sizeof(struct equiv_cpu_entry);
arch/x86/kernel/cpu/microcode/amd.c
353
if (!et || !et->num_entries)
arch/x86/kernel/cpu/microcode/amd.c
356
for (i = 0; i < et->num_entries; i++) {
arch/x86/kernel/cpu/microcode/amd.c
607
table.num_entries = hdr[2] / sizeof(struct equiv_cpu_entry);
arch/x86/kernel/cpu/microcode/amd.c
93
unsigned int num_entries;
arch/x86/kernel/ldt.c
149
static struct ldt_struct *alloc_ldt_struct(unsigned int num_entries)
arch/x86/kernel/ldt.c
154
if (num_entries > LDT_ENTRIES)
arch/x86/kernel/ldt.c
162
alloc_size = num_entries * LDT_ENTRY_SIZE;
arch/x86/kernel/ldt.c
183
new_ldt->nr_entries = num_entries;
arch/x86/kernel/unwind_orc.c
310
unsigned int num_entries = orc_ip_size / sizeof(int);
arch/x86/kernel/unwind_orc.c
314
num_entries != orc_size / sizeof(*orc));
arch/x86/kernel/unwind_orc.c
324
sort(orc_ip, num_entries, sizeof(int), orc_sort_cmp, orc_sort_swap);
arch/x86/kernel/unwind_orc.c
329
mod->arch.num_orcs = num_entries;
arch/x86/kernel/unwind_orc.c
337
size_t num_entries = orc_ip_size / sizeof(int);
arch/x86/kernel/unwind_orc.c
341
if (!num_entries || orc_ip_size % sizeof(int) != 0 ||
arch/x86/kernel/unwind_orc.c
343
num_entries != orc_size / sizeof(struct orc_entry)) {
arch/x86/kernel/unwind_orc.c
358
num_entries,
arch/x86/kernel/unwind_orc.c
369
orc = __orc_find(__start_orc_unwind_ip, __start_orc_unwind, num_entries,
arch/x86/kernel/unwind_orc.c
85
unsigned int num_entries, unsigned long ip)
arch/x86/kernel/unwind_orc.c
88
int *last = ip_table + num_entries - 1;
arch/x86/kernel/unwind_orc.c
91
if (!num_entries)
arch/x86/kvm/cpuid.c
1950
__u32 num_entries, unsigned int ioctl_type)
arch/x86/kvm/cpuid.c
1966
for (i = 0; i < num_entries; i++) {
arch/x86/platform/efi/efi_64.c
110
unsigned num_entries;
arch/x86/platform/efi/efi_64.c
119
num_entries = pgd_index(EFI_VA_END) - pgd_index(PAGE_OFFSET);
arch/x86/platform/efi/efi_64.c
120
memcpy(pgd_efi, pgd_k, sizeof(pgd_t) * num_entries);
arch/x86/platform/efi/efi_64.c
127
num_entries = p4d_index(EFI_VA_END);
arch/x86/platform/efi/efi_64.c
128
memcpy(p4d_efi, p4d_k, sizeof(p4d_t) * num_entries);
arch/x86/platform/efi/efi_64.c
142
num_entries = pud_index(EFI_VA_END);
arch/x86/platform/efi/efi_64.c
143
memcpy(pud_efi, pud_k, sizeof(pud_t) * num_entries);
arch/x86/platform/efi/efi_64.c
148
num_entries = PTRS_PER_PUD - pud_index(EFI_VA_START);
arch/x86/platform/efi/efi_64.c
149
memcpy(pud_efi, pud_k, sizeof(pud_t) * num_entries);
arch/x86/platform/efi/memmap.c
59
int __init efi_memmap_alloc(unsigned int num_entries,
arch/x86/platform/efi/memmap.c
65
data->size = num_entries * efi.memmap.desc_size;
arch/x86/platform/efi/quirks.c
258
int num_entries;
arch/x86/platform/efi/quirks.c
280
num_entries = efi_memmap_split_count(&md, &mr.range);
arch/x86/platform/efi/quirks.c
281
num_entries += efi.memmap.nr_map;
arch/x86/platform/efi/quirks.c
283
if (efi_memmap_alloc(num_entries, &data) != 0) {
arch/x86/platform/efi/quirks.c
418
int num_entries = 0;
arch/x86/platform/efi/quirks.c
441
num_entries++;
arch/x86/platform/efi/quirks.c
447
num_entries++;
arch/x86/platform/efi/quirks.c
501
if (!num_entries)
arch/x86/platform/efi/quirks.c
504
if (efi_memmap_alloc(num_entries, &data) != 0) {
block/sed-opal.c
1014
int num_entries = 0;
block/sed-opal.c
1068
num_entries++;
block/sed-opal.c
1075
resp->num = num_entries;
drivers/accel/amdxdna/amdxdna_gem.c
556
if (va_tbl.num_entries) {
drivers/accel/amdxdna/amdxdna_gem.c
560
dma_buf = amdxdna_get_ubuf(dev, flags, va_tbl.num_entries,
drivers/accel/amdxdna/amdxdna_ubuf.c
137
u32 num_entries, void __user *va_entries)
drivers/accel/amdxdna/amdxdna_ubuf.c
159
va_ent = kvzalloc_objs(*va_ent, num_entries);
drivers/accel/amdxdna/amdxdna_ubuf.c
165
if (copy_from_user(va_ent, va_entries, sizeof(*va_ent) * num_entries)) {
drivers/accel/amdxdna/amdxdna_ubuf.c
171
for (i = 0, exp_info.size = 0; i < num_entries; i++) {
drivers/accel/amdxdna/amdxdna_ubuf.c
202
for (i = 0; i < num_entries; i++) {
drivers/accel/amdxdna/amdxdna_ubuf.h
17
u32 num_entries, void __user *va_entries);
drivers/acpi/cppc_acpi.c
432
if (pdomain->num_entries != ACPI_PSD_REV0_ENTRIES) {
drivers/acpi/cppc_acpi.c
768
cpc_ptr->num_entries = num_ent;
drivers/acpi/cppc_acpi.c
912
for (i = 2; i < cpc_ptr->num_entries; i++) {
drivers/acpi/cppc_acpi.c
955
for (i = 2; i < cpc_ptr->num_entries; i++) {
drivers/acpi/processor_perflib.c
566
if (pdomain->num_entries != ACPI_PSD_REV0_ENTRIES) {
drivers/acpi/processor_throttling.c
607
if (pdomain->num_entries != ACPI_TSD_REV0_ENTRIES) {
drivers/atm/nicstar.c
1531
for (i = 0; i < scq->num_entries; i++) {
drivers/atm/nicstar.c
1774
if (scq->num_entries == VBR_SCQ_NUM_ENTRIES) {
drivers/atm/nicstar.c
1924
if (pos >= scq->num_entries) {
drivers/atm/nicstar.c
1931
if (++i == scq->num_entries)
drivers/atm/nicstar.c
1950
if (++i == scq->num_entries)
drivers/atm/nicstar.c
886
scq->num_entries = size / NS_SCQE_SIZE;
drivers/atm/nicstar.c
889
scq->last = scq->base + (scq->num_entries - 1);
drivers/atm/nicstar.c
905
if (scq->num_entries == VBR_SCQ_NUM_ENTRIES)
drivers/atm/nicstar.c
906
for (i = 0; i < scq->num_entries; i++) {
drivers/atm/nicstar.c
919
for (i = 0; i < scq->num_entries; i++)
drivers/atm/nicstar.c
922
for (i = 0; i < scq->num_entries; i++) {
drivers/atm/nicstar.c
933
2 * (scq->num_entries == VBR_SCQ_NUM_ENTRIES ?
drivers/atm/nicstar.h
669
unsigned num_entries;
drivers/block/sunvdc.c
734
dr->num_entries = VDC_TX_RING_SIZE;
drivers/block/sunvdc.c
748
(dr->entry_size * dr->num_entries),
drivers/block/sunvdc.c
752
dr->num_entries = 0;
drivers/char/agp/agp.h
66
int num_entries;
drivers/char/agp/agp.h
73
int num_entries;
drivers/char/agp/agp.h
80
int num_entries;
drivers/char/agp/agp.h
87
int num_entries;
drivers/char/agp/agp.h
93
int num_entries;
drivers/char/agp/alpha-agp.c
169
aper_size->num_entries = agp->aperture.size / PAGE_SIZE;
drivers/char/agp/alpha-agp.c
170
aper_size->page_order = __ffs(aper_size->num_entries / 1024);
drivers/char/agp/alpha-agp.c
88
int num_entries, status;
drivers/char/agp/alpha-agp.c
95
num_entries = A_SIZE_FIX(temp)->num_entries;
drivers/char/agp/alpha-agp.c
96
if ((pg_start + mem->page_count) > num_entries)
drivers/char/agp/amd-k7-agp.c
137
retval = amd_create_gatt_pages(value->num_entries / 1024);
drivers/char/agp/amd-k7-agp.c
156
for (i = 0; i < value->num_entries / 1024; i++, addr += 0x00400000) {
drivers/char/agp/amd-k7-agp.c
162
for (i = 0; i < value->num_entries; i++) {
drivers/char/agp/amd-k7-agp.c
286
int i, j, num_entries;
drivers/char/agp/amd-k7-agp.c
290
num_entries = A_SIZE_LVL2(agp_bridge->current_size)->num_entries;
drivers/char/agp/amd-k7-agp.c
296
if ((pg_start + mem->page_count) > num_entries)
drivers/char/agp/amd64-agp.c
47
int i, j, num_entries;
drivers/char/agp/amd64-agp.c
53
num_entries = agp_num_entries();
drivers/char/agp/amd64-agp.c
64
if (((unsigned long)pg_start + mem->page_count) > num_entries)
drivers/char/agp/ati-agp.c
261
int i, j, num_entries;
drivers/char/agp/ati-agp.c
266
num_entries = A_SIZE_LVL2(agp_bridge->current_size)->num_entries;
drivers/char/agp/ati-agp.c
275
if ((pg_start + mem->page_count) > num_entries)
drivers/char/agp/ati-agp.c
348
retval = ati_create_gatt_pages(value->num_entries / 1024);
drivers/char/agp/ati-agp.c
384
for (i = 0; i < value->num_entries / 1024; i++, addr += 0x00400000) {
drivers/char/agp/ati-agp.c
390
for (i = 0; i < value->num_entries; i++) {
drivers/char/agp/efficeon-agp.c
198
int num_entries, l1_pages;
drivers/char/agp/efficeon-agp.c
200
num_entries = A_SIZE_LVL2(agp_bridge->current_size)->num_entries;
drivers/char/agp/efficeon-agp.c
202
printk(KERN_DEBUG PFX "efficeon_create_gatt_table(%d)\n", num_entries);
drivers/char/agp/efficeon-agp.c
205
BUG_ON(num_entries & 0x3ff);
drivers/char/agp/efficeon-agp.c
206
l1_pages = num_entries >> 10;
drivers/char/agp/efficeon-agp.c
238
int i, count = mem->page_count, num_entries;
drivers/char/agp/efficeon-agp.c
245
num_entries = A_SIZE_LVL2(agp_bridge->current_size)->num_entries;
drivers/char/agp/efficeon-agp.c
246
if ((pg_start + mem->page_count) > num_entries)
drivers/char/agp/efficeon-agp.c
287
int i, count = mem->page_count, num_entries;
drivers/char/agp/efficeon-agp.c
291
num_entries = A_SIZE_LVL2(agp_bridge->current_size)->num_entries;
drivers/char/agp/efficeon-agp.c
293
if ((pg_start + mem->page_count) > num_entries)
drivers/char/agp/generic.c
1027
int num_entries;
drivers/char/agp/generic.c
1045
num_entries = A_SIZE_8(temp)->num_entries;
drivers/char/agp/generic.c
1048
num_entries = A_SIZE_16(temp)->num_entries;
drivers/char/agp/generic.c
1051
num_entries = A_SIZE_32(temp)->num_entries;
drivers/char/agp/generic.c
1054
num_entries = A_SIZE_FIX(temp)->num_entries;
drivers/char/agp/generic.c
1060
num_entries = 0;
drivers/char/agp/generic.c
1064
num_entries -= agp_memory_reserved/PAGE_SIZE;
drivers/char/agp/generic.c
1065
if (num_entries < 0) num_entries = 0;
drivers/char/agp/generic.c
1076
if (((pg_start + mem->page_count) > num_entries) ||
drivers/char/agp/generic.c
1111
int mask_type, num_entries;
drivers/char/agp/generic.c
1123
num_entries = agp_num_entries();
drivers/char/agp/generic.c
1124
if (((pg_start + mem->page_count) > num_entries) ||
drivers/char/agp/generic.c
321
int num_entries;
drivers/char/agp/generic.c
328
num_entries = A_SIZE_8(temp)->num_entries;
drivers/char/agp/generic.c
331
num_entries = A_SIZE_16(temp)->num_entries;
drivers/char/agp/generic.c
334
num_entries = A_SIZE_32(temp)->num_entries;
drivers/char/agp/generic.c
337
num_entries = A_SIZE_LVL2(temp)->num_entries;
drivers/char/agp/generic.c
340
num_entries = A_SIZE_FIX(temp)->num_entries;
drivers/char/agp/generic.c
343
num_entries = 0;
drivers/char/agp/generic.c
347
num_entries -= agp_memory_reserved>>PAGE_SHIFT;
drivers/char/agp/generic.c
348
if (num_entries<0)
drivers/char/agp/generic.c
349
num_entries = 0;
drivers/char/agp/generic.c
350
return num_entries;
drivers/char/agp/generic.c
855
int num_entries;
drivers/char/agp/generic.c
867
page_order = num_entries = 0;
drivers/char/agp/generic.c
875
num_entries =
drivers/char/agp/generic.c
876
A_SIZE_8(temp)->num_entries;
drivers/char/agp/generic.c
880
num_entries = A_SIZE_16(temp)->num_entries;
drivers/char/agp/generic.c
884
num_entries = A_SIZE_32(temp)->num_entries;
drivers/char/agp/generic.c
890
page_order = num_entries = 0;
drivers/char/agp/generic.c
921
num_entries = ((struct aper_size_info_fixed *) temp)->num_entries;
drivers/char/agp/generic.c
959
for (i = 0; i < num_entries; i++) {
drivers/char/agp/intel-gtt.c
101
unsigned int num_entries,
drivers/char/agp/intel-gtt.c
107
DBG("try mapping %lu pages\n", (unsigned long)num_entries);
drivers/char/agp/intel-gtt.c
109
if (sg_alloc_table(st, num_entries, GFP_KERNEL))
drivers/char/agp/intel-gtt.c
112
for_each_sg(st->sgl, sg, num_entries, i)
drivers/char/agp/intel-gtt.c
917
unsigned int num_entries,
drivers/char/agp/intel-gtt.c
923
for (i = 0, j = first_entry; i < num_entries; i++, j++) {
drivers/char/agp/intel-gtt.c
983
void intel_gmch_gtt_clear_range(unsigned int first_entry, unsigned int num_entries)
drivers/char/agp/intel-gtt.c
987
for (i = first_entry; i < (first_entry + num_entries); i++) {
drivers/char/agp/nvidia-agp.c
133
nvidia_private.num_active_entries = current_size->num_entries;
drivers/char/agp/sworks-agp.c
161
retval = serverworks_create_gatt_pages(value->num_entries / 1024);
drivers/char/agp/sworks-agp.c
181
for (i = 0; i < value->num_entries / 1024; i++)
drivers/char/agp/sworks-agp.c
319
int i, j, num_entries;
drivers/char/agp/sworks-agp.c
323
num_entries = A_SIZE_LVL2(agp_bridge->current_size)->num_entries;
drivers/char/agp/sworks-agp.c
328
if ((pg_start + mem->page_count) > num_entries) {
drivers/char/agp/uninorth-agp.c
152
int i, num_entries;
drivers/char/agp/uninorth-agp.c
170
num_entries = A_SIZE_32(temp)->num_entries;
drivers/char/agp/uninorth-agp.c
172
if ((pg_start + mem->page_count) > num_entries)
drivers/char/agp/uninorth-agp.c
375
int num_entries;
drivers/char/agp/uninorth-agp.c
387
size = page_order = num_entries = 0;
drivers/char/agp/uninorth-agp.c
392
num_entries = A_SIZE_32(temp)->num_entries;
drivers/char/agp/uninorth-agp.c
435
for (i = 0; i < num_entries; i++)
drivers/comedi/drivers/cb_pcidas64.c
1354
unsigned int num_entries)
drivers/comedi/drivers/cb_pcidas64.c
1363
if (num_entries < increment_size)
drivers/comedi/drivers/cb_pcidas64.c
1364
num_entries = increment_size;
drivers/comedi/drivers/cb_pcidas64.c
1365
if (num_entries > fifo->max_segment_length)
drivers/comedi/drivers/cb_pcidas64.c
1366
num_entries = fifo->max_segment_length;
drivers/comedi/drivers/cb_pcidas64.c
1369
num_increments = DIV_ROUND_CLOSEST(num_entries, increment_size);
drivers/crypto/intel/qat/qat_common/adf_accel_devices.h
84
u32 num_entries;
drivers/crypto/intel/qat/qat_common/adf_isr.c
186
if (pci_dev_info->msix_entries.num_entries > 1) {
drivers/crypto/intel/qat/qat_common/adf_isr.c
285
accel_dev->accel_pci_dev.msix_entries.num_entries = msix_num_entries;
drivers/cxl/core/features.c
116
int num_entries;
drivers/cxl/core/features.c
155
num_entries = le16_to_cpu(mbox_out->num_entries);
drivers/cxl/core/features.c
160
if (num_entries * feat_size != retrieved)
drivers/cxl/core/features.c
164
for (int i = 0; i < num_entries; i++) {
drivers/cxl/core/features.c
168
entry += num_entries;
drivers/cxl/core/features.c
173
remain_feats += copy_feats - num_entries;
drivers/cxl/core/features.c
174
start += num_entries;
drivers/cxl/core/features.c
447
feat_out->num_entries = cpu_to_le16(requested);
drivers/firmware/arm_ffa/driver.c
688
u32 idx, frag_len, length, buf_sz = 0, num_entries = sg_nents(args->sg);
drivers/firmware/arm_ffa/driver.c
715
composite->addr_range_cnt = num_entries;
drivers/firmware/arm_ffa/driver.c
718
length = composite_offset + CONSTITUENTS_OFFSET(num_entries);
drivers/firmware/efi/memattr.c
176
for (i = ret = 0; ret == 0 && i < tbl->num_entries; i++) {
drivers/firmware/efi/memattr.c
51
size = tbl->num_entries * tbl->desc_size;
drivers/firmware/efi/memattr.c
54
tbl->version, tbl->desc_size, tbl->num_entries);
drivers/fwctl/pds/main.c
115
num_endpoints = le32_to_cpu(pdsfc->endpoints->num_entries);
drivers/fwctl/pds/main.c
132
num_endpoints = le32_to_cpu(pdsfc->endpoints->num_entries);
drivers/fwctl/pds/main.c
194
num_endpoints = le32_to_cpu(pdsfc->endpoints->num_entries);
drivers/fwctl/pds/main.c
220
u32 num_entries;
drivers/fwctl/pds/main.c
253
num_entries = le32_to_cpu(data->num_entries);
drivers/fwctl/pds/main.c
254
dev_dbg(dev, "num_entries %d\n", num_entries);
drivers/fwctl/pds/main.c
255
for (i = 0; i < num_entries; i++) {
drivers/fwctl/pds/main.c
288
u32 num_entries;
drivers/fwctl/pds/main.c
306
num_entries = le32_to_cpu(pdsfc->endpoints->num_entries);
drivers/fwctl/pds/main.c
307
for (i = 0; i < num_entries; i++) {
drivers/fwctl/pds/main.c
336
num_entries = le32_to_cpu(ep_info->operations->num_entries);
drivers/fwctl/pds/main.c
337
for (i = 0; i < num_entries; i++) {
drivers/gpio/gpio-virtuser.c
1383
size_t num_entries = gpio_virtuser_get_lookup_count(dev);
drivers/gpio/gpio-virtuser.c
1391
kzalloc_flex(*table, table, num_entries + 1);
drivers/gpu/drm/amd/amdgpu/amdgpu.h
744
uint8_t num_entries;
drivers/gpu/drm/amd/amdgpu/amdgpu_atombios.c
1456
u8 frev, crev, num_entries, t_mem_id, num_ranges = 0;
drivers/gpu/drm/amd/amdgpu/amdgpu_atombios.c
1483
num_entries = (u8)((le16_to_cpu(reg_block->usRegIndexTblSize)) /
drivers/gpu/drm/amd/amdgpu/amdgpu_atombios.c
1485
if (num_entries > VBIOS_MC_REGISTER_ARRAY_SIZE)
drivers/gpu/drm/amd/amdgpu/amdgpu_atombios.c
1487
while (i < num_entries) {
drivers/gpu/drm/amd/amdgpu/amdgpu_atombios.c
1526
reg_table->num_entries = num_ranges;
drivers/gpu/drm/amd/amdgpu/amdgpu_atombios.h
113
u8 num_entries;
drivers/gpu/drm/amd/amdgpu/amdgpu_atombios.h
93
u8 num_entries;
drivers/gpu/drm/amd/amdgpu/amdgpu_atomfirmware.c
318
uma_info->num_entries = nr_uma_options;
drivers/gpu/drm/amd/amdgpu/amdgpu_bo_list.c
135
trace_amdgpu_cs_bo_status(list->num_entries, total_size);
drivers/gpu/drm/amd/amdgpu/amdgpu_bo_list.c
144
for (i = first_userptr; i < num_entries; ++i)
drivers/gpu/drm/amd/amdgpu/amdgpu_bo_list.c
71
size_t num_entries, struct amdgpu_bo_list **result)
drivers/gpu/drm/amd/amdgpu/amdgpu_bo_list.c
73
unsigned last_entry = 0, first_userptr = num_entries;
drivers/gpu/drm/amd/amdgpu/amdgpu_bo_list.c
80
list = kvzalloc_flex(*list, entries, num_entries);
drivers/gpu/drm/amd/amdgpu/amdgpu_bo_list.c
86
list->num_entries = num_entries;
drivers/gpu/drm/amd/amdgpu/amdgpu_bo_list.c
89
for (i = 0; i < num_entries; ++i) {
drivers/gpu/drm/amd/amdgpu/amdgpu_bo_list.h
52
unsigned num_entries;
drivers/gpu/drm/amd/amdgpu/amdgpu_bo_list.h
58
struct amdgpu_bo_list_entry entries[] __counted_by(num_entries);
drivers/gpu/drm/amd/amdgpu/amdgpu_bo_list.h
70
size_t num_entries,
drivers/gpu/drm/amd/amdgpu/amdgpu_bo_list.h
75
e != &list->entries[list->num_entries]; \
drivers/gpu/drm/amd/amdgpu/amdgpu_bo_list.h
80
e != &list->entries[list->num_entries]; \
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
413
if (!uma_info || !uma_info->num_entries)
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
416
for (int i = 0; i < uma_info->num_entries; i++) {
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
472
if (val >= uma_info->num_entries)
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
475
val = array_index_nospec(val, uma_info->num_entries);
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
551
uma_info->num_entries = 0;
drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
1098
vm_entries = kvcalloc(args->num_entries, sizeof(*vm_entries), GFP_KERNEL);
drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
1103
if (num_mappings < args->num_entries) {
drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
1113
if (num_mappings < args->num_entries) {
drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
1124
if (num_mappings > 0 && num_mappings <= args->num_entries)
drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
1128
args->num_entries = num_mappings;
drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
1176
if (args->num_entries < num_bos) {
drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
1177
args->num_entries = num_bos;
drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
1182
args->num_entries = 0;
drivers/gpu/drm/amd/amdgpu/amdgpu_gem.c
1215
args->num_entries = bo_index;
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
1173
uint64_t tmp, num_entries, addr;
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
1175
num_entries = cursor.size >> AMDGPU_GPU_PAGE_SHIFT;
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
1179
if (num_entries > AMDGPU_GPU_PAGES_IN_CPU_PAGE) {
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
1186
tmp = num_entries /
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
1197
num_entries = count *
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
1215
tmp = start + num_entries;
drivers/gpu/drm/amd/amdgpu/amdgpu_vm.c
1220
amdgpu_res_next(&cursor, num_entries * AMDGPU_GPU_PAGE_SIZE);
drivers/gpu/drm/amd/amdgpu/amdgpu_vm_pt.c
211
unsigned int shift, num_entries;
drivers/gpu/drm/amd/amdgpu/amdgpu_vm_pt.c
220
num_entries = amdgpu_vm_pt_num_entries(adev, cursor->level - 1);
drivers/gpu/drm/amd/amdgpu/amdgpu_vm_pt.c
223
if (cursor->entry == &parent->entries[num_entries - 1])
drivers/gpu/drm/amd/amdgpu/amdgpu_vm_pt.c
446
unsigned int num_entries;
drivers/gpu/drm/amd/amdgpu/amdgpu_vm_pt.c
463
num_entries = amdgpu_vm_pt_num_entries(adev, level);
drivers/gpu/drm/amd/amdgpu/amdgpu_vm_pt.c
465
num_entries = 0;
drivers/gpu/drm/amd/amdgpu/amdgpu_vm_pt.c
467
bp.bo_ptr_size = struct_size((*vmbo), entries, num_entries);
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
684
int num_entries = 0;
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
694
num_entries++;
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
698
r = sg_alloc_table(*sgt, num_entries, GFP_KERNEL);
drivers/gpu/drm/amd/amdkfd/kfd_crat.c
1826
uint32_t *num_entries,
drivers/gpu/drm/amd/amdkfd/kfd_crat.c
1838
*num_entries = 0;
drivers/gpu/drm/amd/amdkfd/kfd_crat.c
1861
(*num_entries)++;
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_color.c
620
gamma->num_entries = lut_size;
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_color.c
658
gamma->num_entries = lut_size;
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_color.c
714
gamma->num_entries = lut_size;
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_color.c
805
gamma->num_entries = lut_size;
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_color.c
841
gamma->num_entries = lut_size;
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_debugfs.c
915
uint32_t tbuf_size, max_entries, num_entries, first_entry, i;
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_debugfs.c
932
num_entries =
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_debugfs.c
938
if (num_entries > max_entries)
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_debugfs.c
941
first_entry = num_entries % max_entries;
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_debugfs.c
942
num_entries = min(num_entries, max_entries);
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_debugfs.c
951
for (i = first_entry; i < num_entries; ++i) {
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn21/rn_clk_mgr.c
609
.num_entries = 4,
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn21/rn_clk_mgr.c
664
bw_params->clk_table.num_entries = j + 1;
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn21/rn_clk_mgr.c
666
for (i = 0; i < bw_params->clk_table.num_entries; i++, j--) {
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn21/rn_clk_mgr.c
681
if (i >= bw_params->clk_table.num_entries) {
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn30/dcn30_clk_mgr.c
271
clk_mgr_base->bw_params->clk_table.entries[clk_mgr_base->bw_params->clk_table.num_entries - 1].memclk_mhz);
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn30/dcn30_clk_mgr.c
369
clk_mgr_base->bw_params->clk_table.entries[clk_mgr_base->bw_params->clk_table.num_entries - 1].memclk_mhz);
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn30/dcn30_clk_mgr.c
385
clk_mgr_base->bw_params->clk_table.entries[clk_mgr_base->bw_params->clk_table.num_entries - 1].memclk_mhz);
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn30/dcn30_clk_mgr.c
419
clk_mgr_base->bw_params->clk_table.num_entries = num_levels ? num_levels : 1;
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn301/vg_clk_mgr.c
525
.num_entries = 4,
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn301/vg_clk_mgr.c
597
bw_params->clk_table.num_entries = j + 1;
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn301/vg_clk_mgr.c
599
for (i = 0; i < bw_params->clk_table.num_entries - 1; i++, j--) {
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn301/vg_clk_mgr.c
622
if (i >= bw_params->clk_table.num_entries) {
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn31/dcn31_clk_mgr.c
339
.num_entries = 4,
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn31/dcn31_clk_mgr.c
584
bw_params->clk_table.num_entries = j + 1;
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn31/dcn31_clk_mgr.c
595
for (i = 0; i < bw_params->clk_table.num_entries; i++, j--) {
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn31/dcn31_clk_mgr.c
623
if (i >= bw_params->clk_table.num_entries) {
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn314/dcn314_clk_mgr.c
405
.num_entries = 4,
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn314/dcn314_clk_mgr.c
629
struct clk_limit_table_entry def_max = bw_params->clk_table.entries[bw_params->clk_table.num_entries - 1];
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn314/dcn314_clk_mgr.c
670
for (j = bw_params->clk_table.num_entries - 1; j > 0; j--)
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn314/dcn314_clk_mgr.c
706
bw_params->clk_table.num_entries = i--;
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn314/dcn314_clk_mgr.c
721
for (i = 0; i < bw_params->clk_table.num_entries; i++) {
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn314/dcn314_clk_mgr.c
751
if (i >= bw_params->clk_table.num_entries) {
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn315/dcn315_clk_mgr.c
299
.num_entries = 5,
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn315/dcn315_clk_mgr.c
488
struct clk_limit_table_entry def_max = bw_params->clk_table.entries[bw_params->clk_table.num_entries - 1];
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn315/dcn315_clk_mgr.c
504
for (j = bw_params->clk_table.num_entries - 1; j > 0; j--)
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn315/dcn315_clk_mgr.c
536
bw_params->clk_table.num_entries = i;
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn315/dcn315_clk_mgr.c
541
for (i = 0; i < bw_params->clk_table.num_entries; i++) {
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn315/dcn315_clk_mgr.c
575
if (i >= bw_params->clk_table.num_entries) {
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn316/dcn316_clk_mgr.c
265
.num_entries = 5,
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn316/dcn316_clk_mgr.c
511
bw_params->clk_table.num_entries = j + 1;
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn316/dcn316_clk_mgr.c
522
for (i = 0; i < bw_params->clk_table.num_entries; i++, j--) {
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn316/dcn316_clk_mgr.c
555
if (i >= bw_params->clk_table.num_entries) {
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn32/dcn32_clk_mgr.c
1054
clk_mgr_base->bw_params->clk_table.num_entries = num_levels ? num_levels : 1;
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn35/dcn35_clk_mgr.c
1023
struct clk_limit_table_entry def_max = bw_params->clk_table.entries[bw_params->clk_table.num_entries - 1];
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn35/dcn35_clk_mgr.c
1083
for (j = bw_params->clk_table.num_entries - 1; j > 0; j--)
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn35/dcn35_clk_mgr.c
1123
bw_params->clk_table.num_entries = i--;
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn35/dcn35_clk_mgr.c
1149
for (i = 0; i < bw_params->clk_table.num_entries; i++) {
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn35/dcn35_clk_mgr.c
1180
if (i >= bw_params->clk_table.num_entries) {
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn35/dcn35_clk_mgr.c
758
.num_entries = 4,
drivers/gpu/drm/amd/display/dc/clk_mgr/dcn401/dcn401_clk_mgr.c
1409
clk_mgr_base->bw_params->clk_table.num_entries = num_levels ? num_levels : 1;
drivers/gpu/drm/amd/display/dc/core/dc.c
3156
surface->gamma_correction.num_entries =
drivers/gpu/drm/amd/display/dc/core/dc.c
3157
srf_update->gamma->num_entries;
drivers/gpu/drm/amd/display/dc/core/dc.c
5851
for (i = 0; i < dc->clk_mgr->bw_params->clk_table.num_entries; i++) {
drivers/gpu/drm/amd/display/dc/core/dc.c
6905
if (plane_state->gamma_correction.type != GAMMA_CS_TFM_1D && plane_state->gamma_correction.num_entries > 0) {
drivers/gpu/drm/amd/display/dc/dc_hw_types.h
545
unsigned int num_entries;
drivers/gpu/drm/amd/display/dc/dce/dce_ipp.c
196
for (i = 0; i < gamma->num_entries; i++) {
drivers/gpu/drm/amd/display/dc/dml/dcn20/dcn20_fpu.c
2290
vlevel_max = bw_params->clk_table.num_entries - 1;
drivers/gpu/drm/amd/display/dc/dml/dcn20/dcn20_fpu.c
2393
for (i = clk_table->num_entries; i > 1; i--)
drivers/gpu/drm/amd/display/dc/dml/dcn20/dcn20_fpu.c
2396
clk_table->num_entries++;
drivers/gpu/drm/amd/display/dc/dml/dcn20/dcn20_fpu.c
2415
ASSERT(clk_table->num_entries);
drivers/gpu/drm/amd/display/dc/dml/dcn20/dcn20_fpu.c
2419
for (i = 0; i < clk_table->num_entries; i++) {
drivers/gpu/drm/amd/display/dc/dml/dcn20/dcn20_fpu.c
2452
if (clk_table->num_entries) {
drivers/gpu/drm/amd/display/dc/dml/dcn20/dcn20_fpu.c
2453
dcn2_1_soc.num_states = clk_table->num_entries + 1;
drivers/gpu/drm/amd/display/dc/dml/dcn30/dcn30_fpu.c
417
int min_dram_speed_mts_offset = dc->clk_mgr->bw_params->clk_table.num_entries - 1;
drivers/gpu/drm/amd/display/dc/dml/dcn301/dcn301_fpu.c
340
ASSERT(clk_table->num_entries);
drivers/gpu/drm/amd/display/dc/dml/dcn301/dcn301_fpu.c
343
for (i = 0; i < clk_table->num_entries; ++i) {
drivers/gpu/drm/amd/display/dc/dml/dcn301/dcn301_fpu.c
350
for (i = 0; i < clk_table->num_entries; i++) {
drivers/gpu/drm/amd/display/dc/dml/dcn301/dcn301_fpu.c
381
if (clk_table->num_entries) {
drivers/gpu/drm/amd/display/dc/dml/dcn301/dcn301_fpu.c
382
dcn3_01_soc.num_states = clk_table->num_entries;
drivers/gpu/drm/amd/display/dc/dml/dcn301/dcn301_fpu.c
440
vlevel_max = bw_params->clk_table.num_entries - 1;
drivers/gpu/drm/amd/display/dc/dml/dcn302/dcn302_fpu.c
258
num_uclk_states = bw_params->clk_table.num_entries;
drivers/gpu/drm/amd/display/dc/dml/dcn303/dcn303_fpu.c
252
num_uclk_states = bw_params->clk_table.num_entries;
drivers/gpu/drm/amd/display/dc/dml/dcn31/dcn31_fpu.c
607
ASSERT(clk_table->num_entries);
drivers/gpu/drm/amd/display/dc/dml/dcn31/dcn31_fpu.c
610
for (i = 0; i < clk_table->num_entries; ++i) {
drivers/gpu/drm/amd/display/dc/dml/dcn31/dcn31_fpu.c
617
for (i = 0; i < clk_table->num_entries; i++) {
drivers/gpu/drm/amd/display/dc/dml/dcn31/dcn31_fpu.c
651
if (clk_table->num_entries)
drivers/gpu/drm/amd/display/dc/dml/dcn31/dcn31_fpu.c
652
dcn3_1_soc.num_states = clk_table->num_entries;
drivers/gpu/drm/amd/display/dc/dml/dcn31/dcn31_fpu.c
683
ASSERT(clk_table->num_entries);
drivers/gpu/drm/amd/display/dc/dml/dcn31/dcn31_fpu.c
686
for (i = 0; i < clk_table->num_entries; ++i) {
drivers/gpu/drm/amd/display/dc/dml/dcn31/dcn31_fpu.c
693
for (i = 0; i < clk_table->num_entries; i++) {
drivers/gpu/drm/amd/display/dc/dml/dcn31/dcn31_fpu.c
712
dcn3_15_soc.num_states = clk_table->num_entries;
drivers/gpu/drm/amd/display/dc/dml/dcn31/dcn31_fpu.c
746
ASSERT(clk_table->num_entries);
drivers/gpu/drm/amd/display/dc/dml/dcn31/dcn31_fpu.c
749
for (i = 0; i < clk_table->num_entries; ++i) {
drivers/gpu/drm/amd/display/dc/dml/dcn31/dcn31_fpu.c
756
for (i = 0; i < clk_table->num_entries; i++) {
drivers/gpu/drm/amd/display/dc/dml/dcn31/dcn31_fpu.c
791
if (clk_table->num_entries)
drivers/gpu/drm/amd/display/dc/dml/dcn31/dcn31_fpu.c
792
dcn3_16_soc.num_states = clk_table->num_entries;
drivers/gpu/drm/amd/display/dc/dml/dcn314/dcn314_fpu.c
205
ASSERT(clk_table->num_entries);
drivers/gpu/drm/amd/display/dc/dml/dcn314/dcn314_fpu.c
208
for (i = 0; i < clk_table->num_entries; ++i) {
drivers/gpu/drm/amd/display/dc/dml/dcn314/dcn314_fpu.c
215
for (i = 0; i < clk_table->num_entries; i++) {
drivers/gpu/drm/amd/display/dc/dml/dcn314/dcn314_fpu.c
223
if (clk_table->num_entries == 1) {
drivers/gpu/drm/amd/display/dc/dml/dcn314/dcn314_fpu.c
232
if (clk_table->num_entries == 1 &&
drivers/gpu/drm/amd/display/dc/dml/dcn314/dcn314_fpu.c
256
for (i = 0; i < clk_table->num_entries; i++)
drivers/gpu/drm/amd/display/dc/dml/dcn314/dcn314_fpu.c
258
if (clk_table->num_entries) {
drivers/gpu/drm/amd/display/dc/dml/dcn314/dcn314_fpu.c
259
dcn3_14_soc.num_states = clk_table->num_entries;
drivers/gpu/drm/amd/display/dc/dml/dcn32/dcn32_fpu.c
2656
static void remove_entry_from_table_at_index(struct _vcs_dpi_voltage_scaling_st *table, unsigned int *num_entries,
drivers/gpu/drm/amd/display/dc/dml/dcn32/dcn32_fpu.c
2661
if (*num_entries == 0)
drivers/gpu/drm/amd/display/dc/dml/dcn32/dcn32_fpu.c
2664
for (i = index; i < *num_entries - 1; i++) {
drivers/gpu/drm/amd/display/dc/dml/dcn32/dcn32_fpu.c
2667
memset(&table[--(*num_entries)], 0, sizeof(struct _vcs_dpi_voltage_scaling_st));
drivers/gpu/drm/amd/display/dc/dml/dcn32/dcn32_fpu.c
2720
static void sort_entries_with_same_bw(struct _vcs_dpi_voltage_scaling_st *table, unsigned int *num_entries)
drivers/gpu/drm/amd/display/dc/dml/dcn32/dcn32_fpu.c
2726
for (int i = 0; i < (*num_entries - 1); i++) {
drivers/gpu/drm/amd/display/dc/dml/dcn32/dcn32_fpu.c
2732
while ((i < (*num_entries - 1)) && (table[i+1].net_bw_in_kbytes_sec == current_bw))
drivers/gpu/drm/amd/display/dc/dml/dcn32/dcn32_fpu.c
2755
static void remove_inconsistent_entries(struct _vcs_dpi_voltage_scaling_st *table, unsigned int *num_entries)
drivers/gpu/drm/amd/display/dc/dml/dcn32/dcn32_fpu.c
2757
for (int i = 0; i < (*num_entries - 1); i++) {
drivers/gpu/drm/amd/display/dc/dml/dcn32/dcn32_fpu.c
2761
remove_entry_from_table_at_index(table, num_entries, i);
drivers/gpu/drm/amd/display/dc/dml/dcn32/dcn32_fpu.c
2803
struct _vcs_dpi_voltage_scaling_st *table, unsigned int *num_entries)
drivers/gpu/drm/amd/display/dc/dml/dcn32/dcn32_fpu.c
2882
*num_entries = 0;
drivers/gpu/drm/amd/display/dc/dml/dcn32/dcn32_fpu.c
2899
insert_entry_into_table_sorted(table, num_entries, &entry);
drivers/gpu/drm/amd/display/dc/dml/dcn32/dcn32_fpu.c
2909
insert_entry_into_table_sorted(table, num_entries, &entry);
drivers/gpu/drm/amd/display/dc/dml/dcn32/dcn32_fpu.c
2919
insert_entry_into_table_sorted(table, num_entries, &entry);
drivers/gpu/drm/amd/display/dc/dml/dcn32/dcn32_fpu.c
2931
insert_entry_into_table_sorted(table, num_entries, &entry);
drivers/gpu/drm/amd/display/dc/dml/dcn32/dcn32_fpu.c
2942
insert_entry_into_table_sorted(table, num_entries, &entry);
drivers/gpu/drm/amd/display/dc/dml/dcn32/dcn32_fpu.c
2950
for (i = *num_entries - 1; i >= 0 ; i--) {
drivers/gpu/drm/amd/display/dc/dml/dcn32/dcn32_fpu.c
2954
remove_entry_from_table_at_index(table, num_entries, i);
drivers/gpu/drm/amd/display/dc/dml/dcn32/dcn32_fpu.c
2966
insert_entry_into_table_sorted(table, num_entries, &max_dc_limits_entry);
drivers/gpu/drm/amd/display/dc/dml/dcn32/dcn32_fpu.c
2968
sort_entries_with_same_bw(table, num_entries);
drivers/gpu/drm/amd/display/dc/dml/dcn32/dcn32_fpu.c
2969
remove_inconsistent_entries(table, num_entries);
drivers/gpu/drm/amd/display/dc/dml/dcn32/dcn32_fpu.c
2978
for (i = *num_entries - 1; i >= 0 ; i--) {
drivers/gpu/drm/amd/display/dc/dml/dcn32/dcn32_fpu.c
2989
for (i = *num_entries - 1; i >= 0 ; i--) {
drivers/gpu/drm/amd/display/dc/dml/dcn32/dcn32_fpu.c
3000
for (i = *num_entries - 1; i >= 0 ; i--) {
drivers/gpu/drm/amd/display/dc/dml/dcn32/dcn32_fpu.c
3008
for (i = *num_entries - 1; i >= 0 ; i--) {
drivers/gpu/drm/amd/display/dc/dml/dcn32/dcn32_fpu.c
3016
while (i < *num_entries - 1) {
drivers/gpu/drm/amd/display/dc/dml/dcn32/dcn32_fpu.c
3020
remove_entry_from_table_at_index(table, num_entries, i + 1);
drivers/gpu/drm/amd/display/dc/dml/dcn32/dcn32_fpu.c
3026
for (i = *num_entries - 1; i >= 0 ; i--) {
drivers/gpu/drm/amd/display/dc/dml/dcn32/dcn32_fpu.c
3206
num_uclk_states = bw_params->clk_table.num_entries;
drivers/gpu/drm/amd/display/dc/dml/dcn32/dcn32_fpu.c
3315
if (dc->clk_mgr->bw_params->clk_table.num_entries > 1) {
drivers/gpu/drm/amd/display/dc/dml/dcn32/dcn32_fpu.c
3318
dc->dml2_options.bbox_overrides.clks_table.num_states = dc->clk_mgr->bw_params->clk_table.num_entries;
drivers/gpu/drm/amd/display/dc/dml/dcn32/dcn32_fpu.c
418
unsigned int *num_entries,
drivers/gpu/drm/amd/display/dc/dml/dcn32/dcn32_fpu.c
426
if (*num_entries == 0) {
drivers/gpu/drm/amd/display/dc/dml/dcn32/dcn32_fpu.c
428
(*num_entries)++;
drivers/gpu/drm/amd/display/dc/dml/dcn32/dcn32_fpu.c
432
if (index >= *num_entries)
drivers/gpu/drm/amd/display/dc/dml/dcn32/dcn32_fpu.c
436
for (i = *num_entries; i > index; i--)
drivers/gpu/drm/amd/display/dc/dml/dcn32/dcn32_fpu.c
440
(*num_entries)++;
drivers/gpu/drm/amd/display/dc/dml/dcn321/dcn321_fpu.c
211
unsigned int *num_entries,
drivers/gpu/drm/amd/display/dc/dml/dcn321/dcn321_fpu.c
219
if (*num_entries == 0) {
drivers/gpu/drm/amd/display/dc/dml/dcn321/dcn321_fpu.c
221
(*num_entries)++;
drivers/gpu/drm/amd/display/dc/dml/dcn321/dcn321_fpu.c
225
if (index >= *num_entries)
drivers/gpu/drm/amd/display/dc/dml/dcn321/dcn321_fpu.c
229
for (i = *num_entries; i > index; i--)
drivers/gpu/drm/amd/display/dc/dml/dcn321/dcn321_fpu.c
233
(*num_entries)++;
drivers/gpu/drm/amd/display/dc/dml/dcn321/dcn321_fpu.c
237
static void remove_entry_from_table_at_index(struct _vcs_dpi_voltage_scaling_st *table, unsigned int *num_entries,
drivers/gpu/drm/amd/display/dc/dml/dcn321/dcn321_fpu.c
242
if (*num_entries == 0)
drivers/gpu/drm/amd/display/dc/dml/dcn321/dcn321_fpu.c
245
for (i = index; i < *num_entries - 1; i++) {
drivers/gpu/drm/amd/display/dc/dml/dcn321/dcn321_fpu.c
248
memset(&table[--(*num_entries)], 0, sizeof(struct _vcs_dpi_voltage_scaling_st));
drivers/gpu/drm/amd/display/dc/dml/dcn321/dcn321_fpu.c
262
static void sort_entries_with_same_bw(struct _vcs_dpi_voltage_scaling_st *table, unsigned int *num_entries)
drivers/gpu/drm/amd/display/dc/dml/dcn321/dcn321_fpu.c
268
for (int i = 0; i < (*num_entries - 1); i++) {
drivers/gpu/drm/amd/display/dc/dml/dcn321/dcn321_fpu.c
274
while ((i < (*num_entries - 1)) && (table[i+1].net_bw_in_kbytes_sec == current_bw))
drivers/gpu/drm/amd/display/dc/dml/dcn321/dcn321_fpu.c
297
static void remove_inconsistent_entries(struct _vcs_dpi_voltage_scaling_st *table, unsigned int *num_entries)
drivers/gpu/drm/amd/display/dc/dml/dcn321/dcn321_fpu.c
299
for (int i = 0; i < (*num_entries - 1); i++) {
drivers/gpu/drm/amd/display/dc/dml/dcn321/dcn321_fpu.c
303
remove_entry_from_table_at_index(table, num_entries, i);
drivers/gpu/drm/amd/display/dc/dml/dcn321/dcn321_fpu.c
345
struct _vcs_dpi_voltage_scaling_st *table, unsigned int *num_entries)
drivers/gpu/drm/amd/display/dc/dml/dcn321/dcn321_fpu.c
424
*num_entries = 0;
drivers/gpu/drm/amd/display/dc/dml/dcn321/dcn321_fpu.c
441
dcn321_insert_entry_into_table_sorted(table, num_entries, &entry);
drivers/gpu/drm/amd/display/dc/dml/dcn321/dcn321_fpu.c
451
dcn321_insert_entry_into_table_sorted(table, num_entries, &entry);
drivers/gpu/drm/amd/display/dc/dml/dcn321/dcn321_fpu.c
461
dcn321_insert_entry_into_table_sorted(table, num_entries, &entry);
drivers/gpu/drm/amd/display/dc/dml/dcn321/dcn321_fpu.c
473
dcn321_insert_entry_into_table_sorted(table, num_entries, &entry);
drivers/gpu/drm/amd/display/dc/dml/dcn321/dcn321_fpu.c
484
dcn321_insert_entry_into_table_sorted(table, num_entries, &entry);
drivers/gpu/drm/amd/display/dc/dml/dcn321/dcn321_fpu.c
492
for (i = *num_entries - 1; i >= 0 ; i--) {
drivers/gpu/drm/amd/display/dc/dml/dcn321/dcn321_fpu.c
496
remove_entry_from_table_at_index(table, num_entries, i);
drivers/gpu/drm/amd/display/dc/dml/dcn321/dcn321_fpu.c
508
dcn321_insert_entry_into_table_sorted(table, num_entries, &max_dc_limits_entry);
drivers/gpu/drm/amd/display/dc/dml/dcn321/dcn321_fpu.c
510
sort_entries_with_same_bw(table, num_entries);
drivers/gpu/drm/amd/display/dc/dml/dcn321/dcn321_fpu.c
511
remove_inconsistent_entries(table, num_entries);
drivers/gpu/drm/amd/display/dc/dml/dcn321/dcn321_fpu.c
522
for (i = *num_entries - 1; i >= 0 ; i--) {
drivers/gpu/drm/amd/display/dc/dml/dcn321/dcn321_fpu.c
533
for (i = *num_entries - 1; i >= 0 ; i--) {
drivers/gpu/drm/amd/display/dc/dml/dcn321/dcn321_fpu.c
544
for (i = *num_entries - 1; i >= 0 ; i--) {
drivers/gpu/drm/amd/display/dc/dml/dcn321/dcn321_fpu.c
552
for (i = *num_entries - 1; i >= 0 ; i--) {
drivers/gpu/drm/amd/display/dc/dml/dcn321/dcn321_fpu.c
560
while (i < *num_entries - 1) {
drivers/gpu/drm/amd/display/dc/dml/dcn321/dcn321_fpu.c
564
remove_entry_from_table_at_index(table, num_entries, i + 1);
drivers/gpu/drm/amd/display/dc/dml/dcn321/dcn321_fpu.c
570
for (i = *num_entries - 1; i >= 0 ; i--) {
drivers/gpu/drm/amd/display/dc/dml/dcn321/dcn321_fpu.c
756
num_uclk_states = bw_params->clk_table.num_entries;
drivers/gpu/drm/amd/display/dc/dml/dcn321/dcn321_fpu.c
864
if (dc->clk_mgr->bw_params->clk_table.num_entries > 1) {
drivers/gpu/drm/amd/display/dc/dml/dcn321/dcn321_fpu.c
867
dc->dml2_options.bbox_overrides.clks_table.num_states = dc->clk_mgr->bw_params->clk_table.num_entries;
drivers/gpu/drm/amd/display/dc/dml/dcn35/dcn35_fpu.c
244
ASSERT(clk_table->num_entries);
drivers/gpu/drm/amd/display/dc/dml/dcn35/dcn35_fpu.c
247
for (i = 0; i < clk_table->num_entries; ++i) {
drivers/gpu/drm/amd/display/dc/dml/dcn35/dcn35_fpu.c
254
for (i = 0; i < clk_table->num_entries; i++) {
drivers/gpu/drm/amd/display/dc/dml/dcn35/dcn35_fpu.c
264
if (clk_table->num_entries == 1) {
drivers/gpu/drm/amd/display/dc/dml/dcn35/dcn35_fpu.c
273
if (clk_table->num_entries == 1 &&
drivers/gpu/drm/amd/display/dc/dml/dcn35/dcn35_fpu.c
316
if (clk_table->num_entries)
drivers/gpu/drm/amd/display/dc/dml/dcn35/dcn35_fpu.c
317
dcn3_5_soc.num_states = clk_table->num_entries;
drivers/gpu/drm/amd/display/dc/dml/dcn35/dcn35_fpu.c
353
if (clk_table->num_entries > 2) {
drivers/gpu/drm/amd/display/dc/dml/dcn35/dcn35_fpu.c
355
for (i = 0; i < clk_table->num_entries; i++) {
drivers/gpu/drm/amd/display/dc/dml/dcn35/dcn35_fpu.c
357
clk_table->num_entries;
drivers/gpu/drm/amd/display/dc/dml/dcn35/dcn35_fpu.c
375
clk_table->num_entries;
drivers/gpu/drm/amd/display/dc/dml/dcn35/dcn35_fpu.c
377
clk_table->num_entries;
drivers/gpu/drm/amd/display/dc/dml/dcn35/dcn35_fpu.c
379
clk_table->num_entries;
drivers/gpu/drm/amd/display/dc/dml/dcn35/dcn35_fpu.c
381
clk_table->num_entries;
drivers/gpu/drm/amd/display/dc/dml/dcn35/dcn35_fpu.c
383
clk_table->num_entries;
drivers/gpu/drm/amd/display/dc/dml/dcn35/dcn35_fpu.c
385
clk_table->num_entries;
drivers/gpu/drm/amd/display/dc/dml/dcn35/dcn35_fpu.c
387
clk_table->num_entries;
drivers/gpu/drm/amd/display/dc/dml/dcn351/dcn351_fpu.c
278
ASSERT(clk_table->num_entries);
drivers/gpu/drm/amd/display/dc/dml/dcn351/dcn351_fpu.c
281
for (i = 0; i < clk_table->num_entries; ++i) {
drivers/gpu/drm/amd/display/dc/dml/dcn351/dcn351_fpu.c
288
for (i = 0; i < clk_table->num_entries; i++) {
drivers/gpu/drm/amd/display/dc/dml/dcn351/dcn351_fpu.c
298
if (clk_table->num_entries == 1) {
drivers/gpu/drm/amd/display/dc/dml/dcn351/dcn351_fpu.c
307
if (clk_table->num_entries == 1 &&
drivers/gpu/drm/amd/display/dc/dml/dcn351/dcn351_fpu.c
350
if (clk_table->num_entries)
drivers/gpu/drm/amd/display/dc/dml/dcn351/dcn351_fpu.c
351
dcn3_51_soc.num_states = clk_table->num_entries;
drivers/gpu/drm/amd/display/dc/dml/dcn351/dcn351_fpu.c
387
if (clk_table->num_entries > 2) {
drivers/gpu/drm/amd/display/dc/dml/dcn351/dcn351_fpu.c
389
for (i = 0; i < clk_table->num_entries; i++) {
drivers/gpu/drm/amd/display/dc/dml/dcn351/dcn351_fpu.c
391
clk_table->num_entries;
drivers/gpu/drm/amd/display/dc/dml/dcn351/dcn351_fpu.c
408
clk_table->num_entries;
drivers/gpu/drm/amd/display/dc/dml/dcn351/dcn351_fpu.c
410
clk_table->num_entries;
drivers/gpu/drm/amd/display/dc/dml/dcn351/dcn351_fpu.c
412
clk_table->num_entries;
drivers/gpu/drm/amd/display/dc/dml/dcn351/dcn351_fpu.c
414
clk_table->num_entries;
drivers/gpu/drm/amd/display/dc/dml/dcn351/dcn351_fpu.c
416
clk_table->num_entries;
drivers/gpu/drm/amd/display/dc/dml/dcn351/dcn351_fpu.c
418
clk_table->num_entries;
drivers/gpu/drm/amd/display/dc/dml/dcn351/dcn351_fpu.c
420
clk_table->num_entries;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
7088
for (i = 0; i < dram_bw_table->num_entries; i++)
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
7982
mode_lib->ms.max_dram_bw_mbps = ((double)min_clk_table->dram_bw_table.entries[min_clk_table->dram_bw_table.num_entries - 1].pre_derate_dram_bw_kbps / 1000);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_dpmm/dml2_dpmm_dcn4.c
22
for (i = 0; i < dram_bw_table->num_entries; i++)
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_mcg/dml2_mcg_dcn4.c
106
for (i = 0; i < (int)min_table->dram_bw_table.num_entries; i++) {
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_mcg/dml2_mcg_dcn4.c
109
min_table->dram_bw_table.num_entries = i;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_mcg/dml2_mcg_dcn4.c
115
for (i = 0; i < (int)min_table->dram_bw_table.num_entries - 1; i++) {
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_mcg/dml2_mcg_dcn4.c
121
for (j = i + 1; j < min_table->dram_bw_table.num_entries; j++) {
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_mcg/dml2_mcg_dcn4.c
126
min_table->dram_bw_table.num_entries--;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_mcg/dml2_mcg_dcn4.c
142
min_table->dram_bw_table.num_entries = soc_bb->clk_table.uclk.num_clk_values;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_mcg/dml2_mcg_dcn4.c
67
min_table->dram_bw_table.num_entries = soc_bb->clk_table.uclk.num_clk_values;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_mcg/dml2_mcg_dcn4.c
70
for (i = min_table->dram_bw_table.num_entries - 1; i > 0; i--) {
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_mcg/dml2_mcg_dcn4.c
82
for (i = 0; i < (int)min_table->dram_bw_table.num_entries; i++) {
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_top/dml2_top_soc15.c
1164
pmo_init_params.mcg_clock_table_size = dml->min_clk_table.dram_bw_table.num_entries;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_top/dml2_top_soc15.c
15
out->stage1.min_clk_index_for_latency = dml->min_clk_table.dram_bw_table.num_entries - 1; //dml->min_clk_table.clean_me_up.soc_bb.num_states - 1;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/inc/dml2_internal_shared_types.h
27
unsigned int num_entries;
drivers/gpu/drm/amd/display/dc/dml2_0/dml2_wrapper_fpu.c
229
static void copy_dummy_pstate_table(struct dummy_pstate_entry *dest, struct dummy_pstate_entry *src, unsigned int num_entries)
drivers/gpu/drm/amd/display/dc/dml2_0/dml2_wrapper_fpu.c
231
for (int i = 0; i < num_entries; i++) {
drivers/gpu/drm/amd/display/dc/dpp/dcn10/dcn10_dpp_cm.c
857
for (i = 0; i < gamma->num_entries; i++) {
drivers/gpu/drm/amd/display/dc/hwss/dcn30/dcn30_hwseq.c
1201
dc->clk_mgr->funcs->set_max_memclk(dc->clk_mgr, dc->clk_mgr->bw_params->clk_table.entries[dc->clk_mgr->bw_params->clk_table.num_entries - 1].memclk_mhz);
drivers/gpu/drm/amd/display/dc/hwss/dcn32/dcn32_hwseq.c
1811
dc->clk_mgr->funcs->set_max_memclk(dc->clk_mgr, dc->clk_mgr->bw_params->clk_table.entries[dc->clk_mgr->bw_params->clk_table.num_entries - 1].memclk_mhz);
drivers/gpu/drm/amd/display/dc/hwss/dcn401/dcn401_hwseq.c
1397
dc->clk_mgr->funcs->set_max_memclk(dc->clk_mgr, dc->clk_mgr->bw_params->clk_table.entries[dc->clk_mgr->bw_params->clk_table.num_entries - 1].memclk_mhz);
drivers/gpu/drm/amd/display/dc/inc/hw/clk_mgr.h
145
unsigned int num_entries; /* highest populated dpm level for back compatibility */
drivers/gpu/drm/amd/display/dc/resource/dcn30/dcn30_resource.c
2189
num_uclk_states = bw_params->clk_table.num_entries;
drivers/gpu/drm/amd/display/dc/sspl/dc_spl_filters.c
10
int num_entries = NUM_PHASES_COEFF * num_taps;
drivers/gpu/drm/amd/display/dc/sspl/dc_spl_filters.c
13
for (i = 0; i < num_entries; i++)
drivers/gpu/drm/amd/display/dc/sspl/dc_spl_scl_easf_filters.c
2385
unsigned int num_entries)
drivers/gpu/drm/amd/display/dc/sspl/dc_spl_scl_easf_filters.c
2391
lookup_table_index_ptr = (lookup_table_base_ptr + num_entries - 1);
drivers/gpu/drm/amd/display/dc/sspl/dc_spl_scl_easf_filters.c
2394
while (count < num_entries) {
drivers/gpu/drm/amd/display/dc/sspl/dc_spl_scl_easf_filters.c
2414
unsigned int num_entries = sizeof(easf_v_bf3_mode_lookup) /
drivers/gpu/drm/amd/display/dc/sspl/dc_spl_scl_easf_filters.c
2417
easf_v_bf3_mode_lookup, num_entries);
drivers/gpu/drm/amd/display/dc/sspl/dc_spl_scl_easf_filters.c
2423
unsigned int num_entries = sizeof(easf_h_bf3_mode_lookup) /
drivers/gpu/drm/amd/display/dc/sspl/dc_spl_scl_easf_filters.c
2426
easf_h_bf3_mode_lookup, num_entries);
drivers/gpu/drm/amd/display/dc/sspl/dc_spl_scl_easf_filters.c
2432
unsigned int num_entries;
drivers/gpu/drm/amd/display/dc/sspl/dc_spl_scl_easf_filters.c
2435
num_entries = sizeof(easf_reducer_gain6_4tap_lookup) /
drivers/gpu/drm/amd/display/dc/sspl/dc_spl_scl_easf_filters.c
2438
easf_reducer_gain6_4tap_lookup, num_entries);
drivers/gpu/drm/amd/display/dc/sspl/dc_spl_scl_easf_filters.c
2440
num_entries = sizeof(easf_reducer_gain6_6tap_lookup) /
drivers/gpu/drm/amd/display/dc/sspl/dc_spl_scl_easf_filters.c
2443
easf_reducer_gain6_6tap_lookup, num_entries);
drivers/gpu/drm/amd/display/dc/sspl/dc_spl_scl_easf_filters.c
2451
unsigned int num_entries;
drivers/gpu/drm/amd/display/dc/sspl/dc_spl_scl_easf_filters.c
2454
num_entries = sizeof(easf_reducer_gain4_4tap_lookup) /
drivers/gpu/drm/amd/display/dc/sspl/dc_spl_scl_easf_filters.c
2457
easf_reducer_gain4_4tap_lookup, num_entries);
drivers/gpu/drm/amd/display/dc/sspl/dc_spl_scl_easf_filters.c
2459
num_entries = sizeof(easf_reducer_gain4_6tap_lookup) /
drivers/gpu/drm/amd/display/dc/sspl/dc_spl_scl_easf_filters.c
2462
easf_reducer_gain4_6tap_lookup, num_entries);
drivers/gpu/drm/amd/display/dc/sspl/dc_spl_scl_easf_filters.c
2470
unsigned int num_entries;
drivers/gpu/drm/amd/display/dc/sspl/dc_spl_scl_easf_filters.c
2473
num_entries = sizeof(easf_gain_ring6_4tap_lookup) /
drivers/gpu/drm/amd/display/dc/sspl/dc_spl_scl_easf_filters.c
2476
easf_gain_ring6_4tap_lookup, num_entries);
drivers/gpu/drm/amd/display/dc/sspl/dc_spl_scl_easf_filters.c
2478
num_entries = sizeof(easf_gain_ring6_6tap_lookup) /
drivers/gpu/drm/amd/display/dc/sspl/dc_spl_scl_easf_filters.c
2481
easf_gain_ring6_6tap_lookup, num_entries);
drivers/gpu/drm/amd/display/dc/sspl/dc_spl_scl_easf_filters.c
2489
unsigned int num_entries;
drivers/gpu/drm/amd/display/dc/sspl/dc_spl_scl_easf_filters.c
2492
num_entries = sizeof(easf_gain_ring4_4tap_lookup) /
drivers/gpu/drm/amd/display/dc/sspl/dc_spl_scl_easf_filters.c
2495
easf_gain_ring4_4tap_lookup, num_entries);
drivers/gpu/drm/amd/display/dc/sspl/dc_spl_scl_easf_filters.c
2497
num_entries = sizeof(easf_gain_ring4_6tap_lookup) /
drivers/gpu/drm/amd/display/dc/sspl/dc_spl_scl_easf_filters.c
2500
easf_gain_ring4_6tap_lookup, num_entries);
drivers/gpu/drm/amd/display/dc/sspl/dc_spl_scl_easf_filters.c
2509
unsigned int num_entries;
drivers/gpu/drm/amd/display/dc/sspl/dc_spl_scl_easf_filters.c
2512
num_entries = sizeof(easf_3tap_dntilt_uptilt_offset_lookup) /
drivers/gpu/drm/amd/display/dc/sspl/dc_spl_scl_easf_filters.c
2515
easf_3tap_dntilt_uptilt_offset_lookup, num_entries);
drivers/gpu/drm/amd/display/dc/sspl/dc_spl_scl_easf_filters.c
2523
unsigned int num_entries;
drivers/gpu/drm/amd/display/dc/sspl/dc_spl_scl_easf_filters.c
2526
num_entries = sizeof(easf_3tap_uptilt_maxval_lookup) /
drivers/gpu/drm/amd/display/dc/sspl/dc_spl_scl_easf_filters.c
2529
easf_3tap_uptilt_maxval_lookup, num_entries);
drivers/gpu/drm/amd/display/dc/sspl/dc_spl_scl_easf_filters.c
2537
unsigned int num_entries;
drivers/gpu/drm/amd/display/dc/sspl/dc_spl_scl_easf_filters.c
2540
num_entries = sizeof(easf_3tap_dntilt_slope_lookup) /
drivers/gpu/drm/amd/display/dc/sspl/dc_spl_scl_easf_filters.c
2543
easf_3tap_dntilt_slope_lookup, num_entries);
drivers/gpu/drm/amd/display/dc/sspl/dc_spl_scl_easf_filters.c
2551
unsigned int num_entries;
drivers/gpu/drm/amd/display/dc/sspl/dc_spl_scl_easf_filters.c
2554
num_entries = sizeof(easf_3tap_uptilt1_slope_lookup) /
drivers/gpu/drm/amd/display/dc/sspl/dc_spl_scl_easf_filters.c
2557
easf_3tap_uptilt1_slope_lookup, num_entries);
drivers/gpu/drm/amd/display/dc/sspl/dc_spl_scl_easf_filters.c
2565
unsigned int num_entries;
drivers/gpu/drm/amd/display/dc/sspl/dc_spl_scl_easf_filters.c
2568
num_entries = sizeof(easf_3tap_uptilt2_slope_lookup) /
drivers/gpu/drm/amd/display/dc/sspl/dc_spl_scl_easf_filters.c
2571
easf_3tap_uptilt2_slope_lookup, num_entries);
drivers/gpu/drm/amd/display/dc/sspl/dc_spl_scl_easf_filters.c
2579
unsigned int num_entries;
drivers/gpu/drm/amd/display/dc/sspl/dc_spl_scl_easf_filters.c
2582
num_entries = sizeof(easf_3tap_uptilt2_offset_lookup) /
drivers/gpu/drm/amd/display/dc/sspl/dc_spl_scl_easf_filters.c
2585
easf_3tap_uptilt2_offset_lookup, num_entries);
drivers/gpu/drm/amd/display/modules/color/color_gamma.c
1284
struct pwl_float_data *rgb_last = rgb + ramp->num_entries - 1;
drivers/gpu/drm/amd/display/modules/color/color_gamma.c
1296
} while (i != ramp->num_entries);
drivers/gpu/drm/amd/display/modules/color/color_gamma.c
1310
} while (i != ramp->num_entries);
drivers/gpu/drm/amd/display/modules/color/color_gamma.c
1349
for (i = 0 ; i < ramp->num_entries; i++) {
drivers/gpu/drm/amd/display/modules/color/color_gamma.c
1374
for (i = 0 ; i < ramp->num_entries; i++) {
drivers/gpu/drm/amd/display/modules/color/color_gamma.c
1582
max_entries += ramp->num_entries;
drivers/gpu/drm/amd/display/modules/color/color_gamma.c
1718
ramp->num_entries + _EXTRA_POINTS);
drivers/gpu/drm/amd/display/modules/color/color_gamma.c
1723
ramp->num_entries + _EXTRA_POINTS);
drivers/gpu/drm/amd/display/modules/color/color_gamma.c
1733
ramp->num_entries,
drivers/gpu/drm/amd/display/modules/color/color_gamma.c
1941
ramp->num_entries + _EXTRA_POINTS);
drivers/gpu/drm/amd/display/modules/color/color_gamma.c
1945
axis_x = kvzalloc_objs(*axis_x, ramp->num_entries + 3);
drivers/gpu/drm/amd/display/modules/color/color_gamma.c
1955
ramp->num_entries,
drivers/gpu/drm/amd/display/modules/color/color_gamma.c
621
const uint32_t max_number = ramp->num_entries + 3;
drivers/gpu/drm/amd/display/modules/color/color_gamma.c
728
if (index_left >= ramp->num_entries + 3) {
drivers/gpu/drm/amd/display/modules/color/color_gamma.c
733
if (index_right >= ramp->num_entries + 3) {
drivers/gpu/drm/amd/display/modules/power/power_helpers.c
249
unsigned int num_entries = NUM_BL_CURVE_SEGS;
drivers/gpu/drm/amd/display/modules/power/power_helpers.c
254
table->backlight_thresholds[num_entries-1] = 0xFFFF;
drivers/gpu/drm/amd/display/modules/power/power_helpers.c
255
table->backlight_offsets[num_entries-1] =
drivers/gpu/drm/amd/display/modules/power/power_helpers.c
266
for (i = 1; i+1 < num_entries; i++) {
drivers/gpu/drm/amd/display/modules/power/power_helpers.c
267
lut_index = (params.backlight_lut_array_size - 1) * i / (num_entries - 1);
drivers/gpu/drm/amd/display/modules/power/power_helpers.c
271
cpu_to_be16(DIV_ROUNDUP((i * 65536), num_entries));
drivers/gpu/drm/amd/display/modules/power/power_helpers.c
281
unsigned int num_entries = NUM_BL_CURVE_SEGS;
drivers/gpu/drm/amd/display/modules/power/power_helpers.c
286
table->backlight_thresholds[num_entries-1] = 0xFFFF;
drivers/gpu/drm/amd/display/modules/power/power_helpers.c
287
table->backlight_offsets[num_entries-1] =
drivers/gpu/drm/amd/display/modules/power/power_helpers.c
298
for (i = 1; i+1 < num_entries; i++) {
drivers/gpu/drm/amd/display/modules/power/power_helpers.c
299
lut_index = DIV_ROUNDUP((i * params.backlight_lut_array_size), num_entries);
drivers/gpu/drm/amd/display/modules/power/power_helpers.c
303
cpu_to_be16(DIV_ROUNDUP((i * 65536), num_entries)) :
drivers/gpu/drm/amd/display/modules/power/power_helpers.c
304
cpu_to_le16(DIV_ROUNDUP((i * 65536), num_entries));
drivers/gpu/drm/amd/pm/legacy-dpm/kv_dpm.c
109
for (i = 0; i < vid_mapping_table->num_entries; i++) {
drivers/gpu/drm/amd/pm/legacy-dpm/kv_dpm.c
114
return vid_mapping_table->entries[vid_mapping_table->num_entries - 1].vid_2bit;
drivers/gpu/drm/amd/pm/legacy-dpm/kv_dpm.c
192
vid_mapping_table->num_entries = i;
drivers/gpu/drm/amd/pm/legacy-dpm/kv_dpm.c
86
for (i = 0; i < vid_mapping_table->num_entries; i++) {
drivers/gpu/drm/amd/pm/legacy-dpm/kv_dpm.c
90
return vid_mapping_table->entries[vid_mapping_table->num_entries - 1].vid_7bit;
drivers/gpu/drm/amd/pm/legacy-dpm/kv_dpm.h
43
u32 num_entries;
drivers/gpu/drm/amd/pm/legacy-dpm/si_dpm.c
5894
for (k = 0; k < table->num_entries; k++)
drivers/gpu/drm/amd/pm/legacy-dpm/si_dpm.c
5905
for (k = 0; k < table->num_entries; k++) {
drivers/gpu/drm/amd/pm/legacy-dpm/si_dpm.c
5919
for (k = 0; k < table->num_entries; k++)
drivers/gpu/drm/amd/pm/legacy-dpm/si_dpm.c
5929
for(k = 0; k < table->num_entries; k++)
drivers/gpu/drm/amd/pm/legacy-dpm/si_dpm.c
6004
for (j = 1; j < table->num_entries; j++) {
drivers/gpu/drm/amd/pm/legacy-dpm/si_dpm.c
6031
if (table->num_entries > MAX_AC_TIMING_ENTRIES)
drivers/gpu/drm/amd/pm/legacy-dpm/si_dpm.c
6038
for (i = 0; i < table->num_entries; i++) {
drivers/gpu/drm/amd/pm/legacy-dpm/si_dpm.c
6046
si_table->num_entries = table->num_entries;
drivers/gpu/drm/amd/pm/legacy-dpm/si_dpm.c
6123
u32 num_entries, u32 valid_flag)
drivers/gpu/drm/amd/pm/legacy-dpm/si_dpm.c
6127
for(i = 0, j = 0; j < num_entries; j++) {
drivers/gpu/drm/amd/pm/legacy-dpm/si_dpm.c
6142
for (i = 0; i < si_pi->mc_reg_table.num_entries; i++) {
drivers/gpu/drm/amd/pm/legacy-dpm/si_dpm.c
6147
if ((i == si_pi->mc_reg_table.num_entries) && (i > 0))
drivers/gpu/drm/amd/pm/legacy-dpm/si_dpm.h
381
u8 num_entries;
drivers/gpu/drm/amd/pm/powerplay/hwmgr/ppatomctrl.c
106
uint8_t num_entries = (uint8_t)((le16_to_cpu(reg_block->usRegIndexTblSize))
drivers/gpu/drm/amd/pm/powerplay/hwmgr/ppatomctrl.c
110
num_entries--; /* subtract 1 data end mark entry */
drivers/gpu/drm/amd/pm/powerplay/hwmgr/ppatomctrl.c
112
PP_ASSERT_WITH_CODE((num_entries <= VBIOS_MC_REGISTER_ARRAY_SIZE),
drivers/gpu/drm/amd/pm/powerplay/hwmgr/ppatomctrl.c
117
(i < num_entries)) {
drivers/gpu/drm/amd/pm/powerplay/hwmgr/ppatomctrl.c
89
table->num_entries = num_ranges;
drivers/gpu/drm/amd/pm/powerplay/hwmgr/ppatomctrl.h
184
uint8_t num_entries;
drivers/gpu/drm/amd/pm/powerplay/hwmgr/ppatomctrl.h
240
uint8_t num_entries; /* number of AC timing entries */
drivers/gpu/drm/amd/pm/powerplay/hwmgr/vega10_processpptables.c
731
uint8_t num_entries;
drivers/gpu/drm/amd/pm/powerplay/hwmgr/vega10_processpptables.c
752
num_entries = clk_dep_table->ucNumEntries + 1 > NUM_DSPCLK_LEVELS ?
drivers/gpu/drm/amd/pm/powerplay/hwmgr/vega10_processpptables.c
755
num_entries = clk_dep_table->ucNumEntries;
drivers/gpu/drm/amd/pm/powerplay/hwmgr/vega10_processpptables.c
758
clk_table = kzalloc_flex(*clk_table, entries, num_entries);
drivers/gpu/drm/amd/pm/powerplay/hwmgr/vega10_processpptables.c
762
clk_table->count = (uint32_t)num_entries;
drivers/gpu/drm/amd/pm/powerplay/hwmgr/vega10_processpptables.c
771
if (i < num_entries) {
drivers/gpu/drm/amd/pm/powerplay/inc/smu_ucode_xfer_vi.h
97
uint32_t num_entries;
drivers/gpu/drm/amd/pm/powerplay/smumgr/ci_smumgr.c
1750
uint32_t num_entries, uint32_t valid_flag)
drivers/gpu/drm/amd/pm/powerplay/smumgr/ci_smumgr.c
1754
for (i = 0, j = 0; j < num_entries; j++) {
drivers/gpu/drm/amd/pm/powerplay/smumgr/ci_smumgr.c
1771
for (i = 0; i < smu_data->mc_reg_table.num_entries; i++) {
drivers/gpu/drm/amd/pm/powerplay/smumgr/ci_smumgr.c
1778
if ((i == smu_data->mc_reg_table.num_entries) && (i > 0))
drivers/gpu/drm/amd/pm/powerplay/smumgr/ci_smumgr.c
2561
PP_ASSERT_WITH_CODE((table->num_entries <= MAX_AC_TIMING_ENTRIES),
drivers/gpu/drm/amd/pm/powerplay/smumgr/ci_smumgr.c
2569
for (i = 0; i < table->num_entries; i++) {
drivers/gpu/drm/amd/pm/powerplay/smumgr/ci_smumgr.c
2578
ni_table->num_entries = table->num_entries;
drivers/gpu/drm/amd/pm/powerplay/smumgr/ci_smumgr.c
2600
for (k = 0; k < table->num_entries; k++) {
drivers/gpu/drm/amd/pm/powerplay/smumgr/ci_smumgr.c
2612
for (k = 0; k < table->num_entries; k++) {
drivers/gpu/drm/amd/pm/powerplay/smumgr/ci_smumgr.c
2627
for (k = 0; k < table->num_entries; k++) {
drivers/gpu/drm/amd/pm/powerplay/smumgr/ci_smumgr.c
2640
for (k = 0; k < table->num_entries; k++) {
drivers/gpu/drm/amd/pm/powerplay/smumgr/ci_smumgr.c
2664
for (j = 1; j < table->num_entries; j++) {
drivers/gpu/drm/amd/pm/powerplay/smumgr/ci_smumgr.h
55
uint8_t num_entries;
drivers/gpu/drm/amd/pm/powerplay/smumgr/iceland_smumgr.c
1716
uint32_t num_entries, uint32_t valid_flag)
drivers/gpu/drm/amd/pm/powerplay/smumgr/iceland_smumgr.c
1720
for (i = 0, j = 0; j < num_entries; j++) {
drivers/gpu/drm/amd/pm/powerplay/smumgr/iceland_smumgr.c
1736
for (i = 0; i < smu_data->mc_reg_table.num_entries; i++) {
drivers/gpu/drm/amd/pm/powerplay/smumgr/iceland_smumgr.c
1743
if ((i == smu_data->mc_reg_table.num_entries) && (i > 0))
drivers/gpu/drm/amd/pm/powerplay/smumgr/iceland_smumgr.c
2488
PP_ASSERT_WITH_CODE((table->num_entries <= MAX_AC_TIMING_ENTRIES),
drivers/gpu/drm/amd/pm/powerplay/smumgr/iceland_smumgr.c
2496
for (i = 0; i < table->num_entries; i++) {
drivers/gpu/drm/amd/pm/powerplay/smumgr/iceland_smumgr.c
2505
ni_table->num_entries = table->num_entries;
drivers/gpu/drm/amd/pm/powerplay/smumgr/iceland_smumgr.c
2527
for (k = 0; k < table->num_entries; k++) {
drivers/gpu/drm/amd/pm/powerplay/smumgr/iceland_smumgr.c
2539
for (k = 0; k < table->num_entries; k++) {
drivers/gpu/drm/amd/pm/powerplay/smumgr/iceland_smumgr.c
2555
for (k = 0; k < table->num_entries; k++) {
drivers/gpu/drm/amd/pm/powerplay/smumgr/iceland_smumgr.c
2568
for (k = 0; k < table->num_entries; k++) {
drivers/gpu/drm/amd/pm/powerplay/smumgr/iceland_smumgr.c
2591
for (j = 1; j < table->num_entries; j++) {
drivers/gpu/drm/amd/pm/powerplay/smumgr/iceland_smumgr.h
54
uint8_t num_entries; /* number of entries in mc_reg_table_entry used*/
drivers/gpu/drm/amd/pm/powerplay/smumgr/smu7_smumgr.c
340
toc->num_entries = 0;
drivers/gpu/drm/amd/pm/powerplay/smumgr/smu7_smumgr.c
344
UCODE_ID_RLC_G, &toc->entry[toc->num_entries++]),
drivers/gpu/drm/amd/pm/powerplay/smumgr/smu7_smumgr.c
347
UCODE_ID_CP_CE, &toc->entry[toc->num_entries++]),
drivers/gpu/drm/amd/pm/powerplay/smumgr/smu7_smumgr.c
350
UCODE_ID_CP_PFP, &toc->entry[toc->num_entries++]),
drivers/gpu/drm/amd/pm/powerplay/smumgr/smu7_smumgr.c
353
UCODE_ID_CP_ME, &toc->entry[toc->num_entries++]),
drivers/gpu/drm/amd/pm/powerplay/smumgr/smu7_smumgr.c
356
UCODE_ID_CP_MEC, &toc->entry[toc->num_entries++]),
drivers/gpu/drm/amd/pm/powerplay/smumgr/smu7_smumgr.c
359
UCODE_ID_CP_MEC_JT1, &toc->entry[toc->num_entries++]),
drivers/gpu/drm/amd/pm/powerplay/smumgr/smu7_smumgr.c
362
UCODE_ID_CP_MEC_JT2, &toc->entry[toc->num_entries++]),
drivers/gpu/drm/amd/pm/powerplay/smumgr/smu7_smumgr.c
365
UCODE_ID_SDMA0, &toc->entry[toc->num_entries++]),
drivers/gpu/drm/amd/pm/powerplay/smumgr/smu7_smumgr.c
368
UCODE_ID_SDMA1, &toc->entry[toc->num_entries++]),
drivers/gpu/drm/amd/pm/powerplay/smumgr/smu7_smumgr.c
372
UCODE_ID_MEC_STORAGE, &toc->entry[toc->num_entries++]),
drivers/gpu/drm/amd/pm/powerplay/smumgr/tonga_smumgr.c
2094
uint32_t num_entries, uint32_t valid_flag)
drivers/gpu/drm/amd/pm/powerplay/smumgr/tonga_smumgr.c
2098
for (i = 0, j = 0; j < num_entries; j++) {
drivers/gpu/drm/amd/pm/powerplay/smumgr/tonga_smumgr.c
2115
for (i = 0; i < smu_data->mc_reg_table.num_entries; i++) {
drivers/gpu/drm/amd/pm/powerplay/smumgr/tonga_smumgr.c
2122
if ((i == smu_data->mc_reg_table.num_entries) && (i > 0))
drivers/gpu/drm/amd/pm/powerplay/smumgr/tonga_smumgr.c
2952
PP_ASSERT_WITH_CODE((table->num_entries <= MAX_AC_TIMING_ENTRIES),
drivers/gpu/drm/amd/pm/powerplay/smumgr/tonga_smumgr.c
2960
for (i = 0; i < table->num_entries; i++) {
drivers/gpu/drm/amd/pm/powerplay/smumgr/tonga_smumgr.c
2969
ni_table->num_entries = table->num_entries;
drivers/gpu/drm/amd/pm/powerplay/smumgr/tonga_smumgr.c
2992
for (k = 0; k < table->num_entries; k++) {
drivers/gpu/drm/amd/pm/powerplay/smumgr/tonga_smumgr.c
3004
for (k = 0; k < table->num_entries; k++) {
drivers/gpu/drm/amd/pm/powerplay/smumgr/tonga_smumgr.c
3019
for (k = 0; k < table->num_entries; k++)
drivers/gpu/drm/amd/pm/powerplay/smumgr/tonga_smumgr.c
3031
for (k = 0; k < table->num_entries; k++) {
drivers/gpu/drm/amd/pm/powerplay/smumgr/tonga_smumgr.c
3055
for (j = 1; j < table->num_entries; j++) {
drivers/gpu/drm/amd/pm/powerplay/smumgr/tonga_smumgr.h
56
uint8_t num_entries; /* number of entries in mc_reg_table_entry used*/
drivers/gpu/drm/bridge/sil-sii8620.c
1742
d->num_entries = 1;
drivers/gpu/drm/bridge/sil-sii8620.c
737
d->num_entries = 1;
drivers/gpu/drm/i915/display/intel_bios.c
240
if (ptrs->num_entries != 3)
drivers/gpu/drm/i915/display/intel_bios.c
416
ptrs->num_entries++;
drivers/gpu/drm/i915/display/intel_bios.c
418
ptrs->num_entries++;
drivers/gpu/drm/i915/display/intel_bios.c
420
ptrs->num_entries++;
drivers/gpu/drm/i915/display/intel_bios.c
422
if (size != 0 || ptrs->num_entries != 3) {
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
1012
.num_entries = ARRAY_SIZE(_dg2_snps_trans),
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
1037
.num_entries = ARRAY_SIZE(_dg2_snps_trans_uhbr),
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
1055
.num_entries = ARRAY_SIZE(_mtl_c10_trans_dp14),
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
106
.num_entries = ARRAY_SIZE(_bdw_trans_dp),
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
1104
.num_entries = ARRAY_SIZE(_mtl_c20_trans_hdmi),
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
1110
.num_entries = ARRAY_SIZE(_mtl_c20_trans_dp14),
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
1116
.num_entries = ARRAY_SIZE(_mtl_c20_trans_uhbr),
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
1169
.num_entries = ARRAY_SIZE(_xe3plpd_lt_trans_dp14),
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
1174
.num_entries = ARRAY_SIZE(_xe3plpd_lt_trans_uhbr),
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
1179
.num_entries = ARRAY_SIZE(_xe3plpd_lt_trans_edp),
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
1204
intel_get_buf_trans(const struct intel_ddi_buf_trans *trans, int *num_entries)
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
1206
*num_entries = trans->num_entries;
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
123
.num_entries = ARRAY_SIZE(_bdw_trans_fdi),
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
142
.num_entries = ARRAY_SIZE(_bdw_trans_hdmi),
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
161
.num_entries = ARRAY_SIZE(_skl_trans_dp),
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
179
.num_entries = ARRAY_SIZE(_skl_u_trans_dp),
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
197
.num_entries = ARRAY_SIZE(_skl_y_trans_dp),
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
215
.num_entries = ARRAY_SIZE(_kbl_trans_dp),
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
233
.num_entries = ARRAY_SIZE(_kbl_u_trans_dp),
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
251
.num_entries = ARRAY_SIZE(_kbl_y_trans_dp),
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
273
.num_entries = ARRAY_SIZE(_skl_trans_edp),
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
295
.num_entries = ARRAY_SIZE(_skl_u_trans_edp),
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
317
.num_entries = ARRAY_SIZE(_skl_y_trans_edp),
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
33
.num_entries = ARRAY_SIZE(_hsw_trans_dp),
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
337
.num_entries = ARRAY_SIZE(_skl_trans_hdmi),
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
358
.num_entries = ARRAY_SIZE(_skl_y_trans_hdmi),
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
378
.num_entries = ARRAY_SIZE(_bxt_trans_dp),
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
397
.num_entries = ARRAY_SIZE(_bxt_trans_edp),
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
419
.num_entries = ARRAY_SIZE(_bxt_trans_hdmi),
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
440
.num_entries = ARRAY_SIZE(_icl_combo_phy_trans_dp_hbr2_edp_hbr3),
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
459
.num_entries = ARRAY_SIZE(_icl_combo_phy_trans_edp_hbr2),
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
475
.num_entries = ARRAY_SIZE(_icl_combo_phy_trans_hdmi),
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
495
.num_entries = ARRAY_SIZE(_ehl_combo_phy_trans_dp),
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
50
.num_entries = ARRAY_SIZE(_hsw_trans_fdi),
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
514
.num_entries = ARRAY_SIZE(_ehl_combo_phy_trans_edp_hbr2),
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
533
.num_entries = ARRAY_SIZE(_jsl_combo_phy_trans_edp_hbr),
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
552
.num_entries = ARRAY_SIZE(_jsl_combo_phy_trans_edp_hbr2),
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
571
.num_entries = ARRAY_SIZE(_dg1_combo_phy_trans_dp_rbr_hbr),
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
590
.num_entries = ARRAY_SIZE(_dg1_combo_phy_trans_dp_hbr2_hbr3),
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
609
.num_entries = ARRAY_SIZE(_icl_mg_phy_trans_rbr_hbr),
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
628
.num_entries = ARRAY_SIZE(_icl_mg_phy_trans_hbr2_hbr3),
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
647
.num_entries = ARRAY_SIZE(_icl_mg_phy_trans_hdmi),
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
667
.num_entries = ARRAY_SIZE(_tgl_dkl_phy_trans_dp_hbr),
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
686
.num_entries = ARRAY_SIZE(_tgl_dkl_phy_trans_dp_hbr2),
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
705
.num_entries = ARRAY_SIZE(_tgl_dkl_phy_trans_hdmi),
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
71
.num_entries = ARRAY_SIZE(_hsw_trans_hdmi),
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
725
.num_entries = ARRAY_SIZE(_tgl_combo_phy_trans_dp_hbr),
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
744
.num_entries = ARRAY_SIZE(_tgl_combo_phy_trans_dp_hbr2),
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
763
.num_entries = ARRAY_SIZE(_tgl_uy_combo_phy_trans_dp_hbr2),
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
785
.num_entries = ARRAY_SIZE(_tgl_combo_phy_trans_edp_hbr2_hobl),
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
804
.num_entries = ARRAY_SIZE(_rkl_combo_phy_trans_dp_hbr),
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
823
.num_entries = ARRAY_SIZE(_rkl_combo_phy_trans_dp_hbr2_hbr3),
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
842
.num_entries = ARRAY_SIZE(_adls_combo_phy_trans_dp_hbr2_hbr3),
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
861
.num_entries = ARRAY_SIZE(_adls_combo_phy_trans_edp_hbr2),
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
880
.num_entries = ARRAY_SIZE(_adls_combo_phy_trans_edp_hbr3),
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
89
.num_entries = ARRAY_SIZE(_bdw_trans_edp),
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
899
.num_entries = ARRAY_SIZE(_adlp_combo_phy_trans_dp_hbr),
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
946
.num_entries = ARRAY_SIZE(_adlp_combo_phy_trans_dp_hbr2_hbr3),
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
951
.num_entries = ARRAY_SIZE(_adlp_combo_phy_trans_dp_hbr2_edp_hbr3),
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
956
.num_entries = ARRAY_SIZE(_adlp_combo_phy_trans_edp_hbr2),
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
975
.num_entries = ARRAY_SIZE(_adlp_dkl_phy_trans_dp_hbr),
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.c
994
.num_entries = ARRAY_SIZE(_adlp_dkl_phy_trans_dp_hbr2_hbr3),
drivers/gpu/drm/i915/display/intel_ddi_buf_trans.h
73
u8 num_entries;
drivers/gpu/drm/i915/display/intel_dmc.c
1200
u32 num_entries, max_entries;
drivers/gpu/drm/i915/display/intel_dmc.c
1230
num_entries = package_header->num_entries;
drivers/gpu/drm/i915/display/intel_dmc.c
1231
if (WARN_ON(num_entries > max_entries))
drivers/gpu/drm/i915/display/intel_dmc.c
1232
num_entries = max_entries;
drivers/gpu/drm/i915/display/intel_dmc.c
1236
dmc_set_fw_offset(dmc, fw_info, num_entries, si,
drivers/gpu/drm/i915/display/intel_dmc.c
332
u32 num_entries;
drivers/gpu/drm/i915/display/intel_dmc.c
989
unsigned int num_entries,
drivers/gpu/drm/i915/display/intel_dmc.c
997
for (i = 0; i < num_entries; i++) {
drivers/gpu/drm/i915/display/intel_vbt_defs.h
1131
u16 num_entries;
drivers/gpu/drm/i915/display/intel_vbt_defs.h
1150
u8 num_entries;
drivers/gpu/drm/i915/display/intel_vbt_defs.h
1164
u16 num_entries;
drivers/gpu/drm/i915/display/intel_vbt_defs.h
1183
u8 num_entries;
drivers/gpu/drm/i915/display/intel_vbt_defs.h
1238
u8 num_entries;
drivers/gpu/drm/i915/display/intel_vbt_defs.h
1658
u16 num_entries; /* ???-216 */
drivers/gpu/drm/i915/display/intel_vbt_defs.h
1667
u16 num_entries; /* 217+ */
drivers/gpu/drm/i915/display/intel_vbt_defs.h
605
u16 num_entries; /* ALM only */
drivers/gpu/drm/i915/display/intel_vbt_defs.h
884
u16 num_entries;
drivers/gpu/drm/i915/display/intel_vbt_defs.h
924
u8 num_entries;
drivers/gpu/drm/i915/display/intel_vbt_defs.h
948
u8 num_entries;
drivers/gpu/drm/i915/display/intel_vbt_defs.h
963
u8 num_entries;
drivers/gpu/drm/i915/display/skl_watermark.c
2412
int num_entries, int ignore_idx)
drivers/gpu/drm/i915/display/skl_watermark.c
2416
for (i = 0; i < num_entries; i++) {
drivers/gpu/drm/i915/display/skl_watermark.h
35
int num_entries, int ignore_idx);
drivers/gpu/drm/i915/gt/gen6_ppgtt.c
82
unsigned int num_entries = length / I915_GTT_PAGE_SIZE;
drivers/gpu/drm/i915/gt/gen6_ppgtt.c
84
while (num_entries) {
drivers/gpu/drm/i915/gt/gen6_ppgtt.c
87
const unsigned int count = min(num_entries, GEN6_PTES - pte);
drivers/gpu/drm/i915/gt/gen6_ppgtt.c
90
num_entries -= count;
drivers/gpu/drm/i915/gt/intel_engine_cs.c
2144
const u8 num_entries = execlists->csb_size;
drivers/gpu/drm/i915/gt/intel_engine_cs.c
2160
read, write, num_entries);
drivers/gpu/drm/i915/gt/intel_engine_cs.c
2162
if (read >= num_entries)
drivers/gpu/drm/i915/gt/intel_engine_cs.c
2164
if (write >= num_entries)
drivers/gpu/drm/i915/gt/intel_engine_cs.c
2167
write += num_entries;
drivers/gpu/drm/i915/gt/intel_engine_cs.c
2169
idx = ++read % num_entries;
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
1818
const u8 num_entries = execlists->csb_size;
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
1882
if (++head == num_entries)
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
2010
drm_clflush_virt_range(&buf[0], num_entries * sizeof(buf[0]));
drivers/gpu/drm/i915/gt/intel_ggtt.c
341
struct sg_table *pages, u32 num_entries,
drivers/gpu/drm/i915/gt/intel_ggtt.c
353
if (!num_entries)
drivers/gpu/drm/i915/gt/intel_ggtt.c
363
while (num_entries) {
drivers/gpu/drm/i915/gt/intel_ggtt.c
370
u32 n_ptes = min_t(u32, 511, num_entries);
drivers/gpu/drm/i915/gt/intel_ggtt.c
428
num_entries -= n_ptes;
drivers/gpu/drm/i915/gt/intel_ggtt.c
580
unsigned int num_entries = length / I915_GTT_PAGE_SIZE;
drivers/gpu/drm/i915/gt/intel_ggtt.c
587
if (WARN(num_entries > max_entries,
drivers/gpu/drm/i915/gt/intel_ggtt.c
589
first_entry, num_entries, max_entries))
drivers/gpu/drm/i915/gt/intel_ggtt.c
590
num_entries = max_entries;
drivers/gpu/drm/i915/gt/intel_ggtt.c
592
for (i = 0; i < num_entries; i++)
drivers/gpu/drm/i915/gt/intel_ggtt.c
601
unsigned int num_entries = length / I915_GTT_PAGE_SIZE;
drivers/gpu/drm/i915/gt/intel_ggtt.c
605
if (WARN(num_entries > max_entries,
drivers/gpu/drm/i915/gt/intel_ggtt.c
607
first_entry, num_entries, max_entries))
drivers/gpu/drm/i915/gt/intel_ggtt.c
608
num_entries = max_entries;
drivers/gpu/drm/i915/gt/intel_ggtt.c
611
NULL, num_entries, scratch_pte))
drivers/gpu/drm/i915/gt/intel_ggtt.c
761
unsigned int num_entries = length / I915_GTT_PAGE_SIZE;
drivers/gpu/drm/i915/gt/intel_ggtt.c
767
if (WARN(num_entries > max_entries,
drivers/gpu/drm/i915/gt/intel_ggtt.c
769
first_entry, num_entries, max_entries))
drivers/gpu/drm/i915/gt/intel_ggtt.c
770
num_entries = max_entries;
drivers/gpu/drm/i915/gt/intel_ggtt.c
773
for (i = 0; i < num_entries; i++)
drivers/gpu/drm/i915/gt/uc/intel_guc_capture.c
413
struct guc_mmio_reg *ptr, u16 num_entries)
drivers/gpu/drm/i915/gt/uc/intel_guc_capture.c
428
for (i = 0; i < num_entries && i < match->num_regs; ++i) {
drivers/gpu/drm/i915/gt/uc/intel_guc_capture.c
437
for (i = match->num_regs, j = 0; i < num_entries &&
drivers/gpu/drm/i915/gt/uc/intel_guc_capture.c
446
if (i < num_entries)
drivers/gpu/drm/i915/gt/uc/intel_guc_capture.c
447
guc_dbg(guc, "Got short capture reglist init: %d out %d.\n", i, num_entries);
drivers/gpu/drm/i915/gvt/gtt.c
2732
u32 num_entries;
drivers/gpu/drm/i915/gvt/gtt.c
2738
num_entries = vgpu_aperture_sz(vgpu) >> PAGE_SHIFT;
drivers/gpu/drm/i915/gvt/gtt.c
2739
while (num_entries--) {
drivers/gpu/drm/i915/gvt/gtt.c
2748
num_entries = vgpu_hidden_sz(vgpu) >> PAGE_SHIFT;
drivers/gpu/drm/i915/gvt/gtt.c
2749
while (num_entries--) {
drivers/gpu/drm/nouveau/dispnv50/crc.c
137
while (crc->entry_idx < func->num_entries) {
drivers/gpu/drm/nouveau/dispnv50/crc.h
59
short num_entries;
drivers/gpu/drm/nouveau/dispnv50/crc907d.c
140
.num_entries = CRC907D_MAX_ENTRIES,
drivers/gpu/drm/nouveau/dispnv50/crcc37d.c
125
.num_entries = CRCC37D_MAX_ENTRIES,
drivers/gpu/drm/nouveau/dispnv50/crcc57d.c
56
.num_entries = CRCC37D_MAX_ENTRIES,
drivers/gpu/drm/nouveau/dispnv50/crcca7d.c
96
.num_entries = CRCC37D_MAX_ENTRIES,
drivers/gpu/drm/nouveau/nouveau_bios.c
706
uint8_t version, headerlen, entrylen, num_entries;
drivers/gpu/drm/nouveau/nouveau_bios.c
730
num_entries = bios->data[load_table_ptr + 3];
drivers/gpu/drm/nouveau/nouveau_bios.c
732
if (headerlen != 4 || entrylen != 4 || num_entries != 2) {
drivers/gpu/drm/radeon/btc_dpm.c
1876
for (j = 1; j < table->num_entries; j++) {
drivers/gpu/drm/radeon/btc_dpm.c
1899
for (k = 0; k < table->num_entries; k++) {
drivers/gpu/drm/radeon/btc_dpm.c
1912
for (k = 0; k < table->num_entries; k++) {
drivers/gpu/drm/radeon/btc_dpm.c
1928
for (k = 0; k < table->num_entries; k++) {
drivers/gpu/drm/radeon/btc_dpm.c
1968
if (table->num_entries > MAX_AC_TIMING_ENTRIES)
drivers/gpu/drm/radeon/btc_dpm.c
1975
for (i = 0; i < table->num_entries; i++) {
drivers/gpu/drm/radeon/btc_dpm.c
1982
eg_table->num_entries = table->num_entries;
drivers/gpu/drm/radeon/ci_dpm.c
4301
for (k = 0; k < table->num_entries; k++) {
drivers/gpu/drm/radeon/ci_dpm.c
4312
for (k = 0; k < table->num_entries; k++) {
drivers/gpu/drm/radeon/ci_dpm.c
4325
for (k = 0; k < table->num_entries; k++) {
drivers/gpu/drm/radeon/ci_dpm.c
4338
for (k = 0; k < table->num_entries; k++) {
drivers/gpu/drm/radeon/ci_dpm.c
4435
for (j = 1; j < table->num_entries; j++) {
drivers/gpu/drm/radeon/ci_dpm.c
4464
if (table->num_entries > MAX_AC_TIMING_ENTRIES)
drivers/gpu/drm/radeon/ci_dpm.c
4472
for (i = 0; i < table->num_entries; i++) {
drivers/gpu/drm/radeon/ci_dpm.c
4479
ci_table->num_entries = table->num_entries;
drivers/gpu/drm/radeon/ci_dpm.c
4502
for (k = 0; k < table->num_entries; k++) {
drivers/gpu/drm/radeon/ci_dpm.c
4511
for (k = 0; k < table->num_entries; k++) {
drivers/gpu/drm/radeon/ci_dpm.c
4520
for (k = 0; k < table->num_entries; k++) {
drivers/gpu/drm/radeon/ci_dpm.c
4529
for (k = 0; k < table->num_entries; k++) {
drivers/gpu/drm/radeon/ci_dpm.c
4536
for (k = 0; k < table->num_entries; k++) {
drivers/gpu/drm/radeon/ci_dpm.c
4548
for (k = 0; k < table->num_entries; k++) {
drivers/gpu/drm/radeon/ci_dpm.c
4656
u32 num_entries, u32 valid_flag)
drivers/gpu/drm/radeon/ci_dpm.c
4660
for (i = 0, j = 0; j < num_entries; j++) {
drivers/gpu/drm/radeon/ci_dpm.c
4675
for (i = 0; i < pi->mc_reg_table.num_entries; i++) {
drivers/gpu/drm/radeon/ci_dpm.c
4680
if ((i == pi->mc_reg_table.num_entries) && (i > 0))
drivers/gpu/drm/radeon/ci_dpm.h
84
u8 num_entries;
drivers/gpu/drm/radeon/cypress_dpm.c
1051
for (i = 0; i < range_table->num_entries; i++) {
drivers/gpu/drm/radeon/cypress_dpm.c
1059
eg_pi->mc_reg_table.num_entries = range_table->num_entries;
drivers/gpu/drm/radeon/cypress_dpm.c
1063
for (j = 1; j < range_table->num_entries; j++) {
drivers/gpu/drm/radeon/cypress_dpm.c
818
u32 num_entries, u32 valid_flag)
drivers/gpu/drm/radeon/cypress_dpm.c
822
for (i = 0, j = 0; j < num_entries; j++) {
drivers/gpu/drm/radeon/cypress_dpm.c
837
for (i = 0; i < eg_pi->mc_reg_table.num_entries; i++) {
drivers/gpu/drm/radeon/cypress_dpm.c
843
if ((i == eg_pi->mc_reg_table.num_entries) && (i > 0))
drivers/gpu/drm/radeon/cypress_dpm.h
36
u8 num_entries;
drivers/gpu/drm/radeon/kv_dpm.c
407
for (i = 0; i < vid_mapping_table->num_entries; i++) {
drivers/gpu/drm/radeon/kv_dpm.c
411
return vid_mapping_table->entries[vid_mapping_table->num_entries - 1].vid_7bit;
drivers/gpu/drm/radeon/kv_dpm.c
430
for (i = 0; i < vid_mapping_table->num_entries; i++) {
drivers/gpu/drm/radeon/kv_dpm.c
435
return vid_mapping_table->entries[vid_mapping_table->num_entries - 1].vid_2bit;
drivers/gpu/drm/radeon/ni_dpm.c
2729
for (k = 0; k < table->num_entries; k++)
drivers/gpu/drm/radeon/ni_dpm.c
2740
for(k = 0; k < table->num_entries; k++) {
drivers/gpu/drm/radeon/ni_dpm.c
2755
for (k = 0; k < table->num_entries; k++)
drivers/gpu/drm/radeon/ni_dpm.c
2828
for (j = 1; j < table->num_entries; j++) {
drivers/gpu/drm/radeon/ni_dpm.c
2855
if (table->num_entries > MAX_AC_TIMING_ENTRIES)
drivers/gpu/drm/radeon/ni_dpm.c
2862
for (i = 0; i < table->num_entries; i++) {
drivers/gpu/drm/radeon/ni_dpm.c
2869
ni_table->num_entries = table->num_entries;
drivers/gpu/drm/radeon/ni_dpm.c
2948
u32 num_entries, u32 valid_flag)
drivers/gpu/drm/radeon/ni_dpm.c
2952
for (i = 0, j = 0; j < num_entries; j++) {
drivers/gpu/drm/radeon/ni_dpm.c
2967
for (i = 0; i < ni_pi->mc_reg_table.num_entries; i++) {
drivers/gpu/drm/radeon/ni_dpm.c
2972
if ((i == ni_pi->mc_reg_table.num_entries) && (i > 0))
drivers/gpu/drm/radeon/ni_dpm.h
54
u8 num_entries;
drivers/gpu/drm/radeon/radeon_atombios.c
3936
mclk_range_table->num_entries = (u8)
drivers/gpu/drm/radeon/radeon_atombios.c
3940
for (i = 0; i < mclk_range_table->num_entries; i++) {
drivers/gpu/drm/radeon/radeon_atombios.c
3978
u8 frev, crev, num_entries, t_mem_id, num_ranges = 0;
drivers/gpu/drm/radeon/radeon_atombios.c
4005
num_entries = (u8)((le16_to_cpu(reg_block->usRegIndexTblSize)) /
drivers/gpu/drm/radeon/radeon_atombios.c
4007
if (num_entries > VBIOS_MC_REGISTER_ARRAY_SIZE)
drivers/gpu/drm/radeon/radeon_atombios.c
4009
while (i < num_entries) {
drivers/gpu/drm/radeon/radeon_atombios.c
4046
reg_table->num_entries = num_ranges;
drivers/gpu/drm/radeon/radeon_mode.h
611
u8 num_entries;
drivers/gpu/drm/radeon/radeon_mode.h
631
u8 num_entries;
drivers/gpu/drm/radeon/si_dpm.c
5314
for (k = 0; k < table->num_entries; k++)
drivers/gpu/drm/radeon/si_dpm.c
5325
for (k = 0; k < table->num_entries; k++) {
drivers/gpu/drm/radeon/si_dpm.c
5339
for (k = 0; k < table->num_entries; k++)
drivers/gpu/drm/radeon/si_dpm.c
5351
for(k = 0; k < table->num_entries; k++)
drivers/gpu/drm/radeon/si_dpm.c
5429
for (j = 1; j < table->num_entries; j++) {
drivers/gpu/drm/radeon/si_dpm.c
5456
if (table->num_entries > MAX_AC_TIMING_ENTRIES)
drivers/gpu/drm/radeon/si_dpm.c
5463
for (i = 0; i < table->num_entries; i++) {
drivers/gpu/drm/radeon/si_dpm.c
5471
si_table->num_entries = table->num_entries;
drivers/gpu/drm/radeon/si_dpm.c
5548
u32 num_entries, u32 valid_flag)
drivers/gpu/drm/radeon/si_dpm.c
5552
for(i = 0, j = 0; j < num_entries; j++) {
drivers/gpu/drm/radeon/si_dpm.c
5567
for (i = 0; i < si_pi->mc_reg_table.num_entries; i++) {
drivers/gpu/drm/radeon/si_dpm.c
5572
if ((i == si_pi->mc_reg_table.num_entries) && (i > 0))
drivers/gpu/drm/radeon/si_dpm.h
109
u8 num_entries;
drivers/gpu/drm/radeon/sumo_dpm.c
1533
for (i = 0; i < vid_mapping_table->num_entries; i++) {
drivers/gpu/drm/radeon/sumo_dpm.c
1538
return vid_mapping_table->entries[vid_mapping_table->num_entries - 1].vid_7bit;
drivers/gpu/drm/radeon/sumo_dpm.c
1548
for (i = 0; i < vid_mapping_table->num_entries; i++) {
drivers/gpu/drm/radeon/sumo_dpm.c
1553
return vid_mapping_table->entries[vid_mapping_table->num_entries - 1].vid_2bit;
drivers/gpu/drm/radeon/sumo_dpm.c
1646
vid_mapping_table->num_entries = i;
drivers/gpu/drm/radeon/sumo_dpm.h
66
u32 num_entries;
drivers/gpu/drm/vmwgfx/vmwgfx_cotable.c
605
u32 num_entries;
drivers/gpu/drm/vmwgfx/vmwgfx_cotable.c
621
num_entries = PAGE_SIZE / co_info[type].size;
drivers/gpu/drm/vmwgfx/vmwgfx_cotable.c
622
if (num_entries < co_info[type].min_initial_entries) {
drivers/gpu/drm/xe/xe_guc_capture.c
504
u16 num_entries)
drivers/gpu/drm/xe/xe_guc_capture.c
520
for (list_idx = 0; ptr_idx < num_entries && list_idx < list_num; ++list_idx, ++ptr_idx) {
drivers/gpu/drm/xe/xe_guc_capture.c
530
ptr_idx < num_entries && list_idx < match->num_regs;
drivers/gpu/drm/xe/xe_guc_capture.c
538
if (ptr_idx < num_entries)
drivers/gpu/drm/xe/xe_guc_capture.c
540
ptr_idx, num_entries);
drivers/gpu/drm/xe/xe_lmtt.c
126
unsigned int num_entries = lmtt->ops->lmtt_pte_num(pd->level);
drivers/gpu/drm/xe/xe_lmtt.c
130
for (n = 0; n < num_entries; n++)
drivers/gpu/drm/xe/xe_lmtt.c
326
unsigned int num_entries = pd->level ? lmtt->ops->lmtt_pte_num(pd->level) : 0;
drivers/gpu/drm/xe/xe_lmtt.c
330
for (i = 0; i < num_entries; i++) {
drivers/gpu/drm/xe/xe_lmtt.c
62
unsigned int num_entries = level ? lmtt->ops->lmtt_pte_num(level) : 0;
drivers/gpu/drm/xe/xe_lmtt.c
67
pt = kzalloc_flex(*pt, entries, num_entries);
drivers/gpu/drm/xe/xe_migrate.c
1736
for (j = 0; j < pt_op->num_entries; j++) {
drivers/gpu/drm/xe/xe_migrate.c
1782
num_updates += pt_op->num_entries;
drivers/gpu/drm/xe/xe_migrate.c
1783
for (j = 0; j < pt_op->num_entries; ++j) {
drivers/gpu/drm/xe/xe_migrate.c
1851
for (; j < pt_op->num_entries; ++j, ++current_update, ++idx) {
drivers/gpu/drm/xe/xe_migrate.c
1888
for (j = 0; j < pt_op->num_entries; ++j) {
drivers/gpu/drm/xe/xe_migrate.c
1906
for (j = 0; j < pt_op->num_entries; ++j)
drivers/gpu/drm/xe/xe_migrate.c
192
u32 num_entries = NUM_PT_SLOTS, num_level = vm->pt_root[id]->level;
drivers/gpu/drm/xe/xe_migrate.c
211
num_entries * XE_PAGE_SIZE,
drivers/gpu/drm/xe/xe_migrate.c
223
map_ofs = (num_entries - num_setup) * XE_PAGE_SIZE;
drivers/gpu/drm/xe/xe_migrate.c
226
for (i = 0, level = 0; i < num_entries; level++) {
drivers/gpu/drm/xe/xe_page_reclaim.c
109
prl->num_entries = 0;
drivers/gpu/drm/xe/xe_page_reclaim.c
128
prl->num_entries = 0;
drivers/gpu/drm/xe/xe_page_reclaim.c
66
int prl_size = min(prl->num_entries + 1, XE_PAGE_RECLAIM_MAX_ENTRIES) *
drivers/gpu/drm/xe/xe_page_reclaim.c
97
prl->num_entries = XE_PAGE_RECLAIM_INVALID_LIST;
drivers/gpu/drm/xe/xe_page_reclaim.h
46
int num_entries;
drivers/gpu/drm/xe/xe_page_reclaim.h
58
return !prl->entries && prl->num_entries == 0;
drivers/gpu/drm/xe/xe_page_reclaim.h
71
prl->num_entries != XE_PAGE_RECLAIM_INVALID_LIST;
drivers/gpu/drm/xe/xe_pt.c
1003
u32 num_entries)
drivers/gpu/drm/xe/xe_pt.c
1007
for (i = 0; i < num_entries; i++) {
drivers/gpu/drm/xe/xe_pt.c
1059
u32 num_entries, struct llist_head *deferred)
drivers/gpu/drm/xe/xe_pt.c
1065
for (i = 0; i < num_entries; i++) {
drivers/gpu/drm/xe/xe_pt.c
1086
u32 num_entries, bool rebind)
drivers/gpu/drm/xe/xe_pt.c
1092
for (i = num_entries - 1; i >= 0; --i) {
drivers/gpu/drm/xe/xe_pt.c
1116
u32 num_entries, bool rebind)
drivers/gpu/drm/xe/xe_pt.c
1122
for (i = 0; i < num_entries; i++) {
drivers/gpu/drm/xe/xe_pt.c
1148
u32 num_entries)
drivers/gpu/drm/xe/xe_pt.c
1152
for (i = 0; i < num_entries; i++)
drivers/gpu/drm/xe/xe_pt.c
1160
u32 *num_entries, bool invalidate_on_bind)
drivers/gpu/drm/xe/xe_pt.c
1164
*num_entries = 0;
drivers/gpu/drm/xe/xe_pt.c
1165
err = xe_pt_stage_bind(tile, vma, range, entries, num_entries,
drivers/gpu/drm/xe/xe_pt.c
1168
xe_tile_assert(tile, *num_entries);
drivers/gpu/drm/xe/xe_pt.c
1175
unsigned int num_entries, bool bind)
drivers/gpu/drm/xe/xe_pt.c
1181
num_entries);
drivers/gpu/drm/xe/xe_pt.c
1182
for (i = 0; i < num_entries; i++) {
drivers/gpu/drm/xe/xe_pt.c
1593
int num_entries = prl->num_entries;
drivers/gpu/drm/xe/xe_pt.c
1598
xe_tile_assert(tile, num_entries < XE_PAGE_RECLAIM_MAX_ENTRIES - 1);
drivers/gpu/drm/xe/xe_pt.c
1628
reclaim_entries[num_entries].qw =
drivers/gpu/drm/xe/xe_pt.c
1633
prl->num_entries++;
drivers/gpu/drm/xe/xe_pt.c
1636
xe_child->level, pte, reclamation_size, num_entries);
drivers/gpu/drm/xe/xe_pt.c
1706
if (xe_walk->prl->num_entries < XE_PAGE_RECLAIM_MAX_ENTRIES - 1) {
drivers/gpu/drm/xe/xe_pt.c
1856
u32 num_entries)
drivers/gpu/drm/xe/xe_pt.c
1862
for (i = num_entries - 1; i >= 0; --i) {
drivers/gpu/drm/xe/xe_pt.c
1882
u32 num_entries)
drivers/gpu/drm/xe/xe_pt.c
1888
for (i = 0; i < num_entries; ++i) {
drivers/gpu/drm/xe/xe_pt.c
1915
for (i = 0; i < pt_op->num_entries; i++) {
drivers/gpu/drm/xe/xe_pt.c
1967
&pt_op->num_entries, invalidate_on_bind);
drivers/gpu/drm/xe/xe_pt.c
1969
xe_tile_assert(tile, pt_op->num_entries <=
drivers/gpu/drm/xe/xe_pt.c
1972
pt_op->num_entries, true);
drivers/gpu/drm/xe/xe_pt.c
2002
pt_op->num_entries, pt_op->rebind);
drivers/gpu/drm/xe/xe_pt.c
2004
xe_pt_cancel_bind(vma, pt_op->entries, pt_op->num_entries);
drivers/gpu/drm/xe/xe_pt.c
2029
&pt_op->num_entries, false);
drivers/gpu/drm/xe/xe_pt.c
2031
xe_tile_assert(tile, pt_op->num_entries <=
drivers/gpu/drm/xe/xe_pt.c
2034
pt_op->num_entries, true);
drivers/gpu/drm/xe/xe_pt.c
2044
pt_op->num_entries, pt_op->rebind);
drivers/gpu/drm/xe/xe_pt.c
2046
xe_pt_cancel_bind(vma, pt_op->entries, pt_op->num_entries);
drivers/gpu/drm/xe/xe_pt.c
2090
pt_op->num_entries = xe_pt_stage_unbind(tile, xe_vma_vm(vma),
drivers/gpu/drm/xe/xe_pt.c
2094
pt_op->num_entries, false);
drivers/gpu/drm/xe/xe_pt.c
2101
xe_pt_commit_prepare_unbind(vma, pt_op->entries, pt_op->num_entries);
drivers/gpu/drm/xe/xe_pt.c
2112
XE_WARN_ON(!pt_op->num_entries);
drivers/gpu/drm/xe/xe_pt.c
2120
if (pt_op->num_entries > 1)
drivers/gpu/drm/xe/xe_pt.c
2152
pt_op->num_entries = xe_pt_stage_unbind(tile, vm, NULL, range,
drivers/gpu/drm/xe/xe_pt.c
2156
pt_op->num_entries, false);
drivers/gpu/drm/xe/xe_pt.c
2167
pt_op->num_entries);
drivers/gpu/drm/xe/xe_pt.c
2612
pt_op->num_entries, &pt_update_ops->deferred);
drivers/gpu/drm/xe/xe_pt.c
2710
xe_pt_free_bind(pt_op->entries, pt_op->num_entries);
drivers/gpu/drm/xe/xe_pt.c
2740
pt_op->num_entries,
drivers/gpu/drm/xe/xe_pt.c
2744
pt_op->num_entries);
drivers/gpu/drm/xe/xe_pt.c
700
u32 *num_entries, bool clear_pt)
drivers/gpu/drm/xe/xe_pt.c
783
*num_entries = xe_walk.wupd.num_used_entries;
drivers/gpu/drm/xe/xe_pt_types.h
87
u32 num_entries;
drivers/gpu/drm/xe/xe_ttm_vram_mgr.c
368
int num_entries = 0;
drivers/gpu/drm/xe/xe_ttm_vram_mgr.c
381
num_entries++;
drivers/gpu/drm/xe/xe_ttm_vram_mgr.c
386
r = sg_alloc_table(*sgt, num_entries, GFP_KERNEL);
drivers/gpu/drm/xe/xe_vsec.c
29
.num_entries = 2,
drivers/gpu/drm/xe/xe_vsec.c
39
.num_entries = 2,
drivers/infiniband/core/cache.c
1225
ssize_t num_entries = 0, ret;
drivers/infiniband/core/cache.c
1237
if (num_entries >= max_entries) {
drivers/infiniband/core/cache.c
1255
num_entries++;
drivers/infiniband/core/cache.c
1261
return num_entries;
drivers/infiniband/core/cq.c
83
static int __poll_cq(struct ib_cq *cq, int num_entries, struct ib_wc *wc)
drivers/infiniband/core/cq.c
87
rc = ib_poll_cq(cq, num_entries, wc);
drivers/infiniband/core/cq.c
88
trace_cq_poll(cq, num_entries, rc);
drivers/infiniband/core/uverbs_std_types_device.c
307
size_t num_entries, size_t user_entry_size)
drivers/infiniband/core/uverbs_std_types_device.c
318
entries, sizeof(*entries) * num_entries);
drivers/infiniband/core/uverbs_std_types_device.c
328
for (i = 0; i < num_entries; i++) {
drivers/infiniband/core/uverbs_std_types_device.c
353
ssize_t num_entries;
drivers/infiniband/core/uverbs_std_types_device.c
386
num_entries = rdma_query_gid_table(ib_dev, entries, max_entries);
drivers/infiniband/core/uverbs_std_types_device.c
387
if (num_entries < 0)
drivers/infiniband/core/uverbs_std_types_device.c
390
ret = copy_gid_entries_to_user(attrs, entries, num_entries,
drivers/infiniband/core/uverbs_std_types_device.c
397
&num_entries, sizeof(num_entries));
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3853
int bnxt_re_poll_cq(struct ib_cq *ib_cq, int num_entries, struct ib_wc *wc)
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3875
budget = min_t(u32, num_entries, cq->max_cql);
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3876
num_entries = budget;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3991
return num_entries - budget;
drivers/infiniband/hw/bnxt_re/ib_verbs.h
252
int bnxt_re_poll_cq(struct ib_cq *cq, int num_entries, struct ib_wc *wc);
drivers/infiniband/hw/cxgb4/cq.c
949
int c4iw_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc)
drivers/infiniband/hw/cxgb4/cq.c
959
for (npolled = 0; npolled < num_entries; ++npolled) {
drivers/infiniband/hw/cxgb4/iw_cxgb4.h
990
int c4iw_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc);
drivers/infiniband/hw/erdma/erdma_cq.c
201
int erdma_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc)
drivers/infiniband/hw/erdma/erdma_cq.c
209
for (npolled = 0; npolled < num_entries;) {
drivers/infiniband/hw/erdma/erdma_verbs.h
468
int erdma_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc);
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
3705
static int hns_roce_v2_poll_cq(struct ib_cq *ibcq, int num_entries,
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
3956
int num_entries, struct ib_wc *wc)
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
3965
left = min_t(unsigned int, (unsigned int)num_entries, left);
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
3980
static int hns_roce_v2_sw_poll_cq(struct hns_roce_cq *hr_cq, int num_entries,
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
3988
num_entries - npolled, wc + npolled);
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
3989
if (npolled >= num_entries)
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
3995
num_entries - npolled, wc + npolled);
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
3996
if (npolled >= num_entries)
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
4278
static int hns_roce_v2_poll_cq(struct ib_cq *ibcq, int num_entries,
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
4297
npolled = hns_roce_v2_sw_poll_cq(hr_cq, num_entries, wc);
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
4301
for (npolled = 0; npolled < num_entries; ++npolled) {
drivers/infiniband/hw/irdma/verbs.c
4385
static int __irdma_poll_cq(struct irdma_cq *iwcq, int num_entries, struct ib_wc *entry)
drivers/infiniband/hw/irdma/verbs.c
4404
while (npolled < num_entries) {
drivers/infiniband/hw/irdma/verbs.c
4428
while (npolled < num_entries) {
drivers/infiniband/hw/irdma/verbs.c
4475
static int irdma_poll_cq(struct ib_cq *ibcq, int num_entries,
drivers/infiniband/hw/irdma/verbs.c
4485
ret = __irdma_poll_cq(iwcq, num_entries, entry);
drivers/infiniband/hw/mana/cq.c
322
int mana_ib_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc)
drivers/infiniband/hw/mana/cq.c
333
for (i = 0; i < num_entries; i++) {
drivers/infiniband/hw/mana/cq.c
340
num_polled = mana_process_completions(cq, num_entries, wc);
drivers/infiniband/hw/mana/mana_ib.h
736
int mana_ib_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc);
drivers/infiniband/hw/mana/qp.c
223
resp.num_entries = i;
drivers/infiniband/hw/mlx4/cq.c
612
static void mlx4_ib_qp_sw_comp(struct mlx4_ib_qp *qp, int num_entries,
drivers/infiniband/hw/mlx4/cq.c
625
for (i = 0; i < cur && *npolled < num_entries; i++) {
drivers/infiniband/hw/mlx4/cq.c
636
static void mlx4_ib_poll_sw_comp(struct mlx4_ib_cq *cq, int num_entries,
drivers/infiniband/hw/mlx4/cq.c
646
mlx4_ib_qp_sw_comp(qp, num_entries, wc + *npolled, npolled, 1);
drivers/infiniband/hw/mlx4/cq.c
647
if (*npolled >= num_entries)
drivers/infiniband/hw/mlx4/cq.c
652
mlx4_ib_qp_sw_comp(qp, num_entries, wc + *npolled, npolled, 0);
drivers/infiniband/hw/mlx4/cq.c
653
if (*npolled >= num_entries)
drivers/infiniband/hw/mlx4/cq.c
882
int mlx4_ib_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc)
drivers/infiniband/hw/mlx4/cq.c
892
mlx4_ib_poll_sw_comp(cq, num_entries, wc, &npolled);
drivers/infiniband/hw/mlx4/cq.c
896
for (npolled = 0; npolled < num_entries; ++npolled) {
drivers/infiniband/hw/mlx4/mlx4_ib.h
776
int mlx4_ib_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc);
drivers/infiniband/hw/mlx5/cq.c
397
static void sw_comp(struct mlx5_ib_qp *qp, int num_entries, struct ib_wc *wc,
drivers/infiniband/hw/mlx5/cq.c
412
for (i = 0; i < cur && np < num_entries; i++) {
drivers/infiniband/hw/mlx5/cq.c
430
static void mlx5_ib_poll_sw_comp(struct mlx5_ib_cq *cq, int num_entries,
drivers/infiniband/hw/mlx5/cq.c
438
sw_comp(qp, num_entries, wc + *npolled, npolled, true);
drivers/infiniband/hw/mlx5/cq.c
439
if (*npolled >= num_entries)
drivers/infiniband/hw/mlx5/cq.c
444
sw_comp(qp, num_entries, wc + *npolled, npolled, false);
drivers/infiniband/hw/mlx5/cq.c
445
if (*npolled >= num_entries)
drivers/infiniband/hw/mlx5/cq.c
584
static int poll_soft_wc(struct mlx5_ib_cq *cq, int num_entries,
drivers/infiniband/hw/mlx5/cq.c
592
if (npolled >= num_entries)
drivers/infiniband/hw/mlx5/cq.c
610
int mlx5_ib_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc)
drivers/infiniband/hw/mlx5/cq.c
624
soft_polled = poll_soft_wc(cq, num_entries, wc, true);
drivers/infiniband/hw/mlx5/cq.c
626
mlx5_ib_poll_sw_comp(cq, num_entries - soft_polled,
drivers/infiniband/hw/mlx5/cq.c
632
soft_polled = poll_soft_wc(cq, num_entries, wc, false);
drivers/infiniband/hw/mlx5/cq.c
634
for (npolled = 0; npolled < num_entries - soft_polled; npolled++) {
drivers/infiniband/hw/mlx5/fs.c
720
int num_entries;
drivers/infiniband/hw/mlx5/fs.c
757
num_entries = MLX5_FS_MAX_ENTRIES;
drivers/infiniband/hw/mlx5/fs.c
764
build_leftovers_ft_param(&priority, &num_entries, &num_groups);
drivers/infiniband/hw/mlx5/fs.c
779
num_entries = 1;
drivers/infiniband/hw/mlx5/fs.c
789
max_table_size = min_t(int, num_entries, max_table_size);
drivers/infiniband/hw/mlx5/mlx5_ib.h
1375
int mlx5_ib_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc);
drivers/infiniband/hw/mlx5/wr.c
496
sblock_ctrl->num_entries = cpu_to_be16(2);
drivers/infiniband/hw/mthca/mthca_cq.c
654
int mthca_poll_cq(struct ib_cq *ibcq, int num_entries,
drivers/infiniband/hw/mthca/mthca_cq.c
669
while (npolled < num_entries) {
drivers/infiniband/hw/mthca/mthca_dev.h
483
int mthca_poll_cq(struct ib_cq *ibcq, int num_entries,
drivers/infiniband/hw/ocrdma/ocrdma_hw.c
1614
static int ocrdma_build_q_conf(u32 *num_entries, int entry_size,
drivers/infiniband/hw/ocrdma/ocrdma_hw.c
1620
*num_entries = roundup_pow_of_two(*num_entries);
drivers/infiniband/hw/ocrdma/ocrdma_hw.c
1621
mem_size = *num_entries * entry_size;
drivers/infiniband/hw/ocrdma/ocrdma_hw.c
1634
*num_entries = mem_size / entry_size;
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2754
static int ocrdma_poll_hwcq(struct ocrdma_cq *cq, int num_entries,
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2767
while (num_entries) {
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2798
num_entries -= 1;
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2814
static int ocrdma_add_err_cqe(struct ocrdma_cq *cq, int num_entries,
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2819
while (num_entries) {
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2835
num_entries -= 1;
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2840
int ocrdma_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc)
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2842
int cqes_to_poll = num_entries;
drivers/infiniband/hw/ocrdma/ocrdma_verbs.h
51
int ocrdma_poll_cq(struct ib_cq *, int num_entries, struct ib_wc *wc);
drivers/infiniband/hw/qedr/qedr_roce_cm.c
668
int qedr_gsi_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc)
drivers/infiniband/hw/qedr/qedr_roce_cm.c
679
while (i < num_entries && qp->rq.cons != qp->rq.gsi_cons) {
drivers/infiniband/hw/qedr/qedr_roce_cm.c
707
while (i < num_entries && qp->sq.cons != qp->sq.gsi_cons) {
drivers/infiniband/hw/qedr/qedr_roce_cm.c
723
num_entries, i, qp->rq.cons, qp->rq.gsi_cons, qp->sq.cons,
drivers/infiniband/hw/qedr/qedr_roce_cm.h
48
int qedr_gsi_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc);
drivers/infiniband/hw/qedr/verbs.c
4056
struct qedr_cq *cq, int num_entries,
drivers/infiniband/hw/qedr/verbs.c
4062
while (num_entries && qp->sq.wqe_cons != hw_cons) {
drivers/infiniband/hw/qedr/verbs.c
4098
num_entries--;
drivers/infiniband/hw/qedr/verbs.c
4112
int num_entries, struct ib_wc *wc,
drivers/infiniband/hw/qedr/verbs.c
4119
cnt = process_req(dev, qp, cq, num_entries, wc, req->sq_cons,
drivers/infiniband/hw/qedr/verbs.c
4127
cnt = process_req(dev, qp, cq, num_entries, wc, req->sq_cons,
drivers/infiniband/hw/qedr/verbs.c
4133
cnt = process_req(dev, qp, cq, num_entries, wc,
drivers/infiniband/hw/qedr/verbs.c
4137
if (cnt < num_entries) {
drivers/infiniband/hw/qedr/verbs.c
4337
int num_entries, struct ib_wc *wc, u16 hw_cons)
drivers/infiniband/hw/qedr/verbs.c
4341
while (num_entries && qp->rq.wqe_cons != hw_cons) {
drivers/infiniband/hw/qedr/verbs.c
4350
num_entries--;
drivers/infiniband/hw/qedr/verbs.c
4371
struct qedr_cq *cq, int num_entries,
drivers/infiniband/hw/qedr/verbs.c
4384
struct qedr_cq *cq, int num_entries,
drivers/infiniband/hw/qedr/verbs.c
4391
cnt = process_resp_flush(qp, cq, num_entries, wc,
drivers/infiniband/hw/qedr/verbs.c
4412
int qedr_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc)
drivers/infiniband/hw/qedr/verbs.c
4430
return qedr_gsi_poll_cq(ibcq, num_entries, wc);
drivers/infiniband/hw/qedr/verbs.c
4435
while (num_entries && is_valid_cqe(cq, cqe)) {
drivers/infiniband/hw/qedr/verbs.c
4452
cnt = qedr_poll_cq_req(dev, qp, cq, num_entries, wc,
drivers/infiniband/hw/qedr/verbs.c
4457
cnt = qedr_poll_cq_resp(dev, qp, cq, num_entries, wc,
drivers/infiniband/hw/qedr/verbs.c
4461
cnt = qedr_poll_cq_resp_srq(dev, qp, cq, num_entries,
drivers/infiniband/hw/qedr/verbs.c
4470
num_entries -= cnt;
drivers/infiniband/hw/qedr/verbs.h
90
int qedr_poll_cq(struct ib_cq *, int num_entries, struct ib_wc *wc);
drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c
387
int pvrdma_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc)
drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c
394
if (num_entries < 1 || wc == NULL)
drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c
398
for (npolled = 0; npolled < num_entries; ++npolled) {
drivers/infiniband/hw/vmw_pvrdma/pvrdma_verbs.h
379
int pvrdma_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc);
drivers/infiniband/sw/rdmavt/cq.c
478
int rvt_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *entry)
drivers/infiniband/sw/rdmavt/cq.c
496
for (npolled = 0; npolled < num_entries; ++npolled, ++entry) {
drivers/infiniband/sw/rdmavt/cq.h
17
int rvt_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *entry);
drivers/infiniband/sw/rdmavt/qp.c
167
wss->num_entries = table_bits / BITS_PER_LONG;
drivers/infiniband/sw/rdmavt/qp.c
176
wss->entries = kcalloc_node(wss->num_entries, sizeof(*wss->entries),
drivers/infiniband/sw/rdmavt/qp.c
231
& (wss->num_entries - 1);
drivers/infiniband/sw/rxe/rxe_verbs.c
1165
static int rxe_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc)
drivers/infiniband/sw/rxe/rxe_verbs.c
1173
for (i = 0; i < num_entries; i++) {
drivers/infiniband/sw/siw/siw_verbs.h
65
int siw_poll_cq(struct ib_cq *base_cq, int num_entries, struct ib_wc *wc);
drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c
166
lnum_entries = be16_to_cpu(tbl->num_entries);
drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.c
222
lnum_entries = be16_to_cpu(tbl->num_entries);
drivers/infiniband/ulp/opa_vnic/opa_vnic_encap.h
268
__be16 num_entries;
drivers/infiniband/ulp/opa_vnic/opa_vnic_vema.c
165
u16 offset, num_entries;
drivers/infiniband/ulp/opa_vnic/opa_vnic_vema.c
170
num_entries = be16_to_cpu(mac_tbl->num_entries);
drivers/infiniband/ulp/opa_vnic/opa_vnic_vema.c
172
return ((num_entries <= req_entries) &&
drivers/infiniband/ulp/opa_vnic/opa_vnic_vema.c
173
(offset + num_entries <= OPA_VNIC_MAC_TBL_MAX_ENTRIES));
drivers/infiniband/ulp/opa_vnic/opa_vnic_vema.c
371
mac_tbl_out->num_entries = mac_tbl_in->num_entries;
drivers/iommu/io-pgtable-arm-v7s.c
314
static void __arm_v7s_pte_sync(arm_v7s_iopte *ptep, int num_entries,
drivers/iommu/io-pgtable-arm-v7s.c
321
num_entries * sizeof(*ptep), DMA_TO_DEVICE);
drivers/iommu/io-pgtable-arm-v7s.c
324
int num_entries, struct io_pgtable_cfg *cfg)
drivers/iommu/io-pgtable-arm-v7s.c
328
for (i = 0; i < num_entries; i++)
drivers/iommu/io-pgtable-arm-v7s.c
331
__arm_v7s_pte_sync(ptep, num_entries, cfg);
drivers/iommu/io-pgtable-arm-v7s.c
396
int lvl, int num_entries, arm_v7s_iopte *ptep)
drivers/iommu/io-pgtable-arm-v7s.c
402
for (i = 0; i < num_entries; i++)
drivers/iommu/io-pgtable-arm-v7s.c
422
if (num_entries > 1)
drivers/iommu/io-pgtable-arm-v7s.c
427
__arm_v7s_set_pte(ptep, pte, num_entries, cfg);
drivers/iommu/io-pgtable-arm-v7s.c
466
int num_entries = size >> ARM_V7S_LVL_SHIFT(lvl);
drivers/iommu/io-pgtable-arm-v7s.c
472
if (num_entries)
drivers/iommu/io-pgtable-arm-v7s.c
474
lvl, num_entries, ptep);
drivers/iommu/io-pgtable-arm-v7s.c
564
int idx, i = 0, num_entries = size >> ARM_V7S_LVL_SHIFT(lvl);
drivers/iommu/io-pgtable-arm-v7s.c
576
} while (++i < num_entries);
drivers/iommu/io-pgtable-arm-v7s.c
588
if (num_entries <= 1 && arm_v7s_pte_is_cont(pte[0], lvl)) {
drivers/iommu/io-pgtable-arm-v7s.c
594
if (num_entries) {
drivers/iommu/io-pgtable-arm-v7s.c
597
__arm_v7s_set_pte(ptep, 0, num_entries, &iop->cfg);
drivers/iommu/io-pgtable-arm-v7s.c
599
for (i = 0; i < num_entries; i++) {
drivers/iommu/io-pgtable-arm.c
316
static void __arm_lpae_sync_pte(arm_lpae_iopte *ptep, int num_entries,
drivers/iommu/io-pgtable-arm.c
320
sizeof(*ptep) * num_entries, DMA_TO_DEVICE);
drivers/iommu/io-pgtable-arm.c
323
static void __arm_lpae_clear_pte(arm_lpae_iopte *ptep, struct io_pgtable_cfg *cfg, int num_entries)
drivers/iommu/io-pgtable-arm.c
325
for (int i = 0; i < num_entries; i++)
drivers/iommu/io-pgtable-arm.c
328
if (!cfg->coherent_walk && num_entries)
drivers/iommu/io-pgtable-arm.c
329
__arm_lpae_sync_pte(ptep, num_entries, cfg);
drivers/iommu/io-pgtable-arm.c
339
int lvl, int num_entries, arm_lpae_iopte *ptep)
drivers/iommu/io-pgtable-arm.c
351
for (i = 0; i < num_entries; i++)
drivers/iommu/io-pgtable-arm.c
355
__arm_lpae_sync_pte(ptep, num_entries, cfg);
drivers/iommu/io-pgtable-arm.c
360
arm_lpae_iopte prot, int lvl, int num_entries,
drivers/iommu/io-pgtable-arm.c
365
for (i = 0; i < num_entries; i++)
drivers/iommu/io-pgtable-arm.c
386
__arm_lpae_init_pte(data, paddr, prot, lvl, num_entries, ptep);
drivers/iommu/io-pgtable-arm.c
431
int ret = 0, num_entries, max_entries, map_idx_start;
drivers/iommu/io-pgtable-arm.c
440
num_entries = min_t(int, pgcount, max_entries);
drivers/iommu/io-pgtable-arm.c
441
ret = arm_lpae_init_pte(data, iova, paddr, prot, lvl, num_entries, ptep);
drivers/iommu/io-pgtable-arm.c
443
*mapped += num_entries * size;
drivers/iommu/io-pgtable-arm.c
629
int i = 0, num_entries, max_entries, unmap_idx_start;
drivers/iommu/io-pgtable-arm.c
646
num_entries = min_t(int, pgcount, max_entries);
drivers/iommu/io-pgtable-arm.c
649
for (i = 0; i < num_entries; i++) {
drivers/iommu/io-pgtable-dart.c
114
dart_iopte prot, int num_entries,
drivers/iommu/io-pgtable-dart.c
121
for (i = 0; i < num_entries; i++)
drivers/iommu/io-pgtable-dart.c
134
for (i = 0; i < num_entries; i++)
drivers/iommu/io-pgtable-dart.c
233
int ret = 0, tbl, num_entries, max_entries, map_idx_start;
drivers/iommu/io-pgtable-dart.c
278
num_entries = min_t(int, pgcount, max_entries);
drivers/iommu/io-pgtable-dart.c
280
ret = dart_init_pte(data, iova, paddr, prot, num_entries, ptep);
drivers/iommu/io-pgtable-dart.c
282
*mapped += num_entries * pgsize;
drivers/iommu/io-pgtable-dart.c
299
int i = 0, num_entries, max_entries, unmap_idx_start;
drivers/iommu/io-pgtable-dart.c
315
num_entries = min_t(int, pgcount, max_entries);
drivers/iommu/io-pgtable-dart.c
317
while (i < num_entries) {
drivers/iommu/omap-iommu.c
461
unsigned long offset, int num_entries)
drivers/iommu/omap-iommu.c
463
size_t size = num_entries * sizeof(u32);
drivers/media/i2c/ccs/ccs-core.c
2698
desc->num_entries++;
drivers/media/i2c/ds90ub913.c
405
for (i = 0; i < source_fd.num_entries; i++) {
drivers/media/i2c/ds90ub913.c
410
if (i == source_fd.num_entries) {
drivers/media/i2c/ds90ub913.c
417
fd->entry[fd->num_entries].stream = route->source_stream;
drivers/media/i2c/ds90ub913.c
418
fd->entry[fd->num_entries].flags = source_fd.entry[i].flags;
drivers/media/i2c/ds90ub913.c
419
fd->entry[fd->num_entries].length = source_fd.entry[i].length;
drivers/media/i2c/ds90ub913.c
420
fd->entry[fd->num_entries].pixelcode =
drivers/media/i2c/ds90ub913.c
423
fd->num_entries++;
drivers/media/i2c/ds90ub953.c
455
for (i = 0; i < source_fd.num_entries; i++) {
drivers/media/i2c/ds90ub953.c
469
fd->entry[fd->num_entries].stream = route->source_stream;
drivers/media/i2c/ds90ub953.c
470
fd->entry[fd->num_entries].flags = source_entry->flags;
drivers/media/i2c/ds90ub953.c
471
fd->entry[fd->num_entries].length = source_entry->length;
drivers/media/i2c/ds90ub953.c
472
fd->entry[fd->num_entries].pixelcode = source_entry->pixelcode;
drivers/media/i2c/ds90ub953.c
473
fd->entry[fd->num_entries].bus.csi2.vc =
drivers/media/i2c/ds90ub953.c
475
fd->entry[fd->num_entries].bus.csi2.dt =
drivers/media/i2c/ds90ub953.c
478
fd->num_entries++;
drivers/media/i2c/ds90ub960.c
3498
if (desc.num_entries == 0)
drivers/media/i2c/ds90ub960.c
3503
for (unsigned int i = 1; i < desc.num_entries; i++) {
drivers/media/i2c/ds90ub960.c
3931
for (i = 0; i < source_fd.num_entries; i++) {
drivers/media/i2c/ds90ub960.c
3945
fd->entry[fd->num_entries].stream = route->source_stream;
drivers/media/i2c/ds90ub960.c
3946
fd->entry[fd->num_entries].flags = source_entry->flags;
drivers/media/i2c/ds90ub960.c
3947
fd->entry[fd->num_entries].length = source_entry->length;
drivers/media/i2c/ds90ub960.c
3948
fd->entry[fd->num_entries].pixelcode = source_entry->pixelcode;
drivers/media/i2c/ds90ub960.c
3950
fd->entry[fd->num_entries].bus.csi2.vc = vc_map[nport];
drivers/media/i2c/ds90ub960.c
3953
fd->entry[fd->num_entries].bus.csi2.dt =
drivers/media/i2c/ds90ub960.c
3974
fd->entry[fd->num_entries].bus.csi2.dt =
drivers/media/i2c/ds90ub960.c
3978
fd->num_entries++;
drivers/media/i2c/s5c73m3/s5c73m3-core.c
1168
fd->num_entries = 2;
drivers/media/i2c/s5c73m3/s5c73m3-core.c
1169
for (i = 0; i < fd->num_entries; i++)
drivers/media/i2c/s5c73m3/s5c73m3-core.c
1189
fd->num_entries = 2;
drivers/media/i2c/s5c73m3/s5c73m3-core.c
1192
for (i = 0; i < fd->num_entries; i++)
drivers/media/i2c/vd56g3.c
894
fd->num_entries = 1;
drivers/media/i2c/vgxy61.c
1190
fd->num_entries = 1;
drivers/media/pci/intel/ipu6/ipu6-isys-csi2.c
626
for (i = 0; i < desc.num_entries; i++) {
drivers/media/platform/broadcom/bcm2835-unicam.c
914
for (unsigned int i = 0; i < fd.num_entries; ++i) {
drivers/media/platform/nxp/imx-mipi-csis.c
1200
fd->num_entries = 1;
drivers/media/platform/nxp/imx8-isi/imx8-isi-crossbar.c
55
if (fd.num_entries != 1) {
drivers/media/platform/qcom/venus/hfi_helper.h
1087
u32 num_entries;
drivers/media/platform/qcom/venus/hfi_helper.h
1266
u32 num_entries;
drivers/media/platform/qcom/venus/hfi_parser.c
71
u32 num_entries = mode->num_entries;
drivers/media/platform/qcom/venus/hfi_parser.c
74
if (num_entries > MAX_ALLOC_MODE_ENTRIES)
drivers/media/platform/qcom/venus/hfi_parser.c
79
while (num_entries--) {
drivers/media/platform/raspberrypi/rp1-cfe/cfe.c
834
for (i = 0; i < remote_desc.num_entries; i++) {
drivers/media/platform/raspberrypi/rp1-cfe/cfe.c
839
if (i == remote_desc.num_entries) {
drivers/media/platform/renesas/rzg2l-cru/rzg2l-video.c
428
if (!fd.num_entries) {
drivers/media/platform/renesas/vsp1/vsp1_dl.c
117
unsigned int num_entries;
drivers/media/platform/renesas/vsp1/vsp1_dl.c
256
unsigned int num_entries, size_t extra_size)
drivers/media/platform/renesas/vsp1/vsp1_dl.c
274
dlb_size = num_entries * sizeof(struct vsp1_dl_entry) + extra_size;
drivers/media/platform/renesas/vsp1/vsp1_dl.c
298
dlb->max_entries = num_entries;
drivers/media/platform/renesas/vsp1/vsp1_dl.c
370
dlb->num_entries = 0;
drivers/media/platform/renesas/vsp1/vsp1_dl.c
389
if (WARN_ONCE(dlb->num_entries >= dlb->max_entries,
drivers/media/platform/renesas/vsp1/vsp1_dl.c
393
dlb->entries[dlb->num_entries].addr = reg;
drivers/media/platform/renesas/vsp1/vsp1_dl.c
394
dlb->entries[dlb->num_entries].data = data;
drivers/media/platform/renesas/vsp1/vsp1_dl.c
395
dlb->num_entries++;
drivers/media/platform/renesas/vsp1/vsp1_dl.c
670
dl->body0->num_entries = 0;
drivers/media/platform/renesas/vsp1/vsp1_dl.c
783
hdr->num_bytes = dl->body0->num_entries
drivers/media/platform/renesas/vsp1/vsp1_dl.c
791
hdr->num_bytes = dlb->num_entries
drivers/media/platform/renesas/vsp1/vsp1_dl.h
70
unsigned int num_entries, size_t extra_size);
drivers/media/platform/rockchip/rkisp1/rkisp1-isp.c
113
if (fd.num_entries != 1) {
drivers/media/platform/samsung/exynos4-is/fimc-capture.c
872
if (num_planes != fd.num_entries)
drivers/media/platform/ti/cal/cal-camerarx.c
908
for (i = 0; i < remote_desc.num_entries; i++) {
drivers/media/platform/ti/cal/cal-camerarx.c
913
if (i == remote_desc.num_entries) {
drivers/media/platform/ti/cal/cal-camerarx.c
921
fd->num_entries = 1;
drivers/media/platform/ti/cal/cal.c
465
if (fd.num_entries != 1)
drivers/media/v4l2-core/v4l2-subdev.c
375
for (i = 0; i < fd->num_entries; i++) {
drivers/net/dsa/hirschmann/hellcreek.c
1588
for (i = 1; i <= schedule->num_entries; ++i) {
drivers/net/dsa/hirschmann/hellcreek.c
1592
if (i == schedule->num_entries)
drivers/net/dsa/hirschmann/hellcreek.c
1601
if (i == schedule->num_entries)
drivers/net/dsa/hirschmann/hellcreek.c
1850
for (i = 0; i < schedule->num_entries; ++i)
drivers/net/dsa/ocelot/felix_vsc9959.c
1106
n = taprio->num_entries;
drivers/net/dsa/ocelot/felix_vsc9959.c
1460
if (taprio->num_entries > VSC9959_TAS_GCL_ENTRY_MAX) {
drivers/net/dsa/ocelot/felix_vsc9959.c
1508
QSYS_PARAM_CFG_REG_3_LIST_LENGTH(taprio->num_entries),
drivers/net/dsa/ocelot/felix_vsc9959.c
1513
for (i = 0; i < taprio->num_entries; i++)
drivers/net/dsa/ocelot/felix_vsc9959.c
1579
for (i = 0; i < taprio->num_entries; i++)
drivers/net/dsa/ocelot/felix_vsc9959.c
1747
u32 num_entries;
drivers/net/dsa/ocelot/felix_vsc9959.c
1748
struct action_gate_entry entries[] __counted_by(num_entries);
drivers/net/dsa/ocelot/felix_vsc9959.c
2107
sgi->num_entries = entry->gate.num_entries;
drivers/net/dsa/ocelot/felix_vsc9959.c
2111
entry->gate.num_entries * sizeof(struct action_gate_entry));
drivers/net/dsa/ocelot/felix_vsc9959.c
2147
if (sgi->num_entries > VSC9959_PSFP_GATE_LIST_NUM)
drivers/net/dsa/ocelot/felix_vsc9959.c
2160
ANA_SG_CONFIG_REG_3_LIST_LENGTH(sgi->num_entries) |
drivers/net/dsa/ocelot/felix_vsc9959.c
2168
for (i = 0; i < sgi->num_entries; i++) {
drivers/net/dsa/ocelot/felix_vsc9959.c
2274
size = struct_size(sgi, entries, a->gate.num_entries);
drivers/net/dsa/sja1105/sja1105.h
235
int num_entries;
drivers/net/dsa/sja1105/sja1105_flower.c
427
act->gate.num_entries,
drivers/net/dsa/sja1105/sja1105_tas.c
172
int num_entries = 0;
drivers/net/dsa/sja1105/sja1105_tas.c
213
num_entries += tas_data->offload[port]->num_entries;
drivers/net/dsa/sja1105/sja1105_tas.c
219
num_entries += gating_cfg->num_entries;
drivers/net/dsa/sja1105/sja1105_tas.c
231
table->entries = kcalloc(num_entries, table->ops->unpacked_entry_size,
drivers/net/dsa/sja1105/sja1105_tas.c
235
table->entry_count = num_entries;
drivers/net/dsa/sja1105/sja1105_tas.c
283
schedule_end_idx = k + offload->num_entries - 1;
drivers/net/dsa/sja1105/sja1105_tas.c
310
for (i = 0; i < offload->num_entries; i++, k++) {
drivers/net/dsa/sja1105/sja1105_tas.c
329
schedule_end_idx = k + gating_cfg->num_entries - 1;
drivers/net/dsa/sja1105/sja1105_tas.c
425
i < offload->num_entries;
drivers/net/dsa/sja1105/sja1105_tas.c
431
j < admin->num_entries;
drivers/net/dsa/sja1105/sja1105_tas.c
470
size_t num_entries = gating_cfg->num_entries;
drivers/net/dsa/sja1105/sja1105_tas.c
480
dummy = kzalloc_flex(*dummy, entries, num_entries);
drivers/net/dsa/sja1105/sja1105_tas.c
486
dummy->num_entries = num_entries;
drivers/net/dsa/sja1105/sja1105_tas.c
547
for (i = 0; i < admin->num_entries; i++) {
drivers/net/dsa/sja1105/sja1105_tas.h
37
int num_entries;
drivers/net/dsa/sja1105/sja1105_vl.c
122
gating_cfg->num_entries = 0;
drivers/net/dsa/sja1105/sja1105_vl.c
143
for (i = 0; i < rule->vl.num_entries; i++) {
drivers/net/dsa/sja1105/sja1105_vl.c
46
gating_cfg->num_entries++;
drivers/net/dsa/sja1105/sja1105_vl.c
585
u32 num_entries, struct action_gate_entry *entries)
drivers/net/dsa/sja1105/sja1105_vl.c
637
rule->vl.num_entries = num_entries;
drivers/net/dsa/sja1105/sja1105_vl.c
639
num_entries);
drivers/net/dsa/sja1105/sja1105_vl.c
645
for (i = 0; i < num_entries; i++) {
drivers/net/dsa/sja1105/sja1105_vl.h
24
u32 num_entries, struct action_gate_entry *entries);
drivers/net/dsa/sja1105/sja1105_vl.h
56
u64 cycle_time_ext, u32 num_entries,
drivers/net/ethernet/3com/typhoon.c
329
typhoon_inc_index(u32 *index, const int count, const int num_entries)
drivers/net/ethernet/3com/typhoon.c
336
*index %= num_entries * sizeof(struct cmd_desc);
drivers/net/ethernet/agere/et131x.c
1593
for (entry = 0; entry < fbr->num_entries; entry++) {
drivers/net/ethernet/agere/et131x.c
1603
writel(fbr->num_entries - 1, num_des);
drivers/net/ethernet/agere/et131x.c
1610
writel(((fbr->num_entries * LO_MARK_PERCENT_FOR_RX) / 100) - 1,
drivers/net/ethernet/agere/et131x.c
1895
rx_ring->fbr[0]->num_entries = 512;
drivers/net/ethernet/agere/et131x.c
1897
rx_ring->fbr[1]->num_entries = 512;
drivers/net/ethernet/agere/et131x.c
1900
rx_ring->fbr[0]->num_entries = 1024;
drivers/net/ethernet/agere/et131x.c
1902
rx_ring->fbr[1]->num_entries = 512;
drivers/net/ethernet/agere/et131x.c
1905
rx_ring->fbr[0]->num_entries = 768;
drivers/net/ethernet/agere/et131x.c
1907
rx_ring->fbr[1]->num_entries = 128;
drivers/net/ethernet/agere/et131x.c
1910
rx_ring->psr_entries = rx_ring->fbr[0]->num_entries +
drivers/net/ethernet/agere/et131x.c
1911
rx_ring->fbr[1]->num_entries;
drivers/net/ethernet/agere/et131x.c
1916
bufsize = sizeof(struct fbr_desc) * fbr->num_entries;
drivers/net/ethernet/agere/et131x.c
1933
for (i = 0; i < fbr->num_entries / FBR_CHUNKS; i++) {
drivers/net/ethernet/agere/et131x.c
2032
for (ii = 0; ii < fbr->num_entries / FBR_CHUNKS; ii++) {
drivers/net/ethernet/agere/et131x.c
2045
bufsize = sizeof(struct fbr_desc) * fbr->num_entries;
drivers/net/ethernet/agere/et131x.c
2135
if (buff_index < fbr->num_entries) {
drivers/net/ethernet/agere/et131x.c
2157
fbr->num_entries - 1);
drivers/net/ethernet/agere/et131x.c
2233
if (ring_index > 1 || buff_index > fbr->num_entries - 1) {
drivers/net/ethernet/agere/et131x.c
281
u32 num_entries;
drivers/net/ethernet/broadcom/bnge/bnge_hwrm_lib.c
372
req->num_entries = cpu_to_le32(ctx_pg->entries);
drivers/net/ethernet/broadcom/bnge/bnge_resc.c
369
u16 aux_msix, tx_cp, num_entries;
drivers/net/ethernet/broadcom/bnge/bnge_resc.c
389
num_entries = irqs_demand;
drivers/net/ethernet/broadcom/bnge/bnge_resc.c
391
num_entries = max;
drivers/net/ethernet/broadcom/bnge/bnge_resc.c
392
bd->irq_tbl = kzalloc_objs(*bd->irq_tbl, num_entries);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
8895
__le32 *num_entries;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
8998
for (i = 0, num_entries = &req->tqm_sp_num_entries,
drivers/net/ethernet/broadcom/bnxt/bnxt.c
9005
i++, num_entries++, pg_attr++, pg_dir++, ena <<= 1) {
drivers/net/ethernet/broadcom/bnxt/bnxt.c
9010
*num_entries = cpu_to_le32(ctx_pg->entries);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
9221
req->num_entries = cpu_to_le32(ctx_pg->entries);
drivers/net/ethernet/brocade/bna/bfa_msgq.c
238
if (ntohs(cmd->msg_hdr->num_entries) <=
drivers/net/ethernet/brocade/bna/bfa_msgq.c
462
int num_entries;
drivers/net/ethernet/brocade/bna/bfa_msgq.c
474
num_entries = ntohs(msghdr->num_entries);
drivers/net/ethernet/brocade/bna/bfa_msgq.c
481
BFA_MSGQ_INDX_ADD(rspq->consumer_index, num_entries,
drivers/net/ethernet/brocade/bna/bfa_msgq.c
624
if (ntohs(cmd->msg_hdr->num_entries) <=
drivers/net/ethernet/brocade/bna/bfi.h
390
u16 num_entries;
drivers/net/ethernet/brocade/bna/bna_enet.c
1164
pause_req->mh.num_entries = htons(
drivers/net/ethernet/brocade/bna/bna_enet.c
1626
attr_req->mh.num_entries = htons(
drivers/net/ethernet/brocade/bna/bna_enet.c
1848
stats_req->mh.num_entries = htons(
drivers/net/ethernet/brocade/bna/bna_enet.c
412
admin_up_req->mh.num_entries = htons(
drivers/net/ethernet/brocade/bna/bna_enet.c
429
admin_down_req->mh.num_entries = htons(
drivers/net/ethernet/brocade/bna/bna_enet.c
446
lpbk_up_req->mh.num_entries = htons(
drivers/net/ethernet/brocade/bna/bna_enet.c
467
lpbk_down_req->mh.num_entries = htons(
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1622
cfg_req->mh.num_entries = htons(
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1721
req->mh.num_entries = htons(
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
185
req->mh.num_entries = htons(
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
201
req->mh.num_entries = htons(
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
217
req->mh.num_entries = htons(
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
232
req->mh.num_entries = htons(
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
247
req->mh.num_entries = htons(
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
264
req->mh.num_entries = htons(
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
287
req->mh.num_entries = htons(
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
302
req->mh.num_entries = htons(
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3094
cfg_req->mh.num_entries = htons(
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3141
req->mh.num_entries = htons(
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
319
req->mh.num_entries = htons(
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
338
req->mh.num_entries = htons(
drivers/net/ethernet/cadence/macb_main.c
4188
if (conf->num_entries > bp->num_queues) {
drivers/net/ethernet/cadence/macb_main.c
4190
conf->num_entries, bp->num_queues);
drivers/net/ethernet/cadence/macb_main.c
4213
enst_queue = kcalloc(conf->num_entries, sizeof(*enst_queue), GFP_KERNEL);
drivers/net/ethernet/cadence/macb_main.c
4218
for (i = 0; i < conf->num_entries; i++) {
drivers/net/ethernet/cadence/macb_main.c
4295
conf->num_entries, conf->base_time, conf->cycle_time);
drivers/net/ethernet/cadence/macb_main.c
4304
for (i = 0; i < conf->num_entries; i++) {
drivers/net/ethernet/cadence/macb_main.c
4320
conf->num_entries, hweight32(configured_queues));
drivers/net/ethernet/emulex/benet/be_cmds.c
102
int num_entries = ARRAY_SIZE(cmd_priv_map);
drivers/net/ethernet/emulex/benet/be_cmds.c
105
for (i = 0; i < num_entries; i++)
drivers/net/ethernet/engleder/tsnep_selftests.c
381
qopt->num_entries = 7;
drivers/net/ethernet/engleder/tsnep_selftests.c
405
qopt->num_entries = 8;
drivers/net/ethernet/engleder/tsnep_selftests.c
434
qopt->num_entries = 10;
drivers/net/ethernet/engleder/tsnep_selftests.c
468
qopt->num_entries = 2;
drivers/net/ethernet/engleder/tsnep_selftests.c
501
qopt->num_entries = 3;
drivers/net/ethernet/engleder/tsnep_selftests.c
513
qopt->num_entries = 2;
drivers/net/ethernet/engleder/tsnep_selftests.c
527
qopt->num_entries = 4;
drivers/net/ethernet/engleder/tsnep_selftests.c
539
qopt->num_entries = 2;
drivers/net/ethernet/engleder/tsnep_selftests.c
551
qopt->num_entries = 3;
drivers/net/ethernet/engleder/tsnep_selftests.c
567
qopt->num_entries = 4;
drivers/net/ethernet/engleder/tsnep_selftests.c
583
qopt->num_entries = 5;
drivers/net/ethernet/engleder/tsnep_selftests.c
621
qopt->num_entries = 2;
drivers/net/ethernet/engleder/tsnep_selftests.c
631
qopt->num_entries = 2;
drivers/net/ethernet/engleder/tsnep_selftests.c
644
qopt->num_entries = 2;
drivers/net/ethernet/engleder/tsnep_selftests.c
655
qopt->num_entries = 2;
drivers/net/ethernet/engleder/tsnep_selftests.c
670
qopt->num_entries = 3;
drivers/net/ethernet/engleder/tsnep_selftests.c
681
qopt->num_entries = 2;
drivers/net/ethernet/engleder/tsnep_selftests.c
696
qopt->num_entries = 3;
drivers/net/ethernet/engleder/tsnep_selftests.c
706
qopt->num_entries = 2;
drivers/net/ethernet/engleder/tsnep_selftests.c
721
qopt->num_entries = 4;
drivers/net/ethernet/engleder/tsnep_selftests.c
738
qopt->num_entries = 5;
drivers/net/ethernet/engleder/tsnep_tc.c
101
if (i == (qopt->num_entries - 1))
drivers/net/ethernet/engleder/tsnep_tc.c
107
gcl->count = qopt->num_entries;
drivers/net/ethernet/engleder/tsnep_tc.c
18
if (qopt->num_entries > TSNEP_MAX_GCL_NUM)
drivers/net/ethernet/engleder/tsnep_tc.c
21
for (i = 0; i < qopt->num_entries; i++) {
drivers/net/ethernet/engleder/tsnep_tc.c
99
for (i = 0; i < qopt->num_entries; i++) {
drivers/net/ethernet/freescale/dpaa2/dpsw-cmd.h
414
__le16 num_entries;
drivers/net/ethernet/freescale/dpaa2/dpsw.c
968
u64 iova_addr, u32 iova_size, u16 *num_entries)
drivers/net/ethernet/freescale/dpaa2/dpsw.c
990
*num_entries = le16_to_cpu(rsp_params->num_entries);
drivers/net/ethernet/freescale/dpaa2/dpsw.h
517
u64 iova_addr, u32 iova_size, u16 *num_entries);
drivers/net/ethernet/freescale/enetc/enetc_pf.c
300
int num_entries, vf_entries, i;
drivers/net/ethernet/freescale/enetc/enetc_pf.c
305
num_entries = ENETC_PRFSCAPR_GET_NUM_RFS(val);
drivers/net/ethernet/freescale/enetc/enetc_pf.c
306
vf_entries = num_entries / (pf->total_vfs + 1);
drivers/net/ethernet/freescale/enetc/enetc_pf.c
311
num_entries - vf_entries * pf->total_vfs);
drivers/net/ethernet/freescale/enetc/enetc_qos.c
1239
if (entryg->gate.num_entries >= priv->psfp_cap.max_psfp_gatelist) {
drivers/net/ethernet/freescale/enetc/enetc_qos.c
1245
entries_size = struct_size(sgi, entries, entryg->gate.num_entries);
drivers/net/ethernet/freescale/enetc/enetc_qos.c
1257
sgi->num_entries = entryg->gate.num_entries;
drivers/net/ethernet/freescale/enetc/enetc_qos.c
1260
for (i = 0; i < entryg->gate.num_entries; i++) {
drivers/net/ethernet/freescale/enetc/enetc_qos.c
443
u32 num_entries;
drivers/net/ethernet/freescale/enetc/enetc_qos.c
446
struct action_gate_entry entries[] __counted_by(num_entries);
drivers/net/ethernet/freescale/enetc/enetc_qos.c
67
if (admin_conf->num_entries > enetc_get_max_gcl_len(hw))
drivers/net/ethernet/freescale/enetc/enetc_qos.c
763
if (!sgi->num_entries)
drivers/net/ethernet/freescale/enetc/enetc_qos.c
766
if (sgi->num_entries > priv->psfp_cap.max_psfp_gatelist ||
drivers/net/ethernet/freescale/enetc/enetc_qos.c
78
gcl_len = admin_conf->num_entries;
drivers/net/ethernet/freescale/enetc/enetc_qos.c
795
sgcl_config->acl_len = (sgi->num_entries - 1) & 0x3;
drivers/net/ethernet/freescale/enetc/enetc_qos.c
797
data_size = struct_size(sgcl_data, sgcl, sgi->num_entries);
drivers/net/ethernet/freescale/enetc/enetc_qos.c
813
for (i = 0; i < sgi->num_entries; i++) {
drivers/net/ethernet/freescale/ucc_geth.c
287
u8 num_entries,
drivers/net/ethernet/freescale/ucc_geth.c
297
for (i = 0; i < num_entries; i++) {
drivers/net/ethernet/freescale/ucc_geth.c
326
u8 num_entries,
drivers/net/ethernet/freescale/ucc_geth.c
334
for (i = 0; i < num_entries; i++) {
drivers/net/ethernet/freescale/ucc_geth.c
360
u8 num_entries,
drivers/net/ethernet/freescale/ucc_geth.c
369
for (i = 0; i < num_entries; i++) {
drivers/net/ethernet/google/gve/gve.h
132
u32 num_entries;
drivers/net/ethernet/google/gve/gve_adminq.c
1165
u32 num_entries = qpl->num_entries;
drivers/net/ethernet/google/gve/gve_adminq.c
1166
u32 size = num_entries * sizeof(qpl->page_buses[0]);
drivers/net/ethernet/google/gve/gve_adminq.c
1178
for (i = 0; i < num_entries; i++)
drivers/net/ethernet/google/gve/gve_adminq.c
1184
.num_pages = cpu_to_be32(num_entries),
drivers/net/ethernet/google/gve/gve_main.c
1097
qpl->num_entries = 0;
drivers/net/ethernet/google/gve/gve_main.c
1112
qpl->num_entries++;
drivers/net/ethernet/google/gve/gve_main.c
1144
for (i = 0; i < qpl->num_entries; i++)
drivers/net/ethernet/google/gve/gve_main.c
782
priv->num_registered_pages -= qpl->num_entries;
drivers/net/ethernet/google/gve/gve_main.c
795
pages = qpl->num_entries;
drivers/net/ethernet/google/gve/gve_tx.c
41
fifo->base = vmap(fifo->qpl->pages, fifo->qpl->num_entries, VM_MAP,
drivers/net/ethernet/google/gve/gve_tx.c
49
fifo->size = fifo->qpl->num_entries * PAGE_SIZE;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
276
tx->dqo.qpl->num_entries;
drivers/net/ethernet/ibm/ehea/ehea.h
406
int num_entries;
drivers/net/ethernet/ibm/ehea/ehea.h
419
int num_entries;
drivers/net/ethernet/ibm/ehea/ehea_main.c
227
ehea_fw_handles.num_entries = i;
drivers/net/ethernet/ibm/ehea/ehea_main.c
316
ehea_bcmc_regs.num_entries = i;
drivers/net/ethernet/ibm/ehea/ehea_main.c
3293
for (i = 0; i < ehea_fw_handles.num_entries; i++)
drivers/net/ethernet/ibm/ehea/ehea_main.c
3299
for (i = 0; i < ehea_bcmc_regs.num_entries; i++)
drivers/net/ethernet/ibm/ibmvnic.c
2266
int *num_entries, u8 hdr_field)
drivers/net/ethernet/ibm/ibmvnic.c
2272
*num_entries += create_hdr_descs(hdr_field, skb_mac_header(skb),
drivers/net/ethernet/ibm/ibmvnic.c
2437
int num_entries = 1;
drivers/net/ethernet/ibm/ibmvnic.c
2596
tx_buff->num_entries = 1;
drivers/net/ethernet/ibm/ibmvnic.c
2608
build_hdr_descs_arr(skb, indir_arr, &num_entries, *hdrs);
drivers/net/ethernet/ibm/ibmvnic.c
2610
tx_crq.v1.n_crq_elem = num_entries;
drivers/net/ethernet/ibm/ibmvnic.c
2611
tx_buff->num_entries = num_entries;
drivers/net/ethernet/ibm/ibmvnic.c
2613
if (num_entries + ind_bufp->index > cur_max_ind_descs) {
drivers/net/ethernet/ibm/ibmvnic.c
2621
num_entries * sizeof(struct ibmvnic_generic_scrq));
drivers/net/ethernet/ibm/ibmvnic.c
2623
ind_bufp->index += num_entries;
drivers/net/ethernet/ibm/ibmvnic.c
2635
if (atomic_add_return(num_entries, &tx_scrq->used)
drivers/net/ethernet/ibm/ibmvnic.c
4255
int num_entries = 0;
drivers/net/ethernet/ibm/ibmvnic.c
4268
num_entries += txbuff->num_entries;
drivers/net/ethernet/ibm/ibmvnic.c
4292
if (atomic_sub_return(num_entries, &scrq->used) <=
drivers/net/ethernet/ibm/ibmvnic.c
4736
u64 remote_handle, u64 ioba, u64 num_entries)
drivers/net/ethernet/ibm/ibmvnic.c
4746
ioba, num_entries);
drivers/net/ethernet/ibm/ibmvnic.h
853
int num_entries;
drivers/net/ethernet/intel/e1000e/ich8lan.c
2074
u32 num_entries;
drivers/net/ethernet/intel/e1000e/ich8lan.c
2082
num_entries = hw->mac.rar_entry_count;
drivers/net/ethernet/intel/e1000e/ich8lan.c
2086
num_entries = 1;
drivers/net/ethernet/intel/e1000e/ich8lan.c
2090
num_entries = wlock_mac + 1;
drivers/net/ethernet/intel/e1000e/ich8lan.c
2094
return num_entries;
drivers/net/ethernet/intel/i40e/i40e.h
206
u16 num_entries;
drivers/net/ethernet/intel/i40e/i40e_adminq_cmd.h
1094
__le16 num_entries;
drivers/net/ethernet/intel/i40e/i40e_adminq_cmd.h
477
u8 num_entries; /* reserved for command */
drivers/net/ethernet/intel/i40e/i40e_main.c
12139
pf->irq_pile->num_entries = vectors;
drivers/net/ethernet/intel/i40e/i40e_main.c
12804
pf->qp_pile->num_entries = pf->hw.func_caps.num_tx_qp;
drivers/net/ethernet/intel/i40e/i40e_main.c
247
if (pile->list[pile->num_entries - 1] & I40E_PILE_VALID_BIT) {
drivers/net/ethernet/intel/i40e/i40e_main.c
250
pile->num_entries - 1);
drivers/net/ethernet/intel/i40e/i40e_main.c
253
pile->list[pile->num_entries - 1] = id | I40E_PILE_VALID_BIT;
drivers/net/ethernet/intel/i40e/i40e_main.c
254
return pile->num_entries - 1;
drivers/net/ethernet/intel/i40e/i40e_main.c
258
while (i < pile->num_entries) {
drivers/net/ethernet/intel/i40e/i40e_main.c
266
for (j = 0; (j < needed) && ((i+j) < pile->num_entries); j++) {
drivers/net/ethernet/intel/i40e/i40e_main.c
300
if (!pile || index >= pile->num_entries)
drivers/net/ethernet/intel/i40e/i40e_main.c
304
i < pile->num_entries && pile->list[i] == valid_id;
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2748
i < pile->num_entries; i++) {
drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c
2759
for (i = 0; i < pile->num_entries; i++) {
drivers/net/ethernet/intel/ice/ice_adminq_cmd.h
196
__le16 num_entries; /* Number of Resource entries */
drivers/net/ethernet/intel/ice/ice_adminq_cmd.h
2022
u8 num_entries;
drivers/net/ethernet/intel/ice/ice_common.c
2257
cmd->num_entries = cpu_to_le16(1);
drivers/net/ethernet/intel/ice/ice_common.c
4715
cmd->num_entries = num_qgrps;
drivers/net/ethernet/intel/ice/ice_controlq.c
262
ice_cfg_cq_regs(struct ice_hw *hw, struct ice_ctl_q_ring *ring, u16 num_entries)
drivers/net/ethernet/intel/ice/ice_controlq.c
269
wr32(hw, ring->len, (num_entries | ring->len_ena_mask));
drivers/net/ethernet/intel/ice/ice_irq.c
18
pf->irq_tracker.num_entries = max_vectors;
drivers/net/ethernet/intel/ice/ice_irq.c
24
ice_init_virt_irq_tracker(struct ice_pf *pf, u32 base, u32 num_entries)
drivers/net/ethernet/intel/ice/ice_irq.c
254
pf->virt_irq_tracker.num_entries,
drivers/net/ethernet/intel/ice/ice_irq.c
257
if (res >= pf->virt_irq_tracker.num_entries)
drivers/net/ethernet/intel/ice/ice_irq.c
26
pf->virt_irq_tracker.bm = bitmap_zalloc(num_entries, GFP_KERNEL);
drivers/net/ethernet/intel/ice/ice_irq.c
30
pf->virt_irq_tracker.num_entries = num_entries;
drivers/net/ethernet/intel/ice/ice_irq.c
77
struct xa_limit limit = { .max = pf->irq_tracker.num_entries - 1,
drivers/net/ethernet/intel/ice/ice_irq.h
14
u16 num_entries; /* total vectors available */
drivers/net/ethernet/intel/ice/ice_irq.h
20
u32 num_entries;
drivers/net/ethernet/intel/ice/ice_ptp.c
2385
unsigned int num_entries,
drivers/net/ethernet/intel/ice/ice_ptp.c
2398
for (i = 0; i < num_entries; i++) {
drivers/net/ethernet/intel/ice/ice_ptp.c
2474
unsigned int num_entries;
drivers/net/ethernet/intel/ice/ice_ptp.c
2477
err = ice_ptp_read_sdp_ac(&pf->hw, entries, &num_entries);
drivers/net/ethernet/intel/ice/ice_ptp.c
2495
err = ice_ptp_parse_sdp_entries(pf, entries, num_entries, desc);
drivers/net/ethernet/intel/ice/ice_ptp_hw.c
4837
int ice_ptp_read_sdp_ac(struct ice_hw *hw, __le16 *entries, uint *num_entries)
drivers/net/ethernet/intel/ice/ice_ptp_hw.c
4874
*num_entries = le16_to_cpu(data);
drivers/net/ethernet/intel/ice/ice_ptp_hw.c
4878
err = ice_aq_read_nvm(hw, 0, offset, *num_entries * sizeof(data),
drivers/net/ethernet/intel/ice/ice_ptp_hw.h
357
int ice_ptp_read_sdp_ac(struct ice_hw *hw, __le16 *entries, uint *num_entries);
drivers/net/ethernet/intel/ice/ice_sriov.c
383
msix_avail_for_sriov = pf->virt_irq_tracker.num_entries;
drivers/net/ethernet/intel/ice/ice_sriov.c
873
return pf->virt_irq_tracker.num_entries;
drivers/net/ethernet/intel/idpf/idpf_ptp.h
148
u16 num_entries;
drivers/net/ethernet/intel/idpf/idpf_txrx.c
2014
for (u32 i = 0; i < tx_tstamp_caps->num_entries; i++) {
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
4237
u16 num_entries;
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
4250
num_entries = le16_to_cpu(ma_list->num_mac_addr);
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
4252
if (xn->reply_sz < struct_size(ma_list, mac_addr_list, num_entries))
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
4268
for (i = 0; i < num_entries; i++)
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
4368
u32 entries_size, buf_size, num_entries;
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
4370
num_entries = min_t(u32, total_filters,
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
4372
entries_size = sizeof(struct virtchnl2_mac_addr) * num_entries;
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
4373
buf_size = struct_size(ma_list, mac_addr_list, num_entries);
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
4375
if (!ma_list || num_entries != IDPF_NUM_FILTERS_PER_MSG) {
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
4385
ma_list->num_mac_addr = cpu_to_le16(num_entries);
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
4394
k += num_entries;
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
4395
total_filters -= num_entries;
drivers/net/ethernet/intel/idpf/idpf_virtchnl_ptp.c
384
tstamp_caps->num_entries = num_latches;
drivers/net/ethernet/intel/idpf/idpf_virtchnl_ptp.c
394
for (u16 i = 0; i < tstamp_caps->num_entries; i++) {
drivers/net/ethernet/intel/idpf/idpf_virtchnl_ptp.c
463
for (u16 i = 0; i < caps->num_entries; i++) {
drivers/net/ethernet/intel/idpf/idpf_virtchnl_ptp.c
640
tx_tstamp_caps->num_entries);
drivers/net/ethernet/intel/igc/igc_main.c
6376
for (n = 0; n < qopt->num_entries; n++) {
drivers/net/ethernet/intel/igc/igc_main.c
6522
for (n = 0; n < qopt->num_entries; n++) {
drivers/net/ethernet/intel/igc/igc_main.c
6539
n + 1 == qopt->num_entries)
drivers/net/ethernet/intel/libie/fwlog.c
202
struct libie_fwlog_module_entry *entries, u16 num_entries,
drivers/net/ethernet/intel/libie/fwlog.c
211
fw_modules = kzalloc_objs(*fw_modules, num_entries);
drivers/net/ethernet/intel/libie/fwlog.c
215
for (i = 0; i < num_entries; i++) {
drivers/net/ethernet/intel/libie/fwlog.c
229
cmd->ops.cfg.mdl_cnt = cpu_to_le16(num_entries);
drivers/net/ethernet/intel/libie/fwlog.c
237
sizeof(*fw_modules) * num_entries);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c
1488
int entry, num_entries, max_entries;
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c
1500
num_entries = min_t(int, profile->cam_entries, max_entries);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c
1501
for (entry = 0; entry < num_entries; entry++)
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c
1506
num_entries = min_t(int, profile->action_entries, max_entries);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c
1507
for (entry = 0; entry < num_entries; entry++)
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c
1512
num_entries = min_t(int, profile->action_entries, profile->cam_entries);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c
1513
entry_mask = enable_mask(num_entries);
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c
1519
if (num_entries > 64) {
drivers/net/ethernet/marvell/octeontx2/af/rvu_npc.c
1522
enable_mask(num_entries - 64));
drivers/net/ethernet/mellanox/mlx4/main.c
3173
int num_entries = dev->caps.num_ports;
drivers/net/ethernet/mellanox/mlx4/main.c
3176
priv->steer = kzalloc_objs(struct mlx4_steer, num_entries);
drivers/net/ethernet/mellanox/mlx4/main.c
3180
for (i = 0; i < num_entries; i++)
drivers/net/ethernet/mellanox/mlx4/main.c
3193
int num_entries = dev->caps.num_ports;
drivers/net/ethernet/mellanox/mlx4/main.c
3196
for (i = 0; i < num_entries; i++) {
drivers/net/ethernet/mellanox/mlx4/mr.c
433
int num_entries)
drivers/net/ethernet/mellanox/mlx4/mr.c
435
return mlx4_cmd(dev, mailbox->dma, num_entries, 0, MLX4_CMD_WRITE_MTT,
drivers/net/ethernet/mellanox/mlx5/core/diag/en_tc_tracepoint.h
31
f->rule->action.num_entries : 0)
drivers/net/ethernet/mellanox/mlx5/core/diag/en_tc_tracepoint.h
35
f->rule->action.num_entries : 0);
drivers/net/ethernet/mellanox/mlx5/core/diag/en_tc_tracepoint.h
39
f->rule->action.num_entries);
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
499
u32 num_entries, max_num_entries;
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
507
&num_entries) ||
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
508
num_entries > max_num_entries))
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
513
err = mlx5e_create_umr_mkey(mdev, num_entries, rq->mpwqe.page_shift,
drivers/net/ethernet/mellanox/mlx5/core/en_tc.c
4067
i < flow_action->num_entries - 1)) {
drivers/net/ethernet/mellanox/mlx5/core/lib/st.c
30
u16 num_entries;
drivers/net/ethernet/mellanox/mlx5/core/lib/st.c
51
num_entries = pcie_tph_get_st_table_size(pdev);
drivers/net/ethernet/mellanox/mlx5/core/lib/st.c
53
if (num_entries < 2)
drivers/net/ethernet/mellanox/mlx5/core/lib/st.c
74
st->index_limit.max = num_entries - 1;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/bwc.c
15
return min(ctx->send_queue[queue_id].num_entries / 2,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/debug.c
236
send_queue->num_entries,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/debug.c
238
send_queue->num_entries,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/debug.c
313
ctx->send_queue->num_entries,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
1018
if (!queue->num_entries)
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
1033
queue->num_entries = roundup_pow_of_two(queue_size);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
1037
queue->num_entries);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
1043
queue->completed.mask = queue->num_entries - 1;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
13
unsigned int idx = send_sq->head_dep_idx++ & (queue->num_entries - 1);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
40
dep_wqe = &send_sq->dep_wqe[send_sq->tail_dep_idx++ & (queue->num_entries - 1)];
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
703
buf_sz = queue->num_entries * MAX_WQES_PER_RULE;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
704
sq->dep_wqe = kzalloc_objs(*sq->dep_wqe, queue->num_entries);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
716
sq->buf_mask = (queue->num_entries * MAX_WQES_PER_RULE) - 1;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
849
buf_sz = queue->num_entries * MAX_WQES_PER_RULE;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
960
MLX5_SET(cqc, cqc_data, log_cq_size, ilog2(queue->num_entries));
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.h
141
u16 num_entries;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.h
235
return queue->used_entries >= queue->num_entries;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_icm_pool.c
348
u32 i, num_entries;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_icm_pool.c
352
num_entries = mlx5dr_icm_pool_chunk_size_to_entries(hot_chunk->size);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_icm_pool.c
354
hot_chunk->seg, ilog2(num_entries));
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
29
u32 num_entries = mlx5dr_icm_pool_get_chunk_num_of_entries(htbl->chunk);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
37
if (num_entries == 1 || htbl->byte_mask == 0)
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
387
u32 num_entries =
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
390
ste_ctx->set_hit_addr(hw_ste, icm_addr, num_entries);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
50
index = crc32 & (num_entries - 1);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
502
u32 num_entries;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
520
num_entries = mlx5dr_icm_pool_get_chunk_num_of_entries(chunk);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
522
for (i = 0; i < num_entries; i++) {
drivers/net/ethernet/mellanox/mlxsw/spectrum_nve.c
310
WARN_ON(mc_record->num_entries);
drivers/net/ethernet/mellanox/mlxsw/spectrum_nve.c
322
unsigned int num_entries = mc_record->num_entries;
drivers/net/ethernet/mellanox/mlxsw/spectrum_nve.c
326
num_entries < nve->num_max_mc_entries[proto])
drivers/net/ethernet/mellanox/mlxsw/spectrum_nve.c
336
if (mc_record->num_entries != 0)
drivers/net/ethernet/mellanox/mlxsw/spectrum_nve.c
367
unsigned int num_entries = 0;
drivers/net/ethernet/mellanox/mlxsw/spectrum_nve.c
382
next_kvdl_index, mc_record->num_entries);
drivers/net/ethernet/mellanox/mlxsw/spectrum_nve.c
392
num_entries++);
drivers/net/ethernet/mellanox/mlxsw/spectrum_nve.c
395
WARN_ON(num_entries != mc_record->num_entries);
drivers/net/ethernet/mellanox/mlxsw/spectrum_nve.c
448
mc_record->num_entries++;
drivers/net/ethernet/mellanox/mlxsw/spectrum_nve.c
458
if (mc_record->num_entries != 1 ||
drivers/net/ethernet/mellanox/mlxsw/spectrum_nve.c
471
mc_record->num_entries--;
drivers/net/ethernet/mellanox/mlxsw/spectrum_nve.c
483
mc_record->num_entries--;
drivers/net/ethernet/mellanox/mlxsw/spectrum_nve.c
488
if (mc_record->num_entries != 0) {
drivers/net/ethernet/mellanox/mlxsw/spectrum_nve.c
626
if (mc_record->num_entries != 1)
drivers/net/ethernet/mellanox/mlxsw/spectrum_nve.c
65
unsigned int num_entries;
drivers/net/ethernet/mellanox/mlxsw/spectrum_nve.c
697
WARN_ON(mc_record->num_entries);
drivers/net/ethernet/mellanox/mlxsw/spectrum_qdisc.c
2004
if (f->rule->action.num_entries != 1) {
drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c
2481
u8 num_entries;
drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c
2484
num_entries = mlxsw_reg_rauhtd_ipv4_rec_num_entries_get(rauhtd_pl,
drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c
2487
num_entries++;
drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c
2490
for (i = 0; i < num_entries; i++) {
drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c
2526
u8 num_rec, last_rec_index, num_entries;
drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c
2537
num_entries = mlxsw_reg_rauhtd_ipv4_rec_num_entries_get(rauhtd_pl,
drivers/net/ethernet/mellanox/mlxsw/spectrum_router.c
2539
if (++num_entries == MLXSW_REG_RAUHTD_IPV4_ENT_PER_REC)
drivers/net/ethernet/microchip/lan966x/lan966x_taprio.c
231
if (qopt->num_entries > LAN966X_TAPRIO_NUM_GCL)
drivers/net/ethernet/microchip/lan966x/lan966x_taprio.c
239
for (i = 0; i < qopt->num_entries; ++i) {
drivers/net/ethernet/microchip/lan966x/lan966x_taprio.c
335
if (lan966x_taprio_gcl_free_get(port, free_list) < qopt->num_entries)
drivers/net/ethernet/microchip/lan966x/lan966x_taprio.c
351
for (i = 0; i < qopt->num_entries; ++i) {
drivers/net/ethernet/microchip/lan966x/lan966x_taprio.c
357
if (i == qopt->num_entries - 1)
drivers/net/ethernet/microchip/sparx5/sparx5_main.h
674
u32 num_entries; /* PSFPAdminControlListLength */
drivers/net/ethernet/microchip/sparx5/sparx5_psfp.c
156
ANA_AC_SG_CONFIG_REG_3_LIST_LENGTH_SET(sg->num_entries) |
drivers/net/ethernet/microchip/sparx5/sparx5_psfp.c
170
for (i = 0; i < sg->num_entries; i++) {
drivers/net/ethernet/microchip/sparx5/sparx5_tc_flower.c
724
if (act->gate.num_entries >= SPX5_PSFP_GCE_CNT) {
drivers/net/ethernet/microchip/sparx5/sparx5_tc_flower.c
731
sg->num_entries = act->gate.num_entries;
drivers/net/ethernet/microchip/sparx5/sparx5_tc_flower.c
735
for (i = 0; i < sg->num_entries; i++) {
drivers/net/ethernet/microchip/sparx5/sparx5_tc_flower.c
793
sg->num_entries = 1;
drivers/net/ethernet/netronome/nfp/flower/action.c
1264
if (current_act_idx == flow_act->num_entries)
drivers/net/ethernet/netronome/nfp/flower/action.c
269
int num_act = rule->action.num_entries;
drivers/net/ethernet/netronome/nfp/flower/conntrack.c
1494
entry->rule = flow_rule_alloc(flow->rule->action.num_entries);
drivers/net/ethernet/netronome/nfp/flower/conntrack.c
1538
entry->rule->action.num_entries = flow->rule->action.num_entries;
drivers/net/ethernet/netronome/nfp/flower/conntrack.c
722
num_actions += rules[i]->action.num_entries;
drivers/net/ethernet/netronome/nfp/flower/conntrack.c
732
if (rules[num_rules - 1]->action.num_entries != 0)
drivers/net/ethernet/netronome/nfp/flower/conntrack.c
760
for (i = 0; i < rules[j]->action.num_entries; i++) {
drivers/net/ethernet/netronome/nfp/flower/conntrack.c
804
a_rule->action.num_entries = offset;
drivers/net/ethernet/netronome/nfp/flower/qos_conf.c
170
u32 action_num = flow->rule->action.num_entries;
drivers/net/ethernet/netronome/nfp/flower/qos_conf.c
698
u32 action_num = fl_act->action.num_entries;
drivers/net/ethernet/netronome/nfp/nfp_shared_buf.c
100
entry_sz = nfp_cpp_area_size(sb_desc_area) / num_entries;
drivers/net/ethernet/netronome/nfp/nfp_shared_buf.c
102
pf->shared_bufs = kmalloc_objs(pf->shared_bufs[0], num_entries);
drivers/net/ethernet/netronome/nfp/nfp_shared_buf.c
108
for (i = 0; i < num_entries; i++) {
drivers/net/ethernet/netronome/nfp/nfp_shared_buf.c
124
pf->num_shared_bufs = num_entries;
drivers/net/ethernet/netronome/nfp/nfp_shared_buf.c
81
unsigned int i, num_entries, entry_sz;
drivers/net/ethernet/netronome/nfp/nfp_shared_buf.c
92
num_entries = n;
drivers/net/ethernet/netronome/nfp/nfp_shared_buf.c
95
num_entries * sizeof(pf->shared_bufs[0]),
drivers/net/ethernet/qlogic/netxen/netxen_nic.h
426
uint32_t num_entries;
drivers/net/ethernet/qlogic/netxen/netxen_nic_init.c
568
__le32 entries = cpu_to_le32(directory->num_entries);
drivers/net/ethernet/qlogic/netxen/netxen_nic_init.c
598
entries = cpu_to_le32(directory->num_entries);
drivers/net/ethernet/qlogic/netxen/netxen_nic_init.c
694
entries = cpu_to_le32(ptab_descr->num_entries);
drivers/net/ethernet/qlogic/qed/qed_dbg_hsi.h
336
u16 num_entries;
drivers/net/ethernet/qlogic/qed/qed_dcbx.c
336
int num_entries;
drivers/net/ethernet/qlogic/qed/qed_dcbx.c
349
num_entries = QED_MFW_GET_FIELD(p_app->flags, DCBX_APP_NUM_ENTRIES);
drivers/net/ethernet/qlogic/qed/qed_dcbx.c
352
num_entries, dcbx_version);
drivers/net/ethernet/qlogic/qed/qed_debug.c
2698
u32 num_entries;
drivers/net/ethernet/qlogic/qed/qed_debug.c
2704
num_entries = cond_hdr->data_size / MEM_DUMP_ENTRY_SIZE_DWORDS;
drivers/net/ethernet/qlogic/qed/qed_debug.c
2722
for (i = 0; i < num_entries;
drivers/net/ethernet/qlogic/qed/qed_debug.c
3140
u32 rss_addr, num_entries, total_dwords;
drivers/net/ethernet/qlogic/qed/qed_debug.c
3147
num_entries = rss_defs->num_entries[dev_data->chip_id];
drivers/net/ethernet/qlogic/qed/qed_debug.c
3148
total_dwords = (num_entries * rss_defs->entry_width) / 32;
drivers/net/ethernet/qlogic/qed/qed_debug.c
341
u32 num_entries[MAX_CHIP_IDS];
drivers/net/ethernet/qlogic/qed/qed_debug.c
3771
reg->num_entries > 1 || reg->start_entry > 0 ? 1 : 0);
drivers/net/ethernet/qlogic/qed/qed_debug.c
3897
if (cond_regs[reg_id].num_entries > num_reg_entries)
drivers/net/ethernet/qlogic/qed/qed_debug.c
3898
num_reg_entries = cond_regs[reg_id].num_entries;
drivers/net/ethernet/qlogic/qed/qed_debug.c
3942
if (reg->num_entries > 1 ||
drivers/net/ethernet/qlogic/qlcnic/qlcnic.h
226
__le32 num_entries;
drivers/net/ethernet/qlogic/qlcnic/qlcnic.h
411
u32 num_entries;
drivers/net/ethernet/qlogic/qlcnic/qlcnic.h
428
u32 num_entries;
drivers/net/ethernet/qlogic/qlcnic/qlcnic.h
459
u32 num_entries;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_init.c
748
entries = le32_to_cpu(directory->num_entries);
drivers/net/ethernet/qlogic/qlcnic/qlcnic_init.c
777
entries = le32_to_cpu(directory->num_entries);
drivers/net/ethernet/qlogic/qlcnic/qlcnic_init.c
870
entries = le32_to_cpu(ptab_descr->num_entries);
drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c
1333
no_entries = fw_dump->num_entries;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c
292
fw_dump->num_entries = hdr->num_entries;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_minidump.c
346
fw_dump->num_entries = hdr->num_entries;
drivers/net/ethernet/sfc/ef10.c
3795
size_t num_entries = 0;
drivers/net/ethernet/sfc/ef10.c
3826
num_entries++) = entry;
drivers/net/ethernet/sfc/ef10.c
3838
EFX_WORD_1, num_entries);
drivers/net/ethernet/sfc/ef10.c
3842
inlen = MC_CMD_SET_TUNNEL_ENCAP_UDP_PORTS_IN_LEN(num_entries);
drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c
1390
rule->action.num_entries = 1;
drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c
1519
rule->action.num_entries = 1;
drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c
1004
size = qopt->num_entries;
drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c
984
if (qopt->num_entries > dep)
drivers/net/ethernet/sun/niu.c
3030
static u64 hash_addr_regval(unsigned long index, unsigned long num_entries)
drivers/net/ethernet/sun/niu.c
3032
return (u64)index | (num_entries == 1 ? HASH_TBL_ADDR_AUTOINC : 0);
drivers/net/ethernet/sun/niu.c
3037
unsigned long index, unsigned long num_entries,
drivers/net/ethernet/sun/niu.c
3040
u64 val = hash_addr_regval(index, num_entries);
drivers/net/ethernet/sun/niu.c
3044
index + num_entries > FCRAM_SIZE)
drivers/net/ethernet/sun/niu.c
3048
for (i = 0; i < num_entries; i++)
drivers/net/ethernet/sun/niu.c
3056
unsigned long index, unsigned long num_entries,
drivers/net/ethernet/sun/niu.c
3059
u64 val = hash_addr_regval(index, num_entries);
drivers/net/ethernet/sun/niu.c
3063
index + (num_entries * 8) > FCRAM_SIZE)
drivers/net/ethernet/sun/niu.c
3067
for (i = 0; i < num_entries; i++)
drivers/net/ethernet/sun/sunvnet_common.c
1698
(dr->entry_size * dr->num_entries),
drivers/net/ethernet/sun/sunvnet_common.c
1702
dr->num_entries = 0;
drivers/net/ethernet/sun/sunvnet_common.c
1746
dr->num_entries = VNET_TX_RING_SIZE;
drivers/net/ethernet/ti/am65-cpsw-qos.c
679
for (i = 0; i < taprio->num_entries; i++) {
drivers/net/ethernet/ti/am65-cpsw-qos.c
734
for (i = 0; i < est_new->taprio.num_entries; i++) {
drivers/net/ethernet/ti/am65-cpsw-qos.c
747
if (!fetch_cnt && i < est_new->taprio.num_entries - 1) {
drivers/net/ethernet/ti/am65-cpsw-qos.c
852
for (i = 0; i < from->num_entries; i++)
drivers/net/ethernet/ti/am65-cpsw-qos.c
888
struct_size(est_new, taprio.entries, taprio->num_entries),
drivers/net/wireless/ath/ath11k/ce.c
563
params.num_entries = ce_ring->nentries;
drivers/net/wireless/ath/ath11k/dbring.c
238
int ring_num, int num_entries)
drivers/net/wireless/ath/ath11k/dbring.c
243
ring_num, ar->pdev_idx, num_entries);
drivers/net/wireless/ath/ath11k/dbring.h
77
int ring_num, int num_entries);
drivers/net/wireless/ath/ath11k/dp.c
226
int mac_id, int num_entries)
drivers/net/wireless/ath/ath11k/dp.c
237
if (num_entries > max_entries)
drivers/net/wireless/ath/ath11k/dp.c
238
num_entries = max_entries;
drivers/net/wireless/ath/ath11k/dp.c
240
ring->size = (num_entries * entry_sz) + HAL_RING_BASE_ALIGN - 1;
drivers/net/wireless/ath/ath11k/dp.c
273
params.num_entries = num_entries;
drivers/net/wireless/ath/ath11k/dp.c
285
params.low_threshold = num_entries >> 3;
drivers/net/wireless/ath/ath11k/dp.h
1675
int mac_id, int num_entries);
drivers/net/wireless/ath/ath11k/dp_rx.c
484
int num_entries;
drivers/net/wireless/ath/ath11k/dp_rx.c
486
num_entries = rx_ring->refill_buf_ring.size /
drivers/net/wireless/ath/ath11k/dp_rx.c
489
rx_ring->bufs_max = num_entries;
drivers/net/wireless/ath/ath11k/dp_rx.c
490
ath11k_dp_rxbufs_replenish(ar->ab, dp->mac_id, rx_ring, num_entries,
drivers/net/wireless/ath/ath11k/dp_tx.c
939
params.num_entries * ring_entry_sz);
drivers/net/wireless/ath/ath11k/hal.c
1021
srng->num_entries = params->num_entries;
drivers/net/wireless/ath/ath11k/hal.c
1022
srng->ring_size = srng->entry_size * srng->num_entries;
drivers/net/wireless/ath/ath11k/hal.c
1039
(srng->entry_size * srng->num_entries) << 2);
drivers/net/wireless/ath/ath11k/hal.c
301
(srng->entry_size * srng->num_entries));
drivers/net/wireless/ath/ath11k/hal.c
381
(srng->entry_size * srng->num_entries));
drivers/net/wireless/ath/ath11k/hal.c
393
(srng->entry_size * srng->num_entries));
drivers/net/wireless/ath/ath11k/hal.c
516
params->num_entries = srng->num_entries;
drivers/net/wireless/ath/ath11k/hal.h
501
int num_entries;
drivers/net/wireless/ath/ath11k/hal.h
549
u32 num_entries;
drivers/net/wireless/ath/ath11k/hal_rx.c
800
for (i = 0; i < params.num_entries; i++) {
drivers/net/wireless/ath/ath11k/hal_tx.c
153
for (i = 0; i < params.num_entries; i++) {
drivers/net/wireless/ath/ath12k/ce.c
286
params.num_entries = ce_ring->nentries;
drivers/net/wireless/ath/ath12k/dbring.c
188
int ring_num, int num_entries)
drivers/net/wireless/ath/ath12k/dbring.c
193
ring_num, ar->pdev_idx, num_entries);
drivers/net/wireless/ath/ath12k/dbring.h
71
int ring_num, int num_entries);
drivers/net/wireless/ath/ath12k/dp.c
237
int mac_id, int num_entries)
drivers/net/wireless/ath/ath12k/dp.c
247
if (num_entries > max_entries)
drivers/net/wireless/ath/ath12k/dp.c
248
num_entries = max_entries;
drivers/net/wireless/ath/ath12k/dp.c
250
ring->size = (num_entries * entry_sz) + HAL_RING_BASE_ALIGN - 1;
drivers/net/wireless/ath/ath12k/dp.c
263
params.num_entries = num_entries;
drivers/net/wireless/ath/ath12k/dp.c
274
params.low_threshold = num_entries >> 3;
drivers/net/wireless/ath/ath12k/dp.c
280
params.low_threshold = num_entries >> 3;
drivers/net/wireless/ath/ath12k/dp.h
692
int mac_id, int num_entries);
drivers/net/wireless/ath/ath12k/dp_htt.c
790
cmd->info1 |= le32_encode_bits(params.num_entries * ring_entry_sz,
drivers/net/wireless/ath/ath12k/dp_rx.c
215
int num_entries;
drivers/net/wireless/ath/ath12k/dp_rx.c
217
num_entries = rx_ring->refill_buf_ring.size /
drivers/net/wireless/ath/ath12k/dp_rx.c
220
rx_ring->bufs_max = num_entries;
drivers/net/wireless/ath/ath12k/dp_rx.c
224
num_entries);
drivers/net/wireless/ath/ath12k/dp_rx.c
226
ath12k_dp_mon_buf_replenish(ab, rx_ring, num_entries);
drivers/net/wireless/ath/ath12k/hal.c
245
params->num_entries = srng->num_entries;
drivers/net/wireless/ath/ath12k/hal.c
597
srng->num_entries = params->num_entries;
drivers/net/wireless/ath/ath12k/hal.c
598
srng->ring_size = srng->entry_size * srng->num_entries;
drivers/net/wireless/ath/ath12k/hal.c
617
(srng->entry_size * srng->num_entries) << 2);
drivers/net/wireless/ath/ath12k/hal.h
483
int num_entries;
drivers/net/wireless/ath/ath12k/hal.h
863
u32 num_entries;
drivers/net/wireless/ath/ath12k/wifi7/hal.c
169
u32_encode_bits((srng->entry_size * srng->num_entries),
drivers/net/wireless/ath/ath12k/wifi7/hal.c
251
u32_encode_bits((srng->entry_size * srng->num_entries),
drivers/net/wireless/ath/ath12k/wifi7/hal_rx.c
884
for (i = 0; i < params.num_entries; i++) {
drivers/net/wireless/ath/ath12k/wifi7/hal_rx.c
909
for (i = 0; i < params.num_entries; i++) {
drivers/net/wireless/ath/ath6kl/debug.c
1132
u16 num_entries;
drivers/net/wireless/ath/ath6kl/debug.c
1138
num_entries = le16_to_cpu(tbl->num_entries);
drivers/net/wireless/ath/ath6kl/debug.c
1139
if (struct_size(tbl, info, num_entries) > len)
drivers/net/wireless/ath/ath6kl/debug.c
1168
u16 num_entries, i;
drivers/net/wireless/ath/ath6kl/debug.c
1195
num_entries = le16_to_cpu(tbl->num_entries);
drivers/net/wireless/ath/ath6kl/debug.c
1197
buf_len = 100 + num_entries * 100;
drivers/net/wireless/ath/ath6kl/debug.c
1207
for (i = 0; i < num_entries; i++) {
drivers/net/wireless/ath/ath6kl/htc_pipe.c
655
int num_entries,
drivers/net/wireless/ath/ath6kl/htc_pipe.c
664
for (i = 0; i < num_entries; i++, rpt++) {
drivers/net/wireless/ath/ath6kl/wmi.h
1921
__le16 num_entries;
drivers/net/wireless/intel/iwlwifi/mld/tests/agg.c
243
.num_entries = 1,
drivers/net/wireless/intel/iwlwifi/mld/tests/agg.c
269
.num_entries = 1,
drivers/net/wireless/intel/iwlwifi/mld/tests/agg.c
296
.num_entries = 1,
drivers/net/wireless/intel/iwlwifi/mld/tests/agg.c
322
.num_entries = 3,
drivers/net/wireless/intel/iwlwifi/mld/tests/agg.c
350
.num_entries = 3,
drivers/net/wireless/intel/iwlwifi/mld/tests/agg.c
380
.num_entries = 1,
drivers/net/wireless/intel/iwlwifi/mld/tests/agg.c
411
.num_entries = 1,
drivers/net/wireless/intel/iwlwifi/mld/tests/agg.c
42
u16 num_entries;
drivers/net/wireless/intel/iwlwifi/mld/tests/agg.c
547
for (int i = 0; i < param->reorder_buf_state.num_entries; i++) {
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
725
size_t num_entries = trans->mac_cfg->gen2 ?
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
737
tfd_sz = trans_pcie->txqs.tfd.size * num_entries;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
775
for (i = 0; i < num_entries; i++) {
drivers/net/wireless/marvell/mwifiex/cfg80211.c
3448
int ret = 0, num_entries = 1;
drivers/net/wireless/marvell/mwifiex/cfg80211.c
3453
num_entries++;
drivers/net/wireless/marvell/mwifiex/cfg80211.c
3455
mef_entry = kzalloc_objs(*mef_entry, num_entries);
drivers/net/wireless/marvell/mwifiex/cfg80211.c
3462
mef_cfg.num_entries = num_entries;
drivers/net/wireless/marvell/mwifiex/fw.h
2153
__le16 num_entries;
drivers/net/wireless/marvell/mwifiex/ioctl.h
417
u16 num_entries;
drivers/net/wireless/marvell/mwifiex/sta_cmd.c
1416
mef_cfg->num_entries = cpu_to_le16(mef->num_entries);
drivers/net/wireless/marvell/mwifiex/sta_cmd.c
1419
for (i = 0; i < mef->num_entries; i++) {
drivers/net/wireless/quantenna/qtnfmac/qlink.h
809
__le32 num_entries;
drivers/net/wireless/quantenna/qtnfmac/qlink_util.c
183
qacl->num_entries = cpu_to_le32(acl->n_acl_entries);
drivers/net/wwan/iosm/iosm_ipc_coredump.c
101
for (i = 0; i < num_entries; i++) {
drivers/net/wwan/iosm/iosm_ipc_coredump.c
66
u32 byte_read, num_entries, file_size;
drivers/net/wwan/iosm/iosm_ipc_coredump.c
95
num_entries = le32_to_cpu(cd_table->list.num_entries);
drivers/net/wwan/iosm/iosm_ipc_coredump.c
96
if (num_entries == 0 || num_entries > IOSM_NOF_CD_REGION) {
drivers/net/wwan/iosm/iosm_ipc_coredump.h
37
__le32 num_entries;
drivers/parisc/iosapic.c
222
static struct irt_entry *iosapic_alloc_irt(int num_entries)
drivers/parisc/iosapic.c
224
return kzalloc_objs(struct irt_entry, num_entries);
drivers/parisc/iosapic.c
259
unsigned long num_entries = 0UL;
drivers/parisc/iosapic.c
265
status = pdc_pat_get_irt_size(&num_entries, cell_num);
drivers/parisc/iosapic.c
267
DBG("returned %ld, entries: %lu\n", status, num_entries);
drivers/parisc/iosapic.c
270
BUG_ON(num_entries == 0);
drivers/parisc/iosapic.c
278
table = iosapic_alloc_irt(num_entries);
drivers/parisc/iosapic.c
299
status = pdc_pci_irt_size(&num_entries, 0);
drivers/parisc/iosapic.c
307
BUG_ON(num_entries == 0);
drivers/parisc/iosapic.c
309
table = iosapic_alloc_irt(num_entries);
drivers/parisc/iosapic.c
317
status = pdc_pci_irt(num_entries, 0, table);
drivers/parisc/iosapic.c
332
num_entries,
drivers/parisc/iosapic.c
335
for (i = 0 ; i < num_entries ; i++, p++) {
drivers/parisc/iosapic.c
347
return num_entries;
drivers/pci/hotplug/ibmphp.h
213
u16 num_entries;
drivers/pci/hotplug/ibmphp_ebda.c
234
u16 ebda_seg, num_entries, next_offset, offset, blk_id, sub_addr, re, rc_id, re_id, base;
drivers/pci/hotplug/ibmphp_ebda.c
332
num_entries = readw(io_mem + sub_addr);
drivers/pci/hotplug/ibmphp_ebda.c
341
rsrc_list_ptr->num_entries = num_entries;
drivers/pci/hotplug/ibmphp_ebda.c
346
debug("num of rsrc: %x\n", num_entries);
drivers/pci/hotplug/ibmphp_ebda.c
930
for (rsrc = 0; rsrc < rsrc_list_ptr->num_entries; rsrc++) {
drivers/pci/tph.c
438
int num_entries, i, offset;
drivers/pci/tph.c
457
num_entries = pcie_tph_get_st_table_size(pdev);
drivers/pci/tph.c
458
for (i = 0; i < num_entries; i++) {
drivers/pci/tph.c
468
int num_entries, i, offset;
drivers/pci/tph.c
489
num_entries = pcie_tph_get_st_table_size(pdev);
drivers/pci/tph.c
490
for (i = 0; i < num_entries; i++) {
drivers/pci/tph.c
506
int num_entries;
drivers/pci/tph.c
513
num_entries = pcie_tph_get_st_table_size(pdev);
drivers/pci/tph.c
514
save_size = sizeof(u32) + num_entries * sizeof(u16);
drivers/phy/cadence/phy-cadence-torrent.c
444
u32 num_entries;
drivers/phy/cadence/phy-cadence-torrent.c
475
for (i = 0; i < tbl->num_entries; i++) {
drivers/phy/cadence/phy-cadence-torrent.c
5356
.num_entries = ARRAY_SIZE(link_cmn_vals_entries),
drivers/phy/cadence/phy-cadence-torrent.c
5360
.num_entries = ARRAY_SIZE(xcvr_diag_vals_entries),
drivers/phy/cadence/phy-cadence-torrent.c
5364
.num_entries = ARRAY_SIZE(pcs_cmn_vals_entries),
drivers/phy/cadence/phy-cadence-torrent.c
5368
.num_entries = ARRAY_SIZE(cmn_vals_entries),
drivers/phy/cadence/phy-cadence-torrent.c
5372
.num_entries = ARRAY_SIZE(cdns_tx_ln_vals_entries),
drivers/phy/cadence/phy-cadence-torrent.c
5376
.num_entries = ARRAY_SIZE(cdns_rx_ln_vals_entries),
drivers/phy/cadence/phy-cadence-torrent.c
5490
.num_entries = ARRAY_SIZE(link_cmn_vals_entries),
drivers/phy/cadence/phy-cadence-torrent.c
5494
.num_entries = ARRAY_SIZE(xcvr_diag_vals_entries),
drivers/phy/cadence/phy-cadence-torrent.c
5498
.num_entries = ARRAY_SIZE(pcs_cmn_vals_entries),
drivers/phy/cadence/phy-cadence-torrent.c
5502
.num_entries = ARRAY_SIZE(j721e_phy_pma_cmn_vals_entries),
drivers/phy/cadence/phy-cadence-torrent.c
5506
.num_entries = ARRAY_SIZE(cmn_vals_entries),
drivers/phy/cadence/phy-cadence-torrent.c
5510
.num_entries = ARRAY_SIZE(ti_tx_ln_vals_entries),
drivers/phy/cadence/phy-cadence-torrent.c
5514
.num_entries = ARRAY_SIZE(cdns_rx_ln_vals_entries),
drivers/phy/cadence/phy-cadence-torrent.c
5818
.num_entries = ARRAY_SIZE(link_cmn_vals_entries),
drivers/phy/cadence/phy-cadence-torrent.c
5822
.num_entries = ARRAY_SIZE(xcvr_diag_vals_entries),
drivers/phy/cadence/phy-cadence-torrent.c
5826
.num_entries = ARRAY_SIZE(pcs_cmn_vals_entries),
drivers/phy/cadence/phy-cadence-torrent.c
5830
.num_entries = ARRAY_SIZE(j721e_phy_pma_cmn_vals_entries),
drivers/phy/cadence/phy-cadence-torrent.c
5834
.num_entries = ARRAY_SIZE(ti_j7200_cmn_vals_entries),
drivers/phy/cadence/phy-cadence-torrent.c
5838
.num_entries = ARRAY_SIZE(ti_j7200_tx_ln_vals_entries),
drivers/phy/cadence/phy-cadence-torrent.c
5842
.num_entries = ARRAY_SIZE(ti_j7200_rx_ln_vals_entries),
drivers/platform/x86/hp/hp-wmi.c
2561
if (fan_table->header.num_entries == 0 ||
drivers/platform/x86/hp/hp-wmi.c
2563
sizeof(struct victus_s_fan_table_entry) * fan_table->header.num_entries > sizeof(fan_data))
drivers/platform/x86/hp/hp-wmi.c
2567
max_rpm = fan_table->entries[fan_table->header.num_entries - 1].cpu_rpm;
drivers/platform/x86/hp/hp-wmi.c
466
u8 num_entries;
drivers/platform/x86/intel/pmc/ssram_telemetry.c
49
header.num_entries = readb(dvsec + INTEL_DVSEC_ENTRIES);
drivers/platform/x86/intel/pmt/crashlog.c
119
int num_entries;
drivers/platform/x86/intel/pmt/crashlog.c
544
for (i = 0; i < priv->num_entries; i++) {
drivers/platform/x86/intel/pmt/crashlog.c
568
struct intel_pmt_entry *entry = &priv->entry[priv->num_entries].entry;
drivers/platform/x86/intel/pmt/crashlog.c
576
priv->num_entries++;
drivers/platform/x86/intel/pmt/telemetry.c
368
for (i = 0; i < priv->num_entries; i++) {
drivers/platform/x86/intel/pmt/telemetry.c
392
struct intel_pmt_entry *entry = &priv->entry[priv->num_entries];
drivers/platform/x86/intel/pmt/telemetry.c
402
priv->num_entries++;
drivers/platform/x86/intel/pmt/telemetry.c
57
int num_entries;
drivers/platform/x86/intel/vsec.c
292
if (!header->num_entries) {
drivers/platform/x86/intel/vsec.c
306
res = kzalloc_objs(*res, header->num_entries);
drivers/platform/x86/intel/vsec.c
323
for (i = 0, tmp = res; i < header->num_entries; i++, tmp++) {
drivers/platform/x86/intel/vsec.c
337
intel_vsec_dev->num_resources = header->num_entries;
drivers/platform/x86/intel/vsec.c
505
pci_read_config_byte(pdev, pos + INTEL_DVSEC_ENTRIES, &header.num_entries);
drivers/platform/x86/intel/vsec.c
553
pci_read_config_byte(pdev, pos + INTEL_DVSEC_ENTRIES, &header.num_entries);
drivers/platform/x86/intel/vsec.c
703
.num_entries = 1,
drivers/platform/x86/intel/vsec_tpmi.c
394
pfs->pfs_header.tpmi_id, pfs->pfs_header.num_entries,
drivers/platform/x86/intel/vsec_tpmi.c
428
for (count = 0; count < pfs->pfs_header.num_entries; ++count) {
drivers/platform/x86/intel/vsec_tpmi.c
482
if (punit >= pfs->pfs_header.num_entries) {
drivers/platform/x86/intel/vsec_tpmi.c
625
res = kzalloc_objs(*res, pfs->pfs_header.num_entries);
drivers/platform/x86/intel/vsec_tpmi.c
637
for (i = 0, tmp = res; i < pfs->pfs_header.num_entries; i++, tmp++) {
drivers/platform/x86/intel/vsec_tpmi.c
647
feature_vsec_dev->num_resources = pfs->pfs_header.num_entries;
drivers/platform/x86/intel/vsec_tpmi.c
79
u64 num_entries:8;
drivers/regulator/ti-abb-regulator.c
505
int num_entries, min_uV = INT_MAX, max_uV = 0;
drivers/regulator/ti-abb-regulator.c
513
num_entries = of_property_count_u32_elems(dev->of_node, pname);
drivers/regulator/ti-abb-regulator.c
514
if (num_entries < 0) {
drivers/regulator/ti-abb-regulator.c
516
return num_entries;
drivers/regulator/ti-abb-regulator.c
519
if (!num_entries || (num_entries % num_values)) {
drivers/regulator/ti-abb-regulator.c
524
num_entries /= num_values;
drivers/regulator/ti-abb-regulator.c
526
info = devm_kcalloc(dev, num_entries, sizeof(*info), GFP_KERNEL);
drivers/regulator/ti-abb-regulator.c
532
volt_table = devm_kcalloc(dev, num_entries, sizeof(unsigned int),
drivers/regulator/ti-abb-regulator.c
537
abb->rdesc.n_voltages = num_entries;
drivers/regulator/ti-abb-regulator.c
542
for (i = 0; i < num_entries; i++, info++, volt_table++) {
drivers/remoteproc/qcom_pil_info.c
24
size_t num_entries;
drivers/remoteproc/qcom_pil_info.c
59
_reloc.num_entries = (u32)resource_size(&imem) / PIL_RELOC_ENTRY_SIZE;
drivers/remoteproc/qcom_pil_info.c
86
for (i = 0; i < _reloc.num_entries; i++) {
drivers/rpmsg/qcom_glink_rpm.c
220
int num_entries;
drivers/rpmsg/qcom_glink_rpm.c
241
num_entries = le32_to_cpu(toc->count);
drivers/rpmsg/qcom_glink_rpm.c
242
if (num_entries > RPM_TOC_MAX_ENTRIES) {
drivers/rpmsg/qcom_glink_rpm.c
247
for (i = 0; i < num_entries; i++) {
drivers/s390/net/qeth_core_mpc.h
744
__u8 num_entries;
drivers/s390/net/qeth_core_mpc.h
781
__u16 num_entries;
drivers/s390/net/qeth_l2_main.c
1342
if (qports->num_entries == 0) {
drivers/s390/net/qeth_l2_main.c
1434
for (i = 0; i < data->ac_event.num_entries; i++) {
drivers/s390/net/qeth_l2_main.c
1486
for (i = 0; i < data->ac_event.num_entries; i++) {
drivers/s390/net/qeth_l2_main.c
1523
hostevs->num_entries;
drivers/s390/net/qeth_l2_main.c
1737
if (qports->num_entries > 0) {
drivers/scsi/aic7xxx/aic79xx.h
1459
u_int num_entries,
drivers/scsi/aic7xxx/aic79xx_core.c
9584
ahd_print_register(const ahd_reg_parse_entry_t *table, u_int num_entries,
drivers/scsi/aic7xxx/aic79xx_core.c
9605
for (entry = 0; entry < num_entries; entry++) {
drivers/scsi/aic7xxx/aic79xx_core.c
9619
if (entry >= num_entries)
drivers/scsi/aic7xxx/aic7xxx.h
1261
u_int num_entries,
drivers/scsi/aic7xxx/aic7xxx_core.c
7053
ahc_print_register(const ahc_reg_parse_entry_t *table, u_int num_entries,
drivers/scsi/aic7xxx/aic7xxx_core.c
7074
for (entry = 0; entry < num_entries; entry++) {
drivers/scsi/aic7xxx/aic7xxx_core.c
7087
if (entry >= num_entries)
drivers/scsi/aic7xxx/aicasm/aicasm_symbol.c
384
symbol_node_t *regnode, u_int num_entries)
drivers/scsi/aic7xxx/aicasm/aicasm_symbol.c
397
if (num_entries != 0)
drivers/scsi/aic7xxx/aicasm/aicasm_symbol.c
414
num_entries != 0 ? regnode->symbol->name : "NULL",
drivers/scsi/aic7xxx/aicasm/aicasm_symbol.c
415
num_entries != 0 ? "_parse_table" : "",
drivers/scsi/aic7xxx/aicasm/aicasm_symbol.c
416
num_entries,
drivers/scsi/aic7xxx/aicasm/aicasm_symbol.c
549
int num_entries;
drivers/scsi/aic7xxx/aicasm/aicasm_symbol.c
551
num_entries = 0;
drivers/scsi/aic7xxx/aicasm/aicasm_symbol.c
557
if (num_entries == 0)
drivers/scsi/aic7xxx/aicasm/aicasm_symbol.c
562
num_entries++;
drivers/scsi/aic7xxx/aicasm/aicasm_symbol.c
566
curnode, num_entries);
drivers/scsi/be2iscsi/be_main.c
1759
if (nbuf == pasync_ctx->num_entries) {
drivers/scsi/be2iscsi/be_main.c
1764
if (++pi == pasync_ctx->num_entries)
drivers/scsi/be2iscsi/be_main.c
2739
pasync_ctx->num_entries = BEISCSI_ASYNC_HDQ_SIZE(phba,
drivers/scsi/be2iscsi/be_main.c
3735
nbufs = phwi_context->pasync_ctx[ulp_num]->num_entries;
drivers/scsi/be2iscsi/be_main.h
592
u16 num_entries;
drivers/scsi/bfa/bfa_fc.h
930
u32 num_entries:16;
drivers/scsi/bfa/bfa_fcbuild.c
560
rpsc_acc->num_entries = cpu_to_be16(1);
drivers/scsi/bfa/bfa_fcs_lport.c
3124
u32 num_entries;
drivers/scsi/bfa/bfa_fcs_lport.c
3145
num_entries = be32_to_cpu(gmal_resp->ms_len);
drivers/scsi/bfa/bfa_fcs_lport.c
3146
if (num_entries == 0) {
drivers/scsi/bfa/bfa_fcs_lport.c
3158
while (num_entries > 0) {
drivers/scsi/bfa/bfa_fcs_lport.c
3178
--num_entries;
drivers/scsi/bfa/bfa_fcs_lport.c
5536
int num_entries;
drivers/scsi/bfa/bfa_fcs_lport.c
5541
num_entries =
drivers/scsi/bfa/bfa_fcs_lport.c
5545
bfa_trc(port->fcs, num_entries);
drivers/scsi/bfa/bfa_fcs_lport.c
5551
for (i = 0; i < num_entries; i++) {
drivers/scsi/bfa/bfi.h
563
u16 num_entries;
drivers/scsi/ipr.c
1741
int i, num_entries;
drivers/scsi/ipr.c
1760
num_entries = min_t(u32, be32_to_cpu(error->num_entries),
drivers/scsi/ipr.c
1763
for (i = 0; i < num_entries; i++, array_entry++) {
drivers/scsi/ipr.c
2205
for (i = 0, fabric = error->desc; i < error->num_entries; i++) {
drivers/scsi/ipr.c
2229
int i, num_entries;
drivers/scsi/ipr.c
2247
num_entries = min_t(u32, error->num_entries,
drivers/scsi/ipr.c
2250
for (i = 0; i < num_entries; i++, array_entry++) {
drivers/scsi/ipr.c
2299
for (i = 0, fabric = error->desc; i < error->num_entries; i++) {
drivers/scsi/ipr.c
2975
driver_dump->hdr.num_entries++;
drivers/scsi/ipr.c
2996
driver_dump->hdr.num_entries++;
drivers/scsi/ipr.c
3017
driver_dump->hdr.num_entries++;
drivers/scsi/ipr.c
3038
driver_dump->hdr.num_entries++;
drivers/scsi/ipr.c
3055
u32 num_entries, max_num_entries, start_off, end_off;
drivers/scsi/ipr.c
3091
driver_dump->hdr.num_entries = 1;
drivers/scsi/ipr.c
3142
num_entries = be32_to_cpu(sdt->hdr.num_entries_used);
drivers/scsi/ipr.c
3144
if (num_entries > max_num_entries)
drivers/scsi/ipr.c
3145
num_entries = max_num_entries;
drivers/scsi/ipr.c
3150
dump->driver_dump.hdr.len += num_entries * sizeof(struct ipr_sdt_entry);
drivers/scsi/ipr.c
3156
for (i = 0; i < num_entries; i++) {
drivers/scsi/ipr.c
6653
for (i = 0; i < mode_page->num_entries; i++) {
drivers/scsi/ipr.c
6714
i < mode_page->num_entries;
drivers/scsi/ipr.c
7022
entries = be16_to_cpu(ioa_cfg->u.cfg_table64->hdr64.num_entries);
drivers/scsi/ipr.c
7024
entries = ioa_cfg->u.cfg_table->hdr.num_entries;
drivers/scsi/ipr.h
1032
__be16 num_entries;
drivers/scsi/ipr.h
1046
__be16 num_entries;
drivers/scsi/ipr.h
1056
cfg < ((fabric)->elem + be16_to_cpu((fabric)->num_entries)); \
drivers/scsi/ipr.h
1062
u8 num_entries;
drivers/scsi/ipr.h
1069
u8 num_entries;
drivers/scsi/ipr.h
1203
__be32 num_entries;
drivers/scsi/ipr.h
1270
u16 num_entries;
drivers/scsi/ipr.h
1589
u32 num_entries;
drivers/scsi/ipr.h
437
u8 num_entries;
drivers/scsi/ipr.h
444
__be16 num_entries;
drivers/scsi/ipr.h
741
u8 num_entries;
drivers/scsi/ipr.h
942
__be32 num_entries;
drivers/scsi/ipr.h
958
u8 num_entries;
drivers/scsi/megaraid/megaraid_sas_fusion.c
3730
int num_entries = 0;
drivers/scsi/megaraid/megaraid_sas_fusion.c
3742
num_entries = complete_cmd_fusion(instance, queue_num, NULL);
drivers/scsi/megaraid/megaraid_sas_fusion.c
3745
return num_entries;
drivers/scsi/megaraid/megaraid_sas_fusion.c
3802
int num_entries;
drivers/scsi/megaraid/megaraid_sas_fusion.c
3812
num_entries = complete_cmd_fusion(instance, irq_ctx->MSIxIndex, irq_ctx);
drivers/scsi/megaraid/megaraid_sas_fusion.c
3813
if (num_entries < budget) {
drivers/scsi/megaraid/megaraid_sas_fusion.c
3820
return num_entries;
drivers/scsi/mpi3mr/mpi/mpi30_ioc.h
515
u8 num_entries;
drivers/scsi/mpi3mr/mpi/mpi30_ioc.h
519
struct mpi3_event_sas_topo_phy_entry phy_entry[] __counted_by(num_entries);
drivers/scsi/mpi3mr/mpi/mpi30_ioc.h
605
u8 num_entries;
drivers/scsi/mpi3mr/mpi/mpi30_ioc.h
610
struct mpi3_event_pcie_topo_port_entry port_entry[] __counted_by(num_entries);
drivers/scsi/mpi3mr/mpi3mr_app.c
1337
u16 num_entries, sz, entry_sz = mrioc->logdata_entry_sz;
drivers/scsi/mpi3mr/mpi3mr_app.c
1342
num_entries = job->request_payload.payload_len / entry_sz;
drivers/scsi/mpi3mr/mpi3mr_app.c
1343
if (num_entries > MPI3MR_BSG_LOGDATA_MAX_ENTRIES)
drivers/scsi/mpi3mr/mpi3mr_app.c
1344
num_entries = MPI3MR_BSG_LOGDATA_MAX_ENTRIES;
drivers/scsi/mpi3mr/mpi3mr_app.c
1345
sz = num_entries * entry_sz;
drivers/scsi/mpi3mr/mpi3mr_fw.c
639
int num_entries = 0;
drivers/scsi/mpi3mr/mpi3mr_fw.c
648
num_entries = mpi3mr_process_op_reply_q(mrioc,
drivers/scsi/mpi3mr/mpi3mr_fw.c
651
return num_entries;
drivers/scsi/mpi3mr/mpi3mr_os.c
1773
event_data->start_phy_num, event_data->num_entries);
drivers/scsi/mpi3mr/mpi3mr_os.c
1774
for (i = 0; i < event_data->num_entries; i++) {
drivers/scsi/mpi3mr/mpi3mr_os.c
1854
for (i = 0; i < event_data->num_entries; i++) {
drivers/scsi/mpi3mr/mpi3mr_os.c
1950
event_data->start_port_num, event_data->num_entries);
drivers/scsi/mpi3mr/mpi3mr_os.c
1951
for (i = 0; i < event_data->num_entries; i++) {
drivers/scsi/mpi3mr/mpi3mr_os.c
2012
for (i = 0; i < event_data->num_entries; i++) {
drivers/scsi/mpi3mr/mpi3mr_os.c
2751
for (i = 0; i < topo_evt->num_entries; i++) {
drivers/scsi/mpi3mr/mpi3mr_os.c
2817
for (i = 0; i < topo_evt->num_entries; i++) {
drivers/scsi/mpt3sas/mpt3sas_base.c
1880
int num_entries = 0;
drivers/scsi/mpt3sas/mpt3sas_base.c
1889
num_entries = _base_process_reply_queue(reply_q);
drivers/scsi/mpt3sas/mpt3sas_base.c
1892
return num_entries;
drivers/scsi/mpt3sas/mpt3sas_base.c
1927
int num_entries = 0;
drivers/scsi/mpt3sas/mpt3sas_base.c
1935
num_entries = _base_process_reply_queue(reply_q);
drivers/scsi/mpt3sas/mpt3sas_base.c
1936
if (num_entries < budget) {
drivers/scsi/mpt3sas/mpt3sas_base.c
1950
return num_entries;
drivers/scsi/pmcraid.c
5026
for (i = 0; i < le16_to_cpu(pinstance->cfg_table->num_entries); i++) {
drivers/scsi/pmcraid.h
361
__le16 num_entries;
drivers/scsi/qla2xxx/qla_bsg.c
125
for (i = 0; i < pri_cfg->num_entries; i++) {
drivers/scsi/qla2xxx/qla_fw.h
2246
uint16_t num_entries; /* Number of entries */
drivers/scsi/qla2xxx/qla_gbl.h
436
u8 *num_entries);
drivers/scsi/qla2xxx/qla_init.c
9723
entries = ha->fcp_prio_cfg->num_entries;
drivers/scsi/qla2xxx/qla_mbx.c
3121
u8 *num_entries)
drivers/scsi/qla2xxx/qla_mbx.c
3161
if (num_entries)
drivers/scsi/qla2xxx/qla_mbx.c
3162
*num_entries = pmap[0];
drivers/scsi/qla2xxx/qla_nx.c
1513
uint32_t entries = le32_to_cpu(directory->num_entries);
drivers/scsi/qla2xxx/qla_nx.c
1791
entries = le32_to_cpu(ptab_desc->num_entries);
drivers/scsi/qla2xxx/qla_nx.h
807
__le32 num_entries;
drivers/scsi/qla2xxx/qla_sup.c
3695
len = ha->fcp_prio_cfg->num_entries * sizeof(struct qla_fcp_prio_entry);
drivers/scsi/qla4xxx/ql4_os.c
149
uint32_t *num_entries, char *buf);
drivers/scsi/qla4xxx/ql4_os.c
768
uint32_t *num_entries, char *buf)
drivers/scsi/qla4xxx/ql4_os.c
784
__func__, *num_entries, chap_tbl_idx);
drivers/scsi/qla4xxx/ql4_os.c
817
if (valid_chap_entries == *num_entries)
drivers/scsi/qla4xxx/ql4_os.c
825
*num_entries = valid_chap_entries;
drivers/scsi/scsi_debug.c
9090
int *num_entries;
drivers/scsi/scsi_debug.c
9139
(*data->num_entries)++;
drivers/scsi/scsi_debug.c
9145
int num_entries = 0;
drivers/scsi/scsi_debug.c
9148
.num_entries = &num_entries,
drivers/scsi/scsi_debug.c
9154
if (num_entries > 0)
drivers/scsi/scsi_debug.c
9155
atomic_add(num_entries, &sdeb_mq_poll_count);
drivers/scsi/scsi_debug.c
9156
return num_entries;
drivers/scsi/scsi_transport_iscsi.c
3318
chap_buf_size = (ev->u.get_chap.num_entries * sizeof(*chap_rec));
drivers/scsi/scsi_transport_iscsi.c
3346
evchap->u.get_chap.num_entries = ev->u.get_chap.num_entries;
drivers/scsi/scsi_transport_iscsi.c
3351
&evchap->u.get_chap.num_entries, buf);
drivers/soc/qcom/smem.c
1029
for (i = 0; i < le32_to_cpu(ptable->num_entries); i++) {
drivers/soc/qcom/smem.c
170
__le32 num_entries;
drivers/soc/qcom/smem.c
905
info = (struct smem_info *)&ptable->entry[le32_to_cpu(ptable->num_entries)];
drivers/soc/qcom/smem.c
980
for (i = 0; i < le32_to_cpu(ptable->num_entries); i++) {
drivers/soc/qcom/smsm.c
484
u32 num_entries;
drivers/soc/qcom/smsm.c
495
smsm->num_entries = SMSM_DEFAULT_NUM_ENTRIES;
drivers/soc/qcom/smsm.c
500
smsm->num_entries = info->num_entries;
drivers/soc/qcom/smsm.c
505
smsm->num_entries, smsm->num_hosts);
drivers/soc/qcom/smsm.c
533
smsm->num_entries,
drivers/soc/qcom/smsm.c
576
smsm->num_entries * sizeof(u32));
drivers/soc/qcom/smsm.c
590
size = smsm->num_entries * smsm->num_hosts * sizeof(u32);
drivers/soc/qcom/smsm.c
622
if (ret || id >= smsm->num_entries) {
drivers/soc/qcom/smsm.c
649
for (id = 0; id < smsm->num_entries; id++)
drivers/soc/qcom/smsm.c
667
for (id = 0; id < smsm->num_entries; id++)
drivers/soc/qcom/smsm.c
83
u32 num_entries;
drivers/staging/media/ipu7/ipu7-isys-csi-phy.c
330
for (i = 0; i < desc.num_entries; i++) {
drivers/staging/media/ipu7/ipu7-isys-csi2.c
517
for (i = 0; i < desc.num_entries; i++) {
drivers/staging/media/ipu7/ipu7-isys-csi2.c
537
for (i = 0; i < desc.num_entries; i++) {
drivers/target/sbp/sbp_target.c
1785
int num_luns, num_entries, idx = 0, mgt_agt_addr, ret;
drivers/target/sbp/sbp_target.c
1810
num_entries = ARRAY_SIZE(sbp_unit_directory_template) + 4 + num_luns;
drivers/target/sbp/sbp_target.c
1813
num_entries++;
drivers/target/sbp/sbp_target.c
1816
data = kcalloc((num_entries + 4), sizeof(u32), GFP_KERNEL);
drivers/target/sbp/sbp_target.c
1821
data[idx++] = num_entries << 16;
drivers/tee/optee/optee_private.h
300
size_t num_entries;
drivers/tee/optee/smc_abi.c
371
static size_t get_pages_list_size(size_t num_entries)
drivers/tee/optee/smc_abi.c
373
int pages = DIV_ROUND_UP(num_entries, PAGELIST_ENTRIES_PER_PAGE);
drivers/tee/optee/smc_abi.c
378
static u64 *optee_allocate_pages_list(size_t num_entries)
drivers/tee/optee/smc_abi.c
380
return alloc_pages_exact(get_pages_list_size(num_entries), GFP_KERNEL);
drivers/tee/optee/smc_abi.c
383
static void optee_free_pages_list(void *list, size_t num_entries)
drivers/tee/optee/smc_abi.c
385
free_pages_exact(list, get_pages_list_size(num_entries));
drivers/tee/optee/smc_abi.c
736
call_ctx->num_entries = page_count;
drivers/tee/optee/smc_abi.c
774
call_ctx->num_entries);
drivers/tee/optee/smc_abi.c
776
call_ctx->num_entries = 0;
drivers/usb/cdns3/cdnsp-gadget.h
1304
unsigned int num_entries;
drivers/usb/cdns3/cdnsp-mem.c
1039
erst->num_entries = evt_ring->num_segs;
drivers/usb/cdns3/cdnsp-mem.c
1055
size_t size = sizeof(struct cdnsp_erst_entry) * (erst->num_entries);
drivers/usb/host/xhci-dbgcap.c
430
erst->num_entries = 1;
drivers/usb/host/xhci-dbgcap.c
576
writel(dbc->erst.num_entries, &dbc->regs->ersts);
drivers/usb/host/xhci-mem.c
1807
erst->num_entries = evt_ring->num_segs;
drivers/usb/host/xhci-mem.c
1852
erst_size = array_size(sizeof(struct xhci_erst_entry), ir->erst.num_entries);
drivers/usb/host/xhci.h
1395
unsigned int num_entries;
drivers/xen/xen-acpi-processor.c
431
if (acpi_psd[i].num_entries) {
fs/btrfs/accessors.h
550
num_entries, 64);
fs/btrfs/free-space-cache.c
745
u64 num_entries;
fs/btrfs/free-space-cache.c
772
num_entries = btrfs_free_space_entries(leaf, header);
fs/btrfs/free-space-cache.c
791
if (!num_entries)
fs/btrfs/free-space-cache.c
812
while (num_entries) {
fs/btrfs/free-space-cache.c
869
num_entries--;
fs/btrfs/raid56.c
199
unsigned int num_entries = 1U << BTRFS_STRIPE_HASH_TABLE_BITS;
fs/btrfs/raid56.c
211
table = kvzalloc_flex(*table, table, num_entries);
fs/btrfs/raid56.c
220
for (unsigned int i = 0; i < num_entries; i++) {
fs/btrfs/relocation.c
3864
struct btrfs_key *entries, unsigned int num_entries)
fs/btrfs/relocation.c
3874
data_sizes = kzalloc(sizeof(u32) * min_t(u32, num_entries, max_items), GFP_NOFS);
fs/btrfs/relocation.c
3882
batch.nr = min_t(u32, num_entries, max_items);
fs/btrfs/relocation.c
3887
if (num_entries <= max_items)
fs/btrfs/relocation.c
3890
num_entries -= max_items;
fs/btrfs/relocation.c
4434
unsigned int max_entries, num_entries;
fs/btrfs/relocation.c
4554
num_entries = 0;
fs/btrfs/relocation.c
4557
entries[num_entries].objectid = bg->start;
fs/btrfs/relocation.c
4558
entries[num_entries].type = BTRFS_IDENTITY_REMAP_KEY;
fs/btrfs/relocation.c
4559
entries[num_entries].offset = bg->length;
fs/btrfs/relocation.c
4560
num_entries++;
fs/btrfs/relocation.c
4563
entries[num_entries].objectid = bg->start;
fs/btrfs/relocation.c
4564
entries[num_entries].type = BTRFS_IDENTITY_REMAP_KEY;
fs/btrfs/relocation.c
4565
entries[num_entries].offset = space_runs[0].start - bg->start;
fs/btrfs/relocation.c
4566
num_entries++;
fs/btrfs/relocation.c
4570
entries[num_entries].objectid = space_runs[i - 1].end;
fs/btrfs/relocation.c
4571
entries[num_entries].type = BTRFS_IDENTITY_REMAP_KEY;
fs/btrfs/relocation.c
4572
entries[num_entries].offset =
fs/btrfs/relocation.c
4574
num_entries++;
fs/btrfs/relocation.c
4578
entries[num_entries].objectid =
fs/btrfs/relocation.c
4580
entries[num_entries].type = BTRFS_IDENTITY_REMAP_KEY;
fs/btrfs/relocation.c
4581
entries[num_entries].offset =
fs/btrfs/relocation.c
4583
num_entries++;
fs/btrfs/relocation.c
4586
if (num_entries == 0)
fs/btrfs/relocation.c
4590
bg->identity_remap_count = num_entries;
fs/btrfs/relocation.c
4592
ret = add_remap_tree_entries(trans, path, entries, num_entries);
fs/ceph/mds_client.c
2536
unsigned int num_entries;
fs/ceph/mds_client.c
2541
num_entries = ci->i_files + ci->i_subdirs;
fs/ceph/mds_client.c
2543
num_entries = max(num_entries, 1U);
fs/ceph/mds_client.c
2544
num_entries = min(num_entries, opt->max_readdir);
fs/ceph/mds_client.c
2546
bytes_count = (u64)size * num_entries;
fs/ceph/mds_client.c
2563
num_entries = (PAGE_SIZE << order) / size;
fs/ceph/mds_client.c
2564
num_entries = min(num_entries, opt->max_readdir);
fs/ceph/mds_client.c
2567
req->r_num_caps = num_entries + 1;
fs/ceph/mds_client.c
2568
req->r_args.readdir.max_entries = cpu_to_le32(num_entries);
fs/exfat/dir.c
1005
int num_entries = exfat_calc_num_entries(p_uniname);
fs/exfat/dir.c
1008
if (num_entries < 0)
fs/exfat/dir.c
1009
return num_entries;
fs/exfat/dir.c
1048
dentry, num_entries,
fs/exfat/dir.c
486
void exfat_init_ext_entry(struct exfat_entry_set_cache *es, int num_entries,
fs/exfat/dir.c
494
ep->dentry.file.num_ext = (unsigned char)(num_entries - 1);
fs/exfat/dir.c
50
for (i = ES_IDX_FIRST_FILENAME; i < es.num_entries; i++) {
fs/exfat/dir.c
500
for (i = ES_IDX_FIRST_FILENAME; i < num_entries; i++) {
fs/exfat/dir.c
515
for (i = order; i < es->num_entries; i++) {
fs/exfat/dir.c
524
if (order < es->num_entries)
fs/exfat/dir.c
534
for (i = ES_IDX_FILE; i < es->num_entries; i++) {
fs/exfat/dir.c
761
unsigned int num_entries)
fs/exfat/dir.c
789
if (num_entries == ES_ALL_ENTRIES) {
fs/exfat/dir.c
798
num_entries = ep->dentry.file.num_ext + 1;
fs/exfat/dir.c
801
es->num_entries = num_entries;
fs/exfat/dir.c
803
num_bh = EXFAT_B_TO_BLK_ROUND_UP(off + num_entries * DENTRY_SIZE, sb);
fs/exfat/dir.c
842
int entry, unsigned int num_entries)
fs/exfat/dir.c
848
ret = __exfat_get_dentry_set(es, sb, p_dir, entry, num_entries);
fs/exfat/dir.c
853
for (i = ES_IDX_STREAM; i < es->num_entries; i++) {
fs/exfat/dir.c
878
for (i = 0; i < es->num_entries; i++) {
fs/exfat/dir.c
903
es->num_entries = EXFAT_B_TO_DEN(EXFAT_BLK_TO_B(es->num_bh, es->sb) - es->start_off);
fs/exfat/dir.c
904
for (; i < es->num_entries; i++) {
fs/exfat/dir.c
929
int entry, unsigned int num_entries)
fs/exfat/dir.c
933
ret = __exfat_get_dentry_set(es, sb, p_dir, entry, num_entries);
fs/exfat/dir.c
952
int dentry, int num_entries, int entry_type)
fs/exfat/dir.c
968
if (candi_empty->count == num_entries ||
fs/exfat/exfat_fs.h
197
unsigned int num_entries;
fs/exfat/exfat_fs.h
481
struct exfat_chain *p_dir, int num_entries,
fs/exfat/exfat_fs.h
498
void exfat_init_ext_entry(struct exfat_entry_set_cache *es, int num_entries,
fs/exfat/exfat_fs.h
514
unsigned int num_entries);
fs/exfat/exfat_fs.h
519
unsigned int num_entries);
fs/exfat/namei.c
1035
if (old_es.num_entries < num_new_entries) {
fs/exfat/namei.c
205
int num_entries, struct exfat_entry_set_cache *es)
fs/exfat/namei.c
225
num_entries > hint_femp->count)
fs/exfat/namei.c
235
while (dentry + num_entries <= total_entries &&
fs/exfat/namei.c
239
ret = exfat_get_empty_dentry_set(es, sb, &clu, i, num_entries);
fs/exfat/namei.c
304
struct exfat_chain *p_dir, int num_entries,
fs/exfat/namei.c
327
num_entries, es)) < 0) {
fs/exfat/namei.c
470
int ret, dentry, num_entries;
fs/exfat/namei.c
484
num_entries = exfat_calc_num_entries(&uniname);
fs/exfat/namei.c
485
if (num_entries < 0) {
fs/exfat/namei.c
486
ret = num_entries;
fs/exfat/namei.c
491
dentry = exfat_find_empty_entry(inode, &info->dir, num_entries, &es);
fs/exfat/namei.c
512
exfat_init_ext_entry(&es, num_entries, &uniname);
fs/nfs/flexfilelayout/flexfilelayout.c
2549
unsigned int num_entries)
fs/nfs/flexfilelayout/flexfilelayout.c
2553
for (i = 0; i < num_entries; i++) {
fs/ocfs2/dir.c
2341
int dx_inline, u32 num_entries,
fs/ocfs2/dir.c
2388
dx_root->dr_num_entries = cpu_to_le32(num_entries);
fs/ocfs2/xattr.c
1456
int num_entries)
fs/ocfs2/xattr.c
1465
(num_entries * sizeof(struct ocfs2_xattr_entry)) -
fs/smb/client/cached_dir.c
387
cfids->num_entries--;
fs/smb/client/cached_dir.c
457
cfid->cfids->num_entries--;
fs/smb/client/cached_dir.c
49
if (cfids->num_entries >= max_cached_dirs) {
fs/smb/client/cached_dir.c
57
cfids->num_entries++;
fs/smb/client/cached_dir.c
612
cfids->num_entries--;
fs/smb/client/cached_dir.c
683
cfids->num_entries--;
fs/smb/client/cached_dir.c
775
cfids->num_entries--;
fs/smb/client/cached_dir.h
64
int num_entries;
fs/smb/client/cifs_debug.c
326
cfids->num_entries,
fs/smb/client/smb2pdu.c
5636
srch_inf->entries_in_buffer = num_entries(
include/acpi/actbl1.h
118
u32 num_entries;
include/acpi/cppc_acpi.h
72
int num_entries;
include/acpi/processor.h
104
u64 num_entries;
include/acpi/processor.h
145
u64 num_entries;
include/drm/bridge/mhl.h
326
u8 num_entries;
include/drm/bridge/mhl.h
335
u8 num_entries;
include/drm/intel/intel-gtt.h
30
void intel_gmch_gtt_clear_range(unsigned int first_entry, unsigned int num_entries);
include/linux/bnge/hsi.h
3951
__le16 num_entries;
include/linux/bnge/hsi.h
4022
__le32 num_entries;
include/linux/bnge/hsi.h
4190
__le32 num_entries;
include/linux/bnxt/hsi.h
3888
__le32 num_entries;
include/linux/bnxt/hsi.h
4011
__le32 num_entries;
include/linux/bnxt/hsi.h
8960
__le32 num_entries;
include/linux/bnxt/hsi.h
9004
__le32 num_entries;
include/linux/efi.h
601
u32 num_entries;
include/linux/habanalabs/cpucp_if.h
1237
__u8 num_entries;
include/linux/intel_vsec.h
58
u8 num_entries;
include/linux/mlx5/qp.h
503
__be16 num_entries;
include/linux/pds/pds_adminq.h
1365
__le32 num_entries;
include/linux/pds/pds_adminq.h
1366
u8 entries[] __counted_by_le(num_entries);
include/linux/platform_data/cros_ec_commands.h
3017
uint32_t num_entries;
include/media/v4l2-subdev.h
401
unsigned short num_entries;
include/net/flow_offload.h
326
u32 num_entries;
include/net/flow_offload.h
337
unsigned int num_entries;
include/net/flow_offload.h
338
struct flow_action_entry entries[] __counted_by(num_entries);
include/net/flow_offload.h
343
return action->num_entries;
include/net/flow_offload.h
354
return action->num_entries == 1;
include/net/flow_offload.h
360
return entry == &action->entries[action->num_entries - 1];
include/net/flow_offload.h
365
__i < (__actions)->num_entries; \
include/net/nfc/nci.h
416
__u8 num_entries;
include/net/pkt_sched.h
237
size_t num_entries;
include/net/tc_act/tc_gate.h
110
u32 num_entries;
include/net/tc_act/tc_gate.h
113
num_entries = p->num_entries;
include/net/tc_act/tc_gate.h
115
return num_entries;
include/net/tc_act/tc_gate.h
124
u32 num_entries;
include/net/tc_act/tc_gate.h
128
num_entries = p->num_entries;
include/net/tc_act/tc_gate.h
133
if (i != num_entries)
include/net/tc_act/tc_gate.h
136
oe = kzalloc_objs(*oe, num_entries, GFP_ATOMIC);
include/net/tc_act/tc_gate.h
33
size_t num_entries;
include/rdma/ib_verbs.h
2405
int (*poll_cq)(struct ib_cq *cq, int num_entries, struct ib_wc *wc);
include/rdma/ib_verbs.h
4111
static inline int ib_poll_cq(struct ib_cq *cq, int num_entries,
include/rdma/ib_verbs.h
4114
return cq->device->ops.poll_cq(cq, num_entries, wc);
include/rdma/rdma_vt.h
190
int num_entries;
include/scsi/iscsi_if.h
201
uint32_t num_entries; /* number of CHAP entries
include/scsi/scsi_transport_iscsi.h
141
uint32_t *num_entries, char *buf);
include/uapi/cxl/features.h
101
__le16 num_entries;
include/uapi/cxl/features.h
105
struct cxl_feat_entry ents[] __counted_by_le(num_entries);
include/uapi/drm/amdgpu_drm.h
830
__u32 num_entries;
include/uapi/drm/amdgpu_drm.h
842
__u32 num_entries;
include/uapi/drm/amdxdna_accel.h
186
__u32 num_entries;
include/uapi/linux/btrfs_tree.h
749
__le64 num_entries;
include/uapi/linux/i2o-dev.h
194
__u16 num_entries;
include/uapi/linux/netfilter_arp/arp_tables.h
150
unsigned int num_entries;
include/uapi/linux/netfilter_arp/arp_tables.h
166
unsigned int num_entries;
include/uapi/linux/netfilter_ipv4/ip_tables.h
172
unsigned int num_entries;
include/uapi/linux/netfilter_ipv4/ip_tables.h
188
unsigned int num_entries;
include/uapi/linux/netfilter_ipv6/ip6_tables.h
212
unsigned int num_entries;
include/uapi/linux/netfilter_ipv6/ip6_tables.h
228
unsigned int num_entries;
include/uapi/rdma/mana-abi.h
86
__aligned_u64 num_entries;
include/xen/interface/platform.h
388
uint64_t num_entries;
kernel/bpf/hashtab.c
222
u32 num_entries = htab->map.max_entries;
kernel/bpf/hashtab.c
226
num_entries += num_possible_cpus();
kernel/bpf/hashtab.c
228
for (i = 0; i < num_entries; i++) {
kernel/bpf/hashtab.c
2319
u64 num_entries, usage;
kernel/bpf/hashtab.c
2325
num_entries = map->max_entries;
kernel/bpf/hashtab.c
2327
num_entries += num_possible_cpus();
kernel/bpf/hashtab.c
2329
usage += htab->elem_size * num_entries;
kernel/bpf/hashtab.c
2332
usage += value_size * num_possible_cpus() * num_entries;
kernel/bpf/hashtab.c
2338
num_entries = htab->use_percpu_counter ?
kernel/bpf/hashtab.c
2341
usage += (htab->elem_size + LLIST_NODE_SZ) * num_entries;
kernel/bpf/hashtab.c
2343
usage += (LLIST_NODE_SZ + sizeof(void *)) * num_entries;
kernel/bpf/hashtab.c
2344
usage += value_size * num_possible_cpus() * num_entries;
kernel/bpf/hashtab.c
240
u32 num_entries = htab->map.max_entries;
kernel/bpf/hashtab.c
246
num_entries += num_possible_cpus();
kernel/bpf/hashtab.c
247
for (i = 0; i < num_entries; i++) {
kernel/bpf/hashtab.c
316
u32 num_entries = htab->map.max_entries;
kernel/bpf/hashtab.c
320
num_entries += num_possible_cpus();
kernel/bpf/hashtab.c
322
htab->elems = bpf_map_area_alloc((u64)htab->elem_size * num_entries,
kernel/bpf/hashtab.c
330
for (i = 0; i < num_entries; i++) {
kernel/bpf/hashtab.c
360
htab->elem_size, num_entries);
kernel/bpf/hashtab.c
364
htab->elem_size, num_entries);
kernel/kcsan/report.c
278
static int get_stack_skipnr(const unsigned long stack_entries[], int num_entries)
kernel/kcsan/report.c
284
for (skip = 0; skip < num_entries; ++skip) {
kernel/kcsan/report.c
316
replace_stack_entry(unsigned long stack_entries[], int num_entries, unsigned long ip,
kernel/kcsan/report.c
328
for (skip = 0; skip < num_entries; ++skip) {
kernel/kcsan/report.c
345
return get_stack_skipnr(stack_entries, num_entries);
kernel/kcsan/report.c
349
sanitize_stack_entries(unsigned long stack_entries[], int num_entries, unsigned long ip,
kernel/kcsan/report.c
352
return ip ? replace_stack_entry(stack_entries, num_entries, ip, replaced) :
kernel/kcsan/report.c
353
get_stack_skipnr(stack_entries, num_entries);
kernel/kcsan/report.c
369
print_stack_trace(unsigned long stack_entries[], int num_entries, unsigned long reordered_to)
kernel/kcsan/report.c
372
stack_trace_print(stack_entries, num_entries, 0);
mm/kfence/core.c
209
static u32 get_alloc_stack_hash(unsigned long *stack_entries, size_t num_entries)
mm/kfence/core.c
211
num_entries = min(num_entries, UNIQUE_ALLOC_STACK_DEPTH);
mm/kfence/core.c
212
num_entries = filter_irq_stacks(stack_entries, num_entries);
mm/kfence/core.c
213
return jhash(stack_entries, num_entries * sizeof(stack_entries[0]), stack_hash_seed);
mm/kfence/report.c
100
if (fallback < num_entries)
mm/kfence/report.c
104
return skipnr < num_entries ? skipnr : 0;
mm/kfence/report.c
50
static int get_stack_skipnr(const unsigned long stack_entries[], int num_entries,
mm/kfence/report.c
74
for (skipnr = 0; skipnr < num_entries; skipnr++) {
mm/kmsan/report.c
36
int num_entries)
mm/kmsan/report.c
41
for (skip = 0; skip < num_entries; ++skip) {
net/batman-adv/translation-table.c
302
atomic_add(v, &vlan->tt.num_entries);
net/batman-adv/translation-table.c
3246
u16 num_entries)
net/batman-adv/translation-table.c
3258
_batadv_tt_update_changes(bat_priv, orig_node, tt_change, num_entries,
net/batman-adv/translation-table.c
3324
u8 *resp_src, u16 num_entries)
net/batman-adv/translation-table.c
3334
resp_src, tt_data->ttvn, num_entries,
net/batman-adv/translation-table.c
3348
resp_src, num_entries);
net/batman-adv/translation-table.c
3350
batadv_tt_update_changes(bat_priv, orig_node, num_entries,
net/batman-adv/translation-table.c
347
if (atomic_add_return(v, &vlan->tt.num_entries) == 0) {
net/batman-adv/translation-table.c
3944
u16 num_entries, num_vlan;
net/batman-adv/translation-table.c
3961
num_entries = batadv_tt_entries(tvlv_value_len);
net/batman-adv/translation-table.c
3964
tt_change, num_entries, tt_data->ttvn);
net/batman-adv/translation-table.c
521
tt_local_entries += atomic_read(&vlan->tt.num_entries);
net/batman-adv/translation-table.c
800
u16 num_entries = 0;
net/batman-adv/translation-table.c
810
num_entries += atomic_read(&vlan->tt.num_entries);
net/batman-adv/translation-table.c
817
*tt_len = batadv_tt_len(num_entries);
net/batman-adv/translation-table.c
889
vlan_entries = atomic_read(&vlan->tt.num_entries);
net/batman-adv/translation-table.c
918
vlan_entries = atomic_read(&vlan->tt.num_entries);
net/batman-adv/types.h
331
atomic_t num_entries;
net/bluetooth/hci_sync.c
2475
u8 *num_entries)
net/bluetooth/hci_sync.c
2489
if (*num_entries >= hdev->le_accept_list_size)
net/bluetooth/hci_sync.c
2514
*num_entries += 1;
net/bluetooth/hci_sync.c
2757
u8 num_entries = 0;
net/bluetooth/hci_sync.c
2802
num_entries = 1;
net/bluetooth/hci_sync.c
2804
&num_entries);
net/bluetooth/hci_sync.c
2836
num_entries++;
net/bluetooth/hci_sync.c
2861
&num_entries);
net/bluetooth/hci_sync.c
2883
&num_entries);
net/bridge/br_multicast_eht.c
171
eht_host->num_entries--;
net/bridge/br_multicast_eht.c
313
if (!allow_zero_src && eht_host->num_entries >= PG_SRC_ENT_LIMIT)
net/bridge/br_multicast_eht.c
335
eht_host->num_entries++;
net/bridge/br_private_mcast_eht.h
22
unsigned int num_entries;
net/core/drop_monitor.c
359
for (i = 0; i < hw_entries->num_entries; i++) {
net/core/drop_monitor.c
458
for (i = 0; i < hw_entries->num_entries; i++) {
net/core/drop_monitor.c
466
if (WARN_ON_ONCE(hw_entries->num_entries == dm_hit_limit))
net/core/drop_monitor.c
469
hw_entry = &hw_entries->entries[hw_entries->num_entries];
net/core/drop_monitor.c
473
hw_entries->num_entries++;
net/core/drop_monitor.c
72
u32 num_entries;
net/core/flow_offload.c
19
rule->action.num_entries = num_actions;
net/core/flow_offload.c
39
fl_action->action.num_entries = num_actions;
net/ipv4/netfilter/arp_tables.c
1056
u32 num_entries;
net/ipv4/netfilter/arp_tables.c
1175
info->number = compatr->num_entries;
net/ipv4/netfilter/arp_tables.c
1179
ret = xt_compat_init_offsets(NFPROTO_ARP, compatr->num_entries);
net/ipv4/netfilter/arp_tables.c
1193
if (j != compatr->num_entries)
net/ipv4/netfilter/arp_tables.c
1203
newinfo->number = compatr->num_entries;
net/ipv4/netfilter/arp_tables.c
531
newinfo->number = repl->num_entries;
net/ipv4/netfilter/arp_tables.c
553
if (i < repl->num_entries)
net/ipv4/netfilter/arp_tables.c
562
if (i != repl->num_entries)
net/ipv4/netfilter/arp_tables.c
827
info.num_entries = private->number;
net/ipv4/netfilter/ip_tables.c
1208
u32 num_entries;
net/ipv4/netfilter/ip_tables.c
1407
info->number = compatr->num_entries;
net/ipv4/netfilter/ip_tables.c
1411
ret = xt_compat_init_offsets(AF_INET, compatr->num_entries);
net/ipv4/netfilter/ip_tables.c
1425
if (j != compatr->num_entries)
net/ipv4/netfilter/ip_tables.c
1435
newinfo->number = compatr->num_entries;
net/ipv4/netfilter/ip_tables.c
670
newinfo->number = repl->num_entries;
net/ipv4/netfilter/ip_tables.c
691
if (i < repl->num_entries)
net/ipv4/netfilter/ip_tables.c
700
if (i != repl->num_entries)
net/ipv4/netfilter/ip_tables.c
982
info.num_entries = private->number;
net/ipv6/netfilter/ip6_tables.c
1224
u32 num_entries;
net/ipv6/netfilter/ip6_tables.c
1422
info->number = compatr->num_entries;
net/ipv6/netfilter/ip6_tables.c
1426
ret = xt_compat_init_offsets(AF_INET6, compatr->num_entries);
net/ipv6/netfilter/ip6_tables.c
1440
if (j != compatr->num_entries)
net/ipv6/netfilter/ip6_tables.c
1450
newinfo->number = compatr->num_entries;
net/ipv6/netfilter/ip6_tables.c
687
newinfo->number = repl->num_entries;
net/ipv6/netfilter/ip6_tables.c
708
if (i < repl->num_entries)
net/ipv6/netfilter/ip6_tables.c
717
if (i != repl->num_entries)
net/ipv6/netfilter/ip6_tables.c
998
info.num_entries = private->number;
net/netfilter/nf_flow_table_offload.c
223
if (unlikely(flow_rule->rule->action.num_entries >= NF_FLOW_RULE_ACTION_MAX))
net/netfilter/nf_flow_table_offload.c
226
i = flow_rule->rule->action.num_entries++;
net/netfilter/nf_flow_table_offload.c
829
flow_rule->rule->action.num_entries = 0;
net/netfilter/nf_flow_table_offload.c
848
for (i = 0; i < flow_rule->rule->action.num_entries; i++) {
net/netfilter/xt_repldata.h
35
tbl->repl.num_entries = nhooks + 1; \
net/nfc/nci/ntf.c
61
if (skb->len < offsetofend(struct nci_core_conn_credit_ntf, num_entries))
net/nfc/nci/ntf.c
66
pr_debug("num_entries %d\n", ntf->num_entries);
net/nfc/nci/ntf.c
68
if (ntf->num_entries > NCI_MAX_NUM_CONN)
net/nfc/nci/ntf.c
69
ntf->num_entries = NCI_MAX_NUM_CONN;
net/nfc/nci/ntf.c
71
if (skb->len < offsetofend(struct nci_core_conn_credit_ntf, num_entries) +
net/nfc/nci/ntf.c
72
ntf->num_entries * sizeof(struct conn_credit_entry))
net/nfc/nci/ntf.c
76
for (i = 0; i < ntf->num_entries; i++) {
net/sched/act_ct.c
250
int num_entries = action->num_entries;
net/sched/act_ct.c
281
for (i = num_entries; i < action->num_entries; i++)
net/sched/act_ct.c
283
action->num_entries = num_entries;
net/sched/act_ct.c
72
int i = flow_action->num_entries++;
net/sched/act_gate.c
258
dst->num_entries = i;
net/sched/act_gate.c
297
sched->num_entries = i;
net/sched/act_gate.c
722
entry->gate.num_entries = tcf_gate_num_entries(act);
net/sched/sch_taprio.c
1122
sched->num_entries = i;
net/sched/sch_taprio.c
1168
if (new->cycle_time < new->num_entries * length_to_duration(q, ETH_ZLEN)) {
net/sched/sch_taprio.c
1375
static struct tc_taprio_qopt_offload *taprio_offload_alloc(int num_entries)
net/sched/sch_taprio.c
1379
__offload = kzalloc_flex(*__offload, offload.entries, num_entries);
net/sched/sch_taprio.c
1483
offload->num_entries = i;
net/sched/sch_taprio.c
1543
offload = taprio_offload_alloc(sched->num_entries);
net/sched/sch_taprio.c
1913
if (new_admin->num_entries == 0) {
net/sched/sch_taprio.c
81
size_t num_entries;
net/sctp/protocol.c
1468
int num_entries;
net/sctp/protocol.c
1571
num_entries = (1UL << order) * PAGE_SIZE /
net/sctp/protocol.c
1579
sctp_port_hashsize = rounddown_pow_of_two(num_entries);
net/sctp/protocol.c
1591
num_entries);
net/smc/smc_llc.c
1765
int num_entries;
net/smc/smc_llc.c
1773
num_entries = llc->rtoken[0].num_rkeys;
net/smc/smc_llc.c
1774
if (num_entries > SMC_LLC_RKEYS_PER_MSG)
net/smc/smc_llc.c
1783
for (i = 1; i <= min_t(u8, num_entries, SMC_LLC_RKEYS_PER_MSG - 1); i++)
scripts/sorttable.c
171
unsigned int num_entries = *orc_ip_size / sizeof(int);
scripts/sorttable.c
172
unsigned int orc_size = num_entries * sizeof(struct orc_entry);
scripts/sorttable.c
196
for (i = 0; i < num_entries; i++) {
scripts/sorttable.c
202
qsort(idxs, num_entries, sizeof(int), orc_sort_cmp);
scripts/sorttable.c
204
for (i = 0; i < num_entries; i++) {
scripts/sorttable.c
744
int num_entries = shdr_size(extab_sec) / extable_ent_size;
scripts/sorttable.c
745
qsort(extab_image, num_entries,
sound/hda/codecs/side-codecs/cs35l41_hda.c
504
for (i = 0; i < le32_to_cpu(params->num_entries); i++) {
sound/hda/codecs/side-codecs/cs35l41_hda.c
62
__le32 num_entries;
sound/hda/codecs/side-codecs/cs35l41_hda_property.c
150
unsigned int num_entries = 0;
sound/hda/codecs/side-codecs/cs35l41_hda_property.c
161
reset_index = num_entries;
sound/hda/codecs/side-codecs/cs35l41_hda_property.c
162
num_entries++;
sound/hda/codecs/side-codecs/cs35l41_hda_property.c
166
spkid_index = num_entries;
sound/hda/codecs/side-codecs/cs35l41_hda_property.c
167
num_entries++;
sound/hda/codecs/side-codecs/cs35l41_hda_property.c
171
csgpio_index = num_entries;
sound/hda/codecs/side-codecs/cs35l41_hda_property.c
172
num_entries++;
sound/hda/codecs/side-codecs/cs35l41_hda_property.c
175
if (!num_entries)
sound/hda/codecs/side-codecs/cs35l41_hda_property.c
179
num_entries++;
sound/hda/codecs/side-codecs/cs35l41_hda_property.c
181
gpio_mapping = devm_kcalloc(physdev, num_entries, sizeof(struct acpi_gpio_mapping),
sound/soc/codecs/cs-amp-lib.c
479
unsigned long num_entries;
sound/soc/codecs/cs-amp-lib.c
541
num_entries = max(num_amps, amp_index + 1);
sound/soc/codecs/cs-amp-lib.c
542
if (!data || (data->count < num_entries)) {
sound/soc/codecs/cs-amp-lib.c
544
unsigned int new_data_size = struct_size(data, data, num_entries);
sound/soc/codecs/cs-amp-lib.c
558
data->count = num_entries;
sound/soc/intel/avs/topology.c
443
void **dict, u32 *num_entries, size_t entry_size,
sound/soc/intel/avs/topology.c
456
*num_entries = le32_to_cpu(tuple->value);
sound/soc/intel/avs/topology.c
457
*dict = devm_kcalloc(comp->card->dev, *num_entries, entry_size, GFP_KERNEL);
sound/soc/intel/avs/topology.c
467
void *dict, u32 num_entries, size_t entry_size,
sound/soc/intel/avs/topology.c
474
for (i = 0; i < num_entries; i++) {
sound/soc/intel/avs/topology.c
500
void **dict, u32 *num_entries, size_t entry_size,
sound/soc/intel/avs/topology.c
506
ret = parse_dictionary_header(comp, tuples, dict, num_entries,
sound/soc/intel/avs/topology.c
516
*num_entries, entry_size,
sound/soc/intel/catpt/messages.c
33
u8 num_entries;
sound/soc/intel/catpt/messages.c
73
input.num_entries = num_modules;
sound/soc/intel/catpt/messages.h
186
u8 num_entries;
sound/soc/intel/catpt/pcm.c
1151
for (j = 0; j < template->num_entries; j++) {
sound/soc/intel/catpt/pcm.c
19
u8 num_entries;
sound/soc/intel/catpt/pcm.c
26
.num_entries = 1,
sound/soc/intel/catpt/pcm.c
33
.num_entries = 1,
sound/soc/intel/catpt/pcm.c
40
.num_entries = 1,
sound/soc/intel/catpt/pcm.c
405
stream->template->num_entries,
sound/soc/intel/catpt/pcm.c
47
.num_entries = 1,
sound/soc/intel/catpt/pcm.c
54
.num_entries = 1,
sound/soc/intel/catpt/pcm.c
61
.num_entries = 1,
sound/soc/sdca/sdca_functions.c
2100
int num_files, num_entries;
sound/soc/sdca/sdca_functions.c
2105
num_entries = fwnode_property_count_u32(function_node, fileset_name);
sound/soc/sdca/sdca_functions.c
2106
if (num_entries <= 0) {
sound/soc/sdca/sdca_functions.c
2108
function_node, filesets_list[i], num_entries);
sound/soc/sdca/sdca_functions.c
2110
} else if (num_entries % mult_fileset != 0) {
sound/soc/sdca/sdca_functions.c
2118
files = devm_kcalloc(dev, num_entries / mult_fileset,
sound/soc/sdca/sdca_functions.c
2123
u32 *fileset_entries __free(kfree) = kcalloc(num_entries, sizeof(u32),
sound/soc/sdca/sdca_functions.c
2129
fileset_entries, num_entries);
sound/soc/sdca/sdca_functions.c
2131
for (j = 0, num_files = 0; j < num_entries; num_files++) {
tools/perf/pmu-events/empty-pmu-events.c
17
uint32_t num_entries;
tools/perf/pmu-events/empty-pmu-events.c
2612
.num_entries = ARRAY_SIZE(pmu_events__common_default_core),
tools/perf/pmu-events/empty-pmu-events.c
2617
.num_entries = ARRAY_SIZE(pmu_events__common_software),
tools/perf/pmu-events/empty-pmu-events.c
2622
.num_entries = ARRAY_SIZE(pmu_events__common_tool),
tools/perf/pmu-events/empty-pmu-events.c
2651
.num_entries = ARRAY_SIZE(pmu_metrics__common_default_core),
tools/perf/pmu-events/empty-pmu-events.c
2686
.num_entries = ARRAY_SIZE(pmu_events__test_soc_cpu_default_core),
tools/perf/pmu-events/empty-pmu-events.c
2691
.num_entries = ARRAY_SIZE(pmu_events__test_soc_cpu_hisi_sccl_ddrc),
tools/perf/pmu-events/empty-pmu-events.c
2696
.num_entries = ARRAY_SIZE(pmu_events__test_soc_cpu_hisi_sccl_l3c),
tools/perf/pmu-events/empty-pmu-events.c
2701
.num_entries = ARRAY_SIZE(pmu_events__test_soc_cpu_uncore_cbox),
tools/perf/pmu-events/empty-pmu-events.c
2706
.num_entries = ARRAY_SIZE(pmu_events__test_soc_cpu_uncore_imc),
tools/perf/pmu-events/empty-pmu-events.c
2711
.num_entries = ARRAY_SIZE(pmu_events__test_soc_cpu_uncore_imc_free_running),
tools/perf/pmu-events/empty-pmu-events.c
2738
.num_entries = ARRAY_SIZE(pmu_metrics__test_soc_cpu_default_core),
tools/perf/pmu-events/empty-pmu-events.c
2757
.num_entries = ARRAY_SIZE(pmu_events__test_soc_sys_uncore_sys_ccn_pmu),
tools/perf/pmu-events/empty-pmu-events.c
2762
.num_entries = ARRAY_SIZE(pmu_events__test_soc_sys_uncore_sys_cmn_pmu),
tools/perf/pmu-events/empty-pmu-events.c
2767
.num_entries = ARRAY_SIZE(pmu_events__test_soc_sys_uncore_sys_ddr_pmu),
tools/perf/pmu-events/empty-pmu-events.c
2927
for (uint32_t i = 0; i < pmu->num_entries; i++) {
tools/perf/pmu-events/empty-pmu-events.c
2947
int low = 0, high = pmu->num_entries - 1;
tools/perf/pmu-events/empty-pmu-events.c
3038
count += table_pmu->num_entries;
tools/perf/pmu-events/empty-pmu-events.c
3053
for (uint32_t i = 0; i < pmu->num_entries; i++) {
tools/perf/pmu-events/empty-pmu-events.c
3073
int low = 0, high = pmu->num_entries - 1;
tools/perf/util/dwarf-regs-arch/dwarf-regs-x86.c
145
static int get_regnum(const struct dwarf_regs_idx *entries, size_t num_entries, const char *name)
tools/perf/util/dwarf-regs-arch/dwarf-regs-x86.c
151
for (size_t i = 0; i < num_entries; i++) {
tools/testing/cxl/test/mem.c
1527
out->num_entries = 0;
tools/testing/cxl/test/mem.c
1531
out->num_entries = cpu_to_le16(MAX_CXL_TEST_FEATS);