arch/arm/net/bpf_jit_32.c
1446
BUILD_BUG_ON(imm8m(offsetof(struct bpf_array, ptrs)) < 0);
arch/arm/net/bpf_jit_32.c
1447
off = imm8m(offsetof(struct bpf_array, ptrs));
arch/arm64/net/bpf_jit_comp.c
666
off = offsetof(struct bpf_array, ptrs);
arch/loongarch/net/bpf_jit.c
327
off = offsetof(struct bpf_array, ptrs);
arch/mips/net/bpf_jit_comp32.c
1323
off = offsetof(struct bpf_array, ptrs);
arch/mips/net/bpf_jit_comp64.c
505
off = offsetof(struct bpf_array, ptrs);
arch/parisc/net/bpf_jit_comp32.c
985
BUILD_BUG_ON(sizeof(bpfa.ptrs[0]) != 4);
arch/parisc/net/bpf_jit_comp32.c
987
off = offsetof(struct bpf_array, ptrs);
arch/parisc/net/bpf_jit_comp64.c
387
BUILD_BUG_ON(sizeof(bpfa.ptrs[0]) != 8);
arch/parisc/net/bpf_jit_comp64.c
389
off = offsetof(struct bpf_array, ptrs);
arch/powerpc/kernel/eeh_driver.c
906
void **ptrs = (void **) pe->stack_trace;
arch/powerpc/kernel/eeh_driver.c
915
pr_err("EEH: [%p] %pS\n", ptrs[i], ptrs[i]);
arch/powerpc/net/bpf_jit_comp32.c
258
EMIT(PPC_RAW_LWZ(_R3, _R3, offsetof(struct bpf_array, ptrs)));
arch/powerpc/net/bpf_jit_comp64.c
627
offsetof(struct bpf_array, ptrs)));
arch/riscv/net/bpf_jit_comp32.c
815
off = offsetof(struct bpf_array, ptrs);
arch/riscv/net/bpf_jit_comp64.c
390
off = offsetof(struct bpf_array, ptrs);
arch/s390/net/bpf_jit_comp.c
1892
REG_1, offsetof(struct bpf_array, ptrs));
arch/sparc/net/bpf_jit_comp_64.c
879
off = offsetof(struct bpf_array, ptrs);
arch/x86/net/bpf_jit_comp.c
764
offsetof(struct bpf_array, ptrs));
arch/x86/net/bpf_jit_comp.c
890
target = array->ptrs[poke->tail_call.key];
arch/x86/net/bpf_jit_comp32.c
1355
EMIT3_off32(0x8B, 0x94, 0x90, offsetof(struct bpf_array, ptrs));
crypto/async_tx/async_raid6_recov.c
412
void **ptrs = scribble ? scribble : (void **) blocks;
crypto/async_tx/async_raid6_recov.c
417
ptrs[i] = raid6_get_zero_page();
crypto/async_tx/async_raid6_recov.c
419
ptrs[i] = page_address(blocks[i]) + offs[i];
crypto/async_tx/async_raid6_recov.c
421
raid6_2data_recov(disks, bytes, faila, failb, ptrs);
crypto/async_tx/async_raid6_recov.c
495
void **ptrs = scribble ? scribble : (void **) blocks;
crypto/async_tx/async_raid6_recov.c
500
ptrs[i] = raid6_get_zero_page();
crypto/async_tx/async_raid6_recov.c
502
ptrs[i] = page_address(blocks[i]) + offs[i];
crypto/async_tx/async_raid6_recov.c
504
raid6_datap_recov(disks, bytes, faila, ptrs);
crypto/async_tx/raid6test.c
104
faila, ptrs, offs, &submit);
crypto/async_tx/raid6test.c
109
faila, failb, ptrs, offs, &submit);
crypto/async_tx/raid6test.c
114
tx = async_syndrome_val(ptrs, offs,
crypto/async_tx/raid6test.c
58
struct page **ptrs, unsigned int *offs)
crypto/async_tx/raid6test.c
72
tx = async_gen_syndrome(ptrs, offs,
crypto/async_tx/raid6test.c
88
blocks[count++] = ptrs[i];
crypto/async_tx/raid6test.c
90
dest = ptrs[faila];
crypto/async_tx/raid6test.c
96
tx = async_gen_syndrome(ptrs, offs,
drivers/gpu/drm/i915/display/intel_bios.c
208
const struct bdb_lfp_data_ptrs *ptrs;
drivers/gpu/drm/i915/display/intel_bios.c
211
ptrs = bdb_find_section(display, BDB_LFP_DATA_PTRS);
drivers/gpu/drm/i915/display/intel_bios.c
212
if (!ptrs)
drivers/gpu/drm/i915/display/intel_bios.c
216
if (ptrs->panel_name.table_size)
drivers/gpu/drm/i915/display/intel_bios.c
217
size = max(size, ptrs->panel_name.offset +
drivers/gpu/drm/i915/display/intel_bios.c
224
const struct bdb_lfp_data_ptrs *ptrs)
drivers/gpu/drm/i915/display/intel_bios.c
240
if (ptrs->num_entries != 3)
drivers/gpu/drm/i915/display/intel_bios.c
243
fp_timing_size = ptrs->ptr[0].fp_timing.table_size;
drivers/gpu/drm/i915/display/intel_bios.c
244
dvo_timing_size = ptrs->ptr[0].dvo_timing.table_size;
drivers/gpu/drm/i915/display/intel_bios.c
245
panel_pnp_id_size = ptrs->ptr[0].panel_pnp_id.table_size;
drivers/gpu/drm/i915/display/intel_bios.c
246
panel_name_size = ptrs->panel_name.table_size;
drivers/gpu/drm/i915/display/intel_bios.c
259
lfp_data_size = ptrs->ptr[1].fp_timing.offset - ptrs->ptr[0].fp_timing.offset;
drivers/gpu/drm/i915/display/intel_bios.c
265
if (ptrs->ptr[i].fp_timing.table_size != fp_timing_size ||
drivers/gpu/drm/i915/display/intel_bios.c
266
ptrs->ptr[i].dvo_timing.table_size != dvo_timing_size ||
drivers/gpu/drm/i915/display/intel_bios.c
267
ptrs->ptr[i].panel_pnp_id.table_size != panel_pnp_id_size)
drivers/gpu/drm/i915/display/intel_bios.c
270
if (ptrs->ptr[i].fp_timing.offset - ptrs->ptr[i-1].fp_timing.offset != lfp_data_size ||
drivers/gpu/drm/i915/display/intel_bios.c
271
ptrs->ptr[i].dvo_timing.offset - ptrs->ptr[i-1].dvo_timing.offset != lfp_data_size ||
drivers/gpu/drm/i915/display/intel_bios.c
272
ptrs->ptr[i].panel_pnp_id.offset - ptrs->ptr[i-1].panel_pnp_id.offset != lfp_data_size)
drivers/gpu/drm/i915/display/intel_bios.c
288
if (ptrs->ptr[0].fp_timing.offset + fp_timing_size != ptrs->ptr[0].dvo_timing.offset ||
drivers/gpu/drm/i915/display/intel_bios.c
289
ptrs->ptr[0].dvo_timing.offset + dvo_timing_size != ptrs->ptr[0].panel_pnp_id.offset ||
drivers/gpu/drm/i915/display/intel_bios.c
290
ptrs->ptr[0].panel_pnp_id.offset + panel_pnp_id_size != lfp_data_size)
drivers/gpu/drm/i915/display/intel_bios.c
295
if (ptrs->ptr[i].fp_timing.offset + fp_timing_size > data_block_size ||
drivers/gpu/drm/i915/display/intel_bios.c
296
ptrs->ptr[i].dvo_timing.offset + dvo_timing_size > data_block_size ||
drivers/gpu/drm/i915/display/intel_bios.c
297
ptrs->ptr[i].panel_pnp_id.offset + panel_pnp_id_size > data_block_size)
drivers/gpu/drm/i915/display/intel_bios.c
301
if (ptrs->panel_name.offset + 16 * panel_name_size > data_block_size)
drivers/gpu/drm/i915/display/intel_bios.c
306
const u16 *t = data_block + ptrs->ptr[i].fp_timing.offset +
drivers/gpu/drm/i915/display/intel_bios.c
319
struct bdb_lfp_data_ptrs *ptrs = ptrs_block;
drivers/gpu/drm/i915/display/intel_bios.c
326
if (ptrs->ptr[i].fp_timing.offset < offset ||
drivers/gpu/drm/i915/display/intel_bios.c
327
ptrs->ptr[i].dvo_timing.offset < offset ||
drivers/gpu/drm/i915/display/intel_bios.c
328
ptrs->ptr[i].panel_pnp_id.offset < offset)
drivers/gpu/drm/i915/display/intel_bios.c
331
ptrs->ptr[i].fp_timing.offset -= offset;
drivers/gpu/drm/i915/display/intel_bios.c
332
ptrs->ptr[i].dvo_timing.offset -= offset;
drivers/gpu/drm/i915/display/intel_bios.c
333
ptrs->ptr[i].panel_pnp_id.offset -= offset;
drivers/gpu/drm/i915/display/intel_bios.c
336
if (ptrs->panel_name.table_size) {
drivers/gpu/drm/i915/display/intel_bios.c
337
if (ptrs->panel_name.offset < offset)
drivers/gpu/drm/i915/display/intel_bios.c
340
ptrs->panel_name.offset -= offset;
drivers/gpu/drm/i915/display/intel_bios.c
343
return validate_lfp_data_ptrs(bdb, ptrs);
drivers/gpu/drm/i915/display/intel_bios.c
370
struct bdb_lfp_data_ptrs *ptrs;
drivers/gpu/drm/i915/display/intel_bios.c
398
ptrs_block = kzalloc(sizeof(*ptrs) + 3, GFP_KERNEL);
drivers/gpu/drm/i915/display/intel_bios.c
403
*(u16 *)(ptrs_block + 1) = sizeof(*ptrs);
drivers/gpu/drm/i915/display/intel_bios.c
404
ptrs = ptrs_block + 3;
drivers/gpu/drm/i915/display/intel_bios.c
407
size = make_lfp_data_ptr(&ptrs->ptr[0].panel_pnp_id, table_size, size);
drivers/gpu/drm/i915/display/intel_bios.c
410
size = make_lfp_data_ptr(&ptrs->ptr[0].dvo_timing, table_size, size);
drivers/gpu/drm/i915/display/intel_bios.c
413
size = make_lfp_data_ptr(&ptrs->ptr[0].fp_timing, table_size, size);
drivers/gpu/drm/i915/display/intel_bios.c
415
if (ptrs->ptr[0].fp_timing.table_size)
drivers/gpu/drm/i915/display/intel_bios.c
416
ptrs->num_entries++;
drivers/gpu/drm/i915/display/intel_bios.c
417
if (ptrs->ptr[0].dvo_timing.table_size)
drivers/gpu/drm/i915/display/intel_bios.c
418
ptrs->num_entries++;
drivers/gpu/drm/i915/display/intel_bios.c
419
if (ptrs->ptr[0].panel_pnp_id.table_size)
drivers/gpu/drm/i915/display/intel_bios.c
420
ptrs->num_entries++;
drivers/gpu/drm/i915/display/intel_bios.c
422
if (size != 0 || ptrs->num_entries != 3) {
drivers/gpu/drm/i915/display/intel_bios.c
430
next_lfp_data_ptr(&ptrs->ptr[i].fp_timing, &ptrs->ptr[i-1].fp_timing, size);
drivers/gpu/drm/i915/display/intel_bios.c
431
next_lfp_data_ptr(&ptrs->ptr[i].dvo_timing, &ptrs->ptr[i-1].dvo_timing, size);
drivers/gpu/drm/i915/display/intel_bios.c
432
next_lfp_data_ptr(&ptrs->ptr[i].panel_pnp_id, &ptrs->ptr[i-1].panel_pnp_id, size);
drivers/gpu/drm/i915/display/intel_bios.c
438
ptrs->panel_name.table_size = table_size;
drivers/gpu/drm/i915/display/intel_bios.c
439
ptrs->panel_name.offset = size * 16;
drivers/gpu/drm/i915/display/intel_bios.c
445
ptrs->ptr[i].fp_timing.offset += offset;
drivers/gpu/drm/i915/display/intel_bios.c
446
ptrs->ptr[i].dvo_timing.offset += offset;
drivers/gpu/drm/i915/display/intel_bios.c
447
ptrs->ptr[i].panel_pnp_id.offset += offset;
drivers/gpu/drm/i915/display/intel_bios.c
450
if (ptrs->panel_name.table_size)
drivers/gpu/drm/i915/display/intel_bios.c
451
ptrs->panel_name.offset += offset;
drivers/gpu/drm/i915/display/intel_bios.c
590
const struct bdb_lfp_data_ptrs *ptrs,
drivers/gpu/drm/i915/display/intel_bios.c
593
return (const void *)data + ptrs->ptr[index].dvo_timing.offset;
drivers/gpu/drm/i915/display/intel_bios.c
598
const struct bdb_lfp_data_ptrs *ptrs,
drivers/gpu/drm/i915/display/intel_bios.c
601
return (const void *)data + ptrs->ptr[index].fp_timing.offset;
drivers/gpu/drm/i915/display/intel_bios.c
606
const struct bdb_lfp_data_ptrs *ptrs,
drivers/gpu/drm/i915/display/intel_bios.c
612
return (const void *)data + ptrs->ptr[index].panel_pnp_id.offset;
drivers/gpu/drm/i915/display/intel_bios.c
617
const struct bdb_lfp_data_ptrs *ptrs)
drivers/gpu/drm/i915/display/intel_bios.c
619
if (ptrs->panel_name.table_size)
drivers/gpu/drm/i915/display/intel_bios.c
620
return (const void *)data + ptrs->panel_name.offset;
drivers/gpu/drm/i915/display/intel_bios.c
663
const struct bdb_lfp_data_ptrs *ptrs;
drivers/gpu/drm/i915/display/intel_bios.c
680
ptrs = bdb_find_section(display, BDB_LFP_DATA_PTRS);
drivers/gpu/drm/i915/display/intel_bios.c
681
if (!ptrs)
drivers/gpu/drm/i915/display/intel_bios.c
690
get_lfp_pnp_id(data, ptrs, i);
drivers/gpu/drm/i915/display/intel_bios.c
886
const struct bdb_lfp_data_ptrs *ptrs;
drivers/gpu/drm/i915/display/intel_bios.c
891
ptrs = bdb_find_section(display, BDB_LFP_DATA_PTRS);
drivers/gpu/drm/i915/display/intel_bios.c
892
if (!ptrs)
drivers/gpu/drm/i915/display/intel_bios.c
900
parse_lfp_panel_dtd(display, panel, data, ptrs);
drivers/gpu/drm/i915/display/intel_bios.c
902
pnp_id = get_lfp_pnp_id(data, ptrs, panel_type);
drivers/gpu/drm/i915/display/intel_bios.c
907
tail = get_lfp_data_tail(data, ptrs);
drivers/gpu/drm/nouveau/include/nvkm/core/gpuobj.h
13
const struct nvkm_gpuobj_func *ptrs;
drivers/gpu/drm/nouveau/include/nvkm/core/memory.h
24
const struct nvkm_memory_ptrs *ptrs;
drivers/gpu/drm/nouveau/include/nvkm/core/memory.h
76
#define nvkm_ro32(o,a) (o)->ptrs->rd32((o), (a))
drivers/gpu/drm/nouveau/include/nvkm/core/memory.h
77
#define nvkm_wo32(o,a,d) (o)->ptrs->wr32((o), (a), (d))
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c
395
node->base.base.memory.ptrs = &gk20a_instobj_ptrs;
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/gk20a.c
442
node->base.base.memory.ptrs = &gk20a_instobj_ptrs;
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv04.c
133
iobj->base.memory.ptrs = &nv04_instobj_ptrs;
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv40.c
132
iobj->base.memory.ptrs = &nv40_instobj_ptrs;
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
211
iobj->base.memory.ptrs = NULL;
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
253
iobj->base.memory.ptrs = &nv50_instobj_fast;
drivers/gpu/drm/nouveau/nvkm/subdev/instmem/nv50.c
255
iobj->base.memory.ptrs = &nv50_instobj_slow;
drivers/gpu/drm/xe/xe_guc_submit.c
2480
q->lrc[i]->ring.tail = job->ptrs[i].head;
drivers/gpu/drm/xe/xe_guc_submit.c
2541
q->lrc[i]->ring.tail = job->ptrs[i].head;
drivers/gpu/drm/xe/xe_ring_ops.c
442
i = emit_bb_start(job->ptrs[0].batch_addr, BIT(8), dw, i);
drivers/gpu/drm/xe/xe_ring_ops.c
448
i = emit_bb_start(job->ptrs[1].batch_addr, BIT(8), dw, i);
drivers/gpu/drm/xe/xe_ring_ops.c
468
job->ptrs[0].batch_addr,
drivers/gpu/drm/xe/xe_ring_ops.c
469
&job->ptrs[0].head,
drivers/gpu/drm/xe/xe_ring_ops.c
479
&job->ptrs[0].head,
drivers/gpu/drm/xe/xe_ring_ops.c
486
job->ptrs[i].batch_addr,
drivers/gpu/drm/xe/xe_ring_ops.c
487
&job->ptrs[i].head,
drivers/gpu/drm/xe/xe_ring_ops.c
498
job->ptrs[i].batch_addr,
drivers/gpu/drm/xe/xe_ring_ops.c
499
&job->ptrs[i].head,
drivers/gpu/drm/xe/xe_ring_ops.c
509
job->ptrs[i].batch_addr,
drivers/gpu/drm/xe/xe_ring_ops.c
510
&job->ptrs[i].head,
drivers/gpu/drm/xe/xe_sched_job.c
130
job->ptrs[i].lrc_fence = fence;
drivers/gpu/drm/xe/xe_sched_job.c
140
job->ptrs[i].chain_fence = chain;
drivers/gpu/drm/xe/xe_sched_job.c
148
job->ptrs[i].batch_addr = batch_addr[i];
drivers/gpu/drm/xe/xe_sched_job.c
273
fence = job->ptrs[i].lrc_fence;
drivers/gpu/drm/xe/xe_sched_job.c
275
job->ptrs[i].lrc_fence = NULL;
drivers/gpu/drm/xe/xe_sched_job.c
283
chain = job->ptrs[i - 1].chain_fence;
drivers/gpu/drm/xe/xe_sched_job.c
285
job->ptrs[i - 1].chain_fence = NULL;
drivers/gpu/drm/xe/xe_sched_job.c
333
xe_device_uncanonicalize_addr(xe, job->ptrs[i].batch_addr);
drivers/gpu/drm/xe/xe_sched_job.c
88
struct xe_job_ptrs *ptrs = &job->ptrs[i];
drivers/gpu/drm/xe/xe_sched_job.c
90
if (ptrs->lrc_fence)
drivers/gpu/drm/xe/xe_sched_job.c
91
xe_lrc_free_seqno_fence(ptrs->lrc_fence);
drivers/gpu/drm/xe/xe_sched_job.c
92
dma_fence_chain_free(ptrs->chain_fence);
drivers/gpu/drm/xe/xe_sched_job_types.h
73
struct xe_job_ptrs ptrs[];
drivers/gpu/drm/xe/xe_trace.h
259
__entry->batch_addr = (u64)job->ptrs[0].batch_addr;
drivers/media/common/videobuf2/frame_vector.c
144
pages = (struct page **)(vec->ptrs);
drivers/media/common/videobuf2/frame_vector.c
162
int size = struct_size(vec, ptrs, nr_frames);
drivers/media/common/videobuf2/frame_vector.c
58
(struct page **)(vec->ptrs));
drivers/net/ethernet/8390/apne.c
383
short *ptrs;
drivers/net/ethernet/8390/apne.c
403
ptrs = (short*)hdr;
drivers/net/ethernet/8390/apne.c
405
*ptrs++ = inw(NE_BASE + NE_DATAPORT);
drivers/net/ethernet/8390/apne.c
429
short *ptrs;
drivers/net/ethernet/8390/apne.c
448
ptrs = (short*)buf;
drivers/net/ethernet/8390/apne.c
450
*ptrs++ = inw(NE_BASE + NE_DATAPORT);
drivers/net/ethernet/8390/apne.c
471
short *ptrs;
drivers/net/ethernet/8390/apne.c
501
ptrs = (short*)buf;
drivers/net/ethernet/8390/apne.c
503
outw(*ptrs++, NE_BASE+NE_DATAPORT);
drivers/net/ethernet/8390/hydra.c
205
short *ptrs;
drivers/net/ethernet/8390/hydra.c
208
ptrs = (short *)hdr;
drivers/net/ethernet/8390/hydra.c
210
*(ptrs++) = z_readw(hdr_start);
drivers/net/ethernet/8390/hydra.c
213
*(ptrs++) = z_readw(hdr_start);
drivers/net/ethernet/8390/zorro8390.c
111
short *ptrs;
drivers/net/ethernet/8390/zorro8390.c
132
ptrs = (short *)hdr;
drivers/net/ethernet/8390/zorro8390.c
134
*ptrs++ = z_readw(NE_BASE + NE_DATAPORT);
drivers/net/ethernet/8390/zorro8390.c
154
short *ptrs;
drivers/net/ethernet/8390/zorro8390.c
173
ptrs = (short *)buf;
drivers/net/ethernet/8390/zorro8390.c
175
*ptrs++ = z_readw(NE_BASE + NE_DATAPORT);
drivers/net/ethernet/8390/zorro8390.c
189
short *ptrs;
drivers/net/ethernet/8390/zorro8390.c
220
ptrs = (short *)buf;
drivers/net/ethernet/8390/zorro8390.c
222
z_writew(*ptrs++, NE_BASE + NE_DATAPORT);
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
128
u64 ptrs[NPA_MAX_BURST];
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
139
__cn10k_aura_freeptr(pfvf, cq->cq_idx, ptrs,
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
144
ptrs[num_ptrs] = pool->xsk_pool ?
drivers/net/ethernet/marvell/octeontx2/nic/cn10k.c
149
__cn10k_aura_freeptr(pfvf, cq->cq_idx, ptrs,
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.h
782
u64 *ptrs, u64 num_ptrs)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.h
802
ptrs[0] = (count_eot << 32) | (aura & 0xFFFFF);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.h
815
memcpy((u64 *)lmt_info->lmt_addr, ptrs, sizeof(u64) * num_ptrs);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.h
823
u64 ptrs[2];
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.h
825
ptrs[1] = buf;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.h
828
__cn10k_aura_freeptr(pfvf, aura, ptrs, 2);
drivers/platform/goldfish/goldfish_pipe.c
306
command->rw_params.ptrs[0] = (u64)(xaddr | (address & ~PAGE_MASK));
drivers/platform/goldfish/goldfish_pipe.c
316
command->rw_params.ptrs[buffer_idx] = (u64)xaddr;
drivers/platform/goldfish/goldfish_pipe.c
98
u64 ptrs[MAX_BUFFERS_PER_COMMAND];
drivers/tty/mips_ejtag_fdc.c
216
static struct fdc_word mips_ejtag_fdc_encode(const u8 **ptrs,
drivers/tty/mips_ejtag_fdc.c
221
const u8 **ptrs_end = ptrs + ranges;
drivers/tty/mips_ejtag_fdc.c
223
for (; ptrs < ptrs_end; ++ptrs) {
drivers/tty/mips_ejtag_fdc.c
224
const u8 *ptr = *(ptrs++);
drivers/tty/mips_ejtag_fdc.c
420
const u8 *ptrs[2];
drivers/tty/mips_ejtag_fdc.c
428
ptrs[0] = dport->port.xmit_buf + dport->xmit_tail;
drivers/tty/mips_ejtag_fdc.c
432
ptrs[1] = dport->port.xmit_buf;
drivers/tty/mips_ejtag_fdc.c
434
word = mips_ejtag_fdc_encode(ptrs, sizes, 1 + !!sizes[1]);
drivers/tty/mips_ejtag_fdc.c
438
min_t(int, word.bytes, sizes[0]), ptrs[0],
drivers/tty/mips_ejtag_fdc.c
439
max_t(int, 0, word.bytes - sizes[0]), ptrs[1]);
fs/btrfs/accessors.h
399
ptr = offsetof(struct btrfs_node, ptrs) +
fs/btrfs/accessors.h
409
ptr = offsetof(struct btrfs_node, ptrs) +
fs/btrfs/accessors.h
418
ptr = offsetof(struct btrfs_node, ptrs) +
fs/btrfs/accessors.h
428
ptr = offsetof(struct btrfs_node, ptrs) +
fs/btrfs/accessors.h
435
return offsetof(struct btrfs_node, ptrs) +
fs/btrfs/ctree.c
759
p = offsetof(struct btrfs_node, ptrs);
fs/ext2/inode.c
166
int ptrs = EXT2_ADDR_PER_BLOCK(inode->i_sb);
fs/ext2/inode.c
169
indirect_blocks = ptrs,
fs/ext2/inode.c
183
final = ptrs;
fs/ext2/inode.c
187
offsets[n++] = i_block & (ptrs - 1);
fs/ext2/inode.c
188
final = ptrs;
fs/ext2/inode.c
189
} else if (((i_block -= double_blocks) >> (ptrs_bits * 2)) < ptrs) {
fs/ext2/inode.c
192
offsets[n++] = (i_block >> ptrs_bits) & (ptrs - 1);
fs/ext2/inode.c
193
offsets[n++] = i_block & (ptrs - 1);
fs/ext2/inode.c
194
final = ptrs;
fs/ext2/inode.c
200
*boundary = final - 1 - (i_block & (ptrs - 1));
fs/ext4/indirect.c
101
offsets[n++] = (i_block >> ptrs_bits) & (ptrs - 1);
fs/ext4/indirect.c
102
offsets[n++] = i_block & (ptrs - 1);
fs/ext4/indirect.c
103
final = ptrs;
fs/ext4/indirect.c
110
*boundary = final - 1 - (i_block & (ptrs - 1));
fs/ext4/indirect.c
78
int ptrs = EXT4_ADDR_PER_BLOCK(inode->i_sb);
fs/ext4/indirect.c
81
indirect_blocks = ptrs,
fs/ext4/indirect.c
92
final = ptrs;
fs/ext4/indirect.c
96
offsets[n++] = i_block & (ptrs - 1);
fs/ext4/indirect.c
97
final = ptrs;
fs/ext4/indirect.c
98
} else if (((i_block -= double_blocks) >> (ptrs_bits * 2)) < ptrs) {
fs/gfs2/bmap.c
492
unsigned int ptrs;
fs/gfs2/bmap.c
496
ptrs = (hgt >= 1 ? sdp->sd_inptrs : sdp->sd_diptrs) - start;
fs/gfs2/bmap.c
497
len = ptrs * factor;
fs/gfs2/bmap.c
499
ptrs = DIV_ROUND_UP_ULL(max_len, factor);
fs/gfs2/bmap.c
500
status = walker(mp, ptrs);
fs/gfs2/bmap.c
506
ptrs = mp->mp_list[hgt] - start;
fs/gfs2/bmap.c
507
len = ptrs * factor;
fs/gfs2/bmap.c
552
unsigned int ptrs)
fs/gfs2/bmap.c
559
end = start + ptrs;
fs/nilfs2/bmap.c
106
__u64 ptrs[NILFS_BMAP_SMALL_HIGH + 1];
fs/nilfs2/bmap.c
113
bmap, keys, ptrs, NILFS_BMAP_SMALL_HIGH + 1);
fs/nilfs2/bmap.c
117
bmap, key, ptr, keys, ptrs, n);
fs/nilfs2/bmap.c
158
__u64 ptrs[NILFS_BMAP_LARGE_LOW + 1];
fs/nilfs2/bmap.c
165
bmap, keys, ptrs, NILFS_BMAP_LARGE_LOW + 1);
fs/nilfs2/bmap.c
169
bmap, key, keys, ptrs, n);
fs/nilfs2/btree.c
164
const __u64 *keys, const __u64 *ptrs)
fs/nilfs2/btree.c
1694
__u64 *keys, __u64 *ptrs, int nitems)
fs/nilfs2/btree.c
1733
ptrs[i] = le64_to_cpu(dptrs[i]);
fs/nilfs2/btree.c
178
dptrs[i] = cpu_to_le64(ptrs[i]);
fs/nilfs2/btree.c
1801
const __u64 *keys, const __u64 *ptrs,
fs/nilfs2/btree.c
1829
nilfs_btree_node_init(node, 0, 1, n, ncblk, keys, ptrs);
fs/nilfs2/btree.c
1851
keys, ptrs);
fs/nilfs2/btree.c
1882
const __u64 *keys, const __u64 *ptrs, int n)
fs/nilfs2/btree.c
1906
nilfs_btree_commit_convert_and_insert(btree, key, ptr, keys, ptrs, n,
fs/nilfs2/direct.c
208
__u64 *keys, __u64 *ptrs, int nitems)
fs/nilfs2/direct.c
221
ptrs[n] = ptr;
fs/nilfs2/direct.c
229
__u64 key, __u64 *keys, __u64 *ptrs, int n)
fs/nilfs2/direct.c
250
cpu_to_le64(ptrs[j]) :
fs/ufs/inode.c
50
int ptrs = uspi->s_apb;
fs/ufs/inode.c
53
indirect_blocks = ptrs,
fs/ufs/inode.c
58
UFSD("ptrs=uspi->s_apb = %d,double_blocks=%ld \n",ptrs,double_blocks);
fs/ufs/inode.c
67
offsets[n++] = i_block & (ptrs - 1);
fs/ufs/inode.c
68
} else if (((i_block -= double_blocks) >> (ptrs_bits * 2)) < ptrs) {
fs/ufs/inode.c
71
offsets[n++] = (i_block >> ptrs_bits) & (ptrs - 1);
fs/ufs/inode.c
72
offsets[n++] = i_block & (ptrs - 1);
fs/xfs/libxfs/xfs_iext_tree.c
1051
xfs_iext_destroy_node(node->ptrs[i], level - 1);
fs/xfs/libxfs/xfs_iext_tree.c
119
void *ptrs[KEYS_PER_NODE];
fs/xfs/libxfs/xfs_iext_tree.c
168
node = node->ptrs[0];
fs/xfs/libxfs/xfs_iext_tree.c
187
if (!node->ptrs[i])
fs/xfs/libxfs/xfs_iext_tree.c
189
node = node->ptrs[i - 1];
fs/xfs/libxfs/xfs_iext_tree.c
320
node = node->ptrs[i - 1];
fs/xfs/libxfs/xfs_iext_tree.c
415
node->ptrs[0] = prev;
fs/xfs/libxfs/xfs_iext_tree.c
422
node->ptrs[0] = prev;
fs/xfs/libxfs/xfs_iext_tree.c
450
node = node->ptrs[i - 1];
fs/xfs/libxfs/xfs_iext_tree.c
480
new->ptrs[i] = node->ptrs[nr_keep + i];
fs/xfs/libxfs/xfs_iext_tree.c
483
node->ptrs[nr_keep + i] = NULL;
fs/xfs/libxfs/xfs_iext_tree.c
533
node->ptrs[i] = node->ptrs[i - 1];
fs/xfs/libxfs/xfs_iext_tree.c
536
node->ptrs[pos] = ptr;
fs/xfs/libxfs/xfs_iext_tree.c
704
struct xfs_iext_node *prev = parent->ptrs[*pos - 1];
fs/xfs/libxfs/xfs_iext_tree.c
710
prev->ptrs[nr_prev + i] = node->ptrs[i];
fs/xfs/libxfs/xfs_iext_tree.c
717
struct xfs_iext_node *next = parent->ptrs[*pos + 1];
fs/xfs/libxfs/xfs_iext_tree.c
728
node->ptrs[nr_entries + i] = next->ptrs[i];
fs/xfs/libxfs/xfs_iext_tree.c
752
ASSERT(node->ptrs[pos]);
fs/xfs/libxfs/xfs_iext_tree.c
753
ASSERT(node->ptrs[pos] == victim);
fs/xfs/libxfs/xfs_iext_tree.c
760
node->ptrs[i] = node->ptrs[i + 1];
fs/xfs/libxfs/xfs_iext_tree.c
763
node->ptrs[nr_entries] = NULL;
fs/xfs/libxfs/xfs_iext_tree.c
784
ASSERT(parent->ptrs[pos] == node);
fs/xfs/libxfs/xfs_iext_tree.c
798
ifp->if_data = node->ptrs[0];
include/linux/async_tx.h
194
struct page **ptrs, unsigned int *offs,
include/linux/async_tx.h
199
struct page **ptrs, unsigned int *offs,
include/linux/bpf.h
2248
DECLARE_FLEX_ARRAY(void *, ptrs) __aligned(8);
include/linux/raid/pq.h
165
void **ptrs);
include/linux/raid/pq.h
167
void **ptrs);
include/linux/raid/pq.h
169
void **ptrs);
include/media/frame_vector.h
11
void *ptrs[]; /* Array of pinned pfns / pages. Use
include/media/frame_vector.h
37
return (struct page **)(vec->ptrs);
include/media/frame_vector.h
44
return (unsigned long *)(vec->ptrs);
include/uapi/linux/btrfs_tree.h
585
struct btrfs_key_ptr ptrs[];
kernel/bpf/arraymap.c
1308
ee = READ_ONCE(array->ptrs[i]);
kernel/bpf/arraymap.c
869
BUG_ON(array->ptrs[i] != NULL);
kernel/bpf/arraymap.c
920
old_ptr = xchg(array->ptrs + index, new_ptr);
kernel/bpf/arraymap.c
924
old_ptr = xchg(array->ptrs + index, new_ptr);
kernel/bpf/arraymap.c
943
old_ptr = xchg(array->ptrs + index, NULL);
kernel/bpf/arraymap.c
947
old_ptr = xchg(array->ptrs + index, NULL);
kernel/bpf/core.c
2092
prog = READ_ONCE(array->ptrs[index]);
kernel/bpf/helpers.c
2706
cgrp = READ_ONCE(array->ptrs[idx]);
kernel/bpf/reuseport_array.c
127
sk = rcu_dereference(array->ptrs[i]);
kernel/bpf/reuseport_array.c
13
struct sock __rcu *ptrs[];
kernel/bpf/reuseport_array.c
137
RCU_INIT_POINTER(array->ptrs[i], NULL);
kernel/bpf/reuseport_array.c
155
array = bpf_map_area_alloc(struct_size(array, ptrs, attr->max_entries), numa_node);
kernel/bpf/reuseport_array.c
271
rcu_access_pointer(array->ptrs[index]),
kernel/bpf/reuseport_array.c
285
osk = rcu_dereference_protected(array->ptrs[index],
kernel/bpf/reuseport_array.c
293
sk_user_data = (uintptr_t)&array->ptrs[index] | SK_USER_DATA_NOCOPY |
kernel/bpf/reuseport_array.c
296
rcu_assign_pointer(array->ptrs[index], nsk);
kernel/bpf/reuseport_array.c
339
return struct_size(array, ptrs, map->max_entries);
kernel/bpf/reuseport_array.c
58
return rcu_dereference(array->ptrs[index]);
kernel/bpf/reuseport_array.c
72
if (!rcu_access_pointer(array->ptrs[index]))
kernel/bpf/reuseport_array.c
77
sk = rcu_dereference_protected(array->ptrs[index],
kernel/bpf/reuseport_array.c
82
RCU_INIT_POINTER(array->ptrs[index], NULL);
kernel/trace/bpf_trace.c
552
ee = READ_ONCE(array->ptrs[index]);
kernel/trace/bpf_trace.c
630
ee = READ_ONCE(array->ptrs[index]);
lib/raid6/avx2.c
141
static void raid6_avx22_gen_syndrome(int disks, size_t bytes, void **ptrs)
lib/raid6/avx2.c
143
u8 **dptr = (u8 **)ptrs;
lib/raid6/avx2.c
193
size_t bytes, void **ptrs)
lib/raid6/avx2.c
195
u8 **dptr = (u8 **)ptrs;
lib/raid6/avx2.c
273
static void raid6_avx24_gen_syndrome(int disks, size_t bytes, void **ptrs)
lib/raid6/avx2.c
275
u8 **dptr = (u8 **)ptrs;
lib/raid6/avx2.c
34
static void raid6_avx21_gen_syndrome(int disks, size_t bytes, void **ptrs)
lib/raid6/avx2.c
354
size_t bytes, void **ptrs)
lib/raid6/avx2.c
356
u8 **dptr = (u8 **)ptrs;
lib/raid6/avx2.c
36
u8 **dptr = (u8 **)ptrs;
lib/raid6/avx2.c
83
size_t bytes, void **ptrs)
lib/raid6/avx2.c
85
u8 **dptr = (u8 **)ptrs;
lib/raid6/avx512.c
100
size_t bytes, void **ptrs)
lib/raid6/avx512.c
102
u8 **dptr = (u8 **)ptrs;
lib/raid6/avx512.c
169
static void raid6_avx5122_gen_syndrome(int disks, size_t bytes, void **ptrs)
lib/raid6/avx512.c
171
u8 **dptr = (u8 **)ptrs;
lib/raid6/avx512.c
232
size_t bytes, void **ptrs)
lib/raid6/avx512.c
234
u8 **dptr = (u8 **)ptrs;
lib/raid6/avx512.c
328
static void raid6_avx5124_gen_syndrome(int disks, size_t bytes, void **ptrs)
lib/raid6/avx512.c
330
u8 **dptr = (u8 **)ptrs;
lib/raid6/avx512.c
42
static void raid6_avx5121_gen_syndrome(int disks, size_t bytes, void **ptrs)
lib/raid6/avx512.c
422
size_t bytes, void **ptrs)
lib/raid6/avx512.c
424
u8 **dptr = (u8 **)ptrs;
lib/raid6/avx512.c
44
u8 **dptr = (u8 **)ptrs;
lib/raid6/loongarch_simd.c
115
size_t bytes, void **ptrs)
lib/raid6/loongarch_simd.c
117
u8 **dptr = (u8 **)ptrs;
lib/raid6/loongarch_simd.c
264
static void raid6_lasx_gen_syndrome(int disks, size_t bytes, void **ptrs)
lib/raid6/loongarch_simd.c
266
u8 **dptr = (u8 **)ptrs;
lib/raid6/loongarch_simd.c
324
size_t bytes, void **ptrs)
lib/raid6/loongarch_simd.c
326
u8 **dptr = (u8 **)ptrs;
lib/raid6/loongarch_simd.c
33
static void raid6_lsx_gen_syndrome(int disks, size_t bytes, void **ptrs)
lib/raid6/loongarch_simd.c
35
u8 **dptr = (u8 **)ptrs;
lib/raid6/mmx.c
35
static void raid6_mmx1_gen_syndrome(int disks, size_t bytes, void **ptrs)
lib/raid6/mmx.c
37
u8 **dptr = (u8 **)ptrs;
lib/raid6/mmx.c
83
static void raid6_mmx2_gen_syndrome(int disks, size_t bytes, void **ptrs)
lib/raid6/mmx.c
85
u8 **dptr = (u8 **)ptrs;
lib/raid6/neon.c
30
size_t bytes, void **ptrs) \
lib/raid6/neon.c
36
(unsigned long)bytes, ptrs); \
lib/raid6/neon.c
40
size_t bytes, void **ptrs) \
lib/raid6/neon.c
46
start, stop, (unsigned long)bytes, ptrs);\
lib/raid6/neon.h
11
unsigned long bytes, void **ptrs);
lib/raid6/neon.h
12
void raid6_neon8_gen_syndrome_real(int disks, unsigned long bytes, void **ptrs);
lib/raid6/neon.h
14
unsigned long bytes, void **ptrs);
lib/raid6/neon.h
3
void raid6_neon1_gen_syndrome_real(int disks, unsigned long bytes, void **ptrs);
lib/raid6/neon.h
5
unsigned long bytes, void **ptrs);
lib/raid6/neon.h
6
void raid6_neon2_gen_syndrome_real(int disks, unsigned long bytes, void **ptrs);
lib/raid6/neon.h
8
unsigned long bytes, void **ptrs);
lib/raid6/neon.h
9
void raid6_neon4_gen_syndrome_real(int disks, unsigned long bytes, void **ptrs);
lib/raid6/recov.c
107
void raid6_dual_recov(int disks, size_t bytes, int faila, int failb, void **ptrs)
lib/raid6/recov.c
118
raid6_call.gen_syndrome(disks, bytes, ptrs);
lib/raid6/recov.c
127
raid6_datap_recov(disks, bytes, faila, ptrs);
lib/raid6/recov.c
130
raid6_2data_recov(disks, bytes, faila, failb, ptrs);
lib/raid6/recov.c
20
int failb, void **ptrs)
lib/raid6/recov.c
27
p = (u8 *)ptrs[disks-2];
lib/raid6/recov.c
28
q = (u8 *)ptrs[disks-1];
lib/raid6/recov.c
33
dp = (u8 *)ptrs[faila];
lib/raid6/recov.c
34
ptrs[faila] = raid6_get_zero_page();
lib/raid6/recov.c
35
ptrs[disks-2] = dp;
lib/raid6/recov.c
36
dq = (u8 *)ptrs[failb];
lib/raid6/recov.c
37
ptrs[failb] = raid6_get_zero_page();
lib/raid6/recov.c
38
ptrs[disks-1] = dq;
lib/raid6/recov.c
40
raid6_call.gen_syndrome(disks, bytes, ptrs);
lib/raid6/recov.c
43
ptrs[faila] = dp;
lib/raid6/recov.c
44
ptrs[failb] = dq;
lib/raid6/recov.c
45
ptrs[disks-2] = p;
lib/raid6/recov.c
46
ptrs[disks-1] = q;
lib/raid6/recov.c
64
void **ptrs)
lib/raid6/recov.c
69
p = (u8 *)ptrs[disks-2];
lib/raid6/recov.c
70
q = (u8 *)ptrs[disks-1];
lib/raid6/recov.c
74
dq = (u8 *)ptrs[faila];
lib/raid6/recov.c
75
ptrs[faila] = raid6_get_zero_page();
lib/raid6/recov.c
76
ptrs[disks-1] = dq;
lib/raid6/recov.c
78
raid6_call.gen_syndrome(disks, bytes, ptrs);
lib/raid6/recov.c
81
ptrs[faila] = dq;
lib/raid6/recov.c
82
ptrs[disks-1] = q;
lib/raid6/recov_avx2.c
17
int failb, void **ptrs)
lib/raid6/recov_avx2.c
187
void **ptrs)
lib/raid6/recov_avx2.c
193
p = (u8 *)ptrs[disks-2];
lib/raid6/recov_avx2.c
194
q = (u8 *)ptrs[disks-1];
lib/raid6/recov_avx2.c
198
dq = (u8 *)ptrs[faila];
lib/raid6/recov_avx2.c
199
ptrs[faila] = raid6_get_zero_page();
lib/raid6/recov_avx2.c
200
ptrs[disks-1] = dq;
lib/raid6/recov_avx2.c
202
raid6_call.gen_syndrome(disks, bytes, ptrs);
lib/raid6/recov_avx2.c
205
ptrs[faila] = dq;
lib/raid6/recov_avx2.c
206
ptrs[disks-1] = q;
lib/raid6/recov_avx2.c
24
p = (u8 *)ptrs[disks-2];
lib/raid6/recov_avx2.c
25
q = (u8 *)ptrs[disks-1];
lib/raid6/recov_avx2.c
30
dp = (u8 *)ptrs[faila];
lib/raid6/recov_avx2.c
31
ptrs[faila] = raid6_get_zero_page();
lib/raid6/recov_avx2.c
32
ptrs[disks-2] = dp;
lib/raid6/recov_avx2.c
33
dq = (u8 *)ptrs[failb];
lib/raid6/recov_avx2.c
34
ptrs[failb] = raid6_get_zero_page();
lib/raid6/recov_avx2.c
35
ptrs[disks-1] = dq;
lib/raid6/recov_avx2.c
37
raid6_call.gen_syndrome(disks, bytes, ptrs);
lib/raid6/recov_avx2.c
40
ptrs[faila] = dp;
lib/raid6/recov_avx2.c
41
ptrs[failb] = dq;
lib/raid6/recov_avx2.c
42
ptrs[disks-2] = p;
lib/raid6/recov_avx2.c
43
ptrs[disks-1] = q;
lib/raid6/recov_avx512.c
226
void **ptrs)
lib/raid6/recov_avx512.c
23
int failb, void **ptrs)
lib/raid6/recov_avx512.c
232
p = (u8 *)ptrs[disks-2];
lib/raid6/recov_avx512.c
233
q = (u8 *)ptrs[disks-1];
lib/raid6/recov_avx512.c
240
dq = (u8 *)ptrs[faila];
lib/raid6/recov_avx512.c
241
ptrs[faila] = raid6_get_zero_page();
lib/raid6/recov_avx512.c
242
ptrs[disks-1] = dq;
lib/raid6/recov_avx512.c
244
raid6_call.gen_syndrome(disks, bytes, ptrs);
lib/raid6/recov_avx512.c
247
ptrs[faila] = dq;
lib/raid6/recov_avx512.c
248
ptrs[disks-1] = q;
lib/raid6/recov_avx512.c
30
p = (u8 *)ptrs[disks-2];
lib/raid6/recov_avx512.c
31
q = (u8 *)ptrs[disks-1];
lib/raid6/recov_avx512.c
39
dp = (u8 *)ptrs[faila];
lib/raid6/recov_avx512.c
40
ptrs[faila] = raid6_get_zero_page();
lib/raid6/recov_avx512.c
41
ptrs[disks-2] = dp;
lib/raid6/recov_avx512.c
42
dq = (u8 *)ptrs[failb];
lib/raid6/recov_avx512.c
43
ptrs[failb] = raid6_get_zero_page();
lib/raid6/recov_avx512.c
44
ptrs[disks-1] = dq;
lib/raid6/recov_avx512.c
46
raid6_call.gen_syndrome(disks, bytes, ptrs);
lib/raid6/recov_avx512.c
49
ptrs[faila] = dp;
lib/raid6/recov_avx512.c
50
ptrs[failb] = dq;
lib/raid6/recov_avx512.c
51
ptrs[disks-2] = p;
lib/raid6/recov_avx512.c
52
ptrs[disks-1] = q;
lib/raid6/recov_loongarch_simd.c
187
void **ptrs)
lib/raid6/recov_loongarch_simd.c
192
p = (u8 *)ptrs[disks - 2];
lib/raid6/recov_loongarch_simd.c
193
q = (u8 *)ptrs[disks - 1];
lib/raid6/recov_loongarch_simd.c
199
dq = (u8 *)ptrs[faila];
lib/raid6/recov_loongarch_simd.c
200
ptrs[faila] = raid6_get_zero_page();
lib/raid6/recov_loongarch_simd.c
201
ptrs[disks - 1] = dq;
lib/raid6/recov_loongarch_simd.c
203
raid6_call.gen_syndrome(disks, bytes, ptrs);
lib/raid6/recov_loongarch_simd.c
206
ptrs[faila] = dq;
lib/raid6/recov_loongarch_simd.c
207
ptrs[disks - 1] = q;
lib/raid6/recov_loongarch_simd.c
30
int failb, void **ptrs)
lib/raid6/recov_loongarch_simd.c
304
int failb, void **ptrs)
lib/raid6/recov_loongarch_simd.c
310
p = (u8 *)ptrs[disks - 2];
lib/raid6/recov_loongarch_simd.c
311
q = (u8 *)ptrs[disks - 1];
lib/raid6/recov_loongarch_simd.c
318
dp = (u8 *)ptrs[faila];
lib/raid6/recov_loongarch_simd.c
319
ptrs[faila] = raid6_get_zero_page();
lib/raid6/recov_loongarch_simd.c
320
ptrs[disks - 2] = dp;
lib/raid6/recov_loongarch_simd.c
321
dq = (u8 *)ptrs[failb];
lib/raid6/recov_loongarch_simd.c
322
ptrs[failb] = raid6_get_zero_page();
lib/raid6/recov_loongarch_simd.c
323
ptrs[disks - 1] = dq;
lib/raid6/recov_loongarch_simd.c
325
raid6_call.gen_syndrome(disks, bytes, ptrs);
lib/raid6/recov_loongarch_simd.c
328
ptrs[faila] = dp;
lib/raid6/recov_loongarch_simd.c
329
ptrs[failb] = dq;
lib/raid6/recov_loongarch_simd.c
330
ptrs[disks - 2] = p;
lib/raid6/recov_loongarch_simd.c
331
ptrs[disks - 1] = q;
lib/raid6/recov_loongarch_simd.c
36
p = (u8 *)ptrs[disks - 2];
lib/raid6/recov_loongarch_simd.c
37
q = (u8 *)ptrs[disks - 1];
lib/raid6/recov_loongarch_simd.c
426
void **ptrs)
lib/raid6/recov_loongarch_simd.c
431
p = (u8 *)ptrs[disks - 2];
lib/raid6/recov_loongarch_simd.c
432
q = (u8 *)ptrs[disks - 1];
lib/raid6/recov_loongarch_simd.c
438
dq = (u8 *)ptrs[faila];
lib/raid6/recov_loongarch_simd.c
439
ptrs[faila] = raid6_get_zero_page();
lib/raid6/recov_loongarch_simd.c
44
dp = (u8 *)ptrs[faila];
lib/raid6/recov_loongarch_simd.c
440
ptrs[disks - 1] = dq;
lib/raid6/recov_loongarch_simd.c
442
raid6_call.gen_syndrome(disks, bytes, ptrs);
lib/raid6/recov_loongarch_simd.c
445
ptrs[faila] = dq;
lib/raid6/recov_loongarch_simd.c
446
ptrs[disks - 1] = q;
lib/raid6/recov_loongarch_simd.c
45
ptrs[faila] = raid6_get_zero_page();
lib/raid6/recov_loongarch_simd.c
46
ptrs[disks - 2] = dp;
lib/raid6/recov_loongarch_simd.c
47
dq = (u8 *)ptrs[failb];
lib/raid6/recov_loongarch_simd.c
48
ptrs[failb] = raid6_get_zero_page();
lib/raid6/recov_loongarch_simd.c
49
ptrs[disks - 1] = dq;
lib/raid6/recov_loongarch_simd.c
51
raid6_call.gen_syndrome(disks, bytes, ptrs);
lib/raid6/recov_loongarch_simd.c
54
ptrs[faila] = dp;
lib/raid6/recov_loongarch_simd.c
55
ptrs[failb] = dq;
lib/raid6/recov_loongarch_simd.c
56
ptrs[disks - 2] = p;
lib/raid6/recov_loongarch_simd.c
57
ptrs[disks - 1] = q;
lib/raid6/recov_neon.c
23
int failb, void **ptrs)
lib/raid6/recov_neon.c
29
p = (u8 *)ptrs[disks - 2];
lib/raid6/recov_neon.c
30
q = (u8 *)ptrs[disks - 1];
lib/raid6/recov_neon.c
37
dp = (u8 *)ptrs[faila];
lib/raid6/recov_neon.c
38
ptrs[faila] = raid6_get_zero_page();
lib/raid6/recov_neon.c
39
ptrs[disks - 2] = dp;
lib/raid6/recov_neon.c
40
dq = (u8 *)ptrs[failb];
lib/raid6/recov_neon.c
41
ptrs[failb] = raid6_get_zero_page();
lib/raid6/recov_neon.c
42
ptrs[disks - 1] = dq;
lib/raid6/recov_neon.c
44
raid6_call.gen_syndrome(disks, bytes, ptrs);
lib/raid6/recov_neon.c
47
ptrs[faila] = dp;
lib/raid6/recov_neon.c
48
ptrs[failb] = dq;
lib/raid6/recov_neon.c
49
ptrs[disks - 2] = p;
lib/raid6/recov_neon.c
50
ptrs[disks - 1] = q;
lib/raid6/recov_neon.c
62
void **ptrs)
lib/raid6/recov_neon.c
67
p = (u8 *)ptrs[disks - 2];
lib/raid6/recov_neon.c
68
q = (u8 *)ptrs[disks - 1];
lib/raid6/recov_neon.c
74
dq = (u8 *)ptrs[faila];
lib/raid6/recov_neon.c
75
ptrs[faila] = raid6_get_zero_page();
lib/raid6/recov_neon.c
76
ptrs[disks - 1] = dq;
lib/raid6/recov_neon.c
78
raid6_call.gen_syndrome(disks, bytes, ptrs);
lib/raid6/recov_neon.c
81
ptrs[faila] = dq;
lib/raid6/recov_neon.c
82
ptrs[disks - 1] = q;
lib/raid6/recov_rvv.c
146
int failb, void **ptrs)
lib/raid6/recov_rvv.c
152
p = (u8 *)ptrs[disks - 2];
lib/raid6/recov_rvv.c
153
q = (u8 *)ptrs[disks - 1];
lib/raid6/recov_rvv.c
160
dp = (u8 *)ptrs[faila];
lib/raid6/recov_rvv.c
161
ptrs[faila] = raid6_get_zero_page();
lib/raid6/recov_rvv.c
162
ptrs[disks - 2] = dp;
lib/raid6/recov_rvv.c
163
dq = (u8 *)ptrs[failb];
lib/raid6/recov_rvv.c
164
ptrs[failb] = raid6_get_zero_page();
lib/raid6/recov_rvv.c
165
ptrs[disks - 1] = dq;
lib/raid6/recov_rvv.c
167
raid6_call.gen_syndrome(disks, bytes, ptrs);
lib/raid6/recov_rvv.c
170
ptrs[faila] = dp;
lib/raid6/recov_rvv.c
171
ptrs[failb] = dq;
lib/raid6/recov_rvv.c
172
ptrs[disks - 2] = p;
lib/raid6/recov_rvv.c
173
ptrs[disks - 1] = q;
lib/raid6/recov_rvv.c
186
void **ptrs)
lib/raid6/recov_rvv.c
191
p = (u8 *)ptrs[disks - 2];
lib/raid6/recov_rvv.c
192
q = (u8 *)ptrs[disks - 1];
lib/raid6/recov_rvv.c
198
dq = (u8 *)ptrs[faila];
lib/raid6/recov_rvv.c
199
ptrs[faila] = raid6_get_zero_page();
lib/raid6/recov_rvv.c
200
ptrs[disks - 1] = dq;
lib/raid6/recov_rvv.c
202
raid6_call.gen_syndrome(disks, bytes, ptrs);
lib/raid6/recov_rvv.c
205
ptrs[faila] = dq;
lib/raid6/recov_rvv.c
206
ptrs[disks - 1] = q;
lib/raid6/recov_s390xc.c
23
int failb, void **ptrs)
lib/raid6/recov_s390xc.c
30
p = (u8 *)ptrs[disks-2];
lib/raid6/recov_s390xc.c
31
q = (u8 *)ptrs[disks-1];
lib/raid6/recov_s390xc.c
36
dp = (u8 *)ptrs[faila];
lib/raid6/recov_s390xc.c
37
ptrs[faila] = raid6_get_zero_page();
lib/raid6/recov_s390xc.c
38
ptrs[disks-2] = dp;
lib/raid6/recov_s390xc.c
39
dq = (u8 *)ptrs[failb];
lib/raid6/recov_s390xc.c
40
ptrs[failb] = raid6_get_zero_page();
lib/raid6/recov_s390xc.c
41
ptrs[disks-1] = dq;
lib/raid6/recov_s390xc.c
43
raid6_call.gen_syndrome(disks, bytes, ptrs);
lib/raid6/recov_s390xc.c
46
ptrs[faila] = dp;
lib/raid6/recov_s390xc.c
47
ptrs[failb] = dq;
lib/raid6/recov_s390xc.c
48
ptrs[disks-2] = p;
lib/raid6/recov_s390xc.c
49
ptrs[disks-1] = q;
lib/raid6/recov_s390xc.c
72
void **ptrs)
lib/raid6/recov_s390xc.c
78
p = (u8 *)ptrs[disks-2];
lib/raid6/recov_s390xc.c
79
q = (u8 *)ptrs[disks-1];
lib/raid6/recov_s390xc.c
83
dq = (u8 *)ptrs[faila];
lib/raid6/recov_s390xc.c
84
ptrs[faila] = raid6_get_zero_page();
lib/raid6/recov_s390xc.c
85
ptrs[disks-1] = dq;
lib/raid6/recov_s390xc.c
87
raid6_call.gen_syndrome(disks, bytes, ptrs);
lib/raid6/recov_s390xc.c
90
ptrs[faila] = dq;
lib/raid6/recov_s390xc.c
91
ptrs[disks-1] = q;
lib/raid6/recov_ssse3.c
17
int failb, void **ptrs)
lib/raid6/recov_ssse3.c
192
void **ptrs)
lib/raid6/recov_ssse3.c
200
p = (u8 *)ptrs[disks-2];
lib/raid6/recov_ssse3.c
201
q = (u8 *)ptrs[disks-1];
lib/raid6/recov_ssse3.c
205
dq = (u8 *)ptrs[faila];
lib/raid6/recov_ssse3.c
206
ptrs[faila] = raid6_get_zero_page();
lib/raid6/recov_ssse3.c
207
ptrs[disks-1] = dq;
lib/raid6/recov_ssse3.c
209
raid6_call.gen_syndrome(disks, bytes, ptrs);
lib/raid6/recov_ssse3.c
212
ptrs[faila] = dq;
lib/raid6/recov_ssse3.c
213
ptrs[disks-1] = q;
lib/raid6/recov_ssse3.c
26
p = (u8 *)ptrs[disks-2];
lib/raid6/recov_ssse3.c
27
q = (u8 *)ptrs[disks-1];
lib/raid6/recov_ssse3.c
32
dp = (u8 *)ptrs[faila];
lib/raid6/recov_ssse3.c
33
ptrs[faila] = raid6_get_zero_page();
lib/raid6/recov_ssse3.c
34
ptrs[disks-2] = dp;
lib/raid6/recov_ssse3.c
35
dq = (u8 *)ptrs[failb];
lib/raid6/recov_ssse3.c
36
ptrs[failb] = raid6_get_zero_page();
lib/raid6/recov_ssse3.c
37
ptrs[disks-1] = dq;
lib/raid6/recov_ssse3.c
39
raid6_call.gen_syndrome(disks, bytes, ptrs);
lib/raid6/recov_ssse3.c
42
ptrs[faila] = dp;
lib/raid6/recov_ssse3.c
43
ptrs[failb] = dq;
lib/raid6/recov_ssse3.c
44
ptrs[disks-2] = p;
lib/raid6/recov_ssse3.c
45
ptrs[disks-1] = q;
lib/raid6/rvv.c
18
static void raid6_rvv1_gen_syndrome_real(int disks, unsigned long bytes, void **ptrs)
lib/raid6/rvv.c
193
static void raid6_rvv2_gen_syndrome_real(int disks, unsigned long bytes, void **ptrs)
lib/raid6/rvv.c
195
u8 **dptr = (u8 **)ptrs;
lib/raid6/rvv.c
20
u8 **dptr = (u8 **)ptrs;
lib/raid6/rvv.c
287
unsigned long bytes, void **ptrs)
lib/raid6/rvv.c
289
u8 **dptr = (u8 **)ptrs;
lib/raid6/rvv.c
417
static void raid6_rvv4_gen_syndrome_real(int disks, unsigned long bytes, void **ptrs)
lib/raid6/rvv.c
419
u8 **dptr = (u8 **)ptrs;
lib/raid6/rvv.c
545
unsigned long bytes, void **ptrs)
lib/raid6/rvv.c
547
u8 **dptr = (u8 **)ptrs;
lib/raid6/rvv.c
731
static void raid6_rvv8_gen_syndrome_real(int disks, unsigned long bytes, void **ptrs)
lib/raid6/rvv.c
733
u8 **dptr = (u8 **)ptrs;
lib/raid6/rvv.c
927
unsigned long bytes, void **ptrs)
lib/raid6/rvv.c
929
u8 **dptr = (u8 **)ptrs;
lib/raid6/rvv.c
93
unsigned long bytes, void **ptrs)
lib/raid6/rvv.c
95
u8 **dptr = (u8 **)ptrs;
lib/raid6/rvv.h
29
size_t bytes, void **ptrs) \
lib/raid6/rvv.h
35
(unsigned long)bytes, ptrs); \
lib/raid6/rvv.h
40
size_t bytes, void **ptrs) \
lib/raid6/rvv.h
47
start, stop, (unsigned long)bytes, ptrs); \
lib/raid6/sse1.c
101
u8 **dptr = (u8 **)ptrs;
lib/raid6/sse1.c
40
static void raid6_sse11_gen_syndrome(int disks, size_t bytes, void **ptrs)
lib/raid6/sse1.c
42
u8 **dptr = (u8 **)ptrs;
lib/raid6/sse1.c
99
static void raid6_sse12_gen_syndrome(int disks, size_t bytes, void **ptrs)
lib/raid6/sse2.c
146
static void raid6_sse22_gen_syndrome(int disks, size_t bytes, void **ptrs)
lib/raid6/sse2.c
148
u8 **dptr = (u8 **)ptrs;
lib/raid6/sse2.c
199
size_t bytes, void **ptrs)
lib/raid6/sse2.c
201
u8 **dptr = (u8 **)ptrs;
lib/raid6/sse2.c
278
static void raid6_sse24_gen_syndrome(int disks, size_t bytes, void **ptrs)
lib/raid6/sse2.c
280
u8 **dptr = (u8 **)ptrs;
lib/raid6/sse2.c
36
static void raid6_sse21_gen_syndrome(int disks, size_t bytes, void **ptrs)
lib/raid6/sse2.c
365
size_t bytes, void **ptrs)
lib/raid6/sse2.c
367
u8 **dptr = (u8 **)ptrs;
lib/raid6/sse2.c
38
u8 **dptr = (u8 **)ptrs;
lib/raid6/sse2.c
88
size_t bytes, void **ptrs)
lib/raid6/sse2.c
90
u8 **dptr = (u8 **)ptrs;
lib/test_bpf.c
15452
if (progs->ptrs[i])
lib/test_bpf.c
15453
bpf_prog_free(progs->ptrs[i]);
lib/test_bpf.c
15464
progs = kzalloc_flex(*progs, ptrs, ntests + 1);
lib/test_bpf.c
15558
progs->ptrs[which] = fp;
lib/test_bpf.c
15582
struct bpf_prog *fp = progs->ptrs[i];
net/core/filter.c
4991
cgrp = READ_ONCE(array->ptrs[idx]);
net/xdp/xsk.c
1600
ring->producer = offsetof(struct xdp_rxtx_ring, ptrs.producer);
net/xdp/xsk.c
1601
ring->consumer = offsetof(struct xdp_rxtx_ring, ptrs.consumer);
net/xdp/xsk.c
1607
ring->producer = offsetof(struct xdp_umem_ring, ptrs.producer);
net/xdp/xsk.c
1608
ring->consumer = offsetof(struct xdp_umem_ring, ptrs.consumer);
net/xdp/xsk.c
1695
ptrs.flags);
net/xdp/xsk.c
1697
ptrs.flags);
net/xdp/xsk.c
1699
ptrs.flags);
net/xdp/xsk.c
1701
ptrs.flags);
net/xdp/xsk_queue.h
30
struct xdp_ring ptrs;
net/xdp/xsk_queue.h
36
struct xdp_ring ptrs;
net/xfrm/xfrm_state.c
1117
static void xfrm_hash_ptrs_get(const struct net *net, struct xfrm_hash_state_ptrs *ptrs)
net/xfrm/xfrm_state.c
1124
ptrs->bydst = xfrm_state_deref_check(net->xfrm.state_bydst, net);
net/xfrm/xfrm_state.c
1125
ptrs->bysrc = xfrm_state_deref_check(net->xfrm.state_bysrc, net);
net/xfrm/xfrm_state.c
1126
ptrs->byspi = xfrm_state_deref_check(net->xfrm.state_byspi, net);
net/xfrm/xfrm_state.c
1127
ptrs->hmask = net->xfrm.state_hmask;
tools/bpf/bpftool/skeleton/profiler.bpf.c
48
struct bpf_perf_event_value___local *ptrs[MAX_NUM_METRICS];
tools/bpf/bpftool/skeleton/profiler.bpf.c
56
ptrs[i] = bpf_map_lookup_elem(&fentry_readings, &flag);
tools/bpf/bpftool/skeleton/profiler.bpf.c
57
if (!ptrs[i])
tools/bpf/bpftool/skeleton/profiler.bpf.c
69
*(ptrs[i]) = reading;
tools/testing/selftests/bpf/progs/cpumask_success.c
30
struct kptr_nested_mid ptrs[2];
tools/testing/selftests/bpf/progs/cpumask_success.c
677
r = _global_mask_array_rcu(&global_mask_nested_deep.ptrs[0].m.mask,
tools/testing/selftests/bpf/progs/cpumask_success.c
678
&global_mask_nested_deep.ptrs[1].m.mask);