Symbol: meta
arch/arm64/kernel/stacktrace.c
194
struct frame_record_meta *meta;
arch/arm64/kernel/stacktrace.c
197
info = unwind_find_stack(&state->common, fp, sizeof(*meta));
arch/arm64/kernel/stacktrace.c
201
meta = (struct frame_record_meta *)fp;
arch/arm64/kernel/stacktrace.c
202
switch (READ_ONCE(meta->type)) {
arch/arm64/kernel/stacktrace.c
204
if (meta == &task_pt_regs(tsk)->stackframe)
arch/arm64/net/bpf_jit_comp.c
2692
u64 meta = func_meta | (cookie_bargs_off << BPF_TRAMP_COOKIE_INDEX_SHIFT);
arch/arm64/net/bpf_jit_comp.c
2694
store_func_meta(ctx, meta, func_meta_off);
arch/arm64/net/bpf_jit_comp.c
2740
u64 meta = func_meta | (cookie_bargs_off << BPF_TRAMP_COOKIE_INDEX_SHIFT);
arch/arm64/net/bpf_jit_comp.c
2742
store_func_meta(ctx, meta, func_meta_off);
arch/riscv/net/bpf_jit_comp64.c
1009
u64 meta = func_meta | ((u64)cur_cookie << BPF_TRAMP_COOKIE_INDEX_SHIFT);
arch/riscv/net/bpf_jit_comp64.c
1011
emit_store_stack_imm64(RV_REG_T1, -func_meta_off, meta, ctx);
block/bio-integrity.c
370
static void bio_uio_meta_to_bip(struct bio *bio, struct uio_meta *meta)
block/bio-integrity.c
374
if (meta->flags & IO_INTEGRITY_CHK_GUARD)
block/bio-integrity.c
376
if (meta->flags & IO_INTEGRITY_CHK_APPTAG)
block/bio-integrity.c
378
if (meta->flags & IO_INTEGRITY_CHK_REFTAG)
block/bio-integrity.c
381
bip->app_tag = meta->app_tag;
block/bio-integrity.c
384
int bio_integrity_map_iter(struct bio *bio, struct uio_meta *meta)
block/bio-integrity.c
397
it = meta->iter;
block/bio-integrity.c
405
if (meta->flags && (meta->flags & ~IO_INTEGRITY_VALID_FLAGS))
block/bio-integrity.c
411
bio_uio_meta_to_bip(bio, meta);
block/bio-integrity.c
412
bip_set_seed(bio_integrity(bio), meta->seed);
block/bio-integrity.c
413
iov_iter_advance(&meta->iter, integrity_bytes);
block/bio-integrity.c
414
meta->seed += bio_integrity_intervals(bi, bio_sectors(bio));
drivers/accel/amdxdna/aie2_pci.c
676
struct amdxdna_drm_query_aie_metadata *meta;
drivers/accel/amdxdna/aie2_pci.c
682
meta = kzalloc_obj(*meta);
drivers/accel/amdxdna/aie2_pci.c
683
if (!meta)
drivers/accel/amdxdna/aie2_pci.c
686
meta->col_size = ndev->metadata.size;
drivers/accel/amdxdna/aie2_pci.c
687
meta->cols = ndev->metadata.cols;
drivers/accel/amdxdna/aie2_pci.c
688
meta->rows = ndev->metadata.rows;
drivers/accel/amdxdna/aie2_pci.c
690
meta->version.major = ndev->metadata.version.major;
drivers/accel/amdxdna/aie2_pci.c
691
meta->version.minor = ndev->metadata.version.minor;
drivers/accel/amdxdna/aie2_pci.c
693
meta->core.row_count = ndev->metadata.core.row_count;
drivers/accel/amdxdna/aie2_pci.c
694
meta->core.row_start = ndev->metadata.core.row_start;
drivers/accel/amdxdna/aie2_pci.c
695
meta->core.dma_channel_count = ndev->metadata.core.dma_channel_count;
drivers/accel/amdxdna/aie2_pci.c
696
meta->core.lock_count = ndev->metadata.core.lock_count;
drivers/accel/amdxdna/aie2_pci.c
697
meta->core.event_reg_count = ndev->metadata.core.event_reg_count;
drivers/accel/amdxdna/aie2_pci.c
699
meta->mem.row_count = ndev->metadata.mem.row_count;
drivers/accel/amdxdna/aie2_pci.c
700
meta->mem.row_start = ndev->metadata.mem.row_start;
drivers/accel/amdxdna/aie2_pci.c
701
meta->mem.dma_channel_count = ndev->metadata.mem.dma_channel_count;
drivers/accel/amdxdna/aie2_pci.c
702
meta->mem.lock_count = ndev->metadata.mem.lock_count;
drivers/accel/amdxdna/aie2_pci.c
703
meta->mem.event_reg_count = ndev->metadata.mem.event_reg_count;
drivers/accel/amdxdna/aie2_pci.c
705
meta->shim.row_count = ndev->metadata.shim.row_count;
drivers/accel/amdxdna/aie2_pci.c
706
meta->shim.row_start = ndev->metadata.shim.row_start;
drivers/accel/amdxdna/aie2_pci.c
707
meta->shim.dma_channel_count = ndev->metadata.shim.dma_channel_count;
drivers/accel/amdxdna/aie2_pci.c
708
meta->shim.lock_count = ndev->metadata.shim.lock_count;
drivers/accel/amdxdna/aie2_pci.c
709
meta->shim.event_reg_count = ndev->metadata.shim.event_reg_count;
drivers/accel/amdxdna/aie2_pci.c
711
if (copy_to_user(u64_to_user_ptr(args->buffer), meta, sizeof(*meta)))
drivers/accel/amdxdna/aie2_pci.c
714
kfree(meta);
drivers/block/drbd/drbd_int.h
643
struct drbd_socket meta; /* ping/ack (metadata) packets */
drivers/block/drbd/drbd_main.c
1042
sock = &peer_device->connection->meta;
drivers/block/drbd/drbd_main.c
1056
sock = &connection->meta;
drivers/block/drbd/drbd_main.c
1301
sock = &connection->meta;
drivers/block/drbd/drbd_main.c
1327
sock = &peer_device->connection->meta;
drivers/block/drbd/drbd_main.c
1456
drop_it = connection->meta.socket == sock
drivers/block/drbd/drbd_main.c
1847
sock == connection->meta.socket ? "msock" : "sock",
drivers/block/drbd/drbd_main.c
2556
if (drbd_alloc_socket(&connection->meta))
drivers/block/drbd/drbd_main.c
2584
mutex_init(&connection->meta.mutex);
drivers/block/drbd/drbd_main.c
2610
drbd_free_socket(&connection->meta);
drivers/block/drbd/drbd_main.c
2627
drbd_free_socket(&connection->meta);
drivers/block/drbd/drbd_main.c
2912
if (connection->meta.socket)
drivers/block/drbd/drbd_main.c
2913
drbd_free_one_sock(&connection->meta);
drivers/block/drbd/drbd_main.c
673
sock = &connection->meta;
drivers/block/drbd/drbd_main.c
683
sock = &connection->meta;
drivers/block/drbd/drbd_receiver.c
5433
connection->meta.socket->sk->sk_rcvtimeo = connection->net_conf->ping_int*HZ;
drivers/block/drbd/drbd_receiver.c
5734
connection->meta.socket->sk->sk_rcvtimeo = t;
drivers/block/drbd/drbd_receiver.c
5774
void *buf = connection->meta.rbuf;
drivers/block/drbd/drbd_receiver.c
5795
rv = drbd_recv_short(connection->meta.socket, buf, expect-received, 0);
drivers/block/drbd/drbd_receiver.c
5848
if (decode_header(connection, connection->meta.rbuf, &pi))
drivers/block/drbd/drbd_receiver.c
5879
buf = connection->meta.rbuf;
drivers/block/drbd/drbd_receiver.c
5916
tcp_sock_set_cork(connection->meta.socket->sk, true);
drivers/block/drbd/drbd_receiver.c
5929
tcp_sock_set_cork(connection->meta.socket->sk, false);
drivers/block/drbd/drbd_receiver.c
757
msock.sbuf = connection->meta.sbuf;
drivers/block/drbd/drbd_receiver.c
758
msock.rbuf = connection->meta.rbuf;
drivers/block/drbd/drbd_receiver.c
874
connection->meta.socket = msock.socket;
drivers/gpio/gpio-aggregator.c
1442
struct gpio_aggregator_pdev_meta meta = { .init_via_sysfs = true };
drivers/gpio/gpio-aggregator.c
1492
pdev = platform_device_register_data(NULL, DRV_NAME, aggr->id, &meta, sizeof(meta));
drivers/gpio/gpio-aggregator.c
1585
struct gpio_aggregator_pdev_meta *meta;
drivers/gpio/gpio-aggregator.c
1601
meta = dev_get_platdata(&pdev->dev);
drivers/gpio/gpio-aggregator.c
1602
if (meta && meta->init_via_sysfs)
drivers/gpio/gpio-sloppy-logic-analyzer.c
229
char *meta = NULL;
drivers/gpio/gpio-sloppy-logic-analyzer.c
274
new_meta = devm_krealloc(dev, meta, meta_len + add_len, GFP_KERNEL);
drivers/gpio/gpio-sloppy-logic-analyzer.c
278
meta = new_meta;
drivers/gpio/gpio-sloppy-logic-analyzer.c
279
meta_len += snprintf(meta + meta_len, add_len, "probe%02u=%s\n",
drivers/gpio/gpio-sloppy-logic-analyzer.c
286
priv->meta.data = meta;
drivers/gpio/gpio-sloppy-logic-analyzer.c
287
priv->meta.size = meta_len;
drivers/gpio/gpio-sloppy-logic-analyzer.c
289
debugfs_create_blob("meta_data", 0400, priv->debug_dir, &priv->meta);
drivers/gpio/gpio-sloppy-logic-analyzer.c
49
struct debugfs_blob_wrapper meta;
drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c
379
struct drm_amdgpu_info_uq_metadata_gfx *meta)
drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c
387
meta->shadow_size = shadow.shadow_size;
drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c
388
meta->shadow_alignment = shadow.shadow_alignment;
drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c
389
meta->csa_size = shadow.csa_size;
drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c
390
meta->csa_alignment = shadow.csa_alignment;
drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c
399
struct drm_amdgpu_info_uq_metadata_compute *meta)
drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c
407
meta->eop_size = shadow.eop_size;
drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c
408
meta->eop_alignment = shadow.eop_alignment;
drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c
417
struct drm_amdgpu_info_uq_metadata_sdma *meta)
drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c
425
meta->csa_size = csa.size;
drivers/gpu/drm/amd/amdgpu/amdgpu_kms.c
426
meta->csa_alignment = csa.alignment;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4.c
157
const struct dml2_implicit_svp_meta *meta)
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4.c
161
phantom->timing.v_total = meta->v_total;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4.c
162
phantom->timing.v_active = meta->v_active;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4.c
163
phantom->timing.v_front_porch = meta->v_front_porch;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_utils.c
577
const struct dml2_implicit_svp_meta *meta)
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_utils.c
581
phantom->timing.v_total = meta->v_total;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_utils.c
582
phantom->timing.v_active = meta->v_active;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_utils.c
583
phantom->timing.v_front_porch = meta->v_front_porch;
drivers/gpu/drm/amd/display/dmub/src/dmub_srv.c
119
const union dmub_fw_meta *meta;
drivers/gpu/drm/amd/display/dmub/src/dmub_srv.c
127
meta = (const union dmub_fw_meta *)(blob + blob_size - meta_offset -
drivers/gpu/drm/amd/display/dmub/src/dmub_srv.c
130
if (meta->info.magic_value != DMUB_FW_META_MAGIC)
drivers/gpu/drm/amd/display/dmub/src/dmub_srv.c
133
return &meta->info;
drivers/gpu/drm/etnaviv/etnaviv_perfmon.c
471
const struct etnaviv_pm_domain_meta *meta = &doms_meta[i];
drivers/gpu/drm/etnaviv/etnaviv_perfmon.c
473
if (gpu->identity.features & meta->feature)
drivers/gpu/drm/etnaviv/etnaviv_perfmon.c
474
num += meta->nr_domains;
drivers/gpu/drm/etnaviv/etnaviv_perfmon.c
487
const struct etnaviv_pm_domain_meta *meta = &doms_meta[i];
drivers/gpu/drm/etnaviv/etnaviv_perfmon.c
489
if (!(gpu->identity.features & meta->feature))
drivers/gpu/drm/etnaviv/etnaviv_perfmon.c
492
if (index - offset >= meta->nr_domains) {
drivers/gpu/drm/etnaviv/etnaviv_perfmon.c
493
offset += meta->nr_domains;
drivers/gpu/drm/etnaviv/etnaviv_perfmon.c
497
domain = meta->domains + (index - offset);
drivers/gpu/drm/etnaviv/etnaviv_perfmon.c
559
const struct etnaviv_pm_domain_meta *meta = &doms_meta[exec_state];
drivers/gpu/drm/etnaviv/etnaviv_perfmon.c
562
if (r->domain >= meta->nr_domains)
drivers/gpu/drm/etnaviv/etnaviv_perfmon.c
565
dom = meta->domains + r->domain;
drivers/gpu/drm/etnaviv/etnaviv_perfmon.c
576
const struct etnaviv_pm_domain_meta *meta = &doms_meta[exec_state];
drivers/gpu/drm/etnaviv/etnaviv_perfmon.c
582
dom = meta->domains + pmr->domain;
drivers/gpu/drm/imagination/pvr_device.c
687
if (PVR_HAS_FEATURE(pvr_dev, meta))
drivers/gpu/drm/imagination/pvr_device_info.c
154
FEATURE_MAPPING_VALUE(META, meta),
drivers/gpu/drm/imagination/pvr_device_info.h
99
u64 meta;
drivers/gpu/drm/msm/disp/dpu1/dpu_formats.c
124
if (!meta)
drivers/gpu/drm/msm/disp/dpu1/dpu_formats.c
258
bool meta;
drivers/gpu/drm/msm/disp/dpu1/dpu_formats.c
263
meta = MSM_FORMAT_IS_UBWC(fmt);
drivers/gpu/drm/msm/disp/dpu1/dpu_formats.c
294
if (!meta)
drivers/gpu/drm/msm/disp/dpu1/dpu_formats.c
325
if (!meta)
drivers/gpu/drm/msm/disp/dpu1/dpu_formats.c
61
bool meta = MSM_FORMAT_IS_UBWC(fmt);
drivers/gpu/drm/msm/disp/dpu1/dpu_formats.c
98
if (!meta)
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
155
struct nvkm_gsp_mem meta;
drivers/gpu/drm/nouveau/nvkm/falcon/fw.c
316
u32 loc, sig, cnt, *meta;
drivers/gpu/drm/nouveau/nvkm/falcon/fw.c
325
meta = (u32 *)(blob->data + hshdr->meta_data_offset);
drivers/gpu/drm/nouveau/nvkm/falcon/fw.c
353
fw->fuse_ver = meta[0];
drivers/gpu/drm/nouveau/nvkm/falcon/fw.c
354
fw->engine_id = meta[1];
drivers/gpu/drm/nouveau/nvkm/falcon/fw.c
355
fw->ucode_id = meta[2];
drivers/gpu/drm/nouveau/nvkm/subdev/acr/lsfw.c
221
u32 sig, cnt, *meta;
drivers/gpu/drm/nouveau/nvkm/subdev/acr/lsfw.c
229
meta = (u32 *)(hsbl->data + hshdr->meta_data_offset);
drivers/gpu/drm/nouveau/nvkm/subdev/acr/lsfw.c
233
lsfw->fuse_ver = meta[0];
drivers/gpu/drm/nouveau/nvkm/subdev/acr/lsfw.c
234
lsfw->engine_id = meta[1];
drivers/gpu/drm/nouveau/nvkm/subdev/acr/lsfw.c
235
lsfw->ucode_id = meta[2];
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/ga102.c
50
u32 loc, sig, cnt, *meta;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/ga102.c
55
meta = (u32 *)(blob->data + hshdr->meta_data_offset);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/ga102.c
83
fw->fuse_ver = meta[0];
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/ga102.c
84
fw->engine_id = meta[1];
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/ga102.c
85
fw->ucode_id = meta[2];
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/gh100.c
133
GspFwWprMeta *meta;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/gh100.c
136
ret = nvkm_gsp_mem_ctor(gsp, sizeof(*meta), &gsp->wpr_meta);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/gh100.c
144
meta = gsp->wpr_meta.data;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/gh100.c
146
meta->magic = GSP_FW_WPR_META_MAGIC;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/gh100.c
147
meta->revision = GSP_FW_WPR_META_REVISION;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/gh100.c
149
meta->sizeOfRadix3Elf = gsp->fw.len;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/gh100.c
150
meta->sysmemAddrOfRadix3Elf = gsp->radix3.lvl0.addr;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/gh100.c
152
meta->sizeOfBootloader = gsp->boot.fw.size;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/gh100.c
153
meta->sysmemAddrOfBootloader = gsp->boot.fw.addr;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/gh100.c
154
meta->bootloaderCodeOffset = gsp->boot.code_offset;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/gh100.c
155
meta->bootloaderDataOffset = gsp->boot.data_offset;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/gh100.c
156
meta->bootloaderManifestOffset = gsp->boot.manifest_offset;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/gh100.c
158
meta->sysmemAddrOfSignature = gsp->sig.addr;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/gh100.c
159
meta->sizeOfSignature = gsp->sig.size;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/gh100.c
161
meta->nonWprHeapSize = gsp->fb.heap.size;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/gh100.c
162
meta->gspFwHeapSize = tu102_gsp_wpr_heap_size(gsp);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/gh100.c
163
meta->frtsSize = 0x100000;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/gh100.c
164
meta->vgaWorkspaceSize = gsp->fb.bios.vga_workspace.size;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/gh100.c
165
meta->pmuReservedSize = gsp->rm->wpr->rsvd_size_pmu;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/gh100.c
73
const bool resume = gsp->sr.meta.data != NULL;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/gh100.c
74
struct nvkm_gsp_mem *meta;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/gh100.c
85
meta = &gsp->wpr_meta;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/gh100.c
88
meta = &gsp->sr.meta;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/gh100.c
93
args->bootGspRmParams.gspRmDescOffset = meta->addr;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/gh100.c
94
args->bootGspRmParams.gspRmDescSize = meta->size;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
1718
GspFwWprMeta *meta = gsp->wpr_meta.data;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
1720
return meta->gspFwWprEnd - meta->gspFwWprStart;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
1741
ret = nvkm_gsp_mem_ctor(gsp, sizeof(*sr), &gsp->sr.meta);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
1745
sr = gsp->sr.meta.data;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
1753
nvkm_gsp_mem_dtor(&gsp->sr.meta);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
1797
if (gsp->sr.meta.data) {
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
1800
nvkm_gsp_mem_dtor(&gsp->sr.meta);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/fbsr.c
83
ctrl->sysmemAddrOfSuspendResumeData = gsp->sr.meta.addr;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/gsp.c
108
GspFwWprMeta *meta = gsp->wpr_meta.data;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/gsp.c
110
meta->nonWprHeapOffset = rpc->fwWprLayoutOffset.nonWprHeapOffset;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/gsp.c
111
meta->frtsOffset = rpc->fwWprLayoutOffset.frtsOffset;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/gsp.c
20
GspFwWprMeta *meta = gsp->wpr_meta.data;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/gsp.c
22
return (meta->frtsOffset + meta->frtsSize) -
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r570/gsp.c
23
(meta->nonWprHeapOffset + meta->nonWprHeapSize);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/tu102.c
179
mbox0 = lower_32_bits(gsp->sr.meta.addr);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/tu102.c
180
mbox1 = upper_32_bits(gsp->sr.meta.addr);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/tu102.c
194
if (!gsp->sr.meta.data) {
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/tu102.c
200
mbox0 = lower_32_bits(gsp->sr.meta.addr);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/tu102.c
201
mbox1 = upper_32_bits(gsp->sr.meta.addr);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/tu102.c
215
GspFwWprMeta *meta;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/tu102.c
218
ret = nvkm_gsp_mem_ctor(gsp, sizeof(*meta), &gsp->wpr_meta);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/tu102.c
222
meta = gsp->wpr_meta.data;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/tu102.c
224
meta->magic = GSP_FW_WPR_META_MAGIC;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/tu102.c
225
meta->revision = GSP_FW_WPR_META_REVISION;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/tu102.c
227
meta->sysmemAddrOfRadix3Elf = gsp->radix3.lvl0.addr;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/tu102.c
228
meta->sizeOfRadix3Elf = gsp->fb.wpr2.elf.size;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/tu102.c
230
meta->sysmemAddrOfBootloader = gsp->boot.fw.addr;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/tu102.c
231
meta->sizeOfBootloader = gsp->boot.fw.size;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/tu102.c
232
meta->bootloaderCodeOffset = gsp->boot.code_offset;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/tu102.c
233
meta->bootloaderDataOffset = gsp->boot.data_offset;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/tu102.c
234
meta->bootloaderManifestOffset = gsp->boot.manifest_offset;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/tu102.c
236
meta->sysmemAddrOfSignature = gsp->sig.addr;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/tu102.c
237
meta->sizeOfSignature = gsp->sig.size;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/tu102.c
239
meta->gspFwRsvdStart = gsp->fb.heap.addr;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/tu102.c
240
meta->nonWprHeapOffset = gsp->fb.heap.addr;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/tu102.c
241
meta->nonWprHeapSize = gsp->fb.heap.size;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/tu102.c
242
meta->gspFwWprStart = gsp->fb.wpr2.addr;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/tu102.c
243
meta->gspFwHeapOffset = gsp->fb.wpr2.heap.addr;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/tu102.c
244
meta->gspFwHeapSize = gsp->fb.wpr2.heap.size;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/tu102.c
245
meta->gspFwOffset = gsp->fb.wpr2.elf.addr;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/tu102.c
246
meta->bootBinOffset = gsp->fb.wpr2.boot.addr;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/tu102.c
247
meta->frtsOffset = gsp->fb.wpr2.frts.addr;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/tu102.c
248
meta->frtsSize = gsp->fb.wpr2.frts.size;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/tu102.c
249
meta->gspFwWprEnd = ALIGN_DOWN(gsp->fb.bios.vga_workspace.addr, 0x20000);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/tu102.c
250
meta->fbSize = gsp->fb.size;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/tu102.c
251
meta->vgaWorkspaceOffset = gsp->fb.bios.vga_workspace.addr;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/tu102.c
252
meta->vgaWorkspaceSize = gsp->fb.bios.vga_workspace.size;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/tu102.c
253
meta->bootCount = 0;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/tu102.c
254
meta->partitionRpcAddr = 0;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/tu102.c
255
meta->partitionRpcRequestOffset = 0;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/tu102.c
256
meta->partitionRpcReplyOffset = 0;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/tu102.c
257
meta->verified = 0;
drivers/infiniband/hw/efa/efa_io_defs.h
238
struct efa_io_tx_meta_desc meta;
drivers/infiniband/hw/ionic/ionic_datapath.c
1004
struct ionic_sq_meta *meta;
drivers/infiniband/hw/ionic/ionic_datapath.c
101
meta = &qp->sq_meta[qp->sq.cons];
drivers/infiniband/hw/ionic/ionic_datapath.c
1013
meta = &qp->sq_meta[qp->sq.prod];
drivers/infiniband/hw/ionic/ionic_datapath.c
1018
meta->ibop = IB_WC_RDMA_WRITE;
drivers/infiniband/hw/ionic/ionic_datapath.c
1022
meta->ibop = IB_WC_COMP_SWAP;
drivers/infiniband/hw/ionic/ionic_datapath.c
1030
meta->ibop = IB_WC_FETCH_ADD;
drivers/infiniband/hw/ionic/ionic_datapath.c
1048
return ionic_prep_common(qp, &wr->wr, meta, wqe);
drivers/infiniband/hw/ionic/ionic_datapath.c
1055
struct ionic_sq_meta *meta;
drivers/infiniband/hw/ionic/ionic_datapath.c
1061
meta = &qp->sq_meta[qp->sq.prod];
drivers/infiniband/hw/ionic/ionic_datapath.c
1069
meta->len = 0;
drivers/infiniband/hw/ionic/ionic_datapath.c
1070
meta->ibop = IB_WC_LOCAL_INV;
drivers/infiniband/hw/ionic/ionic_datapath.c
1072
ionic_prep_base(qp, wr, meta, wqe);
drivers/infiniband/hw/ionic/ionic_datapath.c
108
wc->wr_id = meta->wrid;
drivers/infiniband/hw/ionic/ionic_datapath.c
1082
struct ionic_sq_meta *meta;
drivers/infiniband/hw/ionic/ionic_datapath.c
1094
meta = &qp->sq_meta[qp->sq.prod];
drivers/infiniband/hw/ionic/ionic_datapath.c
1115
meta->len = 0;
drivers/infiniband/hw/ionic/ionic_datapath.c
1116
meta->ibop = IB_WC_REG_MR;
drivers/infiniband/hw/ionic/ionic_datapath.c
1118
ionic_prep_base(qp, &wr->wr, meta, wqe);
drivers/infiniband/hw/ionic/ionic_datapath.c
1180
struct ionic_rq_meta *meta;
drivers/infiniband/hw/ionic/ionic_datapath.c
1191
meta = qp->rq_meta_head;
drivers/infiniband/hw/ionic/ionic_datapath.c
1192
if (unlikely(meta == IONIC_META_LAST) ||
drivers/infiniband/hw/ionic/ionic_datapath.c
1193
unlikely(meta == IONIC_META_POSTED))
drivers/infiniband/hw/ionic/ionic_datapath.c
1206
meta->wrid = wr->wr_id;
drivers/infiniband/hw/ionic/ionic_datapath.c
1208
wqe->base.wqe_id = meta - qp->rq_meta;
drivers/infiniband/hw/ionic/ionic_datapath.c
1216
qp->rq_meta_head = meta->next;
drivers/infiniband/hw/ionic/ionic_datapath.c
1217
meta->next = IONIC_META_POSTED;
drivers/infiniband/hw/ionic/ionic_datapath.c
135
struct ionic_rq_meta *meta;
drivers/infiniband/hw/ionic/ionic_datapath.c
172
meta = &qp->rq_meta[cqe->recv.wqe_id];
drivers/infiniband/hw/ionic/ionic_datapath.c
173
if (unlikely(meta->next != IONIC_META_POSTED)) {
drivers/infiniband/hw/ionic/ionic_datapath.c
180
meta->next = qp->rq_meta_head;
drivers/infiniband/hw/ionic/ionic_datapath.c
181
qp->rq_meta_head = meta;
drivers/infiniband/hw/ionic/ionic_datapath.c
185
wc->wr_id = meta->wrid;
drivers/infiniband/hw/ionic/ionic_datapath.c
264
struct ionic_sq_meta *meta;
drivers/infiniband/hw/ionic/ionic_datapath.c
273
meta = &qp->sq_meta[qp->sq.cons];
drivers/infiniband/hw/ionic/ionic_datapath.c
276
if (meta->remote && meta->seq == qp->sq_msn_cons)
drivers/infiniband/hw/ionic/ionic_datapath.c
280
if (!meta->remote && !meta->local_comp)
drivers/infiniband/hw/ionic/ionic_datapath.c
289
struct ionic_sq_meta *meta;
drivers/infiniband/hw/ionic/ionic_datapath.c
299
meta = &qp->sq_meta[qp->sq.cons];
drivers/infiniband/hw/ionic/ionic_datapath.c
302
if (meta->remote && meta->seq == qp->sq_msn_cons)
drivers/infiniband/hw/ionic/ionic_datapath.c
306
if (!meta->remote && !meta->local_comp)
drivers/infiniband/hw/ionic/ionic_datapath.c
312
} while (!meta->signal && meta->ibsts == IB_WC_SUCCESS);
drivers/infiniband/hw/ionic/ionic_datapath.c
316
wc->status = meta->ibsts;
drivers/infiniband/hw/ionic/ionic_datapath.c
317
wc->wr_id = meta->wrid;
drivers/infiniband/hw/ionic/ionic_datapath.c
320
if (meta->ibsts == IB_WC_SUCCESS) {
drivers/infiniband/hw/ionic/ionic_datapath.c
321
wc->byte_len = meta->len;
drivers/infiniband/hw/ionic/ionic_datapath.c
322
wc->opcode = meta->ibop;
drivers/infiniband/hw/ionic/ionic_datapath.c
324
wc->vendor_err = meta->len;
drivers/infiniband/hw/ionic/ionic_datapath.c
33
struct ionic_rq_meta *meta;
drivers/infiniband/hw/ionic/ionic_datapath.c
369
struct ionic_sq_meta *meta;
drivers/infiniband/hw/ionic/ionic_datapath.c
395
meta = &qp->sq_meta[cqe_idx];
drivers/infiniband/hw/ionic/ionic_datapath.c
396
meta->len = be32_to_cpu(cqe->status_length);
drivers/infiniband/hw/ionic/ionic_datapath.c
397
meta->ibsts = ionic_to_ib_status(meta->len);
drivers/infiniband/hw/ionic/ionic_datapath.c
410
struct ionic_sq_meta *meta;
drivers/infiniband/hw/ionic/ionic_datapath.c
434
meta = &qp->sq_meta[cqe_idx];
drivers/infiniband/hw/ionic/ionic_datapath.c
435
meta->local_comp = true;
drivers/infiniband/hw/ionic/ionic_datapath.c
438
meta->len = st_len;
drivers/infiniband/hw/ionic/ionic_datapath.c
439
meta->ibsts = ionic_to_ib_status(st_len);
drivers/infiniband/hw/ionic/ionic_datapath.c
440
meta->remote = false;
drivers/infiniband/hw/ionic/ionic_datapath.c
53
meta = &qp->rq_meta[wqe->base.wqe_id];
drivers/infiniband/hw/ionic/ionic_datapath.c
54
if (unlikely(meta->next != IONIC_META_POSTED)) {
drivers/infiniband/hw/ionic/ionic_datapath.c
66
wc->wr_id = meta->wrid;
drivers/infiniband/hw/ionic/ionic_datapath.c
69
meta->next = qp->rq_meta_head;
drivers/infiniband/hw/ionic/ionic_datapath.c
70
qp->rq_meta_head = meta;
drivers/infiniband/hw/ionic/ionic_datapath.c
806
struct ionic_sq_meta *meta,
drivers/infiniband/hw/ionic/ionic_datapath.c
809
meta->wrid = wr->wr_id;
drivers/infiniband/hw/ionic/ionic_datapath.c
810
meta->ibsts = IB_WC_SUCCESS;
drivers/infiniband/hw/ionic/ionic_datapath.c
811
meta->signal = false;
drivers/infiniband/hw/ionic/ionic_datapath.c
812
meta->local_comp = false;
drivers/infiniband/hw/ionic/ionic_datapath.c
824
meta->signal = true;
drivers/infiniband/hw/ionic/ionic_datapath.c
827
meta->seq = qp->sq_msn_prod;
drivers/infiniband/hw/ionic/ionic_datapath.c
828
meta->remote =
drivers/infiniband/hw/ionic/ionic_datapath.c
833
if (meta->remote) {
drivers/infiniband/hw/ionic/ionic_datapath.c
834
qp->sq_msn_idx[meta->seq] = qp->sq.prod;
drivers/infiniband/hw/ionic/ionic_datapath.c
843
struct ionic_sq_meta *meta,
drivers/infiniband/hw/ionic/ionic_datapath.c
868
meta->len = signed_len;
drivers/infiniband/hw/ionic/ionic_datapath.c
871
ionic_prep_base(qp, wr, meta, wqe);
drivers/infiniband/hw/ionic/ionic_datapath.c
890
struct ionic_sq_meta *meta;
drivers/infiniband/hw/ionic/ionic_datapath.c
893
meta = &qp->sq_meta[qp->sq.prod];
drivers/infiniband/hw/ionic/ionic_datapath.c
898
meta->ibop = IB_WC_SEND;
drivers/infiniband/hw/ionic/ionic_datapath.c
917
return ionic_prep_common(qp, wr, meta, wqe);
drivers/infiniband/hw/ionic/ionic_datapath.c
924
struct ionic_sq_meta *meta;
drivers/infiniband/hw/ionic/ionic_datapath.c
93
struct ionic_sq_meta *meta;
drivers/infiniband/hw/ionic/ionic_datapath.c
933
meta = &qp->sq_meta[qp->sq.prod];
drivers/infiniband/hw/ionic/ionic_datapath.c
942
meta->ibop = IB_WC_SEND;
drivers/infiniband/hw/ionic/ionic_datapath.c
956
return ionic_prep_common(qp, &wr->wr, meta, wqe);
drivers/infiniband/hw/ionic/ionic_datapath.c
963
struct ionic_sq_meta *meta;
drivers/infiniband/hw/ionic/ionic_datapath.c
966
meta = &qp->sq_meta[qp->sq.prod];
drivers/infiniband/hw/ionic/ionic_datapath.c
971
meta->ibop = IB_WC_RDMA_WRITE;
drivers/infiniband/hw/ionic/ionic_datapath.c
977
meta->ibop = IB_WC_RDMA_READ;
drivers/infiniband/hw/ionic/ionic_datapath.c
997
return ionic_prep_common(qp, &wr->wr, meta, wqe);
drivers/md/bcache/sysfs.c
1056
size_t unused = 0, available = 0, dirty = 0, meta = 0;
drivers/md/bcache/sysfs.c
1076
meta++;
drivers/md/bcache/sysfs.c
1118
meta * 100 / (size_t) ca->sb.nbuckets, sum,
drivers/md/dm-pcache/pcache_internal.h
100
if (!latest || pcache_meta_seq_after(meta->seq, seq_latest)) {
drivers/md/dm-pcache/pcache_internal.h
101
seq_latest = meta->seq;
drivers/md/dm-pcache/pcache_internal.h
82
struct pcache_meta_header *meta, *latest = NULL;
drivers/md/dm-pcache/pcache_internal.h
86
meta = meta_ret;
drivers/md/dm-pcache/pcache_internal.h
90
if (copy_mc_to_kernel(meta, meta_addr, meta_size)) {
drivers/md/dm-pcache/pcache_internal.h
96
if (meta->crc != pcache_meta_crc(meta, meta_size))
drivers/media/common/videobuf2/videobuf2-v4l2.c
790
requested_sizes[0] = f->fmt.meta.buffersize;
drivers/media/i2c/ds90ub960.c
3576
if (ub960_fmt->meta) {
drivers/media/i2c/ds90ub960.c
624
bool meta;
drivers/media/i2c/et8ek8/et8ek8_driver.c
360
struct et8ek8_meta_reglist *meta,
drivers/media/i2c/et8ek8/et8ek8_driver.c
363
struct et8ek8_reglist **next = &meta->reglist[0].ptr;
drivers/media/i2c/et8ek8/et8ek8_driver.c
376
struct et8ek8_meta_reglist *meta,
drivers/media/i2c/et8ek8/et8ek8_driver.c
381
reglist = et8ek8_reglist_find_type(meta, type);
drivers/media/i2c/et8ek8/et8ek8_driver.c
389
struct et8ek8_meta_reglist *meta)
drivers/media/i2c/et8ek8/et8ek8_driver.c
391
return &meta->reglist[0].ptr;
drivers/media/i2c/et8ek8/et8ek8_driver.c
403
struct et8ek8_meta_reglist *meta,
drivers/media/i2c/et8ek8/et8ek8_driver.c
406
struct et8ek8_reglist **list = et8ek8_reglist_first(meta);
drivers/media/i2c/et8ek8/et8ek8_driver.c
457
struct et8ek8_meta_reglist *meta,
drivers/media/i2c/et8ek8/et8ek8_driver.c
462
struct et8ek8_reglist **list = et8ek8_reglist_first(meta);
drivers/media/i2c/et8ek8/et8ek8_driver.c
510
struct et8ek8_meta_reglist *meta)
drivers/media/i2c/et8ek8/et8ek8_driver.c
514
dev_info(&client->dev, "meta_reglist version %s\n", meta->version);
drivers/media/i2c/et8ek8/et8ek8_driver.c
516
while (meta->reglist[nlists].ptr)
drivers/media/i2c/et8ek8/et8ek8_driver.c
522
sort(&meta->reglist[0].ptr, nlists, sizeof(meta->reglist[0].ptr),
drivers/media/i2c/et8ek8/et8ek8_driver.c
531
list = meta->reglist[nlists].ptr;
drivers/media/i2c/et8ek8/et8ek8_driver.c
541
(void *)meta->reglist[nlists].ptr);
drivers/media/pci/intel/ipu6/ipu6-isys-video.c
1251
.fmt.meta = {
drivers/media/pci/intel/ipu6/ipu6-isys-video.c
1285
av->meta_fmt = format_meta.fmt.meta;
drivers/media/pci/intel/ipu6/ipu6-isys-video.c
237
f->fmt.meta = av->meta_fmt;
drivers/media/pci/intel/ipu6/ipu6-isys-video.c
304
ipu6_isys_try_fmt_cap(av, f->type, &f->fmt.meta.dataformat,
drivers/media/pci/intel/ipu6/ipu6-isys-video.c
305
&f->fmt.meta.width, &f->fmt.meta.height,
drivers/media/pci/intel/ipu6/ipu6-isys-video.c
306
&f->fmt.meta.bytesperline,
drivers/media/pci/intel/ipu6/ipu6-isys-video.c
307
&f->fmt.meta.buffersize);
drivers/media/pci/intel/ipu6/ipu6-isys-video.c
342
av->meta_fmt = f->fmt.meta;
drivers/media/platform/amlogic/c3/isp/c3-isp-params.c
699
f->fmt.meta = params->vfmt.fmt.meta;
drivers/media/platform/amlogic/c3/isp/c3-isp-params.c
773
params->vfmt.fmt.meta.buffersize);
drivers/media/platform/amlogic/c3/isp/c3-isp-params.c
795
buf->cfg = kvmalloc(params->vfmt.fmt.meta.buffersize, GFP_KERNEL);
drivers/media/platform/amlogic/c3/isp/c3-isp-params.c
843
params->vfmt.fmt.meta.dataformat = V4L2_META_FMT_C3ISP_PARAMS;
drivers/media/platform/amlogic/c3/isp/c3-isp-params.c
844
params->vfmt.fmt.meta.buffersize = sizeof(struct c3_isp_params_cfg);
drivers/media/platform/amlogic/c3/isp/c3-isp-stats.c
115
f->fmt.meta = stats->vfmt.fmt.meta;
drivers/media/platform/amlogic/c3/isp/c3-isp-stats.c
182
unsigned int size = stats->vfmt.fmt.meta.buffersize;
drivers/media/platform/amlogic/c3/isp/c3-isp-stats.c
244
stats->vfmt.fmt.meta.dataformat = V4L2_META_FMT_C3ISP_STATS;
drivers/media/platform/amlogic/c3/isp/c3-isp-stats.c
245
stats->vfmt.fmt.meta.buffersize = sizeof(struct c3_isp_stats_info);
drivers/media/platform/arm/mali-c55/mali-c55-params.c
493
f->fmt.meta = mfmt;
drivers/media/platform/arm/mali-c55/mali-c55-stats.c
52
f->fmt.meta = mfmt;
drivers/media/platform/broadcom/bcm2835-unicam.c
1571
: node->fmt.fmt.meta.buffersize;
drivers/media/platform/broadcom/bcm2835-unicam.c
1593
: node->fmt.fmt.meta.buffersize;
drivers/media/platform/broadcom/bcm2835-unicam.c
1958
f->fmt.meta = node->fmt.fmt.meta;
drivers/media/platform/broadcom/bcm2835-unicam.c
1964
__unicam_try_fmt_meta(struct unicam_node *node, struct v4l2_meta_format *meta)
drivers/media/platform/broadcom/bcm2835-unicam.c
1972
fmtinfo = unicam_find_format_by_fourcc(meta->dataformat,
drivers/media/platform/broadcom/bcm2835-unicam.c
1976
meta->dataformat = fmtinfo->fourcc;
drivers/media/platform/broadcom/bcm2835-unicam.c
1979
unicam_calc_meta_size_bpl(node->dev, fmtinfo, meta);
drivers/media/platform/broadcom/bcm2835-unicam.c
1989
__unicam_try_fmt_meta(node, &f->fmt.meta);
drivers/media/platform/broadcom/bcm2835-unicam.c
2001
__unicam_try_fmt_meta(node, &f->fmt.meta);
drivers/media/platform/broadcom/bcm2835-unicam.c
2171
const struct v4l2_meta_format *fmt = &node->fmt.fmt.meta;
drivers/media/platform/broadcom/bcm2835-unicam.c
2222
struct v4l2_meta_format *fmt = &node->fmt.fmt.meta;
drivers/media/platform/broadcom/bcm2835-unicam.c
585
struct v4l2_meta_format *meta)
drivers/media/platform/broadcom/bcm2835-unicam.c
587
v4l_bound_align_image(&meta->width, UNICAM_META_MIN_WIDTH,
drivers/media/platform/broadcom/bcm2835-unicam.c
589
&meta->height, UNICAM_META_MIN_HEIGHT,
drivers/media/platform/broadcom/bcm2835-unicam.c
592
meta->bytesperline = ALIGN(meta->width * fmtinfo->depth / 8,
drivers/media/platform/broadcom/bcm2835-unicam.c
594
meta->buffersize = meta->height * meta->bytesperline;
drivers/media/platform/chips-media/coda/coda-bit.c
1850
struct coda_buffer_meta *meta, *tmp;
drivers/media/platform/chips-media/coda/coda-bit.c
1857
list_for_each_entry_safe(meta, tmp, &ctx->buffer_meta_list, list) {
drivers/media/platform/chips-media/coda/coda-bit.c
1858
if (ctx->bitstream_fifo.kfifo.out >= meta->end) {
drivers/media/platform/chips-media/coda/coda-bit.c
1860
meta->sequence, meta->start, meta->end);
drivers/media/platform/chips-media/coda/coda-bit.c
1862
list_del(&meta->list);
drivers/media/platform/chips-media/coda/coda-bit.c
1865
kfree(meta);
drivers/media/platform/chips-media/coda/coda-bit.c
2169
struct coda_buffer_meta *meta;
drivers/media/platform/chips-media/coda/coda-bit.c
2266
meta = list_first_entry_or_null(&ctx->buffer_meta_list,
drivers/media/platform/chips-media/coda/coda-bit.c
2269
if (meta && ctx->codec->src_fourcc == V4L2_PIX_FMT_JPEG) {
drivers/media/platform/chips-media/coda/coda-bit.c
2272
if (meta->end == ctx->bitstream_fifo.kfifo.in) {
drivers/media/platform/chips-media/coda/coda-bit.c
2277
pad = ((0 - meta->end) & 0xff) + 256;
drivers/media/platform/chips-media/coda/coda-bit.c
2294
trace_coda_dec_pic_run(ctx, meta);
drivers/media/platform/chips-media/coda/coda-bit.c
2307
struct coda_buffer_meta *meta;
drivers/media/platform/chips-media/coda/coda-bit.c
2458
meta = list_first_entry(&ctx->buffer_meta_list,
drivers/media/platform/chips-media/coda/coda-bit.c
2460
list_del(&meta->list);
drivers/media/platform/chips-media/coda/coda-bit.c
2470
if ((sequence & 0xffff) != (meta->sequence & 0xffff)) {
drivers/media/platform/chips-media/coda/coda-bit.c
2474
meta->sequence);
drivers/media/platform/chips-media/coda/coda-bit.c
2476
decoded_frame->meta = *meta;
drivers/media/platform/chips-media/coda/coda-bit.c
2477
kfree(meta);
drivers/media/platform/chips-media/coda/coda-bit.c
2481
memset(&decoded_frame->meta, 0,
drivers/media/platform/chips-media/coda/coda-bit.c
2483
decoded_frame->meta.sequence = sequence;
drivers/media/platform/chips-media/coda/coda-bit.c
2484
decoded_frame->meta.last = false;
drivers/media/platform/chips-media/coda/coda-bit.c
2488
trace_coda_dec_pic_done(ctx, &decoded_frame->meta);
drivers/media/platform/chips-media/coda/coda-bit.c
2527
meta = &ready_frame->meta;
drivers/media/platform/chips-media/coda/coda-bit.c
2528
if (meta->last && !coda_reorder_enable(ctx)) {
drivers/media/platform/chips-media/coda/coda-bit.c
2545
dst_buf->timecode = meta->timecode;
drivers/media/platform/chips-media/coda/coda-bit.c
2546
dst_buf->vb2_buf.timestamp = meta->timestamp;
drivers/media/platform/chips-media/coda/coda-bit.c
2548
trace_coda_dec_rot_done(ctx, dst_buf, meta);
drivers/media/platform/chips-media/coda/coda-bit.c
2561
decoded_frame->meta.sequence,
drivers/media/platform/chips-media/coda/coda-bit.c
2563
ready_frame->meta.sequence,
drivers/media/platform/chips-media/coda/coda-bit.c
2571
ready_frame->meta.sequence,
drivers/media/platform/chips-media/coda/coda-bit.c
2580
decoded_frame->meta.sequence,
drivers/media/platform/chips-media/coda/coda-bit.c
326
struct coda_buffer_meta *meta;
drivers/media/platform/chips-media/coda/coda-bit.c
346
meta = list_first_entry(&ctx->buffer_meta_list,
drivers/media/platform/chips-media/coda/coda-bit.c
357
if (coda_bitstream_can_fetch_past(ctx, meta->end))
drivers/media/platform/chips-media/coda/coda-bit.c
400
meta = kmalloc_obj(*meta);
drivers/media/platform/chips-media/coda/coda-bit.c
401
if (meta) {
drivers/media/platform/chips-media/coda/coda-bit.c
402
meta->sequence = src_buf->sequence;
drivers/media/platform/chips-media/coda/coda-bit.c
403
meta->timecode = src_buf->timecode;
drivers/media/platform/chips-media/coda/coda-bit.c
404
meta->timestamp = src_buf->vb2_buf.timestamp;
drivers/media/platform/chips-media/coda/coda-bit.c
405
meta->start = start;
drivers/media/platform/chips-media/coda/coda-bit.c
406
meta->end = ctx->bitstream_fifo.kfifo.in;
drivers/media/platform/chips-media/coda/coda-bit.c
407
meta->last = src_buf->flags & V4L2_BUF_FLAG_LAST;
drivers/media/platform/chips-media/coda/coda-bit.c
408
if (meta->last)
drivers/media/platform/chips-media/coda/coda-bit.c
411
list_add_tail(&meta->list,
drivers/media/platform/chips-media/coda/coda-bit.c
416
trace_coda_bit_queue(ctx, src_buf, meta);
drivers/media/platform/chips-media/coda/coda-common.c
1160
struct coda_buffer_meta *meta;
drivers/media/platform/chips-media/coda/coda-common.c
1170
meta = list_last_entry(&ctx->buffer_meta_list, struct coda_buffer_meta,
drivers/media/platform/chips-media/coda/coda-common.c
1172
meta->last = true;
drivers/media/platform/chips-media/coda/coda-common.c
1592
struct coda_buffer_meta *meta;
drivers/media/platform/chips-media/coda/coda-common.c
1617
meta = list_first_entry(&ctx->buffer_meta_list,
drivers/media/platform/chips-media/coda/coda-common.c
1619
if (!coda_bitstream_can_fetch_past(ctx, meta->end) &&
drivers/media/platform/chips-media/coda/coda-common.c
1623
meta->end, ctx->bitstream_fifo.kfifo.in);
drivers/media/platform/chips-media/coda/coda-common.c
2143
struct coda_buffer_meta *meta;
drivers/media/platform/chips-media/coda/coda-common.c
2151
meta = list_first_entry(&ctx->buffer_meta_list,
drivers/media/platform/chips-media/coda/coda-common.c
2153
list_del(&meta->list);
drivers/media/platform/chips-media/coda/coda-common.c
2154
kfree(meta);
drivers/media/platform/chips-media/coda/coda.h
218
struct coda_buffer_meta meta;
drivers/media/platform/chips-media/coda/trace.h
100
__entry->start = meta->start & ctx->bitstream_fifo.kfifo.mask;
drivers/media/platform/chips-media/coda/trace.h
101
__entry->end = meta->end & ctx->bitstream_fifo.kfifo.mask;
drivers/media/platform/chips-media/coda/trace.h
112
struct coda_buffer_meta *meta),
drivers/media/platform/chips-media/coda/trace.h
113
TP_ARGS(ctx, buf, meta)
drivers/media/platform/chips-media/coda/trace.h
117
TP_PROTO(struct coda_ctx *ctx, struct coda_buffer_meta *meta),
drivers/media/platform/chips-media/coda/trace.h
119
TP_ARGS(ctx, meta),
drivers/media/platform/chips-media/coda/trace.h
130
__entry->start = meta ? (meta->start &
drivers/media/platform/chips-media/coda/trace.h
132
__entry->end = meta ? (meta->end &
drivers/media/platform/chips-media/coda/trace.h
142
TP_PROTO(struct coda_ctx *ctx, struct coda_buffer_meta *meta),
drivers/media/platform/chips-media/coda/trace.h
143
TP_ARGS(ctx, meta)
drivers/media/platform/chips-media/coda/trace.h
147
TP_PROTO(struct coda_ctx *ctx, struct coda_buffer_meta *meta),
drivers/media/platform/chips-media/coda/trace.h
148
TP_ARGS(ctx, meta)
drivers/media/platform/chips-media/coda/trace.h
153
struct coda_buffer_meta *meta),
drivers/media/platform/chips-media/coda/trace.h
154
TP_ARGS(ctx, buf, meta)
drivers/media/platform/chips-media/coda/trace.h
85
struct coda_buffer_meta *meta),
drivers/media/platform/chips-media/coda/trace.h
87
TP_ARGS(ctx, buf, meta),
drivers/media/platform/raspberrypi/pisp_be/pisp_be.c
1204
f->fmt.meta.dataformat = V4L2_META_FMT_RPI_BE_CFG;
drivers/media/platform/raspberrypi/pisp_be/pisp_be.c
1205
f->fmt.meta.buffersize = sizeof(struct pisp_be_tiles_config);
drivers/media/platform/raspberrypi/pisp_be/pisp_be.c
1274
NODE_NAME(node), &f->fmt.meta.dataformat);
drivers/media/platform/raspberrypi/pisp_be/pisp_be.c
1373
f->fmt.meta.dataformat = V4L2_META_FMT_RPI_BE_CFG;
drivers/media/platform/raspberrypi/pisp_be/pisp_be.c
1374
f->fmt.meta.buffersize = sizeof(struct pisp_be_tiles_config);
drivers/media/platform/raspberrypi/pisp_be/pisp_be.c
792
node->format.fmt.meta.buffersize;
drivers/media/platform/raspberrypi/pisp_be/pisp_be.c
805
node->format.fmt.meta.buffersize;
drivers/media/platform/raspberrypi/pisp_be/pisp_be.c
826
node->format.fmt.meta.buffersize;
drivers/media/platform/raspberrypi/rp1-cfe/cfe.c
1022
node->meta_fmt.fmt.meta.buffersize;
drivers/media/platform/raspberrypi/rp1-cfe/cfe.c
1051
node->meta_fmt.fmt.meta.buffersize;
drivers/media/platform/raspberrypi/rp1-cfe/cfe.c
1510
node_desc[node->id].name, f->fmt.meta.width,
drivers/media/platform/raspberrypi/rp1-cfe/cfe.c
1511
f->fmt.meta.height, &f->fmt.meta.dataformat);
drivers/media/platform/raspberrypi/rp1-cfe/cfe.c
1516
node_desc[node->id].name, f->fmt.meta.buffersize,
drivers/media/platform/raspberrypi/rp1-cfe/cfe.c
1517
&f->fmt.meta.dataformat);
drivers/media/platform/raspberrypi/rp1-cfe/cfe.c
1528
fmt = find_format_by_pix(f->fmt.meta.dataformat);
drivers/media/platform/raspberrypi/rp1-cfe/cfe.c
1532
f->fmt.meta.dataformat = fmt->fourcc;
drivers/media/platform/raspberrypi/rp1-cfe/cfe.c
1538
f->fmt.meta.dataformat = V4L2_META_FMT_RPI_FE_STATS;
drivers/media/platform/raspberrypi/rp1-cfe/cfe.c
1539
f->fmt.meta.buffersize = sizeof(struct pisp_statistics);
drivers/media/platform/raspberrypi/rp1-cfe/cfe.c
1542
f->fmt.meta.dataformat = V4L2_META_FMT_RPI_FE_CFG;
drivers/media/platform/raspberrypi/rp1-cfe/cfe.c
1543
f->fmt.meta.buffersize = sizeof(struct pisp_fe_config);
drivers/media/platform/raspberrypi/rp1-cfe/cfe.c
1587
&node->meta_fmt.fmt.meta.dataformat);
drivers/media/platform/raspberrypi/rp1-cfe/cfe.c
1818
struct v4l2_meta_format *meta_fmt = &node->meta_fmt.fmt.meta;
drivers/media/platform/raspberrypi/rp1-cfe/cfe.c
511
v4l_bound_align_image(&f->fmt.meta.width, MIN_META_WIDTH, MAX_WIDTH, 2,
drivers/media/platform/raspberrypi/rp1-cfe/cfe.c
512
&f->fmt.meta.height, MIN_META_HEIGHT, MAX_HEIGHT,
drivers/media/platform/raspberrypi/rp1-cfe/cfe.c
515
f->fmt.meta.bytesperline = (f->fmt.meta.width * fmt->depth) >> 3;
drivers/media/platform/raspberrypi/rp1-cfe/cfe.c
516
f->fmt.meta.buffersize = f->fmt.meta.height * f->fmt.pix.bytesperline;
drivers/media/platform/raspberrypi/rp1-cfe/cfe.c
519
&f->fmt.meta.dataformat, f->fmt.meta.width, f->fmt.meta.height,
drivers/media/platform/raspberrypi/rp1-cfe/cfe.c
520
f->fmt.meta.bytesperline, f->fmt.meta.buffersize);
drivers/media/platform/raspberrypi/rp1-cfe/cfe.c
543
size = node->meta_fmt.fmt.meta.buffersize;
drivers/media/platform/renesas/vsp1/vsp1_histo.c
428
struct v4l2_meta_format *meta = &format->fmt.meta;
drivers/media/platform/renesas/vsp1/vsp1_histo.c
433
memset(meta, 0, sizeof(*meta));
drivers/media/platform/renesas/vsp1/vsp1_histo.c
435
meta->dataformat = histo->meta_format;
drivers/media/platform/renesas/vsp1/vsp1_histo.c
436
meta->buffersize = histo->data_size;
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2522
struct v4l2_meta_format *meta = &f->fmt.meta;
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2527
*meta = *params->metafmt;
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2536
struct v4l2_meta_format *meta = &f->fmt.meta;
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2541
*meta = *rkisp1_params_get_format_info(meta->dataformat);
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2551
struct v4l2_meta_format *meta = &f->fmt.meta;
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2559
params->metafmt = rkisp1_params_get_format_info(meta->dataformat);
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2560
*meta = *params->metafmt;
drivers/media/platform/rockchip/rkisp1/rkisp1-stats.c
30
f->pixelformat = stats->vdev_fmt.fmt.meta.dataformat;
drivers/media/platform/rockchip/rkisp1/rkisp1-stats.c
39
struct v4l2_meta_format *meta = &f->fmt.meta;
drivers/media/platform/rockchip/rkisp1/rkisp1-stats.c
400
stats->vdev_fmt.fmt.meta.dataformat =
drivers/media/platform/rockchip/rkisp1/rkisp1-stats.c
402
stats->vdev_fmt.fmt.meta.buffersize =
drivers/media/platform/rockchip/rkisp1/rkisp1-stats.c
44
memset(meta, 0, sizeof(*meta));
drivers/media/platform/rockchip/rkisp1/rkisp1-stats.c
45
meta->dataformat = stats->vdev_fmt.fmt.meta.dataformat;
drivers/media/platform/rockchip/rkisp1/rkisp1-stats.c
46
meta->buffersize = stats->vdev_fmt.fmt.meta.buffersize;
drivers/media/platform/ti/cal/cal-video.c
440
if (cal_formats[i].meta)
drivers/media/platform/ti/cal/cal-video.c
470
if (!fmtinfo || fmtinfo->meta)
drivers/media/platform/ti/cal/cal.h
91
bool meta;
drivers/media/test-drivers/vivid/vivid-meta-cap.c
147
struct v4l2_meta_format *meta = &f->fmt.meta;
drivers/media/test-drivers/vivid/vivid-meta-cap.c
152
meta->dataformat = V4L2_META_FMT_UVC;
drivers/media/test-drivers/vivid/vivid-meta-cap.c
153
meta->buffersize = sizeof(struct vivid_uvc_meta_buf);
drivers/media/test-drivers/vivid/vivid-meta-cap.c
160
struct vivid_uvc_meta_buf *meta = vb2_plane_vaddr(&buf->vb.vb2_buf, 0);
drivers/media/test-drivers/vivid/vivid-meta-cap.c
166
memset(meta, 1, vb2_plane_size(&buf->vb.vb2_buf, 0));
drivers/media/test-drivers/vivid/vivid-meta-cap.c
168
meta->ns = ktime_get_ns();
drivers/media/test-drivers/vivid/vivid-meta-cap.c
169
meta->sof = buf->vb.sequence * 30;
drivers/media/test-drivers/vivid/vivid-meta-cap.c
170
meta->length = sizeof(*meta) - offsetof(struct vivid_uvc_meta_buf, length);
drivers/media/test-drivers/vivid/vivid-meta-cap.c
171
meta->flags = UVC_STREAM_EOH | UVC_STREAM_EOF;
drivers/media/test-drivers/vivid/vivid-meta-cap.c
174
meta->flags |= UVC_STREAM_FID;
drivers/media/test-drivers/vivid/vivid-meta-cap.c
177
__func__, meta->ns, meta->sof, meta->length, meta->flags);
drivers/media/test-drivers/vivid/vivid-meta-cap.c
179
meta->flags |= UVC_STREAM_PTS;
drivers/media/test-drivers/vivid/vivid-meta-cap.c
180
meta->buf[0] = div_u64(soe, VIVID_META_CLOCK_UNIT);
drivers/media/test-drivers/vivid/vivid-meta-cap.c
182
dprintk(dev, 2, " pts: %u\n", *(__u32 *)(meta->buf));
drivers/media/test-drivers/vivid/vivid-meta-cap.c
186
meta->flags |= UVC_STREAM_SCR;
drivers/media/test-drivers/vivid/vivid-meta-cap.c
187
meta->buf[buf_off] = div_u64((soe + dev->cap_frame_eof_offset),
drivers/media/test-drivers/vivid/vivid-meta-cap.c
190
meta->buf[buf_off + 4] = (buf->vb.sequence * 30) % 1000;
drivers/media/test-drivers/vivid/vivid-meta-cap.c
192
*(__u32 *)(meta->buf + buf_off),
drivers/media/test-drivers/vivid/vivid-meta-cap.c
193
*(__u16 *)(meta->buf + buf_off + 4));
drivers/media/test-drivers/vivid/vivid-meta-out.c
147
struct v4l2_meta_format *meta = &f->fmt.meta;
drivers/media/test-drivers/vivid/vivid-meta-out.c
152
meta->dataformat = V4L2_META_FMT_VIVID;
drivers/media/test-drivers/vivid/vivid-meta-out.c
153
meta->buffersize = sizeof(struct vivid_meta_out_buf);
drivers/media/test-drivers/vivid/vivid-meta-out.c
160
struct vivid_meta_out_buf *meta = vb2_plane_vaddr(&buf->vb.vb2_buf, 0);
drivers/media/test-drivers/vivid/vivid-meta-out.c
162
v4l2_ctrl_s_ctrl(dev->brightness, meta->brightness);
drivers/media/test-drivers/vivid/vivid-meta-out.c
163
v4l2_ctrl_s_ctrl(dev->contrast, meta->contrast);
drivers/media/test-drivers/vivid/vivid-meta-out.c
164
v4l2_ctrl_s_ctrl(dev->saturation, meta->saturation);
drivers/media/test-drivers/vivid/vivid-meta-out.c
165
v4l2_ctrl_s_ctrl(dev->hue, meta->hue);
drivers/media/test-drivers/vivid/vivid-meta-out.c
168
__func__, meta->brightness, meta->contrast,
drivers/media/test-drivers/vivid/vivid-meta-out.c
169
meta->saturation, meta->hue);
drivers/media/usb/uvc/uvc_driver.c
2005
vb2_video_unregister_device(&stream->meta.queue.vdev);
drivers/media/usb/uvc/uvc_metadata.c
102
if (vb2_is_busy(&stream->meta.queue.queue))
drivers/media/usb/uvc/uvc_metadata.c
105
stream->meta.format = fmt->dataformat;
drivers/media/usb/uvc/uvc_metadata.c
229
struct uvc_video_queue *queue = &stream->meta.queue;
drivers/media/usb/uvc/uvc_metadata.c
231
stream->meta.format = V4L2_META_FMT_UVC;
drivers/media/usb/uvc/uvc_metadata.c
47
struct v4l2_meta_format *fmt = &format->fmt.meta;
drivers/media/usb/uvc/uvc_metadata.c
52
fmt->dataformat = stream->meta.format;
drivers/media/usb/uvc/uvc_metadata.c
64
struct v4l2_meta_format *fmt = &format->fmt.meta;
drivers/media/usb/uvc/uvc_metadata.c
90
struct v4l2_meta_format *fmt = &format->fmt.meta;
drivers/media/usb/uvc/uvc_video.c
1419
struct uvc_meta_buf *meta;
drivers/media/usb/uvc/uvc_video.c
1443
if (stream->meta.format == V4L2_META_FMT_UVC)
drivers/media/usb/uvc/uvc_video.c
1451
length + sizeof(meta->ns) + sizeof(meta->sof)) {
drivers/media/usb/uvc/uvc_video.c
1456
meta = (struct uvc_meta_buf *)((u8 *)meta_buf->mem + meta_buf->bytesused);
drivers/media/usb/uvc/uvc_video.c
1461
put_unaligned(ktime_to_ns(time), &meta->ns);
drivers/media/usb/uvc/uvc_video.c
1462
put_unaligned(sof, &meta->sof);
drivers/media/usb/uvc/uvc_video.c
1467
meta->length = mem[0];
drivers/media/usb/uvc/uvc_video.c
1468
meta->flags = mem[1];
drivers/media/usb/uvc/uvc_video.c
1469
memcpy(meta->buf, &mem[2], length - 2);
drivers/media/usb/uvc/uvc_video.c
1470
meta_buf->bytesused += length + sizeof(meta->ns) + sizeof(meta->sof);
drivers/media/usb/uvc/uvc_video.c
1474
__func__, ktime_to_ns(time), meta->sof, meta->length,
drivers/media/usb/uvc/uvc_video.c
1475
meta->flags,
drivers/media/usb/uvc/uvc_video.c
1476
has_pts ? *(u32 *)meta->buf : 0,
drivers/media/usb/uvc/uvc_video.c
1516
*meta_buf = uvc_queue_next_buffer(&stream->meta.queue,
drivers/media/usb/uvc/uvc_video.c
1692
struct uvc_video_queue *qmeta = &stream->meta.queue;
drivers/media/usb/uvc/uvc_video.c
1693
struct vb2_queue *vb2_qmeta = stream->meta.queue.vdev.queue;
drivers/media/usb/uvc/uvcvideo.h
485
} meta;
drivers/media/v4l2-core/v4l2-compat-ioctl32.c
103
struct v4l2_meta_format meta;
drivers/media/v4l2-core/v4l2-compat-ioctl32.c
167
return copy_from_user(&p64->fmt.meta, &p32->fmt.meta,
drivers/media/v4l2-core/v4l2-compat-ioctl32.c
168
sizeof(p64->fmt.meta)) ? -EFAULT : 0;
drivers/media/v4l2-core/v4l2-compat-ioctl32.c
217
return copy_to_user(&p32->fmt.meta, &p64->fmt.meta,
drivers/media/v4l2-core/v4l2-compat-ioctl32.c
218
sizeof(p64->fmt.meta)) ? -EFAULT : 0;
drivers/media/v4l2-core/v4l2-ioctl.c
1843
memset_after(p, 0, fmt.meta);
drivers/media/v4l2-core/v4l2-ioctl.c
1848
memset_after(p, 0, fmt.meta);
drivers/media/v4l2-core/v4l2-ioctl.c
1951
memset_after(p, 0, fmt.meta);
drivers/media/v4l2-core/v4l2-ioctl.c
1956
memset_after(p, 0, fmt.meta);
drivers/media/v4l2-core/v4l2-ioctl.c
278
const struct v4l2_meta_format *meta;
drivers/media/v4l2-core/v4l2-ioctl.c
344
meta = &p->fmt.meta;
drivers/media/v4l2-core/v4l2-ioctl.c
345
pixelformat = meta->dataformat;
drivers/media/v4l2-core/v4l2-ioctl.c
347
&pixelformat, meta->buffersize, meta->width,
drivers/media/v4l2-core/v4l2-ioctl.c
348
meta->height, meta->bytesperline);
drivers/misc/mei/client.c
1741
struct mei_ext_meta_hdr *meta;
drivers/misc/mei/client.c
1761
hdr_len += sizeof(*meta);
drivers/misc/mei/client.c
1781
meta = (struct mei_ext_meta_hdr *)mei_hdr->extension;
drivers/misc/mei/client.c
1782
meta->size = 0;
drivers/misc/mei/client.c
1783
next_ext = (struct mei_ext_hdr *)meta->hdrs;
drivers/misc/mei/client.c
1785
meta->count++;
drivers/misc/mei/client.c
1786
meta->size += mei_ext_hdr_set_vtag(next_ext, cb->vtag);
drivers/misc/mei/client.c
1791
meta->count++;
drivers/misc/mei/client.c
1792
meta->size += mei_ext_hdr_set_gsc(next_ext, cb->ext_hdr);
drivers/misc/mei/hw.h
295
static inline struct mei_ext_hdr *mei_ext_begin(struct mei_ext_meta_hdr *meta)
drivers/misc/mei/hw.h
297
return (struct mei_ext_hdr *)meta->hdrs;
drivers/misc/mei/hw.h
308
static inline bool mei_ext_last(struct mei_ext_meta_hdr *meta,
drivers/misc/mei/hw.h
311
return (u8 *)ext >= (u8 *)meta + sizeof(*meta) + (meta->size * 4);
drivers/misc/mei/interrupt.c
111
ext_len = sizeof(*meta) + mei_slots2data(meta->size);
drivers/misc/mei/interrupt.c
128
struct mei_ext_hdr *ext = mei_ext_begin(meta);
drivers/misc/mei/interrupt.c
151
} while (!mei_ext_last(meta, ext));
drivers/misc/mei/interrupt.c
95
struct mei_ext_meta_hdr *meta,
drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c
1427
int last, int meta)
drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c
1437
status = this->auxiliary_virt + ALIGN(meta, 4);
drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c
1592
int meta, n, page_size;
drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c
1632
meta = geo->metadata_size;
drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c
1635
col = meta + ecc_parity_size
drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c
1638
col = meta + (size + ecc_parity_size) * first;
drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c
1640
meta = 0;
drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c
1647
if (geo->ecc_for_meta && meta)
drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c
1648
page_size = meta + ecc_parity_size
drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c
1651
page_size = meta + (size + ecc_parity_size) * n;
drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c
1657
BF_BCH_FLASH0LAYOUT0_META_SIZE(meta) |
drivers/mtd/nand/raw/gpmi-nand/gpmi-nand.c
1677
max_bitflips = gpmi_count_bitflips(chip, buf, first, last, meta);
drivers/net/dsa/hirschmann/hellcreek.c
717
u16 meta = 0;
drivers/net/dsa/hirschmann/hellcreek.c
730
meta |= entry->portmask << HR_FDBWRM0_PORTMASK_SHIFT;
drivers/net/dsa/hirschmann/hellcreek.c
732
meta |= HR_FDBWRM0_OBT;
drivers/net/dsa/hirschmann/hellcreek.c
734
meta |= HR_FDBWRM0_PASS_BLOCKED;
drivers/net/dsa/hirschmann/hellcreek.c
736
meta |= HR_FDBWRM0_REPRIO_EN;
drivers/net/dsa/hirschmann/hellcreek.c
737
meta |= entry->reprio_tc << HR_FDBWRM0_REPRIO_TC_SHIFT;
drivers/net/dsa/hirschmann/hellcreek.c
739
hellcreek_write(hellcreek, meta, HR_FDBWRM0);
drivers/net/dsa/hirschmann/hellcreek.c
765
u16 meta, mac;
drivers/net/dsa/hirschmann/hellcreek.c
768
meta = hellcreek_read(hellcreek, HR_FDBMDRD);
drivers/net/dsa/hirschmann/hellcreek.c
782
entry->portmask = (meta & HR_FDBMDRD_PORTMASK_MASK) >>
drivers/net/dsa/hirschmann/hellcreek.c
784
entry->age = (meta & HR_FDBMDRD_AGE_MASK) >>
drivers/net/dsa/hirschmann/hellcreek.c
786
entry->is_obt = !!(meta & HR_FDBMDRD_OBT);
drivers/net/dsa/hirschmann/hellcreek.c
787
entry->pass_blocked = !!(meta & HR_FDBMDRD_PASS_BLOCKED);
drivers/net/dsa/hirschmann/hellcreek.c
788
entry->is_static = !!(meta & HR_FDBMDRD_STATIC);
drivers/net/dsa/hirschmann/hellcreek.c
789
entry->reprio_tc = (meta & HR_FDBMDRD_REPRIO_TC_MASK) >>
drivers/net/dsa/hirschmann/hellcreek.c
791
entry->reprio_en = !!(meta & HR_FDBMDRD_REPRIO_EN);
drivers/net/ethernet/intel/ice/ice_ddp.c
1947
struct ice_meta_sect *meta;
drivers/net/ethernet/intel/ice/ice_ddp.c
1953
meta = ice_pkg_enum_section((struct ice_seg *)seg_hdr, &state,
drivers/net/ethernet/intel/ice/ice_ddp.c
1955
if (!meta) {
drivers/net/ethernet/intel/ice/ice_ddp.c
1961
hw->pkg_ver = meta->ver;
drivers/net/ethernet/intel/ice/ice_ddp.c
1962
memcpy(hw->pkg_name, meta->name, sizeof(meta->name));
drivers/net/ethernet/intel/ice/ice_ddp.c
1965
meta->ver.major, meta->ver.minor, meta->ver.update,
drivers/net/ethernet/intel/ice/ice_ddp.c
1966
meta->ver.draft, meta->name);
drivers/net/ethernet/intel/igc/igc.h
603
struct xsk_tx_metadata *meta;
drivers/net/ethernet/intel/igc/igc_main.c
2967
xsk_tx_metadata_to_compl(meta_req->meta, &tstamp->xsk_meta);
drivers/net/ethernet/intel/igc/igc_main.c
3065
struct xsk_tx_metadata *meta = NULL;
drivers/net/ethernet/intel/igc/igc_main.c
3077
meta = xsk_buff_get_metadata(pool, xdp_desc.addr);
drivers/net/ethernet/intel/igc/igc_main.c
3083
meta_req.meta = meta;
drivers/net/ethernet/intel/igc/igc_main.c
3085
xsk_tx_metadata_request(meta, &igc_xsk_tx_metadata_ops,
drivers/net/ethernet/mellanox/mlx5/core/en/tc_ct.c
724
struct flow_action_entry *meta;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_ct.c
729
meta = mlx5_tc_ct_get_ct_metadata_action(flow_rule);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_ct.c
730
if (!meta)
drivers/net/ethernet/mellanox/mlx5/core/en/tc_ct.c
732
ctinfo = meta->ct_metadata.cookie & NFCT_INFOMASK;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_ct.c
734
err = mlx5_get_label_mapping(ct_priv, meta->ct_metadata.labels,
drivers/net/ethernet/mellanox/mlx5/core/en/tc_ct.c
750
ct_state |= meta->ct_metadata.orig_dir ? 0 : MLX5_CT_STATE_REPLY_BIT;
drivers/net/ethernet/mellanox/mlx5/core/en/tc_ct.c
753
meta->ct_metadata.mark,
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
448
int check_result, struct xsk_tx_metadata *meta);
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
452
int check_result, struct xsk_tx_metadata *meta)
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
470
return mlx5e_xmit_xdp_frame(sq, xdptxd, 0, meta);
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
500
xsk_tx_metadata_request(meta, &mlx5e_xsk_tx_metadata_ops, &session->wqe->eth);
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
531
int check_result, struct xsk_tx_metadata *meta)
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
645
xsk_tx_metadata_request(meta, &mlx5e_xsk_tx_metadata_ops, eseg);
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.h
116
struct xsk_tx_metadata *meta));
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.h
120
struct xsk_tx_metadata *meta));
drivers/net/ethernet/mellanox/mlx5/core/en/xsk/tx.c
102
meta = xsk_buff_get_metadata(pool, desc.addr);
drivers/net/ethernet/mellanox/mlx5/core/en/xsk/tx.c
108
check_result, meta);
drivers/net/ethernet/mellanox/mlx5/core/en/xsk/tx.c
119
xsk_tx_metadata_to_compl(meta, &compl);
drivers/net/ethernet/mellanox/mlx5/core/en/xsk/tx.c
69
struct xsk_tx_metadata *meta = NULL;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
197
struct skb_shared_info *shinfo, __le64 *meta,
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
219
*meta |= cpu_to_le64(FIELD_PREP(FBNIC_TWD_L3_OHLEN_MASK,
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
242
*meta |= cpu_to_le64(FIELD_PREP(FBNIC_TWD_L3_TYPE_MASK, l3_type) |
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
259
fbnic_tx_offloads(struct fbnic_ring *ring, struct sk_buff *skb, __le64 *meta)
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
265
*meta |= cpu_to_le64(FBNIC_TWD_FLAG_REQ_TS);
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
273
*meta |= cpu_to_le64(FIELD_PREP(FBNIC_TWD_CSUM_OFFSET_MASK,
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
277
if (fbnic_tx_lso(ring, skb, shinfo, meta, &l2len, &i3len))
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
280
*meta |= cpu_to_le64(FBNIC_TWD_FLAG_REQ_CSO);
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
286
*meta |= cpu_to_le64(FIELD_PREP(FBNIC_TWD_L2_HLEN_MASK, l2len / 2) |
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
314
fbnic_tx_map(struct fbnic_ring *ring, struct sk_buff *skb, __le64 *meta)
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
370
FBNIC_XMIT_CB(skb)->desc_count = ((twd - meta) + 1) & ring->size_mask;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
381
*meta |= cpu_to_le64(FBNIC_TWD_FLAG_REQ_COMPLETION);
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
414
__le64 *meta = &ring->desc[ring->tail];
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
430
*meta = cpu_to_le64(FBNIC_TWD_FLAG_DEST_MAC);
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
438
if (fbnic_tx_offloads(ring, skb, meta))
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
441
if (fbnic_tx_map(ring, skb, meta))
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1161
mem_op_stack(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1165
s32 off = nfp_prog->stack_frame_depth + meta->insn.off + ptr_off;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1174
if (meta->ptr_not_const ||
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1175
meta->flags & FLAG_INSN_PTR_CALLER_STACK_FRAME) {
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1180
stack_off_reg = ur_load_imm_any(nfp_prog, meta->insn.off,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1223
nop_cnt = narrow_ld && meta->flags & FLAG_INSN_DO_ZEXT ? 2 : 3;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1228
wrp_zext(nfp_prog, meta, gpr);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1294
wrp_alu64_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1297
const struct bpf_insn *insn = &meta->insn;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1301
meta->flags |= FLAG_INSN_SKIP_NOOP;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1312
wrp_alu64_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1315
u8 dst = meta->insn.dst_reg * 2, src = meta->insn.src_reg * 2;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1325
wrp_alu32_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1328
const struct bpf_insn *insn = &meta->insn;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1332
wrp_zext(nfp_prog, meta, dst);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1338
wrp_alu32_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1341
u8 dst = meta->insn.dst_reg * 2, src = meta->insn.src_reg * 2;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1344
wrp_zext(nfp_prog, meta, dst);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1358
wrp_test_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1361
const struct bpf_insn *insn = &meta->insn;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1365
if (is_mbpf_jmp64(meta))
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1386
static const struct jmp_code_map *nfp_jmp_code_get(struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1390
op = BPF_OP(meta->insn.code) >> 4;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1400
static int cmp_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1402
const struct bpf_insn *insn = &meta->insn;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1409
code = nfp_jmp_code_get(meta);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1413
alu_op = meta->jump_neg_op ? ALU_OP_ADD : ALU_OP_SUB;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1414
carry_op = meta->jump_neg_op ? ALU_OP_ADD_C : ALU_OP_SUB_C;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1422
if (is_mbpf_jmp64(meta)) {
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1437
static int cmp_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1439
const struct bpf_insn *insn = &meta->insn;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1443
code = nfp_jmp_code_get(meta);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1457
if (is_mbpf_jmp64(meta))
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1501
wrp_mul(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1505
const struct bpf_insn *insn = &meta->insn;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1513
lopnd_max = meta->umax_dst;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1516
ropnd_max = meta->umax_src;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1600
static int adjust_head(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1610
if (WARN_ON_ONCE(nfp_prog->adjust_head_location != meta->n))
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1680
static int adjust_tail(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1726
map_call_stack_common(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1734
lm_off += meta->arg2.reg.var_off.value + meta->arg2.reg.off;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1735
load_lm_ptr = meta->arg2.var_off || lm_off;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1740
if (meta->func_id == BPF_FUNC_map_update_elem)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1743
emit_br_relo(nfp_prog, BR_UNC, BR_OFF_RELO + meta->func_id,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1767
nfp_get_prandom_u32(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1779
nfp_perf_event_output(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1784
ptr_type = ur_load_imm_any(nfp_prog, meta->arg1.type, imm_a(nfp_prog));
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1788
emit_br_relo(nfp_prog, BR_UNC, BR_OFF_RELO + meta->func_id,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1804
nfp_queue_select(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1811
emit_alu(nfp_prog, reg_none(), reg_a(meta->insn.src_reg * 2),
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1820
pv_qsel_val(nfp_prog), 0x1, reg_b(meta->insn.src_reg * 2),
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1836
static int mov_reg64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1838
const struct bpf_insn *insn = &meta->insn;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1859
static int mov_imm64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1861
u64 imm = meta->insn.imm; /* sign extend */
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1863
wrp_immed(nfp_prog, reg_both(meta->insn.dst_reg * 2), imm & ~0U);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1864
wrp_immed(nfp_prog, reg_both(meta->insn.dst_reg * 2 + 1), imm >> 32);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1869
static int xor_reg64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1871
return wrp_alu64_reg(nfp_prog, meta, ALU_OP_XOR);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1874
static int xor_imm64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1876
return wrp_alu64_imm(nfp_prog, meta, ALU_OP_XOR, !meta->insn.imm);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1879
static int and_reg64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1881
return wrp_alu64_reg(nfp_prog, meta, ALU_OP_AND);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1884
static int and_imm64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1886
return wrp_alu64_imm(nfp_prog, meta, ALU_OP_AND, !~meta->insn.imm);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1889
static int or_reg64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1891
return wrp_alu64_reg(nfp_prog, meta, ALU_OP_OR);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1894
static int or_imm64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1896
return wrp_alu64_imm(nfp_prog, meta, ALU_OP_OR, !meta->insn.imm);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1899
static int add_reg64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1901
const struct bpf_insn *insn = &meta->insn;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1913
static int add_imm64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1915
const struct bpf_insn *insn = &meta->insn;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1924
static int sub_reg64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1926
const struct bpf_insn *insn = &meta->insn;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1938
static int sub_imm64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1940
const struct bpf_insn *insn = &meta->insn;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1949
static int mul_reg64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1951
return wrp_mul(nfp_prog, meta, true, true);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1954
static int mul_imm64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1956
return wrp_mul(nfp_prog, meta, true, false);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1959
static int div_imm64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1961
const struct bpf_insn *insn = &meta->insn;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1966
static int div_reg64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1971
return wrp_div_imm(nfp_prog, meta->insn.dst_reg * 2, meta->umin_src);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1974
static int neg_reg64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
1976
const struct bpf_insn *insn = &meta->insn;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2019
static int shl_imm64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2021
const struct bpf_insn *insn = &meta->insn;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2058
static int shl_reg64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2060
const struct bpf_insn *insn = &meta->insn;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2065
umin = meta->umin_src;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2066
umax = meta->umax_src;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2133
static int shr_imm64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2135
const struct bpf_insn *insn = &meta->insn;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2170
static int shr_reg64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2172
const struct bpf_insn *insn = &meta->insn;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2177
umin = meta->umin_src;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2178
umax = meta->umax_src;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2245
static int ashr_imm64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2247
const struct bpf_insn *insn = &meta->insn;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2287
static int ashr_reg64(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2289
const struct bpf_insn *insn = &meta->insn;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2294
umin = meta->umin_src;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2295
umax = meta->umax_src;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2326
static int mov_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2328
const struct bpf_insn *insn = &meta->insn;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2336
static int mov_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2338
const struct bpf_insn *insn = &meta->insn;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2346
static int xor_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2348
return wrp_alu32_reg(nfp_prog, meta, ALU_OP_XOR);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2351
static int xor_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2353
return wrp_alu32_imm(nfp_prog, meta, ALU_OP_XOR);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2356
static int and_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2358
return wrp_alu32_reg(nfp_prog, meta, ALU_OP_AND);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2361
static int and_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2363
return wrp_alu32_imm(nfp_prog, meta, ALU_OP_AND);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2366
static int or_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2368
return wrp_alu32_reg(nfp_prog, meta, ALU_OP_OR);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2371
static int or_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2373
return wrp_alu32_imm(nfp_prog, meta, ALU_OP_OR);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2376
static int add_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2378
return wrp_alu32_reg(nfp_prog, meta, ALU_OP_ADD);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2381
static int add_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2383
return wrp_alu32_imm(nfp_prog, meta, ALU_OP_ADD);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2386
static int sub_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2388
return wrp_alu32_reg(nfp_prog, meta, ALU_OP_SUB);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2391
static int sub_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2393
return wrp_alu32_imm(nfp_prog, meta, ALU_OP_SUB);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2396
static int mul_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2398
return wrp_mul(nfp_prog, meta, false, true);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2401
static int mul_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2403
return wrp_mul(nfp_prog, meta, false, false);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2406
static int div_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2408
return div_reg64(nfp_prog, meta);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2411
static int div_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2413
return div_imm64(nfp_prog, meta);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2416
static int neg_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2418
u8 dst = meta->insn.dst_reg * 2;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2421
wrp_zext(nfp_prog, meta, dst);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2427
__ashr_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta, u8 dst,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2437
wrp_zext(nfp_prog, meta, dst);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2442
static int ashr_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2444
const struct bpf_insn *insn = &meta->insn;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2449
umin = meta->umin_src;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2450
umax = meta->umax_src;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2452
return __ashr_imm(nfp_prog, meta, dst, umin);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2461
wrp_zext(nfp_prog, meta, dst);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2466
static int ashr_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2468
const struct bpf_insn *insn = &meta->insn;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2471
return __ashr_imm(nfp_prog, meta, dst, insn->imm);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2475
__shr_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta, u8 dst,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2481
wrp_zext(nfp_prog, meta, dst);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2485
static int shr_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2487
const struct bpf_insn *insn = &meta->insn;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2490
return __shr_imm(nfp_prog, meta, dst, insn->imm);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2493
static int shr_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2495
const struct bpf_insn *insn = &meta->insn;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2500
umin = meta->umin_src;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2501
umax = meta->umax_src;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2503
return __shr_imm(nfp_prog, meta, dst, umin);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2509
wrp_zext(nfp_prog, meta, dst);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2514
__shl_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta, u8 dst,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2520
wrp_zext(nfp_prog, meta, dst);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2524
static int shl_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2526
const struct bpf_insn *insn = &meta->insn;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2529
return __shl_imm(nfp_prog, meta, dst, insn->imm);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2532
static int shl_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2534
const struct bpf_insn *insn = &meta->insn;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2539
umin = meta->umin_src;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2540
umax = meta->umax_src;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2542
return __shl_imm(nfp_prog, meta, dst, umin);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2546
wrp_zext(nfp_prog, meta, dst);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2550
static int end_reg32(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2552
const struct bpf_insn *insn = &meta->insn;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2579
static int imm_ld8_part2(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2581
struct nfp_insn_meta *prev = nfp_meta_prev(meta);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2587
imm_hi = meta->insn.imm;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2600
static int imm_ld8(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2602
meta->double_cb = imm_ld8_part2;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2606
static int data_ld1(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2608
return construct_data_ld(nfp_prog, meta, meta->insn.imm, 1);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2611
static int data_ld2(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2613
return construct_data_ld(nfp_prog, meta, meta->insn.imm, 2);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2616
static int data_ld4(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2618
return construct_data_ld(nfp_prog, meta, meta->insn.imm, 4);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2621
static int data_ind_ld1(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2623
return construct_data_ind_ld(nfp_prog, meta, meta->insn.imm,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2624
meta->insn.src_reg * 2, 1);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2627
static int data_ind_ld2(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2629
return construct_data_ind_ld(nfp_prog, meta, meta->insn.imm,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2630
meta->insn.src_reg * 2, 2);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2633
static int data_ind_ld4(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2635
return construct_data_ind_ld(nfp_prog, meta, meta->insn.imm,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2636
meta->insn.src_reg * 2, 4);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2640
mem_ldx_stack(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2643
return mem_op_stack(nfp_prog, meta, size, ptr_off,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2644
meta->insn.dst_reg * 2, meta->insn.src_reg * 2,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2648
static int mem_ldx_skb(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2651
swreg dst = reg_both(meta->insn.dst_reg * 2);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2653
switch (meta->insn.off) {
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2674
wrp_immed(nfp_prog, reg_both(meta->insn.dst_reg * 2 + 1), 0);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2679
static int mem_ldx_xdp(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2682
swreg dst = reg_both(meta->insn.dst_reg * 2);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2684
switch (meta->insn.off) {
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2700
wrp_immed(nfp_prog, reg_both(meta->insn.dst_reg * 2 + 1), 0);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2706
mem_ldx_data(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2711
tmp_reg = re_load_imm_any(nfp_prog, meta->insn.off, imm_b(nfp_prog));
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2713
return data_ld_host_order_addr32(nfp_prog, meta, meta->insn.src_reg * 2,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2714
tmp_reg, meta->insn.dst_reg * 2, size);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2718
mem_ldx_emem(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2723
tmp_reg = re_load_imm_any(nfp_prog, meta->insn.off, imm_b(nfp_prog));
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2725
return data_ld_host_order_addr40(nfp_prog, meta, meta->insn.src_reg * 2,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2726
tmp_reg, meta->insn.dst_reg * 2, size);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2731
struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2733
s16 range_start = meta->pkt_cache.range_start;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2734
s16 range_end = meta->pkt_cache.range_end;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2740
src_base = reg_a(meta->insn.src_reg * 2);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2757
struct nfp_insn_meta *meta,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2760
s16 range_start = meta->pkt_cache.range_start;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2761
s16 insn_off = meta->insn.off - range_start;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2763
u8 dst_gpr = meta->insn.dst_reg * 2;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2786
wrp_zext(nfp_prog, meta, dst_gpr);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2794
wrp_zext(nfp_prog, meta, dst_gpr);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2811
struct nfp_insn_meta *meta,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2817
idx = (meta->insn.off - meta->pkt_cache.range_start) / REG_WIDTH;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2818
dst_gpr = meta->insn.dst_reg * 2;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2825
wrp_zext(nfp_prog, meta, dst_gpr);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2828
wrp_zext(nfp_prog, meta, dst_gpr);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2841
struct nfp_insn_meta *meta, unsigned int size)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2843
u8 off = meta->insn.off - meta->pkt_cache.range_start;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2846
return mem_ldx_data_from_pktcache_aligned(nfp_prog, meta, size);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2848
return mem_ldx_data_from_pktcache_unaligned(nfp_prog, meta, size);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2852
mem_ldx(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2855
if (meta->ldst_gather_len)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2856
return nfp_cpp_memcpy(nfp_prog, meta);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2858
if (meta->ptr.type == PTR_TO_CTX) {
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2860
return mem_ldx_xdp(nfp_prog, meta, size);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2862
return mem_ldx_skb(nfp_prog, meta, size);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2865
if (meta->ptr.type == PTR_TO_PACKET) {
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2866
if (meta->pkt_cache.range_end) {
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2867
if (meta->pkt_cache.do_init)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2868
mem_ldx_data_init_pktcache(nfp_prog, meta);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2870
return mem_ldx_data_from_pktcache(nfp_prog, meta, size);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2872
return mem_ldx_data(nfp_prog, meta, size);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2876
if (meta->ptr.type == PTR_TO_STACK)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2877
return mem_ldx_stack(nfp_prog, meta, size,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2878
meta->ptr.off + meta->ptr.var_off.value);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2880
if (meta->ptr.type == PTR_TO_MAP_VALUE)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2881
return mem_ldx_emem(nfp_prog, meta, size);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2886
static int mem_ldx1(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2888
return mem_ldx(nfp_prog, meta, 1);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2891
static int mem_ldx2(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2893
return mem_ldx(nfp_prog, meta, 2);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2896
static int mem_ldx4(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2898
return mem_ldx(nfp_prog, meta, 4);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2901
static int mem_ldx8(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2903
return mem_ldx(nfp_prog, meta, 8);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2907
mem_st_data(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2910
u64 imm = meta->insn.imm; /* sign extend */
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2913
off_reg = re_load_imm_any(nfp_prog, meta->insn.off, imm_b(nfp_prog));
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2915
return data_st_host_order(nfp_prog, meta->insn.dst_reg * 2, off_reg,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2919
static int mem_st(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2922
if (meta->ptr.type == PTR_TO_PACKET)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2923
return mem_st_data(nfp_prog, meta, size);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2928
static int mem_st1(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2930
return mem_st(nfp_prog, meta, 1);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2933
static int mem_st2(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2935
return mem_st(nfp_prog, meta, 2);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2938
static int mem_st4(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2940
return mem_st(nfp_prog, meta, 4);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2943
static int mem_st8(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2945
return mem_st(nfp_prog, meta, 8);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2949
mem_stx_data(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2954
off_reg = re_load_imm_any(nfp_prog, meta->insn.off, imm_b(nfp_prog));
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2956
return data_stx_host_order(nfp_prog, meta->insn.dst_reg * 2, off_reg,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2957
meta->insn.src_reg * 2, size);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2961
mem_stx_stack(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2964
return mem_op_stack(nfp_prog, meta, size, ptr_off,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2965
meta->insn.src_reg * 2, meta->insn.dst_reg * 2,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2969
static int mem_stx_xdp(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2971
switch (meta->insn.off) {
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2973
return nfp_queue_select(nfp_prog, meta);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2981
mem_stx(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2984
if (meta->ptr.type == PTR_TO_PACKET)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2985
return mem_stx_data(nfp_prog, meta, size);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2987
if (meta->ptr.type == PTR_TO_STACK)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2988
return mem_stx_stack(nfp_prog, meta, size,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2989
meta->ptr.off + meta->ptr.var_off.value);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2994
static int mem_stx1(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2996
return mem_stx(nfp_prog, meta, 1);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
2999
static int mem_stx2(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3001
return mem_stx(nfp_prog, meta, 2);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3004
static int mem_stx4(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3006
if (meta->ptr.type == PTR_TO_CTX)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3008
return mem_stx_xdp(nfp_prog, meta);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3009
return mem_stx(nfp_prog, meta, 4);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3012
static int mem_stx8(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3014
return mem_stx(nfp_prog, meta, 8);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3018
mem_xadd(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta, bool is64)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3020
u8 dst_gpr = meta->insn.dst_reg * 2;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3021
u8 src_gpr = meta->insn.src_reg * 2;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3025
off = ur_load_imm_any(nfp_prog, meta->insn.off, imm_b(nfp_prog));
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3035
if (meta->insn.off) {
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3039
if (meta->xadd_maybe_16bit) {
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3043
if (meta->xadd_over_16bit)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3045
if (meta->xadd_maybe_16bit && meta->xadd_over_16bit) {
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3051
if (meta->xadd_maybe_16bit && meta->xadd_over_16bit) {
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3059
emit_br(nfp_prog, BR_BLO, full_add, meta->insn.off ? 2 : 0);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3064
if (!meta->insn.off) {
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3077
if (meta->xadd_maybe_16bit) {
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3088
if (meta->xadd_over_16bit)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3096
if (meta->xadd_over_16bit) {
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3112
static int mem_atomic4(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3114
if (meta->insn.imm != BPF_ADD)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3117
return mem_xadd(nfp_prog, meta, false);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3120
static int mem_atomic8(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3122
if (meta->insn.imm != BPF_ADD)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3125
return mem_xadd(nfp_prog, meta, true);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3128
static int jump(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3130
emit_br(nfp_prog, BR_UNC, meta->insn.off, 0);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3135
static int jeq_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3137
const struct bpf_insn *insn = &meta->insn;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3164
static int jeq32_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3166
const struct bpf_insn *insn = &meta->insn;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3177
static int jset_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3179
const struct bpf_insn *insn = &meta->insn;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3190
if (is_mbpf_jmp64(meta) && imm >> 32) {
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3199
static int jne_imm(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3201
const struct bpf_insn *insn = &meta->insn;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3203
bool is_jmp32 = is_mbpf_jmp32(meta);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3233
static int jeq_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3235
const struct bpf_insn *insn = &meta->insn;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3239
if (is_mbpf_jmp64(meta)) {
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3251
static int jset_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3253
return wrp_test_reg(nfp_prog, meta, ALU_OP_AND, BR_BNE);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3256
static int jne_reg(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3258
return wrp_test_reg(nfp_prog, meta, ALU_OP_XOR, BR_BNE);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3262
bpf_to_bpf_call(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3311
if (!meta->jmp_dst) {
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3315
if (nfp_prog->subprog[meta->jmp_dst->subprog_idx].needs_reg_push) {
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3323
emit_br(nfp_prog, BR_UNC, meta->insn.imm, 1);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3341
meta->num_insns_after_br = nfp_prog_current_offset(nfp_prog);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3342
meta->num_insns_after_br -= offset_br;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3347
static int helper_call(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3349
switch (meta->insn.imm) {
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3351
return adjust_head(nfp_prog, meta);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3353
return adjust_tail(nfp_prog, meta);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3357
return map_call_stack_common(nfp_prog, meta);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3359
return nfp_get_prandom_u32(nfp_prog, meta);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3361
return nfp_perf_event_output(nfp_prog, meta);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3368
static int call(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3370
if (is_mbpf_pseudo_call(meta))
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3371
return bpf_to_bpf_call(nfp_prog, meta);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3373
return helper_call(nfp_prog, meta);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3376
static bool nfp_is_main_function(struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3378
return meta->subprog_idx == 0;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3381
static int goto_out(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3389
nfp_subprog_epilogue(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3391
if (nfp_prog->subprog[meta->subprog_idx].needs_reg_push) {
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3413
static int jmp_exit(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3415
if (nfp_is_main_function(meta))
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3416
return goto_out(nfp_prog, meta);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3418
return nfp_subprog_epilogue(nfp_prog, meta);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3541
nfp_fixup_immed_relo(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3556
struct nfp_insn_meta *meta, *jmp_dst;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3560
list_for_each_entry(meta, &nfp_prog->insns, l) {
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3561
if (meta->flags & FLAG_INSN_SKIP_MASK)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3563
if (!is_mbpf_jmp(meta))
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3565
if (meta->insn.code == (BPF_JMP | BPF_EXIT) &&
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3566
!nfp_is_main_function(meta))
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3568
if (is_mbpf_helper_call(meta))
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3571
if (list_is_last(&meta->l, &nfp_prog->insns))
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3574
br_idx = list_next_entry(meta, l)->off - 1;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3581
if (is_mbpf_pseudo_call(meta))
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3582
br_idx -= meta->num_insns_after_br;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3586
br_idx, meta->insn.code, nfp_prog->prog[br_idx]);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3590
if (meta->insn.code == (BPF_JMP | BPF_EXIT))
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3595
RELO_BR_REL && !is_mbpf_pseudo_call(meta))
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3598
if (!meta->jmp_dst) {
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3603
jmp_dst = meta->jmp_dst;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3610
if (is_mbpf_pseudo_call(meta) &&
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3612
err = nfp_fixup_immed_relo(nfp_prog, meta,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3622
for (idx = meta->off; idx <= br_idx; idx++) {
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3640
nfp_subprog_prologue(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3647
nfp_start_subprog(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3649
unsigned int depth = nfp_prog->subprog[meta->subprog_idx].stack_depth;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3652
nfp_subprog_prologue(nfp_prog, meta);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3655
bool nfp_is_subprog_start(struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3657
return meta->flags & FLAG_INSN_IS_SUBPROG_START;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3833
struct nfp_insn_meta *meta;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3844
list_for_each_entry(meta, &nfp_prog->insns, l) {
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3845
instr_cb_t cb = instr_cb[meta->insn.code];
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3847
meta->off = nfp_prog_current_offset(nfp_prog);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3849
if (nfp_is_subprog_start(meta)) {
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3850
nfp_start_subprog(nfp_prog, meta);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3855
if (meta->flags & FLAG_INSN_SKIP_MASK) {
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3860
if (nfp_meta_has_prev(nfp_prog, meta) &&
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3861
nfp_meta_prev(meta)->double_cb)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3862
cb = nfp_meta_prev(meta)->double_cb;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3865
err = cb(nfp_prog, meta);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3890
struct nfp_insn_meta *meta;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3892
list_for_each_entry(meta, &nfp_prog->insns, l) {
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3893
struct bpf_insn insn = meta->insn;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3903
meta->flags |= FLAG_INSN_SKIP_PREC_DEPENDENT;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3906
if (!(meta->flags & FLAG_INSN_SKIP_MASK))
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3916
struct nfp_insn_meta *meta;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3918
list_for_each_entry(meta, &nfp_prog->insns, l) {
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3919
struct bpf_insn insn = meta->insn;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3921
if (meta->flags & FLAG_INSN_SKIP_MASK)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3924
if (!is_mbpf_alu(meta) && !is_mbpf_jmp(meta))
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3931
if (is_mbpf_jmp(meta)) {
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3937
meta->jump_neg_op = true;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3950
meta->insn.code = insn.code | BPF_K;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
3953
meta->insn.imm = -insn.imm;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
42
nfp_meta_has_prev(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
4278
struct nfp_insn_meta *meta, *range_node = NULL;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
4285
list_for_each_entry(meta, &nfp_prog->insns, l) {
drivers/net/ethernet/netronome/nfp/bpf/jit.c
4286
if (meta->flags & FLAG_INSN_IS_JUMP_DST)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
4289
if (meta->flags & FLAG_INSN_SKIP_MASK)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
4292
insn = &meta->insn;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
4294
if (is_mbpf_store_pkt(meta) ||
drivers/net/ethernet/netronome/nfp/bpf/jit.c
4296
is_mbpf_classic_store_pkt(meta) ||
drivers/net/ethernet/netronome/nfp/bpf/jit.c
4297
is_mbpf_classic_load(meta)) {
drivers/net/ethernet/netronome/nfp/bpf/jit.c
4302
if (!is_mbpf_load(meta))
drivers/net/ethernet/netronome/nfp/bpf/jit.c
4305
if (meta->ptr.type != PTR_TO_PACKET || meta->ldst_gather_len) {
drivers/net/ethernet/netronome/nfp/bpf/jit.c
4328
if (meta->ptr.id == range_ptr_id &&
drivers/net/ethernet/netronome/nfp/bpf/jit.c
4329
meta->ptr.off == range_ptr_off) {
drivers/net/ethernet/netronome/nfp/bpf/jit.c
4361
range_node = meta;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
4374
list_for_each_entry(meta, &nfp_prog->insns, l) {
drivers/net/ethernet/netronome/nfp/bpf/jit.c
4375
if (meta->flags & FLAG_INSN_SKIP_MASK)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
4378
if (is_mbpf_load_pkt(meta) && !meta->ldst_gather_len) {
drivers/net/ethernet/netronome/nfp/bpf/jit.c
4379
if (meta->pkt_cache.do_init) {
drivers/net/ethernet/netronome/nfp/bpf/jit.c
4380
range_start = meta->pkt_cache.range_start;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
4381
range_end = meta->pkt_cache.range_end;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
4383
meta->pkt_cache.range_start = range_start;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
4384
meta->pkt_cache.range_end = range_end;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
44
return meta->l.prev != &nfp_prog->insns;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
4493
struct nfp_insn_meta *meta;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
4496
list_for_each_entry(meta, &nfp_prog->insns, l) {
drivers/net/ethernet/netronome/nfp/bpf/jit.c
4498
u64 code = meta->insn.code;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
4502
if (!is_mbpf_jmp(meta))
drivers/net/ethernet/netronome/nfp/bpf/jit.c
4506
if (is_mbpf_helper_call(meta))
drivers/net/ethernet/netronome/nfp/bpf/jit.c
4515
dst_idx = meta->n + 1 + meta->insn.imm;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
4517
dst_idx = meta->n + 1 + meta->insn.off;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
4519
dst_meta = nfp_bpf_goto_meta(nfp_prog, meta, dst_idx);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
4525
meta->jmp_dst = dst_meta;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
626
wrp_zext(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta, u8 dst)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
628
if (meta->flags & FLAG_INSN_DO_ZEXT)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
733
static int nfp_cpp_memcpy(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
735
bool descending_seq = meta->ldst_gather_len < 0;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
736
s16 len = abs(meta->ldst_gather_len);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
742
off = re_load_imm_any(nfp_prog, meta->insn.off, imm_b(nfp_prog));
drivers/net/ethernet/netronome/nfp/bpf/jit.c
743
src_40bit_addr = meta->ptr.type == PTR_TO_MAP_VALUE;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
744
src_base = reg_a(meta->insn.src_reg * 2);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
748
addr40_offset(nfp_prog, meta->insn.src_reg * 2, off, &src_base,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
765
off = re_load_imm_any(nfp_prog, meta->paired_st->off, imm_b(nfp_prog));
drivers/net/ethernet/netronome/nfp/bpf/jit.c
770
reg_a(meta->paired_st->dst_reg * 2), off, len - 1,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
775
reg_a(meta->paired_st->dst_reg * 2), off, xfer_num - 1,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
782
reg_a(meta->paired_st->dst_reg * 2), off,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
789
reg_a(meta->paired_st->dst_reg * 2), off,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
796
reg_a(meta->paired_st->dst_reg * 2), off, 7,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
799
off = re_load_imm_any(nfp_prog, meta->paired_st->off + 32,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
802
reg_a(meta->paired_st->dst_reg * 2), off, len - 33,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
813
reg_a(meta->paired_st->dst_reg * 2), off,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
815
new_off = meta->paired_st->off + (xfer_num - 1) * 4;
drivers/net/ethernet/netronome/nfp/bpf/jit.c
818
xfer_num - 1, reg_a(meta->paired_st->dst_reg * 2), off,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
834
else if (BPF_SIZE(meta->insn.code) != BPF_DW)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
839
switch (BPF_SIZE(meta->insn.code)) {
drivers/net/ethernet/netronome/nfp/bpf/jit.c
841
wrp_reg_subpart(nfp_prog, reg_both(meta->insn.dst_reg * 2),
drivers/net/ethernet/netronome/nfp/bpf/jit.c
846
wrp_reg_subpart(nfp_prog, reg_both(meta->insn.dst_reg * 2),
drivers/net/ethernet/netronome/nfp/bpf/jit.c
850
wrp_mov(nfp_prog, reg_both(meta->insn.dst_reg * 2),
drivers/net/ethernet/netronome/nfp/bpf/jit.c
854
wrp_mov(nfp_prog, reg_both(meta->insn.dst_reg * 2),
drivers/net/ethernet/netronome/nfp/bpf/jit.c
856
wrp_mov(nfp_prog, reg_both(meta->insn.dst_reg * 2 + 1),
drivers/net/ethernet/netronome/nfp/bpf/jit.c
861
if (BPF_SIZE(meta->insn.code) != BPF_DW)
drivers/net/ethernet/netronome/nfp/bpf/jit.c
862
wrp_immed(nfp_prog, reg_both(meta->insn.dst_reg * 2 + 1), 0);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
868
data_ld(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta, swreg offset,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
892
wrp_zext(nfp_prog, meta, dst_gpr);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
898
data_ld_host_order(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
923
wrp_zext(nfp_prog, meta, dst_gpr);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
929
data_ld_host_order_addr32(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
932
return data_ld_host_order(nfp_prog, meta, dst_gpr, reg_a(src_gpr),
drivers/net/ethernet/netronome/nfp/bpf/jit.c
937
data_ld_host_order_addr40(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
944
return data_ld_host_order(nfp_prog, meta, dst_gpr, rega, regb,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
949
construct_data_ind_ld(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
966
return data_ld(nfp_prog, meta, imm_b(nfp_prog), 0, size);
drivers/net/ethernet/netronome/nfp/bpf/jit.c
970
construct_data_ld(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta,
drivers/net/ethernet/netronome/nfp/bpf/jit.c
982
return data_ld(nfp_prog, meta, tmp_reg, 0, size);
drivers/net/ethernet/netronome/nfp/bpf/main.h
241
#define nfp_meta_next(meta) list_next_entry(meta, l)
drivers/net/ethernet/netronome/nfp/bpf/main.h
242
#define nfp_meta_prev(meta) list_prev_entry(meta, l)
drivers/net/ethernet/netronome/nfp/bpf/main.h
351
static inline u8 mbpf_class(const struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/main.h
353
return BPF_CLASS(meta->insn.code);
drivers/net/ethernet/netronome/nfp/bpf/main.h
356
static inline u8 mbpf_src(const struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/main.h
358
return BPF_SRC(meta->insn.code);
drivers/net/ethernet/netronome/nfp/bpf/main.h
361
static inline u8 mbpf_op(const struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/main.h
363
return BPF_OP(meta->insn.code);
drivers/net/ethernet/netronome/nfp/bpf/main.h
366
static inline u8 mbpf_mode(const struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/main.h
368
return BPF_MODE(meta->insn.code);
drivers/net/ethernet/netronome/nfp/bpf/main.h
371
static inline bool is_mbpf_alu(const struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/main.h
373
return mbpf_class(meta) == BPF_ALU64 || mbpf_class(meta) == BPF_ALU;
drivers/net/ethernet/netronome/nfp/bpf/main.h
376
static inline bool is_mbpf_load(const struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/main.h
378
return (meta->insn.code & ~BPF_SIZE_MASK) == (BPF_LDX | BPF_MEM);
drivers/net/ethernet/netronome/nfp/bpf/main.h
381
static inline bool is_mbpf_jmp32(const struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/main.h
383
return mbpf_class(meta) == BPF_JMP32;
drivers/net/ethernet/netronome/nfp/bpf/main.h
386
static inline bool is_mbpf_jmp64(const struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/main.h
388
return mbpf_class(meta) == BPF_JMP;
drivers/net/ethernet/netronome/nfp/bpf/main.h
391
static inline bool is_mbpf_jmp(const struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/main.h
393
return is_mbpf_jmp32(meta) || is_mbpf_jmp64(meta);
drivers/net/ethernet/netronome/nfp/bpf/main.h
396
static inline bool is_mbpf_store(const struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/main.h
398
return (meta->insn.code & ~BPF_SIZE_MASK) == (BPF_STX | BPF_MEM);
drivers/net/ethernet/netronome/nfp/bpf/main.h
401
static inline bool is_mbpf_load_pkt(const struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/main.h
403
return is_mbpf_load(meta) && meta->ptr.type == PTR_TO_PACKET;
drivers/net/ethernet/netronome/nfp/bpf/main.h
406
static inline bool is_mbpf_store_pkt(const struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/main.h
408
return is_mbpf_store(meta) && meta->ptr.type == PTR_TO_PACKET;
drivers/net/ethernet/netronome/nfp/bpf/main.h
411
static inline bool is_mbpf_classic_load(const struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/main.h
413
u8 code = meta->insn.code;
drivers/net/ethernet/netronome/nfp/bpf/main.h
419
static inline bool is_mbpf_classic_store(const struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/main.h
421
u8 code = meta->insn.code;
drivers/net/ethernet/netronome/nfp/bpf/main.h
426
static inline bool is_mbpf_classic_store_pkt(const struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/main.h
428
return is_mbpf_classic_store(meta) && meta->ptr.type == PTR_TO_PACKET;
drivers/net/ethernet/netronome/nfp/bpf/main.h
431
static inline bool is_mbpf_atomic(const struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/main.h
433
return (meta->insn.code & ~BPF_SIZE_MASK) == (BPF_STX | BPF_ATOMIC);
drivers/net/ethernet/netronome/nfp/bpf/main.h
436
static inline bool is_mbpf_mul(const struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/main.h
438
return is_mbpf_alu(meta) && mbpf_op(meta) == BPF_MUL;
drivers/net/ethernet/netronome/nfp/bpf/main.h
441
static inline bool is_mbpf_div(const struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/main.h
443
return is_mbpf_alu(meta) && mbpf_op(meta) == BPF_DIV;
drivers/net/ethernet/netronome/nfp/bpf/main.h
446
static inline bool is_mbpf_cond_jump(const struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/main.h
450
if (is_mbpf_jmp32(meta))
drivers/net/ethernet/netronome/nfp/bpf/main.h
453
if (!is_mbpf_jmp64(meta))
drivers/net/ethernet/netronome/nfp/bpf/main.h
456
op = mbpf_op(meta);
drivers/net/ethernet/netronome/nfp/bpf/main.h
460
static inline bool is_mbpf_helper_call(const struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/main.h
462
struct bpf_insn insn = meta->insn;
drivers/net/ethernet/netronome/nfp/bpf/main.h
468
static inline bool is_mbpf_pseudo_call(const struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/main.h
470
struct bpf_insn insn = meta->insn;
drivers/net/ethernet/netronome/nfp/bpf/main.h
559
bool nfp_is_subprog_start(struct nfp_insn_meta *meta);
drivers/net/ethernet/netronome/nfp/bpf/main.h
586
nfp_bpf_goto_meta(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta,
drivers/net/ethernet/netronome/nfp/bpf/offload.c
152
struct nfp_insn_meta *meta;
drivers/net/ethernet/netronome/nfp/bpf/offload.c
156
meta = kzalloc_obj(*meta);
drivers/net/ethernet/netronome/nfp/bpf/offload.c
157
if (!meta)
drivers/net/ethernet/netronome/nfp/bpf/offload.c
160
meta->insn = prog[i];
drivers/net/ethernet/netronome/nfp/bpf/offload.c
161
meta->n = i;
drivers/net/ethernet/netronome/nfp/bpf/offload.c
162
if (is_mbpf_alu(meta)) {
drivers/net/ethernet/netronome/nfp/bpf/offload.c
163
meta->umin_src = U64_MAX;
drivers/net/ethernet/netronome/nfp/bpf/offload.c
164
meta->umin_dst = U64_MAX;
drivers/net/ethernet/netronome/nfp/bpf/offload.c
167
list_add_tail(&meta->l, &nfp_prog->insns);
drivers/net/ethernet/netronome/nfp/bpf/offload.c
178
struct nfp_insn_meta *meta, *tmp;
drivers/net/ethernet/netronome/nfp/bpf/offload.c
182
list_for_each_entry_safe(meta, tmp, &nfp_prog->insns, l) {
drivers/net/ethernet/netronome/nfp/bpf/offload.c
183
list_del(&meta->l);
drivers/net/ethernet/netronome/nfp/bpf/offload.c
184
kfree(meta);
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
158
struct nfp_insn_meta *meta,
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
172
struct nfp_insn_meta *meta)
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
178
u32 func_id = meta->insn.imm;
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
191
nfp_record_adjust_head(bpf, nfp_prog, meta, reg2);
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
20
nfp_bpf_goto_meta(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta,
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
202
if (!nfp_bpf_map_call_ok("map_lookup", env, meta,
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
205
meta->func_id ? &meta->arg2 : NULL))
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
210
if (!nfp_bpf_map_call_ok("map_update", env, meta,
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
213
meta->func_id ? &meta->arg2 : NULL) ||
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
220
if (!nfp_bpf_map_call_ok("map_delete", env, meta,
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
223
meta->func_id ? &meta->arg2 : NULL))
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
25
backward = meta->n - insn_idx;
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
26
forward = insn_idx - meta->n;
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
288
if (!meta->func_id)
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
291
if (reg1->type != meta->arg1.type) {
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
293
meta->arg1.type, reg1->type);
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
30
meta = nfp_prog_last_meta(nfp_prog);
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
303
meta->func_id = func_id;
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
304
meta->arg1 = *reg1;
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
305
meta->arg2.reg = *reg2;
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
34
meta = nfp_prog_first_meta(nfp_prog);
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
344
struct nfp_insn_meta *meta,
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
351
meta->flags |= FLAG_INSN_PTR_CALLER_STACK_FRAME;
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
358
if (meta->ptr.type == NOT_INIT)
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
361
old_off = meta->ptr.off + meta->ptr.var_off.value;
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
364
meta->ptr_not_const |= old_off != new_off;
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
366
if (!meta->ptr_not_const)
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
39
meta = nfp_meta_next(meta);
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
417
nfp_bpf_map_mark_used(struct bpf_verifier_env *env, struct nfp_insn_meta *meta,
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
42
meta = nfp_meta_prev(meta);
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
431
off = reg->var_off.value + meta->insn.off + reg->off;
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
432
size = BPF_LDST_BYTES(&meta->insn);
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
44
return meta;
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
451
nfp_bpf_check_ptr(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta,
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
466
err = nfp_bpf_check_stack_access(nfp_prog, meta, reg, env);
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
472
if (is_mbpf_load(meta)) {
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
473
err = nfp_bpf_map_mark_used(env, meta, reg,
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
478
if (is_mbpf_store(meta)) {
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
482
if (is_mbpf_atomic(meta)) {
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
483
err = nfp_bpf_map_mark_used(env, meta, reg,
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
49
struct nfp_insn_meta *meta,
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
490
if (meta->ptr.type != NOT_INIT && meta->ptr.type != reg->type) {
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
492
meta->ptr.type, reg->type);
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
496
meta->ptr = *reg;
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
502
nfp_bpf_check_store(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta,
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
505
const struct bpf_reg_state *reg = cur_regs(env) + meta->insn.dst_reg;
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
510
switch (meta->insn.off) {
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
522
return nfp_bpf_check_ptr(nfp_prog, meta, env, meta->insn.dst_reg);
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
526
nfp_bpf_check_atomic(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta,
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
529
const struct bpf_reg_state *sreg = cur_regs(env) + meta->insn.src_reg;
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
530
const struct bpf_reg_state *dreg = cur_regs(env) + meta->insn.dst_reg;
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
532
if (meta->insn.imm != BPF_ADD) {
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
533
pr_vlog(env, "atomic op not implemented: %d\n", meta->insn.imm);
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
547
meta->xadd_over_16bit |=
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
549
meta->xadd_maybe_16bit |=
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
552
return nfp_bpf_check_ptr(nfp_prog, meta, env, meta->insn.dst_reg);
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
556
nfp_bpf_check_alu(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta,
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
560
cur_regs(env) + meta->insn.src_reg;
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
562
cur_regs(env) + meta->insn.dst_reg;
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
564
meta->umin_src = min(meta->umin_src, sreg->umin_value);
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
565
meta->umax_src = max(meta->umax_src, sreg->umax_value);
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
566
meta->umin_dst = min(meta->umin_dst, dreg->umin_value);
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
567
meta->umax_dst = max(meta->umax_dst, dreg->umax_value);
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
581
if (is_mbpf_mul(meta)) {
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
582
if (meta->umax_dst > U32_MAX) {
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
586
if (mbpf_src(meta) == BPF_X && meta->umax_src > U32_MAX) {
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
590
if (mbpf_class(meta) == BPF_ALU64 &&
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
591
mbpf_src(meta) == BPF_K && meta->insn.imm < 0) {
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
607
if (is_mbpf_div(meta)) {
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
608
if (meta->umax_dst > U32_MAX) {
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
612
if (mbpf_src(meta) == BPF_X) {
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
613
if (meta->umin_src != meta->umax_src) {
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
617
if (meta->umax_src > U32_MAX) {
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
622
if (mbpf_src(meta) == BPF_K && meta->insn.imm < 0) {
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
635
struct nfp_insn_meta *meta = nfp_prog->verifier_meta;
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
637
meta = nfp_bpf_goto_meta(nfp_prog, meta, insn_idx);
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
638
nfp_prog->verifier_meta = meta;
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
640
if (!nfp_bpf_supported_opcode(meta->insn.code)) {
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
642
meta->insn.code);
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
646
if (meta->insn.src_reg >= MAX_BPF_REG ||
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
647
meta->insn.dst_reg >= MAX_BPF_REG) {
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
652
if (is_mbpf_helper_call(meta))
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
653
return nfp_bpf_check_helper_call(nfp_prog, env, meta);
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
654
if (meta->insn.code == (BPF_JMP | BPF_EXIT))
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
657
if (is_mbpf_load(meta))
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
658
return nfp_bpf_check_ptr(nfp_prog, meta, env,
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
659
meta->insn.src_reg);
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
660
if (is_mbpf_store(meta))
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
661
return nfp_bpf_check_store(nfp_prog, meta, env);
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
663
if (is_mbpf_atomic(meta))
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
664
return nfp_bpf_check_atomic(nfp_prog, meta, env);
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
666
if (is_mbpf_alu(meta))
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
667
return nfp_bpf_check_alu(nfp_prog, meta, env);
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
676
struct nfp_insn_meta *meta;
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
679
list_for_each_entry(meta, &nfp_prog->insns, l) {
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
680
if (nfp_is_subprog_start(meta))
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
682
meta->subprog_idx = index;
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
684
if (meta->insn.dst_reg >= BPF_REG_6 &&
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
685
meta->insn.dst_reg <= BPF_REG_9)
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
700
struct nfp_insn_meta *meta = nfp_prog_first_meta(nfp_prog);
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
705
unsigned short idx = meta->subprog_idx;
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
71
if (nfp_prog->adjust_head_location != meta->n)
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
720
for (; meta != nfp_prog_last_meta(nfp_prog) && meta->subprog_idx == idx;
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
721
meta = nfp_meta_next(meta)) {
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
722
if (!is_mbpf_pseudo_call(meta))
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
728
ret_insn[frame] = nfp_meta_next(meta);
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
732
meta = nfp_bpf_goto_meta(nfp_prog, meta,
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
733
meta->n + 1 + meta->insn.imm);
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
734
idx = meta->subprog_idx;
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
74
if (meta->arg2.reg.var_off.value != imm)
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
747
meta = ret_insn[frame];
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
755
struct nfp_insn_meta *meta;
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
757
list_for_each_entry(meta, &nfp_prog->insns, l) {
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
758
if (aux[meta->n].zext_dst)
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
759
meta->flags |= FLAG_INSN_DO_ZEXT;
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
78
location = meta->n;
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
812
struct nfp_insn_meta *meta = nfp_prog->verifier_meta;
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
814
meta = nfp_bpf_goto_meta(nfp_prog, meta, aux_data[off].orig_idx);
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
815
nfp_prog->verifier_meta = meta;
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
818
if (is_mbpf_cond_jump(meta) &&
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
825
meta->jmp_dst = list_next_entry(meta, l);
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
826
meta->jump_neg_op = false;
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
827
} else if (meta->jmp_dst->n != aux_data[tgt_off].orig_idx) {
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
829
off, meta->jmp_dst->n,
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
837
meta->insn.code, insn->code);
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
845
struct nfp_insn_meta *meta = nfp_prog->verifier_meta;
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
848
meta = nfp_bpf_goto_meta(nfp_prog, meta, aux_data[off].orig_idx);
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
851
if (WARN_ON_ONCE(&meta->l == &nfp_prog->insns))
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
855
if (meta->flags & FLAG_INSN_SKIP_VERIFIER_OPT)
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
858
meta->flags |= FLAG_INSN_SKIP_VERIFIER_OPT;
drivers/net/ethernet/netronome/nfp/bpf/verifier.c
859
meta = list_next_entry(meta, l);
drivers/net/ethernet/netronome/nfp/crypto/crypto.h
62
int nfp_net_ipsec_rx(struct nfp_meta_parsed *meta, struct sk_buff *skb);
drivers/net/ethernet/netronome/nfp/crypto/ipsec.c
607
int nfp_net_ipsec_rx(struct nfp_meta_parsed *meta, struct sk_buff *skb)
drivers/net/ethernet/netronome/nfp/crypto/ipsec.c
618
saidx = meta->ipsec_saidx - 1;
drivers/net/ethernet/netronome/nfp/flower/action.c
1062
nfp_fl->meta.shortcut = cpu_to_be32(NFP_FL_SC_ACT_DROP);
drivers/net/ethernet/netronome/nfp/flower/action.c
1090
nfp_fl->meta.shortcut = cpu_to_be32(NFP_FL_SC_ACT_POPV);
drivers/net/ethernet/netronome/nfp/flower/action.c
1103
nfp_fl->meta.shortcut = cpu_to_be32(NFP_FL_SC_ACT_NULL);
drivers/net/ethernet/netronome/nfp/flower/action.c
1133
nfp_fl->meta.shortcut = cpu_to_be32(NFP_FL_SC_ACT_NULL);
drivers/net/ethernet/netronome/nfp/flower/action.c
1175
nfp_fl->meta.shortcut = cpu_to_be32(NFP_FL_SC_ACT_NULL);
drivers/net/ethernet/netronome/nfp/flower/action.c
1190
nfp_fl->meta.shortcut = cpu_to_be32(NFP_FL_SC_ACT_NULL);
drivers/net/ethernet/netronome/nfp/flower/action.c
1203
nfp_fl->meta.shortcut = cpu_to_be32(NFP_FL_SC_ACT_NULL);
drivers/net/ethernet/netronome/nfp/flower/action.c
1290
nfp_flow->meta.act_len = 0;
drivers/net/ethernet/netronome/nfp/flower/action.c
1317
nfp_flow->meta.shortcut = cpu_to_be32(NFP_FL_SC_ACT_NULL);
drivers/net/ethernet/netronome/nfp/flower/action.c
1319
nfp_flow->meta.act_len = act_len;
drivers/net/ethernet/netronome/nfp/flower/action.c
166
nfp_flow->meta.shortcut = cpu_to_be32(NFP_FL_SC_ACT_NULL);
drivers/net/ethernet/netronome/nfp/flower/action.c
260
nfp_flow->meta.shortcut = output->port;
drivers/net/ethernet/netronome/nfp/flower/conntrack.c
2000
ctx_id = be32_to_cpu(nfp_flow->meta.host_ctx_id);
drivers/net/ethernet/netronome/nfp/flower/main.h
434
struct nfp_fl_rule_metadata meta;
drivers/net/ethernet/netronome/nfp/flower/metadata.c
323
nfp_flow->meta.host_ctx_id = cpu_to_be32(stats_cxt);
drivers/net/ethernet/netronome/nfp/flower/metadata.c
324
nfp_flow->meta.host_cookie = cpu_to_be64(cookie);
drivers/net/ethernet/netronome/nfp/flower/metadata.c
350
nfp_flow->meta.mask_len,
drivers/net/ethernet/netronome/nfp/flower/metadata.c
351
&nfp_flow->meta.flags, &new_mask_id)) {
drivers/net/ethernet/netronome/nfp/flower/metadata.c
357
nfp_flow->meta.flow_version = cpu_to_be64(priv->flower_version);
drivers/net/ethernet/netronome/nfp/flower/metadata.c
378
nfp_flow->meta.mask_len,
drivers/net/ethernet/netronome/nfp/flower/metadata.c
395
nfp_flow->meta.flags &= ~NFP_FL_META_FLAG_MANAGE_MASK;
drivers/net/ethernet/netronome/nfp/flower/metadata.c
396
nfp_flow->meta.flow_version = cpu_to_be64(priv->flower_version);
drivers/net/ethernet/netronome/nfp/flower/metadata.c
412
nfp_flow->meta.mask_len, &nfp_flow->meta.flags,
drivers/net/ethernet/netronome/nfp/flower/metadata.c
419
temp_ctx_id = be32_to_cpu(nfp_flow->meta.host_ctx_id);
drivers/net/ethernet/netronome/nfp/flower/offload.c
1022
parent_ctx = (u64)(be32_to_cpu(sub_flow1->meta.host_ctx_id)) << 32;
drivers/net/ethernet/netronome/nfp/flower/offload.c
1023
parent_ctx |= (u64)(be32_to_cpu(sub_flow2->meta.host_ctx_id));
drivers/net/ethernet/netronome/nfp/flower/offload.c
103
key_len = nfp_flow->meta.key_len;
drivers/net/ethernet/netronome/nfp/flower/offload.c
1034
merge_key_ls.key_size = sub_flow1->meta.key_len;
drivers/net/ethernet/netronome/nfp/flower/offload.c
104
mask_len = nfp_flow->meta.mask_len;
drivers/net/ethernet/netronome/nfp/flower/offload.c
1044
sub_flow1->meta.key_len);
drivers/net/ethernet/netronome/nfp/flower/offload.c
1046
sub_flow1->meta.mask_len);
drivers/net/ethernet/netronome/nfp/flower/offload.c
105
act_len = nfp_flow->meta.act_len;
drivers/net/ethernet/netronome/nfp/flower/offload.c
112
nfp_flow->meta.key_len >>= NFP_FL_LW_SIZ;
drivers/net/ethernet/netronome/nfp/flower/offload.c
113
nfp_flow->meta.mask_len >>= NFP_FL_LW_SIZ;
drivers/net/ethernet/netronome/nfp/flower/offload.c
114
nfp_flow->meta.act_len >>= NFP_FL_LW_SIZ;
drivers/net/ethernet/netronome/nfp/flower/offload.c
121
memcpy(msg, &nfp_flow->meta, meta_len);
drivers/net/ethernet/netronome/nfp/flower/offload.c
1294
if (act_offset != flow->meta.act_len) {
drivers/net/ethernet/netronome/nfp/flower/offload.c
130
nfp_flow->meta.key_len <<= NFP_FL_LW_SIZ;
drivers/net/ethernet/netronome/nfp/flower/offload.c
131
nfp_flow->meta.mask_len <<= NFP_FL_LW_SIZ;
drivers/net/ethernet/netronome/nfp/flower/offload.c
132
nfp_flow->meta.act_len <<= NFP_FL_LW_SIZ;
drivers/net/ethernet/netronome/nfp/flower/offload.c
1501
u32 ctx_id = be32_to_cpu(link->sub_flow.flow->meta.host_ctx_id);
drivers/net/ethernet/netronome/nfp/flower/offload.c
1639
ctx_id = be32_to_cpu(merge_flow->meta.host_ctx_id);
drivers/net/ethernet/netronome/nfp/flower/offload.c
1657
ctx_id = be32_to_cpu(sub_flow->meta.host_ctx_id);
drivers/net/ethernet/netronome/nfp/flower/offload.c
1710
ctx_id = be32_to_cpu(nfp_flow->meta.host_ctx_id);
drivers/net/ethernet/netronome/nfp/flower/offload.c
556
flow_pay->meta.key_len = key_layer->key_size;
drivers/net/ethernet/netronome/nfp/flower/offload.c
561
flow_pay->meta.mask_len = key_layer->key_size;
drivers/net/ethernet/netronome/nfp/flower/offload.c
572
flow_pay->meta.flags = 0;
drivers/net/ethernet/netronome/nfp/flower/offload.c
608
while (act_off < flow->meta.act_len) {
drivers/net/ethernet/netronome/nfp/flower/offload.c
890
sub1_act_len = sub_flow1->meta.act_len - sizeof(struct nfp_fl_output);
drivers/net/ethernet/netronome/nfp/flower/offload.c
891
sub2_act_len = sub_flow2->meta.act_len;
drivers/net/ethernet/netronome/nfp/flower/offload.c
901
merge_flow->meta.shortcut = cpu_to_be32(NFP_FL_SC_ACT_NULL);
drivers/net/ethernet/netronome/nfp/flower/offload.c
903
merge_flow->meta.shortcut = sub_flow2->meta.shortcut;
drivers/net/ethernet/netronome/nfp/flower/offload.c
905
merge_flow->meta.act_len = sub1_act_len + sub2_act_len;
drivers/net/ethernet/netronome/nfp/flower/offload.c
948
merge_flow->meta.act_len -= sizeof(*post_tun_push_vlan);
drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c
1445
payload.host_ctx_id = flow->meta.host_ctx_id;
drivers/net/ethernet/netronome/nfp/flower/tunnel_conf.c
354
ext->host_ctx = flow_pay->meta.host_ctx_id;
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
1007
if (xdp_prog && !meta.portid) {
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
1050
if (likely(!meta.portid)) {
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
1052
} else if (meta.portid == NFP_META_PORT_ID_CTRL) {
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
1064
netdev = nfp_app_dev_get(nn->app, meta.portid,
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
1094
skb->mark = meta.mark;
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
1095
skb_set_hash(skb, meta.hash, meta.hash_type);
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
1100
nfp_nfd3_rx_csum(dp, r_vec, rxd, &meta, skb);
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
1111
if (unlikely(!nfp_net_vlan_strip(skb, rxd, &meta))) {
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
1117
if (meta.ipsec_saidx != 0 && unlikely(nfp_net_ipsec_rx(&meta, skb))) {
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
664
const struct nfp_meta_parsed *meta, struct sk_buff *skb)
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
671
if (meta->csum_type) {
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
672
skb->ip_summed = meta->csum_type;
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
673
skb->csum = meta->csum;
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
709
nfp_nfd3_set_hash(struct net_device *netdev, struct nfp_meta_parsed *meta,
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
719
meta->hash_type = PKT_HASH_TYPE_L3;
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
722
meta->hash_type = PKT_HASH_TYPE_L4;
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
726
meta->hash = get_unaligned_be32(hash);
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
730
nfp_nfd3_set_hash_desc(struct net_device *netdev, struct nfp_meta_parsed *meta,
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
738
nfp_nfd3_set_hash(netdev, meta, get_unaligned_be32(&rx_hash->hash_type),
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
743
nfp_nfd3_parse_meta(struct net_device *netdev, struct nfp_meta_parsed *meta,
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
755
nfp_nfd3_set_hash(netdev, meta,
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
761
meta->mark = get_unaligned_be32(data);
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
767
meta->vlan.stripped = true;
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
768
meta->vlan.tpid = FIELD_GET(NFP_NET_META_VLAN_TPID_MASK,
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
770
meta->vlan.tci = FIELD_GET(NFP_NET_META_VLAN_TCI_MASK,
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
776
meta->portid = get_unaligned_be32(data);
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
780
meta->csum_type = CHECKSUM_COMPLETE;
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
781
meta->csum =
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
796
meta->ipsec_saidx = get_unaligned_be32(data) + 1;
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
928
struct nfp_meta_parsed meta;
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
946
memset(&meta, 0, sizeof(meta));
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
993
nfp_nfd3_set_hash_desc(dp->netdev, &meta,
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
996
if (unlikely(nfp_nfd3_parse_meta(dp->netdev, &meta,
drivers/net/ethernet/netronome/nfp/nfd3/nfd3.h
90
const struct nfp_meta_parsed *meta, struct sk_buff *skb);
drivers/net/ethernet/netronome/nfp/nfd3/nfd3.h
92
nfp_nfd3_parse_meta(struct net_device *netdev, struct nfp_meta_parsed *meta,
drivers/net/ethernet/netronome/nfp/nfd3/xsk.c
132
struct nfp_meta_parsed meta;
drivers/net/ethernet/netronome/nfp/nfd3/xsk.c
158
memset(&meta, 0, sizeof(meta));
drivers/net/ethernet/netronome/nfp/nfd3/xsk.c
191
if (unlikely(nfp_nfd3_parse_meta(dp->netdev, &meta,
drivers/net/ethernet/netronome/nfp/nfd3/xsk.c
201
if (unlikely(meta.portid)) {
drivers/net/ethernet/netronome/nfp/nfd3/xsk.c
204
if (meta.portid != NFP_META_PORT_ID_CTRL) {
drivers/net/ethernet/netronome/nfp/nfd3/xsk.c
206
xrxbuf, &meta,
drivers/net/ethernet/netronome/nfp/nfd3/xsk.c
226
nfp_nfd3_xsk_rx_skb(rx_ring, rxd, xrxbuf, &meta, pkt_len,
drivers/net/ethernet/netronome/nfp/nfd3/xsk.c
59
const struct nfp_meta_parsed *meta,
drivers/net/ethernet/netronome/nfp/nfd3/xsk.c
69
if (likely(!meta->portid)) {
drivers/net/ethernet/netronome/nfp/nfd3/xsk.c
74
netdev = nfp_app_dev_get(nn->app, meta->portid, NULL);
drivers/net/ethernet/netronome/nfp/nfd3/xsk.c
89
skb->mark = meta->mark;
drivers/net/ethernet/netronome/nfp/nfd3/xsk.c
90
skb_set_hash(skb, meta->hash, meta->hash_type);
drivers/net/ethernet/netronome/nfp/nfd3/xsk.c
95
nfp_nfd3_rx_csum(dp, r_vec, rxd, meta, skb);
drivers/net/ethernet/netronome/nfp/nfd3/xsk.c
97
if (unlikely(!nfp_net_vlan_strip(skb, rxd, meta))) {
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
1050
struct nfp_meta_parsed meta;
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
1068
memset(&meta, 0, sizeof(meta));
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
1115
if (unlikely(nfp_nfdk_parse_meta(dp->netdev, &meta,
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
1126
if (xdp_prog && !meta.portid) {
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
1169
if (likely(!meta.portid)) {
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
1171
} else if (meta.portid == NFP_META_PORT_ID_CTRL) {
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
1183
netdev = nfp_app_dev_get(nn->app, meta.portid,
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
1213
skb->mark = meta.mark;
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
1214
skb_set_hash(skb, meta.hash, meta.hash_type);
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
1219
nfp_nfdk_rx_csum(dp, r_vec, rxd, &meta, skb);
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
1221
if (unlikely(!nfp_net_vlan_strip(skb, rxd, &meta))) {
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
1227
if (meta.ipsec_saidx != 0 && unlikely(nfp_net_ipsec_rx(&meta, skb))) {
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
676
struct nfp_net_rx_desc *rxd, struct nfp_meta_parsed *meta,
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
684
if (meta->csum_type) {
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
685
skb->ip_summed = meta->csum_type;
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
686
skb->csum = meta->csum;
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
722
nfp_nfdk_set_hash(struct net_device *netdev, struct nfp_meta_parsed *meta,
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
732
meta->hash_type = PKT_HASH_TYPE_L3;
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
735
meta->hash_type = PKT_HASH_TYPE_L4;
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
739
meta->hash = get_unaligned_be32(hash);
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
743
nfp_nfdk_parse_meta(struct net_device *netdev, struct nfp_meta_parsed *meta,
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
755
nfp_nfdk_set_hash(netdev, meta,
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
761
meta->mark = get_unaligned_be32(data);
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
767
meta->vlan.stripped = true;
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
768
meta->vlan.tpid = FIELD_GET(NFP_NET_META_VLAN_TPID_MASK,
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
770
meta->vlan.tci = FIELD_GET(NFP_NET_META_VLAN_TCI_MASK,
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
776
meta->portid = get_unaligned_be32(data);
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
780
meta->csum_type = CHECKSUM_COMPLETE;
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
781
meta->csum =
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
796
meta->ipsec_saidx = get_unaligned_be32(data) + 1;
drivers/net/ethernet/netronome/nfp/nfp_net_dp.c
443
const struct nfp_meta_parsed *meta)
drivers/net/ethernet/netronome/nfp/nfp_net_dp.c
450
} else if (meta->vlan.stripped) {
drivers/net/ethernet/netronome/nfp/nfp_net_dp.c
451
if (meta->vlan.tpid == NFP_NET_VLAN_CTAG)
drivers/net/ethernet/netronome/nfp/nfp_net_dp.c
453
else if (meta->vlan.tpid == NFP_NET_VLAN_STAG)
drivers/net/ethernet/netronome/nfp/nfp_net_dp.c
458
tci = meta->vlan.tci;
drivers/net/ethernet/netronome/nfp/nfp_net_dp.h
110
const struct nfp_meta_parsed *meta);
drivers/net/ethernet/qlogic/qed/qed_debug.c
6775
struct mcp_trace_meta *meta;
drivers/net/ethernet/qlogic/qed/qed_debug.c
6779
meta = &dev_user_data->mcp_trace_meta;
drivers/net/ethernet/qlogic/qed/qed_debug.c
6783
if (meta->is_allocated)
drivers/net/ethernet/qlogic/qed/qed_debug.c
6786
memset(meta, 0, sizeof(*meta));
drivers/net/ethernet/qlogic/qed/qed_debug.c
6794
meta->modules_num = qed_read_byte_from_buf(meta_buf_bytes, &offset);
drivers/net/ethernet/qlogic/qed/qed_debug.c
6795
meta->modules = kcalloc(meta->modules_num, sizeof(char *),
drivers/net/ethernet/qlogic/qed/qed_debug.c
6797
if (!meta->modules)
drivers/net/ethernet/qlogic/qed/qed_debug.c
6801
for (i = 0; i < meta->modules_num; i++) {
drivers/net/ethernet/qlogic/qed/qed_debug.c
6804
*(meta->modules + i) = kzalloc(module_len, GFP_KERNEL);
drivers/net/ethernet/qlogic/qed/qed_debug.c
6805
if (!(*(meta->modules + i))) {
drivers/net/ethernet/qlogic/qed/qed_debug.c
6807
meta->modules_num = i ? i - 1 : 0;
drivers/net/ethernet/qlogic/qed/qed_debug.c
6812
*(meta->modules + i));
drivers/net/ethernet/qlogic/qed/qed_debug.c
6814
(*(meta->modules + i))[MCP_TRACE_MAX_MODULE_LEN] = '\0';
drivers/net/ethernet/qlogic/qed/qed_debug.c
6823
meta->formats_num = qed_read_dword_from_buf(meta_buf_bytes, &offset);
drivers/net/ethernet/qlogic/qed/qed_debug.c
6824
meta->formats = kzalloc_objs(struct mcp_trace_format, meta->formats_num);
drivers/net/ethernet/qlogic/qed/qed_debug.c
6825
if (!meta->formats)
drivers/net/ethernet/qlogic/qed/qed_debug.c
6829
for (i = 0; i < meta->formats_num; i++) {
drivers/net/ethernet/qlogic/qed/qed_debug.c
6830
struct mcp_trace_format *format_ptr = &meta->formats[i];
drivers/net/ethernet/qlogic/qed/qed_debug.c
6840
meta->formats_num = i ? i - 1 : 0;
drivers/net/ethernet/qlogic/qed/qed_debug.c
6849
meta->is_allocated = true;
drivers/net/ethernet/qlogic/qed/qed_debug.c
6873
struct mcp_trace_meta *meta;
drivers/net/ethernet/qlogic/qed/qed_debug.c
6878
meta = &dev_user_data->mcp_trace_meta;
drivers/net/ethernet/qlogic/qed/qed_debug.c
6881
if (!meta->is_allocated)
drivers/net/ethernet/qlogic/qed/qed_debug.c
6903
if (format_idx >= meta->formats_num) {
drivers/net/ethernet/qlogic/qed/qed_debug.c
6917
format_ptr = &meta->formats[format_idx];
drivers/net/ethernet/qlogic/qed/qed_debug.c
6964
meta->modules[format_module]);
drivers/net/ethernet/qlogic/qed/qed_debug.c
7619
struct mcp_trace_meta *meta;
drivers/net/ethernet/qlogic/qed/qed_debug.c
7623
meta = &dev_user_data->mcp_trace_meta;
drivers/net/ethernet/qlogic/qed/qed_debug.c
7624
if (!meta->is_allocated)
drivers/net/ethernet/qlogic/qed/qed_debug.c
7628
if (meta->modules) {
drivers/net/ethernet/qlogic/qed/qed_debug.c
7629
for (i = 0; i < meta->modules_num; i++)
drivers/net/ethernet/qlogic/qed/qed_debug.c
7630
kfree(meta->modules[i]);
drivers/net/ethernet/qlogic/qed/qed_debug.c
7631
kfree(meta->modules);
drivers/net/ethernet/qlogic/qed/qed_debug.c
7635
if (meta->formats) {
drivers/net/ethernet/qlogic/qed/qed_debug.c
7636
for (i = 0; i < meta->formats_num; i++)
drivers/net/ethernet/qlogic/qed/qed_debug.c
7637
kfree(meta->formats[i].format_str);
drivers/net/ethernet/qlogic/qed/qed_debug.c
7638
kfree(meta->formats);
drivers/net/ethernet/qlogic/qed/qed_debug.c
7641
meta->is_allocated = false;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2666
struct xsk_tx_metadata *meta = NULL;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2697
meta = xsk_buff_get_metadata(pool, xdp_desc.addr);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2730
xsk_tx_metadata_request(meta, &stmmac_xsk_tx_metadata_ops,
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2744
xsk_tx_metadata_to_compl(meta,
drivers/net/wireless/ath/ath11k/dbring.c
357
handler_data.meta = ev->meta_data[num_buff_reaped];
drivers/net/wireless/ath/ath11k/dbring.h
23
struct wmi_dma_buf_release_meta_data meta;
drivers/net/wireless/ath/ath11k/spectral.c
466
struct wmi_dma_buf_release_meta_data *meta,
drivers/net/wireless/ath/ath11k/spectral.c
494
memcpy(&report->meta, meta, sizeof(*meta));
drivers/net/wireless/ath/ath11k/spectral.c
630
chan_width_mhz = summary->meta.ch_width;
drivers/net/wireless/ath/ath11k/spectral.c
661
fft_sample->noise = __cpu_to_be32(summary->meta.noise_floor[search.chain_idx]);
drivers/net/wireless/ath/ath11k/spectral.c
663
freq = summary->meta.freq1;
drivers/net/wireless/ath/ath11k/spectral.c
666
freq = summary->meta.freq2;
drivers/net/wireless/ath/ath11k/spectral.c
77
struct wmi_dma_buf_release_meta_data meta;
drivers/net/wireless/ath/ath11k/spectral.c
782
ath11k_spectral_pull_summary(ar, &param->meta,
drivers/net/wireless/ath/ath12k/dbring.c
297
handler_data.meta = ev->meta_data[num_buff_reaped];
drivers/net/wireless/ath/ath12k/dbring.h
23
struct ath12k_wmi_dma_buf_release_meta_data_params meta;
drivers/net/wireless/ath/ath6kl/txrx.c
1305
struct wmi_rx_meta_v2 *meta;
drivers/net/wireless/ath/ath6kl/txrx.c
1533
meta = (struct wmi_rx_meta_v2 *) skb->data;
drivers/net/wireless/ath/ath6kl/txrx.c
1534
if (meta->csum_flags & 0x1) {
drivers/net/wireless/ath/ath6kl/txrx.c
1536
skb->csum = (__force __wsum) meta->csum;
drivers/net/wireless/ath/ath6kl/txrx.c
368
void *meta;
drivers/net/wireless/ath/ath6kl/txrx.c
421
meta = &meta_v2;
drivers/net/wireless/ath/ath6kl/txrx.c
424
meta = NULL;
drivers/net/wireless/ath/ath6kl/txrx.c
430
meta, vif->fw_vif_idx);
drivers/net/wireless/broadcom/b43/dma.c
1007
kfree(ring->meta);
drivers/net/wireless/broadcom/b43/dma.c
1198
struct b43_dmadesc_meta *meta;
drivers/net/wireless/broadcom/b43/dma.c
1238
desc = ops->idx2desc(ring, slot, &meta);
drivers/net/wireless/broadcom/b43/dma.c
1239
memset(meta, 0, sizeof(*meta));
drivers/net/wireless/broadcom/b43/dma.c
1241
meta->skb = skb;
drivers/net/wireless/broadcom/b43/dma.c
1242
meta->is_last_fragment = true;
drivers/net/wireless/broadcom/b43/dma.c
1245
meta->dmaaddr = map_descbuffer(ring, skb->data, skb->len, 1);
drivers/net/wireless/broadcom/b43/dma.c
1247
if (b43_dma_mapping_error(ring, meta->dmaaddr, skb->len, 1)) {
drivers/net/wireless/broadcom/b43/dma.c
1257
meta->dmaaddr = map_descbuffer(ring, priv_info->bouncebuffer, skb->len, 1);
drivers/net/wireless/broadcom/b43/dma.c
1258
if (b43_dma_mapping_error(ring, meta->dmaaddr, skb->len, 1)) {
drivers/net/wireless/broadcom/b43/dma.c
1268
ops->fill_descriptor(ring, desc, meta->dmaaddr, skb->len, 0, 1, 1);
drivers/net/wireless/broadcom/b43/dma.c
1418
struct b43_dmadesc_meta *meta;
drivers/net/wireless/broadcom/b43/dma.c
1475
ops->idx2desc(ring, slot, &meta);
drivers/net/wireless/broadcom/b43/dma.c
1477
if (b43_dma_ptr_is_poisoned(meta->skb)) {
drivers/net/wireless/broadcom/b43/dma.c
1484
if (meta->skb) {
drivers/net/wireless/broadcom/b43/dma.c
1486
b43_get_priv_tx_info(IEEE80211_SKB_CB(meta->skb));
drivers/net/wireless/broadcom/b43/dma.c
1488
unmap_descbuffer(ring, meta->dmaaddr,
drivers/net/wireless/broadcom/b43/dma.c
1489
meta->skb->len, 1);
drivers/net/wireless/broadcom/b43/dma.c
1493
unmap_descbuffer(ring, meta->dmaaddr,
drivers/net/wireless/broadcom/b43/dma.c
1497
if (meta->is_last_fragment) {
drivers/net/wireless/broadcom/b43/dma.c
1500
if (unlikely(!meta->skb)) {
drivers/net/wireless/broadcom/b43/dma.c
1510
info = IEEE80211_SKB_CB(meta->skb);
drivers/net/wireless/broadcom/b43/dma.c
1533
ieee80211_tx_status_skb(dev->wl->hw, meta->skb);
drivers/net/wireless/broadcom/b43/dma.c
1537
meta->skb = B43_DMA_PTR_POISON;
drivers/net/wireless/broadcom/b43/dma.c
1542
if (unlikely(meta->skb)) {
drivers/net/wireless/broadcom/b43/dma.c
1553
if (meta->is_last_fragment && !skip) {
drivers/net/wireless/broadcom/b43/dma.c
1585
struct b43_dmadesc_meta *meta;
drivers/net/wireless/broadcom/b43/dma.c
1592
desc = ops->idx2desc(ring, *slot, &meta);
drivers/net/wireless/broadcom/b43/dma.c
1594
sync_descbuffer_for_cpu(ring, meta->dmaaddr, ring->rx_buffersize);
drivers/net/wireless/broadcom/b43/dma.c
1595
skb = meta->skb;
drivers/net/wireless/broadcom/b43/dma.c
1608
dmaaddr = meta->dmaaddr;
drivers/net/wireless/broadcom/b43/dma.c
1616
dmaaddr = meta->dmaaddr;
drivers/net/wireless/broadcom/b43/dma.c
1629
desc = ops->idx2desc(ring, *slot, &meta);
drivers/net/wireless/broadcom/b43/dma.c
1631
b43_poison_rx_buffer(ring, meta->skb);
drivers/net/wireless/broadcom/b43/dma.c
1632
sync_descbuffer_for_device(ring, meta->dmaaddr,
drivers/net/wireless/broadcom/b43/dma.c
1646
dmaaddr = meta->dmaaddr;
drivers/net/wireless/broadcom/b43/dma.c
1647
err = setup_rx_descbuffer(ring, desc, meta, GFP_ATOMIC);
drivers/net/wireless/broadcom/b43/dma.c
166
struct b43_dmadesc_meta **meta)
drivers/net/wireless/broadcom/b43/dma.c
170
*meta = &(ring->meta[slot]);
drivers/net/wireless/broadcom/b43/dma.c
396
struct b43_dmadesc_meta *meta)
drivers/net/wireless/broadcom/b43/dma.c
398
if (meta->skb) {
drivers/net/wireless/broadcom/b43/dma.c
400
ieee80211_free_txskb(ring->dev->wl->hw, meta->skb);
drivers/net/wireless/broadcom/b43/dma.c
402
dev_kfree_skb_any(meta->skb);
drivers/net/wireless/broadcom/b43/dma.c
403
meta->skb = NULL;
drivers/net/wireless/broadcom/b43/dma.c
594
struct b43_dmadesc_meta *meta, gfp_t gfp_flags)
drivers/net/wireless/broadcom/b43/dma.c
625
meta->skb = skb;
drivers/net/wireless/broadcom/b43/dma.c
626
meta->dmaaddr = dmaaddr;
drivers/net/wireless/broadcom/b43/dma.c
640
struct b43_dmadesc_meta *meta;
drivers/net/wireless/broadcom/b43/dma.c
643
desc = ring->ops->idx2desc(ring, i, &meta);
drivers/net/wireless/broadcom/b43/dma.c
645
err = setup_rx_descbuffer(ring, desc, meta, GFP_KERNEL);
drivers/net/wireless/broadcom/b43/dma.c
660
desc = ring->ops->idx2desc(ring, i, &meta);
drivers/net/wireless/broadcom/b43/dma.c
662
unmap_descbuffer(ring, meta->dmaaddr, ring->rx_buffersize, 0);
drivers/net/wireless/broadcom/b43/dma.c
663
dev_kfree_skb(meta->skb);
drivers/net/wireless/broadcom/b43/dma.c
74
struct b43_dmadesc_meta **meta)
drivers/net/wireless/broadcom/b43/dma.c
776
struct b43_dmadesc_meta *meta;
drivers/net/wireless/broadcom/b43/dma.c
78
*meta = &(ring->meta[slot]);
drivers/net/wireless/broadcom/b43/dma.c
783
ring->ops->idx2desc(ring, i, &meta);
drivers/net/wireless/broadcom/b43/dma.c
785
if (!meta->skb || b43_dma_ptr_is_poisoned(meta->skb)) {
drivers/net/wireless/broadcom/b43/dma.c
790
unmap_descbuffer(ring, meta->dmaaddr,
drivers/net/wireless/broadcom/b43/dma.c
791
meta->skb->len, 1);
drivers/net/wireless/broadcom/b43/dma.c
793
unmap_descbuffer(ring, meta->dmaaddr,
drivers/net/wireless/broadcom/b43/dma.c
796
free_descriptor_buffer(ring, meta);
drivers/net/wireless/broadcom/b43/dma.c
849
ring->meta = kzalloc_objs(struct b43_dmadesc_meta, ring->nr_slots);
drivers/net/wireless/broadcom/b43/dma.c
850
if (!ring->meta)
drivers/net/wireless/broadcom/b43/dma.c
853
ring->meta->skb = B43_DMA_PTR_POISON;
drivers/net/wireless/broadcom/b43/dma.c
946
kfree(ring->meta);
drivers/net/wireless/broadcom/b43/dma.h
202
meta);
drivers/net/wireless/broadcom/b43/dma.h
232
struct b43_dmadesc_meta *meta;
drivers/net/wireless/broadcom/b43legacy/dma.c
1000
if (b43legacy_dma_mapping_error(ring, meta->dmaaddr, skb->len, 1)) {
drivers/net/wireless/broadcom/b43legacy/dma.c
1008
op32_fill_descriptor(ring, desc, meta->dmaaddr,
drivers/net/wireless/broadcom/b43legacy/dma.c
1104
struct b43legacy_dmadesc_meta *meta;
drivers/net/wireless/broadcom/b43legacy/dma.c
1132
op32_idx2desc(ring, slot, &meta);
drivers/net/wireless/broadcom/b43legacy/dma.c
1134
if (meta->skb)
drivers/net/wireless/broadcom/b43legacy/dma.c
1135
unmap_descbuffer(ring, meta->dmaaddr,
drivers/net/wireless/broadcom/b43legacy/dma.c
1136
meta->skb->len, 1);
drivers/net/wireless/broadcom/b43legacy/dma.c
1138
unmap_descbuffer(ring, meta->dmaaddr,
drivers/net/wireless/broadcom/b43legacy/dma.c
1142
if (meta->is_last_fragment) {
drivers/net/wireless/broadcom/b43legacy/dma.c
1144
BUG_ON(!meta->skb);
drivers/net/wireless/broadcom/b43legacy/dma.c
1145
info = IEEE80211_SKB_CB(meta->skb);
drivers/net/wireless/broadcom/b43legacy/dma.c
1182
ieee80211_tx_status_irqsafe(dev->wl->hw, meta->skb);
drivers/net/wireless/broadcom/b43legacy/dma.c
1184
meta->skb = NULL;
drivers/net/wireless/broadcom/b43legacy/dma.c
1189
B43legacy_WARN_ON(meta->skb != NULL);
drivers/net/wireless/broadcom/b43legacy/dma.c
1195
if (meta->is_last_fragment)
drivers/net/wireless/broadcom/b43legacy/dma.c
1223
struct b43legacy_dmadesc_meta *meta;
drivers/net/wireless/broadcom/b43legacy/dma.c
1230
desc = op32_idx2desc(ring, *slot, &meta);
drivers/net/wireless/broadcom/b43legacy/dma.c
1232
sync_descbuffer_for_cpu(ring, meta->dmaaddr, ring->rx_buffersize);
drivers/net/wireless/broadcom/b43legacy/dma.c
1233
skb = meta->skb;
drivers/net/wireless/broadcom/b43legacy/dma.c
1250
sync_descbuffer_for_device(ring, meta->dmaaddr,
drivers/net/wireless/broadcom/b43legacy/dma.c
1267
sync_descbuffer_for_device(ring, meta->dmaaddr,
drivers/net/wireless/broadcom/b43legacy/dma.c
1282
desc = op32_idx2desc(ring, *slot, &meta);
drivers/net/wireless/broadcom/b43legacy/dma.c
1284
sync_descbuffer_for_device(ring, meta->dmaaddr,
drivers/net/wireless/broadcom/b43legacy/dma.c
1298
dmaaddr = meta->dmaaddr;
drivers/net/wireless/broadcom/b43legacy/dma.c
1299
err = setup_rx_descbuffer(ring, desc, meta, GFP_ATOMIC);
drivers/net/wireless/broadcom/b43legacy/dma.c
285
struct b43legacy_dmadesc_meta *meta,
drivers/net/wireless/broadcom/b43legacy/dma.c
288
if (meta->skb) {
drivers/net/wireless/broadcom/b43legacy/dma.c
290
dev_kfree_skb_irq(meta->skb);
drivers/net/wireless/broadcom/b43legacy/dma.c
292
dev_kfree_skb(meta->skb);
drivers/net/wireless/broadcom/b43legacy/dma.c
293
meta->skb = NULL;
drivers/net/wireless/broadcom/b43legacy/dma.c
34
struct b43legacy_dmadesc_meta **meta)
drivers/net/wireless/broadcom/b43legacy/dma.c
38
*meta = &(ring->meta[slot]);
drivers/net/wireless/broadcom/b43legacy/dma.c
421
struct b43legacy_dmadesc_meta *meta,
drivers/net/wireless/broadcom/b43legacy/dma.c
454
meta->skb = skb;
drivers/net/wireless/broadcom/b43legacy/dma.c
455
meta->dmaaddr = dmaaddr;
drivers/net/wireless/broadcom/b43legacy/dma.c
474
struct b43legacy_dmadesc_meta *meta;
drivers/net/wireless/broadcom/b43legacy/dma.c
477
desc = op32_idx2desc(ring, i, &meta);
drivers/net/wireless/broadcom/b43legacy/dma.c
479
err = setup_rx_descbuffer(ring, desc, meta, GFP_KERNEL);
drivers/net/wireless/broadcom/b43legacy/dma.c
494
desc = op32_idx2desc(ring, i, &meta);
drivers/net/wireless/broadcom/b43legacy/dma.c
496
unmap_descbuffer(ring, meta->dmaaddr, ring->rx_buffersize, 0);
drivers/net/wireless/broadcom/b43legacy/dma.c
497
dev_kfree_skb(meta->skb);
drivers/net/wireless/broadcom/b43legacy/dma.c
563
struct b43legacy_dmadesc_meta *meta;
drivers/net/wireless/broadcom/b43legacy/dma.c
569
op32_idx2desc(ring, i, &meta);
drivers/net/wireless/broadcom/b43legacy/dma.c
571
if (!meta->skb) {
drivers/net/wireless/broadcom/b43legacy/dma.c
576
unmap_descbuffer(ring, meta->dmaaddr,
drivers/net/wireless/broadcom/b43legacy/dma.c
577
meta->skb->len, 1);
drivers/net/wireless/broadcom/b43legacy/dma.c
579
unmap_descbuffer(ring, meta->dmaaddr,
drivers/net/wireless/broadcom/b43legacy/dma.c
581
free_descriptor_buffer(ring, meta, 0);
drivers/net/wireless/broadcom/b43legacy/dma.c
623
ring->meta = kzalloc_objs(struct b43legacy_dmadesc_meta, nr_slots);
drivers/net/wireless/broadcom/b43legacy/dma.c
624
if (!ring->meta)
drivers/net/wireless/broadcom/b43legacy/dma.c
698
kfree(ring->meta);
drivers/net/wireless/broadcom/b43legacy/dma.c
723
kfree(ring->meta);
drivers/net/wireless/broadcom/b43legacy/dma.c
933
struct b43legacy_dmadesc_meta *meta;
drivers/net/wireless/broadcom/b43legacy/dma.c
972
desc = op32_idx2desc(ring, slot, &meta);
drivers/net/wireless/broadcom/b43legacy/dma.c
973
memset(meta, 0, sizeof(*meta));
drivers/net/wireless/broadcom/b43legacy/dma.c
975
meta->skb = skb;
drivers/net/wireless/broadcom/b43legacy/dma.c
976
meta->is_last_fragment = true;
drivers/net/wireless/broadcom/b43legacy/dma.c
978
meta->dmaaddr = map_descbuffer(ring, skb->data, skb->len, 1);
drivers/net/wireless/broadcom/b43legacy/dma.c
980
if (b43legacy_dma_mapping_error(ring, meta->dmaaddr, skb->len, 1)) {
drivers/net/wireless/broadcom/b43legacy/dma.c
998
meta->skb = skb;
drivers/net/wireless/broadcom/b43legacy/dma.c
999
meta->dmaaddr = map_descbuffer(ring, skb->data, skb->len, 1);
drivers/net/wireless/broadcom/b43legacy/dma.h
126
struct b43legacy_dmadesc_meta *meta;
drivers/net/wireless/intel/iwlegacy/3945-mac.c
525
out_meta = &txq->meta[idx];
drivers/net/wireless/intel/iwlegacy/3945.c
656
dma_unmap_addr(&txq->meta[idx], mapping),
drivers/net/wireless/intel/iwlegacy/3945.c
657
dma_unmap_len(&txq->meta[idx], len),
drivers/net/wireless/intel/iwlegacy/4965-mac.c
1774
out_meta = &txq->meta[q->write_ptr];
drivers/net/wireless/intel/iwlegacy/4965-mac.c
3935
dma_unmap_addr(&txq->meta[idx], mapping),
drivers/net/wireless/intel/iwlegacy/4965-mac.c
3936
dma_unmap_len(&txq->meta[idx], len),
drivers/net/wireless/intel/iwlegacy/common.c
2798
kfree(txq->meta);
drivers/net/wireless/intel/iwlegacy/common.c
2800
txq->meta = NULL;
drivers/net/wireless/intel/iwlegacy/common.c
2823
if (txq->meta[i].flags & CMD_MAPPED) {
drivers/net/wireless/intel/iwlegacy/common.c
2825
dma_unmap_addr(&txq->meta[i], mapping),
drivers/net/wireless/intel/iwlegacy/common.c
2826
dma_unmap_len(&txq->meta[i], len),
drivers/net/wireless/intel/iwlegacy/common.c
2828
txq->meta[i].flags = 0;
drivers/net/wireless/intel/iwlegacy/common.c
2835
if (txq->meta[i].flags & CMD_MAPPED) {
drivers/net/wireless/intel/iwlegacy/common.c
2837
dma_unmap_addr(&txq->meta[i], mapping),
drivers/net/wireless/intel/iwlegacy/common.c
2838
dma_unmap_len(&txq->meta[i], len),
drivers/net/wireless/intel/iwlegacy/common.c
2840
txq->meta[i].flags = 0;
drivers/net/wireless/intel/iwlegacy/common.c
2874
kfree(txq->meta);
drivers/net/wireless/intel/iwlegacy/common.c
2876
txq->meta = NULL;
drivers/net/wireless/intel/iwlegacy/common.c
3023
txq->meta =
drivers/net/wireless/intel/iwlegacy/common.c
3028
if (!txq->meta || !txq->cmd)
drivers/net/wireless/intel/iwlegacy/common.c
3068
kfree(txq->meta);
drivers/net/wireless/intel/iwlegacy/common.c
3069
txq->meta = NULL;
drivers/net/wireless/intel/iwlegacy/common.c
3091
memset(txq->meta, 0, sizeof(struct il_cmd_meta) * actual_slots);
drivers/net/wireless/intel/iwlegacy/common.c
3156
out_meta = &txq->meta[idx];
drivers/net/wireless/intel/iwlegacy/common.c
3289
struct il_cmd_meta *meta;
drivers/net/wireless/intel/iwlegacy/common.c
3307
meta = &txq->meta[cmd_idx];
drivers/net/wireless/intel/iwlegacy/common.c
3311
dma_unmap_single(&il->pci_dev->dev, dma_unmap_addr(meta, mapping),
drivers/net/wireless/intel/iwlegacy/common.c
3312
dma_unmap_len(meta, len), DMA_BIDIRECTIONAL);
drivers/net/wireless/intel/iwlegacy/common.c
3315
if (meta->flags & CMD_WANT_SKB) {
drivers/net/wireless/intel/iwlegacy/common.c
3316
meta->source->reply_page = (unsigned long)rxb_addr(rxb);
drivers/net/wireless/intel/iwlegacy/common.c
3318
} else if (meta->callback)
drivers/net/wireless/intel/iwlegacy/common.c
3319
meta->callback(il, cmd, pkt);
drivers/net/wireless/intel/iwlegacy/common.c
3325
if (!(meta->flags & CMD_ASYNC)) {
drivers/net/wireless/intel/iwlegacy/common.c
3333
meta->flags = 0;
drivers/net/wireless/intel/iwlegacy/common.c
364
il->txq[il->cmd_queue].meta[cmd_idx].flags &= ~CMD_WANT_SKB;
drivers/net/wireless/intel/iwlegacy/common.h
154
struct il_cmd_meta *meta;
drivers/net/wireless/intel/iwlwifi/iwl-trans.h
720
struct iwl_cmd_meta meta;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/internal.h
758
struct iwl_cmd_meta *meta,
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx-gen2.c
1303
out_meta = &txq->entries[idx].meta;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx-gen2.c
133
meta = NULL;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx-gen2.c
147
if (meta)
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx-gen2.c
63
u16 len, struct iwl_cmd_meta *meta,
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx-gen2.c
638
struct iwl_cmd_meta *meta,
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx-gen2.c
653
if (meta->sg_offset)
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx-gen2.c
658
if (meta->tbs & BIT(i))
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx-gen2.c
686
iwl_txq_gen2_tfd_unmap(trans, &txq->entries[idx].meta,
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx-gen2.c
771
out_meta = &txq->entries[idx].meta;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx-gen2.c
79
if (meta)
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx-gen2.c
80
meta->tbs |= BIT(ret);
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx-gen2.c
829
struct iwl_cmd_meta *cmd_meta = &txq->entries[idx].meta;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
1459
out_meta = &txq->entries[idx].meta;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
1632
struct iwl_cmd_meta *meta;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
1649
meta = &txq->entries[cmd_index].meta;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
1654
iwl_txq_gen2_tfd_unmap(trans, meta,
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
1657
iwl_txq_gen1_tfd_unmap(trans, meta, txq, index);
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
1660
if (meta->flags & CMD_WANT_SKB) {
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
1663
meta->source->resp_pkt = pkt;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
1664
meta->source->_rx_page_addr = (unsigned long)page_address(p);
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
1665
meta->source->_rx_page_order = trans_pcie->rx_page_order;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
1668
if (meta->flags & CMD_BLOCK_TXQS)
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
1673
if (!(meta->flags & CMD_ASYNC)) {
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
1685
meta->flags = 0;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
2192
out_meta = &txq->entries[txq->write_ptr].meta;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
2403
struct iwl_cmd_meta *cmd_meta = &txq->entries[read_ptr].meta;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
2635
txq->entries[cmd_idx].meta.flags &= ~CMD_WANT_SKB;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
292
struct iwl_cmd_meta *meta,
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
309
if (meta->sg_offset)
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
315
if (meta->tbs & BIT(i))
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
329
meta->tbs = 0;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
361
iwl_txq_gen2_tfd_unmap(trans, &txq->entries[idx].meta,
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
364
iwl_txq_gen1_tfd_unmap(trans, &txq->entries[idx].meta,
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
402
&txq->entries[txq->read_ptr].meta;
drivers/net/wireless/quantenna/qtnfmac/core.c
822
struct qtnf_frame_meta_info *meta;
drivers/net/wireless/quantenna/qtnfmac/core.c
830
meta = (struct qtnf_frame_meta_info *)
drivers/net/wireless/quantenna/qtnfmac/core.c
831
(skb_tail_pointer(skb) - sizeof(*meta));
drivers/net/wireless/quantenna/qtnfmac/core.c
833
if (unlikely(!qtnf_is_frame_meta_magic_valid(meta))) {
drivers/net/wireless/quantenna/qtnfmac/core.c
835
meta->magic_s, meta->magic_e);
drivers/net/wireless/quantenna/qtnfmac/core.c
839
if (unlikely(meta->macid >= QTNF_MAX_MAC)) {
drivers/net/wireless/quantenna/qtnfmac/core.c
840
pr_err_ratelimited("invalid mac(%u)\n", meta->macid);
drivers/net/wireless/quantenna/qtnfmac/core.c
844
if (unlikely(meta->ifidx >= QTNF_MAX_INTF)) {
drivers/net/wireless/quantenna/qtnfmac/core.c
845
pr_err_ratelimited("invalid vif(%u)\n", meta->ifidx);
drivers/net/wireless/quantenna/qtnfmac/core.c
849
mac = bus->mac[meta->macid];
drivers/net/wireless/quantenna/qtnfmac/core.c
852
pr_err_ratelimited("mac(%d) does not exist\n", meta->macid);
drivers/net/wireless/quantenna/qtnfmac/core.c
856
vif = &mac->iflist[meta->ifidx];
drivers/net/wireless/quantenna/qtnfmac/core.c
859
pr_err_ratelimited("vif(%u) does not exists\n", meta->ifidx);
drivers/net/wireless/quantenna/qtnfmac/core.c
867
meta->macid, meta->ifidx);
drivers/net/wireless/quantenna/qtnfmac/core.c
871
__skb_trim(skb, skb->len - sizeof(*meta));
drivers/net/wireless/quantenna/qtnfmac/pcie/pearl_pcie.c
620
struct qtnf_frame_meta_info *meta;
drivers/net/wireless/quantenna/qtnfmac/pcie/pearl_pcie.c
621
int tail_need = sizeof(*meta) - skb_tailroom(skb);
drivers/net/wireless/quantenna/qtnfmac/pcie/pearl_pcie.c
630
meta = skb_put(skb, sizeof(*meta));
drivers/net/wireless/quantenna/qtnfmac/pcie/pearl_pcie.c
631
meta->magic_s = HBM_FRAME_META_MAGIC_PATTERN_S;
drivers/net/wireless/quantenna/qtnfmac/pcie/pearl_pcie.c
632
meta->magic_e = HBM_FRAME_META_MAGIC_PATTERN_E;
drivers/net/wireless/quantenna/qtnfmac/pcie/pearl_pcie.c
633
meta->macid = macid;
drivers/net/wireless/quantenna/qtnfmac/pcie/pearl_pcie.c
634
meta->ifidx = vifid;
drivers/net/wireless/quantenna/qtnfmac/pcie/pearl_pcie.c
638
__skb_trim(skb, skb->len - sizeof(*meta));
drivers/staging/media/ipu3/ipu3-css.h
94
struct v4l2_meta_format meta;
drivers/staging/media/ipu3/ipu3-v4l2.c
1064
f->fmt.meta.dataformat = V4L2_META_FMT_IPU3_PARAMS;
drivers/staging/media/ipu3/ipu3-v4l2.c
1066
imgu_css_meta_fmt_set(&f->fmt.meta);
drivers/staging/media/ipu3/ipu3-v4l2.c
1071
f->fmt.meta.dataformat = V4L2_META_FMT_IPU3_STAT_3A;
drivers/staging/media/ipu3/ipu3-v4l2.c
1073
imgu_css_meta_fmt_set(&f->fmt.meta);
drivers/staging/media/ipu3/ipu3-v4l2.c
359
need_bytes = node->vdev_fmt.fmt.meta.buffersize;
drivers/staging/media/ipu3/ipu3-v4l2.c
404
size = fmt->fmt.meta.buffersize;
drivers/staging/media/ipu3/ipu3.c
110
const struct v4l2_meta_format *meta;
drivers/staging/media/ipu3/ipu3.c
129
meta = &imgu_pipe->nodes[node].vdev_fmt.fmt.meta;
drivers/staging/media/ipu3/ipu3.c
133
size = meta->buffersize;
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_params.c
299
unsigned int size = isp_dev->params.format.fmt.meta.buffersize;
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_params.c
315
unsigned int size = isp_dev->params.format.fmt.meta.buffersize;
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_params.c
405
&isp_dev->params.format.fmt.meta;
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_params.c
465
struct v4l2_meta_format *params_format = &format->fmt.meta;
drivers/ufs/core/ufs_bsg.c
85
rpmb_req_type = be16_to_cpu(rpmb_request->ehs_req.meta.req_resp_type);
fs/afs/dir_edit.c
210
static void afs_edit_init_block(union afs_xdr_dir_block *meta,
fs/afs/dir_edit.c
221
memset(block->meta.alloc_ctrs,
fs/afs/dir_edit.c
223
sizeof(block->meta.alloc_ctrs));
fs/afs/dir_edit.c
224
meta->meta.alloc_ctrs[0] =
fs/afs/dir_edit.c
229
meta->meta.alloc_ctrs[block_num] =
fs/afs/dir_edit.c
245
union afs_xdr_dir_block *meta, *block;
fs/afs/dir_edit.c
261
meta = afs_dir_get_block(&iter, 0);
fs/afs/dir_edit.c
262
if (!meta)
fs/afs/dir_edit.c
284
meta->meta.alloc_ctrs[b] < iter.nr_slots)
fs/afs/dir_edit.c
297
(b < AFS_DIR_BLOCKS_WITH_CTR) ? meta->meta.alloc_ctrs[b] : 99,
fs/afs/dir_edit.c
304
afs_edit_init_block(meta, block, b);
fs/afs/dir_edit.c
328
afs_edit_init_block(meta, meta, 0);
fs/afs/dir_edit.c
354
meta->meta.alloc_ctrs[b] -= iter.nr_slots;
fs/afs/dir_edit.c
359
de->u.hash_next = meta->meta.hashtable[iter.bucket];
fs/afs/dir_edit.c
360
meta->meta.hashtable[iter.bucket] = htons(entry);
fs/afs/dir_edit.c
370
kunmap_local(meta);
fs/afs/dir_edit.c
396
union afs_xdr_dir_block *meta, *block, *pblock;
fs/afs/dir_edit.c
418
meta = afs_dir_find_block(&iter, 0);
fs/afs/dir_edit.c
419
if (!meta)
fs/afs/dir_edit.c
456
meta->meta.alloc_ctrs[b] += iter.nr_slots;
fs/afs/dir_edit.c
467
__be16 prev_next = meta->meta.hashtable[iter.bucket];
fs/afs/dir_edit.c
476
meta->meta.hashtable[iter.bucket] = next;
fs/afs/dir_edit.c
506
kunmap_local(meta);
fs/afs/dir_edit.c
606
union afs_xdr_dir_block *meta;
fs/afs/dir_edit.c
618
meta = afs_dir_get_block(&iter, 0);
fs/afs/dir_edit.c
619
if (!meta)
fs/afs/dir_edit.c
622
afs_edit_init_block(meta, meta, 0);
fs/afs/dir_edit.c
624
de = &meta->dirents[slot];
fs/afs/dir_edit.c
633
de = &meta->dirents[slot];
fs/afs/dir_edit.c
641
afs_set_contig_bits(meta, AFS_DIR_RESV_BLOCKS0, 2);
fs/afs/dir_edit.c
642
meta->meta.alloc_ctrs[0] -= 2;
fs/afs/dir_edit.c
643
kunmap_local(meta);
fs/afs/dir_search.c
126
const union afs_xdr_dir_block *meta;
fs/afs/dir_search.c
130
meta = afs_dir_find_block(iter, 0);
fs/afs/dir_search.c
131
if (!meta)
fs/afs/dir_search.c
134
entry = ntohs(meta->meta.hashtable[iter->bucket & (AFS_DIR_HASHTBL_SIZE - 1)]);
fs/afs/xdr_fs.h
93
} meta;
fs/binfmt_elf.c
2070
struct core_vma_metadata *meta = cprm->vma_meta + i;
fs/binfmt_elf.c
2075
phdr.p_vaddr = meta->start;
fs/binfmt_elf.c
2077
phdr.p_filesz = meta->dump_size;
fs/binfmt_elf.c
2078
phdr.p_memsz = meta->end - meta->start;
fs/binfmt_elf.c
2081
if (meta->flags & VM_READ)
fs/binfmt_elf.c
2083
if (meta->flags & VM_WRITE)
fs/binfmt_elf.c
2085
if (meta->flags & VM_EXEC)
fs/binfmt_elf.c
2108
struct core_vma_metadata *meta = cprm->vma_meta + i;
fs/binfmt_elf.c
2110
if (!dump_user_range(cprm, meta->start, meta->dump_size))
fs/binfmt_elf_fdpic.c
1460
struct core_vma_metadata *meta = vma_meta + i;
fs/binfmt_elf_fdpic.c
1462
if (!dump_user_range(cprm, meta->start, meta->dump_size))
fs/binfmt_elf_fdpic.c
1587
struct core_vma_metadata *meta = cprm->vma_meta + i;
fs/binfmt_elf_fdpic.c
1591
sz = meta->end - meta->start;
fs/binfmt_elf_fdpic.c
1595
phdr.p_vaddr = meta->start;
fs/binfmt_elf_fdpic.c
1597
phdr.p_filesz = meta->dump_size;
fs/binfmt_elf_fdpic.c
1601
if (meta->flags & VM_READ)
fs/binfmt_elf_fdpic.c
1603
if (meta->flags & VM_WRITE)
fs/binfmt_elf_fdpic.c
1605
if (meta->flags & VM_EXEC)
fs/btrfs/accessors.h
814
read_eb_member(eb, bi, struct btrfs_balance_item, meta, ba);
fs/btrfs/accessors.h
821
write_eb_member(eb, bi, struct btrfs_balance_item, meta, ba);
fs/btrfs/block-group.c
71
bctl->meta.flags & BTRFS_BALANCE_ARGS_CONVERT) {
fs/btrfs/block-group.c
72
target = BTRFS_BLOCK_GROUP_METADATA | bctl->meta.target;
fs/btrfs/ioctl.c
3304
memcpy(&bargs->meta, &bctl->meta, sizeof(bargs->meta));
fs/btrfs/ioctl.c
3439
memcpy(&bctl->meta, &bargs->meta, sizeof(bctl->meta));
fs/btrfs/volumes.c
3818
btrfs_cpu_balance_args_to_disk(&disk_bargs, &bctl->meta);
fs/btrfs/volumes.c
3887
if (bctl->meta.flags & BTRFS_BALANCE_ARGS_CONVERT)
fs/btrfs/volumes.c
3888
bctl->meta.flags |= BTRFS_BALANCE_ARGS_SOFT;
fs/btrfs/volumes.c
3909
if (!(bctl->meta.flags & BTRFS_BALANCE_ARGS_USAGE) &&
fs/btrfs/volumes.c
3910
!(bctl->meta.flags & BTRFS_BALANCE_ARGS_USAGE_RANGE) &&
fs/btrfs/volumes.c
3911
!(bctl->meta.flags & BTRFS_BALANCE_ARGS_CONVERT)) {
fs/btrfs/volumes.c
3912
bctl->meta.flags |= BTRFS_BALANCE_ARGS_USAGE;
fs/btrfs/volumes.c
3913
bctl->meta.usage = 90;
fs/btrfs/volumes.c
4131
bargs = &bctl->meta;
fs/btrfs/volumes.c
4327
u64 limit_meta = bctl->meta.limit;
fs/btrfs/volumes.c
4354
bctl->meta.limit = limit_meta;
fs/btrfs/volumes.c
4454
count_meta < bctl->meta.limit_min)
fs/btrfs/volumes.c
4729
describe_balance_args(&bctl->meta, tmp_buf, sizeof(tmp_buf));
fs/btrfs/volumes.c
4785
memcmp(&bctl->data, &bctl->meta, sizeof(bctl->data))) {
fs/btrfs/volumes.c
4810
!validate_convert_profile(fs_info, &bctl->meta, allowed, "metadata") ||
fs/btrfs/volumes.c
4832
((bctl->meta.flags & BTRFS_BALANCE_ARGS_CONVERT) &&
fs/btrfs/volumes.c
4834
!(bctl->meta.target & allowed)))
fs/btrfs/volumes.c
4840
meta_target = (bctl->meta.flags & BTRFS_BALANCE_ARGS_CONVERT) ?
fs/btrfs/volumes.c
4841
bctl->meta.target : fs_info->avail_metadata_alloc_bits;
fs/btrfs/volumes.c
5029
btrfs_disk_balance_args_to_cpu(&bctl->meta, &disk_bargs);
fs/btrfs/volumes.h
641
struct btrfs_balance_args meta;
fs/ceph/metric.c
100
meta = (struct ceph_metric_metadata_latency *)(write + 1);
fs/ceph/metric.c
101
meta->header.type = cpu_to_le32(CLIENT_METRIC_TYPE_METADATA_LATENCY);
fs/ceph/metric.c
102
meta->header.ver = 2;
fs/ceph/metric.c
103
meta->header.compat = 1;
fs/ceph/metric.c
104
meta->header.data_len = cpu_to_le32(sizeof(*meta) - header_len);
fs/ceph/metric.c
106
ktime_to_ceph_timespec(&meta->lat, sum);
fs/ceph/metric.c
107
ktime_to_ceph_timespec(&meta->avg, m->metric[METRIC_METADATA].latency_avg);
fs/ceph/metric.c
108
meta->sq_sum = cpu_to_le64(m->metric[METRIC_METADATA].latency_sq_sum);
fs/ceph/metric.c
109
meta->count = cpu_to_le64(m->metric[METRIC_METADATA].total);
fs/ceph/metric.c
113
dlease = (struct ceph_metric_dlease *)(meta + 1);
fs/ceph/metric.c
24
struct ceph_metric_metadata_latency *meta;
fs/ceph/metric.c
49
+ sizeof(*meta) + sizeof(*dlease) + sizeof(*files)
fs/f2fs/segment.c
477
unsigned int meta = get_pages(sbi, F2FS_DIRTY_META);
fs/f2fs/segment.c
484
nodes >= threshold || meta >= threshold ||
fs/f2fs/segment.c
487
return dents + qdata + nodes + meta + imeta > global_threshold;
fs/gfs2/bmap.c
1473
bool meta, u32 *btotal)
fs/gfs2/bmap.c
1540
if (meta)
fs/gfs2/bmap.c
1582
__gfs2_free_blocks(ip, rgd, bstart, (u32)blen, meta);
fs/gfs2/bmap.c
1590
__gfs2_free_blocks(ip, rgd, bstart, (u32)blen, meta);
fs/gfs2/meta_io.c
341
void gfs2_remove_from_journal(struct buffer_head *bh, int meta)
fs/gfs2/meta_io.c
353
if (meta == REMOVE_META)
fs/gfs2/meta_io.h
62
void gfs2_remove_from_journal(struct buffer_head *bh, int meta);
fs/gfs2/rgrp.c
2518
u64 bstart, u32 blen, int meta)
fs/gfs2/rgrp.c
2532
if (meta || ip->i_depth || gfs2_is_jdata(ip))
fs/gfs2/rgrp.h
50
u64 bstart, u32 blen, int meta);
fs/ocfs2/ocfs2_trace.h
1738
TP_PROTO(const char *name, int meta, int clusters, int credits),
fs/ocfs2/ocfs2_trace.h
1739
TP_ARGS(name, meta, clusters, credits),
fs/ocfs2/ocfs2_trace.h
1742
__field(int, meta)
fs/ocfs2/ocfs2_trace.h
1748
__entry->meta = meta;
fs/ocfs2/ocfs2_trace.h
1752
TP_printk("%s %d %d %d", __get_str(name), __entry->meta,
fs/squashfs/file.c
133
meta = &msblk->meta_index[msblk->next_meta_index];
fs/squashfs/file.c
137
meta->inode_number = inode->i_ino;
fs/squashfs/file.c
138
meta->offset = offset;
fs/squashfs/file.c
139
meta->skip = skip;
fs/squashfs/file.c
140
meta->entries = 0;
fs/squashfs/file.c
141
meta->locked = 1;
fs/squashfs/file.c
145
return meta;
fs/squashfs/file.c
149
static void release_meta_index(struct inode *inode, struct meta_index *meta)
fs/squashfs/file.c
153
meta->locked = 0;
fs/squashfs/file.c
235
struct meta_index *meta;
fs/squashfs/file.c
248
meta = locate_meta_index(inode, offset + 1, index);
fs/squashfs/file.c
250
if (meta == NULL) {
fs/squashfs/file.c
251
meta = empty_meta_index(inode, offset + 1, skip);
fs/squashfs/file.c
252
if (meta == NULL)
fs/squashfs/file.c
255
offset = index < meta->offset + meta->entries ? index :
fs/squashfs/file.c
256
meta->offset + meta->entries - 1;
fs/squashfs/file.c
257
meta_entry = &meta->meta_entry[offset - meta->offset];
fs/squashfs/file.c
263
"meta->entries %d\n", offset, meta->offset,
fs/squashfs/file.c
264
meta->entries);
fs/squashfs/file.c
275
for (i = meta->offset + meta->entries; i <= index &&
fs/squashfs/file.c
276
i < meta->offset + SQUASHFS_META_ENTRIES; i++) {
fs/squashfs/file.c
282
if (meta->entries == 0)
fs/squashfs/file.c
287
meta->inode_number = 0;
fs/squashfs/file.c
293
meta_entry = &meta->meta_entry[i - meta->offset];
fs/squashfs/file.c
298
meta->entries++;
fs/squashfs/file.c
303
meta->offset, meta->entries);
fs/squashfs/file.c
305
release_meta_index(inode, meta);
fs/squashfs/file.c
320
release_meta_index(inode, meta);
fs/squashfs/file.c
52
struct meta_index *meta = NULL;
fs/squashfs/file.c
70
meta = &msblk->meta_index[i];
fs/squashfs/file.c
71
offset = meta->offset;
fs/squashfs/file.c
75
if (meta)
fs/squashfs/file.c
76
meta->locked = 1;
fs/squashfs/file.c
81
return meta;
fs/squashfs/file.c
92
struct meta_index *meta = NULL;
include/linux/bio-integrity.h
79
int bio_integrity_map_iter(struct bio *bio, struct uio_meta *meta);
include/linux/bio-integrity.h
98
static inline int bio_integrity_map_iter(struct bio *bio, struct uio_meta *meta)
include/linux/bpf.h
2344
u64 bpf_event_output(struct bpf_map *map, u64 flags, void *meta, u64 meta_size,
include/linux/bpf.h
2874
__bpf_md_ptr(struct bpf_iter_meta *, meta);
include/linux/bpf.h
2888
struct bpf_prog *bpf_iter_get_info(struct bpf_iter_meta *meta, bool in_stop);
include/linux/buffer_head.h
136
BUFFER_FNS(Meta, meta)
include/linux/maple_tree.h
110
struct maple_metadata meta;
include/linux/maple_tree.h
129
struct maple_metadata meta;
include/linux/skbuff.h
4599
u8 *meta, *meta_end;
include/linux/skbuff.h
4608
meta = meta_end - meta_len;
include/linux/skbuff.h
4616
memmove(meta + len, meta, meta_len + n);
include/net/ip6_fib.h
604
__bpf_md_ptr(struct bpf_iter_meta *, meta);
include/net/libeth/xsk.h
208
xsk_tx_metadata_request(ctx.meta, tmo, &desc);
include/net/netfilter/nf_flow_table.h
22
struct flow_dissector_key_meta meta;
include/net/netfilter/nf_tables_offload.h
55
struct flow_dissector_key_meta meta;
include/net/xdp_sock.h
140
static inline void xsk_tx_metadata_to_compl(struct xsk_tx_metadata *meta,
include/net/xdp_sock.h
143
if (!meta)
include/net/xdp_sock.h
146
if (meta->flags & XDP_TXMD_FLAGS_TIMESTAMP)
include/net/xdp_sock.h
147
compl->tx_timestamp = &meta->completion.tx_timestamp;
include/net/xdp_sock.h
162
static inline void xsk_tx_metadata_request(const struct xsk_tx_metadata *meta,
include/net/xdp_sock.h
166
if (!meta)
include/net/xdp_sock.h
170
if (meta->flags & XDP_TXMD_FLAGS_LAUNCH_TIME)
include/net/xdp_sock.h
171
ops->tmo_request_launch_time(meta->request.launch_time,
include/net/xdp_sock.h
175
if (meta->flags & XDP_TXMD_FLAGS_TIMESTAMP)
include/net/xdp_sock.h
179
if (meta->flags & XDP_TXMD_FLAGS_CHECKSUM)
include/net/xdp_sock.h
180
ops->tmo_request_checksum(meta->request.csum_start,
include/net/xdp_sock.h
181
meta->request.csum_offset, priv);
include/net/xdp_sock.h
228
static inline void xsk_tx_metadata_to_compl(struct xsk_tx_metadata *meta,
include/net/xdp_sock.h
233
static inline void xsk_tx_metadata_request(struct xsk_tx_metadata *meta,
include/net/xdp_sock_drv.h
263
xsk_buff_valid_tx_metadata(const struct xsk_tx_metadata *meta)
include/net/xdp_sock_drv.h
265
return !(meta->flags & ~XDP_TXMD_FLAGS_VALID);
include/net/xdp_sock_drv.h
271
struct xsk_tx_metadata *meta;
include/net/xdp_sock_drv.h
276
meta = data - pool->tx_metadata_len;
include/net/xdp_sock_drv.h
277
if (unlikely(!xsk_buff_valid_tx_metadata(meta)))
include/net/xdp_sock_drv.h
280
return meta;
include/net/xdp_sock_drv.h
472
static inline bool xsk_buff_valid_tx_metadata(struct xsk_tx_metadata *meta)
include/net/xsk_buff_pool.h
149
struct xsk_tx_metadata *meta;
include/uapi/linux/btrfs.h
484
struct btrfs_balance_args meta; /* in/out */
include/uapi/linux/btrfs_tree.h
1065
struct btrfs_disk_balance_args meta;
include/uapi/linux/fanotify.h
266
#define FAN_EVENT_NEXT(meta, len) ((len) -= (meta)->event_len, \
include/uapi/linux/fanotify.h
267
(struct fanotify_event_metadata*)(((char *)(meta)) + \
include/uapi/linux/fanotify.h
268
(meta)->event_len))
include/uapi/linux/fanotify.h
270
#define FAN_EVENT_OK(meta, len) ((long)(len) >= (long)FAN_EVENT_METADATA_LEN && \
include/uapi/linux/fanotify.h
271
(long)(meta)->event_len >= (long)FAN_EVENT_METADATA_LEN && \
include/uapi/linux/fanotify.h
272
(long)(meta)->event_len <= (long)(len))
include/uapi/linux/videodev2.h
2530
struct v4l2_meta_format meta; /* V4L2_BUF_TYPE_META_CAPTURE */
include/uapi/scsi/scsi_bsg_ufs.h
184
struct ufs_arpmb_meta meta;
io_uring/rw.c
220
io->meta_state.seed = io->meta.seed;
io_uring/rw.c
221
iov_iter_save_state(&io->meta.iter, &io->meta_state.iter_meta);
io_uring/rw.c
227
io->meta.seed = io->meta_state.seed;
io_uring/rw.c
228
iov_iter_restore(&io->meta.iter, &io->meta_state.iter_meta);
io_uring/rw.c
247
io->meta.flags = pi_attr.flags;
io_uring/rw.c
248
io->meta.app_tag = pi_attr.app_tag;
io_uring/rw.c
249
io->meta.seed = pi_attr.seed;
io_uring/rw.c
251
pi_attr.len, &io->meta.iter);
io_uring/rw.c
905
kiocb->private = &io->meta;
io_uring/rw.h
28
struct uio_meta meta;
kernel/bpf/arraymap.c
653
struct bpf_iter_meta meta;
kernel/bpf/arraymap.c
659
meta.seq = seq;
kernel/bpf/arraymap.c
660
prog = bpf_iter_get_info(&meta, v == NULL);
kernel/bpf/arraymap.c
664
ctx.meta = &meta;
kernel/bpf/bpf_iter.c
657
struct bpf_prog *bpf_iter_get_info(struct bpf_iter_meta *meta, bool in_stop)
kernel/bpf/bpf_iter.c
663
seq = meta->seq;
kernel/bpf/bpf_iter.c
674
meta->session_id = iter_priv->session_id;
kernel/bpf/bpf_iter.c
675
meta->seq_num = iter_priv->seq_num;
kernel/bpf/btf.c
4214
struct btf_struct_meta *meta;
kernel/bpf/btf.c
4236
meta = btf_find_struct_meta(btf, btf_id);
kernel/bpf/btf.c
4237
if (!meta)
kernel/bpf/btf.c
4239
rec->fields[i].graph_root.value_rec = meta->record;
kernel/bpf/btf.c
4281
if (meta->record->field_mask & BPF_GRAPH_ROOT)
kernel/bpf/btf.c
7318
struct btf_struct_meta *meta;
kernel/bpf/btf.c
7322
meta = btf_find_struct_meta(btf, id);
kernel/bpf/btf.c
7323
if (!meta)
kernel/bpf/btf.c
7325
rec = meta->record;
kernel/bpf/cgroup_iter.c
130
struct bpf_iter_meta meta;
kernel/bpf/cgroup_iter.c
138
ctx.meta = &meta;
kernel/bpf/cgroup_iter.c
140
meta.seq = seq;
kernel/bpf/cgroup_iter.c
141
prog = bpf_iter_get_info(&meta, in_stop);
kernel/bpf/cgroup_iter.c
284
DEFINE_BPF_ITER_FUNC(cgroup, struct bpf_iter_meta *meta,
kernel/bpf/cgroup_iter.c
47
__bpf_md_ptr(struct bpf_iter_meta *, meta);
kernel/bpf/core.c
3063
bpf_event_output(struct bpf_map *map, u64 flags, void *meta, u64 meta_size,
kernel/bpf/dmabuf_iter.c
135
DEFINE_BPF_ITER_FUNC(dmabuf, struct bpf_iter_meta *meta, struct dma_buf *dmabuf)
kernel/bpf/dmabuf_iter.c
50
__bpf_md_ptr(struct bpf_iter_meta *, meta);
kernel/bpf/dmabuf_iter.c
56
struct bpf_iter_meta meta = {
kernel/bpf/dmabuf_iter.c
60
.meta = &meta,
kernel/bpf/dmabuf_iter.c
63
struct bpf_prog *prog = bpf_iter_get_info(&meta, in_stop);
kernel/bpf/hashtab.c
2167
struct bpf_iter_meta meta;
kernel/bpf/hashtab.c
2173
meta.seq = seq;
kernel/bpf/hashtab.c
2174
prog = bpf_iter_get_info(&meta, elem == NULL);
kernel/bpf/hashtab.c
2176
ctx.meta = &meta;
kernel/bpf/helpers.c
2307
struct btf_struct_meta *meta = meta__ign;
kernel/bpf/helpers.c
2314
if (meta)
kernel/bpf/helpers.c
2315
bpf_obj_init(meta->record, p);
kernel/bpf/helpers.c
2352
struct btf_struct_meta *meta = meta__ign;
kernel/bpf/helpers.c
2355
__bpf_obj_drop_impl(p, meta ? meta->record : NULL, false);
kernel/bpf/helpers.c
2366
struct btf_struct_meta *meta = meta__ign;
kernel/bpf/helpers.c
2372
ref = (struct bpf_refcount *)(p__refcounted_kptr + meta->record->refcount_off);
kernel/bpf/helpers.c
2414
struct btf_struct_meta *meta = meta__ign;
kernel/bpf/helpers.c
2416
return __bpf_list_add(n, head, false, meta ? meta->record : NULL, off);
kernel/bpf/helpers.c
2424
struct btf_struct_meta *meta = meta__ign;
kernel/bpf/helpers.c
2426
return __bpf_list_add(n, head, true, meta ? meta->record : NULL, off);
kernel/bpf/helpers.c
2542
struct btf_struct_meta *meta = meta__ign;
kernel/bpf/helpers.c
2545
return __bpf_rbtree_add(root, n, (void *)less, meta ? meta->record : NULL, off);
kernel/bpf/kmem_cache_iter.c
106
__bpf_md_ptr(struct bpf_iter_meta *, meta);
kernel/bpf/kmem_cache_iter.c
154
struct bpf_iter_meta meta;
kernel/bpf/kmem_cache_iter.c
156
.meta = &meta,
kernel/bpf/kmem_cache_iter.c
162
meta.seq = seq;
kernel/bpf/kmem_cache_iter.c
163
prog = bpf_iter_get_info(&meta, true);
kernel/bpf/kmem_cache_iter.c
181
struct bpf_iter_meta meta;
kernel/bpf/kmem_cache_iter.c
183
.meta = &meta,
kernel/bpf/kmem_cache_iter.c
189
meta.seq = seq;
kernel/bpf/kmem_cache_iter.c
190
prog = bpf_iter_get_info(&meta, false);
kernel/bpf/kmem_cache_iter.c
217
DEFINE_BPF_ITER_FUNC(kmem_cache, struct bpf_iter_meta *meta,
kernel/bpf/link_iter.c
38
__bpf_md_ptr(struct bpf_iter_meta *, meta);
kernel/bpf/link_iter.c
42
DEFINE_BPF_ITER_FUNC(bpf_link, struct bpf_iter_meta *meta, struct bpf_link *link)
kernel/bpf/link_iter.c
47
struct bpf_iter_meta meta;
kernel/bpf/link_iter.c
51
ctx.meta = &meta;
kernel/bpf/link_iter.c
53
meta.seq = seq;
kernel/bpf/link_iter.c
54
prog = bpf_iter_get_info(&meta, in_stop);
kernel/bpf/map_iter.c
164
DEFINE_BPF_ITER_FUNC(bpf_map_elem, struct bpf_iter_meta *meta,
kernel/bpf/map_iter.c
38
__bpf_md_ptr(struct bpf_iter_meta *, meta);
kernel/bpf/map_iter.c
42
DEFINE_BPF_ITER_FUNC(bpf_map, struct bpf_iter_meta *meta, struct bpf_map *map)
kernel/bpf/map_iter.c
47
struct bpf_iter_meta meta;
kernel/bpf/map_iter.c
51
ctx.meta = &meta;
kernel/bpf/map_iter.c
53
meta.seq = seq;
kernel/bpf/map_iter.c
54
prog = bpf_iter_get_info(&meta, in_stop);
kernel/bpf/preload/iterators/iterators.bpf.c
101
struct seq_file *seq = ctx->meta->seq;
kernel/bpf/preload/iterators/iterators.bpf.c
102
__u64 seq_num = ctx->meta->seq_num;
kernel/bpf/preload/iterators/iterators.bpf.c
22
struct bpf_iter_meta *meta;
kernel/bpf/preload/iterators/iterators.bpf.c
54
struct bpf_iter_meta *meta;
kernel/bpf/preload/iterators/iterators.bpf.c
81
struct seq_file *seq = ctx->meta->seq;
kernel/bpf/preload/iterators/iterators.bpf.c
82
__u64 seq_num = ctx->meta->seq_num;
kernel/bpf/prog_iter.c
38
__bpf_md_ptr(struct bpf_iter_meta *, meta);
kernel/bpf/prog_iter.c
42
DEFINE_BPF_ITER_FUNC(bpf_prog, struct bpf_iter_meta *meta, struct bpf_prog *prog)
kernel/bpf/prog_iter.c
47
struct bpf_iter_meta meta;
kernel/bpf/prog_iter.c
51
ctx.meta = &meta;
kernel/bpf/prog_iter.c
53
meta.seq = seq;
kernel/bpf/prog_iter.c
54
prog = bpf_iter_get_info(&meta, in_stop);
kernel/bpf/task_iter.c
169
__bpf_md_ptr(struct bpf_iter_meta *, meta);
kernel/bpf/task_iter.c
173
DEFINE_BPF_ITER_FUNC(task, struct bpf_iter_meta *meta, struct task_struct *task)
kernel/bpf/task_iter.c
178
struct bpf_iter_meta meta;
kernel/bpf/task_iter.c
182
meta.seq = seq;
kernel/bpf/task_iter.c
183
prog = bpf_iter_get_info(&meta, in_stop);
kernel/bpf/task_iter.c
187
ctx.meta = &meta;
kernel/bpf/task_iter.c
333
__bpf_md_ptr(struct bpf_iter_meta *, meta);
kernel/bpf/task_iter.c
339
DEFINE_BPF_ITER_FUNC(task_file, struct bpf_iter_meta *meta,
kernel/bpf/task_iter.c
348
struct bpf_iter_meta meta;
kernel/bpf/task_iter.c
351
meta.seq = seq;
kernel/bpf/task_iter.c
352
prog = bpf_iter_get_info(&meta, in_stop);
kernel/bpf/task_iter.c
356
ctx.meta = &meta;
kernel/bpf/task_iter.c
598
__bpf_md_ptr(struct bpf_iter_meta *, meta);
kernel/bpf/task_iter.c
603
DEFINE_BPF_ITER_FUNC(task_vma, struct bpf_iter_meta *meta,
kernel/bpf/task_iter.c
610
struct bpf_iter_meta meta;
kernel/bpf/task_iter.c
613
meta.seq = seq;
kernel/bpf/task_iter.c
614
prog = bpf_iter_get_info(&meta, in_stop);
kernel/bpf/task_iter.c
618
ctx.meta = &meta;
kernel/bpf/verifier.c
10005
key_size = meta->map.ptr->key_size;
kernel/bpf/verifier.c
10009
if (can_elide_value_nullness(meta->map.ptr->map_type)) {
kernel/bpf/verifier.c
10010
err = get_constant_map_key(env, reg, key_size, &meta->const_map_key);
kernel/bpf/verifier.c
10012
meta->const_map_key = -1;
kernel/bpf/verifier.c
10027
if (!meta->map.ptr) {
kernel/bpf/verifier.c
10032
meta->raw_mode = arg_type & MEM_UNINIT;
kernel/bpf/verifier.c
10033
err = check_helper_mem_access(env, regno, meta->map.ptr->value_size,
kernel/bpf/verifier.c
10035
false, meta);
kernel/bpf/verifier.c
10042
meta->ret_btf = reg->btf;
kernel/bpf/verifier.c
10043
meta->ret_btf_id = reg->btf_id;
kernel/bpf/verifier.c
10050
if (meta->func_id == BPF_FUNC_spin_lock) {
kernel/bpf/verifier.c
10054
} else if (meta->func_id == BPF_FUNC_spin_unlock) {
kernel/bpf/verifier.c
10064
err = process_timer_helper(env, regno, meta);
kernel/bpf/verifier.c
10069
meta->subprogno = reg->subprogno;
kernel/bpf/verifier.c
10075
meta->raw_mode = arg_type & MEM_UNINIT;
kernel/bpf/verifier.c
10079
false, meta);
kernel/bpf/verifier.c
10090
false, meta);
kernel/bpf/verifier.c
10096
true, meta);
kernel/bpf/verifier.c
10109
meta->mem_size = reg->var_off.value;
kernel/bpf/verifier.c
10122
err = process_kptr_func(env, regno, meta);
kernel/bpf/verifier.c
1047
static bool is_kfunc_rcu_protected(struct bpf_kfunc_call_arg_meta *meta);
kernel/bpf/verifier.c
1050
struct bpf_kfunc_call_arg_meta *meta,
kernel/bpf/verifier.c
1071
if (is_kfunc_rcu_protected(meta)) {
kernel/bpf/verifier.c
10762
struct bpf_call_arg_meta meta;
kernel/bpf/verifier.c
10768
memset(&meta, 0, sizeof(meta)); /* leave func_id as zero */
kernel/bpf/verifier.c
10769
err = check_reg_type(env, regno, arg->arg_type, &arg->btf_id, &meta);
kernel/bpf/verifier.c
11331
struct bpf_call_arg_meta *meta)
kernel/bpf/verifier.c
11344
ret_reg->smax_value = meta->msize_max_value;
kernel/bpf/verifier.c
11345
ret_reg->s32_max_value = meta->msize_max_value;
kernel/bpf/verifier.c
11367
record_func_map(struct bpf_verifier_env *env, struct bpf_call_arg_meta *meta,
kernel/bpf/verifier.c
11371
struct bpf_map *map = meta->map.ptr;
kernel/bpf/verifier.c
11404
bpf_map_ptr_store(aux, meta->map.ptr,
kernel/bpf/verifier.c
11405
!meta->map.ptr->bypass_spec_v1, false);
kernel/bpf/verifier.c
11406
else if (aux->map_ptr_state.map_ptr != meta->map.ptr)
kernel/bpf/verifier.c
11407
bpf_map_ptr_store(aux, meta->map.ptr,
kernel/bpf/verifier.c
11408
!meta->map.ptr->bypass_spec_v1, true);
kernel/bpf/verifier.c
11413
record_func_key(struct bpf_verifier_env *env, struct bpf_call_arg_meta *meta,
kernel/bpf/verifier.c
11418
struct bpf_map *map = meta->map.ptr;
kernel/bpf/verifier.c
11649
struct bpf_call_arg_meta meta;
kernel/bpf/verifier.c
11691
memset(&meta, 0, sizeof(meta));
kernel/bpf/verifier.c
11692
meta.pkt_access = fn->pkt_access;
kernel/bpf/verifier.c
11728
meta.func_id = func_id;
kernel/bpf/verifier.c
11731
err = check_func_arg(env, i, &meta, fn, insn_idx);
kernel/bpf/verifier.c
11736
err = record_func_map(env, &meta, func_id, insn_idx);
kernel/bpf/verifier.c
11740
err = record_func_key(env, &meta, func_id, insn_idx);
kernel/bpf/verifier.c
11747
for (i = 0; i < meta.access_size; i++) {
kernel/bpf/verifier.c
11748
err = check_mem_access(env, insn_idx, meta.regno, i, BPF_B,
kernel/bpf/verifier.c
11756
if (meta.release_regno) {
kernel/bpf/verifier.c
11758
if (arg_type_is_dynptr(fn->arg_type[meta.release_regno - BPF_REG_1])) {
kernel/bpf/verifier.c
11759
err = unmark_stack_slots_dynptr(env, &regs[meta.release_regno]);
kernel/bpf/verifier.c
11760
} else if (func_id == BPF_FUNC_kptr_xchg && meta.ref_obj_id) {
kernel/bpf/verifier.c
11761
u32 ref_obj_id = meta.ref_obj_id;
kernel/bpf/verifier.c
11780
} else if (meta.ref_obj_id) {
kernel/bpf/verifier.c
11781
err = release_reference(env, meta.ref_obj_id);
kernel/bpf/verifier.c
11782
} else if (register_is_null(&regs[meta.release_regno])) {
kernel/bpf/verifier.c
11811
err = push_callback_call(env, insn, insn_idx, meta.subprogno,
kernel/bpf/verifier.c
11815
err = push_callback_call(env, insn, insn_idx, meta.subprogno,
kernel/bpf/verifier.c
11819
err = push_callback_call(env, insn, insn_idx, meta.subprogno,
kernel/bpf/verifier.c
11826
update_loop_inline_state(env, meta.subprogno);
kernel/bpf/verifier.c
11834
err = push_callback_call(env, insn, insn_idx, meta.subprogno,
kernel/bpf/verifier.c
1185
struct bpf_kfunc_call_arg_meta *meta,
kernel/bpf/verifier.c
11872
if (meta.dynptr_id) {
kernel/bpf/verifier.c
11876
if (meta.ref_obj_id) {
kernel/bpf/verifier.c
11893
meta.dynptr_id = id;
kernel/bpf/verifier.c
11894
meta.ref_obj_id = ref_obj_id;
kernel/bpf/verifier.c
11938
err = push_callback_call(env, insn, insn_idx, meta.subprogno,
kernel/bpf/verifier.c
11974
if (meta.map.ptr == NULL) {
kernel/bpf/verifier.c
11980
can_elide_value_nullness(meta.map.ptr->map_type) &&
kernel/bpf/verifier.c
11981
meta.const_map_key >= 0 &&
kernel/bpf/verifier.c
11982
meta.const_map_key < meta.map.ptr->max_entries)
kernel/bpf/verifier.c
11985
regs[BPF_REG_0].map_ptr = meta.map.ptr;
kernel/bpf/verifier.c
11986
regs[BPF_REG_0].map_uid = meta.map.uid;
kernel/bpf/verifier.c
11989
btf_record_has_field(meta.map.ptr->record, BPF_SPIN_LOCK | BPF_RES_SPIN_LOCK)) {
kernel/bpf/verifier.c
12008
regs[BPF_REG_0].mem_size = meta.mem_size;
kernel/bpf/verifier.c
12015
t = btf_type_skip_modifiers(meta.ret_btf, meta.ret_btf_id, NULL);
kernel/bpf/verifier.c
12022
ret = btf_resolve_size(meta.ret_btf, t, &tsize);
kernel/bpf/verifier.c
12024
tname = btf_name_by_offset(meta.ret_btf, t->name_off);
kernel/bpf/verifier.c
12044
regs[BPF_REG_0].btf = meta.ret_btf;
kernel/bpf/verifier.c
12045
regs[BPF_REG_0].btf_id = meta.ret_btf_id;
kernel/bpf/verifier.c
12057
ret_btf = meta.kptr_field->kptr.btf;
kernel/bpf/verifier.c
12058
ret_btf_id = meta.kptr_field->kptr.btf_id;
kernel/bpf/verifier.c
12061
if (meta.kptr_field->type == BPF_KPTR_PERCPU)
kernel/bpf/verifier.c
12092
if (helper_multiple_ref_obj_use(func_id, meta.map.ptr)) {
kernel/bpf/verifier.c
12099
regs[BPF_REG_0].dynptr_id = meta.dynptr_id;
kernel/bpf/verifier.c
12103
regs[BPF_REG_0].ref_obj_id = meta.ref_obj_id;
kernel/bpf/verifier.c
12104
} else if (is_acquire_function(func_id, meta.map.ptr)) {
kernel/bpf/verifier.c
12115
err = do_refine_retval_range(env, regs, fn->ret_type, func_id, &meta);
kernel/bpf/verifier.c
12119
err = check_map_func_compatibility(env, meta.map.ptr, func_id);
kernel/bpf/verifier.c
12200
static bool is_kfunc_acquire(struct bpf_kfunc_call_arg_meta *meta)
kernel/bpf/verifier.c
12202
return meta->kfunc_flags & KF_ACQUIRE;
kernel/bpf/verifier.c
12205
static bool is_kfunc_release(struct bpf_kfunc_call_arg_meta *meta)
kernel/bpf/verifier.c
12207
return meta->kfunc_flags & KF_RELEASE;
kernel/bpf/verifier.c
12210
static bool is_kfunc_sleepable(struct bpf_kfunc_call_arg_meta *meta)
kernel/bpf/verifier.c
12212
return meta->kfunc_flags & KF_SLEEPABLE;
kernel/bpf/verifier.c
12215
static bool is_kfunc_destructive(struct bpf_kfunc_call_arg_meta *meta)
kernel/bpf/verifier.c
12217
return meta->kfunc_flags & KF_DESTRUCTIVE;
kernel/bpf/verifier.c
12220
static bool is_kfunc_rcu(struct bpf_kfunc_call_arg_meta *meta)
kernel/bpf/verifier.c
12222
return meta->kfunc_flags & KF_RCU;
kernel/bpf/verifier.c
12225
static bool is_kfunc_rcu_protected(struct bpf_kfunc_call_arg_meta *meta)
kernel/bpf/verifier.c
12227
return meta->kfunc_flags & KF_RCU_PROTECTED;
kernel/bpf/verifier.c
12643
static bool is_kfunc_ret_null(struct bpf_kfunc_call_arg_meta *meta)
kernel/bpf/verifier.c
12645
if (meta->func_id == special_kfunc_list[KF_bpf_refcount_acquire_impl] &&
kernel/bpf/verifier.c
12646
meta->arg_owning_ref) {
kernel/bpf/verifier.c
12650
return meta->kfunc_flags & KF_RET_NULL;
kernel/bpf/verifier.c
12653
static bool is_kfunc_bpf_rcu_read_lock(struct bpf_kfunc_call_arg_meta *meta)
kernel/bpf/verifier.c
12655
return meta->func_id == special_kfunc_list[KF_bpf_rcu_read_lock];
kernel/bpf/verifier.c
12658
static bool is_kfunc_bpf_rcu_read_unlock(struct bpf_kfunc_call_arg_meta *meta)
kernel/bpf/verifier.c
12660
return meta->func_id == special_kfunc_list[KF_bpf_rcu_read_unlock];
kernel/bpf/verifier.c
12663
static bool is_kfunc_bpf_preempt_disable(struct bpf_kfunc_call_arg_meta *meta)
kernel/bpf/verifier.c
12665
return meta->func_id == special_kfunc_list[KF_bpf_preempt_disable];
kernel/bpf/verifier.c
12668
static bool is_kfunc_bpf_preempt_enable(struct bpf_kfunc_call_arg_meta *meta)
kernel/bpf/verifier.c
12670
return meta->func_id == special_kfunc_list[KF_bpf_preempt_enable];
kernel/bpf/verifier.c
12673
static bool is_kfunc_pkt_changing(struct bpf_kfunc_call_arg_meta *meta)
kernel/bpf/verifier.c
12675
return meta->func_id == special_kfunc_list[KF_bpf_xdp_pull_data];
kernel/bpf/verifier.c
12680
struct bpf_kfunc_call_arg_meta *meta,
kernel/bpf/verifier.c
12690
if (meta->func_id == special_kfunc_list[KF_bpf_cast_to_kern_ctx] ||
kernel/bpf/verifier.c
12691
meta->func_id == special_kfunc_list[KF_bpf_session_is_return] ||
kernel/bpf/verifier.c
12692
meta->func_id == special_kfunc_list[KF_bpf_session_cookie])
kernel/bpf/verifier.c
12696
(is_kfunc_arg_mem_size(meta->btf, &args[argno + 1], &regs[regno + 1]) ||
kernel/bpf/verifier.c
12697
is_kfunc_arg_const_mem_size(meta->btf, &args[argno + 1], &regs[regno + 1])))
kernel/bpf/verifier.c
12705
if (btf_is_prog_ctx_type(&env->log, meta->btf, t, resolve_prog_type(env->prog), argno))
kernel/bpf/verifier.c
12708
if (is_kfunc_arg_nullable(meta->btf, &args[argno]) && register_is_null(reg) &&
kernel/bpf/verifier.c
12712
if (is_kfunc_arg_alloc_obj(meta->btf, &args[argno]))
kernel/bpf/verifier.c
12715
if (is_kfunc_arg_refcounted_kptr(meta->btf, &args[argno]))
kernel/bpf/verifier.c
12718
if (is_kfunc_arg_dynptr(meta->btf, &args[argno]))
kernel/bpf/verifier.c
12721
if (is_kfunc_arg_iter(meta, argno, &args[argno]))
kernel/bpf/verifier.c
12724
if (is_kfunc_arg_list_head(meta->btf, &args[argno]))
kernel/bpf/verifier.c
12727
if (is_kfunc_arg_list_node(meta->btf, &args[argno]))
kernel/bpf/verifier.c
12730
if (is_kfunc_arg_rbtree_root(meta->btf, &args[argno]))
kernel/bpf/verifier.c
12733
if (is_kfunc_arg_rbtree_node(meta->btf, &args[argno]))
kernel/bpf/verifier.c
12736
if (is_kfunc_arg_const_str(meta->btf, &args[argno]))
kernel/bpf/verifier.c
12739
if (is_kfunc_arg_map(meta->btf, &args[argno]))
kernel/bpf/verifier.c
12742
if (is_kfunc_arg_wq(meta->btf, &args[argno]))
kernel/bpf/verifier.c
12745
if (is_kfunc_arg_timer(meta->btf, &args[argno]))
kernel/bpf/verifier.c
12748
if (is_kfunc_arg_task_work(meta->btf, &args[argno]))
kernel/bpf/verifier.c
12751
if (is_kfunc_arg_irq_flag(meta->btf, &args[argno]))
kernel/bpf/verifier.c
12754
if (is_kfunc_arg_res_spin_lock(meta->btf, &args[argno]))
kernel/bpf/verifier.c
12760
meta->func_name, argno, btf_type_str(ref_t), ref_tname);
kernel/bpf/verifier.c
12766
if (is_kfunc_arg_callback(env, meta->btf, &args[argno]))
kernel/bpf/verifier.c
12774
if (!btf_type_is_scalar(ref_t) && !__btf_type_is_scalar_struct(env, meta->btf, ref_t, 0) &&
kernel/bpf/verifier.c
12787
struct bpf_kfunc_call_arg_meta *meta,
kernel/bpf/verifier.c
12830
if ((is_kfunc_release(meta) && reg->ref_obj_id) ||
kernel/bpf/verifier.c
12831
btf_type_ids_nocast_alias(&env->log, reg_btf, reg_ref_id, meta->btf, ref_id))
kernel/bpf/verifier.c
12834
WARN_ON_ONCE(is_kfunc_release(meta) &&
kernel/bpf/verifier.c
12840
struct_same = btf_struct_ids_match(&env->log, reg_btf, reg_ref_id, reg->off, meta->btf, ref_id, strict_type_match);
kernel/bpf/verifier.c
12848
meta->func_name, argno, btf_type_str(ref_t), ref_tname, argno + 1,
kernel/bpf/verifier.c
12856
struct bpf_kfunc_call_arg_meta *meta)
kernel/bpf/verifier.c
12862
if (meta->func_id == special_kfunc_list[KF_bpf_local_irq_save] ||
kernel/bpf/verifier.c
12863
meta->func_id == special_kfunc_list[KF_bpf_res_spin_lock_irqsave]) {
kernel/bpf/verifier.c
12865
if (meta->func_id == special_kfunc_list[KF_bpf_res_spin_lock_irqsave])
kernel/bpf/verifier.c
12867
} else if (meta->func_id == special_kfunc_list[KF_bpf_local_irq_restore] ||
kernel/bpf/verifier.c
12868
meta->func_id == special_kfunc_list[KF_bpf_res_spin_unlock_irqrestore]) {
kernel/bpf/verifier.c
12870
if (meta->func_id == special_kfunc_list[KF_bpf_res_spin_unlock_irqrestore])
kernel/bpf/verifier.c
12887
err = mark_stack_slot_irq_flag(env, meta, reg, env->insn_idx, kfunc_class);
kernel/bpf/verifier.c
13186
struct bpf_kfunc_call_arg_meta *meta,
kernel/bpf/verifier.c
13195
if (meta->btf != btf_vmlinux) {
kernel/bpf/verifier.c
13200
if (!check_kfunc_is_graph_root_api(env, head_field_type, meta->func_id))
kernel/bpf/verifier.c
13236
struct bpf_kfunc_call_arg_meta *meta)
kernel/bpf/verifier.c
13238
return __process_kf_arg_ptr_to_graph_root(env, reg, regno, meta, BPF_LIST_HEAD,
kernel/bpf/verifier.c
13239
&meta->arg_list_head.field);
kernel/bpf/verifier.c
13244
struct bpf_kfunc_call_arg_meta *meta)
kernel/bpf/verifier.c
13246
return __process_kf_arg_ptr_to_graph_root(env, reg, regno, meta, BPF_RB_ROOT,
kernel/bpf/verifier.c
13247
&meta->arg_rbtree_root.field);
kernel/bpf/verifier.c
13253
struct bpf_kfunc_call_arg_meta *meta,
kernel/bpf/verifier.c
13263
if (meta->btf != btf_vmlinux) {
kernel/bpf/verifier.c
13268
if (!check_kfunc_is_graph_node_api(env, node_field_type, meta->func_id))
kernel/bpf/verifier.c
13301
meta->arg_btf = reg->btf;
kernel/bpf/verifier.c
13302
meta->arg_btf_id = reg->btf_id;
kernel/bpf/verifier.c
13317
struct bpf_kfunc_call_arg_meta *meta)
kernel/bpf/verifier.c
13319
return __process_kf_arg_ptr_to_graph_node(env, reg, regno, meta,
kernel/bpf/verifier.c
13321
&meta->arg_list_head.field);
kernel/bpf/verifier.c
13326
struct bpf_kfunc_call_arg_meta *meta)
kernel/bpf/verifier.c
13328
return __process_kf_arg_ptr_to_graph_node(env, reg, regno, meta,
kernel/bpf/verifier.c
13330
&meta->arg_rbtree_root.field);
kernel/bpf/verifier.c
13355
static int check_kfunc_args(struct bpf_verifier_env *env, struct bpf_kfunc_call_arg_meta *meta,
kernel/bpf/verifier.c
13358
const char *func_name = meta->func_name, *ref_tname;
kernel/bpf/verifier.c
13359
const struct btf *btf = meta->btf;
kernel/bpf/verifier.c
13365
args = (const struct btf_param *)(meta->func_proto + 1);
kernel/bpf/verifier.c
13366
nargs = btf_type_vlen(meta->func_proto);
kernel/bpf/verifier.c
13391
if (meta->arg_prog) {
kernel/bpf/verifier.c
13395
meta->arg_prog = true;
kernel/bpf/verifier.c
13406
if (is_kfunc_arg_constant(meta->btf, &args[i])) {
kernel/bpf/verifier.c
13407
if (meta->arg_constant.found) {
kernel/bpf/verifier.c
13418
meta->arg_constant.found = true;
kernel/bpf/verifier.c
13419
meta->arg_constant.value = reg->var_off.value;
kernel/bpf/verifier.c
13421
meta->r0_rdonly = true;
kernel/bpf/verifier.c
13428
if (meta->r0_size) {
kernel/bpf/verifier.c
13438
meta->r0_size = reg->var_off.value;
kernel/bpf/verifier.c
13452
!is_kfunc_arg_nullable(meta->btf, &args[i])) {
kernel/bpf/verifier.c
13458
if (is_kfunc_release(meta) && meta->ref_obj_id) {
kernel/bpf/verifier.c
13461
meta->ref_obj_id);
kernel/bpf/verifier.c
13464
meta->ref_obj_id = reg->ref_obj_id;
kernel/bpf/verifier.c
13465
if (is_kfunc_release(meta))
kernel/bpf/verifier.c
13466
meta->release_regno = regno;
kernel/bpf/verifier.c
13472
kf_arg_type = get_kfunc_ptr_arg_type(env, meta, t, ref_t, ref_tname, args, i, nargs);
kernel/bpf/verifier.c
13484
if (meta->map.ptr && (reg->map_ptr->record->wq_off >= 0 ||
kernel/bpf/verifier.c
13498
if (meta->map.ptr != reg->map_ptr ||
kernel/bpf/verifier.c
13499
meta->map.uid != reg->map_uid) {
kernel/bpf/verifier.c
13503
meta->map.uid, reg->map_uid);
kernel/bpf/verifier.c
13508
meta->map.uid, reg->map_uid);
kernel/bpf/verifier.c
13512
meta->map.ptr = reg->map_ptr;
kernel/bpf/verifier.c
13513
meta->map.uid = reg->map_uid;
kernel/bpf/verifier.c
13518
if (!is_kfunc_rcu(meta)) {
kernel/bpf/verifier.c
13551
if (is_kfunc_release(meta) && reg->ref_obj_id)
kernel/bpf/verifier.c
13565
if (meta->func_id == special_kfunc_list[KF_bpf_cast_to_kern_ctx]) {
kernel/bpf/verifier.c
13569
meta->ret_btf_id = ret;
kernel/bpf/verifier.c
13574
if (meta->func_id != special_kfunc_list[KF_bpf_obj_drop_impl]) {
kernel/bpf/verifier.c
13579
if (meta->func_id != special_kfunc_list[KF_bpf_percpu_obj_drop_impl]) {
kernel/bpf/verifier.c
13591
if (meta->btf == btf_vmlinux) {
kernel/bpf/verifier.c
13592
meta->arg_btf = reg->btf;
kernel/bpf/verifier.c
13593
meta->arg_btf_id = reg->btf_id;
kernel/bpf/verifier.c
13607
if (meta->func_id == special_kfunc_list[KF_bpf_dynptr_from_skb]) {
kernel/bpf/verifier.c
13609
} else if (meta->func_id == special_kfunc_list[KF_bpf_dynptr_from_xdp]) {
kernel/bpf/verifier.c
13611
} else if (meta->func_id == special_kfunc_list[KF_bpf_dynptr_from_skb_meta]) {
kernel/bpf/verifier.c
13613
} else if (meta->func_id == special_kfunc_list[KF_bpf_dynptr_from_file]) {
kernel/bpf/verifier.c
13615
} else if (meta->func_id == special_kfunc_list[KF_bpf_dynptr_file_discard]) {
kernel/bpf/verifier.c
13617
meta->release_regno = regno;
kernel/bpf/verifier.c
13618
} else if (meta->func_id == special_kfunc_list[KF_bpf_dynptr_clone] &&
kernel/bpf/verifier.c
13620
enum bpf_dynptr_type parent_type = meta->initialized_dynptr.type;
kernel/bpf/verifier.c
13628
clone_ref_obj_id = meta->initialized_dynptr.ref_obj_id;
kernel/bpf/verifier.c
13646
meta->initialized_dynptr.id = id;
kernel/bpf/verifier.c
13647
meta->initialized_dynptr.type = dynptr_get_type(env, reg);
kernel/bpf/verifier.c
13648
meta->initialized_dynptr.ref_obj_id = dynptr_ref_obj_id(env, reg);
kernel/bpf/verifier.c
13654
if (meta->func_id == special_kfunc_list[KF_bpf_iter_css_task_new]) {
kernel/bpf/verifier.c
13660
ret = process_iter_arg(env, regno, insn_idx, meta);
kernel/bpf/verifier.c
13674
ret = process_kf_arg_ptr_to_list_head(env, reg, regno, meta);
kernel/bpf/verifier.c
13688
ret = process_kf_arg_ptr_to_rbtree_root(env, reg, regno, meta);
kernel/bpf/verifier.c
13701
ret = process_kf_arg_ptr_to_list_node(env, reg, regno, meta);
kernel/bpf/verifier.c
13706
if (meta->func_id == special_kfunc_list[KF_bpf_rbtree_add_impl]) {
kernel/bpf/verifier.c
13726
ret = process_kf_arg_ptr_to_rbtree_node(env, reg, regno, meta);
kernel/bpf/verifier.c
13747
ret = process_kf_arg_ptr_to_btf_id(env, reg, ref_t, ref_tname, ref_id, meta, i);
kernel/bpf/verifier.c
13769
if (!register_is_null(buff_reg) || !is_kfunc_arg_nullable(meta->btf, buff_arg)) {
kernel/bpf/verifier.c
13777
if (is_kfunc_arg_const_mem_size(meta->btf, size_arg, size_reg)) {
kernel/bpf/verifier.c
13778
if (meta->arg_constant.found) {
kernel/bpf/verifier.c
13786
meta->arg_constant.found = true;
kernel/bpf/verifier.c
13787
meta->arg_constant.value = size_reg->var_off.value;
kernel/bpf/verifier.c
13799
meta->subprogno = reg->subprogno;
kernel/bpf/verifier.c
13807
meta->arg_owning_ref = true;
kernel/bpf/verifier.c
13820
meta->arg_btf = reg->btf;
kernel/bpf/verifier.c
13821
meta->arg_btf_id = reg->btf_id;
kernel/bpf/verifier.c
13837
ret = check_map_field_pointer(env, regno, BPF_WORKQUEUE, &meta->map);
kernel/bpf/verifier.c
13846
ret = process_timer_kfunc(env, regno, meta);
kernel/bpf/verifier.c
13855
ret = check_map_field_pointer(env, regno, BPF_TASK_WORK, &meta->map);
kernel/bpf/verifier.c
13864
ret = process_irq_flag(env, regno, meta);
kernel/bpf/verifier.c
13877
if (!is_bpf_res_spin_lock_kfunc(meta->func_id))
kernel/bpf/verifier.c
13879
if (meta->func_id == special_kfunc_list[KF_bpf_res_spin_lock] ||
kernel/bpf/verifier.c
13880
meta->func_id == special_kfunc_list[KF_bpf_res_spin_lock_irqsave])
kernel/bpf/verifier.c
13882
if (meta->func_id == special_kfunc_list[KF_bpf_res_spin_lock_irqsave] ||
kernel/bpf/verifier.c
13883
meta->func_id == special_kfunc_list[KF_bpf_res_spin_unlock_irqrestore])
kernel/bpf/verifier.c
13893
if (is_kfunc_release(meta) && !meta->release_regno) {
kernel/bpf/verifier.c
13905
struct bpf_kfunc_call_arg_meta *meta)
kernel/bpf/verifier.c
13914
memset(meta, 0, sizeof(*meta));
kernel/bpf/verifier.c
13915
meta->btf = kfunc.btf;
kernel/bpf/verifier.c
13916
meta->func_id = kfunc.id;
kernel/bpf/verifier.c
13917
meta->func_proto = kfunc.proto;
kernel/bpf/verifier.c
13918
meta->func_name = kfunc.name;
kernel/bpf/verifier.c
13923
meta->kfunc_flags = *kfunc.flags;
kernel/bpf/verifier.c
13933
static int check_special_kfunc(struct bpf_verifier_env *env, struct bpf_kfunc_call_arg_meta *meta,
kernel/bpf/verifier.c
13940
if (meta->btf != btf_vmlinux)
kernel/bpf/verifier.c
13943
if (meta->func_id == special_kfunc_list[KF_bpf_obj_new_impl] ||
kernel/bpf/verifier.c
13944
meta->func_id == special_kfunc_list[KF_bpf_percpu_obj_new_impl]) {
kernel/bpf/verifier.c
13949
if (meta->func_id == special_kfunc_list[KF_bpf_obj_new_impl] && !bpf_global_ma_set)
kernel/bpf/verifier.c
13952
if (((u64)(u32)meta->arg_constant.value) != meta->arg_constant.value) {
kernel/bpf/verifier.c
13958
ret_btf_id = meta->arg_constant.value;
kernel/bpf/verifier.c
13972
if (meta->func_id == special_kfunc_list[KF_bpf_percpu_obj_new_impl]) {
kernel/bpf/verifier.c
14002
if (meta->func_id == special_kfunc_list[KF_bpf_percpu_obj_new_impl]) {
kernel/bpf/verifier.c
14018
if (meta->func_id == special_kfunc_list[KF_bpf_percpu_obj_new_impl])
kernel/bpf/verifier.c
14023
} else if (meta->func_id == special_kfunc_list[KF_bpf_refcount_acquire_impl]) {
kernel/bpf/verifier.c
14026
regs[BPF_REG_0].btf = meta->arg_btf;
kernel/bpf/verifier.c
14027
regs[BPF_REG_0].btf_id = meta->arg_btf_id;
kernel/bpf/verifier.c
14030
btf_find_struct_meta(meta->arg_btf,
kernel/bpf/verifier.c
14031
meta->arg_btf_id);
kernel/bpf/verifier.c
14033
struct btf_field *field = meta->arg_list_head.field;
kernel/bpf/verifier.c
14037
struct btf_field *field = meta->arg_rbtree_root.field;
kernel/bpf/verifier.c
14040
} else if (meta->func_id == special_kfunc_list[KF_bpf_cast_to_kern_ctx]) {
kernel/bpf/verifier.c
14044
regs[BPF_REG_0].btf_id = meta->ret_btf_id;
kernel/bpf/verifier.c
14045
} else if (meta->func_id == special_kfunc_list[KF_bpf_rdonly_cast]) {
kernel/bpf/verifier.c
14046
ret_t = btf_type_by_id(desc_btf, meta->arg_constant.value);
kernel/bpf/verifier.c
14049
meta->arg_constant.value);
kernel/bpf/verifier.c
14055
regs[BPF_REG_0].btf_id = meta->arg_constant.value;
kernel/bpf/verifier.c
14065
} else if (meta->func_id == special_kfunc_list[KF_bpf_dynptr_slice] ||
kernel/bpf/verifier.c
14066
meta->func_id == special_kfunc_list[KF_bpf_dynptr_slice_rdwr]) {
kernel/bpf/verifier.c
14067
enum bpf_type_flag type_flag = get_dynptr_type_flag(meta->initialized_dynptr.type);
kernel/bpf/verifier.c
14071
if (!meta->arg_constant.found) {
kernel/bpf/verifier.c
14076
regs[BPF_REG_0].mem_size = meta->arg_constant.value;
kernel/bpf/verifier.c
14081
if (meta->func_id == special_kfunc_list[KF_bpf_dynptr_slice]) {
kernel/bpf/verifier.c
14091
if (!meta->initialized_dynptr.id) {
kernel/bpf/verifier.c
14095
regs[BPF_REG_0].dynptr_id = meta->initialized_dynptr.id;
kernel/bpf/verifier.c
14118
struct bpf_kfunc_call_arg_meta meta;
kernel/bpf/verifier.c
14128
err = fetch_kfunc_arg_meta(env, insn->imm, insn->off, &meta);
kernel/bpf/verifier.c
14129
if (err == -EACCES && meta.func_name)
kernel/bpf/verifier.c
14130
verbose(env, "calling kernel function %s is not allowed\n", meta.func_name);
kernel/bpf/verifier.c
14133
desc_btf = meta.btf;
kernel/bpf/verifier.c
14134
func_name = meta.func_name;
kernel/bpf/verifier.c
14137
insn_aux->is_iter_next = is_iter_next_kfunc(&meta);
kernel/bpf/verifier.c
14169
if (is_kfunc_destructive(&meta) && !capable(CAP_SYS_BOOT)) {
kernel/bpf/verifier.c
14174
sleepable = is_kfunc_sleepable(&meta);
kernel/bpf/verifier.c
14185
err = check_kfunc_args(env, &meta, insn_idx);
kernel/bpf/verifier.c
14189
if (meta.func_id == special_kfunc_list[KF_bpf_rbtree_add_impl]) {
kernel/bpf/verifier.c
14190
err = push_callback_call(env, insn, insn_idx, meta.subprogno,
kernel/bpf/verifier.c
14194
func_name, meta.func_id);
kernel/bpf/verifier.c
14199
if (meta.func_id == special_kfunc_list[KF_bpf_session_cookie]) {
kernel/bpf/verifier.c
14200
meta.r0_size = sizeof(u64);
kernel/bpf/verifier.c
14201
meta.r0_rdonly = false;
kernel/bpf/verifier.c
14204
if (is_bpf_wq_set_callback_kfunc(meta.func_id)) {
kernel/bpf/verifier.c
14205
err = push_callback_call(env, insn, insn_idx, meta.subprogno,
kernel/bpf/verifier.c
14209
func_name, meta.func_id);
kernel/bpf/verifier.c
14214
if (is_task_work_add_kfunc(meta.func_id)) {
kernel/bpf/verifier.c
14215
err = push_callback_call(env, insn, insn_idx, meta.subprogno,
kernel/bpf/verifier.c
14219
func_name, meta.func_id);
kernel/bpf/verifier.c
14224
rcu_lock = is_kfunc_bpf_rcu_read_lock(&meta);
kernel/bpf/verifier.c
14225
rcu_unlock = is_kfunc_bpf_rcu_read_unlock(&meta);
kernel/bpf/verifier.c
14227
preempt_disable = is_kfunc_bpf_preempt_disable(&meta);
kernel/bpf/verifier.c
14228
preempt_enable = is_kfunc_bpf_preempt_enable(&meta);
kernel/bpf/verifier.c
14280
if (is_kfunc_rcu_protected(&meta) && !in_rcu_cs(env)) {
kernel/bpf/verifier.c
14288
if (meta.release_regno) {
kernel/bpf/verifier.c
14289
struct bpf_reg_state *reg = &regs[meta.release_regno];
kernel/bpf/verifier.c
14291
if (meta.initialized_dynptr.ref_obj_id) {
kernel/bpf/verifier.c
14297
func_name, meta.func_id);
kernel/bpf/verifier.c
14303
if (meta.func_id == special_kfunc_list[KF_bpf_list_push_front_impl] ||
kernel/bpf/verifier.c
14304
meta.func_id == special_kfunc_list[KF_bpf_list_push_back_impl] ||
kernel/bpf/verifier.c
14305
meta.func_id == special_kfunc_list[KF_bpf_rbtree_add_impl]) {
kernel/bpf/verifier.c
14308
insn_aux->kptr_struct_meta = btf_find_struct_meta(meta.arg_btf, meta.arg_btf_id);
kernel/bpf/verifier.c
14312
func_name, meta.func_id);
kernel/bpf/verifier.c
14319
func_name, meta.func_id);
kernel/bpf/verifier.c
14324
if (meta.func_id == special_kfunc_list[KF_bpf_throw]) {
kernel/bpf/verifier.c
14327
func_name, meta.func_id);
kernel/bpf/verifier.c
14350
t = btf_type_skip_modifiers(desc_btf, meta.func_proto->type, NULL);
kernel/bpf/verifier.c
14352
if (is_kfunc_acquire(&meta) && !btf_type_is_struct_ptr(meta.btf, t)) {
kernel/bpf/verifier.c
14354
if (meta.btf != btf_vmlinux ||
kernel/bpf/verifier.c
14355
(meta.func_id != special_kfunc_list[KF_bpf_obj_new_impl] &&
kernel/bpf/verifier.c
14356
meta.func_id != special_kfunc_list[KF_bpf_percpu_obj_new_impl] &&
kernel/bpf/verifier.c
14357
meta.func_id != special_kfunc_list[KF_bpf_refcount_acquire_impl])) {
kernel/bpf/verifier.c
14365
if (meta.btf == btf_vmlinux && (meta.func_id == special_kfunc_list[KF_bpf_res_spin_lock] ||
kernel/bpf/verifier.c
14366
meta.func_id == special_kfunc_list[KF_bpf_res_spin_lock_irqsave]))
kernel/bpf/verifier.c
14371
err = check_special_kfunc(env, &meta, regs, insn_aux, ptr_type, desc_btf);
kernel/bpf/verifier.c
14379
if (!meta.r0_size) {
kernel/bpf/verifier.c
14383
meta.r0_size = sz;
kernel/bpf/verifier.c
14384
meta.r0_rdonly = true;
kernel/bpf/verifier.c
14387
if (!meta.r0_size) {
kernel/bpf/verifier.c
14400
regs[BPF_REG_0].mem_size = meta.r0_size;
kernel/bpf/verifier.c
14402
if (meta.r0_rdonly)
kernel/bpf/verifier.c
14406
if (meta.ref_obj_id)
kernel/bpf/verifier.c
14407
regs[BPF_REG_0].ref_obj_id = meta.ref_obj_id;
kernel/bpf/verifier.c
14409
if (is_kfunc_rcu_protected(&meta))
kernel/bpf/verifier.c
14414
if (meta.func_id == special_kfunc_list[KF_bpf_get_kmem_cache])
kernel/bpf/verifier.c
14416
else if (is_kfunc_rcu_protected(&meta) ||
kernel/bpf/verifier.c
14417
(is_iter_next_kfunc(&meta) &&
kernel/bpf/verifier.c
14418
(get_iter_from_state(env->cur_state, &meta)
kernel/bpf/verifier.c
14446
if (is_kfunc_ret_null(&meta)) {
kernel/bpf/verifier.c
14452
if (is_kfunc_acquire(&meta)) {
kernel/bpf/verifier.c
14457
if (is_kfunc_ret_null(&meta))
kernel/bpf/verifier.c
14467
if (meta.btf == btf_vmlinux) {
kernel/bpf/verifier.c
14468
if (meta.func_id == special_kfunc_list[KF_bpf_obj_drop_impl] ||
kernel/bpf/verifier.c
14469
meta.func_id == special_kfunc_list[KF_bpf_percpu_obj_drop_impl]) {
kernel/bpf/verifier.c
14471
btf_find_struct_meta(meta.arg_btf,
kernel/bpf/verifier.c
14472
meta.arg_btf_id);
kernel/bpf/verifier.c
14477
if (is_kfunc_pkt_changing(&meta))
kernel/bpf/verifier.c
14480
nargs = btf_type_vlen(meta.func_proto);
kernel/bpf/verifier.c
14481
args = (const struct btf_param *)(meta.func_proto + 1);
kernel/bpf/verifier.c
14493
if (is_iter_next_kfunc(&meta)) {
kernel/bpf/verifier.c
14494
err = process_iter_next_call(env, insn_idx, &meta);
kernel/bpf/verifier.c
14499
if (meta.func_id == special_kfunc_list[KF_bpf_session_cookie])
kernel/bpf/verifier.c
18356
struct bpf_kfunc_call_arg_meta meta;
kernel/bpf/verifier.c
18381
err = fetch_kfunc_arg_meta(env, call->imm, call->off, &meta);
kernel/bpf/verifier.c
18385
cs->num_params = btf_type_vlen(meta.func_proto);
kernel/bpf/verifier.c
18386
cs->fastcall = meta.kfunc_flags & KF_FASTCALL;
kernel/bpf/verifier.c
18387
cs->is_void = btf_type_is_void(btf_type_by_id(meta.btf, meta.func_proto->type));
kernel/bpf/verifier.c
18887
struct bpf_kfunc_call_arg_meta meta;
kernel/bpf/verifier.c
18889
ret = fetch_kfunc_arg_meta(env, insn->imm, insn->off, &meta);
kernel/bpf/verifier.c
18890
if (ret == 0 && is_iter_next_kfunc(&meta)) {
kernel/bpf/verifier.c
18910
if (ret == 0 && is_kfunc_sleepable(&meta))
kernel/bpf/verifier.c
18912
if (ret == 0 && is_kfunc_pkt_changing(&meta))
kernel/bpf/verifier.c
428
struct btf_struct_meta *meta;
kernel/bpf/verifier.c
433
meta = btf_find_struct_meta(reg->btf, reg->btf_id);
kernel/bpf/verifier.c
434
if (meta)
kernel/bpf/verifier.c
435
rec = meta->record;
kernel/bpf/verifier.c
5699
struct bpf_call_arg_meta *meta);
kernel/bpf/verifier.c
6100
struct btf_struct_meta *meta;
kernel/bpf/verifier.c
6105
meta = btf_find_struct_meta(kptr_field->kptr.btf,
kernel/bpf/verifier.c
6108
return meta ? meta->record : NULL;
kernel/bpf/verifier.c
6304
const struct bpf_call_arg_meta *meta,
kernel/bpf/verifier.c
6328
if (meta)
kernel/bpf/verifier.c
6329
return meta->pkt_access;
kernel/bpf/verifier.c
7307
struct bpf_iter_meta *meta;
kernel/bpf/verifier.c
8209
enum bpf_access_type type, struct bpf_call_arg_meta *meta)
kernel/bpf/verifier.c
8254
if (meta && meta->raw_mode)
kernel/bpf/verifier.c
8255
meta = NULL;
kernel/bpf/verifier.c
8261
if (meta && meta->raw_mode) {
kernel/bpf/verifier.c
8287
meta->access_size = access_size;
kernel/bpf/verifier.c
8288
meta->regno = regno;
kernel/bpf/verifier.c
8355
struct bpf_call_arg_meta *meta)
kernel/bpf/verifier.c
8408
zero_size_allowed, access_type, meta);
kernel/bpf/verifier.c
8453
struct bpf_call_arg_meta *meta)
kernel/bpf/verifier.c
8465
meta->msize_max_value = reg->umax_value;
kernel/bpf/verifier.c
8473
meta = NULL;
kernel/bpf/verifier.c
8493
access_type, zero_size_allowed, meta);
kernel/bpf/verifier.c
8533
struct bpf_call_arg_meta meta;
kernel/bpf/verifier.c
8538
memset(&meta, 0, sizeof(meta));
kernel/bpf/verifier.c
8545
err = check_mem_size_reg(env, reg, regno, BPF_READ, true, &meta);
kernel/bpf/verifier.c
8546
err = err ?: check_mem_size_reg(env, reg, regno, BPF_WRITE, true, &meta);
kernel/bpf/verifier.c
8764
struct bpf_call_arg_meta *meta)
kernel/bpf/verifier.c
8766
return process_timer_func(env, regno, &meta->map);
kernel/bpf/verifier.c
8770
struct bpf_kfunc_call_arg_meta *meta)
kernel/bpf/verifier.c
8772
return process_timer_func(env, regno, &meta->map);
kernel/bpf/verifier.c
8776
struct bpf_call_arg_meta *meta)
kernel/bpf/verifier.c
8794
meta->map.ptr = map_ptr;
kernel/bpf/verifier.c
8819
meta->kptr_field = kptr_field;
kernel/bpf/verifier.c
8935
static bool is_iter_kfunc(struct bpf_kfunc_call_arg_meta *meta)
kernel/bpf/verifier.c
8937
return meta->kfunc_flags & (KF_ITER_NEW | KF_ITER_NEXT | KF_ITER_DESTROY);
kernel/bpf/verifier.c
8940
static bool is_iter_new_kfunc(struct bpf_kfunc_call_arg_meta *meta)
kernel/bpf/verifier.c
8942
return meta->kfunc_flags & KF_ITER_NEW;
kernel/bpf/verifier.c
8945
static bool is_iter_next_kfunc(struct bpf_kfunc_call_arg_meta *meta)
kernel/bpf/verifier.c
8947
return meta->kfunc_flags & KF_ITER_NEXT;
kernel/bpf/verifier.c
8950
static bool is_iter_destroy_kfunc(struct bpf_kfunc_call_arg_meta *meta)
kernel/bpf/verifier.c
8952
return meta->kfunc_flags & KF_ITER_DESTROY;
kernel/bpf/verifier.c
8955
static bool is_kfunc_arg_iter(struct bpf_kfunc_call_arg_meta *meta, int arg_idx,
kernel/bpf/verifier.c
8961
if (is_iter_kfunc(meta))
kernel/bpf/verifier.c
8965
return btf_param_match_suffix(meta->btf, arg, "__iter");
kernel/bpf/verifier.c
8969
struct bpf_kfunc_call_arg_meta *meta)
kernel/bpf/verifier.c
8986
btf_id = btf_check_iter_arg(meta->btf, meta->func_proto, regno - 1);
kernel/bpf/verifier.c
8991
t = btf_type_by_id(meta->btf, btf_id);
kernel/bpf/verifier.c
8994
if (is_iter_new_kfunc(meta)) {
kernel/bpf/verifier.c
8998
iter_type_str(meta->btf, btf_id), regno - 1);
kernel/bpf/verifier.c
9009
err = mark_stack_slots_iter(env, meta, reg, insn_idx, meta->btf, btf_id, nr_slots);
kernel/bpf/verifier.c
9016
err = is_iter_reg_valid_init(env, reg, meta->btf, btf_id, nr_slots);
kernel/bpf/verifier.c
9022
iter_type_str(meta->btf, btf_id), regno - 1);
kernel/bpf/verifier.c
9025
verbose(env, "expected an RCU CS when using %s\n", meta->func_name);
kernel/bpf/verifier.c
9040
meta->iter.spi = spi;
kernel/bpf/verifier.c
9041
meta->iter.frameno = reg->frameno;
kernel/bpf/verifier.c
9042
meta->ref_obj_id = iter_ref_obj_id(env, reg, spi);
kernel/bpf/verifier.c
9044
if (is_iter_destroy_kfunc(meta)) {
kernel/bpf/verifier.c
9140
struct bpf_kfunc_call_arg_meta *meta)
kernel/bpf/verifier.c
9142
int iter_frameno = meta->iter.frameno;
kernel/bpf/verifier.c
9143
int iter_spi = meta->iter.spi;
kernel/bpf/verifier.c
9227
struct bpf_kfunc_call_arg_meta *meta)
kernel/bpf/verifier.c
9235
cur_iter = get_iter_from_state(cur_st, meta);
kernel/bpf/verifier.c
9263
queued_iter = get_iter_from_state(queued_st, meta);
kernel/bpf/verifier.c
9304
const struct bpf_call_arg_meta *meta,
kernel/bpf/verifier.c
9307
if (!meta->map.ptr) {
kernel/bpf/verifier.c
9313
switch (meta->map.ptr->map_type) {
kernel/bpf/verifier.c
9324
if (meta->func_id == BPF_FUNC_map_peek_elem)
kernel/bpf/verifier.c
9447
struct bpf_call_arg_meta *meta)
kernel/bpf/verifier.c
9480
if (meta->func_id == BPF_FUNC_kptr_xchg && type_is_alloc(type) && regno == BPF_REG_2) {
kernel/bpf/verifier.c
9508
func_id_name(meta->func_id),
kernel/bpf/verifier.c
9528
meta->func_id != BPF_FUNC_sk_release;
kernel/bpf/verifier.c
9544
if (meta->func_id == BPF_FUNC_kptr_xchg) {
kernel/bpf/verifier.c
9545
if (map_kptr_match_type(env, meta->kptr_field, reg, regno))
kernel/bpf/verifier.c
9568
if (meta->func_id != BPF_FUNC_spin_lock && meta->func_id != BPF_FUNC_spin_unlock &&
kernel/bpf/verifier.c
9569
meta->func_id != BPF_FUNC_kptr_xchg) {
kernel/bpf/verifier.c
9574
if (meta->func_id == BPF_FUNC_kptr_xchg && regno == BPF_REG_2) {
kernel/bpf/verifier.c
9575
if (map_kptr_match_type(env, meta->kptr_field, reg, regno))
kernel/bpf/verifier.c
9862
struct bpf_call_arg_meta *meta,
kernel/bpf/verifier.c
9891
!may_access_direct_pkt_data(env, meta, BPF_READ)) {
kernel/bpf/verifier.c
9897
err = resolve_map_arg_type(env, meta, &arg_type);
kernel/bpf/verifier.c
9913
err = check_reg_type(env, regno, arg_type, arg_btf_id, meta);
kernel/bpf/verifier.c
9947
if (meta->release_regno) {
kernel/bpf/verifier.c
9951
meta->release_regno = regno;
kernel/bpf/verifier.c
9955
if (meta->ref_obj_id) {
kernel/bpf/verifier.c
9958
meta->ref_obj_id);
kernel/bpf/verifier.c
9961
meta->ref_obj_id = reg->ref_obj_id;
kernel/bpf/verifier.c
9967
if (meta->map.ptr) {
kernel/bpf/verifier.c
9980
if (meta->map.ptr != reg->map_ptr ||
kernel/bpf/verifier.c
9981
meta->map.uid != reg->map_uid) {
kernel/bpf/verifier.c
9984
meta->map.uid, reg->map_uid);
kernel/bpf/verifier.c
9988
meta->map.ptr = reg->map_ptr;
kernel/bpf/verifier.c
9989
meta->map.uid = reg->map_uid;
kernel/bpf/verifier.c
9996
if (!meta->map.ptr) {
kernel/kallsyms.c
798
__bpf_md_ptr(struct bpf_iter_meta *, meta);
kernel/kallsyms.c
805
struct bpf_iter_meta meta;
kernel/kallsyms.c
808
meta.seq = m;
kernel/kallsyms.c
809
prog = bpf_iter_get_info(&meta, in_stop);
kernel/kallsyms.c
813
ctx.meta = &meta;
kernel/kallsyms.c
852
DEFINE_BPF_ITER_FUNC(ksym, struct bpf_iter_meta *meta, struct kallsym_iter *ksym)
kernel/trace/bpf_trace.c
713
u64 bpf_event_output(struct bpf_map *map, u64 flags, void *meta, u64 meta_size,
kernel/trace/bpf_trace.c
727
.data = meta,
kernel/trace/ring_buffer.c
1313
struct ring_buffer_cpu_meta *meta = cpu_buffer->ring_meta;
kernel/trace/ring_buffer.c
1314
meta->head_buffer = (unsigned long)head->page;
kernel/trace/ring_buffer.c
1629
struct ring_buffer_cpu_meta *meta;
kernel/trace/ring_buffer.c
1634
bmeta = buffer->meta;
kernel/trace/ring_buffer.c
1639
meta = (struct ring_buffer_cpu_meta *)ptr;
kernel/trace/ring_buffer.c
1643
nr_subbufs = meta->nr_subbufs;
kernel/trace/ring_buffer.c
1676
static void *rb_subbufs_from_meta(struct ring_buffer_cpu_meta *meta)
kernel/trace/ring_buffer.c
1678
int subbuf_size = meta->subbuf_size;
kernel/trace/ring_buffer.c
1681
ptr = (unsigned long)meta;
kernel/trace/ring_buffer.c
1682
ptr = rb_range_align_subbuf(ptr, subbuf_size, meta->nr_subbufs);
kernel/trace/ring_buffer.c
1692
struct ring_buffer_cpu_meta *meta;
kernel/trace/ring_buffer.c
1696
meta = rb_range_meta(cpu_buffer->buffer, 0, cpu_buffer->cpu);
kernel/trace/ring_buffer.c
1697
if (!meta)
kernel/trace/ring_buffer.c
1700
if (WARN_ON_ONCE(idx >= meta->nr_subbufs))
kernel/trace/ring_buffer.c
1703
subbuf_size = meta->subbuf_size;
kernel/trace/ring_buffer.c
1706
idx = meta->buffers[idx];
kernel/trace/ring_buffer.c
1708
ptr = (unsigned long)rb_subbufs_from_meta(meta);
kernel/trace/ring_buffer.c
1729
buffer->meta = bmeta;
kernel/trace/ring_buffer.c
1786
static bool rb_cpu_meta_valid(struct ring_buffer_cpu_meta *meta, int cpu,
kernel/trace/ring_buffer.c
1799
buffers_start = meta->first_buffer;
kernel/trace/ring_buffer.c
1800
buffers_end = meta->first_buffer + (subbuf_size * meta->nr_subbufs);
kernel/trace/ring_buffer.c
1803
if (meta->head_buffer < buffers_start ||
kernel/trace/ring_buffer.c
1804
meta->head_buffer >= buffers_end) {
kernel/trace/ring_buffer.c
1809
if (meta->commit_buffer < buffers_start ||
kernel/trace/ring_buffer.c
1810
meta->commit_buffer >= buffers_end) {
kernel/trace/ring_buffer.c
1815
subbuf = rb_subbufs_from_meta(meta);
kernel/trace/ring_buffer.c
1817
bitmap_clear(subbuf_mask, 0, meta->nr_subbufs);
kernel/trace/ring_buffer.c
1820
for (i = 0; i < meta->nr_subbufs; i++) {
kernel/trace/ring_buffer.c
1821
if (meta->buffers[i] < 0 ||
kernel/trace/ring_buffer.c
1822
meta->buffers[i] >= meta->nr_subbufs) {
kernel/trace/ring_buffer.c
1832
if (test_bit(meta->buffers[i], subbuf_mask)) {
kernel/trace/ring_buffer.c
1837
set_bit(meta->buffers[i], subbuf_mask);
kernel/trace/ring_buffer.c
1844
static int rb_meta_subbuf_idx(struct ring_buffer_cpu_meta *meta, void *subbuf);
kernel/trace/ring_buffer.c
1915
struct ring_buffer_cpu_meta *meta = cpu_buffer->ring_meta;
kernel/trace/ring_buffer.c
1923
if (!meta || !meta->head_buffer)
kernel/trace/ring_buffer.c
1948
for (i = 0; i < meta->nr_subbufs + 1; i++, rb_dec_page(&head_page)) {
kernel/trace/ring_buffer.c
2011
meta->head_buffer = (unsigned long)head_page->page;
kernel/trace/ring_buffer.c
2015
meta->buffers[0] = rb_meta_subbuf_idx(meta, bpage->page);
kernel/trace/ring_buffer.c
2018
for (i = 1, bpage = head_page; i < meta->nr_subbufs;
kernel/trace/ring_buffer.c
2020
meta->buffers[i] = rb_meta_subbuf_idx(meta, bpage->page);
kernel/trace/ring_buffer.c
2030
if (meta->commit_buffer == (unsigned long)cpu_buffer->reader_page->page) {
kernel/trace/ring_buffer.c
2037
for (i = 0; i < meta->nr_subbufs + 1; i++, rb_inc_page(&head_page)) {
kernel/trace/ring_buffer.c
2076
meta->head_buffer = 0;
kernel/trace/ring_buffer.c
2077
meta->commit_buffer = 0;
kernel/trace/ring_buffer.c
2084
for (i = 0; i < meta->nr_subbufs - 1; i++, rb_inc_page(&head_page)) {
kernel/trace/ring_buffer.c
2092
struct ring_buffer_cpu_meta *meta;
kernel/trace/ring_buffer.c
2110
meta = rb_range_meta(buffer, nr_pages, cpu);
kernel/trace/ring_buffer.c
2112
if (valid && rb_cpu_meta_valid(meta, cpu, buffer, nr_pages, subbuf_mask)) {
kernel/trace/ring_buffer.c
2114
subbuf = rb_subbufs_from_meta(meta);
kernel/trace/ring_buffer.c
2115
delta = (unsigned long)subbuf - meta->first_buffer;
kernel/trace/ring_buffer.c
2116
meta->first_buffer += delta;
kernel/trace/ring_buffer.c
2117
meta->head_buffer += delta;
kernel/trace/ring_buffer.c
2118
meta->commit_buffer += delta;
kernel/trace/ring_buffer.c
2127
memset(meta, 0, next_meta - (void *)meta);
kernel/trace/ring_buffer.c
2129
meta->nr_subbufs = nr_pages + 1;
kernel/trace/ring_buffer.c
2130
meta->subbuf_size = PAGE_SIZE;
kernel/trace/ring_buffer.c
2132
subbuf = rb_subbufs_from_meta(meta);
kernel/trace/ring_buffer.c
2134
meta->first_buffer = (unsigned long)subbuf;
kernel/trace/ring_buffer.c
2144
for (i = 0; i < meta->nr_subbufs; i++) {
kernel/trace/ring_buffer.c
2145
meta->buffers[i] = i;
kernel/trace/ring_buffer.c
2147
subbuf += meta->subbuf_size;
kernel/trace/ring_buffer.c
2156
struct ring_buffer_cpu_meta *meta = cpu_buffer->ring_meta;
kernel/trace/ring_buffer.c
2159
if (!meta)
kernel/trace/ring_buffer.c
2162
if (*pos > meta->nr_subbufs)
kernel/trace/ring_buffer.c
2181
struct ring_buffer_cpu_meta *meta = cpu_buffer->ring_meta;
kernel/trace/ring_buffer.c
2186
rb_meta_subbuf_idx(meta, (void *)meta->head_buffer));
kernel/trace/ring_buffer.c
2188
rb_meta_subbuf_idx(meta, (void *)meta->commit_buffer));
kernel/trace/ring_buffer.c
2189
seq_printf(m, "subbuf_size: %d\n", meta->subbuf_size);
kernel/trace/ring_buffer.c
2190
seq_printf(m, "nr_subbufs: %d\n", meta->nr_subbufs);
kernel/trace/ring_buffer.c
2195
seq_printf(m, "buffer[%ld]: %d\n", val, meta->buffers[val]);
kernel/trace/ring_buffer.c
2230
struct ring_buffer_cpu_meta *meta = cpu_buffer->ring_meta;
kernel/trace/ring_buffer.c
2232
if (meta->head_buffer == (unsigned long)bpage->page)
kernel/trace/ring_buffer.c
2235
if (meta->commit_buffer == (unsigned long)bpage->page) {
kernel/trace/ring_buffer.c
2245
struct ring_buffer_cpu_meta *meta = NULL;
kernel/trace/ring_buffer.c
2274
meta = rb_range_meta(buffer, nr_pages, cpu_buffer->cpu);
kernel/trace/ring_buffer.c
2290
if (meta) {
kernel/trace/ring_buffer.c
2296
if (meta->head_buffer)
kernel/trace/ring_buffer.c
2357
struct ring_buffer_cpu_meta *meta;
kernel/trace/ring_buffer.c
2414
meta = cpu_buffer->ring_meta;
kernel/trace/ring_buffer.c
2415
if (!meta || !meta->head_buffer ||
kernel/trace/ring_buffer.c
2417
if (meta && meta->head_buffer &&
kernel/trace/ring_buffer.c
2435
meta->commit_buffer = meta->head_buffer;
kernel/trace/ring_buffer.c
2655
struct ring_buffer_meta *meta;
kernel/trace/ring_buffer.c
2658
if (!buffer || !buffer->meta)
kernel/trace/ring_buffer.c
2661
meta = buffer->meta;
kernel/trace/ring_buffer.c
2663
ptr = (void *)ALIGN((unsigned long)meta + sizeof(*meta), sizeof(long));
kernel/trace/ring_buffer.c
2666
*size = (void *)meta + meta->buffers_offset - ptr;
kernel/trace/ring_buffer.c
3282
static int rb_meta_subbuf_idx(struct ring_buffer_cpu_meta *meta, void *subbuf)
kernel/trace/ring_buffer.c
3286
subbuf_array = (void *)meta + sizeof(int) * meta->nr_subbufs;
kernel/trace/ring_buffer.c
3287
subbuf_array = (void *)ALIGN((unsigned long)subbuf_array, meta->subbuf_size);
kernel/trace/ring_buffer.c
3288
return (subbuf - subbuf_array) / meta->subbuf_size;
kernel/trace/ring_buffer.c
3294
struct ring_buffer_cpu_meta *meta = cpu_buffer->ring_meta;
kernel/trace/ring_buffer.c
3305
(void)cmpxchg(&meta->head_buffer, old_head, new_head);
kernel/trace/ring_buffer.c
3311
struct ring_buffer_cpu_meta *meta = cpu_buffer->ring_meta;
kernel/trace/ring_buffer.c
3320
meta->buffers[0] = rb_meta_subbuf_idx(meta, new_reader);
kernel/trace/ring_buffer.c
3321
meta->buffers[id] = rb_meta_subbuf_idx(meta, old_reader);
kernel/trace/ring_buffer.c
3940
struct ring_buffer_cpu_meta *meta = cpu_buffer->ring_meta;
kernel/trace/ring_buffer.c
3941
meta->commit_buffer = (unsigned long)cpu_buffer->commit_page->page;
kernel/trace/ring_buffer.c
593
struct ring_buffer_meta *meta;
kernel/trace/ring_buffer.c
6143
struct trace_buffer_meta *meta = cpu_buffer->meta_page;
kernel/trace/ring_buffer.c
6145
if (!meta)
kernel/trace/ring_buffer.c
6148
meta->reader.read = cpu_buffer->reader_page->read;
kernel/trace/ring_buffer.c
6149
meta->reader.id = rb_page_id(cpu_buffer, cpu_buffer->reader_page,
kernel/trace/ring_buffer.c
6152
meta->reader.lost_events = cpu_buffer->lost_events;
kernel/trace/ring_buffer.c
6154
meta->entries = local_read(&cpu_buffer->entries);
kernel/trace/ring_buffer.c
6155
meta->overrun = local_read(&cpu_buffer->overrun);
kernel/trace/ring_buffer.c
6156
meta->read = cpu_buffer->read;
kernel/trace/ring_buffer.c
6212
struct ring_buffer_cpu_meta *meta = cpu_buffer->ring_meta;
kernel/trace/ring_buffer.c
6213
meta->commit_buffer = meta->head_buffer;
kernel/trace/ring_buffer.c
7039
struct trace_buffer_meta *meta = cpu_buffer->meta_page;
kernel/trace/ring_buffer.c
7068
meta->meta_struct_len = sizeof(*meta);
kernel/trace/ring_buffer.c
7069
meta->nr_subbufs = nr_subbufs;
kernel/trace/ring_buffer.c
7070
meta->subbuf_size = cpu_buffer->buffer->subbuf_size + BUF_PAGE_HDR_SIZE;
kernel/trace/ring_buffer.c
7071
meta->meta_page_size = meta->subbuf_size;
kernel/trace/trace_syscalls.c
1334
struct syscall_metadata *meta;
kernel/trace/trace_syscalls.c
1350
meta = find_syscall_meta(addr);
kernel/trace/trace_syscalls.c
1351
if (!meta)
kernel/trace/trace_syscalls.c
1354
meta->syscall_nr = i;
kernel/trace/trace_syscalls.c
1357
syscalls_metadata[i] = meta;
kernel/trace/trace_syscalls.c
1359
ret = xa_store(&syscalls_metadata_sparse, i, meta,
kernel/trace/trace_syscalls.c
529
struct syscall_metadata *meta = call->data;
kernel/trace/trace_syscalls.c
537
for (i = 0; i < meta->nb_args; i++) {
kernel/trace/trace_syscalls.c
538
ret = trace_define_field(call, meta->types[i],
kernel/trace/trace_syscalls.c
539
meta->args[i], offset,
kernel/trace/trace_syscalls.c
547
if (ret || !meta->user_mask)
kernel/trace/trace_syscalls.c
550
mask = meta->user_mask;
kernel/trace/trace_syscalls.c
562
len = strlen(meta->args[idx]) + sizeof("___val");
kernel/trace/trace_syscalls.c
565
meta->user_mask = 0;
kernel/trace/trace_syscalls.c
569
snprintf(arg, len, "__%s_val", meta->args[idx]);
lib/maple_tree.c
6775
pr_cont("| %02X %02X| ", node->meta.end, node->meta.gap);
lib/maple_tree.c
804
return &mn->ma64.meta;
lib/maple_tree.c
806
return &mn->mr64.meta;
lib/maple_tree.c
820
struct maple_metadata *meta = ma_meta(mn, mt);
lib/maple_tree.c
822
meta->gap = offset;
lib/maple_tree.c
823
meta->end = end;
lib/maple_tree.c
835
struct maple_metadata *meta;
lib/maple_tree.c
853
meta = ma_meta(mn, type);
lib/maple_tree.c
859
meta->gap = 0;
lib/maple_tree.c
860
meta->end = 0;
lib/maple_tree.c
871
struct maple_metadata *meta = ma_meta(mn, mt);
lib/maple_tree.c
873
return meta->end;
lib/maple_tree.c
882
return mn->ma64.meta.gap;
lib/maple_tree.c
895
struct maple_metadata *meta = ma_meta(mn, mt);
lib/maple_tree.c
897
meta->gap = offset;
lib/parser.c
36
char *meta;
lib/parser.c
44
meta = strchr(p, '%');
lib/parser.c
45
if (!meta)
lib/parser.c
48
if (strncmp(p, s, meta-p))
lib/parser.c
51
s += meta - p;
lib/parser.c
52
p = meta + 1;
mm/kasan/generic.c
501
static void release_alloc_meta(struct kasan_alloc_meta *meta)
mm/kasan/generic.c
504
__memset(meta, 0, sizeof(*meta));
mm/kasan/generic.c
507
static void release_free_meta(const void *object, struct kasan_free_meta *meta)
mm/kasan/quarantine.c
190
struct kasan_free_meta *meta = kasan_get_free_meta(cache, object);
mm/kasan/quarantine.c
196
if (!meta)
mm/kasan/quarantine.c
214
qlist_put(q, &meta->quarantine_link, cache->size);
mm/kfence/core.c
1079
struct kfence_metadata *meta;
mm/kfence/core.c
1089
meta = &kfence_metadata[i];
mm/kfence/core.c
1098
if (READ_ONCE(meta->cache) != s || !kfence_obj_allocated(meta))
mm/kfence/core.c
1101
raw_spin_lock_irqsave(&meta->lock, flags);
mm/kfence/core.c
1102
in_use = meta->cache == s && kfence_obj_allocated(meta);
mm/kfence/core.c
1103
raw_spin_unlock_irqrestore(&meta->lock, flags);
mm/kfence/core.c
1120
kfence_guarded_free((void *)meta->addr, meta, /*zombie=*/true);
mm/kfence/core.c
1125
meta = &kfence_metadata[i];
mm/kfence/core.c
1128
if (READ_ONCE(meta->cache) != s || READ_ONCE(meta->state) != KFENCE_OBJECT_FREED)
mm/kfence/core.c
1131
raw_spin_lock_irqsave(&meta->lock, flags);
mm/kfence/core.c
1132
if (meta->cache == s && meta->state == KFENCE_OBJECT_FREED)
mm/kfence/core.c
1133
meta->cache = NULL;
mm/kfence/core.c
1134
raw_spin_unlock_irqrestore(&meta->lock, flags);
mm/kfence/core.c
1216
const struct kfence_metadata *meta = addr_to_metadata((unsigned long)addr);
mm/kfence/core.c
1222
return meta ? meta->size : 0;
mm/kfence/core.c
1227
const struct kfence_metadata *meta = addr_to_metadata((unsigned long)addr);
mm/kfence/core.c
1233
return meta ? (void *)meta->addr : NULL;
mm/kfence/core.c
1238
struct kfence_metadata *meta = addr_to_metadata((unsigned long)addr);
mm/kfence/core.c
1241
KFENCE_WARN_ON(meta->obj_exts.objcg);
mm/kfence/core.c
1251
if (unlikely(meta->cache && (meta->cache->flags & SLAB_TYPESAFE_BY_RCU))) {
mm/kfence/core.c
1254
raw_spin_lock_irqsave(&meta->lock, flags);
mm/kfence/core.c
1255
metadata_update_state(meta, KFENCE_OBJECT_RCU_FREEING, NULL, 0);
mm/kfence/core.c
1256
raw_spin_unlock_irqrestore(&meta->lock, flags);
mm/kfence/core.c
1257
call_rcu(&meta->rcu_head, rcu_guarded_free);
mm/kfence/core.c
1259
kfence_guarded_free(addr, meta, false);
mm/kfence/core.c
1281
struct kfence_metadata *meta;
mm/kfence/core.c
1284
meta = addr_to_metadata(addr - PAGE_SIZE);
mm/kfence/core.c
1285
if (meta && kfence_obj_allocated(meta)) {
mm/kfence/core.c
1286
to_report = meta;
mm/kfence/core.c
1288
distance = addr - data_race(meta->addr + meta->size);
mm/kfence/core.c
1291
meta = addr_to_metadata(addr + PAGE_SIZE);
mm/kfence/core.c
1292
if (meta && kfence_obj_allocated(meta)) {
mm/kfence/core.c
1294
if (!to_report || distance > data_race(meta->addr) - addr)
mm/kfence/core.c
1295
to_report = meta;
mm/kfence/core.c
257
static inline unsigned long metadata_to_pageaddr(const struct kfence_metadata *meta)
mm/kfence/core.c
258
__must_hold(&meta->lock)
mm/kfence/core.c
260
unsigned long offset = (meta - kfence_metadata + 1) * PAGE_SIZE * 2;
mm/kfence/core.c
266
if (KFENCE_WARN_ON(meta < kfence_metadata ||
mm/kfence/core.c
267
meta >= kfence_metadata + CONFIG_KFENCE_NUM_OBJECTS))
mm/kfence/core.c
274
if (KFENCE_WARN_ON(ALIGN_DOWN(meta->addr, PAGE_SIZE) != pageaddr))
mm/kfence/core.c
280
static inline bool kfence_obj_allocated(const struct kfence_metadata *meta)
mm/kfence/core.c
282
enum kfence_object_state state = READ_ONCE(meta->state);
mm/kfence/core.c
292
metadata_update_state(struct kfence_metadata *meta, enum kfence_object_state next,
mm/kfence/core.c
294
__must_hold(&meta->lock)
mm/kfence/core.c
297
next == KFENCE_OBJECT_ALLOCATED ? &meta->alloc_track : &meta->free_track;
mm/kfence/core.c
299
lockdep_assert_held(&meta->lock);
mm/kfence/core.c
302
if (READ_ONCE(meta->state) == KFENCE_OBJECT_RCU_FREEING)
mm/kfence/core.c
326
WRITE_ONCE(meta->state, next);
mm/kfence/core.c
338
struct kfence_metadata *meta;
mm/kfence/core.c
346
meta = addr_to_metadata((unsigned long)addr);
mm/kfence/core.c
347
raw_spin_lock_irqsave(&meta->lock, flags);
mm/kfence/core.c
348
kfence_report_error((unsigned long)addr, false, NULL, meta, KFENCE_ERROR_CORRUPTION);
mm/kfence/core.c
349
raw_spin_unlock_irqrestore(&meta->lock, flags);
mm/kfence/core.c
354
static inline void set_canary(const struct kfence_metadata *meta)
mm/kfence/core.c
356
const unsigned long pageaddr = ALIGN_DOWN(meta->addr, PAGE_SIZE);
mm/kfence/core.c
363
for (; addr < meta->addr; addr += sizeof(u64))
mm/kfence/core.c
366
addr = ALIGN_DOWN(meta->addr + meta->size, sizeof(u64));
mm/kfence/core.c
372
check_canary(const struct kfence_metadata *meta)
mm/kfence/core.c
374
const unsigned long pageaddr = ALIGN_DOWN(meta->addr, PAGE_SIZE);
mm/kfence/core.c
387
for (; meta->addr - addr >= sizeof(u64); addr += sizeof(u64)) {
mm/kfence/core.c
397
for (; addr < meta->addr; addr++) {
mm/kfence/core.c
403
for (addr = meta->addr + meta->size; addr % sizeof(u64) != 0; addr++) {
mm/kfence/core.c
422
struct kfence_metadata *meta = NULL;
mm/kfence/core.c
433
meta = list_entry(kfence_freelist.next, struct kfence_metadata, list);
mm/kfence/core.c
434
list_del_init(&meta->list);
mm/kfence/core.c
437
if (!meta) {
mm/kfence/core.c
442
if (unlikely(!raw_spin_trylock_irqsave(&meta->lock, flags))) {
mm/kfence/core.c
454
list_add_tail(&meta->list, &kfence_freelist);
mm/kfence/core.c
460
meta->addr = metadata_to_pageaddr(meta);
mm/kfence/core.c
462
if (meta->state == KFENCE_OBJECT_FREED)
mm/kfence/core.c
463
kfence_unprotect(meta->addr);
mm/kfence/core.c
475
meta->addr += PAGE_SIZE - size;
mm/kfence/core.c
476
meta->addr = ALIGN_DOWN(meta->addr, cache->align);
mm/kfence/core.c
479
addr = (void *)meta->addr;
mm/kfence/core.c
482
metadata_update_state(meta, KFENCE_OBJECT_ALLOCATED, stack_entries, num_stack_entries);
mm/kfence/core.c
484
WRITE_ONCE(meta->cache, cache);
mm/kfence/core.c
485
meta->size = size;
mm/kfence/core.c
486
meta->alloc_stack_hash = alloc_stack_hash;
mm/kfence/core.c
487
raw_spin_unlock_irqrestore(&meta->lock, flags);
mm/kfence/core.c
497
set_canary(meta);
mm/kfence/core.c
510
kfence_protect(meta->addr); /* Random "faults" by protecting the object. */
mm/kfence/core.c
518
static void kfence_guarded_free(void *addr, struct kfence_metadata *meta, bool zombie)
mm/kfence/core.c
525
raw_spin_lock_irqsave(&meta->lock, flags);
mm/kfence/core.c
527
if (!kfence_obj_allocated(meta) || meta->addr != (unsigned long)addr) {
mm/kfence/core.c
530
kfence_report_error((unsigned long)addr, false, NULL, meta,
mm/kfence/core.c
532
raw_spin_unlock_irqrestore(&meta->lock, flags);
mm/kfence/core.c
545
if (meta->unprotected_page) {
mm/kfence/core.c
546
memzero_explicit((void *)ALIGN_DOWN(meta->unprotected_page, PAGE_SIZE), PAGE_SIZE);
mm/kfence/core.c
547
kfence_protect(meta->unprotected_page);
mm/kfence/core.c
548
meta->unprotected_page = 0;
mm/kfence/core.c
552
metadata_update_state(meta, KFENCE_OBJECT_FREED, NULL, 0);
mm/kfence/core.c
553
init = slab_want_init_on_free(meta->cache);
mm/kfence/core.c
554
alloc_stack_hash = meta->alloc_stack_hash;
mm/kfence/core.c
555
raw_spin_unlock_irqrestore(&meta->lock, flags);
mm/kfence/core.c
560
check_canary(meta);
mm/kfence/core.c
568
memzero_explicit(addr, meta->size);
mm/kfence/core.c
577
KFENCE_WARN_ON(!list_empty(&meta->list));
mm/kfence/core.c
578
list_add_tail(&meta->list, &kfence_freelist);
mm/kfence/core.c
591
struct kfence_metadata *meta = container_of(h, struct kfence_metadata, rcu_head);
mm/kfence/core.c
593
kfence_guarded_free((void *)meta->addr, meta, false);
mm/kfence/core.c
650
struct kfence_metadata *meta = &kfence_metadata_init[i];
mm/kfence/core.c
653
INIT_LIST_HEAD(&meta->list);
mm/kfence/core.c
654
raw_spin_lock_init(&meta->lock);
mm/kfence/core.c
655
meta->state = KFENCE_OBJECT_UNUSED;
mm/kfence/core.c
657
meta->addr = i;
mm/kfence/core.c
783
struct kfence_metadata *meta = &kfence_metadata[(long)v - 1];
mm/kfence/core.c
786
raw_spin_lock_irqsave(&meta->lock, flags);
mm/kfence/core.c
787
kfence_print_object(seq, meta);
mm/kfence/core.c
788
raw_spin_unlock_irqrestore(&meta->lock, flags);
mm/kfence/core.c
824
struct kfence_metadata *meta = &kfence_metadata[i];
mm/kfence/core.c
826
if (kfence_obj_allocated(meta))
mm/kfence/core.c
827
check_canary(meta);
mm/kfence/kfence.h
144
const struct kfence_metadata *meta, enum kfence_error_type type);
mm/kfence/kfence.h
146
void kfence_print_object(struct seq_file *seq, const struct kfence_metadata *meta) __must_hold(&meta->lock);
mm/kfence/report.c
107
static void kfence_print_stack(struct seq_file *seq, const struct kfence_metadata *meta,
mm/kfence/report.c
109
__must_hold(&meta->lock)
mm/kfence/report.c
111
const struct kfence_track *track = show_alloc ? &meta->alloc_track : &meta->free_track;
mm/kfence/report.c
119
show_alloc ? "allocated" : meta->state == KFENCE_OBJECT_RCU_FREEING ?
mm/kfence/report.c
136
void kfence_print_object(struct seq_file *seq, const struct kfence_metadata *meta)
mm/kfence/report.c
138
const int size = abs(meta->size);
mm/kfence/report.c
139
const unsigned long start = meta->addr;
mm/kfence/report.c
140
const struct kmem_cache *const cache = meta->cache;
mm/kfence/report.c
142
lockdep_assert_held(&meta->lock);
mm/kfence/report.c
144
if (meta->state == KFENCE_OBJECT_UNUSED) {
mm/kfence/report.c
145
seq_con_printf(seq, "kfence-#%td unused\n", meta - kfence_metadata);
mm/kfence/report.c
150
meta - kfence_metadata, (void *)start, (void *)(start + size - 1),
mm/kfence/report.c
153
kfence_print_stack(seq, meta, true);
mm/kfence/report.c
155
if (meta->state == KFENCE_OBJECT_FREED || meta->state == KFENCE_OBJECT_RCU_FREEING) {
mm/kfence/report.c
157
kfence_print_stack(seq, meta, false);
mm/kfence/report.c
166
const struct kfence_metadata *meta)
mm/kfence/report.c
172
end = (const u8 *)(address < meta->addr ? min(show_until_addr, meta->addr)
mm/kfence/report.c
193
const struct kfence_metadata *meta, enum kfence_error_type type)
mm/kfence/report.c
196
const ptrdiff_t object_index = meta ? meta - kfence_metadata : -1;
mm/kfence/report.c
208
if (WARN_ON(type != KFENCE_ERROR_INVALID && !meta))
mm/kfence/report.c
226
const bool left_of_object = address < meta->addr;
mm/kfence/report.c
232
left_of_object ? meta->addr - address : address - meta->addr,
mm/kfence/report.c
245
print_diff_canary(address, 16, meta);
mm/kfence/report.c
264
if (meta) {
mm/kfence/report.c
265
lockdep_assert_held(&meta->lock);
mm/kfence/report.c
267
kfence_print_object(NULL, meta);
mm/kfence/report.c
301
struct kfence_metadata *meta = addr_to_metadata((unsigned long)object);
mm/kfence/report.c
304
if (!meta)
mm/kfence/report.c
314
if (WARN_ON(meta->state == KFENCE_OBJECT_UNUSED))
mm/kfence/report.c
317
raw_spin_lock_irqsave(&meta->lock, flags);
mm/kfence/report.c
320
kpp->kp_slab_cache = meta->cache;
mm/kfence/report.c
321
kpp->kp_objp = (void *)meta->addr;
mm/kfence/report.c
322
kfence_to_kp_stack(&meta->alloc_track, kpp->kp_stack);
mm/kfence/report.c
323
if (meta->state == KFENCE_OBJECT_FREED || meta->state == KFENCE_OBJECT_RCU_FREEING)
mm/kfence/report.c
324
kfence_to_kp_stack(&meta->free_track, kpp->kp_free_stack);
mm/kfence/report.c
328
raw_spin_unlock_irqrestore(&meta->lock, flags);
net/bridge/netfilter/nft_meta_bridge.c
122
const struct nft_meta *meta = nft_expr_priv(expr);
net/bridge/netfilter/nft_meta_bridge.c
123
u32 *sreg = &regs->data[meta->sreg];
net/bridge/netfilter/nft_meta_bridge.c
127
switch (meta->key) {
net/core/bpf_sk_storage.c
785
__bpf_md_ptr(struct bpf_iter_meta *, meta);
net/core/bpf_sk_storage.c
791
DEFINE_BPF_ITER_FUNC(bpf_sk_storage_map, struct bpf_iter_meta *meta,
net/core/bpf_sk_storage.c
801
struct bpf_iter_meta meta;
net/core/bpf_sk_storage.c
805
meta.seq = seq;
net/core/bpf_sk_storage.c
806
prog = bpf_iter_get_info(&meta, selem == NULL);
net/core/bpf_sk_storage.c
808
ctx.meta = &meta;
net/core/filter.c
4290
void *meta = xdp->data_meta + offset;
net/core/filter.c
4291
unsigned long metalen = xdp->data - meta;
net/core/filter.c
4295
if (unlikely(meta < xdp_frame_end ||
net/core/filter.c
4296
meta > xdp->data))
net/core/filter.c
4301
xdp->data_meta = meta;
net/core/filter.c
4687
u64, flags, void *, meta, u64, meta_size)
net/core/filter.c
4696
return bpf_event_output(map, flags, meta, meta_size, skb, skb_size,
net/core/filter.c
5100
u64, flags, void *, meta, u64, meta_size)
net/core/filter.c
5110
return bpf_event_output(map, flags, meta, meta_size, xdp,
net/core/flow_dissector.c
280
struct flow_dissector_key_meta *meta;
net/core/flow_dissector.c
285
meta = skb_flow_dissector_target(flow_dissector,
net/core/flow_dissector.c
288
meta->ingress_ifindex = skb->skb_iif;
net/core/flow_dissector.c
295
meta->l2_miss = ext->l2_miss;
net/core/skbuff.c
6290
void *meta;
net/core/skbuff.c
6305
meta = skb_metadata_end(skb) - meta_len;
net/core/skbuff.c
6306
memmove(meta + VLAN_HLEN, meta, meta_len);
net/core/sock_map.c
1374
struct bpf_iter_meta meta;
net/core/sock_map.c
1377
meta.seq = seq;
net/core/sock_map.c
1378
prog = bpf_iter_get_info(&meta, !elem);
net/core/sock_map.c
1382
ctx.meta = &meta;
net/core/sock_map.c
711
__bpf_md_ptr(struct bpf_iter_meta *, meta);
net/core/sock_map.c
717
DEFINE_BPF_ITER_FUNC(sockmap, struct bpf_iter_meta *meta,
net/core/sock_map.c
761
struct bpf_iter_meta meta;
net/core/sock_map.c
764
meta.seq = seq;
net/core/sock_map.c
765
prog = bpf_iter_get_info(&meta, !v);
net/core/sock_map.c
769
ctx.meta = &meta;
net/dsa/tag_sja1105.c
103
struct sja1105_meta *meta)
net/dsa/tag_sja1105.c
117
packing(buf, &meta->tstamp, 31, 0, 4, UNPACK, 0);
net/dsa/tag_sja1105.c
118
packing(buf + 4, &meta->dmac_byte_3, 7, 0, 1, UNPACK, 0);
net/dsa/tag_sja1105.c
119
packing(buf + 5, &meta->dmac_byte_4, 7, 0, 1, UNPACK, 0);
net/dsa/tag_sja1105.c
120
packing(buf + 6, &meta->source_port, 7, 0, 1, UNPACK, 0);
net/dsa/tag_sja1105.c
121
packing(buf + 7, &meta->switch_id, 7, 0, 1, UNPACK, 0);
net/dsa/tag_sja1105.c
347
const struct sja1105_meta *meta)
net/dsa/tag_sja1105.c
351
hdr->h_dest[3] = meta->dmac_byte_3;
net/dsa/tag_sja1105.c
352
hdr->h_dest[4] = meta->dmac_byte_4;
net/dsa/tag_sja1105.c
353
SJA1105_SKB_CB(skb)->tstamp = meta->tstamp;
net/dsa/tag_sja1105.c
378
struct sja1105_meta *meta,
net/dsa/tag_sja1105.c
454
sja1105_transfer_meta(skb, meta);
net/dsa/tag_sja1105.c
479
struct sja1105_meta meta = {0};
net/dsa/tag_sja1105.c
496
sja1105_meta_unpack(skb, &meta);
net/dsa/tag_sja1105.c
497
source_port = meta.source_port;
net/dsa/tag_sja1105.c
498
switch_id = meta.switch_id;
net/dsa/tag_sja1105.c
522
return sja1105_rcv_meta_state_machine(skb, &meta, is_link_local,
net/ipv4/tcp_ipv4.c
2971
__bpf_md_ptr(struct bpf_iter_meta *, meta);
net/ipv4/tcp_ipv4.c
2976
static int tcp_prog_seq_show(struct bpf_prog *prog, struct bpf_iter_meta *meta,
net/ipv4/tcp_ipv4.c
2981
meta->seq_num--; /* skip SEQ_START_TOKEN */
net/ipv4/tcp_ipv4.c
2982
ctx.meta = meta;
net/ipv4/tcp_ipv4.c
3298
struct bpf_iter_meta meta;
net/ipv4/tcp_ipv4.c
3326
meta.seq = seq;
net/ipv4/tcp_ipv4.c
3327
prog = bpf_iter_get_info(&meta, false);
net/ipv4/tcp_ipv4.c
3328
ret = tcp_prog_seq_show(prog, &meta, v, uid);
net/ipv4/tcp_ipv4.c
3340
struct bpf_iter_meta meta;
net/ipv4/tcp_ipv4.c
3344
meta.seq = seq;
net/ipv4/tcp_ipv4.c
3345
prog = bpf_iter_get_info(&meta, true);
net/ipv4/tcp_ipv4.c
3347
(void)tcp_prog_seq_show(prog, &meta, v, 0);
net/ipv4/tcp_ipv4.c
3646
DEFINE_BPF_ITER_FUNC(tcp, struct bpf_iter_meta *meta,
net/ipv4/udp.c
3465
__bpf_md_ptr(struct bpf_iter_meta *, meta);
net/ipv4/udp.c
3652
static int udp_prog_seq_show(struct bpf_prog *prog, struct bpf_iter_meta *meta,
net/ipv4/udp.c
3657
meta->seq_num--; /* skip SEQ_START_TOKEN */
net/ipv4/udp.c
3658
ctx.meta = meta;
net/ipv4/udp.c
3668
struct bpf_iter_meta meta;
net/ipv4/udp.c
3685
meta.seq = seq;
net/ipv4/udp.c
3686
prog = bpf_iter_get_info(&meta, false);
net/ipv4/udp.c
3687
ret = udp_prog_seq_show(prog, &meta, v, uid, state->bucket);
net/ipv4/udp.c
3714
struct bpf_iter_meta meta;
net/ipv4/udp.c
3718
meta.seq = seq;
net/ipv4/udp.c
3719
prog = bpf_iter_get_info(&meta, true);
net/ipv4/udp.c
3721
(void)udp_prog_seq_show(prog, &meta, v, 0, 0);
net/ipv4/udp.c
3972
DEFINE_BPF_ITER_FUNC(udp, struct bpf_iter_meta *meta,
net/ipv6/ip6_fib.c
2784
struct bpf_iter_meta *meta,
net/ipv6/ip6_fib.c
2789
ctx.meta = meta;
net/ipv6/ip6_fib.c
2797
struct bpf_iter_meta meta;
net/ipv6/ip6_fib.c
2801
meta.seq = seq;
net/ipv6/ip6_fib.c
2802
prog = bpf_iter_get_info(&meta, false);
net/ipv6/ip6_fib.c
2806
ret = ipv6_route_prog_seq_show(prog, &meta, v);
net/ipv6/ip6_fib.c
2814
struct bpf_iter_meta meta;
net/ipv6/ip6_fib.c
2818
meta.seq = seq;
net/ipv6/ip6_fib.c
2819
prog = bpf_iter_get_info(&meta, true);
net/ipv6/ip6_fib.c
2821
(void)ipv6_route_prog_seq_show(prog, &meta, v);
net/ipv6/route.c
6831
DEFINE_BPF_ITER_FUNC(ipv6_route, struct bpf_iter_meta *meta, struct fib6_info *rt)
net/netfilter/nf_flow_table_offload.c
102
NF_FLOW_DISSECTOR(match, FLOW_DISSECTOR_KEY_META, meta);
net/netfilter/nf_flow_table_offload.c
116
key->meta.ingress_ifindex = tuple->tc.iifidx;
net/netfilter/nf_flow_table_offload.c
118
key->meta.ingress_ifindex = tuple->iifidx;
net/netfilter/nf_flow_table_offload.c
120
mask->meta.ingress_ifindex = 0xffffffff;
net/netfilter/nft_cmp.c
169
reg->offset == offsetof(struct nft_flow_key, meta.ingress_iftype) &&
net/netfilter/nft_inner.c
40
struct nft_meta meta;
net/netfilter/nft_meta.c
422
const struct nft_meta *meta = nft_expr_priv(expr);
net/netfilter/nft_meta.c
424
u32 *sreg = &regs->data[meta->sreg];
net/netfilter/nft_meta.c
428
switch (meta->key) {
net/netfilter/nft_meta.c
731
NFT_OFFLOAD_MATCH_EXACT(FLOW_DISSECTOR_KEY_META, meta,
net/netfilter/nft_meta.c
735
NFT_OFFLOAD_MATCH_EXACT(FLOW_DISSECTOR_KEY_META, meta,
net/netfilter/nft_meta.c
749
const struct nft_meta *meta;
net/netfilter/nft_meta.c
756
meta = nft_expr_priv(track->regs[priv->dreg].selector);
net/netfilter/nft_meta.c
757
if (priv->key != meta->key ||
net/netfilter/nft_meta.c
758
priv->dreg != meta->dreg) {
net/netlink/af_netlink.c
2722
__bpf_md_ptr(struct bpf_iter_meta *, meta);
net/netlink/af_netlink.c
2726
DEFINE_BPF_ITER_FUNC(netlink, struct bpf_iter_meta *meta, struct netlink_sock *sk)
net/netlink/af_netlink.c
2729
struct bpf_iter_meta *meta,
net/netlink/af_netlink.c
2734
meta->seq_num--; /* skip SEQ_START_TOKEN */
net/netlink/af_netlink.c
2735
ctx.meta = meta;
net/netlink/af_netlink.c
2742
struct bpf_iter_meta meta;
net/netlink/af_netlink.c
2745
meta.seq = seq;
net/netlink/af_netlink.c
2746
prog = bpf_iter_get_info(&meta, false);
net/netlink/af_netlink.c
2751
return netlink_prog_seq_show(prog, &meta, v);
net/netlink/af_netlink.c
2758
struct bpf_iter_meta meta;
net/netlink/af_netlink.c
2762
meta.seq = seq;
net/netlink/af_netlink.c
2763
prog = bpf_iter_get_info(&meta, true);
net/netlink/af_netlink.c
2765
(void)netlink_prog_seq_show(prog, &meta, v);
net/sched/cls_flower.c
1881
key->meta.ingress_ifindex = err;
net/sched/cls_flower.c
1882
mask->meta.ingress_ifindex = 0xffffffff;
net/sched/cls_flower.c
1885
fl_set_key_val(tb, &key->meta.l2_miss, TCA_FLOWER_L2_MISS,
net/sched/cls_flower.c
1886
&mask->meta.l2_miss, TCA_FLOWER_UNSPEC,
net/sched/cls_flower.c
1887
sizeof(key->meta.l2_miss));
net/sched/cls_flower.c
2172
FLOW_DISSECTOR_KEY_META, meta);
net/sched/cls_flower.c
2332
return mask->meta.l2_miss;
net/sched/cls_flower.c
3442
if (mask->meta.ingress_ifindex) {
net/sched/cls_flower.c
3445
dev = __dev_get_by_index(net, key->meta.ingress_ifindex);
net/sched/cls_flower.c
3450
if (fl_dump_key_val(skb, &key->meta.l2_miss,
net/sched/cls_flower.c
3451
TCA_FLOWER_L2_MISS, &mask->meta.l2_miss,
net/sched/cls_flower.c
3452
TCA_FLOWER_UNSPEC, sizeof(key->meta.l2_miss)))
net/sched/cls_flower.c
55
struct flow_dissector_key_meta meta;
net/sched/em_meta.c
850
struct meta_match *meta = (struct meta_match *) m->data;
net/sched/em_meta.c
853
if (meta_get(skb, info, &meta->lvalue, &l_value) < 0 ||
net/sched/em_meta.c
854
meta_get(skb, info, &meta->rvalue, &r_value) < 0)
net/sched/em_meta.c
857
r = meta_type_ops(&meta->lvalue)->compare(&l_value, &r_value);
net/sched/em_meta.c
859
switch (meta->lvalue.hdr.op) {
net/sched/em_meta.c
871
static void meta_delete(struct meta_match *meta)
net/sched/em_meta.c
873
if (meta) {
net/sched/em_meta.c
874
const struct meta_type_ops *ops = meta_type_ops(&meta->lvalue);
net/sched/em_meta.c
877
ops->destroy(&meta->lvalue);
net/sched/em_meta.c
878
ops->destroy(&meta->rvalue);
net/sched/em_meta.c
882
kfree(meta);
net/sched/em_meta.c
912
struct meta_match *meta = NULL;
net/sched/em_meta.c
930
meta = kzalloc_obj(*meta);
net/sched/em_meta.c
931
if (meta == NULL) {
net/sched/em_meta.c
936
memcpy(&meta->lvalue.hdr, &hdr->left, sizeof(hdr->left));
net/sched/em_meta.c
937
memcpy(&meta->rvalue.hdr, &hdr->right, sizeof(hdr->right));
net/sched/em_meta.c
939
if (!meta_is_supported(&meta->lvalue) ||
net/sched/em_meta.c
940
!meta_is_supported(&meta->rvalue)) {
net/sched/em_meta.c
945
if (meta_change_data(&meta->lvalue, tb[TCA_EM_META_LVALUE]) < 0 ||
net/sched/em_meta.c
946
meta_change_data(&meta->rvalue, tb[TCA_EM_META_RVALUE]) < 0)
net/sched/em_meta.c
949
m->datalen = sizeof(*meta);
net/sched/em_meta.c
950
m->data = (unsigned long) meta;
net/sched/em_meta.c
954
if (err && meta)
net/sched/em_meta.c
955
meta_delete(meta);
net/sched/em_meta.c
967
struct meta_match *meta = (struct meta_match *) em->data;
net/sched/em_meta.c
972
memcpy(&hdr.left, &meta->lvalue.hdr, sizeof(hdr.left));
net/sched/em_meta.c
973
memcpy(&hdr.right, &meta->rvalue.hdr, sizeof(hdr.right));
net/sched/em_meta.c
978
ops = meta_type_ops(&meta->lvalue);
net/sched/em_meta.c
979
if (ops->dump(skb, &meta->lvalue, TCA_EM_META_LVALUE) < 0 ||
net/sched/em_meta.c
980
ops->dump(skb, &meta->rvalue, TCA_EM_META_RVALUE) < 0)
net/unix/af_unix.c
3595
__bpf_md_ptr(struct bpf_iter_meta *, meta);
net/unix/af_unix.c
3600
static int unix_prog_seq_show(struct bpf_prog *prog, struct bpf_iter_meta *meta,
net/unix/af_unix.c
3605
meta->seq_num--; /* skip SEQ_START_TOKEN */
net/unix/af_unix.c
3606
ctx.meta = meta;
net/unix/af_unix.c
3730
struct bpf_iter_meta meta;
net/unix/af_unix.c
3748
meta.seq = seq;
net/unix/af_unix.c
3749
prog = bpf_iter_get_info(&meta, false);
net/unix/af_unix.c
3750
ret = unix_prog_seq_show(prog, &meta, v, uid);
net/unix/af_unix.c
3759
struct bpf_iter_meta meta;
net/unix/af_unix.c
3763
meta.seq = seq;
net/unix/af_unix.c
3764
prog = bpf_iter_get_info(&meta, true);
net/unix/af_unix.c
3766
(void)unix_prog_seq_show(prog, &meta, v, 0);
net/unix/af_unix.c
3846
DEFINE_BPF_ITER_FUNC(unix, struct bpf_iter_meta *meta,
net/xdp/xsk.c
685
struct xsk_tx_metadata *meta = NULL;
net/xdp/xsk.c
690
meta = buffer - pool->tx_metadata_len;
net/xdp/xsk.c
691
if (unlikely(!xsk_buff_valid_tx_metadata(meta)))
net/xdp/xsk.c
694
if (meta->flags & XDP_TXMD_FLAGS_CHECKSUM) {
net/xdp/xsk.c
695
if (unlikely(meta->request.csum_start +
net/xdp/xsk.c
696
meta->request.csum_offset +
net/xdp/xsk.c
700
skb->csum_start = hr + meta->request.csum_start;
net/xdp/xsk.c
701
skb->csum_offset = meta->request.csum_offset;
net/xdp/xsk.c
713
if (meta->flags & XDP_TXMD_FLAGS_LAUNCH_TIME)
net/xdp/xsk.c
714
skb->skb_mstamp_ns = meta->request.launch_time;
net/xdp/xsk.c
715
xsk_tx_metadata_to_compl(meta, &skb_shinfo(skb)->xsk_meta);
net/xdp/xsk_buff_pool.c
773
ret.meta = __xsk_buff_get_metadata(pool, __xp_raw_get_data(pool, addr));
samples/bpf/xdp2skb_meta_kern.c
34
struct meta_info *meta;
samples/bpf/xdp2skb_meta_kern.c
41
ret = bpf_xdp_adjust_meta(ctx, -(int)sizeof(*meta));
samples/bpf/xdp2skb_meta_kern.c
53
meta = (void *)(unsigned long)ctx->data_meta;
samples/bpf/xdp2skb_meta_kern.c
54
if (meta + 1 > data)
samples/bpf/xdp2skb_meta_kern.c
57
meta->mark = 42;
samples/bpf/xdp2skb_meta_kern.c
67
struct meta_info *meta = data_meta;
samples/bpf/xdp2skb_meta_kern.c
70
if (meta + 1 > data) {
samples/bpf/xdp2skb_meta_kern.c
77
ctx->mark = meta->mark; /* Transfer XDP-mark to SKB-mark */
tools/bpf/bpftool/skeleton/pid_iter.bpf.c
119
bpf_seq_write(ctx->meta->seq, &e, sizeof(e));
tools/include/uapi/linux/fanotify.h
266
#define FAN_EVENT_NEXT(meta, len) ((len) -= (meta)->event_len, \
tools/include/uapi/linux/fanotify.h
267
(struct fanotify_event_metadata*)(((char *)(meta)) + \
tools/include/uapi/linux/fanotify.h
268
(meta)->event_len))
tools/include/uapi/linux/fanotify.h
270
#define FAN_EVENT_OK(meta, len) ((long)(len) >= (long)FAN_EVENT_METADATA_LEN && \
tools/include/uapi/linux/fanotify.h
271
(long)(meta)->event_len >= (long)FAN_EVENT_METADATA_LEN && \
tools/include/uapi/linux/fanotify.h
272
(long)(meta)->event_len <= (long)(len))
tools/sched_ext/include/scx/common.bpf.h
302
void *bpf_obj_new_impl(__u64 local_type_id, void *meta) __ksym;
tools/sched_ext/include/scx/common.bpf.h
303
void bpf_obj_drop_impl(void *kptr, void *meta) __ksym;
tools/sched_ext/include/scx/common.bpf.h
310
void *meta, __u64 off) __ksym;
tools/sched_ext/include/scx/common.bpf.h
315
void *meta, __u64 off) __ksym;
tools/sched_ext/include/scx/common.bpf.h
324
void *meta, __u64 off) __ksym;
tools/sched_ext/include/scx/common.bpf.h
329
void *bpf_refcount_acquire_impl(void *kptr, void *meta) __ksym;
tools/testing/selftests/bpf/bpf_experimental.h
121
void *meta, __u64 off) __ksym;
tools/testing/selftests/bpf/bpf_experimental.h
146
extern void *bpf_percpu_obj_new_impl(__u64 local_type_id, void *meta) __ksym;
tools/testing/selftests/bpf/bpf_experimental.h
160
extern void bpf_percpu_obj_drop_impl(void *kptr, void *meta) __ksym;
tools/testing/selftests/bpf/bpf_experimental.h
23
extern void *bpf_obj_new_impl(__u64 local_type_id, void *meta) __ksym;
tools/testing/selftests/bpf/bpf_experimental.h
37
extern void bpf_obj_drop_impl(void *kptr, void *meta) __ksym;
tools/testing/selftests/bpf/bpf_experimental.h
51
extern void *bpf_refcount_acquire_impl(void *kptr, void *meta) __ksym;
tools/testing/selftests/bpf/bpf_experimental.h
67
void *meta, __u64 off) __ksym;
tools/testing/selftests/bpf/bpf_experimental.h
83
void *meta, __u64 off) __ksym;
tools/testing/selftests/bpf/prog_tests/kfree_skb.c
22
struct meta *meta = (struct meta *)data;
tools/testing/selftests/bpf/prog_tests/kfree_skb.c
23
struct ipv6_packet *pkt_v6 = data + sizeof(*meta);
tools/testing/selftests/bpf/prog_tests/kfree_skb.c
26
if (CHECK(size != 72 + sizeof(*meta), "check_size", "size %u != %zu\n",
tools/testing/selftests/bpf/prog_tests/kfree_skb.c
27
size, 72 + sizeof(*meta)))
tools/testing/selftests/bpf/prog_tests/kfree_skb.c
29
if (CHECK(meta->ifindex != 1, "check_meta_ifindex",
tools/testing/selftests/bpf/prog_tests/kfree_skb.c
30
"meta->ifindex = %d\n", meta->ifindex))
tools/testing/selftests/bpf/prog_tests/kfree_skb.c
33
if (CHECK(meta->cb8_0 != cb.cb8[0], "check_cb8_0", "cb8_0 %x != %x\n",
tools/testing/selftests/bpf/prog_tests/kfree_skb.c
34
meta->cb8_0, cb.cb8[0]))
tools/testing/selftests/bpf/prog_tests/kfree_skb.c
36
if (CHECK(meta->cb32_0 != cb.cb32[0], "check_cb32_0",
tools/testing/selftests/bpf/prog_tests/kfree_skb.c
38
meta->cb32_0, cb.cb32[0]))
tools/testing/selftests/bpf/prog_tests/task_local_data.h
109
struct tld_meta_u *meta;
tools/testing/selftests/bpf/prog_tests/task_local_data.h
129
struct tld_meta_u *meta, *uninit = NULL;
tools/testing/selftests/bpf/prog_tests/task_local_data.h
132
meta = (struct tld_meta_u *)aligned_alloc(TLD_PAGE_SIZE, TLD_PAGE_SIZE);
tools/testing/selftests/bpf/prog_tests/task_local_data.h
133
if (!meta) {
tools/testing/selftests/bpf/prog_tests/task_local_data.h
138
memset(meta, 0, TLD_PAGE_SIZE);
tools/testing/selftests/bpf/prog_tests/task_local_data.h
139
meta->size = TLD_DYN_DATA_SIZE;
tools/testing/selftests/bpf/prog_tests/task_local_data.h
141
if (!atomic_compare_exchange_strong(&tld_meta_p, &uninit, meta)) {
tools/testing/selftests/bpf/prog_tests/task_local_data.h
142
free(meta);
tools/testing/selftests/bpf/prog_tests/task_local_data.h
198
map_val.meta = TLD_READ_ONCE(tld_meta_p);
tools/testing/selftests/bpf/prog_tests/test_xsk.c
769
struct xdp_info *meta = data - sizeof(struct xdp_info);
tools/testing/selftests/bpf/prog_tests/test_xsk.c
771
if (meta->count != pkt->pkt_nb) {
tools/testing/selftests/bpf/prog_tests/test_xsk.c
774
(unsigned long long)meta->count);
tools/testing/selftests/bpf/prog_tests/xdp_bpf2bpf.c
22
struct meta *meta = (struct meta *)data;
tools/testing/selftests/bpf/prog_tests/xdp_bpf2bpf.c
23
struct ipv4_packet *trace_pkt_v4 = data + sizeof(*meta);
tools/testing/selftests/bpf/prog_tests/xdp_bpf2bpf.c
24
unsigned char *raw_pkt = data + sizeof(*meta);
tools/testing/selftests/bpf/prog_tests/xdp_bpf2bpf.c
27
ASSERT_GE(size, sizeof(pkt_v4) + sizeof(*meta), "check_size");
tools/testing/selftests/bpf/prog_tests/xdp_bpf2bpf.c
28
ASSERT_EQ(meta->ifindex, if_nametoindex("lo"), "check_meta_ifindex");
tools/testing/selftests/bpf/prog_tests/xdp_bpf2bpf.c
29
ASSERT_EQ(meta->pkt_len, tst_ctx->pkt_size, "check_meta_pkt_len");
tools/testing/selftests/bpf/prog_tests/xdp_bpf2bpf.c
33
if (meta->pkt_len > sizeof(pkt_v4)) {
tools/testing/selftests/bpf/prog_tests/xdp_bpf2bpf.c
34
for (int i = 0; i < meta->pkt_len - sizeof(pkt_v4); i++)
tools/testing/selftests/bpf/prog_tests/xdp_metadata.c
138
struct xsk_tx_metadata *meta;
tools/testing/selftests/bpf/prog_tests/xdp_metadata.c
156
meta = data - sizeof(struct xsk_tx_metadata);
tools/testing/selftests/bpf/prog_tests/xdp_metadata.c
157
memset(meta, 0, sizeof(*meta));
tools/testing/selftests/bpf/prog_tests/xdp_metadata.c
158
meta->flags = XDP_TXMD_FLAGS_TIMESTAMP;
tools/testing/selftests/bpf/prog_tests/xdp_metadata.c
188
meta->flags |= XDP_TXMD_FLAGS_CHECKSUM;
tools/testing/selftests/bpf/prog_tests/xdp_metadata.c
189
meta->request.csum_start = sizeof(*eth) + sizeof(*iph);
tools/testing/selftests/bpf/prog_tests/xdp_metadata.c
190
meta->request.csum_offset = offsetof(struct udphdr, check);
tools/testing/selftests/bpf/prog_tests/xdp_metadata.c
229
struct xsk_tx_metadata *meta;
tools/testing/selftests/bpf/prog_tests/xdp_metadata.c
240
meta = data - sizeof(struct xsk_tx_metadata);
tools/testing/selftests/bpf/prog_tests/xdp_metadata.c
242
ASSERT_NEQ(meta->completion.tx_timestamp, 0, "tx_timestamp");
tools/testing/selftests/bpf/prog_tests/xdp_metadata.c
263
struct xdp_meta *meta;
tools/testing/selftests/bpf/prog_tests/xdp_metadata.c
305
meta = data - sizeof(struct xdp_meta);
tools/testing/selftests/bpf/prog_tests/xdp_metadata.c
307
if (!ASSERT_NEQ(meta->rx_timestamp, 0, "rx_timestamp"))
tools/testing/selftests/bpf/prog_tests/xdp_metadata.c
310
if (!ASSERT_NEQ(meta->rx_hash, 0, "rx_hash"))
tools/testing/selftests/bpf/prog_tests/xdp_metadata.c
314
if (!ASSERT_NEQ(meta->rx_hash_type & XDP_RSS_TYPE_L4, 0, "rx_hash_type"))
tools/testing/selftests/bpf/prog_tests/xdp_metadata.c
317
if (!ASSERT_EQ(meta->rx_vlan_tci & VLAN_VID_MASK, VLAN_ID, "rx_vlan_tci"))
tools/testing/selftests/bpf/prog_tests/xdp_metadata.c
320
if (!ASSERT_EQ(meta->rx_vlan_proto, VLAN_PID, "rx_vlan_proto"))
tools/testing/selftests/bpf/prog_tests/xdp_metadata.c
325
ASSERT_EQ(meta->rx_hash_type, 0, "rx_hash_type");
tools/testing/selftests/bpf/progs/bpf_iter_bpf_array_map.c
35
bpf_seq_write(ctx->meta->seq, key, sizeof(__u32));
tools/testing/selftests/bpf/progs/bpf_iter_bpf_array_map.c
36
bpf_seq_write(ctx->meta->seq, val, sizeof(__u64));
tools/testing/selftests/bpf/progs/bpf_iter_bpf_hash_map.c
119
if (ctx->meta->seq_num == 0)
tools/testing/selftests/bpf/progs/bpf_iter_bpf_hash_map.c
120
BPF_SEQ_PRINTF(ctx->meta->seq, "map dump starts\n");
tools/testing/selftests/bpf/progs/bpf_iter_bpf_hash_map.c
45
struct seq_file *seq = ctx->meta->seq;
tools/testing/selftests/bpf/progs/bpf_iter_bpf_hash_map.c
46
__u32 seq_num = ctx->meta->seq_num;
tools/testing/selftests/bpf/progs/bpf_iter_bpf_link.c
11
struct seq_file *seq = ctx->meta->seq;
tools/testing/selftests/bpf/progs/bpf_iter_bpf_map.c
11
struct seq_file *seq = ctx->meta->seq;
tools/testing/selftests/bpf/progs/bpf_iter_bpf_map.c
12
__u64 seq_num = ctx->meta->seq_num;
tools/testing/selftests/bpf/progs/bpf_iter_ipv6_route.c
14
struct seq_file *seq = ctx->meta->seq;
tools/testing/selftests/bpf/progs/bpf_iter_ksym.c
31
struct seq_file *seq = ctx->meta->seq;
tools/testing/selftests/bpf/progs/bpf_iter_ksym.c
33
__u32 seq_num = ctx->meta->seq_num;
tools/testing/selftests/bpf/progs/bpf_iter_netlink.c
17
struct seq_file *seq = ctx->meta->seq;
tools/testing/selftests/bpf/progs/bpf_iter_netlink.c
27
if (ctx->meta->seq_num == 0)
tools/testing/selftests/bpf/progs/bpf_iter_task_btf.c
18
struct seq_file *seq = ctx->meta->seq;
tools/testing/selftests/bpf/progs/bpf_iter_task_btf.c
27
if (ctx->meta->seq_num == 0)
tools/testing/selftests/bpf/progs/bpf_iter_task_file.c
16
struct seq_file *seq = ctx->meta->seq;
tools/testing/selftests/bpf/progs/bpf_iter_task_file.c
24
if (ctx->meta->seq_num == 0) {
tools/testing/selftests/bpf/progs/bpf_iter_task_stack.c
15
struct seq_file *seq = ctx->meta->seq;
tools/testing/selftests/bpf/progs/bpf_iter_task_stack.c
43
struct seq_file *seq = ctx->meta->seq;
tools/testing/selftests/bpf/progs/bpf_iter_task_vmas.c
29
struct seq_file *seq = ctx->meta->seq;
tools/testing/selftests/bpf/progs/bpf_iter_tasks.c
193
if (ctx->meta->seq_num == 0)
tools/testing/selftests/bpf/progs/bpf_iter_tasks.c
23
struct seq_file *seq = ctx->meta->seq;
tools/testing/selftests/bpf/progs/bpf_iter_tasks.c
37
if (ctx->meta->seq_num == 0)
tools/testing/selftests/bpf/progs/bpf_iter_tasks.c
52
struct seq_file *seq = ctx->meta->seq;
tools/testing/selftests/bpf/progs/bpf_iter_tcp4.c
199
struct seq_file *seq = ctx->meta->seq;
tools/testing/selftests/bpf/progs/bpf_iter_tcp4.c
209
seq_num = ctx->meta->seq_num;
tools/testing/selftests/bpf/progs/bpf_iter_tcp6.c
215
struct seq_file *seq = ctx->meta->seq;
tools/testing/selftests/bpf/progs/bpf_iter_tcp6.c
225
seq_num = ctx->meta->seq_num;
tools/testing/selftests/bpf/progs/bpf_iter_test_kern3.c
11
struct seq_file *seq = ctx->meta->seq;
tools/testing/selftests/bpf/progs/bpf_iter_test_kern4.c
18
struct seq_file *seq = ctx->meta->seq;
tools/testing/selftests/bpf/progs/bpf_iter_test_kern4.c
30
seq_num = ctx->meta->seq_num;
tools/testing/selftests/bpf/progs/bpf_iter_test_kern_common.h
12
struct seq_file *seq = ctx->meta->seq;
tools/testing/selftests/bpf/progs/bpf_iter_udp4.c
27
struct seq_file *seq = ctx->meta->seq;
tools/testing/selftests/bpf/progs/bpf_iter_udp4.c
38
seq_num = ctx->meta->seq_num;
tools/testing/selftests/bpf/progs/bpf_iter_udp6.c
34
struct seq_file *seq = ctx->meta->seq;
tools/testing/selftests/bpf/progs/bpf_iter_udp6.c
46
seq_num = ctx->meta->seq_num;
tools/testing/selftests/bpf/progs/bpf_iter_unix.c
35
seq = ctx->meta->seq;
tools/testing/selftests/bpf/progs/bpf_iter_unix.c
36
seq_num = ctx->meta->seq_num;
tools/testing/selftests/bpf/progs/bpf_iter_vma_offset.c
19
struct seq_file *seq = ctx->meta->seq;
tools/testing/selftests/bpf/progs/cgroup_hierarchical_stats.c
134
int BPF_PROG(dumper, struct bpf_iter_meta *meta, struct cgroup *cgrp)
tools/testing/selftests/bpf/progs/cgroup_hierarchical_stats.c
136
struct seq_file *seq = meta->seq;
tools/testing/selftests/bpf/progs/cgroup_iter.c
19
struct seq_file *seq = ctx->meta->seq;
tools/testing/selftests/bpf/progs/cgroup_iter.c
29
if (ctx->meta->seq_num == 0)
tools/testing/selftests/bpf/progs/dmabuf_iter.c
34
struct seq_file *seq = ctx->meta->seq;
tools/testing/selftests/bpf/progs/dynptr_fail.c
1117
struct bpf_dynptr meta;
tools/testing/selftests/bpf/progs/dynptr_fail.c
1120
bpf_dynptr_from_skb_meta(skb, 0, &meta);
tools/testing/selftests/bpf/progs/dynptr_fail.c
1122
md = bpf_dynptr_slice(&meta, 0, NULL, sizeof(*md));
tools/testing/selftests/bpf/progs/dynptr_fail.c
1240
struct bpf_dynptr data, meta;
tools/testing/selftests/bpf/progs/dynptr_fail.c
1244
bpf_dynptr_from_skb_meta(skb, 0, &meta);
tools/testing/selftests/bpf/progs/dynptr_fail.c
1250
bpf_dynptr_write(&meta, 0, "x", 1, 0);
tools/testing/selftests/bpf/progs/dynptr_fail.c
1263
struct bpf_dynptr data, meta;
tools/testing/selftests/bpf/progs/dynptr_fail.c
1267
bpf_dynptr_from_skb_meta(skb, 0, &meta);
tools/testing/selftests/bpf/progs/dynptr_fail.c
1273
bpf_dynptr_write(&meta, 0, "x", 1, 0);
tools/testing/selftests/bpf/progs/dynptr_fail.c
1286
struct bpf_dynptr data, meta;
tools/testing/selftests/bpf/progs/dynptr_fail.c
1290
bpf_dynptr_from_skb_meta(skb, 0, &meta);
tools/testing/selftests/bpf/progs/dynptr_fail.c
1292
md = bpf_dynptr_slice(&meta, 0, NULL, sizeof(*md));
tools/testing/selftests/bpf/progs/dynptr_fail.c
1309
struct bpf_dynptr data, meta;
tools/testing/selftests/bpf/progs/dynptr_fail.c
1313
bpf_dynptr_from_skb_meta(skb, 0, &meta);
tools/testing/selftests/bpf/progs/dynptr_fail.c
1315
md = bpf_dynptr_slice_rdwr(&meta, 0, NULL, sizeof(*md));
tools/testing/selftests/bpf/progs/dynptr_fail.c
1332
struct bpf_dynptr meta;
tools/testing/selftests/bpf/progs/dynptr_fail.c
1335
bpf_dynptr_from_skb_meta(skb, 0, &meta);
tools/testing/selftests/bpf/progs/dynptr_fail.c
1337
md = bpf_dynptr_slice(&meta, 0, NULL, sizeof(*md));
tools/testing/selftests/bpf/progs/dynptr_fail.c
1355
struct bpf_dynptr meta;
tools/testing/selftests/bpf/progs/dynptr_fail.c
1358
bpf_dynptr_from_skb_meta(skb, 0, &meta);
tools/testing/selftests/bpf/progs/dynptr_fail.c
1360
md = bpf_dynptr_slice_rdwr(&meta, 0, NULL, sizeof(*md));
tools/testing/selftests/bpf/progs/dynptr_fail.c
1378
struct bpf_dynptr meta;
tools/testing/selftests/bpf/progs/dynptr_fail.c
1381
bpf_dynptr_from_skb_meta(skb, 0, &meta);
tools/testing/selftests/bpf/progs/dynptr_fail.c
1383
md = bpf_dynptr_slice(&meta, 0, NULL, sizeof(*md));
tools/testing/selftests/bpf/progs/dynptr_fail.c
1387
bpf_dynptr_write(&meta, 0, "x", 1, 0);
tools/testing/selftests/bpf/progs/dynptr_fail.c
1400
struct bpf_dynptr meta;
tools/testing/selftests/bpf/progs/dynptr_fail.c
1403
bpf_dynptr_from_skb_meta(skb, 0, &meta);
tools/testing/selftests/bpf/progs/dynptr_fail.c
1405
md = bpf_dynptr_slice_rdwr(&meta, 0, NULL, sizeof(*md));
tools/testing/selftests/bpf/progs/dynptr_fail.c
1409
bpf_dynptr_write(&meta, 0, "x", 1, 0);
tools/testing/selftests/bpf/progs/dynptr_fail.c
1485
struct bpf_dynptr meta;
tools/testing/selftests/bpf/progs/dynptr_fail.c
1488
bpf_dynptr_from_skb_meta(ctx, 0, &meta);
tools/testing/selftests/bpf/progs/dynptr_fail.c
1908
struct bpf_dynptr clone, meta;
tools/testing/selftests/bpf/progs/dynptr_fail.c
1911
bpf_dynptr_from_skb_meta(skb, 0, &meta);
tools/testing/selftests/bpf/progs/dynptr_fail.c
1912
bpf_dynptr_clone(&meta, &clone);
tools/testing/selftests/bpf/progs/dynptr_fail.c
277
struct bpf_dynptr meta;
tools/testing/selftests/bpf/progs/dynptr_fail.c
280
bpf_dynptr_from_skb_meta(skb, 0, &meta);
tools/testing/selftests/bpf/progs/dynptr_fail.c
282
md = bpf_dynptr_slice_rdwr(&meta, 0, NULL, sizeof(*md));
tools/testing/selftests/bpf/progs/dynptr_success.c
217
struct bpf_dynptr meta;
tools/testing/selftests/bpf/progs/dynptr_success.c
222
ret = bpf_dynptr_from_skb_meta(skb, 0, &meta);
tools/testing/selftests/bpf/progs/dynptr_success.c
228
md = bpf_dynptr_data(&meta, 0, sizeof(*md));
tools/testing/selftests/bpf/progs/dynptr_success.c
241
struct bpf_dynptr meta;
tools/testing/selftests/bpf/progs/dynptr_success.c
246
ret = bpf_dynptr_from_skb_meta(skb, INVALID_FLAGS, &meta);
tools/testing/selftests/bpf/progs/dynptr_success.c
251
ret = bpf_dynptr_from_skb_meta(skb, 0, &meta);
tools/testing/selftests/bpf/progs/dynptr_success.c
256
ret = bpf_dynptr_read(&buf, 0, &meta, 0, INVALID_FLAGS);
tools/testing/selftests/bpf/progs/dynptr_success.c
261
ret = bpf_dynptr_write(&meta, 0, &buf, 0, INVALID_FLAGS);
tools/testing/selftests/bpf/progs/iters_css_task.c
58
struct seq_file *seq = ctx->meta->seq;
tools/testing/selftests/bpf/progs/iters_css_task.c
70
if (ctx->meta->seq_num == 0)
tools/testing/selftests/bpf/progs/kfree_skb.c
100
bpf_printk("cb8_0:%x cb32_0:%x\n", meta.cb8_0, meta.cb32_0);
tools/testing/selftests/bpf/progs/kfree_skb.c
102
if (users != 1 || pkt_data != bpf_htons(0x86dd) || meta.ifindex != 1)
tools/testing/selftests/bpf/progs/kfree_skb.c
108
&meta, sizeof(meta));
tools/testing/selftests/bpf/progs/kfree_skb.c
69
struct meta meta = {};
tools/testing/selftests/bpf/progs/kfree_skb.c
84
meta.ifindex = _(dev->ifindex);
tools/testing/selftests/bpf/progs/kfree_skb.c
85
meta.cb8_0 = cb8[8];
tools/testing/selftests/bpf/progs/kfree_skb.c
86
meta.cb32_0 = cb32[2];
tools/testing/selftests/bpf/progs/kfree_skb.c
99
meta.ifindex, data, pkt_data);
tools/testing/selftests/bpf/progs/kmem_cache_iter.c
41
struct seq_file *seq = ctx->meta->seq;
tools/testing/selftests/bpf/progs/map_percpu_stats.c
15
struct seq_file *seq = ctx->meta->seq;
tools/testing/selftests/bpf/progs/profiler.h
110
struct var_metadata_t meta;
tools/testing/selftests/bpf/progs/profiler.h
123
struct var_metadata_t meta;
tools/testing/selftests/bpf/progs/profiler.h
131
struct var_metadata_t meta;
tools/testing/selftests/bpf/progs/profiler.h
87
struct var_metadata_t meta;
tools/testing/selftests/bpf/progs/profiler.h
96
struct var_metadata_t meta;
tools/testing/selftests/bpf/progs/profiler.inc.h
173
if (arr_struct->array[i].meta.pid == spid)
tools/testing/selftests/bpf/progs/profiler.inc.h
362
void* payload = populate_var_metadata(&kill_data->meta, task, spid, kill_data->payload);
tools/testing/selftests/bpf/progs/profiler.inc.h
367
kill_data->meta.type = KILL_EVENT;
tools/testing/selftests/bpf/progs/profiler.inc.h
406
if (arr_struct->array[i].meta.pid == 0) {
tools/testing/selftests/bpf/progs/profiler.inc.h
464
struct var_metadata_t* meta)
tools/testing/selftests/bpf/progs/profiler.inc.h
468
meta->bpf_stats_num_perf_events =
tools/testing/selftests/bpf/progs/profiler.inc.h
471
meta->bpf_stats_start_ktime_ns = bpf_stat_ctx->start_time_ns;
tools/testing/selftests/bpf/progs/profiler.inc.h
472
meta->cpu_id = bpf_get_smp_processor_id();
tools/testing/selftests/bpf/progs/profiler.inc.h
562
sysctl_data->meta.type = SYSCTL_EVENT;
tools/testing/selftests/bpf/progs/profiler.inc.h
563
void* payload = populate_var_metadata(&sysctl_data->meta, task, pid, sysctl_data->payload);
tools/testing/selftests/bpf/progs/profiler.inc.h
587
bpf_stats_pre_submit_var_perf_event(&stats_ctx, &sysctl_data->meta);
tools/testing/selftests/bpf/progs/profiler.inc.h
662
bpf_stats_pre_submit_var_perf_event(&stats_ctx, &kill_data->meta);
tools/testing/selftests/bpf/progs/profiler.inc.h
700
proc_exec_data->meta.type = EXEC_EVENT;
tools/testing/selftests/bpf/progs/profiler.inc.h
704
void* payload = populate_var_metadata(&proc_exec_data->meta, task, pid,
tools/testing/selftests/bpf/progs/profiler.inc.h
743
bpf_stats_pre_submit_var_perf_event(&stats_ctx, &proc_exec_data->meta);
tools/testing/selftests/bpf/progs/profiler.inc.h
789
filemod_data->meta.type = FILEMOD_EVENT;
tools/testing/selftests/bpf/progs/profiler.inc.h
799
void* payload = populate_var_metadata(&filemod_data->meta, task, pid,
tools/testing/selftests/bpf/progs/profiler.inc.h
808
bpf_stats_pre_submit_var_perf_event(&stats_ctx, &filemod_data->meta);
tools/testing/selftests/bpf/progs/profiler.inc.h
842
filemod_data->meta.type = FILEMOD_EVENT;
tools/testing/selftests/bpf/progs/profiler.inc.h
852
void* payload = populate_var_metadata(&filemod_data->meta, task, pid,
tools/testing/selftests/bpf/progs/profiler.inc.h
868
bpf_stats_pre_submit_var_perf_event(&stats_ctx, &filemod_data->meta);
tools/testing/selftests/bpf/progs/profiler.inc.h
897
filemod_data->meta.type = FILEMOD_EVENT;
tools/testing/selftests/bpf/progs/profiler.inc.h
907
void* payload = populate_var_metadata(&filemod_data->meta, task, pid,
tools/testing/selftests/bpf/progs/profiler.inc.h
922
bpf_stats_pre_submit_var_perf_event(&stats_ctx, &filemod_data->meta);
tools/testing/selftests/bpf/progs/profiler.inc.h
944
fork_data->meta.type = FORK_EVENT;
tools/testing/selftests/bpf/progs/profiler.inc.h
946
void* payload = populate_var_metadata(&fork_data->meta, child,
tools/testing/selftests/bpf/progs/profiler.inc.h
951
bpf_stats_pre_submit_var_perf_event(&stats_ctx, &fork_data->meta);
tools/testing/selftests/bpf/progs/sock_iter_batch.c
127
bpf_seq_write(ctx->meta->seq, &idx, sizeof(idx));
tools/testing/selftests/bpf/progs/sock_iter_batch.c
128
bpf_seq_write(ctx->meta->seq, &sock_cookie, sizeof(sock_cookie));
tools/testing/selftests/bpf/progs/sock_iter_batch.c
67
bpf_seq_write(ctx->meta->seq, &idx, sizeof(idx));
tools/testing/selftests/bpf/progs/sock_iter_batch.c
68
bpf_seq_write(ctx->meta->seq, &sock_cookie, sizeof(sock_cookie));
tools/testing/selftests/bpf/progs/sock_iter_batch.c
89
bpf_seq_write(ctx->meta->seq, &sock_cookie, sizeof(sock_cookie));
tools/testing/selftests/bpf/progs/task_local_data.bpf.h
100
struct tld_meta_u meta[0];
tools/testing/selftests/bpf/progs/task_local_data.bpf.h
176
if (!tld_obj->data_map || !tld_obj->data_map->data || !tld_obj->data_map->meta)
tools/testing/selftests/bpf/progs/task_local_data.bpf.h
180
cnt = tld_obj->data_map->meta->cnt;
tools/testing/selftests/bpf/progs/task_local_data.bpf.h
181
metadata = tld_obj->data_map->meta->metadata;
tools/testing/selftests/bpf/progs/task_local_data.bpf.h
221
if (likely((tld_obj)->data_map->meta) && \
tools/testing/selftests/bpf/progs/task_local_data.bpf.h
222
cnt < (tld_obj)->data_map->meta->cnt) { \
tools/testing/selftests/bpf/progs/task_local_data.bpf.h
95
struct tld_meta_u __uptr *meta;
tools/testing/selftests/bpf/progs/test_xdp_bpf2bpf.c
47
struct meta meta;
tools/testing/selftests/bpf/progs/test_xdp_bpf2bpf.c
49
meta.ifindex = xdp->rxq->dev->ifindex;
tools/testing/selftests/bpf/progs/test_xdp_bpf2bpf.c
50
meta.pkt_len = bpf_xdp_get_buff_len((struct xdp_md *)xdp);
tools/testing/selftests/bpf/progs/test_xdp_bpf2bpf.c
52
((__u64) meta.pkt_len << 32) |
tools/testing/selftests/bpf/progs/test_xdp_bpf2bpf.c
54
&meta, sizeof(meta));
tools/testing/selftests/bpf/progs/test_xdp_meta.c
106
struct bpf_dynptr data, meta;
tools/testing/selftests/bpf/progs/test_xdp_meta.c
114
bpf_dynptr_from_skb_meta(ctx, 0, &meta);
tools/testing/selftests/bpf/progs/test_xdp_meta.c
115
bpf_dynptr_write(&meta, 0, src, META_SIZE, 0);
tools/testing/selftests/bpf/progs/test_xdp_meta.c
124
struct bpf_dynptr meta;
tools/testing/selftests/bpf/progs/test_xdp_meta.c
127
bpf_dynptr_from_skb_meta(ctx, 0, &meta);
tools/testing/selftests/bpf/progs/test_xdp_meta.c
128
meta_have = bpf_dynptr_slice(&meta, 0, NULL, META_SIZE);
tools/testing/selftests/bpf/progs/test_xdp_meta.c
144
struct bpf_dynptr data, meta;
tools/testing/selftests/bpf/progs/test_xdp_meta.c
152
bpf_dynptr_from_skb_meta(ctx, 0, &meta);
tools/testing/selftests/bpf/progs/test_xdp_meta.c
153
dst = bpf_dynptr_slice_rdwr(&meta, 0, NULL, META_SIZE);
tools/testing/selftests/bpf/progs/test_xdp_meta.c
168
struct bpf_dynptr meta;
tools/testing/selftests/bpf/progs/test_xdp_meta.c
174
bpf_dynptr_from_skb_meta(ctx, 0, &meta);
tools/testing/selftests/bpf/progs/test_xdp_meta.c
175
bpf_dynptr_read(dst, chunk_len, &meta, 0, 0);
tools/testing/selftests/bpf/progs/test_xdp_meta.c
179
bpf_dynptr_adjust(&meta, chunk_len, bpf_dynptr_size(&meta));
tools/testing/selftests/bpf/progs/test_xdp_meta.c
180
bpf_dynptr_read(dst, chunk_len, &meta, 0, 0);
tools/testing/selftests/bpf/progs/test_xdp_meta.c
184
bpf_dynptr_read(dst, chunk_len, &meta, chunk_len, 0);
tools/testing/selftests/bpf/progs/test_xdp_meta.c
188
src = bpf_dynptr_slice(&meta, 2 * chunk_len, NULL, chunk_len);
tools/testing/selftests/bpf/progs/test_xdp_meta.c
207
struct bpf_dynptr meta;
tools/testing/selftests/bpf/progs/test_xdp_meta.c
214
bpf_dynptr_from_skb_meta(ctx, 0, &meta);
tools/testing/selftests/bpf/progs/test_xdp_meta.c
215
bpf_dynptr_write(&meta, 0, src, chunk_len, 0);
tools/testing/selftests/bpf/progs/test_xdp_meta.c
219
bpf_dynptr_adjust(&meta, chunk_len, bpf_dynptr_size(&meta));
tools/testing/selftests/bpf/progs/test_xdp_meta.c
220
bpf_dynptr_write(&meta, 0, src, chunk_len, 0);
tools/testing/selftests/bpf/progs/test_xdp_meta.c
224
bpf_dynptr_write(&meta, chunk_len, src, chunk_len, 0);
tools/testing/selftests/bpf/progs/test_xdp_meta.c
228
dst = bpf_dynptr_slice_rdwr(&meta, 2 * chunk_len, NULL, chunk_len);
tools/testing/selftests/bpf/progs/test_xdp_meta.c
240
struct bpf_dynptr meta;
tools/testing/selftests/bpf/progs/test_xdp_meta.c
244
err = bpf_dynptr_from_skb_meta(ctx, 0, &meta);
tools/testing/selftests/bpf/progs/test_xdp_meta.c
249
err = bpf_dynptr_read(&md, sizeof(md), &meta, META_SIZE, 0);
tools/testing/selftests/bpf/progs/test_xdp_meta.c
254
err = bpf_dynptr_write(&meta, META_SIZE, &md, sizeof(md), 0);
tools/testing/selftests/bpf/progs/test_xdp_meta.c
259
err = bpf_dynptr_adjust(&meta, 0, META_SIZE + 1);
tools/testing/selftests/bpf/progs/test_xdp_meta.c
264
err = bpf_dynptr_adjust(&meta, META_SIZE + 1, META_SIZE + 1);
tools/testing/selftests/bpf/progs/test_xdp_meta.c
269
p = bpf_dynptr_slice(&meta, META_SIZE, NULL, sizeof(*p));
tools/testing/selftests/bpf/progs/test_xdp_meta.c
274
p = bpf_dynptr_slice_rdwr(&meta, META_SIZE, NULL, sizeof(*p));
tools/testing/selftests/bpf/progs/test_xdp_meta.c
288
__u8 *meta;
tools/testing/selftests/bpf/progs/test_xdp_meta.c
301
meta = ctx_ptr(ctx, data_meta);
tools/testing/selftests/bpf/progs/test_xdp_meta.c
302
if (meta + META_SIZE > ctx_ptr(ctx, data))
tools/testing/selftests/bpf/progs/test_xdp_meta.c
305
__builtin_memset(meta, 0, META_SIZE);
tools/testing/selftests/bpf/progs/test_xdp_meta.c
410
struct bpf_dynptr data, meta;
tools/testing/selftests/bpf/progs/test_xdp_meta.c
422
bpf_dynptr_from_skb_meta(ctx, 0, &meta);
tools/testing/selftests/bpf/progs/test_xdp_meta.c
423
bpf_dynptr_read(meta_have, META_SIZE, &meta, 0, 0);
tools/testing/selftests/bpf/progs/test_xdp_meta.c
439
struct bpf_dynptr data, meta;
tools/testing/selftests/bpf/progs/test_xdp_meta.c
451
bpf_dynptr_from_skb_meta(ctx, 0, &meta);
tools/testing/selftests/bpf/progs/test_xdp_meta.c
452
meta_have = bpf_dynptr_slice_rdwr(&meta, 0, NULL, META_SIZE);
tools/testing/selftests/bpf/progs/test_xdp_meta.c
472
struct bpf_dynptr data, meta;
tools/testing/selftests/bpf/progs/test_xdp_meta.c
486
bpf_dynptr_from_skb_meta(ctx, 0, &meta);
tools/testing/selftests/bpf/progs/test_xdp_meta.c
487
if (bpf_dynptr_is_rdonly(&meta))
tools/testing/selftests/bpf/progs/test_xdp_meta.c
490
err = bpf_dynptr_read(meta_have, META_SIZE, &meta, 0, 0);
tools/testing/selftests/bpf/progs/test_xdp_meta.c
497
err = bpf_dynptr_read(meta_have, META_SIZE, &meta, 0, 0);
tools/testing/selftests/bpf/progs/test_xdp_meta.c
514
struct bpf_dynptr data, meta;
tools/testing/selftests/bpf/progs/test_xdp_meta.c
528
bpf_dynptr_from_skb_meta(ctx, 0, &meta);
tools/testing/selftests/bpf/progs/test_xdp_meta.c
529
if (bpf_dynptr_is_rdonly(&meta))
tools/testing/selftests/bpf/progs/test_xdp_meta.c
532
err = bpf_dynptr_read(meta_have, META_SIZE, &meta, 0, 0);
tools/testing/selftests/bpf/progs/test_xdp_meta.c
537
bpf_dynptr_write(&meta, 0, &meta_have[0], 1, 0);
tools/testing/selftests/bpf/progs/test_xdp_meta.c
539
err = bpf_dynptr_read(meta_have, META_SIZE, &meta, 0, 0);
tools/testing/selftests/bpf/progs/test_xdp_meta.c
89
struct bpf_dynptr meta;
tools/testing/selftests/bpf/progs/test_xdp_meta.c
91
bpf_dynptr_from_skb_meta(ctx, 0, &meta);
tools/testing/selftests/bpf/progs/test_xdp_meta.c
92
bpf_dynptr_read(meta_have, META_SIZE, &meta, 0, 0);
tools/testing/selftests/bpf/progs/verifier_arena.c
426
struct seq_file *seq = ctx->meta->seq;
tools/testing/selftests/bpf/progs/verifier_bits_iter.c
23
int BPF_PROG(no_destroy, struct bpf_iter_meta *meta, struct cgroup *cgrp)
tools/testing/selftests/bpf/progs/verifier_bits_iter.c
36
int BPF_PROG(next_uninit, struct bpf_iter_meta *meta, struct cgroup *cgrp)
tools/testing/selftests/bpf/progs/verifier_bits_iter.c
47
int BPF_PROG(destroy_uninit, struct bpf_iter_meta *meta, struct cgroup *cgrp)
tools/testing/selftests/bpf/progs/xdp_hw_metadata.c
100
&meta->rx_hash_type);
tools/testing/selftests/bpf/progs/xdp_hw_metadata.c
102
meta->rx_hash_err = err;
tools/testing/selftests/bpf/progs/xdp_hw_metadata.c
104
meta->hint_valid |= XDP_META_FIELD_RSS;
tools/testing/selftests/bpf/progs/xdp_hw_metadata.c
106
err = bpf_xdp_metadata_rx_vlan_tag(ctx, &meta->rx_vlan_proto,
tools/testing/selftests/bpf/progs/xdp_hw_metadata.c
107
&meta->rx_vlan_tci);
tools/testing/selftests/bpf/progs/xdp_hw_metadata.c
109
meta->rx_vlan_tag_err = err;
tools/testing/selftests/bpf/progs/xdp_hw_metadata.c
111
meta->hint_valid |= XDP_META_FIELD_VLAN_TAG;
tools/testing/selftests/bpf/progs/xdp_hw_metadata.c
34
struct xdp_meta *meta;
tools/testing/selftests/bpf/progs/xdp_hw_metadata.c
83
meta = data_meta;
tools/testing/selftests/bpf/progs/xdp_hw_metadata.c
85
if (meta + 1 > data) {
tools/testing/selftests/bpf/progs/xdp_hw_metadata.c
90
meta->hint_valid = 0;
tools/testing/selftests/bpf/progs/xdp_hw_metadata.c
92
meta->xdp_timestamp = bpf_ktime_get_tai_ns();
tools/testing/selftests/bpf/progs/xdp_hw_metadata.c
93
err = bpf_xdp_metadata_rx_timestamp(ctx, &meta->rx_timestamp);
tools/testing/selftests/bpf/progs/xdp_hw_metadata.c
95
meta->rx_timestamp_err = err;
tools/testing/selftests/bpf/progs/xdp_hw_metadata.c
97
meta->hint_valid |= XDP_META_FIELD_TS;
tools/testing/selftests/bpf/progs/xdp_hw_metadata.c
99
err = bpf_xdp_metadata_rx_hash(ctx, &meta->rx_hash,
tools/testing/selftests/bpf/progs/xdp_metadata.c
100
&meta->rx_vlan_tci);
tools/testing/selftests/bpf/progs/xdp_metadata.c
45
struct xdp_meta *meta;
tools/testing/selftests/bpf/progs/xdp_metadata.c
86
meta = data_meta;
tools/testing/selftests/bpf/progs/xdp_metadata.c
96
meta->rx_timestamp = 1;
tools/testing/selftests/bpf/progs/xdp_metadata.c
98
bpf_xdp_metadata_rx_hash(ctx, &meta->rx_hash, &meta->rx_hash_type);
tools/testing/selftests/bpf/progs/xdp_metadata.c
99
bpf_xdp_metadata_rx_vlan_tag(ctx, &meta->rx_vlan_proto,
tools/testing/selftests/bpf/progs/xsk_xdp_progs.c
41
struct xdp_info *meta;
tools/testing/selftests/bpf/progs/xsk_xdp_progs.c
55
meta = data_meta;
tools/testing/selftests/bpf/progs/xsk_xdp_progs.c
56
meta->count = count++;
tools/testing/selftests/bpf/xdp_hw_metadata.c
224
struct xdp_meta *meta;
tools/testing/selftests/bpf/xdp_hw_metadata.c
226
meta = data - sizeof(*meta);
tools/testing/selftests/bpf/xdp_hw_metadata.c
228
if (meta->hint_valid & XDP_META_FIELD_RSS)
tools/testing/selftests/bpf/xdp_hw_metadata.c
230
meta->rx_hash, meta->rx_hash_type);
tools/testing/selftests/bpf/xdp_hw_metadata.c
232
printf("No rx_hash, err=%d\n", meta->rx_hash_err);
tools/testing/selftests/bpf/xdp_hw_metadata.c
234
if (meta->hint_valid & XDP_META_FIELD_TS) {
tools/testing/selftests/bpf/xdp_hw_metadata.c
238
last_hw_rx_timestamp = meta->rx_timestamp;
tools/testing/selftests/bpf/xdp_hw_metadata.c
239
last_xdp_rx_timestamp = meta->xdp_timestamp;
tools/testing/selftests/bpf/xdp_hw_metadata.c
242
meta->rx_timestamp, ref_tstamp);
tools/testing/selftests/bpf/xdp_hw_metadata.c
244
meta->xdp_timestamp, ref_tstamp);
tools/testing/selftests/bpf/xdp_hw_metadata.c
246
printf("No rx_timestamp, err=%d\n", meta->rx_timestamp_err);
tools/testing/selftests/bpf/xdp_hw_metadata.c
249
if (meta->hint_valid & XDP_META_FIELD_VLAN_TAG) {
tools/testing/selftests/bpf/xdp_hw_metadata.c
250
printf("rx_vlan_proto: 0x%X\n", ntohs(meta->rx_vlan_proto));
tools/testing/selftests/bpf/xdp_hw_metadata.c
252
print_vlan_tci(meta->rx_vlan_tci);
tools/testing/selftests/bpf/xdp_hw_metadata.c
255
meta->rx_vlan_tag_err);
tools/testing/selftests/bpf/xdp_hw_metadata.c
306
struct xsk_tx_metadata *meta;
tools/testing/selftests/bpf/xdp_hw_metadata.c
316
meta = data - sizeof(struct xsk_tx_metadata);
tools/testing/selftests/bpf/xdp_hw_metadata.c
320
if (meta->completion.tx_timestamp) {
tools/testing/selftests/bpf/xdp_hw_metadata.c
327
meta->completion.tx_timestamp);
tools/testing/selftests/bpf/xdp_hw_metadata.c
330
meta->completion.tx_timestamp, ref_tstamp);
tools/testing/selftests/bpf/xdp_hw_metadata.c
334
last_hw_rx_timestamp, meta->completion.tx_timestamp);
tools/testing/selftests/bpf/xdp_hw_metadata.c
354
struct xsk_tx_metadata *meta;
tools/testing/selftests/bpf/xdp_hw_metadata.c
376
meta = data - sizeof(struct xsk_tx_metadata);
tools/testing/selftests/bpf/xdp_hw_metadata.c
377
memset(meta, 0, sizeof(*meta));
tools/testing/selftests/bpf/xdp_hw_metadata.c
378
meta->flags = XDP_TXMD_FLAGS_TIMESTAMP;
tools/testing/selftests/bpf/xdp_hw_metadata.c
415
meta->flags |= XDP_TXMD_FLAGS_CHECKSUM;
tools/testing/selftests/bpf/xdp_hw_metadata.c
417
meta->request.csum_start = sizeof(*eth) + sizeof(*iph);
tools/testing/selftests/bpf/xdp_hw_metadata.c
419
meta->request.csum_start = sizeof(*eth) + sizeof(*ip6h);
tools/testing/selftests/bpf/xdp_hw_metadata.c
420
meta->request.csum_offset = offsetof(struct udphdr, check);
tools/testing/selftests/bpf/xdp_hw_metadata.c
424
meta->request.csum_start, meta->request.csum_offset);
tools/testing/selftests/bpf/xdp_hw_metadata.c
428
meta->flags |= XDP_TXMD_FLAGS_LAUNCH_TIME;
tools/testing/selftests/bpf/xdp_hw_metadata.c
429
meta->request.launch_time = last_hw_rx_timestamp +
tools/testing/selftests/bpf/xdp_hw_metadata.c
431
last_launch_time = meta->request.launch_time;
tools/testing/selftests/bpf/xdp_hw_metadata.c
434
meta->request.launch_time);
tools/testing/selftests/filesystems/mount-notify/mount-notify_test.c
120
struct fanotify_event_metadata *meta;
tools/testing/selftests/filesystems/mount-notify/mount-notify_test.c
145
meta = self->next;
tools/testing/selftests/filesystems/mount-notify/mount-notify_test.c
146
ASSERT_TRUE(FAN_EVENT_OK(meta, self->rem));
tools/testing/selftests/filesystems/mount-notify/mount-notify_test.c
148
thislen = meta->event_len;
tools/testing/selftests/filesystems/mount-notify/mount-notify_test.c
152
*mask = meta->mask;
tools/testing/selftests/filesystems/mount-notify/mount-notify_test.c
153
thislen -= sizeof(*meta);
tools/testing/selftests/filesystems/mount-notify/mount-notify_test.c
155
mnt = ((void *) meta) + meta->event_len - thislen;
tools/testing/selftests/filesystems/mount-notify/mount-notify_test_ns.c
147
struct fanotify_event_metadata *meta;
tools/testing/selftests/filesystems/mount-notify/mount-notify_test_ns.c
172
meta = self->next;
tools/testing/selftests/filesystems/mount-notify/mount-notify_test_ns.c
173
ASSERT_TRUE(FAN_EVENT_OK(meta, self->rem));
tools/testing/selftests/filesystems/mount-notify/mount-notify_test_ns.c
175
thislen = meta->event_len;
tools/testing/selftests/filesystems/mount-notify/mount-notify_test_ns.c
179
*mask = meta->mask;
tools/testing/selftests/filesystems/mount-notify/mount-notify_test_ns.c
180
thislen -= sizeof(*meta);
tools/testing/selftests/filesystems/mount-notify/mount-notify_test_ns.c
182
mnt = ((void *) meta) + meta->event_len - thislen;
tools/testing/selftests/ring-buffer/map_test.c
100
desc->meta = (struct trace_buffer_meta *)map;
tools/testing/selftests/ring-buffer/map_test.c
103
if (page_size < desc->meta->meta_struct_len) {
tools/testing/selftests/ring-buffer/map_test.c
104
int meta_page_size = desc->meta->meta_page_size;
tools/testing/selftests/ring-buffer/map_test.c
106
munmap(desc->meta, page_size);
tools/testing/selftests/ring-buffer/map_test.c
116
munmap(desc->meta, desc->meta->meta_page_size);
tools/testing/selftests/ring-buffer/map_test.c
188
ASSERT_EQ(desc->meta->entries, 0);
tools/testing/selftests/ring-buffer/map_test.c
189
ASSERT_EQ(desc->meta->overrun, 0);
tools/testing/selftests/ring-buffer/map_test.c
190
ASSERT_EQ(desc->meta->read, 0);
tools/testing/selftests/ring-buffer/map_test.c
192
ASSERT_EQ(desc->meta->reader.id, 0);
tools/testing/selftests/ring-buffer/map_test.c
193
ASSERT_EQ(desc->meta->reader.read, 0);
tools/testing/selftests/ring-buffer/map_test.c
196
ASSERT_EQ(desc->meta->reader.id, 0);
tools/testing/selftests/ring-buffer/map_test.c
204
ASSERT_EQ(desc->meta->entries, 16);
tools/testing/selftests/ring-buffer/map_test.c
205
ASSERT_EQ(desc->meta->overrun, 0);
tools/testing/selftests/ring-buffer/map_test.c
206
ASSERT_EQ(desc->meta->read, 16);
tools/testing/selftests/ring-buffer/map_test.c
208
ASSERT_EQ(desc->meta->reader.id, 1);
tools/testing/selftests/ring-buffer/map_test.c
220
meta_len = desc->meta->meta_page_size;
tools/testing/selftests/ring-buffer/map_test.c
221
data_len = desc->meta->subbuf_size * desc->meta->nr_subbufs;
tools/testing/selftests/ring-buffer/map_test.c
230
data_len -= desc->meta->subbuf_size;
tools/testing/selftests/ring-buffer/map_test.c
237
meta_len += desc->meta->subbuf_size * 2;
tools/testing/selftests/ring-buffer/map_test.c
243
data_len = desc->meta->subbuf_size * desc->meta->nr_subbufs;
tools/testing/selftests/ring-buffer/map_test.c
245
desc->cpu_fd, data_len + (desc->meta->subbuf_size * 2));
tools/testing/selftests/ring-buffer/map_test.c
249
if (desc->meta->meta_page_size > getpagesize()) {
tools/testing/selftests/ring-buffer/map_test.c
250
data_len = desc->meta->meta_page_size;
tools/testing/selftests/ring-buffer/map_test.c
255
for (int i = desc->meta->meta_struct_len;
tools/testing/selftests/ring-buffer/map_test.c
256
i < desc->meta->meta_page_size; i += sizeof(int))
tools/testing/selftests/ring-buffer/map_test.c
75
struct trace_buffer_meta *meta;