arch/mips/include/asm/octeon/cvmx-pip-defs.h
2342
uint64_t frag:32;
arch/mips/include/asm/octeon/cvmx-pip-defs.h
2346
uint64_t frag:32;
arch/mips/include/asm/octeon/cvmx-pip-defs.h
2355
uint64_t frag:32;
arch/mips/include/asm/octeon/cvmx-pip-defs.h
2359
uint64_t frag:32;
arch/mips/include/asm/octeon/cvmx-pip-defs.h
2712
uint64_t frag:32;
arch/mips/include/asm/octeon/cvmx-pip-defs.h
2716
uint64_t frag:32;
arch/mips/include/asm/octeon/cvmx-pip.h
422
status->runt_crc_packets = stat8.s.frag;
arch/powerpc/mm/book3s64/mmu_context.c
259
void *frag;
arch/powerpc/mm/book3s64/mmu_context.c
261
frag = mm->context.pte_frag;
arch/powerpc/mm/book3s64/mmu_context.c
262
if (frag)
arch/powerpc/mm/book3s64/mmu_context.c
263
pte_frag_destroy(frag);
arch/powerpc/mm/book3s64/mmu_context.c
265
frag = mm->context.pmd_frag;
arch/powerpc/mm/book3s64/mmu_context.c
266
if (frag)
arch/powerpc/mm/book3s64/mmu_context.c
267
pmd_frag_destroy(frag);
arch/powerpc/mm/mmu_context.c
112
void *frag = pte_frag_get(&mm->context);
arch/powerpc/mm/mmu_context.c
114
if (frag)
arch/powerpc/mm/mmu_context.c
115
pte_frag_destroy(frag);
arch/s390/kernel/perf_cpum_cf.c
975
raw.frag.size = cpuhw->usedss;
arch/s390/kernel/perf_cpum_cf.c
976
raw.frag.data = cpuhw->stop;
arch/s390/kernel/perf_pai.c
630
raw.frag.size = rawsize;
arch/s390/kernel/perf_pai.c
631
raw.frag.data = cpump->save;
arch/um/drivers/vector_kern.c
290
int nr_frags, frag;
arch/um/drivers/vector_kern.c
310
for (frag = 0; frag < nr_frags; frag++) {
arch/um/drivers/vector_kern.c
311
skb_frag = &skb_shinfo(skb)->frags[frag];
arch/um/drivers/vector_kern.c
604
int err, nr_frags, frag;
arch/um/drivers/vector_kern.c
637
for (frag = 0; frag < nr_frags; frag++) {
arch/um/drivers/vector_kern.c
638
skb_frag = &skb_shinfo(result)->frags[frag];
arch/x86/events/amd/ibs.c
1353
.frag = {
arch/x86/kvm/x86.c
11833
struct kvm_mmio_fragment *frag;
arch/x86/kvm/x86.c
11839
frag = &vcpu->mmio_fragments[vcpu->mmio_cur_fragment];
arch/x86/kvm/x86.c
11840
len = min(8u, frag->len);
arch/x86/kvm/x86.c
11842
memcpy(frag->data, run->mmio.data, len);
arch/x86/kvm/x86.c
11844
if (frag->len <= 8) {
arch/x86/kvm/x86.c
11846
frag++;
arch/x86/kvm/x86.c
11850
frag->data += len;
arch/x86/kvm/x86.c
11851
frag->gpa += len;
arch/x86/kvm/x86.c
11852
frag->len -= len;
arch/x86/kvm/x86.c
11866
run->mmio.phys_addr = frag->gpa;
arch/x86/kvm/x86.c
11868
memcpy(run->mmio.data, frag->data, min(8u, frag->len));
arch/x86/kvm/x86.c
11869
run->mmio.len = min(8u, frag->len);
arch/x86/kvm/x86.c
14243
struct kvm_mmio_fragment *frag;
arch/x86/kvm/x86.c
14249
frag = &vcpu->mmio_fragments[vcpu->mmio_cur_fragment];
arch/x86/kvm/x86.c
14250
len = min(8u, frag->len);
arch/x86/kvm/x86.c
14252
memcpy(frag->data, run->mmio.data, len);
arch/x86/kvm/x86.c
14254
if (frag->len <= 8) {
arch/x86/kvm/x86.c
14256
frag++;
arch/x86/kvm/x86.c
14260
frag->data += len;
arch/x86/kvm/x86.c
14261
frag->gpa += len;
arch/x86/kvm/x86.c
14262
frag->len -= len;
arch/x86/kvm/x86.c
14274
run->mmio.phys_addr = frag->gpa;
arch/x86/kvm/x86.c
14275
run->mmio.len = min(8u, frag->len);
arch/x86/kvm/x86.c
14278
memcpy(run->mmio.data, frag->data, min(8u, frag->len));
arch/x86/kvm/x86.c
14290
struct kvm_mmio_fragment *frag;
arch/x86/kvm/x86.c
14304
frag = vcpu->mmio_fragments;
arch/x86/kvm/x86.c
14306
frag->len = bytes;
arch/x86/kvm/x86.c
14307
frag->gpa = gpa;
arch/x86/kvm/x86.c
14308
frag->data = data;
arch/x86/kvm/x86.c
14314
vcpu->run->mmio.len = min(8u, frag->len);
arch/x86/kvm/x86.c
14316
memcpy(vcpu->run->mmio.data, frag->data, min(8u, frag->len));
arch/x86/kvm/x86.c
14329
struct kvm_mmio_fragment *frag;
arch/x86/kvm/x86.c
14343
frag = vcpu->mmio_fragments;
arch/x86/kvm/x86.c
14345
frag->len = bytes;
arch/x86/kvm/x86.c
14346
frag->gpa = gpa;
arch/x86/kvm/x86.c
14347
frag->data = data;
arch/x86/kvm/x86.c
14353
vcpu->run->mmio.len = min(8u, frag->len);
arch/x86/kvm/x86.c
8162
struct kvm_mmio_fragment *frag = &vcpu->mmio_fragments[0];
arch/x86/kvm/x86.c
8164
memcpy(vcpu->run->mmio.data, frag->data, min(8u, frag->len));
arch/x86/kvm/x86.c
8191
struct kvm_mmio_fragment *frag;
arch/x86/kvm/x86.c
8226
frag = &vcpu->mmio_fragments[vcpu->mmio_nr_fragments++];
arch/x86/kvm/x86.c
8227
frag->gpa = gpa;
arch/x86/kvm/x86.c
8228
frag->data = val;
arch/x86/kvm/x86.c
8229
frag->len = bytes;
block/partitions/ldm.c
1218
struct frag *f;
block/partitions/ldm.c
1242
f = list_entry (item, struct frag, list);
block/partitions/ldm.c
1293
kfree (list_entry (item, struct frag, list));
block/partitions/ldm.c
1309
struct frag *f;
block/partitions/ldm.c
1315
f = list_entry (item, struct frag, list);
drivers/atm/he.c
2555
skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/atm/he.c
2579
frag, 0, skb_frag_size(frag), DMA_TO_DEVICE);
drivers/atm/he.c
2580
tpd->iovec[slot].len = skb_frag_size(frag);
drivers/block/aoe/aoecmd.c
298
int frag = 0;
drivers/block/aoe/aoecmd.c
302
skb_fill_page_desc(skb, frag++, bv.bv_page,
drivers/block/aoe/aoecmd.c
592
int frag;
drivers/block/aoe/aoecmd.c
609
for (frag = 0, n = f->iter.bi_size; n > 0; ++frag, n -= m) {
drivers/block/aoe/aoecmd.c
614
skb_fill_page_desc(skb, frag, empty_page, 0, m);
drivers/gpu/drm/amd/amdgpu/amdgpu_vm_pt.c
737
unsigned int *frag, uint64_t *frag_end)
drivers/gpu/drm/amd/amdgpu/amdgpu_vm_pt.c
769
*frag = 0;
drivers/gpu/drm/amd/amdgpu/amdgpu_vm_pt.c
775
*frag = min_t(unsigned int, ffs(start) - 1, fls64(end - start) - 1);
drivers/gpu/drm/amd/amdgpu/amdgpu_vm_pt.c
776
if (*frag >= max_frag) {
drivers/gpu/drm/amd/amdgpu/amdgpu_vm_pt.c
777
*frag = max_frag;
drivers/gpu/drm/amd/amdgpu/amdgpu_vm_pt.c
780
*frag_end = start + (1 << *frag);
drivers/gpu/drm/amd/amdgpu/amdgpu_vm_pt.c
805
unsigned int frag;
drivers/gpu/drm/amd/amdgpu/amdgpu_vm_pt.c
809
amdgpu_vm_pte_fragment(params, frag_start, end, flags, &frag,
drivers/gpu/drm/amd/amdgpu/amdgpu_vm_pt.c
843
} else if (frag < shift) {
drivers/gpu/drm/amd/amdgpu/amdgpu_vm_pt.c
850
} else if (frag >= parent_shift) {
drivers/gpu/drm/amd/amdgpu/amdgpu_vm_pt.c
897
uint64_t upd_flags = flags | AMDGPU_PTE_FRAG(frag);
drivers/gpu/drm/amd/amdgpu/amdgpu_vm_pt.c
920
flags, &frag, &frag_end);
drivers/gpu/drm/amd/amdgpu/amdgpu_vm_pt.c
921
if (frag < shift)
drivers/gpu/drm/amd/amdgpu/amdgpu_vm_pt.c
942
} else if (frag >= shift) {
drivers/hsi/clients/ssi_protocol.c
170
skb_frag_t *frag;
drivers/hsi/clients/ssi_protocol.c
181
frag = &skb_shinfo(skb)->frags[i];
drivers/hsi/clients/ssi_protocol.c
182
sg_set_page(sg, skb_frag_page(frag), skb_frag_size(frag),
drivers/hsi/clients/ssi_protocol.c
183
skb_frag_off(frag));
drivers/infiniband/hw/hfi1/ipoib_rx.c
26
void *frag;
drivers/infiniband/hw/hfi1/ipoib_rx.c
30
frag = napi_alloc_frag(skb_size);
drivers/infiniband/hw/hfi1/ipoib_rx.c
32
if (unlikely(!frag))
drivers/infiniband/hw/hfi1/ipoib_rx.c
35
skb = build_skb(frag, skb_size);
drivers/infiniband/hw/hfi1/ipoib_rx.c
38
skb_free_frag(frag);
drivers/infiniband/hw/hfi1/ipoib_tx.c
215
const skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/infiniband/hw/hfi1/ipoib_tx.c
219
skb_frag_page(frag),
drivers/infiniband/hw/hfi1/ipoib_tx.c
220
skb_frag_off(frag),
drivers/infiniband/hw/hfi1/ipoib_tx.c
221
skb_frag_size(frag),
drivers/infiniband/hw/hfi1/vnic_sdma.c
63
skb_frag_t *frag = &skb_shinfo(tx->skb)->frags[i];
drivers/infiniband/hw/hfi1/vnic_sdma.c
68
skb_frag_page(frag),
drivers/infiniband/hw/hfi1/vnic_sdma.c
69
skb_frag_off(frag),
drivers/infiniband/hw/hfi1/vnic_sdma.c
70
skb_frag_size(frag),
drivers/infiniband/ulp/ipoib/ipoib_cm.c
541
skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/infiniband/ulp/ipoib/ipoib_cm.c
545
skb_fill_page_desc(toskb, i, skb_frag_page(frag),
drivers/infiniband/ulp/ipoib/ipoib_cm.c
551
skb_frag_size_set(frag, size);
drivers/infiniband/ulp/ipoib/ipoib_ib.c
294
const skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/infiniband/ulp/ipoib/ipoib_ib.c
296
skb_frag_page(frag),
drivers/infiniband/ulp/ipoib/ipoib_ib.c
297
skb_frag_off(frag),
drivers/infiniband/ulp/ipoib/ipoib_ib.c
298
skb_frag_size(frag),
drivers/infiniband/ulp/ipoib/ipoib_ib.c
307
const skb_frag_t *frag = &skb_shinfo(skb)->frags[i - 1];
drivers/infiniband/ulp/ipoib/ipoib_ib.c
309
ib_dma_unmap_page(ca, mapping[i - !off], skb_frag_size(frag), DMA_TO_DEVICE);
drivers/infiniband/ulp/ipoib/ipoib_ib.c
334
const skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/infiniband/ulp/ipoib/ipoib_ib.c
337
skb_frag_size(frag), DMA_TO_DEVICE);
drivers/misc/mei/vsc-fw-loader.c
417
struct vsc_img_frag *frag;
drivers/misc/mei/vsc-fw-loader.c
434
frag = &fw_loader->frags[csi_image_map[i].frag_index];
drivers/misc/mei/vsc-fw-loader.c
436
frag->data = sign->image;
drivers/misc/mei/vsc-fw-loader.c
437
frag->size = le32_to_cpu(sign->image_size);
drivers/misc/mei/vsc-fw-loader.c
438
frag->location = le32_to_cpu(img->image_location[i]);
drivers/misc/mei/vsc-fw-loader.c
439
frag->type = csi_image_map[i].image_type;
drivers/misc/mei/vsc-fw-loader.c
497
struct vsc_img_frag *frag, *last_frag;
drivers/misc/mei/vsc-fw-loader.c
511
frag = &fw_loader->frags[frag_index];
drivers/misc/mei/vsc-fw-loader.c
513
frag->data = sign->image;
drivers/misc/mei/vsc-fw-loader.c
514
frag->size = le32_to_cpu(sign->image_size);
drivers/misc/mei/vsc-fw-loader.c
515
frag->location = le32_to_cpu(img->image_location[i]);
drivers/misc/mei/vsc-fw-loader.c
516
frag->type = ace_image_map[i].image_type;
drivers/misc/mei/vsc-fw-loader.c
518
if (!frag->location) {
drivers/misc/mei/vsc-fw-loader.c
520
frag->location =
drivers/misc/mei/vsc-fw-loader.c
540
struct vsc_img_frag *frag = &fw_loader->frags[VSC_IMG_SKU_CFG_FRAG];
drivers/misc/mei/vsc-fw-loader.c
565
frag->data = image->data;
drivers/misc/mei/vsc-fw-loader.c
566
frag->size = image->size;
drivers/misc/mei/vsc-fw-loader.c
567
frag->type = VSC_IMG_SKU_CFG_TYPE;
drivers/misc/mei/vsc-fw-loader.c
568
frag->location = VSC_SKU_CFG_LOCATION;
drivers/misc/mei/vsc-fw-loader.c
582
struct vsc_img_frag *frag = &fw_loader->frags[VSC_IMG_BOOTLOADER_FRAG];
drivers/misc/mei/vsc-fw-loader.c
602
cmd->data.dl_start.img_type = frag->type;
drivers/misc/mei/vsc-fw-loader.c
603
cmd->data.dl_start.img_len = cpu_to_le32(frag->size);
drivers/misc/mei/vsc-fw-loader.c
604
cmd->data.dl_start.img_loc = cpu_to_le32(frag->location);
drivers/misc/mei/vsc-fw-loader.c
613
p = frag->data;
drivers/misc/mei/vsc-fw-loader.c
614
remain = frag->size;
drivers/misc/mei/vsc-fw-loader.c
650
struct vsc_img_frag *frag = &fw_loader->frags[i];
drivers/misc/mei/vsc-fw-loader.c
652
cmd->data.dl_set.payload[index++] = cpu_to_le32(frag->location);
drivers/misc/mei/vsc-fw-loader.c
653
cmd->data.dl_set.payload[index++] = cpu_to_le32(frag->size);
drivers/misc/mei/vsc-fw-loader.c
664
struct vsc_img_frag *frag = &fw_loader->frags[i];
drivers/misc/mei/vsc-fw-loader.c
670
cmd->data.dl_start.img_type = frag->type;
drivers/misc/mei/vsc-fw-loader.c
671
cmd->data.dl_start.img_len = cpu_to_le32(frag->size);
drivers/misc/mei/vsc-fw-loader.c
672
cmd->data.dl_start.img_loc = cpu_to_le32(frag->location);
drivers/misc/mei/vsc-fw-loader.c
682
p = frag->data;
drivers/misc/mei/vsc-fw-loader.c
683
remain = frag->size;
drivers/net/ethernet/3com/3c59x.c
2160
vp->tx_ring[entry].frag[0].addr = cpu_to_le32(dma_addr);
drivers/net/ethernet/3com/3c59x.c
2161
vp->tx_ring[entry].frag[0].length = cpu_to_le32(skb->len | LAST_FRAG);
drivers/net/ethernet/3com/3c59x.c
2170
vp->tx_ring[entry].frag[0].addr = cpu_to_le32(dma_addr);
drivers/net/ethernet/3com/3c59x.c
2171
vp->tx_ring[entry].frag[0].length = cpu_to_le32(skb_headlen(skb));
drivers/net/ethernet/3com/3c59x.c
2174
skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/3com/3c59x.c
2176
dma_addr = skb_frag_dma_map(vp->gendev, frag,
drivers/net/ethernet/3com/3c59x.c
2178
skb_frag_size(frag),
drivers/net/ethernet/3com/3c59x.c
2183
le32_to_cpu(vp->tx_ring[entry].frag[i+1].addr),
drivers/net/ethernet/3com/3c59x.c
2184
le32_to_cpu(vp->tx_ring[entry].frag[i+1].length),
drivers/net/ethernet/3com/3c59x.c
2188
le32_to_cpu(vp->tx_ring[entry].frag[0].addr),
drivers/net/ethernet/3com/3c59x.c
2189
le32_to_cpu(vp->tx_ring[entry].frag[0].length),
drivers/net/ethernet/3com/3c59x.c
2195
vp->tx_ring[entry].frag[i+1].addr =
drivers/net/ethernet/3com/3c59x.c
2199
vp->tx_ring[entry].frag[i+1].length = cpu_to_le32(skb_frag_size(frag)|LAST_FRAG);
drivers/net/ethernet/3com/3c59x.c
2201
vp->tx_ring[entry].frag[i+1].length = cpu_to_le32(skb_frag_size(frag));
drivers/net/ethernet/3com/3c59x.c
2437
le32_to_cpu(vp->tx_ring[entry].frag[0].addr),
drivers/net/ethernet/3com/3c59x.c
2438
le32_to_cpu(vp->tx_ring[entry].frag[0].length)&0xFFF,
drivers/net/ethernet/3com/3c59x.c
2443
le32_to_cpu(vp->tx_ring[entry].frag[i].addr),
drivers/net/ethernet/3com/3c59x.c
2444
le32_to_cpu(vp->tx_ring[entry].frag[i].length)&0xFFF,
drivers/net/ethernet/3com/3c59x.c
2772
le32_to_cpu(vp->tx_ring[i].frag[k].addr),
drivers/net/ethernet/3com/3c59x.c
2773
le32_to_cpu(vp->tx_ring[i].frag[k].length)&0xFFF,
drivers/net/ethernet/3com/3c59x.c
2810
length = le32_to_cpu(vp->tx_ring[i].frag[0].length);
drivers/net/ethernet/3com/3c59x.c
570
} frag[1+MAX_SKB_FRAGS];
drivers/net/ethernet/3com/typhoon.c
1524
skb_dma = (dma_addr_t) le32_to_cpu(tx->frag.addr);
drivers/net/ethernet/3com/typhoon.c
791
txd->frag.addr = cpu_to_le32(skb_dma);
drivers/net/ethernet/3com/typhoon.c
792
txd->frag.addrHi = 0;
drivers/net/ethernet/3com/typhoon.c
802
txd->frag.addr = cpu_to_le32(skb_dma);
drivers/net/ethernet/3com/typhoon.c
803
txd->frag.addrHi = 0;
drivers/net/ethernet/3com/typhoon.c
807
const skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/3com/typhoon.c
814
len = skb_frag_size(frag);
drivers/net/ethernet/3com/typhoon.c
815
frag_addr = skb_frag_address(frag);
drivers/net/ethernet/3com/typhoon.c
820
txd->frag.addr = cpu_to_le32(skb_dma);
drivers/net/ethernet/3com/typhoon.c
821
txd->frag.addrHi = 0;
drivers/net/ethernet/3com/typhoon.h
173
} frag;
drivers/net/ethernet/aeroflex/greth.c
198
skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/aeroflex/greth.c
203
skb_frag_size(frag),
drivers/net/ethernet/aeroflex/greth.c
519
skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/aeroflex/greth.c
526
status |= skb_frag_size(frag) & GRETH_BD_LEN;
drivers/net/ethernet/aeroflex/greth.c
540
dma_addr = skb_frag_dma_map(greth->dev, frag, 0, skb_frag_size(frag),
drivers/net/ethernet/aeroflex/greth.c
716
skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/aeroflex/greth.c
721
skb_frag_size(frag),
drivers/net/ethernet/agere/et131x.c
2423
u32 frag = 0;
drivers/net/ethernet/agere/et131x.c
2457
desc[frag].len_vlan = skb_headlen(skb);
drivers/net/ethernet/agere/et131x.c
2466
desc[frag].addr_lo = lower_32_bits(dma_addr);
drivers/net/ethernet/agere/et131x.c
2467
desc[frag].addr_hi = upper_32_bits(dma_addr);
drivers/net/ethernet/agere/et131x.c
2468
frag++;
drivers/net/ethernet/agere/et131x.c
2470
desc[frag].len_vlan = skb_headlen(skb) / 2;
drivers/net/ethernet/agere/et131x.c
2479
desc[frag].addr_lo = lower_32_bits(dma_addr);
drivers/net/ethernet/agere/et131x.c
2480
desc[frag].addr_hi = upper_32_bits(dma_addr);
drivers/net/ethernet/agere/et131x.c
2481
frag++;
drivers/net/ethernet/agere/et131x.c
2483
desc[frag].len_vlan = skb_headlen(skb) / 2;
drivers/net/ethernet/agere/et131x.c
2493
desc[frag].addr_lo = lower_32_bits(dma_addr);
drivers/net/ethernet/agere/et131x.c
2494
desc[frag].addr_hi = upper_32_bits(dma_addr);
drivers/net/ethernet/agere/et131x.c
2495
frag++;
drivers/net/ethernet/agere/et131x.c
2498
desc[frag].len_vlan = skb_frag_size(&frags[i - 1]);
drivers/net/ethernet/agere/et131x.c
2502
desc[frag].len_vlan,
drivers/net/ethernet/agere/et131x.c
2507
desc[frag].addr_lo = lower_32_bits(dma_addr);
drivers/net/ethernet/agere/et131x.c
2508
desc[frag].addr_hi = upper_32_bits(dma_addr);
drivers/net/ethernet/agere/et131x.c
2509
frag++;
drivers/net/ethernet/agere/et131x.c
2516
desc[frag - 1].flags =
drivers/net/ethernet/agere/et131x.c
2520
desc[frag - 1].flags = TXDESC_FLAG_LASTPKT;
drivers/net/ethernet/agere/et131x.c
2523
desc[frag - 1].flags =
drivers/net/ethernet/agere/et131x.c
2534
if (thiscopy >= frag) {
drivers/net/ethernet/agere/et131x.c
2536
thiscopy = frag;
drivers/net/ethernet/agere/et131x.c
2538
remainder = frag - thiscopy;
drivers/net/ethernet/agere/et131x.c
2600
frag--;
drivers/net/ethernet/agere/et131x.c
2601
dma_addr = desc[frag].addr_lo;
drivers/net/ethernet/agere/et131x.c
2602
dma_addr |= (u64)desc[frag].addr_hi << 32;
drivers/net/ethernet/agere/et131x.c
2604
desc[frag].len_vlan, DMA_TO_DEVICE);
drivers/net/ethernet/agere/et131x.c
2609
while (frag--) {
drivers/net/ethernet/agere/et131x.c
2610
dma_addr = desc[frag].addr_lo;
drivers/net/ethernet/agere/et131x.c
2611
dma_addr |= (u64)desc[frag].addr_hi << 32;
drivers/net/ethernet/agere/et131x.c
2613
desc[frag].len_vlan, DMA_TO_DEVICE);
drivers/net/ethernet/airoha/airoha_eth.c
1989
skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/airoha/airoha_eth.c
2018
data = skb_frag_address(frag);
drivers/net/ethernet/airoha/airoha_eth.c
2019
len = skb_frag_size(frag);
drivers/net/ethernet/alteon/acenic.c
2454
const skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/alteon/acenic.c
2460
mapping = skb_frag_dma_map(&ap->pdev->dev, frag, 0,
drivers/net/ethernet/alteon/acenic.c
2461
skb_frag_size(frag),
drivers/net/ethernet/alteon/acenic.c
2464
flagsize = skb_frag_size(frag) << 16;
drivers/net/ethernet/alteon/acenic.c
2483
dma_unmap_len_set(info, maplen, skb_frag_size(frag));
drivers/net/ethernet/amazon/ena/ena_eth_com.c
374
ena_rx_ctx->frag =
drivers/net/ethernet/amazon/ena/ena_eth_com.c
381
ena_rx_ctx->l4_csum_err, ena_rx_ctx->hash, ena_rx_ctx->frag, cdesc->status);
drivers/net/ethernet/amazon/ena/ena_eth_com.h
47
bool frag;
drivers/net/ethernet/amazon/ena/ena_netdev.c
1080
if (ena_rx_ctx->frag) {
drivers/net/ethernet/amazon/ena/ena_netdev.c
1141
if (ena_rx_ctx->frag)
drivers/net/ethernet/amazon/ena/ena_netdev.c
2566
const skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/amazon/ena/ena_netdev.c
2568
frag_len = skb_frag_size(frag);
drivers/net/ethernet/amazon/ena/ena_netdev.c
2575
dma = skb_frag_dma_map(tx_ring->dev, frag, delta,
drivers/net/ethernet/amd/xgbe/xgbe-desc.c
425
skb_frag_t *frag;
drivers/net/ethernet/amd/xgbe/xgbe-desc.c
503
frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/amd/xgbe/xgbe-desc.c
506
for (datalen = skb_frag_size(frag); datalen; ) {
drivers/net/ethernet/amd/xgbe/xgbe-desc.c
510
skb_dma = skb_frag_dma_map(pdata->dev, frag, offset,
drivers/net/ethernet/amd/xgbe/xgbe-drv.c
1510
skb_frag_t *frag;
drivers/net/ethernet/amd/xgbe/xgbe-drv.c
1569
frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/amd/xgbe/xgbe-drv.c
1570
for (len = skb_frag_size(frag); len; ) {
drivers/net/ethernet/apm/xgene/xgene_enet_cle.c
13
static void xgene_cle_sband_to_hw(u8 frag, enum xgene_cle_prot_version ver,
drivers/net/ethernet/apm/xgene/xgene_enet_cle.c
17
*reg = SET_VAL(SB_IPFRAG, frag) |
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
231
skb_frag_t *frag;
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
248
frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
249
dma_unmap_page(dev, frag_dma_addr[i], skb_frag_size(frag),
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
414
skb_frag_t *frag;
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
466
frag = &skb_shinfo(skb)->frags[fidx];
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
467
size = skb_frag_size(frag);
drivers/net/ethernet/apm/xgene/xgene_enet_main.c
470
pbuf_addr = skb_frag_dma_map(dev, frag, 0, size,
drivers/net/ethernet/aquantia/atlantic/aq_nic.c
612
skb_frag_t *frag = &sinfo->frags[frag_count];
drivers/net/ethernet/aquantia/atlantic/aq_nic.c
613
unsigned int frag_len = skb_frag_size(frag);
drivers/net/ethernet/aquantia/atlantic/aq_nic.c
624
frag_pa = skb_frag_dma_map(dev, frag, buff_offset,
drivers/net/ethernet/aquantia/atlantic/aq_nic.c
758
skb_frag_t *frag = &skb_shinfo(skb)->frags[frag_count];
drivers/net/ethernet/aquantia/atlantic/aq_nic.c
760
frag_len = skb_frag_size(frag);
drivers/net/ethernet/aquantia/atlantic/aq_nic.c
769
frag,
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
33
skb_frag_t *frag = &sinfo->frags[i];
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
35
page_ref_inc(skb_frag_page(frag));
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
494
skb_frag_t *frag;
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
499
frag = &sinfo->frags[sinfo->nr_frags++];
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
507
skb_frag_fill_page_desc(frag, buff_->rxdata.page,
drivers/net/ethernet/atheros/alx/main.c
1492
skb_frag_t *frag = &skb_shinfo(skb)->frags[f];
drivers/net/ethernet/atheros/alx/main.c
1500
maplen = skb_frag_size(frag);
drivers/net/ethernet/atheros/alx/main.c
1501
dma = skb_frag_dma_map(txq->dev, frag, 0,
drivers/net/ethernet/atheros/atl1c/atl1c_main.c
2226
skb_frag_t *frag = &skb_shinfo(skb)->frags[f];
drivers/net/ethernet/atheros/atl1c/atl1c_main.c
2232
buffer_info->length = skb_frag_size(frag);
drivers/net/ethernet/atheros/atl1c/atl1c_main.c
2234
frag, 0,
drivers/net/ethernet/atheros/atl1e/atl1e_main.c
1787
const skb_frag_t *frag = &skb_shinfo(skb)->frags[f];
drivers/net/ethernet/atheros/atl1e/atl1e_main.c
1791
buf_len = skb_frag_size(frag);
drivers/net/ethernet/atheros/atl1e/atl1e_main.c
1808
frag,
drivers/net/ethernet/atheros/atlx/atl1.c
2278
const skb_frag_t *frag = &skb_shinfo(skb)->frags[f];
drivers/net/ethernet/atheros/atlx/atl1.c
2281
buf_len = skb_frag_size(frag);
drivers/net/ethernet/atheros/atlx/atl1.c
2294
frag, i * ATL1_MAX_TX_BUF_LEN,
drivers/net/ethernet/atheros/atlx/atl2.h
297
unsigned frag:1; /* received a packet less than 64 bytes
drivers/net/ethernet/broadcom/asp2/bcmasp_intf.c
244
skb_frag_t *frag;
drivers/net/ethernet/broadcom/asp2/bcmasp_intf.c
276
frag = &skb_shinfo(skb)->frags[i - 1];
drivers/net/ethernet/broadcom/asp2/bcmasp_intf.c
277
size = skb_frag_size(frag);
drivers/net/ethernet/broadcom/asp2/bcmasp_intf.c
278
mapping = skb_frag_dma_map(kdev, frag, 0, size,
drivers/net/ethernet/broadcom/bgmac.c
175
skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/broadcom/bgmac.c
176
int len = skb_frag_size(frag);
drivers/net/ethernet/broadcom/bgmac.c
180
slot->dma_addr = skb_frag_dma_map(dma_dev, frag, 0,
drivers/net/ethernet/broadcom/bnge/bnge_netdev.c
459
skb_frag_t *frag = &skb_shinfo(skb)->frags[k];
drivers/net/ethernet/broadcom/bnge/bnge_netdev.c
464
skb_frag_size(frag),
drivers/net/ethernet/broadcom/bnge/bnge_txrx.c
1049
skb_frag_t *frag;
drivers/net/ethernet/broadcom/bnge/bnge_txrx.c
1075
frag = &skb_shinfo(skb)->frags[j];
drivers/net/ethernet/broadcom/bnge/bnge_txrx.c
1081
skb_frag_size(frag),
drivers/net/ethernet/broadcom/bnge/bnge_txrx.c
1447
skb_frag_t *frag;
drivers/net/ethernet/broadcom/bnge/bnge_txrx.c
1535
frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/broadcom/bnge/bnge_txrx.c
1540
len = skb_frag_size(frag);
drivers/net/ethernet/broadcom/bnge/bnge_txrx.c
1541
mapping = skb_frag_dma_map(bd->dev, frag, 0, len,
drivers/net/ethernet/broadcom/bnge/bnge_txrx.c
1548
netmem_dma_unmap_addr_set(skb_frag_netmem(frag), tx_buf,
drivers/net/ethernet/broadcom/bnge/bnge_txrx.c
1602
frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/broadcom/bnge/bnge_txrx.c
1605
skb_frag_size(frag),
drivers/net/ethernet/broadcom/bnx2.c
3081
skb_frag_t *frag =
drivers/net/ethernet/broadcom/bnx2.c
3083
skb_frag_size_sub(frag, tail);
drivers/net/ethernet/broadcom/bnx2.c
6670
const skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/broadcom/bnx2.c
6676
len = skb_frag_size(frag);
drivers/net/ethernet/broadcom/bnx2.c
6677
mapping = skb_frag_dma_map(&bp->pdev->dev, frag, 0, len,
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
4090
skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
4092
mapping = skb_frag_dma_map(&bp->pdev->dev, frag, 0,
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
4093
skb_frag_size(frag), DMA_TO_DEVICE);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
4119
tx_data_bd->nbytes = cpu_to_le16(skb_frag_size(frag));
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
4120
le16_add_cpu(&pkt_size, skb_frag_size(frag));
drivers/net/ethernet/broadcom/bnxt/bnxt.c
1178
skb_frag_t *frag;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
1208
frag = &skb_shinfo(skb)->frags[0];
drivers/net/ethernet/broadcom/bnxt/bnxt.c
1209
skb_frag_size_sub(frag, payload);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
1210
skb_frag_off_add(frag, payload);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
1291
skb_frag_t *frag = &shinfo->frags[i];
drivers/net/ethernet/broadcom/bnxt/bnxt.c
1293
skb_frag_fill_netmem_desc(frag, cons_rx_buf->netmem,
drivers/net/ethernet/broadcom/bnxt/bnxt.c
3454
skb_frag_t *frag = &skb_shinfo(skb)->frags[j];
drivers/net/ethernet/broadcom/bnxt/bnxt.c
3460
skb_frag_size(frag),
drivers/net/ethernet/broadcom/bnxt/bnxt.c
485
skb_frag_t *frag;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
606
frag = &skb_shinfo(skb)->frags[j];
drivers/net/ethernet/broadcom/bnxt/bnxt.c
607
fptr = skb_frag_address_safe(frag);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
611
memcpy(pdata, fptr, skb_frag_size(frag));
drivers/net/ethernet/broadcom/bnxt/bnxt.c
612
pdata += skb_frag_size(frag);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
714
frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/broadcom/bnxt/bnxt.c
718
len = skb_frag_size(frag);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
719
mapping = skb_frag_dma_map(&pdev->dev, frag, 0, len,
drivers/net/ethernet/broadcom/bnxt/bnxt.c
726
netmem_dma_unmap_addr_set(skb_frag_netmem(frag), tx_buf,
drivers/net/ethernet/broadcom/bnxt/bnxt.c
784
frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/broadcom/bnxt/bnxt.c
787
skb_frag_size(frag),
drivers/net/ethernet/broadcom/bnxt/bnxt.c
817
skb_frag_t *frag;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
857
frag = &skb_shinfo(skb)->frags[j];
drivers/net/ethernet/broadcom/bnxt/bnxt.c
863
skb_frag_size(frag),
drivers/net/ethernet/broadcom/bnxt/bnxt_xdp.c
60
skb_frag_t *frag = &sinfo->frags[i];
drivers/net/ethernet/broadcom/bnxt/bnxt_xdp.c
70
frag_tx_buf->page = skb_frag_page(frag);
drivers/net/ethernet/broadcom/bnxt/bnxt_xdp.c
74
frag_len = skb_frag_size(frag);
drivers/net/ethernet/broadcom/bnxt/bnxt_xdp.c
77
frag_mapping = page_pool_get_dma_addr(skb_frag_page(frag)) +
drivers/net/ethernet/broadcom/bnxt/bnxt_xdp.c
78
skb_frag_off(frag);
drivers/net/ethernet/broadcom/genet/bcmgenet.c
2125
skb_frag_t *frag;
drivers/net/ethernet/broadcom/genet/bcmgenet.c
2176
frag = &skb_shinfo(skb)->frags[i - 1];
drivers/net/ethernet/broadcom/genet/bcmgenet.c
2177
size = skb_frag_size(frag);
drivers/net/ethernet/broadcom/genet/bcmgenet.c
2178
mapping = skb_frag_dma_map(kdev, frag, 0, size,
drivers/net/ethernet/broadcom/tg3.c
7837
const skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/broadcom/tg3.c
7844
skb_frag_size(frag), DMA_TO_DEVICE);
drivers/net/ethernet/broadcom/tg3.c
8144
skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/broadcom/tg3.c
8146
len = skb_frag_size(frag);
drivers/net/ethernet/broadcom/tg3.c
8147
mapping = skb_frag_dma_map(&tp->pdev->dev, frag, 0,
drivers/net/ethernet/brocade/bna/bnad.c
3027
const skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/brocade/bna/bnad.c
3028
u32 size = skb_frag_size(frag);
drivers/net/ethernet/brocade/bna/bnad.c
3050
dma_addr = skb_frag_dma_map(&bnad->pcidev->dev, frag,
drivers/net/ethernet/cadence/macb_main.c
1420
unsigned int frag;
drivers/net/ethernet/cadence/macb_main.c
1422
for (frag = begin; frag != end; frag++) {
drivers/net/ethernet/cadence/macb_main.c
1423
struct macb_dma_desc *desc = macb_rx_desc(queue, frag);
drivers/net/ethernet/cadence/macb_main.c
1532
unsigned int frag;
drivers/net/ethernet/cadence/macb_main.c
1556
for (frag = first_frag; ; frag++) {
drivers/net/ethernet/cadence/macb_main.c
1557
desc = macb_rx_desc(queue, frag);
drivers/net/ethernet/cadence/macb_main.c
1559
if (frag == last_frag)
drivers/net/ethernet/cadence/macb_main.c
1574
for (frag = first_frag; ; frag++) {
drivers/net/ethernet/cadence/macb_main.c
1578
if (unlikely(frag != last_frag)) {
drivers/net/ethernet/cadence/macb_main.c
1585
macb_rx_buffer(queue, frag),
drivers/net/ethernet/cadence/macb_main.c
1588
desc = macb_rx_desc(queue, frag);
drivers/net/ethernet/cadence/macb_main.c
1591
if (frag == last_frag)
drivers/net/ethernet/cadence/macb_main.c
2137
const skb_frag_t *frag = &skb_shinfo(skb)->frags[f];
drivers/net/ethernet/cadence/macb_main.c
2139
len = skb_frag_size(frag);
drivers/net/ethernet/cadence/macb_main.c
2145
mapping = skb_frag_dma_map(&bp->pdev->dev, frag,
drivers/net/ethernet/cadence/macb_main.c
2277
const skb_frag_t *frag = &skb_shinfo(skb)->frags[f];
drivers/net/ethernet/cadence/macb_main.c
2279
if (!IS_ALIGNED(skb_frag_size(frag), MACB_TX_LEN_ALIGN))
drivers/net/ethernet/calxeda/xgmac.c
1103
skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/calxeda/xgmac.c
1105
len = skb_frag_size(frag);
drivers/net/ethernet/calxeda/xgmac.c
1107
paddr = skb_frag_dma_map(priv->device, frag, 0, len,
drivers/net/ethernet/cavium/liquidio/lio_main.c
1446
skb_frag_t *frag = &skb_shinfo(skb)->frags[i - 1];
drivers/net/ethernet/cavium/liquidio/lio_main.c
1450
skb_frag_size(frag), DMA_TO_DEVICE);
drivers/net/ethernet/cavium/liquidio/lio_main.c
1489
skb_frag_t *frag = &skb_shinfo(skb)->frags[i - 1];
drivers/net/ethernet/cavium/liquidio/lio_main.c
1493
skb_frag_size(frag), DMA_TO_DEVICE);
drivers/net/ethernet/cavium/liquidio/lio_main.c
2394
skb_frag_t *frag;
drivers/net/ethernet/cavium/liquidio/lio_main.c
2429
frag = &skb_shinfo(skb)->frags[i - 1];
drivers/net/ethernet/cavium/liquidio/lio_main.c
2433
frag, 0, skb_frag_size(frag),
drivers/net/ethernet/cavium/liquidio/lio_main.c
2443
frag = &skb_shinfo(skb)->frags[j - 1];
drivers/net/ethernet/cavium/liquidio/lio_main.c
2446
skb_frag_size(frag),
drivers/net/ethernet/cavium/liquidio/lio_main.c
2454
add_sg_size(&g->sg[(i >> 2)], skb_frag_size(frag),
drivers/net/ethernet/cavium/liquidio/lio_vf_main.c
1492
skb_frag_t *frag;
drivers/net/ethernet/cavium/liquidio/lio_vf_main.c
1527
frag = &skb_shinfo(skb)->frags[i - 1];
drivers/net/ethernet/cavium/liquidio/lio_vf_main.c
1531
frag, 0, skb_frag_size(frag),
drivers/net/ethernet/cavium/liquidio/lio_vf_main.c
1540
frag = &skb_shinfo(skb)->frags[j - 1];
drivers/net/ethernet/cavium/liquidio/lio_vf_main.c
1543
skb_frag_size(frag),
drivers/net/ethernet/cavium/liquidio/lio_vf_main.c
1551
add_sg_size(&g->sg[(i >> 2)], skb_frag_size(frag),
drivers/net/ethernet/cavium/liquidio/lio_vf_main.c
841
skb_frag_t *frag = &skb_shinfo(skb)->frags[i - 1];
drivers/net/ethernet/cavium/liquidio/lio_vf_main.c
845
skb_frag_size(frag), DMA_TO_DEVICE);
drivers/net/ethernet/cavium/liquidio/lio_vf_main.c
885
skb_frag_t *frag = &skb_shinfo(skb)->frags[i - 1];
drivers/net/ethernet/cavium/liquidio/lio_vf_main.c
889
skb_frag_size(frag), DMA_TO_DEVICE);
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1591
const skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1594
size = skb_frag_size(frag);
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1596
skb_frag_page(frag),
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1597
skb_frag_off(frag), size,
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1665
int frag;
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1687
for (frag = 0; frag < cqe_rx->rb_cnt; frag++) {
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1688
payload_len = rb_lens[frag_num(frag)];
drivers/net/ethernet/cavium/thunder/nicvf_queues.c
1696
if (!frag) {
drivers/net/ethernet/chelsio/cxgb/sge.c
1125
const skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/chelsio/cxgb/sge.c
1126
len = skb_frag_size(frag);
drivers/net/ethernet/chelsio/cxgb/sge.c
1257
skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/chelsio/cxgb/sge.c
1267
mapping = skb_frag_dma_map(&adapter->pdev->dev, frag, 0,
drivers/net/ethernet/chelsio/cxgb/sge.c
1268
skb_frag_size(frag), DMA_TO_DEVICE);
drivers/net/ethernet/chelsio/cxgb/sge.c
1270
desc_len = skb_frag_size(frag);
drivers/net/ethernet/chelsio/cxgb/sge.c
1280
dma_unmap_len_set(ce, dma_len, skb_frag_size(frag));
drivers/net/ethernet/chelsio/cxgb3/sge.c
1020
const skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/chelsio/cxgb3/sge.c
1022
sgp->len[j] = cpu_to_be32(skb_frag_size(frag));
drivers/net/ethernet/chelsio/cxgb4/cxgb4.h
1340
uint32_t frag:FRAG_BITWIDTH; /* IP fragmentation header */
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
1076
if (fs->val.frag && !fs->mask.frag)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
1077
fs->mask.frag |= ~0;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
1228
ntuple_mask |= (u64)fs->mask.frag << tp->frag_shift;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
1287
if (tp->frag_shift >= 0 && fs->mask.frag)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
1288
ntuple |= (u64)(fs->val.frag) << tp->frag_shift;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
267
unsupported(fconf, FRAGMENTATION_F, fs->val.frag, fs->mask.frag) ||
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
877
(FW_FILTER_WR_FRAG_V(f->fs.val.frag) |
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
878
FW_FILTER_WR_FRAGM_V(f->fs.mask.frag) |
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_flower.c
864
fs->val.frag = match.key->flags & FLOW_DIS_IS_FRAGMENT;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_flower.c
865
fs->mask.frag = true;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32_parse.h
64
f->val.frag = 1;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32_parse.h
65
f->mask.frag = 1;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32_parse.h
67
f->val.frag = 0;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_u32_parse.h
68
f->mask.frag = 1;
drivers/net/ethernet/chelsio/cxgb4/sge.c
900
skb_frag_t *frag;
drivers/net/ethernet/chelsio/cxgb4/sge.c
913
frag = &si->frags[frag_idx];
drivers/net/ethernet/chelsio/cxgb4/sge.c
914
frag_size = skb_frag_size(frag);
drivers/net/ethernet/chelsio/cxgb4/sge.c
919
frag = &si->frags[frag_idx];
drivers/net/ethernet/chelsio/cxgb4/sge.c
920
frag_size = skb_frag_size(frag);
drivers/net/ethernet/chelsio/cxgb4/sge.c
923
frag_size = min(len, skb_frag_size(frag) - start);
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
1548
int frag;
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
1550
frag = gl->nfrags - 1;
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
1551
while (frag--)
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
1552
put_page(gl->frags[frag].page);
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
1779
u32 bufsz, frag;
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
1804
for (frag = 0, fp = gl.frags; /**/; frag++, fp++) {
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
1805
BUG_ON(frag >= MAX_SKB_FRAGS);
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
1817
gl.nfrags = frag+1;
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
1839
restore_rx_bufs(&gl, &rxq->fl, frag);
drivers/net/ethernet/chelsio/inline_crypto/ch_ktls/chcr_ktls.c
29
skb_frag_t *frag;
drivers/net/ethernet/chelsio/inline_crypto/ch_ktls/chcr_ktls.c
40
frag = &si->frags[frag_idx];
drivers/net/ethernet/chelsio/inline_crypto/ch_ktls/chcr_ktls.c
41
frag_size = skb_frag_size(frag);
drivers/net/ethernet/chelsio/inline_crypto/ch_ktls/chcr_ktls.c
45
frag = &si->frags[frag_idx];
drivers/net/ethernet/chelsio/inline_crypto/ch_ktls/chcr_ktls.c
46
frag_size = skb_frag_size(frag);
drivers/net/ethernet/chelsio/inline_crypto/ch_ktls/chcr_ktls.c
48
frag_size = min(len, skb_frag_size(frag) - start);
drivers/net/ethernet/cisco/enic/enic_main.c
535
const skb_frag_t *frag;
drivers/net/ethernet/cisco/enic/enic_main.c
539
for (frag = skb_shinfo(skb)->frags; len_left; frag++) {
drivers/net/ethernet/cisco/enic/enic_main.c
540
len_left -= skb_frag_size(frag);
drivers/net/ethernet/cisco/enic/enic_main.c
541
dma_addr = skb_frag_dma_map(&enic->pdev->dev, frag, 0,
drivers/net/ethernet/cisco/enic/enic_main.c
542
skb_frag_size(frag),
drivers/net/ethernet/cisco/enic/enic_main.c
546
enic_queue_wq_desc_cont(wq, skb, dma_addr, skb_frag_size(frag),
drivers/net/ethernet/cisco/enic/enic_main.c
677
skb_frag_t *frag;
drivers/net/ethernet/cisco/enic/enic_main.c
712
for (frag = skb_shinfo(skb)->frags; len_left; frag++) {
drivers/net/ethernet/cisco/enic/enic_main.c
713
len_left -= skb_frag_size(frag);
drivers/net/ethernet/cisco/enic/enic_main.c
714
frag_len_left = skb_frag_size(frag);
drivers/net/ethernet/cisco/enic/enic_main.c
720
dma_addr = skb_frag_dma_map(&enic->pdev->dev, frag,
drivers/net/ethernet/cortina/gemini.c
1139
short frag, last_frag = skb_si->nr_frags - 1;
drivers/net/ethernet/cortina/gemini.c
1223
frag = -1;
drivers/net/ethernet/cortina/gemini.c
1224
while (frag <= last_frag) {
drivers/net/ethernet/cortina/gemini.c
1225
if (frag == -1) {
drivers/net/ethernet/cortina/gemini.c
1229
skb_frag = skb_si->frags + frag;
drivers/net/ethernet/cortina/gemini.c
1234
if (frag == last_frag) {
drivers/net/ethernet/cortina/gemini.c
1253
frag++;
drivers/net/ethernet/dlink/sundance.c
1030
np->rx_ring[i].frag.length = 0;
drivers/net/ethernet/dlink/sundance.c
1051
np->rx_ring[i].frag.addr = cpu_to_le32(addr);
drivers/net/ethernet/dlink/sundance.c
1052
np->rx_ring[i].frag.length = cpu_to_le32(np->rx_buf_sz | LastFrag);
drivers/net/ethernet/dlink/sundance.c
1107
txdesc->frag.addr = cpu_to_le32(addr);
drivers/net/ethernet/dlink/sundance.c
1108
txdesc->frag.length = cpu_to_le32 (skb->len | LastFrag);
drivers/net/ethernet/dlink/sundance.c
1157
le32_to_cpu(np->tx_ring[i].frag.addr),
drivers/net/ethernet/dlink/sundance.c
1277
le32_to_cpu(np->tx_ring[entry].frag.addr),
drivers/net/ethernet/dlink/sundance.c
1281
np->tx_ring[entry].frag.addr = 0;
drivers/net/ethernet/dlink/sundance.c
1282
np->tx_ring[entry].frag.length = 0;
drivers/net/ethernet/dlink/sundance.c
1296
le32_to_cpu(np->tx_ring[entry].frag.addr),
drivers/net/ethernet/dlink/sundance.c
1300
np->tx_ring[entry].frag.addr = 0;
drivers/net/ethernet/dlink/sundance.c
1301
np->tx_ring[entry].frag.length = 0;
drivers/net/ethernet/dlink/sundance.c
1378
le32_to_cpu(desc->frag.addr),
drivers/net/ethernet/dlink/sundance.c
1382
le32_to_cpu(desc->frag.addr),
drivers/net/ethernet/dlink/sundance.c
1387
le32_to_cpu(desc->frag.addr),
drivers/net/ethernet/dlink/sundance.c
1442
np->rx_ring[entry].frag.addr = cpu_to_le32(addr);
drivers/net/ethernet/dlink/sundance.c
1445
np->rx_ring[entry].frag.length =
drivers/net/ethernet/dlink/sundance.c
1876
i, np->tx_ring[i].status, np->tx_ring[i].frag.addr,
drivers/net/ethernet/dlink/sundance.c
1877
np->tx_ring[i].frag.length);
drivers/net/ethernet/dlink/sundance.c
1882
i, np->rx_ring[i].status, np->rx_ring[i].frag.addr,
drivers/net/ethernet/dlink/sundance.c
1883
np->rx_ring[i].frag.length);
drivers/net/ethernet/dlink/sundance.c
1898
le32_to_cpu(np->rx_ring[i].frag.addr),
drivers/net/ethernet/dlink/sundance.c
1903
np->rx_ring[i].frag.addr = cpu_to_le32(0xBADF00D0); /* poison */
drivers/net/ethernet/dlink/sundance.c
1910
le32_to_cpu(np->tx_ring[i].frag.addr),
drivers/net/ethernet/dlink/sundance.c
343
struct desc_frag { __le32 addr, length; } frag;
drivers/net/ethernet/dlink/sundance.c
983
le32_to_cpu(np->tx_ring[i].frag.addr),
drivers/net/ethernet/dlink/sundance.c
984
le32_to_cpu(np->tx_ring[i].frag.length));
drivers/net/ethernet/emulex/benet/be_main.c
1015
const skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/emulex/benet/be_main.c
1016
len = skb_frag_size(frag);
drivers/net/ethernet/emulex/benet/be_main.c
1018
busaddr = skb_frag_dma_map(dev, frag, 0, len, DMA_TO_DEVICE);
drivers/net/ethernet/engleder/tsnep_main.c
450
static int tsnep_tx_map_frag(skb_frag_t *frag, struct tsnep_tx_entry *entry,
drivers/net/ethernet/engleder/tsnep_main.c
456
len = skb_frag_size(frag);
drivers/net/ethernet/engleder/tsnep_main.c
458
*dma = skb_frag_dma_map(dmadev, frag, 0, len, DMA_TO_DEVICE);
drivers/net/ethernet/engleder/tsnep_main.c
464
void *fragdata = skb_frag_address_safe(frag);
drivers/net/ethernet/engleder/tsnep_main.c
469
struct page *page = skb_frag_page(frag);
drivers/net/ethernet/engleder/tsnep_main.c
472
memcpy(&entry->desc->tx, fragdata + skb_frag_off(frag),
drivers/net/ethernet/engleder/tsnep_main.c
514
skb_frag_t *frag = &skb_shinfo(skb)->frags[i - 1];
drivers/net/ethernet/engleder/tsnep_main.c
516
len = skb_frag_size(frag);
drivers/net/ethernet/engleder/tsnep_main.c
517
mapped = tsnep_tx_map_frag(frag, entry, dmadev, &dma);
drivers/net/ethernet/engleder/tsnep_main.c
632
skb_frag_t *frag;
drivers/net/ethernet/engleder/tsnep_main.c
639
frag = NULL;
drivers/net/ethernet/engleder/tsnep_main.c
644
data = unlikely(frag) ? skb_frag_address(frag) :
drivers/net/ethernet/engleder/tsnep_main.c
652
page = unlikely(frag) ? skb_frag_page(frag) :
drivers/net/ethernet/engleder/tsnep_main.c
655
if (unlikely(frag))
drivers/net/ethernet/engleder/tsnep_main.c
656
dma += skb_frag_off(frag);
drivers/net/ethernet/engleder/tsnep_main.c
673
frag = &shinfo->frags[i];
drivers/net/ethernet/engleder/tsnep_main.c
674
len = skb_frag_size(frag);
drivers/net/ethernet/faraday/ftgmac100.c
828
skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/faraday/ftgmac100.c
830
len = skb_frag_size(frag);
drivers/net/ethernet/faraday/ftgmac100.c
833
map = skb_frag_dma_map(priv->dev, frag, 0, len,
drivers/net/ethernet/freescale/dpaa/dpaa_eth.c
2004
skb_frag_t *frag;
drivers/net/ethernet/freescale/dpaa/dpaa_eth.c
2049
frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/freescale/dpaa/dpaa_eth.c
2050
frag_len = skb_frag_size(frag);
drivers/net/ethernet/freescale/dpaa/dpaa_eth.c
2051
WARN_ON(!skb_frag_page(frag));
drivers/net/ethernet/freescale/dpaa/dpaa_eth.c
2052
addr = skb_frag_dma_map(priv->tx_dma_dev, frag, 0,
drivers/net/ethernet/freescale/dpaa/dpaa_eth.c
2156
skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/freescale/dpaa/dpaa_eth.c
2159
if (!IS_ALIGNED(skb_frag_off(frag), DPAA_A050385_ALIGN))
drivers/net/ethernet/freescale/dpaa/dpaa_eth.c
2163
if (!IS_ALIGNED(skb_frag_size(frag), DPAA_A050385_ALIGN) &&
drivers/net/ethernet/freescale/enetc/enetc.c
1722
skb_frag_t *frag;
drivers/net/ethernet/freescale/enetc/enetc.c
1749
for (f = 0, frag = &shinfo->frags[0]; f < shinfo->nr_frags;
drivers/net/ethernet/freescale/enetc/enetc.c
1750
f++, frag++) {
drivers/net/ethernet/freescale/enetc/enetc.c
1751
data = skb_frag_address(frag);
drivers/net/ethernet/freescale/enetc/enetc.c
1752
len = skb_frag_size(frag);
drivers/net/ethernet/freescale/enetc/enetc.c
1848
skb_frag_t *frag;
drivers/net/ethernet/freescale/enetc/enetc.c
1864
frag = &shinfo->frags[shinfo->nr_frags];
drivers/net/ethernet/freescale/enetc/enetc.c
1865
skb_frag_fill_page_desc(frag, rx_swbd->page, rx_swbd->page_offset,
drivers/net/ethernet/freescale/enetc/enetc.c
333
skb_frag_t *frag;
drivers/net/ethernet/freescale/enetc/enetc.c
443
frag = &skb_shinfo(skb)->frags[0];
drivers/net/ethernet/freescale/enetc/enetc.c
444
for (f = 0; f < skb_shinfo(skb)->nr_frags; f++, frag++) {
drivers/net/ethernet/freescale/enetc/enetc.c
445
len = skb_frag_size(frag);
drivers/net/ethernet/freescale/enetc/enetc.c
446
dma = skb_frag_dma_map(tx_ring->dev, frag, 0, len,
drivers/net/ethernet/freescale/enetc/enetc.c
774
skb_frag_t *frag;
drivers/net/ethernet/freescale/enetc/enetc.c
802
frag = &skb_shinfo(skb)->frags[0];
drivers/net/ethernet/freescale/enetc/enetc.c
803
for (f = 0; f < skb_shinfo(skb)->nr_frags; f++, frag++) {
drivers/net/ethernet/freescale/enetc/enetc.c
807
len = skb_frag_size(frag);
drivers/net/ethernet/freescale/enetc/enetc.c
808
dma = skb_frag_dma_map(tx_ring->dev, frag);
drivers/net/ethernet/freescale/fec_main.c
516
int frag, frag_len;
drivers/net/ethernet/freescale/fec_main.c
525
for (frag = 0; frag < nr_frags; frag++) {
drivers/net/ethernet/freescale/fec_main.c
526
this_frag = &skb_shinfo(skb)->frags[frag];
drivers/net/ethernet/freescale/fec_main.c
533
frag_len = skb_frag_size(&skb_shinfo(skb)->frags[frag]);
drivers/net/ethernet/freescale/fec_main.c
536
if (frag == nr_frags - 1) {
drivers/net/ethernet/freescale/fec_main.c
588
for (i = 0; i < frag; i++) {
drivers/net/ethernet/freescale/fs_enet/fs_enet-main.c
457
skb_frag_t *frag;
drivers/net/ethernet/freescale/fs_enet/fs_enet-main.c
466
frag = skb_shinfo(skb)->frags;
drivers/net/ethernet/freescale/fs_enet/fs_enet-main.c
467
for (i = 0; i < nr_frags; i++, frag++) {
drivers/net/ethernet/freescale/fs_enet/fs_enet-main.c
468
if (!IS_ALIGNED(skb_frag_off(frag), 4)) {
drivers/net/ethernet/freescale/fs_enet/fs_enet-main.c
519
frag = skb_shinfo(skb)->frags;
drivers/net/ethernet/freescale/fs_enet/fs_enet-main.c
534
len = skb_frag_size(frag);
drivers/net/ethernet/freescale/fs_enet/fs_enet-main.c
535
CBDW_BUFADDR(bdp, skb_frag_dma_map(fep->dev, frag, 0, len,
drivers/net/ethernet/freescale/fs_enet/fs_enet-main.c
542
frag++;
drivers/net/ethernet/freescale/gianfar.c
1756
skb_frag_t *frag;
drivers/net/ethernet/freescale/gianfar.c
1869
frag = &skb_shinfo(skb)->frags[0];
drivers/net/ethernet/freescale/gianfar.c
1870
for (i = 0; i < nr_frags; i++, frag++) {
drivers/net/ethernet/freescale/gianfar.c
1876
size = skb_frag_size(frag);
drivers/net/ethernet/freescale/gianfar.c
1885
bufaddr = skb_frag_dma_map(priv->dev, frag, 0,
drivers/net/ethernet/google/gve/gve_tx.c
701
const skb_frag_t *frag = &shinfo->frags[i];
drivers/net/ethernet/google/gve/gve_tx.c
705
len = skb_frag_size(frag);
drivers/net/ethernet/google/gve/gve_tx.c
706
addr = skb_frag_dma_map(tx->dev, frag, 0, len, DMA_TO_DEVICE);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
713
const skb_frag_t *frag = &shinfo->frags[i];
drivers/net/ethernet/google/gve/gve_tx_dqo.c
715
u32 len = skb_frag_size(frag);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
718
addr = skb_frag_dma_map(tx->dev, frag, 0, len, DMA_TO_DEVICE);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
723
netmem_dma_unmap_addr_set(skb_frag_netmem(frag), pkt,
drivers/net/ethernet/hisilicon/hix5hd2_gmac.c
722
skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/hisilicon/hix5hd2_gmac.c
723
int len = skb_frag_size(frag);
drivers/net/ethernet/hisilicon/hix5hd2_gmac.c
725
addr = skb_frag_dma_map(priv->dev, frag, 0, len, DMA_TO_DEVICE);
drivers/net/ethernet/hisilicon/hns/hns_enet.c
248
skb_frag_t *frag;
drivers/net/ethernet/hisilicon/hns/hns_enet.c
255
frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/hisilicon/hns/hns_enet.c
256
size = skb_frag_size(frag);
drivers/net/ethernet/hisilicon/hns/hns_enet.c
334
skb_frag_t *frag;
drivers/net/ethernet/hisilicon/hns/hns_enet.c
372
frag = &skb_shinfo(skb)->frags[i - 1];
drivers/net/ethernet/hisilicon/hns/hns_enet.c
373
size = skb_frag_size(frag);
drivers/net/ethernet/hisilicon/hns/hns_enet.c
374
dma = skb_frag_dma_map(dev, frag, 0, size, DMA_TO_DEVICE);
drivers/net/ethernet/hisilicon/hns/hns_enet.c
380
priv->ops.fill_desc(ring, skb_frag_page(frag), size, dma,
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
1764
skb_frag_t *frag = (skb_frag_t *)priv;
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
1766
size = skb_frag_size(frag);
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
1770
dma = skb_frag_dma_map(dev, frag, 0, size, DMA_TO_DEVICE);
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
1808
skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
1809
size = skb_frag_size(frag);
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
2042
skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
2044
ret = hns3_map_and_fill_desc(ring, frag, DESC_TYPE_PAGE);
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
3780
void *frag = napi_alloc_frag(frag_size);
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
3782
if (unlikely(!frag)) {
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
3791
memcpy(frag, desc_cb->buf + frag_offset, frag_size);
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
3792
skb_add_rx_frag(skb, i, virt_to_page(frag),
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
3793
offset_in_page(frag), frag_size, frag_size);
drivers/net/ethernet/huawei/hinic/hinic_tx.c
138
skb_frag_t *frag;
drivers/net/ethernet/huawei/hinic/hinic_tx.c
152
frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/huawei/hinic/hinic_tx.c
154
dma_addr = skb_frag_dma_map(&pdev->dev, frag, 0,
drivers/net/ethernet/huawei/hinic/hinic_tx.c
155
skb_frag_size(frag),
drivers/net/ethernet/huawei/hinic/hinic_tx.c
162
hinic_set_sge(&sges[i + 1], dma_addr, skb_frag_size(frag));
drivers/net/ethernet/huawei/hinic3/hinic3_rx.c
301
skb_frag_t *frag = &skb_shinfo(skb)->frags[0];
drivers/net/ethernet/huawei/hinic3/hinic3_rx.c
305
va = skb_frag_address(frag);
drivers/net/ethernet/huawei/hinic3/hinic3_rx.c
316
skb_frag_size_sub(frag, pull_len);
drivers/net/ethernet/huawei/hinic3/hinic3_rx.c
317
skb_frag_off_add(frag, pull_len);
drivers/net/ethernet/huawei/hinic3/hinic3_tx.c
111
frag = &(skb_shinfo(skb)->frags[i]);
drivers/net/ethernet/huawei/hinic3/hinic3_tx.c
116
dma_info[idx].dma = skb_frag_dma_map(&pdev->dev, frag, 0,
drivers/net/ethernet/huawei/hinic3/hinic3_tx.c
117
skb_frag_size(frag),
drivers/net/ethernet/huawei/hinic3/hinic3_tx.c
123
dma_info[idx].len = skb_frag_size(frag);
drivers/net/ethernet/huawei/hinic3/hinic3_tx.c
94
skb_frag_t *frag;
drivers/net/ethernet/ibm/ehea/ehea_main.c
1645
skb_frag_t *frag;
drivers/net/ethernet/ibm/ehea/ehea_main.c
1659
frag = &skb_shinfo(skb)->frags[0];
drivers/net/ethernet/ibm/ehea/ehea_main.c
1663
sg1entry->len = skb_frag_size(frag);
drivers/net/ethernet/ibm/ehea/ehea_main.c
1665
ehea_map_vaddr(skb_frag_address(frag));
drivers/net/ethernet/ibm/ehea/ehea_main.c
1672
frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/ibm/ehea/ehea_main.c
1676
sgentry->len = skb_frag_size(frag);
drivers/net/ethernet/ibm/ehea/ehea_main.c
1677
sgentry->vaddr = ehea_map_vaddr(skb_frag_address(frag));
drivers/net/ethernet/ibm/emac/core.c
1530
skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/ibm/emac/core.c
1531
len = skb_frag_size(frag);
drivers/net/ethernet/ibm/emac/core.c
1536
pd = skb_frag_dma_map(&dev->ofdev->dev, frag, 0, len,
drivers/net/ethernet/ibm/ibmveth.c
1299
const skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/ibm/ibmveth.c
1302
skb_frag_address_safe(frag), skb_frag_size(frag));
drivers/net/ethernet/ibm/ibmveth.c
1303
total_bytes += skb_frag_size(frag);
drivers/net/ethernet/ibm/ibmvnic.c
2519
const skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/ibm/ibmvnic.c
2521
memcpy(dst + cur, skb_frag_address(frag),
drivers/net/ethernet/ibm/ibmvnic.c
2522
skb_frag_size(frag));
drivers/net/ethernet/ibm/ibmvnic.c
2523
cur += skb_frag_size(frag);
drivers/net/ethernet/intel/e1000/e1000_main.c
2895
const skb_frag_t *frag = &skb_shinfo(skb)->frags[f];
drivers/net/ethernet/intel/e1000/e1000_main.c
2897
len = skb_frag_size(frag);
drivers/net/ethernet/intel/e1000/e1000_main.c
2919
page_to_phys(skb_frag_page(frag));
drivers/net/ethernet/intel/e1000/e1000_main.c
2929
buffer_info->dma = skb_frag_dma_map(&pdev->dev, frag,
drivers/net/ethernet/intel/e1000e/netdev.c
5612
const skb_frag_t *frag = &skb_shinfo(skb)->frags[f];
drivers/net/ethernet/intel/e1000e/netdev.c
5614
len = skb_frag_size(frag);
drivers/net/ethernet/intel/e1000e/netdev.c
5628
buffer_info->dma = skb_frag_dma_map(&pdev->dev, frag,
drivers/net/ethernet/intel/fm10k/fm10k_main.c
1066
skb_frag_t *frag = &skb_shinfo(skb)->frags[f];
drivers/net/ethernet/intel/fm10k/fm10k_main.c
1068
count += TXD_USE_COUNT(skb_frag_size(frag));
drivers/net/ethernet/intel/fm10k/fm10k_main.c
938
skb_frag_t *frag;
drivers/net/ethernet/intel/fm10k/fm10k_main.c
962
for (frag = &skb_shinfo(skb)->frags[0];; frag++) {
drivers/net/ethernet/intel/fm10k/fm10k_main.c
990
size = skb_frag_size(frag);
drivers/net/ethernet/intel/fm10k/fm10k_main.c
993
dma = skb_frag_dma_map(tx_ring->dev, frag, 0, size,
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3489
const skb_frag_t *frag, *stale;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3501
frag = &skb_shinfo(skb)->frags[0];
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3512
sum += skb_frag_size(frag++);
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3513
sum += skb_frag_size(frag++);
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3514
sum += skb_frag_size(frag++);
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3515
sum += skb_frag_size(frag++);
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3516
sum += skb_frag_size(frag++);
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3524
sum += skb_frag_size(frag++);
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3576
skb_frag_t *frag;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3596
for (frag = &skb_shinfo(skb)->frags[0];; frag++) {
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3646
size = skb_frag_size(frag);
drivers/net/ethernet/intel/i40e/i40e_txrx.c
3649
dma = skb_frag_dma_map(tx_ring->dev, frag, 0, size,
drivers/net/ethernet/intel/i40e/i40e_txrx.h
506
const skb_frag_t *frag = &skb_shinfo(skb)->frags[0];
drivers/net/ethernet/intel/i40e/i40e_txrx.h
516
size = skb_frag_size(frag++);
drivers/net/ethernet/intel/i40e/i40e_xsk.c
322
skb_frag_t *frag = &sinfo->frags[i];
drivers/net/ethernet/intel/i40e/i40e_xsk.c
333
memcpy(addr, skb_frag_page(frag), skb_frag_size(frag));
drivers/net/ethernet/intel/i40e/i40e_xsk.c
336
addr, 0, skb_frag_size(frag));
drivers/net/ethernet/intel/iavf/iavf_txrx.c
2048
const skb_frag_t *frag, *stale;
drivers/net/ethernet/intel/iavf/iavf_txrx.c
2060
frag = &skb_shinfo(skb)->frags[0];
drivers/net/ethernet/intel/iavf/iavf_txrx.c
2071
sum += skb_frag_size(frag++);
drivers/net/ethernet/intel/iavf/iavf_txrx.c
2072
sum += skb_frag_size(frag++);
drivers/net/ethernet/intel/iavf/iavf_txrx.c
2073
sum += skb_frag_size(frag++);
drivers/net/ethernet/intel/iavf/iavf_txrx.c
2074
sum += skb_frag_size(frag++);
drivers/net/ethernet/intel/iavf/iavf_txrx.c
2075
sum += skb_frag_size(frag++);
drivers/net/ethernet/intel/iavf/iavf_txrx.c
2083
sum += skb_frag_size(frag++);
drivers/net/ethernet/intel/iavf/iavf_txrx.c
2156
skb_frag_t *frag;
drivers/net/ethernet/intel/iavf/iavf_txrx.c
2175
for (frag = &skb_shinfo(skb)->frags[0];; frag++) {
drivers/net/ethernet/intel/iavf/iavf_txrx.c
2223
size = skb_frag_size(frag);
drivers/net/ethernet/intel/iavf/iavf_txrx.c
2226
dma = skb_frag_dma_map(tx_ring->dev, frag, 0, size,
drivers/net/ethernet/intel/iavf/iavf_txrx.h
333
const skb_frag_t *frag = &skb_shinfo(skb)->frags[0];
drivers/net/ethernet/intel/iavf/iavf_txrx.h
343
size = skb_frag_size(frag++);
drivers/net/ethernet/intel/ice/ice_fdir.c
859
u8 *pkt, bool frag, bool tun)
drivers/net/ethernet/intel/ice/ice_fdir.c
951
if (frag)
drivers/net/ethernet/intel/ice/ice_fdir.h
223
u8 *pkt, bool frag, bool tun);
drivers/net/ethernet/intel/ice/ice_txrx.c
1406
skb_frag_t *frag;
drivers/net/ethernet/intel/ice/ice_txrx.c
1429
for (frag = &skb_shinfo(skb)->frags[0];; frag++) {
drivers/net/ethernet/intel/ice/ice_txrx.c
1480
size = skb_frag_size(frag);
drivers/net/ethernet/intel/ice/ice_txrx.c
1483
dma = skb_frag_dma_map(tx_ring->dev, frag, 0, size,
drivers/net/ethernet/intel/ice/ice_txrx.c
1986
const skb_frag_t *frag = &skb_shinfo(skb)->frags[0];
drivers/net/ethernet/intel/ice/ice_txrx.c
1996
size = skb_frag_size(frag++);
drivers/net/ethernet/intel/ice/ice_txrx.c
2017
const skb_frag_t *frag, *stale;
drivers/net/ethernet/intel/ice/ice_txrx.c
2029
frag = &skb_shinfo(skb)->frags[0];
drivers/net/ethernet/intel/ice/ice_txrx.c
2040
sum += skb_frag_size(frag++);
drivers/net/ethernet/intel/ice/ice_txrx.c
2041
sum += skb_frag_size(frag++);
drivers/net/ethernet/intel/ice/ice_txrx.c
2042
sum += skb_frag_size(frag++);
drivers/net/ethernet/intel/ice/ice_txrx.c
2043
sum += skb_frag_size(frag++);
drivers/net/ethernet/intel/ice/ice_txrx.c
2044
sum += skb_frag_size(frag++);
drivers/net/ethernet/intel/ice/ice_txrx.c
2052
sum += skb_frag_size(frag++);
drivers/net/ethernet/intel/ice/ice_txrx_lib.c
383
u32 frag = 0;
drivers/net/ethernet/intel/ice/ice_txrx_lib.c
433
if (frag == nr_frags)
drivers/net/ethernet/intel/ice/ice_txrx_lib.c
439
page = skb_frag_page(&sinfo->frags[frag]);
drivers/net/ethernet/intel/ice/ice_txrx_lib.c
440
offset = skb_frag_off(&sinfo->frags[frag]);
drivers/net/ethernet/intel/ice/ice_txrx_lib.c
441
data = skb_frag_address(&sinfo->frags[frag]);
drivers/net/ethernet/intel/ice/ice_txrx_lib.c
442
size = skb_frag_size(&sinfo->frags[frag]);
drivers/net/ethernet/intel/ice/ice_txrx_lib.c
443
frag++;
drivers/net/ethernet/intel/ice/ice_xsk.c
444
u32 frag = 0;
drivers/net/ethernet/intel/ice/ice_xsk.c
480
if (frag == nr_frags)
drivers/net/ethernet/intel/ice/ice_xsk.c
487
size = skb_frag_size(&sinfo->frags[frag]);
drivers/net/ethernet/intel/ice/ice_xsk.c
488
frag++;
drivers/net/ethernet/intel/idpf/idpf_lib.c
2383
const skb_frag_t *frag, *stale;
drivers/net/ethernet/intel/idpf/idpf_lib.c
2395
frag = &shinfo->frags[0];
drivers/net/ethernet/intel/idpf/idpf_lib.c
2406
sum += skb_frag_size(frag++);
drivers/net/ethernet/intel/idpf/idpf_lib.c
2407
sum += skb_frag_size(frag++);
drivers/net/ethernet/intel/idpf/idpf_lib.c
2408
sum += skb_frag_size(frag++);
drivers/net/ethernet/intel/idpf/idpf_lib.c
2409
sum += skb_frag_size(frag++);
drivers/net/ethernet/intel/idpf/idpf_lib.c
2410
sum += skb_frag_size(frag++);
drivers/net/ethernet/intel/idpf/idpf_lib.c
2418
sum += skb_frag_size(frag++);
drivers/net/ethernet/intel/idpf/idpf_singleq_txrx.c
255
skb_frag_t *frag;
drivers/net/ethernet/intel/idpf/idpf_singleq_txrx.c
270
for (frag = &skb_shinfo(skb)->frags[0];; frag++) {
drivers/net/ethernet/intel/idpf/idpf_singleq_txrx.c
327
size = skb_frag_size(frag);
drivers/net/ethernet/intel/idpf/idpf_singleq_txrx.c
330
dma = skb_frag_dma_map(tx_q->dev, frag, 0, size,
drivers/net/ethernet/intel/idpf/idpf_txrx.c
2652
skb_frag_t *frag;
drivers/net/ethernet/intel/idpf/idpf_txrx.c
2671
for (frag = &skb_shinfo(skb)->frags[0];; frag++) {
drivers/net/ethernet/intel/idpf/idpf_txrx.c
2785
size = skb_frag_size(frag);
drivers/net/ethernet/intel/idpf/idpf_txrx.c
2788
dma = skb_frag_dma_map(tx_q->dev, frag, 0, size,
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
3027
bool ipv4, bool frag)
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
3037
if (frag) {
drivers/net/ethernet/intel/idpf/idpf_virtchnl.c
3050
if (frag)
drivers/net/ethernet/intel/igb/igb_main.c
6280
skb_frag_t *frag;
drivers/net/ethernet/intel/igb/igb_main.c
6298
for (frag = &skb_shinfo(skb)->frags[0];; frag++) {
drivers/net/ethernet/intel/igb/igb_main.c
6339
size = skb_frag_size(frag);
drivers/net/ethernet/intel/igb/igb_main.c
6342
dma = skb_frag_dma_map(tx_ring->dev, frag, 0,
drivers/net/ethernet/intel/igbvf/netdev.c
2160
const skb_frag_t *frag;
drivers/net/ethernet/intel/igbvf/netdev.c
2167
frag = &skb_shinfo(skb)->frags[f];
drivers/net/ethernet/intel/igbvf/netdev.c
2168
len = skb_frag_size(frag);
drivers/net/ethernet/intel/igbvf/netdev.c
2175
buffer_info->dma = skb_frag_dma_map(&pdev->dev, frag, 0, len,
drivers/net/ethernet/intel/igc/igc_main.c
1321
skb_frag_t *frag;
drivers/net/ethernet/intel/igc/igc_main.c
1339
for (frag = &skb_shinfo(skb)->frags[0];; frag++) {
drivers/net/ethernet/intel/igc/igc_main.c
1380
size = skb_frag_size(frag);
drivers/net/ethernet/intel/igc/igc_main.c
1383
dma = skb_frag_dma_map(tx_ring->dev, frag, 0,
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
1997
skb_frag_t *frag = &skb_shinfo(skb)->frags[0];
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
2006
va = skb_frag_address(frag);
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
2018
skb_frag_size_sub(frag, pull_len);
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
2019
skb_frag_off_add(frag, pull_len);
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
2047
skb_frag_t *frag = &skb_shinfo(skb)->frags[0];
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
2051
skb_frag_off(frag),
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
2052
skb_frag_size(frag),
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
8951
skb_frag_t *frag;
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
8980
for (frag = &skb_shinfo(skb)->frags[0];; frag++) {
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
9022
size = min_t(unsigned int, data_len, skb_frag_size(frag));
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
9024
size = skb_frag_size(frag);
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
9028
dma = skb_frag_dma_map(tx_ring->dev, frag, 0, size,
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
4001
skb_frag_t *frag;
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
4019
for (frag = &skb_shinfo(skb)->frags[0];; frag++) {
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
4060
size = skb_frag_size(frag);
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
4063
dma = skb_frag_dma_map(tx_ring->dev, frag, 0, size,
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
4187
skb_frag_t *frag = &skb_shinfo(skb)->frags[f];
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
4189
count += TXD_USE_COUNT(skb_frag_size(frag));
drivers/net/ethernet/jme.c
2013
const skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/jme.c
2019
skb_frag_page(frag), skb_frag_off(frag),
drivers/net/ethernet/jme.c
2020
skb_frag_size(frag), hidma);
drivers/net/ethernet/marvell/mv643xx_eth.c
658
int frag;
drivers/net/ethernet/marvell/mv643xx_eth.c
660
for (frag = 0; frag < skb_shinfo(skb)->nr_frags; frag++) {
drivers/net/ethernet/marvell/mv643xx_eth.c
661
const skb_frag_t *fragp = &skb_shinfo(skb)->frags[frag];
drivers/net/ethernet/marvell/mv643xx_eth.c
895
int frag;
drivers/net/ethernet/marvell/mv643xx_eth.c
897
for (frag = 0; frag < nr_frags; frag++) {
drivers/net/ethernet/marvell/mv643xx_eth.c
902
this_frag = &skb_shinfo(skb)->frags[frag];
drivers/net/ethernet/marvell/mv643xx_eth.c
913
if (frag == nr_frags - 1) {
drivers/net/ethernet/marvell/mvneta.c
2108
skb_frag_t *frag = NULL;
drivers/net/ethernet/marvell/mvneta.c
2113
frag = &sinfo->frags[i - 1];
drivers/net/ethernet/marvell/mvneta.c
2114
len = skb_frag_size(frag);
drivers/net/ethernet/marvell/mvneta.c
2122
data = unlikely(frag) ? skb_frag_address(frag)
drivers/net/ethernet/marvell/mvneta.c
2133
page = unlikely(frag) ? skb_frag_page(frag)
drivers/net/ethernet/marvell/mvneta.c
2136
if (unlikely(frag))
drivers/net/ethernet/marvell/mvneta.c
2137
dma_addr += skb_frag_off(frag);
drivers/net/ethernet/marvell/mvneta.c
2377
skb_frag_t *frag = &sinfo->frags[sinfo->nr_frags++];
drivers/net/ethernet/marvell/mvneta.c
2379
skb_frag_fill_page_desc(frag, page,
drivers/net/ethernet/marvell/mvneta.c
2873
skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/marvell/mvneta.c
2874
void *addr = skb_frag_address(frag);
drivers/net/ethernet/marvell/mvneta.c
2877
tx_desc->data_size = skb_frag_size(frag);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
4199
skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
4200
void *addr = skb_frag_address(frag);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
4205
mvpp2_txdesc_size_set(port, tx_desc, skb_frag_size(frag));
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
4208
skb_frag_size(frag),
drivers/net/ethernet/marvell/octeon_ep/octep_main.c
867
skb_frag_t *frag;
drivers/net/ethernet/marvell/octeon_ep/octep_main.c
926
frag = &shinfo->frags[0];
drivers/net/ethernet/marvell/octeon_ep/octep_main.c
928
len = skb_frag_size(frag);
drivers/net/ethernet/marvell/octeon_ep/octep_main.c
929
dma = skb_frag_dma_map(iq->dev, frag, 0,
drivers/net/ethernet/marvell/octeon_ep/octep_main.c
937
frag++;
drivers/net/ethernet/marvell/octeon_ep/octep_pfvf_mbox.c
334
if (cmd.s_data.frag != OCTEP_PFVF_MBOX_MORE_FRAG_FLAG) {
drivers/net/ethernet/marvell/octeon_ep/octep_pfvf_mbox.h
107
u64 frag:1;
drivers/net/ethernet/marvell/octeon_ep_vf/octep_vf_main.c
622
skb_frag_t *frag;
drivers/net/ethernet/marvell/octeon_ep_vf/octep_vf_main.c
682
frag = &shinfo->frags[0];
drivers/net/ethernet/marvell/octeon_ep_vf/octep_vf_main.c
684
len = skb_frag_size(frag);
drivers/net/ethernet/marvell/octeon_ep_vf/octep_vf_main.c
685
dma = skb_frag_dma_map(iq->dev, frag, 0,
drivers/net/ethernet/marvell/octeon_ep_vf/octep_vf_main.c
693
frag++;
drivers/net/ethernet/marvell/octeon_ep_vf/octep_vf_mbox.c
185
cmd.s_data.frag = 0;
drivers/net/ethernet/marvell/octeon_ep_vf/octep_vf_mbox.c
201
cmd.s_data.frag = 1;
drivers/net/ethernet/marvell/octeon_ep_vf/octep_vf_mbox.c
226
cmd.s_data.frag = 1;
drivers/net/ethernet/marvell/octeon_ep_vf/octep_vf_mbox.h
97
u64 frag:1;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
2119
const skb_frag_t *frag;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
2135
frag = &skb_shinfo(skb)->frags[seg - 1];
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
2136
page = skb_frag_page(frag);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
2137
offset = skb_frag_off(frag);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_common.c
2138
*len = skb_frag_size(frag);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
884
const skb_frag_t *frag;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
890
frag = &skb_shinfo(skb)->frags[seg];
drivers/net/ethernet/marvell/octeontx2/nic/otx2_txrx.c
891
offset = seg_addr - (u64)skb_frag_address(frag);
drivers/net/ethernet/marvell/skge.c
2795
const skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/marvell/skge.c
2797
map = skb_frag_dma_map(&hw->pdev->dev, frag, 0,
drivers/net/ethernet/marvell/skge.c
2798
skb_frag_size(frag), DMA_TO_DEVICE);
drivers/net/ethernet/marvell/skge.c
2810
dma_unmap_len_set(e, maplen, skb_frag_size(frag));
drivers/net/ethernet/marvell/skge.c
2812
tf->control = BMU_OWN | BMU_SW | control | skb_frag_size(frag);
drivers/net/ethernet/marvell/sky2.c
1221
const skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/marvell/sky2.c
1223
re->frag_addr[i] = skb_frag_dma_map(&pdev->dev, frag, 0,
drivers/net/ethernet/marvell/sky2.c
1224
skb_frag_size(frag),
drivers/net/ethernet/marvell/sky2.c
1934
const skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/marvell/sky2.c
1936
mapping = skb_frag_dma_map(&hw->pdev->dev, frag, 0,
drivers/net/ethernet/marvell/sky2.c
1937
skb_frag_size(frag), DMA_TO_DEVICE);
drivers/net/ethernet/marvell/sky2.c
1953
dma_unmap_len_set(re, maplen, skb_frag_size(frag));
drivers/net/ethernet/marvell/sky2.c
1957
le->length = cpu_to_le16(skb_frag_size(frag));
drivers/net/ethernet/marvell/sky2.c
2500
skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/marvell/sky2.c
2504
__skb_frag_unref(frag, false);
drivers/net/ethernet/marvell/sky2.c
2509
skb_frag_size_set(frag, size);
drivers/net/ethernet/mediatek/mtk_eth_soc.c
1610
skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/mediatek/mtk_eth_soc.c
1612
int frag_size = skb_frag_size(frag);
drivers/net/ethernet/mediatek/mtk_eth_soc.c
1635
txd_info.addr = skb_frag_dma_map(eth->dma_dev, frag,
drivers/net/ethernet/mediatek/mtk_eth_soc.c
1715
skb_frag_t *frag;
drivers/net/ethernet/mediatek/mtk_eth_soc.c
1719
frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/mediatek/mtk_eth_soc.c
1720
nfrags += DIV_ROUND_UP(skb_frag_size(frag),
drivers/net/ethernet/mellanox/mlx4/en_rx.c
84
struct mlx4_en_rx_alloc *frag)
drivers/net/ethernet/mellanox/mlx4/en_rx.c
86
if (frag->page)
drivers/net/ethernet/mellanox/mlx4/en_rx.c
87
page_pool_put_full_page(ring->pp, frag->page, false);
drivers/net/ethernet/mellanox/mlx4/en_rx.c
91
memset(frag, 0, sizeof(*frag));
drivers/net/ethernet/mellanox/mlx4/en_tx.c
815
const skb_frag_t *frag = &shinfo->frags[i_frag];
drivers/net/ethernet/mellanox/mlx4/en_tx.c
816
byte_count = skb_frag_size(frag);
drivers/net/ethernet/mellanox/mlx4/en_tx.c
817
dma = skb_frag_dma_map(ddev, frag,
drivers/net/ethernet/mellanox/mlx5/core/alloc.c
87
struct mlx5_buf_list *frag = &buf->frags[i];
drivers/net/ethernet/mellanox/mlx5/core/alloc.c
90
frag->buf = mlx5_dma_zalloc_coherent_node(dev, frag_sz,
drivers/net/ethernet/mellanox/mlx5/core/alloc.c
91
&frag->map, node);
drivers/net/ethernet/mellanox/mlx5/core/alloc.c
92
if (!frag->buf)
drivers/net/ethernet/mellanox/mlx5/core/alloc.c
94
if (frag->map & ((1 << buf->page_shift) - 1)) {
drivers/net/ethernet/mellanox/mlx5/core/alloc.c
98
&frag->map, buf->page_shift);
drivers/net/ethernet/mellanox/mlx5/core/diag/fs_tracepoint.c
152
PRINT_MASKED_VAL_L2(u8, frag, frag, p, "%d");
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
938
MLX5_SET(fte_match_set_lyr_2_4, headers_c, frag, 1);
drivers/net/ethernet/mellanox/mlx5/core/en/tc_tun.c
939
MLX5_SET(fte_match_set_lyr_2_4, headers_v, frag, 0);
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
133
skb_frag_t *frag = &xdptxdf.sinfo->frags[i];
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
137
addr = page_pool_get_dma_addr(skb_frag_page(frag)) +
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
138
skb_frag_off(frag);
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
139
len = skb_frag_size(frag);
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
162
skb_frag_t *frag = &xdptxdf.sinfo->frags[i];
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
166
{ .page.page = skb_frag_page(frag) });
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
473
skb_frag_t *frag = &xdptxdf->sinfo->frags[0];
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
475
tmp.data = skb_frag_address(frag);
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
476
tmp.len = skb_frag_size(frag);
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
478
page_pool_get_dma_addr(skb_frag_page(frag)) +
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
479
skb_frag_off(frag);
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
623
skb_frag_t *frag = &xdptxdf->sinfo->frags[i];
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
627
page_pool_get_dma_addr(skb_frag_page(frag)) +
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
628
skb_frag_off(frag);
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
631
dseg->byte_count = cpu_to_be32(skb_frag_size(frag));
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
685
skb_frag_t *frag = &sinfo->frags[j];
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
691
skb_frag_size(frag), DMA_TO_DEVICE);
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
888
skb_frag_t *frag = &xdptxdf.sinfo->frags[j];
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
890
dma_arr[j] = dma_map_single(sq->pdev, skb_frag_address(frag),
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
891
skb_frag_size(frag), DMA_TO_DEVICE);
drivers/net/ethernet/mellanox/mlx5/core/en/xsk/rx.c
185
struct mlx5e_wqe_frag_info *frag;
drivers/net/ethernet/mellanox/mlx5/core/en/xsk/rx.c
191
frag = &rq->wqe.frags[j];
drivers/net/ethernet/mellanox/mlx5/core/en/xsk/rx.c
193
addr = xsk_buff_xdp_get_frame_dma(*frag->xskp);
drivers/net/ethernet/mellanox/mlx5/core/en/xsk/rx.c
195
frag->flags &= ~BIT(MLX5E_WQE_FRAG_SKIP_RELEASE);
drivers/net/ethernet/mellanox/mlx5/core/en/xsk/rx.c
208
struct mlx5e_wqe_frag_info *frag;
drivers/net/ethernet/mellanox/mlx5/core/en/xsk/rx.c
214
frag = &rq->wqe.frags[j];
drivers/net/ethernet/mellanox/mlx5/core/en/xsk/rx.c
216
*frag->xskp = xsk_buff_alloc(rq->xsk_pool);
drivers/net/ethernet/mellanox/mlx5/core/en/xsk/rx.c
217
if (unlikely(!*frag->xskp))
drivers/net/ethernet/mellanox/mlx5/core/en/xsk/rx.c
220
addr = xsk_buff_xdp_get_frame_dma(*frag->xskp);
drivers/net/ethernet/mellanox/mlx5/core/en/xsk/rx.c
222
frag->flags &= ~BIT(MLX5E_WQE_FRAG_SKIP_RELEASE);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1610
MLX5_SET_TO_ONES(fte_match_param, spec->match_criteria, outer_headers.frag);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1611
MLX5_SET(fte_match_param, spec->match_value, outer_headers.frag, 0);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ktls_tx.c
661
skb_frag_t *frag = &record->frags[i];
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ktls_tx.c
663
page_ref_inc(skb_frag_page(frag));
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ktls_tx.c
664
remaining -= skb_frag_size(frag);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ktls_tx.c
665
info->frags[i++] = *frag;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ktls_tx.c
715
tx_post_resync_dump(struct mlx5e_txqsq *sq, skb_frag_t *frag, u32 tisn)
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ktls_tx.c
738
fsz = skb_frag_size(frag);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ktls_tx.c
739
dma_addr = skb_frag_dma_map(sq->pdev, frag, 0, fsz,
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ktls_tx.c
747
mlx5e_dma_push_netmem(sq, skb_frag_netmem(frag), dma_addr, fsz);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ktls_tx.c
749
tx_fill_wi(sq, pi, MLX5E_KTLS_DUMP_WQEBBS, fsz, skb_frag_page(frag));
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
535
struct mlx5e_wqe_frag_info *frag =
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
539
for (f = 0; f < rq->wqe.info.num_frags; f++, frag++) {
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
547
*frag = next_frag;
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
551
prev = frag;
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
306
struct mlx5e_wqe_frag_info *frag)
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
310
if (!frag->offset)
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
317
frag->frag_page);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
322
static bool mlx5e_frag_can_release(struct mlx5e_wqe_frag_info *frag)
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
329
return (frag->flags & CAN_RELEASE_MASK) == CAN_RELEASE_VALUE;
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
333
struct mlx5e_wqe_frag_info *frag)
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
335
if (mlx5e_frag_can_release(frag))
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
336
mlx5e_page_release_fragmented(rq->page_pool, frag->frag_page);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
347
struct mlx5e_wqe_frag_info *frag = get_frag(rq, ix);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
351
for (i = 0; i < rq->wqe.info.num_frags; i++, frag++) {
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
355
err = mlx5e_get_rx_frag(rq, frag);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
359
frag->flags &= ~BIT(MLX5E_WQE_FRAG_SKIP_RELEASE);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
362
addr = page_pool_get_dma_addr_netmem(frag->frag_page->netmem);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
363
wqe->data[i].addr = cpu_to_be64(addr + frag->offset + headroom);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
370
mlx5e_put_rx_frag(rq, --frag);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
488
struct mlx5e_wqe_frag_info *frag;
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
490
frag = get_frag(rq, j);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
491
for (int k = 0; k < rq->wqe.info.num_frags; k++, frag++)
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
492
frag->flags |= BIT(MLX5E_WQE_FRAG_SKIP_RELEASE);
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
504
skb_frag_t *frag;
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
518
frag = &sinfo->frags[sinfo->nr_frags++];
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
519
skb_frag_fill_netmem_desc(frag, netmem, frag_offset, len);
drivers/net/ethernet/mellanox/mlx5/core/en_tc.c
2929
MLX5_SET(fte_match_set_lyr_2_4, headers_c, frag, 1);
drivers/net/ethernet/mellanox/mlx5/core/en_tc.c
2930
MLX5_SET(fte_match_set_lyr_2_4, headers_v, frag,
drivers/net/ethernet/mellanox/mlx5/core/en_tx.c
205
skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/mellanox/mlx5/core/en_tx.c
206
int fsz = skb_frag_size(frag);
drivers/net/ethernet/mellanox/mlx5/core/en_tx.c
208
dma_addr = skb_frag_dma_map(sq->pdev, frag, 0, fsz,
drivers/net/ethernet/mellanox/mlx5/core/en_tx.c
217
mlx5e_dma_push_netmem(sq, skb_frag_netmem(frag), dma_addr, fsz);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/definer.c
1047
if (HWS_IS_FLD_SET(match_param, inner_headers.frag)) {
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/definer.c
1050
inner_headers.frag, eth_l2_inner.ip_fragmented);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/definer.c
1058
inner_headers.frag, eth_l4_inner.ip_fragmented);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/definer.c
1061
inner_headers.frag, eth_l2_src_inner.ip_fragmented);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/definer.c
871
if (HWS_IS_FLD_SET(match_param, outer_headers.frag)) {
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/definer.c
878
outer_headers.frag, eth_l4_outer.ip_fragmented);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/definer.c
881
outer_headers.frag, eth_l2_src_outer.ip_fragmented);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
18
return (spec->ip_protocol || spec->frag || spec->tcp_flags ||
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
879
spec->frag = IFC_GET_CLR(fte_match_set_lyr_2_4, mask, frag, clr);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1078
DR_STE_SET_TAG(eth_l2_tnl, bit_mask, ip_fragmented, mask, frag);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1107
DR_STE_SET_TAG(eth_l2_tnl, tag, ip_fragmented, spec, frag);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1188
DR_STE_SET_TAG(eth_l4, tag, fragmented, spec, frag);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
859
DR_STE_SET_TAG(eth_l3_ipv4_5_tuple, tag, fragmented, spec, frag);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
892
DR_STE_SET_TAG(eth_l2_src, bit_mask, ip_fragmented, mask, frag);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
943
DR_STE_SET_TAG(eth_l2_src, tag, ip_fragmented, spec, frag);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1097
DR_STE_SET_TAG(eth_l3_ipv4_5_tuple_v1, tag, fragmented, spec, frag);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1128
DR_STE_SET_TAG(eth_l2_src_v1, bit_mask, ip_fragmented, mask, frag);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1178
DR_STE_SET_TAG(eth_l2_src_v1, tag, ip_fragmented, spec, frag);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1304
DR_STE_SET_TAG(eth_l2_tnl_v1, bit_mask, ip_fragmented, mask, frag);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1332
DR_STE_SET_TAG(eth_l2_tnl_v1, tag, ip_fragmented, spec, frag);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1407
DR_STE_SET_TAG(eth_l4_v1, tag, fragmented, spec, frag);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
566
u32 frag:1; /* Packet is an IP fragment */
drivers/net/ethernet/mellanox/mlxsw/pci.c
2199
const skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/mellanox/mlxsw/pci.c
2202
skb_frag_address(frag),
drivers/net/ethernet/mellanox/mlxsw/pci.c
2203
skb_frag_size(frag),
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
1076
skb_frag_t *frag = NULL;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
1085
frag = &shinfo->frags[0];
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
1120
offset = skb_frag_off(frag);
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
1121
page = skb_frag_page(frag);
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
1124
size = skb_frag_size(frag);
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
1126
frag++;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
319
skb_frag_t *frag;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
339
for (frag = &skb_shinfo(skb)->frags[0];; frag++) {
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
358
size = skb_frag_size(frag);
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
364
is_net_iov = skb_frag_is_net_iov(frag);
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
365
dma = skb_frag_dma_map(dev, frag, 0, size, DMA_TO_DEVICE);
drivers/net/ethernet/micrel/ksz884x.c
4284
int frag;
drivers/net/ethernet/micrel/ksz884x.c
4294
frag = 0;
drivers/net/ethernet/micrel/ksz884x.c
4296
this_frag = &skb_shinfo(skb)->frags[frag];
drivers/net/ethernet/micrel/ksz884x.c
4314
frag++;
drivers/net/ethernet/micrel/ksz884x.c
4315
if (frag == last_frag)
drivers/net/ethernet/microchip/lan743x_main.c
2024
const skb_frag_t *frag = &(skb_shinfo(skb)->frags[j]);
drivers/net/ethernet/microchip/lan743x_main.c
2026
if (lan743x_tx_frame_add_fragment(tx, frag, frame_length)) {
drivers/net/ethernet/microsoft/mana/mana_en.c
194
skb_frag_t *frag;
drivers/net/ethernet/microsoft/mana/mana_en.c
229
frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/microsoft/mana/mana_en.c
230
da = skb_frag_dma_map(dev, frag, 0, skb_frag_size(frag),
drivers/net/ethernet/microsoft/mana/mana_en.c
235
mana_add_sge(tp, ash, sg_i, da, skb_frag_size(frag),
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1289
skb_frag_t *frag;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1308
frag = skb_shinfo(skb)->frags;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1309
skb_frag_off_add(frag, VLAN_HLEN);
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1310
skb_frag_size_sub(frag, VLAN_HLEN);
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2651
skb_frag_t *frag;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2853
frag = &skb_shinfo(skb)->frags[frag_idx];
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2855
len = skb_frag_size(frag);
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2856
bus = skb_frag_dma_map(&mgp->pdev->dev, frag, 0, len,
drivers/net/ethernet/natsemi/ns83820.c
1050
skb_frag_t *frag;
drivers/net/ethernet/natsemi/ns83820.c
1095
frag = skb_shinfo(skb)->frags;
drivers/net/ethernet/natsemi/ns83820.c
1097
frag = NULL;
drivers/net/ethernet/natsemi/ns83820.c
1145
buf = skb_frag_dma_map(&dev->pci_dev->dev, frag, 0,
drivers/net/ethernet/natsemi/ns83820.c
1146
skb_frag_size(frag), DMA_TO_DEVICE);
drivers/net/ethernet/natsemi/ns83820.c
1148
(long long)buf, (long) page_to_pfn(frag->page),
drivers/net/ethernet/natsemi/ns83820.c
1149
frag->page_offset);
drivers/net/ethernet/natsemi/ns83820.c
1150
len = skb_frag_size(frag);
drivers/net/ethernet/natsemi/ns83820.c
1151
frag++;
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
1008
void *orig_data = rxbuf->frag + pkt_off;
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
1013
rxbuf->frag + NFP_NET_RX_BUF_HEADROOM,
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
1044
nfp_nfd3_rx_give_one(dp, rx_ring, rxbuf->frag,
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
1055
nfp_app_ctrl_rx_raw(nn->app, rxbuf->frag + pkt_off,
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
1057
nfp_nfd3_rx_give_one(dp, rx_ring, rxbuf->frag,
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
1076
skb = napi_build_skb(rxbuf->frag, true_bufsz);
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
1354
if (unlikely(!nfp_ctrl_meta_ok(nn, rxbuf->frag + meta_off, meta_len))) {
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
1361
skb = build_skb(rxbuf->frag, dp->fl_bufsz);
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
264
const skb_frag_t *frag;
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
347
frag = &skb_shinfo(skb)->frags[f];
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
348
fsize = skb_frag_size(frag);
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
350
dma_addr = skb_frag_dma_map(dp->dev, frag, 0,
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
389
frag = &skb_shinfo(skb)->frags[f];
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
391
skb_frag_size(frag), DMA_TO_DEVICE);
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
442
const skb_frag_t *frag;
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
467
frag = &skb_shinfo(skb)->frags[fidx];
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
469
skb_frag_size(frag), DMA_TO_DEVICE);
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
554
void *frag;
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
557
frag = napi_alloc_frag(dp->fl_bufsz);
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
558
if (unlikely(!frag))
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
566
frag = page_address(page);
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
569
*dma_addr = nfp_net_dma_map_rx(dp, frag);
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
571
nfp_net_free_frag(frag, dp->xdp_prog);
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
576
return frag;
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
589
void *frag, dma_addr_t dma_addr)
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
598
rx_ring->rxbufs[wr_idx].frag = frag;
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
635
nfp_nfd3_rx_give_one(dp, rx_ring, rx_ring->rxbufs[i].frag,
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
827
if (skb && rxbuf && skb->head == rxbuf->frag)
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
828
page_ref_inc(virt_to_head_page(rxbuf->frag));
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
830
nfp_nfd3_rx_give_one(dp, rx_ring, rxbuf->frag, rxbuf->dma_addr);
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
868
nfp_nfd3_rx_give_one(dp, rx_ring, txbuf->frag, txbuf->dma_addr);
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
870
txbuf->frag = rxbuf->frag;
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
994
rxbuf->frag + meta_off, rxd);
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
997
rxbuf->frag + meta_off,
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
998
rxbuf->frag + pkt_off,
drivers/net/ethernet/netronome/nfp/nfd3/nfd3.h
71
void *frag;
drivers/net/ethernet/netronome/nfp/nfd3/rings.c
168
if (!tx_ring->txbufs[i].frag)
drivers/net/ethernet/netronome/nfp/nfd3/rings.c
172
__free_page(virt_to_page(tx_ring->txbufs[i].frag));
drivers/net/ethernet/netronome/nfp/nfd3/rings.c
187
txbufs[i].frag = nfp_net_rx_alloc_one(dp, &txbufs[i].dma_addr);
drivers/net/ethernet/netronome/nfp/nfd3/rings.c
188
if (!txbufs[i].frag) {
drivers/net/ethernet/netronome/nfp/nfd3/rings.c
45
const skb_frag_t *frag;
drivers/net/ethernet/netronome/nfp/nfd3/rings.c
64
frag = &skb_shinfo(skb)->frags[tx_buf->fidx];
drivers/net/ethernet/netronome/nfp/nfd3/rings.c
66
skb_frag_size(frag), DMA_TO_DEVICE);
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
1116
rxbuf->frag + meta_off,
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
1117
rxbuf->frag + pkt_off,
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
1127
void *orig_data = rxbuf->frag + pkt_off;
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
1132
rxbuf->frag + NFP_NET_RX_BUF_HEADROOM,
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
1163
nfp_nfdk_rx_give_one(dp, rx_ring, rxbuf->frag,
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
1174
nfp_app_ctrl_rx_raw(nn->app, rxbuf->frag + pkt_off,
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
1176
nfp_nfdk_rx_give_one(dp, rx_ring, rxbuf->frag,
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
1195
skb = napi_build_skb(rxbuf->frag, true_bufsz);
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
123
const skb_frag_t *frag, *fend;
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
132
frag = skb_shinfo(skb)->frags;
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
133
fend = frag + nr_frags;
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
134
for (; frag < fend; frag++)
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
135
n_descs += DIV_ROUND_UP(skb_frag_size(frag),
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
1516
if (unlikely(!nfp_ctrl_meta_ok(nn, rxbuf->frag + meta_off, meta_len))) {
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
1523
skb = build_skb(rxbuf->frag, dp->fl_bufsz);
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
260
const skb_frag_t *frag, *fend;
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
354
frag = skb_shinfo(skb)->frags;
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
355
fend = frag + nr_frags;
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
370
if (frag >= fend)
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
373
dma_len = skb_frag_size(frag);
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
374
dma_addr = skb_frag_dma_map(dp->dev, frag, 0, dma_len,
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
382
frag++;
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
447
frag = skb_shinfo(skb)->frags;
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
450
skb_frag_size(frag), DMA_TO_DEVICE);
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
452
frag++;
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
496
const skb_frag_t *frag, *fend;
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
518
frag = skb_shinfo(skb)->frags;
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
519
fend = frag + skb_shinfo(skb)->nr_frags;
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
520
for (; frag < fend; frag++) {
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
521
size = skb_frag_size(frag);
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
525
skb_frag_size(frag), DMA_TO_DEVICE);
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
574
void *frag;
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
577
frag = napi_alloc_frag(dp->fl_bufsz);
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
578
if (unlikely(!frag))
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
586
frag = page_address(page);
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
589
*dma_addr = nfp_net_dma_map_rx(dp, frag);
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
591
nfp_net_free_frag(frag, dp->xdp_prog);
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
596
return frag;
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
609
void *frag, dma_addr_t dma_addr)
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
618
rx_ring->rxbufs[wr_idx].frag = frag;
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
648
nfp_nfdk_rx_give_one(dp, rx_ring, rx_ring->rxbufs[i].frag,
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
827
if (skb && rxbuf && skb->head == rxbuf->frag)
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
828
page_ref_inc(virt_to_head_page(rxbuf->frag));
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
830
nfp_nfdk_rx_give_one(dp, rx_ring, rxbuf->frag, rxbuf->dma_addr);
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
955
txbuf[0].val = (unsigned long)rxbuf->frag | NFDK_TX_BUF_INFO_SOP;
drivers/net/ethernet/netronome/nfp/nfdk/nfdk.h
94
void *frag;
drivers/net/ethernet/netronome/nfp/nfdk/rings.c
17
const skb_frag_t *frag, *fend;
drivers/net/ethernet/netronome/nfp/nfdk/rings.c
41
frag = skb_shinfo(skb)->frags;
drivers/net/ethernet/netronome/nfp/nfdk/rings.c
42
fend = frag + nr_frags;
drivers/net/ethernet/netronome/nfp/nfdk/rings.c
43
for (; frag < fend; frag++) {
drivers/net/ethernet/netronome/nfp/nfdk/rings.c
44
size = skb_frag_size(frag);
drivers/net/ethernet/netronome/nfp/nfdk/rings.c
46
skb_frag_size(frag), DMA_TO_DEVICE);
drivers/net/ethernet/netronome/nfp/nfp_net.h
286
void *frag;
drivers/net/ethernet/netronome/nfp/nfp_net_debugfs.c
19
void *frag;
drivers/net/ethernet/netronome/nfp/nfp_net_debugfs.c
47
frag = READ_ONCE(rx_ring->rxbufs[i].frag);
drivers/net/ethernet/netronome/nfp/nfp_net_debugfs.c
48
if (frag)
drivers/net/ethernet/netronome/nfp/nfp_net_debugfs.c
49
seq_printf(file, " frag=%p", frag);
drivers/net/ethernet/netronome/nfp/nfp_net_dp.c
145
if (!rx_ring->rxbufs[i].frag)
drivers/net/ethernet/netronome/nfp/nfp_net_dp.c
149
nfp_net_free_frag(rx_ring->rxbufs[i].frag, dp->xdp_prog);
drivers/net/ethernet/netronome/nfp/nfp_net_dp.c
151
rx_ring->rxbufs[i].frag = NULL;
drivers/net/ethernet/netronome/nfp/nfp_net_dp.c
173
rxbufs[i].frag = nfp_net_rx_alloc_one(dp, &rxbufs[i].dma_addr);
drivers/net/ethernet/netronome/nfp/nfp_net_dp.c
174
if (!rxbufs[i].frag) {
drivers/net/ethernet/netronome/nfp/nfp_net_dp.c
19
void *frag;
drivers/net/ethernet/netronome/nfp/nfp_net_dp.c
22
frag = netdev_alloc_frag(dp->fl_bufsz);
drivers/net/ethernet/netronome/nfp/nfp_net_dp.c
27
frag = page ? page_address(page) : NULL;
drivers/net/ethernet/netronome/nfp/nfp_net_dp.c
29
if (!frag) {
drivers/net/ethernet/netronome/nfp/nfp_net_dp.c
34
*dma_addr = nfp_net_dma_map_rx(dp, frag);
drivers/net/ethernet/netronome/nfp/nfp_net_dp.c
36
nfp_net_free_frag(frag, dp->xdp_prog);
drivers/net/ethernet/netronome/nfp/nfp_net_dp.c
41
return frag;
drivers/net/ethernet/netronome/nfp/nfp_net_dp.h
11
return dma_map_single_attrs(dp->dev, frag + NFP_NET_RX_BUF_HEADROOM,
drivers/net/ethernet/netronome/nfp/nfp_net_dp.h
71
static inline void nfp_net_free_frag(void *frag, bool xdp)
drivers/net/ethernet/netronome/nfp/nfp_net_dp.h
74
skb_free_frag(frag);
drivers/net/ethernet/netronome/nfp/nfp_net_dp.h
76
__free_page(virt_to_page(frag));
drivers/net/ethernet/netronome/nfp/nfp_net_dp.h
9
static inline dma_addr_t nfp_net_dma_map_rx(struct nfp_net_dp *dp, void *frag)
drivers/net/ethernet/ni/nixge.c
513
skb_frag_t *frag;
drivers/net/ethernet/ni/nixge.c
545
frag = &skb_shinfo(skb)->frags[ii];
drivers/net/ethernet/ni/nixge.c
547
cur_phys = skb_frag_dma_map(ndev->dev.parent, frag, 0,
drivers/net/ethernet/ni/nixge.c
548
skb_frag_size(frag),
drivers/net/ethernet/ni/nixge.c
554
cur_p->cntrl = skb_frag_size(frag);
drivers/net/ethernet/ni/nixge.c
558
tx_skb->size = skb_frag_size(frag);
drivers/net/ethernet/nvidia/forcedeth.c
2270
const skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/nvidia/forcedeth.c
2271
u32 frag_size = skb_frag_size(frag);
drivers/net/ethernet/nvidia/forcedeth.c
2281
frag, offset,
drivers/net/ethernet/nvidia/forcedeth.c
2448
skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/nvidia/forcedeth.c
2449
u32 frag_size = skb_frag_size(frag);
drivers/net/ethernet/nvidia/forcedeth.c
2458
frag, offset,
drivers/net/ethernet/pasemi/pasemi_mac.c
1452
skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/pasemi/pasemi_mac.c
1454
map[i + 1] = skb_frag_dma_map(&mac->dma_pdev->dev, frag, 0,
drivers/net/ethernet/pasemi/pasemi_mac.c
1455
skb_frag_size(frag), DMA_TO_DEVICE);
drivers/net/ethernet/pasemi/pasemi_mac.c
1456
map_size[i+1] = skb_frag_size(frag);
drivers/net/ethernet/pasemi/pasemi_mac.c
254
const skb_frag_t *frag = &skb_shinfo(skb)->frags[f];
drivers/net/ethernet/pasemi/pasemi_mac.c
256
dma_unmap_page(&pdev->dev, dmas[f + 1], skb_frag_size(frag),
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
1096
const skb_frag_t *frag,
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
1102
dma_addr = skb_frag_dma_map(dev, frag, offset, len, DMA_TO_DEVICE);
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
1119
skb_frag_t *frag;
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
1129
frag = skb_shinfo(skb)->frags;
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
1131
for (frag_idx = 0; frag_idx < nfrags; frag_idx++, frag++) {
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
1132
dma_addr = ionic_tx_map_frag(q, frag, 0, skb_frag_size(frag));
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
1136
buf_info->len = skb_frag_size(frag);
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
1682
skb_frag_t *frag;
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
1717
frag = skb_shinfo(skb)->frags;
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
1743
frag_rem = skb_frag_size(frag);
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
1744
frag++;
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
19
const skb_frag_t *frag,
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
345
skb_frag_t *frag;
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
350
frag = sinfo->frags;
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
352
for (i = 0; i < sinfo->nr_frags; i++, frag++, bi++) {
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
354
struct page *pg = skb_frag_page(frag);
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
357
skb_frag_off(frag);
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
359
skb_frag_size(frag),
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
362
dma_addr = ionic_tx_map_frag(q, frag, 0,
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
363
skb_frag_size(frag));
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
370
bi->len = skb_frag_size(frag);
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
371
bi->page = skb_frag_page(frag);
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
504
skb_frag_t *frag;
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
518
frag = &sinfo->frags[sinfo->nr_frags];
drivers/net/ethernet/pensando/ionic/ionic_txrx.c
525
skb_frag_fill_page_desc(frag, bi->page, 0, frag_len);
drivers/net/ethernet/qlogic/netxen/netxen_nic_init.c
1740
struct netxen_skb_frag *frag;
drivers/net/ethernet/qlogic/netxen/netxen_nic_init.c
1753
frag = &buffer->frag_array[0];
drivers/net/ethernet/qlogic/netxen/netxen_nic_init.c
1754
dma_unmap_single(&pdev->dev, frag->dma, frag->length,
drivers/net/ethernet/qlogic/netxen/netxen_nic_init.c
1756
frag->dma = 0ULL;
drivers/net/ethernet/qlogic/netxen/netxen_nic_init.c
1758
frag++; /* Get the next frag */
drivers/net/ethernet/qlogic/netxen/netxen_nic_init.c
1759
dma_unmap_page(&pdev->dev, frag->dma,
drivers/net/ethernet/qlogic/netxen/netxen_nic_init.c
1760
frag->length, DMA_TO_DEVICE);
drivers/net/ethernet/qlogic/netxen/netxen_nic_init.c
1761
frag->dma = 0ULL;
drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c
1967
skb_frag_t *frag;
drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c
1983
frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c
1986
map = skb_frag_dma_map(&pdev->dev, frag, 0, skb_frag_size(frag),
drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c
1992
nf->length = skb_frag_size(frag);
drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c
2030
skb_frag_t *frag;
drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c
2044
frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c
2045
delta += skb_frag_size(frag);
drivers/net/ethernet/qlogic/qed/qed_ll2.c
2680
const skb_frag_t *frag;
drivers/net/ethernet/qlogic/qed/qed_ll2.c
2742
frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/qlogic/qed/qed_ll2.c
2744
mapping = skb_frag_dma_map(&cdev->pdev->dev, frag, 0,
drivers/net/ethernet/qlogic/qed/qed_ll2.c
2745
skb_frag_size(frag), DMA_TO_DEVICE);
drivers/net/ethernet/qlogic/qed/qed_ll2.c
2757
skb_frag_size(frag));
drivers/net/ethernet/qlogic/qede/qede_fp.c
245
skb_frag_t *frag, struct eth_tx_bd *bd)
drivers/net/ethernet/qlogic/qede/qede_fp.c
250
mapping = skb_frag_dma_map(txq->dev, frag, 0,
drivers/net/ethernet/qlogic/qede/qede_fp.c
251
skb_frag_size(frag), DMA_TO_DEVICE);
drivers/net/ethernet/qlogic/qede/qede_fp.c
256
BD_SET_UNMAP_ADDR_LEN(bd, mapping, skb_frag_size(frag));
drivers/net/ethernet/qlogic/qla3xxx.c
2345
skb_frag_t *frag = &skb_shinfo(skb)->frags[completed_segs];
drivers/net/ethernet/qlogic/qla3xxx.c
2380
map = skb_frag_dma_map(&qdev->pdev->dev, frag, 0, skb_frag_size(frag),
drivers/net/ethernet/qlogic/qla3xxx.c
2393
oal_entry->len = cpu_to_le32(skb_frag_size(frag));
drivers/net/ethernet/qlogic/qla3xxx.c
2395
dma_unmap_len_set(&tx_cb->map[seg], maplen, skb_frag_size(frag));
drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c
583
skb_frag_t *frag;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c
599
frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c
601
map = skb_frag_dma_map(&pdev->dev, frag, 0, skb_frag_size(frag),
drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c
607
nf->length = skb_frag_size(frag);
drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c
894
struct qlcnic_skb_frag *frag;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c
905
frag = &buffer->frag_array[0];
drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c
906
dma_unmap_single(&pdev->dev, frag->dma, frag->length,
drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c
908
frag->dma = 0ULL;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c
910
frag++;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c
911
dma_unmap_page(&pdev->dev, frag->dma,
drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c
912
frag->length, DMA_TO_DEVICE);
drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c
913
frag->dma = 0ULL;
drivers/net/ethernet/qualcomm/emac/emac-mac.c
1388
skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/qualcomm/emac/emac-mac.c
1391
tpbuf->length = skb_frag_size(frag);
drivers/net/ethernet/qualcomm/emac/emac-mac.c
1393
frag, 0, tpbuf->length,
drivers/net/ethernet/realtek/8139cp.c
718
int frag, index;
drivers/net/ethernet/realtek/8139cp.c
721
for (frag = 0; frag+first < entry_last; frag++) {
drivers/net/ethernet/realtek/8139cp.c
722
index = first+frag;
drivers/net/ethernet/realtek/8139cp.c
725
this_frag = &skb_shinfo(skb)->frags[frag];
drivers/net/ethernet/realtek/8139cp.c
805
int frag, first_entry = entry;
drivers/net/ethernet/realtek/8139cp.c
819
for (frag = 0; frag < skb_shinfo(skb)->nr_frags; frag++) {
drivers/net/ethernet/realtek/8139cp.c
820
const skb_frag_t *this_frag = &skb_shinfo(skb)->frags[frag];
drivers/net/ethernet/realtek/8139cp.c
839
if (frag == skb_shinfo(skb)->nr_frags - 1)
drivers/net/ethernet/realtek/r8169_main.c
4360
const skb_frag_t *frag = info->frags + cur_frag;
drivers/net/ethernet/realtek/r8169_main.c
4361
void *addr = skb_frag_address(frag);
drivers/net/ethernet/realtek/r8169_main.c
4362
u32 len = skb_frag_size(frag);
drivers/net/ethernet/realtek/rtase/rtase_main.c
1261
const skb_frag_t *frag = &info->frags[cur_frag];
drivers/net/ethernet/realtek/rtase/rtase_main.c
1269
len = skb_frag_size(frag);
drivers/net/ethernet/realtek/rtase/rtase_main.c
1270
addr = skb_frag_address(frag);
drivers/net/ethernet/rocker/rocker_main.c
1853
struct rocker_tlv *frag;
drivers/net/ethernet/rocker/rocker_main.c
1861
frag = rocker_tlv_nest_start(desc_info, ROCKER_TLV_TX_FRAG);
drivers/net/ethernet/rocker/rocker_main.c
1862
if (!frag)
drivers/net/ethernet/rocker/rocker_main.c
1870
rocker_tlv_nest_end(desc_info, frag);
drivers/net/ethernet/rocker/rocker_main.c
1874
rocker_tlv_nest_cancel(desc_info, frag);
drivers/net/ethernet/rocker/rocker_main.c
1912
const skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/rocker/rocker_main.c
1915
skb_frag_address(frag),
drivers/net/ethernet/rocker/rocker_main.c
1916
skb_frag_size(frag));
drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c
1357
const skb_frag_t *frag = &skb_shinfo(skb)->frags[frag_num];
drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c
1358
int len = skb_frag_size(frag);
drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c
1362
tx_desc->tdes01 = skb_frag_dma_map(priv->device, frag, 0, len,
drivers/net/ethernet/sfc/tx_tso.c
209
skb_frag_t *frag)
drivers/net/ethernet/sfc/tx_tso.c
211
st->unmap_addr = skb_frag_dma_map(&efx->pci_dev->dev, frag, 0,
drivers/net/ethernet/sfc/tx_tso.c
212
skb_frag_size(frag), DMA_TO_DEVICE);
drivers/net/ethernet/sfc/tx_tso.c
214
st->unmap_len = skb_frag_size(frag);
drivers/net/ethernet/sfc/tx_tso.c
215
st->in_len = skb_frag_size(frag);
drivers/net/ethernet/spacemit/k1_emac.c
691
const skb_frag_t *frag;
drivers/net/ethernet/spacemit/k1_emac.c
705
frag = &skb_shinfo(skb)->frags[frag_idx - 1];
drivers/net/ethernet/spacemit/k1_emac.c
706
len = skb_frag_size(frag);
drivers/net/ethernet/spacemit/k1_emac.c
707
addr = skb_frag_dma_map(dev, frag, 0, len, DMA_TO_DEVICE);
drivers/net/ethernet/stmicro/stmmac/dwmac5.c
436
struct stmmac_tc_entry *entry, *frag;
drivers/net/ethernet/stmicro/stmmac/dwmac5.c
464
frag = entry->frag_ptr;
drivers/net/ethernet/stmicro/stmmac/dwmac5.c
467
if (frag) {
drivers/net/ethernet/stmicro/stmmac/dwmac5.c
481
if (frag && !frag->in_hw) {
drivers/net/ethernet/stmicro/stmmac/dwmac5.c
482
ret = dwmac5_rxp_update_single_entry(ioaddr, frag, nve);
drivers/net/ethernet/stmicro/stmmac/dwmac5.c
485
frag->table_pos = nve++;
drivers/net/ethernet/stmicro/stmmac/dwmac5.c
486
frag->in_hw = true;
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c
1059
struct stmmac_tc_entry *entry, *frag;
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c
1087
frag = entry->frag_ptr;
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c
1090
if (frag) {
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c
1104
if (frag && !frag->in_hw) {
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c
1105
ret = dwxgmac3_rxp_update_single_entry(ioaddr, frag, nve);
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c
1108
frag->table_pos = nve++;
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c
1109
frag->in_hw = true;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4490
const skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4492
des = skb_frag_dma_map(priv->device, frag, 0,
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4493
skb_frag_size(frag),
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4498
stmmac_tso_allocator(priv, des, skb_frag_size(frag),
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4502
tx_q->tx_skbuff_dma[tx_q->cur_tx].len = skb_frag_size(frag);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4750
const skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4751
int len = skb_frag_size(frag);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4764
des = skb_frag_dma_map(priv->device, frag, 0, len,
drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c
129
frag = tc_find_entry(priv, cls, true);
drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c
130
if (!frag) {
drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c
135
entry->frag_ptr = frag;
drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c
143
frag->val.match_en = (mask >> (rem * 8)) &
drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c
145
frag->val.match_data = (data >> (rem * 8)) &
drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c
147
frag->val.frame_offset = real_off + 1;
drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c
148
frag->prio = prio;
drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c
149
frag->is_frag = true;
drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c
158
ret = tc_fill_actions(entry, frag, cls);
drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c
165
if (frag)
drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c
166
frag->in_use = false;
drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c
58
struct stmmac_tc_entry *frag,
drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c
69
if (frag)
drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c
70
action_entry = frag;
drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c
94
struct stmmac_tc_entry *entry, *frag = NULL;
drivers/net/ethernet/sun/cassini.c
1828
int frag;
drivers/net/ethernet/sun/cassini.c
1848
for (frag = 0; frag <= skb_shinfo(skb)->nr_frags; frag++) {
drivers/net/ethernet/sun/cassini.c
1961
skb_frag_t *frag = skb_shinfo(skb)->frags;
drivers/net/ethernet/sun/cassini.c
2002
skb_frag_fill_page_desc(frag, page->buffer, off, hlen - swivel);
drivers/net/ethernet/sun/cassini.c
2003
__skb_frag_ref(frag);
drivers/net/ethernet/sun/cassini.c
2024
frag++;
drivers/net/ethernet/sun/cassini.c
2026
skb_frag_fill_page_desc(frag, page->buffer, 0, hlen);
drivers/net/ethernet/sun/cassini.c
2027
__skb_frag_ref(frag);
drivers/net/ethernet/sun/cassini.c
2716
int entry, nr_frags, frag, tabort, tentry;
drivers/net/ethernet/sun/cassini.c
2770
for (frag = 0; frag < nr_frags; frag++) {
drivers/net/ethernet/sun/cassini.c
2771
const skb_frag_t *fragp = &skb_shinfo(skb)->frags[frag];
drivers/net/ethernet/sun/cassini.c
2792
(frag + 1 == nr_frags));
drivers/net/ethernet/sun/cassini.c
3840
int frag;
drivers/net/ethernet/sun/cassini.c
3848
for (frag = 0; frag <= skb_shinfo(skb)->nr_frags; frag++) {
drivers/net/ethernet/sun/cassini.c
3860
if (frag != skb_shinfo(skb)->nr_frags) {
drivers/net/ethernet/sun/niu.c
6713
const skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/sun/niu.c
6715
len = skb_frag_size(frag);
drivers/net/ethernet/sun/niu.c
6716
mapping = np->ops->map_page(np->device, skb_frag_page(frag),
drivers/net/ethernet/sun/niu.c
6717
skb_frag_off(frag), len,
drivers/net/ethernet/sun/niu.c
6747
const skb_frag_t *frag;
drivers/net/ethernet/sun/niu.c
6750
frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/sun/niu.c
6752
skb_frag_size(frag), DMA_TO_DEVICE);
drivers/net/ethernet/sun/sungem.c
1030
int frag, first_entry = entry;
drivers/net/ethernet/sun/sungem.c
1046
for (frag = 0; frag < skb_shinfo(skb)->nr_frags; frag++) {
drivers/net/ethernet/sun/sungem.c
1047
const skb_frag_t *this_frag = &skb_shinfo(skb)->frags[frag];
drivers/net/ethernet/sun/sungem.c
1056
if (frag == skb_shinfo(skb)->nr_frags - 1)
drivers/net/ethernet/sun/sungem.c
1583
int frag;
drivers/net/ethernet/sun/sungem.c
1588
for (frag = 0; frag <= skb_shinfo(skb)->nr_frags; frag++) {
drivers/net/ethernet/sun/sungem.c
1597
if (frag != skb_shinfo(skb)->nr_frags)
drivers/net/ethernet/sun/sungem.c
642
int frag;
drivers/net/ethernet/sun/sungem.c
667
for (frag = 0; frag <= skb_shinfo(skb)->nr_frags; frag++) {
drivers/net/ethernet/sun/sunhme.c
1181
int frag;
drivers/net/ethernet/sun/sunhme.c
1185
for (frag = 0; frag <= skb_shinfo(skb)->nr_frags; frag++) {
drivers/net/ethernet/sun/sunhme.c
1188
if (!frag)
drivers/net/ethernet/sun/sunhme.c
1199
if (frag != skb_shinfo(skb)->nr_frags)
drivers/net/ethernet/sun/sunhme.c
1707
int frag;
drivers/net/ethernet/sun/sunhme.c
1727
for (frag = 0; frag <= skb_shinfo(skb)->nr_frags; frag++) {
drivers/net/ethernet/sun/sunhme.c
1732
if (!frag)
drivers/net/ethernet/sun/sunhme.c
2015
int frag, first_entry = entry;
drivers/net/ethernet/sun/sunhme.c
2027
for (frag = 0; frag < skb_shinfo(skb)->nr_frags; frag++) {
drivers/net/ethernet/sun/sunhme.c
2028
const skb_frag_t *this_frag = &skb_shinfo(skb)->frags[frag];
drivers/net/ethernet/sun/sunhme.c
2040
if (frag == skb_shinfo(skb)->nr_frags - 1)
drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c
502
skb_frag_t *frag;
drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c
579
frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c
582
for (datalen = skb_frag_size(frag); datalen; ) {
drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c
586
skb_dma = skb_frag_dma_map(pdata->dev, frag, offset,
drivers/net/ethernet/synopsys/dwc-xlgmac-net.c
119
skb_frag_t *frag;
drivers/net/ethernet/synopsys/dwc-xlgmac-net.c
181
frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/synopsys/dwc-xlgmac-net.c
182
for (len = skb_frag_size(frag); len; ) {
drivers/net/ethernet/tehuti/tehuti.c
1488
const skb_frag_t *frag;
drivers/net/ethernet/tehuti/tehuti.c
1490
frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/tehuti/tehuti.c
1491
db->wptr->len = skb_frag_size(frag);
drivers/net/ethernet/tehuti/tehuti.c
1492
db->wptr->addr.dma = skb_frag_dma_map(&priv->pdev->dev, frag,
drivers/net/ethernet/tehuti/tehuti.c
1493
0, skb_frag_size(frag),
drivers/net/ethernet/tehuti/tn40.c
606
skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/tehuti/tn40.c
608
size = skb_frag_size(frag);
drivers/net/ethernet/tehuti/tn40.c
609
dma = skb_frag_dma_map(&priv->pdev->dev, frag, 0,
drivers/net/ethernet/ti/am65-cpsw-nuss.c
1672
skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/ti/am65-cpsw-nuss.c
1673
u32 frag_size = skb_frag_size(frag);
drivers/net/ethernet/ti/am65-cpsw-nuss.c
1684
buf_dma = skb_frag_dma_map(tx_chn->dma_dev, frag, 0, frag_size,
drivers/net/ethernet/ti/icssg/icssg_common.c
1261
skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/ti/icssg/icssg_common.c
1262
u32 frag_size = skb_frag_size(frag);
drivers/net/ethernet/ti/icssg/icssg_common.c
1271
buf_dma = skb_frag_dma_map(tx_chn->dma_dev, frag, 0, frag_size,
drivers/net/ethernet/ti/netcp_core.c
1116
skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/ti/netcp_core.c
1117
struct page *page = skb_frag_page(frag);
drivers/net/ethernet/ti/netcp_core.c
1118
u32 page_offset = skb_frag_off(frag);
drivers/net/ethernet/ti/netcp_core.c
1119
u32 buf_len = skb_frag_size(frag);
drivers/net/ethernet/tundra/tsi108_eth.c
690
const skb_frag_t *frag = &skb_shinfo(skb)->frags[i - 1];
drivers/net/ethernet/tundra/tsi108_eth.c
693
skb_frag_dma_map(&data->pdev->dev, frag,
drivers/net/ethernet/tundra/tsi108_eth.c
694
0, skb_frag_size(frag),
drivers/net/ethernet/tundra/tsi108_eth.c
696
data->txring[tx].len = skb_frag_size(frag);
drivers/net/ethernet/via/via-velocity.c
2581
const skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/net/ethernet/via/via-velocity.c
2584
frag, 0,
drivers/net/ethernet/via/via-velocity.c
2585
skb_frag_size(frag),
drivers/net/ethernet/via/via-velocity.c
2590
td_ptr->td_buf[i + 1].size = cpu_to_le16(skb_frag_size(frag));
drivers/net/ethernet/wangxun/libwx/wx_lib.c
1044
skb_frag_t *frag;
drivers/net/ethernet/wangxun/libwx/wx_lib.c
1058
for (frag = &skb_shinfo(skb)->frags[0];; frag++) {
drivers/net/ethernet/wangxun/libwx/wx_lib.c
1099
size = skb_frag_size(frag);
drivers/net/ethernet/wangxun/libwx/wx_lib.c
1103
dma = skb_frag_dma_map(tx_ring->dev, frag, 0, size,
drivers/net/ethernet/wangxun/libwx/wx_lib.c
171
skb_frag_t *frag = &skb_shinfo(skb)->frags[0];
drivers/net/ethernet/wangxun/libwx/wx_lib.c
175
skb_frag_off(frag),
drivers/net/ethernet/wangxun/libwx/wx_lib.c
176
skb_frag_size(frag),
drivers/net/ethernet/wangxun/libwx/wx_lib.c
438
skb_frag_t *frag = &skb_shinfo(skb)->frags[0];
drivers/net/ethernet/wangxun/libwx/wx_lib.c
446
va = skb_frag_address(frag);
drivers/net/ethernet/wangxun/libwx/wx_lib.c
457
skb_frag_size_sub(frag, pull_len);
drivers/net/ethernet/wangxun/libwx/wx_lib.c
458
skb_frag_off_add(frag, pull_len);
drivers/net/ethernet/xilinx/ll_temac_main.c
853
skb_frag_t *frag;
drivers/net/ethernet/xilinx/ll_temac_main.c
856
frag = &skb_shinfo(skb)->frags[0];
drivers/net/ethernet/xilinx/ll_temac_main.c
903
skb_frag_address(frag),
drivers/net/ethernet/xilinx/ll_temac_main.c
904
skb_frag_size(frag),
drivers/net/ethernet/xilinx/ll_temac_main.c
911
--frag;
drivers/net/ethernet/xilinx/ll_temac_main.c
914
skb_frag_size(frag),
drivers/net/ethernet/xilinx/ll_temac_main.c
928
cur_p->len = cpu_to_be32(skb_frag_size(frag));
drivers/net/ethernet/xilinx/ll_temac_main.c
930
frag++;
drivers/net/ethernet/xilinx/xilinx_axienet_main.c
1053
skb_frag_t *frag;
drivers/net/ethernet/xilinx/xilinx_axienet_main.c
1107
frag = &skb_shinfo(skb)->frags[ii];
drivers/net/ethernet/xilinx/xilinx_axienet_main.c
1109
skb_frag_address(frag),
drivers/net/ethernet/xilinx/xilinx_axienet_main.c
1110
skb_frag_size(frag),
drivers/net/ethernet/xilinx/xilinx_axienet_main.c
1122
cur_p->cntrl = skb_frag_size(frag);
drivers/net/hyperv/netvsc_drv.c
352
skb_frag_t *frag = skb_shinfo(skb)->frags + i;
drivers/net/hyperv/netvsc_drv.c
354
u64 pfn = page_to_hvpfn(skb_frag_page(frag));
drivers/net/hyperv/netvsc_drv.c
355
u32 offset = skb_frag_off(frag);
drivers/net/hyperv/netvsc_drv.c
358
cur_pb->len = skb_frag_size(frag);
drivers/net/hyperv/netvsc_drv.c
370
skb_frag_t *frag = skb_shinfo(skb)->frags + i;
drivers/net/hyperv/netvsc_drv.c
371
unsigned long size = skb_frag_size(frag);
drivers/net/hyperv/netvsc_drv.c
372
unsigned long offset = skb_frag_off(frag);
drivers/net/netconsole.c
1874
int frag, left = len;
drivers/net/netconsole.c
1877
frag = min(left, MAX_PRINT_CHUNK);
drivers/net/netconsole.c
1878
send_udp(nt, tmp, frag);
drivers/net/netconsole.c
1879
tmp += frag;
drivers/net/netconsole.c
1880
left -= frag;
drivers/net/ppp/ppp_generic.c
1964
struct sk_buff *frag;
drivers/net/ppp/ppp_generic.c
2134
frag = alloc_skb(flen + hdrlen + (flen == 0), GFP_ATOMIC);
drivers/net/ppp/ppp_generic.c
2135
if (!frag)
drivers/net/ppp/ppp_generic.c
2137
q = skb_put(frag, flen + hdrlen);
drivers/net/ppp/ppp_generic.c
2156
!chan->ops->start_xmit(chan, frag))
drivers/net/ppp/ppp_generic.c
2157
skb_queue_tail(&pch->file.xq, frag);
drivers/net/thunderbolt/main.c
1097
const skb_frag_t *frag = &skb_shinfo(skb)->frags[frag_num];
drivers/net/thunderbolt/main.c
1099
*len = skb_frag_size(frag);
drivers/net/thunderbolt/main.c
1100
return kmap_local_page(skb_frag_page(frag)) + skb_frag_off(frag);
drivers/net/thunderbolt/main.c
1113
unsigned int frag = 0;
drivers/net/thunderbolt/main.c
1165
if (frag < skb_shinfo(skb)->nr_frags) {
drivers/net/thunderbolt/main.c
1167
src = tbnet_kmap_frag(skb, frag++, &len);
drivers/net/thunderbolt/main.c
1202
if (frag < skb_shinfo(skb)->nr_frags) {
drivers/net/thunderbolt/main.c
1203
src = tbnet_kmap_frag(skb, frag++, &len);
drivers/net/tun.c
1422
void *frag;
drivers/net/tun.c
1428
frag = netdev_alloc_frag(fragsz);
drivers/net/tun.c
1429
if (!frag) {
drivers/net/tun.c
1433
page = virt_to_head_page(frag);
drivers/net/tun.c
1435
frag - page_address(page), fragsz);
drivers/net/virtio_net.c
1661
skb_frag_t *frag = &shinfo->frags[i];
drivers/net/virtio_net.c
1663
sg_set_page(&sq->sg[i + 1], skb_frag_page(frag),
drivers/net/virtio_net.c
1664
skb_frag_size(frag), skb_frag_off(frag));
drivers/net/virtio_net.c
2212
skb_frag_t *frag;
drivers/net/virtio_net.c
2261
frag = &shinfo->frags[shinfo->nr_frags++];
drivers/net/virtio_net.c
2262
skb_frag_fill_page_desc(frag, page, offset, len);
drivers/net/vmxnet3/vmxnet3_drv.c
1125
const skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/net/vmxnet3/vmxnet3_drv.c
1127
count += VMXNET3_TXD_NEEDED(skb_frag_size(frag));
drivers/net/vmxnet3/vmxnet3_drv.c
782
skb_frag_t *frag = skb_shinfo(skb)->frags + skb_shinfo(skb)->nr_frags;
drivers/net/vmxnet3/vmxnet3_drv.c
786
skb_frag_fill_page_desc(frag, rbi->page, 0, rcd->len);
drivers/net/vmxnet3/vmxnet3_drv.c
877
const skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/net/vmxnet3/vmxnet3_drv.c
881
len = skb_frag_size(frag);
drivers/net/vmxnet3/vmxnet3_drv.c
892
tbi->dma_addr = skb_frag_dma_map(&adapter->pdev->dev, frag,
drivers/net/wireless/admtek/adm8211.c
1704
txhdr->frag = cpu_to_le16(0x0FFF);
drivers/net/wireless/admtek/adm8211.h
475
__le16 frag;
drivers/net/wireless/ath/ath10k/htt_rx.c
1931
bool frag)
drivers/net/wireless/ath/ath10k/htt_rx.c
2058
if (frag && !fill_crypt_header && is_decrypted &&
drivers/net/wireless/ath/ath10k/htt_rx.c
2065
if (frag)
drivers/net/wireless/ath/ath10k/htt_rx.c
2083
if (frag && !fill_crypt_header &&
drivers/net/wireless/ath/ath10k/htt_rx.c
2105
if (frag && !fill_crypt_header &&
drivers/net/wireless/ath/ath10k/htt_rx.c
2768
u32 tid, frag, seq, rx_desc_info;
drivers/net/wireless/ath/ath10k/htt_rx.c
2827
frag = sc & IEEE80211_SCTL_FRAG;
drivers/net/wireless/ath/ath10k/htt_rx.c
2885
if (frag == 0) {
drivers/net/wireless/ath/ath10k/htt_rx.c
3273
bool frag;
drivers/net/wireless/ath/ath10k/htt_rx.c
3290
frag = !!(resp->rx_in_ord_ind.info & HTT_RX_IN_ORD_IND_INFO_FRAG_MASK);
drivers/net/wireless/ath/ath10k/htt_rx.c
3294
vdev_id, peer_id, tid, offload, frag, msdu_count);
drivers/net/wireless/ath/ath10k/htt_rx.c
3338
NULL, peer_id, frag);
drivers/net/wireless/ath/ath9k/ath9k.h
314
struct sk_buff *frag;
drivers/net/wireless/ath/ath9k/main.c
953
if (sc->rx.frag) {
drivers/net/wireless/ath/ath9k/main.c
954
dev_kfree_skb_any(sc->rx.frag);
drivers/net/wireless/ath/ath9k/main.c
955
sc->rx.frag = NULL;
drivers/net/wireless/ath/ath9k/recv.c
1108
if (sc->rx.frag)
drivers/net/wireless/ath/ath9k/recv.c
1109
hdr_skb = sc->rx.frag;
drivers/net/wireless/ath/ath9k/recv.c
1164
if (sc->rx.frag) {
drivers/net/wireless/ath/ath9k/recv.c
1166
dev_kfree_skb_any(sc->rx.frag);
drivers/net/wireless/ath/ath9k/recv.c
1171
sc->rx.frag = skb;
drivers/net/wireless/ath/ath9k/recv.c
1175
if (sc->rx.frag) {
drivers/net/wireless/ath/ath9k/recv.c
1184
sc->rx.frag = NULL;
drivers/net/wireless/ath/ath9k/recv.c
1215
if (sc->rx.frag) {
drivers/net/wireless/ath/ath9k/recv.c
1216
dev_kfree_skb_any(sc->rx.frag);
drivers/net/wireless/ath/ath9k/recv.c
1217
sc->rx.frag = NULL;
drivers/net/wireless/ath/wil6210/debugfs.c
1051
const skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/net/wireless/ath/wil6210/debugfs.c
1053
len = skb_frag_size(frag);
drivers/net/wireless/ath/wil6210/debugfs.c
1054
p = skb_frag_address_safe(frag);
drivers/net/wireless/ath/wil6210/txrx.c
1755
const skb_frag_t *frag = NULL;
drivers/net/wireless/ath/wil6210/txrx.c
1844
frag = &skb_shinfo(skb)->frags[f];
drivers/net/wireless/ath/wil6210/txrx.c
1845
len = skb_frag_size(frag);
drivers/net/wireless/ath/wil6210/txrx.c
1865
pa = skb_frag_dma_map(dev, frag,
drivers/net/wireless/ath/wil6210/txrx.c
1866
skb_frag_size(frag) - len,
drivers/net/wireless/ath/wil6210/txrx.c
2088
const skb_frag_t *frag = &skb_shinfo(skb)->frags[f];
drivers/net/wireless/ath/wil6210/txrx.c
2089
int len = skb_frag_size(frag);
drivers/net/wireless/ath/wil6210/txrx.c
2097
pa = skb_frag_dma_map(dev, frag, 0, skb_frag_size(frag),
drivers/net/wireless/ath/wil6210/txrx_edma.c
1351
skb_frag_t *frag, struct wil_ring *ring,
drivers/net/wireless/ath/wil6210/txrx_edma.c
1366
if (!frag) {
drivers/net/wireless/ath/wil6210/txrx_edma.c
1370
pa = skb_frag_dma_map(dev, frag, 0, len, DMA_TO_DEVICE);
drivers/net/wireless/ath/wil6210/txrx_edma.c
1481
skb_frag_t *frag = &skb_shinfo(skb)->frags[f];
drivers/net/wireless/ath/wil6210/txrx_edma.c
1482
int len = skb_frag_size(frag);
drivers/net/wireless/ath/wil6210/txrx_edma.c
1491
frag, ring, skb, is_ipv4,
drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c
6153
struct sk_buff *p, struct scb *scb, uint frag,
drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c
6472
if (frag == 0)
drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c
6700
if (!(tx_info->flags & IEEE80211_TX_CTL_AMPDU) && frag == 0) {
drivers/net/wireless/intel/ipw2x00/ipw2100.c
7246
if (!wrqu->frag.fixed)
drivers/net/wireless/intel/ipw2x00/ipw2100.c
7249
if (wrqu->frag.disabled) {
drivers/net/wireless/intel/ipw2x00/ipw2100.c
7253
if (wrqu->frag.value < MIN_FRAG_THRESHOLD ||
drivers/net/wireless/intel/ipw2x00/ipw2100.c
7254
wrqu->frag.value > MAX_FRAG_THRESHOLD)
drivers/net/wireless/intel/ipw2x00/ipw2100.c
7257
priv->ieee->fts = wrqu->frag.value & ~0x1;
drivers/net/wireless/intel/ipw2x00/ipw2100.c
7275
wrqu->frag.value = priv->frag_threshold & ~FRAG_DISABLED;
drivers/net/wireless/intel/ipw2x00/ipw2100.c
7276
wrqu->frag.fixed = 0; /* no auto select */
drivers/net/wireless/intel/ipw2x00/ipw2100.c
7277
wrqu->frag.disabled = (priv->frag_threshold & FRAG_DISABLED) ? 1 : 0;
drivers/net/wireless/intel/ipw2x00/ipw2100.c
7279
IPW_DEBUG_WX("GET Frag Threshold -> %d\n", wrqu->frag.value);
drivers/net/wireless/intel/ipw2x00/ipw2200.c
2491
static int ipw_send_frag_threshold(struct ipw_priv *priv, u16 frag)
drivers/net/wireless/intel/ipw2x00/ipw2200.c
2494
.frag_threshold = cpu_to_le16(frag),
drivers/net/wireless/intel/ipw2x00/ipw2200.c
8086
u16 frag = WLAN_GET_SEQ_FRAG(sc);
drivers/net/wireless/intel/ipw2x00/ipw2200.c
8113
entry->frag_num = frag;
drivers/net/wireless/intel/ipw2x00/ipw2200.c
8134
if (*last_frag == frag)
drivers/net/wireless/intel/ipw2x00/ipw2200.c
8136
if (*last_frag + 1 != frag)
drivers/net/wireless/intel/ipw2x00/ipw2200.c
8142
*last_frag = frag;
drivers/net/wireless/intel/ipw2x00/ipw2200.c
9296
if (wrqu->frag.disabled || !wrqu->frag.fixed)
drivers/net/wireless/intel/ipw2x00/ipw2200.c
9299
if (wrqu->frag.value < MIN_FRAG_THRESHOLD ||
drivers/net/wireless/intel/ipw2x00/ipw2200.c
9300
wrqu->frag.value > MAX_FRAG_THRESHOLD) {
drivers/net/wireless/intel/ipw2x00/ipw2200.c
9305
priv->ieee->fts = wrqu->frag.value & ~0x1;
drivers/net/wireless/intel/ipw2x00/ipw2200.c
9308
ipw_send_frag_threshold(priv, wrqu->frag.value);
drivers/net/wireless/intel/ipw2x00/ipw2200.c
9310
IPW_DEBUG_WX("SET Frag Threshold -> %d\n", wrqu->frag.value);
drivers/net/wireless/intel/ipw2x00/ipw2200.c
9320
wrqu->frag.value = priv->ieee->fts;
drivers/net/wireless/intel/ipw2x00/ipw2200.c
9321
wrqu->frag.fixed = 0; /* no auto select */
drivers/net/wireless/intel/ipw2x00/ipw2200.c
9322
wrqu->frag.disabled = (wrqu->frag.value == DEFAULT_FTS);
drivers/net/wireless/intel/ipw2x00/ipw2200.c
9324
IPW_DEBUG_WX("GET Frag Threshold -> %d\n", wrqu->frag.value);
drivers/net/wireless/intel/ipw2x00/libipw_rx.c
114
entry->last_frag = frag;
drivers/net/wireless/intel/ipw2x00/libipw_rx.c
121
entry = libipw_frag_cache_find(ieee, seq, frag, hdr->addr2,
drivers/net/wireless/intel/ipw2x00/libipw_rx.c
124
entry->last_frag = frag;
drivers/net/wireless/intel/ipw2x00/libipw_rx.c
332
unsigned int frag;
drivers/net/wireless/intel/ipw2x00/libipw_rx.c
359
frag = WLAN_GET_SEQ_FRAG(sc);
drivers/net/wireless/intel/ipw2x00/libipw_rx.c
53
unsigned int frag,
drivers/net/wireless/intel/ipw2x00/libipw_rx.c
575
if ((frag != 0) || (fc & IEEE80211_FCTL_MOREFRAGS)) {
drivers/net/wireless/intel/ipw2x00/libipw_rx.c
578
LIBIPW_DEBUG_FRAG("Rx Fragment received (%u)\n", frag);
drivers/net/wireless/intel/ipw2x00/libipw_rx.c
585
WLAN_GET_SEQ_SEQ(sc), frag);
drivers/net/wireless/intel/ipw2x00/libipw_rx.c
590
if (frag != 0)
drivers/net/wireless/intel/ipw2x00/libipw_rx.c
601
if (frag == 0) {
drivers/net/wireless/intel/ipw2x00/libipw_rx.c
72
(entry->last_frag + 1 == frag || frag == -1) &&
drivers/net/wireless/intel/ipw2x00/libipw_rx.c
87
unsigned int frag, seq;
drivers/net/wireless/intel/ipw2x00/libipw_rx.c
91
frag = WLAN_GET_SEQ_FRAG(sc);
drivers/net/wireless/intel/ipw2x00/libipw_rx.c
94
if (frag == 0) {
drivers/net/wireless/intel/ipw2x00/libipw_tx.c
139
struct sk_buff *frag, int hdr_len)
drivers/net/wireless/intel/ipw2x00/libipw_tx.c
153
res = crypt->ops->encrypt_mpdu(frag, hdr_len, crypt->priv);
drivers/net/wireless/intel/ipw2x00/libipw_tx.c
158
ieee->dev->name, frag->len);
drivers/net/wireless/intel/iwlwifi/fw/dbg.c
1308
struct iwl_dram_data *frag;
drivers/net/wireless/intel/iwlwifi/fw/dbg.c
1311
frag = &fwrt->trans->dbg.fw_mon_ini[alloc_id].frags[idx];
drivers/net/wireless/intel/iwlwifi/fw/dbg.c
1313
range->dram_base_addr = cpu_to_le64(frag->physical);
drivers/net/wireless/intel/iwlwifi/fw/dbg.c
1314
range->range_data_size = cpu_to_le32(frag->size);
drivers/net/wireless/intel/iwlwifi/fw/dbg.c
1316
memcpy(range->data, frag->block, frag->size);
drivers/net/wireless/intel/iwlwifi/fw/dbg.c
2063
struct iwl_dram_data *frag = &fw_mon->frags[i];
drivers/net/wireless/intel/iwlwifi/fw/dbg.c
2065
if (!frag->size)
drivers/net/wireless/intel/iwlwifi/fw/dbg.c
2068
size += sizeof(struct iwl_fw_ini_error_dump_range) + frag->size;
drivers/net/wireless/intel/iwlwifi/iwl-dbg-tlv.c
403
struct iwl_dram_data *frag = &fw_mon->frags[i];
drivers/net/wireless/intel/iwlwifi/iwl-dbg-tlv.c
405
dma_free_coherent(trans->dev, frag->size, frag->block,
drivers/net/wireless/intel/iwlwifi/iwl-dbg-tlv.c
406
frag->physical);
drivers/net/wireless/intel/iwlwifi/iwl-dbg-tlv.c
408
frag->physical = 0;
drivers/net/wireless/intel/iwlwifi/iwl-dbg-tlv.c
409
frag->block = NULL;
drivers/net/wireless/intel/iwlwifi/iwl-dbg-tlv.c
410
frag->size = 0;
drivers/net/wireless/intel/iwlwifi/iwl-dbg-tlv.c
541
struct iwl_dram_data *frag, u32 pages)
drivers/net/wireless/intel/iwlwifi/iwl-dbg-tlv.c
546
if (!frag || frag->size || !pages)
drivers/net/wireless/intel/iwlwifi/iwl-dbg-tlv.c
572
frag->physical = physical;
drivers/net/wireless/intel/iwlwifi/iwl-dbg-tlv.c
573
frag->block = block;
drivers/net/wireless/intel/iwlwifi/iwl-dbg-tlv.c
574
frag->size = pages * PAGE_SIZE;
drivers/net/wireless/intel/iwlwifi/iwl-dbg-tlv.c
707
struct iwl_buf_alloc_frag *frag = &data.frags[j];
drivers/net/wireless/intel/iwlwifi/iwl-dbg-tlv.c
711
frag->addr = cpu_to_le64(fw_mon_frag->physical);
drivers/net/wireless/intel/iwlwifi/iwl-dbg-tlv.c
712
frag->size = cpu_to_le32(fw_mon_frag->size);
drivers/net/wireless/intel/iwlwifi/iwl-dbg-tlv.c
779
struct iwl_buf_alloc_frag *frag = &data->frags[j];
drivers/net/wireless/intel/iwlwifi/iwl-dbg-tlv.c
782
frag->addr = cpu_to_le64(fw_mon_frag->physical);
drivers/net/wireless/intel/iwlwifi/iwl-dbg-tlv.c
783
frag->size = cpu_to_le32(fw_mon_frag->size);
drivers/net/wireless/intel/iwlwifi/pcie/ctxt-info-v2.c
76
struct iwl_dram_data *frag =
drivers/net/wireless/intel/iwlwifi/pcie/ctxt-info-v2.c
79
dbg_cfg->hwm_base_addr = cpu_to_le64(frag->physical);
drivers/net/wireless/intel/iwlwifi/pcie/ctxt-info-v2.c
80
dbg_cfg->hwm_size = cpu_to_le32(frag->size);
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/trans.c
831
struct iwl_dram_data *frag;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/trans.c
851
frag = &trans->dbg.fw_mon_ini[alloc_id].frags[0];
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/trans.c
857
frag->physical >> MON_BUFF_SHIFT_VER2);
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/trans.c
859
(frag->physical + frag->size - 256) >>
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx-gen2.c
363
const skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx-gen2.c
365
unsigned int fragsz = skb_frag_size(frag);
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx-gen2.c
371
tb_phys = skb_frag_dma_map(trans->dev, frag, 0,
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx-gen2.c
374
skb_frag_address(frag),
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx-gen2.c
398
struct sk_buff *frag;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx-gen2.c
457
skb_walk_frags(skb, frag) {
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx-gen2.c
460
tb_phys = dma_map_single(trans->dev, frag->data,
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx-gen2.c
461
skb_headlen(frag), DMA_TO_DEVICE);
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx-gen2.c
463
frag->data,
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx-gen2.c
464
skb_headlen(frag), NULL,
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx-gen2.c
468
if (iwl_txq_gen2_tx_add_frags(trans, frag, tfd, out_meta))
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
1716
const skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
1720
if (!skb_frag_size(frag))
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
1723
tb_phys = skb_frag_dma_map(trans->dev, frag, 0,
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
1724
skb_frag_size(frag), DMA_TO_DEVICE);
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
1728
trace_iwlwifi_dev_tx_tb(trans->dev, skb, skb_frag_address(frag),
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
1729
tb_phys, skb_frag_size(frag));
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
1731
skb_frag_size(frag), false);
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
2254
struct sk_buff *frag;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
2260
skb_walk_frags(skb, frag) {
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
2261
if (unlikely(iwl_fill_data_tbs(trans, frag, txq, 0,
drivers/net/wireless/mediatek/mt76/testmode.c
147
struct sk_buff *frag;
drivers/net/wireless/mediatek/mt76/testmode.c
155
frag = alloc_skb(frag_len, GFP_KERNEL);
drivers/net/wireless/mediatek/mt76/testmode.c
156
if (!frag) {
drivers/net/wireless/mediatek/mt76/testmode.c
162
get_random_bytes(__skb_put(frag, frag_len), frag_len);
drivers/net/wireless/mediatek/mt76/testmode.c
163
head->len += frag->len;
drivers/net/wireless/mediatek/mt76/testmode.c
164
head->data_len += frag->len;
drivers/net/wireless/mediatek/mt76/testmode.c
166
*frag_tail = frag;
drivers/net/wireless/mediatek/mt7601u/dma.c
32
u32 true_len, hdr_len = 0, copy, frag;
drivers/net/wireless/mediatek/mt7601u/dma.c
56
frag = true_len - copy;
drivers/net/wireless/mediatek/mt7601u/dma.c
61
if (frag) {
drivers/net/wireless/mediatek/mt7601u/dma.c
63
frag, truesize);
drivers/net/wireless/realtek/rtl8xxxu/rtl8xxxu.h
168
u32 frag:4;
drivers/net/wireless/realtek/rtl8xxxu/rtl8xxxu.h
236
u32 frag:4;
drivers/net/wireless/realtek/rtl8xxxu/rtl8xxxu.h
312
u32 frag:4;
drivers/net/wireless/realtek/rtl8xxxu/rtl8xxxu.h
375
u32 frag:4; /* 16 */
drivers/net/wireless/realtek/rtlwifi/rtl8188ee/trx.h
755
u32 frag:4;
drivers/net/wireless/realtek/rtlwifi/rtl8192ce/trx.h
485
u32 frag:4;
drivers/net/wireless/realtek/rtlwifi/rtl8192cu/mac.h
97
u32 frag:4;
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/trx.h
691
u32 frag:4;
drivers/net/wireless/realtek/rtlwifi/rtl8723ae/trx.h
487
u32 frag:4;
drivers/net/wireless/realtek/rtlwifi/rtl8723be/trx.h
600
u32 frag:4;
drivers/net/wireless/realtek/rtlwifi/rtl8821ae/trx.h
606
u32 frag:4;
drivers/net/wireless/realtek/rtw89/core.h
1122
u8 frag;
drivers/net/wwan/t7xx/t7xx_hif_dpmaif_tx.c
289
skb_frag_t *frag = shinfo->frags + wr_cnt - 1;
drivers/net/wwan/t7xx/t7xx_hif_dpmaif_tx.c
291
data_len = skb_frag_size(frag);
drivers/net/wwan/t7xx/t7xx_hif_dpmaif_tx.c
292
data_addr = skb_frag_address(frag);
drivers/net/xen-netback/netback.c
146
static u16 frag_get_pending_idx(skb_frag_t *frag)
drivers/net/xen-netback/netback.c
148
return (u16)skb_frag_off(frag);
drivers/net/xen-netback/netback.c
151
static void frag_set_pending_idx(skb_frag_t *frag, u16 pending_idx)
drivers/net/xen-netback/netback.c
153
skb_frag_off_set(frag, pending_idx);
drivers/net/xen-netback/netback.c
696
skb_frag_t *frag = shinfo->frags + i;
drivers/net/xen-netback/netback.c
701
pending_idx = frag_get_pending_idx(frag);
drivers/net/xen-netback/rx.c
245
int frag; /* frag == -1 => frag_iter->head */
drivers/net/xen-netback/rx.c
269
pkt->frag = -1;
drivers/net/xen-netback/rx.c
342
pkt->frag++;
drivers/net/xen-netback/rx.c
345
if (pkt->frag >= nr_frags) {
drivers/net/xen-netback/rx.c
351
pkt->frag = -1;
drivers/net/xen-netback/rx.c
366
if (pkt->frag == -1) {
drivers/net/xen-netback/rx.c
370
skb_frag_t *frag = &skb_shinfo(frag_iter)->frags[pkt->frag];
drivers/net/xen-netback/rx.c
372
frag_data = skb_frag_address(frag);
drivers/net/xen-netback/rx.c
373
frag_len = skb_frag_size(frag);
drivers/net/xen-netfront.c
573
skb_frag_t *frag = skb_shinfo(skb)->frags + i;
drivers/net/xen-netfront.c
574
unsigned long size = skb_frag_size(frag);
drivers/net/xen-netfront.c
575
unsigned long offset = skb_frag_off(frag);
drivers/net/xen-netfront.c
827
skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
drivers/net/xen-netfront.c
828
xennet_make_txreqs(&info, skb_frag_page(frag),
drivers/net/xen-netfront.c
829
skb_frag_off(frag),
drivers/net/xen-netfront.c
830
skb_frag_size(frag));
drivers/nfc/pn533/pn533.c
2210
struct sk_buff *frag;
drivers/nfc/pn533/pn533.c
2221
frag = pn533_alloc_skb(dev, frag_size);
drivers/nfc/pn533/pn533.c
2222
if (!frag) {
drivers/nfc/pn533/pn533.c
2229
skb_reserve(frag, 1);
drivers/nfc/pn533/pn533.c
2233
*(u8 *)skb_push(frag, sizeof(u8)) =
drivers/nfc/pn533/pn533.c
2236
*(u8 *)skb_push(frag, sizeof(u8)) = 1; /* TG */
drivers/nfc/pn533/pn533.c
2239
skb_put_data(frag, skb->data, frag_size);
drivers/nfc/pn533/pn533.c
2245
skb_queue_tail(&dev->fragment_skb, frag);
drivers/s390/net/qeth_core_main.c
3900
skb_frag_t *frag = &skb_shinfo(skb)->frags[cnt];
drivers/s390/net/qeth_core_main.c
3903
(addr_t)skb_frag_address(frag),
drivers/s390/net/qeth_core_main.c
3904
(addr_t)skb_frag_address(frag) + skb_frag_size(frag));
drivers/s390/net/qeth_core_main.c
4115
skb_frag_t *frag = &skb_shinfo(skb)->frags[cnt];
drivers/s390/net/qeth_core_main.c
4117
data = skb_frag_address(frag);
drivers/s390/net/qeth_core_main.c
4118
length = skb_frag_size(frag);
drivers/scsi/bnx2fc/bnx2fc_fcoe.c
344
skb_frag_t *frag;
drivers/scsi/bnx2fc/bnx2fc_fcoe.c
349
frag = &skb_shinfo(skb)->frags[skb_shinfo(skb)->nr_frags - 1];
drivers/scsi/bnx2fc/bnx2fc_fcoe.c
350
cp = kmap_atomic(skb_frag_page(frag)) + skb_frag_off(frag);
drivers/scsi/cxgbi/libcxgbi.c
2210
struct page_frag *frag = tdata->frags;
drivers/scsi/cxgbi/libcxgbi.c
2219
for (i = 0; i < tdata->nr_frags; i++, frag++) {
drivers/scsi/cxgbi/libcxgbi.c
2220
char *src = kmap_atomic(frag->page);
drivers/scsi/cxgbi/libcxgbi.c
2222
memcpy(dst, src + frag->offset, frag->size);
drivers/scsi/cxgbi/libcxgbi.c
2223
dst += frag->size;
drivers/scsi/cxgbi/libcxgbi.c
2233
for (i = 0; i < tdata->nr_frags; i++, frag++) {
drivers/scsi/cxgbi/libcxgbi.c
2234
get_page(frag->page);
drivers/scsi/cxgbi/libcxgbi.c
2235
skb_fill_page_desc(skb, i, frag->page,
drivers/scsi/cxgbi/libcxgbi.c
2236
frag->offset, frag->size);
drivers/scsi/fcoe/fcoe.c
1502
skb_frag_t *frag;
drivers/scsi/fcoe/fcoe.c
1507
frag = &skb_shinfo(skb)->frags[skb_shinfo(skb)->nr_frags - 1];
drivers/scsi/fcoe/fcoe.c
1508
cp = kmap_atomic(skb_frag_page(frag)) + skb_frag_off(frag);
drivers/scsi/fcoe/fcoe_transport.c
312
skb_frag_t *frag;
drivers/scsi/fcoe/fcoe_transport.c
321
frag = &skb_shinfo(skb)->frags[i];
drivers/scsi/fcoe/fcoe_transport.c
322
off = skb_frag_off(frag);
drivers/scsi/fcoe/fcoe_transport.c
323
len = skb_frag_size(frag);
drivers/scsi/fcoe/fcoe_transport.c
327
skb_frag_page(frag) + (off >> PAGE_SHIFT));
drivers/scsi/qedf/qedf_main.c
1177
skb_frag_t *frag;
drivers/scsi/qedf/qedf_main.c
1183
frag = &skb_shinfo(skb)->frags[skb_shinfo(skb)->nr_frags - 1];
drivers/scsi/qedf/qedf_main.c
1184
cp = kmap_atomic(skb_frag_page(frag)) + skb_frag_off(frag);
drivers/staging/rtl8723bs/hal/rtl8723bs_recv.c
59
pattrib->frag_num = (u8)prxreport->frag;
drivers/staging/rtl8723bs/include/rtl8723b_recv.h
50
u32 frag:4;
drivers/tty/serial/esp32_uart.c
380
u32 frag = (sclk * 16) / baud - div * 16;
drivers/tty/serial/esp32_uart.c
383
div | FIELD_PREP(UART_CLKDIV_FRAG, frag));
fs/adfs/map.c
104
frag, start, fragend);
fs/adfs/map.c
123
u32 frag;
fs/adfs/map.c
129
frag = GET_FRAG_ID(map, start, idmask);
fs/adfs/map.c
135
if (frag == 0)
fs/adfs/map.c
139
start += frag;
fs/adfs/map.c
141
frag = GET_FRAG_ID(map, start, idmask);
fs/adfs/map.c
148
} while (frag >= idlen + 1);
fs/adfs/map.c
150
if (frag != 0)
fs/adfs/map.c
76
u32 frag;
fs/adfs/map.c
78
frag = GET_FRAG_ID(map, 8, idmask & 0x7fff);
fs/adfs/map.c
79
freelink = frag ? 8 + frag : 0;
fs/adfs/map.c
82
frag = GET_FRAG_ID(map, start, idmask);
fs/adfs/map.c
89
freelink += frag & 0x7fff;
fs/adfs/map.c
90
} else if (frag == frag_id) {
fs/ceph/dir.c
312
return !ceph_frag_contains_value(dfi->frag, fpos_hash(pos));
fs/ceph/dir.c
314
return dfi->frag != fpos_frag(pos);
fs/ceph/dir.c
327
unsigned frag = -1;
fs/ceph/dir.c
400
if (frag == (unsigned)-1)
fs/ceph/dir.c
401
frag = ceph_choose_frag(ci, fpos_hash(ctx->pos),
fs/ceph/dir.c
404
frag = fpos_frag(ctx->pos);
fs/ceph/dir.c
408
inode, ceph_vinop(inode), frag, dfi->last_name);
fs/ceph/dir.c
421
req->r_direct_hash = ceph_frag_value(frag);
fs/ceph/dir.c
449
req->r_args.readdir.frag = cpu_to_le32(frag);
fs/ceph/dir.c
463
inode, ceph_vinop(inode), err, frag,
fs/ceph/dir.c
469
if (le32_to_cpu(rinfo->dir_dir->frag) != frag) {
fs/ceph/dir.c
470
frag = le32_to_cpu(rinfo->dir_dir->frag);
fs/ceph/dir.c
474
ctx->pos = ceph_make_fpos(frag,
fs/ceph/dir.c
480
dfi->frag = frag;
fs/ceph/dir.c
488
} else if (ceph_frag_is_leftmost(frag) &&
fs/ceph/dir.c
525
inode, ceph_vinop(inode), dfi->frag, rinfo->dir_nr, ctx->pos,
fs/ceph/dir.c
581
frag = dfi->frag;
fs/ceph/dir.c
586
if (!ceph_frag_is_rightmost(dfi->frag)) {
fs/ceph/dir.c
587
frag = ceph_frag_next(dfi->frag);
fs/ceph/dir.c
589
loff_t new_pos = ceph_make_fpos(ceph_frag_value(frag),
fs/ceph/dir.c
595
ctx->pos = ceph_make_fpos(frag, dfi->next_offset,
fs/ceph/dir.c
601
ceph_vinop(inode), frag);
fs/ceph/dir.c
663
} else if (dfi->frag != fpos_frag(new_pos)) {
fs/ceph/inode.c
1970
u32 frag = le32_to_cpu(req->r_args.readdir.frag);
fs/ceph/inode.c
1992
le32_to_cpu(rinfo->dir_dir->frag) != frag) {
fs/ceph/inode.c
1993
doutc(cl, "got new frag %x -> %x\n", frag,
fs/ceph/inode.c
1994
le32_to_cpu(rinfo->dir_dir->frag));
fs/ceph/inode.c
1995
frag = le32_to_cpu(rinfo->dir_dir->frag);
fs/ceph/inode.c
2008
if (ceph_frag_is_leftmost(frag) &&
fs/ceph/inode.c
2043
rde->offset = ceph_make_fpos(frag, fpos_offset++, false);
fs/ceph/inode.c
308
struct ceph_inode_frag *frag;
fs/ceph/inode.c
314
frag = rb_entry(parent, struct ceph_inode_frag, node);
fs/ceph/inode.c
315
c = ceph_frag_compare(f, frag->frag);
fs/ceph/inode.c
321
return frag;
fs/ceph/inode.c
324
frag = kmalloc_obj(*frag, GFP_NOFS);
fs/ceph/inode.c
325
if (!frag)
fs/ceph/inode.c
328
frag->frag = f;
fs/ceph/inode.c
329
frag->split_by = 0;
fs/ceph/inode.c
330
frag->mds = -1;
fs/ceph/inode.c
331
frag->ndist = 0;
fs/ceph/inode.c
333
rb_link_node(&frag->node, parent, p);
fs/ceph/inode.c
334
rb_insert_color(&frag->node, &ci->i_fragtree);
fs/ceph/inode.c
337
return frag;
fs/ceph/inode.c
348
struct ceph_inode_frag *frag =
fs/ceph/inode.c
350
int c = ceph_frag_compare(f, frag->frag);
fs/ceph/inode.c
356
return frag;
fs/ceph/inode.c
371
struct ceph_inode_frag *frag;
fs/ceph/inode.c
380
frag = __ceph_find_frag(ci, t);
fs/ceph/inode.c
381
if (!frag)
fs/ceph/inode.c
383
if (frag->split_by == 0) {
fs/ceph/inode.c
385
memcpy(pfrag, frag, sizeof(*pfrag));
fs/ceph/inode.c
392
nway = 1 << frag->split_by;
fs/ceph/inode.c
394
frag->split_by, nway);
fs/ceph/inode.c
396
n = ceph_frag_make_child(t, frag->split_by, i);
fs/ceph/inode.c
429
struct ceph_inode_frag *frag;
fs/ceph/inode.c
430
u32 id = le32_to_cpu(dirinfo->frag);
fs/ceph/inode.c
448
frag = __ceph_find_frag(ci, id);
fs/ceph/inode.c
449
if (!frag)
fs/ceph/inode.c
451
if (frag->split_by == 0) {
fs/ceph/inode.c
455
rb_erase(&frag->node, &ci->i_fragtree);
fs/ceph/inode.c
456
kfree(frag);
fs/ceph/inode.c
461
frag->mds = -1;
fs/ceph/inode.c
462
frag->ndist = 0;
fs/ceph/inode.c
469
frag = __get_or_create_frag(ci, id);
fs/ceph/inode.c
470
if (IS_ERR(frag)) {
fs/ceph/inode.c
475
le32_to_cpu(dirinfo->frag));
fs/ceph/inode.c
480
frag->mds = mds;
fs/ceph/inode.c
481
frag->ndist = min_t(u32, ndist, CEPH_MAX_DIRFRAG_REP);
fs/ceph/inode.c
482
for (i = 0; i < frag->ndist; i++)
fs/ceph/inode.c
483
frag->dist[i] = le32_to_cpu(dirinfo->dist[i]);
fs/ceph/inode.c
485
ceph_vinop(inode), frag->frag, frag->ndist);
fs/ceph/inode.c
496
return ceph_frag_compare(le32_to_cpu(ls->frag),
fs/ceph/inode.c
497
le32_to_cpu(rs->frag));
fs/ceph/inode.c
500
static bool is_frag_child(u32 f, struct ceph_inode_frag *frag)
fs/ceph/inode.c
502
if (!frag)
fs/ceph/inode.c
504
if (ceph_frag_bits(f) != ceph_frag_bits(frag->frag) + frag->split_by)
fs/ceph/inode.c
506
return ceph_frag_contains_value(frag->frag, ceph_frag_value(f));
fs/ceph/inode.c
515
struct ceph_inode_frag *frag, *prev_frag = NULL;
fs/ceph/inode.c
527
id = le32_to_cpu(fragtree->splits[i].frag);
fs/ceph/inode.c
532
frag = rb_entry(rb_node, struct ceph_inode_frag, node);
fs/ceph/inode.c
533
if (frag->frag != ceph_frag_make(0, 0) || rb_next(rb_node))
fs/ceph/inode.c
537
id = le32_to_cpu(dirinfo->frag);
fs/ceph/inode.c
552
id = le32_to_cpu(fragtree->splits[i].frag);
fs/ceph/inode.c
560
frag = NULL;
fs/ceph/inode.c
562
frag = rb_entry(rb_node, struct ceph_inode_frag, node);
fs/ceph/inode.c
563
if (ceph_frag_compare(frag->frag, id) >= 0) {
fs/ceph/inode.c
564
if (frag->frag != id)
fs/ceph/inode.c
565
frag = NULL;
fs/ceph/inode.c
572
if (frag->split_by > 0 ||
fs/ceph/inode.c
573
!is_frag_child(frag->frag, prev_frag)) {
fs/ceph/inode.c
574
rb_erase(&frag->node, &ci->i_fragtree);
fs/ceph/inode.c
575
if (frag->split_by > 0)
fs/ceph/inode.c
577
kfree(frag);
fs/ceph/inode.c
579
frag = NULL;
fs/ceph/inode.c
581
if (!frag) {
fs/ceph/inode.c
582
frag = __get_or_create_frag(ci, id);
fs/ceph/inode.c
583
if (IS_ERR(frag))
fs/ceph/inode.c
586
if (frag->split_by == 0)
fs/ceph/inode.c
588
frag->split_by = split_by;
fs/ceph/inode.c
589
doutc(cl, " frag %x split by %d\n", frag->frag, frag->split_by);
fs/ceph/inode.c
590
prev_frag = frag;
fs/ceph/inode.c
593
frag = rb_entry(rb_node, struct ceph_inode_frag, node);
fs/ceph/inode.c
596
if (frag->split_by > 0 ||
fs/ceph/inode.c
597
!is_frag_child(frag->frag, prev_frag)) {
fs/ceph/inode.c
598
rb_erase(&frag->node, &ci->i_fragtree);
fs/ceph/inode.c
599
if (frag->split_by > 0)
fs/ceph/inode.c
601
kfree(frag);
fs/ceph/inode.c
738
struct ceph_inode_frag *frag;
fs/ceph/inode.c
775
frag = rb_entry(n, struct ceph_inode_frag, node);
fs/ceph/inode.c
777
kfree(frag);
fs/ceph/mds_client.c
1370
struct ceph_inode_frag frag;
fs/ceph/mds_client.c
1373
ceph_choose_frag(ci, hash, &frag, &found);
fs/ceph/mds_client.c
1375
if (mode == USE_ANY_MDS && frag.ndist > 0) {
fs/ceph/mds_client.c
1380
r %= frag.ndist;
fs/ceph/mds_client.c
1381
mds = frag.dist[r];
fs/ceph/mds_client.c
1383
inode, ceph_vinop(inode), frag.frag,
fs/ceph/mds_client.c
1384
mds, (int)r, frag.ndist);
fs/ceph/mds_client.c
1394
if (frag.mds >= 0) {
fs/ceph/mds_client.c
1396
mds = frag.mds;
fs/ceph/mds_client.c
1398
inode, ceph_vinop(inode), frag.frag, mds);
fs/ceph/super.h
300
u32 frag;
fs/ceph/super.h
875
u32 frag;
fs/configfs/dir.c
1300
struct configfs_fragment *frag;
fs/configfs/dir.c
1319
frag = new_fragment();
fs/configfs/dir.c
1320
if (!frag) {
fs/configfs/dir.c
1427
ret = configfs_attach_group(parent_item, item, dentry, frag);
fs/configfs/dir.c
1429
ret = configfs_attach_item(parent_item, item, dentry, frag);
fs/configfs/dir.c
1466
put_fragment(frag);
fs/configfs/dir.c
1478
struct configfs_fragment *frag;
fs/configfs/dir.c
1536
frag = sd->s_frag;
fs/configfs/dir.c
1537
if (down_write_killable(&frag->frag_sem)) {
fs/configfs/dir.c
1544
frag->frag_dead = true;
fs/configfs/dir.c
1545
up_write(&frag->frag_sem);
fs/configfs/dir.c
171
void put_fragment(struct configfs_fragment *frag)
fs/configfs/dir.c
173
if (frag && atomic_dec_and_test(&frag->frag_count))
fs/configfs/dir.c
174
kfree(frag);
fs/configfs/dir.c
1759
struct configfs_fragment *frag;
fs/configfs/dir.c
1762
frag = new_fragment();
fs/configfs/dir.c
1763
if (!frag)
fs/configfs/dir.c
177
struct configfs_fragment *get_fragment(struct configfs_fragment *frag)
fs/configfs/dir.c
1773
ret = create_default_group(parent_group, group, frag);
fs/configfs/dir.c
1781
put_fragment(frag);
fs/configfs/dir.c
1788
put_fragment(frag);
fs/configfs/dir.c
179
if (likely(frag))
fs/configfs/dir.c
180
atomic_inc(&frag->frag_count);
fs/configfs/dir.c
1805
struct configfs_fragment *frag = sd->s_frag;
fs/configfs/dir.c
1807
down_write(&frag->frag_sem);
fs/configfs/dir.c
1808
frag->frag_dead = true;
fs/configfs/dir.c
1809
up_write(&frag->frag_sem);
fs/configfs/dir.c
181
return frag;
fs/configfs/dir.c
1882
struct configfs_fragment *frag;
fs/configfs/dir.c
1884
frag = new_fragment();
fs/configfs/dir.c
1885
if (!frag)
fs/configfs/dir.c
189
struct configfs_fragment *frag)
fs/configfs/dir.c
1890
put_fragment(frag);
fs/configfs/dir.c
1913
dentry, frag);
fs/configfs/dir.c
1933
put_fragment(frag);
fs/configfs/dir.c
1944
struct configfs_fragment *frag = sd->s_frag;
fs/configfs/dir.c
1951
down_write(&frag->frag_sem);
fs/configfs/dir.c
1952
frag->frag_dead = true;
fs/configfs/dir.c
1953
up_write(&frag->frag_sem);
fs/configfs/dir.c
208
sd->s_frag = get_fragment(frag);
fs/configfs/dir.c
254
umode_t mode, int type, struct configfs_fragment *frag)
fs/configfs/dir.c
258
sd = configfs_new_dirent(parent_sd, element, type, frag);
fs/configfs/dir.c
293
struct configfs_fragment *frag)
fs/configfs/dir.c
304
frag);
fs/configfs/dir.c
647
struct configfs_fragment *frag);
fs/configfs/dir.c
696
struct configfs_fragment *frag)
fs/configfs/dir.c
712
&group->cg_item, child, frag);
fs/configfs/dir.c
727
struct configfs_fragment *frag)
fs/configfs/dir.c
733
ret = create_default_group(group, new_group, frag);
fs/configfs/dir.c
848
struct configfs_fragment *frag)
fs/configfs/dir.c
852
ret = configfs_create_dir(item, dentry, frag);
fs/configfs/dir.c
883
struct configfs_fragment *frag)
fs/configfs/dir.c
888
ret = configfs_attach_item(parent_item, item, dentry, frag);
fs/configfs/dir.c
904
ret = populate_groups(to_config_group(item), frag);
fs/configfs/file.c
109
struct configfs_fragment *frag = to_frag(file);
fs/configfs/file.c
125
down_read(&frag->frag_sem);
fs/configfs/file.c
126
if (!frag->frag_dead)
fs/configfs/file.c
130
up_read(&frag->frag_sem);
fs/configfs/file.c
150
down_read(&frag->frag_sem);
fs/configfs/file.c
151
if (!frag->frag_dead)
fs/configfs/file.c
156
up_read(&frag->frag_sem);
fs/configfs/file.c
202
struct configfs_fragment *frag = to_frag(file);
fs/configfs/file.c
205
down_read(&frag->frag_sem);
fs/configfs/file.c
206
if (!frag->frag_dead)
fs/configfs/file.c
208
up_read(&frag->frag_sem);
fs/configfs/file.c
293
struct configfs_fragment *frag = to_frag(file);
fs/configfs/file.c
304
down_read(&frag->frag_sem);
fs/configfs/file.c
305
if (unlikely(frag->frag_dead))
fs/configfs/file.c
367
up_read(&frag->frag_sem);
fs/configfs/file.c
373
up_read(&frag->frag_sem);
fs/configfs/file.c
406
struct configfs_fragment *frag = to_frag(file);
fs/configfs/file.c
408
down_read(&frag->frag_sem);
fs/configfs/file.c
409
if (!frag->frag_dead) {
fs/configfs/file.c
415
up_read(&frag->frag_sem);
fs/configfs/file.c
58
struct configfs_fragment *frag = to_frag(file);
fs/configfs/file.c
66
down_read(&frag->frag_sem);
fs/configfs/file.c
67
if (!frag->frag_dead)
fs/configfs/file.c
69
up_read(&frag->frag_sem);
fs/jffs2/debug.c
105
if ((frag->ofs+frag->size) & (PAGE_SIZE-1) && frag_next(frag)
fs/jffs2/debug.c
106
&& frag_next(frag)->size < PAGE_SIZE && frag_next(frag)->node) {
fs/jffs2/debug.c
108
ref_offset(fn->raw), frag->ofs, frag->ofs+frag->size);
fs/jffs2/debug.c
77
struct jffs2_node_frag *frag;
fs/jffs2/debug.c
80
for (frag = frag_first(&f->fragtree); frag; frag = frag_next(frag)) {
fs/jffs2/debug.c
81
struct jffs2_full_dnode *fn = frag->node;
fs/jffs2/debug.c
98
if (frag->ofs & (PAGE_SIZE-1) && frag_prev(frag)
fs/jffs2/debug.c
99
&& frag_prev(frag)->size < PAGE_SIZE && frag_prev(frag)->node) {
fs/jffs2/gc.c
1018
struct jffs2_node_frag *frag;
fs/jffs2/gc.c
1083
frag = frag_last(&f->fragtree);
fs/jffs2/gc.c
1084
if (frag)
fs/jffs2/gc.c
1087
ilen = frag->ofs + frag->size;
fs/jffs2/gc.c
1139
for (frag = jffs2_lookup_node_frag(&f->fragtree, fn->ofs);
fs/jffs2/gc.c
1140
frag; frag = frag_next(frag)) {
fs/jffs2/gc.c
1141
if (frag->ofs > fn->size + fn->ofs)
fs/jffs2/gc.c
1143
if (frag->node == fn) {
fs/jffs2/gc.c
1144
frag->node = new_fn;
fs/jffs2/gc.c
1193
struct jffs2_node_frag *frag;
fs/jffs2/gc.c
1199
frag = jffs2_lookup_node_frag(&f->fragtree, start);
fs/jffs2/gc.c
1203
BUG_ON(frag->ofs != start);
fs/jffs2/gc.c
1206
while((frag = frag_prev(frag)) && frag->ofs >= min) {
fs/jffs2/gc.c
1210
if (frag->ofs > min) {
fs/jffs2/gc.c
1212
frag->ofs, frag->ofs+frag->size);
fs/jffs2/gc.c
1213
start = frag->ofs;
fs/jffs2/gc.c
1217
if (!frag->node || !frag->node->raw) {
fs/jffs2/gc.c
1219
frag->ofs, frag->ofs+frag->size);
fs/jffs2/gc.c
1227
struct jffs2_raw_node_ref *raw = frag->node->raw;
fs/jffs2/gc.c
1234
frag->ofs,
fs/jffs2/gc.c
1235
frag->ofs + frag->size,
fs/jffs2/gc.c
1237
start = frag->ofs;
fs/jffs2/gc.c
1242
frag->ofs,
fs/jffs2/gc.c
1243
frag->ofs + frag->size,
fs/jffs2/gc.c
1249
frag->ofs,
fs/jffs2/gc.c
1250
frag->ofs + frag->size,
fs/jffs2/gc.c
1252
start = frag->ofs;
fs/jffs2/gc.c
1260
frag = jffs2_lookup_node_frag(&f->fragtree, end-1);
fs/jffs2/gc.c
1262
while((frag = frag_next(frag)) && frag->ofs+frag->size <= max) {
fs/jffs2/gc.c
1266
if (frag->ofs+frag->size < max) {
fs/jffs2/gc.c
1268
frag->ofs, frag->ofs+frag->size);
fs/jffs2/gc.c
1269
end = frag->ofs + frag->size;
fs/jffs2/gc.c
1273
if (!frag->node || !frag->node->raw) {
fs/jffs2/gc.c
1275
frag->ofs, frag->ofs+frag->size);
fs/jffs2/gc.c
1283
struct jffs2_raw_node_ref *raw = frag->node->raw;
fs/jffs2/gc.c
1290
frag->ofs,
fs/jffs2/gc.c
1291
frag->ofs + frag->size,
fs/jffs2/gc.c
1293
end = frag->ofs + frag->size;
fs/jffs2/gc.c
1298
frag->ofs,
fs/jffs2/gc.c
1299
frag->ofs + frag->size,
fs/jffs2/gc.c
1305
frag->ofs,
fs/jffs2/gc.c
1306
frag->ofs + frag->size,
fs/jffs2/gc.c
1308
end = frag->ofs + frag->size;
fs/jffs2/gc.c
500
struct jffs2_node_frag *frag;
fs/jffs2/gc.c
534
for (frag = frag_first(&f->fragtree); frag; frag = frag_next(frag)) {
fs/jffs2/gc.c
535
if (frag->node && frag->node->raw == raw) {
fs/jffs2/gc.c
536
fn = frag->node;
fs/jffs2/gc.c
537
end = frag->ofs + frag->size;
fs/jffs2/gc.c
539
start = frag->ofs;
fs/jffs2/gc.c
540
if (nrfrags == frag->node->frags)
fs/jffs2/gc.c
549
frag->node->raw = f->inocache->nodes;
fs/jffs2/nodelist.c
530
struct jffs2_node_frag *frag = NULL;
fs/jffs2/nodelist.c
537
frag = rb_entry(next, struct jffs2_node_frag, rb);
fs/jffs2/nodelist.c
539
if (frag->ofs + frag->size <= offset) {
fs/jffs2/nodelist.c
541
if (!prev || frag->ofs > prev->ofs)
fs/jffs2/nodelist.c
542
prev = frag;
fs/jffs2/nodelist.c
543
next = frag->rb.rb_right;
fs/jffs2/nodelist.c
544
} else if (frag->ofs > offset) {
fs/jffs2/nodelist.c
545
next = frag->rb.rb_left;
fs/jffs2/nodelist.c
547
return frag;
fs/jffs2/nodelist.c
567
struct jffs2_node_frag *frag, *next;
fs/jffs2/nodelist.c
570
rbtree_postorder_for_each_entry_safe(frag, next, root, rb) {
fs/jffs2/nodelist.c
571
if (frag->node && !(--frag->node->frags)) {
fs/jffs2/nodelist.c
575
jffs2_mark_node_obsolete(c, frag->node->raw);
fs/jffs2/nodelist.c
577
jffs2_free_full_dnode(frag->node);
fs/jffs2/nodelist.c
580
jffs2_free_node_frag(frag);
fs/jffs2/nodelist.c
61
struct jffs2_node_frag *frag = jffs2_lookup_node_frag(list, size);
fs/jffs2/nodelist.c
66
if (frag && frag->ofs != size) {
fs/jffs2/nodelist.c
67
if (frag->ofs+frag->size > size) {
fs/jffs2/nodelist.c
68
frag->size = size - frag->ofs;
fs/jffs2/nodelist.c
70
frag = frag_next(frag);
fs/jffs2/nodelist.c
72
while (frag && frag->ofs >= size) {
fs/jffs2/nodelist.c
73
struct jffs2_node_frag *next = frag_next(frag);
fs/jffs2/nodelist.c
75
frag_erase(frag, list);
fs/jffs2/nodelist.c
76
jffs2_obsolete_node_frag(c, frag);
fs/jffs2/nodelist.c
77
frag = next;
fs/jffs2/nodelist.c
83
frag = frag_last(list);
fs/jffs2/nodelist.c
86
if (!frag)
fs/jffs2/nodelist.c
88
if (frag->ofs + frag->size < size)
fs/jffs2/nodelist.c
89
return frag->ofs + frag->size;
fs/jffs2/nodelist.c
93
if (frag->node && (frag->ofs & (PAGE_SIZE - 1)) == 0) {
fs/jffs2/nodelist.c
95
frag->ofs, frag->ofs + frag->size);
fs/jffs2/nodelist.c
96
frag->node->raw->flash_offset = ref_offset(frag->node->raw) | REF_PRISTINE;
fs/jffs2/nodelist.h
347
#define frag_next(frag) rb_entry(rb_next(&(frag)->rb), struct jffs2_node_frag, rb)
fs/jffs2/nodelist.h
348
#define frag_prev(frag) rb_entry(rb_prev(&(frag)->rb), struct jffs2_node_frag, rb)
fs/jffs2/nodelist.h
349
#define frag_parent(frag) rb_entry(rb_parent(&(frag)->rb), struct jffs2_node_frag, rb)
fs/jffs2/nodelist.h
350
#define frag_left(frag) rb_entry((frag)->rb.rb_left, struct jffs2_node_frag, rb)
fs/jffs2/nodelist.h
351
#define frag_right(frag) rb_entry((frag)->rb.rb_right, struct jffs2_node_frag, rb)
fs/jffs2/nodelist.h
352
#define frag_erase(frag, list) rb_erase(&frag->rb, list)
fs/jffs2/read.c
161
struct jffs2_node_frag *frag;
fs/jffs2/read.c
167
frag = jffs2_lookup_node_frag(&f->fragtree, offset);
fs/jffs2/read.c
177
if (unlikely(!frag || frag->ofs > offset ||
fs/jffs2/read.c
178
frag->ofs + frag->size <= offset)) {
fs/jffs2/read.c
180
if (frag && frag->ofs > offset) {
fs/jffs2/read.c
182
f->inocache->ino, frag->ofs, offset);
fs/jffs2/read.c
183
holesize = min(holesize, frag->ofs - offset);
fs/jffs2/read.c
191
} else if (unlikely(!frag->node)) {
fs/jffs2/read.c
192
uint32_t holeend = min(end, frag->ofs + frag->size);
fs/jffs2/read.c
194
offset, holeend, frag->ofs,
fs/jffs2/read.c
195
frag->ofs + frag->size);
fs/jffs2/read.c
199
frag = frag_next(frag);
fs/jffs2/read.c
205
fragofs = offset - frag->ofs;
fs/jffs2/read.c
206
readlen = min(frag->size - fragofs, end - offset);
fs/jffs2/read.c
208
frag->ofs+fragofs,
fs/jffs2/read.c
209
frag->ofs + fragofs+readlen,
fs/jffs2/read.c
210
ref_offset(frag->node->raw),
fs/jffs2/read.c
211
ref_flags(frag->node->raw));
fs/jffs2/read.c
212
ret = jffs2_read_dnode(c, f, frag->node, buf, fragofs + frag->ofs - frag->node->ofs, readlen);
fs/jffs2/read.c
222
frag = frag_next(frag);
fs/jffs2/wbuf.c
187
struct jffs2_node_frag *frag;
fs/jffs2/wbuf.c
202
frag = jffs2_lookup_node_frag(&f->fragtree, je32_to_cpu(node->i.offset));
fs/jffs2/wbuf.c
203
BUG_ON(!frag);
fs/jffs2/wbuf.c
205
while (!frag->node || frag->node->raw != raw) {
fs/jffs2/wbuf.c
206
frag = frag_next(frag);
fs/jffs2/wbuf.c
207
BUG_ON(!frag);
fs/jffs2/wbuf.c
209
dbg_noderef("Will replace ->raw in full_dnode at %p\n", frag->node);
fs/jffs2/wbuf.c
210
return &frag->node->raw;
fs/squashfs/inode.c
137
unsigned int frag_offset, frag;
fs/squashfs/inode.c
148
frag = le32_to_cpu(sqsh_ino->fragment);
fs/squashfs/inode.c
149
if (frag != SQUASHFS_INVALID_FRAG) {
fs/squashfs/inode.c
159
frag_size = squashfs_frag_lookup(sb, frag, &frag_blk);
fs/squashfs/inode.c
189
unsigned int frag_offset, frag;
fs/squashfs/inode.c
204
frag = le32_to_cpu(sqsh_ino->fragment);
fs/squashfs/inode.c
205
if (frag != SQUASHFS_INVALID_FRAG) {
fs/squashfs/inode.c
215
frag_size = squashfs_frag_lookup(sb, frag, &frag_blk);
fs/ufs/balloc.c
847
struct ufs_cg_private_info * ucpi, unsigned frag, int cnt)
fs/ufs/balloc.c
851
unsigned blkno = ufs_fragstoblks(frag);
fs/ufs/inode.c
383
unsigned frag = fragment & uspi->s_fpbmask;
fs/ufs/inode.c
435
phys64 += frag;
fs/ufs/inode.c
445
map_bh(bh_result, sb, phys64 + frag);
fs/ufs/util.h
454
struct ufs_cg_private_info *ucpi, unsigned int frag)
fs/ufs/util.h
457
u8 *p = ubh_get_addr(ubh, ucpi->c_freeoff + (frag >> 3));
fs/ufs/util.h
464
mask = 0x0f << (frag & 4);
fs/ufs/util.h
467
mask = 0x03 << (frag & 6);
fs/ufs/util.h
470
mask = 0x01 << (frag & 7);
fs/ufs/util.h
477
struct ufs_cg_private_info *ucpi, unsigned int frag)
fs/ufs/util.h
480
u8 *p = ubh_get_addr(ubh, ucpi->c_freeoff + (frag >> 3));
fs/ufs/util.h
487
*p &= ~(0x0f << (frag & 4));
fs/ufs/util.h
490
*p &= ~(0x03 << (frag & 6));
fs/ufs/util.h
493
*p &= ~(0x01 << (frag & 7));
fs/ufs/util.h
499
struct ufs_cg_private_info *ucpi, unsigned int frag)
fs/ufs/util.h
502
u8 *p = ubh_get_addr(ubh, ucpi->c_freeoff + (frag >> 3));
fs/ufs/util.h
509
*p |= 0x0f << (frag & 4);
fs/ufs/util.h
512
*p |= 0x03 << (frag & 6);
fs/ufs/util.h
515
*p |= 0x01 << (frag & 7);
fs/xfs/scrub/refcount.c
115
struct xchk_refcnt_frag *frag;
fs/xfs/scrub/refcount.c
145
frag = kmalloc_obj(struct xchk_refcnt_frag, XCHK_GFP_FLAGS);
fs/xfs/scrub/refcount.c
146
if (!frag)
fs/xfs/scrub/refcount.c
148
memcpy(&frag->rm, rec, sizeof(frag->rm));
fs/xfs/scrub/refcount.c
149
list_add_tail(&frag->list, &refchk->fragments);
fs/xfs/scrub/refcount.c
167
struct xchk_refcnt_frag *frag;
fs/xfs/scrub/refcount.c
192
list_for_each_entry(frag, &refchk->fragments, list) {
fs/xfs/scrub/refcount.c
193
if (frag->rm.rm_startblock < bno)
fs/xfs/scrub/refcount.c
195
bno = frag->rm.rm_startblock;
fs/xfs/scrub/refcount.c
203
list_for_each_entry_safe(frag, n, &refchk->fragments, list) {
fs/xfs/scrub/refcount.c
204
if (frag->rm.rm_startblock > refchk->bno || nr > target_nr)
fs/xfs/scrub/refcount.c
206
bno = frag->rm.rm_startblock + frag->rm.rm_blockcount;
fs/xfs/scrub/refcount.c
209
list_move_tail(&frag->list, &worklist);
fs/xfs/scrub/refcount.c
224
list_for_each_entry_safe(frag, n, &worklist, list) {
fs/xfs/scrub/refcount.c
225
bno = frag->rm.rm_startblock + frag->rm.rm_blockcount;
fs/xfs/scrub/refcount.c
231
list_del(&frag->list);
fs/xfs/scrub/refcount.c
232
kfree(frag);
fs/xfs/scrub/refcount.c
237
list_for_each_entry_safe(frag, n, &refchk->fragments, list) {
fs/xfs/scrub/refcount.c
238
bno = frag->rm.rm_startblock + frag->rm.rm_blockcount;
fs/xfs/scrub/refcount.c
239
if (frag->rm.rm_startblock != rbno)
fs/xfs/scrub/refcount.c
241
list_move_tail(&frag->list, &worklist);
fs/xfs/scrub/refcount.c
272
list_for_each_entry_safe(frag, n, &worklist, list) {
fs/xfs/scrub/refcount.c
273
list_del(&frag->list);
fs/xfs/scrub/refcount.c
274
kfree(frag);
fs/xfs/scrub/refcount.c
276
list_for_each_entry_safe(frag, n, &refchk->fragments, list) {
fs/xfs/scrub/refcount.c
277
list_del(&frag->list);
fs/xfs/scrub/refcount.c
278
kfree(frag);
fs/xfs/scrub/refcount.c
297
struct xchk_refcnt_frag *frag;
fs/xfs/scrub/refcount.c
323
list_for_each_entry_safe(frag, n, &refchk.fragments, list) {
fs/xfs/scrub/refcount.c
324
list_del(&frag->list);
fs/xfs/scrub/refcount.c
325
kfree(frag);
fs/xfs/scrub/rtrefcount.c
129
struct xchk_rtrefcnt_frag *frag;
fs/xfs/scrub/rtrefcount.c
159
frag = kmalloc_obj(struct xchk_rtrefcnt_frag, XCHK_GFP_FLAGS);
fs/xfs/scrub/rtrefcount.c
160
if (!frag)
fs/xfs/scrub/rtrefcount.c
162
memcpy(&frag->rm, rec, sizeof(frag->rm));
fs/xfs/scrub/rtrefcount.c
163
list_add_tail(&frag->list, &refchk->fragments);
fs/xfs/scrub/rtrefcount.c
181
struct xchk_rtrefcnt_frag *frag;
fs/xfs/scrub/rtrefcount.c
206
list_for_each_entry(frag, &refchk->fragments, list) {
fs/xfs/scrub/rtrefcount.c
207
if (frag->rm.rm_startblock < bno)
fs/xfs/scrub/rtrefcount.c
209
bno = frag->rm.rm_startblock;
fs/xfs/scrub/rtrefcount.c
217
list_for_each_entry_safe(frag, n, &refchk->fragments, list) {
fs/xfs/scrub/rtrefcount.c
218
if (frag->rm.rm_startblock > refchk->bno || nr > target_nr)
fs/xfs/scrub/rtrefcount.c
220
bno = frag->rm.rm_startblock + frag->rm.rm_blockcount;
fs/xfs/scrub/rtrefcount.c
223
list_move_tail(&frag->list, &worklist);
fs/xfs/scrub/rtrefcount.c
238
list_for_each_entry_safe(frag, n, &worklist, list) {
fs/xfs/scrub/rtrefcount.c
239
bno = frag->rm.rm_startblock + frag->rm.rm_blockcount;
fs/xfs/scrub/rtrefcount.c
245
list_del(&frag->list);
fs/xfs/scrub/rtrefcount.c
246
kfree(frag);
fs/xfs/scrub/rtrefcount.c
251
list_for_each_entry_safe(frag, n, &refchk->fragments, list) {
fs/xfs/scrub/rtrefcount.c
252
bno = frag->rm.rm_startblock + frag->rm.rm_blockcount;
fs/xfs/scrub/rtrefcount.c
253
if (frag->rm.rm_startblock != rbno)
fs/xfs/scrub/rtrefcount.c
255
list_move_tail(&frag->list, &worklist);
fs/xfs/scrub/rtrefcount.c
286
list_for_each_entry_safe(frag, n, &worklist, list) {
fs/xfs/scrub/rtrefcount.c
287
list_del(&frag->list);
fs/xfs/scrub/rtrefcount.c
288
kfree(frag);
fs/xfs/scrub/rtrefcount.c
290
list_for_each_entry_safe(frag, n, &refchk->fragments, list) {
fs/xfs/scrub/rtrefcount.c
291
list_del(&frag->list);
fs/xfs/scrub/rtrefcount.c
292
kfree(frag);
fs/xfs/scrub/rtrefcount.c
311
struct xchk_rtrefcnt_frag *frag;
fs/xfs/scrub/rtrefcount.c
335
list_for_each_entry_safe(frag, n, &refchk.fragments, list) {
fs/xfs/scrub/rtrefcount.c
336
list_del(&frag->list);
fs/xfs/scrub/rtrefcount.c
337
kfree(frag);
include/linux/ceph/ceph_fs.h
434
__le32 frag; /* which dir fragment */
include/linux/ceph/ceph_fs.h
555
__le32 frag; /* this frag splits... */
include/linux/ceph/ceph_fs.h
608
__le32 frag; /* fragment */
include/linux/mlx5/driver.h
946
unsigned int frag;
include/linux/mlx5/driver.h
949
frag = ix >> fbc->log_frag_strides;
include/linux/mlx5/driver.h
951
return fbc->frags[frag].buf + ((fbc->frag_sz_m1 & ix) << fbc->log_stride);
include/linux/mlx5/mlx5_ifc.h
592
u8 frag[0x1];
include/linux/perf_event.h
1396
struct perf_raw_frag *frag = &raw->frag;
include/linux/perf_event.h
1406
sum += frag->size;
include/linux/perf_event.h
1407
if (perf_raw_frag_last(frag))
include/linux/perf_event.h
1409
frag = frag->next;
include/linux/perf_event.h
1414
frag->pad = raw->size - sum;
include/linux/perf_event.h
86
struct perf_raw_frag frag;
include/linux/perf_event.h
90
static __always_inline bool perf_raw_frag_last(const struct perf_raw_frag *frag)
include/linux/perf_event.h
92
return frag->pad < sizeof(u64);
include/linux/skbuff.h
2550
static inline void skb_frag_fill_netmem_desc(skb_frag_t *frag,
include/linux/skbuff.h
2554
frag->netmem = netmem;
include/linux/skbuff.h
2555
frag->offset = off;
include/linux/skbuff.h
2556
skb_frag_size_set(frag, size);
include/linux/skbuff.h
2559
static inline void skb_frag_fill_page_desc(skb_frag_t *frag,
include/linux/skbuff.h
2563
skb_frag_fill_netmem_desc(frag, page_to_netmem(page), off, size);
include/linux/skbuff.h
2570
skb_frag_t *frag = &shinfo->frags[i];
include/linux/skbuff.h
2572
skb_frag_fill_netmem_desc(frag, netmem, off, size);
include/linux/skbuff.h
3627
static inline unsigned int skb_frag_off(const skb_frag_t *frag)
include/linux/skbuff.h
3629
return frag->offset;
include/linux/skbuff.h
3637
static inline void skb_frag_off_add(skb_frag_t *frag, int delta)
include/linux/skbuff.h
3639
frag->offset += delta;
include/linux/skbuff.h
3647
static inline void skb_frag_off_set(skb_frag_t *frag, unsigned int offset)
include/linux/skbuff.h
3649
frag->offset = offset;
include/linux/skbuff.h
3664
static inline bool skb_frag_is_net_iov(const skb_frag_t *frag)
include/linux/skbuff.h
3666
return netmem_is_net_iov(frag->netmem);
include/linux/skbuff.h
3676
static inline struct net_iov *skb_frag_net_iov(const skb_frag_t *frag)
include/linux/skbuff.h
3678
if (!skb_frag_is_net_iov(frag))
include/linux/skbuff.h
3681
return netmem_to_net_iov(frag->netmem);
include/linux/skbuff.h
3691
static inline struct page *skb_frag_page(const skb_frag_t *frag)
include/linux/skbuff.h
3693
if (skb_frag_is_net_iov(frag))
include/linux/skbuff.h
3696
return netmem_to_page(frag->netmem);
include/linux/skbuff.h
3705
static inline netmem_ref skb_frag_netmem(const skb_frag_t *frag)
include/linux/skbuff.h
3707
return frag->netmem;
include/linux/skbuff.h
371
static inline unsigned int skb_frag_size(const skb_frag_t *frag)
include/linux/skbuff.h
3722
static inline void *skb_frag_address(const skb_frag_t *frag)
include/linux/skbuff.h
3724
if (!skb_frag_page(frag))
include/linux/skbuff.h
3727
return page_address(skb_frag_page(frag)) + skb_frag_off(frag);
include/linux/skbuff.h
373
return frag->len;
include/linux/skbuff.h
3737
static inline void *skb_frag_address_safe(const skb_frag_t *frag)
include/linux/skbuff.h
3739
struct page *page = skb_frag_page(frag);
include/linux/skbuff.h
3749
return ptr + skb_frag_off(frag);
include/linux/skbuff.h
3777
const skb_frag_t *frag,
include/linux/skbuff.h
3781
if (skb_frag_is_net_iov(frag)) {
include/linux/skbuff.h
3782
return netmem_to_net_iov(frag->netmem)->desc.dma_addr +
include/linux/skbuff.h
3783
offset + frag->offset;
include/linux/skbuff.h
3785
return dma_map_page(dev, skb_frag_page(frag),
include/linux/skbuff.h
3786
skb_frag_off(frag) + offset, size, dir);
include/linux/skbuff.h
3789
#define skb_frag_dma_map(dev, frag, ...) \
include/linux/skbuff.h
3791
COUNT_ARGS(__VA_ARGS__))(dev, frag, ##__VA_ARGS__)
include/linux/skbuff.h
3793
#define __skb_frag_dma_map1(dev, frag, offset, uf, uo) ({ \
include/linux/skbuff.h
3794
const skb_frag_t *uf = (frag); \
include/linux/skbuff.h
3800
#define _skb_frag_dma_map1(dev, frag, offset) \
include/linux/skbuff.h
3801
__skb_frag_dma_map1(dev, frag, offset, __UNIQUE_ID(frag_), \
include/linux/skbuff.h
3803
#define _skb_frag_dma_map0(dev, frag) \
include/linux/skbuff.h
3804
_skb_frag_dma_map1(dev, frag, 0)
include/linux/skbuff.h
3805
#define _skb_frag_dma_map2(dev, frag, offset, size) \
include/linux/skbuff.h
3806
__skb_frag_dma_map(dev, frag, offset, size, DMA_TO_DEVICE)
include/linux/skbuff.h
3807
#define _skb_frag_dma_map3(dev, frag, offset, size, dir) \
include/linux/skbuff.h
3808
__skb_frag_dma_map(dev, frag, offset, size, dir)
include/linux/skbuff.h
381
static inline void skb_frag_size_set(skb_frag_t *frag, unsigned int size)
include/linux/skbuff.h
383
frag->len = size;
include/linux/skbuff.h
391
static inline void skb_frag_size_add(skb_frag_t *frag, int delta)
include/linux/skbuff.h
393
frag->len += delta;
include/linux/skbuff.h
3959
const skb_frag_t *frag = &skb_shinfo(skb)->frags[i - 1];
include/linux/skbuff.h
3961
return netmem == skb_frag_netmem(frag) &&
include/linux/skbuff.h
3962
off == skb_frag_off(frag) + skb_frag_size(frag);
include/linux/skbuff.h
401
static inline void skb_frag_size_sub(skb_frag_t *frag, int delta)
include/linux/skbuff.h
403
frag->len -= delta;
include/linux/skbuff_ref.h
18
static __always_inline void __skb_frag_ref(skb_frag_t *frag)
include/linux/skbuff_ref.h
20
get_netmem(skb_frag_netmem(frag));
include/linux/skbuff_ref.h
54
static __always_inline void __skb_frag_unref(skb_frag_t *frag, bool recycle)
include/linux/skbuff_ref.h
56
skb_page_unref(skb_frag_netmem(frag), recycle);
include/net/bluetooth/bluetooth.h
589
struct sk_buff *skb, **frag;
include/net/bluetooth/bluetooth.h
600
frag = &skb_shinfo(skb)->frag_list;
include/net/bluetooth/bluetooth.h
611
*frag = tmp;
include/net/bluetooth/bluetooth.h
612
frag = &(*frag)->next;
include/net/ip.h
178
struct sk_buff *frag;
include/net/ip.h
190
struct sk_buff *skb = iter->frag;
include/net/ip.h
192
iter->frag = skb->next;
include/net/ipv6.h
157
struct sk_buff *frag;
include/net/ipv6.h
171
struct sk_buff *skb = iter->frag;
include/net/ipv6.h
173
iter->frag = skb->next;
include/net/libeth/xdp.h
340
skb_frag_t frag;
include/net/libeth/xdp.h
363
static_assert(offsetof(struct libeth_xdp_tx_frame, frag.len) ==
include/net/libeth/xdp.h
623
const skb_frag_t *frag)
include/net/libeth/xdp.h
625
bq->bulk[bq->count++].frag = *frag;
include/net/libeth/xdp.h
719
skb_frag_t *frag = &frm.frag;
include/net/libeth/xdp.h
725
skb_frag_fill_netmem_desc(frag, virt_to_netmem(frm.data),
include/net/libeth/xdp.h
732
netmem = skb_frag_netmem(frag);
include/net/libeth/xdp.h
735
skb_frag_off(frag),
include/net/libeth/xdp.h
736
.len = skb_frag_size(frag) & LIBETH_XDP_TX_LEN,
include/net/libeth/xdp.h
737
.flags = skb_frag_size(frag) & LIBETH_XDP_TX_FLAGS,
include/net/libeth/xdp.h
908
const skb_frag_t *frag,
include/net/libeth/xdp.h
913
dma = skb_frag_dma_map(dev, frag);
include/net/libeth/xdp.h
919
__libeth_xdp_tx_len(skb_frag_size(frag)),
include/net/libeth/xsk.h
47
struct libeth_xdp_buff *frag)
include/net/libeth/xsk.h
50
.xsk = frag,
include/net/libeth/xsk.h
51
__libeth_xdp_tx_len(frag->base.data_end - frag->data),
include/net/page_pool/types.h
181
__cacheline_group_begin_aligned(frag, PAGE_POOL_FRAG_GROUP_ALIGN);
include/net/page_pool/types.h
185
__cacheline_group_end_aligned(frag, PAGE_POOL_FRAG_GROUP_ALIGN);
include/net/xdp_sock_drv.h
164
struct xdp_buff_xsk *frag;
include/net/xdp_sock_drv.h
171
frag = container_of(xdp, struct xdp_buff_xsk, xdp);
include/net/xdp_sock_drv.h
172
list_add_tail(&frag->list_node, &frag->pool->xskb_list);
include/net/xdp_sock_drv.h
181
struct xdp_buff_xsk *frag;
include/net/xdp_sock_drv.h
183
frag = list_first_entry_or_null(&xskb->pool->xskb_list,
include/net/xdp_sock_drv.h
185
if (frag) {
include/net/xdp_sock_drv.h
186
list_del_init(&frag->list_node);
include/net/xdp_sock_drv.h
187
ret = &frag->xdp;
include/net/xdp_sock_drv.h
203
struct xdp_buff_xsk *frag;
include/net/xdp_sock_drv.h
205
frag = list_first_entry(&xskb->pool->xskb_list, struct xdp_buff_xsk,
include/net/xdp_sock_drv.h
207
return &frag->xdp;
include/net/xdp_sock_drv.h
213
struct xdp_buff_xsk *frag;
include/net/xdp_sock_drv.h
215
frag = list_last_entry(&xskb->pool->xskb_list, struct xdp_buff_xsk,
include/net/xdp_sock_drv.h
217
return &frag->xdp;
include/uapi/linux/icmp.h
102
} frag;
include/uapi/linux/wireless.h
918
struct iw_param frag; /* Fragmentation threshold */
io_uring/net.c
1409
int frag = shinfo->nr_frags;
io_uring/net.c
1415
if (!frag)
io_uring/net.c
1424
while (bi.bi_size && frag < MAX_SKB_FRAGS) {
io_uring/net.c
1429
__skb_fill_page_desc_noacc(shinfo, frag++, v.bv_page,
io_uring/net.c
1436
shinfo->nr_frags = frag;
io_uring/zcrx.c
1323
const skb_frag_t *frag, int off, int len)
io_uring/zcrx.c
1325
struct page *page = skb_frag_page(frag);
io_uring/zcrx.c
1327
return io_zcrx_copy_chunk(req, ifq, page, off + skb_frag_off(frag), len);
io_uring/zcrx.c
1331
const skb_frag_t *frag, int off, int len)
io_uring/zcrx.c
1336
if (unlikely(!skb_frag_is_net_iov(frag)))
io_uring/zcrx.c
1337
return io_zcrx_copy_frag(req, ifq, frag, off, len);
io_uring/zcrx.c
1339
niov = netmem_to_net_iov(frag->netmem);
io_uring/zcrx.c
1345
if (!io_zcrx_queue_cqe(req, niov, ifq, off + skb_frag_off(frag), len))
io_uring/zcrx.c
1404
const skb_frag_t *frag;
io_uring/zcrx.c
1409
frag = &skb_shinfo(skb)->frags[i];
io_uring/zcrx.c
1410
end = start + skb_frag_size(frag);
io_uring/zcrx.c
1418
ret = io_zcrx_recv_frag(req, ifq, frag, off, copy);
kernel/events/core.c
11293
void *record = raw->frag.data;
kernel/events/core.c
11395
.frag = {
kernel/events/core.c
8233
struct perf_raw_frag *frag = &raw->frag;
kernel/events/core.c
8237
if (frag->copy) {
kernel/events/core.c
8238
__output_custom(handle, frag->copy,
kernel/events/core.c
8239
frag->data, frag->size);
kernel/events/core.c
8241
__output_copy(handle, frag->data,
kernel/events/core.c
8242
frag->size);
kernel/events/core.c
8244
if (perf_raw_frag_last(frag))
kernel/events/core.c
8246
frag = frag->next;
kernel/events/core.c
8248
if (frag->pad)
kernel/events/core.c
8249
__output_skip(handle, NULL, frag->pad);
kernel/trace/bpf_trace.c
662
.frag = {
kernel/trace/bpf_trace.c
716
struct perf_raw_frag frag = {
kernel/trace/bpf_trace.c
722
.frag = {
kernel/trace/bpf_trace.c
724
.next = ctx_size ? &frag : NULL,
lib/kunit/string-stream.c
19
struct string_stream_fragment *frag;
lib/kunit/string-stream.c
21
frag = kzalloc_obj(*frag, gfp);
lib/kunit/string-stream.c
22
if (!frag)
lib/kunit/string-stream.c
25
frag->fragment = kmalloc(len, gfp);
lib/kunit/string-stream.c
26
if (!frag->fragment) {
lib/kunit/string-stream.c
27
kfree(frag);
lib/kunit/string-stream.c
31
return frag;
lib/kunit/string-stream.c
34
static void string_stream_fragment_destroy(struct string_stream_fragment *frag)
lib/kunit/string-stream.c
36
list_del(&frag->node);
lib/kunit/string-stream.c
37
kfree(frag->fragment);
lib/kunit/string-stream.c
38
kfree(frag);
net/appletalk/ddp.c
968
const skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
net/appletalk/ddp.c
971
end = start + skb_frag_size(frag);
net/appletalk/ddp.c
977
vaddr = kmap_atomic(skb_frag_page(frag));
net/appletalk/ddp.c
978
sum = atalk_sum_partial(vaddr + skb_frag_off(frag) +
net/bluetooth/af_bluetooth.c
478
struct sk_buff *frag;
net/bluetooth/af_bluetooth.c
483
skb_walk_frags(skb, frag) {
net/bluetooth/af_bluetooth.c
484
if (chunk <= frag->len) {
net/bluetooth/af_bluetooth.c
488
__skb_pull(frag, chunk);
net/bluetooth/af_bluetooth.c
490
} else if (frag->len) {
net/bluetooth/af_bluetooth.c
492
chunk -= frag->len;
net/bluetooth/af_bluetooth.c
493
skb->len -= frag->len;
net/bluetooth/af_bluetooth.c
494
skb->data_len -= frag->len;
net/bluetooth/af_bluetooth.c
495
__skb_pull(frag, frag->len);
net/bluetooth/iso.c
1472
struct sk_buff *skb, **frag;
net/bluetooth/iso.c
1514
frag = &skb_shinfo(skb)->frag_list;
net/bluetooth/iso.c
1524
*frag = tmp;
net/bluetooth/iso.c
1531
BT_DBG("frag %p len %d", *frag, tmp->len);
net/bluetooth/iso.c
1533
frag = &(*frag)->next;
net/bluetooth/l2cap_core.c
2215
struct sk_buff **frag;
net/bluetooth/l2cap_core.c
2225
frag = &skb_shinfo(skb)->frag_list;
net/bluetooth/l2cap_core.c
2236
*frag = tmp;
net/bluetooth/l2cap_core.c
2238
if (!copy_from_iter_full(skb_put(*frag, count), count,
net/bluetooth/l2cap_core.c
2245
skb->len += (*frag)->len;
net/bluetooth/l2cap_core.c
2246
skb->data_len += (*frag)->len;
net/bluetooth/l2cap_core.c
2248
frag = &(*frag)->next;
net/bluetooth/l2cap_core.c
2990
struct sk_buff *skb, **frag;
net/bluetooth/l2cap_core.c
3030
frag = &skb_shinfo(skb)->frag_list;
net/bluetooth/l2cap_core.c
3034
*frag = bt_skb_alloc(count, GFP_KERNEL);
net/bluetooth/l2cap_core.c
3035
if (!*frag)
net/bluetooth/l2cap_core.c
3038
skb_put_data(*frag, data, count);
net/bluetooth/l2cap_core.c
3043
frag = &(*frag)->next;
net/bluetooth/rfcomm/core.c
556
static int rfcomm_dlc_send_frag(struct rfcomm_dlc *d, struct sk_buff *frag)
net/bluetooth/rfcomm/core.c
558
int len = frag->len;
net/bluetooth/rfcomm/core.c
565
rfcomm_make_uih(frag, d->addr);
net/bluetooth/rfcomm/core.c
566
__skb_queue_tail(&d->tx_queue, frag);
net/bluetooth/rfcomm/core.c
574
struct sk_buff *frag, *next;
net/bluetooth/rfcomm/core.c
580
frag = skb_shinfo(skb)->frag_list;
net/bluetooth/rfcomm/core.c
587
if (len < 0 || !frag)
net/bluetooth/rfcomm/core.c
590
for (; frag; frag = next) {
net/bluetooth/rfcomm/core.c
593
next = frag->next;
net/bluetooth/rfcomm/core.c
595
ret = rfcomm_dlc_send_frag(d, frag);
net/bluetooth/rfcomm/core.c
597
dev_kfree_skb_irq(frag);
net/bpf/test_run.c
1362
skb_frag_t *frag;
net/bpf/test_run.c
1376
frag = &sinfo->frags[sinfo->nr_frags++];
net/bpf/test_run.c
1380
skb_frag_fill_page_desc(frag, page, 0, data_len);
net/bpf/test_run.c
471
skb_frag_t *frag = &sinfo->frags[i];
net/bpf/test_run.c
479
skb_frag_size(frag));
net/bpf/test_run.c
482
skb_frag_address(frag),
net/bridge/netfilter/nf_conntrack_bridge.c
60
struct sk_buff *frag;
net/bridge/netfilter/nf_conntrack_bridge.c
69
skb_walk_frags(skb, frag) {
net/bridge/netfilter/nf_conntrack_bridge.c
70
if (frag->len > mtu)
net/bridge/netfilter/nf_conntrack_bridge.c
73
if (skb_shared(frag) ||
net/bridge/netfilter/nf_conntrack_bridge.c
74
skb_headroom(frag) < hlen + ll_rs)
net/bridge/netfilter/nf_conntrack_bridge.c
81
if (iter.frag)
net/bridge/netfilter/nf_conntrack_bridge.c
86
if (err || !iter.frag)
net/bridge/netfilter/nf_conntrack_bridge.c
95
kfree_skb_list(iter.frag);
net/core/datagram.c
417
const skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
net/core/datagram.c
421
end = start + skb_frag_size(frag);
net/core/datagram.c
431
skb_frag_foreach_page(frag,
net/core/datagram.c
432
skb_frag_off(frag) + offset - start,
net/core/datagram.c
571
const skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
net/core/datagram.c
575
end = start + skb_frag_size(frag);
net/core/datagram.c
581
copied = copy_page_from_iter(skb_frag_page(frag),
net/core/datagram.c
582
skb_frag_off(frag) + offset - start,
net/core/datagram.c
638
int frag = skb_shinfo(skb)->nr_frags;
net/core/datagram.c
650
if (frag == MAX_SKB_FRAGS)
net/core/datagram.c
654
MAX_SKB_FRAGS - frag, &start);
net/core/datagram.c
678
if (frag) {
net/core/datagram.c
679
skb_frag_t *last = &skb_shinfo(skb)->frags[frag - 1];
net/core/datagram.c
698
skb_fill_page_desc_noacc(skb, frag++, head, start, size);
net/core/dev.c
3706
skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
net/core/dev.c
3707
struct page *page = skb_frag_page(frag);
net/core/filter.c
12406
skb_frag_t *frag = &sinfo->frags[i];
net/core/filter.c
12407
u32 shrink = min_t(u32, delta, skb_frag_size(frag));
net/core/filter.c
12409
memcpy(xdp->data_end, skb_frag_address(frag), shrink);
net/core/filter.c
12414
if (bpf_xdp_shrink_data(xdp, frag, shrink, false))
net/core/filter.c
4158
skb_frag_t *frag = &sinfo->frags[sinfo->nr_frags - 1];
net/core/filter.c
4165
tailroom = rxq->frag_size - skb_frag_size(frag) -
net/core/filter.c
4166
skb_frag_off(frag) % rxq->frag_size;
net/core/filter.c
4171
memset(skb_frag_address(frag) + skb_frag_size(frag), 0, offset);
net/core/filter.c
4172
skb_frag_size_add(frag, offset);
net/core/filter.c
4198
static bool bpf_xdp_shrink_data(struct xdp_buff *xdp, skb_frag_t *frag,
net/core/filter.c
4202
bool release = skb_frag_size(frag) == shrink;
net/core/filter.c
4203
netmem_ref netmem = skb_frag_netmem(frag);
net/core/filter.c
4215
skb_frag_off_add(frag, shrink);
net/core/filter.c
4216
skb_frag_size_sub(frag, shrink);
net/core/filter.c
4231
skb_frag_t *frag = &sinfo->frags[i];
net/core/filter.c
4232
int shrink = min_t(int, offset, skb_frag_size(frag));
net/core/filter.c
4236
if (bpf_xdp_shrink_data(xdp, frag, shrink, true))
net/core/gro.c
127
skb_frag_t *frag;
net/core/gro.c
139
frag = pinfo->frags + nr_frags;
net/core/gro.c
142
*--frag = *--frag2;
net/core/gro.c
145
skb_frag_off_add(frag, offset);
net/core/gro.c
146
skb_frag_size_sub(frag, offset);
net/core/gro.c
160
skb_frag_t *frag = pinfo->frags + nr_frags;
net/core/gro.c
174
skb_frag_fill_page_desc(frag, page, first_offset, first_size);
net/core/gro.c
176
memcpy(frag + 1, skbinfo->frags, sizeof(*frag) * skbinfo->nr_frags);
net/core/page_pool.c
186
CACHELINE_ASSERT_GROUP_MEMBER(struct page_pool, frag, frag_users);
net/core/page_pool.c
187
CACHELINE_ASSERT_GROUP_MEMBER(struct page_pool, frag, frag_page);
net/core/page_pool.c
188
CACHELINE_ASSERT_GROUP_MEMBER(struct page_pool, frag, frag_offset);
net/core/page_pool.c
189
CACHELINE_ASSERT_GROUP_SIZE(struct page_pool, frag,
net/core/skbuff.c
1380
skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
net/core/skbuff.c
1385
if (skb_frag_is_net_iov(frag)) {
net/core/skbuff.c
1387
len -= skb_frag_size(frag);
net/core/skbuff.c
1393
skb_frag_foreach_page(frag, skb_frag_off(frag),
net/core/skbuff.c
1394
skb_frag_size(frag), p, p_off, p_len,
net/core/skbuff.c
2738
struct sk_buff *frag;
net/core/skbuff.c
2773
for (fragp = &skb_shinfo(skb)->frag_list; (frag = *fragp);
net/core/skbuff.c
2774
fragp = &frag->next) {
net/core/skbuff.c
2775
int end = offset + frag->len;
net/core/skbuff.c
2777
if (skb_shared(frag)) {
net/core/skbuff.c
2780
nfrag = skb_clone(frag, GFP_ATOMIC);
net/core/skbuff.c
2784
nfrag->next = frag->next;
net/core/skbuff.c
2785
consume_skb(frag);
net/core/skbuff.c
2786
frag = nfrag;
net/core/skbuff.c
2787
*fragp = frag;
net/core/skbuff.c
2796
unlikely((err = pskb_trim(frag, len - offset))))
net/core/skbuff.c
2799
if (frag->next)
net/core/skbuff.c
2800
skb_drop_list(&frag->next);
net/core/skbuff.c
2969
skb_frag_t *frag = &skb_shinfo(skb)->frags[k];
net/core/skbuff.c
2971
*frag = skb_shinfo(skb)->frags[i];
net/core/skbuff.c
2973
skb_frag_off_add(frag, eat);
net/core/skbuff.c
2974
skb_frag_size_sub(frag, eat);
net/core/skbuff.c
3345
skb_frag_t *frag = &skb_shinfo(skb)->frags[fragidx];
net/core/skbuff.c
3347
if (offset < skb_frag_size(frag))
net/core/skbuff.c
3350
offset -= skb_frag_size(frag);
net/core/skbuff.c
3354
skb_frag_t *frag = &skb_shinfo(skb)->frags[fragidx];
net/core/skbuff.c
3356
slen = min_t(size_t, len, skb_frag_size(frag) - offset);
net/core/skbuff.c
3367
bvec_set_page(&bvec, skb_frag_page(frag), slen,
net/core/skbuff.c
3368
skb_frag_off(frag) + offset);
net/core/skbuff.c
3462
skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
net/core/skbuff.c
3467
end = start + skb_frag_size(frag);
net/core/skbuff.c
3476
skb_frag_foreach_page(frag,
net/core/skbuff.c
3477
skb_frag_off(frag) + offset - start,
net/core/skbuff.c
3543
skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
net/core/skbuff.c
3547
end = start + skb_frag_size(frag);
net/core/skbuff.c
3557
skb_frag_foreach_page(frag,
net/core/skbuff.c
3558
skb_frag_off(frag) + offset - start,
net/core/skbuff.c
3634
skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
net/core/skbuff.c
3643
skb_frag_foreach_page(frag,
net/core/skbuff.c
3644
skb_frag_off(frag) + offset - start,
net/core/skbuff.c
3711
skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
net/core/skbuff.c
3715
end = start + skb_frag_size(frag);
net/core/skbuff.c
3723
skb_frag_foreach_page(frag,
net/core/skbuff.c
3724
skb_frag_off(frag) + offset - start,
net/core/skbuff.c
4433
skb_frag_t *frag;
net/core/skbuff.c
4460
frag = &skb_shinfo(st->cur_skb)->frags[st->frag_idx];
net/core/skbuff.c
4463
pg_off = skb_frag_off(frag);
net/core/skbuff.c
4464
pg_sz = skb_frag_size(frag);
net/core/skbuff.c
4466
if (skb_frag_must_loop(skb_frag_page(frag))) {
net/core/skbuff.c
4476
st->frag_data = kmap_atomic(skb_frag_page(frag) + pg_idx);
net/core/skbuff.c
4491
if (st->frag_off == skb_frag_size(frag)) {
net/core/skbuff.c
4780
skb_frag_t *frag;
net/core/skbuff.c
4867
frag = skb_shinfo(head_skb)->frags;
net/core/skbuff.c
4896
frag = skb_shinfo(list_skb)->frags;
net/core/skbuff.c
4903
size = skb_frag_size(frag);
net/core/skbuff.c
4909
frag++;
net/core/skbuff.c
5001
frag = skb_shinfo(list_skb)->frags;
net/core/skbuff.c
5010
frag--;
net/core/skbuff.c
5025
*nskb_frag = (i < 0) ? skb_head_frag_to_page_desc(frag_skb) : *frag;
net/core/skbuff.c
5038
frag++;
net/core/skbuff.c
5243
skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
net/core/skbuff.c
5249
sg_set_page(&sg[elt], skb_frag_page(frag), copy,
net/core/skbuff.c
5250
skb_frag_off(frag) + offset - start);
net/core/skbuff.c
909
skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
net/core/skbuff.c
913
skb_frag_size_add(frag, size);
net/core/tso.c
54
skb_frag_t *frag = &skb_shinfo(skb)->frags[tso->next_frag_idx];
net/core/tso.c
57
tso->size = skb_frag_size(frag);
net/core/tso.c
58
tso->data = skb_frag_address(frag);
net/core/tso.c
80
skb_frag_t *frag = &skb_shinfo(skb)->frags[tso->next_frag_idx];
net/core/tso.c
83
tso->size = skb_frag_size(frag);
net/core/tso.c
84
tso->data = skb_frag_address(frag);
net/core/xdp.c
527
skb_frag_t *frag = &sinfo->frags[i];
net/core/xdp.c
529
bq->q[bq->count++] = skb_frag_netmem(frag);
net/core/xdp.c
700
const skb_frag_t *frag = &xinfo->frags[i];
net/core/xdp.c
701
u32 len = skb_frag_size(frag);
net/core/xdp.c
711
memcpy(page_address(page) + offset, skb_frag_address(frag),
net/ieee802154/6lowpan/reassembly.c
47
struct inet_frag_queue *frag = timer_container_of(frag, t, timer);
net/ieee802154/6lowpan/reassembly.c
51
fq = container_of(frag, struct frag_queue, q);
net/ieee802154/6lowpan/tx.c
105
struct sk_buff *frag;
net/ieee802154/6lowpan/tx.c
108
frag = alloc_skb(wdev->needed_headroom + wdev->needed_tailroom + size,
net/ieee802154/6lowpan/tx.c
111
if (likely(frag)) {
net/ieee802154/6lowpan/tx.c
112
frag->dev = wdev;
net/ieee802154/6lowpan/tx.c
113
frag->priority = skb->priority;
net/ieee802154/6lowpan/tx.c
114
skb_reserve(frag, wdev->needed_headroom);
net/ieee802154/6lowpan/tx.c
115
skb_reset_network_header(frag);
net/ieee802154/6lowpan/tx.c
116
*mac_cb(frag) = *mac_cb(skb);
net/ieee802154/6lowpan/tx.c
119
skb_put_data(frag, skb_mac_header(skb), skb->mac_len);
net/ieee802154/6lowpan/tx.c
121
rc = wpan_dev_hard_header(frag, wdev,
net/ieee802154/6lowpan/tx.c
125
kfree_skb(frag);
net/ieee802154/6lowpan/tx.c
130
frag = ERR_PTR(-ENOMEM);
net/ieee802154/6lowpan/tx.c
133
return frag;
net/ieee802154/6lowpan/tx.c
141
struct sk_buff *frag;
net/ieee802154/6lowpan/tx.c
145
frag = lowpan_alloc_frag(skb, frag_hdrlen + len, wpan_hdr, frag1);
net/ieee802154/6lowpan/tx.c
146
if (IS_ERR(frag))
net/ieee802154/6lowpan/tx.c
147
return PTR_ERR(frag);
net/ieee802154/6lowpan/tx.c
149
skb_put_data(frag, frag_hdr, frag_hdrlen);
net/ieee802154/6lowpan/tx.c
150
skb_put_data(frag, skb_network_header(skb) + offset, len);
net/ieee802154/6lowpan/tx.c
152
raw_dump_table(__func__, " fragment dump", frag->data, frag->len);
net/ieee802154/6lowpan/tx.c
154
return dev_queue_xmit(frag);
net/ipv4/esp4_offload.c
171
__be16 frag;
net/ipv4/esp4_offload.c
174
ipv6_skip_exthdr(skb, 0, &proto, &frag);
net/ipv4/icmp.c
1149
info = ntohs(icmph->un.frag.mtu);
net/ipv4/ip_fragment.c
126
struct inet_frag_queue *frag = timer_container_of(frag, t, timer);
net/ipv4/ip_fragment.c
133
qp = container_of(frag, struct ipq, q);
net/ipv4/ip_output.c
603
iter->frag = skb_shinfo(skb)->frag_list;
net/ipv4/ip_output.c
622
struct sk_buff *frag;
net/ipv4/ip_output.c
624
frag = iter->frag;
net/ipv4/ip_output.c
625
frag->ip_summed = CHECKSUM_NONE;
net/ipv4/ip_output.c
626
skb_reset_transport_header(frag);
net/ipv4/ip_output.c
627
__skb_push(frag, hlen);
net/ipv4/ip_output.c
628
skb_reset_network_header(frag);
net/ipv4/ip_output.c
629
memcpy(skb_network_header(frag), iph, hlen);
net/ipv4/ip_output.c
630
iter->iph = ip_hdr(frag);
net/ipv4/ip_output.c
632
iph->tot_len = htons(frag->len);
net/ipv4/ip_output.c
633
ip_copy_metadata(frag, skb);
net/ipv4/ip_output.c
636
if (frag->next)
net/ipv4/ip_output.c
806
struct sk_buff *frag, *frag2;
net/ipv4/ip_output.c
816
skb_walk_frags(skb, frag) {
net/ipv4/ip_output.c
818
if (frag->len > mtu ||
net/ipv4/ip_output.c
819
((frag->len & 7) && frag->next) ||
net/ipv4/ip_output.c
820
skb_headroom(frag) < hlen + ll_rs)
net/ipv4/ip_output.c
824
if (skb_shared(frag))
net/ipv4/ip_output.c
827
BUG_ON(frag->sk);
net/ipv4/ip_output.c
829
frag->sk = skb->sk;
net/ipv4/ip_output.c
830
frag->destructor = sock_wfree;
net/ipv4/ip_output.c
832
skb->truesize -= frag->truesize;
net/ipv4/ip_output.c
841
if (iter.frag) {
net/ipv4/ip_output.c
844
IPCB(iter.frag)->flags = IPCB(skb)->flags;
net/ipv4/ip_output.c
851
IPCB(iter.frag)->opt.optlen =
net/ipv4/ip_output.c
853
ip_options_fragment(iter.frag);
net/ipv4/ip_output.c
863
if (err || !iter.frag)
net/ipv4/ip_output.c
874
kfree_skb_list(iter.frag);
net/ipv4/ip_output.c
881
if (frag2 == frag)
net/ipv4/ip_tunnel_core.c
245
.un.frag.__unused = 0,
net/ipv4/ip_tunnel_core.c
246
.un.frag.mtu = htons(mtu),
net/ipv4/tcp.c
1911
skb_frag_t *frag;
net/ipv4/tcp.c
1920
frag = skb_shinfo(skb)->frags;
net/ipv4/tcp.c
1922
if (skb_frag_size(frag) > offset_skb) {
net/ipv4/tcp.c
1924
return frag;
net/ipv4/tcp.c
1926
offset_skb -= skb_frag_size(frag);
net/ipv4/tcp.c
1927
++frag;
net/ipv4/tcp.c
1930
return frag;
net/ipv4/tcp.c
1933
static bool can_map_frag(const skb_frag_t *frag)
net/ipv4/tcp.c
1937
if (skb_frag_size(frag) != PAGE_SIZE || skb_frag_off(frag))
net/ipv4/tcp.c
1940
page = skb_frag_page(frag);
net/ipv4/tcp.c
1948
static int find_next_mappable_frag(const skb_frag_t *frag,
net/ipv4/tcp.c
1953
if (likely(can_map_frag(frag)))
net/ipv4/tcp.c
1956
while (offset < remaining_in_skb && !can_map_frag(frag)) {
net/ipv4/tcp.c
1957
offset += skb_frag_size(frag);
net/ipv4/tcp.c
1958
++frag;
net/ipv4/tcp.c
1969
skb_frag_t *frag;
net/ipv4/tcp.c
1975
frag = skb_advance_to_frag(skb, offset, &frag_offset);
net/ipv4/tcp.c
1976
if (!frag)
net/ipv4/tcp.c
1983
if (frag == &info->frags[info->nr_frags - 1])
net/ipv4/tcp.c
1987
partial_frag_remainder = skb_frag_size(frag) - frag_offset;
net/ipv4/tcp.c
1989
++frag;
net/ipv4/tcp.c
1996
mappable_offset = find_next_mappable_frag(frag, zc->recv_skip_hint);
net/ipv4/tcp.c
2589
skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
net/ipv4/tcp.c
2601
if (!skb_frag_net_iov(frag)) {
net/ipv4/tcp.c
2607
niov = skb_frag_net_iov(frag);
net/ipv4/tcp.c
2613
end = start + skb_frag_size(frag);
net/ipv4/tcp.c
2620
skb_frag_off(frag) + offset -
net/ipv4/tcp.c
2644
tcp_xa_pool.netmems[tcp_xa_pool.idx++] = skb_frag_netmem(frag);
net/ipv6/esp6_offload.c
143
__be16 frag;
net/ipv6/esp6_offload.c
145
ipv6_skip_exthdr(skb, sizeof(struct ipv6hdr), &proto, &frag);
net/ipv6/esp6_offload.c
213
__be16 frag;
net/ipv6/esp6_offload.c
216
ipv6_skip_exthdr(skb, 0, &proto, &frag);
net/ipv6/ip6_output.c
1016
frag = ip6_frag_next(skb, &state);
net/ipv6/ip6_output.c
1017
if (IS_ERR(frag)) {
net/ipv6/ip6_output.c
1018
err = PTR_ERR(frag);
net/ipv6/ip6_output.c
1025
skb_set_delivery_time(frag, tstamp, tstamp_type);
net/ipv6/ip6_output.c
1026
err = output(net, sk, frag);
net/ipv6/ip6_output.c
711
iter->frag = skb_shinfo(skb)->frag_list;
net/ipv6/ip6_output.c
742
struct sk_buff *frag = iter->frag;
net/ipv6/ip6_output.c
746
frag->ip_summed = CHECKSUM_NONE;
net/ipv6/ip6_output.c
747
skb_reset_transport_header(frag);
net/ipv6/ip6_output.c
748
fh = __skb_push(frag, sizeof(struct frag_hdr));
net/ipv6/ip6_output.c
749
__skb_push(frag, hlen);
net/ipv6/ip6_output.c
750
skb_reset_network_header(frag);
net/ipv6/ip6_output.c
751
memcpy(skb_network_header(frag), iter->tmp_hdr, hlen);
net/ipv6/ip6_output.c
756
if (frag->next)
net/ipv6/ip6_output.c
759
ipv6_hdr(frag)->payload_len = htons(frag->len - sizeof(struct ipv6hdr));
net/ipv6/ip6_output.c
760
ip6_copy_metadata(frag, skb);
net/ipv6/ip6_output.c
788
struct sk_buff *frag;
net/ipv6/ip6_output.c
802
frag = alloc_skb(len + state->hlen + sizeof(struct frag_hdr) +
net/ipv6/ip6_output.c
804
if (!frag)
net/ipv6/ip6_output.c
811
ip6_copy_metadata(frag, skb);
net/ipv6/ip6_output.c
812
skb_reserve(frag, state->hroom);
net/ipv6/ip6_output.c
813
skb_put(frag, len + state->hlen + sizeof(struct frag_hdr));
net/ipv6/ip6_output.c
814
skb_reset_network_header(frag);
net/ipv6/ip6_output.c
815
fh = (struct frag_hdr *)(skb_network_header(frag) + state->hlen);
net/ipv6/ip6_output.c
816
frag->transport_header = (frag->network_header + state->hlen +
net/ipv6/ip6_output.c
824
skb_set_owner_w(frag, skb->sk);
net/ipv6/ip6_output.c
829
skb_copy_from_linear_data(skb, skb_network_header(frag), state->hlen);
net/ipv6/ip6_output.c
831
fragnexthdr_offset = skb_network_header(frag);
net/ipv6/ip6_output.c
845
BUG_ON(skb_copy_bits(skb, state->ptr, skb_transport_header(frag),
net/ipv6/ip6_output.c
852
ipv6_hdr(frag)->payload_len = htons(frag->len - sizeof(struct ipv6hdr));
net/ipv6/ip6_output.c
857
return frag;
net/ipv6/ip6_output.c
864
struct sk_buff *frag;
net/ipv6/ip6_output.c
931
skb_walk_frags(skb, frag) {
net/ipv6/ip6_output.c
933
if (frag->len > mtu ||
net/ipv6/ip6_output.c
934
((frag->len & 7) && frag->next) ||
net/ipv6/ip6_output.c
935
skb_headroom(frag) < (hlen + hroom + sizeof(struct frag_hdr)))
net/ipv6/ip6_output.c
939
if (skb_shared(frag))
net/ipv6/ip6_output.c
942
BUG_ON(frag->sk);
net/ipv6/ip6_output.c
944
frag->sk = skb->sk;
net/ipv6/ip6_output.c
945
frag->destructor = sock_wfree;
net/ipv6/ip6_output.c
947
skb->truesize -= frag->truesize;
net/ipv6/ip6_output.c
961
if (iter.frag)
net/ipv6/ip6_output.c
970
if (err || !iter.frag)
net/ipv6/ip6_output.c
985
kfree_skb_list(iter.frag);
net/ipv6/ip6_output.c
994
if (frag2 == frag)
net/ipv6/netfilter.c
196
if (iter.frag)
net/ipv6/netfilter.c
201
if (err || !iter.frag)
net/ipv6/netfilter.c
211
kfree_skb_list(iter.frag);
net/ipv6/netfilter/nf_conntrack_reasm.c
136
struct inet_frag_queue *frag = timer_container_of(frag, t, timer);
net/ipv6/netfilter/nf_conntrack_reasm.c
139
fq = container_of(frag, struct frag_queue, q);
net/ipv6/reassembly.c
76
struct inet_frag_queue *frag = timer_container_of(frag, t, timer);
net/ipv6/reassembly.c
79
fq = container_of(frag, struct frag_queue, q);
net/iucv/af_iucv.c
1067
skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
net/iucv/af_iucv.c
1069
iba[i + 1].address = virt_to_dma32(skb_frag_address(frag));
net/iucv/af_iucv.c
1070
iba[i + 1].length = (u32)skb_frag_size(frag);
net/iucv/af_iucv.c
1168
skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
net/iucv/af_iucv.c
1170
iba[i + 1].address = virt_to_dma32(skb_frag_address(frag));
net/iucv/af_iucv.c
1171
iba[i + 1].length = (u32)skb_frag_size(frag);
net/mac80211/rx.c
2268
unsigned int frag, unsigned int seq, int rx_queue,
net/mac80211/rx.c
2284
entry->last_frag = frag;
net/mac80211/rx.c
2293
unsigned int frag, unsigned int seq,
net/mac80211/rx.c
2311
entry->last_frag + 1 != frag)
net/mac80211/rx.c
2353
unsigned int frag, seq;
net/mac80211/rx.c
2365
frag = sc & IEEE80211_SCTL_FRAG;
net/mac80211/rx.c
2370
if (likely(!ieee80211_has_morefrags(fc) && frag == 0))
net/mac80211/rx.c
2389
if (frag == 0) {
net/mac80211/rx.c
2391
entry = ieee80211_reassemble_add(cache, frag, seq,
net/mac80211/rx.c
2425
entry = ieee80211_reassemble_find(cache, frag, seq,
net/mac80211/rx.c
2477
entry->last_frag = frag;
net/mptcp/protocol.h
678
int frag = (subflow->lent_mem_frag + size) & (PAGE_SIZE - 1);
net/mptcp/protocol.h
680
subflow->lent_mem_frag = frag;
net/netfilter/ipvs/ip_vs_core.c
1737
u32 mtu = ntohs(ic->un.frag.mtu);
net/netfilter/nf_log_syslog.c
454
ntohs(ich->un.frag.mtu));
net/openvswitch/conntrack.c
1053
if (key->ip.frag != OVS_FRAG_TYPE_NONE) {
net/openvswitch/conntrack.c
458
key->ip.frag = OVS_FRAG_TYPE_NONE;
net/openvswitch/datapath.c
389
later_key.ip.frag = OVS_FRAG_TYPE_LATER;
net/openvswitch/flow.c
408
key->ip.frag = OVS_FRAG_TYPE_LATER;
net/openvswitch/flow.c
412
key->ip.frag = OVS_FRAG_TYPE_FIRST;
net/openvswitch/flow.c
414
key->ip.frag = OVS_FRAG_TYPE_NONE;
net/openvswitch/flow.c
709
key->ip.frag = OVS_FRAG_TYPE_LATER;
net/openvswitch/flow.c
715
key->ip.frag = OVS_FRAG_TYPE_FIRST;
net/openvswitch/flow.c
717
key->ip.frag = OVS_FRAG_TYPE_NONE;
net/openvswitch/flow.c
836
if (key->ip.frag == OVS_FRAG_TYPE_LATER) {
net/openvswitch/flow.c
841
key->ip.frag = OVS_FRAG_TYPE_FIRST;
net/openvswitch/flow.h
105
u8 frag; /* One of OVS_FRAG_TYPE_*. */
net/openvswitch/flow_netlink.c
1542
SW_FLOW_KEY_PUT(match, ip.frag,
net/openvswitch/flow_netlink.c
1575
SW_FLOW_KEY_PUT(match, ip.frag,
net/openvswitch/flow_netlink.c
2099
ipv4_key->ipv4_frag = output->ip.frag;
net/openvswitch/flow_netlink.c
2116
ipv6_key->ipv6_frag = output->ip.frag;
net/openvswitch/flow_netlink.c
212
if (match->key->ip.frag != OVS_FRAG_TYPE_LATER) {
net/openvswitch/flow_netlink.c
2158
swkey->ip.frag != OVS_FRAG_TYPE_LATER) {
net/openvswitch/flow_netlink.c
249
if (match->key->ip.frag != OVS_FRAG_TYPE_LATER) {
net/openvswitch/flow_netlink.c
2902
if (ipv4_key->ipv4_frag != flow_key->ip.frag)
net/openvswitch/flow_netlink.c
2929
if (ipv6_key->ipv6_frag != flow_key->ip.frag)
net/rds/ib_recv.c
160
struct rds_page_frag *frag;
net/rds/ib_recv.c
179
list_for_each_entry_safe(frag, frag_tmp, &list, f_cache_entry) {
net/rds/ib_recv.c
180
list_del(&frag->f_cache_entry);
net/rds/ib_recv.c
181
WARN_ON(!list_empty(&frag->f_item));
net/rds/ib_recv.c
182
kmem_cache_free(rds_ib_frag_slab, frag);
net/rds/ib_recv.c
194
struct rds_page_frag *frag)
net/rds/ib_recv.c
196
rdsdebug("frag %p page %p\n", frag, sg_page(&frag->f_sg));
net/rds/ib_recv.c
198
rds_ib_recv_cache_put(&frag->f_cache_entry, &ic->i_cache_frags);
net/rds/ib_recv.c
207
struct rds_page_frag *frag;
net/rds/ib_recv.c
214
list_for_each_entry_safe(frag, pos, &ibinc->ii_frags, f_item) {
net/rds/ib_recv.c
215
list_del_init(&frag->f_item);
net/rds/ib_recv.c
216
rds_ib_frag_free(ic, frag);
net/rds/ib_recv.c
279
struct rds_page_frag *frag;
net/rds/ib_recv.c
285
frag = container_of(cache_item, struct rds_page_frag, f_cache_entry);
net/rds/ib_recv.c
289
frag = kmem_cache_alloc(rds_ib_frag_slab, slab_mask);
net/rds/ib_recv.c
290
if (!frag)
net/rds/ib_recv.c
293
sg_init_table(&frag->f_sg, 1);
net/rds/ib_recv.c
294
ret = rds_page_remainder_alloc(&frag->f_sg,
net/rds/ib_recv.c
297
kmem_cache_free(rds_ib_frag_slab, frag);
net/rds/ib_recv.c
303
INIT_LIST_HEAD(&frag->f_item);
net/rds/ib_recv.c
305
return frag;
net/rds/ib_recv.c
537
struct rds_page_frag *frag;
net/rds/ib_recv.c
545
frag = list_entry(ibinc->ii_frags.next, struct rds_page_frag, f_item);
net/rds/ib_recv.c
550
frag = list_entry(frag->f_item.next,
net/rds/ib_recv.c
560
ret = copy_page_to_iter(sg_page(&frag->f_sg),
net/rds/ib_recv.c
561
frag->f_sg.offset + frag_off,
net/rds/ib_recv.c
790
struct rds_page_frag *frag;
net/rds/ib_recv.c
805
frag = list_entry(ibinc->ii_frags.next, struct rds_page_frag, f_item);
net/rds/ib_recv.c
817
addr = kmap_atomic(sg_page(&frag->f_sg));
net/rds/ib_recv.c
819
src = addr + frag->f_sg.offset + frag_off;
net/rds/ib_recv.c
839
frag = list_entry(frag->f_item.next,
net/rxrpc/output.c
645
enum rxrpc_tx_point frag;
net/rxrpc/output.c
683
frag = rxrpc_tx_point_call_data_frag;
net/rxrpc/output.c
686
frag = rxrpc_tx_point_call_data_nofrag;
net/rxrpc/output.c
720
trace_rxrpc_tx_packet(call->debug_id, whdr, frag);
net/rxrpc/output.c
724
trace_rxrpc_tx_fail(call->debug_id, txb->serial, ret, frag);
net/rxrpc/output.c
726
trace_rxrpc_tx_packet(call->debug_id, whdr, frag);
net/sched/act_ct.c
805
static int tcf_ct_ipv4_is_fragment(struct sk_buff *skb, bool *frag)
net/sched/act_ct.c
815
*frag = ip_is_fragment(ip_hdr(skb));
net/sched/act_ct.c
819
static int tcf_ct_ipv6_is_fragment(struct sk_buff *skb, bool *frag)
net/sched/act_ct.c
835
*frag = flags & IP6_FH_F_FRAG;
net/sched/act_ct.c
845
bool frag;
net/sched/act_ct.c
855
err = tcf_ct_ipv4_is_fragment(skb, &frag);
net/sched/act_ct.c
857
err = tcf_ct_ipv6_is_fragment(skb, &frag);
net/sched/act_ct.c
858
if (err || !frag)
net/sctp/associola.c
1382
int frag = sctp_mtu_payload(sctp_sk(asoc->base.sk), asoc->pathmtu,
net/sctp/associola.c
1386
frag = min_t(int, frag, asoc->user_frag);
net/sctp/associola.c
1388
frag = min_t(int, frag, SCTP_MAX_CHUNK_LEN -
net/sctp/associola.c
1391
asoc->frag_point = SCTP_TRUNC4(frag);
net/sctp/chunk.c
238
u8 frag = SCTP_DATA_MIDDLE_FRAG;
net/sctp/chunk.c
242
frag |= SCTP_DATA_FIRST_FRAG;
net/sctp/chunk.c
252
frag |= SCTP_DATA_LAST_FRAG;
net/sctp/chunk.c
260
frag |= SCTP_DATA_SACK_IMM;
net/sctp/chunk.c
263
chunk = asoc->stream.si->make_datafrag(asoc, sinfo, len, frag,
net/sctp/input.c
661
info = ntohs(hdr->un.frag.mtu);
net/sctp/socket.c
9484
struct sk_buff *frag;
net/sctp/socket.c
9490
skb_walk_frags(skb, frag)
net/sctp/socket.c
9491
sctp_skb_set_owner_r_frag(frag, sk);
net/sctp/ulpevent.c
1081
struct sk_buff *skb, *frag;
net/sctp/ulpevent.c
1097
skb_walk_frags(skb, frag)
net/sctp/ulpevent.c
1098
sctp_ulpevent_receive_data(sctp_skb2event(frag), asoc);
net/sctp/ulpevent.c
1106
struct sk_buff *skb, *frag;
net/sctp/ulpevent.c
1123
skb_walk_frags(skb, frag) {
net/sctp/ulpevent.c
1128
sctp_ulpevent_release_frag_data(sctp_skb2event(frag));
net/sctp/ulpevent.c
1139
struct sk_buff *skb, *frag;
net/sctp/ulpevent.c
1147
skb_walk_frags(skb, frag) {
net/sctp/ulpevent.c
1152
sctp_ulpevent_release_frag_data(sctp_skb2event(frag));
net/tipc/msg.c
127
struct sk_buff *frag = *buf;
net/tipc/msg.c
134
if (!frag)
net/tipc/msg.c
137
msg = buf_msg(frag);
net/tipc/msg.c
139
frag->next = NULL;
net/tipc/msg.c
140
skb_pull(frag, msg_hdr_sz(msg));
net/tipc/msg.c
145
if (skb_has_frag_list(frag) && __skb_linearize(frag))
net/tipc/msg.c
148
frag = skb_unshare(frag, GFP_ATOMIC);
net/tipc/msg.c
149
if (unlikely(!frag))
net/tipc/msg.c
151
head = *headbuf = frag;
net/tipc/msg.c
164
if (skb_try_coalesce(head, frag, &headstolen, &delta)) {
net/tipc/msg.c
165
kfree_skb_partial(frag, headstolen);
net/tipc/msg.c
169
skb_shinfo(head)->frag_list = frag;
net/tipc/msg.c
171
tail->next = frag;
net/tipc/msg.c
172
head->truesize += frag->truesize;
net/tipc/msg.c
173
head->data_len += frag->len;
net/tipc/msg.c
174
head->len += frag->len;
net/tipc/msg.c
175
TIPC_SKB_CB(head)->tail = frag;
net/tipc/msg.c
767
struct sk_buff *frag = NULL;
net/tipc/msg.c
784
frag = skb_clone(skb, GFP_ATOMIC);
net/tipc/msg.c
785
if (!frag)
net/tipc/msg.c
787
frag->next = NULL;
net/tipc/msg.c
788
if (tipc_buf_append(&head, &frag))
net/tipc/msg.c
793
__skb_queue_tail(rcvq, frag);
net/tls/tls_device.c
261
skb_frag_t *frag;
net/tls/tls_device.c
263
frag = &record->frags[record->num_frags - 1];
net/tls/tls_device.c
264
if (skb_frag_page(frag) == pfrag->page &&
net/tls/tls_device.c
265
skb_frag_off(frag) + skb_frag_size(frag) == pfrag->offset) {
net/tls/tls_device.c
266
skb_frag_size_add(frag, size);
net/tls/tls_device.c
268
++frag;
net/tls/tls_device.c
269
skb_frag_fill_page_desc(frag, pfrag->page, pfrag->offset,
net/tls/tls_device.c
287
skb_frag_t *frag;
net/tls/tls_device.c
300
frag = &record->frags[i];
net/tls/tls_device.c
302
sg_set_page(&offload_ctx->sg_tx_data[i], skb_frag_page(frag),
net/tls/tls_device.c
303
skb_frag_size(frag), skb_frag_off(frag));
net/tls/tls_device.c
304
sk_mem_charge(sk, skb_frag_size(frag));
net/tls/tls_device.c
305
get_page(skb_frag_page(frag));
net/tls/tls_device.c
347
skb_frag_t *frag;
net/tls/tls_device.c
353
frag = &record->frags[0];
net/tls/tls_device.c
354
skb_frag_fill_page_desc(frag, pfrag->page, pfrag->offset,
net/tls/tls_device_fallback.c
260
skb_frag_t *frag = &record->frags[i];
net/tls/tls_device_fallback.c
262
__skb_frag_ref(frag);
net/tls/tls_device_fallback.c
263
sg_set_page(sg_in + i, skb_frag_page(frag),
net/tls/tls_device_fallback.c
264
skb_frag_size(frag), skb_frag_off(frag));
net/tls/tls_device_fallback.c
266
remaining -= skb_frag_size(frag);
net/tls/tls_strp.c
216
skb_frag_t *frag;
net/tls/tls_strp.c
224
frag = &skb_shinfo(skb)->frags[nfrag];
net/tls/tls_strp.c
230
chunk = min_t(size_t, len, PAGE_SIZE - skb_frag_size(frag));
net/tls/tls_strp.c
232
skb_frag_address(frag) +
net/tls/tls_strp.c
233
skb_frag_size(frag),
net/tls/tls_strp.c
238
skb_frag_size_add(frag, chunk);
net/tls/tls_strp.c
251
skb_frag_size_add(frag, -over);
net/tls/tls_strp.c
256
frag++;
net/tls/tls_strp.c
268
chunk = min_t(size_t, chunk, PAGE_SIZE - skb_frag_size(frag));
net/tls/tls_strp.c
270
skb_frag_address(frag) +
net/tls/tls_strp.c
271
skb_frag_size(frag),
net/tls/tls_strp.c
276
skb_frag_size_add(frag, chunk);
net/tls/tls_strp.c
277
frag++;
net/tls/tls_strp.c
54
skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
net/tls/tls_strp.c
57
skb_frag_address(frag),
net/tls/tls_strp.c
58
skb_frag_size(frag)));
net/tls/tls_strp.c
59
offset += skb_frag_size(frag);
net/wireless/util.c
670
const skb_frag_t *frag = &sh->frags[0];
net/wireless/util.c
683
frag_page = skb_frag_page(frag);
net/wireless/util.c
684
frag_ptr = skb_frag_address(frag);
net/wireless/util.c
685
frag_size = skb_frag_size(frag);
net/wireless/util.c
686
frag++;
net/wireless/util.c
698
frag_len = skb_frag_size(frag);
net/wireless/util.c
700
__frame_add_frag(frame, skb_frag_page(frag),
net/wireless/util.c
701
skb_frag_address(frag), cur_len, frag_len);
net/wireless/util.c
703
frag++;
net/wireless/wext-compat.c
290
struct iw_param *frag = &wrqu->frag;
net/wireless/wext-compat.c
298
if (frag->disabled || !frag->fixed) {
net/wireless/wext-compat.c
300
} else if (frag->value < 256) {
net/wireless/wext-compat.c
304
wdev->wiphy->frag_threshold = frag->value & ~0x1;
net/wireless/wext-compat.c
317
struct iw_param *frag = &wrqu->frag;
net/wireless/wext-compat.c
320
frag->value = wdev->wiphy->frag_threshold;
net/wireless/wext-compat.c
321
frag->disabled = frag->value == (u32) -1;
net/wireless/wext-compat.c
322
frag->fixed = 1;
net/xdp/xsk.c
213
u32 *from_len, skb_frag_t **frag, u32 rem)
net/xdp/xsk.c
226
*from = skb_frag_address(*frag);
net/xdp/xsk.c
227
*from_len = skb_frag_size((*frag)++);
net/xdp/xsk.c
247
skb_frag_t *frag;
net/xdp/xsk.c
287
frag = &sinfo->frags[0];
net/xdp/xsk.c
297
copied = xsk_copy_xdp(copy_to, ©_from, to_len, &from_len, &frag, rem);
net/xfrm/xfrm_ipcomp.c
71
skb_frag_t *frag;
net/xfrm/xfrm_ipcomp.c
74
frag = skb_shinfo(skb)->frags + skb_shinfo(skb)->nr_frags;
net/xfrm/xfrm_ipcomp.c
82
skb_frag_fill_page_desc(frag, page, 0, len);
net/xfrm/xfrm_iptfs.c
297
static void iptfs_skb_head_to_frag(const struct sk_buff *skb, skb_frag_t *frag)
net/xfrm/xfrm_iptfs.c
302
skb_frag_fill_page_desc(frag, page, skb->data - addr, skb_headlen(skb));
net/xfrm/xfrm_iptfs.c
338
skb_frag_t *frag, *from;
net/xfrm/xfrm_iptfs.c
352
frag = &walk->frags[walk->nr_frags++];
net/xfrm/xfrm_iptfs.c
353
iptfs_skb_head_to_frag(skb, frag);
net/xfrm/xfrm_iptfs.c
354
frag->offset += initial_offset;
net/xfrm/xfrm_iptfs.c
355
frag->len -= initial_offset;
net/xfrm/xfrm_iptfs.c
356
walk->total += frag->len;
net/xfrm/xfrm_iptfs.c
369
frag = &walk->frags[walk->nr_frags++];
net/xfrm/xfrm_iptfs.c
370
*frag = *from;
net/xfrm/xfrm_iptfs.c
372
frag->offset += initial_offset;
net/xfrm/xfrm_iptfs.c
373
frag->len -= initial_offset;
net/xfrm/xfrm_iptfs.c
376
walk->total += frag->len;
net/xfrm/xfrm_iptfs.c
426
skb_frag_t *frag = &walk->frags[fragi];
net/xfrm/xfrm_iptfs.c
428
fraglen = frag->len;
net/xfrm/xfrm_iptfs.c
472
skb_frag_t *frag = &walk->frags[walk->fragi];
net/xfrm/xfrm_iptfs.c
475
*tofrag = *frag;
net/xfrm/xfrm_iptfs.c
496
walk->past += frag->len; /* careful, use src bv_len */
samples/bpf/xdp_adjust_tail_kern.c
99
icmp_hdr->un.frag.mtu = htons(max_pcktsz - sizeof(struct ethhdr));
sound/pci/cs4281.c
1744
cdma->frag++;
sound/pci/cs4281.c
1745
if ((val & BA0_HDSR_DHTC) && !(cdma->frag & 1)) {
sound/pci/cs4281.c
1746
cdma->frag--;
sound/pci/cs4281.c
1750
if ((val & BA0_HDSR_DTC) && (cdma->frag & 1)) {
sound/pci/cs4281.c
1751
cdma->frag--;
sound/pci/cs4281.c
431
int frag; /* period number */
sound/pci/cs4281.c
733
dma->frag = 0; /* for workaround */
sound/pci/rme9652/rme9652.c
343
unsigned int offset, frag;
sound/pci/rme9652/rme9652.c
364
frag = status & RME9652_buffer_id;
sound/pci/rme9652/rme9652.c
368
if (frag)
sound/pci/rme9652/rme9652.c
372
} else if (!frag)
sound/pci/rme9652/rme9652.c
379
if (!frag)
sound/pci/rme9652/rme9652.c
383
} else if (frag)
tools/testing/selftests/drivers/net/gro.c
923
struct ip6_frag *frag = (void *)(extpkt + tcp_offset);
tools/testing/selftests/drivers/net/gro.c
939
frag->ip6f_nxt = IPPROTO_TCP;
tools/testing/selftests/net/tcp_ao/icmps-accept.c
205
randomize_buffer(&packet.icmph.un.frag.mtu,
tools/testing/selftests/net/tcp_ao/icmps-accept.c
206
sizeof(packet.icmph.un.frag.mtu));
tools/testing/selftests/net/tcp_ao/icmps-discard.c
205
randomize_buffer(&packet.icmph.un.frag.mtu,
tools/testing/selftests/net/tcp_ao/icmps-discard.c
206
sizeof(packet.icmph.un.frag.mtu));