Symbol: copy
arch/arc/kernel/smp.c
353
unsigned long __maybe_unused copy;
arch/arc/kernel/smp.c
365
copy = pending = xchg(this_cpu_ptr(&ipi_data), 0);
arch/arc/kernel/smp.c
373
pr_info("IPI with bogus msg %ld in %ld\n", msg, copy);
arch/loongarch/kernel/inst.c
245
struct insn_copy *copy = data;
arch/loongarch/kernel/inst.c
247
if (smp_processor_id() == copy->cpu) {
arch/loongarch/kernel/inst.c
248
ret = copy_to_kernel_nofault(copy->dst, copy->src, copy->len);
arch/loongarch/kernel/inst.c
255
flush_icache_range((unsigned long)copy->dst, (unsigned long)copy->dst + copy->len);
arch/loongarch/kernel/inst.c
265
struct insn_copy copy = {
arch/loongarch/kernel/inst.c
287
ret = stop_machine_cpuslocked(text_copy_cb, &copy, cpu_online_mask);
arch/m68k/fpsp040/fpsp.h
72
| of these registers, it should modify the saved copy and let
arch/mips/boot/elf2ecoff.c
598
copy(outfile, infile, ph[i].p_offset,
arch/powerpc/kvm/book3s_64_mmu_radix.c
1360
goto copy;
arch/powerpc/kvm/book3s_64_mmu_radix.c
1406
copy:
arch/s390/kernel/debug.c
867
debug_info_t *copy;
arch/s390/kernel/debug.c
875
copy = debug_info_alloc("", pages_per_area, nr_areas, id->buf_size,
arch/s390/kernel/debug.c
877
if (!copy) {
arch/s390/kernel/debug.c
892
debug_events_append(copy, id);
arch/s390/kernel/debug.c
893
debug_areas_swap(id, copy);
arch/s390/kernel/debug.c
897
copy->areas = NULL;
arch/s390/kernel/debug.c
898
copy->active_pages = NULL;
arch/s390/kernel/debug.c
899
copy->active_entries = NULL;
arch/s390/kernel/debug.c
900
debug_info_free(copy);
arch/s390/lib/csum-partial.c
19
static __always_inline __wsum csum_copy(void *dst, const void *src, int len, __wsum sum, bool copy)
arch/s390/lib/csum-partial.c
24
if (copy)
arch/s390/lib/csum-partial.c
35
if (copy) {
arch/s390/lib/csum-partial.c
48
if (copy) {
arch/s390/lib/csum-partial.c
59
if (copy) {
arch/s390/lib/csum-partial.c
69
if (copy)
arch/x86/boot/video.c
279
int copy = (xs < saved.x) ? xs : saved.x;
arch/x86/boot/video.c
280
copy_to_fs(dst, src, copy*sizeof(u16));
arch/x86/boot/video.c
281
dst += copy*sizeof(u16);
arch/x86/tools/insn_decoder_test.c
121
char copy[BUFSIZE], *s, *tab1, *tab2;
arch/x86/tools/insn_decoder_test.c
133
strcpy(copy, line);
arch/x86/tools/insn_decoder_test.c
134
tab1 = strchr(copy, '\t');
arch/x86/tools/vdso2c.h
39
BITSFUNC(copy)(outfile, data + offset, len);
arch/x86/tools/vdso2c.h
8
static void BITSFUNC(copy)(FILE *outfile, const unsigned char *data, size_t len)
block/bio-integrity.c
312
bool copy, is_p2p = false;
block/bio-integrity.c
331
copy = iov_iter_alignment(iter) &
block/bio-integrity.c
347
copy = true;
block/bio-integrity.c
351
if (copy)
block/blk-map.c
503
bool copy = false, map_bvec = false;
block/blk-map.c
510
copy = true;
block/blk-map.c
512
copy = true;
block/blk-map.c
516
copy = true;
block/blk-map.c
518
copy = queue_virt_boundary(q) & iov_iter_gap_alignment(iter);
block/blk-map.c
527
copy = true;
block/blk-map.c
532
if (copy)
drivers/base/test/property-entry-test.c
353
struct property_entry *copy;
drivers/base/test/property-entry-test.c
356
copy = property_entries_dup(entries);
drivers/base/test/property-entry-test.c
357
KUNIT_ASSERT_NOT_ERR_OR_NULL(test, copy);
drivers/base/test/property-entry-test.c
359
KUNIT_EXPECT_TRUE(test, copy[0].is_inline);
drivers/base/test/property-entry-test.c
360
data_ptr = (u8 *)&copy[0].value;
drivers/base/test/property-entry-test.c
364
KUNIT_EXPECT_FALSE(test, copy[1].is_inline);
drivers/base/test/property-entry-test.c
365
data_ptr = copy[1].pointer;
drivers/base/test/property-entry-test.c
369
property_entries_free(copy);
drivers/base/test/property-entry-test.c
385
struct property_entry *copy;
drivers/base/test/property-entry-test.c
388
copy = property_entries_dup(entries);
drivers/base/test/property-entry-test.c
389
KUNIT_ASSERT_NOT_ERR_OR_NULL(test, copy);
drivers/base/test/property-entry-test.c
391
KUNIT_EXPECT_TRUE(test, copy[0].is_inline);
drivers/base/test/property-entry-test.c
392
KUNIT_EXPECT_STREQ(test, copy[0].value.str[0], "a");
drivers/base/test/property-entry-test.c
394
KUNIT_EXPECT_FALSE(test, copy[1].is_inline);
drivers/base/test/property-entry-test.c
395
data_ptr = copy[1].pointer;
drivers/base/test/property-entry-test.c
399
KUNIT_EXPECT_TRUE(test, copy[2].is_inline);
drivers/base/test/property-entry-test.c
400
KUNIT_EXPECT_STREQ(test, copy[2].value.str[0], "");
drivers/base/test/property-entry-test.c
402
property_entries_free(copy);
drivers/clk/samsung/clk-exynos3250.c
757
#define E3250_CPU_DIV1(hpm, copy) \
drivers/clk/samsung/clk-exynos3250.c
758
(((hpm) << 4) | ((copy) << 0))
drivers/clk/samsung/clk-exynos4.c
1200
#define E4210_CPU_DIV1(hpm, copy) \
drivers/clk/samsung/clk-exynos4.c
1201
(((hpm) << 4) | ((copy) << 0))
drivers/clk/samsung/clk-exynos4.c
1231
#define E4412_CPU_DIV1(cores, hpm, copy) \
drivers/clk/samsung/clk-exynos4.c
1232
(((cores) << 8) | ((hpm) << 4) | ((copy) << 0))
drivers/clk/samsung/clk-exynos5250.c
756
#define E5250_CPU_DIV1(hpm, copy) \
drivers/clk/samsung/clk-exynos5250.c
757
(((hpm) << 4) | (copy))
drivers/clk/samsung/clk-exynos5433.c
3682
#define E5433_APOLLO_DIV1(hpm, copy) \
drivers/clk/samsung/clk-exynos5433.c
3683
(((hpm) << 4) | ((copy) << 0))
drivers/clk/samsung/clk-exynos5433.c
3921
#define E5433_ATLAS_DIV1(hpm, copy) \
drivers/clk/samsung/clk-exynos5433.c
3922
(((hpm) << 4) | ((copy) << 0))
drivers/comedi/comedi_fops.c
1902
struct comedi_cmd *cmd, bool *copy, void *file)
drivers/comedi/comedi_fops.c
1957
*copy = true;
drivers/comedi/comedi_fops.c
2013
struct comedi_cmd *cmd, bool *copy, void *file)
drivers/comedi/comedi_fops.c
2045
*copy = true;
drivers/comedi/comedi_fops.c
2370
bool copy = false;
drivers/comedi/comedi_fops.c
2376
rc = do_cmd_ioctl(dev, &cmd, &copy, file);
drivers/comedi/comedi_fops.c
2377
if (copy && copy_to_user((void __user *)arg, &cmd, sizeof(cmd)))
drivers/comedi/comedi_fops.c
2383
bool copy = false;
drivers/comedi/comedi_fops.c
2389
rc = do_cmdtest_ioctl(dev, &cmd, &copy, file);
drivers/comedi/comedi_fops.c
2390
if (copy && copy_to_user((void __user *)arg, &cmd, sizeof(cmd)))
drivers/comedi/comedi_fops.c
3239
bool copy = false;
drivers/comedi/comedi_fops.c
3251
rc = do_cmd_ioctl(dev, &cmd, &copy, file);
drivers/comedi/comedi_fops.c
3254
if (copy) {
drivers/comedi/comedi_fops.c
3269
bool copy = false;
drivers/comedi/comedi_fops.c
3281
rc = do_cmdtest_ioctl(dev, &cmd, &copy, file);
drivers/comedi/comedi_fops.c
3284
if (copy) {
drivers/comedi/drivers/jr3_pci.c
647
char copy[ARRAY_SIZE(sensor0->copyright) + 1];
drivers/comedi/drivers/jr3_pci.c
651
copy[i] = (char)(get_u16(&sensor0->copyright[i]) >> 8);
drivers/comedi/drivers/jr3_pci.c
652
copy[i] = '\0';
drivers/comedi/drivers/jr3_pci.c
653
dev_dbg(dev->class_dev, "Firmware copyright: %s\n", copy);
drivers/crypto/hifn_795x.c
1395
unsigned int copy, drest = *drestp, nbytes = *nbytesp;
drivers/crypto/hifn_795x.c
1402
copy = min3(drest, size, dst->length);
drivers/crypto/hifn_795x.c
1404
size -= copy;
drivers/crypto/hifn_795x.c
1405
drest -= copy;
drivers/crypto/hifn_795x.c
1406
nbytes -= copy;
drivers/crypto/hifn_795x.c
1409
__func__, copy, size, drest, nbytes);
drivers/crypto/hifn_795x.c
1425
unsigned int nbytes = req->cryptlen, offset, copy, diff;
drivers/crypto/hifn_795x.c
1453
copy = slen & ~(HIFN_D_DST_DALIGN - 1);
drivers/crypto/hifn_795x.c
1482
copy += diff + nbytes;
drivers/crypto/hifn_795x.c
1493
t->length = copy;
drivers/crypto/hifn_795x.c
1601
unsigned int srest = *srestp, nbytes = *nbytesp, copy;
drivers/crypto/hifn_795x.c
1609
copy = min3(srest, dst->length, size);
drivers/crypto/hifn_795x.c
1612
memcpy(daddr + dst->offset + offset, saddr, copy);
drivers/crypto/hifn_795x.c
1615
nbytes -= copy;
drivers/crypto/hifn_795x.c
1616
size -= copy;
drivers/crypto/hifn_795x.c
1617
srest -= copy;
drivers/crypto/hifn_795x.c
1618
saddr += copy;
drivers/crypto/hifn_795x.c
1622
__func__, copy, size, srest, nbytes);
drivers/crypto/inside-secure/eip93/eip93-common.c
151
const u32 len, const bool copy)
drivers/crypto/inside-secure/eip93/eip93-common.c
171
if (copy)
drivers/dma-buf/dma-buf.c
898
struct dma_buf_sg_table_wrapper *copy;
drivers/dma-buf/dma-buf.c
903
copy = container_of(*sg_table, typeof(*copy), wrapper);
drivers/dma-buf/dma-buf.c
904
*sg_table = copy->original;
drivers/dma-buf/dma-buf.c
905
sg_free_table(&copy->wrapper);
drivers/dma-buf/dma-buf.c
906
kfree(copy);
drivers/dma/altera-msgdma.c
338
size_t copy;
drivers/dma/altera-msgdma.c
357
copy = min_t(size_t, len, MSGDMA_MAX_TRANS_LEN);
drivers/dma/altera-msgdma.c
359
msgdma_desc_config(desc, dma_dst, dma_src, copy,
drivers/dma/altera-msgdma.c
361
len -= copy;
drivers/dma/altera-msgdma.c
362
dma_src += copy;
drivers/dma/altera-msgdma.c
363
dma_dst += copy;
drivers/dma/fsldma.c
766
size_t copy;
drivers/dma/fsldma.c
785
copy = min(len, (size_t)FSL_DMA_BCR_MAX_CNT);
drivers/dma/fsldma.c
787
set_desc_cnt(chan, &new->hw, copy);
drivers/dma/fsldma.c
800
len -= copy;
drivers/dma/fsldma.c
801
dma_src += copy;
drivers/dma/fsldma.c
802
dma_dst += copy;
drivers/dma/ioat/prep.c
123
size_t copy = min_t(size_t, len, 1 << ioat_chan->xfercap_log);
drivers/dma/ioat/prep.c
128
hw->size = copy;
drivers/dma/ioat/prep.c
133
len -= copy;
drivers/dma/ioat/prep.c
134
dst += copy;
drivers/dma/ioat/prep.c
135
src += copy;
drivers/dma/k3dma.c
494
size_t copy = 0;
drivers/dma/k3dma.c
518
copy = min_t(size_t, len, DMA_MAX_SIZE);
drivers/dma/k3dma.c
519
k3_dma_fill_desc(ds, dst, src, copy, num++, c->ccfg);
drivers/dma/k3dma.c
521
src += copy;
drivers/dma/k3dma.c
522
dst += copy;
drivers/dma/k3dma.c
523
len -= copy;
drivers/dma/mmp_pdma.c
609
size_t copy = 0;
drivers/dma/mmp_pdma.c
632
copy = min_t(size_t, len, PDMA_MAX_DESC_BYTES);
drivers/dma/mmp_pdma.c
636
new->desc.dcmd = chan->dcmd | (DCMD_LENGTH & copy);
drivers/dma/mmp_pdma.c
650
len -= copy;
drivers/dma/mmp_pdma.c
653
dma_src += copy;
drivers/dma/mmp_pdma.c
655
dma_dst += copy;
drivers/dma/mmp_pdma.c
657
dma_src += copy;
drivers/dma/mmp_pdma.c
658
dma_dst += copy;
drivers/dma/pxa_dma.c
923
size_t copy;
drivers/dma/pxa_dma.c
947
copy = min_t(size_t, len, PDMA_MAX_DESC_BYTES);
drivers/dma/pxa_dma.c
948
hw_desc->dcmd = dcmd | (PXA_DCMD_LENGTH & copy);
drivers/dma/pxa_dma.c
951
len -= copy;
drivers/dma/pxa_dma.c
952
dma_src += copy;
drivers/dma/pxa_dma.c
953
dma_dst += copy;
drivers/dma/xilinx/xilinx_dma.c
1256
size_t copy;
drivers/dma/xilinx/xilinx_dma.c
1258
copy = min_t(size_t, size - done,
drivers/dma/xilinx/xilinx_dma.c
1261
if ((copy + done < size) &&
drivers/dma/xilinx/xilinx_dma.c
1267
copy = rounddown(copy,
drivers/dma/xilinx/xilinx_dma.c
1270
return copy;
drivers/dma/xilinx/xilinx_dma.c
2229
size_t copy;
drivers/dma/xilinx/xilinx_dma.c
2260
copy = xilinx_dma_calc_copysize(chan, vecs[i].len,
drivers/dma/xilinx/xilinx_dma.c
2266
hw->control = copy;
drivers/dma/xilinx/xilinx_dma.c
2272
sg_used += copy;
drivers/dma/xilinx/xilinx_dma.c
2325
size_t copy;
drivers/dma/xilinx/xilinx_dma.c
2357
copy = xilinx_dma_calc_copysize(chan, sg_dma_len(sg),
drivers/dma/xilinx/xilinx_dma.c
2365
hw->control = copy;
drivers/dma/xilinx/xilinx_dma.c
2373
sg_used += copy;
drivers/dma/xilinx/xilinx_dma.c
2425
size_t copy, sg_used;
drivers/dma/xilinx/xilinx_dma.c
2465
copy = xilinx_dma_calc_copysize(chan, period_len,
drivers/dma/xilinx/xilinx_dma.c
2470
hw->control = copy;
drivers/dma/xilinx/xilinx_dma.c
2476
sg_used += copy;
drivers/dma/xilinx/xilinx_dma.c
2535
size_t copy;
drivers/dma/xilinx/xilinx_dma.c
2567
copy = min_t(size_t, sg_dma_len(sg) - sg_used,
drivers/dma/xilinx/xilinx_dma.c
2574
hw->control = copy;
drivers/dma/xilinx/xilinx_dma.c
2581
sg_used += copy;
drivers/dma/xilinx/zynqmp_dma.c
838
size_t copy;
drivers/dma/xilinx/zynqmp_dma.c
859
copy = min_t(size_t, len, ZYNQMP_DMA_MAX_TRANS_LEN);
drivers/dma/xilinx/zynqmp_dma.c
862
dma_dst, copy, prev);
drivers/dma/xilinx/zynqmp_dma.c
864
len -= copy;
drivers/dma/xilinx/zynqmp_dma.c
865
dma_src += copy;
drivers/dma/xilinx/zynqmp_dma.c
866
dma_dst += copy;
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_mst_types.c
66
uint8_t copy[16];
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_mst_types.c
83
memcpy(copy, msg->buffer, msg->size);
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_mst_types.c
84
payload.data = copy;
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
1152
struct atmel_hlcdc_plane_state *copy;
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
1154
copy = kmemdup(state, sizeof(*state), GFP_KERNEL);
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
1155
if (!copy)
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
1158
if (atmel_hlcdc_plane_alloc_dscrs(p, copy)) {
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
1159
kfree(copy);
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
1163
__drm_atomic_helper_plane_duplicate_state(p, &copy->base);
drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_plane.c
1165
return &copy->base;
drivers/gpu/drm/drm_print.c
100
if (copy > iterator->remain)
drivers/gpu/drm/drm_print.c
101
copy = iterator->remain;
drivers/gpu/drm/drm_print.c
106
str + (iterator->start - iterator->offset), copy);
drivers/gpu/drm/drm_print.c
108
iterator->offset = iterator->start + copy;
drivers/gpu/drm/drm_print.c
109
iterator->remain -= copy;
drivers/gpu/drm/drm_print.c
89
ssize_t copy;
drivers/gpu/drm/drm_print.c
98
copy = len - (iterator->start - iterator->offset);
drivers/gpu/drm/exynos/exynos_drm_plane.c
148
struct exynos_drm_plane_state *copy;
drivers/gpu/drm/exynos/exynos_drm_plane.c
151
copy = kzalloc_obj(*exynos_state);
drivers/gpu/drm/exynos/exynos_drm_plane.c
152
if (!copy)
drivers/gpu/drm/exynos/exynos_drm_plane.c
155
__drm_atomic_helper_plane_duplicate_state(plane, &copy->base);
drivers/gpu/drm/exynos/exynos_drm_plane.c
156
return &copy->base;
drivers/gpu/drm/i915/gt/selftest_migrate.c
441
return copy(migrate, __migrate_copy, sz, prng);
drivers/gpu/drm/i915/gt/selftest_migrate.c
447
return copy(migrate, __global_copy, sz, prng);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
1213
void (*copy)(void *dst, const void *src, size_t size);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
1259
tests[i].copy(dst_addr, src_addr, size);
drivers/gpu/drm/nouveau/nouveau_bo.c
1048
&drm->ttm.copy);
drivers/gpu/drm/nouveau/nouveau_bo.c
1050
ret = mthd->init(chan, drm->ttm.copy.handle);
drivers/gpu/drm/nouveau/nouveau_bo.c
1052
nvif_object_dtor(&drm->ttm.copy);
drivers/gpu/drm/nouveau/nouveau_dmem.c
686
switch (drm->ttm.copy.oclass) {
drivers/gpu/drm/nouveau/nouveau_drm.c
331
nvif_object_dtor(&drm->ttm.copy);
drivers/gpu/drm/nouveau/nouveau_drv.h
242
struct nvif_object copy;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
1117
struct nvkm_ram_data *copy;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
1133
copy = &ram->base.former;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
1136
copy = &ram->base.target;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
1139
xits->bios.ramcfg_11_02_04 = copy->bios.ramcfg_11_02_04;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
1140
xits->bios.ramcfg_11_02_03 = copy->bios.ramcfg_11_02_03;
drivers/gpu/drm/nouveau/nvkm/subdev/fb/ramgk104.c
1141
xits->bios.timing_20_30_07 = copy->bios.timing_20_30_07;
drivers/gpu/drm/qxl/qxl_dev.h
689
struct qxl_copy copy;
drivers/gpu/drm/qxl/qxl_dev.h
718
struct qxl_copy copy;
drivers/gpu/drm/qxl/qxl_draw.c
229
drawable->u.copy.src_area.top = 0;
drivers/gpu/drm/qxl/qxl_draw.c
230
drawable->u.copy.src_area.bottom = height;
drivers/gpu/drm/qxl/qxl_draw.c
231
drawable->u.copy.src_area.left = 0;
drivers/gpu/drm/qxl/qxl_draw.c
232
drawable->u.copy.src_area.right = width;
drivers/gpu/drm/qxl/qxl_draw.c
234
drawable->u.copy.rop_descriptor = SPICE_ROPD_OP_PUT;
drivers/gpu/drm/qxl/qxl_draw.c
235
drawable->u.copy.scale_mode = 0;
drivers/gpu/drm/qxl/qxl_draw.c
236
drawable->u.copy.mask.flags = 0;
drivers/gpu/drm/qxl/qxl_draw.c
237
drawable->u.copy.mask.pos.x = 0;
drivers/gpu/drm/qxl/qxl_draw.c
238
drawable->u.copy.mask.pos.y = 0;
drivers/gpu/drm/qxl/qxl_draw.c
239
drawable->u.copy.mask.bitmap = 0;
drivers/gpu/drm/qxl/qxl_draw.c
241
drawable->u.copy.src_bitmap = qxl_bo_physical_address(qdev, dimage->bo, 0);
drivers/gpu/drm/radeon/cik.c
3652
int ring_index = rdev->asic->copy.blit_ring_index;
drivers/gpu/drm/radeon/cik.c
3868
if (rdev->asic->copy.copy_ring_index == RADEON_RING_TYPE_GFX_INDEX)
drivers/gpu/drm/radeon/cik.c
4107
if (rdev->asic->copy.copy_ring_index == RADEON_RING_TYPE_GFX_INDEX)
drivers/gpu/drm/radeon/cik_sdma.c
254
if ((rdev->asic->copy.copy_ring_index == R600_RING_TYPE_DMA_INDEX) ||
drivers/gpu/drm/radeon/cik_sdma.c
255
(rdev->asic->copy.copy_ring_index == CAYMAN_RING_TYPE_DMA1_INDEX))
drivers/gpu/drm/radeon/cik_sdma.c
432
if ((rdev->asic->copy.copy_ring_index == R600_RING_TYPE_DMA_INDEX) ||
drivers/gpu/drm/radeon/cik_sdma.c
433
(rdev->asic->copy.copy_ring_index == CAYMAN_RING_TYPE_DMA1_INDEX))
drivers/gpu/drm/radeon/cik_sdma.c
585
int ring_index = rdev->asic->copy.dma_ring_index;
drivers/gpu/drm/radeon/evergreen_dma.c
114
int ring_index = rdev->asic->copy.dma_ring_index;
drivers/gpu/drm/radeon/ni.c
1440
if (rdev->asic->copy.copy_ring_index == RADEON_RING_TYPE_GFX_INDEX)
drivers/gpu/drm/radeon/ni.c
1714
if (rdev->asic->copy.copy_ring_index == RADEON_RING_TYPE_GFX_INDEX)
drivers/gpu/drm/radeon/ni_dma.c
160
if ((rdev->asic->copy.copy_ring_index == R600_RING_TYPE_DMA_INDEX) ||
drivers/gpu/drm/radeon/ni_dma.c
161
(rdev->asic->copy.copy_ring_index == CAYMAN_RING_TYPE_DMA1_INDEX))
drivers/gpu/drm/radeon/ni_dma.c
256
if ((rdev->asic->copy.copy_ring_index == R600_RING_TYPE_DMA_INDEX) ||
drivers/gpu/drm/radeon/ni_dma.c
257
(rdev->asic->copy.copy_ring_index == CAYMAN_RING_TYPE_DMA1_INDEX))
drivers/gpu/drm/radeon/r600.c
2425
if (rdev->asic->copy.copy_ring_index == RADEON_RING_TYPE_GFX_INDEX)
drivers/gpu/drm/radeon/r600.c
2774
if (rdev->asic->copy.copy_ring_index == RADEON_RING_TYPE_GFX_INDEX)
drivers/gpu/drm/radeon/r600.c
2971
int ring_index = rdev->asic->copy.blit_ring_index;
drivers/gpu/drm/radeon/r600_dma.c
102
if (rdev->asic->copy.copy_ring_index == R600_RING_TYPE_DMA_INDEX)
drivers/gpu/drm/radeon/r600_dma.c
179
if (rdev->asic->copy.copy_ring_index == R600_RING_TYPE_DMA_INDEX)
drivers/gpu/drm/radeon/r600_dma.c
450
int ring_index = rdev->asic->copy.dma_ring_index;
drivers/gpu/drm/radeon/radeon.h
1915
struct radeon_fence *(*copy)(struct radeon_device *rdev,
drivers/gpu/drm/radeon/radeon.h
1922
} copy;
drivers/gpu/drm/radeon/radeon.h
2727
#define radeon_copy_blit(rdev, s, d, np, resv) (rdev)->asic->copy.blit((rdev), (s), (d), (np), (resv))
drivers/gpu/drm/radeon/radeon.h
2728
#define radeon_copy_dma(rdev, s, d, np, resv) (rdev)->asic->copy.dma((rdev), (s), (d), (np), (resv))
drivers/gpu/drm/radeon/radeon.h
2729
#define radeon_copy(rdev, s, d, np, resv) (rdev)->asic->copy.copy((rdev), (s), (d), (np), (resv))
drivers/gpu/drm/radeon/radeon.h
2730
#define radeon_copy_blit_ring_index(rdev) (rdev)->asic->copy.blit_ring_index
drivers/gpu/drm/radeon/radeon.h
2731
#define radeon_copy_dma_ring_index(rdev) (rdev)->asic->copy.dma_ring_index
drivers/gpu/drm/radeon/radeon.h
2732
#define radeon_copy_ring_index(rdev) (rdev)->asic->copy.copy_ring_index
drivers/gpu/drm/radeon/radeon_asic.c
1050
.copy = {
drivers/gpu/drm/radeon/radeon_asic.c
1055
.copy = &r600_copy_cpdma,
drivers/gpu/drm/radeon/radeon_asic.c
1143
.copy = {
drivers/gpu/drm/radeon/radeon_asic.c
1148
.copy = &r600_copy_cpdma,
drivers/gpu/drm/radeon/radeon_asic.c
1249
.copy = {
drivers/gpu/drm/radeon/radeon_asic.c
1254
.copy = &rv770_copy_dma,
drivers/gpu/drm/radeon/radeon_asic.c
1369
.copy = {
drivers/gpu/drm/radeon/radeon_asic.c
1374
.copy = &evergreen_copy_dma,
drivers/gpu/drm/radeon/radeon_asic.c
1463
.copy = {
drivers/gpu/drm/radeon/radeon_asic.c
1468
.copy = &evergreen_copy_dma,
drivers/gpu/drm/radeon/radeon_asic.c
1557
.copy = {
drivers/gpu/drm/radeon/radeon_asic.c
1562
.copy = &evergreen_copy_dma,
drivers/gpu/drm/radeon/radeon_asic.c
1705
.copy = {
drivers/gpu/drm/radeon/radeon_asic.c
1710
.copy = &evergreen_copy_dma,
drivers/gpu/drm/radeon/radeon_asic.c
1825
.copy = {
drivers/gpu/drm/radeon/radeon_asic.c
1830
.copy = &evergreen_copy_dma,
drivers/gpu/drm/radeon/radeon_asic.c
1963
.copy = {
drivers/gpu/drm/radeon/radeon_asic.c
1968
.copy = &si_copy_dma,
drivers/gpu/drm/radeon/radeon_asic.c
2133
.copy = {
drivers/gpu/drm/radeon/radeon_asic.c
2138
.copy = &cik_copy_dma,
drivers/gpu/drm/radeon/radeon_asic.c
2246
.copy = {
drivers/gpu/drm/radeon/radeon_asic.c
2251
.copy = &cik_copy_dma,
drivers/gpu/drm/radeon/radeon_asic.c
227
.copy = {
drivers/gpu/drm/radeon/radeon_asic.c
232
.copy = &r100_copy_blit,
drivers/gpu/drm/radeon/radeon_asic.c
295
.copy = {
drivers/gpu/drm/radeon/radeon_asic.c
300
.copy = &r100_copy_blit,
drivers/gpu/drm/radeon/radeon_asic.c
391
.copy = {
drivers/gpu/drm/radeon/radeon_asic.c
396
.copy = &r100_copy_blit,
drivers/gpu/drm/radeon/radeon_asic.c
459
.copy = {
drivers/gpu/drm/radeon/radeon_asic.c
464
.copy = &r100_copy_blit,
drivers/gpu/drm/radeon/radeon_asic.c
527
.copy = {
drivers/gpu/drm/radeon/radeon_asic.c
532
.copy = &r100_copy_blit,
drivers/gpu/drm/radeon/radeon_asic.c
595
.copy = {
drivers/gpu/drm/radeon/radeon_asic.c
600
.copy = &r100_copy_blit,
drivers/gpu/drm/radeon/radeon_asic.c
663
.copy = {
drivers/gpu/drm/radeon/radeon_asic.c
668
.copy = &r100_copy_blit,
drivers/gpu/drm/radeon/radeon_asic.c
731
.copy = {
drivers/gpu/drm/radeon/radeon_asic.c
736
.copy = &r200_copy_dma,
drivers/gpu/drm/radeon/radeon_asic.c
799
.copy = {
drivers/gpu/drm/radeon/radeon_asic.c
804
.copy = &r100_copy_blit,
drivers/gpu/drm/radeon/radeon_asic.c
867
.copy = {
drivers/gpu/drm/radeon/radeon_asic.c
872
.copy = &r100_copy_blit,
drivers/gpu/drm/radeon/radeon_asic.c
964
.copy = {
drivers/gpu/drm/radeon/radeon_asic.c
969
.copy = &r600_copy_cpdma,
drivers/gpu/drm/radeon/radeon_benchmark.c
122
if (rdev->asic->copy.dma) {
drivers/gpu/drm/radeon/radeon_benchmark.c
133
if (rdev->asic->copy.blit) {
drivers/gpu/drm/radeon/radeon_test.c
256
if (rdev->asic->copy.dma)
drivers/gpu/drm/radeon/radeon_test.c
258
if (rdev->asic->copy.blit)
drivers/gpu/drm/radeon/radeon_ttm.c
226
rdev->asic->copy.copy != NULL) {
drivers/gpu/drm/radeon/rv770.c
1082
if (rdev->asic->copy.copy_ring_index == RADEON_RING_TYPE_GFX_INDEX)
drivers/gpu/drm/radeon/rv770_dma.c
49
int ring_index = rdev->asic->copy.dma_ring_index;
drivers/gpu/drm/radeon/si.c
3445
if (rdev->asic->copy.copy_ring_index == RADEON_RING_TYPE_GFX_INDEX)
drivers/gpu/drm/radeon/si.c
3748
if (rdev->asic->copy.copy_ring_index == RADEON_RING_TYPE_GFX_INDEX)
drivers/gpu/drm/radeon/si_dma.c
237
int ring_index = rdev->asic->copy.dma_ring_index;
drivers/gpu/drm/renesas/rcar-du/rcar_du_crtc.c
969
struct rcar_du_crtc_state *copy;
drivers/gpu/drm/renesas/rcar-du/rcar_du_crtc.c
975
copy = kmemdup(state, sizeof(*state), GFP_KERNEL);
drivers/gpu/drm/renesas/rcar-du/rcar_du_crtc.c
976
if (copy == NULL)
drivers/gpu/drm/renesas/rcar-du/rcar_du_crtc.c
979
__drm_atomic_helper_crtc_duplicate_state(crtc, &copy->state);
drivers/gpu/drm/renesas/rcar-du/rcar_du_crtc.c
981
return &copy->state;
drivers/gpu/drm/renesas/rcar-du/rcar_du_plane.c
693
struct rcar_du_plane_state *copy;
drivers/gpu/drm/renesas/rcar-du/rcar_du_plane.c
699
copy = kmemdup(state, sizeof(*state), GFP_KERNEL);
drivers/gpu/drm/renesas/rcar-du/rcar_du_plane.c
700
if (copy == NULL)
drivers/gpu/drm/renesas/rcar-du/rcar_du_plane.c
703
__drm_atomic_helper_plane_duplicate_state(plane, &copy->state);
drivers/gpu/drm/renesas/rcar-du/rcar_du_plane.c
705
return &copy->state;
drivers/gpu/drm/renesas/rcar-du/rcar_du_vsp.c
402
struct rcar_du_vsp_plane_state *copy;
drivers/gpu/drm/renesas/rcar-du/rcar_du_vsp.c
407
copy = kzalloc_obj(*copy);
drivers/gpu/drm/renesas/rcar-du/rcar_du_vsp.c
408
if (copy == NULL)
drivers/gpu/drm/renesas/rcar-du/rcar_du_vsp.c
411
__drm_atomic_helper_plane_duplicate_state(plane, &copy->state);
drivers/gpu/drm/renesas/rcar-du/rcar_du_vsp.c
413
return &copy->state;
drivers/gpu/drm/renesas/rcar-du/rcar_du_writeback.c
102
copy = kzalloc_obj(*copy);
drivers/gpu/drm/renesas/rcar-du/rcar_du_writeback.c
103
if (!copy)
drivers/gpu/drm/renesas/rcar-du/rcar_du_writeback.c
106
__drm_atomic_helper_connector_duplicate_state(connector, &copy->state);
drivers/gpu/drm/renesas/rcar-du/rcar_du_writeback.c
108
return &copy->state;
drivers/gpu/drm/renesas/rcar-du/rcar_du_writeback.c
97
struct rcar_du_wb_conn_state *copy;
drivers/gpu/drm/renesas/rz-du/rzg2l_du_crtc.c
308
struct rzg2l_du_crtc_state *copy;
drivers/gpu/drm/renesas/rz-du/rzg2l_du_crtc.c
314
copy = kmemdup(state, sizeof(*state), GFP_KERNEL);
drivers/gpu/drm/renesas/rz-du/rzg2l_du_crtc.c
315
if (!copy)
drivers/gpu/drm/renesas/rz-du/rzg2l_du_crtc.c
318
__drm_atomic_helper_crtc_duplicate_state(crtc, &copy->state);
drivers/gpu/drm/renesas/rz-du/rzg2l_du_crtc.c
320
return &copy->state;
drivers/gpu/drm/renesas/rz-du/rzg2l_du_vsp.c
247
struct rzg2l_du_vsp_plane_state *copy;
drivers/gpu/drm/renesas/rz-du/rzg2l_du_vsp.c
252
copy = kzalloc_obj(*copy);
drivers/gpu/drm/renesas/rz-du/rzg2l_du_vsp.c
253
if (!copy)
drivers/gpu/drm/renesas/rz-du/rzg2l_du_vsp.c
256
__drm_atomic_helper_plane_duplicate_state(plane, &copy->state);
drivers/gpu/drm/renesas/rz-du/rzg2l_du_vsp.c
258
return &copy->state;
drivers/gpu/drm/renesas/shmobile/shmob_drm_plane.c
232
struct shmob_drm_plane_state *copy;
drivers/gpu/drm/renesas/shmobile/shmob_drm_plane.c
238
copy = kmemdup(state, sizeof(*state), GFP_KERNEL);
drivers/gpu/drm/renesas/shmobile/shmob_drm_plane.c
239
if (copy == NULL)
drivers/gpu/drm/renesas/shmobile/shmob_drm_plane.c
242
__drm_atomic_helper_plane_duplicate_state(plane, &copy->base);
drivers/gpu/drm/renesas/shmobile/shmob_drm_plane.c
244
return &copy->base;
drivers/gpu/drm/sun4i/sun4i_layer.c
41
struct sun4i_layer_state *copy;
drivers/gpu/drm/sun4i/sun4i_layer.c
43
copy = kzalloc_obj(*copy);
drivers/gpu/drm/sun4i/sun4i_layer.c
44
if (!copy)
drivers/gpu/drm/sun4i/sun4i_layer.c
47
__drm_atomic_helper_plane_duplicate_state(plane, &copy->state);
drivers/gpu/drm/sun4i/sun4i_layer.c
48
copy->uses_frontend = orig->uses_frontend;
drivers/gpu/drm/sun4i/sun4i_layer.c
50
return &copy->state;
drivers/gpu/drm/tegra/dc.c
1407
struct tegra_dc_state *copy;
drivers/gpu/drm/tegra/dc.c
1409
copy = kmalloc_obj(*copy);
drivers/gpu/drm/tegra/dc.c
1410
if (!copy)
drivers/gpu/drm/tegra/dc.c
1413
__drm_atomic_helper_crtc_duplicate_state(crtc, &copy->base);
drivers/gpu/drm/tegra/dc.c
1414
copy->clk = state->clk;
drivers/gpu/drm/tegra/dc.c
1415
copy->pclk = state->pclk;
drivers/gpu/drm/tegra/dc.c
1416
copy->div = state->div;
drivers/gpu/drm/tegra/dc.c
1417
copy->planes = state->planes;
drivers/gpu/drm/tegra/dc.c
1419
return &copy->base;
drivers/gpu/drm/tegra/dsi.c
792
struct tegra_dsi_state *copy;
drivers/gpu/drm/tegra/dsi.c
794
copy = kmemdup(state, sizeof(*state), GFP_KERNEL);
drivers/gpu/drm/tegra/dsi.c
795
if (!copy)
drivers/gpu/drm/tegra/dsi.c
799
&copy->base);
drivers/gpu/drm/tegra/dsi.c
801
return &copy->base;
drivers/gpu/drm/tegra/plane.c
55
struct tegra_plane_state *copy;
drivers/gpu/drm/tegra/plane.c
58
copy = kmalloc_obj(*copy);
drivers/gpu/drm/tegra/plane.c
59
if (!copy)
drivers/gpu/drm/tegra/plane.c
62
__drm_atomic_helper_plane_duplicate_state(plane, &copy->base);
drivers/gpu/drm/tegra/plane.c
63
copy->tiling = state->tiling;
drivers/gpu/drm/tegra/plane.c
64
copy->format = state->format;
drivers/gpu/drm/tegra/plane.c
65
copy->swap = state->swap;
drivers/gpu/drm/tegra/plane.c
66
copy->reflect_x = state->reflect_x;
drivers/gpu/drm/tegra/plane.c
67
copy->reflect_y = state->reflect_y;
drivers/gpu/drm/tegra/plane.c
68
copy->opaque = state->opaque;
drivers/gpu/drm/tegra/plane.c
69
copy->total_peak_memory_bandwidth = state->total_peak_memory_bandwidth;
drivers/gpu/drm/tegra/plane.c
70
copy->peak_memory_bandwidth = state->peak_memory_bandwidth;
drivers/gpu/drm/tegra/plane.c
71
copy->avg_memory_bandwidth = state->avg_memory_bandwidth;
drivers/gpu/drm/tegra/plane.c
74
copy->blending[i] = state->blending[i];
drivers/gpu/drm/tegra/plane.c
77
copy->iova[i] = DMA_MAPPING_ERROR;
drivers/gpu/drm/tegra/plane.c
78
copy->map[i] = NULL;
drivers/gpu/drm/tegra/plane.c
81
return &copy->base;
drivers/gpu/drm/tegra/sor.c
1752
struct tegra_sor_state *copy;
drivers/gpu/drm/tegra/sor.c
1754
copy = kmemdup(state, sizeof(*state), GFP_KERNEL);
drivers/gpu/drm/tegra/sor.c
1755
if (!copy)
drivers/gpu/drm/tegra/sor.c
1758
__drm_atomic_helper_connector_duplicate_state(connector, &copy->base);
drivers/gpu/drm/tegra/sor.c
1760
return &copy->base;
drivers/gpu/drm/v3d/v3d_drv.h
467
struct v3d_copy_query_results_info copy;
drivers/gpu/drm/v3d/v3d_sched.c
541
struct v3d_copy_query_results_info *copy = &job->copy;
drivers/gpu/drm/v3d/v3d_sched.c
551
data = ((u8 *)bo->vaddr) + copy->offset;
drivers/gpu/drm/v3d/v3d_sched.c
557
write_result = available || copy->do_partial;
drivers/gpu/drm/v3d/v3d_sched.c
560
write_to_buffer(data, 0, copy->do_64bit, *((u64 *)query_addr));
drivers/gpu/drm/v3d/v3d_sched.c
563
if (copy->availability_bit)
drivers/gpu/drm/v3d/v3d_sched.c
564
write_to_buffer(data, 1, copy->do_64bit, available ? 1u : 0u);
drivers/gpu/drm/v3d/v3d_sched.c
566
data += copy->stride;
drivers/gpu/drm/v3d/v3d_sched.c
629
if (job->copy.do_64bit) {
drivers/gpu/drm/v3d/v3d_sched.c
647
struct v3d_copy_query_results_info *copy = &job->copy;
drivers/gpu/drm/v3d/v3d_sched.c
655
data = ((u8 *)bo->vaddr) + copy->offset;
drivers/gpu/drm/v3d/v3d_sched.c
661
write_result = available || copy->do_partial;
drivers/gpu/drm/v3d/v3d_sched.c
665
if (copy->availability_bit)
drivers/gpu/drm/v3d/v3d_sched.c
667
copy->do_64bit, available ? 1u : 0u);
drivers/gpu/drm/v3d/v3d_sched.c
669
data += copy->stride;
drivers/gpu/drm/v3d/v3d_submit.c
585
struct drm_v3d_copy_timestamp_query copy;
drivers/gpu/drm/v3d/v3d_submit.c
593
if (copy_from_user(&copy, ext, sizeof(copy)))
drivers/gpu/drm/v3d/v3d_submit.c
596
if (copy.pad)
drivers/gpu/drm/v3d/v3d_submit.c
602
copy.count);
drivers/gpu/drm/v3d/v3d_submit.c
606
offsets = u64_to_user_ptr(copy.offsets);
drivers/gpu/drm/v3d/v3d_submit.c
607
syncs = u64_to_user_ptr(copy.syncs);
drivers/gpu/drm/v3d/v3d_submit.c
609
for (i = 0; i < copy.count; i++) {
drivers/gpu/drm/v3d/v3d_submit.c
631
query_info->count = copy.count;
drivers/gpu/drm/v3d/v3d_submit.c
633
job->copy.do_64bit = copy.do_64bit;
drivers/gpu/drm/v3d/v3d_submit.c
634
job->copy.do_partial = copy.do_partial;
drivers/gpu/drm/v3d/v3d_submit.c
635
job->copy.availability_bit = copy.availability_bit;
drivers/gpu/drm/v3d/v3d_submit.c
636
job->copy.offset = copy.offset;
drivers/gpu/drm/v3d/v3d_submit.c
637
job->copy.stride = copy.stride;
drivers/gpu/drm/v3d/v3d_submit.c
752
struct drm_v3d_copy_performance_query copy;
drivers/gpu/drm/v3d/v3d_submit.c
758
if (copy_from_user(&copy, ext, sizeof(copy)))
drivers/gpu/drm/v3d/v3d_submit.c
761
if (copy.pad)
drivers/gpu/drm/v3d/v3d_submit.c
767
kvmalloc_objs(struct v3d_performance_query, copy.count);
drivers/gpu/drm/v3d/v3d_submit.c
772
copy.count,
drivers/gpu/drm/v3d/v3d_submit.c
773
copy.nperfmons,
drivers/gpu/drm/v3d/v3d_submit.c
774
u64_to_user_ptr(copy.syncs),
drivers/gpu/drm/v3d/v3d_submit.c
775
u64_to_user_ptr(copy.kperfmon_ids),
drivers/gpu/drm/v3d/v3d_submit.c
780
query_info->count = copy.count;
drivers/gpu/drm/v3d/v3d_submit.c
781
query_info->nperfmons = copy.nperfmons;
drivers/gpu/drm/v3d/v3d_submit.c
782
query_info->ncounters = copy.ncounters;
drivers/gpu/drm/v3d/v3d_submit.c
784
job->copy.do_64bit = copy.do_64bit;
drivers/gpu/drm/v3d/v3d_submit.c
785
job->copy.do_partial = copy.do_partial;
drivers/gpu/drm/v3d/v3d_submit.c
786
job->copy.availability_bit = copy.availability_bit;
drivers/gpu/drm/v3d/v3d_submit.c
787
job->copy.offset = copy.offset;
drivers/gpu/drm/v3d/v3d_submit.c
788
job->copy.stride = copy.stride;
drivers/gpu/drm/xe/xe_guc_log.c
131
snapshot->copy = kzalloc_objs(*snapshot->copy, snapshot->num_chunks,
drivers/gpu/drm/xe/xe_guc_log.c
133
if (!snapshot->copy)
drivers/gpu/drm/xe/xe_guc_log.c
140
snapshot->copy[i] = kmalloc(size, atomic ? GFP_ATOMIC : GFP_KERNEL);
drivers/gpu/drm/xe/xe_guc_log.c
141
if (!snapshot->copy[i])
drivers/gpu/drm/xe/xe_guc_log.c
150
kfree(snapshot->copy[i]);
drivers/gpu/drm/xe/xe_guc_log.c
151
kfree(snapshot->copy);
drivers/gpu/drm/xe/xe_guc_log.c
171
if (snapshot->copy) {
drivers/gpu/drm/xe/xe_guc_log.c
173
kfree(snapshot->copy[i]);
drivers/gpu/drm/xe/xe_guc_log.c
174
kfree(snapshot->copy);
drivers/gpu/drm/xe/xe_guc_log.c
208
xe_map_memcpy_from(xe, snapshot->copy[i], &log->bo->vmap,
drivers/gpu/drm/xe/xe_guc_log.c
258
xe_print_blob_ascii85(p, prefix, suffix, snapshot->copy[i], 0, size);
drivers/gpu/drm/xe/xe_guc_log.c
445
lfd_output_binary(p, snapshot->copy[chunk_id] + chunk_from, size);
drivers/gpu/drm/xe/xe_guc_log.c
522
buf32 = snapshot->copy[chunk_id] + chunk_from;
drivers/gpu/drm/xe/xe_guc_log.c
566
xe_guc_log_load_lic(snapshot->copy[0], &config);
drivers/gpu/drm/xe/xe_guc_log_types.h
24
void **copy;
drivers/infiniband/hw/hfi1/fault.c
101
copy = min(len, datalen - 1);
drivers/infiniband/hw/hfi1/fault.c
102
if (copy_from_user(data, buf, copy)) {
drivers/infiniband/hw/hfi1/fault.c
94
size_t copy, datalen = 1280;
drivers/input/misc/yealink.c
114
} master, copy;
drivers/input/misc/yealink.c
337
if (val != yld->copy.b[ix])
drivers/input/misc/yealink.c
351
yld->copy.b[ix] = val;
drivers/input/misc/yealink.c
389
yld->copy.b[ix] = val;
drivers/input/misc/yealink.c
512
yld->copy.b[i] = ~yld->master.b[i];
drivers/md/bcache/bset.c
920
goto copy;
drivers/md/bcache/bset.c
925
goto copy;
drivers/md/bcache/bset.c
928
copy: bkey_copy(m, k);
drivers/md/dm-cache-target.c
1359
copy(mg, is_policy_promote);
drivers/md/dm-cache-target.c
2588
const char **copy;
drivers/md/dm-cache-target.c
2590
copy = kcalloc(argc, sizeof(*copy), GFP_KERNEL);
drivers/md/dm-cache-target.c
2591
if (!copy)
drivers/md/dm-cache-target.c
2594
copy[i] = kstrdup(argv[i], GFP_KERNEL);
drivers/md/dm-cache-target.c
2595
if (!copy[i]) {
drivers/md/dm-cache-target.c
2597
kfree(copy[i]);
drivers/md/dm-cache-target.c
2598
kfree(copy);
drivers/md/dm-cache-target.c
2604
cache->ctr_args = copy;
drivers/md/dm-clone-target.c
1728
const char **copy;
drivers/md/dm-clone-target.c
1730
copy = kcalloc(argc, sizeof(*copy), GFP_KERNEL);
drivers/md/dm-clone-target.c
1731
if (!copy)
drivers/md/dm-clone-target.c
1735
copy[i] = kstrdup(argv[i], GFP_KERNEL);
drivers/md/dm-clone-target.c
1737
if (!copy[i]) {
drivers/md/dm-clone-target.c
1739
kfree(copy[i]);
drivers/md/dm-clone-target.c
1740
kfree(copy);
drivers/md/dm-clone-target.c
1746
clone->ctr_args = copy;
drivers/md/dm-thin-metadata.c
1336
struct dm_block *copy, *sblock;
drivers/md/dm-thin-metadata.c
1355
&sb_validator, &copy, &inc);
drivers/md/dm-thin-metadata.c
1361
held_root = dm_block_location(copy);
drivers/md/dm-thin-metadata.c
1362
disk_super = dm_block_data(copy);
drivers/md/dm-thin-metadata.c
1368
dm_tm_unlock(pmd->tm, copy);
drivers/md/dm-thin-metadata.c
1385
dm_tm_unlock(pmd->tm, copy);
drivers/md/dm-thin-metadata.c
1418
struct dm_block *sblock, *copy;
drivers/md/dm-thin-metadata.c
1436
r = dm_tm_read_lock(pmd->tm, held_root, &sb_validator, &copy);
drivers/md/dm-thin-metadata.c
1440
disk_super = dm_block_data(copy);
drivers/md/dm-thin-metadata.c
1445
dm_tm_unlock(pmd->tm, copy);
drivers/media/pci/bt8xx/bttv-driver.c
1698
struct v4l2_tuner copy = *t;
drivers/media/pci/bt8xx/bttv-driver.c
1700
btv->audio_mode_gpio(btv, &copy, 1);
drivers/media/pci/solo6x10/solo6x10-g723.c
240
.copy = snd_solo_pcm_copy,
drivers/media/platform/samsung/s3c-camif/camif-core.c
175
scaler->copy = 0;
drivers/media/platform/samsung/s3c-camif/camif-core.h
168
u8 copy;
drivers/media/platform/samsung/s3c-camif/camif-regs.c
453
if (scaler->copy)
drivers/media/test-drivers/vicodec/codec-fwht.c
842
s16 copy[8 * 8];
drivers/media/test-drivers/vicodec/codec-fwht.c
876
memcpy(cf->de_fwht, copy, sizeof(copy));
drivers/media/test-drivers/vicodec/codec-fwht.c
899
memcpy(copy, cf->de_fwht, sizeof(copy));
drivers/media/test-drivers/vicodec/vicodec-core.c
358
u32 copy;
drivers/media/test-drivers/vicodec/vicodec-core.c
367
copy = sizeof(magic) - ctx->comp_magic_cnt;
drivers/media/test-drivers/vicodec/vicodec-core.c
368
if (*pp + sz - p < copy)
drivers/media/test-drivers/vicodec/vicodec-core.c
369
copy = *pp + sz - p;
drivers/media/test-drivers/vicodec/vicodec-core.c
371
memcpy(header + ctx->comp_magic_cnt, p, copy);
drivers/media/test-drivers/vicodec/vicodec-core.c
372
ctx->comp_magic_cnt += copy;
drivers/media/test-drivers/vicodec/vicodec-core.c
374
p += copy;
drivers/media/test-drivers/vicodec/vicodec-core.c
388
u32 copy = sizeof(struct fwht_cframe_hdr) - ctx->header_size;
drivers/media/test-drivers/vicodec/vicodec-core.c
390
if (*pp + sz - p < copy)
drivers/media/test-drivers/vicodec/vicodec-core.c
391
copy = *pp + sz - p;
drivers/media/test-drivers/vicodec/vicodec-core.c
393
memcpy(header + ctx->header_size, p, copy);
drivers/media/test-drivers/vicodec/vicodec-core.c
394
p += copy;
drivers/media/test-drivers/vicodec/vicodec-core.c
395
ctx->header_size += copy;
drivers/media/test-drivers/vicodec/vicodec-core.c
623
u32 copy = max_to_copy - ctx->comp_size;
drivers/media/test-drivers/vicodec/vicodec-core.c
625
if (copy > p_src + sz - p)
drivers/media/test-drivers/vicodec/vicodec-core.c
626
copy = p_src + sz - p;
drivers/media/test-drivers/vicodec/vicodec-core.c
629
p, copy);
drivers/media/test-drivers/vicodec/vicodec-core.c
630
p += copy;
drivers/media/test-drivers/vicodec/vicodec-core.c
631
ctx->comp_size += copy;
drivers/misc/mei/bus.c
177
goto copy;
drivers/misc/mei/bus.c
232
copy:
drivers/mtd/ubi/vtbl.c
286
int copy, void *vtbl)
drivers/mtd/ubi/vtbl.c
293
dbg_gen("create volume table (copy #%d)", copy + 1);
drivers/mtd/ubi/vtbl.c
313
vid_hdr->lnum = cpu_to_be32(copy);
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_io.c
1034
int copy = 0;
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_io.c
1038
copy = mss - skb->len;
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_io.c
1064
copy <= 0) {
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_io.c
1087
copy = mss;
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_io.c
1089
if (copy > size)
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_io.c
1090
copy = size;
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_io.c
1093
err = skb_splice_from_iter(skb, &msg->msg_iter, copy);
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_io.c
1099
copy = err;
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_io.c
1100
sk_wmem_queued_add(sk, copy);
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_io.c
1102
copy = min(copy, skb_tailroom(skb));
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_io.c
1104
copy = min_t(int, copy, csk->tlshws.txleft);
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_io.c
1106
&msg->msg_iter, copy);
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_io.c
1121
goto copy;
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_io.c
1154
copy:
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_io.c
1155
if (copy > pg_size - off)
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_io.c
1156
copy = pg_size - off;
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_io.c
1158
copy = min_t(int, copy, csk->tlshws.txleft);
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_io.c
1162
off, copy);
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_io.c
1174
copy);
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_io.c
1176
skb_fill_page_desc(skb, i, page, off, copy);
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_io.c
1177
if (off + copy < pg_size) {
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_io.c
1185
TCP_OFF(sk) = off + copy;
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_io.c
1189
tp->write_seq += copy;
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_io.c
1190
copied += copy;
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_io.c
1191
size -= copy;
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_io.c
1194
csk->tlshws.txleft -= copy;
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_io.c
883
int off, int copy)
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_io.c
888
off, copy, skb->len);
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_io.c
892
skb->len += copy;
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_io.c
893
skb->data_len += copy;
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_io.c
894
skb->truesize += copy;
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_io.c
895
sk->sk_wmem_queued += copy;
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
4270
struct list_head chg, copy;
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
4274
INIT_LIST_HEAD(&copy);
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
4320
status = ice_get_profs_vsig(hw, blk, vsig, &copy);
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
4325
status = ice_rem_prof_from_list(hw, &copy, hdl);
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
4329
if (list_empty(&copy)) {
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
4335
} else if (!ice_find_dup_props_vsig(hw, blk, &copy,
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
4353
&copy, &vsig,
drivers/net/ethernet/intel/ice/ice_flex_pipe.c
4379
list_for_each_entry_safe(del1, tmp1, &copy, list) {
drivers/net/ethernet/intel/ice/ice_main.c
6858
struct ice_vsi_tx_stats *copy)
drivers/net/ethernet/intel/ice/ice_main.c
6865
copy->pkts = u64_stats_read(&stats->pkts);
drivers/net/ethernet/intel/ice/ice_main.c
6866
copy->bytes = u64_stats_read(&stats->bytes);
drivers/net/ethernet/intel/ice/ice_main.c
6867
copy->tx_restart_q = u64_stats_read(&stats->tx_restart_q);
drivers/net/ethernet/intel/ice/ice_main.c
6868
copy->tx_busy = u64_stats_read(&stats->tx_busy);
drivers/net/ethernet/intel/ice/ice_main.c
6869
copy->tx_linearize = u64_stats_read(&stats->tx_linearize);
drivers/net/ethernet/intel/ice/ice_main.c
6883
struct ice_vsi_rx_stats *copy)
drivers/net/ethernet/intel/ice/ice_main.c
6890
copy->pkts = u64_stats_read(&stats->pkts);
drivers/net/ethernet/intel/ice/ice_main.c
6891
copy->bytes = u64_stats_read(&stats->bytes);
drivers/net/ethernet/intel/ice/ice_main.c
6892
copy->rx_non_eop_descs =
drivers/net/ethernet/intel/ice/ice_main.c
6894
copy->rx_page_failed = u64_stats_read(&stats->rx_page_failed);
drivers/net/ethernet/intel/ice/ice_main.c
6895
copy->rx_buf_failed = u64_stats_read(&stats->rx_buf_failed);
drivers/net/ethernet/intel/ice/ice_main.c
6910
struct ice_vsi_tx_stats copy = {};
drivers/net/ethernet/intel/ice/ice_main.c
6920
ice_fetch_u64_tx_stats(ring, &copy);
drivers/net/ethernet/intel/ice/ice_main.c
6922
vsi_stats->pkts += copy.pkts;
drivers/net/ethernet/intel/ice/ice_main.c
6923
vsi_stats->bytes += copy.bytes;
drivers/net/ethernet/intel/ice/ice_main.c
6924
vsi_stats->tx_restart_q += copy.tx_restart_q;
drivers/net/ethernet/intel/ice/ice_main.c
6925
vsi_stats->tx_busy += copy.tx_busy;
drivers/net/ethernet/intel/ice/ice_main.c
6926
vsi_stats->tx_linearize += copy.tx_linearize;
drivers/net/ethernet/intel/ice/ice_main.c
6941
struct ice_vsi_rx_stats copy = {};
drivers/net/ethernet/intel/ice/ice_main.c
6951
ice_fetch_u64_rx_stats(ring, &copy);
drivers/net/ethernet/intel/ice/ice_main.c
6953
vsi_stats->pkts += copy.pkts;
drivers/net/ethernet/intel/ice/ice_main.c
6954
vsi_stats->bytes += copy.bytes;
drivers/net/ethernet/intel/ice/ice_main.c
6955
vsi_stats->rx_non_eop_descs += copy.rx_non_eop_descs;
drivers/net/ethernet/intel/ice/ice_main.c
6956
vsi_stats->rx_page_failed += copy.rx_page_failed;
drivers/net/ethernet/intel/ice/ice_main.c
6957
vsi_stats->rx_buf_failed += copy.rx_buf_failed;
drivers/net/ethernet/intel/idpf/idpf_txrx.c
3464
u32 copy = data_len <= L1_CACHE_BYTES ? data_len : ETH_HLEN;
drivers/net/ethernet/intel/idpf/idpf_txrx.c
3470
!libeth_rx_sync_for_cpu(buf, copy))
drivers/net/ethernet/intel/idpf/idpf_txrx.c
3480
memcpy(dst, src, LARGEST_ALIGN(copy));
drivers/net/ethernet/intel/idpf/idpf_txrx.c
3481
buf->offset += copy;
drivers/net/ethernet/intel/idpf/idpf_txrx.c
3483
return copy;
drivers/net/ethernet/mellanox/mlx4/en_tx.c
572
u32 copy = (ring->size - index) << LOG_TXBB_SIZE;
drivers/net/ethernet/mellanox/mlx4/en_tx.c
575
for (i = desc_size - copy - 4; i >= 0; i -= 4) {
drivers/net/ethernet/mellanox/mlx4/en_tx.c
580
*((u32 *) (ring->bounce_buf + copy + i));
drivers/net/ethernet/mellanox/mlx4/en_tx.c
583
for (i = copy - 4; i >= 4 ; i -= 4) {
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1372
int copy;
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1377
copy = min_t(int, size, sizeof(to->first.data));
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1378
memcpy(to->first.data, from, copy);
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1379
size -= copy;
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1380
from += copy;
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1389
copy = min_t(int, size, MLX5_CMD_DATA_BLOCK_SIZE);
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1391
memcpy(block->data, from, copy);
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1392
from += copy;
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1393
size -= copy;
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1405
int copy;
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1410
copy = min_t(int, size, sizeof(from->first.data));
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1411
memcpy(to, from->first.data, copy);
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1412
size -= copy;
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1413
to += copy;
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1422
copy = min_t(int, size, MLX5_CMD_DATA_BLOCK_SIZE);
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1425
memcpy(to, block->data, copy);
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1426
to += copy;
drivers/net/ethernet/mellanox/mlx5/core/cmd.c
1427
size -= copy;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/prm.h
333
struct mlx5_ifc_copy_action_in_bits copy;
drivers/net/ethernet/netronome/nfp/nfp_devlink.c
15
struct nfp_eth_table_port *copy)
drivers/net/ethernet/netronome/nfp/nfp_devlink.c
23
memcpy(copy, eth_port, sizeof(*eth_port));
drivers/net/ethernet/netronome/nfp/nfp_devlink.c
31
struct nfp_eth_table_port *copy)
drivers/net/ethernet/netronome/nfp/nfp_devlink.c
35
return nfp_devlink_fill_eth_port(port, copy);
drivers/net/ethernet/sfc/efx_channels.c
1347
.copy = efx_copy_channel,
drivers/net/ethernet/sfc/efx_channels.c
865
if (channel->type->copy)
drivers/net/ethernet/sfc/efx_channels.c
866
channel = channel->type->copy(channel);
drivers/net/ethernet/sfc/efx_channels.c
884
if (!channel->type->copy)
drivers/net/ethernet/sfc/efx_channels.c
898
if (channel && channel->type->copy) {
drivers/net/ethernet/sfc/falcon/efx.c
776
if (channel->type->copy)
drivers/net/ethernet/sfc/falcon/efx.c
799
if (channel->type->copy)
drivers/net/ethernet/sfc/falcon/efx.c
800
channel = channel->type->copy(channel);
drivers/net/ethernet/sfc/falcon/efx.c
822
if (!channel->type->copy)
drivers/net/ethernet/sfc/falcon/efx.c
834
if (channel && channel->type->copy) {
drivers/net/ethernet/sfc/falcon/efx.c
872
.copy = ef4_copy_channel,
drivers/net/ethernet/sfc/falcon/net_driver.h
481
struct ef4_channel *(*copy)(const struct ef4_channel *);
drivers/net/ethernet/sfc/net_driver.h
628
struct efx_channel *(*copy)(const struct efx_channel *);
drivers/net/ethernet/sfc/ptp.c
2207
.copy = efx_copy_channel,
drivers/net/ethernet/sfc/siena/efx_channels.c
1366
.copy = efx_copy_channel,
drivers/net/ethernet/sfc/siena/efx_channels.c
874
if (channel->type->copy)
drivers/net/ethernet/sfc/siena/efx_channels.c
897
if (channel->type->copy)
drivers/net/ethernet/sfc/siena/efx_channels.c
898
channel = channel->type->copy(channel);
drivers/net/ethernet/sfc/siena/efx_channels.c
919
if (!channel->type->copy)
drivers/net/ethernet/sfc/siena/efx_channels.c
932
if (channel && channel->type->copy) {
drivers/net/ethernet/sfc/siena/net_driver.h
591
struct efx_channel *(*copy)(const struct efx_channel *);
drivers/net/ethernet/sfc/siena/siena_sriov.c
405
struct efx_memcpy_req copy[4];
drivers/net/ethernet/sfc/siena/siena_sriov.c
417
memset(copy, '\0', sizeof(copy));
drivers/net/ethernet/sfc/siena/siena_sriov.c
419
copy[0].from_buf = &status->generation_start;
drivers/net/ethernet/sfc/siena/siena_sriov.c
420
copy[0].to_rid = vf->pci_rid;
drivers/net/ethernet/sfc/siena/siena_sriov.c
421
copy[0].to_addr = vf->status_addr + offsetof(struct vfdi_status,
drivers/net/ethernet/sfc/siena/siena_sriov.c
423
copy[0].length = sizeof(status->generation_start);
drivers/net/ethernet/sfc/siena/siena_sriov.c
429
copy[1].from_rid = efx->pci_dev->devfn;
drivers/net/ethernet/sfc/siena/siena_sriov.c
430
copy[1].from_addr = nic_data->vfdi_status.dma_addr + data_offset;
drivers/net/ethernet/sfc/siena/siena_sriov.c
431
copy[1].to_rid = vf->pci_rid;
drivers/net/ethernet/sfc/siena/siena_sriov.c
432
copy[1].to_addr = vf->status_addr + data_offset;
drivers/net/ethernet/sfc/siena/siena_sriov.c
433
copy[1].length = status->length - data_offset;
drivers/net/ethernet/sfc/siena/siena_sriov.c
445
copy[pos].from_buf = NULL;
drivers/net/ethernet/sfc/siena/siena_sriov.c
446
copy[pos].from_rid = efx->pci_dev->devfn;
drivers/net/ethernet/sfc/siena/siena_sriov.c
447
copy[pos].from_addr = epp->addr;
drivers/net/ethernet/sfc/siena/siena_sriov.c
448
copy[pos].to_rid = vf->pci_rid;
drivers/net/ethernet/sfc/siena/siena_sriov.c
449
copy[pos].to_addr = vf->peer_page_addrs[count];
drivers/net/ethernet/sfc/siena/siena_sriov.c
450
copy[pos].length = EFX_PAGE_SIZE;
drivers/net/ethernet/sfc/siena/siena_sriov.c
452
if (++pos == ARRAY_SIZE(copy)) {
drivers/net/ethernet/sfc/siena/siena_sriov.c
453
efx_siena_sriov_memcpy(efx, copy, ARRAY_SIZE(copy));
drivers/net/ethernet/sfc/siena/siena_sriov.c
460
copy[pos].from_buf = &status->generation_end;
drivers/net/ethernet/sfc/siena/siena_sriov.c
461
copy[pos].to_rid = vf->pci_rid;
drivers/net/ethernet/sfc/siena/siena_sriov.c
462
copy[pos].to_addr = vf->status_addr + offsetof(struct vfdi_status,
drivers/net/ethernet/sfc/siena/siena_sriov.c
464
copy[pos].length = sizeof(status->generation_end);
drivers/net/ethernet/sfc/siena/siena_sriov.c
465
efx_siena_sriov_memcpy(efx, copy, pos + 1);
drivers/net/ethernet/sfc/siena/siena_sriov.c
877
struct efx_memcpy_req copy[2];
drivers/net/ethernet/sfc/siena/siena_sriov.c
881
memset(copy, '\0', sizeof(copy));
drivers/net/ethernet/sfc/siena/siena_sriov.c
882
copy[0].from_rid = vf->pci_rid;
drivers/net/ethernet/sfc/siena/siena_sriov.c
883
copy[0].from_addr = vf->req_addr;
drivers/net/ethernet/sfc/siena/siena_sriov.c
884
copy[0].to_rid = efx->pci_dev->devfn;
drivers/net/ethernet/sfc/siena/siena_sriov.c
885
copy[0].to_addr = vf->buf.dma_addr;
drivers/net/ethernet/sfc/siena/siena_sriov.c
886
copy[0].length = EFX_PAGE_SIZE;
drivers/net/ethernet/sfc/siena/siena_sriov.c
887
rc = efx_siena_sriov_memcpy(efx, copy, 1);
drivers/net/ethernet/sfc/siena/siena_sriov.c
921
memset(copy, '\0', sizeof(copy));
drivers/net/ethernet/sfc/siena/siena_sriov.c
922
copy[0].from_buf = &req->rc;
drivers/net/ethernet/sfc/siena/siena_sriov.c
923
copy[0].to_rid = vf->pci_rid;
drivers/net/ethernet/sfc/siena/siena_sriov.c
924
copy[0].to_addr = vf->req_addr + offsetof(struct vfdi_req, rc);
drivers/net/ethernet/sfc/siena/siena_sriov.c
925
copy[0].length = sizeof(req->rc);
drivers/net/ethernet/sfc/siena/siena_sriov.c
926
copy[1].from_buf = &req->op;
drivers/net/ethernet/sfc/siena/siena_sriov.c
927
copy[1].to_rid = vf->pci_rid;
drivers/net/ethernet/sfc/siena/siena_sriov.c
928
copy[1].to_addr = vf->req_addr + offsetof(struct vfdi_req, op);
drivers/net/ethernet/sfc/siena/siena_sriov.c
929
copy[1].length = sizeof(req->op);
drivers/net/ethernet/sfc/siena/siena_sriov.c
931
(void)efx_siena_sriov_memcpy(efx, copy, ARRAY_SIZE(copy));
drivers/net/virtio_net.c
851
unsigned int copy, hdr_len, hdr_padded_len;
drivers/net/virtio_net.c
893
copy = len;
drivers/net/virtio_net.c
895
copy = ETH_HLEN;
drivers/net/virtio_net.c
896
skb_put_data(skb, p, copy);
drivers/net/virtio_net.c
898
len -= copy;
drivers/net/virtio_net.c
899
offset += copy;
drivers/net/wireless/ath/ath12k/qmi.c
3329
goto copy;
drivers/net/wireless/ath/ath12k/qmi.c
3344
copy:
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx-gen2.c
1230
int copy = IWL_FIRST_TB_SIZE - copy_size;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx-gen2.c
1232
if (copy > cmdlen[i])
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx-gen2.c
1233
copy = cmdlen[i];
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx-gen2.c
1234
cmdlen[i] -= copy;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx-gen2.c
1235
cmddata[i] += copy;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx-gen2.c
1236
copy_size += copy;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx-gen2.c
1326
int copy;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx-gen2.c
1334
copy = cmd->len[i];
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx-gen2.c
1336
memcpy((u8 *)out_cmd + cmd_pos, cmd->data[i], copy);
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx-gen2.c
1337
cmd_pos += copy;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx-gen2.c
1338
copy_size += copy;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx-gen2.c
1347
copy = min_t(int, TFD_MAX_PAYLOAD_SIZE - cmd_pos, cmd->len[i]);
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx-gen2.c
1349
memcpy((u8 *)out_cmd + cmd_pos, cmd->data[i], copy);
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx-gen2.c
1350
cmd_pos += copy;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx-gen2.c
1354
copy = IWL_FIRST_TB_SIZE - copy_size;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx-gen2.c
1356
if (copy > cmd->len[i])
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx-gen2.c
1357
copy = cmd->len[i];
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx-gen2.c
1358
copy_size += copy;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
1387
int copy = IWL_FIRST_TB_SIZE - copy_size;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
1389
if (copy > cmdlen[i])
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
1390
copy = cmdlen[i];
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
1391
cmdlen[i] -= copy;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
1392
cmddata[i] += copy;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
1393
copy_size += copy;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
1494
int copy;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
1502
copy = cmd->len[i];
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
1504
memcpy((u8 *)out_cmd + cmd_pos, cmd->data[i], copy);
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
1505
cmd_pos += copy;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
1506
copy_size += copy;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
1515
copy = min_t(int, TFD_MAX_PAYLOAD_SIZE - cmd_pos, cmd->len[i]);
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
1517
memcpy((u8 *)out_cmd + cmd_pos, cmd->data[i], copy);
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
1518
cmd_pos += copy;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
1522
copy = IWL_FIRST_TB_SIZE - copy_size;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
1524
if (copy > cmd->len[i])
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
1525
copy = cmd->len[i];
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
1526
copy_size += copy;
drivers/net/wireless/mediatek/mt7601u/dma.c
32
u32 true_len, hdr_len = 0, copy, frag;
drivers/net/wireless/mediatek/mt7601u/dma.c
55
copy = (true_len <= skb_tailroom(skb)) ? true_len : hdr_len + 8;
drivers/net/wireless/mediatek/mt7601u/dma.c
56
frag = true_len - copy;
drivers/net/wireless/mediatek/mt7601u/dma.c
58
skb_put_data(skb, data, copy);
drivers/net/wireless/mediatek/mt7601u/dma.c
59
data += copy;
drivers/net/wwan/iosm/iosm_ipc_task_queue.c
146
void *copy = msg;
drivers/net/wwan/iosm/iosm_ipc_task_queue.c
150
copy = kmemdup(msg, size, GFP_ATOMIC);
drivers/net/wwan/iosm/iosm_ipc_task_queue.c
151
if (!copy)
drivers/net/wwan/iosm/iosm_ipc_task_queue.c
157
ret = ipc_task_queue_add_task(imem, arg, copy, func,
drivers/net/wwan/iosm/iosm_ipc_task_queue.c
162
copy, size, is_copy);
drivers/net/wwan/iosm/iosm_ipc_task_queue.c
164
kfree(copy);
drivers/nvdimm/bus.c
1100
u32 in_size, copy;
drivers/nvdimm/bus.c
1109
copy = min_t(u32, ND_CMD_MAX_ENVELOPE - in_len, in_size);
drivers/nvdimm/bus.c
1111
copy = 0;
drivers/nvdimm/bus.c
1112
if (copy && copy_from_user(&in_env[in_len], p + in_len, copy))
drivers/nvdimm/bus.c
1132
u32 copy;
drivers/nvdimm/bus.c
1140
copy = min_t(u32, ND_CMD_MAX_ENVELOPE - out_len, out_size);
drivers/nvdimm/bus.c
1142
copy = 0;
drivers/nvdimm/bus.c
1143
if (copy && copy_from_user(&out_env[out_len],
drivers/nvdimm/bus.c
1144
p + in_len + out_len, copy)) {
drivers/nvme/target/admin-cmd.c
514
goto copy;
drivers/nvme/target/admin-cmd.c
527
copy:
drivers/pinctrl/pinctrl-mcp23s08_spi.c
124
copy = devm_kmemdup(dev, info->regmap, sizeof(*info->regmap), GFP_KERNEL);
drivers/pinctrl/pinctrl-mcp23s08_spi.c
125
if (!copy)
drivers/pinctrl/pinctrl-mcp23s08_spi.c
128
copy->name = name;
drivers/pinctrl/pinctrl-mcp23s08_spi.c
130
mcp->regmap = devm_regmap_init(dev, &mcp23sxx_spi_regmap, mcp, copy);
drivers/pinctrl/pinctrl-mcp23s08_spi.c
86
struct regmap_config *copy;
drivers/s390/block/dasd_devmap.c
2048
struct dasd_copy_relation *copy;
drivers/s390/block/dasd_devmap.c
2057
if (!devmap->copy)
drivers/s390/block/dasd_devmap.c
2060
copy = devmap->copy;
drivers/s390/block/dasd_devmap.c
2063
if (copy->entry[i].configured && copy->entry[i].primary) {
drivers/s390/block/dasd_devmap.c
2064
strscpy(prim_busid, copy->entry[i].busid,
drivers/s390/block/dasd_devmap.c
2074
if (copy->entry[i].configured && !copy->entry[i].primary)
drivers/s390/block/dasd_devmap.c
2076
copy->entry[i].busid);
drivers/s390/block/dasd_devmap.c
2085
struct dasd_copy_relation *copy,
drivers/s390/block/dasd_devmap.c
2093
if (copy->entry[i].configured &&
drivers/s390/block/dasd_devmap.c
2094
strncmp(copy->entry[i].busid, busid, DASD_BUS_ID_SIZE) == 0)
drivers/s390/block/dasd_devmap.c
2097
if (!copy->entry[i].configured)
drivers/s390/block/dasd_devmap.c
2103
copy->entry[i].configured = true;
drivers/s390/block/dasd_devmap.c
2104
strscpy(copy->entry[i].busid, busid, DASD_BUS_ID_SIZE);
drivers/s390/block/dasd_devmap.c
2106
copy->active = &copy->entry[i];
drivers/s390/block/dasd_devmap.c
2107
copy->entry[i].primary = true;
drivers/s390/block/dasd_devmap.c
2109
if (!devmap->copy)
drivers/s390/block/dasd_devmap.c
2110
devmap->copy = copy;
drivers/s390/block/dasd_devmap.c
2115
static void dasd_devmap_del_copy_relation(struct dasd_copy_relation *copy,
drivers/s390/block/dasd_devmap.c
2123
if (copy->entry[i].configured &&
drivers/s390/block/dasd_devmap.c
2124
strncmp(copy->entry[i].busid, busid, DASD_BUS_ID_SIZE) == 0)
drivers/s390/block/dasd_devmap.c
2127
if (i == DASD_CP_ENTRIES || !copy->entry[i].configured) {
drivers/s390/block/dasd_devmap.c
2132
copy->entry[i].configured = false;
drivers/s390/block/dasd_devmap.c
2133
memset(copy->entry[i].busid, 0, DASD_BUS_ID_SIZE);
drivers/s390/block/dasd_devmap.c
2134
if (copy->active == &copy->entry[i]) {
drivers/s390/block/dasd_devmap.c
2135
copy->active = NULL;
drivers/s390/block/dasd_devmap.c
2136
copy->entry[i].primary = false;
drivers/s390/block/dasd_devmap.c
2143
struct dasd_copy_relation *copy;
drivers/s390/block/dasd_devmap.c
2152
if (!devmap->copy)
drivers/s390/block/dasd_devmap.c
2155
copy = devmap->copy;
drivers/s390/block/dasd_devmap.c
2158
if (!copy->entry[i].configured)
drivers/s390/block/dasd_devmap.c
2161
if (copy->entry[i].device == copy->active->device)
drivers/s390/block/dasd_devmap.c
2164
if (copy->entry[i].device)
drivers/s390/block/dasd_devmap.c
2169
if (strlen(copy->entry[i].busid) == 0)
drivers/s390/block/dasd_devmap.c
2171
if (copy->entry[i].device) {
drivers/s390/block/dasd_devmap.c
2172
dasd_put_device(copy->entry[i].device);
drivers/s390/block/dasd_devmap.c
2173
copy->entry[i].device->copy = NULL;
drivers/s390/block/dasd_devmap.c
2174
copy->entry[i].device = NULL;
drivers/s390/block/dasd_devmap.c
2176
devmap = dasd_find_busid_locked(copy->entry[i].busid);
drivers/s390/block/dasd_devmap.c
2177
devmap->copy = NULL;
drivers/s390/block/dasd_devmap.c
2178
memset(copy->entry[i].busid, 0, DASD_BUS_ID_SIZE);
drivers/s390/block/dasd_devmap.c
2180
kfree(copy);
drivers/s390/block/dasd_devmap.c
2233
struct dasd_copy_relation *copy;
drivers/s390/block/dasd_devmap.c
2272
if (prim_devmap->copy) {
drivers/s390/block/dasd_devmap.c
2273
copy = prim_devmap->copy;
drivers/s390/block/dasd_devmap.c
2274
} else if (sec_devmap->copy) {
drivers/s390/block/dasd_devmap.c
2275
copy = sec_devmap->copy;
drivers/s390/block/dasd_devmap.c
2277
copy = kzalloc_obj(*copy);
drivers/s390/block/dasd_devmap.c
2278
if (!copy)
drivers/s390/block/dasd_devmap.c
2282
rc = dasd_devmap_set_copy_relation(prim_devmap, copy, prim_busid, true);
drivers/s390/block/dasd_devmap.c
2287
rc = dasd_devmap_set_copy_relation(sec_devmap, copy, sec_busid, false);
drivers/s390/block/dasd_devmap.c
2295
if (prim_devmap->device && !prim_devmap->device->copy) {
drivers/s390/block/dasd_devmap.c
2306
dasd_devmap_del_copy_relation(copy, prim_busid);
drivers/s390/block/dasd_devmap.c
2307
dasd_devmap_del_copy_relation(copy, sec_busid);
drivers/s390/block/dasd_devmap.c
2320
struct dasd_copy_relation *copy;
drivers/s390/block/dasd_devmap.c
2328
if (!device->copy) {
drivers/s390/block/dasd_devmap.c
2332
copy = device->copy;
drivers/s390/block/dasd_devmap.c
2334
if (copy->active->device == device) {
drivers/s390/block/dasd_devmap.c
2339
if (copy->entry[i].device == device) {
drivers/s390/block/dasd_devmap.c
50
struct dasd_copy_relation *copy;
drivers/s390/block/dasd_devmap.c
661
struct dasd_copy_relation *copy)
drivers/s390/block/dasd_devmap.c
714
if (entry == &copy->entry[j])
drivers/s390/block/dasd_devmap.c
717
tmp_dev = copy->entry[j].device;
drivers/s390/block/dasd_devmap.c
740
struct dasd_copy_relation *copy;
drivers/s390/block/dasd_devmap.c
743
if (!device->copy)
drivers/s390/block/dasd_devmap.c
746
copy = device->copy;
drivers/s390/block/dasd_devmap.c
748
if (copy->entry[i].device == device)
drivers/s390/block/dasd_devmap.c
749
copy->entry[i].device = NULL;
drivers/s390/block/dasd_devmap.c
752
device->copy = NULL;
drivers/s390/block/dasd_devmap.c
764
struct dasd_copy_relation *copy;
drivers/s390/block/dasd_devmap.c
777
copy = devmap->copy;
drivers/s390/block/dasd_devmap.c
779
if (!copy)
drivers/s390/block/dasd_devmap.c
800
if (copy->entry[i].configured &&
drivers/s390/block/dasd_devmap.c
802
copy->entry[i].busid, DASD_BUS_ID_SIZE) == 0) {
drivers/s390/block/dasd_devmap.c
803
entry = &copy->entry[i];
drivers/s390/block/dasd_devmap.c
813
if (dasd_devmap_check_copy_relation(device, entry, data, copy)) {
drivers/s390/block/dasd_devmap.c
820
copy->entry[i].device = device;
drivers/s390/block/dasd_devmap.c
821
device->copy = copy;
drivers/s390/block/dasd_eckd.c
2008
if (!device->copy)
drivers/s390/block/dasd_eckd.c
2011
if (device->copy->active->device == device)
drivers/s390/block/dasd_eckd.c
4067
char *copy = kmem_cache_alloc(dasd_page_cache,
drivers/s390/block/dasd_eckd.c
4069
if (copy && rq_data_dir(req) == WRITE)
drivers/s390/block/dasd_eckd.c
4070
memcpy(copy + bv.bv_offset, dst, bv.bv_len);
drivers/s390/block/dasd_eckd.c
4071
if (copy)
drivers/s390/block/dasd_eckd.c
4072
dst = copy + bv.bv_offset;
drivers/s390/block/dasd_eckd.c
6093
*copy_relation_find_device(struct dasd_copy_relation *copy,
drivers/s390/block/dasd_eckd.c
6099
if (copy->entry[i].configured &&
drivers/s390/block/dasd_eckd.c
6100
strncmp(copy->entry[i].busid, busid, DASD_BUS_ID_SIZE) == 0)
drivers/s390/block/dasd_eckd.c
6101
return copy->entry[i].device;
drivers/s390/block/dasd_eckd.c
6109
static void copy_pair_set_active(struct dasd_copy_relation *copy, char *new_busid,
drivers/s390/block/dasd_eckd.c
6115
if (copy->entry[i].configured &&
drivers/s390/block/dasd_eckd.c
6116
strncmp(copy->entry[i].busid, new_busid,
drivers/s390/block/dasd_eckd.c
6118
copy->active = &copy->entry[i];
drivers/s390/block/dasd_eckd.c
6119
copy->entry[i].primary = true;
drivers/s390/block/dasd_eckd.c
6120
} else if (copy->entry[i].configured &&
drivers/s390/block/dasd_eckd.c
6121
strncmp(copy->entry[i].busid, old_busid,
drivers/s390/block/dasd_eckd.c
6123
copy->entry[i].primary = false;
drivers/s390/block/dasd_eckd.c
6140
struct dasd_copy_relation *copy;
drivers/s390/block/dasd_eckd.c
6145
copy = device->copy;
drivers/s390/block/dasd_eckd.c
6146
if (!copy)
drivers/s390/block/dasd_eckd.c
6148
primary = copy->active->device;
drivers/s390/block/dasd_eckd.c
6155
secondary = copy_relation_find_device(copy, sec_busid);
drivers/s390/block/dasd_eckd.c
6176
copy_pair_set_active(copy, sec_busid, prim_busid);
drivers/s390/block/dasd_fba.c
501
char *copy = kmem_cache_alloc(dasd_page_cache,
drivers/s390/block/dasd_fba.c
503
if (copy && rq_data_dir(req) == WRITE)
drivers/s390/block/dasd_fba.c
504
memcpy(copy + bv.bv_offset, dst, bv.bv_len);
drivers/s390/block/dasd_fba.c
505
if (copy)
drivers/s390/block/dasd_fba.c
506
dst = copy + bv.bv_offset;
drivers/s390/block/dasd_int.h
612
struct dasd_copy_relation *copy;
drivers/s390/block/dasd_ioctl.c
410
if (!device->copy) {
drivers/s390/cio/cio_inject.c
36
struct crw *copy;
drivers/s390/cio/cio_inject.c
39
copy = kmemdup(crw, sizeof(*crw), GFP_KERNEL);
drivers/s390/cio/cio_inject.c
40
if (!copy)
drivers/s390/cio/cio_inject.c
45
kfree(copy);
drivers/s390/cio/cio_inject.c
48
crw_inject_data = copy;
drivers/scsi/cxgbi/libcxgbi.c
1772
unsigned int copy;
drivers/scsi/cxgbi/libcxgbi.c
1786
copy = min(datalen, sglen);
drivers/scsi/cxgbi/libcxgbi.c
1790
frags[i - 1].size += copy;
drivers/scsi/cxgbi/libcxgbi.c
1801
frags[i].size = copy;
drivers/scsi/cxgbi/libcxgbi.c
1804
datalen -= copy;
drivers/scsi/cxgbi/libcxgbi.c
1805
sgoffset += copy;
drivers/scsi/cxgbi/libcxgbi.c
1806
sglen -= copy;
drivers/scsi/isci/request.c
664
bool copy)
drivers/scsi/isci/request.c
719
return sci_stp_pio_request_construct(ireq, copy);
drivers/scsi/isci/request.c
755
bool copy = false;
drivers/scsi/isci/request.c
760
copy = (task->data_dir == DMA_NONE) ? false : true;
drivers/scsi/isci/request.c
765
copy);
drivers/scsi/iscsi_tcp.c
305
unsigned int offset, copy;
drivers/scsi/iscsi_tcp.c
309
copy = segment->size - offset;
drivers/scsi/iscsi_tcp.c
322
bvec_set_page(&bv, sg_page(sg), copy, offset);
drivers/scsi/iscsi_tcp.c
324
bvec_set_virt(&bv, segment->data + offset, copy);
drivers/scsi/iscsi_tcp.c
326
iov_iter_bvec(&msg.msg_iter, ITER_SOURCE, &bv, 1, copy);
drivers/scsi/libiscsi_tcp.c
291
unsigned int copy = 0, copied = 0;
drivers/scsi/libiscsi_tcp.c
293
while (!iscsi_tcp_segment_done(tcp_conn, segment, 1, copy)) {
drivers/scsi/libiscsi_tcp.c
300
copy = min(len - copied, segment->size - segment->copied);
drivers/scsi/libiscsi_tcp.c
301
ISCSI_DBG_TCP(tcp_conn->iscsi_conn, "copying %d\n", copy);
drivers/scsi/libiscsi_tcp.c
302
memcpy(segment->data + segment->copied, ptr + copied, copy);
drivers/scsi/libiscsi_tcp.c
303
copied += copy;
drivers/spi/spi-qpic-snand.c
1374
bool copy = false, copy_ftr = false;
drivers/spi/spi-qpic-snand.c
1390
copy = true;
drivers/spi/spi-qpic-snand.c
1418
if (copy) {
drivers/staging/axis-fifo/axis-fifo.c
127
unsigned int copy;
drivers/staging/axis-fifo/axis-fifo.c
168
copy = min(words_available, READ_BUF_SIZE);
drivers/staging/axis-fifo/axis-fifo.c
170
for (i = 0; i < copy; i++) {
drivers/staging/axis-fifo/axis-fifo.c
174
words_available -= copy;
drivers/staging/axis-fifo/axis-fifo.c
177
copy * sizeof(u32))) {
drivers/staging/axis-fifo/axis-fifo.c
182
copied += copy;
drivers/staging/media/atomisp/pci/sh_css_internal.h
508
} copy;
drivers/staging/media/atomisp/pci/sh_css_sp.c
196
pipe->copy.bin.bytes_available = out_frame->data_bytes;
drivers/staging/media/atomisp/pci/sh_css_sp.c
262
pipe->copy.raw.height = out_frame->frame_info.res.height;
drivers/staging/media/atomisp/pci/sh_css_sp.c
263
pipe->copy.raw.width = out_frame->frame_info.res.width;
drivers/staging/media/atomisp/pci/sh_css_sp.c
264
pipe->copy.raw.padded_width = out_frame->frame_info.padded_width;
drivers/staging/media/atomisp/pci/sh_css_sp.c
265
pipe->copy.raw.raw_bit_depth = out_frame->frame_info.raw_bit_depth;
drivers/staging/media/atomisp/pci/sh_css_sp.c
266
pipe->copy.raw.max_input_width = max_input_width;
drivers/staging/media/atomisp/pci/sh_css_sp.c
333
pipe->copy.raw.height = out_frame->frame_info.res.height;
drivers/staging/media/atomisp/pci/sh_css_sp.c
334
pipe->copy.raw.width = out_frame->frame_info.res.width;
drivers/staging/media/atomisp/pci/sh_css_sp.c
335
pipe->copy.raw.padded_width = out_frame->frame_info.padded_width;
drivers/staging/media/atomisp/pci/sh_css_sp.c
336
pipe->copy.raw.raw_bit_depth = out_frame->frame_info.raw_bit_depth;
drivers/staging/media/atomisp/pci/sh_css_sp.c
337
pipe->copy.raw.max_input_width = max_input_width;
drivers/staging/media/ipu3/ipu3-abi.h
1808
} copy;
drivers/thunderbolt/ctl.c
1061
req->copy = tb_cfg_copy;
drivers/thunderbolt/ctl.c
515
if (req->copy(req, pkg))
drivers/thunderbolt/ctl.c
925
req->copy = tb_cfg_copy;
drivers/thunderbolt/ctl.c
985
req->copy = tb_cfg_copy;
drivers/thunderbolt/ctl.h
89
bool (*copy)(struct tb_cfg_request *req, const struct ctl_pkg *pkg);
drivers/thunderbolt/dma_port.c
110
req->copy = dma_port_copy;
drivers/thunderbolt/dma_port.c
153
req->copy = dma_port_copy;
drivers/thunderbolt/icm.c
2266
req->copy = icm_copy;
drivers/thunderbolt/icm.c
314
req->copy = icm_copy;
drivers/thunderbolt/xdomain.c
145
req->copy = tb_xdomain_copy;
drivers/thunderbolt/xdomain.c
185
req->copy = tb_xdomain_copy;
drivers/usb/chipidea/udc.c
588
bool copy, unsigned int *bounced)
drivers/usb/chipidea/udc.c
627
if (copy)
drivers/usb/chipidea/udc.c
635
static void sglist_do_debounce(struct ci_hw_req *hwreq, bool copy)
drivers/usb/chipidea/udc.c
646
if (copy) {
drivers/usb/gadget/configfs.c
117
char *copy = *s_copy;
drivers/usb/gadget/configfs.c
125
if (copy) {
drivers/usb/gadget/configfs.c
126
str = copy;
drivers/usb/usbip/stub_tx.c
242
unsigned int copy = urb->actual_length;
drivers/usb/usbip/stub_tx.c
246
if (copy == 0)
drivers/usb/usbip/stub_tx.c
249
size = min(copy, sg->length);
drivers/usb/usbip/stub_tx.c
254
copy -= size;
drivers/usb/usbip/usbip_common.c
768
int copy;
drivers/usb/usbip/usbip_common.c
794
copy = size;
drivers/usb/usbip/usbip_common.c
798
if (copy < sg->length)
drivers/usb/usbip/usbip_common.c
799
recv_size = copy;
drivers/usb/usbip/usbip_common.c
809
copy -= recv;
drivers/usb/usbip/usbip_common.c
812
if (!copy)
drivers/vhost/vringh.c
265
int (*copy)(const struct vringh *vrh,
drivers/vhost/vringh.c
280
err = copy(vrh, dst, src, part);
drivers/vhost/vringh.c
301
int (*copy)(const struct vringh *vrh,
drivers/vhost/vringh.c
330
&slowrange, copy);
drivers/vhost/vringh.c
332
err = copy(vrh, &desc, &descs[i], sizeof(desc));
drivers/video/fbdev/leo.c
114
u32 copy;
drivers/w1/slaves/w1_ds2430.c
210
int copy;
drivers/w1/slaves/w1_ds2430.c
235
copy = W1_F14_SCRATCH_SIZE -
drivers/w1/slaves/w1_ds2430.c
238
if (copy > len)
drivers/w1/slaves/w1_ds2430.c
239
copy = len;
drivers/w1/slaves/w1_ds2430.c
241
memcpy(&tmp[addr & W1_F14_SCRATCH_MASK], buf, copy);
drivers/w1/slaves/w1_ds2430.c
249
copy = W1_F14_SCRATCH_SIZE;
drivers/w1/slaves/w1_ds2430.c
250
if (w1_f14_write(sl, addr, copy, buf) < 0) {
drivers/w1/slaves/w1_ds2430.c
255
buf += copy;
drivers/w1/slaves/w1_ds2430.c
256
addr += copy;
drivers/w1/slaves/w1_ds2430.c
257
len -= copy;
drivers/w1/slaves/w1_ds2431.c
209
int copy;
drivers/w1/slaves/w1_ds2431.c
234
copy = W1_F2D_SCRATCH_SIZE -
drivers/w1/slaves/w1_ds2431.c
237
if (copy > len)
drivers/w1/slaves/w1_ds2431.c
238
copy = len;
drivers/w1/slaves/w1_ds2431.c
240
memcpy(&tmp[addr & W1_F2D_SCRATCH_MASK], buf, copy);
drivers/w1/slaves/w1_ds2431.c
248
copy = W1_F2D_SCRATCH_SIZE;
drivers/w1/slaves/w1_ds2431.c
249
if (w1_f2d_write(sl, addr, copy, buf) < 0) {
drivers/w1/slaves/w1_ds2431.c
254
buf += copy;
drivers/w1/slaves/w1_ds2431.c
255
addr += copy;
drivers/w1/slaves/w1_ds2431.c
256
len -= copy;
drivers/w1/slaves/w1_ds2805.c
208
int copy;
drivers/w1/slaves/w1_ds2805.c
233
copy = W1_F0D_SCRATCH_SIZE -
drivers/w1/slaves/w1_ds2805.c
236
if (copy > len)
drivers/w1/slaves/w1_ds2805.c
237
copy = len;
drivers/w1/slaves/w1_ds2805.c
239
memcpy(&tmp[addr & W1_F0D_SCRATCH_MASK], buf, copy);
drivers/w1/slaves/w1_ds2805.c
247
copy = W1_F0D_SCRATCH_SIZE;
drivers/w1/slaves/w1_ds2805.c
248
if (w1_f0d_write(sl, addr, copy, buf) < 0) {
drivers/w1/slaves/w1_ds2805.c
253
buf += copy;
drivers/w1/slaves/w1_ds2805.c
254
addr += copy;
drivers/w1/slaves/w1_ds2805.c
255
len -= copy;
drivers/xen/gntdev.c
959
struct ioctl_gntdev_grant_copy copy;
drivers/xen/gntdev.c
964
if (copy_from_user(&copy, u, sizeof(copy)))
drivers/xen/gntdev.c
981
for (i = 0; i < copy.count; i++) {
drivers/xen/gntdev.c
984
if (copy_from_user(&seg, &copy.segments[i], sizeof(seg))) {
drivers/xen/gntdev.c
990
ret = gntdev_grant_copy_seg(batch, &seg, &copy.segments[i].status);
fs/debugfs/file.c
1020
char *str, *copy = NULL;
fs/debugfs/file.c
1030
copy = kmalloc(len, GFP_KERNEL);
fs/debugfs/file.c
1031
if (!copy) {
fs/debugfs/file.c
1036
copy_len = strscpy(copy, str, len);
fs/debugfs/file.c
1039
kfree(copy);
fs/debugfs/file.c
1043
copy[copy_len] = '\n';
fs/debugfs/file.c
1045
ret = simple_read_from_buffer(user_buf, count, ppos, copy, len);
fs/debugfs/file.c
1046
kfree(copy);
fs/fat/dir.c
1221
unsigned long size, copy;
fs/fat/dir.c
1242
i = n = copy = 0;
fs/fat/dir.c
1254
copy = min(size, sb->s_blocksize);
fs/fat/dir.c
1257
memcpy(bhs[n]->b_data, slots, copy);
fs/fat/dir.c
1261
slots += copy;
fs/fat/dir.c
1262
size -= copy;
fs/fat/dir.c
1270
memset(bhs[n]->b_data + copy, 0, sb->s_blocksize - copy);
fs/fat/dir.c
1271
offset = copy - sizeof(struct msdos_dir_entry);
fs/fat/dir.c
1359
int copy = umin(sb->s_blocksize - offset, size);
fs/fat/dir.c
1360
memcpy(bhs[i]->b_data + offset, slots, copy);
fs/fat/dir.c
1363
slots += copy;
fs/fat/dir.c
1364
size -= copy;
fs/fat/dir.c
1370
int copy = umin(sb->s_blocksize - offset, size);
fs/fat/dir.c
1371
memcpy(bhs[i]->b_data + offset, slots, copy);
fs/fat/fatent.c
381
int err, n, copy;
fs/fat/fatent.c
384
for (copy = 1; copy < sbi->fats; copy++) {
fs/fat/fatent.c
385
sector_t backup_fat = sbi->fat_length * copy;
fs/fuse/dev.c
1157
unsigned int copy = count;
fs/fuse/dev.c
1161
copy = PAGE_SIZE - offset_in_page(offset);
fs/fuse/dev.c
1163
bytes_copied = fuse_copy_do(cs, &buf, &copy);
fs/gfs2/xattr.c
698
unsigned int copy;
fs/gfs2/xattr.c
718
copy = data_len > sdp->sd_jbsize ? sdp->sd_jbsize :
fs/gfs2/xattr.c
720
memcpy(bh->b_data + mh_size, data, copy);
fs/gfs2/xattr.c
721
if (copy < sdp->sd_jbsize)
fs/gfs2/xattr.c
722
memset(bh->b_data + mh_size + copy, 0,
fs/gfs2/xattr.c
723
sdp->sd_jbsize - copy);
fs/gfs2/xattr.c
726
data += copy;
fs/gfs2/xattr.c
727
data_len -= copy;
fs/namespace.c
4031
char *copy;
fs/namespace.c
4037
copy = kmalloc(PAGE_SIZE, GFP_KERNEL);
fs/namespace.c
4038
if (!copy)
fs/namespace.c
4041
left = copy_from_user(copy, data, PAGE_SIZE);
fs/namespace.c
4052
copy[offset] = c;
fs/namespace.c
4058
kfree(copy);
fs/namespace.c
4062
return copy;
fs/nfs/callback_proc.c
718
struct nfs4_copy_state *copy, *tmp_copy;
fs/nfs/callback_proc.c
721
copy = kzalloc_obj(struct nfs4_copy_state);
fs/nfs/callback_proc.c
722
if (!copy)
fs/nfs/callback_proc.c
743
memcpy(&copy->stateid, &args->coa_stateid, NFS4_STATEID_SIZE);
fs/nfs/callback_proc.c
744
nfs4_copy_cb_args(copy, args);
fs/nfs/callback_proc.c
745
list_add_tail(&copy->copies, &cps->clp->pending_cb_stateids);
fs/nfs/callback_proc.c
747
kfree(copy);
fs/nfs/nfs42proc.c
222
struct nfs4_copy_state *copy)
fs/nfs/nfs42proc.c
225
list_del_init(&copy->copies);
fs/nfs/nfs42proc.c
229
list_del_init(&copy->src_copies);
fs/nfs/nfs42proc.c
242
struct nfs4_copy_state *copy, *tmp_copy = NULL, *iter;
fs/nfs/nfs42proc.c
250
copy = kzalloc_obj(struct nfs4_copy_state);
fs/nfs/nfs42proc.c
251
if (!copy)
fs/nfs/nfs42proc.c
267
kfree(copy);
fs/nfs/nfs42proc.c
268
copy = tmp_copy;
fs/nfs/nfs42proc.c
272
memcpy(&copy->stateid, &res->write_res.stateid, NFS4_STATEID_SIZE);
fs/nfs/nfs42proc.c
273
init_completion(&copy->completion);
fs/nfs/nfs42proc.c
274
copy->parent_dst_state = dst_ctx->state;
fs/nfs/nfs42proc.c
275
copy->parent_src_state = src_ctx->state;
fs/nfs/nfs42proc.c
277
list_add_tail(&copy->copies, &dst_server->ss_copies);
fs/nfs/nfs42proc.c
282
list_add_tail(&copy->src_copies, &src_server->ss_src_copies);
fs/nfs/nfs42proc.c
287
status = wait_for_completion_interruptible_timeout(&copy->completion,
fs/nfs/nfs42proc.c
291
nfs4_copy_dequeue_callback(dst_server, src_server, copy);
fs/nfs/nfs42proc.c
294
} else if (copy->flags || copy->error == NFS4ERR_PARTNER_NO_AUTH) {
fs/nfs/nfs42proc.c
300
res->write_res.count = copy->count;
fs/nfs/nfs42proc.c
302
memcpy(&res->write_res.verifier, &copy->verf, sizeof(copy->verf));
fs/nfs/nfs42proc.c
303
status = -copy->error;
fs/nfs/nfs42proc.c
306
kfree(copy);
fs/nfs/nfs42proc.c
309
nfs42_do_offload_cancel_async(dst, &copy->stateid);
fs/nfs/nfs42proc.c
317
status = nfs42_proc_offload_status(dst, &copy->stateid, &copied);
fs/nfs/nfs42proc.c
320
nfs4_copy_dequeue_callback(dst_server, src_server, copy);
fs/nfs/nfs4state.c
1497
struct nfs4_copy_state *copy;
fs/nfs/nfs4state.c
1504
list_for_each_entry(copy, &sp->so_server->ss_copies, copies) {
fs/nfs/nfs4state.c
1507
&copy->parent_dst_state->stateid)))
fs/nfs/nfs4state.c
1509
copy->flags = 1;
fs/nfs/nfs4state.c
1513
complete(&copy->completion);
fs/nfs/nfs4state.c
1516
list_for_each_entry(copy, &sp->so_server->ss_src_copies, src_copies) {
fs/nfs/nfs4state.c
1519
&copy->parent_src_state->stateid)))
fs/nfs/nfs4state.c
1521
copy->flags = 1;
fs/nfs/nfs4state.c
1524
complete(&copy->completion);
fs/nfsd/nfs4proc.c
1437
struct nfsd4_copy *copy;
fs/nfsd/nfs4proc.c
1441
list_for_each_entry(copy, &clp->async_copies, copies) {
fs/nfsd/nfs4proc.c
1442
if (!test_bit(NFSD4_COPY_F_COMPLETED, &copy->cp_flags) &&
fs/nfsd/nfs4proc.c
1443
!test_bit(NFSD4_COPY_F_STOPPED, &copy->cp_flags)) {
fs/nfsd/nfs4proc.c
1459
struct nfsd4_copy *copy;
fs/nfsd/nfs4proc.c
1468
copy = list_entry(pos, struct nfsd4_copy, copies);
fs/nfsd/nfs4proc.c
1469
if (test_bit(NFSD4_COPY_F_OFFLOAD_DONE, &copy->cp_flags)) {
fs/nfsd/nfs4proc.c
1470
if (--copy->cp_ttl) {
fs/nfsd/nfs4proc.c
1471
list_del_init(&copy->copies);
fs/nfsd/nfs4proc.c
1472
list_add(&copy->copies, &reaplist);
fs/nfsd/nfs4proc.c
1481
copy = list_first_entry(&reaplist, struct nfsd4_copy, copies);
fs/nfsd/nfs4proc.c
1482
list_del_init(&copy->copies);
fs/nfsd/nfs4proc.c
1483
cleanup_async_copy(copy);
fs/nfsd/nfs4proc.c
1487
static void nfs4_put_copy(struct nfsd4_copy *copy)
fs/nfsd/nfs4proc.c
1489
if (!refcount_dec_and_test(&copy->refcount))
fs/nfsd/nfs4proc.c
1491
kfree(copy->cp_src);
fs/nfsd/nfs4proc.c
1492
kfree(copy);
fs/nfsd/nfs4proc.c
1495
static void release_copy_files(struct nfsd4_copy *copy);
fs/nfsd/nfs4proc.c
1497
static void nfsd4_stop_copy(struct nfsd4_copy *copy)
fs/nfsd/nfs4proc.c
1499
trace_nfsd_copy_async_cancel(copy);
fs/nfsd/nfs4proc.c
1500
if (!test_and_set_bit(NFSD4_COPY_F_STOPPED, &copy->cp_flags)) {
fs/nfsd/nfs4proc.c
1501
kthread_stop(copy->copy_task);
fs/nfsd/nfs4proc.c
1502
if (!test_bit(NFSD4_COPY_F_CB_ERROR, &copy->cp_flags))
fs/nfsd/nfs4proc.c
1503
copy->nfserr = nfs_ok;
fs/nfsd/nfs4proc.c
1504
set_bit(NFSD4_COPY_F_COMPLETED, &copy->cp_flags);
fs/nfsd/nfs4proc.c
1514
release_copy_files(copy);
fs/nfsd/nfs4proc.c
1515
nfs4_put_copy(copy);
fs/nfsd/nfs4proc.c
1520
struct nfsd4_copy *copy = NULL;
fs/nfsd/nfs4proc.c
1524
copy = list_first_entry(&clp->async_copies, struct nfsd4_copy,
fs/nfsd/nfs4proc.c
1526
refcount_inc(&copy->refcount);
fs/nfsd/nfs4proc.c
1527
copy->cp_clp = NULL;
fs/nfsd/nfs4proc.c
1528
if (!list_empty(&copy->copies))
fs/nfsd/nfs4proc.c
1529
list_del_init(&copy->copies);
fs/nfsd/nfs4proc.c
1532
return copy;
fs/nfsd/nfs4proc.c
1537
struct nfsd4_copy *copy;
fs/nfsd/nfs4proc.c
1539
while ((copy = nfsd4_unhash_copy(clp)) != NULL)
fs/nfsd/nfs4proc.c
1540
nfsd4_stop_copy(copy);
fs/nfsd/nfs4proc.c
1543
static bool nfsd4_copy_on_sb(const struct nfsd4_copy *copy,
fs/nfsd/nfs4proc.c
1546
if (copy->nf_src &&
fs/nfsd/nfs4proc.c
1547
file_inode(copy->nf_src->nf_file)->i_sb == sb)
fs/nfsd/nfs4proc.c
1549
if (copy->nf_dst &&
fs/nfsd/nfs4proc.c
1550
file_inode(copy->nf_dst->nf_file)->i_sb == sb)
fs/nfsd/nfs4proc.c
1563
struct nfsd4_copy *copy, *tmp;
fs/nfsd/nfs4proc.c
1574
list_for_each_entry_safe(copy, tmp,
fs/nfsd/nfs4proc.c
1576
if (nfsd4_copy_on_sb(copy, sb)) {
fs/nfsd/nfs4proc.c
1577
refcount_inc(&copy->refcount);
fs/nfsd/nfs4proc.c
1588
copy->nfserr = nfserr_admin_revoked;
fs/nfsd/nfs4proc.c
1590
&copy->cp_flags);
fs/nfsd/nfs4proc.c
1591
list_move(&copy->copies, &to_cancel);
fs/nfsd/nfs4proc.c
1599
list_for_each_entry_safe(copy, tmp, &to_cancel, copies) {
fs/nfsd/nfs4proc.c
1600
struct nfs4_client *clp = copy->cp_clp;
fs/nfsd/nfs4proc.c
1602
list_del_init(&copy->copies);
fs/nfsd/nfs4proc.c
1603
nfsd4_stop_copy(copy);
fs/nfsd/nfs4proc.c
1802
struct nfsd4_copy *copy)
fs/nfsd/nfs4proc.c
1805
stateid_t *s_stid = &copy->cp_src_stateid;
fs/nfsd/nfs4proc.c
1810
&copy->cp_dst_stateid,
fs/nfsd/nfs4proc.c
1811
WR_STATE, &copy->nf_dst, NULL);
fs/nfsd/nfs4proc.c
1815
status = nfsd4_interssc_connect(copy->cp_src, rqstp, &copy->ss_nsui);
fs/nfsd/nfs4proc.c
1821
copy->c_fh.size = s_fh->fh_handle.fh_size;
fs/nfsd/nfs4proc.c
1822
memcpy(copy->c_fh.data, &s_fh->fh_handle.fh_raw, copy->c_fh.size);
fs/nfsd/nfs4proc.c
1823
copy->stateid.seqid = cpu_to_be32(s_stid->si_generation);
fs/nfsd/nfs4proc.c
1824
memcpy(copy->stateid.other, (void *)&s_stid->si_opaque,
fs/nfsd/nfs4proc.c
1860
struct nfsd4_copy *copy)
fs/nfsd/nfs4proc.c
1882
struct nfsd4_copy *copy)
fs/nfsd/nfs4proc.c
1884
return nfsd4_verify_copy(rqstp, cstate, &copy->cp_src_stateid,
fs/nfsd/nfs4proc.c
1885
&copy->nf_src, &copy->cp_dst_stateid,
fs/nfsd/nfs4proc.c
1886
&copy->nf_dst);
fs/nfsd/nfs4proc.c
1893
struct nfsd4_copy *copy =
fs/nfsd/nfs4proc.c
1896
set_bit(NFSD4_COPY_F_OFFLOAD_DONE, &copy->cp_flags);
fs/nfsd/nfs4proc.c
1924
static void nfsd4_init_copy_res(struct nfsd4_copy *copy, bool sync)
fs/nfsd/nfs4proc.c
1926
copy->cp_res.wr_stable_how =
fs/nfsd/nfs4proc.c
1927
test_bit(NFSD4_COPY_F_COMMITTED, &copy->cp_flags) ?
fs/nfsd/nfs4proc.c
1929
nfsd4_copy_set_sync(copy, sync);
fs/nfsd/nfs4proc.c
1932
static ssize_t _nfsd_copy_file_range(struct nfsd4_copy *copy,
fs/nfsd/nfs4proc.c
1938
u64 bytes_total = copy->cp_count;
fs/nfsd/nfs4proc.c
1939
u64 src_pos = copy->cp_src_pos;
fs/nfsd/nfs4proc.c
1940
u64 dst_pos = copy->cp_dst_pos;
fs/nfsd/nfs4proc.c
1956
copy->cp_res.wr_bytes_written += bytes_copied;
fs/nfsd/nfs4proc.c
1959
} while (bytes_total > 0 && nfsd4_copy_is_async(copy));
fs/nfsd/nfs4proc.c
1961
if (nfsd4_copy_is_async(copy) && copy->cp_res.wr_bytes_written > 0) {
fs/nfsd/nfs4proc.c
1963
end = copy->cp_dst_pos + copy->cp_res.wr_bytes_written - 1;
fs/nfsd/nfs4proc.c
1964
status = vfs_fsync_range(dst, copy->cp_dst_pos, end, 0);
fs/nfsd/nfs4proc.c
1968
set_bit(NFSD4_COPY_F_COMMITTED, &copy->cp_flags);
fs/nfsd/nfs4proc.c
1973
static __be32 nfsd4_do_copy(struct nfsd4_copy *copy,
fs/nfsd/nfs4proc.c
1980
bytes = _nfsd_copy_file_range(copy, dst, src);
fs/nfsd/nfs4proc.c
1985
if (bytes < 0 && !copy->cp_res.wr_bytes_written)
fs/nfsd/nfs4proc.c
1988
nfsd4_init_copy_res(copy, sync);
fs/nfsd/nfs4proc.c
2015
static void release_copy_files(struct nfsd4_copy *copy)
fs/nfsd/nfs4proc.c
2017
if (copy->nf_src) {
fs/nfsd/nfs4proc.c
2018
nfsd_file_put(copy->nf_src);
fs/nfsd/nfs4proc.c
2019
copy->nf_src = NULL;
fs/nfsd/nfs4proc.c
2021
if (copy->nf_dst) {
fs/nfsd/nfs4proc.c
2022
nfsd_file_put(copy->nf_dst);
fs/nfsd/nfs4proc.c
2023
copy->nf_dst = NULL;
fs/nfsd/nfs4proc.c
2027
static void cleanup_async_copy(struct nfsd4_copy *copy)
fs/nfsd/nfs4proc.c
2029
nfs4_free_copy_state(copy);
fs/nfsd/nfs4proc.c
2030
release_copy_files(copy);
fs/nfsd/nfs4proc.c
2031
if (copy->cp_clp) {
fs/nfsd/nfs4proc.c
2032
spin_lock(&copy->cp_clp->async_lock);
fs/nfsd/nfs4proc.c
2033
if (!list_empty(&copy->copies))
fs/nfsd/nfs4proc.c
2034
list_del_init(&copy->copies);
fs/nfsd/nfs4proc.c
2035
spin_unlock(&copy->cp_clp->async_lock);
fs/nfsd/nfs4proc.c
2037
nfs4_put_copy(copy);
fs/nfsd/nfs4proc.c
2040
static void nfsd4_send_cb_offload(struct nfsd4_copy *copy)
fs/nfsd/nfs4proc.c
2042
struct nfsd4_cb_offload *cbo = &copy->cp_cb_offload;
fs/nfsd/nfs4proc.c
2043
struct nfs4_client *clp = copy->cp_clp;
fs/nfsd/nfs4proc.c
2050
set_bit(NFSD4_COPY_F_OFFLOAD_DONE, &copy->cp_flags);
fs/nfsd/nfs4proc.c
2054
memcpy(&cbo->co_res, &copy->cp_res, sizeof(copy->cp_res));
fs/nfsd/nfs4proc.c
2055
memcpy(&cbo->co_fh, &copy->fh, sizeof(copy->fh));
fs/nfsd/nfs4proc.c
2056
cbo->co_nfserr = copy->nfserr;
fs/nfsd/nfs4proc.c
2071
&cbo->co_fh, copy->cp_count, copy->nfserr);
fs/nfsd/nfs4proc.c
2084
struct nfsd4_copy *copy = (struct nfsd4_copy *)data;
fs/nfsd/nfs4proc.c
2087
trace_nfsd_copy_async(copy);
fs/nfsd/nfs4proc.c
2088
if (nfsd4_ssc_is_inter(copy)) {
fs/nfsd/nfs4proc.c
2091
filp = nfs42_ssc_open(copy->ss_nsui->nsui_vfsmount,
fs/nfsd/nfs4proc.c
2092
&copy->c_fh, &copy->stateid);
fs/nfsd/nfs4proc.c
2104
nfserr = nfsd4_do_copy(copy, filp, copy->nf_dst->nf_file,
fs/nfsd/nfs4proc.c
2106
nfsd4_cleanup_inter_ssc(copy->ss_nsui, filp, copy->nf_dst);
fs/nfsd/nfs4proc.c
2108
nfserr = nfsd4_do_copy(copy, copy->nf_src->nf_file,
fs/nfsd/nfs4proc.c
2109
copy->nf_dst->nf_file, false);
fs/nfsd/nfs4proc.c
2113
if (!test_bit(NFSD4_COPY_F_CB_ERROR, &copy->cp_flags))
fs/nfsd/nfs4proc.c
2114
copy->nfserr = nfserr;
fs/nfsd/nfs4proc.c
2117
set_bit(NFSD4_COPY_F_STOPPED, &copy->cp_flags);
fs/nfsd/nfs4proc.c
2119
set_bit(NFSD4_COPY_F_COMPLETED, &copy->cp_flags);
fs/nfsd/nfs4proc.c
2120
trace_nfsd_copy_async_done(copy);
fs/nfsd/nfs4proc.c
2121
nfsd4_send_cb_offload(copy);
fs/nfsd/nfs4proc.c
2122
atomic_dec(&copy->cp_nn->pending_async_copies);
fs/nfsd/nfs4proc.c
2132
struct nfsd4_copy *copy = &u->copy;
fs/nfsd/nfs4proc.c
2136
result = &copy->cp_res;
fs/nfsd/nfs4proc.c
2139
copy->cp_clp = cstate->clp;
fs/nfsd/nfs4proc.c
2140
if (nfsd4_ssc_is_inter(copy)) {
fs/nfsd/nfs4proc.c
2141
trace_nfsd_copy_inter(copy);
fs/nfsd/nfs4proc.c
2142
if (!inter_copy_offload_enable || nfsd4_copy_is_sync(copy)) {
fs/nfsd/nfs4proc.c
2146
status = nfsd4_setup_inter_ssc(rqstp, cstate, copy);
fs/nfsd/nfs4proc.c
2148
trace_nfsd_copy_done(copy, status);
fs/nfsd/nfs4proc.c
2152
trace_nfsd_copy_intra(copy);
fs/nfsd/nfs4proc.c
2153
status = nfsd4_setup_intra_ssc(rqstp, cstate, copy);
fs/nfsd/nfs4proc.c
2155
trace_nfsd_copy_done(copy, status);
fs/nfsd/nfs4proc.c
2160
memcpy(&copy->fh, &cstate->current_fh.fh_handle,
fs/nfsd/nfs4proc.c
2162
if (nfsd4_copy_is_async(copy)) {
fs/nfsd/nfs4proc.c
2177
if (!nfs4_init_copy_state(nn, copy))
fs/nfsd/nfs4proc.c
2179
memcpy(&result->cb_stateid, &copy->cp_stateid.cs_stid,
fs/nfsd/nfs4proc.c
2181
dup_copy_fields(copy, async_copy);
fs/nfsd/nfs4proc.c
2198
status = nfsd4_do_copy(copy, copy->nf_src->nf_file,
fs/nfsd/nfs4proc.c
2199
copy->nf_dst->nf_file, true);
fs/nfsd/nfs4proc.c
2202
trace_nfsd_copy_done(copy, status);
fs/nfsd/nfs4proc.c
2203
release_copy_files(copy);
fs/nfsd/nfs4proc.c
2209
if (nfsd4_ssc_is_inter(copy)) {
fs/nfsd/nfs4proc.c
2215
refcount_dec(&copy->ss_nsui->nsui_refcnt);
fs/nfsd/nfs4proc.c
2226
struct nfsd4_copy *copy;
fs/nfsd/nfs4proc.c
2230
list_for_each_entry(copy, &clp->async_copies, copies) {
fs/nfsd/nfs4proc.c
2231
if (memcmp(&copy->cp_stateid.cs_stid, stateid, NFS4_STATEID_SIZE))
fs/nfsd/nfs4proc.c
2233
return copy;
fs/nfsd/nfs4proc.c
2241
struct nfsd4_copy *copy;
fs/nfsd/nfs4proc.c
2244
copy = find_async_copy_locked(clp, stateid);
fs/nfsd/nfs4proc.c
2245
if (copy)
fs/nfsd/nfs4proc.c
2246
refcount_inc(&copy->refcount);
fs/nfsd/nfs4proc.c
2248
return copy;
fs/nfsd/nfs4proc.c
2257
struct nfsd4_copy *copy;
fs/nfsd/nfs4proc.c
2260
copy = find_async_copy(clp, &os->stateid);
fs/nfsd/nfs4proc.c
2261
if (!copy) {
fs/nfsd/nfs4proc.c
2266
nfsd4_stop_copy(copy);
fs/nfsd/nfs4proc.c
2345
struct nfsd4_copy *copy;
fs/nfsd/nfs4proc.c
2350
copy = find_async_copy_locked(clp, &os->stateid);
fs/nfsd/nfs4proc.c
2351
if (copy) {
fs/nfsd/nfs4proc.c
2352
os->count = copy->cp_res.wr_bytes_written;
fs/nfsd/nfs4proc.c
2353
if (test_bit(NFSD4_COPY_F_COMPLETED, &copy->cp_flags)) {
fs/nfsd/nfs4proc.c
2355
os->status = copy->nfserr;
fs/nfsd/nfs4proc.c
2999
struct nfsd4_copy *copy;
fs/nfsd/nfs4proc.c
3015
copy = (struct nfsd4_copy *)&op->u;
fs/nfsd/nfs4proc.c
3021
if (nfsd4_ssc_is_inter(copy))
fs/nfsd/nfs4proc.c
60
static void cleanup_async_copy(struct nfsd4_copy *copy);
fs/nfsd/nfs4state.c
1002
copy->cp_stateid.cs_stid.si_opaque.so_id);
fs/nfsd/nfs4state.c
967
int nfs4_init_copy_state(struct nfsd_net *nn, struct nfsd4_copy *copy)
fs/nfsd/nfs4state.c
969
return nfs4_init_cp_state(nn, &copy->cp_stateid, NFS4_COPY_STID);
fs/nfsd/nfs4state.c
993
void nfs4_free_copy_state(struct nfsd4_copy *copy)
fs/nfsd/nfs4state.c
997
if (copy->cp_stateid.cs_type != NFS4_COPY_STID)
fs/nfsd/nfs4state.c
999
nn = net_generic(copy->cp_clp->net, nfsd_net_id);
fs/nfsd/nfs4xdr.c
2147
struct nfsd4_copy *copy = &u->copy;
fs/nfsd/nfs4xdr.c
2152
memset(copy, 0, sizeof(*copy));
fs/nfsd/nfs4xdr.c
2153
status = nfsd4_decode_stateid4(argp, &copy->cp_src_stateid);
fs/nfsd/nfs4xdr.c
2156
status = nfsd4_decode_stateid4(argp, &copy->cp_dst_stateid);
fs/nfsd/nfs4xdr.c
2159
if (xdr_stream_decode_u64(argp->xdr, &copy->cp_src_pos) < 0)
fs/nfsd/nfs4xdr.c
2161
if (xdr_stream_decode_u64(argp->xdr, &copy->cp_dst_pos) < 0)
fs/nfsd/nfs4xdr.c
2163
if (xdr_stream_decode_u64(argp->xdr, &copy->cp_count) < 0)
fs/nfsd/nfs4xdr.c
2170
nfsd4_copy_set_sync(copy, sync);
fs/nfsd/nfs4xdr.c
2174
copy->cp_src = svcxdr_tmpalloc(argp, sizeof(*copy->cp_src));
fs/nfsd/nfs4xdr.c
2175
if (copy->cp_src == NULL)
fs/nfsd/nfs4xdr.c
2178
__set_bit(NFSD4_COPY_F_INTRA, &copy->cp_flags);
fs/nfsd/nfs4xdr.c
2183
status = nfsd4_decode_nl4_server(argp, copy->cp_src);
fs/nfsd/nfs4xdr.c
5640
const struct nfsd4_copy *copy)
fs/nfsd/nfs4xdr.c
5642
const struct nfsd42_write_res *write = &copy->cp_res;
fs/nfsd/nfs4xdr.c
5643
u32 count = nfsd4_copy_is_sync(copy) ? 0 : 1;
fs/nfsd/nfs4xdr.c
5667
const struct nfsd4_copy *copy)
fs/nfsd/nfs4xdr.c
5676
return nfsd4_encode_bool(xdr, nfsd4_copy_is_sync(copy));
fs/nfsd/nfs4xdr.c
5683
struct nfsd4_copy *copy = &u->copy;
fs/nfsd/nfs4xdr.c
5685
nfserr = nfsd4_encode_write_response4(resp->xdr, copy);
fs/nfsd/nfs4xdr.c
5688
return nfsd4_encode_copy_requirements4(resp->xdr, copy);
fs/nfsd/state.h
802
int nfs4_init_copy_state(struct nfsd_net *nn, struct nfsd4_copy *copy);
fs/nfsd/state.h
803
void nfs4_free_copy_state(struct nfsd4_copy *copy);
fs/nfsd/trace.h
2245
const struct nfsd4_copy *copy
fs/nfsd/trace.h
2247
TP_ARGS(copy),
fs/nfsd/trace.h
2269
const stateid_t *src_stp = &copy->cp_src_stateid;
fs/nfsd/trace.h
2270
const stateid_t *dst_stp = &copy->cp_dst_stateid;
fs/nfsd/trace.h
2271
const stateid_t *cb_stp = &copy->cp_res.cb_stateid;
fs/nfsd/trace.h
2273
__entry->intra = test_bit(NFSD4_COPY_F_INTRA, &copy->cp_flags);
fs/nfsd/trace.h
2274
__entry->async = !test_bit(NFSD4_COPY_F_SYNCHRONOUS, &copy->cp_flags);
fs/nfsd/trace.h
2287
__entry->src_cp_pos = copy->cp_src_pos;
fs/nfsd/trace.h
2288
__entry->dst_cp_pos = copy->cp_dst_pos;
fs/nfsd/trace.h
2289
__entry->cp_count = copy->cp_count;
fs/nfsd/trace.h
2290
__assign_sockaddr(addr, &copy->cp_clp->cl_addr,
fs/nfsd/trace.h
2311
TP_PROTO(const struct nfsd4_copy *copy), \
fs/nfsd/trace.h
2312
TP_ARGS(copy))
fs/nfsd/trace.h
2320
const struct nfsd4_copy *copy,
fs/nfsd/trace.h
2323
TP_ARGS(copy, status),
fs/nfsd/trace.h
2332
__entry->intra = test_bit(NFSD4_COPY_F_INTRA, &copy->cp_flags);
fs/nfsd/trace.h
2333
__entry->async = !test_bit(NFSD4_COPY_F_SYNCHRONOUS, &copy->cp_flags);
fs/nfsd/trace.h
2334
__assign_sockaddr(addr, &copy->cp_clp->cl_addr,
fs/nfsd/trace.h
2344
const struct nfsd4_copy *copy
fs/nfsd/trace.h
2346
TP_ARGS(copy),
fs/nfsd/trace.h
2369
const stateid_t *src_stp = &copy->cp_src_stateid;
fs/nfsd/trace.h
2370
const stateid_t *dst_stp = &copy->cp_dst_stateid;
fs/nfsd/trace.h
2371
const stateid_t *cb_stp = &copy->cp_res.cb_stateid;
fs/nfsd/trace.h
2373
__entry->status = be32_to_cpu(copy->nfserr);
fs/nfsd/trace.h
2374
__entry->intra = test_bit(NFSD4_COPY_F_INTRA, &copy->cp_flags);
fs/nfsd/trace.h
2375
__entry->async = !test_bit(NFSD4_COPY_F_SYNCHRONOUS, &copy->cp_flags);
fs/nfsd/trace.h
2388
__entry->src_cp_pos = copy->cp_src_pos;
fs/nfsd/trace.h
2389
__entry->dst_cp_pos = copy->cp_dst_pos;
fs/nfsd/trace.h
2390
__entry->cp_count = copy->cp_count;
fs/nfsd/trace.h
2391
__assign_sockaddr(addr, &copy->cp_clp->cl_addr,
fs/nfsd/trace.h
2414
TP_PROTO(const struct nfsd4_copy *copy), \
fs/nfsd/trace.h
2415
TP_ARGS(copy))
fs/nfsd/xdr4.h
769
static inline void nfsd4_copy_set_sync(struct nfsd4_copy *copy, bool sync)
fs/nfsd/xdr4.h
772
set_bit(NFSD4_COPY_F_SYNCHRONOUS, &copy->cp_flags);
fs/nfsd/xdr4.h
774
clear_bit(NFSD4_COPY_F_SYNCHRONOUS, &copy->cp_flags);
fs/nfsd/xdr4.h
777
static inline bool nfsd4_copy_is_sync(const struct nfsd4_copy *copy)
fs/nfsd/xdr4.h
779
return test_bit(NFSD4_COPY_F_SYNCHRONOUS, &copy->cp_flags);
fs/nfsd/xdr4.h
782
static inline bool nfsd4_copy_is_async(const struct nfsd4_copy *copy)
fs/nfsd/xdr4.h
784
return !test_bit(NFSD4_COPY_F_SYNCHRONOUS, &copy->cp_flags);
fs/nfsd/xdr4.h
787
static inline bool nfsd4_ssc_is_inter(const struct nfsd4_copy *copy)
fs/nfsd/xdr4.h
789
return !test_bit(NFSD4_COPY_F_INTRA, &copy->cp_flags);
fs/nfsd/xdr4.h
883
struct nfsd4_copy copy;
fs/ntfs3/frecord.c
2230
u32 pages_per_frame, int copy)
fs/ntfs3/frecord.c
2454
if (!err && copy) {
fs/ntfs3/ntfs_fs.h
602
u32 pages_per_frame, int copy);
fs/pnode.c
314
struct mount *m, *n, *copy, *this;
fs/pnode.c
323
copy = source_mnt;
fs/pnode.c
339
copy = find_master(n, copy, source_mnt);
fs/pnode.c
340
this = copy_tree(copy, copy->mnt.mnt_root, type);
fs/pnode.c
349
copy = this;
fs/smb/server/transport_tcp.c
127
int copy = min(bytes, iov->iov_len);
fs/smb/server/transport_tcp.c
129
bytes -= copy;
fs/smb/server/transport_tcp.c
130
base += copy;
fs/udf/directory.c
333
int copy;
fs/udf/directory.c
338
copy = min(off + len, len1) - off;
fs/udf/directory.c
339
memcpy(buf1 + off, src, copy);
fs/udf/directory.c
340
src += copy;
fs/udf/directory.c
341
len -= copy;
fs/udf/directory.c
354
int copy;
fs/udf/directory.c
360
copy = min(off + len, len1) - off;
fs/udf/directory.c
361
crc = crc_itu_t(crc, buf1 + off, copy);
fs/udf/directory.c
362
len -= copy;
fs/userfaultfd.c
1617
if (unlikely(put_user(ret, &user_uffdio_copy->copy)))
fs/userfaultfd.c
1648
if (unlikely(put_user(ret, &user_uffdio_copy->copy)))
include/linux/bitmap.h
280
unsigned int copy = BITS_TO_LONGS(count);
include/linux/bitmap.h
282
memcpy(to, from, copy * sizeof(long));
include/linux/bitmap.h
284
to[copy - 1] &= BITMAP_LAST_WORD_MASK(count);
include/linux/bitmap.h
285
memset(to + copy, 0, bitmap_size(size) - copy * sizeof(long));
include/linux/cdrom.h
197
__u8 copy : 1;
include/linux/cdrom.h
212
__u8 copy : 1;
include/linux/perf_event.h
80
perf_copy_f copy;
include/linux/rbtree_augmented.h
129
.copy = RBNAME ## _copy, \
include/linux/rbtree_augmented.h
273
augment->copy(node, successor);
include/linux/rbtree_augmented.h
29
void (*copy)(struct rb_node *old, struct rb_node *new);
include/linux/rbtree_augmented.h
299
augment->copy(node, successor);
include/linux/regset.h
272
unsigned int copy = (end_pos < 0 ? *count
include/linux/regset.h
276
memcpy(data, *kbuf, copy);
include/linux/regset.h
277
*kbuf += copy;
include/linux/regset.h
278
} else if (__copy_from_user(data, *ubuf, copy))
include/linux/regset.h
281
*ubuf += copy;
include/linux/regset.h
282
*pos += copy;
include/linux/regset.h
283
*count -= copy;
include/linux/regset.h
299
unsigned int copy = (end_pos < 0 ? *count
include/linux/regset.h
302
*kbuf += copy;
include/linux/regset.h
304
*ubuf += copy;
include/linux/regset.h
305
*pos += copy;
include/linux/regset.h
306
*count -= copy;
include/linux/skbuff.h
5133
bool copy)
include/linux/skbuff.h
5140
if (copy)
include/linux/skmsg.h
249
if (test_bit(msg->sg.start, msg->sg.copy)) {
include/linux/skmsg.h
268
__set_bit(msg->sg.end, msg->sg.copy);
include/linux/skmsg.h
277
__set_bit(i, msg->sg.copy);
include/linux/skmsg.h
279
__clear_bit(i, msg->sg.copy);
include/linux/skmsg.h
32
DECLARE_BITMAP(copy, MAX_MSG_FRAGS + 2);
include/net/sock.h
2290
int copy, int offset)
include/net/sock.h
2294
if (!csum_and_copy_from_iter_full(to, copy, &csum, from))
include/net/sock.h
2298
if (!copy_from_iter_full_nocache(to, copy, from))
include/net/sock.h
2300
} else if (!copy_from_iter_full(to, copy, from))
include/net/sock.h
2307
struct iov_iter *from, int copy)
include/net/sock.h
2311
err = skb_do_copy_data_nocache(sk, skb, from, skb_put(skb, copy),
include/net/sock.h
2312
copy, offset);
include/net/sock.h
2322
int off, int copy)
include/net/sock.h
2327
copy, skb->len);
include/net/sock.h
2331
skb_len_add(skb, copy);
include/net/sock.h
2332
sk_wmem_queued_add(sk, copy);
include/net/sock.h
2333
sk_mem_charge(sk, copy);
include/net/tcp.h
376
int tcp_wmem_schedule(struct sock *sk, int copy);
include/sound/compress_driver.h
165
int (*copy)(struct snd_compr_stream *stream, char __user *buf,
include/sound/pcm-indirect.h
110
snd_pcm_indirect_copy_t copy)
include/sound/pcm-indirect.h
137
copy(substream, rec, bytes);
include/sound/pcm-indirect.h
36
snd_pcm_indirect_copy_t copy)
include/sound/pcm-indirect.h
64
copy(substream, rec, bytes);
include/sound/pcm.h
73
int (*copy)(struct snd_pcm_substream *substream, int channel,
include/sound/soc-component.h
140
int (*copy)(struct snd_soc_component *component,
include/sound/soc-component.h
51
int (*copy)(struct snd_soc_component *component,
include/trace/events/smbus.h
159
goto copy;
include/trace/events/smbus.h
163
goto copy;
include/trace/events/smbus.h
168
copy:
include/trace/events/smbus.h
50
goto copy;
include/trace/events/smbus.h
54
goto copy;
include/trace/events/smbus.h
59
copy:
include/uapi/linux/cdrom.h
886
__u8 copy : 1;
include/uapi/linux/cdrom.h
898
__u8 copy : 1;
include/uapi/linux/userfaultfd.h
293
__s64 copy;
io_uring/zcrx.c
1366
int i, copy, end, off;
io_uring/zcrx.c
1413
copy = end - offset;
io_uring/zcrx.c
1414
if (copy > len)
io_uring/zcrx.c
1415
copy = len;
io_uring/zcrx.c
1418
ret = io_zcrx_recv_frag(req, ifq, frag, off, copy);
io_uring/zcrx.c
1424
if (len == 0 || ret != copy)
io_uring/zcrx.c
1438
copy = end - offset;
io_uring/zcrx.c
1439
if (copy > len)
io_uring/zcrx.c
1440
copy = len;
io_uring/zcrx.c
1444
ret = io_zcrx_recv_skb(desc, frag_iter, off, copy);
io_uring/zcrx.c
1451
if (len == 0 || ret != copy)
ipc/msg.c
1047
struct msg_msg *copy;
ipc/msg.c
1052
copy = load_msg(buf, bufsz);
ipc/msg.c
1053
if (!IS_ERR(copy))
ipc/msg.c
1054
copy->m_ts = bufsz;
ipc/msg.c
1055
return copy;
ipc/msg.c
1058
static inline void free_copy(struct msg_msg *copy)
ipc/msg.c
1060
if (copy)
ipc/msg.c
1061
free_msg(copy);
ipc/msg.c
1069
static inline void free_copy(struct msg_msg *copy)
ipc/msg.c
1104
struct msg_msg *msg, *copy = NULL;
ipc/msg.c
1115
copy = prepare_copy(buf, min_t(size_t, bufsz, ns->msg_ctlmax));
ipc/msg.c
1116
if (IS_ERR(copy))
ipc/msg.c
1117
return PTR_ERR(copy);
ipc/msg.c
1125
free_copy(copy);
ipc/msg.c
1159
msg = copy_msg(msg, copy);
ipc/msg.c
1254
free_copy(copy);
kernel/audit.c
849
struct sk_buff *copy;
kernel/audit.c
869
copy = skb_copy(skb, GFP_KERNEL);
kernel/audit.c
870
if (!copy)
kernel/audit.c
872
nlh = nlmsg_hdr(copy);
kernel/audit.c
875
nlmsg_multicast(sock, copy, 0, AUDIT_NLGRP_READLOG, GFP_KERNEL);
kernel/bpf/memalloc.c
726
struct bpf_mem_alloc *copy;
kernel/bpf/memalloc.c
736
copy = kmemdup(ma, sizeof(*ma), GFP_KERNEL);
kernel/bpf/memalloc.c
737
if (!copy) {
kernel/bpf/memalloc.c
745
INIT_WORK(&copy->work, free_mem_alloc_deferred);
kernel/bpf/memalloc.c
746
queue_work(system_dfl_wq, &copy->work);
kernel/events/core.c
8237
if (frag->copy) {
kernel/events/core.c
8238
__output_custom(handle, frag->copy,
kernel/fork.c
3256
struct files_struct *old, *copy = NULL;
kernel/fork.c
3259
error = unshare_fd(CLONE_FILES, &copy);
kernel/fork.c
3260
if (error || !copy)
kernel/fork.c
3265
task->files = copy;
kernel/gcov/fs.c
417
char *copy;
kernel/gcov/fs.c
419
copy = kstrdup(path, GFP_KERNEL);
kernel/gcov/fs.c
420
if (!copy)
kernel/gcov/fs.c
422
old_ext = strrchr(copy, '.');
kernel/gcov/fs.c
426
target = kasprintf(GFP_KERNEL, "%s/%s.%s", dir, copy, ext);
kernel/gcov/fs.c
428
target = kasprintf(GFP_KERNEL, "%s.%s", copy, ext);
kernel/gcov/fs.c
429
kfree(copy);
kernel/sched/sched.h
640
# define u64_u32_load_copy(var, copy) var
kernel/sched/sched.h
641
# define u64_u32_store_copy(var, copy, val) (var = val)
kernel/sched/sched.h
643
# define u64_u32_load_copy(var, copy) \
kernel/sched/sched.h
647
__val_copy = copy; \
kernel/sched/sched.h
657
# define u64_u32_store_copy(var, copy, val) \
kernel/sched/sched.h
666
copy = __val; \
kernel/sys.c
1334
size_t copy;
kernel/sys.c
1344
copy = clamp_t(size_t, len, 1, sizeof(buf));
kernel/sys.c
1345
copy = scnprintf(buf, copy, "2.6.%u%s", v, rest);
kernel/sys.c
1346
ret = copy_to_user(release, buf, copy + 1);
kernel/trace/bpf_trace.c
717
.copy = ctx_copy,
kernel/trace/bpf_trace.c
904
struct path copy;
kernel/trace/bpf_trace.c
916
len = copy_from_kernel_nofault(&copy, path, sizeof(*path));
kernel/trace/bpf_trace.c
920
p = d_path(&copy, buf, sz);
kernel/trace/trace_events_user.c
2276
struct iov_iter copy;
kernel/trace/trace_events_user.c
2291
copy = *i;
kernel/trace/trace_events_user.c
2294
probe_func(user, &copy, tpdata, &faulted);
lib/crypto/sha3.c
272
size_t copy = min(out_len, block_size - squeeze_offset);
lib/crypto/sha3.c
274
memcpy(out, &ctx->state.bytes[squeeze_offset], copy);
lib/crypto/sha3.c
275
out += copy;
lib/crypto/sha3.c
276
out_len -= copy;
lib/crypto/sha3.c
277
squeeze_offset += copy;
lib/kunit/attributes.c
402
struct kunit_suite *copy;
lib/kunit/attributes.c
407
copy = kmemdup(suite, sizeof(*copy), GFP_KERNEL);
lib/kunit/attributes.c
408
if (!copy)
lib/kunit/attributes.c
415
kfree(copy);
lib/kunit/attributes.c
466
kfree(copy);
lib/kunit/attributes.c
471
copy->test_cases = filtered;
lib/kunit/attributes.c
473
return copy;
lib/kunit/executor.c
120
struct kunit_suite *copy;
lib/kunit/executor.c
130
copy = kmemdup(suite, sizeof(*copy), GFP_KERNEL);
lib/kunit/executor.c
131
if (!copy)
lib/kunit/executor.c
136
kfree(copy);
lib/kunit/executor.c
146
copy->test_cases = filtered;
lib/kunit/executor.c
147
return copy;
lib/kunit/executor.c
174
struct kunit_suite **copy, **copy_start, *filtered_suite, *new_filtered_suite;
lib/kunit/executor.c
182
copy = kzalloc_objs(*copy, max);
lib/kunit/executor.c
183
if (!copy) { /* won't be able to run anything, return an empty set */
lib/kunit/executor.c
186
copy_start = copy;
lib/kunit/executor.c
248
*copy++ = filtered_suite;
lib/kunit/executor.c
251
filtered.end = copy;
lib/kunit/executor.c
255
for (suites = copy_start; suites < copy; suites++) {
lib/maple_tree.c
6487
unsigned long copy = *index;
lib/maple_tree.c
6520
if (MT_WARN_ON(mt, (*index) && ((*index) <= copy)))
lib/maple_tree.c
6522
*index, copy);
lib/rbtree.c
430
.copy = dummy_copy,
lib/test_bitmap.c
1002
bitmap_clear(copy, s, e - s);
lib/test_bitmap.c
1006
expect_eq_bitmap(tmp, copy, 500);
lib/test_bitmap.c
780
DECLARE_BITMAP(copy, 500);
lib/test_bitmap.c
793
bitmap_zero(copy, 500);
lib/test_bitmap.c
796
bitmap_set(copy, bit, 1);
lib/test_bitmap.c
798
expect_eq_bitmap(orig, copy, 500);
lib/test_bitmap.c
805
DECLARE_BITMAP(copy, 500);
lib/test_bitmap.c
809
bitmap_zero(copy, 500);
lib/test_bitmap.c
819
bitmap_set(copy, bit, 1);
lib/test_bitmap.c
821
expect_eq_bitmap(orig, copy, 500);
lib/test_bitmap.c
827
DECLARE_BITMAP(copy, 500);
lib/test_bitmap.c
842
bitmap_zero(copy, 500);
lib/test_bitmap.c
846
bitmap_set(copy, bit, 1);
lib/test_bitmap.c
850
expect_eq_bitmap(tmp, copy, 500);
lib/test_bitmap.c
857
DECLARE_BITMAP(copy, 500);
lib/test_bitmap.c
861
bitmap_fill(copy, 500);
lib/test_bitmap.c
871
bitmap_clear(copy, bit, 1);
lib/test_bitmap.c
873
expect_eq_bitmap(orig, copy, 500);
lib/test_bitmap.c
879
DECLARE_BITMAP(copy, 500);
lib/test_bitmap.c
894
bitmap_fill(copy, 500);
lib/test_bitmap.c
898
bitmap_clear(copy, bit, 1);
lib/test_bitmap.c
902
expect_eq_bitmap(tmp, copy, 500);
lib/test_bitmap.c
909
DECLARE_BITMAP(copy, 500);
lib/test_bitmap.c
913
bitmap_zero(copy, 500);
lib/test_bitmap.c
923
bitmap_set(copy, s, e-s);
lib/test_bitmap.c
925
expect_eq_bitmap(orig, copy, 500);
lib/test_bitmap.c
931
DECLARE_BITMAP(copy, 500);
lib/test_bitmap.c
935
bitmap_fill(copy, 500);
lib/test_bitmap.c
945
bitmap_clear(copy, s, e-s);
lib/test_bitmap.c
947
expect_eq_bitmap(orig, copy, 500);
lib/test_bitmap.c
953
DECLARE_BITMAP(copy, 500);
lib/test_bitmap.c
968
bitmap_zero(copy, 500);
lib/test_bitmap.c
972
bitmap_set(copy, s, e - s);
lib/test_bitmap.c
976
expect_eq_bitmap(tmp, copy, 500);
lib/test_bitmap.c
983
DECLARE_BITMAP(copy, 500);
lib/test_bitmap.c
998
bitmap_fill(copy, 500);
lib/tests/fortify_kunit.c
1030
char *copy;
lib/tests/fortify_kunit.c
1036
copy = kmemdup(src, len, GFP_KERNEL);
lib/tests/fortify_kunit.c
1037
KUNIT_EXPECT_NOT_NULL(test, copy);
lib/tests/fortify_kunit.c
1039
kfree(copy);
lib/tests/fortify_kunit.c
1042
copy = kmemdup(src, len - 1, GFP_KERNEL);
lib/tests/fortify_kunit.c
1043
KUNIT_EXPECT_NOT_NULL(test, copy);
lib/tests/fortify_kunit.c
1045
kfree(copy);
lib/tests/fortify_kunit.c
1048
copy = kmemdup(src, 1, GFP_KERNEL);
lib/tests/fortify_kunit.c
1049
KUNIT_EXPECT_NOT_NULL(test, copy);
lib/tests/fortify_kunit.c
1051
kfree(copy);
lib/tests/fortify_kunit.c
1054
copy = kmemdup(src, len + 1, GFP_KERNEL);
lib/tests/fortify_kunit.c
1055
KUNIT_EXPECT_PTR_EQ(test, copy, ZERO_SIZE_PTR);
lib/tests/fortify_kunit.c
1057
kfree(copy);
lib/tests/fortify_kunit.c
1060
copy = kmemdup(src, len * 2, GFP_KERNEL);
lib/tests/fortify_kunit.c
1061
KUNIT_EXPECT_PTR_EQ(test, copy, ZERO_SIZE_PTR);
lib/tests/fortify_kunit.c
1063
kfree(copy);
lib/tests/fortify_kunit.c
1066
copy = kmemdup(src + 1, len, GFP_KERNEL);
lib/tests/fortify_kunit.c
1067
KUNIT_EXPECT_PTR_EQ(test, copy, ZERO_SIZE_PTR);
lib/tests/fortify_kunit.c
1069
kfree(copy);
lib/vsprintf.c
2896
int copy = read;
lib/vsprintf.c
2897
if (copy > end - str)
lib/vsprintf.c
2898
copy = end - str;
lib/vsprintf.c
2899
memcpy(str, old_fmt, copy);
lib/vsprintf.c
3328
int copy = read;
lib/vsprintf.c
3329
if (copy > end - str)
lib/vsprintf.c
3330
copy = end - str;
lib/vsprintf.c
3331
memcpy(str, old_fmt, copy);
lib/vsprintf.c
3376
int copy, len;
lib/vsprintf.c
3393
len = copy = strlen(args);
lib/vsprintf.c
3394
if (copy > end - str)
lib/vsprintf.c
3395
copy = end - str;
lib/vsprintf.c
3396
memcpy(str, args, copy);
lib/zlib_inflate/inflate.c
123
unsigned copy, dist;
lib/zlib_inflate/inflate.c
128
copy = out - strm->avail_out;
lib/zlib_inflate/inflate.c
129
if (copy >= state->wsize) {
lib/zlib_inflate/inflate.c
136
if (dist > copy) dist = copy;
lib/zlib_inflate/inflate.c
137
memcpy(state->window + state->write, strm->next_out - copy, dist);
lib/zlib_inflate/inflate.c
138
copy -= dist;
lib/zlib_inflate/inflate.c
139
if (copy) {
lib/zlib_inflate/inflate.c
140
memcpy(state->window, strm->next_out - copy, copy);
lib/zlib_inflate/inflate.c
141
state->write = copy;
lib/zlib_inflate/inflate.c
340
unsigned copy; /* number of stored or match bytes to copy */
lib/zlib_inflate/inflate.c
451
copy = state->length;
lib/zlib_inflate/inflate.c
452
if (copy) {
lib/zlib_inflate/inflate.c
453
if (copy > have) copy = have;
lib/zlib_inflate/inflate.c
454
if (copy > left) copy = left;
lib/zlib_inflate/inflate.c
455
if (copy == 0) goto inf_leave;
lib/zlib_inflate/inflate.c
456
memcpy(put, next, copy);
lib/zlib_inflate/inflate.c
457
have -= copy;
lib/zlib_inflate/inflate.c
458
next += copy;
lib/zlib_inflate/inflate.c
459
left -= copy;
lib/zlib_inflate/inflate.c
460
put += copy;
lib/zlib_inflate/inflate.c
461
state->length -= copy;
lib/zlib_inflate/inflate.c
527
copy = 3 + BITS(2);
lib/zlib_inflate/inflate.c
534
copy = 3 + BITS(3);
lib/zlib_inflate/inflate.c
541
copy = 11 + BITS(7);
lib/zlib_inflate/inflate.c
544
if (state->have + copy > state->nlen + state->ndist) {
lib/zlib_inflate/inflate.c
549
while (copy--)
lib/zlib_inflate/inflate.c
675
copy = out - left;
lib/zlib_inflate/inflate.c
676
if (state->offset > copy) { /* copy from window */
lib/zlib_inflate/inflate.c
677
copy = state->offset - copy;
lib/zlib_inflate/inflate.c
678
if (copy > state->write) {
lib/zlib_inflate/inflate.c
679
copy -= state->write;
lib/zlib_inflate/inflate.c
680
from = state->window + (state->wsize - copy);
lib/zlib_inflate/inflate.c
683
from = state->window + (state->write - copy);
lib/zlib_inflate/inflate.c
684
if (copy > state->length) copy = state->length;
lib/zlib_inflate/inflate.c
688
copy = state->length;
lib/zlib_inflate/inflate.c
690
if (copy > left) copy = left;
lib/zlib_inflate/inflate.c
691
left -= copy;
lib/zlib_inflate/inflate.c
692
state->length -= copy;
lib/zlib_inflate/inflate.c
695
} while (--copy);
mm/mempolicy.c
1697
unsigned long copy = ALIGN(maxnode-1, 64) / 8;
mm/mempolicy.c
1704
if (copy > nbytes) {
mm/mempolicy.c
1705
if (copy > PAGE_SIZE)
mm/mempolicy.c
1707
if (clear_user((char __user *)mask + nbytes, copy - nbytes))
mm/mempolicy.c
1709
copy = nbytes;
mm/mempolicy.c
1717
return copy_to_user(mask, nodes_addr(*nodes), copy) ? -EFAULT : 0;
mm/process_vm_access.c
36
size_t copy = PAGE_SIZE - offset;
mm/process_vm_access.c
39
if (copy > len)
mm/process_vm_access.c
40
copy = len;
mm/process_vm_access.c
43
copied = copy_page_from_iter(page, offset, copy, iter);
mm/process_vm_access.c
45
copied = copy_page_to_iter(page, offset, copy, iter);
mm/process_vm_access.c
48
if (copied < copy && iov_iter_count(iter))
net/appletalk/ddp.c
1000
if ((len -= copy) == 0)
net/appletalk/ddp.c
1002
offset += copy;
net/appletalk/ddp.c
952
int i, copy;
net/appletalk/ddp.c
955
if ((copy = start - offset) > 0) {
net/appletalk/ddp.c
956
if (copy > len)
net/appletalk/ddp.c
957
copy = len;
net/appletalk/ddp.c
958
sum = atalk_sum_partial(skb->data + offset, copy, sum);
net/appletalk/ddp.c
959
if ((len -= copy) == 0)
net/appletalk/ddp.c
962
offset += copy;
net/appletalk/ddp.c
972
if ((copy = end - offset) > 0) {
net/appletalk/ddp.c
975
if (copy > len)
net/appletalk/ddp.c
976
copy = len;
net/appletalk/ddp.c
979
offset - start, copy, sum);
net/appletalk/ddp.c
982
if (!(len -= copy))
net/appletalk/ddp.c
984
offset += copy;
net/appletalk/ddp.c
995
if ((copy = end - offset) > 0) {
net/appletalk/ddp.c
996
if (copy > len)
net/appletalk/ddp.c
997
copy = len;
net/appletalk/ddp.c
999
copy, sum);
net/ax25/af_ax25.c
262
struct sk_buff *copy;
net/ax25/af_ax25.c
271
if ((copy = skb_clone(skb, GFP_ATOMIC)) == NULL)
net/ax25/af_ax25.c
273
if (sock_queue_rcv_skb(s->sk, copy) != 0)
net/ax25/af_ax25.c
274
kfree_skb(copy);
net/core/datagram.c
395
int i, copy = start - offset, start_off = offset, n;
net/core/datagram.c
399
if (copy > 0) {
net/core/datagram.c
400
if (copy > len)
net/core/datagram.c
401
copy = len;
net/core/datagram.c
403
skb->data + offset, copy, data, to);
net/core/datagram.c
405
if (n != copy)
net/core/datagram.c
407
if ((len -= copy) == 0)
net/core/datagram.c
422
if ((copy = end - offset) > 0) {
net/core/datagram.c
427
if (copy > len)
net/core/datagram.c
428
copy = len;
net/core/datagram.c
433
copy, p, p_off, p_len, copied) {
net/core/datagram.c
441
if (n != copy)
net/core/datagram.c
443
if (!(len -= copy))
net/core/datagram.c
455
if ((copy = end - offset) > 0) {
net/core/datagram.c
456
if (copy > len)
net/core/datagram.c
457
copy = len;
net/core/datagram.c
459
to, copy, fault_short, cb, data))
net/core/datagram.c
461
if ((len -= copy) == 0)
net/core/datagram.c
463
offset += copy;
net/core/datagram.c
554
int i, copy = start - offset;
net/core/datagram.c
558
if (copy > 0) {
net/core/datagram.c
559
if (copy > len)
net/core/datagram.c
560
copy = len;
net/core/datagram.c
561
if (copy_from_iter(skb->data + offset, copy, from) != copy)
net/core/datagram.c
563
if ((len -= copy) == 0)
net/core/datagram.c
565
offset += copy;
net/core/datagram.c
576
if ((copy = end - offset) > 0) {
net/core/datagram.c
579
if (copy > len)
net/core/datagram.c
580
copy = len;
net/core/datagram.c
583
copy, from);
net/core/datagram.c
584
if (copied != copy)
net/core/datagram.c
587
if (!(len -= copy))
net/core/datagram.c
589
offset += copy;
net/core/datagram.c
600
if ((copy = end - offset) > 0) {
net/core/datagram.c
601
if (copy > len)
net/core/datagram.c
602
copy = len;
net/core/datagram.c
605
from, copy))
net/core/datagram.c
607
if ((len -= copy) == 0)
net/core/datagram.c
609
offset += copy;
net/core/datagram.c
786
int copy = min_t(int, skb_headlen(skb), iov_iter_count(from));
net/core/datagram.c
789
if (skb_copy_datagram_from_iter(skb, 0, from, copy))
net/core/dev.c
2854
bool active = false, copy = false;
net/core/dev.c
2899
copy = true;
net/core/dev.c
2916
map = copy ? xmap_dereference(dev_maps->attr_map[tci]) : NULL;
net/core/dev.c
2962
if (copy)
net/core/dev.c
2979
if (copy) {
net/core/dev.c
3017
copy ? old_dev_maps : NULL,
net/core/dev.c
3038
map = copy ?
net/core/filter.c
2664
u32 len = 0, offset = 0, copy = 0, poffset = 0, bytes = end - start;
net/core/filter.c
2691
if (!test_bit(i, msg->sg.copy) && bytes_sg_total <= len)
net/core/filter.c
2705
copy += sk_msg_elem(msg, i)->length;
net/core/filter.c
2707
if (bytes_sg_total <= copy)
net/core/filter.c
2712
if (unlikely(bytes_sg_total > copy))
net/core/filter.c
2716
get_order(copy));
net/core/filter.c
2736
sg_set_page(&msg->sg.data[first_sge], page, copy, 0);
net/core/filter.c
2792
u32 new, i = 0, l = 0, space, copy = 0, offset = 0;
net/core/filter.c
2826
copy = msg->sg.data[i].length;
net/core/filter.c
2829
get_order(copy + len));
net/core/filter.c
2833
if (copy) {
net/core/filter.c
2914
__clear_bit(new, msg->sg.copy);
net/core/filter.c
2915
sg_set_page(&msg->sg.data[new], page, len + copy, 0);
net/core/pktgen.c
1031
size_t copy = min_t(size_t, count + 1, 1024);
net/core/pktgen.c
1032
char *tp = strndup_user(user_buffer, copy);
net/core/skbuff.c
2050
u32 copy, done = 0;
net/core/skbuff.c
2058
copy = min_t(u32, psize - d_off, p_len - done);
net/core/skbuff.c
2060
vaddr + p_off + done, copy);
net/core/skbuff.c
2061
done += copy;
net/core/skbuff.c
2062
d_off += copy;
net/core/skbuff.c
3017
int i, copy;
net/core/skbuff.c
3023
if ((copy = start - offset) > 0) {
net/core/skbuff.c
3024
if (copy > len)
net/core/skbuff.c
3025
copy = len;
net/core/skbuff.c
3026
skb_copy_from_linear_data_offset(skb, offset, to, copy);
net/core/skbuff.c
3027
if ((len -= copy) == 0)
net/core/skbuff.c
3029
offset += copy;
net/core/skbuff.c
3030
to += copy;
net/core/skbuff.c
3043
if ((copy = end - offset) > 0) {
net/core/skbuff.c
3048
if (copy > len)
net/core/skbuff.c
3049
copy = len;
net/core/skbuff.c
3053
copy, p, p_off, p_len, copied) {
net/core/skbuff.c
3059
if ((len -= copy) == 0)
net/core/skbuff.c
3061
offset += copy;
net/core/skbuff.c
3062
to += copy;
net/core/skbuff.c
3073
if ((copy = end - offset) > 0) {
net/core/skbuff.c
3074
if (copy > len)
net/core/skbuff.c
3075
copy = len;
net/core/skbuff.c
3076
if (skb_copy_bits(frag_iter, offset - start, to, copy))
net/core/skbuff.c
3078
if ((len -= copy) == 0)
net/core/skbuff.c
3080
offset += copy;
net/core/skbuff.c
3081
to += copy;
net/core/skbuff.c
3446
int i, copy;
net/core/skbuff.c
3451
if ((copy = start - offset) > 0) {
net/core/skbuff.c
3452
if (copy > len)
net/core/skbuff.c
3453
copy = len;
net/core/skbuff.c
3454
skb_copy_to_linear_data_offset(skb, offset, from, copy);
net/core/skbuff.c
3455
if ((len -= copy) == 0)
net/core/skbuff.c
3457
offset += copy;
net/core/skbuff.c
3458
from += copy;
net/core/skbuff.c
3471
if ((copy = end - offset) > 0) {
net/core/skbuff.c
3476
if (copy > len)
net/core/skbuff.c
3477
copy = len;
net/core/skbuff.c
3481
copy, p, p_off, p_len, copied) {
net/core/skbuff.c
3487
if ((len -= copy) == 0)
net/core/skbuff.c
3489
offset += copy;
net/core/skbuff.c
3490
from += copy;
net/core/skbuff.c
3501
if ((copy = end - offset) > 0) {
net/core/skbuff.c
3502
if (copy > len)
net/core/skbuff.c
3503
copy = len;
net/core/skbuff.c
3505
from, copy))
net/core/skbuff.c
3507
if ((len -= copy) == 0)
net/core/skbuff.c
3509
offset += copy;
net/core/skbuff.c
3510
from += copy;
net/core/skbuff.c
3526
int i, copy = start - offset;
net/core/skbuff.c
3531
if (copy > 0) {
net/core/skbuff.c
3532
if (copy > len)
net/core/skbuff.c
3533
copy = len;
net/core/skbuff.c
3534
csum = csum_partial(skb->data + offset, copy, csum);
net/core/skbuff.c
3535
if ((len -= copy) == 0)
net/core/skbuff.c
3537
offset += copy;
net/core/skbuff.c
3538
pos = copy;
net/core/skbuff.c
3551
if ((copy = end - offset) > 0) {
net/core/skbuff.c
3557
if (copy > len)
net/core/skbuff.c
3558
copy = len;
net/core/skbuff.c
3562
copy, p, p_off, p_len, copied) {
net/core/skbuff.c
3570
if (!(len -= copy))
net/core/skbuff.c
3572
offset += copy;
net/core/skbuff.c
3583
if ((copy = end - offset) > 0) {
net/core/skbuff.c
3585
if (copy > len)
net/core/skbuff.c
3586
copy = len;
net/core/skbuff.c
3587
csum2 = skb_checksum(frag_iter, offset - start, copy,
net/core/skbuff.c
3590
if ((len -= copy) == 0)
net/core/skbuff.c
3592
offset += copy;
net/core/skbuff.c
3593
pos += copy;
net/core/skbuff.c
3609
int i, copy = start - offset;
net/core/skbuff.c
3615
if (copy > 0) {
net/core/skbuff.c
3616
if (copy > len)
net/core/skbuff.c
3617
copy = len;
net/core/skbuff.c
3619
copy);
net/core/skbuff.c
3620
if ((len -= copy) == 0)
net/core/skbuff.c
3622
offset += copy;
net/core/skbuff.c
3623
to += copy;
net/core/skbuff.c
3624
pos = copy;
net/core/skbuff.c
3636
if ((copy = end - offset) > 0) {
net/core/skbuff.c
3643
if (copy > len)
net/core/skbuff.c
3644
copy = len;
net/core/skbuff.c
3648
copy, p, p_off, p_len, copied) {
net/core/skbuff.c
3658
if (!(len -= copy))
net/core/skbuff.c
3660
offset += copy;
net/core/skbuff.c
3661
to += copy;
net/core/skbuff.c
3673
if ((copy = end - offset) > 0) {
net/core/skbuff.c
3674
if (copy > len)
net/core/skbuff.c
3675
copy = len;
net/core/skbuff.c
3678
to, copy);
net/core/skbuff.c
3680
if ((len -= copy) == 0)
net/core/skbuff.c
3682
offset += copy;
net/core/skbuff.c
3683
to += copy;
net/core/skbuff.c
3684
pos += copy;
net/core/skbuff.c
3697
int i, copy = start - offset;
net/core/skbuff.c
3700
if (copy > 0) {
net/core/skbuff.c
3701
copy = min(copy, len);
net/core/skbuff.c
3702
crc = crc32c(crc, skb->data + offset, copy);
net/core/skbuff.c
3703
len -= copy;
net/core/skbuff.c
3706
offset += copy;
net/core/skbuff.c
3719
copy = end - offset;
net/core/skbuff.c
3720
if (copy > 0) {
net/core/skbuff.c
3725
copy = min(copy, len);
net/core/skbuff.c
3728
copy, p, p_off, p_len, copied) {
net/core/skbuff.c
3733
len -= copy;
net/core/skbuff.c
3736
offset += copy;
net/core/skbuff.c
3747
copy = end - offset;
net/core/skbuff.c
3748
if (copy > 0) {
net/core/skbuff.c
3749
copy = min(copy, len);
net/core/skbuff.c
3750
crc = skb_crc32c(frag_iter, offset - start, copy, crc);
net/core/skbuff.c
3751
len -= copy;
net/core/skbuff.c
3754
offset += copy;
net/core/skbuff.c
5222
int i, copy = start - offset;
net/core/skbuff.c
5229
if (copy > 0) {
net/core/skbuff.c
5230
if (copy > len)
net/core/skbuff.c
5231
copy = len;
net/core/skbuff.c
5232
sg_set_buf(sg, skb->data + offset, copy);
net/core/skbuff.c
5234
if ((len -= copy) == 0)
net/core/skbuff.c
5236
offset += copy;
net/core/skbuff.c
5245
if ((copy = end - offset) > 0) {
net/core/skbuff.c
5250
if (copy > len)
net/core/skbuff.c
5251
copy = len;
net/core/skbuff.c
5252
sg_set_page(&sg[elt], skb_frag_page(frag), copy,
net/core/skbuff.c
5255
if (!(len -= copy))
net/core/skbuff.c
5257
offset += copy;
net/core/skbuff.c
5268
if ((copy = end - offset) > 0) {
net/core/skbuff.c
5272
if (copy > len)
net/core/skbuff.c
5273
copy = len;
net/core/skbuff.c
5275
copy, recursion_level + 1);
net/core/skbuff.c
5279
if ((len -= copy) == 0)
net/core/skbuff.c
5281
offset += copy;
net/core/skmsg.c
372
u32 copy, buf_size, copied = 0;
net/core/skmsg.c
388
copy = (buf_size > bytes) ? bytes : buf_size;
net/core/skmsg.c
390
msg->sg.copybreak += copy;
net/core/skmsg.c
392
ret = copy_from_iter_nocache(to, copy, from);
net/core/skmsg.c
394
ret = copy_from_iter(to, copy, from);
net/core/skmsg.c
395
if (ret != copy) {
net/core/skmsg.c
399
bytes -= copy;
net/core/skmsg.c
400
copied += copy;
net/core/skmsg.c
435
int copy;
net/core/skmsg.c
438
copy = sge->length;
net/core/skmsg.c
440
if (copied + copy > len)
net/core/skmsg.c
441
copy = len - copied;
net/core/skmsg.c
442
if (copy)
net/core/skmsg.c
443
copy = copy_page_to_iter(page, sge->offset, copy, iter);
net/core/skmsg.c
444
if (!copy) {
net/core/skmsg.c
449
copied += copy;
net/core/skmsg.c
451
*copied_from_self += copy;
net/core/skmsg.c
454
sge->offset += copy;
net/core/skmsg.c
455
sge->length -= copy;
net/core/skmsg.c
457
sk_mem_uncharge(sk, copy);
net/core/skmsg.c
458
atomic_sub(copy, &sk->sk_rmem_alloc);
net/core/skmsg.c
460
msg_rx->sg.size -= copy;
net/core/skmsg.c
461
sk_psock_msg_len_add(psock, -copy);
net/core/skmsg.c
472
if (copy != sge->length)
net/ipv4/ip_output.c
1073
copy = mtu - skb->len;
net/ipv4/ip_output.c
1074
if (copy < length)
net/ipv4/ip_output.c
1075
copy = maxfraglen - skb->len;
net/ipv4/ip_output.c
1076
if (copy <= 0) {
net/ipv4/ip_output.c
1167
copy = datalen - transhdrlen - fraggap - pagedlen;
net/ipv4/ip_output.c
1171
if (copy > 0 &&
net/ipv4/ip_output.c
1174
copy, fraggap, skb) < 0) {
net/ipv4/ip_output.c
1179
copy = 0;
net/ipv4/ip_output.c
1182
offset += copy;
net/ipv4/ip_output.c
1183
length -= copy + transhdrlen;
net/ipv4/ip_output.c
1210
if (copy > length)
net/ipv4/ip_output.c
1211
copy = length;
net/ipv4/ip_output.c
1214
skb_tailroom(skb) >= copy) {
net/ipv4/ip_output.c
1219
from, skb_put(skb, copy),
net/ipv4/ip_output.c
1220
offset, copy, off, skb) < 0) {
net/ipv4/ip_output.c
1229
if (WARN_ON_ONCE(copy > msg->msg_iter.count))
net/ipv4/ip_output.c
1232
err = skb_splice_from_iter(skb, &msg->msg_iter, copy);
net/ipv4/ip_output.c
1235
copy = err;
net/ipv4/ip_output.c
1236
wmem_alloc_delta += copy;
net/ipv4/ip_output.c
1256
copy = min_t(int, copy, pfrag->size - pfrag->offset);
net/ipv4/ip_output.c
1260
offset, copy, skb->len, skb) < 0)
net/ipv4/ip_output.c
1263
pfrag->offset += copy;
net/ipv4/ip_output.c
1264
skb_frag_size_add(&skb_shinfo(skb)->frags[i - 1], copy);
net/ipv4/ip_output.c
1265
skb_len_add(skb, copy);
net/ipv4/ip_output.c
1266
wmem_alloc_delta += copy;
net/ipv4/ip_output.c
1268
err = skb_zerocopy_iter_dgram(skb, from, copy);
net/ipv4/ip_output.c
1272
offset += copy;
net/ipv4/ip_output.c
1273
length -= copy;
net/ipv4/ip_output.c
967
int copy;
net/ipv4/raw.c
457
int copy = min(rfv->hlen - offset, len);
net/ipv4/raw.c
460
memcpy(to, rfv->hdr.c + offset, copy);
net/ipv4/raw.c
465
to, copy),
net/ipv4/raw.c
469
offset += copy;
net/ipv4/raw.c
470
to += copy;
net/ipv4/raw.c
471
len -= copy;
net/ipv4/tcp.c
1040
int tcp_wmem_schedule(struct sock *sk, int copy)
net/ipv4/tcp.c
1044
if (likely(sk_wmem_schedule(sk, copy)))
net/ipv4/tcp.c
1045
return copy;
net/ipv4/tcp.c
1053
sk_forced_mem_schedule(sk, min(left, copy));
net/ipv4/tcp.c
1054
return min(copy, sk->sk_forward_alloc);
net/ipv4/tcp.c
1249
int copy = 0;
net/ipv4/tcp.c
1253
copy = size_goal - skb->len;
net/ipv4/tcp.c
1257
if (copy <= 0 || !tcp_skb_can_collapse_to(skb)) {
net/ipv4/tcp.c
1281
copy = size_goal;
net/ipv4/tcp.c
1292
if (copy > msg_data_left(msg))
net/ipv4/tcp.c
1293
copy = msg_data_left(msg);
net/ipv4/tcp.c
1312
copy = min_t(int, copy, pfrag->size - pfrag->offset);
net/ipv4/tcp.c
1320
copy = tcp_wmem_schedule(sk, copy);
net/ipv4/tcp.c
1321
if (!copy)
net/ipv4/tcp.c
1327
copy);
net/ipv4/tcp.c
1333
skb_frag_size_add(&skb_shinfo(skb)->frags[i - 1], copy);
net/ipv4/tcp.c
1336
pfrag->offset, copy);
net/ipv4/tcp.c
1339
pfrag->offset += copy;
net/ipv4/tcp.c
1348
copy = tcp_wmem_schedule(sk, copy);
net/ipv4/tcp.c
1349
if (!copy)
net/ipv4/tcp.c
1353
err = skb_zerocopy_iter_stream(sk, skb, msg, copy, uarg,
net/ipv4/tcp.c
1361
copy = err;
net/ipv4/tcp.c
1366
copy = tcp_wmem_schedule(sk, copy);
net/ipv4/tcp.c
1367
if (!copy)
net/ipv4/tcp.c
1370
err = skb_splice_from_iter(skb, &msg->msg_iter, copy);
net/ipv4/tcp.c
1378
copy = err;
net/ipv4/tcp.c
1383
sk_wmem_queued_add(sk, copy);
net/ipv4/tcp.c
1384
sk_mem_charge(sk, copy);
net/ipv4/tcp.c
1390
WRITE_ONCE(tp->write_seq, tp->write_seq + copy);
net/ipv4/tcp.c
1391
TCP_SKB_CB(skb)->end_seq += copy;
net/ipv4/tcp.c
1394
copied += copy;
net/ipv4/tcp.c
2538
int i, copy, n;
net/ipv4/tcp.c
2553
copy = start - offset;
net/ipv4/tcp.c
2554
if (copy > 0) {
net/ipv4/tcp.c
2555
copy = min(copy, remaining_len);
net/ipv4/tcp.c
2557
n = copy_to_iter(skb->data + offset, copy,
net/ipv4/tcp.c
2559
if (n != copy) {
net/ipv4/tcp.c
2564
offset += copy;
net/ipv4/tcp.c
2565
remaining_len -= copy;
net/ipv4/tcp.c
2571
dmabuf_cmsg.frag_size = copy;
net/ipv4/tcp.c
2579
sent += copy;
net/ipv4/tcp.c
2614
copy = end - offset;
net/ipv4/tcp.c
2616
if (copy > 0) {
net/ipv4/tcp.c
2617
copy = min(copy, remaining_len);
net/ipv4/tcp.c
2623
dmabuf_cmsg.frag_size = copy;
net/ipv4/tcp.c
2633
offset += copy;
net/ipv4/tcp.c
2634
remaining_len -= copy;
net/ipv4/tcp.c
2646
sent += copy;
net/ipv4/tcp_bpf.c
538
u32 copy, osize;
net/ipv4/tcp_bpf.c
545
copy = msg_data_left(msg);
net/ipv4/tcp_bpf.c
556
err = sk_msg_alloc(sk, msg_tx, msg_tx->sg.size + copy, msg_tx->sg.end - 1);
net/ipv4/tcp_bpf.c
561
copy = msg_tx->sg.size - osize;
net/ipv4/tcp_bpf.c
565
copy);
net/ipv4/tcp_input.c
5848
int copy = min_t(int, SKB_MAX_ORDER(0, 0), end - start);
net/ipv4/tcp_input.c
5851
nskb = alloc_skb(copy, GFP_ATOMIC);
net/ipv4/tcp_input.c
5866
while (copy > 0) {
net/ipv4/tcp_input.c
5872
size = min(copy, size);
net/ipv4/tcp_input.c
5876
copy -= size;
net/ipv4/tcp_output.c
2670
int copy, len;
net/ipv4/tcp_output.c
2755
copy = min_t(int, skb->len, probe_size - len);
net/ipv4/tcp_output.c
2757
if (skb->len <= copy) {
net/ipv4/tcp_output.c
2762
__pskb_trim_head(skb, copy);
net/ipv4/tcp_output.c
2764
TCP_SKB_CB(skb)->seq += copy;
net/ipv4/tcp_output.c
2767
len += copy;
net/ipv6/ip6_output.c
1436
int copy;
net/ipv6/ip6_output.c
1585
copy = (cork->length <= mtu ? mtu : maxfraglen) - skb->len;
net/ipv6/ip6_output.c
1586
if (copy < length)
net/ipv6/ip6_output.c
1587
copy = maxfraglen - skb->len;
net/ipv6/ip6_output.c
1589
if (copy <= 0) {
net/ipv6/ip6_output.c
1654
copy = datalen - transhdrlen - fraggap - pagedlen;
net/ipv6/ip6_output.c
1658
if (copy < 0 && !(flags & MSG_SPLICE_PAGES)) {
net/ipv6/ip6_output.c
1703
if (copy > 0 &&
net/ipv6/ip6_output.c
1706
copy, fraggap, skb) < 0) {
net/ipv6/ip6_output.c
1711
copy = 0;
net/ipv6/ip6_output.c
1714
offset += copy;
net/ipv6/ip6_output.c
1715
length -= copy + transhdrlen;
net/ipv6/ip6_output.c
1742
if (copy > length)
net/ipv6/ip6_output.c
1743
copy = length;
net/ipv6/ip6_output.c
1746
skb_tailroom(skb) >= copy) {
net/ipv6/ip6_output.c
1751
from, skb_put(skb, copy),
net/ipv6/ip6_output.c
1752
offset, copy, off, skb) < 0) {
net/ipv6/ip6_output.c
1761
if (WARN_ON_ONCE(copy > msg->msg_iter.count))
net/ipv6/ip6_output.c
1764
err = skb_splice_from_iter(skb, &msg->msg_iter, copy);
net/ipv6/ip6_output.c
1767
copy = err;
net/ipv6/ip6_output.c
1768
wmem_alloc_delta += copy;
net/ipv6/ip6_output.c
1788
copy = min_t(int, copy, pfrag->size - pfrag->offset);
net/ipv6/ip6_output.c
1792
offset, copy, skb->len, skb) < 0)
net/ipv6/ip6_output.c
1795
pfrag->offset += copy;
net/ipv6/ip6_output.c
1796
skb_frag_size_add(&skb_shinfo(skb)->frags[i - 1], copy);
net/ipv6/ip6_output.c
1797
skb->len += copy;
net/ipv6/ip6_output.c
1798
skb->data_len += copy;
net/ipv6/ip6_output.c
1799
skb->truesize += copy;
net/ipv6/ip6_output.c
1800
wmem_alloc_delta += copy;
net/ipv6/ip6_output.c
1802
err = skb_zerocopy_iter_dgram(skb, from, copy);
net/ipv6/ip6_output.c
1806
offset += copy;
net/ipv6/ip6_output.c
1807
length -= copy;
net/ipv6/raw.c
716
int copy = min(rfv->hlen - offset, len);
net/ipv6/raw.c
719
memcpy(to, rfv->c + offset, copy);
net/ipv6/raw.c
724
to, copy),
net/ipv6/raw.c
728
offset += copy;
net/ipv6/raw.c
729
to += copy;
net/ipv6/raw.c
730
len -= copy;
net/iucv/iucv.c
1103
size_t copy;
net/iucv/iucv.c
1116
copy = min_t(size_t, size, array->length);
net/iucv/iucv.c
1117
memcpy(dma32_to_virt(array->address), rmmsg, copy);
net/iucv/iucv.c
1118
rmmsg += copy;
net/iucv/iucv.c
1119
size -= copy;
net/kcm/kcmsock.c
753
size_t copy, copied = 0;
net/kcm/kcmsock.c
836
copy = msg_data_left(msg);
net/kcm/kcmsock.c
837
if (!sk_wmem_schedule(sk, copy))
net/kcm/kcmsock.c
840
err = skb_splice_from_iter(skb, &msg->msg_iter, copy);
net/kcm/kcmsock.c
847
copy = err;
net/kcm/kcmsock.c
849
sk_wmem_queued_add(sk, copy);
net/kcm/kcmsock.c
850
sk_mem_charge(sk, copy);
net/kcm/kcmsock.c
853
head->truesize += copy;
net/kcm/kcmsock.c
855
copy = min_t(int, msg_data_left(msg),
net/kcm/kcmsock.c
857
if (!sk_wmem_schedule(sk, copy))
net/kcm/kcmsock.c
863
copy);
net/kcm/kcmsock.c
870
&skb_shinfo(skb)->frags[i - 1], copy);
net/kcm/kcmsock.c
873
pfrag->offset, copy);
net/kcm/kcmsock.c
877
pfrag->offset += copy;
net/kcm/kcmsock.c
880
copied += copy;
net/kcm/kcmsock.c
882
head->len += copy;
net/kcm/kcmsock.c
883
head->data_len += copy;
net/mac80211/mesh_pathtbl.c
193
bool copy)
net/mac80211/mesh_pathtbl.c
224
if (copy)
net/mac80211/mesh_pathtbl.c
234
if (!copy)
net/mac80211/mesh_pathtbl.c
969
bool copy = false;
net/mac80211/mesh_pathtbl.c
977
mesh_path_move_to_queue(gate, from_mpath, copy);
net/mac80211/mesh_pathtbl.c
979
copy = true;
net/mac80211/tx.c
5784
struct sk_buff *copy;
net/mac80211/tx.c
5800
copy = skb_copy(bcn, GFP_ATOMIC);
net/mac80211/tx.c
5801
if (!copy)
net/mac80211/tx.c
5804
ieee80211_tx_monitor(hw_to_local(hw), copy, 1, NULL);
net/mptcp/protocol.c
1282
size_t copy;
net/mptcp/protocol.c
1299
copy = info->size_goal;
net/mptcp/protocol.c
1302
if (skb && copy > skb->len) {
net/mptcp/protocol.c
1323
copy -= skb->len;
net/mptcp/protocol.c
1336
copy = mptcp_check_allowed_size(msk, ssk, data_seq, copy);
net/mptcp/protocol.c
1337
if (copy == 0) {
net/mptcp/protocol.c
1352
copy = 1;
net/mptcp/protocol.c
1355
copy = min_t(size_t, copy, info->limit - info->sent);
net/mptcp/protocol.c
1356
if (!sk_wmem_schedule(ssk, copy)) {
net/mptcp/protocol.c
1362
skb_frag_size_add(&skb_shinfo(skb)->frags[i - 1], copy);
net/mptcp/protocol.c
1365
skb_fill_page_desc(skb, i, dfrag->page, offset, copy);
net/mptcp/protocol.c
1368
skb->len += copy;
net/mptcp/protocol.c
1369
skb->data_len += copy;
net/mptcp/protocol.c
1370
skb->truesize += copy;
net/mptcp/protocol.c
1371
sk_wmem_queued_add(ssk, copy);
net/mptcp/protocol.c
1372
sk_mem_charge(ssk, copy);
net/mptcp/protocol.c
1373
WRITE_ONCE(tcp_sk(ssk)->write_seq, tcp_sk(ssk)->write_seq + copy);
net/mptcp/protocol.c
1374
TCP_SKB_CB(skb)->end_seq += copy;
net/mptcp/protocol.c
1380
mpext->data_len += copy;
net/mptcp/protocol.c
1387
mpext->data_len = copy;
net/mptcp/protocol.c
1397
mptcp_subflow_ctx(ssk)->rel_write_seq += copy;
net/mptcp/protocol.c
1400
mptcp_update_data_checksum(skb, copy);
net/mptcp/protocol.c
1406
mptcp_update_data_checksum(skb, copy);
net/mptcp/protocol.c
1410
mptcp_subflow_ctx(ssk)->rel_write_seq += copy;
net/mptcp/protocol.c
1411
return copy;
net/mptcp/protocol.c
1812
static int do_copy_data_nocache(struct sock *sk, int copy,
net/mptcp/protocol.c
1816
if (!copy_from_iter_full_nocache(to, copy, from))
net/mptcp/protocol.c
1818
} else if (!copy_from_iter_full(to, copy, from)) {
net/netfilter/ipset/ip_set_core.c
2309
goto copy;
net/netfilter/ipset/ip_set_core.c
2326
goto copy;
net/netfilter/ipset/ip_set_core.c
2344
goto copy;
net/netfilter/ipset/ip_set_core.c
2351
copy:
net/netfilter/nft_set_pipapo.c
715
unsigned int new_bucket_size, copy;
net/netfilter/nft_set_pipapo.c
732
copy = f->bsize;
net/netfilter/nft_set_pipapo.c
734
copy = new_bucket_size;
net/netfilter/nft_set_pipapo.c
749
memcpy(new_p, old_p, copy * sizeof(*new_p));
net/netfilter/nft_set_pipapo.c
750
new_p += copy;
net/netfilter/nft_set_pipapo.c
751
old_p += copy;
net/phonet/pep.c
1288
goto copy;
net/phonet/pep.c
1306
copy:
net/rds/message.c
80
LIST_HEAD(copy);
net/rds/message.c
84
list_splice(&q->zcookie_head, &copy);
net/rds/message.c
88
list_for_each_entry_safe(info, tmp, &copy, rs_zcookie_next) {
net/rds/recv.c
494
LIST_HEAD(copy);
net/rds/recv.c
516
list_move(&notifier->n_list, &copy);
net/rds/recv.c
524
while (!list_empty(&copy)) {
net/rds/recv.c
525
notifier = list_entry(copy.next, struct rds_notifier, n_list);
net/rds/recv.c
544
if (!list_empty(&copy)) {
net/rds/recv.c
546
list_splice(&copy, &rs->rs_notify_queue);
net/rxrpc/recvmsg.c
253
int copy, ret = -EAGAIN, ret2;
net/rxrpc/recvmsg.c
303
copy = rx_pkt_len;
net/rxrpc/recvmsg.c
304
if (copy > remain)
net/rxrpc/recvmsg.c
305
copy = remain;
net/rxrpc/recvmsg.c
306
if (copy > 0) {
net/rxrpc/recvmsg.c
308
copy);
net/rxrpc/recvmsg.c
315
rx_pkt_offset += copy;
net/rxrpc/recvmsg.c
316
rx_pkt_len -= copy;
net/rxrpc/recvmsg.c
317
*_offset += copy;
net/rxrpc/sendmsg.c
419
size_t copy = umin(txb->space, msg_data_left(msg));
net/rxrpc/sendmsg.c
421
_debug("add %zu", copy);
net/rxrpc/sendmsg.c
423
copy, &msg->msg_iter))
net/rxrpc/sendmsg.c
426
txb->space -= copy;
net/rxrpc/sendmsg.c
427
txb->len += copy;
net/rxrpc/sendmsg.c
428
txb->offset += copy;
net/rxrpc/sendmsg.c
429
copied += copy;
net/rxrpc/sendmsg.c
431
call->tx_total_len -= copy;
net/smc/smc_rx.c
396
goto copy;
net/smc/smc_rx.c
403
goto copy;
net/smc/smc_rx.c
441
copy:
net/sunrpc/xdr.c
244
size_t copy;
net/sunrpc/xdr.c
267
copy = len;
net/sunrpc/xdr.c
268
if (copy > (PAGE_SIZE - pgto_base))
net/sunrpc/xdr.c
269
copy = PAGE_SIZE - pgto_base;
net/sunrpc/xdr.c
270
if (copy > (PAGE_SIZE - pgfrom_base))
net/sunrpc/xdr.c
271
copy = PAGE_SIZE - pgfrom_base;
net/sunrpc/xdr.c
276
memcpy(vto + pgto_base, vfrom + pgfrom_base, copy);
net/sunrpc/xdr.c
279
memmove(vto + pgto_base, vto + pgfrom_base, copy);
net/sunrpc/xdr.c
283
pgto_base += copy;
net/sunrpc/xdr.c
284
pgfrom_base += copy;
net/sunrpc/xdr.c
286
} while ((len -= copy) != 0);
net/sunrpc/xdr.c
309
size_t copy;
net/sunrpc/xdr.c
336
copy = len;
net/sunrpc/xdr.c
337
if (copy > pgto_base)
net/sunrpc/xdr.c
338
copy = pgto_base;
net/sunrpc/xdr.c
339
if (copy > pgfrom_base)
net/sunrpc/xdr.c
340
copy = pgfrom_base;
net/sunrpc/xdr.c
341
pgto_base -= copy;
net/sunrpc/xdr.c
342
pgfrom_base -= copy;
net/sunrpc/xdr.c
347
memcpy(vto + pgto_base, vfrom + pgfrom_base, copy);
net/sunrpc/xdr.c
350
memmove(vto + pgto_base, vto + pgfrom_base, copy);
net/sunrpc/xdr.c
354
} while ((len -= copy) != 0);
net/sunrpc/xdr.c
372
size_t copy;
net/sunrpc/xdr.c
381
copy = PAGE_SIZE - pgbase;
net/sunrpc/xdr.c
382
if (copy > len)
net/sunrpc/xdr.c
383
copy = len;
net/sunrpc/xdr.c
386
memcpy(vto + pgbase, p, copy);
net/sunrpc/xdr.c
389
len -= copy;
net/sunrpc/xdr.c
393
pgbase += copy;
net/sunrpc/xdr.c
399
p += copy;
net/sunrpc/xdr.c
419
size_t copy;
net/sunrpc/xdr.c
428
copy = PAGE_SIZE - pgbase;
net/sunrpc/xdr.c
429
if (copy > len)
net/sunrpc/xdr.c
430
copy = len;
net/sunrpc/xdr.c
433
memcpy(p, vfrom + pgbase, copy);
net/sunrpc/xdr.c
436
pgbase += copy;
net/sunrpc/xdr.c
441
p += copy;
net/sunrpc/xdr.c
443
} while ((len -= copy) != 0);
net/tipc/socket.c
1887
int rc, err, hlen, dlen, copy;
net/tipc/socket.c
1934
copy = min_t(int, dlen - offset, buflen);
net/tipc/socket.c
1935
rc = skb_copy_datagram_msg(skb, hlen + offset, m, copy);
net/tipc/socket.c
1938
if (unlikely(offset + copy < dlen)) {
net/tipc/socket.c
1941
skb_cb->bytes_read = offset + copy;
net/tipc/socket.c
1952
copy = 0;
net/tipc/socket.c
1965
copy = 0;
net/tipc/socket.c
1995
return rc ? rc : copy;
net/tipc/socket.c
2019
int offset, required, copy, copied = 0;
net/tipc/socket.c
2066
copy = min_t(int, dlen - offset, buflen - copied);
net/tipc/socket.c
2067
rc = skb_copy_datagram_msg(skb, hlen + offset, m, copy);
net/tipc/socket.c
2070
copied += copy;
net/tipc/socket.c
2071
offset += copy;
net/tls/tls_device.c
436
int copy, rc = 0;
net/tls/tls_device.c
495
copy = min_t(size_t, size, max_open_record_len - record->len);
net/tls/tls_device.c
496
if (copy && (flags & MSG_SPLICE_PAGES)) {
net/tls/tls_device.c
502
copy, 1, 0, &off);
net/tls/tls_device.c
508
copy = rc;
net/tls/tls_device.c
511
iov_iter_revert(iter, copy);
net/tls/tls_device.c
517
zc_pfrag.size = copy;
net/tls/tls_device.c
518
tls_append_frag(record, &zc_pfrag, copy);
net/tls/tls_device.c
519
} else if (copy) {
net/tls/tls_device.c
520
copy = min_t(size_t, copy, pfrag->size - pfrag->offset);
net/tls/tls_device.c
523
pfrag->offset, copy,
net/tls/tls_device.c
527
tls_append_frag(record, pfrag, copy);
net/tls/tls_device.c
530
size -= copy;
net/tls/tls_device.c
887
int err, offset, copy, data_len, pos;
net/tls/tls_device.c
928
copy = min_t(int, skb_pagelen(skb) - offset, data_len);
net/tls/tls_device.c
931
err = skb_store_bits(skb, offset, buf, copy);
net/tls/tls_device.c
936
offset += copy;
net/tls/tls_device.c
937
buf += copy;
net/tls/tls_device.c
954
copy = min_t(int, skb_iter->len - frag_pos,
net/tls/tls_device.c
958
err = skb_store_bits(skb_iter, frag_pos, buf, copy);
net/tls/tls_device.c
963
offset += copy;
net/tls/tls_device.c
964
buf += copy;
net/wireless/util.c
1907
unsigned int copy;
net/wireless/util.c
1933
copy = min_t(unsigned int, attr_remaining, iedatalen);
net/wireless/util.c
1934
if (copy && desired_attr) {
net/wireless/util.c
1935
desired_len += copy;
net/wireless/util.c
1937
memcpy(out, iedata, min(bufsize, copy));
net/wireless/util.c
1938
out += min(bufsize, copy);
net/wireless/util.c
1939
bufsize -= min(bufsize, copy);
net/wireless/util.c
1943
if (copy == attr_remaining)
net/wireless/util.c
1947
attr_remaining -= copy;
net/wireless/util.c
1951
iedatalen -= copy;
net/wireless/util.c
1952
iedata += copy;
net/wireless/util.c
1965
copy = min_t(unsigned int, attr_len, iedatalen);
net/wireless/util.c
1968
desired_len += copy;
net/wireless/util.c
1970
memcpy(out, iedata, min(bufsize, copy));
net/wireless/util.c
1971
out += min(bufsize, copy);
net/wireless/util.c
1972
bufsize -= min(bufsize, copy);
net/wireless/util.c
1975
if (copy == attr_len)
net/wireless/util.c
1979
iedata += copy;
net/wireless/util.c
1980
iedatalen -= copy;
net/wireless/util.c
1981
attr_remaining = attr_len - copy;
net/xdp/xsk.c
724
u32 hr, len, ts, offset, copy, copied;
net/xdp/xsk.c
785
copy = min_t(u32, PAGE_SIZE - offset, len - copied);
net/xdp/xsk.c
786
skb_fill_page_desc(skb, i, page, offset, copy);
net/xdp/xsk.c
788
copied += copy;
net/xdp/xsk.c
789
addr += copy;
security/ipe/fs.c
140
char *copy = NULL;
security/ipe/fs.c
148
copy = memdup_user_nul(data, len);
security/ipe/fs.c
149
if (IS_ERR(copy)) {
security/ipe/fs.c
150
rc = PTR_ERR(copy);
security/ipe/fs.c
151
copy = NULL;
security/ipe/fs.c
155
p = ipe_new_policy(NULL, 0, copy, len);
security/ipe/fs.c
166
kfree(copy);
security/ipe/policy_fs.c
302
char *copy = NULL;
security/ipe/policy_fs.c
310
copy = memdup_user(data, len);
security/ipe/policy_fs.c
311
if (IS_ERR(copy)) {
security/ipe/policy_fs.c
312
rc = PTR_ERR(copy);
security/ipe/policy_fs.c
313
copy = NULL;
security/ipe/policy_fs.c
319
rc = ipe_update_policy(root, NULL, 0, copy, len);
security/ipe/policy_fs.c
323
kfree(copy);
security/selinux/ss/hashtab.c
141
int (*copy)(struct hashtab_node *new,
security/selinux/ss/hashtab.c
166
rc = copy(tmp, orig_cur, args);
security/selinux/ss/hashtab.h
140
int (*copy)(struct hashtab_node *new,
security/smack/smack_lsm.c
4667
char *copy;
security/smack/smack_lsm.c
4674
copy = kstrdup(skp->smk_known, GFP_KERNEL);
security/smack/smack_lsm.c
4675
if (copy == NULL)
security/smack/smack_lsm.c
4677
length = strlen(copy) + 1;
security/smack/smack_lsm.c
4679
*_buffer = copy;
sound/core/compress_offload.c
293
size_t copy;
sound/core/compress_offload.c
307
copy = runtime->buffer_size - app_pointer;
sound/core/compress_offload.c
308
if (copy_from_user(dstn, buf, copy))
sound/core/compress_offload.c
310
if (copy_from_user(runtime->buffer, buf + copy, count - copy))
sound/core/compress_offload.c
350
if (stream->ops->copy) {
sound/core/compress_offload.c
352
retval = stream->ops->copy(stream, cbuf, avail);
sound/core/compress_offload.c
406
if (stream->ops->copy)
sound/core/compress_offload.c
407
retval = stream->ops->copy(stream, buf, avail);
sound/core/compress_offload.c
589
if (stream->ops->copy) {
sound/core/init.c
787
size_t copy = count > sizeof(card->id) - 1 ?
sound/core/init.c
792
for (idx = 0; idx < copy; idx++) {
sound/core/init.c
797
memcpy(buf1, buf, copy);
sound/core/init.c
798
buf1[copy] = '\0';
sound/core/pcm_lib.c
2201
if (snd_BUG_ON(!substream->ops->copy && !runtime->dma_area))
sound/core/pcm_lib.c
2307
if (substream->ops->copy)
sound/core/pcm_lib.c
2308
transfer = substream->ops->copy;
sound/core/pcm_native.c
854
if (runtime->dma_area && !substream->ops->copy) {
sound/drivers/dummy.c
659
.copy = dummy_pcm_copy,
sound/isa/gus/gus_pcm.c
791
.copy = snd_gf1_pcm_playback_copy,
sound/isa/sb/emu8000_pcm.c
626
.copy = emu8k_pcm_copy,
sound/pci/es1938.c
961
.copy = snd_es1938_capture_copy,
sound/pci/korg1212/korg1212.c
1596
.copy = snd_korg1212_playback_copy,
sound/pci/korg1212/korg1212.c
1609
.copy = snd_korg1212_capture_copy,
sound/pci/nm256/nm256.c
878
.copy = snd_nm256_playback_copy,
sound/pci/nm256/nm256.c
892
.copy = snd_nm256_capture_copy,
sound/pci/rme32.c
1144
.copy = snd_rme32_playback_copy,
sound/pci/rme32.c
1156
.copy = snd_rme32_capture_copy,
sound/pci/rme32.c
1167
.copy = snd_rme32_playback_copy,
sound/pci/rme32.c
1179
.copy = snd_rme32_capture_copy,
sound/pci/rme96.c
1473
.copy = snd_rme96_playback_copy,
sound/pci/rme96.c
1485
.copy = snd_rme96_capture_copy,
sound/pci/rme96.c
1496
.copy = snd_rme96_playback_copy,
sound/pci/rme96.c
1508
.copy = snd_rme96_capture_copy,
sound/pci/rme9652/hdsp.c
4860
.copy = snd_hdsp_playback_copy,
sound/pci/rme9652/hdsp.c
4872
.copy = snd_hdsp_capture_copy,
sound/pci/rme9652/rme9652.c
2297
.copy = snd_rme9652_playback_copy,
sound/pci/rme9652/rme9652.c
2309
.copy = snd_rme9652_capture_copy,
sound/sh/sh_dac_audio.c
212
.copy = snd_sh_dac_pcm_copy,
sound/soc/codecs/cs47l15.c
1344
.copy = &wm_adsp_compr_copy,
sound/soc/codecs/cs47l24.c
1192
.copy = wm_adsp_compr_copy,
sound/soc/codecs/cs47l35.c
1628
.copy = &wm_adsp_compr_copy,
sound/soc/codecs/cs47l85.c
2572
.copy = &wm_adsp_compr_copy,
sound/soc/codecs/cs47l90.c
2487
.copy = &wm_adsp_compr_copy,
sound/soc/codecs/cs47l92.c
1952
.copy = &wm_adsp_compr_copy,
sound/soc/codecs/cs48l32.c
3306
.copy = &wm_adsp_compr_copy,
sound/soc/codecs/wm5102.c
2016
.copy = wm_adsp_compr_copy,
sound/soc/codecs/wm5110.c
2373
.copy = wm_adsp_compr_copy,
sound/soc/intel/avs/probes.c
271
.copy = avs_probe_compr_copy,
sound/soc/mediatek/common/mtk-btcvsd.c
1270
.copy = mtk_pcm_btcvsd_copy,
sound/soc/qcom/lpass-platform.c
1274
.copy = lpass_platform_copy,
sound/soc/qcom/qdsp6/q6apm-dai.c
755
size_t copy;
sound/soc/qcom/qdsp6/q6apm-dai.c
782
copy = prtd->pcm_size - app_pointer;
sound/soc/qcom/qdsp6/q6apm-dai.c
783
if (copy_from_user(dstn, buf, copy))
sound/soc/qcom/qdsp6/q6apm-dai.c
785
if (copy_from_user(prtd->dma_buffer.area, buf + copy, count - copy))
sound/soc/qcom/qdsp6/q6apm-dai.c
826
.copy = q6apm_compr_copy,
sound/soc/qcom/qdsp6/q6asm-dai.c
1068
size_t copy;
sound/soc/qcom/qdsp6/q6asm-dai.c
1092
copy = prtd->pcm_size - app_pointer;
sound/soc/qcom/qdsp6/q6asm-dai.c
1093
if (copy_from_user(dstn, buf, copy))
sound/soc/qcom/qdsp6/q6asm-dai.c
1095
if (copy_from_user(prtd->dma_buffer.area, buf + copy,
sound/soc/qcom/qdsp6/q6asm-dai.c
1096
count - copy))
sound/soc/qcom/qdsp6/q6asm-dai.c
1186
.copy = q6asm_compr_copy,
sound/soc/soc-component.c
588
component->driver->compress_ops->copy) {
sound/soc/soc-component.c
589
ret = component->driver->compress_ops->copy(
sound/soc/soc-component.c
986
if (component->driver->copy)
sound/soc/soc-component.c
988
component->driver->copy(component, substream,
sound/soc/soc-compress.c
647
!component->driver->compress_ops->copy)
sound/soc/soc-compress.c
650
compr->ops->copy = snd_soc_component_compr_copy;
sound/soc/soc-generic-dmaengine-pcm.c
347
.copy = dmaengine_copy,
sound/soc/soc-pcm.c
2973
if (drv->copy)
sound/soc/soc-pcm.c
2974
rtd->ops.copy = snd_soc_pcm_component_copy;
sound/soc/sof/compress.c
397
.copy = sof_compr_copy,
sound/soc/sof/sof-client-probes.c
199
.copy = sof_probes_compr_copy,
sound/soc/sprd/sprd-pcm-compress.c
664
.copy = sprd_platform_compr_copy,
sound/soc/uniphier/aio-compress.c
430
.copy = uniphier_aio_compr_copy,
sound/xen/xen_snd_front_alsa.c
662
.copy = alsa_pb_copy,
sound/xen/xen_snd_front_alsa.c
674
.copy = alsa_cap_copy,
tools/bootconfig/main.c
274
char *copy = strdup(buf);
tools/bootconfig/main.c
278
if (!copy)
tools/bootconfig/main.c
283
show_xbc_error(copy, msg, pos);
tools/bootconfig/main.c
284
free(copy);
tools/include/linux/rbtree_augmented.h
105
.copy = RBNAME ## _copy, \
tools/include/linux/rbtree_augmented.h
236
augment->copy(node, successor);
tools/include/linux/rbtree_augmented.h
262
augment->copy(node, successor);
tools/include/linux/rbtree_augmented.h
31
void (*copy)(struct rb_node *old, struct rb_node *new);
tools/include/uapi/linux/userfaultfd.h
293
__s64 copy;
tools/lib/rbtree.c
429
.copy = dummy_copy,
tools/perf/builtin-inject.c
2395
.fc.copy = feat_copy_cb,
tools/perf/util/header.c
3879
if (fc && fc->copy) {
tools/perf/util/header.c
3886
err = fc->copy(fc, type, &h.fw);
tools/perf/util/header.h
148
int (*copy)(struct feat_copier *fc, int feat, struct feat_writer *fw);
tools/perf/util/ordered-events.h
74
void ordered_events__set_copy_on_queue(struct ordered_events *oe, bool copy)
tools/perf/util/ordered-events.h
76
oe->copy_on_queue = copy;
tools/testing/selftests/arm64/fp/fp-stress.c
311
static void start_fpsimd(struct child_data *child, int cpu, int copy)
tools/testing/selftests/arm64/fp/fp-stress.c
315
ret = asprintf(&child->name, "FPSIMD-%d-%d", cpu, copy);
tools/testing/selftests/arm64/fp/fp-stress.c
324
static void start_kernel(struct child_data *child, int cpu, int copy)
tools/testing/selftests/arm64/fp/fp-stress.c
328
ret = asprintf(&child->name, "KERNEL-%d-%d", cpu, copy);
tools/testing/selftests/bpf/prog_tests/bpf_iter.c
1628
do_read_map_iter_fd(&skel->skeleton, skel->progs.copy, skel->maps.sockmap);
tools/testing/selftests/bpf/prog_tests/sockmap_basic.c
341
link = bpf_program__attach_iter(skel->progs.copy, &opts);
tools/testing/selftests/bpf/progs/bpf_iter_sockmap.c
36
int copy(struct bpf_iter__sockmap *ctx)
tools/testing/selftests/kvm/arm64/page_fault_test.c
303
char *copy;
tools/testing/selftests/kvm/arm64/page_fault_test.c
314
struct uffdio_copy copy;
tools/testing/selftests/kvm/arm64/page_fault_test.c
324
copy.src = (uint64_t)args->copy;
tools/testing/selftests/kvm/arm64/page_fault_test.c
325
copy.dst = addr;
tools/testing/selftests/kvm/arm64/page_fault_test.c
326
copy.len = args->paging_size;
tools/testing/selftests/kvm/arm64/page_fault_test.c
327
copy.mode = 0;
tools/testing/selftests/kvm/arm64/page_fault_test.c
329
ret = ioctl(uffd, UFFDIO_COPY, &copy);
tools/testing/selftests/kvm/arm64/page_fault_test.c
358
args->copy = malloc(args->paging_size);
tools/testing/selftests/kvm/arm64/page_fault_test.c
359
TEST_ASSERT(args->copy, "Failed to allocate data copy.");
tools/testing/selftests/kvm/arm64/page_fault_test.c
360
memcpy(args->copy, args->hva, args->paging_size);
tools/testing/selftests/kvm/arm64/page_fault_test.c
395
free(pt_args.copy);
tools/testing/selftests/kvm/arm64/page_fault_test.c
396
free(data_args.copy);
tools/testing/selftests/kvm/demand_paging_test.c
69
struct uffdio_copy copy;
tools/testing/selftests/kvm/demand_paging_test.c
71
copy.src = (uint64_t)guest_data_prototype;
tools/testing/selftests/kvm/demand_paging_test.c
72
copy.dst = addr;
tools/testing/selftests/kvm/demand_paging_test.c
73
copy.len = demand_paging_size;
tools/testing/selftests/kvm/demand_paging_test.c
74
copy.mode = 0;
tools/testing/selftests/kvm/demand_paging_test.c
76
r = ioctl(uffd, UFFDIO_COPY, &copy);
tools/testing/selftests/mm/uffd-common.c
606
if (uffdio_copy->copy != -EEXIST)
tools/testing/selftests/mm/uffd-common.c
608
(int64_t)uffdio_copy->copy);
tools/testing/selftests/mm/uffd-common.c
611
(int64_t)uffdio_copy->copy);
tools/testing/selftests/mm/uffd-common.c
640
uffdio_copy.copy = 0;
tools/testing/selftests/mm/uffd-common.c
643
if (uffdio_copy.copy != -EEXIST)
tools/testing/selftests/mm/uffd-common.c
645
(int64_t)uffdio_copy.copy);
tools/testing/selftests/mm/uffd-common.c
647
} else if (uffdio_copy.copy != gopts->page_size) {
tools/testing/selftests/mm/uffd-common.c
648
err("UFFDIO_COPY error: %"PRId64, (int64_t)uffdio_copy.copy);
tools/testing/selftests/mm/uffd-unit-tests.c
1283
DEFINE_MMAP_CHANGING_TEST(copy, UFFDIO_COPY, copy)
tools/testing/selftests/openat2/openat2_test.c
115
copy = malloc(misalign + sizeof(how_ext));
tools/testing/selftests/openat2/openat2_test.c
116
how_copy = copy + misalign;
tools/testing/selftests/openat2/openat2_test.c
117
memset(copy, 0xff, misalign);
tools/testing/selftests/openat2/openat2_test.c
150
free(copy);
tools/testing/selftests/openat2/openat2_test.c
99
void *copy = NULL, *how_copy = &how_ext;
tools/testing/selftests/powerpc/tm/tm-signal-pagefault.c
129
uffdio_copy.copy = 0;
tools/testing/selftests/vfio/lib/drivers/dsa/dsa.c
204
dsa->max_copies_per_batch = min(dsa->max_copies_per_batch, ARRAY_SIZE(dsa->copy));
tools/testing/selftests/vfio/lib/drivers/dsa/dsa.c
22
struct dsa_hw_desc copy[1024];
tools/testing/selftests/vfio/lib/drivers/dsa/dsa.c
299
.desc_list_addr = to_iova(device, &dsa->copy[0]),
tools/testing/selftests/vfio/lib/drivers/dsa/dsa.c
317
dsa_copy_desc_init(device, &dsa->copy[0], src, dst, size, interrupt);
tools/testing/selftests/vfio/lib/drivers/dsa/dsa.c
318
dsa_desc_write(device, &dsa->copy[0]);
tools/testing/selftests/vfio/lib/drivers/dsa/dsa.c
329
for (i = 0; i < ARRAY_SIZE(dsa->copy); i++) {
tools/testing/selftests/vfio/lib/drivers/dsa/dsa.c
330
struct dsa_hw_desc *copy_desc = &dsa->copy[i];