arch/arm/mm/cache-uniphier.c
206
unsigned long chunk_size = min_t(unsigned long, size,
arch/arm/mm/cache-uniphier.c
209
__uniphier_cache_maint_common(data, start, chunk_size,
arch/arm/mm/cache-uniphier.c
212
start += chunk_size;
arch/arm/mm/cache-uniphier.c
213
size -= chunk_size;
arch/arm64/kvm/mmu.c
109
u64 chunk_size, min;
arch/arm64/kvm/mmu.c
114
chunk_size = kvm->arch.mmu.split_page_chunk_size;
arch/arm64/kvm/mmu.c
115
min = kvm_mmu_split_nr_page_tables(chunk_size);
arch/arm64/kvm/mmu.c
126
u64 next, chunk_size;
arch/arm64/kvm/mmu.c
130
chunk_size = kvm->arch.mmu.split_page_chunk_size;
arch/arm64/kvm/mmu.c
131
cache_capacity = kvm_mmu_split_nr_page_tables(chunk_size);
arch/arm64/kvm/mmu.c
133
if (chunk_size == 0)
arch/arm64/kvm/mmu.c
155
next = __stage2_range_addr_end(addr, end, chunk_size);
arch/x86/kernel/cpu/microcode/intel.c
104
unsigned int chunk_size;
arch/x86/kernel/cpu/microcode/intel.c
420
ss->chunk_size = min(PAGE_SIZE, ss->ucode_len - ss->offset);
arch/x86/kernel/cpu/microcode/intel.c
438
if (ss->bytes_sent + ss->chunk_size > ss->ucode_len * 2) {
arch/x86/kernel/cpu/microcode/intel.c
514
mbox_size = MBOX_HEADER_SIZE * 2 + ss->chunk_size;
arch/x86/kernel/cpu/microcode/intel.c
517
write_mbox_data(ss->mmio_base, src_chunk, ss->chunk_size);
arch/x86/kernel/cpu/microcode/intel.c
518
ss->bytes_sent += ss->chunk_size;
arch/x86/kernel/cpu/mtrr/cleanup.c
448
u64 chunk_size, u64 gran_size)
arch/x86/kernel/cpu/mtrr/cleanup.c
457
var_state.chunk_sizek = chunk_size >> 10;
arch/x86/kernel/cpu/mtrr/cleanup.c
564
mtrr_calc_range_state(u64 chunk_size, u64 gran_size,
arch/x86/kernel/cpu/mtrr/cleanup.c
581
num_reg = x86_setup_var_mtrrs(range, nr_range, chunk_size, gran_size);
arch/x86/kernel/cpu/mtrr/cleanup.c
589
result[i].chunk_sizek = chunk_size >> 10;
arch/x86/kernel/cpu/mtrr/cleanup.c
661
u64 chunk_size, gran_size;
arch/x86/kernel/cpu/mtrr/cleanup.c
735
for (chunk_size = gran_size; chunk_size < (1ULL<<32);
arch/x86/kernel/cpu/mtrr/cleanup.c
736
chunk_size <<= 1) {
arch/x86/kernel/cpu/mtrr/cleanup.c
741
mtrr_calc_range_state(chunk_size, gran_size,
arch/x86/kernel/cpu/mtrr/cleanup.c
761
chunk_size = result[i].chunk_sizek;
arch/x86/kernel/cpu/mtrr/cleanup.c
762
chunk_size <<= 10;
arch/x86/kernel/cpu/mtrr/cleanup.c
765
x86_setup_var_mtrrs(range, nr_range, chunk_size, gran_size);
arch/x86/platform/olpc/olpc_dt.c
131
const size_t chunk_size = max(PAGE_SIZE, size);
arch/x86/platform/olpc/olpc_dt.c
139
res = memblock_alloc_or_panic(chunk_size, SMP_CACHE_BYTES);
arch/x86/platform/olpc/olpc_dt.c
140
prom_early_allocated += chunk_size;
arch/x86/platform/olpc/olpc_dt.c
141
memset(res, 0, chunk_size);
arch/x86/platform/olpc/olpc_dt.c
142
free_mem = chunk_size;
arch/x86/xen/setup.c
792
phys_addr_t mem_end, addr, size, chunk_size;
arch/x86/xen/setup.c
899
chunk_size = size;
arch/x86/xen/setup.c
907
chunk_size = min(size, mem_end - addr);
arch/x86/xen/setup.c
909
chunk_size = min(size, PFN_PHYS(extra_pages));
arch/x86/xen/setup.c
911
n_pfns = PFN_DOWN(addr + chunk_size) - pfn_s;
arch/x86/xen/setup.c
920
xen_align_and_add_e820_region(addr, chunk_size, type);
arch/x86/xen/setup.c
922
addr += chunk_size;
arch/x86/xen/setup.c
923
size -= chunk_size;
drivers/accel/habanalabs/common/memory.c
1515
static int set_dma_sg(struct scatterlist *sg, u64 bar_address, u64 chunk_size,
drivers/accel/habanalabs/common/memory.c
1521
addr = dma_map_resource(dev, bar_address, chunk_size, dir,
drivers/accel/habanalabs/common/memory.c
1527
sg_set_page(sg, NULL, chunk_size, 0);
drivers/accel/habanalabs/common/memory.c
1529
sg_dma_len(sg) = chunk_size;
drivers/accel/habanalabs/common/memory.c
1538
u64 dma_max_seg_size, curr_page, size, chunk_size, left_size_to_export, left_size_in_page,
drivers/accel/habanalabs/common/memory.c
1620
chunk_size = 0;
drivers/accel/habanalabs/common/memory.c
1624
chunk_size += size;
drivers/accel/habanalabs/common/memory.c
1648
device_address += chunk_size;
drivers/accel/habanalabs/common/memory.c
1661
rc = set_dma_sg(sg, bar_address, chunk_size, dev, dir);
drivers/accel/habanalabs/gaudi/gaudi.c
4575
u32 chunk_size, busy;
drivers/accel/habanalabs/gaudi/gaudi.c
4582
chunk_size =
drivers/accel/habanalabs/gaudi/gaudi.c
4587
cur_addr, cur_addr + chunk_size);
drivers/accel/habanalabs/gaudi/gaudi.c
4598
chunk_size);
drivers/accel/habanalabs/gaudi/gaudi.c
4603
cur_addr += chunk_size;
drivers/accel/habanalabs/gaudi2/gaudi2.c
10675
u32 chunk_size, dcore, edma_idx, sob_offset, sob_addr, comp_val,
drivers/accel/habanalabs/gaudi2/gaudi2.c
10752
chunk_size = min_t(u64, SZ_2G, end_addr - cur_addr);
drivers/accel/habanalabs/gaudi2/gaudi2.c
10758
chunk_size, cur_addr, val);
drivers/accel/habanalabs/gaudi2/gaudi2.c
10763
cur_addr += chunk_size;
drivers/crypto/intel/qat/qat_common/icp_qat_uclo.h
470
u64 chunk_size;
drivers/crypto/intel/qat/qat_common/qat_uclo.c
1382
(*desc))->chunk_size;
drivers/crypto/intel/qat/qat_common/qat_uclo.c
1659
auth_chunk->chunk_size = img_desc.dram_size;
drivers/dma/uniphier-mdmac.c
112
u32 src_mode, src_addr, dest_mode, dest_addr, chunk_size;
drivers/dma/uniphier-mdmac.c
128
chunk_size = sg_dma_len(sg);
drivers/dma/uniphier-mdmac.c
134
writel(chunk_size, mc->reg_ch_base + UNIPHIER_MDMAC_CH_SIZE);
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
309
rsv->size, mm->chunk_size, &rsv->allocated,
drivers/gpu/drm/amd/amdgpu/amdgpu_vram_mgr.c
537
BUG_ON(min_block_size < mm->chunk_size);
drivers/gpu/drm/amd/amdkfd/kfd_device.c
1274
unsigned int chunk_size)
drivers/gpu/drm/amd/amdkfd/kfd_device.c
1276
if (WARN_ON(buf_size < chunk_size))
drivers/gpu/drm/amd/amdkfd/kfd_device.c
1280
if (WARN_ON(chunk_size == 0))
drivers/gpu/drm/amd/amdkfd/kfd_device.c
1283
kfd->gtt_sa_chunk_size = chunk_size;
drivers/gpu/drm/amd/amdkfd/kfd_device.c
1284
kfd->gtt_sa_num_of_chunks = buf_size / chunk_size;
drivers/gpu/drm/amd/amdkfd/kfd_device.c
1307
unsigned int chunk_size)
drivers/gpu/drm/amd/amdkfd/kfd_device.c
1309
return start_addr + bit_num * chunk_size;
drivers/gpu/drm/amd/amdkfd/kfd_device.c
1314
unsigned int chunk_size)
drivers/gpu/drm/amd/amdkfd/kfd_device.c
1316
return (uint32_t *) ((uint64_t) start_addr + bit_num * chunk_size);
drivers/gpu/drm/amd/amdkfd/kfd_device.c
63
unsigned int chunk_size);
drivers/gpu/drm/amd/display/dc/core/dc.c
6693
if (pipe_ctx->rq_regs.rq_regs_l.chunk_size > 0) {
drivers/gpu/drm/amd/display/dc/core/dc.c
6694
state->hubp[i].rq_chunk_size = pipe_ctx->rq_regs.rq_regs_l.chunk_size;
drivers/gpu/drm/amd/display/dc/core/dc.c
6705
if (pipe_ctx->rq_regs.rq_regs_c.chunk_size > 0) {
drivers/gpu/drm/amd/display/dc/core/dc.c
6706
state->hubp[i].rq_chunk_size_c = pipe_ctx->rq_regs.rq_regs_c.chunk_size;
drivers/gpu/drm/amd/display/dc/dcn10/dcn10_hw_sequencer_debug.c
214
rq_regs->crq_expansion_mode, rq_regs->plane1_base_address, rq_regs->rq_regs_l.chunk_size,
drivers/gpu/drm/amd/display/dc/dcn10/dcn10_hw_sequencer_debug.c
218
rq_regs->rq_regs_l.pte_row_height_linear, rq_regs->rq_regs_c.chunk_size, rq_regs->rq_regs_c.min_chunk_size,
drivers/gpu/drm/amd/display/dc/dml/dcn20/display_rq_dlg_calc_20.c
172
rq_regs->chunk_size = dml_log2(rq_sizing->chunk_bytes) - 10;
drivers/gpu/drm/amd/display/dc/dml/dcn20/display_rq_dlg_calc_20v2.c
172
rq_regs->chunk_size = dml_log2(rq_sizing->chunk_bytes) - 10;
drivers/gpu/drm/amd/display/dc/dml/dcn21/display_rq_dlg_calc_21.c
149
rq_regs->chunk_size = dml_log2(rq_sizing->chunk_bytes) - 10;
drivers/gpu/drm/amd/display/dc/dml/dcn30/display_rq_dlg_calc_30.c
96
rq_regs->chunk_size = dml_log2(rq_sizing->chunk_bytes) - 10;
drivers/gpu/drm/amd/display/dc/dml/dcn31/display_rq_dlg_calc_31.c
95
rq_regs->chunk_size = dml_log2(rq_sizing->chunk_bytes) - 10;
drivers/gpu/drm/amd/display/dc/dml/dcn314/display_rq_dlg_calc_314.c
183
rq_regs->chunk_size = dml_log2(rq_sizing->chunk_bytes) - 10;
drivers/gpu/drm/amd/display/dc/dml/dcn32/display_rq_dlg_calc_32.c
98
rq_regs->rq_regs_l.chunk_size = dml_log2(pixel_chunk_bytes) - 10;
drivers/gpu/drm/amd/display/dc/dml/dcn32/display_rq_dlg_calc_32.c
99
rq_regs->rq_regs_c.chunk_size = dml_log2(p1_pixel_chunk_bytes) - 10;
drivers/gpu/drm/amd/display/dc/dml/display_mode_structs.h
699
unsigned int chunk_size;
drivers/gpu/drm/amd/display/dc/dml/display_rq_dlg_helpers.c
149
dml_print("DML_RQ_DLG_CALC: chunk_size = 0x%0x\n", rq_regs->chunk_size);
drivers/gpu/drm/amd/display/dc/dml/dml1_display_rq_dlg_calc.c
214
rq_regs->chunk_size = dml_log2(rq_sizing->chunk_bytes) - 10;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core_structs.h
2012
dml_uint_t chunk_size;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_util.c
222
dml_print("DML: chunk_size = 0x%x\n", rq_regs->chunk_size);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/inc/dml_top_dchub_registers.h
104
uint32_t chunk_size;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12299
rq_regs->rq_regs_l.chunk_size = log_and_substract_if_non_zero(pixel_chunk_bytes, 10);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12300
rq_regs->rq_regs_c.chunk_size = log_and_substract_if_non_zero(p1_pixel_chunk_bytes, 10);
drivers/gpu/drm/amd/display/dc/dml2_0/dml2_translation_helper.c
1435
out->rq_regs.rq_regs_l.chunk_size = rq_regs->rq_regs_l.chunk_size;
drivers/gpu/drm/amd/display/dc/dml2_0/dml2_translation_helper.c
1444
out->rq_regs.rq_regs_c.chunk_size = rq_regs->rq_regs_c.chunk_size;
drivers/gpu/drm/amd/display/dc/dml2_0/dml_display_rq_dlg_calc.c
97
rq_regs->rq_regs_l.chunk_size = (dml_uint_t)(dml_log2((dml_float_t) pixel_chunk_bytes) - 10);
drivers/gpu/drm/amd/display/dc/dml2_0/dml_display_rq_dlg_calc.c
98
rq_regs->rq_regs_c.chunk_size = (dml_uint_t)(dml_log2((dml_float_t) p1_pixel_chunk_bytes) - 10);
drivers/gpu/drm/amd/display/dc/hubp/dcn10/dcn10_hubp.c
1102
CHUNK_SIZE, &rq_regs->rq_regs_l.chunk_size,
drivers/gpu/drm/amd/display/dc/hubp/dcn10/dcn10_hubp.c
1112
CHUNK_SIZE_C, &rq_regs->rq_regs_c.chunk_size,
drivers/gpu/drm/amd/display/dc/hubp/dcn10/dcn10_hubp.c
587
CHUNK_SIZE, rq_regs->rq_regs_l.chunk_size,
drivers/gpu/drm/amd/display/dc/hubp/dcn10/dcn10_hubp.c
596
CHUNK_SIZE_C, rq_regs->rq_regs_c.chunk_size,
drivers/gpu/drm/amd/display/dc/hubp/dcn20/dcn20_hubp.c
1339
CHUNK_SIZE, &rq_regs->rq_regs_l.chunk_size,
drivers/gpu/drm/amd/display/dc/hubp/dcn20/dcn20_hubp.c
1349
CHUNK_SIZE_C, &rq_regs->rq_regs_c.chunk_size,
drivers/gpu/drm/amd/display/dc/hubp/dcn20/dcn20_hubp.c
1388
CHUNK_SIZE, &rq_regs.rq_regs_l.chunk_size,
drivers/gpu/drm/amd/display/dc/hubp/dcn20/dcn20_hubp.c
1397
CHUNK_SIZE_C, &rq_regs.rq_regs_c.chunk_size,
drivers/gpu/drm/amd/display/dc/hubp/dcn20/dcn20_hubp.c
1422
if (rq_regs.rq_regs_l.chunk_size != dml_rq_regs->rq_regs_l.chunk_size)
drivers/gpu/drm/amd/display/dc/hubp/dcn20/dcn20_hubp.c
1424
dml_rq_regs->rq_regs_l.chunk_size, rq_regs.rq_regs_l.chunk_size);
drivers/gpu/drm/amd/display/dc/hubp/dcn20/dcn20_hubp.c
1447
if (rq_regs.rq_regs_c.chunk_size != dml_rq_regs->rq_regs_c.chunk_size)
drivers/gpu/drm/amd/display/dc/hubp/dcn20/dcn20_hubp.c
1449
dml_rq_regs->rq_regs_c.chunk_size, rq_regs.rq_regs_c.chunk_size);
drivers/gpu/drm/amd/display/dc/hubp/dcn20/dcn20_hubp.c
210
CHUNK_SIZE, rq_regs->rq_regs_l.chunk_size,
drivers/gpu/drm/amd/display/dc/hubp/dcn20/dcn20_hubp.c
219
CHUNK_SIZE_C, rq_regs->rq_regs_c.chunk_size,
drivers/gpu/drm/amd/display/dc/hubp/dcn201/dcn201_hubp.c
82
CHUNK_SIZE, rq_regs->rq_regs_l.chunk_size,
drivers/gpu/drm/amd/display/dc/hubp/dcn201/dcn201_hubp.c
89
CHUNK_SIZE_C, rq_regs->rq_regs_c.chunk_size,
drivers/gpu/drm/amd/display/dc/hubp/dcn21/dcn21_hubp.c
151
CHUNK_SIZE, rq_regs->rq_regs_l.chunk_size,
drivers/gpu/drm/amd/display/dc/hubp/dcn21/dcn21_hubp.c
160
CHUNK_SIZE_C, rq_regs->rq_regs_c.chunk_size,
drivers/gpu/drm/amd/display/dc/hubp/dcn21/dcn21_hubp.c
273
CHUNK_SIZE, &rq_regs.rq_regs_l.chunk_size,
drivers/gpu/drm/amd/display/dc/hubp/dcn21/dcn21_hubp.c
282
CHUNK_SIZE_C, &rq_regs.rq_regs_c.chunk_size,
drivers/gpu/drm/amd/display/dc/hubp/dcn21/dcn21_hubp.c
306
if (rq_regs.rq_regs_l.chunk_size != dml_rq_regs->rq_regs_l.chunk_size)
drivers/gpu/drm/amd/display/dc/hubp/dcn21/dcn21_hubp.c
308
dml_rq_regs->rq_regs_l.chunk_size, rq_regs.rq_regs_l.chunk_size);
drivers/gpu/drm/amd/display/dc/hubp/dcn21/dcn21_hubp.c
331
if (rq_regs.rq_regs_c.chunk_size != dml_rq_regs->rq_regs_c.chunk_size)
drivers/gpu/drm/amd/display/dc/hubp/dcn21/dcn21_hubp.c
333
dml_rq_regs->rq_regs_c.chunk_size, rq_regs.rq_regs_c.chunk_size);
drivers/gpu/drm/amd/display/dc/hubp/dcn30/dcn30_hubp.c
451
CHUNK_SIZE, &rq_regs->rq_regs_l.chunk_size,
drivers/gpu/drm/amd/display/dc/hubp/dcn30/dcn30_hubp.c
460
CHUNK_SIZE_C, &rq_regs->rq_regs_c.chunk_size,
drivers/gpu/drm/amd/display/dc/hubp/dcn401/dcn401_hubp.c
234
CHUNK_SIZE, rq_regs->rq_regs_l.chunk_size,
drivers/gpu/drm/amd/display/dc/hubp/dcn401/dcn401_hubp.c
241
CHUNK_SIZE_C, rq_regs->rq_regs_c.chunk_size,
drivers/gpu/drm/amd/display/dc/hubp/dcn401/dcn401_hubp.c
840
CHUNK_SIZE, &rq_regs->rq_regs_l.chunk_size,
drivers/gpu/drm/amd/display/dc/hubp/dcn401/dcn401_hubp.c
847
CHUNK_SIZE_C, &rq_regs->rq_regs_c.chunk_size,
drivers/gpu/drm/amd/display/dc/hwss/dcn10/dcn10_hwseq.c
382
rq_regs->crq_expansion_mode, rq_regs->plane1_base_address, rq_regs->rq_regs_l.chunk_size,
drivers/gpu/drm/amd/display/dc/hwss/dcn10/dcn10_hwseq.c
386
rq_regs->rq_regs_l.pte_row_height_linear, rq_regs->rq_regs_c.chunk_size, rq_regs->rq_regs_c.min_chunk_size,
drivers/gpu/drm/display/drm_dsc_helper.c
155
pps_payload->chunk_size = cpu_to_be16(dsc_cfg->slice_chunk_size);
drivers/gpu/drm/drm_buddy.c
1019
if (!new_size || !IS_ALIGNED(new_size, mm->chunk_size))
drivers/gpu/drm/drm_buddy.c
1032
if (!IS_ALIGNED(new_start, mm->chunk_size))
drivers/gpu/drm/drm_buddy.c
1112
if (size < mm->chunk_size)
drivers/gpu/drm/drm_buddy.c
1115
if (min_block_size < mm->chunk_size)
drivers/gpu/drm/drm_buddy.c
1121
if (!IS_ALIGNED(start | end | size, mm->chunk_size))
drivers/gpu/drm/drm_buddy.c
1150
pages = size >> ilog2(mm->chunk_size);
drivers/gpu/drm/drm_buddy.c
1152
min_order = ilog2(min_block_size) - ilog2(mm->chunk_size);
drivers/gpu/drm/drm_buddy.c
1283
mm->chunk_size >> 10, mm->size >> 20, mm->avail >> 20, mm->clear_avail >> 20);
drivers/gpu/drm/drm_buddy.c
1302
free = count * (mm->chunk_size << order);
drivers/gpu/drm/drm_buddy.c
299
int drm_buddy_init(struct drm_buddy *mm, u64 size, u64 chunk_size)
drivers/gpu/drm/drm_buddy.c
304
if (size < chunk_size)
drivers/gpu/drm/drm_buddy.c
307
if (chunk_size < SZ_4K)
drivers/gpu/drm/drm_buddy.c
310
if (!is_power_of_2(chunk_size))
drivers/gpu/drm/drm_buddy.c
313
size = round_down(size, chunk_size);
drivers/gpu/drm/drm_buddy.c
318
mm->chunk_size = chunk_size;
drivers/gpu/drm/drm_buddy.c
319
mm->max_order = ilog2(size) - ilog2(chunk_size);
drivers/gpu/drm/drm_buddy.c
352
order = ilog2(size) - ilog2(chunk_size);
drivers/gpu/drm/drm_buddy.c
353
root_size = chunk_size << order;
drivers/gpu/drm/drm_buddy.c
362
BUG_ON(drm_buddy_block_size(mm, root) < chunk_size);
drivers/gpu/drm/drm_buddy.c
401
order = ilog2(size) - ilog2(mm->chunk_size);
drivers/gpu/drm/drm_buddy.c
410
root_size = mm->chunk_size << order;
drivers/gpu/drm/drm_buddy.c
437
offset + (mm->chunk_size << block_order));
drivers/gpu/drm/drm_buddy.c
492
order = ilog2(size) - ilog2(mm->chunk_size);
drivers/gpu/drm/drm_buddy.c
496
root_size = mm->chunk_size << order;
drivers/gpu/drm/drm_buddy.c
603
u64 req_size = mm->chunk_size << order;
drivers/gpu/drm/drm_buddy.c
923
pages = modify_size >> ilog2(mm->chunk_size);
drivers/gpu/drm/drm_gpusvm.c
1025
unsigned long chunk_size;
drivers/gpu/drm/drm_gpusvm.c
1079
chunk_size = drm_gpusvm_range_chunk_size(gpusvm, notifier, vas,
drivers/gpu/drm/drm_gpusvm.c
1084
if (chunk_size == LONG_MAX) {
drivers/gpu/drm/drm_gpusvm.c
1089
range = drm_gpusvm_range_alloc(gpusvm, notifier, fault_addr, chunk_size,
drivers/gpu/drm/drm_gpusvm.c
624
unsigned long fault_addr, unsigned long chunk_size,
drivers/gpu/drm/drm_gpusvm.c
640
range->itree.start = ALIGN_DOWN(fault_addr, chunk_size);
drivers/gpu/drm/drm_gpusvm.c
641
range->itree.last = ALIGN(fault_addr + 1, chunk_size) - 1;
drivers/gpu/drm/i915/display/intel_gmbus.c
558
unsigned int chunk_size = len;
drivers/gpu/drm/i915/display/intel_gmbus.c
569
gmbus1_index | GMBUS_CYCLE_WAIT | (chunk_size << GMBUS_BYTE_COUNT_SHIFT) | (addr << GMBUS_SLAVE_ADDR_SHIFT) | GMBUS_SLAVE_WRITE | GMBUS_SW_RDY);
drivers/gpu/drm/i915/i915_ttm_buddy_manager.c
286
u64 chunk_size)
drivers/gpu/drm/i915/i915_ttm_buddy_manager.c
296
err = drm_buddy_init(&bman->mm, size, chunk_size);
drivers/gpu/drm/i915/i915_ttm_buddy_manager.c
302
GEM_BUG_ON(default_page_size < chunk_size);
drivers/gpu/drm/i915/i915_ttm_buddy_manager.c
384
size, mm->chunk_size,
drivers/gpu/drm/i915/i915_ttm_buddy_manager.c
75
GEM_BUG_ON(min_page_size < mm->chunk_size);
drivers/gpu/drm/i915/i915_ttm_buddy_manager.c
83
n_pages = size >> ilog2(mm->chunk_size);
drivers/gpu/drm/i915/i915_ttm_buddy_manager.h
55
u64 default_page_size, u64 chunk_size);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
162
const u32 chunk_size = SZ_32M;
drivers/gpu/drm/i915/selftests/intel_memory_region.c
169
count = avail / chunk_size;
drivers/gpu/drm/i915/selftests/intel_memory_region.c
183
u64 start = order[i] * chunk_size;
drivers/gpu/drm/i915/selftests/intel_memory_region.c
184
u64 size = i915_prandom_u32_max_state(chunk_size, &prng);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
191
offset = igt_random_offset(&prng, 0, chunk_size, size,
drivers/gpu/drm/panthor/panthor_drv.c
1159
args->chunk_size,
drivers/gpu/drm/panthor/panthor_heap.c
134
atomic_sub(heap->chunk_size, &pool->size);
drivers/gpu/drm/panthor/panthor_heap.c
152
chunk->bo = panthor_kernel_bo_create(pool->ptdev, pool->vm, heap->chunk_size,
drivers/gpu/drm/panthor/panthor_heap.c
179
(heap->chunk_size >> 12);
drivers/gpu/drm/panthor/panthor_heap.c
189
atomic_add(heap->chunk_size, &pool->size);
drivers/gpu/drm/panthor/panthor_heap.c
276
u32 chunk_size,
drivers/gpu/drm/panthor/panthor_heap.c
294
if (!IS_ALIGNED(chunk_size, PAGE_SIZE) ||
drivers/gpu/drm/panthor/panthor_heap.c
295
chunk_size < SZ_128K || chunk_size > SZ_8M)
drivers/gpu/drm/panthor/panthor_heap.c
314
heap->chunk_size = chunk_size;
drivers/gpu/drm/panthor/panthor_heap.c
398
atomic_sub(heap->chunk_size, &pool->size);
drivers/gpu/drm/panthor/panthor_heap.c
484
(heap->chunk_size >> 12);
drivers/gpu/drm/panthor/panthor_heap.c
61
u32 chunk_size;
drivers/gpu/drm/panthor/panthor_heap.h
15
u32 chunk_size,
drivers/gpu/drm/qxl/qxl_image.c
38
unsigned int chunk_size)
drivers/gpu/drm/qxl/qxl_image.c
47
ret = qxl_alloc_bo_reserved(qdev, release, chunk_size, &chunk->bo);
drivers/gpu/drm/tests/drm_buddy_test.c
19
static inline u64 get_size(int order, u64 chunk_size)
drivers/gpu/drm/tests/drm_buddy_test.c
21
return (1 << order) * chunk_size;
drivers/gpu/drm/tests/drm_buddy_test.c
637
size = get_size(order, mm.chunk_size);
drivers/gpu/drm/tests/drm_buddy_test.c
651
size = get_size(0, mm.chunk_size);
drivers/gpu/drm/tests/drm_buddy_test.c
661
size = get_size(top, mm.chunk_size);
drivers/gpu/drm/tests/drm_buddy_test.c
672
size = get_size(order, mm.chunk_size);
drivers/gpu/drm/tests/drm_buddy_test.c
708
size = get_size(order, mm.chunk_size);
drivers/gpu/drm/tests/drm_buddy_test.c
721
size = get_size(0, mm.chunk_size);
drivers/gpu/drm/tests/drm_buddy_test.c
733
size = get_size(order, mm.chunk_size);
drivers/gpu/drm/tests/drm_buddy_test.c
749
size = get_size(order, mm.chunk_size);
drivers/gpu/drm/tests/drm_buddy_test.c
764
size = get_size(max_order, mm.chunk_size);
drivers/gpu/drm/tests/drm_buddy_test.c
803
size = get_size(order, mm.chunk_size);
drivers/gpu/drm/tests/drm_buddy_test.c
816
size = get_size(0, mm.chunk_size);
drivers/gpu/drm/tests/drm_buddy_test.c
839
size = mm.chunk_size << mm.max_order;
drivers/gpu/drm/tests/drm_buddy_test.c
841
mm.chunk_size, &allocated, flags));
drivers/gpu/drm/tests/drm_buddy_test.c
851
BIT_ULL(mm.max_order) * mm.chunk_size,
drivers/gpu/drm/tests/drm_buddy_test.c
854
BIT_ULL(mm.max_order) * mm.chunk_size);
drivers/gpu/drm/vmwgfx/vmwgfx_cmd.c
391
uint32_t chunk_size = max - next_cmd;
drivers/gpu/drm/vmwgfx/vmwgfx_cmd.c
396
if (bytes < chunk_size)
drivers/gpu/drm/vmwgfx/vmwgfx_cmd.c
397
chunk_size = bytes;
drivers/gpu/drm/vmwgfx/vmwgfx_cmd.c
401
memcpy(fifo_mem + (next_cmd >> 2), buffer, chunk_size);
drivers/gpu/drm/vmwgfx/vmwgfx_cmd.c
402
rest = bytes - chunk_size;
drivers/gpu/drm/vmwgfx/vmwgfx_cmd.c
404
memcpy(fifo_mem + (min >> 2), buffer + (chunk_size >> 2), rest);
drivers/gpu/drm/xe/xe_gt_sriov_pf_migration.c
606
size_t vram_size, chunk_size;
drivers/gpu/drm/xe/xe_gt_sriov_pf_migration.c
617
chunk_size = min(vram_size - *offset, VF_VRAM_STATE_CHUNK_MAX_SIZE);
drivers/gpu/drm/xe/xe_gt_sriov_pf_migration.c
619
ret = pf_save_vram_chunk(gt, vfid, vram, *offset, chunk_size);
drivers/gpu/drm/xe/xe_gt_sriov_pf_migration.c
623
*offset += chunk_size;
drivers/gpu/drm/xe/xe_gt_sriov_pf_service.c
319
const u32 chunk_size = hxg_sizeof(struct reg_data);
drivers/gpu/drm/xe/xe_gt_sriov_pf_service.c
340
max_chunks = (resp_size - VF2PF_QUERY_RUNTIME_RESPONSE_MSG_MIN_LEN) / chunk_size;
drivers/gpu/drm/xe/xe_ttm_vram_mgr.c
97
if (WARN_ON(min_page_size < mm->chunk_size)) {
drivers/infiniband/hw/hns/hns_roce_hem.c
208
u32 chunk_size;
drivers/infiniband/hw/hns/hns_roce_hem.c
224
chunk_size = table->type < HEM_TYPE_MTT ? mhop->buf_chunk_size :
drivers/infiniband/hw/hns/hns_roce_hem.c
226
table_idx = *obj / (chunk_size / table->obj_size);
drivers/infiniband/hw/irdma/ctrl.c
1483
if (info->chunk_size)
drivers/infiniband/hw/irdma/ctrl.c
1490
FIELD_PREP(IRDMA_CQPSQ_STAG_LPBLSIZE, info->chunk_size) |
drivers/infiniband/hw/irdma/ctrl.c
1549
if (info->chunk_size && info->first_pm_pbl_index >= pble_obj_cnt)
drivers/infiniband/hw/irdma/ctrl.c
1568
if (!info->chunk_size) {
drivers/infiniband/hw/irdma/ctrl.c
1582
FIELD_PREP(IRDMA_CQPSQ_STAG_LPBLSIZE, info->chunk_size) |
drivers/infiniband/hw/irdma/ctrl.c
1742
FIELD_PREP(IRDMAQPSQ_LPBLSIZE, info->chunk_size) |
drivers/infiniband/hw/irdma/type.h
1130
u32 chunk_size;
drivers/infiniband/hw/irdma/type.h
1157
u32 chunk_size;
drivers/infiniband/hw/irdma/type.h
1178
u32 chunk_size;
drivers/infiniband/hw/irdma/verbs.c
3288
stag_info->chunk_size = 1;
drivers/infiniband/hw/irdma/verbs.c
3291
stag_info->chunk_size = 3;
drivers/infiniband/hw/irdma/verbs.c
4120
stag_info.chunk_size = 1;
drivers/infiniband/hw/mthca/mthca_memfree.c
364
unsigned chunk_size;
drivers/infiniband/hw/mthca/mthca_memfree.c
386
chunk_size = MTHCA_TABLE_CHUNK_SIZE;
drivers/infiniband/hw/mthca/mthca_memfree.c
388
chunk_size = nobj * obj_size - i * MTHCA_TABLE_CHUNK_SIZE;
drivers/infiniband/hw/mthca/mthca_memfree.c
390
table->icm[i] = mthca_alloc_icm(dev, chunk_size >> PAGE_SHIFT,
drivers/infiniband/ulp/rtrs/rtrs-clt.c
1096
if (tsize > clt_path->chunk_size) {
drivers/infiniband/ulp/rtrs/rtrs-clt.c
1098
tsize, clt_path->chunk_size);
drivers/infiniband/ulp/rtrs/rtrs-clt.c
1186
if (tsize > clt_path->chunk_size) {
drivers/infiniband/ulp/rtrs/rtrs-clt.c
1189
tsize, clt_path->chunk_size);
drivers/infiniband/ulp/rtrs/rtrs-clt.c
1884
clt_path->chunk_size = clt_path->max_io_size + clt_path->max_hdr_size;
drivers/infiniband/ulp/rtrs/rtrs-clt.c
2452
if ((ilog2(sg_cnt - 1) + 1) + (ilog2(clt_path->chunk_size - 1) + 1) >
drivers/infiniband/ulp/rtrs/rtrs-clt.c
2456
MAX_IMM_PAYL_BITS, sg_cnt, clt_path->chunk_size);
drivers/infiniband/ulp/rtrs/rtrs-clt.c
2471
if (!len || (len % clt_path->chunk_size)) {
drivers/infiniband/ulp/rtrs/rtrs-clt.c
2481
len -= clt_path->chunk_size;
drivers/infiniband/ulp/rtrs/rtrs-clt.c
2482
addr += clt_path->chunk_size;
drivers/infiniband/ulp/rtrs/rtrs-clt.c
2491
if (total_len != clt_path->chunk_size * clt_path->queue_depth) {
drivers/infiniband/ulp/rtrs/rtrs-clt.h
139
u32 chunk_size;
drivers/input/evdev.c
716
size_t chunk_size = compat ? sizeof(compat_long_t) : sizeof(long);
drivers/input/evdev.c
719
if (maxlen % chunk_size)
drivers/input/evdev.c
723
len *= chunk_size;
drivers/input/touchscreen/wdt87xx_i2c.c
418
u32 chunk_id, chunk_size;
drivers/input/touchscreen/wdt87xx_i2c.c
426
chunk_size = get_unaligned_le32(fw->data +
drivers/input/touchscreen/wdt87xx_i2c.c
428
pos += chunk_size + 2 * sizeof(u32); /* chunk ID + size */
drivers/md/dm-exception-store.c
146
unsigned int chunk_size;
drivers/md/dm-exception-store.c
148
if (kstrtouint(chunk_size_arg, 10, &chunk_size)) {
drivers/md/dm-exception-store.c
153
if (!chunk_size) {
drivers/md/dm-exception-store.c
154
store->chunk_size = store->chunk_mask = store->chunk_shift = 0;
drivers/md/dm-exception-store.c
158
return dm_exception_store_set_chunk_size(store, chunk_size, error);
drivers/md/dm-exception-store.c
162
unsigned int chunk_size,
drivers/md/dm-exception-store.c
166
if (!is_power_of_2(chunk_size)) {
drivers/md/dm-exception-store.c
172
if (chunk_size %
drivers/md/dm-exception-store.c
174
chunk_size %
drivers/md/dm-exception-store.c
180
if (chunk_size > INT_MAX >> SECTOR_SHIFT) {
drivers/md/dm-exception-store.c
185
store->chunk_size = chunk_size;
drivers/md/dm-exception-store.c
186
store->chunk_mask = chunk_size - 1;
drivers/md/dm-exception-store.c
187
store->chunk_shift = __ffs(chunk_size);
drivers/md/dm-exception-store.h
122
unsigned int chunk_size;
drivers/md/dm-exception-store.h
185
unsigned int chunk_size,
drivers/md/dm-snap-persistent.c
175
len = ps->store->chunk_size << SECTOR_SHIFT;
drivers/md/dm-snap-persistent.c
237
.sector = ps->store->chunk_size * chunk,
drivers/md/dm-snap-persistent.c
238
.count = ps->store->chunk_size,
drivers/md/dm-snap-persistent.c
297
memset(ps->area, 0, ps->store->chunk_size << SECTOR_SHIFT);
drivers/md/dm-snap-persistent.c
310
unsigned int chunk_size;
drivers/md/dm-snap-persistent.c
318
if (!ps->store->chunk_size) {
drivers/md/dm-snap-persistent.c
319
ps->store->chunk_size = max(DM_CHUNK_SIZE_DEFAULT_SECTORS,
drivers/md/dm-snap-persistent.c
322
ps->store->chunk_mask = ps->store->chunk_size - 1;
drivers/md/dm-snap-persistent.c
323
ps->store->chunk_shift = __ffs(ps->store->chunk_size);
drivers/md/dm-snap-persistent.c
355
chunk_size = le32_to_cpu(dh->chunk_size);
drivers/md/dm-snap-persistent.c
357
if (ps->store->chunk_size == chunk_size)
drivers/md/dm-snap-persistent.c
362
chunk_size, ps->store->chunk_size);
drivers/md/dm-snap-persistent.c
367
r = dm_exception_store_set_chunk_size(ps->store, chunk_size,
drivers/md/dm-snap-persistent.c
371
chunk_size, chunk_err);
drivers/md/dm-snap-persistent.c
387
memset(ps->header_area, 0, ps->store->chunk_size << SECTOR_SHIFT);
drivers/md/dm-snap-persistent.c
393
dh->chunk_size = cpu_to_le32(ps->store->chunk_size);
drivers/md/dm-snap-persistent.c
498
ps->store->chunk_size << SECTOR_SHIFT,
drivers/md/dm-snap-persistent.c
546
memcpy(ps->area, area, ps->store->chunk_size << SECTOR_SHIFT);
drivers/md/dm-snap-persistent.c
580
*sectors_allocated = ps->next_free * store->chunk_size;
drivers/md/dm-snap-persistent.c
589
store->chunk_size;
drivers/md/dm-snap-persistent.c
627
ps->exceptions_per_area = (ps->store->chunk_size << SECTOR_SHIFT) /
drivers/md/dm-snap-persistent.c
680
if (size < ((ps->next_free + 1) * store->chunk_size))
drivers/md/dm-snap-persistent.c
84
__le32 chunk_size;
drivers/md/dm-snap-persistent.c
916
(unsigned long long)store->chunk_size);
drivers/md/dm-snap-transient.c
100
DMEMIT(" N %llu", (unsigned long long)store->chunk_size);
drivers/md/dm-snap-transient.c
48
if (size < (tc->next_free + store->chunk_size))
drivers/md/dm-snap-transient.c
52
tc->next_free += store->chunk_size;
drivers/md/dm-snap.c
1008
sector_t sector, unsigned int chunk_size);
drivers/md/dm-snap.c
1073
io_size = linear_chunks * s->store->chunk_size;
drivers/md/dm-snap.c
1369
s->store->chunk_size = 0;
drivers/md/dm-snap.c
1383
if (!s->store->chunk_size) {
drivers/md/dm-snap.c
1389
r = dm_set_target_max_io_len(ti, s->store->chunk_size);
drivers/md/dm-snap.c
1449
snap_dest->ti->max_io_len = snap_dest->store->chunk_size;
drivers/md/dm-snap.c
1796
src.count = min((sector_t)s->store->chunk_size, dev_size - src.sector);
drivers/md/dm-snap.c
1925
dest.count = s->store->chunk_size;
drivers/md/dm-snap.c
1936
(s->store->chunk_size << SECTOR_SHIFT);
drivers/md/dm-snap.c
2403
limits->discard_granularity = snap->store->chunk_size;
drivers/md/dm-snap.c
2404
limits->max_hw_discard_sectors = snap->store->chunk_size;
drivers/md/dm-snap.c
540
if (l->store->chunk_size < s->store->chunk_size)
drivers/md/dm-snap.c
849
unsigned int chunk_size = rounddown_pow_of_two(UINT_MAX);
drivers/md/dm-snap.c
853
chunk_size = min_not_zero(chunk_size,
drivers/md/dm-snap.c
854
snap->store->chunk_size);
drivers/md/dm-snap.c
856
return (uint32_t) chunk_size;
drivers/md/dm-stripe.c
107
if (kstrtouint(argv[1], 10, &chunk_size) || !chunk_size) {
drivers/md/dm-stripe.c
119
if (sector_div(tmp_len, chunk_size)) {
drivers/md/dm-stripe.c
150
r = dm_set_target_max_io_len(ti, chunk_size);
drivers/md/dm-stripe.c
162
sc->chunk_size = chunk_size;
drivers/md/dm-stripe.c
163
if (chunk_size & (chunk_size - 1))
drivers/md/dm-stripe.c
166
sc->chunk_size_shift = __ffs(chunk_size);
drivers/md/dm-stripe.c
209
chunk_offset = sector_div(chunk, sc->chunk_size);
drivers/md/dm-stripe.c
211
chunk_offset = chunk & (sc->chunk_size - 1);
drivers/md/dm-stripe.c
223
chunk *= sc->chunk_size;
drivers/md/dm-stripe.c
242
*result -= sector_div(sector, sc->chunk_size);
drivers/md/dm-stripe.c
244
*result = sector & ~(sector_t)(sc->chunk_size - 1);
drivers/md/dm-stripe.c
247
*result += sc->chunk_size; /* next chunk */
drivers/md/dm-stripe.c
38
uint32_t chunk_size;
drivers/md/dm-stripe.c
381
(unsigned long long)sc->chunk_size);
drivers/md/dm-stripe.c
390
(unsigned long long)sc->chunk_size);
drivers/md/dm-stripe.c
461
limits->chunk_sectors = sc->chunk_size;
drivers/md/dm-stripe.c
463
if (!check_shl_overflow(sc->chunk_size, SECTOR_SHIFT, &io_min) &&
drivers/md/dm-stripe.c
468
if (max_hw_discard_sectors >= sc->chunk_size) {
drivers/md/dm-stripe.c
471
sc->chunk_size * sc->stripes);
drivers/md/dm-stripe.c
93
uint32_t chunk_size;
drivers/md/dm-unstripe.c
127
sector_div(tmp_sector, uc->chunk_size);
drivers/md/dm-unstripe.c
157
uc->stripes, (unsigned long long)uc->chunk_size, uc->unstripe,
drivers/md/dm-unstripe.c
180
limits->chunk_sectors = uc->chunk_size;
drivers/md/dm-unstripe.c
22
uint32_t chunk_size;
drivers/md/dm-unstripe.c
62
if (kstrtouint(argv[1], 10, &uc->chunk_size) || !uc->chunk_size) {
drivers/md/dm-unstripe.c
88
uc->unstripe_offset = (sector_t)uc->unstripe * uc->chunk_size;
drivers/md/dm-unstripe.c
89
uc->unstripe_width = (sector_t)(uc->stripes - 1) * uc->chunk_size;
drivers/md/dm-unstripe.c
90
uc->chunk_shift = is_power_of_2(uc->chunk_size) ? fls(uc->chunk_size) - 1 : 0;
drivers/md/dm-unstripe.c
93
if (sector_div(tmp_len, uc->chunk_size)) {
drivers/md/dm-unstripe.c
98
if (dm_set_target_max_io_len(ti, uc->chunk_size)) {
drivers/md/dm-vdo/indexer/io-factory.c
220
size_t chunk_size;
drivers/md/dm-vdo/indexer/io-factory.c
227
chunk_size = min(length, bytes_remaining_in_read_buffer(reader));
drivers/md/dm-vdo/indexer/io-factory.c
228
memcpy(data, reader->end, chunk_size);
drivers/md/dm-vdo/indexer/io-factory.c
229
length -= chunk_size;
drivers/md/dm-vdo/indexer/io-factory.c
230
data += chunk_size;
drivers/md/dm-vdo/indexer/io-factory.c
231
reader->end += chunk_size;
drivers/md/dm-vdo/indexer/io-factory.c
245
size_t chunk_size;
drivers/md/dm-vdo/indexer/io-factory.c
256
chunk_size = min(length, bytes_remaining_in_read_buffer(reader));
drivers/md/dm-vdo/indexer/io-factory.c
257
if (memcmp(value, reader->end, chunk_size) != 0) {
drivers/md/dm-vdo/indexer/io-factory.c
262
length -= chunk_size;
drivers/md/dm-vdo/indexer/io-factory.c
263
value += chunk_size;
drivers/md/dm-vdo/indexer/io-factory.c
264
reader->end += chunk_size;
drivers/md/dm-vdo/indexer/io-factory.c
383
size_t chunk_size;
drivers/md/dm-vdo/indexer/io-factory.c
391
chunk_size = min(length, get_remaining_write_space(writer));
drivers/md/dm-vdo/indexer/io-factory.c
393
memset(writer->end, 0, chunk_size);
drivers/md/dm-vdo/indexer/io-factory.c
395
memcpy(writer->end, data, chunk_size);
drivers/md/dm-vdo/indexer/io-factory.c
396
data += chunk_size;
drivers/md/dm-vdo/indexer/io-factory.c
399
length -= chunk_size;
drivers/md/dm-vdo/indexer/io-factory.c
400
writer->end += chunk_size;
drivers/md/md-autodetect.c
196
ainfo.chunk_size = args->chunk;
drivers/md/md.c
1488
mddev->chunk_sectors = sb->chunk_size >> 9;
drivers/md/md.c
1664
sb->chunk_size = mddev->chunk_sectors << 9;
drivers/md/md.c
4490
__ATTR(chunk_size, S_IRUGO|S_IWUSR, chunk_size_show, chunk_size_store);
drivers/md/md.c
7227
info.chunk_size = mddev->chunk_sectors << 9;
drivers/md/md.c
7767
mddev->chunk_sectors = info->chunk_size >> 9;
drivers/md/md.c
7936
mddev->chunk_sectors != info->chunk_size >> 9 ||
drivers/media/dvb-frontends/dib9000.c
235
u32 chunk_size = 126;
drivers/media/dvb-frontends/dib9000.c
261
l = min(len, chunk_size);
drivers/media/dvb-frontends/dib9000.c
319
u32 chunk_size = 126;
drivers/media/dvb-frontends/dib9000.c
345
l = min(len, chunk_size);
drivers/media/dvb-frontends/drxk.h
49
int chunk_size;
drivers/media/dvb-frontends/drxk_hard.c
6737
state->m_chunk_size = config->chunk_size;
drivers/media/i2c/thp7312.c
1681
unsigned int chunk_size;
drivers/media/i2c/thp7312.c
1684
chunk_size = min_t(u32, size, THP7312_FW_DOWNLOAD_UNIT);
drivers/media/i2c/thp7312.c
1687
data, chunk_size);
drivers/media/i2c/thp7312.c
1694
data += chunk_size;
drivers/media/i2c/thp7312.c
1695
size -= chunk_size;
drivers/media/i2c/vgxy61.c
513
const unsigned int chunk_size = get_chunk_size(sensor);
drivers/media/i2c/vgxy61.c
518
sz = min(nb, chunk_size);
drivers/media/usb/dvb-usb-v2/az6007.c
57
.chunk_size = 64,
drivers/media/usb/dvb-usb-v2/az6007.c
70
.chunk_size = 64,
drivers/media/usb/em28xx/em28xx-dvb.c
413
.chunk_size = 56,
drivers/media/usb/em28xx/em28xx-dvb.c
422
.chunk_size = 54,
drivers/media/usb/em28xx/em28xx-dvb.c
435
.chunk_size = 62,
drivers/media/usb/em28xx/em28xx-dvb.c
444
.chunk_size = 58,
drivers/mtd/nand/raw/marvell_nand.c
1265
int chunk_size = lt->data_bytes + lt->spare_bytes + lt->ecc_bytes;
drivers/mtd/nand/raw/marvell_nand.c
1289
nand_change_read_column_op(chip, chunk * chunk_size,
drivers/net/ethernet/chelsio/cxgb3/t3_hw.c
1085
unsigned int chunk_size = min(size, 256U);
drivers/net/ethernet/chelsio/cxgb3/t3_hw.c
1087
ret = t3_write_flash(adapter, addr, chunk_size, fw_data);
drivers/net/ethernet/chelsio/cxgb3/t3_hw.c
1091
addr += chunk_size;
drivers/net/ethernet/chelsio/cxgb3/t3_hw.c
1092
fw_data += chunk_size;
drivers/net/ethernet/chelsio/cxgb3/t3_hw.c
1093
size -= chunk_size;
drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c
3538
u32 chunk_size = min_t(u32, tot_len, CUDBG_CHUNK_SIZE);
drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c
3540
rc = cudbg_get_buff(pdbg_init, dbg_buff, chunk_size,
drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c
3547
memcpy(temp_buff.data, data + cur_off, chunk_size);
drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c
3548
tot_len -= chunk_size;
drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c
3549
cur_off += chunk_size;
drivers/net/ethernet/emulex/benet/be_cmds.c
3056
u32 chunk_size = 0;
drivers/net/ethernet/emulex/benet/be_cmds.c
3081
chunk_size = min_t(u32, image_size, LANCER_FW_DOWNLOAD_CHUNK);
drivers/net/ethernet/emulex/benet/be_cmds.c
3084
memcpy(dest_image_ptr, data_ptr, chunk_size);
drivers/net/ethernet/emulex/benet/be_cmds.c
3087
chunk_size, offset,
drivers/net/ethernet/emulex/benet/be_ethtool.c
266
u32 read_len = 0, total_read_len = 0, chunk_size;
drivers/net/ethernet/emulex/benet/be_ethtool.c
282
chunk_size = min_t(u32, (buf_len - total_read_len),
drivers/net/ethernet/emulex/benet/be_ethtool.c
284
chunk_size = ALIGN(chunk_size, 4);
drivers/net/ethernet/emulex/benet/be_ethtool.c
285
status = lancer_cmd_read_object(adapter, &read_cmd, chunk_size,
drivers/net/ethernet/intel/ice/ice_adminq_cmd.h
925
u8 chunk_size;
drivers/net/ethernet/mellanox/mlx4/icm.c
420
unsigned chunk_size;
drivers/net/ethernet/mellanox/mlx4/icm.c
442
chunk_size = MLX4_TABLE_CHUNK_SIZE;
drivers/net/ethernet/mellanox/mlx4/icm.c
444
chunk_size = PAGE_ALIGN(size -
drivers/net/ethernet/mellanox/mlx4/icm.c
447
table->icm[i] = mlx4_alloc_icm(dev, chunk_size >> PAGE_SHIFT,
drivers/net/ethernet/mellanox/mlx5/core/en/params.c
1129
.chunk_size = 1 << frame_shift,
drivers/net/ethernet/mellanox/mlx5/core/en/params.c
1143
xsk.chunk_size -= 1;
drivers/net/ethernet/mellanox/mlx5/core/en/params.c
1148
xsk.chunk_size = (1 << frame_shift) / 4 * 3;
drivers/net/ethernet/mellanox/mlx5/core/en/params.c
26
u8 req_page_shift = xsk ? order_base_2(xsk->chunk_size) : PAGE_SHIFT;
drivers/net/ethernet/mellanox/mlx5/core/en/params.c
322
if (xsk && mlx5e_rx_get_linear_sz_xsk(params, xsk) > xsk->chunk_size)
drivers/net/ethernet/mellanox/mlx5/core/en/params.c
53
oversized = xsk->chunk_size < (1 << page_shift);
drivers/net/ethernet/mellanox/mlx5/core/en/params.c
54
WARN_ON_ONCE(xsk->chunk_size > (1 << page_shift));
drivers/net/ethernet/mellanox/mlx5/core/en/params.c
562
1 << params->log_rq_mtu_frames, xsk->chunk_size);
drivers/net/ethernet/mellanox/mlx5/core/en/params.c
70
if (xsk->chunk_size % 3 == 0 && is_power_of_2(xsk->chunk_size / 3))
drivers/net/ethernet/mellanox/mlx5/core/en/params.h
11
u16 chunk_size;
drivers/net/ethernet/mellanox/mlx5/core/en/xsk/pool.c
103
const char *recommendation = is_power_of_2(xsk.chunk_size) ?
drivers/net/ethernet/mellanox/mlx5/core/en/xsk/pool.c
107
xsk.chunk_size, recommendation);
drivers/net/ethernet/mellanox/mlx5/core/en/xsk/pool.c
73
xsk->chunk_size = xsk_pool_get_chunk_size(pool);
drivers/net/ethernet/mellanox/mlx5/core/en/xsk/rx.c
104
rq->xsk_pool->chunk_size);
drivers/net/ethernet/mellanox/mlx5/core/en/xsk/rx.c
105
__be32 frame_size = cpu_to_be32(rq->xsk_pool->chunk_size);
drivers/net/ethernet/mellanox/mlx5/core/en/xsk/setup.c
34
if ((size_t)xsk->chunk_size > PAGE_SIZE || xsk->chunk_size < MLX5E_MIN_XSK_CHUNK_SIZE) {
drivers/net/ethernet/mellanox/mlx5/core/en/xsk/setup.c
35
mlx5_core_err(mdev, "XSK chunk size %u out of bounds [%u, %lu]\n", xsk->chunk_size,
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
4608
max_mtu_frame = MLX5E_HW2SW_MTU(new_params, xsk.chunk_size - hr);
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
497
u32 xsk_chunk_size = rq->xsk_pool ? rq->xsk_pool->chunk_size : 0;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_arg.c
131
enum dr_arg_chunk_size chunk_size)
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_arg.c
144
pool->log_chunk_size = chunk_size;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_icm_pool.c
322
enum mlx5dr_icm_chunk_size chunk_size,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_icm_pool.c
329
chunk->size = chunk_size;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_icm_pool.c
386
enum mlx5dr_icm_chunk_size chunk_size,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_icm_pool.c
398
chunk_size, seg);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_icm_pool.c
406
chunk_size);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_icm_pool.c
416
chunk_size);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_icm_pool.c
435
enum mlx5dr_icm_chunk_size chunk_size)
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_icm_pool.c
442
if (chunk_size > pool->max_log_chunk_sz)
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_icm_pool.c
447
ret = dr_icm_handle_buddies_get_mem(pool, chunk_size, &buddy, &seg);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_icm_pool.c
455
dr_icm_chunk_init(chunk, pool, chunk_size, buddy, seg);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_icm_pool.c
460
mlx5dr_buddy_free_mem(buddy, seg, chunk_size);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ptrn.c
82
u32 chunk_size;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ptrn.c
85
chunk_size = ilog2(roundup_pow_of_two(num_of_actions));
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ptrn.c
87
chunk_size = max_t(u32, chunk_size, DR_CHUNK_SIZE_8);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ptrn.c
89
chunk = mlx5dr_icm_alloc_chunk(mgr->ptrn_icm_pool, chunk_size);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
497
enum mlx5dr_icm_chunk_size chunk_size,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
509
chunk = mlx5dr_icm_alloc_chunk(pool, chunk_size);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
641
u32 chunk_size;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
644
chunk_size = ilog2(roundup_pow_of_two(action->rewrite->num_of_actions));
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
647
chunk_size = max_t(u32, chunk_size, DR_CHUNK_SIZE_8);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
650
chunk_size);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
1239
mlx5dr_icm_pool_chunk_size_to_entries(enum mlx5dr_icm_chunk_size chunk_size)
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
1241
return 1 << chunk_size;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
1245
mlx5dr_icm_pool_chunk_size_to_byte(enum mlx5dr_icm_chunk_size chunk_size,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
1252
num_of_entries = mlx5dr_icm_pool_chunk_size_to_entries(chunk_size);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
1393
enum mlx5dr_icm_chunk_size chunk_size);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
230
enum mlx5dr_icm_chunk_size chunk_size,
drivers/net/ethernet/mellanox/mlxsw/i2c.c
337
int off = mlxsw_i2c->cmd.mb_off_in, chunk_size, i, j;
drivers/net/ethernet/mellanox/mlxsw/i2c.c
351
chunk_size = (in_mbox_size > mlxsw_i2c->block_size) ?
drivers/net/ethernet/mellanox/mlxsw/i2c.c
353
write_tran.len = MLXSW_I2C_ADDR_WIDTH + chunk_size;
drivers/net/ethernet/mellanox/mlxsw/i2c.c
356
mlxsw_i2c->block_size * i, chunk_size);
drivers/net/ethernet/mellanox/mlxsw/i2c.c
376
off += chunk_size;
drivers/net/ethernet/mellanox/mlxsw/i2c.c
377
in_mbox_size -= chunk_size;
drivers/net/ethernet/mellanox/mlxsw/i2c.c
416
int num, chunk_size, reg_size, i, j;
drivers/net/ethernet/mellanox/mlxsw/i2c.c
462
chunk_size = (reg_size > mlxsw_i2c->block_size) ?
drivers/net/ethernet/mellanox/mlxsw/i2c.c
464
read_tran[1].len = chunk_size;
drivers/net/ethernet/mellanox/mlxsw/i2c.c
486
off += chunk_size;
drivers/net/ethernet/mellanox/mlxsw/i2c.c
487
reg_size -= chunk_size;
drivers/net/ethernet/mellanox/mlxsw/i2c.c
488
read_tran[1].buf += chunk_size;
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_nsp.c
506
size_t chunk_size, dma_size;
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_nsp.c
513
chunk_size = BIT_ULL(chunk_order);
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_nsp.c
515
nseg = DIV_ROUND_UP(max_size, chunk_size);
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_nsp.c
526
chunks[i].chunk = kmalloc(chunk_size,
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_nsp.c
531
chunks[i].len = min_t(u64, chunk_size, max_size - off);
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_nsp.c
535
coff = min_t(u64, arg->in_size - off, chunk_size);
drivers/net/ethernet/netronome/nfp/nfpcore/nfp_nsp.c
538
memset(chunks[i].chunk + coff, 0, chunk_size - coff);
drivers/net/wireless/ath/wcn36xx/dxe.c
239
int i, chunk_size = pool->chunk_size;
drivers/net/wireless/ath/wcn36xx/dxe.c
250
bd_phy_addr += chunk_size;
drivers/net/wireless/ath/wcn36xx/dxe.c
251
bd_cpu_addr += chunk_size;
drivers/net/wireless/ath/wcn36xx/dxe.c
694
wcn->mgmt_mem_pool.chunk_size = WCN36XX_BD_CHUNK_SIZE +
drivers/net/wireless/ath/wcn36xx/dxe.c
697
s = wcn->mgmt_mem_pool.chunk_size * WCN36XX_DXE_CH_DESC_NUMB_TX_H;
drivers/net/wireless/ath/wcn36xx/dxe.c
709
wcn->data_mem_pool.chunk_size = WCN36XX_BD_CHUNK_SIZE +
drivers/net/wireless/ath/wcn36xx/dxe.c
712
s = wcn->data_mem_pool.chunk_size * WCN36XX_DXE_CH_DESC_NUMB_TX_L;
drivers/net/wireless/ath/wcn36xx/dxe.c
732
dma_free_coherent(wcn->dev, wcn->mgmt_mem_pool.chunk_size *
drivers/net/wireless/ath/wcn36xx/dxe.c
738
dma_free_coherent(wcn->dev, wcn->data_mem_pool.chunk_size *
drivers/net/wireless/ath/wcn36xx/dxe.h
449
int chunk_size;
drivers/net/wireless/intel/iwlwifi/mvm/nvm.c
153
int chunk_size, ret;
drivers/net/wireless/intel/iwlwifi/mvm/nvm.c
155
chunk_size = min(IWL_NVM_DEFAULT_CHUNK_SIZE,
drivers/net/wireless/intel/iwlwifi/mvm/nvm.c
159
chunk_size, data + offset);
drivers/net/wireless/intel/iwlwifi/mvm/nvm.c
163
offset += chunk_size;
drivers/net/wireless/marvell/libertas/if_sdio.c
446
u32 chunk_size;
drivers/net/wireless/marvell/libertas/if_sdio.c
475
chunk_size = min_t(size_t, size, 60);
drivers/net/wireless/marvell/libertas/if_sdio.c
477
*((__le32*)chunk_buffer) = cpu_to_le32(chunk_size);
drivers/net/wireless/marvell/libertas/if_sdio.c
478
memcpy(chunk_buffer + 4, firmware, chunk_size);
drivers/net/wireless/marvell/libertas/if_sdio.c
487
firmware += chunk_size;
drivers/net/wireless/marvell/libertas/if_sdio.c
488
size -= chunk_size;
drivers/net/wireless/marvell/libertas/if_sdio.c
542
u32 chunk_size;
drivers/net/wireless/marvell/libertas/if_sdio.c
613
chunk_size = min_t(size_t, req_size, 512);
drivers/net/wireless/marvell/libertas/if_sdio.c
615
memcpy(chunk_buffer, firmware, chunk_size);
drivers/net/wireless/marvell/libertas/if_sdio.c
621
chunk_buffer, roundup(chunk_size, 32));
drivers/net/wireless/marvell/libertas/if_sdio.c
625
firmware += chunk_size;
drivers/net/wireless/marvell/libertas/if_sdio.c
626
size -= chunk_size;
drivers/net/wireless/marvell/libertas/if_sdio.c
627
req_size -= chunk_size;
drivers/ntb/test/ntb_perf.c
925
u64 total_size, chunk_size;
drivers/ntb/test/ntb_perf.c
930
chunk_size = 1ULL << chunk_order;
drivers/ntb/test/ntb_perf.c
931
chunk_size = min_t(u64, peer->outbuf_size, chunk_size);
drivers/ntb/test/ntb_perf.c
941
ret = perf_copy_chunk(pthr, flt_dst, flt_src, chunk_size);
drivers/ntb/test/ntb_perf.c
948
pthr->copied += chunk_size;
drivers/ntb/test/ntb_perf.c
950
flt_dst += chunk_size;
drivers/ntb/test/ntb_perf.c
951
flt_src += chunk_size;
drivers/nvdimm/btt.c
413
size_t chunk_size = SZ_2M;
drivers/nvdimm/btt.c
416
zerobuf = kzalloc(chunk_size, GFP_KERNEL);
drivers/nvdimm/btt.c
429
size_t size = min(mapsize, chunk_size);
drivers/nvdimm/btt.c
455
size_t chunk_size = SZ_4K, offset = 0;
drivers/nvdimm/btt.c
461
zerobuf = kzalloc(chunk_size, GFP_KERNEL);
drivers/nvdimm/btt.c
473
size_t size = min(logsize, chunk_size);
drivers/nvme/host/pci.c
2575
u32 chunk_size)
drivers/nvme/host/pci.c
2584
tmp = (preferred + chunk_size - 1);
drivers/nvme/host/pci.c
2585
do_div(tmp, chunk_size);
drivers/nvme/host/pci.c
2604
len = min_t(u64, chunk_size, preferred - size);
drivers/nvme/host/pci.c
2640
u64 chunk_size;
drivers/nvme/host/pci.c
2652
for (chunk_size = min_chunk; chunk_size >= hmminds; chunk_size /= 2) {
drivers/nvme/host/pci.c
2653
if (!nvme_alloc_host_mem_multi(dev, preferred, chunk_size)) {
drivers/platform/x86/intel/ifs/ifs.h
188
u32 chunk_size :16;
drivers/platform/x86/intel/ifs/ifs.h
201
u16 chunk_size;
drivers/platform/x86/intel/ifs/ifs.h
370
u32 chunk_size;
drivers/platform/x86/intel/ifs/load.c
123
int i, num_chunks, chunk_size;
drivers/platform/x86/intel/ifs/load.c
136
chunk_size = hashes_status.chunk_size * 1024;
drivers/platform/x86/intel/ifs/load.c
150
linear_addr = base + i * chunk_size;
drivers/platform/x86/intel/ifs/load.c
187
int i, num_chunks, chunk_size;
drivers/platform/x86/intel/ifs/load.c
203
chunk_size = hashes_status.chunk_size * SZ_1K;
drivers/platform/x86/intel/ifs/load.c
213
ifsd->chunk_size = chunk_size;
drivers/platform/x86/intel/ifs/load.c
216
chunk_size = ifsd->chunk_size;
drivers/platform/x86/intel/ifs/load.c
236
linear_addr = base + i * chunk_size;
drivers/rpmsg/qcom_glink_native.c
1011
qcom_glink_rx_advance(glink, ALIGN(sizeof(hdr) + chunk_size, 8));
drivers/rpmsg/qcom_glink_native.c
1485
__le32 chunk_size;
drivers/rpmsg/qcom_glink_native.c
1490
int chunk_size = len;
drivers/rpmsg/qcom_glink_native.c
1526
chunk_size = len - offset;
drivers/rpmsg/qcom_glink_native.c
1527
if (chunk_size > SZ_8K && wait)
drivers/rpmsg/qcom_glink_native.c
1528
chunk_size = SZ_8K;
drivers/rpmsg/qcom_glink_native.c
1533
req.chunk_size = cpu_to_le32(chunk_size);
drivers/rpmsg/qcom_glink_native.c
1534
req.left_size = cpu_to_le32(len - offset - chunk_size);
drivers/rpmsg/qcom_glink_native.c
1538
iid, chunk_size,
drivers/rpmsg/qcom_glink_native.c
1539
len - offset - chunk_size,
drivers/rpmsg/qcom_glink_native.c
1542
ret = qcom_glink_tx(glink, &req, sizeof(req), data + offset, chunk_size, wait);
drivers/rpmsg/qcom_glink_native.c
1550
offset += chunk_size;
drivers/rpmsg/qcom_glink_native.c
904
__le32 chunk_size;
drivers/rpmsg/qcom_glink_native.c
907
unsigned int chunk_size;
drivers/rpmsg/qcom_glink_native.c
920
chunk_size = le32_to_cpu(hdr.chunk_size);
drivers/rpmsg/qcom_glink_native.c
923
if (avail < sizeof(hdr) + chunk_size) {
drivers/rpmsg/qcom_glink_native.c
936
liid, chunk_size, left_size,
drivers/rpmsg/qcom_glink_native.c
952
intent->data = kmalloc(chunk_size + left_size,
drivers/rpmsg/qcom_glink_native.c
960
intent->size = chunk_size + left_size;
drivers/rpmsg/qcom_glink_native.c
981
if (intent->size - intent->offset < chunk_size) {
drivers/rpmsg/qcom_glink_native.c
989
sizeof(hdr), chunk_size);
drivers/rpmsg/qcom_glink_native.c
990
intent->offset += chunk_size;
drivers/rpmsg/qcom_glink_trace.h
281
TP_PROTO(const char *remote, const char *channel, u16 lcid, u16 rcid, u32 iid, u32 chunk_size, u32 left_size, bool cont, bool tx),
drivers/rpmsg/qcom_glink_trace.h
282
TP_ARGS(remote, channel, lcid, rcid, iid, chunk_size, left_size, cont, tx),
drivers/rpmsg/qcom_glink_trace.h
289
__field(u32, chunk_size)
drivers/rpmsg/qcom_glink_trace.h
300
__entry->chunk_size = chunk_size;
drivers/rpmsg/qcom_glink_trace.h
312
__entry->chunk_size,
drivers/rtc/rtc-isl12026.c
327
size_t chunk_size, num_written;
drivers/rtc/rtc-isl12026.c
347
chunk_size = round_down(offset, ISL12026_PAGESIZE) +
drivers/rtc/rtc-isl12026.c
349
chunk_size = min(bytes, chunk_size);
drivers/rtc/rtc-isl12026.c
354
memcpy(payload + 2, v + num_written, chunk_size);
drivers/rtc/rtc-isl12026.c
357
msgs[0].len = chunk_size + 2;
drivers/rtc/rtc-isl12026.c
367
bytes -= chunk_size;
drivers/rtc/rtc-isl12026.c
368
offset += chunk_size;
drivers/rtc/rtc-isl12026.c
369
num_written += chunk_size;
drivers/s390/cio/css.c
1090
size_t chunk_size = chunk->end_addr - chunk->start_addr + 1;
drivers/s390/cio/css.c
1092
dma_free_coherent((struct device *) data, chunk_size,
drivers/s390/cio/css.c
1119
size_t chunk_size;
drivers/s390/cio/css.c
1126
chunk_size = round_up(size, PAGE_SIZE);
drivers/s390/cio/css.c
1127
addr = dma_alloc_coherent(dma_dev, chunk_size, &dma_addr, CIO_DMA_GFP);
drivers/s390/cio/css.c
1130
gen_pool_add_virt(gp_dma, (unsigned long)addr, dma_addr, chunk_size, -1);
drivers/scsi/megaraid/megaraid_sas_fusion.c
793
u32 chunk_size, array_size, offset;
drivers/scsi/megaraid/megaraid_sas_fusion.c
796
chunk_size = fusion->reply_alloc_sz * RDPQ_MAX_INDEX_IN_ONE_CHUNK;
drivers/scsi/megaraid/megaraid_sas_fusion.c
814
chunk_size, 16, 0);
drivers/scsi/megaraid/megaraid_sas_fusion.c
818
chunk_size,
drivers/scsi/megaraid/megaraid_sas_fusion.c
819
roundup_pow_of_two(chunk_size),
drivers/scsi/megaraid/megaraid_sas_fusion.c
861
chunk_size)) {
drivers/scsi/qla2xxx/qla_nx2.c
2902
uint32_t chunk_size, read_size;
drivers/scsi/qla2xxx/qla_nx2.c
2938
dma_desc.cmd.read_data_size = chunk_size = ISP8044_PEX_DMA_READ_SIZE;
drivers/scsi/qla2xxx/qla_nx2.c
2948
chunk_size = (m_hdr->read_data_size - read_size);
drivers/scsi/qla2xxx/qla_nx2.c
2949
dma_desc.cmd.read_data_size = chunk_size;
drivers/scsi/qla2xxx/qla_nx2.c
2966
"(chunk_size 0x%x).\n", __func__, chunk_size);
drivers/scsi/qla2xxx/qla_nx2.c
2973
memcpy(data_ptr, rdmem_buffer, chunk_size);
drivers/scsi/qla2xxx/qla_nx2.c
2974
data_ptr += chunk_size;
drivers/scsi/qla2xxx/qla_nx2.c
2975
read_size += chunk_size;
drivers/scsi/smartpqi/smartpqi_init.c
8968
u32 total_size, u32 chunk_size)
drivers/scsi/smartpqi/smartpqi_init.c
8977
sg_count = DIV_ROUND_UP(total_size, chunk_size);
drivers/scsi/smartpqi/smartpqi_init.c
8990
host_memory_descriptor->host_chunk_virt_address[i] = dma_alloc_coherent(dev, chunk_size, &dma_handle, GFP_KERNEL);
drivers/scsi/smartpqi/smartpqi_init.c
8995
put_unaligned_le32(chunk_size, &mem_descriptor->length);
drivers/scsi/smartpqi/smartpqi_init.c
9000
put_unaligned_le32(sg_count * chunk_size, &host_memory->bytes_allocated);
drivers/scsi/smartpqi/smartpqi_init.c
9007
dma_free_coherent(dev, chunk_size,
drivers/scsi/smartpqi/smartpqi_init.c
9020
u32 chunk_size;
drivers/scsi/smartpqi/smartpqi_init.c
9032
for (chunk_size = total_required_size; chunk_size >= min_chunk_size;) {
drivers/scsi/smartpqi/smartpqi_init.c
9035
chunk_size) == 0)
drivers/scsi/smartpqi/smartpqi_init.c
9037
chunk_size /= 2;
drivers/scsi/smartpqi/smartpqi_init.c
9038
chunk_size = PAGE_ALIGN(chunk_size);
drivers/ufs/core/ufs-sysfs.c
775
return sysfs_emit(buf, "%lu\n", hba->monitor.chunk_size);
drivers/ufs/core/ufs-sysfs.c
791
hba->monitor.chunk_size = value;
drivers/ufs/core/ufshcd.c
2308
(!m->chunk_size || m->chunk_size == cmd->sdb.length) &&
drivers/vfio/pci/mlx5/main.c
390
size_t chunk_size = min_t(size_t, MAX_CHUNK_SIZE, full_size);
drivers/vfio/pci/mlx5/main.c
393
inc_state_size = max(state_size, chunk_size);
drivers/vfio/pci/virtio/migrate.c
810
size_t chunk_size, const char __user **buf,
drivers/vfio/pci/virtio/migrate.c
817
to_copy = min_t(size_t, *len, chunk_size - vhca_buf->length);
drivers/vfio/pci/virtio/migrate.c
827
if (vhca_buf->length == chunk_size) {
drivers/vfio/pci/virtio/migrate.c
829
migf->max_pos += chunk_size;
drivers/virtio/virtio_mem.c
2752
uint64_t chunk_size, cur_start, cur_end, plugged_range_start = 0;
drivers/virtio/virtio_mem.c
2760
chunk_size = max_t(uint64_t, PFN_PHYS(PAGES_PER_SECTION),
drivers/virtio/virtio_mem.c
2773
cur_end = ALIGN_DOWN(cur_start + chunk_size, chunk_size);
fs/btrfs/block-group.c
2816
u64 chunk_offset, u64 chunk_size)
fs/btrfs/block-group.c
2825
map = btrfs_get_chunk_map(fs_info, chunk_offset, chunk_size);
fs/btrfs/space-info.c
232
u64 chunk_size)
fs/btrfs/space-info.c
234
WRITE_ONCE(space_info->chunk_size, chunk_size);
fs/btrfs/space-info.c
437
return data_sinfo->chunk_size;
fs/btrfs/space-info.c
438
data_chunk_size = min(data_sinfo->chunk_size,
fs/btrfs/space-info.h
133
u64 chunk_size;
fs/btrfs/space-info.h
276
u64 chunk_size);
fs/btrfs/sysfs.c
1036
BTRFS_ATTR_PTR(space_info, chunk_size),
fs/btrfs/sysfs.c
756
return sysfs_emit(buf, "%llu\n", READ_ONCE(sinfo->chunk_size));
fs/btrfs/sysfs.c
909
BTRFS_ATTR_RW(space_info, chunk_size, btrfs_chunk_size_show, btrfs_chunk_size_store);
fs/btrfs/volumes.c
5433
u64 chunk_size;
fs/btrfs/volumes.c
5448
ctl->max_chunk_size = READ_ONCE(space_info->chunk_size);
fs/btrfs/volumes.c
5641
ctl->chunk_size = ctl->stripe_size * data_stripes;
fs/btrfs/volumes.c
5677
ctl->chunk_size = ctl->stripe_size * data_stripes;
fs/btrfs/volumes.c
5819
map->chunk_len = ctl->chunk_size;
fs/btrfs/volumes.c
5836
trace_btrfs_chunk_alloc(info, map, start, ctl->chunk_size);
fs/btrfs/volumes.c
5845
ctl->chunk_size);
fs/ext2/dir.c
104
unsigned chunk_size = ext2_chunk_size(dir);
fs/ext2/dir.c
113
if (limit & (chunk_size - 1))
fs/ext2/dir.c
128
if (unlikely(((offs + rec_len - 1) ^ offs) & ~(chunk_size-1)))
fs/ext2/dir.c
482
unsigned chunk_size = ext2_chunk_size(dir);
fs/ext2/dir.c
511
rec_len = chunk_size;
fs/ext2/dir.c
512
de->rec_len = ext2_rec_len_to_disk(chunk_size);
fs/ext2/dir.c
621
unsigned chunk_size = ext2_chunk_size(inode);
fs/ext2/dir.c
629
err = ext2_prepare_chunk(folio, 0, chunk_size);
fs/ext2/dir.c
635
memset(kaddr, 0, chunk_size);
fs/ext2/dir.c
645
de->rec_len = ext2_rec_len_to_disk(chunk_size - EXT2_DIR_REC_LEN(1));
fs/ext2/dir.c
650
ext2_commit_chunk(folio, 0, chunk_size);
fs/ext4/mballoc.c
4560
#define NRL_CHECK_SIZE(req, size, max, chunk_size) \
fs/ext4/mballoc.c
4561
(req <= (size) || max <= (chunk_size))
fs/gfs2/rgrp.c
380
u32 chunk_size;
fs/gfs2/rgrp.c
402
chunk_size = ((ptr == NULL) ? bytes : (ptr - start));
fs/gfs2/rgrp.c
403
chunk_size *= GFS2_NBBY;
fs/gfs2/rgrp.c
404
BUG_ON(len < chunk_size);
fs/gfs2/rgrp.c
405
len -= chunk_size;
fs/gfs2/rgrp.c
407
if (gfs2_rbm_from_block(&rbm, block + chunk_size)) {
fs/minix/dir.c
109
limit = kaddr + minix_last_byte(inode, n) - chunk_size;
fs/minix/dir.c
130
ctx->pos += chunk_size;
fs/minix/dir.c
88
unsigned chunk_size = sbi->s_dirsize;
fs/minix/dir.c
94
ctx->pos = pos = ALIGN(pos, chunk_size);
fs/nfs/blocklayout/blocklayout.h
110
u64 chunk_size;
fs/nfs/blocklayout/blocklayout.h
80
u64 chunk_size;
fs/nfs/blocklayout/dev.c
185
p = xdr_decode_hyper(p, &b->stripe.chunk_size);
fs/nfs/blocklayout/dev.c
263
chunk = div_u64(offset, dev->chunk_size);
fs/nfs/blocklayout/dev.c
268
__func__, chunk_idx, offset, dev->chunk_size);
fs/nfs/blocklayout/dev.c
274
offset = chunk * dev->chunk_size;
fs/nfs/blocklayout/dev.c
277
disk_offset = disk_chunk * dev->chunk_size;
fs/nfs/blocklayout/dev.c
284
map->len = dev->chunk_size;
fs/nfs/blocklayout/dev.c
509
d->chunk_size = v->stripe.chunk_size;
fs/nilfs2/dir.c
114
unsigned int chunk_size = nilfs_chunk_size(dir);
fs/nilfs2/dir.c
122
if (limit & (chunk_size - 1))
fs/nilfs2/dir.c
137
if (((offs + rec_len - 1) ^ offs) & ~(chunk_size-1))
fs/nilfs2/dir.c
433
unsigned int chunk_size = nilfs_chunk_size(dir);
fs/nilfs2/dir.c
462
rec_len = chunk_size;
fs/nilfs2/dir.c
463
de->rec_len = nilfs_rec_len_to_disk(chunk_size);
fs/nilfs2/dir.c
571
unsigned int chunk_size = nilfs_chunk_size(inode);
fs/nilfs2/dir.c
579
err = nilfs_prepare_chunk(folio, 0, chunk_size);
fs/nilfs2/dir.c
585
memset(kaddr, 0, chunk_size);
fs/nilfs2/dir.c
595
de->rec_len = nilfs_rec_len_to_disk(chunk_size - NILFS_DIR_REC_LEN(1));
fs/nilfs2/dir.c
600
nilfs_commit_chunk(folio, mapping, 0, chunk_size);
fs/ufs/dir.c
305
const unsigned int chunk_size = UFS_SB(sb)->s_uspi->s_dirblksize;
fs/ufs/dir.c
335
rec_len = chunk_size;
fs/ufs/dir.c
336
de->d_reclen = cpu_to_fs16(sb, chunk_size);
fs/ufs/dir.c
538
const unsigned int chunk_size = UFS_SB(sb)->s_uspi->s_dirblksize;
fs/ufs/dir.c
546
err = ufs_prepare_chunk(folio, 0, chunk_size);
fs/ufs/dir.c
566
de->d_reclen = cpu_to_fs16(sb, chunk_size - UFS_DIR_REC_LEN(1));
fs/ufs/dir.c
571
ufs_commit_chunk(folio, 0, chunk_size);
include/drm/display/drm_dsc.h
361
__be16 chunk_size;
include/drm/drm_buddy.h
138
return mm->chunk_size << drm_buddy_block_order(block);
include/drm/drm_buddy.h
141
int drm_buddy_init(struct drm_buddy *mm, u64 size, u64 chunk_size);
include/drm/drm_buddy.h
86
u64 chunk_size;
include/net/xdp_sock.h
27
u32 chunk_size;
include/net/xdp_sock_drv.h
46
return pool->chunk_size;
include/net/xsk_buff_pool.h
78
u32 chunk_size;
include/uapi/drm/panthor_drm.h
1008
__u32 chunk_size;
include/uapi/linux/if_xdp.h
87
__u32 chunk_size;
include/uapi/linux/raid/md_p.h
189
__u32 chunk_size; /* 1 chunk size in bytes */
include/uapi/linux/raid/md_u.h
102
int chunk_size; /* 1 chunk size in bytes */
include/uapi/linux/raid/md_u.h
145
int chunk_size; /* in bytes */
include/uapi/linux/xdp_diag.h
66
__u32 chunk_size;
include/ufs/ufshcd.h
780
unsigned long chunk_size;
kernel/padata.c
418
size = roundup(start + 1, ps->chunk_size) - start;
kernel/padata.c
44
unsigned long chunk_size;
kernel/padata.c
481
ps.chunk_size = job->size / (ps.nworks * load_balance_factor);
kernel/padata.c
482
ps.chunk_size = max(ps.chunk_size, job->min_chunk);
kernel/padata.c
483
ps.chunk_size = max(ps.chunk_size, 1ul);
kernel/padata.c
484
ps.chunk_size = roundup(ps.chunk_size, job->align);
lib/genalloc.c
253
end_bit = chunk_size(chunk) >> order;
lib/genalloc.c
302
end_bit = chunk_size(chunk) >> order;
lib/genalloc.c
606
size += chunk_size(chunk);
lib/scatterlist.c
517
unsigned int j, chunk_size;
lib/scatterlist.c
540
chunk_size = ((j - cur_page) << PAGE_SHIFT) - offset;
lib/scatterlist.c
542
min_t(unsigned long, size, chunk_size), offset);
lib/scatterlist.c
544
size -= chunk_size;
mm/kmsan/core.c
261
int chunk_size;
mm/kmsan/core.c
268
chunk_size = min(size - pos,
mm/kmsan/core.c
284
pos += chunk_size;
mm/kmsan/core.c
287
for (int i = 0; i < chunk_size; i++) {
mm/kmsan/core.c
320
pos += chunk_size;
net/devlink/region.c
760
u8 *chunk, u32 chunk_size,
net/devlink/region.c
770
err = nla_put(msg, DEVLINK_ATTR_REGION_CHUNK_DATA, chunk_size, chunk);
net/devlink/region.c
788
typedef int devlink_chunk_fill_t(void *cb_priv, u8 *chunk, u32 chunk_size,
net/devlink/region.c
832
devlink_region_snapshot_fill(void *cb_priv, u8 *chunk, u32 chunk_size,
net/devlink/region.c
838
memcpy(chunk, &snapshot->data[curr_offset], chunk_size);
net/devlink/region.c
844
devlink_region_port_direct_fill(void *cb_priv, u8 *chunk, u32 chunk_size,
net/devlink/region.c
850
curr_offset, chunk_size, chunk);
net/devlink/region.c
854
devlink_region_direct_fill(void *cb_priv, u8 *chunk, u32 chunk_size,
net/devlink/region.c
860
curr_offset, chunk_size, chunk);
net/xdp/xdp_umem.c
161
u32 chunk_size = mr->chunk_size, headroom = mr->headroom;
net/xdp/xdp_umem.c
167
if (chunk_size < XDP_UMEM_MIN_CHUNK_SIZE || chunk_size > PAGE_SIZE) {
net/xdp/xdp_umem.c
180
if (!unaligned_chunks && !is_power_of_2(chunk_size))
net/xdp/xdp_umem.c
199
chunks = div_u64_rem(size, chunk_size, &chunks_rem);
net/xdp/xdp_umem.c
206
if (headroom >= chunk_size - XDP_PACKET_HEADROOM)
net/xdp/xdp_umem.c
217
umem->chunk_size = chunk_size;
net/xdp/xsk.c
1471
__u32 chunk_size;
net/xdp/xsk.c
773
ts = pool->unaligned ? len : pool->chunk_size;
net/xdp/xsk_buff_pool.c
102
xskb->xdp.frame_sz = umem->chunk_size - umem->headroom;
net/xdp/xsk_buff_pool.c
107
xp_init_xskb_addr(xskb, pool, (u64)i * pool->chunk_size);
net/xdp/xsk_buff_pool.c
483
return xp_desc_crosses_non_contig_pg(pool, addr, pool->chunk_size);
net/xdp/xsk_buff_pool.c
490
*addr + pool->chunk_size > pool->addrs_cnt ||
net/xdp/xsk_buff_pool.c
73
pool->chunk_mask = ~((u64)umem->chunk_size - 1);
net/xdp/xsk_buff_pool.c
78
pool->chunk_size = umem->chunk_size;
net/xdp/xsk_buff_pool.c
79
pool->chunk_shift = ffs(umem->chunk_size) - 1;
net/xdp/xsk_buff_pool.c
81
pool->frame_len = umem->chunk_size - umem->headroom -
net/xdp/xsk_diag.c
60
du.chunk_size = umem->chunk_size;
net/xdp/xsk_queue.h
161
offset = addr & (pool->chunk_size - 1);
net/xdp/xsk_queue.h
168
if (offset + len + pool->tx_metadata_len > pool->chunk_size)
net/xdp/xsk_queue.h
191
if (len > pool->chunk_size)
sound/soc/codecs/ntpfw.c
45
size_t chunk_size;
sound/soc/codecs/ntpfw.c
57
chunk_size = be16_to_cpu(chunk->length);
sound/soc/codecs/ntpfw.c
58
if (chunk_size > buf_size) {
sound/soc/codecs/ntpfw.c
63
if (chunk_size % chunk->step) {
sound/soc/sof/intel/hda-stream.c
126
unsigned int chunk_size;
sound/soc/sof/intel/hda-stream.c
128
chunk_size = snd_sgbuf_get_chunk_size(dmab, 0, hstream->bufsize);
sound/soc/sof/intel/hda-stream.c
148
if (chunk_size == hstream->bufsize)
sound/soc/sof/ipc4.c
467
size_t chunk_size;
sound/soc/sof/ipc4.c
513
chunk_size = payload_limit;
sound/soc/sof/ipc4.c
515
chunk_size = remaining;
sound/soc/sof/ipc4.c
527
tx.data_size = chunk_size;
sound/soc/sof/ipc4.c
530
tx_size = chunk_size;
sound/soc/sof/ipc4.c
535
rx.data_size = chunk_size;
sound/soc/sof/ipc4.c
547
rx_size = chunk_size;
sound/soc/sof/ipc4.c
571
if (rx_size < chunk_size) {
sound/soc/sof/ipc4.c
572
chunk_size = rx_size;
sound/soc/sof/ipc4.c
579
offset += chunk_size;
sound/soc/sof/ipc4.c
580
remaining -= chunk_size;
tools/include/uapi/linux/if_xdp.h
87
__u32 chunk_size;
tools/perf/bench/mem-functions.c
269
p.chunk_size = (size_t)perf_atoll((char *)chunk_size_str);
tools/perf/bench/mem-functions.c
270
if ((s64)p.chunk_size < 0 || (s64)p.chunk_size > (s64)p.size) {
tools/perf/bench/mem-functions.c
274
if (!p.chunk_size)
tools/perf/bench/mem-functions.c
275
p.chunk_size = p.size;
tools/perf/bench/mem-functions.c
336
for (size_t off = 0; off < p->size; off += p->chunk_size)
tools/perf/bench/mem-functions.c
337
fn(dst + off, src + off, min(p->chunk_size, p->size - off));
tools/perf/bench/mem-functions.c
440
for (size_t off = 0; off < p->size; off += p->chunk_size)
tools/perf/bench/mem-functions.c
441
fn(dst + off, i, min(p->chunk_size, p->size - off));
tools/perf/bench/mem-functions.c
82
size_t chunk_size;
tools/testing/selftests/bpf/xsk.c
254
mr.chunk_size = umem->config.frame_size;
tools/testing/selftests/mm/mremap_test.c
1079
size_t chunk_size = threshold / num_chunks;
tools/testing/selftests/mm/mremap_test.c
1080
unsigned long shift = i * chunk_size;
tools/testing/selftests/mm/mremap_test.c
1082
if (!memcmp(dest_addr + shift, rand_addr + shift, chunk_size))
tools/testing/selftests/mm/mremap_test.c
1086
for (t = shift; t < shift + chunk_size; ++t) {
tools/testing/selftests/mm/mremap_test.c
1120
size_t chunk_size = c.dest_preamble_size / num_chunks;
tools/testing/selftests/mm/mremap_test.c
1121
unsigned long shift = i * chunk_size;
tools/testing/selftests/mm/mremap_test.c
1124
chunk_size))
tools/testing/selftests/mm/mremap_test.c
1128
for (d = shift; d < shift + chunk_size; ++d) {
tools/testing/selftests/mm/uffd-unit-tests.c
1121
unsigned long chunk_size,
tools/testing/selftests/mm/uffd-unit-tests.c
1150
step_size = chunk_size / gopts->page_size;
tools/testing/selftests/mm/uffd-unit-tests.c
1153
if (chunk_size > gopts->page_size) {
tools/testing/selftests/mm/uffd-unit-tests.c
1154
char *aligned_src = ALIGN_UP(gopts->area_src, chunk_size);
tools/testing/selftests/mm/uffd-unit-tests.c
1155
char *aligned_dst = ALIGN_UP(gopts->area_dst, chunk_size);
tools/testing/selftests/mm/uffd-unit-tests.c
1201
if (chunk_size > gopts->page_size) {
tools/testing/selftests/net/lib/xdp_helper.c
81
umem_reg.chunk_size = 2048;
tools/testing/selftests/net/tcp_mmap.c
191
buffer = mmap_large_buffer(chunk_size, &buffer_sz);
tools/testing/selftests/net/tcp_mmap.c
197
raddr = mmap(NULL, chunk_size + map_align, PROT_READ, flags, fd, 0);
tools/testing/selftests/net/tcp_mmap.c
224
zc.length = min(chunk_size, FILE_SZ - total);
tools/testing/selftests/net/tcp_mmap.c
232
assert(zc.length <= chunk_size);
tools/testing/selftests/net/tcp_mmap.c
245
assert(zc.recv_skip_hint <= chunk_size);
tools/testing/selftests/net/tcp_mmap.c
261
while (sub < chunk_size) {
tools/testing/selftests/net/tcp_mmap.c
262
lu = read(fd, buffer + sub, min(chunk_size - sub,
tools/testing/selftests/net/tcp_mmap.c
319
munmap(raddr, chunk_size + map_align);
tools/testing/selftests/net/tcp_mmap.c
373
rcvlowat = chunk_size;
tools/testing/selftests/net/tcp_mmap.c
498
chunk_size = atol(optarg);
tools/testing/selftests/net/tcp_mmap.c
547
buffer = mmap_large_buffer(chunk_size, &buffer_sz);
tools/testing/selftests/net/tcp_mmap.c
591
size_t offset = total % chunk_size;
tools/testing/selftests/net/tcp_mmap.c
594
if (wr > chunk_size - offset)
tools/testing/selftests/net/tcp_mmap.c
595
wr = chunk_size - offset;
tools/testing/selftests/net/tcp_mmap.c
91
static size_t chunk_size = 512*1024;
tools/testing/selftests/net/tls.c
472
uint16_t chunk_size,
tools/testing/selftests/net/tls.c
485
EXPECT_GE(chunk_size, 1);
tools/testing/selftests/net/tls.c
486
test_payload_size = chunk_size + extra_payload_size;
tools/testing/selftests/net/tls.c
494
ret = sendfile(self->fd, fd, &offset, chunk_size);
tools/testing/selftests/ublk/kublk.h
57
unsigned int chunk_size;
tools/testing/selftests/ublk/stripe.c
292
unsigned chunk_size = ctx->stripe.chunk_size;
tools/testing/selftests/ublk/stripe.c
307
if ((chunk_size & (chunk_size - 1)) || !chunk_size) {
tools/testing/selftests/ublk/stripe.c
308
ublk_err("invalid chunk size %u\n", chunk_size);
tools/testing/selftests/ublk/stripe.c
312
if (chunk_size < 4096 || chunk_size > 512 * 1024) {
tools/testing/selftests/ublk/stripe.c
313
ublk_err("invalid chunk size %u\n", chunk_size);
tools/testing/selftests/ublk/stripe.c
317
chunk_shift = ilog2(chunk_size);
tools/testing/selftests/ublk/stripe.c
372
ctx->stripe.chunk_size = 65536;
tools/testing/selftests/ublk/stripe.c
378
ctx->stripe.chunk_size = strtol(optarg, NULL, 10);
tools/testing/vsock/vsock_test.c
1727
size_t optmem_max, ctl_len, chunk_size;
tools/testing/vsock/vsock_test.c
1761
chunk_size = CMSG_SPACE(ctl_len);
tools/testing/vsock/vsock_test.c
1762
chunk = malloc(chunk_size);
tools/testing/vsock/vsock_test.c
1767
memset(chunk, 0, chunk_size);