chunk_bytes
static void write_mbox_data(void __iomem *mmio_base, u32 *chunk, unsigned int chunk_bytes)
for (i = 0; i < chunk_bytes / sizeof(u32); i++)
unsigned int chunk_bytes;
if (check_shl_overflow(t->chunk_sectors, SECTOR_SHIFT, &chunk_bytes))
chunk_bytes = t->chunk_sectors;
max_pow_of_two_factor(chunk_bytes));
t->atomic_write_hw_max = min(t->atomic_write_hw_max, chunk_bytes);
size_t chunk_count, chunk_bytes;
chunk_bytes = chunk_regs * val_bytes;
ret = _regmap_raw_write_impl(map, reg, val, chunk_bytes, noinc);
val += chunk_bytes;
val_len -= chunk_bytes;
size_t chunk_count, chunk_bytes;
chunk_bytes = chunk_regs * val_bytes;
ret = _regmap_raw_read(map, reg, val, chunk_bytes, false);
val += chunk_bytes;
val_len -= chunk_bytes;
rq_regs->chunk_size = dml_log2(rq_sizing->chunk_bytes) - 10;
if (rq_param->sizing.rq_l.chunk_bytes >= 32 * 1024) { //32kb
rq_sizing_param->chunk_bytes = 8192;
if (rq_sizing_param->chunk_bytes == 64 * 1024)
rq_regs->chunk_size = dml_log2(rq_sizing->chunk_bytes) - 10;
if (rq_param->sizing.rq_l.chunk_bytes >= 32 * 1024) { //32kb
rq_sizing_param->chunk_bytes = 8192;
if (rq_sizing_param->chunk_bytes == 64 * 1024)
rq_regs->chunk_size = dml_log2(rq_sizing->chunk_bytes) - 10;
if (rq_param->sizing.rq_l.chunk_bytes >= 32 * 1024) { //32kb
rq_sizing_param->chunk_bytes = 8192;
if (rq_sizing_param->chunk_bytes == 64 * 1024)
if (rq_param->sizing.rq_l.chunk_bytes >= 32 * 1024 || (rq_param->yuv420 && rq_param->sizing.rq_c.chunk_bytes >= 32 * 1024)) { //32kb
rq_sizing_param->chunk_bytes = 8192;
rq_sizing_param->chunk_bytes = 4096;
if (rq_sizing_param->chunk_bytes == 64 * 1024)
rq_regs->chunk_size = dml_log2(rq_sizing->chunk_bytes) - 10;
if (rq_param->sizing.rq_l.chunk_bytes >= 32 * 1024 || (rq_param->yuv420 && rq_param->sizing.rq_c.chunk_bytes >= 32 * 1024)) { //32kb
rq_sizing_param->chunk_bytes = 8192;
rq_sizing_param->chunk_bytes = 4096;
if (rq_sizing_param->chunk_bytes == 64 * 1024)
rq_regs->chunk_size = dml_log2(rq_sizing->chunk_bytes) - 10;
rq_regs->chunk_size = dml_log2(rq_sizing->chunk_bytes) - 10;
if (rq_param->sizing.rq_l.chunk_bytes >= 32 * 1024 || (rq_param->yuv420 && rq_param->sizing.rq_c.chunk_bytes >= 32 * 1024)) { //32kb
rq_sizing_param->chunk_bytes = 8192;
rq_sizing_param->chunk_bytes = 4096;
if (rq_sizing_param->chunk_bytes == 64 * 1024)
unsigned int chunk_bytes;
dml_print("DML_RQ_DLG_CALC: chunk_bytes = %0d\n", rq_sizing->chunk_bytes);
rq_regs->chunk_size = dml_log2(rq_sizing->chunk_bytes) - 10;
if (rq_param->sizing.rq_l.chunk_bytes >= 32 * 1024) { /*32kb */
rq_sizing_param->chunk_bytes = 8192;
if (rq_sizing_param->chunk_bytes == 64 * 1024)
int bytes, chunk_bytes;
chunk_bytes = min(hostdata->pdma_residual, 512);
bytes = mac_pdma_recv(s, d, chunk_bytes);
__func__, d - dst, len, bytes, chunk_bytes);
int bytes, chunk_bytes;
chunk_bytes = min(hostdata->pdma_residual, 512);
bytes = mac_pdma_send(s, d, chunk_bytes);
__func__, s - src, len, bytes, chunk_bytes);
u32 chunks, chunk_count, chunk_bytes;
chunk_bytes = umin(copy_bytes_left, tcon->max_bytes_chunk);
chunk->Length = cpu_to_le32(chunk_bytes);
src_off += chunk_bytes;
dst_off += chunk_bytes;
copy_bytes_left -= chunk_bytes;
copy_bytes += chunk_bytes;
chunk_bytes = le32_to_cpu(cc_rsp->ChunkBytesWritten);
if (chunk_bytes < tcon->max_bytes_chunk) {
tcon->max_bytes_chunk, chunk_bytes);
tcon->max_bytes_chunk = chunk_bytes;