chunk_size
int chunk_size[10];
if (!TEST_true((test_chunks[idx].chunk_size[i] < (int)sizeof(buf))))
memset(buf, i, test_chunks[idx].chunk_size[i]);
if (!TEST_true(HMAC_Update(ctx, buf, test_chunks[idx].chunk_size[i])))
ssize_t chunk_size = 0;
chunk_size = min(SENDFILE_CHUNK, SENDFILE_SZ - chunk_off);
chunk_size,
!= chunk_size) {
chunk_size))
!= chunk_size) {
chunk_size,
chunk_size))
chunk_off += chunk_size;
int chunk_size;
chunk_size = MIN(size - off, 256);
ret = _pfctl_table_add_addrs_h(h, tbl, &addr[off], chunk_size, &partial_added, flags);
off += chunk_size;
int chunk_size;
chunk_size = MIN(size - off, 256);
ret = _pfctl_table_del_addrs_h(h, tbl, &addr[off], chunk_size,
off += chunk_size;
int chunk_size;
chunk_size = MIN(size - off, 256);
if ((chunk_size + off) == size)
ret = _pfctl_table_set_addrs_h(h, tbl, &addr[off], chunk_size,
off += chunk_size;
int chunk_size;
chunk_size = MIN(size - off, 256);
ret = _pfctl_clr_astats(h, tbl, &addrs[off], chunk_size,
off += chunk_size;
int chunk_size;
chunk_size = MIN(size - off, 256);
ret = _pfctl_test_addrs(h, tbl, &addrs[off], chunk_size,
off += chunk_size;
uint32_t chunk_size;
chunk_size = sizeof(((struct harvest_event *)0)->he_entropy);
size = MIN(cnt, chunk_size);
unsigned int chunk_size = min(size, 256U);
ret = t3_write_flash(adapter, addr, chunk_size, fw_data, 1);
addr += chunk_size;
fw_data += chunk_size;
size -= chunk_size;
unsigned int chunk_size = min(size, 256U);
ret = t3_write_flash(adapter, addr, chunk_size, boot_data, 0);
addr += chunk_size;
boot_data += chunk_size;
size -= chunk_size;
int chunk_size, level = 2, rc = 0;
chunk_size = fastlz_compress_level(level, pin_buff->data,
checksum = update_adler32(1L, result, chunk_size);
if ((chunk_size > 62000) && (cudbg_hdr->reserved[7] < (u32)
chunk_size)) /* 64512 */
cudbg_hdr->reserved[7] = (u32) chunk_size;
rc = write_chunk_header(pout_buff, 17, 1, chunk_size, checksum,
&pout_buff->offset, result, chunk_size);
unsigned long chunk_size;
&chunk_size, &chunk_checksum,
if ((chunk_id == 1) && (chunk_size > 10) &&
(chunk_size < CUDBG_BLOCK_SIZE)) {
chunk_size);
checksum = update_adler32(1L, buffer, chunk_size);
pc_buff->offset -= chunk_size + 16;
if (chunk_size > CUDBG_BLOCK_SIZE) {
rc = get_scratch_buff(pd_buff, chunk_size,
rc = get_scratch_buff(pd_buff, chunk_size,
remaining = chunk_size;
chunk_size);
chunk_size);
chunk_size,
u8 chunk_size;
if (info->chunk_size)
FIELD_PREP(IRDMA_CQPSQ_STAG_LPBLSIZE, info->chunk_size) |
if (info->chunk_size && info->first_pm_pbl_index >= pble_obj_cnt)
if (!info->chunk_size)
FIELD_PREP(IRDMA_CQPSQ_STAG_LPBLSIZE, info->chunk_size) |
FIELD_PREP(IRDMAQPSQ_LPBLSIZE, info->chunk_size) |
u32 chunk_size;
u32 chunk_size;
u32 chunk_size;
stag_info->chunk_size = 1;
stag_info->chunk_size = 3;
stag_info.chunk_size = 3;
stag_info.chunk_size = 1;
unsigned chunk_size;
chunk_size = MLX4_TABLE_CHUNK_SIZE;
chunk_size = PAGE_ALIGN(size -
table->icm[i] = mlx4_alloc_icm(dev, chunk_size >> PAGE_SHIFT,
unsigned chunk_size;
chunk_size = MTHCA_TABLE_CHUNK_SIZE;
chunk_size = nobj * obj_size - i * MTHCA_TABLE_CHUNK_SIZE;
table->icm[i] = mthca_alloc_icm(dev, chunk_size >> PAGE_SHIFT,
uint32_t data_written = 0, chunk_size = 0;
chunk_size = MIN(size, (32*1024));
bcopy(data, dest_image_ptr, chunk_size);
rc = oce_mbox_lancer_write_flashrom(sc, chunk_size, offset,
u64 chunk_size;
auth_chunk->chunk_size = img_desc->dram_size;
size_t chunk_size;
chunk_size = PAGE_SIZE;
chunk_size = RANDOM_FORTUNA_MAX_READ;
chunk_size = MIN(bytecount, chunk_size);
chunk_size = rounddown(chunk_size, RANDOM_BLOCKSIZE);
while (bytecount >= chunk_size && chunk_size > 0) {
randomdev_keystream(p_key, p_counter, buf, chunk_size);
buf += chunk_size;
bytecount -= chunk_size;
if (locked || chunk_size == RANDOM_FORTUNA_MAX_READ) {
if (bytecount < chunk_size) {
chunk_size = bytecount;
chunk_size = rounddown(bytecount,
size_t chunk_size;
if ((rc = efx_nvram_rw_start(enp, type, &chunk_size)) != 0)
size_t chunk_size;
if ((rc = efx_nvram_rw_start(enp, type, &chunk_size)) != 0)
buf = malloc(chunk_size, M_TEMP, M_WAITOK);
size_t len = MIN(chunk_size, total_size);
size_t chunk_size;
chunk_size = vg->vg_extentsize;
chunk_index = offset / chunk_size; /* round downwards */
in_chunk_offset = offset % chunk_size;
min(length, chunk_size - in_chunk_offset);
(off_t)chunk_index * (off_t)chunk_size
sc->chunk_size = md->md_chunk_size;
sc->chunk_count = sc->virsize / sc->chunk_size;
if (sc->chunk_count * (off_t)sc->chunk_size != sc->virsize) {
sc->chunk_count * (off_t)sc->chunk_size);
sc->geom->name, count, sc->chunk_count, sc->chunk_size / 1024);
sc->chunk_size);
size_t chunk_size; /* cached for convenience */
chunk_size = sc->chunk_size;
chunk_index = offset / chunk_size; /* round downwards */
in_chunk_offset = offset % chunk_size;
in_chunk_length = min(length, chunk_size - in_chunk_offset);
(off_t)me->provider_chunk * (off_t)chunk_size
cb->bio_offset = (off_t)me->provider_chunk*(off_t)chunk_size +
sc->chunk_size;
md->md_chunk_size = sc->chunk_size;
size_t chunk_size;
uint32_t size = chunk_size(da, fdesc);
uint32_t size = chunk_size(da, fdesc);
uint32_t size = chunk_size(da, fdesc);
uint32_t chunk_size;
chunk_size = htobe32(16*nsegs);
memcpy(&cmd->data_payload[desc_start+12], &chunk_size, 4);
chunk_size = 0;
chunk_size += segs[i].ds_len;
chunk_size = htobe32(chunk_size);
memcpy(&cmd->data_payload[desc_start+16], &chunk_size, 4);
chunk_size = htobe32(segs[i].ds_len);
&chunk_size, 4);
chunk_size = htobe32(segs[0].ds_len);
memcpy(&cmd->data_payload[desc_start+12], &chunk_size, 4);
handle_chunk(int32_t ds_index, const char *chunk_name, off_t chunk_size)
k_size = chunk_size / 1024;
unsigned u, chunk_size = 16;
if ((u % chunk_size) == chunk_size - 1)
if ((u % chunk_size) != 0)