lib/libcbor/src/cbor/bytestrings.c
109
data->chunks, sizeof(cbor_item_t *), new_chunk_capacity);
lib/libcbor/src/cbor/bytestrings.c
115
data->chunks = new_chunks_data;
lib/libcbor/src/cbor/bytestrings.c
117
data->chunks[data->chunk_count++] = cbor_incref(chunk);
lib/libcbor/src/cbor/bytestrings.c
56
.chunks = NULL,
lib/libcbor/src/cbor/bytestrings.c
83
return ((struct cbor_indefinite_string_data *)item->data)->chunks;
lib/libcbor/src/cbor/common.c
102
((struct cbor_indefinite_string_data *)item->data)->chunks);
lib/libcbor/src/cbor/common.c
116
((struct cbor_indefinite_string_data *)item->data)->chunks);
lib/libcbor/src/cbor/data.h
189
cbor_item_t** chunks;
lib/libcbor/src/cbor/serialization.c
104
cbor_item_t **chunks = cbor_string_chunks_handle(item);
lib/libcbor/src/cbor/serialization.c
107
indef_string_size, cbor_serialized_size(chunks[i]));
lib/libcbor/src/cbor/serialization.c
223
cbor_item_t **chunks = cbor_bytestring_chunks_handle(item);
lib/libcbor/src/cbor/serialization.c
226
chunks[i], buffer + written, buffer_size - written);
lib/libcbor/src/cbor/serialization.c
255
cbor_item_t **chunks = cbor_string_chunks_handle(item);
lib/libcbor/src/cbor/serialization.c
257
size_t chunk_written = cbor_serialize_string(chunks[i], buffer + written,
lib/libcbor/src/cbor/serialization.c
89
cbor_item_t **chunks = cbor_bytestring_chunks_handle(item);
lib/libcbor/src/cbor/serialization.c
92
indef_bytestring_size, cbor_serialized_size(chunks[i]));
lib/libcbor/src/cbor/strings.c
104
data->chunks = new_chunks_data;
lib/libcbor/src/cbor/strings.c
106
data->chunks[data->chunk_count++] = cbor_incref(chunk);
lib/libcbor/src/cbor/strings.c
36
.chunks = NULL,
lib/libcbor/src/cbor/strings.c
74
return ((struct cbor_indefinite_string_data *)item->data)->chunks;
lib/libcbor/src/cbor/strings.c
97
data->chunks, sizeof(cbor_item_t *), new_chunk_capacity);
sbin/disklabel/editor.c
1353
static struct diskchunk chunks[MAXPARTITIONS + 2];
sbin/disklabel/editor.c
1362
chunks[0].start = starting_sector;
sbin/disklabel/editor.c
1363
chunks[0].stop = ending_sector;
sbin/disklabel/editor.c
1364
chunks[1].start = chunks[1].stop = 0;
sbin/disklabel/editor.c
1365
return chunks;
sbin/disklabel/editor.c
1371
chunks[0].start = starting_sector;
sbin/disklabel/editor.c
1372
chunks[0].stop = DL_GETPOFFSET(spp[0]);
sbin/disklabel/editor.c
1390
chunks[numchunks].start = start;
sbin/disklabel/editor.c
1391
chunks[numchunks].stop = stop;
sbin/disklabel/editor.c
1397
chunks[numchunks].start = chunks[numchunks].stop = 0;
sbin/disklabel/editor.c
1398
return chunks;
sys/dev/microcode/bnx/build.c
100
struct chunks chunks_xi_rv2p[] = {
sys/dev/microcode/bnx/build.c
106
struct chunks chunks_xi90_rv2p[] = {
sys/dev/microcode/bnx/build.c
122
struct chunks *chunks, u_int nchunks)
sys/dev/microcode/bnx/build.c
142
hswapn(chunks[i].start, *chunks[i].len);
sys/dev/microcode/bnx/build.c
143
rlen = write(fd, chunks[i].start, *chunks[i].len);
sys/dev/microcode/bnx/build.c
148
if (rlen != *chunks[i].len) {
sys/dev/microcode/bnx/build.c
40
struct chunks chunks_b06[] = {
sys/dev/microcode/bnx/build.c
67
struct chunks chunks_b09[] = {
sys/dev/microcode/bnx/build.c
94
struct chunks chunks_rv2p[] = {
sys/dev/pci/drm/amd/amdgpu/amdgpu_cs.c
1407
kvfree(parser->chunks[i].kdata);
sys/dev/pci/drm/amd/amdgpu/amdgpu_cs.c
1408
kvfree(parser->chunks);
sys/dev/pci/drm/amd/amdgpu/amdgpu_cs.c
188
chunk_array = memdup_array_user(u64_to_user_ptr(cs->in.chunks),
sys/dev/pci/drm/amd/amdgpu/amdgpu_cs.c
195
p->chunks = kvmalloc_array(p->nchunks, sizeof(struct amdgpu_cs_chunk),
sys/dev/pci/drm/amd/amdgpu/amdgpu_cs.c
197
if (!p->chunks) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_cs.c
213
p->chunks[i].chunk_id = user_chunk.chunk_id;
sys/dev/pci/drm/amd/amdgpu/amdgpu_cs.c
214
p->chunks[i].length_dw = user_chunk.length_dw;
sys/dev/pci/drm/amd/amdgpu/amdgpu_cs.c
216
size = p->chunks[i].length_dw;
sys/dev/pci/drm/amd/amdgpu/amdgpu_cs.c
218
p->chunks[i].kdata = vmemdup_array_user(u64_to_user_ptr(user_chunk.chunk_data),
sys/dev/pci/drm/amd/amdgpu/amdgpu_cs.c
221
if (IS_ERR(p->chunks[i].kdata)) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_cs.c
222
ret = PTR_ERR(p->chunks[i].kdata);
sys/dev/pci/drm/amd/amdgpu/amdgpu_cs.c
230
switch (p->chunks[i].chunk_id) {
sys/dev/pci/drm/amd/amdgpu/amdgpu_cs.c
235
ret = amdgpu_cs_p1_ib(p, p->chunks[i].kdata, num_ibs);
sys/dev/pci/drm/amd/amdgpu/amdgpu_cs.c
244
ret = amdgpu_cs_p1_user_fence(p, p->chunks[i].kdata,
sys/dev/pci/drm/amd/amdgpu/amdgpu_cs.c
258
ret = amdgpu_cs_p1_bo_handles(p, p->chunks[i].kdata);
sys/dev/pci/drm/amd/amdgpu/amdgpu_cs.c
328
kvfree(p->chunks[i].kdata);
sys/dev/pci/drm/amd/amdgpu/amdgpu_cs.c
329
kvfree(p->chunks);
sys/dev/pci/drm/amd/amdgpu/amdgpu_cs.c
330
p->chunks = NULL;
sys/dev/pci/drm/amd/amdgpu/amdgpu_cs.c
618
chunk = &p->chunks[i];
sys/dev/pci/drm/amd/amdgpu/amdgpu_cs.h
56
struct amdgpu_cs_chunk *chunks;
sys/dev/pci/drm/apple/dcp-internal.h
163
struct dcp_chunks chunks;
sys/dev/pci/drm/apple/iomfb_template.c
506
if (dcp->chunks.data != NULL) {
sys/dev/pci/drm/apple/iomfb_template.c
511
dcp->chunks.length = *length;
sys/dev/pci/drm/apple/iomfb_template.c
512
dcp->chunks.data = devm_kzalloc(dcp->dev, *length, GFP_KERNEL);
sys/dev/pci/drm/apple/iomfb_template.c
514
if (!dcp->chunks.data) {
sys/dev/pci/drm/apple/iomfb_template.c
525
if (!dcp->chunks.data) {
sys/dev/pci/drm/apple/iomfb_template.c
530
if (req->offset + req->length > dcp->chunks.length) {
sys/dev/pci/drm/apple/iomfb_template.c
535
memcpy(dcp->chunks.data + req->offset, req->data, req->length);
sys/dev/pci/drm/apple/iomfb_template.c
545
if (!dcp->chunks.data) {
sys/dev/pci/drm/apple/iomfb_template.c
553
ret = parse(dcp->chunks.data, dcp->chunks.length, &ctx);
sys/dev/pci/drm/apple/iomfb_template.c
594
devm_kfree(dcp->dev, dcp->chunks.data);
sys/dev/pci/drm/apple/iomfb_template.c
595
dcp->chunks.data = NULL;
sys/dev/pci/drm/include/uapi/drm/amdgpu_drm.h
977
__u64 chunks;
sys/dev/pci/drm/include/uapi/drm/radeon_drm.h
987
__u64 chunks;
sys/dev/pci/drm/radeon/radeon.h
1034
struct radeon_cs_chunk *chunks;
sys/dev/pci/drm/radeon/radeon_cs.c
297
chunk_array_ptr = (uint64_t *)(unsigned long)(cs->chunks);
sys/dev/pci/drm/radeon/radeon_cs.c
304
p->chunks = kvcalloc(p->nchunks, sizeof(struct radeon_cs_chunk), GFP_KERNEL);
sys/dev/pci/drm/radeon/radeon_cs.c
305
if (p->chunks == NULL) {
sys/dev/pci/drm/radeon/radeon_cs.c
318
p->chunks[i].length_dw = user_chunk.length_dw;
sys/dev/pci/drm/radeon/radeon_cs.c
320
p->chunk_relocs = &p->chunks[i];
sys/dev/pci/drm/radeon/radeon_cs.c
323
p->chunk_ib = &p->chunks[i];
sys/dev/pci/drm/radeon/radeon_cs.c
325
if (p->chunks[i].length_dw == 0)
sys/dev/pci/drm/radeon/radeon_cs.c
329
p->chunk_const_ib = &p->chunks[i];
sys/dev/pci/drm/radeon/radeon_cs.c
331
if (p->chunks[i].length_dw == 0)
sys/dev/pci/drm/radeon/radeon_cs.c
335
p->chunk_flags = &p->chunks[i];
sys/dev/pci/drm/radeon/radeon_cs.c
337
if (p->chunks[i].length_dw == 0)
sys/dev/pci/drm/radeon/radeon_cs.c
341
size = p->chunks[i].length_dw;
sys/dev/pci/drm/radeon/radeon_cs.c
343
p->chunks[i].user_ptr = cdata;
sys/dev/pci/drm/radeon/radeon_cs.c
352
p->chunks[i].kdata = kvmalloc_array(size, sizeof(uint32_t), GFP_KERNEL);
sys/dev/pci/drm/radeon/radeon_cs.c
354
if (p->chunks[i].kdata == NULL) {
sys/dev/pci/drm/radeon/radeon_cs.c
357
if (copy_from_user(p->chunks[i].kdata, cdata, size)) {
sys/dev/pci/drm/radeon/radeon_cs.c
361
p->cs_flags = p->chunks[i].kdata[0];
sys/dev/pci/drm/radeon/radeon_cs.c
362
if (p->chunks[i].length_dw > 1)
sys/dev/pci/drm/radeon/radeon_cs.c
363
ring = p->chunks[i].kdata[1];
sys/dev/pci/drm/radeon/radeon_cs.c
364
if (p->chunks[i].length_dw > 2)
sys/dev/pci/drm/radeon/radeon_cs.c
365
priority = (s32)p->chunks[i].kdata[2];
sys/dev/pci/drm/radeon/radeon_cs.c
462
kvfree(parser->chunks[i].kdata);
sys/dev/pci/drm/radeon/radeon_cs.c
463
kvfree(parser->chunks);
sys/dev/softraid_raid1c.c
147
struct sr_chunk **chunks;
sys/dev/softraid_raid1c.c
150
chunks = mallocarray(sd->sd_meta->ssdi.ssd_chunk_no,
sys/dev/softraid_raid1c.c
154
chunks[c] = sd->sd_vol.sv_chunks[c];
sys/dev/softraid_raid1c.c
157
ch_prev = chunks[c - 1];
sys/dev/softraid_raid1c.c
163
chunks[c] = ch_entry;
sys/dev/softraid_raid1c.c
168
sd->sd_vol.sv_chunks = chunks;
sys/kern/subr_hibernate.c
1232
struct hibernate_disk_chunk *chunks;
sys/kern/subr_hibernate.c
1243
chunks = (struct hibernate_disk_chunk *)(pva + HIBERNATE_CHUNK_SIZE);
sys/kern/subr_hibernate.c
1271
hibernate_process_chunk(local_hib, &chunks[fchunks[i]],
sys/kern/subr_hibernate.c
1274
image_cur += chunks[fchunks[i]].compressed_size;
sys/kern/subr_hibernate.c
1441
struct hibernate_disk_chunk *chunks;
sys/kern/subr_hibernate.c
1459
chunks = (struct hibernate_disk_chunk *)(hib->piglet_va +
sys/kern/subr_hibernate.c
1470
chunks[hib->chunk_ctr].base = inaddr;
sys/kern/subr_hibernate.c
1472
chunks[hib->chunk_ctr].end = inaddr +
sys/kern/subr_hibernate.c
1475
chunks[hib->chunk_ctr].end = range_end;
sys/kern/subr_hibernate.c
1486
range_base = chunks[i].base;
sys/kern/subr_hibernate.c
1487
range_end = chunks[i].end;
sys/kern/subr_hibernate.c
1489
chunks[i].offset = blkctr;
sys/kern/subr_hibernate.c
1607
chunks[i].compressed_size = dbtob(blkctr - chunks[i].offset);
sys/kern/subr_hibernate.c
1676
struct hibernate_disk_chunk *chunks;
sys/kern/subr_hibernate.c
1716
chunks = (struct hibernate_disk_chunk *)chunktable;
sys/kern/subr_hibernate.c
1719
compressed_size += chunks[i].compressed_size;
sys/kern/subr_hibernate.c
1740
chunks)) {
sys/kern/subr_hibernate.c
1765
struct hibernate_disk_chunk *chunks)
sys/kern/subr_hibernate.c
1808
chunks[i].flags = 0;
sys/kern/subr_hibernate.c
1815
if (chunks[i].end <= pig_start || chunks[i].base >= pig_end) {
sys/kern/subr_hibernate.c
1818
chunks[i].flags |= HIBERNATE_CHUNK_PLACED;
sys/kern/subr_hibernate.c
1826
if (chunks[i].flags != HIBERNATE_CHUNK_PLACED) {
sys/kern/subr_hibernate.c
1829
chunks[i].flags = HIBERNATE_CHUNK_PLACED;
sys/kern/subr_hibernate.c
1836
blkctr = chunks[fchunks[i]].offset + hib->image_offset;
sys/kern/subr_hibernate.c
1838
compressed_size = chunks[fchunks[i]].compressed_size;
sys/uvm/uvm_amap.c
280
int chunks, log_chunks, chunkperbucket = 1, hashshift = 0;
sys/uvm/uvm_amap.c
290
chunks = roundup((unsigned int)slots, UVM_AMAP_CHUNK) / UVM_AMAP_CHUNK;
sys/uvm/uvm_amap.c
317
for (log_chunks = 1; (chunks >> log_chunks) > 0; log_chunks++)
sys/uvm/uvm_amap.c
352
buckets = howmany(chunks, chunkperbucket);
usr.bin/ssh/libcrux_mlkem768_sha3.h
484
static inline Eurydice_slice chunk_next(Eurydice_chunks *chunks,
usr.bin/ssh/libcrux_mlkem768_sha3.h
486
size_t chunk_size = chunks->slice.len >= chunks->chunk_size
usr.bin/ssh/libcrux_mlkem768_sha3.h
487
? chunks->chunk_size
usr.bin/ssh/libcrux_mlkem768_sha3.h
488
: chunks->slice.len;
usr.bin/ssh/libcrux_mlkem768_sha3.h
490
curr_chunk.ptr = chunks->slice.ptr;
usr.bin/ssh/libcrux_mlkem768_sha3.h
492
chunks->slice.ptr = (char *)(chunks->slice.ptr) + chunk_size * element_size;
usr.bin/ssh/libcrux_mlkem768_sha3.h
493
chunks->slice.len = chunks->slice.len - chunk_size;
usr.sbin/nsd/simdzone/src/haswell/simd.h
113
_mm256_shuffle_epi8(t, simd->chunks[0]), simd->chunks[0]);
usr.sbin/nsd/simdzone/src/haswell/simd.h
115
_mm256_shuffle_epi8(t, simd->chunks[1]), simd->chunks[1]);
usr.sbin/nsd/simdzone/src/haswell/simd.h
27
typedef struct { __m256i chunks[1]; } simd_8x_t;
usr.sbin/nsd/simdzone/src/haswell/simd.h
29
typedef struct { __m128i chunks[1]; } simd_8x16_t;
usr.sbin/nsd/simdzone/src/haswell/simd.h
33
typedef struct { __m256i chunks[2]; } simd_8x64_t;
usr.sbin/nsd/simdzone/src/haswell/simd.h
39
simd->chunks[0] = _mm256_loadu_si256((const __m256i *)(address));
usr.sbin/nsd/simdzone/src/haswell/simd.h
45
_mm256_storeu_si256((__m256i *)address, simd->chunks[0]);
usr.sbin/nsd/simdzone/src/haswell/simd.h
52
const __m256i r = _mm256_cmpeq_epi8(simd->chunks[0], k);
usr.sbin/nsd/simdzone/src/haswell/simd.h
62
_mm256_shuffle_epi8(t, simd->chunks[0]), simd->chunks[0]);
usr.sbin/nsd/simdzone/src/haswell/simd.h
69
simd->chunks[0] = _mm_loadu_si128((const __m128i *)address);
usr.sbin/nsd/simdzone/src/haswell/simd.h
76
const __m128i r = _mm_cmpeq_epi8(simd->chunks[0], k);
usr.sbin/nsd/simdzone/src/haswell/simd.h
88
simd->chunks[0] = _mm256_loadu_si256((const __m256i *)(address));
usr.sbin/nsd/simdzone/src/haswell/simd.h
89
simd->chunks[1] = _mm256_loadu_si256((const __m256i *)(address+32));
usr.sbin/nsd/simdzone/src/haswell/simd.h
97
const __m256i r0 = _mm256_cmpeq_epi8(simd->chunks[0], k);
usr.sbin/nsd/simdzone/src/haswell/simd.h
98
const __m256i r1 = _mm256_cmpeq_epi8(simd->chunks[1], k);
usr.sbin/nsd/simdzone/src/westmere/simd.h
106
const __m128i r0 = _mm_cmpeq_epi8(simd->chunks[0], k);
usr.sbin/nsd/simdzone/src/westmere/simd.h
107
const __m128i r1 = _mm_cmpeq_epi8(simd->chunks[1], k);
usr.sbin/nsd/simdzone/src/westmere/simd.h
108
const __m128i r2 = _mm_cmpeq_epi8(simd->chunks[2], k);
usr.sbin/nsd/simdzone/src/westmere/simd.h
109
const __m128i r3 = _mm_cmpeq_epi8(simd->chunks[3], k);
usr.sbin/nsd/simdzone/src/westmere/simd.h
126
_mm_shuffle_epi8(t, simd->chunks[0]), simd->chunks[0]);
usr.sbin/nsd/simdzone/src/westmere/simd.h
128
_mm_shuffle_epi8(t, simd->chunks[1]), simd->chunks[1]);
usr.sbin/nsd/simdzone/src/westmere/simd.h
130
_mm_shuffle_epi8(t, simd->chunks[2]), simd->chunks[2]);
usr.sbin/nsd/simdzone/src/westmere/simd.h
132
_mm_shuffle_epi8(t, simd->chunks[3]), simd->chunks[3]);
usr.sbin/nsd/simdzone/src/westmere/simd.h
26
typedef struct { __m128i chunks[1]; } simd_8x_t;
usr.sbin/nsd/simdzone/src/westmere/simd.h
30
typedef struct { __m128i chunks[2]; } simd_8x32_t;
usr.sbin/nsd/simdzone/src/westmere/simd.h
32
typedef struct { __m128i chunks[4]; } simd_8x64_t;
usr.sbin/nsd/simdzone/src/westmere/simd.h
37
simd->chunks[0] = _mm_loadu_si128((const __m128i *)address);
usr.sbin/nsd/simdzone/src/westmere/simd.h
43
_mm_storeu_si128((__m128i *)address, simd->chunks[0]);
usr.sbin/nsd/simdzone/src/westmere/simd.h
50
const __m128i r = _mm_cmpeq_epi8(simd->chunks[0], k);
usr.sbin/nsd/simdzone/src/westmere/simd.h
60
_mm_shuffle_epi8(t, simd->chunks[0]), simd->chunks[0]);
usr.sbin/nsd/simdzone/src/westmere/simd.h
70
simd->chunks[0] = _mm_loadu_si128((const __m128i *)(address));
usr.sbin/nsd/simdzone/src/westmere/simd.h
71
simd->chunks[1] = _mm_loadu_si128((const __m128i *)(address+16));
usr.sbin/nsd/simdzone/src/westmere/simd.h
77
_mm_storeu_si128((__m128i *)(address), simd->chunks[0]);
usr.sbin/nsd/simdzone/src/westmere/simd.h
78
_mm_storeu_si128((__m128i *)(address+16), simd->chunks[1]);
usr.sbin/nsd/simdzone/src/westmere/simd.h
85
const __m128i r0 = _mm_cmpeq_epi8(simd->chunks[0], k);
usr.sbin/nsd/simdzone/src/westmere/simd.h
86
const __m128i r1 = _mm_cmpeq_epi8(simd->chunks[1], k);
usr.sbin/nsd/simdzone/src/westmere/simd.h
95
simd->chunks[0] = _mm_loadu_si128((const __m128i *)(address));
usr.sbin/nsd/simdzone/src/westmere/simd.h
96
simd->chunks[1] = _mm_loadu_si128((const __m128i *)(address+16));
usr.sbin/nsd/simdzone/src/westmere/simd.h
97
simd->chunks[2] = _mm_loadu_si128((const __m128i *)(address+32));
usr.sbin/nsd/simdzone/src/westmere/simd.h
98
simd->chunks[3] = _mm_loadu_si128((const __m128i *)(address+48));