Symbol: segments
arch/mips/kernel/segment.c
83
DEFINE_SHOW_ATTRIBUTE(segments);
block/blk-integrity.c
31
unsigned int segments = 0;
block/blk-integrity.c
47
segments++;
block/blk-integrity.c
55
return segments;
block/blk-merge.c
1016
unsigned short segments = blk_rq_nr_discard_segments(req);
block/blk-merge.c
1018
if (segments >= queue_max_discard_segments(q))
block/blk-merge.c
1029
req->nr_phys_segments = segments + 1;
block/blk-merge.c
615
unsigned short segments = blk_rq_nr_discard_segments(req);
block/blk-merge.c
617
if (segments >= queue_max_discard_segments(q))
block/blk-merge.c
623
req->nr_phys_segments = segments + blk_rq_nr_discard_segments(next);
block/blk-mq-dma.c
407
unsigned int segments = 0;
block/blk-mq-dma.c
423
segments++;
block/blk-mq-dma.c
433
BUG_ON(segments > rq->nr_integrity_segments);
block/blk-mq-dma.c
434
BUG_ON(segments > queue_max_integrity_segments(q));
block/blk-mq-dma.c
435
return segments;
drivers/block/virtio_blk.c
162
unsigned short segments = blk_rq_nr_discard_segments(req);
drivers/block/virtio_blk.c
171
range = kmalloc_objs(*range, segments, GFP_ATOMIC);
drivers/block/virtio_blk.c
198
WARN_ON_ONCE(n != segments);
drivers/block/virtio_blk.c
200
bvec_set_virt(&req->special_vec, range, sizeof(*range) * segments);
drivers/block/xen-blkback/blkback.c
1303
struct grant_page **pages = pending_req->segments;
drivers/block/xen-blkback/blkback.c
1473
xen_blkbk_unmap(ring, pending_req->segments,
drivers/block/xen-blkback/blkback.c
707
struct grant_page **pages = req->segments;
drivers/block/xen-blkback/blkback.c
905
rc = xen_blkbk_map(pending_req->ring, pending_req->segments,
drivers/block/xen-blkback/blkback.c
920
struct blkif_request_segment *segments = NULL;
drivers/block/xen-blkback/blkback.c
938
if (segments)
drivers/block/xen-blkback/blkback.c
939
kunmap_atomic(segments);
drivers/block/xen-blkback/blkback.c
940
segments = kmap_atomic(pages[n/SEGS_PER_INDIRECT_FRAME]->page);
drivers/block/xen-blkback/blkback.c
944
pending_req->segments[n]->gref = segments[i].gref;
drivers/block/xen-blkback/blkback.c
946
first_sect = READ_ONCE(segments[i].first_sect);
drivers/block/xen-blkback/blkback.c
947
last_sect = READ_ONCE(segments[i].last_sect);
drivers/block/xen-blkback/blkback.c
959
if (segments)
drivers/block/xen-blkback/blkback.c
960
kunmap_atomic(segments);
drivers/block/xen-blkback/common.h
352
struct grant_page *segments[MAX_INDIRECT_SEGMENTS];
drivers/block/xen-blkback/xenbus.c
1016
req->segments[j] = kzalloc_obj(*req->segments[0]);
drivers/block/xen-blkback/xenbus.c
1017
if (!req->segments[j])
drivers/block/xen-blkback/xenbus.c
1040
if (!req->segments[j])
drivers/block/xen-blkback/xenbus.c
1042
kfree(req->segments[j]);
drivers/block/xen-blkback/xenbus.c
312
kfree(req->segments[j]);
drivers/block/xen-blkfront.c
580
struct blkif_request_segment *segments;
drivers/block/xen-blkfront.c
625
if (setup->segments)
drivers/block/xen-blkfront.c
626
kunmap_atomic(setup->segments);
drivers/block/xen-blkfront.c
631
setup->segments = kmap_atomic(gnt_list_entry->page);
drivers/block/xen-blkfront.c
673
setup->segments[grant_idx % GRANTS_PER_INDIRECT_FRAME] =
drivers/block/xen-blkfront.c
712
.segments = NULL,
drivers/block/xen-blkfront.c
848
if (setup.segments)
drivers/block/xen-blkfront.c
849
kunmap_atomic(setup.segments);
drivers/block/xen-blkfront.c
950
unsigned int segments = info->max_indirect_segments ? :
drivers/block/xen-blkfront.c
971
lim->max_hw_sectors = (segments * XEN_PAGE_SIZE) / 512;
drivers/block/xen-blkfront.c
978
lim->max_segments = segments / GRANTS_PER_PSEG;
drivers/bus/mhi/host/boot.c
373
int segments = DIV_ROUND_UP(alloc_size, seg_size) + 1;
drivers/bus/mhi/host/boot.c
383
img_info->mhi_buf = kzalloc_objs(*img_info->mhi_buf, segments);
drivers/bus/mhi/host/boot.c
389
for (i = 0; i < segments; i++, mhi_buf++) {
drivers/bus/mhi/host/boot.c
393
if (i == segments - 1)
drivers/bus/mhi/host/boot.c
404
img_info->bhi_vec = img_info->mhi_buf[segments - 1].buf;
drivers/bus/mhi/host/boot.c
405
img_info->entries = segments;
drivers/dma/xilinx/xilinx_dma.c
1035
seg = list_first_entry(&desc->segments, struct xilinx_axidma_tx_segment,
drivers/dma/xilinx/xilinx_dma.c
1442
list_for_each_entry(segment, &desc->segments, node) {
drivers/dma/xilinx/xilinx_dma.c
1497
tail_segment = list_last_entry(&tail_desc->segments,
drivers/dma/xilinx/xilinx_dma.c
1525
segment = list_first_entry(&head_desc->segments,
drivers/dma/xilinx/xilinx_dma.c
1566
list_for_each(entry, &desc->segments) {
drivers/dma/xilinx/xilinx_dma.c
1590
tail_segment = list_last_entry(&tail_desc->segments,
drivers/dma/xilinx/xilinx_dma.c
1627
segment = list_first_entry(&head_desc->segments,
drivers/dma/xilinx/xilinx_dma.c
1673
tail_segment = list_last_entry(&tail_desc->segments,
drivers/dma/xilinx/xilinx_dma.c
1760
seg = list_last_entry(&desc->segments,
drivers/dma/xilinx/xilinx_dma.c
1986
tail_segment = list_last_entry(&tail_desc->segments,
drivers/dma/xilinx/xilinx_dma.c
1991
cdma_tail_segment = list_last_entry(&tail_desc->segments,
drivers/dma/xilinx/xilinx_dma.c
1996
axidma_tail_segment = list_last_entry(&tail_desc->segments,
drivers/dma/xilinx/xilinx_dma.c
2002
list_last_entry(&tail_desc->segments,
drivers/dma/xilinx/xilinx_dma.c
2141
list_add_tail(&segment->node, &desc->segments);
drivers/dma/xilinx/xilinx_dma.c
2144
segment = list_first_entry(&desc->segments,
drivers/dma/xilinx/xilinx_dma.c
2199
list_add_tail(&segment->node, &desc->segments);
drivers/dma/xilinx/xilinx_dma.c
2278
list_add_tail(&segment->node, &desc->segments);
drivers/dma/xilinx/xilinx_dma.c
2282
head = list_first_entry(&desc->segments, struct xilinx_axidma_tx_segment, node);
drivers/dma/xilinx/xilinx_dma.c
2288
segment = list_last_entry(&desc->segments,
drivers/dma/xilinx/xilinx_dma.c
2379
list_add_tail(&segment->node, &desc->segments);
drivers/dma/xilinx/xilinx_dma.c
2383
segment = list_first_entry(&desc->segments,
drivers/dma/xilinx/xilinx_dma.c
2390
segment = list_last_entry(&desc->segments,
drivers/dma/xilinx/xilinx_dma.c
2482
list_add_tail(&segment->node, &desc->segments);
drivers/dma/xilinx/xilinx_dma.c
2486
head_segment = list_first_entry(&desc->segments,
drivers/dma/xilinx/xilinx_dma.c
2495
segment = list_last_entry(&desc->segments,
drivers/dma/xilinx/xilinx_dma.c
2586
list_add_tail(&segment->node, &desc->segments);
drivers/dma/xilinx/xilinx_dma.c
2590
segment = list_first_entry(&desc->segments,
drivers/dma/xilinx/xilinx_dma.c
2597
segment = list_last_entry(&desc->segments,
drivers/dma/xilinx/xilinx_dma.c
375
struct list_head segments;
drivers/dma/xilinx/xilinx_dma.c
651
seg = list_first_entry(&desc->segments,
drivers/dma/xilinx/xilinx_dma.c
843
INIT_LIST_HEAD(&desc->segments);
drivers/dma/xilinx/xilinx_dma.c
866
list_for_each_entry_safe(segment, next, &desc->segments, node) {
drivers/dma/xilinx/xilinx_dma.c
872
&desc->segments, node) {
drivers/dma/xilinx/xilinx_dma.c
878
&desc->segments, node) {
drivers/dma/xilinx/xilinx_dma.c
884
&desc->segments, node) {
drivers/dma/xilinx/xilinx_dma.c
994
list_for_each(entry, &desc->segments) {
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_debugfs.c
1298
int i, segments = -EOPNOTSUPP;
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_debugfs.c
1312
pipe_ctx->stream_res.tg->funcs->get_odm_combine_segments(pipe_ctx->stream_res.tg, &segments);
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm_debugfs.c
1314
seq_printf(m, "%d\n", segments);
drivers/gpu/drm/amd/display/dc/optc/dcn32/dcn32_optc.c
103
int segments;
drivers/gpu/drm/amd/display/dc/optc/dcn32/dcn32_optc.c
105
REG_GET(OPTC_DATA_SOURCE_SELECT, OPTC_NUM_OF_INPUT_SEGMENT, &segments);
drivers/gpu/drm/amd/display/dc/optc/dcn32/dcn32_optc.c
107
switch (segments) {
drivers/gpu/drm/arm/malidp_crtc.c
132
delta_in = segments[i].end - segments[i].start;
drivers/gpu/drm/arm/malidp_crtc.c
134
out_start = drm_color_lut_extract(lut[segments[i].start].green,
drivers/gpu/drm/arm/malidp_crtc.c
136
out_end = drm_color_lut_extract(lut[segments[i].end].green, 12);
drivers/gpu/drm/arm/malidp_crtc.c
95
} segments[MALIDP_COEFFTAB_NUM_COEFFS] = {
drivers/md/dm-pcache/cache.c
121
pos->cache_seg = &cache->segments[latest.cache_seg_id];
drivers/md/dm-pcache/cache.c
141
cache->segments = kvzalloc_objs(struct pcache_cache_segment,
drivers/md/dm-pcache/cache.c
143
if (!cache->segments) {
drivers/md/dm-pcache/cache.c
173
kvfree(cache->segments);
drivers/md/dm-pcache/cache.c
181
kvfree(cache->segments);
drivers/md/dm-pcache/cache.c
201
cache->key_head.cache_seg = &cache->segments[0];
drivers/md/dm-pcache/cache.c
274
prev_cache_seg = &cache->segments[i];
drivers/md/dm-pcache/cache.h
198
struct pcache_cache_segment *segments;
drivers/md/dm-pcache/cache_gc.c
84
next_seg = &cache->segments[kset_onmedia->next_cache_seg_id];
drivers/md/dm-pcache/cache_key.c
792
next_seg = &cache->segments[kset_onmedia->next_cache_seg_id];
drivers/md/dm-pcache/cache_key.c
97
key->cache_pos.cache_seg = &cache->segments[key_onmedia->cache_seg_id];
drivers/md/dm-pcache/cache_segment.c
186
struct pcache_cache_segment *cache_seg = &cache->segments[cache_seg_id];
drivers/md/dm-pcache/cache_segment.c
266
cache_seg = &cache->segments[seg_id];
drivers/md/dm-pcache/cache_writeback.c
207
next_seg = &cache->segments[last_kset_onmedia->next_cache_seg_id];
drivers/md/dm-vdo/block-map.c
1379
for (segment = 0; segment < forest->segments; segment++) {
drivers/md/dm-vdo/block-map.c
1385
return &(tree->segments[segment].levels[height - 1][page_index - offset]);
drivers/md/dm-vdo/block-map.c
2358
size_t index = (old_forest == NULL) ? 0 : old_forest->segments;
drivers/md/dm-vdo/block-map.c
2365
forest->segments = index + 1;
drivers/md/dm-vdo/block-map.c
2367
result = vdo_allocate(forest->segments, struct boundary,
drivers/md/dm-vdo/block-map.c
2372
result = vdo_allocate(forest->segments, struct tree_page *,
drivers/md/dm-vdo/block-map.c
2403
int result = vdo_allocate(forest->segments,
drivers/md/dm-vdo/block-map.c
2405
"tree root segments", &tree->segments);
drivers/md/dm-vdo/block-map.c
2410
memcpy(tree->segments, old_forest->trees[root].segments,
drivers/md/dm-vdo/block-map.c
2414
segment = &(tree->segments[index]);
drivers/md/dm-vdo/block-map.c
2444
for (segment = first_page_segment; segment < forest->segments; segment++)
drivers/md/dm-vdo/block-map.c
2450
vdo_free(forest->trees[root].segments);
drivers/md/dm-vdo/block-map.c
2472
old_boundary = &(old_forest->boundaries[old_forest->segments - 1]);
drivers/md/dm-vdo/block-map.c
2490
deforest(forest, forest->segments - 1);
drivers/md/dm-vdo/block-map.c
2507
deforest(map->forest, map->forest->segments);
drivers/md/dm-vdo/block-map.c
2556
&(cursor->tree->segments[0].levels[height][level->page_index]);
drivers/md/dm-vdo/block-map.c
2586
&(cursor->tree->segments[0].levels[height][level->page_index]);
drivers/md/dm-vdo/block-map.c
3088
deforest(forest, forest->segments - 1);
drivers/md/dm-vdo/block-map.c
69
struct block_map_tree_segment *segments;
drivers/md/dm-vdo/block-map.c
74
size_t segments;
drivers/media/dvb-frontends/dib7000p.c
1757
static u32 interpolate_value(u32 value, struct linear_segments *segments,
drivers/media/dvb-frontends/dib7000p.c
1765
if (value >= segments[0].x)
drivers/media/dvb-frontends/dib7000p.c
1766
return segments[0].y;
drivers/media/dvb-frontends/dib7000p.c
1767
if (value < segments[len-1].x)
drivers/media/dvb-frontends/dib7000p.c
1768
return segments[len-1].y;
drivers/media/dvb-frontends/dib7000p.c
1772
if (value == segments[i].x)
drivers/media/dvb-frontends/dib7000p.c
1773
return segments[i].y;
drivers/media/dvb-frontends/dib7000p.c
1774
if (value > segments[i].x)
drivers/media/dvb-frontends/dib7000p.c
1779
dy = segments[i - 1].y - segments[i].y;
drivers/media/dvb-frontends/dib7000p.c
1780
dx = segments[i - 1].x - segments[i].x;
drivers/media/dvb-frontends/dib7000p.c
1782
tmp64 = value - segments[i].x;
drivers/media/dvb-frontends/dib7000p.c
1785
ret = segments[i].y + tmp64;
drivers/media/dvb-frontends/dib8000.c
3978
static u32 interpolate_value(u32 value, struct linear_segments *segments,
drivers/media/dvb-frontends/dib8000.c
3986
if (value >= segments[0].x)
drivers/media/dvb-frontends/dib8000.c
3987
return segments[0].y;
drivers/media/dvb-frontends/dib8000.c
3988
if (value < segments[len-1].x)
drivers/media/dvb-frontends/dib8000.c
3989
return segments[len-1].y;
drivers/media/dvb-frontends/dib8000.c
3993
if (value == segments[i].x)
drivers/media/dvb-frontends/dib8000.c
3994
return segments[i].y;
drivers/media/dvb-frontends/dib8000.c
3995
if (value > segments[i].x)
drivers/media/dvb-frontends/dib8000.c
4000
dy = segments[i - 1].y - segments[i].y;
drivers/media/dvb-frontends/dib8000.c
4001
dx = segments[i - 1].x - segments[i].x;
drivers/media/dvb-frontends/dib8000.c
4003
tmp64 = value - segments[i].x;
drivers/media/dvb-frontends/dib8000.c
4006
ret = segments[i].y + tmp64;
drivers/media/dvb-frontends/mb86a20s.c
1344
static u32 interpolate_value(u32 value, const struct linear_segments *segments,
drivers/media/dvb-frontends/mb86a20s.c
1351
if (value >= segments[0].x)
drivers/media/dvb-frontends/mb86a20s.c
1352
return segments[0].y;
drivers/media/dvb-frontends/mb86a20s.c
1353
if (value < segments[len-1].x)
drivers/media/dvb-frontends/mb86a20s.c
1354
return segments[len-1].y;
drivers/media/dvb-frontends/mb86a20s.c
1358
if (value == segments[i].x)
drivers/media/dvb-frontends/mb86a20s.c
1359
return segments[i].y;
drivers/media/dvb-frontends/mb86a20s.c
1360
if (value > segments[i].x)
drivers/media/dvb-frontends/mb86a20s.c
1365
dy = segments[i].y - segments[i - 1].y;
drivers/media/dvb-frontends/mb86a20s.c
1366
dx = segments[i - 1].x - segments[i].x;
drivers/media/dvb-frontends/mb86a20s.c
1367
tmp64 = value - segments[i].x;
drivers/media/dvb-frontends/mb86a20s.c
1370
ret = segments[i].y - tmp64;
drivers/media/i2c/adv7511-v4l2.c
1574
if ((status & MASK_ADV7511_HPD_DETECT) && ((status & MASK_ADV7511_MSEN_DETECT) || state->edid.segments)) {
drivers/media/i2c/adv7511-v4l2.c
1685
state->edid.segments = segment + 1;
drivers/media/i2c/adv7511-v4l2.c
1687
if (state->edid.blocks > state->edid.segments * 2) {
drivers/media/i2c/adv7511-v4l2.c
1689
v4l2_dbg(1, debug, sd, "%s: request segment %d\n", __func__, state->edid.segments);
drivers/media/i2c/adv7511-v4l2.c
1691
adv7511_wr(sd, 0xc4, state->edid.segments);
drivers/media/i2c/adv7511-v4l2.c
1697
v4l2_dbg(1, debug, sd, "%s: edid complete with %d segment(s)\n", __func__, state->edid.segments);
drivers/media/i2c/adv7511-v4l2.c
1700
state->edid.segments * 256,
drivers/media/i2c/adv7511-v4l2.c
590
edid->segments ? "found" : "no",
drivers/media/i2c/adv7511-v4l2.c
69
u32 segments;
drivers/net/ethernet/sfc/ef100_tx.c
392
unsigned int segments;
drivers/net/ethernet/sfc/ef100_tx.c
401
segments = skb_is_gso(skb) ? skb_shinfo(skb)->gso_segs : 0;
drivers/net/ethernet/sfc/ef100_tx.c
402
if (segments == 1)
drivers/net/ethernet/sfc/ef100_tx.c
403
segments = 0; /* Don't use TSO/GSO for a single segment. */
drivers/net/ethernet/sfc/ef100_tx.c
404
if (segments && !ef100_tx_can_tso(tx_queue, skb)) {
drivers/net/ethernet/sfc/ef100_tx.c
457
rc = efx_tx_map_data(tx_queue, skb, segments);
drivers/net/ethernet/sfc/ef100_tx.c
460
ef100_tx_make_descriptors(tx_queue, skb, segments, efv);
drivers/net/ethernet/sfc/ef100_tx.c
498
if (segments) {
drivers/net/ethernet/sfc/ef100_tx.c
500
tx_queue->tso_packets += segments;
drivers/net/ethernet/sfc/ef100_tx.c
501
tx_queue->tx_packets += segments;
drivers/net/ethernet/sfc/siena/tx.c
145
unsigned int segments;
drivers/net/ethernet/sfc/siena/tx.c
150
segments = skb_is_gso(skb) ? skb_shinfo(skb)->gso_segs : 0;
drivers/net/ethernet/sfc/siena/tx.c
151
if (segments == 1)
drivers/net/ethernet/sfc/siena/tx.c
152
segments = 0; /* Don't use TSO for a single segment. */
drivers/net/ethernet/sfc/siena/tx.c
158
if (segments) {
drivers/net/ethernet/sfc/siena/tx.c
173
if (!data_mapped && (efx_siena_tx_map_data(tx_queue, skb, segments)))
drivers/net/ethernet/sfc/siena/tx_common.c
433
struct sk_buff *segments, *next;
drivers/net/ethernet/sfc/siena/tx_common.c
435
segments = skb_gso_segment(skb, 0);
drivers/net/ethernet/sfc/siena/tx_common.c
436
if (IS_ERR(segments))
drivers/net/ethernet/sfc/siena/tx_common.c
437
return PTR_ERR(segments);
drivers/net/ethernet/sfc/siena/tx_common.c
441
skb_list_walk_safe(segments, skb, next) {
drivers/net/ethernet/sfc/tx.c
319
unsigned int segments;
drivers/net/ethernet/sfc/tx.c
324
segments = skb_is_gso(skb) ? skb_shinfo(skb)->gso_segs : 0;
drivers/net/ethernet/sfc/tx.c
325
if (segments == 1)
drivers/net/ethernet/sfc/tx.c
326
segments = 0; /* Don't use TSO for a single segment. */
drivers/net/ethernet/sfc/tx.c
332
if (segments) {
drivers/net/ethernet/sfc/tx.c
371
if (!data_mapped && (efx_tx_map_data(tx_queue, skb, segments)))
drivers/net/ethernet/sfc/tx.c
382
if (segments) {
drivers/net/ethernet/sfc/tx.c
384
tx_queue->tso_packets += segments;
drivers/net/ethernet/sfc/tx.c
385
tx_queue->tx_packets += segments;
drivers/net/ethernet/sfc/tx_common.c
473
struct sk_buff *segments, *next;
drivers/net/ethernet/sfc/tx_common.c
475
segments = skb_gso_segment(skb, 0);
drivers/net/ethernet/sfc/tx_common.c
476
if (IS_ERR(segments))
drivers/net/ethernet/sfc/tx_common.c
477
return PTR_ERR(segments);
drivers/net/ethernet/sfc/tx_common.c
481
skb_list_walk_safe(segments, skb, next) {
drivers/net/ovpn/io.c
356
struct sk_buff *segments, *curr, *next;
drivers/net/ovpn/io.c
392
segments = skb_gso_segment(skb, 0);
drivers/net/ovpn/io.c
393
if (IS_ERR(segments)) {
drivers/net/ovpn/io.c
394
ret = PTR_ERR(segments);
drivers/net/ovpn/io.c
401
skb = segments;
drivers/nvme/host/core.c
827
unsigned short segments = blk_rq_nr_discard_segments(req), n = 0;
drivers/nvme/host/core.c
865
if (n < segments) {
drivers/nvme/host/core.c
874
if (WARN_ON_ONCE(n != segments)) {
drivers/nvme/host/core.c
885
cmnd->dsm.nr = cpu_to_le32(segments - 1);
drivers/remoteproc/remoteproc_coredump.c
135
struct list_head *segments,
drivers/remoteproc/remoteproc_coredump.c
140
list_for_each_entry(segment, segments, node) {
drivers/s390/block/dcssblk.c
1043
module_param_string(segments, dcssblk_segments, DCSSBLK_PARM_LEN, 0444);
drivers/s390/block/dcssblk.c
1044
MODULE_PARM_DESC(segments, "Name of DCSS segment(s) to be loaded, "
drivers/scsi/mpi3mr/mpi3mr.h
1396
struct segments *trace_buf;
drivers/scsi/mpi3mr/mpi3mr.h
441
struct segments *q_segments;
drivers/scsi/mpi3mr/mpi3mr.h
471
struct segments *q_segments;
drivers/scsi/mpi3mr/mpi3mr_app.c
45
sz = sizeof(struct segments) * seg_count;
drivers/scsi/mpi3mr/mpi3mr_fw.c
1929
struct segments *segments;
drivers/scsi/mpi3mr/mpi3mr_fw.c
1931
segments = mrioc->req_qinfo[q_idx].q_segments;
drivers/scsi/mpi3mr/mpi3mr_fw.c
1932
if (!segments)
drivers/scsi/mpi3mr/mpi3mr_fw.c
1949
if (!segments[j].segment)
drivers/scsi/mpi3mr/mpi3mr_fw.c
1952
size, segments[j].segment, segments[j].segment_dma);
drivers/scsi/mpi3mr/mpi3mr_fw.c
1953
segments[j].segment = NULL;
drivers/scsi/mpi3mr/mpi3mr_fw.c
1973
struct segments *segments;
drivers/scsi/mpi3mr/mpi3mr_fw.c
1975
segments = mrioc->op_reply_qinfo[q_idx].q_segments;
drivers/scsi/mpi3mr/mpi3mr_fw.c
1976
if (!segments)
drivers/scsi/mpi3mr/mpi3mr_fw.c
1993
if (!segments[j].segment)
drivers/scsi/mpi3mr/mpi3mr_fw.c
1996
size, segments[j].segment, segments[j].segment_dma);
drivers/scsi/mpi3mr/mpi3mr_fw.c
1997
segments[j].segment = NULL;
drivers/scsi/mpi3mr/mpi3mr_fw.c
2101
struct segments *segments;
drivers/scsi/mpi3mr/mpi3mr_fw.c
2123
op_reply_q->q_segments = kzalloc_objs(struct segments,
drivers/scsi/mpi3mr/mpi3mr_fw.c
2128
segments = op_reply_q->q_segments;
drivers/scsi/mpi3mr/mpi3mr_fw.c
2130
segments[i].segment =
drivers/scsi/mpi3mr/mpi3mr_fw.c
2132
size, &segments[i].segment_dma, GFP_KERNEL);
drivers/scsi/mpi3mr/mpi3mr_fw.c
2133
if (!segments[i].segment)
drivers/scsi/mpi3mr/mpi3mr_fw.c
2137
(unsigned long)segments[i].segment_dma;
drivers/scsi/mpi3mr/mpi3mr_fw.c
2158
struct segments *segments;
drivers/scsi/mpi3mr/mpi3mr_fw.c
2181
op_req_q->q_segments = kzalloc_objs(struct segments,
drivers/scsi/mpi3mr/mpi3mr_fw.c
2186
segments = op_req_q->q_segments;
drivers/scsi/mpi3mr/mpi3mr_fw.c
2188
segments[i].segment =
drivers/scsi/mpi3mr/mpi3mr_fw.c
2190
size, &segments[i].segment_dma, GFP_KERNEL);
drivers/scsi/mpi3mr/mpi3mr_fw.c
2191
if (!segments[i].segment)
drivers/scsi/mpi3mr/mpi3mr_fw.c
2195
(unsigned long)segments[i].segment_dma;
drivers/scsi/mpi3mr/mpi3mr_fw.c
2558
struct segments *segments = op_req_q->q_segments;
drivers/scsi/mpi3mr/mpi3mr_fw.c
2602
segment_base_addr = segments[pi / op_req_q->segment_qd].segment;
drivers/scsi/mpi3mr/mpi3mr_fw.c
4734
struct segments *segments;
drivers/scsi/mpi3mr/mpi3mr_fw.c
4741
segments = op_reply_q->q_segments;
drivers/scsi/mpi3mr/mpi3mr_fw.c
4743
memset(segments[i].segment, 0, size);
drivers/scsi/mpi3mr/mpi3mr_fw.c
4757
struct segments *segments;
drivers/scsi/mpi3mr/mpi3mr_fw.c
4764
segments = op_req_q->q_segments;
drivers/scsi/mpi3mr/mpi3mr_fw.c
4766
memset(segments[i].segment, 0, size);
drivers/scsi/mpi3mr/mpi3mr_fw.c
523
struct segments *segments = op_reply_q->q_segments;
drivers/scsi/mpi3mr/mpi3mr_fw.c
527
segments[reply_ci / op_reply_q->segment_qd].segment;
drivers/scsi/ncr53c8xx.c
4089
int segments;
drivers/scsi/ncr53c8xx.c
4216
segments = ncr_scatter(np, cp, cp->cmd);
drivers/scsi/ncr53c8xx.c
4217
if (segments < 0) {
drivers/scsi/ncr53c8xx.c
4224
segments = 0;
drivers/scsi/ncr53c8xx.c
4261
if (segments <= MAX_SCATTERL)
drivers/scsi/ncr53c8xx.c
4262
lastp = goalp - 8 - (segments * 16);
drivers/scsi/ncr53c8xx.c
4265
lastp -= (segments - MAX_SCATTERL) * 16;
drivers/scsi/ncr53c8xx.c
4274
if (segments <= MAX_SCATTERL)
drivers/scsi/ncr53c8xx.c
4275
lastp = goalp - 8 - (segments * 16);
drivers/scsi/ncr53c8xx.c
4278
lastp -= (segments - MAX_SCATTERL) * 16;
drivers/scsi/qla2xxx/qla_init.c
8493
uint templates, segments, fragment;
drivers/scsi/qla2xxx/qla_init.c
8500
segments = FA_RISC_CODE_SEGMENTS;
drivers/scsi/qla2xxx/qla_init.c
8502
for (j = 0; j < segments; j++) {
drivers/scsi/qla2xxx/qla_init.c
8610
uint templates, segments, fragment;
drivers/scsi/qla2xxx/qla_init.c
8637
segments = FA_RISC_CODE_SEGMENTS;
drivers/scsi/qla2xxx/qla_init.c
8638
for (j = 0; j < segments; j++) {
drivers/scsi/qla2xxx/qla_init.c
8869
uint templates, segments, fragment;
drivers/scsi/qla2xxx/qla_init.c
8906
segments = FA_RISC_CODE_SEGMENTS;
drivers/scsi/qla2xxx/qla_init.c
8907
for (j = 0; j < segments; j++) {
drivers/scsi/sym53c8xx_2/sym_glue.c
341
cp->segments = sym_scatter(np, cp, cmd);
drivers/scsi/sym53c8xx_2/sym_glue.c
342
if (cp->segments < 0) {
drivers/scsi/sym53c8xx_2/sym_glue.c
350
if (!cp->segments)
drivers/scsi/sym53c8xx_2/sym_glue.c
354
cp->segments = 0;
drivers/scsi/sym53c8xx_2/sym_glue.c
367
lastp = goalp - 8 - (cp->segments * (2*4));
drivers/scsi/sym53c8xx_2/sym_glue.c
372
lastp = goalp - 8 - (cp->segments * (2*4));
drivers/scsi/sym53c8xx_2/sym_hipd.c
3670
dp_sgmin = SYM_CONF_MAX_SG - cp->segments;
drivers/scsi/sym53c8xx_2/sym_hipd.h
737
int segments; /* Number of SG segments */
drivers/staging/octeon/ethernet-rx.c
144
int segments = work->word2.s.bufs;
drivers/staging/octeon/ethernet-rx.c
149
while (segments--) {
drivers/staging/octeon/ethernet.c
178
int segments = work->word2.s.bufs;
drivers/staging/octeon/ethernet.c
181
while (segments--) {
drivers/xen/gntdev.c
984
if (copy_from_user(&seg, &copy.segments[i], sizeof(seg))) {
drivers/xen/gntdev.c
990
ret = gntdev_grant_copy_seg(batch, &seg, &copy.segments[i].status);
fs/f2fs/file.c
3319
end_segno = min(start_segno + range.segments, dev_end_segno);
fs/nfs/nfs42xdr.c
1143
uint32_t segments;
fs/nfs/nfs42xdr.c
1158
segments = be32_to_cpup(p++);
fs/nfs/nfs42xdr.c
1159
if (segments == 0)
fs/nfs/nfs42xdr.c
1162
segs = kmalloc_objs(*segs, segments);
fs/nfs/nfs42xdr.c
1166
for (i = 0; i < segments; i++) {
fs/nfs/nfs42xdr.c
1173
for (i = segments; i > 0; i--)
fs/nfsd/nfs4xdr.c
5803
u32 segments = 0;
fs/nfsd/nfs4xdr.c
5825
segments++;
fs/nfsd/nfs4xdr.c
5829
wire_data[1] = cpu_to_be32(segments);
fs/nfsd/vfs.c
1306
struct nfsd_write_dio_seg segments[3])
fs/nfsd/vfs.c
1336
nfsd_write_dio_seg_init(&segments[nsegs++], bvec,
fs/nfsd/vfs.c
1339
nfsd_write_dio_seg_init(&segments[nsegs], bvec, nvecs,
fs/nfsd/vfs.c
1349
if (iov_iter_bvec_offset(&segments[nsegs].iter) & (mem_align - 1))
fs/nfsd/vfs.c
1351
segments[nsegs].flags |= IOCB_DIRECT;
fs/nfsd/vfs.c
1355
nfsd_write_dio_seg_init(&segments[nsegs++], bvec, nvecs, total,
fs/nfsd/vfs.c
1362
nfsd_write_dio_seg_init(&segments[0], bvec, nvecs, total, 0,
fs/nfsd/vfs.c
1372
struct nfsd_write_dio_seg segments[3];
fs/nfsd/vfs.c
1378
kiocb, *cnt, segments);
fs/nfsd/vfs.c
1382
kiocb->ki_flags = segments[i].flags;
fs/nfsd/vfs.c
1385
segments[i].iter.count);
fs/nfsd/vfs.c
1388
segments[i].iter.count);
fs/nfsd/vfs.c
1397
host_err = vfs_iocb_iter_write(file, kiocb, &segments[i].iter);
fs/nfsd/vfs.c
1401
if (host_err < segments[i].iter.count)
fs/nilfs2/recovery.c
1006
list_splice_tail(&segments, &ri->ri_used_segments);
fs/nilfs2/recovery.c
1014
nilfs_dispose_segment_list(&segments);
fs/nilfs2/recovery.c
876
LIST_HEAD(segments);
fs/nilfs2/recovery.c
955
nilfs_dispose_segment_list(&segments);
fs/nilfs2/recovery.c
993
ret = nilfs_segment_list_add(&segments, segnum);
fs/nilfs2/sysfs.c
436
NILFS_DEV_INT_GROUP_OPS(segments, dev);
fs/nilfs2/sysfs.c
437
NILFS_DEV_INT_GROUP_TYPE(segments, dev);
fs/nilfs2/sysfs.c
438
NILFS_DEV_INT_GROUP_FNS(segments, dev);
fs/nilfs2/sysfs.h
119
NILFS_RO_ATTR(segments, name)
fs/nilfs2/sysfs.h
74
NILFS_DEV_ATTR_STRUCT(segments);
include/linux/agpgart.h
94
struct agp_segment_priv **segments;
include/linux/syscalls.h
603
struct kexec_segment __user *segments,
include/net/seg6.h
87
return &srh->segments[0];
include/sound/sof/ext_manifest4.h
105
struct sof_man4_segment_desc segments[3];
include/uapi/linux/f2fs.h
93
__u32 segments; /* # of segments to flush */
include/uapi/linux/rpl.h
42
} segments;
include/uapi/linux/rpl.h
45
#define rpl_segaddr segments.addr
include/uapi/linux/rpl.h
46
#define rpl_segdata segments.data
include/uapi/linux/seg6.h
33
struct in6_addr segments[];
include/uapi/linux/virtio_net.h
186
__le16 segments;
include/uapi/xen/gntdev.h
202
struct gntdev_grant_copy_segment __user *segments;
kernel/kexec.c
130
ret = kimage_alloc_init(&image, entry, nr_segments, segments, flags);
kernel/kexec.c
24
struct kexec_segment *segments,
kernel/kexec.c
243
struct kexec_segment __user *, segments, unsigned long, flags)
kernel/kexec.c
257
ksegments = memdup_array_user(segments, nr_segments, sizeof(ksegments[0]));
kernel/kexec.c
270
struct compat_kexec_segment __user *, segments,
kernel/kexec.c
292
result = copy_from_user(&in, &segments[i], sizeof(in));
kernel/kexec.c
47
memcpy(image->segment, segments, nr_segments * sizeof(*segments));
kernel/kexec.c
92
struct kexec_segment *segments, unsigned long flags)
lib/assoc_array.c
179
unsigned long segments;
lib/assoc_array.c
199
segments = ops->get_key_chunk(index_key, level);
lib/assoc_array.c
200
pr_devel("segments[%d]: %lx\n", level, segments);
lib/assoc_array.c
207
slot = segments >> (level & ASSOC_ARRAY_KEY_CHUNK_MASK);
lib/assoc_array.c
253
segments = ops->get_key_chunk(index_key, sc_level);
lib/assoc_array.c
256
dissimilarity = segments ^ sc_segments;
net/core/filter.c
6700
ptr + len > (void *)&srh->segments)
net/ipv6/ah6.c
238
int segments, segments_left;
net/ipv6/ah6.c
254
segments = rthdr->hdrlen >> 1;
net/ipv6/ah6.c
257
final_addr = addrs[segments - 1];
net/ipv6/ah6.c
259
addrs += segments - segments_left;
net/ipv6/exthdrs.c
1125
memcpy(sr_phdr->segments + 1, sr_ihdr->segments + 1,
net/ipv6/exthdrs.c
1128
sr_phdr->segments[0] = **addr_p;
net/ipv6/exthdrs.c
1129
*addr_p = &sr_ihdr->segments[sr_ihdr->segments_left];
net/ipv6/exthdrs.c
1380
fl6->daddr = srh->segments[srh->segments_left];
net/ipv6/exthdrs.c
352
addr = hdr->segments + hdr->segments_left;
net/ipv6/exthdrs.c
452
addr = hdr->segments + hdr->segments_left;
net/ipv6/exthdrs.c
562
buf = kcalloc(struct_size(hdr, segments.addr, n + 2), 2, GFP_ATOMIC);
net/ipv6/rpl_iptunnel.c
140
buf = kcalloc(struct_size(srh, segments.addr, srh->segments_left), 2, GFP_ATOMIC);
net/ipv6/seg6_hmac.c
145
memcpy(off, hdr->segments + i, 16);
net/ipv6/seg6_iptunnel.c
185
hdr->daddr = isrh->segments[isrh->first_segment];
net/ipv6/seg6_iptunnel.c
271
hdr->daddr = osrh->segments[first_seg];
net/ipv6/seg6_iptunnel.c
363
isrh->segments[0] = hdr->daddr;
net/ipv6/seg6_iptunnel.c
364
hdr->daddr = isrh->segments[isrh->first_segment];
net/ipv6/seg6_local.c
267
addr = srh->segments + srh->segments_left;
sound/pci/cs46xx/cs46xx_dsp_spos.h
83
struct dsp_segment_desc * segments;
sound/pci/cs46xx/cs46xx_lib.c
357
if (module->segments) {
sound/pci/cs46xx/cs46xx_lib.c
360
kfree(module->segments[i].data);
sound/pci/cs46xx/cs46xx_lib.c
361
kfree(module->segments);
sound/pci/cs46xx/cs46xx_lib.c
436
module->segments =
sound/pci/cs46xx/cs46xx_lib.c
438
if (!module->segments)
sound/pci/cs46xx/cs46xx_lib.c
441
struct dsp_segment_desc *entry = &module->segments[i];
sound/pci/cs46xx/dsp_spos.c
117
if (module->segments[i].segment_type == seg_type) {
sound/pci/cs46xx/dsp_spos.c
118
return (module->segments + i);
sound/pci/cs46xx/dsp_spos.c
533
struct dsp_segment_desc * desc = (ins->modules[i].segments + j);
sound/soc/intel/avs/messages.h
587
struct avs_segment_desc segments[3];
sound/soc/intel/avs/probes.c
35
cfg.base.is_pages = mentry.segments[2].flags.length;
tools/include/uapi/linux/seg6.h
33
struct in6_addr segments[];
tools/testing/selftests/bpf/progs/test_lwt_seg6local.c
46
struct ip6_addr_t segments[0];
tools/testing/selftests/bpf/progs/test_seg6_loop.c
46
struct ip6_addr_t segments[0];