drivers/ata/sata_sil24.c
247
struct sil24_sge sge[SIL24_MAX_SGE];
drivers/ata/sata_sil24.c
253
struct sil24_sge sge[SIL24_MAX_SGE];
drivers/ata/sata_sil24.c
769
struct sil24_sge *sge)
drivers/ata/sata_sil24.c
776
sge->addr = cpu_to_le64(sg_dma_address(sg));
drivers/ata/sata_sil24.c
777
sge->cnt = cpu_to_le32(sg_dma_len(sg));
drivers/ata/sata_sil24.c
778
sge->flags = 0;
drivers/ata/sata_sil24.c
780
last_sge = sge;
drivers/ata/sata_sil24.c
781
sge++;
drivers/ata/sata_sil24.c
838
struct sil24_sge *sge;
drivers/ata/sata_sil24.c
845
sge = cb->ata.sge;
drivers/ata/sata_sil24.c
859
sge = cb->atapi.sge;
drivers/ata/sata_sil24.c
875
sil24_fill_sg(qc, sge);
drivers/crypto/bcm/util.c
28
struct scatterlist **sge, unsigned int *sge_offset)
drivers/crypto/bcm/util.c
45
*sge = sg;
drivers/crypto/bcm/util.h
85
struct scatterlist **sge, unsigned int *sge_offset);
drivers/crypto/chelsio/chcr_algo.c
446
walk->pair = ulp->sge;
drivers/crypto/chelsio/chcr_algo.c
550
adap->sge.uld_txq_info[CXGB4_TX_CRYPTO];
drivers/infiniband/core/mad.c
1025
struct ib_sge *sge;
drivers/infiniband/core/mad.c
1036
sge = mad_send_wr->sg_list;
drivers/infiniband/core/mad.c
1037
sge[0].addr = ib_dma_map_single(mad_agent->device,
drivers/infiniband/core/mad.c
1039
sge[0].length,
drivers/infiniband/core/mad.c
1041
if (unlikely(ib_dma_mapping_error(mad_agent->device, sge[0].addr)))
drivers/infiniband/core/mad.c
1044
mad_send_wr->header_mapping = sge[0].addr;
drivers/infiniband/core/mad.c
1046
sge[1].addr = ib_dma_map_single(mad_agent->device,
drivers/infiniband/core/mad.c
1048
sge[1].length,
drivers/infiniband/core/mad.c
1050
if (unlikely(ib_dma_mapping_error(mad_agent->device, sge[1].addr))) {
drivers/infiniband/core/mad.c
1053
sge[0].length, DMA_TO_DEVICE);
drivers/infiniband/core/mad.c
1056
mad_send_wr->payload_mapping = sge[1].addr;
drivers/infiniband/core/mad.c
1077
sge[0].length, DMA_TO_DEVICE);
drivers/infiniband/core/mad.c
1080
sge[1].length, DMA_TO_DEVICE);
drivers/infiniband/core/rw.c
1025
ib_dma_unmap_bvec(dev, ctx->single.sge.addr,
drivers/infiniband/core/rw.c
1026
ctx->single.sge.length, dir);
drivers/infiniband/core/rw.c
120
reg->sge.addr = reg->mr->iova;
drivers/infiniband/core/rw.c
121
reg->sge.length = reg->mr->length;
drivers/infiniband/core/rw.c
138
reg->wr.wr.sg_list = ®->sge;
drivers/infiniband/core/rw.c
182
remote_addr += reg->sge.length;
drivers/infiniband/core/rw.c
264
remote_addr += reg->sge.length;
drivers/infiniband/core/rw.c
293
struct ib_sge *sge;
drivers/infiniband/core/rw.c
298
ctx->map.sges = sge = kzalloc_objs(*sge, sg_cnt);
drivers/infiniband/core/rw.c
317
rdma_wr->wr.sg_list = sge;
drivers/infiniband/core/rw.c
320
sge->addr = sg_dma_address(sg) + offset;
drivers/infiniband/core/rw.c
321
sge->length = sg_dma_len(sg) - offset;
drivers/infiniband/core/rw.c
322
sge->lkey = qp->pd->local_dma_lkey;
drivers/infiniband/core/rw.c
324
total_len += sge->length;
drivers/infiniband/core/rw.c
325
sge++;
drivers/infiniband/core/rw.c
351
ctx->single.sge.lkey = qp->pd->local_dma_lkey;
drivers/infiniband/core/rw.c
352
ctx->single.sge.addr = sg_dma_address(sg) + offset;
drivers/infiniband/core/rw.c
353
ctx->single.sge.length = sg_dma_len(sg) - offset;
drivers/infiniband/core/rw.c
360
rdma_wr->wr.sg_list = &ctx->single.sge;
drivers/infiniband/core/rw.c
385
ctx->single.sge.lkey = qp->pd->local_dma_lkey;
drivers/infiniband/core/rw.c
386
ctx->single.sge.addr = dma_addr;
drivers/infiniband/core/rw.c
387
ctx->single.sge.length = bv.bv_len;
drivers/infiniband/core/rw.c
394
rdma_wr->wr.sg_list = &ctx->single.sge;
drivers/infiniband/core/rw.c
410
struct ib_sge *sge;
drivers/infiniband/core/rw.c
427
ctx->map.sges = sge = mem;
drivers/infiniband/core/rw.c
441
rdma_wr->wr.sg_list = sge;
drivers/infiniband/core/rw.c
452
sge->addr = dma_addr;
drivers/infiniband/core/rw.c
453
sge->length = bv.bv_len;
drivers/infiniband/core/rw.c
454
sge->lkey = qp->pd->local_dma_lkey;
drivers/infiniband/core/rw.c
457
sge++;
drivers/infiniband/core/rw.c
530
ctx->iova.sge.addr = ctx->iova.state.addr;
drivers/infiniband/core/rw.c
531
ctx->iova.sge.length = mapped_len;
drivers/infiniband/core/rw.c
532
ctx->iova.sge.lkey = qp->pd->local_dma_lkey;
drivers/infiniband/core/rw.c
541
ctx->iova.wr.wr.sg_list = &ctx->iova.sge;
drivers/infiniband/core/rw.c
809
ctx->reg->sge.addr = ctx->reg->mr->iova;
drivers/infiniband/core/rw.c
810
ctx->reg->sge.length = ctx->reg->mr->length;
drivers/infiniband/core/rw.c
812
ctx->reg->sge.length -= ctx->reg->mr->sig_attrs->meta_length;
drivers/infiniband/core/rw.c
815
rdma_wr->wr.sg_list = &ctx->reg->sge;
drivers/infiniband/core/rw.c
852
reg->sge.lkey = reg->mr->lkey;
drivers/infiniband/hw/bnxt_re/bnxt_re.h
98
struct bnxt_qplib_sge sge;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
2425
struct bnxt_qplib_sge sge;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
2542
buf = bnxt_qplib_get_qp1_sq_buf(&qp->qplib_qp, &sge);
drivers/infiniband/hw/bnxt_re/ib_verbs.c
2559
sge.size -= 20;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
2566
sge.size -= 8;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
2570
sge.size -= 4;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
2572
wqe->sg_list[0].addr = sge.addr;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
2573
wqe->sg_list[0].lkey = sge.lkey;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
2574
wqe->sg_list[0].size = sge.size;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
2596
struct bnxt_qplib_sge ref, sge;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
2604
if (!bnxt_qplib_get_qp1_rq_buf(&qp->qplib_qp, &sge))
drivers/infiniband/hw/bnxt_re/ib_verbs.c
2618
wqe->sg_list[0].addr = sge.addr;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
2619
wqe->sg_list[0].lkey = sge.lkey;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
2621
sge.size -= wqe->sg_list[0].size;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
2623
sqp_entry->sge.addr = ref.addr;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
2624
sqp_entry->sge.lkey = ref.lkey;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
2625
sqp_entry->sge.size = ref.size;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3655
r_sge[1].addr = sqp_entry->sge.addr + offset;
drivers/infiniband/hw/bnxt_re/ib_verbs.c
3656
r_sge[1].lkey = sqp_entry->sge.lkey;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
1634
struct bnxt_qplib_sge *sge)
drivers/infiniband/hw/bnxt_re/qplib_fp.c
1639
memset(sge, 0, sizeof(*sge));
drivers/infiniband/hw/bnxt_re/qplib_fp.c
1643
sge->addr = (dma_addr_t)(qp->sq_hdr_buf_map +
drivers/infiniband/hw/bnxt_re/qplib_fp.c
1645
sge->lkey = 0xFFFFFFFF;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
1646
sge->size = qp->sq_hdr_buf_size;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
1647
return qp->sq_hdr_buf + sw_prod * sge->size;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
1665
struct bnxt_qplib_sge *sge)
drivers/infiniband/hw/bnxt_re/qplib_fp.c
1670
memset(sge, 0, sizeof(*sge));
drivers/infiniband/hw/bnxt_re/qplib_fp.c
1674
sge->addr = (dma_addr_t)(qp->rq_hdr_buf_map +
drivers/infiniband/hw/bnxt_re/qplib_fp.c
1676
sge->lkey = 0xFFFFFFFF;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
1677
sge->size = qp->rq_hdr_buf_size;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
1678
return qp->rq_hdr_buf + sw_prod * sge->size;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2170
struct sq_sge *sge;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2172
sge = bnxt_qplib_get_prod_qe(hwq, idx++);
drivers/infiniband/hw/bnxt_re/qplib_fp.c
2173
sge->size = 0;
drivers/infiniband/hw/bnxt_re/qplib_fp.h
570
struct bnxt_qplib_sge *sge);
drivers/infiniband/hw/bnxt_re/qplib_fp.h
572
struct bnxt_qplib_sge *sge);
drivers/infiniband/hw/cxgb4/qp.c
462
flitp = (__be64 *)isglp->sge;
drivers/infiniband/hw/cxgb4/t4fw_ri_api.h
125
struct fw_ri_dsge_pair sge[];
drivers/infiniband/hw/cxgb4/t4fw_ri_api.h
139
struct fw_ri_sge sge[];
drivers/infiniband/hw/erdma/erdma_qp.c
455
struct erdma_sge *sge;
drivers/infiniband/hw/erdma/erdma_qp.c
527
sge = get_queue_entry(qp->kern_qp.sq_buf, idx + 1,
drivers/infiniband/hw/erdma/erdma_qp.c
529
sge->addr = cpu_to_le64(rdma_wr->remote_addr);
drivers/infiniband/hw/erdma/erdma_qp.c
530
sge->key = cpu_to_le32(rdma_wr->rkey);
drivers/infiniband/hw/erdma/erdma_qp.c
531
sge->length = cpu_to_le32(send_wr->sg_list[0].length);
drivers/infiniband/hw/erdma/erdma_qp.c
608
sge = get_queue_entry(qp->kern_qp.sq_buf, idx + 1,
drivers/infiniband/hw/erdma/erdma_qp.c
610
sge->addr = cpu_to_le64(atomic_wr(send_wr)->remote_addr);
drivers/infiniband/hw/erdma/erdma_qp.c
611
sge->key = cpu_to_le32(atomic_wr(send_wr)->rkey);
drivers/infiniband/hw/erdma/erdma_qp.c
612
sge++;
drivers/infiniband/hw/erdma/erdma_qp.c
614
sge->addr = cpu_to_le64(send_wr->sg_list[0].addr);
drivers/infiniband/hw/erdma/erdma_qp.c
615
sge->key = cpu_to_le32(send_wr->sg_list[0].lkey);
drivers/infiniband/hw/erdma/erdma_qp.c
616
sge->length = cpu_to_le32(send_wr->sg_list[0].length);
drivers/infiniband/hw/hfi1/rc.c
158
qp->s_ack_rdma_sge.sge = e->rdma_sge;
drivers/infiniband/hw/hfi1/rc.c
207
qp->s_ack_rdma_sge.sge = e->rdma_sge;
drivers/infiniband/hw/hfi1/rc.c
231
ps->s_txreq->mr = qp->s_ack_rdma_sge.sge.mr;
drivers/infiniband/hw/hfi1/rc.c
234
len = qp->s_ack_rdma_sge.sge.sge_length;
drivers/infiniband/hw/hfi1/rc.c
2987
ok = rvt_rkey_ok(qp, &qp->r_sge.sge, qp->r_len, vaddr,
drivers/infiniband/hw/hfi1/rc.c
2994
qp->r_sge.sge.mr = NULL;
drivers/infiniband/hw/hfi1/rc.c
2995
qp->r_sge.sge.vaddr = NULL;
drivers/infiniband/hw/hfi1/rc.c
2996
qp->r_sge.sge.length = 0;
drivers/infiniband/hw/hfi1/rc.c
2997
qp->r_sge.sge.sge_length = 0;
drivers/infiniband/hw/hfi1/rc.c
3118
if (unlikely(!rvt_rkey_ok(qp, &qp->r_sge.sge, sizeof(u64),
drivers/infiniband/hw/hfi1/rc.c
3123
maddr = (atomic64_t *)qp->r_sge.sge.vaddr;
drivers/infiniband/hw/hfi1/rc.c
3127
(u64)cmpxchg((u64 *)qp->r_sge.sge.vaddr,
drivers/infiniband/hw/hfi1/rc.c
3130
rvt_put_mr(qp->r_sge.sge.mr);
drivers/infiniband/hw/hfi1/rc.c
792
qp->s_sge.sge = wqe->sg_list[0];
drivers/infiniband/hw/hfi1/rc.c
876
qp->s_sge.sge = wqe->sg_list[0];
drivers/infiniband/hw/hfi1/tid_rdma.c
1092
struct rvt_sge *sge = &ss->sge;
drivers/infiniband/hw/hfi1/tid_rdma.c
1098
pages[i++] = virt_to_page(sge->vaddr);
drivers/infiniband/hw/hfi1/tid_rdma.c
1100
sge->vaddr += len;
drivers/infiniband/hw/hfi1/tid_rdma.c
1101
sge->length -= len;
drivers/infiniband/hw/hfi1/tid_rdma.c
1102
sge->sge_length -= len;
drivers/infiniband/hw/hfi1/tid_rdma.c
1103
if (!sge->sge_length) {
drivers/infiniband/hw/hfi1/tid_rdma.c
1105
*sge = ss->sg_list[req->isge - 1];
drivers/infiniband/hw/hfi1/tid_rdma.c
1106
} else if (sge->length == 0 && sge->mr->lkey) {
drivers/infiniband/hw/hfi1/tid_rdma.c
1107
if (++sge->n >= RVT_SEGSZ) {
drivers/infiniband/hw/hfi1/tid_rdma.c
1108
++sge->m;
drivers/infiniband/hw/hfi1/tid_rdma.c
1109
sge->n = 0;
drivers/infiniband/hw/hfi1/tid_rdma.c
1111
sge->vaddr = sge->mr->map[sge->m]->segs[sge->n].vaddr;
drivers/infiniband/hw/hfi1/tid_rdma.c
1112
sge->length = sge->mr->map[sge->m]->segs[sge->n].length;
drivers/infiniband/hw/hfi1/tid_rdma.c
1727
wpriv->ss.sge.vaddr = req_addr;
drivers/infiniband/hw/hfi1/tid_rdma.c
1728
wpriv->ss.sge.sge_length = req_len;
drivers/infiniband/hw/hfi1/tid_rdma.c
1729
wpriv->ss.sge.length = wpriv->ss.sge.sge_length;
drivers/infiniband/hw/hfi1/tid_rdma.c
1734
wpriv->ss.sge.mr = NULL;
drivers/infiniband/hw/hfi1/tid_rdma.c
1735
wpriv->ss.sge.m = 0;
drivers/infiniband/hw/hfi1/tid_rdma.c
1736
wpriv->ss.sge.n = 0;
drivers/infiniband/hw/hfi1/tid_rdma.c
1739
wpriv->ss.total_len = wpriv->ss.sge.sge_length;
drivers/infiniband/hw/hfi1/tid_rdma.c
3269
struct rvt_sge *sge, int num_sge)
drivers/infiniband/hw/hfi1/tid_rdma.c
3273
for (i = 0; i < num_sge; i++, sge++) {
drivers/infiniband/hw/hfi1/tid_rdma.c
3274
trace_hfi1_sge_check_align(qp, i, sge);
drivers/infiniband/hw/hfi1/tid_rdma.c
3275
if ((u64)sge->vaddr & ~PAGE_MASK ||
drivers/infiniband/hw/hfi1/tid_rdma.c
3276
sge->sge_length & ~PAGE_MASK)
drivers/infiniband/hw/hfi1/tid_rdma.c
3768
req->ss.sge = e->rdma_sge;
drivers/infiniband/hw/hfi1/tid_rdma.c
3894
epriv->ss.sge.vaddr = resp_addr;
drivers/infiniband/hw/hfi1/tid_rdma.c
3895
epriv->ss.sge.sge_length = resp_len;
drivers/infiniband/hw/hfi1/tid_rdma.c
3896
epriv->ss.sge.length = epriv->ss.sge.sge_length;
drivers/infiniband/hw/hfi1/tid_rdma.c
3901
epriv->ss.sge.mr = NULL;
drivers/infiniband/hw/hfi1/tid_rdma.c
3902
epriv->ss.sge.m = 0;
drivers/infiniband/hw/hfi1/tid_rdma.c
3903
epriv->ss.sge.n = 0;
drivers/infiniband/hw/hfi1/tid_rdma.c
3906
epriv->ss.total_len = epriv->ss.sge.sge_length;
drivers/infiniband/hw/hfi1/tid_rdma.c
4333
ss.sge = e->rdma_sge;
drivers/infiniband/hw/hfi1/tid_rdma.c
5076
priv->tid_ss.sge = wqe->sg_list[0];
drivers/infiniband/hw/hfi1/trace_tid.h
1324
TP_PROTO(struct rvt_qp *qp, int index, struct rvt_sge *sge),
drivers/infiniband/hw/hfi1/trace_tid.h
1325
TP_ARGS(qp, index, sge),
drivers/infiniband/hw/hfi1/trace_tid.h
1337
__entry->vaddr = (u64)sge->vaddr;
drivers/infiniband/hw/hfi1/trace_tid.h
1338
__entry->sge_length = sge->sge_length;
drivers/infiniband/hw/hfi1/trace_tid.h
1352
TP_PROTO(struct rvt_qp *qp, int index, struct rvt_sge *sge),
drivers/infiniband/hw/hfi1/trace_tid.h
1353
TP_ARGS(qp, index, sge)
drivers/infiniband/hw/hfi1/uc.c
112
qp->s_sge.sge = wqe->sg_list[0];
drivers/infiniband/hw/hfi1/uc.c
454
ok = rvt_rkey_ok(qp, &qp->r_sge.sge, qp->r_len,
drivers/infiniband/hw/hfi1/uc.c
461
qp->r_sge.sge.mr = NULL;
drivers/infiniband/hw/hfi1/uc.c
462
qp->r_sge.sge.vaddr = NULL;
drivers/infiniband/hw/hfi1/uc.c
463
qp->r_sge.sge.length = 0;
drivers/infiniband/hw/hfi1/uc.c
464
qp->r_sge.sge.sge_length = 0;
drivers/infiniband/hw/hfi1/ud.c
179
ssge.sge = *swqe->sg_list;
drivers/infiniband/hw/hfi1/ud.c
181
sge = &ssge.sge;
drivers/infiniband/hw/hfi1/ud.c
183
u32 len = rvt_get_sge_length(sge, length);
drivers/infiniband/hw/hfi1/ud.c
186
rvt_copy_sge(qp, &qp->r_sge, sge->vaddr, len, true, false);
drivers/infiniband/hw/hfi1/ud.c
40
struct rvt_sge *sge;
drivers/infiniband/hw/hfi1/ud.c
514
qp->s_sge.sge = wqe->sg_list[0];
drivers/infiniband/hw/hfi1/verbs.c
1040
void *addr = ss->sge.vaddr;
drivers/infiniband/hw/hfi1/verbs.c
1041
u32 slen = rvt_get_sge_length(&ss->sge, len);
drivers/infiniband/hw/hfi1/verbs.c
661
struct rvt_sge sge = ss->sge;
drivers/infiniband/hw/hfi1/verbs.c
667
len = rvt_get_sge_length(&ss->sge, length);
drivers/infiniband/hw/hfi1/verbs.c
672
ss->sge.vaddr,
drivers/infiniband/hw/hfi1/verbs.c
682
ss->sge = sge;
drivers/infiniband/hw/hns/hns_roce_device.h
638
struct hns_roce_sge sge;
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
215
dseg = hns_roce_get_extend_sge(qp, idx & (qp->sge.sge_cnt - 1));
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
246
idx & (qp->sge.sge_cnt - 1));
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
256
static void set_extend_sge(struct hns_roce_qp *qp, struct ib_sge *sge,
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
263
dseg = hns_roce_get_extend_sge(qp, idx & (qp->sge.sge_cnt - 1));
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
264
if (likely(sge->length)) {
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
265
set_data_seg_v2(dseg, sge);
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
269
sge++;
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
4553
to_hr_hem_entries_shift(hr_qp->sge.sge_cnt,
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
4554
hr_qp->sge.sge_shift));
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
4699
hr_qp->sge.sge_cnt));
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
4753
if (hr_qp->sge.sge_cnt > 0) {
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
4755
hr_qp->sge.offset, &sge_cur_blk, 1);
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
487
curr_idx & (qp->sge.sge_cnt - 1));
drivers/infiniband/hw/hns/hns_roce_hw_v2.c
591
curr_idx & (qp->sge.sge_cnt - 1));
drivers/infiniband/hw/hns/hns_roce_qp.c
1578
return get_wqe(hr_qp, hr_qp->sge.offset + (n << hr_qp->sge.sge_shift));
drivers/infiniband/hw/hns/hns_roce_qp.c
613
hr_qp->sge.sge_shift = HNS_ROCE_SGE_SHIFT;
drivers/infiniband/hw/hns/hns_roce_qp.c
640
hr_qp->sge.sge_cnt = max(total_sge_cnt,
drivers/infiniband/hw/hns/hns_roce_qp.c
718
hr_qp->sge.offset = hr_qp->buff_size;
drivers/infiniband/hw/hns/hns_roce_qp.c
719
buf_size = to_hr_hem_entries_size(hr_qp->sge.sge_cnt,
drivers/infiniband/hw/hns/hns_roce_qp.c
720
hr_qp->sge.sge_shift);
drivers/infiniband/hw/hns/hns_roce_restrack.c
82
if (rdma_nl_put_driver_u32_hex(msg, "ext_sge_sge_cnt", hr_qp->sge.sge_cnt))
drivers/infiniband/hw/ionic/ionic_datapath.c
1044
wqe->atomic.sge.va = cpu_to_be64(wr->wr.sg_list[0].addr);
drivers/infiniband/hw/ionic/ionic_datapath.c
1045
wqe->atomic.sge.len = cpu_to_be32(8);
drivers/infiniband/hw/ionic/ionic_datapath.c
1046
wqe->atomic.sge.lkey = cpu_to_be32(wr->wr.sg_list[0].lkey);
drivers/infiniband/hw/ionic/ionic_fw.h
467
struct ionic_sge sge;
drivers/infiniband/hw/ionic/ionic_fw.h
581
struct ionic_sge *sge = (void *)(1ull << stride_log2);
drivers/infiniband/hw/ionic/ionic_fw.h
586
sge -= 1;
drivers/infiniband/hw/ionic/ionic_fw.h
591
num_sge = sge - &wqe->common.pld.sgl[num_sge];
drivers/infiniband/hw/ionic/ionic_fw.h
633
struct ionic_sge *sge = (void *)(1ull << stride_log2);
drivers/infiniband/hw/ionic/ionic_fw.h
638
sge -= 1;
drivers/infiniband/hw/ionic/ionic_fw.h
643
num_sge = sge - &wqe->recv.pld.sgl[num_sge];
drivers/infiniband/hw/irdma/uk.c
1465
void irdma_get_wqe_shift(struct irdma_uk_attrs *uk_attrs, u32 sge,
drivers/infiniband/hw/irdma/uk.c
1470
if (sge > 1 || inline_data > 8) {
drivers/infiniband/hw/irdma/uk.c
1471
if (sge < 4 && inline_data <= 39)
drivers/infiniband/hw/irdma/uk.c
1473
else if (sge < 8 && inline_data <= 101)
drivers/infiniband/hw/irdma/uk.c
1478
} else if (sge > 1 || inline_data > 16) {
drivers/infiniband/hw/irdma/uk.c
1479
*shift = (sge < 4 && inline_data <= 48) ? 1 : 2;
drivers/infiniband/hw/irdma/uk.c
15
static void irdma_set_fragment(__le64 *wqe, u32 offset, struct ib_sge *sge,
drivers/infiniband/hw/irdma/uk.c
18
if (sge) {
drivers/infiniband/hw/irdma/uk.c
20
FIELD_PREP(IRDMAQPSQ_FRAG_TO, sge->addr));
drivers/infiniband/hw/irdma/uk.c
23
FIELD_PREP(IRDMAQPSQ_FRAG_LEN, sge->length) |
drivers/infiniband/hw/irdma/uk.c
24
FIELD_PREP(IRDMAQPSQ_FRAG_STAG, sge->lkey));
drivers/infiniband/hw/irdma/uk.c
40
struct ib_sge *sge, u8 valid)
drivers/infiniband/hw/irdma/uk.c
42
if (sge) {
drivers/infiniband/hw/irdma/uk.c
44
FIELD_PREP(IRDMAQPSQ_FRAG_TO, sge->addr));
drivers/infiniband/hw/irdma/uk.c
46
FIELD_PREP(IRDMAQPSQ_GEN1_FRAG_LEN, sge->length) |
drivers/infiniband/hw/irdma/uk.c
47
FIELD_PREP(IRDMAQPSQ_GEN1_FRAG_STAG, sge->lkey));
drivers/infiniband/hw/irdma/uk.c
952
struct ib_sge sge = {};
drivers/infiniband/hw/irdma/uk.c
964
sge.lkey = op_info->target_stag;
drivers/infiniband/hw/irdma/uk.c
965
qp->wqe_ops.iw_set_fragment(wqe, 0, &sge, 0);
drivers/infiniband/hw/irdma/user.h
426
void (*iw_set_fragment)(__le64 *wqe, u32 offset, struct ib_sge *sge,
drivers/infiniband/hw/irdma/user.h
589
void irdma_get_wqe_shift(struct irdma_uk_attrs *uk_attrs, u32 sge,
drivers/infiniband/hw/mana/qp.c
400
static u32 mana_ib_wqe_size(u32 sge, u32 oob_size)
drivers/infiniband/hw/mana/qp.c
402
u32 wqe_size = sge * sizeof(struct gdma_sge) + sizeof(struct gdma_wqe) + oob_size;
drivers/infiniband/hw/ocrdma/ocrdma_hw.c
1301
mqe->u.nonemb_req.sge[0].pa_lo = (u32) (dev->stats_mem.pa & 0xffffffff);
drivers/infiniband/hw/ocrdma/ocrdma_hw.c
1302
mqe->u.nonemb_req.sge[0].pa_hi = (u32) upper_32_bits(dev->stats_mem.pa);
drivers/infiniband/hw/ocrdma/ocrdma_hw.c
1303
mqe->u.nonemb_req.sge[0].len = dev->stats_mem.size;
drivers/infiniband/hw/ocrdma/ocrdma_hw.c
1349
mqe->u.nonemb_req.sge[0].pa_lo = (u32) (dma.pa & 0xffffffff);
drivers/infiniband/hw/ocrdma/ocrdma_hw.c
1350
mqe->u.nonemb_req.sge[0].pa_hi = (u32) upper_32_bits(dma.pa);
drivers/infiniband/hw/ocrdma/ocrdma_hw.c
1351
mqe->u.nonemb_req.sge[0].len = dma.size;
drivers/infiniband/hw/ocrdma/ocrdma_hw.c
2882
struct ocrdma_mqe_sge *mqe_sge = cmd.u.nonemb_req.sge;
drivers/infiniband/hw/ocrdma/ocrdma_sli.h
288
struct ocrdma_mqe_sge sge[19];
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
1895
struct ocrdma_sge *sge, int num_sge,
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
1901
sge[i].lrkey = sg_list[i].lkey;
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
1902
sge[i].addr_lo = sg_list[i].addr;
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
1903
sge[i].addr_hi = upper_32_bits(sg_list[i].addr);
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
1904
sge[i].len = sg_list[i].length;
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
1908
memset(sge, 0, sizeof(*sge));
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
1923
struct ocrdma_sge *sge,
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
1937
dpp_addr = (char *)sge;
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
1950
ocrdma_build_sges(hdr, sge, wr->num_sge, wr->sg_list);
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
1964
struct ocrdma_sge *sge;
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
1969
sge = (struct ocrdma_sge *)(hdr + 2);
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
1972
sge = (struct ocrdma_sge *)(hdr + 1);
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
1975
return ocrdma_build_inline_sges(qp, hdr, sge, wr, wqe_size);
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
1983
struct ocrdma_sge *sge = ext_rw + 1;
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
1986
status = ocrdma_build_inline_sges(qp, hdr, sge, wr, wqe_size);
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2000
struct ocrdma_sge *sge = ext_rw + 1;
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2004
ocrdma_build_sges(hdr, sge, wr->num_sge, wr->sg_list);
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2204
struct ocrdma_sge *sge;
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2206
wqe_size = (wr->num_sge * sizeof(*sge)) + sizeof(*rqe);
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2208
wqe_size = sizeof(*sge) + sizeof(*rqe);
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2216
sge = (struct ocrdma_sge *)(rqe + 1);
drivers/infiniband/hw/ocrdma/ocrdma_verbs.c
2217
ocrdma_build_sges(rqe, sge, wr->num_sge, wr->sg_list);
drivers/infiniband/hw/qedr/qedr_hsi_rdma.h
178
struct rdma_srq_sge sge;
drivers/infiniband/hw/qedr/verbs.c
3358
#define RQ_SGE_SET(sge, vaddr, vlength, vflags) \
drivers/infiniband/hw/qedr/verbs.c
3360
DMA_REGPAIR_LE(sge->addr, vaddr); \
drivers/infiniband/hw/qedr/verbs.c
3361
(sge)->length = cpu_to_le32(vlength); \
drivers/infiniband/hw/qedr/verbs.c
3362
(sge)->flags = cpu_to_le32(vflags); \
drivers/infiniband/hw/qedr/verbs.c
3371
#define SRQ_SGE_SET(sge, vaddr, vlength, vlkey) \
drivers/infiniband/hw/qedr/verbs.c
3373
DMA_REGPAIR_LE(sge->addr, vaddr); \
drivers/infiniband/hw/qedr/verbs.c
3374
(sge)->length = cpu_to_le32(vlength); \
drivers/infiniband/hw/qedr/verbs.c
3375
(sge)->l_key = cpu_to_le32(vlkey); \
drivers/infiniband/hw/qedr/verbs.c
3385
struct rdma_sq_sge *sge = qed_chain_produce(&qp->sq.pbl);
drivers/infiniband/hw/qedr/verbs.c
3387
DMA_REGPAIR_LE(sge->addr, wr->sg_list[i].addr);
drivers/infiniband/hw/qedr/verbs.c
3388
sge->l_key = cpu_to_le32(wr->sg_list[i].lkey);
drivers/infiniband/hw/qedr/verbs.c
3389
sge->length = cpu_to_le32(wr->sg_list[i].length);
drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c
678
struct pvrdma_sge *sge;
drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c
835
sge = (struct pvrdma_sge *)(wqe_hdr + 1);
drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c
838
sge->addr = wr->sg_list[i].addr;
drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c
839
sge->length = wr->sg_list[i].length;
drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c
840
sge->lkey = wr->sg_list[i].lkey;
drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c
841
sge++;
drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c
880
struct pvrdma_sge *sge;
drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c
927
sge = (struct pvrdma_sge *)(wqe_hdr + 1);
drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c
929
sge->addr = wr->sg_list[i].addr;
drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c
930
sge->length = wr->sg_list[i].length;
drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c
931
sge->lkey = wr->sg_list[i].lkey;
drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c
932
sge++;
drivers/infiniband/sw/rdmavt/mr.c
497
rval = rvt_mr_has_lkey(ss->sge.mr, lkey);
drivers/infiniband/sw/rdmavt/mr.c
687
struct ib_sge *sge)
drivers/infiniband/sw/rdmavt/mr.c
689
if (last_sge && sge->lkey == last_sge->mr->lkey &&
drivers/infiniband/sw/rdmavt/mr.c
690
((uint64_t)(last_sge->vaddr + last_sge->length) == sge->addr)) {
drivers/infiniband/sw/rdmavt/mr.c
691
if (sge->lkey) {
drivers/infiniband/sw/rdmavt/mr.c
692
if (unlikely((sge->addr - last_sge->mr->user_base +
drivers/infiniband/sw/rdmavt/mr.c
693
sge->length > last_sge->mr->length)))
drivers/infiniband/sw/rdmavt/mr.c
696
last_sge->length += sge->length;
drivers/infiniband/sw/rdmavt/mr.c
698
last_sge->sge_length += sge->length;
drivers/infiniband/sw/rdmavt/mr.c
699
trace_rvt_sge_adjacent(last_sge, sge);
drivers/infiniband/sw/rdmavt/mr.c
723
struct ib_sge *sge, int acc)
drivers/infiniband/sw/rdmavt/mr.c
733
if (sge->lkey == 0) {
drivers/infiniband/sw/rdmavt/mr.c
738
if (rvt_sge_adjacent(last_sge, sge))
drivers/infiniband/sw/rdmavt/mr.c
748
isge->vaddr = (void *)sge->addr;
drivers/infiniband/sw/rdmavt/mr.c
749
isge->length = sge->length;
drivers/infiniband/sw/rdmavt/mr.c
750
isge->sge_length = sge->length;
drivers/infiniband/sw/rdmavt/mr.c
755
if (rvt_sge_adjacent(last_sge, sge))
drivers/infiniband/sw/rdmavt/mr.c
758
mr = rcu_dereference(rkt->table[sge->lkey >> rkt->shift]);
drivers/infiniband/sw/rdmavt/mr.c
766
mr->lkey != sge->lkey || mr->pd != &pd->ibpd))
drivers/infiniband/sw/rdmavt/mr.c
769
off = sge->addr - mr->user_base;
drivers/infiniband/sw/rdmavt/mr.c
770
if (unlikely(sge->addr < mr->user_base ||
drivers/infiniband/sw/rdmavt/mr.c
771
off + sge->length > mr->length ||
drivers/infiniband/sw/rdmavt/mr.c
804
isge->sge_length = sge->length;
drivers/infiniband/sw/rdmavt/mr.c
808
trace_rvt_sge_new(isge, sge);
drivers/infiniband/sw/rdmavt/mr.c
831
int rvt_rkey_ok(struct rvt_qp *qp, struct rvt_sge *sge,
drivers/infiniband/sw/rdmavt/mr.c
857
sge->mr = mr;
drivers/infiniband/sw/rdmavt/mr.c
858
sge->vaddr = (void *)vaddr;
drivers/infiniband/sw/rdmavt/mr.c
859
sge->length = len;
drivers/infiniband/sw/rdmavt/mr.c
860
sge->sge_length = len;
drivers/infiniband/sw/rdmavt/mr.c
861
sge->m = 0;
drivers/infiniband/sw/rdmavt/mr.c
862
sge->n = 0;
drivers/infiniband/sw/rdmavt/mr.c
908
sge->mr = mr;
drivers/infiniband/sw/rdmavt/mr.c
909
sge->vaddr = mr->map[m]->segs[n].vaddr + off;
drivers/infiniband/sw/rdmavt/mr.c
910
sge->length = mr->map[m]->segs[n].length - off;
drivers/infiniband/sw/rdmavt/mr.c
911
sge->sge_length = len;
drivers/infiniband/sw/rdmavt/mr.c
912
sge->m = m;
drivers/infiniband/sw/rdmavt/mr.c
913
sge->n = n;
drivers/infiniband/sw/rdmavt/qp.c
2124
struct rvt_sge *sge = &wqe->sg_list[--j];
drivers/infiniband/sw/rdmavt/qp.c
2126
rvt_put_mr(sge->mr);
drivers/infiniband/sw/rdmavt/qp.c
2250
static struct ib_sge *rvt_cast_sge(struct rvt_wqe_sge *sge)
drivers/infiniband/sw/rdmavt/qp.c
2258
return (struct ib_sge *)sge;
drivers/infiniband/sw/rdmavt/qp.c
2283
ret = rvt_lkey_ok(rkt, pd, j ? &ss->sg_list[j - 1] : &ss->sge,
drivers/infiniband/sw/rdmavt/qp.c
2297
struct rvt_sge *sge = --j ? &ss->sg_list[j - 1] : &ss->sge;
drivers/infiniband/sw/rdmavt/qp.c
2299
rvt_put_mr(sge->mr);
drivers/infiniband/sw/rdmavt/qp.c
2814
struct rvt_sge *sge = &ss->sge;
drivers/infiniband/sw/rdmavt/qp.c
2831
wss_insert(wss, sge->vaddr);
drivers/infiniband/sw/rdmavt/qp.c
2833
wss_insert(wss, (sge->vaddr + PAGE_SIZE));
drivers/infiniband/sw/rdmavt/qp.c
2852
u32 len = rvt_get_sge_length(sge, length);
drivers/infiniband/sw/rdmavt/qp.c
2858
((u8 *)sge->vaddr)[i] = ((u8 *)data)[i];
drivers/infiniband/sw/rdmavt/qp.c
2860
cacheless_memcpy(sge->vaddr, data, len);
drivers/infiniband/sw/rdmavt/qp.c
2862
memcpy(sge->vaddr, data, len);
drivers/infiniband/sw/rdmavt/qp.c
2906
struct rvt_sge *sge;
drivers/infiniband/sw/rdmavt/qp.c
2977
sqp->s_sge.sge = wqe->sg_list[0];
drivers/infiniband/sw/rdmavt/qp.c
3041
if (unlikely(!rvt_rkey_ok(qp, &qp->r_sge.sge, wqe->length,
drivers/infiniband/sw/rdmavt/qp.c
3054
if (unlikely(!rvt_rkey_ok(qp, &sqp->s_sge.sge, wqe->length,
drivers/infiniband/sw/rdmavt/qp.c
3062
qp->r_sge.sge = wqe->sg_list[0];
drivers/infiniband/sw/rdmavt/qp.c
3074
if (unlikely(!rvt_rkey_ok(qp, &qp->r_sge.sge, sizeof(u64),
drivers/infiniband/sw/rdmavt/qp.c
3080
maddr = (atomic64_t *)qp->r_sge.sge.vaddr;
drivers/infiniband/sw/rdmavt/qp.c
3082
*(u64 *)sqp->s_sge.sge.vaddr =
drivers/infiniband/sw/rdmavt/qp.c
3085
(u64)cmpxchg((u64 *)qp->r_sge.sge.vaddr,
drivers/infiniband/sw/rdmavt/qp.c
3087
rvt_put_mr(qp->r_sge.sge.mr);
drivers/infiniband/sw/rdmavt/qp.c
3096
sge = &sqp->s_sge.sge;
drivers/infiniband/sw/rdmavt/qp.c
3098
u32 len = rvt_get_sge_length(sge, sqp->s_len);
drivers/infiniband/sw/rdmavt/qp.c
3101
rvt_copy_sge(qp, &qp->r_sge, sge->vaddr,
drivers/infiniband/sw/rdmavt/qp.c
625
struct rvt_sge *sge = &wqe->sg_list[i];
drivers/infiniband/sw/rdmavt/qp.c
627
if (rvt_mr_has_lkey(sge->mr, lkey))
drivers/infiniband/sw/rdmavt/rc.c
164
ss->sge = wqe->sg_list[0];
drivers/infiniband/sw/rdmavt/trace_mr.h
100
RDI_DEV_ASSIGN(ib_to_rvt(sge->mr->pd->device));
drivers/infiniband/sw/rdmavt/trace_mr.h
101
__entry->mr = sge->mr;
drivers/infiniband/sw/rdmavt/trace_mr.h
102
__entry->sge = sge;
drivers/infiniband/sw/rdmavt/trace_mr.h
104
__entry->vaddr = sge->vaddr;
drivers/infiniband/sw/rdmavt/trace_mr.h
106
__entry->lkey = sge->mr->lkey;
drivers/infiniband/sw/rdmavt/trace_mr.h
107
__entry->sge_length = sge->sge_length;
drivers/infiniband/sw/rdmavt/trace_mr.h
108
__entry->length = sge->length;
drivers/infiniband/sw/rdmavt/trace_mr.h
110
__entry->m = sge->m;
drivers/infiniband/sw/rdmavt/trace_mr.h
111
__entry->n = sge->m;
drivers/infiniband/sw/rdmavt/trace_mr.h
112
__entry->user = ibpd_to_rvtpd(sge->mr->pd)->user;
drivers/infiniband/sw/rdmavt/trace_mr.h
118
__entry->sge,
drivers/infiniband/sw/rdmavt/trace_mr.h
134
TP_PROTO(struct rvt_sge *sge, struct ib_sge *isge),
drivers/infiniband/sw/rdmavt/trace_mr.h
135
TP_ARGS(sge, isge));
drivers/infiniband/sw/rdmavt/trace_mr.h
139
TP_PROTO(struct rvt_sge *sge, struct ib_sge *isge),
drivers/infiniband/sw/rdmavt/trace_mr.h
140
TP_ARGS(sge, isge));
drivers/infiniband/sw/rdmavt/trace_mr.h
82
TP_PROTO(struct rvt_sge *sge, struct ib_sge *isge),
drivers/infiniband/sw/rdmavt/trace_mr.h
83
TP_ARGS(sge, isge),
drivers/infiniband/sw/rdmavt/trace_mr.h
85
RDI_DEV_ENTRY(ib_to_rvt(sge->mr->pd->device))
drivers/infiniband/sw/rdmavt/trace_mr.h
87
__field(struct rvt_sge *, sge)
drivers/infiniband/sw/rxe/rxe_mr.c
424
struct rxe_sge *sge = &dma->sge[dma->cur_sge];
drivers/infiniband/sw/rxe/rxe_mr.c
439
if (sge->length && (offset < sge->length)) {
drivers/infiniband/sw/rxe/rxe_mr.c
440
mr = lookup_mr(pd, access, sge->lkey, RXE_LOOKUP_LOCAL);
drivers/infiniband/sw/rxe/rxe_mr.c
450
if (offset >= sge->length) {
drivers/infiniband/sw/rxe/rxe_mr.c
455
sge++;
drivers/infiniband/sw/rxe/rxe_mr.c
464
if (sge->length) {
drivers/infiniband/sw/rxe/rxe_mr.c
465
mr = lookup_mr(pd, access, sge->lkey,
drivers/infiniband/sw/rxe/rxe_mr.c
476
if (bytes > sge->length - offset)
drivers/infiniband/sw/rxe/rxe_mr.c
477
bytes = sge->length - offset;
drivers/infiniband/sw/rxe/rxe_mr.c
480
iova = sge->addr + offset;
drivers/infiniband/sw/rxe/rxe_mr.c
670
struct rxe_sge *sge = &dma->sge[dma->cur_sge];
drivers/infiniband/sw/rxe/rxe_mr.c
677
if (offset >= sge->length) {
drivers/infiniband/sw/rxe/rxe_mr.c
678
sge++;
drivers/infiniband/sw/rxe/rxe_mr.c
687
if (bytes > sge->length - offset)
drivers/infiniband/sw/rxe/rxe_mr.c
688
bytes = sge->length - offset;
drivers/infiniband/sw/rxe/rxe_resp.c
353
recv_buffer_len += qp->resp.wqe->dma.sge[i].length;
drivers/infiniband/sw/rxe/rxe_verbs.c
1010
memcpy(recv_wqe->dma.sge, ibwr->sg_list,
drivers/infiniband/sw/rxe/rxe_verbs.c
831
struct ib_sge *sge = ibwr->sg_list;
drivers/infiniband/sw/rxe/rxe_verbs.c
835
for (i = 0; i < ibwr->num_sge; i++, sge++) {
drivers/infiniband/sw/rxe/rxe_verbs.c
836
memcpy(p, ib_virt_dma_to_ptr(sge->addr), sge->length);
drivers/infiniband/sw/rxe/rxe_verbs.c
837
p += sge->length;
drivers/infiniband/sw/rxe/rxe_verbs.c
862
memcpy(wqe->dma.sge, ibwr->sg_list,
drivers/infiniband/sw/rxe/rxe_verbs.h
225
struct ib_sge sge[RXE_MAX_SGE];
drivers/infiniband/sw/siw/siw_mem.c
190
int siw_check_sge(struct ib_pd *pd, struct siw_sge *sge, struct siw_mem *mem[],
drivers/infiniband/sw/siw/siw_mem.c
197
if (len + off > sge->length) {
drivers/infiniband/sw/siw/siw_mem.c
202
new = siw_mem_id2obj(sdev, sge->lkey >> 8);
drivers/infiniband/sw/siw/siw_mem.c
204
siw_dbg_pd(pd, "STag unknown: 0x%08x\n", sge->lkey);
drivers/infiniband/sw/siw/siw_mem.c
211
if (unlikely((*mem)->stag != sge->lkey)) {
drivers/infiniband/sw/siw/siw_mem.c
212
siw_dbg_mem((*mem), "STag mismatch: 0x%08x\n", sge->lkey);
drivers/infiniband/sw/siw/siw_mem.c
216
rv = siw_check_mem(pd, *mem, sge->laddr + off, perms, len);
drivers/infiniband/sw/siw/siw_mem.h
18
int siw_check_sge(struct ib_pd *pd, struct siw_sge *sge,
drivers/infiniband/sw/siw/siw_qp.c
253
wqe->sqe.sge[0].length = 0;
drivers/infiniband/sw/siw/siw_qp.c
254
wqe->sqe.sge[0].laddr = 0;
drivers/infiniband/sw/siw/siw_qp.c
255
wqe->sqe.sge[0].lkey = 0;
drivers/infiniband/sw/siw/siw_qp.c
456
rreq->ddp_msn = htonl(wqe->sqe.sge[0].length);
drivers/infiniband/sw/siw/siw_qp.c
461
rreq->read_size = htonl(wqe->sqe.sge[0].length);
drivers/infiniband/sw/siw/siw_qp.c
462
rreq->source_stag = htonl(wqe->sqe.sge[0].lkey);
drivers/infiniband/sw/siw/siw_qp.c
464
cpu_to_be64(wqe->sqe.sge[0].laddr);
drivers/infiniband/sw/siw/siw_qp.c
844
rreq->sge[0].laddr = sqe->sge[0].laddr;
drivers/infiniband/sw/siw/siw_qp.c
845
rreq->sge[0].length = sqe->sge[0].length;
drivers/infiniband/sw/siw/siw_qp.c
846
rreq->sge[0].lkey = sqe->sge[0].lkey;
drivers/infiniband/sw/siw/siw_qp.c
847
rreq->sge[1].lkey = sqe->sge[1].lkey;
drivers/infiniband/sw/siw/siw_qp.c
878
if (wqe->sqe.sge[0].length > SIW_MAX_INLINE) {
drivers/infiniband/sw/siw/siw_qp.c
882
wqe->sqe.sge[0].laddr = (uintptr_t)&wqe->sqe.sge[1];
drivers/infiniband/sw/siw/siw_qp.c
883
wqe->sqe.sge[0].lkey = 0;
drivers/infiniband/sw/siw/siw_qp.c
977
wqe->sqe.sge[0].length = irqe->sge[0].length;
drivers/infiniband/sw/siw/siw_qp.c
978
wqe->sqe.sge[0].laddr = irqe->sge[0].laddr;
drivers/infiniband/sw/siw/siw_qp.c
979
wqe->sqe.sge[0].lkey = irqe->sge[0].lkey;
drivers/infiniband/sw/siw/siw_qp.c
987
wqe->sqe.sge[1].length = irqe->sge[1].length;
drivers/infiniband/sw/siw/siw_qp_rx.c
1265
rv = siw_invalidate_stag(qp->pd, wqe->sqe.sge[0].lkey);
drivers/infiniband/sw/siw/siw_qp_rx.c
176
srx->ddp_stag = wqe->sqe.sge[0].lkey;
drivers/infiniband/sw/siw/siw_qp_rx.c
177
srx->ddp_to = wqe->sqe.sge[0].laddr;
drivers/infiniband/sw/siw/siw_qp_rx.c
367
wqe->rqe.sge[i].laddr = rqe->sge[i].laddr;
drivers/infiniband/sw/siw/siw_qp_rx.c
368
wqe->rqe.sge[i].lkey = rqe->sge[i].lkey;
drivers/infiniband/sw/siw/siw_qp_rx.c
369
wqe->rqe.sge[i].length = rqe->sge[i].length;
drivers/infiniband/sw/siw/siw_qp_rx.c
370
wqe->bytes += wqe->rqe.sge[i].length;
drivers/infiniband/sw/siw/siw_qp_rx.c
471
struct siw_sge *sge;
drivers/infiniband/sw/siw/siw_qp_rx.c
474
sge = &wqe->rqe.sge[frx->sge_idx];
drivers/infiniband/sw/siw/siw_qp_rx.c
476
if (!sge->length) {
drivers/infiniband/sw/siw/siw_qp_rx.c
483
sge_bytes = min(data_bytes, sge->length - frx->sge_off);
drivers/infiniband/sw/siw/siw_qp_rx.c
491
rv = siw_check_sge(pd, sge, mem, IB_ACCESS_LOCAL_WRITE,
drivers/infiniband/sw/siw/siw_qp_rx.c
503
sge->laddr + frx->sge_off, sge_bytes);
drivers/infiniband/sw/siw/siw_qp_rx.c
514
if (frx->sge_off == sge->length) {
drivers/infiniband/sw/siw/siw_qp_rx.c
699
resp->sge[0].length = length;
drivers/infiniband/sw/siw/siw_qp_rx.c
700
resp->sge[0].laddr = laddr;
drivers/infiniband/sw/siw/siw_qp_rx.c
701
resp->sge[0].lkey = lkey;
drivers/infiniband/sw/siw/siw_qp_rx.c
706
resp->sge[1].length = msn;
drivers/infiniband/sw/siw/siw_qp_rx.c
756
wqe->sqe.sge[0].laddr = orqe->sge[0].laddr;
drivers/infiniband/sw/siw/siw_qp_rx.c
757
wqe->sqe.sge[0].lkey = orqe->sge[0].lkey;
drivers/infiniband/sw/siw/siw_qp_rx.c
758
wqe->sqe.sge[0].length = orqe->sge[0].length;
drivers/infiniband/sw/siw/siw_qp_rx.c
761
wqe->bytes = orqe->sge[0].length;
drivers/infiniband/sw/siw/siw_qp_rx.c
788
struct siw_sge *sge;
drivers/infiniband/sw/siw/siw_qp_rx.c
823
sge = wqe->sqe.sge; /* there is only one */
drivers/infiniband/sw/siw/siw_qp_rx.c
830
rv = siw_check_sge(qp->pd, sge, mem, IB_ACCESS_LOCAL_WRITE, 0,
drivers/infiniband/sw/siw/siw_qp_rx.c
849
sge->laddr + wqe->processed, bytes);
drivers/infiniband/sw/siw/siw_qp_tx.c
135
c_tx->pkt.rreq.sink_stag = htonl(wqe->sqe.sge[0].lkey);
drivers/infiniband/sw/siw/siw_qp_tx.c
137
cpu_to_be64(wqe->sqe.sge[0].laddr);
drivers/infiniband/sw/siw/siw_qp_tx.c
140
c_tx->pkt.rreq.read_size = htonl(wqe->sqe.sge[0].length);
drivers/infiniband/sw/siw/siw_qp_tx.c
37
static struct page *siw_get_page(struct siw_mem *mem, struct siw_sge *sge,
drivers/infiniband/sw/siw/siw_qp_tx.c
376
struct siw_sge *sge, unsigned int offset,
drivers/infiniband/sw/siw/siw_qp_tx.c
380
int sge_bytes = min(sge->length - offset, size);
drivers/infiniband/sw/siw/siw_qp_tx.c
382
offset = (sge->laddr + offset) & ~PAGE_MASK;
drivers/infiniband/sw/siw/siw_qp_tx.c
392
sge++;
drivers/infiniband/sw/siw/siw_qp_tx.c
393
sge_bytes = min(sge->length, size - sent);
drivers/infiniband/sw/siw/siw_qp_tx.c
394
offset = sge->laddr & ~PAGE_MASK;
drivers/infiniband/sw/siw/siw_qp_tx.c
41
return siw_get_upage(mem->umem, sge->laddr + offset);
drivers/infiniband/sw/siw/siw_qp_tx.c
43
return siw_get_pblpage(mem, sge->laddr + offset, pbl_idx);
drivers/infiniband/sw/siw/siw_qp_tx.c
441
struct siw_sge *sge = &wqe->sqe.sge[c_tx->sge_idx];
drivers/infiniband/sw/siw/siw_qp_tx.c
470
unsigned int sge_len = min(sge->length - sge_off, data_len);
drivers/infiniband/sw/siw/siw_qp_tx.c
471
unsigned int fp_off = (sge->laddr + sge_off) & ~PAGE_MASK;
drivers/infiniband/sw/siw/siw_qp_tx.c
486
ib_virt_dma_to_ptr(sge->laddr + sge_off);
drivers/infiniband/sw/siw/siw_qp_tx.c
505
p = siw_get_page(mem, sge, sge_off, &pbl_idx);
drivers/infiniband/sw/siw/siw_qp_tx.c
52
struct siw_sge *sge = &wqe->sqe.sge[0];
drivers/infiniband/sw/siw/siw_qp_tx.c
53
u32 bytes = sge->length;
drivers/infiniband/sw/siw/siw_qp_tx.c
538
u64 va = sge->laddr + sge_off;
drivers/infiniband/sw/siw/siw_qp_tx.c
562
if (sge_off == sge->length &&
drivers/infiniband/sw/siw/siw_qp_tx.c
565
sge++;
drivers/infiniband/sw/siw/siw_qp_tx.c
593
rv = siw_0copy_tx(s, page_array, &wqe->sqe.sge[c_tx->sge_idx],
drivers/infiniband/sw/siw/siw_qp_tx.c
62
memcpy(paddr, &wqe->sqe.sge[1], bytes);
drivers/infiniband/sw/siw/siw_qp_tx.c
653
sge = &wqe->sqe.sge[c_tx->sge_idx];
drivers/infiniband/sw/siw/siw_qp_tx.c
654
sge_unsent = sge->length - c_tx->sge_off;
drivers/infiniband/sw/siw/siw_qp_tx.c
660
sge++;
drivers/infiniband/sw/siw/siw_qp_tx.c
661
sge_unsent = sge->length;
drivers/infiniband/sw/siw/siw_qp_tx.c
68
memcpy(paddr, ib_virt_dma_to_ptr(sge->laddr), bytes);
drivers/infiniband/sw/siw/siw_qp_tx.c
70
if (copy_from_user(paddr, u64_to_user_ptr(sge->laddr),
drivers/infiniband/sw/siw/siw_qp_tx.c
74
unsigned int off = sge->laddr & ~PAGE_MASK;
drivers/infiniband/sw/siw/siw_qp_tx.c
766
struct siw_sge *sge = &wqe->sqe.sge[0];
drivers/infiniband/sw/siw/siw_qp_tx.c
772
for (i = 0, len = 0; num_sge; num_sge--, i++, sge++) {
drivers/infiniband/sw/siw/siw_qp_tx.c
776
if (sge->length) {
drivers/infiniband/sw/siw/siw_qp_tx.c
777
int rv = siw_check_sge(pd, sge, &wqe->mem[i], perms, 0,
drivers/infiniband/sw/siw/siw_qp_tx.c
778
sge->length);
drivers/infiniband/sw/siw/siw_qp_tx.c
783
len += sge->length;
drivers/infiniband/sw/siw/siw_qp_tx.c
79
p = siw_get_page(mem, sge, 0, &pbl_idx);
drivers/infiniband/sw/siw/siw_qp_tx.c
831
wqe->bytes = wqe->sqe.sge[0].length;
drivers/infiniband/sw/siw/siw_qp_tx.c
837
wqe->sqe.sge[0].laddr =
drivers/infiniband/sw/siw/siw_qp_tx.c
838
(u64)(uintptr_t)&wqe->sqe.sge[1];
drivers/infiniband/sw/siw/siw_qp_tx.c
93
p = siw_get_page(mem, sge, part, &pbl_idx);
drivers/infiniband/sw/siw/siw_verbs.c
1090
siw_copy_sgl(wr->sg_list, rqe->sge, wr->num_sge);
drivers/infiniband/sw/siw/siw_verbs.c
1808
siw_copy_sgl(wr->sg_list, rqe->sge, wr->num_sge);
drivers/infiniband/sw/siw/siw_verbs.c
655
void *kbuf = &sqe->sge[1];
drivers/infiniband/sw/siw/siw_verbs.c
658
sqe->sge[0].laddr = (uintptr_t)kbuf;
drivers/infiniband/sw/siw/siw_verbs.c
659
sqe->sge[0].lkey = 0;
drivers/infiniband/sw/siw/siw_verbs.c
677
sqe->sge[0].length = max(bytes, 0);
drivers/infiniband/sw/siw/siw_verbs.c
855
siw_copy_sgl(wr->sg_list, sqe->sge,
drivers/infiniband/sw/siw/siw_verbs.c
888
siw_copy_sgl(wr->sg_list, &sqe->sge[0], 1);
drivers/infiniband/sw/siw/siw_verbs.c
904
siw_copy_sgl(wr->sg_list, &sqe->sge[0],
drivers/infiniband/sw/siw/siw_verbs.h
24
static inline void siw_copy_sgl(struct ib_sge *sge, struct siw_sge *siw_sge,
drivers/infiniband/sw/siw/siw_verbs.h
28
siw_sge->laddr = sge->addr;
drivers/infiniband/sw/siw/siw_verbs.h
29
siw_sge->length = sge->length;
drivers/infiniband/sw/siw/siw_verbs.h
30
siw_sge->lkey = sge->lkey;
drivers/infiniband/sw/siw/siw_verbs.h
33
sge++;
drivers/infiniband/ulp/ipoib/ipoib_cm.c
119
struct ib_sge *sge, int id)
drivers/infiniband/ulp/ipoib/ipoib_cm.c
127
sge[i].addr = rx->rx_ring[id].mapping[i];
drivers/infiniband/ulp/ipoib/ipoib_cm.c
330
struct ib_sge *sge)
drivers/infiniband/ulp/ipoib/ipoib_cm.c
336
sge[i].lkey = priv->pd->local_dma_lkey;
drivers/infiniband/ulp/ipoib/ipoib_cm.c
338
sge[0].length = IPOIB_CM_HEAD_SIZE;
drivers/infiniband/ulp/ipoib/ipoib_cm.c
340
sge[i].length = PAGE_SIZE;
drivers/infiniband/ulp/ipoib/ipoib_cm.c
343
wr->sg_list = sge;
drivers/infiniband/ulp/ipoib/ipoib_cm.c
353
struct ib_sge sge[IPOIB_CM_RX_SG];
drivers/infiniband/ulp/ipoib/ipoib_cm.c
369
ipoib_cm_init_rx_wr(dev, &t->wr, t->sge);
drivers/infiniband/ulp/ipoib/ipoib_cm.c
391
ret = ipoib_cm_post_receive_nonsrq(dev, rx, &t->wr, t->sge, i);
drivers/infiniband/ulp/iser/iscsi_iser.h
209
struct ib_sge sge;
drivers/infiniband/ulp/iser/iscsi_iser.h
290
struct ib_sge sge;
drivers/infiniband/ulp/iser/iser_initiator.c
118
hdr->write_va = cpu_to_be64(mem_reg->sge.addr + unsol_sz);
drivers/infiniband/ulp/iser/iser_initiator.c
123
(unsigned long long)mem_reg->sge.addr, unsol_sz);
drivers/infiniband/ulp/iser/iser_initiator.c
129
tx_dsg->addr = mem_reg->sge.addr;
drivers/infiniband/ulp/iser/iser_initiator.c
131
tx_dsg->lkey = mem_reg->sge.lkey;
drivers/infiniband/ulp/iser/iser_initiator.c
447
tx_dsg->addr = mem_reg->sge.addr + buf_offset;
drivers/infiniband/ulp/iser/iser_initiator.c
449
tx_dsg->lkey = mem_reg->sge.lkey;
drivers/infiniband/ulp/iser/iser_initiator.c
71
hdr->read_va = cpu_to_be64(mem_reg->sge.addr);
drivers/infiniband/ulp/iser/iser_initiator.c
75
(unsigned long long)mem_reg->sge.addr);
drivers/infiniband/ulp/iser/iser_memory.c
128
reg->sge.lkey = device->pd->local_dma_lkey;
drivers/infiniband/ulp/iser/iser_memory.c
138
reg->sge.addr = sg_dma_address(&sg[0]);
drivers/infiniband/ulp/iser/iser_memory.c
139
reg->sge.length = sg_dma_len(&sg[0]);
drivers/infiniband/ulp/iser/iser_memory.c
142
" length=0x%x\n", reg->sge.lkey, reg->rkey,
drivers/infiniband/ulp/iser/iser_memory.c
143
reg->sge.addr, reg->sge.length);
drivers/infiniband/ulp/iser/iser_memory.c
293
sig_reg->sge.lkey = mr->lkey;
drivers/infiniband/ulp/iser/iser_memory.c
295
sig_reg->sge.addr = mr->iova;
drivers/infiniband/ulp/iser/iser_memory.c
296
sig_reg->sge.length = mr->length;
drivers/infiniband/ulp/iser/iser_memory.c
299
sig_reg->sge.lkey, sig_reg->rkey, sig_reg->sge.addr,
drivers/infiniband/ulp/iser/iser_memory.c
300
sig_reg->sge.length);
drivers/infiniband/ulp/iser/iser_memory.c
341
reg->sge.lkey = mr->lkey;
drivers/infiniband/ulp/iser/iser_memory.c
343
reg->sge.addr = mr->iova;
drivers/infiniband/ulp/iser/iser_memory.c
344
reg->sge.length = mr->length;
drivers/infiniband/ulp/iser/iser_memory.c
347
reg->sge.lkey, reg->rkey, reg->sge.addr, reg->sge.length);
drivers/infiniband/ulp/iser/iser_verbs.c
813
desc->sge.addr = desc->rsp_dma;
drivers/infiniband/ulp/iser/iser_verbs.c
814
desc->sge.length = ISER_RX_LOGIN_SIZE;
drivers/infiniband/ulp/iser/iser_verbs.c
815
desc->sge.lkey = ib_conn->device->pd->local_dma_lkey;
drivers/infiniband/ulp/iser/iser_verbs.c
819
wr.sg_list = &desc->sge;
drivers/infiniband/ulp/isert/ib_isert.c
885
struct ib_sge sge;
drivers/infiniband/ulp/isert/ib_isert.c
888
memset(&sge, 0, sizeof(struct ib_sge));
drivers/infiniband/ulp/isert/ib_isert.c
889
sge.addr = isert_conn->login_desc->dma_addr +
drivers/infiniband/ulp/isert/ib_isert.c
891
sge.length = ISER_RX_PAYLOAD_SIZE;
drivers/infiniband/ulp/isert/ib_isert.c
892
sge.lkey = isert_conn->device->pd->local_dma_lkey;
drivers/infiniband/ulp/isert/ib_isert.c
895
sge.addr, sge.length, sge.lkey);
drivers/infiniband/ulp/isert/ib_isert.c
901
rx_wr.sg_list = &sge;
drivers/infiniband/ulp/rtrs/rtrs-clt.c
1024
struct ib_sge *sge = req->sge;
drivers/infiniband/ulp/rtrs/rtrs-clt.c
1033
sge[i].addr = req->mr->iova;
drivers/infiniband/ulp/rtrs/rtrs-clt.c
1034
sge[i].length = req->mr->length;
drivers/infiniband/ulp/rtrs/rtrs-clt.c
1035
sge[i].lkey = req->mr->lkey;
drivers/infiniband/ulp/rtrs/rtrs-clt.c
1041
sge[i].addr = sg_dma_address(sg);
drivers/infiniband/ulp/rtrs/rtrs-clt.c
1042
sge[i].length = sg_dma_len(sg);
drivers/infiniband/ulp/rtrs/rtrs-clt.c
1043
sge[i].lkey = clt_path->s.dev->ib_pd->local_dma_lkey;
drivers/infiniband/ulp/rtrs/rtrs-clt.c
1047
sge[i].addr = req->iu->dma_addr;
drivers/infiniband/ulp/rtrs/rtrs-clt.c
1048
sge[i].length = size;
drivers/infiniband/ulp/rtrs/rtrs-clt.c
1049
sge[i].lkey = clt_path->s.dev->ib_pd->local_dma_lkey;
drivers/infiniband/ulp/rtrs/rtrs-clt.c
1062
return rtrs_iu_post_rdma_write_imm(&con->c, req->iu, sge, num_sge,
drivers/infiniband/ulp/rtrs/rtrs-clt.c
1353
kfree(req->sge);
drivers/infiniband/ulp/rtrs/rtrs-clt.c
1385
req->sge = kzalloc_objs(*req->sge, 2);
drivers/infiniband/ulp/rtrs/rtrs-clt.c
1386
if (!req->sge)
drivers/infiniband/ulp/rtrs/rtrs-clt.c
463
struct ib_sge sge;
drivers/infiniband/ulp/rtrs/rtrs-clt.c
472
sge.addr = req->iu->dma_addr;
drivers/infiniband/ulp/rtrs/rtrs-clt.c
473
sge.length = req->sg_size;
drivers/infiniband/ulp/rtrs/rtrs-clt.c
474
sge.lkey = clt_path->s.dev->ib_pd->local_dma_lkey;
drivers/infiniband/ulp/rtrs/rtrs-clt.c
487
return rtrs_iu_post_rdma_write_imm(&con->c, req->iu, &sge, 1,
drivers/infiniband/ulp/rtrs/rtrs-clt.h
105
struct ib_sge *sge;
drivers/infiniband/ulp/rtrs/rtrs-pri.h
317
struct ib_sge *sge, unsigned int num_sge,
drivers/infiniband/ulp/rtrs/rtrs.c
155
struct ib_sge *sge, unsigned int num_sge,
drivers/infiniband/ulp/rtrs/rtrs.c
166
.wr.sg_list = sge,
drivers/infiniband/ulp/rtrs/rtrs.c
180
if (WARN_ONCE(sge[i].length == 0, "sg %d is zero length\n", i))
drivers/infiniband/ulp/srp/ib_srp.c
1666
struct ib_sge *sge = &req->cmd->sge[1];
drivers/infiniband/ulp/srp/ib_srp.c
1675
sge[i].addr = sg_dma_address(sg);
drivers/infiniband/ulp/srp/ib_srp.c
1676
sge[i].length = sg_dma_len(sg);
drivers/infiniband/ulp/srp/ib_srp.c
1677
sge[i].lkey = target->lkey;
drivers/infiniband/ulp/srp/ib_srp.c
1898
iu->sge[0].addr = iu->dma;
drivers/infiniband/ulp/srp/ib_srp.c
1899
iu->sge[0].length = len;
drivers/infiniband/ulp/srp/ib_srp.c
1900
iu->sge[0].lkey = target->lkey;
drivers/infiniband/ulp/srp/ib_srp.c
1906
wr.sg_list = &iu->sge[0];
drivers/infiniband/ulp/srp/ib_srp.h
278
struct ib_sge sge[SRP_MAX_SGE];
drivers/infiniband/ulp/srpt/ib_srpt.c
2876
struct ib_sge sge;
drivers/infiniband/ulp/srpt/ib_srpt.c
2933
sge.addr = ioctx->ioctx.dma;
drivers/infiniband/ulp/srpt/ib_srpt.c
2934
sge.length = resp_len;
drivers/infiniband/ulp/srpt/ib_srpt.c
2935
sge.lkey = sdev->lkey;
drivers/infiniband/ulp/srpt/ib_srpt.c
2940
send_wr.sg_list = &sge;
drivers/message/fusion/mptbase.h
334
SGESimple32_t sge[1];
drivers/message/fusion/mptbase.h
338
SGESimple64_t sge[1];
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
1356
struct eth_rx_sge *sge;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
1358
sge = &fp->rx_sge_ring[RX_SGE_CNT * i - 2];
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
1359
sge->addr_hi =
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
1363
sge->addr_lo =
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
551
struct eth_rx_sge *sge = &fp->rx_sge_ring[index];
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
575
sge->addr_hi = cpu_to_le32(U64_HI(mapping));
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
576
sge->addr_lo = cpu_to_le32(U64_LO(mapping));
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.h
804
struct eth_rx_sge *sge = &fp->rx_sge_ring[index];
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.h
819
sge->addr_hi = 0;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.h
820
sge->addr_lo = 0;
drivers/net/ethernet/broadcom/cnic_defs.h
3097
struct fcoe_bd_ctx sge[3];
drivers/net/ethernet/chelsio/cxgb/common.h
180
struct sge_params sge;
drivers/net/ethernet/chelsio/cxgb/common.h
218
struct sge;
drivers/net/ethernet/chelsio/cxgb/common.h
235
struct sge *sge;
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
170
t1_sched_update_parms(adapter->sge, port_id, 0, sched_speed);
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
225
t1_sge_start(adapter->sge);
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
236
t1_sge_stop(adapter->sge);
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
463
t = t1_sge_get_intr_counts(adapter->sge);
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
464
t1_sge_get_port_stats(adapter->sge, dev->if_port, &ss);
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
724
e->rx_pending = adapter->params.sge.freelQ_size[!jumbo_fl];
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
725
e->rx_jumbo_pending = adapter->params.sge.freelQ_size[jumbo_fl];
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
726
e->tx_pending = adapter->params.sge.cmdQ_size[0];
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
747
adapter->params.sge.freelQ_size[!jumbo_fl] = e->rx_pending;
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
748
adapter->params.sge.freelQ_size[jumbo_fl] = e->rx_jumbo_pending;
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
749
adapter->params.sge.cmdQ_size[0] = e->tx_pending;
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
750
adapter->params.sge.cmdQ_size[1] = e->tx_pending > MAX_CMDQ1_ENTRIES ?
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
761
adapter->params.sge.rx_coalesce_usecs = c->rx_coalesce_usecs;
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
762
adapter->params.sge.coalesce_enable = c->use_adaptive_rx_coalesce;
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
763
adapter->params.sge.sample_interval_usecs = c->rate_sample_interval;
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
764
t1_sge_set_coalesce_params(adapter->sge, &adapter->params.sge);
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
774
c->rx_coalesce_usecs = adapter->params.sge.rx_coalesce_usecs;
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
775
c->rate_sample_interval = adapter->params.sge.sample_interval_usecs;
drivers/net/ethernet/chelsio/cxgb/cxgb2.c
776
c->use_adaptive_rx_coalesce = adapter->params.sge.coalesce_enable;
drivers/net/ethernet/chelsio/cxgb/sge.c
1290
static inline void reclaim_completed_tx(struct sge *sge, struct cmdQ *q)
drivers/net/ethernet/chelsio/cxgb/sge.c
1297
free_cmdQ_buffers(sge, q, reclaim);
drivers/net/ethernet/chelsio/cxgb/sge.c
1309
struct sge *sge = s->sge;
drivers/net/ethernet/chelsio/cxgb/sge.c
1310
struct adapter *adapter = sge->adapter;
drivers/net/ethernet/chelsio/cxgb/sge.c
1311
struct cmdQ *q = &sge->cmdQ[0];
drivers/net/ethernet/chelsio/cxgb/sge.c
1316
reclaim_completed_tx(sge, q);
drivers/net/ethernet/chelsio/cxgb/sge.c
1320
while ((skb = sched_skb(sge, NULL, credits)) != NULL) {
drivers/net/ethernet/chelsio/cxgb/sge.c
1355
static void sge_rx(struct sge *sge, struct freelQ *fl, unsigned int len)
drivers/net/ethernet/chelsio/cxgb/sge.c
1359
struct adapter *adapter = sge->adapter;
drivers/net/ethernet/chelsio/cxgb/sge.c
1363
skb = get_packet(adapter, fl, len - sge->rx_pkt_pad);
drivers/net/ethernet/chelsio/cxgb/sge.c
1365
sge->stats.rx_drops++;
drivers/net/ethernet/chelsio/cxgb/sge.c
1376
st = this_cpu_ptr(sge->port_stats[p->iff]);
drivers/net/ethernet/chelsio/cxgb/sge.c
1410
static void restart_tx_queues(struct sge *sge)
drivers/net/ethernet/chelsio/cxgb/sge.c
1412
struct adapter *adap = sge->adapter;
drivers/net/ethernet/chelsio/cxgb/sge.c
1415
if (!enough_free_Tx_descs(&sge->cmdQ[0]))
drivers/net/ethernet/chelsio/cxgb/sge.c
1421
if (test_and_clear_bit(nd->if_port, &sge->stopped_tx_queues) &&
drivers/net/ethernet/chelsio/cxgb/sge.c
1423
sge->stats.cmdQ_restarted[2]++;
drivers/net/ethernet/chelsio/cxgb/sge.c
1437
struct sge *sge = adapter->sge;
drivers/net/ethernet/chelsio/cxgb/sge.c
1438
struct cmdQ *cmdq = &sge->cmdQ[0];
drivers/net/ethernet/chelsio/cxgb/sge.c
1442
freelQs_empty(sge);
drivers/net/ethernet/chelsio/cxgb/sge.c
1453
if (sge->tx_sched)
drivers/net/ethernet/chelsio/cxgb/sge.c
1454
tasklet_hi_schedule(&sge->tx_sched->sched_tsk);
drivers/net/ethernet/chelsio/cxgb/sge.c
1459
if (unlikely(sge->stopped_tx_queues != 0))
drivers/net/ethernet/chelsio/cxgb/sge.c
1460
restart_tx_queues(sge);
drivers/net/ethernet/chelsio/cxgb/sge.c
1471
struct sge *sge = adapter->sge;
drivers/net/ethernet/chelsio/cxgb/sge.c
1472
struct respQ *q = &sge->respQ;
drivers/net/ethernet/chelsio/cxgb/sge.c
1494
sge->cmdQ[1].processed += cmdq_processed[1];
drivers/net/ethernet/chelsio/cxgb/sge.c
1499
struct freelQ *fl = &sge->freelQ[e->FreelistQid];
drivers/net/ethernet/chelsio/cxgb/sge.c
1505
sge_rx(sge, fl, e->BufferLength);
drivers/net/ethernet/chelsio/cxgb/sge.c
1519
refill_free_list(sge, fl);
drivers/net/ethernet/chelsio/cxgb/sge.c
1521
sge->stats.pure_rsps++;
drivers/net/ethernet/chelsio/cxgb/sge.c
1538
sge->cmdQ[1].processed += cmdq_processed[1];
drivers/net/ethernet/chelsio/cxgb/sge.c
1545
const struct respQ *Q = &adapter->sge->respQ;
drivers/net/ethernet/chelsio/cxgb/sge.c
1561
struct sge *sge = adapter->sge;
drivers/net/ethernet/chelsio/cxgb/sge.c
1562
struct respQ *q = &sge->respQ;
drivers/net/ethernet/chelsio/cxgb/sge.c
1564
const struct freelQ *fl = &sge->freelQ[e->FreelistQid];
drivers/net/ethernet/chelsio/cxgb/sge.c
1590
sge->stats.pure_rsps++;
drivers/net/ethernet/chelsio/cxgb/sge.c
1594
sge->cmdQ[1].processed += cmdq_processed[1];
drivers/net/ethernet/chelsio/cxgb/sge.c
1611
writel(adapter->sge->respQ.cidx,
drivers/net/ethernet/chelsio/cxgb/sge.c
1637
t1_sge_stop(adapter->sge);
drivers/net/ethernet/chelsio/cxgb/sge.c
1655
struct sge *sge = adapter->sge;
drivers/net/ethernet/chelsio/cxgb/sge.c
1666
writel(sge->respQ.cidx, adapter->regs + A_SG_SLEEPING);
drivers/net/ethernet/chelsio/cxgb/sge.c
1679
sge->stats.unhandled_irqs++;
drivers/net/ethernet/chelsio/cxgb/sge.c
1700
struct sge *sge = adapter->sge;
drivers/net/ethernet/chelsio/cxgb/sge.c
1701
struct cmdQ *q = &sge->cmdQ[qid];
drivers/net/ethernet/chelsio/cxgb/sge.c
1706
reclaim_completed_tx(sge, q);
drivers/net/ethernet/chelsio/cxgb/sge.c
1717
set_bit(dev->if_port, &sge->stopped_tx_queues);
drivers/net/ethernet/chelsio/cxgb/sge.c
1718
sge->stats.cmdQ_full[2]++;
drivers/net/ethernet/chelsio/cxgb/sge.c
1728
set_bit(dev->if_port, &sge->stopped_tx_queues);
drivers/net/ethernet/chelsio/cxgb/sge.c
1729
sge->stats.cmdQ_full[2]++;
drivers/net/ethernet/chelsio/cxgb/sge.c
1735
if (sge->tx_sched && !qid && skb->dev) {
drivers/net/ethernet/chelsio/cxgb/sge.c
1741
skb = sched_skb(sge, skb, credits);
drivers/net/ethernet/chelsio/cxgb/sge.c
1811
struct sge *sge = adapter->sge;
drivers/net/ethernet/chelsio/cxgb/sge.c
1812
struct sge_port_stats *st = this_cpu_ptr(sge->port_stats[dev->if_port]);
drivers/net/ethernet/chelsio/cxgb/sge.c
1877
if ((unlikely(!adapter->sge->espibug_skb[dev->if_port]))) {
drivers/net/ethernet/chelsio/cxgb/sge.c
1880
adapter->sge->espibug_skb[dev->if_port] = skb;
drivers/net/ethernet/chelsio/cxgb/sge.c
1925
struct sge *sge = timer_container_of(sge, t, tx_reclaim_timer);
drivers/net/ethernet/chelsio/cxgb/sge.c
1928
struct cmdQ *q = &sge->cmdQ[i];
drivers/net/ethernet/chelsio/cxgb/sge.c
1933
reclaim_completed_tx(sge, q);
drivers/net/ethernet/chelsio/cxgb/sge.c
1935
writel(F_CMDQ0_ENABLE, sge->adapter->regs + A_SG_DOORBELL);
drivers/net/ethernet/chelsio/cxgb/sge.c
1939
mod_timer(&sge->tx_reclaim_timer, jiffies + TX_RECLAIM_PERIOD);
drivers/net/ethernet/chelsio/cxgb/sge.c
1945
int t1_sge_set_coalesce_params(struct sge *sge, struct sge_params *p)
drivers/net/ethernet/chelsio/cxgb/sge.c
1947
sge->fixed_intrtimer = p->rx_coalesce_usecs *
drivers/net/ethernet/chelsio/cxgb/sge.c
1948
core_ticks_per_usec(sge->adapter);
drivers/net/ethernet/chelsio/cxgb/sge.c
1949
writel(sge->fixed_intrtimer, sge->adapter->regs + A_SG_INTRTIMER);
drivers/net/ethernet/chelsio/cxgb/sge.c
1957
int t1_sge_configure(struct sge *sge, struct sge_params *p)
drivers/net/ethernet/chelsio/cxgb/sge.c
1959
if (alloc_rx_resources(sge, p))
drivers/net/ethernet/chelsio/cxgb/sge.c
1961
if (alloc_tx_resources(sge, p)) {
drivers/net/ethernet/chelsio/cxgb/sge.c
1962
free_rx_resources(sge);
drivers/net/ethernet/chelsio/cxgb/sge.c
1965
configure_sge(sge, p);
drivers/net/ethernet/chelsio/cxgb/sge.c
1973
p->large_buf_capacity = jumbo_payload_capacity(sge);
drivers/net/ethernet/chelsio/cxgb/sge.c
1980
void t1_sge_stop(struct sge *sge)
drivers/net/ethernet/chelsio/cxgb/sge.c
1983
writel(0, sge->adapter->regs + A_SG_CONTROL);
drivers/net/ethernet/chelsio/cxgb/sge.c
1984
readl(sge->adapter->regs + A_SG_CONTROL); /* flush */
drivers/net/ethernet/chelsio/cxgb/sge.c
1986
if (is_T2(sge->adapter))
drivers/net/ethernet/chelsio/cxgb/sge.c
1987
timer_delete_sync(&sge->espibug_timer);
drivers/net/ethernet/chelsio/cxgb/sge.c
1989
timer_delete_sync(&sge->tx_reclaim_timer);
drivers/net/ethernet/chelsio/cxgb/sge.c
1990
if (sge->tx_sched)
drivers/net/ethernet/chelsio/cxgb/sge.c
1991
tx_sched_stop(sge);
drivers/net/ethernet/chelsio/cxgb/sge.c
1994
kfree_skb(sge->espibug_skb[i]);
drivers/net/ethernet/chelsio/cxgb/sge.c
2000
void t1_sge_start(struct sge *sge)
drivers/net/ethernet/chelsio/cxgb/sge.c
2002
refill_free_list(sge, &sge->freelQ[0]);
drivers/net/ethernet/chelsio/cxgb/sge.c
2003
refill_free_list(sge, &sge->freelQ[1]);
drivers/net/ethernet/chelsio/cxgb/sge.c
2005
writel(sge->sge_control, sge->adapter->regs + A_SG_CONTROL);
drivers/net/ethernet/chelsio/cxgb/sge.c
2006
doorbell_pio(sge->adapter, F_FL0_ENABLE | F_FL1_ENABLE);
drivers/net/ethernet/chelsio/cxgb/sge.c
2007
readl(sge->adapter->regs + A_SG_CONTROL); /* flush */
drivers/net/ethernet/chelsio/cxgb/sge.c
2009
mod_timer(&sge->tx_reclaim_timer, jiffies + TX_RECLAIM_PERIOD);
drivers/net/ethernet/chelsio/cxgb/sge.c
2011
if (is_T2(sge->adapter))
drivers/net/ethernet/chelsio/cxgb/sge.c
2012
mod_timer(&sge->espibug_timer, jiffies + sge->espibug_timeout);
drivers/net/ethernet/chelsio/cxgb/sge.c
2020
struct sge *sge = timer_container_of(sge, t, espibug_timer);
drivers/net/ethernet/chelsio/cxgb/sge.c
2021
struct adapter *adapter = sge->adapter;
drivers/net/ethernet/chelsio/cxgb/sge.c
2032
struct sk_buff *skb = sge->espibug_skb[i];
drivers/net/ethernet/chelsio/cxgb/sge.c
2058
mod_timer(&sge->espibug_timer, jiffies + sge->espibug_timeout);
drivers/net/ethernet/chelsio/cxgb/sge.c
2063
struct sge *sge = timer_container_of(sge, t, espibug_timer);
drivers/net/ethernet/chelsio/cxgb/sge.c
2064
struct adapter *adapter = sge->adapter;
drivers/net/ethernet/chelsio/cxgb/sge.c
2067
struct sk_buff *skb = sge->espibug_skb[0];
drivers/net/ethernet/chelsio/cxgb/sge.c
2090
mod_timer(&sge->espibug_timer, jiffies + sge->espibug_timeout);
drivers/net/ethernet/chelsio/cxgb/sge.c
2096
struct sge *t1_sge_create(struct adapter *adapter, struct sge_params *p)
drivers/net/ethernet/chelsio/cxgb/sge.c
2098
struct sge *sge = kzalloc_obj(*sge);
drivers/net/ethernet/chelsio/cxgb/sge.c
2101
if (!sge)
drivers/net/ethernet/chelsio/cxgb/sge.c
2104
sge->adapter = adapter;
drivers/net/ethernet/chelsio/cxgb/sge.c
2105
sge->netdev = adapter->port[0].dev;
drivers/net/ethernet/chelsio/cxgb/sge.c
2106
sge->rx_pkt_pad = t1_is_T1B(adapter) ? 0 : 2;
drivers/net/ethernet/chelsio/cxgb/sge.c
2107
sge->jumbo_fl = t1_is_T1B(adapter) ? 1 : 0;
drivers/net/ethernet/chelsio/cxgb/sge.c
2110
sge->port_stats[i] = alloc_percpu(struct sge_port_stats);
drivers/net/ethernet/chelsio/cxgb/sge.c
2111
if (!sge->port_stats[i])
drivers/net/ethernet/chelsio/cxgb/sge.c
2115
timer_setup(&sge->tx_reclaim_timer, sge_tx_reclaim_cb, 0);
drivers/net/ethernet/chelsio/cxgb/sge.c
2117
if (is_T2(sge->adapter)) {
drivers/net/ethernet/chelsio/cxgb/sge.c
2118
timer_setup(&sge->espibug_timer,
drivers/net/ethernet/chelsio/cxgb/sge.c
2123
tx_sched_init(sge);
drivers/net/ethernet/chelsio/cxgb/sge.c
2125
sge->espibug_timeout = 1;
drivers/net/ethernet/chelsio/cxgb/sge.c
2128
sge->espibug_timeout = HZ/100;
drivers/net/ethernet/chelsio/cxgb/sge.c
2134
p->freelQ_size[!sge->jumbo_fl] = SGE_FREEL_SIZE;
drivers/net/ethernet/chelsio/cxgb/sge.c
2135
p->freelQ_size[sge->jumbo_fl] = SGE_JUMBO_FREEL_SIZE;
drivers/net/ethernet/chelsio/cxgb/sge.c
2136
if (sge->tx_sched) {
drivers/net/ethernet/chelsio/cxgb/sge.c
2137
if (board_info(sge->adapter)->board == CHBT_BOARD_CHT204)
drivers/net/ethernet/chelsio/cxgb/sge.c
2147
return sge;
drivers/net/ethernet/chelsio/cxgb/sge.c
2150
free_percpu(sge->port_stats[i]);
drivers/net/ethernet/chelsio/cxgb/sge.c
2153
kfree(sge);
drivers/net/ethernet/chelsio/cxgb/sge.c
233
struct sge *sge;
drivers/net/ethernet/chelsio/cxgb/sge.c
275
static void tx_sched_stop(struct sge *sge)
drivers/net/ethernet/chelsio/cxgb/sge.c
277
struct sched *s = sge->tx_sched;
drivers/net/ethernet/chelsio/cxgb/sge.c
290
unsigned int t1_sched_update_parms(struct sge *sge, unsigned int port,
drivers/net/ethernet/chelsio/cxgb/sge.c
293
struct sched *s = sge->tx_sched;
drivers/net/ethernet/chelsio/cxgb/sge.c
313
if (board_info(sge->adapter)->board == CHBT_BOARD_CHT204) {
drivers/net/ethernet/chelsio/cxgb/sge.c
336
void t1_sched_set_max_avail_bytes(struct sge *sge, unsigned int val)
drivers/net/ethernet/chelsio/cxgb/sge.c
338
struct sched *s = sge->tx_sched;
drivers/net/ethernet/chelsio/cxgb/sge.c
343
t1_sched_update_parms(sge, i, 0, 0);
drivers/net/ethernet/chelsio/cxgb/sge.c
350
void t1_sched_set_drain_bits_per_us(struct sge *sge, unsigned int port,
drivers/net/ethernet/chelsio/cxgb/sge.c
353
struct sched *s = sge->tx_sched;
drivers/net/ethernet/chelsio/cxgb/sge.c
356
t1_sched_update_parms(sge, port, 0, 0);
drivers/net/ethernet/chelsio/cxgb/sge.c
364
static int tx_sched_init(struct sge *sge)
drivers/net/ethernet/chelsio/cxgb/sge.c
375
s->sge = sge;
drivers/net/ethernet/chelsio/cxgb/sge.c
376
sge->tx_sched = s;
drivers/net/ethernet/chelsio/cxgb/sge.c
380
t1_sched_update_parms(sge, i, 1500, 1000);
drivers/net/ethernet/chelsio/cxgb/sge.c
391
static inline int sched_update_avail(struct sge *sge)
drivers/net/ethernet/chelsio/cxgb/sge.c
393
struct sched *s = sge->tx_sched;
drivers/net/ethernet/chelsio/cxgb/sge.c
425
static struct sk_buff *sched_skb(struct sge *sge, struct sk_buff *skb,
drivers/net/ethernet/chelsio/cxgb/sge.c
428
struct sched *s = sge->tx_sched;
drivers/net/ethernet/chelsio/cxgb/sge.c
466
if (update-- && sched_update_avail(sge))
drivers/net/ethernet/chelsio/cxgb/sge.c
474
struct cmdQ *q = &sge->cmdQ[0];
drivers/net/ethernet/chelsio/cxgb/sge.c
478
writel(F_CMDQ0_ENABLE, sge->adapter->regs + A_SG_DOORBELL);
drivers/net/ethernet/chelsio/cxgb/sge.c
518
static void free_rx_resources(struct sge *sge)
drivers/net/ethernet/chelsio/cxgb/sge.c
520
struct pci_dev *pdev = sge->adapter->pdev;
drivers/net/ethernet/chelsio/cxgb/sge.c
523
if (sge->respQ.entries) {
drivers/net/ethernet/chelsio/cxgb/sge.c
524
size = sizeof(struct respQ_e) * sge->respQ.size;
drivers/net/ethernet/chelsio/cxgb/sge.c
525
dma_free_coherent(&pdev->dev, size, sge->respQ.entries,
drivers/net/ethernet/chelsio/cxgb/sge.c
526
sge->respQ.dma_addr);
drivers/net/ethernet/chelsio/cxgb/sge.c
530
struct freelQ *q = &sge->freelQ[i];
drivers/net/ethernet/chelsio/cxgb/sge.c
548
static int alloc_rx_resources(struct sge *sge, struct sge_params *p)
drivers/net/ethernet/chelsio/cxgb/sge.c
550
struct pci_dev *pdev = sge->adapter->pdev;
drivers/net/ethernet/chelsio/cxgb/sge.c
554
struct freelQ *q = &sge->freelQ[i];
drivers/net/ethernet/chelsio/cxgb/sge.c
558
q->dma_offset = sge->rx_pkt_pad ? 0 : NET_IP_ALIGN;
drivers/net/ethernet/chelsio/cxgb/sge.c
578
sge->freelQ[!sge->jumbo_fl].rx_buffer_size = SGE_RX_SM_BUF_SIZE +
drivers/net/ethernet/chelsio/cxgb/sge.c
580
sge->freelQ[!sge->jumbo_fl].dma_offset;
drivers/net/ethernet/chelsio/cxgb/sge.c
584
sge->freelQ[sge->jumbo_fl].rx_buffer_size = size;
drivers/net/ethernet/chelsio/cxgb/sge.c
590
sge->freelQ[!sge->jumbo_fl].recycleq_idx = 0;
drivers/net/ethernet/chelsio/cxgb/sge.c
591
sge->freelQ[sge->jumbo_fl].recycleq_idx = 1;
drivers/net/ethernet/chelsio/cxgb/sge.c
593
sge->respQ.genbit = 1;
drivers/net/ethernet/chelsio/cxgb/sge.c
594
sge->respQ.size = SGE_RESPQ_E_N;
drivers/net/ethernet/chelsio/cxgb/sge.c
595
sge->respQ.credits = 0;
drivers/net/ethernet/chelsio/cxgb/sge.c
596
size = sizeof(struct respQ_e) * sge->respQ.size;
drivers/net/ethernet/chelsio/cxgb/sge.c
597
sge->respQ.entries =
drivers/net/ethernet/chelsio/cxgb/sge.c
598
dma_alloc_coherent(&pdev->dev, size, &sge->respQ.dma_addr,
drivers/net/ethernet/chelsio/cxgb/sge.c
600
if (!sge->respQ.entries)
drivers/net/ethernet/chelsio/cxgb/sge.c
605
free_rx_resources(sge);
drivers/net/ethernet/chelsio/cxgb/sge.c
612
static void free_cmdQ_buffers(struct sge *sge, struct cmdQ *q, unsigned int n)
drivers/net/ethernet/chelsio/cxgb/sge.c
615
struct pci_dev *pdev = sge->adapter->pdev;
drivers/net/ethernet/chelsio/cxgb/sge.c
647
static void free_tx_resources(struct sge *sge)
drivers/net/ethernet/chelsio/cxgb/sge.c
649
struct pci_dev *pdev = sge->adapter->pdev;
drivers/net/ethernet/chelsio/cxgb/sge.c
653
struct cmdQ *q = &sge->cmdQ[i];
drivers/net/ethernet/chelsio/cxgb/sge.c
657
free_cmdQ_buffers(sge, q, q->in_use);
drivers/net/ethernet/chelsio/cxgb/sge.c
671
static int alloc_tx_resources(struct sge *sge, struct sge_params *p)
drivers/net/ethernet/chelsio/cxgb/sge.c
673
struct pci_dev *pdev = sge->adapter->pdev;
drivers/net/ethernet/chelsio/cxgb/sge.c
677
struct cmdQ *q = &sge->cmdQ[i];
drivers/net/ethernet/chelsio/cxgb/sge.c
706
sge->cmdQ[0].stop_thres = sge->adapter->params.nports *
drivers/net/ethernet/chelsio/cxgb/sge.c
711
free_tx_resources(sge);
drivers/net/ethernet/chelsio/cxgb/sge.c
729
struct sge *sge = adapter->sge;
drivers/net/ethernet/chelsio/cxgb/sge.c
732
sge->sge_control |= F_VLAN_XTRACT;
drivers/net/ethernet/chelsio/cxgb/sge.c
734
sge->sge_control &= ~F_VLAN_XTRACT;
drivers/net/ethernet/chelsio/cxgb/sge.c
736
writel(sge->sge_control, adapter->regs + A_SG_CONTROL);
drivers/net/ethernet/chelsio/cxgb/sge.c
745
static void configure_sge(struct sge *sge, struct sge_params *p)
drivers/net/ethernet/chelsio/cxgb/sge.c
747
struct adapter *ap = sge->adapter;
drivers/net/ethernet/chelsio/cxgb/sge.c
750
setup_ring_params(ap, sge->cmdQ[0].dma_addr, sge->cmdQ[0].size,
drivers/net/ethernet/chelsio/cxgb/sge.c
752
setup_ring_params(ap, sge->cmdQ[1].dma_addr, sge->cmdQ[1].size,
drivers/net/ethernet/chelsio/cxgb/sge.c
754
setup_ring_params(ap, sge->freelQ[0].dma_addr,
drivers/net/ethernet/chelsio/cxgb/sge.c
755
sge->freelQ[0].size, A_SG_FL0BASELWR,
drivers/net/ethernet/chelsio/cxgb/sge.c
757
setup_ring_params(ap, sge->freelQ[1].dma_addr,
drivers/net/ethernet/chelsio/cxgb/sge.c
758
sge->freelQ[1].size, A_SG_FL1BASELWR,
drivers/net/ethernet/chelsio/cxgb/sge.c
764
setup_ring_params(ap, sge->respQ.dma_addr, sge->respQ.size,
drivers/net/ethernet/chelsio/cxgb/sge.c
766
writel((u32)sge->respQ.size - 1, ap->regs + A_SG_RSPQUEUECREDIT);
drivers/net/ethernet/chelsio/cxgb/sge.c
768
sge->sge_control = F_CMDQ0_ENABLE | F_CMDQ1_ENABLE | F_FL0_ENABLE |
drivers/net/ethernet/chelsio/cxgb/sge.c
771
V_RX_PKT_OFFSET(sge->rx_pkt_pad);
drivers/net/ethernet/chelsio/cxgb/sge.c
774
sge->sge_control |= F_ENABLE_BIG_ENDIAN;
drivers/net/ethernet/chelsio/cxgb/sge.c
778
sge->intrtimer_nres = SGE_INTRTIMER_NRES * core_ticks_per_usec(ap);
drivers/net/ethernet/chelsio/cxgb/sge.c
780
t1_sge_set_coalesce_params(sge, p);
drivers/net/ethernet/chelsio/cxgb/sge.c
786
static inline unsigned int jumbo_payload_capacity(const struct sge *sge)
drivers/net/ethernet/chelsio/cxgb/sge.c
788
return sge->freelQ[sge->jumbo_fl].rx_buffer_size -
drivers/net/ethernet/chelsio/cxgb/sge.c
789
sge->freelQ[sge->jumbo_fl].dma_offset -
drivers/net/ethernet/chelsio/cxgb/sge.c
796
void t1_sge_destroy(struct sge *sge)
drivers/net/ethernet/chelsio/cxgb/sge.c
800
for_each_port(sge->adapter, i)
drivers/net/ethernet/chelsio/cxgb/sge.c
801
free_percpu(sge->port_stats[i]);
drivers/net/ethernet/chelsio/cxgb/sge.c
803
kfree(sge->tx_sched);
drivers/net/ethernet/chelsio/cxgb/sge.c
804
free_tx_resources(sge);
drivers/net/ethernet/chelsio/cxgb/sge.c
805
free_rx_resources(sge);
drivers/net/ethernet/chelsio/cxgb/sge.c
806
kfree(sge);
drivers/net/ethernet/chelsio/cxgb/sge.c
821
static void refill_free_list(struct sge *sge, struct freelQ *q)
drivers/net/ethernet/chelsio/cxgb/sge.c
823
struct pci_dev *pdev = sge->adapter->pdev;
drivers/net/ethernet/chelsio/cxgb/sge.c
839
skb_reserve(skb, sge->rx_pkt_pad);
drivers/net/ethernet/chelsio/cxgb/sge.c
867
static void freelQs_empty(struct sge *sge)
drivers/net/ethernet/chelsio/cxgb/sge.c
869
struct adapter *adapter = sge->adapter;
drivers/net/ethernet/chelsio/cxgb/sge.c
873
refill_free_list(sge, &sge->freelQ[0]);
drivers/net/ethernet/chelsio/cxgb/sge.c
874
refill_free_list(sge, &sge->freelQ[1]);
drivers/net/ethernet/chelsio/cxgb/sge.c
876
if (sge->freelQ[0].credits > (sge->freelQ[0].size >> 2) &&
drivers/net/ethernet/chelsio/cxgb/sge.c
877
sge->freelQ[1].credits > (sge->freelQ[1].size >> 2)) {
drivers/net/ethernet/chelsio/cxgb/sge.c
879
irqholdoff_reg = sge->fixed_intrtimer;
drivers/net/ethernet/chelsio/cxgb/sge.c
883
irqholdoff_reg = sge->intrtimer_nres;
drivers/net/ethernet/chelsio/cxgb/sge.c
900
void t1_sge_intr_disable(struct sge *sge)
drivers/net/ethernet/chelsio/cxgb/sge.c
902
u32 val = readl(sge->adapter->regs + A_PL_ENABLE);
drivers/net/ethernet/chelsio/cxgb/sge.c
904
writel(val & ~SGE_PL_INTR_MASK, sge->adapter->regs + A_PL_ENABLE);
drivers/net/ethernet/chelsio/cxgb/sge.c
905
writel(0, sge->adapter->regs + A_SG_INT_ENABLE);
drivers/net/ethernet/chelsio/cxgb/sge.c
911
void t1_sge_intr_enable(struct sge *sge)
drivers/net/ethernet/chelsio/cxgb/sge.c
914
u32 val = readl(sge->adapter->regs + A_PL_ENABLE);
drivers/net/ethernet/chelsio/cxgb/sge.c
916
if (sge->adapter->port[0].dev->hw_features & NETIF_F_TSO)
drivers/net/ethernet/chelsio/cxgb/sge.c
918
writel(en, sge->adapter->regs + A_SG_INT_ENABLE);
drivers/net/ethernet/chelsio/cxgb/sge.c
919
writel(val | SGE_PL_INTR_MASK, sge->adapter->regs + A_PL_ENABLE);
drivers/net/ethernet/chelsio/cxgb/sge.c
925
void t1_sge_intr_clear(struct sge *sge)
drivers/net/ethernet/chelsio/cxgb/sge.c
927
writel(SGE_PL_INTR_MASK, sge->adapter->regs + A_PL_CAUSE);
drivers/net/ethernet/chelsio/cxgb/sge.c
928
writel(0xffffffff, sge->adapter->regs + A_SG_INT_CAUSE);
drivers/net/ethernet/chelsio/cxgb/sge.c
934
bool t1_sge_intr_error_handler(struct sge *sge)
drivers/net/ethernet/chelsio/cxgb/sge.c
936
struct adapter *adapter = sge->adapter;
drivers/net/ethernet/chelsio/cxgb/sge.c
943
sge->stats.respQ_empty++;
drivers/net/ethernet/chelsio/cxgb/sge.c
945
sge->stats.respQ_overflow++;
drivers/net/ethernet/chelsio/cxgb/sge.c
950
sge->stats.freelistQ_empty++;
drivers/net/ethernet/chelsio/cxgb/sge.c
951
freelQs_empty(sge);
drivers/net/ethernet/chelsio/cxgb/sge.c
954
sge->stats.pkt_too_big++;
drivers/net/ethernet/chelsio/cxgb/sge.c
959
sge->stats.pkt_mismatch++;
drivers/net/ethernet/chelsio/cxgb/sge.c
972
const struct sge_intr_counts *t1_sge_get_intr_counts(const struct sge *sge)
drivers/net/ethernet/chelsio/cxgb/sge.c
974
return &sge->stats;
drivers/net/ethernet/chelsio/cxgb/sge.c
977
void t1_sge_get_port_stats(const struct sge *sge, int port,
drivers/net/ethernet/chelsio/cxgb/sge.c
984
struct sge_port_stats *st = per_cpu_ptr(sge->port_stats[port], cpu);
drivers/net/ethernet/chelsio/cxgb/sge.h
62
struct sge;
drivers/net/ethernet/chelsio/cxgb/sge.h
64
struct sge *t1_sge_create(struct adapter *, struct sge_params *);
drivers/net/ethernet/chelsio/cxgb/sge.h
65
int t1_sge_configure(struct sge *, struct sge_params *);
drivers/net/ethernet/chelsio/cxgb/sge.h
66
int t1_sge_set_coalesce_params(struct sge *, struct sge_params *);
drivers/net/ethernet/chelsio/cxgb/sge.h
67
void t1_sge_destroy(struct sge *);
drivers/net/ethernet/chelsio/cxgb/sge.h
74
void t1_sge_start(struct sge *);
drivers/net/ethernet/chelsio/cxgb/sge.h
75
void t1_sge_stop(struct sge *);
drivers/net/ethernet/chelsio/cxgb/sge.h
76
bool t1_sge_intr_error_handler(struct sge *sge);
drivers/net/ethernet/chelsio/cxgb/sge.h
77
void t1_sge_intr_enable(struct sge *);
drivers/net/ethernet/chelsio/cxgb/sge.h
78
void t1_sge_intr_disable(struct sge *);
drivers/net/ethernet/chelsio/cxgb/sge.h
79
void t1_sge_intr_clear(struct sge *);
drivers/net/ethernet/chelsio/cxgb/sge.h
80
const struct sge_intr_counts *t1_sge_get_intr_counts(const struct sge *sge);
drivers/net/ethernet/chelsio/cxgb/sge.h
81
void t1_sge_get_port_stats(const struct sge *sge, int port, struct sge_port_stats *);
drivers/net/ethernet/chelsio/cxgb/sge.h
82
unsigned int t1_sched_update_parms(struct sge *, unsigned int, unsigned int,
drivers/net/ethernet/chelsio/cxgb/subr.c
1039
if (adapter->sge)
drivers/net/ethernet/chelsio/cxgb/subr.c
1040
t1_sge_destroy(adapter->sge);
drivers/net/ethernet/chelsio/cxgb/subr.c
1076
adapter->sge = t1_sge_create(adapter, &adapter->params.sge);
drivers/net/ethernet/chelsio/cxgb/subr.c
1077
if (!adapter->sge) {
drivers/net/ethernet/chelsio/cxgb/subr.c
215
if (t1_sge_intr_error_handler(adapter->sge))
drivers/net/ethernet/chelsio/cxgb/subr.c
764
t1_sge_intr_enable(adapter->sge);
drivers/net/ethernet/chelsio/cxgb/subr.c
796
t1_sge_intr_disable(adapter->sge);
drivers/net/ethernet/chelsio/cxgb/subr.c
822
t1_sge_intr_clear(adapter->sge);
drivers/net/ethernet/chelsio/cxgb/subr.c
857
if (t1_sge_intr_error_handler(adapter->sge))
drivers/net/ethernet/chelsio/cxgb/subr.c
999
err = t1_sge_configure(adapter->sge, &adapter->params.sge);
drivers/net/ethernet/chelsio/cxgb3/adapter.h
244
struct sge sge;
drivers/net/ethernet/chelsio/cxgb3/common.h
383
struct sge_params sge;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
1170
struct sge_rspq *q = &adap->sge.qs[i].rspq;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
1276
t3_intr_handler(adap, adap->sge.qs[0].rspq.polling),
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
1657
tot += adapter->sge.qs[i].port_stats[idx];
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
1956
const struct qset_params *q = &adapter->params.sge.qset[pi->first_qset];
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
1990
q = &adapter->params.sge.qset[pi->first_qset];
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2016
qsp = &adapter->params.sge.qset[i];
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2017
qs = &adapter->sge.qs[i];
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2031
struct qset_params *q = adapter->params.sge.qset;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2201
q = &adapter->params.sge.qset[t.qset_idx];
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2219
&adapter->sge.qs[t.qset_idx];
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2234
q = &adapter->params.sge.
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2277
q = &adapter->params.sge.qset[q1 + t.qset_idx];
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2620
struct sge_qset *qs = &adapter->sge.qs[qidx];
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
2768
struct sge_qset *qs = &adapter->sge.qs[0];
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
409
adap->sge.qs[qidx].
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
412
&adap->sge.qs[qidx]);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
416
&adap->sge.qs[qidx]);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
436
&adapter->sge.qs[i]);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
446
while (adap->sge.qs[0].rspq.offload_pkts < init_cnt + n) {
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
459
unsigned long cnt = adap->sge.qs[0].rspq.offload_pkts;
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
595
struct sge_qset *qs = &adap->sge.qs[i];
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
608
struct sge_qset *qs = &adap->sge.qs[i];
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
632
if (adap->sge.qs[i].adap)
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
633
napi_disable(&adap->sge.qs[i].napi);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
640
if (adap->sge.qs[i].adap)
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
641
napi_enable(&adap->sge.qs[i].napi);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
664
pi->qs = &adap->sge.qs[pi->first_qset];
drivers/net/ethernet/chelsio/cxgb3/cxgb3_main.c
669
&adap->params.sge.qset[qset_idx], ntxq, dev,
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
303
spin_lock_irqsave(&adapter->sge.reg_lock, flags);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
306
spin_unlock_irqrestore(&adapter->sge.reg_lock, flags);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
334
spin_lock_irq(&adapter->sge.reg_lock);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
341
spin_unlock_irq(&adapter->sge.reg_lock);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
345
spin_lock_irq(&adapter->sge.reg_lock);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
347
spin_unlock_irq(&adapter->sge.reg_lock);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
352
spin_lock_irq(&adapter->sge.reg_lock);
drivers/net/ethernet/chelsio/cxgb3/cxgb3_offload.c
358
spin_unlock_irq(&adapter->sge.reg_lock);
drivers/net/ethernet/chelsio/cxgb3/sge.c
1565
ret = ctrl_xmit(adap, &adap->sge.qs[0].txq[TXQ_CTRL], skb);
drivers/net/ethernet/chelsio/cxgb3/sge.c
1836
struct sge_qset *qs = &adap->sge.qs[queue_set(skb)];
drivers/net/ethernet/chelsio/cxgb3/sge.c
2100
skb_record_rx_queue(skb, qs - &adap->sge.qs[pi->first_qset]);
drivers/net/ethernet/chelsio/cxgb3/sge.c
2198
skb_record_rx_queue(skb, qs - &adap->sge.qs[pi->first_qset]);
drivers/net/ethernet/chelsio/cxgb3/sge.c
2638
struct sge_rspq *q = &adap->sge.qs[0].rspq;
drivers/net/ethernet/chelsio/cxgb3/sge.c
2642
if (process_responses(adap, &adap->sge.qs[0], -1)) {
drivers/net/ethernet/chelsio/cxgb3/sge.c
2649
process_responses(adap, &adap->sge.qs[1], -1)) {
drivers/net/ethernet/chelsio/cxgb3/sge.c
2650
struct sge_rspq *q1 = &adap->sge.qs[1].rspq;
drivers/net/ethernet/chelsio/cxgb3/sge.c
2683
struct sge_rspq *q = &adap->sge.qs[0].rspq;
drivers/net/ethernet/chelsio/cxgb3/sge.c
2687
new_packets = rspq_check_napi(&adap->sge.qs[0]);
drivers/net/ethernet/chelsio/cxgb3/sge.c
2689
new_packets += rspq_check_napi(&adap->sge.qs[1]);
drivers/net/ethernet/chelsio/cxgb3/sge.c
2721
struct sge_rspq *q0 = &adap->sge.qs[0].rspq;
drivers/net/ethernet/chelsio/cxgb3/sge.c
2722
struct sge_rspq *q1 = &adap->sge.qs[1].rspq;
drivers/net/ethernet/chelsio/cxgb3/sge.c
2759
struct sge_rspq *q0 = &adap->sge.qs[0].rspq;
drivers/net/ethernet/chelsio/cxgb3/sge.c
2776
process_responses_gts(adap, &adap->sge.qs[1].rspq);
drivers/net/ethernet/chelsio/cxgb3/sge.c
2793
struct sge_qset *qs0 = &adap->sge.qs[0];
drivers/net/ethernet/chelsio/cxgb3/sge.c
2811
napi_schedule(&adap->sge.qs[1].napi);
drivers/net/ethernet/chelsio/cxgb3/sge.c
2956
&qs->rspq.lock : &adap->sge.qs[0].rspq.lock;
drivers/net/ethernet/chelsio/cxgb3/sge.c
3027
struct sge_qset *q = &adapter->sge.qs[id];
drivers/net/ethernet/chelsio/cxgb3/sge.c
3108
spin_lock_irq(&adapter->sge.reg_lock);
drivers/net/ethernet/chelsio/cxgb3/sge.c
3152
spin_unlock_irq(&adapter->sge.reg_lock);
drivers/net/ethernet/chelsio/cxgb3/sge.c
3183
spin_unlock_irq(&adapter->sge.reg_lock);
drivers/net/ethernet/chelsio/cxgb3/sge.c
3200
struct sge_qset *q = &adap->sge.qs[i];
drivers/net/ethernet/chelsio/cxgb3/sge.c
3223
struct sge_qset *q = &adap->sge.qs[i];
drivers/net/ethernet/chelsio/cxgb3/sge.c
3243
t3_free_qset(adap, &adap->sge.qs[i]);
drivers/net/ethernet/chelsio/cxgb3/sge.c
3292
struct sge_qset *qs = &adap->sge.qs[i];
drivers/net/ethernet/chelsio/cxgb3/sge.c
3369
spin_lock_init(&adap->sge.reg_lock);
drivers/net/ethernet/chelsio/cxgb3/sge.c
679
spin_lock_irq(&adapter->sge.reg_lock);
drivers/net/ethernet/chelsio/cxgb3/sge.c
681
spin_unlock_irq(&adapter->sge.reg_lock);
drivers/net/ethernet/chelsio/cxgb3/sge.c
692
spin_lock_irq(&adapter->sge.reg_lock);
drivers/net/ethernet/chelsio/cxgb3/sge.c
694
spin_unlock_irq(&adapter->sge.reg_lock);
drivers/net/ethernet/chelsio/cxgb3/sge.c
708
spin_lock_irq(&adapter->sge.reg_lock);
drivers/net/ethernet/chelsio/cxgb3/sge.c
710
spin_unlock_irq(&adapter->sge.reg_lock);
drivers/net/ethernet/chelsio/cxgb3/t3_hw.c
3339
min(adapter->params.sge.max_pkt_size,
drivers/net/ethernet/chelsio/cxgb3/t3_hw.c
3342
min(adapter->params.sge.max_pkt_size, 16384U));
drivers/net/ethernet/chelsio/cxgb3/t3_hw.c
3359
t3_sge_init(adapter, &adapter->params.sge);
drivers/net/ethernet/chelsio/cxgb3/t3_hw.c
3625
t3_sge_prep(adapter, &adapter->params.sge);
drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c
3372
struct sge *s = &padap->sge;
drivers/net/ethernet/chelsio/cxgb4/cxgb4.h
1017
#define for_each_ethrxq(sge, i) for (i = 0; i < (sge)->ethqsets; i++)
drivers/net/ethernet/chelsio/cxgb4/cxgb4.h
1018
#define for_each_ofldtxq(sge, i) for (i = 0; i < (sge)->ofldqsets; i++)
drivers/net/ethernet/chelsio/cxgb4/cxgb4.h
1131
struct sge sge;
drivers/net/ethernet/chelsio/cxgb4/cxgb4.h
1602
return idx < SGE_NTIMERS ? adap->sge.timer_val[idx] : 0;
drivers/net/ethernet/chelsio/cxgb4/cxgb4.h
437
struct sge_params sge;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
2620
const struct sge_uld_txq_info *utxq_info = adap->sge.uld_txq_info[uld];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
2631
const struct sge_uld_rxq_info *urxq_info = adap->sge.uld_rxq_info[uld];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
2661
struct sge *s = &adap->sge;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
2663
eth_entries = DIV_ROUND_UP(adap->sge.ethqsets, 4);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
2806
eohw_entries = DIV_ROUND_UP(adap->sge.eoqsets, 4);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3163
if (adap->sge.eohw_txq)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3164
eohw_entries = DIV_ROUND_UP(adap->sge.eoqsets, 4);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3202
return DIV_ROUND_UP(adap->sge.ethqsets, 4) + mirror_rxq_entries +
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3410
ssize_t size = (adap->sge.egr_sz + 3) / 4 +
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3411
adap->sge.egr_sz / 32 + 2; /* includes ,/\n/\0 */
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3418
adap->sge.egr_sz, adap->sge.blocked_fl);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3432
t = bitmap_zalloc(adap->sge.egr_sz, GFP_KERNEL);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3436
err = bitmap_parse_user(ubuf, count, t, adap->sge.egr_sz);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3442
bitmap_copy(adap->sge.blocked_fl, t, adap->sge.egr_sz);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
1007
txq = &adap->sge.ethtxq[pi->first_qset];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
1013
return adap->sge.dbqtimer_val[txq->dbqtimerix];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
1027
struct sge *s = &adap->sge;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
1062
struct sge *s = &adap->sge;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
1178
const struct sge_rspq *rq = &adap->sge.ethrxq[pi->first_qset].rspq;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
1182
adap->sge.counter_val[rq->pktcnt_idx] : 0;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
283
const struct sge_eth_txq *tx = &adap->sge.ethtxq[p->first_qset];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
284
const struct sge_eth_rxq *rx = &adap->sge.ethrxq[p->first_qset];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
303
if (adap->sge.eohw_txq) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
304
eohw_tx = &adap->sge.eohw_txq[p->first_qset];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
898
const struct sge *s = &pi->adapter->sge;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
918
struct sge *s = &adapter->sge;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
952
struct sge_eth_rxq *q = &adap->sge.ethrxq[pi->first_qset];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
967
struct sge_eth_rxq *q = &adap->sge.ethrxq[pi->first_qset];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
979
struct sge_eth_rxq *q = &adap->sge.ethrxq[pi->first_qset];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
995
return adap->sge.dbqtimer_tick;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
1477
((adapter->sge.fw_evtq.abs_id << 14) | atid),
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
1487
((adapter->sge.fw_evtq.abs_id << 14) | atid),
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
1786
TCB_RSS_INFO_V(adapter->sge.fw_evtq.abs_id), 0, 1);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
348
iq = adapter->sge.ethrxq[pi->first_qset +
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
70
QUEUENO_V(adap->sge.fw_evtq.abs_id) |
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
777
t4_mk_filtdelwr(f->tid, fwr, adapter->sge.fw_evtq.abs_id);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
887
FW_FILTER_WR_RX_RPL_IQ_V(adapter->sge.fw_evtq.abs_id));
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1005
struct sge *s = &adap->sge;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1048
struct sge *s = &adap->sge;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1199
static int closest_timer(const struct sge *s, int time)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1215
static int closest_thres(const struct sge *s, int thres)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1252
new_idx = closest_thres(&adap->sge, cnt);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1267
us = us == 0 ? 6 : closest_timer(&adap->sge, us);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1322
struct sge *s = &adap->sge;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1340
msix = -((int)adap->sge.intrq.abs_id + 1);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1428
struct sge *s = &adap->sge;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1947
chan = rxq_to_chan(&adap->sge, queue);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1990
chan = rxq_to_chan(&adap->sge, queue);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
238
struct sge_eth_txq *txq = &adap->sge.ethtxq[pi->first_qset];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
2444
for_each_ethrxq(&adap->sge, i)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
2445
disable_txq_db(&adap->sge.ethtxq[i].q);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
2448
adap->sge.uld_txq_info[CXGB4_TX_OFLD];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
2451
for_each_ofldtxq(&adap->sge, i) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
2459
disable_txq_db(&adap->sge.ctrlq[i].q);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
2466
for_each_ethrxq(&adap->sge, i)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
2467
enable_txq_db(adap, &adap->sge.ethtxq[i].q);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
2470
adap->sge.uld_txq_info[CXGB4_TX_OFLD];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
2473
for_each_ofldtxq(&adap->sge, i) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
2481
enable_txq_db(adap, &adap->sge.ctrlq[i].q);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
2548
for_each_ethrxq(&adap->sge, i)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
2549
sync_txq_pidx(adap, &adap->sge.ethtxq[i].q);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
2552
adap->sge.uld_txq_info[CXGB4_TX_OFLD];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
2554
for_each_ofldtxq(&adap->sge, i) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
2562
sync_txq_pidx(adap, &adap->sge.ctrlq[i].q);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
2755
struct sge *s = &adap->sge;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
3473
struct sge_eth_rxq *rx = &adap->sge.ethrxq[pi->first_qset];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
4207
ret = t4_cfg_pfvf(adap, adap->mbox, adap->pf, 0, adap->sge.egr_sz, 64,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
4952
adap->sge.dbqtimer_tick = val[0];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
4954
ARRAY_SIZE(adap->sge.dbqtimer_val),
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
4955
adap->sge.dbqtimer_val);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
4976
adap->sge.egr_start = val[0];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
4981
adap->sge.ingr_start = val[5];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
5021
adap->sge.egr_sz = val[0] - adap->sge.egr_start + 1;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
5022
adap->sge.ingr_sz = val[1] - adap->sge.ingr_start + 1;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
5024
adap->sge.egr_map = kzalloc_objs(*adap->sge.egr_map, adap->sge.egr_sz);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
5025
if (!adap->sge.egr_map) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
5030
adap->sge.ingr_map = kzalloc_objs(*adap->sge.ingr_map,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
5031
adap->sge.ingr_sz);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
5032
if (!adap->sge.ingr_map) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
5040
adap->sge.starving_fl = bitmap_zalloc(adap->sge.egr_sz, GFP_KERNEL);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
5041
if (!adap->sge.starving_fl) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
5046
adap->sge.txq_maperr = bitmap_zalloc(adap->sge.egr_sz, GFP_KERNEL);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
5047
if (!adap->sge.txq_maperr) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
5053
adap->sge.blocked_fl = bitmap_zalloc(adap->sge.egr_sz, GFP_KERNEL);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
5054
if (!adap->sge.blocked_fl) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
5404
kfree(adap->sge.egr_map);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
5405
kfree(adap->sge.ingr_map);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
5406
bitmap_free(adap->sge.starving_fl);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
5407
bitmap_free(adap->sge.txq_maperr);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
5409
bitmap_free(adap->sge.blocked_fl);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
5647
struct sge *s = &adap->sge;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
581
txq = q->adap->sge.egr_map[qid - q->adap->sge.egr_start];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
5814
while (n < adap->sge.ethqsets)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
5819
adap->sge.ethqsets--;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
5820
if (adap->sge.ethqsets <= n)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
5899
struct sge *s = &adap->sge;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
6173
kfree(adapter->sge.egr_map);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
6174
kfree(adapter->sge.ingr_map);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
6175
bitmap_free(adapter->sge.starving_fl);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
6176
bitmap_free(adapter->sge.txq_maperr);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
6178
bitmap_free(adapter->sge.blocked_fl);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
719
struct sge *s = &adap->sge;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
759
struct sge *s = &adap->sge;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
866
rxq = &adapter->sge.ethrxq[pi->first_qset];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
907
static unsigned int rxq_to_chan(const struct sge *p, unsigned int qid)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
926
for (i = 0; i < adap->sge.ingr_sz; i++) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
927
struct sge_rspq *q = adap->sge.ingr_map[i];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
939
struct sge *s = &adap->sge;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
972
for (i = 0; i < adap->sge.ingr_sz; i++) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
973
struct sge_rspq *q = adap->sge.ingr_map[i];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
986
adap->sge.nd_msix_idx = -1;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
999
adap->sge.nd_msix_idx = msix;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ptp.c
173
struct sge *s = &adapter->sge;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_mqprio.c
159
adap->sge.eohw_rxq = kzalloc_objs(struct sge_ofld_rxq,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_mqprio.c
160
adap->sge.eoqsets);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_mqprio.c
161
if (!adap->sge.eohw_rxq)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_mqprio.c
164
adap->sge.eohw_txq = kzalloc_objs(struct sge_eohw_txq,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_mqprio.c
165
adap->sge.eoqsets);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_mqprio.c
166
if (!adap->sge.eohw_txq) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_mqprio.c
167
kfree(adap->sge.eohw_rxq);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_mqprio.c
177
msix = -((int)adap->sge.intrq.abs_id + 1);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_mqprio.c
180
eorxq = &adap->sge.eohw_rxq[pi->first_qset + i];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_mqprio.c
181
eotxq = &adap->sge.eohw_txq[pi->first_qset + i];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_mqprio.c
238
eorxq = &adap->sge.eohw_rxq[pi->first_qset + i];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_mqprio.c
252
eorxq = &adap->sge.eohw_rxq[pi->first_qset + i];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_mqprio.c
253
eotxq = &adap->sge.eohw_txq[pi->first_qset + i];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_mqprio.c
263
kfree(adap->sge.eohw_txq);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_mqprio.c
264
kfree(adap->sge.eohw_rxq);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_mqprio.c
282
if (!adap->sge.eohw_rxq[pi->first_qset].rspq.desc)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_mqprio.c
286
eorxq = &adap->sge.eohw_rxq[pi->first_qset + i];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_mqprio.c
287
eotxq = &adap->sge.eohw_txq[pi->first_qset + i];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_mqprio.c
309
kfree(adap->sge.eohw_txq);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_mqprio.c
310
kfree(adap->sge.eohw_rxq);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c
115
struct sge *s = &adap->sge;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c
176
struct sge_uld_rxq_info *rxq_info = adap->sge.uld_rxq_info[uld_type];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c
185
struct sge *s = &adap->sge;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c
214
struct sge_uld_rxq_info *rxq_info = adap->sge.uld_rxq_info[uld_type];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c
217
struct sge *s = &adap->sge;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c
240
struct sge *s = &adap->sge;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c
307
adap->sge.uld_rxq_info[uld_type] = rxq_info;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c
314
struct sge_uld_rxq_info *rxq_info = adap->sge.uld_rxq_info[uld_type];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c
316
adap->sge.uld_rxq_info[uld_type] = NULL;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c
325
struct sge_uld_rxq_info *rxq_info = adap->sge.uld_rxq_info[uld_type];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c
357
struct sge_uld_rxq_info *rxq_info = adap->sge.uld_rxq_info[uld_type];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c
371
struct sge_uld_rxq_info *rxq_info = adap->sge.uld_rxq_info[uld_type];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c
386
struct sge_uld_rxq_info *rxq_info = adap->sge.uld_rxq_info[uld_type];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c
424
struct sge *s = &adap->sge;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c
450
txq_info = adap->sge.uld_txq_info[tx_uld_type];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c
456
adap->sge.uld_txq_info[tx_uld_type] = NULL;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c
468
txq_info = adap->sge.uld_txq_info[tx_uld_type];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c
504
adap->sge.uld_txq_info[tx_uld_type] = txq_info;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c
511
struct sge_uld_rxq_info *rxq_info = adap->sge.uld_rxq_info[uld_type];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c
513
struct sge_uld_txq_info *txq_info = adap->sge.uld_txq_info[tx_uld_type];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c
524
struct sge *s = &adap->sge;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c
548
struct sge *s = &adap->sge;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c
613
lld->udb_density = 1 << adap->params.sge.eq_qpp;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c
614
lld->ucq_density = 1 << adap->params.sge.iq_qpp;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c
615
lld->sge_host_page_size = 1 << (adap->params.sge.hps + 10);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c
624
lld->sge_ingpadboundary = adap->sge.fl_align;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c
625
lld->sge_egrstatuspagesize = adap->sge.stat_len;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.c
626
lld->sge_pktshift = adap->sge.pktshift;
drivers/net/ethernet/chelsio/cxgb4/l2t.c
154
TID_QID_V(adap->sge.fw_evtq.abs_id)));
drivers/net/ethernet/chelsio/cxgb4/sched.c
180
txq = &adap->sge.ethtxq[pi->first_qset + p->queue];
drivers/net/ethernet/chelsio/cxgb4/sched.c
196
txq = &adap->sge.ethtxq[pi->first_qset + p->queue];
drivers/net/ethernet/chelsio/cxgb4/sched.c
232
txq = &adap->sge.ethtxq[pi->first_qset + p->queue];
drivers/net/ethernet/chelsio/cxgb4/sge.c
151
struct sge *s = &adapter->sge;
drivers/net/ethernet/chelsio/cxgb4/sge.c
1532
q = &adap->sge.ptptxq;
drivers/net/ethernet/chelsio/cxgb4/sge.c
1534
q = &adap->sge.ethtxq[qidx + pi->first_qset];
drivers/net/ethernet/chelsio/cxgb4/sge.c
1845
txq = &adapter->sge.ethtxq[pi->first_qset + qidx];
drivers/net/ethernet/chelsio/cxgb4/sge.c
2266
eohw_txq = &adap->sge.eohw_txq[eosw_txq->hwqid];
drivers/net/ethernet/chelsio/cxgb4/sge.c
238
const struct sge *s = &adapter->sge;
drivers/net/ethernet/chelsio/cxgb4/sge.c
2574
rxq = &adap->sge.eohw_rxq[eosw_txq->hwqid];
drivers/net/ethernet/chelsio/cxgb4/sge.c
2673
q = &adap->sge.ethtxq[pi->first_qset];
drivers/net/ethernet/chelsio/cxgb4/sge.c
2836
ret = ctrl_xmit(&adap->sge.ctrlq[0], skb);
drivers/net/ethernet/chelsio/cxgb4/sge.c
2898
set_bit(q->q.cntxt_id - q->adap->sge.egr_start,
drivers/net/ethernet/chelsio/cxgb4/sge.c
2899
q->adap->sge.txq_maperr);
drivers/net/ethernet/chelsio/cxgb4/sge.c
3128
return ctrl_xmit(&adap->sge.ctrlq[idx], skb);
drivers/net/ethernet/chelsio/cxgb4/sge.c
3131
txq_info = adap->sge.uld_txq_info[tx_uld_type];
drivers/net/ethernet/chelsio/cxgb4/sge.c
3258
txq_info = adap->sge.uld_txq_info[CXGB4_TX_OFLD];
drivers/net/ethernet/chelsio/cxgb4/sge.c
3434
struct sge *s = &adapter->sge;
drivers/net/ethernet/chelsio/cxgb4/sge.c
3584
struct sge *s = &adapter->sge;
drivers/net/ethernet/chelsio/cxgb4/sge.c
3633
struct sge *s = &adap->sge;
drivers/net/ethernet/chelsio/cxgb4/sge.c
3670
struct sge *s = &q->adap->sge;
drivers/net/ethernet/chelsio/cxgb4/sge.c
3883
struct sge *s = &adapter->sge;
drivers/net/ethernet/chelsio/cxgb4/sge.c
401
struct sge *s = &adapter->sge;
drivers/net/ethernet/chelsio/cxgb4/sge.c
4150
struct sge_rspq *q = &adap->sge.intrq;
drivers/net/ethernet/chelsio/cxgb4/sge.c
4153
spin_lock(&adap->sge.intrq_lock);
drivers/net/ethernet/chelsio/cxgb4/sge.c
4163
qid -= adap->sge.ingr_start;
drivers/net/ethernet/chelsio/cxgb4/sge.c
4164
napi_schedule(&adap->sge.ingr_map[qid]->napi);
drivers/net/ethernet/chelsio/cxgb4/sge.c
4183
spin_unlock(&adap->sge.intrq_lock);
drivers/net/ethernet/chelsio/cxgb4/sge.c
4237
struct adapter *adap = timer_container_of(adap, t, sge.rx_timer);
drivers/net/ethernet/chelsio/cxgb4/sge.c
4238
struct sge *s = &adap->sge;
drivers/net/ethernet/chelsio/cxgb4/sge.c
4272
struct adapter *adap = timer_container_of(adap, t, sge.tx_timer);
drivers/net/ethernet/chelsio/cxgb4/sge.c
4273
struct sge *s = &adap->sge;
drivers/net/ethernet/chelsio/cxgb4/sge.c
4367
struct sge *s = &adap->sge;
drivers/net/ethernet/chelsio/cxgb4/sge.c
4479
adap->sge.ingr_map[iq->cntxt_id - adap->sge.ingr_start] = iq;
drivers/net/ethernet/chelsio/cxgb4/sge.c
4486
adap->sge.egr_map[fl->cntxt_id - adap->sge.egr_start] = fl;
drivers/net/ethernet/chelsio/cxgb4/sge.c
4565
adap->sge.egr_map[id - adap->sge.egr_start] = q;
drivers/net/ethernet/chelsio/cxgb4/sge.c
4583
struct sge *s = &adap->sge;
drivers/net/ethernet/chelsio/cxgb4/sge.c
4676
struct sge *s = &adap->sge;
drivers/net/ethernet/chelsio/cxgb4/sge.c
4745
struct sge *s = &adap->sge;
drivers/net/ethernet/chelsio/cxgb4/sge.c
4841
struct sge *s = &adap->sge;
drivers/net/ethernet/chelsio/cxgb4/sge.c
4854
struct sge *s = &adap->sge;
drivers/net/ethernet/chelsio/cxgb4/sge.c
4857
adap->sge.ingr_map[rq->cntxt_id - adap->sge.ingr_start] = NULL;
drivers/net/ethernet/chelsio/cxgb4/sge.c
4902
for (i = 0; i < adap->sge.ethqsets; i++) {
drivers/net/ethernet/chelsio/cxgb4/sge.c
4903
eq = &adap->sge.ethrxq[i];
drivers/net/ethernet/chelsio/cxgb4/sge.c
4913
for (i = 0; i < adap->sge.ethqsets; i++) {
drivers/net/ethernet/chelsio/cxgb4/sge.c
4914
eq = &adap->sge.ethrxq[i];
drivers/net/ethernet/chelsio/cxgb4/sge.c
4923
etq = &adap->sge.ethtxq[i];
drivers/net/ethernet/chelsio/cxgb4/sge.c
4936
for (i = 0; i < ARRAY_SIZE(adap->sge.ctrlq); i++) {
drivers/net/ethernet/chelsio/cxgb4/sge.c
4937
struct sge_ctrl_txq *cq = &adap->sge.ctrlq[i];
drivers/net/ethernet/chelsio/cxgb4/sge.c
4948
if (adap->sge.fw_evtq.desc) {
drivers/net/ethernet/chelsio/cxgb4/sge.c
4949
free_rspq_fl(adap, &adap->sge.fw_evtq, NULL);
drivers/net/ethernet/chelsio/cxgb4/sge.c
4950
if (adap->sge.fwevtq_msix_idx >= 0)
drivers/net/ethernet/chelsio/cxgb4/sge.c
4952
adap->sge.fwevtq_msix_idx);
drivers/net/ethernet/chelsio/cxgb4/sge.c
4955
if (adap->sge.nd_msix_idx >= 0)
drivers/net/ethernet/chelsio/cxgb4/sge.c
4956
cxgb4_free_msix_idx_in_bmap(adap, adap->sge.nd_msix_idx);
drivers/net/ethernet/chelsio/cxgb4/sge.c
4958
if (adap->sge.intrq.desc)
drivers/net/ethernet/chelsio/cxgb4/sge.c
4959
free_rspq_fl(adap, &adap->sge.intrq, NULL);
drivers/net/ethernet/chelsio/cxgb4/sge.c
4962
etq = &adap->sge.ptptxq;
drivers/net/ethernet/chelsio/cxgb4/sge.c
4975
memset(adap->sge.egr_map, 0,
drivers/net/ethernet/chelsio/cxgb4/sge.c
4976
adap->sge.egr_sz * sizeof(*adap->sge.egr_map));
drivers/net/ethernet/chelsio/cxgb4/sge.c
4981
adap->sge.ethtxq_rover = 0;
drivers/net/ethernet/chelsio/cxgb4/sge.c
4982
mod_timer(&adap->sge.rx_timer, jiffies + RX_QCHECK_PERIOD);
drivers/net/ethernet/chelsio/cxgb4/sge.c
4983
mod_timer(&adap->sge.tx_timer, jiffies + TX_QCHECK_PERIOD);
drivers/net/ethernet/chelsio/cxgb4/sge.c
4997
struct sge *s = &adap->sge;
drivers/net/ethernet/chelsio/cxgb4/sge.c
5007
txq_info = adap->sge.uld_txq_info[CXGB4_TX_OFLD];
drivers/net/ethernet/chelsio/cxgb4/sge.c
5011
for_each_ofldtxq(&adap->sge, i) {
drivers/net/ethernet/chelsio/cxgb4/sge.c
5021
txq_info = adap->sge.uld_txq_info[CXGB4_TX_CRYPTO];
drivers/net/ethernet/chelsio/cxgb4/sge.c
5025
for_each_ofldtxq(&adap->sge, i) {
drivers/net/ethernet/chelsio/cxgb4/sge.c
5050
struct sge *s = &adap->sge;
drivers/net/ethernet/chelsio/cxgb4/sge.c
5147
struct sge *s = &adap->sge;
drivers/net/ethernet/chelsio/cxgb4/sge.c
538
struct sge *s = &adap->sge;
drivers/net/ethernet/chelsio/cxgb4/sge.c
547
if (test_bit(q->cntxt_id - adap->sge.egr_start, adap->sge.blocked_fl))
drivers/net/ethernet/chelsio/cxgb4/sge.c
625
set_bit(q->cntxt_id - adap->sge.egr_start,
drivers/net/ethernet/chelsio/cxgb4/sge.c
626
adap->sge.starving_fl);
drivers/net/ethernet/chelsio/cxgb4/sge.c
852
to = (u8 *)end > (u8 *)q->stat ? buf : sgl->sge;
drivers/net/ethernet/chelsio/cxgb4/sge.c
866
unsigned int part0 = (u8 *)q->stat - (u8 *)sgl->sge, part1;
drivers/net/ethernet/chelsio/cxgb4/sge.c
869
memcpy(sgl->sge, buf, part0);
drivers/net/ethernet/chelsio/cxgb4/sge.c
941
to = (u8 *)end > (u8 *)q->stat ? buf : sgl->sge;
drivers/net/ethernet/chelsio/cxgb4/sge.c
969
u32 part0 = (u8 *)q->stat - (u8 *)sgl->sge, part1;
drivers/net/ethernet/chelsio/cxgb4/sge.c
972
memcpy(sgl->sge, buf, part0);
drivers/net/ethernet/chelsio/cxgb4/smt.c
201
TID_QID_V(adapter->sge.fw_evtq.abs_id)));
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
9270
page_shift = adapter->params.sge.hps + 10;
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
9276
? adapter->params.sge.eq_qpp
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
9277
: adapter->params.sge.iq_qpp);
drivers/net/ethernet/chelsio/cxgb4/t4_hw.c
9379
struct sge_params *sge_params = &adapter->params.sge;
drivers/net/ethernet/chelsio/cxgb4/t4_msg.h
1542
struct ulptx_sge_pair sge[];
drivers/net/ethernet/chelsio/cxgb4vf/adapter.h
350
#define for_each_ethrxq(sge, iter) \
drivers/net/ethernet/chelsio/cxgb4vf/adapter.h
351
for (iter = 0; iter < (sge)->ethqsets; iter++)
drivers/net/ethernet/chelsio/cxgb4vf/adapter.h
384
struct sge sge;
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
1019
static int closest_timer(const struct sge *s, int us)
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
1035
static int closest_thres(const struct sge *s, int thres)
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
1060
? adapter->sge.timer_val[timer_idx]
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
1097
pktcnt_idx = closest_thres(&adapter->sge, cnt);
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
1116
: closest_timer(&adapter->sge, us));
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
1239
rxq = &adapter->sge.ethrxq[pi->first_qset];
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
1599
const struct sge *s = &pi->adapter->sge;
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
1625
struct sge *s = &adapter->sge;
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
1660
const struct sge_rspq *rspq = &adapter->sge.ethrxq[pi->first_qset].rspq;
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
1665
? adapter->sge.counter_val[rspq->pktcnt_idx]
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
1684
&adapter->sge.ethrxq[pi->first_qset].rspq,
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
1809
const struct sge_eth_txq *txq = &adapter->sge.ethtxq[pi->first_qset];
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
1810
const struct sge_eth_rxq *rxq = &adapter->sge.ethrxq[pi->first_qset];
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
2044
int eth_entries = DIV_ROUND_UP(adapter->sge.ethqsets, QPL);
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
2062
const struct sge_eth_rxq *rxq = &adapter->sge.ethrxq[r * QPL];
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
2063
const struct sge_eth_txq *txq = &adapter->sge.ethtxq[r * QPL];
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
2064
int n = min(QPL, adapter->sge.ethqsets - QPL * r);
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
2086
adapter->sge.counter_val[rxq[qs].rspq.pktcnt_idx]);
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
2099
const struct sge_rspq *evtq = &adapter->sge.fw_evtq;
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
2106
adapter->sge.counter_val[evtq->pktcnt_idx]);
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
2110
const struct sge_rspq *intrq = &adapter->sge.intrq;
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
2117
adapter->sge.counter_val[intrq->pktcnt_idx]);
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
2140
return DIV_ROUND_UP(adapter->sge.ethqsets, QPL) + 1 +
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
2180
int eth_entries = DIV_ROUND_UP(adapter->sge.ethqsets, QPL);
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
2202
const struct sge_eth_rxq *rxq = &adapter->sge.ethrxq[r * QPL];
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
2203
const struct sge_eth_txq *txq = &adapter->sge.ethtxq[r * QPL];
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
2204
int n = min(QPL, adapter->sge.ethqsets - QPL * r);
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
2232
const struct sge_rspq *evtq = &adapter->sge.fw_evtq;
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
2240
const struct sge_rspq *intrq = &adapter->sge.intrq;
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
2269
return DIV_ROUND_UP(adapter->sge.ethqsets, QPL) + 1 +
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
2500
adapter->sge.max_ethqsets = ethqsets;
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
2502
if (adapter->sge.max_ethqsets < adapter->params.nports) {
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
2505
adapter->sge.max_ethqsets, adapter->params.nports);
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
2506
adapter->params.nports = adapter->sge.max_ethqsets;
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
2517
struct sge_params *sge_params = &adapter->params.sge;
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
2518
struct sge *s = &adapter->sge;
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
2670
struct sge *s = &adapter->sge;
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
2697
q10g = (adapter->sge.max_ethqsets - n1g) / n10g;
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
2774
while (n < adapter->sge.ethqsets)
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
2779
adapter->sge.ethqsets--;
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
2780
if (adapter->sge.ethqsets <= n)
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
2807
struct sge *s = &adapter->sge;
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
382
struct sge *s = &adapter->sge;
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
420
struct sge *s = &adapter->sge;
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
453
struct sge *s = &adapter->sge;
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
476
struct sge *s = &adapter->sge;
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
534
struct sge *s = &adapter->sge;
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
590
struct sge *s = &adapter->sge;
drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c
705
struct sge_eth_rxq *rxq = &adapter->sge.ethrxq[pi->first_qset];
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
1195
txq = &adapter->sge.ethtxq[pi->first_qset + qidx];
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
1568
struct sge *s = &adapter->sge;
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
1620
struct sge *s = &adapter->sge;
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
1758
struct sge *s = &adapter->sge;
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
1936
struct sge *s = &adapter->sge;
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
1941
spin_lock(&adapter->sge.intrq_lock);
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
2018
spin_unlock(&adapter->sge.intrq_lock);
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
2065
struct adapter *adapter = timer_container_of(adapter, t, sge.rx_timer);
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
2066
struct sge *s = &adapter->sge;
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
2124
struct adapter *adapter = timer_container_of(adapter, t, sge.tx_timer);
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
2125
struct sge *s = &adapter->sge;
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
2206
struct sge *s = &adapter->sge;
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
2220
rspq != &adapter->sge.intrq) {
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
2222
intr_dest = adapter->sge.intrq.abs_id;
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
2417
struct sge *s = &adapter->sge;
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
2513
struct sge *s = &adapter->sge;
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
2530
struct sge *s = &adapter->sge;
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
2563
struct sge *s = &adapter->sge;
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
2570
for (qs = 0; qs < adapter->sge.ethqsets; qs++, rxq++, txq++) {
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
2594
adapter->sge.ethtxq_rover = 0;
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
2595
mod_timer(&adapter->sge.rx_timer, jiffies + RX_QCHECK_PERIOD);
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
2596
mod_timer(&adapter->sge.tx_timer, jiffies + TX_QCHECK_PERIOD);
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
2609
struct sge *s = &adapter->sge;
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
2628
struct sge_params *sge_params = &adapter->params.sge;
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
2631
struct sge *s = &adapter->sge;
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
264
const struct sge *s = &adapter->sge;
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
325
for (p = sgl->sge; nfrags >= 2; nfrags -= 2) {
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
457
const struct sge *s = &adapter->sge;
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
607
struct sge *s = &adapter->sge;
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
715
set_bit(fl->cntxt_id, adapter->sge.starving_fl);
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
930
to = (u8 *)end > (u8 *)tq->stat ? buf : sgl->sge;
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
944
unsigned int part0 = (u8 *)tq->stat - (u8 *)sgl->sge, part1;
drivers/net/ethernet/chelsio/cxgb4vf/sge.c
947
memcpy(sgl->sge, buf, part0);
drivers/net/ethernet/chelsio/cxgb4vf/t4vf_common.h
262
struct sge_params sge; /* Scatter Gather Engine */
drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c
702
sge_control = adapter->params.sge.sge_control;
drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c
728
sge_control2 = adapter->params.sge.sge_control2;
drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c
783
page_shift = adapter->params.sge.sge_vf_hps + 10;
drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c
789
? adapter->params.sge.sge_vf_eq_qpp
drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c
790
: adapter->params.sge.sge_vf_iq_qpp);
drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c
849
struct sge_params *sge_params = &adapter->params.sge;
drivers/net/ethernet/chelsio/inline_crypto/ch_ipsec/chcr_ipsec.c
426
q = &adap->sge.ethtxq[qidx + pi->first_qset];
drivers/net/ethernet/chelsio/inline_crypto/ch_ipsec/chcr_ipsec.c
474
q = &adap->sge.ethtxq[qidx + pi->first_qset];
drivers/net/ethernet/chelsio/inline_crypto/ch_ipsec/chcr_ipsec.c
517
q = &adap->sge.ethtxq[qidx + pi->first_qset];
drivers/net/ethernet/chelsio/inline_crypto/ch_ipsec/chcr_ipsec.c
573
struct sge_eth_txq *q = &adap->sge.ethtxq[qidx + pi->first_qset];
drivers/net/ethernet/chelsio/inline_crypto/ch_ipsec/chcr_ipsec.c
731
q = &adap->sge.ethtxq[qidx + pi->first_qset];
drivers/net/ethernet/chelsio/inline_crypto/ch_ktls/chcr_ktls.c
1962
q = &adap->sge.ethtxq[qidx + tx_info->first_qset];
drivers/net/ethernet/emulex/benet/be_cmds.c
800
struct be_sge *sge;
drivers/net/ethernet/emulex/benet/be_cmds.c
811
sge = nonembedded_sgl(wrb);
drivers/net/ethernet/emulex/benet/be_cmds.c
812
sge->pa_hi = cpu_to_le32(upper_32_bits(mem->dma));
drivers/net/ethernet/emulex/benet/be_cmds.c
813
sge->pa_lo = cpu_to_le32(mem->dma & 0xFFFFFFFF);
drivers/net/ethernet/emulex/benet/be_cmds.c
814
sge->len = cpu_to_le32(mem->size);
drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c
213
struct dpaa2_sg_entry *sge = &sgt[i];
drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c
220
sg_addr = dpaa2_sg_get_addr(sge);
drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c
225
sg_length = dpaa2_sg_get_len(sge);
drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c
245
sg_offset = dpaa2_sg_get_offset(sge);
drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c
266
if (dpaa2_sg_is_final(sge))
drivers/net/ethernet/huawei/hinic/hinic_common.c
55
void hinic_set_sge(struct hinic_sge *sge, dma_addr_t addr, int len)
drivers/net/ethernet/huawei/hinic/hinic_common.c
57
sge->hi_addr = upper_32_bits(addr);
drivers/net/ethernet/huawei/hinic/hinic_common.c
58
sge->lo_addr = lower_32_bits(addr);
drivers/net/ethernet/huawei/hinic/hinic_common.c
59
sge->len = len;
drivers/net/ethernet/huawei/hinic/hinic_common.c
68
dma_addr_t hinic_sge_to_dma(struct hinic_sge *sge)
drivers/net/ethernet/huawei/hinic/hinic_common.c
70
return (dma_addr_t)((((u64)sge->hi_addr) << 32) | sge->lo_addr);
drivers/net/ethernet/huawei/hinic/hinic_common.h
25
void hinic_set_sge(struct hinic_sge *sge, dma_addr_t addr, int len);
drivers/net/ethernet/huawei/hinic/hinic_common.h
27
dma_addr_t hinic_sge_to_dma(struct hinic_sge *sge);
drivers/net/ethernet/huawei/hinic/hinic_hw_cmdq.c
174
hinic_set_sge(&sge_resp->sge, buf_out->dma_addr, buf_out->size);
drivers/net/ethernet/huawei/hinic/hinic_hw_cmdq.c
231
hinic_set_sge(&wqe_lcmd->buf_desc.sge, buf_in->dma_addr, buf_in->size);
drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c
604
sq_wqe->buf_descs[i].sge = sges[i];
drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c
783
sges[i] = sq_wqe->buf_descs[i].sge;
drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c
923
u16 cons_idx, struct hinic_sge *sge)
drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c
928
sge->hi_addr = be32_to_cpu(rq_wqe->buf_desc.hi_addr);
drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c
929
sge->lo_addr = be32_to_cpu(rq_wqe->buf_desc.lo_addr);
drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c
930
sge->len = HINIC_RQ_CQE_SGE_GET(len, LEN);
drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c
941
struct hinic_rq_wqe *rq_wqe, struct hinic_sge *sge)
drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c
957
hinic_set_sge(&cqe_sect->sge, cqe_dma, sizeof(*cqe));
drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c
959
buf_desc->hi_addr = sge->hi_addr;
drivers/net/ethernet/huawei/hinic/hinic_hw_qp.c
960
buf_desc->lo_addr = sge->lo_addr;
drivers/net/ethernet/huawei/hinic/hinic_hw_qp.h
225
u16 cons_idx, struct hinic_sge *sge);
drivers/net/ethernet/huawei/hinic/hinic_hw_qp.h
228
struct hinic_rq_wqe *wqe, struct hinic_sge *sge);
drivers/net/ethernet/huawei/hinic/hinic_hw_wqe.h
309
struct hinic_sge sge;
drivers/net/ethernet/huawei/hinic/hinic_hw_wqe.h
328
struct hinic_sge sge;
drivers/net/ethernet/huawei/hinic/hinic_hw_wqe.h
378
struct hinic_sge sge;
drivers/net/ethernet/huawei/hinic/hinic_hw_wqe.h
405
struct hinic_sge sge;
drivers/net/ethernet/huawei/hinic/hinic_rx.c
198
struct hinic_sge sge;
drivers/net/ethernet/huawei/hinic/hinic_rx.c
215
hinic_set_sge(&sge, dma_addr, skb->len);
drivers/net/ethernet/huawei/hinic/hinic_rx.c
224
hinic_rq_prepare_wqe(rxq->rq, prod_idx, rq_wqe, &sge);
drivers/net/ethernet/huawei/hinic/hinic_rx.c
247
struct hinic_sge sge;
drivers/net/ethernet/huawei/hinic/hinic_rx.c
254
hinic_rq_get_sge(rq, &hw_wqe->rq_wqe, ci, &sge);
drivers/net/ethernet/huawei/hinic/hinic_rx.c
258
rx_free_skb(rxq, rq->saved_skb[ci], hinic_sge_to_dma(&sge));
drivers/net/ethernet/huawei/hinic/hinic_rx.c
277
struct hinic_sge sge;
drivers/net/ethernet/huawei/hinic/hinic_rx.c
286
hinic_rq_get_sge(rxq->rq, rq_wqe, ci, &sge);
drivers/net/ethernet/huawei/hinic/hinic_rx.c
288
rx_unmap_skb(rxq, hinic_sge_to_dma(&sge));
drivers/net/ethernet/huawei/hinic/hinic_rx.c
365
struct hinic_sge sge;
drivers/net/ethernet/huawei/hinic/hinic_rx.c
389
hinic_rq_get_sge(rxq->rq, rq_wqe, ci, &sge);
drivers/net/ethernet/huawei/hinic/hinic_rx.c
391
rx_unmap_skb(rxq, hinic_sge_to_dma(&sge));
drivers/net/ethernet/huawei/hinic/hinic_rx.c
397
pkt_len = sge.len;
drivers/net/ethernet/huawei/hinic3/hinic3_cmdq.c
305
struct hinic3_sge *sge = &complete->resp.sge;
drivers/net/ethernet/huawei/hinic3/hinic3_cmdq.c
307
hinic3_set_sge(sge, buf_out->dma_addr, cpu_to_le32(CMDQ_BUF_SIZE));
drivers/net/ethernet/huawei/hinic3/hinic3_cmdq.c
321
hinic3_set_sge(&wqe->buf_desc.sge, buf_in->dma_addr,
drivers/net/ethernet/huawei/hinic3/hinic3_cmdq.h
26
struct hinic3_sge sge;
drivers/net/ethernet/huawei/hinic3/hinic3_cmdq.h
46
struct hinic3_sge sge;
drivers/net/ethernet/huawei/hinic3/hinic3_common.h
33
static inline void hinic3_set_sge(struct hinic3_sge *sge, dma_addr_t addr,
drivers/net/ethernet/huawei/hinic3/hinic3_common.h
36
sge->hi_addr = cpu_to_le32(upper_32_bits(addr));
drivers/net/ethernet/huawei/hinic3/hinic3_common.h
37
sge->lo_addr = cpu_to_le32(lower_32_bits(addr));
drivers/net/ethernet/huawei/hinic3/hinic3_common.h
38
sge->len = len;
drivers/net/ethernet/huawei/hinic3/hinic3_common.h
39
sge->rsvd = 0;
drivers/net/ethernet/ibm/ibmvnic.h
806
struct ibmvnic_sge_desc sge;
drivers/net/ethernet/microsoft/mana/hw_channel.c
251
struct gdma_sge *sge;
drivers/net/ethernet/microsoft/mana/hw_channel.c
263
sge = (struct gdma_sge *)(wqe + 8 + dma_oob->inline_oob_size_div4 * 4);
drivers/net/ethernet/microsoft/mana/hw_channel.c
267
rx_req_idx = (sge->address - rq_base_addr) / hwc->max_req_msg_size;
drivers/net/ethernet/microsoft/mana/hw_channel.c
578
struct gdma_sge *sge;
drivers/net/ethernet/microsoft/mana/hw_channel.c
598
sge = &req->sge;
drivers/net/ethernet/microsoft/mana/hw_channel.c
599
sge->address = (u64)req->buf_sge_addr;
drivers/net/ethernet/microsoft/mana/hw_channel.c
60
struct gdma_sge *sge;
drivers/net/ethernet/microsoft/mana/hw_channel.c
600
sge->mem_key = hwc_txq->msg_buf->gpa_mkey;
drivers/net/ethernet/microsoft/mana/hw_channel.c
601
sge->size = req->msg_size;
drivers/net/ethernet/microsoft/mana/hw_channel.c
604
req->wqe_req.sgl = sge;
drivers/net/ethernet/microsoft/mana/hw_channel.c
63
sge = &req->sge;
drivers/net/ethernet/microsoft/mana/hw_channel.c
64
sge->address = (u64)req->buf_sge_addr;
drivers/net/ethernet/microsoft/mana/hw_channel.c
65
sge->mem_key = hwc_rxq->msg_buf->gpa_mkey;
drivers/net/ethernet/microsoft/mana/hw_channel.c
66
sge->size = req->buf_len;
drivers/net/ethernet/microsoft/mana/hw_channel.c
69
req->wqe_req.sgl = sge;
drivers/net/ethernet/qlogic/qed/qed_nvmetcp_fw_funcs.c
39
ctx_data_desc->sge[sge_index].sge_addr.lo =
drivers/net/ethernet/qlogic/qed/qed_nvmetcp_fw_funcs.c
41
ctx_data_desc->sge[sge_index].sge_addr.hi =
drivers/net/ethernet/qlogic/qed/qed_nvmetcp_fw_funcs.c
43
ctx_data_desc->sge[sge_index].sge_len =
drivers/nvme/host/pci.c
1118
static void nvme_pci_sgl_set_data(struct nvme_sgl_desc *sge,
drivers/nvme/host/pci.c
1121
sge->addr = cpu_to_le64(iter->addr);
drivers/nvme/host/pci.c
1122
sge->length = cpu_to_le32(iter->len);
drivers/nvme/host/pci.c
1123
sge->type = NVME_SGL_FMT_DATA_DESC << 4;
drivers/nvme/host/pci.c
1126
static void nvme_pci_sgl_set_seg(struct nvme_sgl_desc *sge,
drivers/nvme/host/pci.c
1129
sge->addr = cpu_to_le64(dma_addr);
drivers/nvme/host/pci.c
1130
sge->length = cpu_to_le32(entries * sizeof(*sge));
drivers/nvme/host/pci.c
1131
sge->type = NVME_SGL_FMT_LAST_SEG_DESC << 4;
drivers/nvme/host/pci.c
869
static void nvme_free_sgls(struct request *req, struct nvme_sgl_desc *sge,
drivers/nvme/host/pci.c
874
unsigned int len = le32_to_cpu(sge->length);
drivers/nvme/host/pci.c
878
if (sge->type == (NVME_SGL_FMT_DATA_DESC << 4)) {
drivers/nvme/host/pci.c
879
dma_unmap_phys(dma_dev, le64_to_cpu(sge->addr), len, dir,
drivers/nvme/host/pci.c
896
struct nvme_sgl_desc *sge = iod->meta_descriptor;
drivers/nvme/host/pci.c
916
nvme_free_sgls(req, sge, &sge[1], attrs);
drivers/nvme/host/rdma.c
1269
struct ib_sge *sge = &req->sge[1];
drivers/nvme/host/rdma.c
1275
sge->addr = sg_dma_address(sgl);
drivers/nvme/host/rdma.c
1276
sge->length = sg_dma_len(sgl);
drivers/nvme/host/rdma.c
1277
sge->lkey = queue->device->pd->local_dma_lkey;
drivers/nvme/host/rdma.c
1278
len += sge->length;
drivers/nvme/host/rdma.c
1279
sge++;
drivers/nvme/host/rdma.c
1592
struct nvme_rdma_qe *qe, struct ib_sge *sge, u32 num_sge,
drivers/nvme/host/rdma.c
1598
sge->addr = qe->dma;
drivers/nvme/host/rdma.c
1599
sge->length = sizeof(struct nvme_command);
drivers/nvme/host/rdma.c
1600
sge->lkey = queue->device->pd->local_dma_lkey;
drivers/nvme/host/rdma.c
1604
wr.sg_list = sge;
drivers/nvme/host/rdma.c
1670
struct ib_sge sge;
drivers/nvme/host/rdma.c
1686
ret = nvme_rdma_post_send(queue, sqe, &sge, 1, NULL);
drivers/nvme/host/rdma.c
2050
err = nvme_rdma_post_send(queue, sqe, req->sge, req->num_sge,
drivers/nvme/host/rdma.c
69
struct ib_sge sge[1 + NVME_RDMA_MAX_INLINE_SEGMENTS];
drivers/nvme/target/rdma.c
255
struct ib_sge *sge;
drivers/nvme/target/rdma.c
262
sge = &c->sge[1];
drivers/nvme/target/rdma.c
264
for (i = 0; i < ndev->inline_page_count; i++, sg++, sge++) {
drivers/nvme/target/rdma.c
265
if (sge->length)
drivers/nvme/target/rdma.c
266
ib_dma_unmap_page(ndev->device, sge->addr,
drivers/nvme/target/rdma.c
267
sge->length, DMA_FROM_DEVICE);
drivers/nvme/target/rdma.c
277
struct ib_sge *sge;
drivers/nvme/target/rdma.c
287
sge = &c->sge[1];
drivers/nvme/target/rdma.c
290
for (i = 0; i < ndev->inline_page_count; i++, sg++, sge++) {
drivers/nvme/target/rdma.c
295
sge->addr = ib_dma_map_page(ndev->device,
drivers/nvme/target/rdma.c
297
if (ib_dma_mapping_error(ndev->device, sge->addr))
drivers/nvme/target/rdma.c
299
sge->length = min_t(int, len, PAGE_SIZE);
drivers/nvme/target/rdma.c
300
sge->lkey = ndev->pd->local_dma_lkey;
drivers/nvme/target/rdma.c
301
len -= sge->length;
drivers/nvme/target/rdma.c
306
for (; i >= 0; i--, sg--, sge--) {
drivers/nvme/target/rdma.c
307
if (sge->length)
drivers/nvme/target/rdma.c
308
ib_dma_unmap_page(ndev->device, sge->addr,
drivers/nvme/target/rdma.c
309
sge->length, DMA_FROM_DEVICE);
drivers/nvme/target/rdma.c
324
c->sge[0].addr = ib_dma_map_single(ndev->device, c->nvme_cmd,
drivers/nvme/target/rdma.c
326
if (ib_dma_mapping_error(ndev->device, c->sge[0].addr))
drivers/nvme/target/rdma.c
329
c->sge[0].length = sizeof(*c->nvme_cmd);
drivers/nvme/target/rdma.c
330
c->sge[0].lkey = ndev->pd->local_dma_lkey;
drivers/nvme/target/rdma.c
338
c->wr.sg_list = c->sge;
drivers/nvme/target/rdma.c
344
ib_dma_unmap_single(ndev->device, c->sge[0].addr,
drivers/nvme/target/rdma.c
358
ib_dma_unmap_single(ndev->device, c->sge[0].addr,
drivers/nvme/target/rdma.c
47
struct ib_sge sge[NVMET_RDMA_MAX_INLINE_SGE + 1];
drivers/nvme/target/rdma.c
499
cmd->sge[0].addr, cmd->sge[0].length,
drivers/nvme/target/rdma.c
972
cmd->cmd->sge[0].addr, cmd->cmd->sge[0].length,
drivers/scsi/aacraid/aachba.c
1252
fibsize = struct_size(readcmd2, sge,
drivers/scsi/aacraid/aachba.c
1383
fibsize = struct_size(writecmd2, sge,
drivers/scsi/aacraid/aachba.c
3941
rio2->sge[i].addrHigh = cpu_to_le32((u32)(addr>>32));
drivers/scsi/aacraid/aachba.c
3942
rio2->sge[i].addrLow = cpu_to_le32((u32)(addr & 0xffffffff));
drivers/scsi/aacraid/aachba.c
3944
rio2->sge[i].length = cur_size;
drivers/scsi/aacraid/aachba.c
3945
rio2->sge[i].flags = 0;
drivers/scsi/aacraid/aachba.c
3962
u32 temp = le32_to_cpu(rio2->sge[i-1].length) -
drivers/scsi/aacraid/aachba.c
3964
rio2->sge[i-1].length = cpu_to_le32(temp);
drivers/scsi/aacraid/aachba.c
3977
if (rio2->sge[j].length % (i*PAGE_SIZE)) {
drivers/scsi/aacraid/aachba.c
3981
nseg_new += (rio2->sge[j].length / (i*PAGE_SIZE));
drivers/scsi/aacraid/aachba.c
4006
struct sge_ieee1212 *sge;
drivers/scsi/aacraid/aachba.c
4013
sge = kmalloc_objs(*sge, nseg_new, GFP_ATOMIC);
drivers/scsi/aacraid/aachba.c
4014
if (sge == NULL)
drivers/scsi/aacraid/aachba.c
4018
for (j = 0; j < rio2->sge[i].length / (pages * PAGE_SIZE); ++j) {
drivers/scsi/aacraid/aachba.c
4019
addr_low = rio2->sge[i].addrLow + j * pages * PAGE_SIZE;
drivers/scsi/aacraid/aachba.c
4020
sge[pos].addrLow = addr_low;
drivers/scsi/aacraid/aachba.c
4021
sge[pos].addrHigh = rio2->sge[i].addrHigh;
drivers/scsi/aacraid/aachba.c
4022
if (addr_low < rio2->sge[i].addrLow)
drivers/scsi/aacraid/aachba.c
4023
sge[pos].addrHigh++;
drivers/scsi/aacraid/aachba.c
4024
sge[pos].length = pages * PAGE_SIZE;
drivers/scsi/aacraid/aachba.c
4025
sge[pos].flags = 0;
drivers/scsi/aacraid/aachba.c
4029
sge[pos] = rio2->sge[nseg-1];
drivers/scsi/aacraid/aachba.c
4030
memcpy(&rio2->sge[1], &sge[1], (nseg_new-1)*sizeof(struct sge_ieee1212));
drivers/scsi/aacraid/aachba.c
4032
kfree(sge);
drivers/scsi/aacraid/aachba.c
4049
struct aac_hba_sgl *sge;
drivers/scsi/aacraid/aachba.c
4058
sge = &hbacmd->sge[2];
drivers/scsi/aacraid/aachba.c
4060
sge = &hbacmd->sge[0];
drivers/scsi/aacraid/aachba.c
4067
sge->addr_hi = cpu_to_le32((u32)(addr>>32));
drivers/scsi/aacraid/aachba.c
4068
sge->addr_lo = cpu_to_le32((u32)(addr & 0xffffffff));
drivers/scsi/aacraid/aachba.c
4070
sge->len = cur_size;
drivers/scsi/aacraid/aachba.c
4071
sge->flags = 0;
drivers/scsi/aacraid/aachba.c
4073
sge++;
drivers/scsi/aacraid/aachba.c
4076
sge--;
drivers/scsi/aacraid/aachba.c
4081
temp = le32_to_cpu(sge->len) - byte_count
drivers/scsi/aacraid/aachba.c
4083
sge->len = cpu_to_le32(temp);
drivers/scsi/aacraid/aachba.c
4089
sge->flags = cpu_to_le32(0x40000000);
drivers/scsi/aacraid/aachba.c
4092
hbacmd->sge[0].flags = cpu_to_le32(0x80000000);
drivers/scsi/aacraid/aachba.c
4094
hbacmd->sge[0].addr_hi = (u32)cpu_to_le32(sg_address >> 32);
drivers/scsi/aacraid/aachba.c
4095
hbacmd->sge[0].addr_lo =
drivers/scsi/aacraid/aacraid.h
1927
struct sge_ieee1212 sge[];
drivers/scsi/aacraid/aacraid.h
227
struct aac_hba_sgl sge[HBA_MAX_SG_SEPARATE+2]; /* SG list space */
drivers/scsi/aacraid/commctrl.c
693
hbacmd->sge[i].addr_hi = cpu_to_le32((u32)(addr>>32));
drivers/scsi/aacraid/commctrl.c
694
hbacmd->sge[i].addr_lo = cpu_to_le32(
drivers/scsi/aacraid/commctrl.c
696
hbacmd->sge[i].len = cpu_to_le32(sg_count[i]);
drivers/scsi/aacraid/commctrl.c
697
hbacmd->sge[i].flags = 0;
drivers/scsi/aacraid/commctrl.c
702
hbacmd->sge[usg32->count-1].flags =
drivers/scsi/aacraid/commsup.c
176
offsetof(struct aac_hba_cmd_req, sge[2]);
drivers/scsi/be2iscsi/be_cmds.c
1262
struct be_sge *sge = nonembedded_sgl(wrb);
drivers/scsi/be2iscsi/be_cmds.c
1282
sge->pa_hi = cpu_to_le32(upper_32_bits(nonemb_cmd.dma));
drivers/scsi/be2iscsi/be_cmds.c
1283
sge->pa_lo = cpu_to_le32(nonemb_cmd.dma & 0xFFFFFFFF);
drivers/scsi/be2iscsi/be_cmds.c
1284
sge->len = cpu_to_le32(nonemb_cmd.size);
drivers/scsi/be2iscsi/be_mgmt.c
1011
struct be_sge *sge;
drivers/scsi/be2iscsi/be_mgmt.c
1035
sge = nonembedded_sgl(wrb);
drivers/scsi/be2iscsi/be_mgmt.c
1041
sge->pa_hi = cpu_to_le32(upper_32_bits(nonemb_cmd->dma));
drivers/scsi/be2iscsi/be_mgmt.c
1042
sge->pa_lo = cpu_to_le32(nonemb_cmd->dma & 0xFFFFFFFF);
drivers/scsi/be2iscsi/be_mgmt.c
1043
sge->len = cpu_to_le32(nonemb_cmd->size);
drivers/scsi/be2iscsi/be_mgmt.c
128
struct be_sge *sge;
drivers/scsi/be2iscsi/be_mgmt.c
1510
struct be_sge *sge;
drivers/scsi/be2iscsi/be_mgmt.c
1547
sge = nonembedded_sgl(wrb);
drivers/scsi/be2iscsi/be_mgmt.c
1548
sge->pa_hi = cpu_to_le32(upper_32_bits(nonemb_cmd.dma));
drivers/scsi/be2iscsi/be_mgmt.c
1549
sge->pa_lo = cpu_to_le32(lower_32_bits(nonemb_cmd.dma));
drivers/scsi/be2iscsi/be_mgmt.c
155
sge = nonembedded_sgl(wrb);
drivers/scsi/be2iscsi/be_mgmt.c
1550
sge->len = cpu_to_le32(nonemb_cmd.size);
drivers/scsi/be2iscsi/be_mgmt.c
198
sge->pa_hi = cpu_to_le32(upper_32_bits(nonemb_cmd->dma));
drivers/scsi/be2iscsi/be_mgmt.c
199
sge->pa_lo = cpu_to_le32(nonemb_cmd->dma & 0xFFFFFFFF);
drivers/scsi/be2iscsi/be_mgmt.c
200
sge->len = cpu_to_le32(nonemb_cmd->size);
drivers/scsi/be2iscsi/be_mgmt.c
230
struct be_sge *sge;
drivers/scsi/be2iscsi/be_mgmt.c
241
sge = nonembedded_sgl(wrb);
drivers/scsi/be2iscsi/be_mgmt.c
243
sge->pa_hi = cpu_to_le32(upper_32_bits(nonemb_cmd->dma));
drivers/scsi/be2iscsi/be_mgmt.c
244
sge->pa_lo = cpu_to_le32(lower_32_bits(nonemb_cmd->dma));
drivers/scsi/be2iscsi/be_mgmt.c
245
sge->len = cpu_to_le32(nonemb_cmd->size);
drivers/scsi/bfa/bfa_fcpim.c
2397
struct bfi_sge_s *sge, *sgpge;
drivers/scsi/bfa/bfa_fcpim.c
2425
sge = &m->sges[0];
drivers/scsi/bfa/bfa_fcpim.c
2434
sge->sga = *(union bfi_addr_u *) &addr;
drivers/scsi/bfa/bfa_fcpim.c
2436
sge->sg_len = pgdlen;
drivers/scsi/bfa/bfa_fcpim.c
2437
sge->flags = (ioim->nsges > BFI_SGE_INLINE) ?
drivers/scsi/bfa/bfa_fcpim.c
2439
bfa_sge_to_be(sge);
drivers/scsi/bfa/bfa_fcpim.c
2440
sge++;
drivers/scsi/bfa/bfa_fcpim.c
2481
sge->sga = ioim->sgpg->sgpg_pa;
drivers/scsi/bfa/bfa_fcpim.c
2483
sge->sga.a32.addr_lo = 0;
drivers/scsi/bfa/bfa_fcpim.c
2484
sge->sga.a32.addr_hi = 0;
drivers/scsi/bfa/bfa_fcpim.c
2486
sge->sg_len = pgdlen;
drivers/scsi/bfa/bfa_fcpim.c
2487
sge->flags = BFI_SGE_PGDLEN;
drivers/scsi/bfa/bfa_fcpim.c
2488
bfa_sge_to_be(sge);
drivers/scsi/bfa/bfad_bsg.c
3210
struct bfa_sge_s *sge;
drivers/scsi/bfa/bfad_bsg.c
3213
sge = drv_fcxp->req_sge + sgeid;
drivers/scsi/bfa/bfad_bsg.c
3214
addr = (u64)(size_t) sge->sg_addr;
drivers/scsi/bfa/bfad_bsg.c
3222
struct bfa_sge_s *sge;
drivers/scsi/bfa/bfad_bsg.c
3224
sge = drv_fcxp->req_sge + sgeid;
drivers/scsi/bfa/bfad_bsg.c
3225
return sge->sg_len;
drivers/scsi/bfa/bfad_bsg.c
3232
struct bfa_sge_s *sge;
drivers/scsi/bfa/bfad_bsg.c
3235
sge = drv_fcxp->rsp_sge + sgeid;
drivers/scsi/bfa/bfad_bsg.c
3236
addr = (u64)(size_t) sge->sg_addr;
drivers/scsi/bfa/bfad_bsg.c
3244
struct bfa_sge_s *sge;
drivers/scsi/bfa/bfad_bsg.c
3246
sge = drv_fcxp->rsp_sge + sgeid;
drivers/scsi/bfa/bfad_bsg.c
3247
return sge->sg_len;
drivers/scsi/csiostor/csio_wr.c
1049
struct csio_sge *sge = &wrm->sge;
drivers/scsi/csiostor/csio_wr.c
1074
bufsz = csio_wr_fl_bufsz(sge, buf);
drivers/scsi/csiostor/csio_wr.c
1095
flq->un.fl.offset += ALIGN(lastlen, sge->csio_fl_align);
drivers/scsi/csiostor/csio_wr.c
118
struct csio_sge *sge = &wrm->sge;
drivers/scsi/csiostor/csio_wr.c
126
buf->len = sge->sge_fl_buf_size[sreg];
drivers/scsi/csiostor/csio_wr.c
1311
struct csio_sge *sge = &wrm->sge;
drivers/scsi/csiostor/csio_wr.c
1417
sge->csio_fl_align = fl_align;
drivers/scsi/csiostor/csio_wr.c
1434
struct csio_sge *sge = &wrm->sge;
drivers/scsi/csiostor/csio_wr.c
1436
csio_sge_thresh_reg = csio_closest_thresh(sge, csio_intr_coalesce_cnt);
drivers/scsi/csiostor/csio_wr.c
1443
csio_sge_timer_reg = csio_closest_timer(sge, csio_intr_coalesce_time);
drivers/scsi/csiostor/csio_wr.c
1456
struct csio_sge *sge = &wrm->sge;
drivers/scsi/csiostor/csio_wr.c
1462
sge->sge_control = csio_rd_reg32(hw, SGE_CONTROL_A);
drivers/scsi/csiostor/csio_wr.c
1464
ingpad = INGPADBOUNDARY_G(sge->sge_control);
drivers/scsi/csiostor/csio_wr.c
1468
sge->csio_fl_align = 32; break;
drivers/scsi/csiostor/csio_wr.c
1470
sge->csio_fl_align = 64; break;
drivers/scsi/csiostor/csio_wr.c
1472
sge->csio_fl_align = 128; break;
drivers/scsi/csiostor/csio_wr.c
1474
sge->csio_fl_align = 256; break;
drivers/scsi/csiostor/csio_wr.c
1476
sge->csio_fl_align = 512; break;
drivers/scsi/csiostor/csio_wr.c
1478
sge->csio_fl_align = 1024; break;
drivers/scsi/csiostor/csio_wr.c
1480
sge->csio_fl_align = 2048; break;
drivers/scsi/csiostor/csio_wr.c
1482
sge->csio_fl_align = 4096; break;
drivers/scsi/csiostor/csio_wr.c
1486
csio_get_flbuf_size(hw, sge, i);
drivers/scsi/csiostor/csio_wr.c
1492
sge->timer_val[0] = (uint16_t)csio_core_ticks_to_us(hw,
drivers/scsi/csiostor/csio_wr.c
1494
sge->timer_val[1] = (uint16_t)csio_core_ticks_to_us(hw,
drivers/scsi/csiostor/csio_wr.c
1496
sge->timer_val[2] = (uint16_t)csio_core_ticks_to_us(hw,
drivers/scsi/csiostor/csio_wr.c
1498
sge->timer_val[3] = (uint16_t)csio_core_ticks_to_us(hw,
drivers/scsi/csiostor/csio_wr.c
1500
sge->timer_val[4] = (uint16_t)csio_core_ticks_to_us(hw,
drivers/scsi/csiostor/csio_wr.c
1502
sge->timer_val[5] = (uint16_t)csio_core_ticks_to_us(hw,
drivers/scsi/csiostor/csio_wr.c
1506
sge->counter_val[0] = THRESHOLD_0_G(ingress_rx_threshold);
drivers/scsi/csiostor/csio_wr.c
1507
sge->counter_val[1] = THRESHOLD_1_G(ingress_rx_threshold);
drivers/scsi/csiostor/csio_wr.c
1508
sge->counter_val[2] = THRESHOLD_2_G(ingress_rx_threshold);
drivers/scsi/csiostor/csio_wr.c
1509
sge->counter_val[3] = THRESHOLD_3_G(ingress_rx_threshold);
drivers/scsi/csiostor/csio_wr.c
1525
struct csio_sge *sge = &wrm->sge;
drivers/scsi/csiostor/csio_wr.c
1534
sge->sge_control = csio_rd_reg32(hw, SGE_CONTROL_A);
drivers/scsi/csiostor/csio_wr.c
1555
csio_wr_reg32(hw, (CSIO_SGE_FLBUF_SIZE2 + sge->csio_fl_align - 1)
drivers/scsi/csiostor/csio_wr.c
1556
& ~(sge->csio_fl_align - 1), SGE_FL_BUFFER_SIZE2_A);
drivers/scsi/csiostor/csio_wr.c
1557
csio_wr_reg32(hw, (CSIO_SGE_FLBUF_SIZE3 + sge->csio_fl_align - 1)
drivers/scsi/csiostor/csio_wr.c
1558
& ~(sge->csio_fl_align - 1), SGE_FL_BUFFER_SIZE3_A);
drivers/scsi/csiostor/csio_wr.c
1566
csio_get_flbuf_size(hw, sge, i);
drivers/scsi/csiostor/csio_wr.c
1569
sge->timer_val[0] = CSIO_SGE_TIMER_VAL_0;
drivers/scsi/csiostor/csio_wr.c
1570
sge->timer_val[1] = CSIO_SGE_TIMER_VAL_1;
drivers/scsi/csiostor/csio_wr.c
1571
sge->timer_val[2] = CSIO_SGE_TIMER_VAL_2;
drivers/scsi/csiostor/csio_wr.c
1572
sge->timer_val[3] = CSIO_SGE_TIMER_VAL_3;
drivers/scsi/csiostor/csio_wr.c
1573
sge->timer_val[4] = CSIO_SGE_TIMER_VAL_4;
drivers/scsi/csiostor/csio_wr.c
1574
sge->timer_val[5] = CSIO_SGE_TIMER_VAL_5;
drivers/scsi/csiostor/csio_wr.c
1576
sge->counter_val[0] = CSIO_SGE_INT_CNT_VAL_0;
drivers/scsi/csiostor/csio_wr.c
1577
sge->counter_val[1] = CSIO_SGE_INT_CNT_VAL_1;
drivers/scsi/csiostor/csio_wr.c
1578
sge->counter_val[2] = CSIO_SGE_INT_CNT_VAL_2;
drivers/scsi/csiostor/csio_wr.c
1579
sge->counter_val[3] = CSIO_SGE_INT_CNT_VAL_3;
drivers/scsi/csiostor/csio_wr.c
1581
csio_wr_reg32(hw, THRESHOLD_0_V(sge->counter_val[0]) |
drivers/scsi/csiostor/csio_wr.c
1582
THRESHOLD_1_V(sge->counter_val[1]) |
drivers/scsi/csiostor/csio_wr.c
1583
THRESHOLD_2_V(sge->counter_val[2]) |
drivers/scsi/csiostor/csio_wr.c
1584
THRESHOLD_3_V(sge->counter_val[3]),
drivers/scsi/csiostor/csio_wr.c
1588
TIMERVALUE0_V(csio_us_to_core_ticks(hw, sge->timer_val[0])) |
drivers/scsi/csiostor/csio_wr.c
1589
TIMERVALUE1_V(csio_us_to_core_ticks(hw, sge->timer_val[1])),
drivers/scsi/csiostor/csio_wr.c
1593
TIMERVALUE2_V(csio_us_to_core_ticks(hw, sge->timer_val[2])) |
drivers/scsi/csiostor/csio_wr.c
1594
TIMERVALUE3_V(csio_us_to_core_ticks(hw, sge->timer_val[3])),
drivers/scsi/csiostor/csio_wr.c
1598
TIMERVALUE4_V(csio_us_to_core_ticks(hw, sge->timer_val[4])) |
drivers/scsi/csiostor/csio_wr.c
1599
TIMERVALUE5_V(csio_us_to_core_ticks(hw, sge->timer_val[5])),
drivers/scsi/csiostor/csio_wr.c
58
csio_get_flbuf_size(struct csio_hw *hw, struct csio_sge *sge, uint32_t reg)
drivers/scsi/csiostor/csio_wr.c
60
sge->sge_fl_buf_size[reg] = csio_rd_reg32(hw, SGE_FL_BUFFER_SIZE0_A +
drivers/scsi/csiostor/csio_wr.c
66
csio_wr_fl_bufsz(struct csio_sge *sge, struct csio_dma_buf *buf)
drivers/scsi/csiostor/csio_wr.c
68
return sge->sge_fl_buf_size[buf->paddr & 0xF];
drivers/scsi/csiostor/csio_wr.c
75
return (hw->wrm.sge.sge_control & EGRSTATUSPAGESIZE_F) ? 128 : 64;
drivers/scsi/csiostor/csio_wr.h
449
struct csio_sge sge; /* SGE params */
drivers/scsi/elx/efct/efct_hw.c
2701
struct sli4_sge *sge = NULL;
drivers/scsi/elx/efct/efct_hw.c
2722
sge = hio->sgl->virt;
drivers/scsi/elx/efct/efct_hw.c
2727
sge0_flags = le32_to_cpu(sge[0].dw2_flags);
drivers/scsi/elx/efct/efct_hw.c
2728
sge1_flags = le32_to_cpu(sge[1].dw2_flags);
drivers/scsi/elx/efct/efct_hw.c
2730
sge[0].buffer_address_high =
drivers/scsi/elx/efct/efct_hw.c
2732
sge[0].buffer_address_low =
drivers/scsi/elx/efct/efct_hw.c
2737
sge[0].buffer_length = cpu_to_le32(len);
drivers/scsi/elx/efct/efct_hw.c
2742
sge[1].buffer_address_high =
drivers/scsi/elx/efct/efct_hw.c
2744
sge[1].buffer_address_low =
drivers/scsi/elx/efct/efct_hw.c
2750
sge[1].buffer_length = cpu_to_le32(receive->size);
drivers/scsi/elx/efct/efct_hw.c
2755
sge[0].dw2_flags = cpu_to_le32(sge0_flags);
drivers/scsi/elx/efct/efct_hw.c
2756
sge[1].dw2_flags = cpu_to_le32(sge1_flags);
drivers/scsi/elx/libefc_sli/sli4.c
1422
struct sli4_sge *sge = sgl->virt;
drivers/scsi/elx/libefc_sli/sli4.c
1437
bptr->u.data.low = sge[0].buffer_address_low;
drivers/scsi/elx/libefc_sli/sli4.c
1438
bptr->u.data.high = sge[0].buffer_address_high;
drivers/scsi/elx/libefc_sli/sli4.c
1567
struct sli4_sge *sge = NULL;
drivers/scsi/elx/libefc_sli/sli4.c
1578
sge = sgl->virt;
drivers/scsi/elx/libefc_sli/sli4.c
1586
(le32_to_cpu(sge[0].buffer_length) &
drivers/scsi/elx/libefc_sli/sli4.c
1589
bptr->u.data.low = sge[0].buffer_address_low;
drivers/scsi/elx/libefc_sli/sli4.c
1590
bptr->u.data.high = sge[0].buffer_address_high;
drivers/scsi/elx/libefc_sli/sli4.c
1602
len = le32_to_cpu(sge[0].buffer_length) +
drivers/scsi/elx/libefc_sli/sli4.c
1603
le32_to_cpu(sge[1].buffer_length);
drivers/scsi/elx/libefc_sli/sli4.c
1634
struct sli4_sge *sge = NULL;
drivers/scsi/elx/libefc_sli/sli4.c
1646
sge = sgl->virt;
drivers/scsi/elx/libefc_sli/sli4.c
1655
(le32_to_cpu(sge[0].buffer_length) &
drivers/scsi/elx/libefc_sli/sli4.c
1658
bptr->u.blp.low = sge[0].buffer_address_low;
drivers/scsi/elx/libefc_sli/sli4.c
1659
bptr->u.blp.high = sge[0].buffer_address_high;
drivers/scsi/elx/libefc_sli/sli4.c
1677
len = le32_to_cpu(sge[0].buffer_length);
drivers/scsi/elx/libefc_sli/sli4.c
1680
sge_flags = le32_to_cpu(sge[1].dw2_flags);
drivers/scsi/elx/libefc_sli/sli4.c
1683
sge[1].dw2_flags = cpu_to_le32(sge_flags);
drivers/scsi/elx/libefc_sli/sli4.c
1686
len = le32_to_cpu(sge[0].buffer_length) +
drivers/scsi/elx/libefc_sli/sli4.c
1687
le32_to_cpu(sge[1].buffer_length);
drivers/scsi/elx/libefc_sli/sli4.c
1717
(le32_to_cpu(sge[first_data_sge].buffer_length) &
drivers/scsi/elx/libefc_sli/sli4.c
1720
sge[first_data_sge].buffer_address_low;
drivers/scsi/elx/libefc_sli/sli4.c
1722
sge[first_data_sge].buffer_address_high;
drivers/scsi/elx/libefc_sli/sli4.c
1737
struct sli4_sge *sge = NULL;
drivers/scsi/elx/libefc_sli/sli4.c
1748
sge = sgl->virt;
drivers/scsi/elx/libefc_sli/sli4.c
1755
(le32_to_cpu(sge[0].buffer_length) & SLI4_BDE_LEN_MASK));
drivers/scsi/elx/libefc_sli/sli4.c
1756
bptr->u.data.low = sge[0].buffer_address_low;
drivers/scsi/elx/libefc_sli/sli4.c
1757
bptr->u.data.high = sge[0].buffer_address_high;
drivers/scsi/elx/libefc_sli/sli4.c
1772
len = le32_to_cpu(sge[0].buffer_length);
drivers/scsi/elx/libefc_sli/sli4.c
1774
sge_flags = le32_to_cpu(sge[1].dw2_flags);
drivers/scsi/elx/libefc_sli/sli4.c
1777
sge[1].dw2_flags = cpu_to_le32(sge_flags);
drivers/scsi/elx/libefc_sli/sli4.c
1780
len = le32_to_cpu(sge[0].buffer_length) +
drivers/scsi/elx/libefc_sli/sli4.c
1781
le32_to_cpu(sge[1].buffer_length);
drivers/scsi/elx/libefc_sli/sli4.c
1812
(le32_to_cpu(sge[first_data_sge].buffer_length) &
drivers/scsi/elx/libefc_sli/sli4.c
1815
bptr->u.data.low = sge[first_data_sge].buffer_address_low;
drivers/scsi/elx/libefc_sli/sli4.c
1816
bptr->u.data.high = sge[first_data_sge].buffer_address_high;
drivers/scsi/elx/libefc_sli/sli4.c
1829
struct sli4_sge *sge = NULL;
drivers/scsi/elx/libefc_sli/sli4.c
1839
sge = sgl->virt;
drivers/scsi/elx/libefc_sli/sli4.c
1848
(le32_to_cpu(sge[0].buffer_length)
drivers/scsi/elx/libefc_sli/sli4.c
1851
bptr->u.data.low = sge[0].buffer_address_low;
drivers/scsi/elx/libefc_sli/sli4.c
1852
bptr->u.data.high = sge[0].buffer_address_high;
drivers/scsi/elx/libefc_sli/sli4.c
1854
trecv->payload_offset_length = sge[0].buffer_length;
drivers/scsi/elx/libefc_sli/sli4.c
1860
params->xmit_len <= le32_to_cpu(sge[2].buffer_length)) {
drivers/scsi/elx/libefc_sli/sli4.c
1864
(le32_to_cpu(sge[2].buffer_length)
drivers/scsi/elx/libefc_sli/sli4.c
1867
bptr->u.data.low = sge[2].buffer_address_low;
drivers/scsi/elx/libefc_sli/sli4.c
1868
bptr->u.data.high = sge[2].buffer_address_high;
drivers/scsi/elx/libefc_sli/sli4.c
1919
(le32_to_cpu(sge[first_data_sge].buffer_length) &
drivers/scsi/elx/libefc_sli/sli4.c
1921
bptr->u.data.low = sge[first_data_sge].buffer_address_low;
drivers/scsi/elx/libefc_sli/sli4.c
1922
bptr->u.data.high = sge[first_data_sge].buffer_address_high;
drivers/scsi/elx/libefc_sli/sli4.c
1971
struct sli4_sge *sge = sgl->virt;
drivers/scsi/elx/libefc_sli/sli4.c
1982
(le32_to_cpu(sge[0].buffer_length) &
drivers/scsi/elx/libefc_sli/sli4.c
1984
bptr->u.data.low = sge[0].buffer_address_low;
drivers/scsi/elx/libefc_sli/sli4.c
1985
bptr->u.data.high = sge[0].buffer_address_high;
drivers/scsi/elx/libefc_sli/sli4.c
2032
struct sli4_sge *sge = NULL;
drivers/scsi/elx/libefc_sli/sli4.c
2042
sge = sgl->virt;
drivers/scsi/elx/libefc_sli/sli4.c
2052
(le32_to_cpu(sge[2].buffer_length) &
drivers/scsi/elx/libefc_sli/sli4.c
2058
bptr->u.data.low = sge[2].buffer_address_low;
drivers/scsi/elx/libefc_sli/sli4.c
2059
bptr->u.data.high = sge[2].buffer_address_high;
drivers/scsi/elx/libefc_sli/sli4.c
2065
params->xmit_len <= le32_to_cpu(sge[2].buffer_length)) {
drivers/scsi/elx/libefc_sli/sli4.c
2070
(le32_to_cpu(sge[2].buffer_length) &
drivers/scsi/elx/libefc_sli/sli4.c
2077
sge[2].buffer_address_low;
drivers/scsi/elx/libefc_sli/sli4.c
2079
sge[2].buffer_address_high;
drivers/scsi/elx/libefc_sli/sli4.c
2128
(le32_to_cpu(sge[first_data_sge].buffer_length) &
drivers/scsi/elx/libefc_sli/sli4.c
2131
sge[first_data_sge].buffer_address_low;
drivers/scsi/elx/libefc_sli/sli4.c
2133
sge[first_data_sge].buffer_address_high;
drivers/scsi/elx/libefc_sli/sli4.c
2156
struct sli4_sge *sge = NULL;
drivers/scsi/elx/libefc_sli/sli4.c
2166
sge = sgl->virt;
drivers/scsi/elx/libefc_sli/sli4.c
2177
bptr->u.data.low = sge[0].buffer_address_low;
drivers/scsi/elx/libefc_sli/sli4.c
2178
bptr->u.data.high = sge[0].buffer_address_high;
drivers/scsi/esas2r/atvda.h
1072
struct atto_vda_sge sge[1];
drivers/scsi/esas2r/atvda.h
1095
struct atto_vda_sge sge[1];
drivers/scsi/esas2r/atvda.h
1098
struct atto_vda_sge sge[1];
drivers/scsi/esas2r/atvda.h
1120
struct atto_vda_sge sge[1];
drivers/scsi/esas2r/atvda.h
1133
struct atto_vda_sge sge[1];
drivers/scsi/esas2r/atvda.h
1147
struct atto_vda_sge sge[1];
drivers/scsi/esas2r/atvda.h
1174
struct atto_vda_sge sge[1];
drivers/scsi/esas2r/atvda.h
1191
struct atto_vda_sge sge;
drivers/scsi/esas2r/atvda.h
1210
struct atto_vda_sge sge[2];
drivers/scsi/esas2r/atvda.h
504
struct atto_vda_sge sge[1];
drivers/scsi/esas2r/atvda.h
557
struct atto_vda_sge sge[1];
drivers/scsi/esas2r/esas2r.h
1175
sgc->sge.a64.limit = (struct atto_vda_sge *)((u8 *)rq->vrq
drivers/scsi/esas2r/esas2r.h
1182
sgc->sge.a64.last =
drivers/scsi/esas2r/esas2r.h
1183
sgc->sge.a64.curr = first;
drivers/scsi/esas2r/esas2r.h
1188
sgc->sge.a64.last =
drivers/scsi/esas2r/esas2r.h
1189
sgc->sge.a64.curr = &rq->vrq->scsi.u.sge[0];
drivers/scsi/esas2r/esas2r.h
1191
(u8)offsetof(struct atto_vda_scsi_req, u.sge);
drivers/scsi/esas2r/esas2r.h
1193
sgc->sge.a64.chain = NULL;
drivers/scsi/esas2r/esas2r.h
619
} sge;
drivers/scsi/esas2r/esas2r_disc.c
901
esas2r_sgc_init(&sgc, a, rq, rq->vrq->ioctl.sge);
drivers/scsi/esas2r/esas2r_flash.c
1312
vrq->data.sge[0].length =
drivers/scsi/esas2r/esas2r_flash.c
1315
vrq->data.sge[0].address = cpu_to_le64(
drivers/scsi/esas2r/esas2r_flash.c
212
esas2r_sgc_init(sgc, a, rq, &rq->vrq->flash.data.sge[0]);
drivers/scsi/esas2r/esas2r_flash.c
915
esas2r_sgc_init(sgc, a, rq, rq->vrq->flash.data.sge);
drivers/scsi/esas2r/esas2r_init.c
1201
data.sge);
drivers/scsi/esas2r/esas2r_io.c
222
if (unlikely(sgc->sge.a64.curr > sgc->sge.a64.limit)) {
drivers/scsi/esas2r/esas2r_io.c
237
sgelen = (u8)((u8 *)sgc->sge.a64.curr
drivers/scsi/esas2r/esas2r_io.c
238
- (u8 *)sgc->sge.a64.last);
drivers/scsi/esas2r/esas2r_io.c
244
memcpy(sgl->virt_addr, sgc->sge.a64.last, sgelen);
drivers/scsi/esas2r/esas2r_io.c
247
sgc->sge.a64.curr =
drivers/scsi/esas2r/esas2r_io.c
252
sgc->sge.a64.limit =
drivers/scsi/esas2r/esas2r_io.c
257
sgc->sge.a64.last->length = cpu_to_le32(
drivers/scsi/esas2r/esas2r_io.c
259
sgc->sge.a64.last->address =
drivers/scsi/esas2r/esas2r_io.c
268
if (sgc->sge.a64.chain) {
drivers/scsi/esas2r/esas2r_io.c
269
sgc->sge.a64.chain->length |=
drivers/scsi/esas2r/esas2r_io.c
271
((u8 *)(sgc->sge.a64.
drivers/scsi/esas2r/esas2r_io.c
280
sge.a64.last -
drivers/scsi/esas2r/esas2r_io.c
299
sgc->sge.a64.chain = sgc->sge.a64.last;
drivers/scsi/esas2r/esas2r_io.c
306
sgc->sge.a64.last = sgc->sge.a64.curr;
drivers/scsi/esas2r/esas2r_io.c
309
sgc->sge.a64.curr->length = cpu_to_le32(SGE_ADDR_64 | len);
drivers/scsi/esas2r/esas2r_io.c
310
sgc->sge.a64.curr->address = cpu_to_le32(addr);
drivers/scsi/esas2r/esas2r_io.c
311
sgc->sge.a64.curr++;
drivers/scsi/esas2r/esas2r_io.c
328
sgc->sge.a64.last->length |= cpu_to_le32(SGE_LAST);
drivers/scsi/esas2r/esas2r_io.c
334
if (sgc->sge.a64.chain) {
drivers/scsi/esas2r/esas2r_io.c
335
sgc->sge.a64.chain->length |= cpu_to_le32(
drivers/scsi/esas2r/esas2r_io.c
336
((u8 *)(sgc->sge.a64.curr) -
drivers/scsi/esas2r/esas2r_io.c
346
((u16)((u8 *)sgc->sge.a64.last - (u8 *)vrq)
drivers/scsi/esas2r/esas2r_io.c
409
if (sgc->sge.prd.sge_cnt == 0) {
drivers/scsi/esas2r/esas2r_io.c
418
sgc->sge.prd.curr->ctl_len = cpu_to_le32(
drivers/scsi/esas2r/esas2r_io.c
420
sgc->sge.prd.curr->address = cpu_to_le64(addr);
drivers/scsi/esas2r/esas2r_io.c
432
if (sgc->sge.prd.chain) {
drivers/scsi/esas2r/esas2r_io.c
439
sgc->sge.prd.chain->ctl_len |= cpu_to_le32(
drivers/scsi/esas2r/esas2r_io.c
440
sgc->sge.prd.sgl_max_cnt);
drivers/scsi/esas2r/esas2r_io.c
467
sgc->sge.prd.chain = sgc->sge.prd.curr;
drivers/scsi/esas2r/esas2r_io.c
469
sgc->sge.prd.chain->ctl_len = cpu_to_le32(PRD_CHAIN);
drivers/scsi/esas2r/esas2r_io.c
470
sgc->sge.prd.chain->address =
drivers/scsi/esas2r/esas2r_io.c
478
sgc->sge.prd.curr =
drivers/scsi/esas2r/esas2r_io.c
482
sgc->sge.prd.sge_cnt = sgc->sge.prd.sgl_max_cnt - 1;
drivers/scsi/esas2r/esas2r_io.c
485
sgc->sge.prd.sge_cnt--;
drivers/scsi/esas2r/esas2r_io.c
487
sgc->sge.prd.curr->ctl_len = cpu_to_le32(PRD_DATA | len);
drivers/scsi/esas2r/esas2r_io.c
488
sgc->sge.prd.curr->address = cpu_to_le64(addr);
drivers/scsi/esas2r/esas2r_io.c
492
sgc->sge.prd.curr++;
drivers/scsi/esas2r/esas2r_io.c
513
if (sgc->sge.prd.chain) {
drivers/scsi/esas2r/esas2r_io.c
514
sgc->sge.prd.chain->ctl_len |=
drivers/scsi/esas2r/esas2r_io.c
515
cpu_to_le32(sgc->sge.prd.sgl_max_cnt
drivers/scsi/esas2r/esas2r_io.c
516
- sgc->sge.prd.sge_cnt
drivers/scsi/esas2r/esas2r_io.c
614
(struct atto_physical_region_description *)sgc->sge.a64.curr;
drivers/scsi/esas2r/esas2r_io.c
616
sgc->sge.prd.sgl_max_cnt = sgl_page_size /
drivers/scsi/esas2r/esas2r_io.c
623
sgc->sge.prd.sge_cnt = 0;
drivers/scsi/esas2r/esas2r_io.c
624
sgc->sge.prd.chain = NULL;
drivers/scsi/esas2r/esas2r_io.c
625
sgc->sge.prd.curr = curr_iblk_chn;
drivers/scsi/esas2r/esas2r_ioctl.c
302
esas2r_sgc_init(sgc, a, rq, rq->vrq->ioctl.sge);
drivers/scsi/esas2r/esas2r_ioctl.c
355
esas2r_sgc_init(sgc, a, rq, rq->vrq->ioctl.sge);
drivers/scsi/esas2r/esas2r_ioctl.c
674
esas2r_sgc_init(sgc, a, rq, rq->vrq->ioctl.sge);
drivers/scsi/esas2r/esas2r_vda.c
119
firstsg = rq->vrq->flash.data.file.sge;
drivers/scsi/esas2r/esas2r_vda.c
130
firstsg = rq->vrq->cli.sge;
drivers/scsi/esas2r/esas2r_vda.c
194
firstsg = rq->vrq->mgt.sge;
drivers/scsi/esas2r/esas2r_vda.c
363
data.sge);
drivers/scsi/esas2r/esas2r_vda.c
394
struct atto_vda_mgmt_req, sge);
drivers/scsi/esas2r/esas2r_vda.c
396
vrq->sge[0].length = cpu_to_le32(SGE_LAST | length);
drivers/scsi/esas2r/esas2r_vda.c
397
vrq->sge[0].address = cpu_to_le64(
drivers/scsi/esas2r/esas2r_vda.c
432
(u8)offsetof(struct atto_vda_ae_req, sge);
drivers/scsi/esas2r/esas2r_vda.c
433
vrq->sge[0].length = cpu_to_le32(SGE_LAST | vrq->length);
drivers/scsi/esas2r/esas2r_vda.c
434
vrq->sge[0].address = cpu_to_le64(
drivers/scsi/esas2r/esas2r_vda.c
461
vrq->sg_list_offset = (u8)offsetof(struct atto_vda_ioctl_req, sge);
drivers/scsi/hisi_sas/hisi_sas.h
603
struct hisi_sas_sge sge[HISI_SAS_SGE_PAGE_CNT];
drivers/scsi/hisi_sas/hisi_sas.h
608
struct hisi_sas_sge sge[HISI_SAS_SGE_DIF_PAGE_CNT];
drivers/scsi/hisi_sas/hisi_sas_v1_hw.c
900
struct hisi_sas_sge *entry = &sge_page->sge[i];
drivers/scsi/hisi_sas/hisi_sas_v2_hw.c
1682
struct hisi_sas_sge *entry = &sge_page->sge[i];
drivers/scsi/hisi_sas/hisi_sas_v3_hw.c
1200
struct hisi_sas_sge *entry = &sge_page->sge[i];
drivers/scsi/hisi_sas/hisi_sas_v3_hw.c
1226
struct hisi_sas_sge *entry = &sge_dif_page->sge[i];
drivers/scsi/lpfc/lpfc_bsg.c
5049
receive_length = nembed_sge->sge[0].length;
drivers/scsi/lpfc/lpfc_bsg.c
5061
nembed_sge->sge[0].pa_hi =
drivers/scsi/lpfc/lpfc_bsg.c
5064
nembed_sge->sge[0].pa_lo =
drivers/scsi/lpfc/lpfc_hbadisc.c
2233
struct lpfc_mbx_sge sge;
drivers/scsi/lpfc/lpfc_hbadisc.c
2242
lpfc_sli4_mbx_sge_get(mboxq, 0, &sge);
drivers/scsi/lpfc/lpfc_hw4.h
1198
struct lpfc_mbx_sge sge[LPFC_SLI4_MBX_SGE_MAX_PAGES];
drivers/scsi/lpfc/lpfc_mbox.c
1744
nembed_sge->sge[sgentry].pa_lo = putPaddrLow(phyaddr);
drivers/scsi/lpfc/lpfc_mbox.c
1745
nembed_sge->sge[sgentry].pa_hi = putPaddrHigh(phyaddr);
drivers/scsi/lpfc/lpfc_mbox.c
1746
nembed_sge->sge[sgentry].length = length;
drivers/scsi/lpfc/lpfc_mbox.c
1760
struct lpfc_mbx_sge *sge)
drivers/scsi/lpfc/lpfc_mbox.c
1766
sge->pa_lo = nembed_sge->sge[sgentry].pa_lo;
drivers/scsi/lpfc/lpfc_mbox.c
1767
sge->pa_hi = nembed_sge->sge[sgentry].pa_hi;
drivers/scsi/lpfc/lpfc_mbox.c
1768
sge->length = nembed_sge->sge[sgentry].length;
drivers/scsi/lpfc/lpfc_mbox.c
1784
struct lpfc_mbx_sge sge;
drivers/scsi/lpfc/lpfc_mbox.c
1805
lpfc_sli4_mbx_sge_get(mbox, sgentry, &sge);
drivers/scsi/lpfc/lpfc_mbox.c
1806
phyaddr = getPaddr(sge.pa_hi, sge.pa_lo);
drivers/scsi/lpfc/lpfc_mbox.c
2086
struct lpfc_mbx_sge sge;
drivers/scsi/lpfc/lpfc_mbox.c
2112
lpfc_sli4_mbx_sge_get(mboxq, 0, &sge);
drivers/scsi/lpfc/lpfc_sli.c
20095
struct lpfc_mbx_sge sge;
drivers/scsi/lpfc/lpfc_sli.c
20126
lpfc_sli4_mbx_sge_get(mboxq, 0, &sge);
drivers/scsi/mpi3mr/mpi/mpi30_init.h
18
struct mpi3_sge_common sge;
drivers/scsi/myrs.c
1721
hw_sge->sge[0].sge_addr = (u64)sg_dma_address(sgl);
drivers/scsi/myrs.c
1722
hw_sge->sge[0].sge_count = (u64)sg_dma_len(sgl);
drivers/scsi/myrs.c
1755
hw_sgl = hw_sge->sge;
drivers/scsi/myrs.c
187
sgl->sge[0].sge_addr = ctlr_info_addr;
drivers/scsi/myrs.c
188
sgl->sge[0].sge_count = mbox->ctlr_info.dma_size;
drivers/scsi/myrs.c
246
sgl->sge[0].sge_addr = ldev_info_addr;
drivers/scsi/myrs.c
247
sgl->sge[0].sge_count = mbox->ldev_info.dma_size;
drivers/scsi/myrs.c
334
sgl->sge[0].sge_addr = pdev_info_addr;
drivers/scsi/myrs.c
335
sgl->sge[0].sge_count = mbox->pdev_info.dma_size;
drivers/scsi/myrs.c
405
sgl->sge[0].sge_addr = devmap_addr;
drivers/scsi/myrs.c
406
sgl->sge[0].sge_count = mbox->pdev_info.dma_size;
drivers/scsi/myrs.c
441
sgl->sge[0].sge_addr = event_addr;
drivers/scsi/myrs.c
442
sgl->sge[0].sge_count = mbox->get_event.dma_size;
drivers/scsi/myrs.c
469
sgl->sge[0].sge_addr = cs->fwstat_addr;
drivers/scsi/myrs.c
470
sgl->sge[0].sge_count = mbox->ctlr_info.dma_size;
drivers/scsi/myrs.h
670
struct myrs_sge sge[2]; /* Bytes 0-31 */
drivers/scsi/qedf/drv_scsi_fw_funcs.c
35
ctx_data_desc->sge[sge_index].sge_addr.lo = val;
drivers/scsi/qedf/drv_scsi_fw_funcs.c
37
ctx_data_desc->sge[sge_index].sge_addr.hi = val;
drivers/scsi/qedf/drv_scsi_fw_funcs.c
39
ctx_data_desc->sge[sge_index].sge_len = val;
drivers/scsi/qedi/qedi_fw_api.c
44
ctx_data_desc->sge[sge_index].sge_addr.lo = val;
drivers/scsi/qedi/qedi_fw_api.c
46
ctx_data_desc->sge[sge_index].sge_addr.hi = val;
drivers/scsi/qedi/qedi_fw_api.c
48
ctx_data_desc->sge[sge_index].sge_len = val;
drivers/scsi/qedi/qedi_iscsi.c
192
struct scsi_sge *sge;
drivers/scsi/qedi/qedi_iscsi.c
196
sizeof(*sge),
drivers/scsi/vmw_pvscsi.c
331
struct PVSCSISGElement *sge;
drivers/scsi/vmw_pvscsi.c
335
sge = &ctx->sgl->sge[0];
drivers/scsi/vmw_pvscsi.c
337
sge[i].addr = sg_dma_address(sg);
drivers/scsi/vmw_pvscsi.c
338
sge[i].length = sg_dma_len(sg);
drivers/scsi/vmw_pvscsi.c
339
sge[i].flags = 0;
drivers/scsi/vmw_pvscsi.c
50
struct PVSCSISGElement sge[PVSCSI_MAX_NUM_SG_ENTRIES_PER_SEGMENT];
drivers/vfio/pci/pds/cmds.c
206
struct pds_lm_sg_elem *sgl, *sge;
drivers/vfio/pci/pds/cmds.c
232
sge = sgl;
drivers/vfio/pci/pds/cmds.c
234
sge->addr = cpu_to_le64(sg_dma_address(sg));
drivers/vfio/pci/pds/cmds.c
235
sge->len = cpu_to_le32(sg_dma_len(sg));
drivers/vfio/pci/pds/cmds.c
236
dev_dbg(dev, "addr = %llx, len = %u\n", sge->addr, sge->len);
drivers/vfio/pci/pds/cmds.c
237
sge++;
fs/smb/client/smbdirect.c
1114
request->sge[0].addr = ib_dma_map_single(
fs/smb/client/smbdirect.c
1117
if (ib_dma_mapping_error(sc->ib.dev, request->sge[0].addr)) {
fs/smb/client/smbdirect.c
1123
request->sge[0].length = sizeof(*packet);
fs/smb/client/smbdirect.c
1124
request->sge[0].lkey = sc->ib.pd->local_dma_lkey;
fs/smb/client/smbdirect.c
1234
i, request->sge[i].addr, request->sge[i].length);
fs/smb/client/smbdirect.c
1237
request->sge[i].addr,
fs/smb/client/smbdirect.c
1238
request->sge[i].length,
fs/smb/client/smbdirect.c
1244
request->wr.sg_list = request->sge;
fs/smb/client/smbdirect.c
1482
memset(request->sge, 0, sizeof(request->sge));
fs/smb/client/smbdirect.c
1491
request->sge[0].addr = ib_dma_map_single(sc->ib.dev,
fs/smb/client/smbdirect.c
1495
if (ib_dma_mapping_error(sc->ib.dev, request->sge[0].addr)) {
fs/smb/client/smbdirect.c
1500
request->sge[0].length = header_length;
fs/smb/client/smbdirect.c
1501
request->sge[0].lkey = sc->ib.pd->local_dma_lkey;
fs/smb/client/smbdirect.c
1509
.sge = request->sge,
fs/smb/client/smbdirect.c
1629
response->sge.addr = ib_dma_map_single(
fs/smb/client/smbdirect.c
1632
if (ib_dma_mapping_error(sc->ib.dev, response->sge.addr))
fs/smb/client/smbdirect.c
1635
response->sge.length = sp->max_recv_size;
fs/smb/client/smbdirect.c
1636
response->sge.lkey = sc->ib.pd->local_dma_lkey;
fs/smb/client/smbdirect.c
1642
recv_wr.sg_list = &response->sge;
fs/smb/client/smbdirect.c
1647
ib_dma_unmap_single(sc->ib.dev, response->sge.addr,
fs/smb/client/smbdirect.c
1648
response->sge.length, DMA_FROM_DEVICE);
fs/smb/client/smbdirect.c
1649
response->sge.length = 0;
fs/smb/client/smbdirect.c
1670
rc, response->sge.addr,
fs/smb/client/smbdirect.c
1671
response->sge.length, response->sge.lkey);
fs/smb/client/smbdirect.c
1784
if (likely(response->sge.length != 0)) {
fs/smb/client/smbdirect.c
1786
response->sge.addr,
fs/smb/client/smbdirect.c
1787
response->sge.length,
fs/smb/client/smbdirect.c
1789
response->sge.length = 0;
fs/smb/client/smbdirect.c
1812
response->sge.length = 0;
fs/smb/client/smbdirect.c
3130
struct ib_sge *sge = &rdma->sge[rdma->nr_sge];
fs/smb/client/smbdirect.c
3138
sge->addr = addr;
fs/smb/client/smbdirect.c
3139
sge->length = len;
fs/smb/client/smbdirect.c
3140
sge->lkey = rdma->local_dma_lkey;
fs/smb/client/smbdirect.c
3335
struct ib_sge *sge = &rdma->sge[rdma->nr_sge--];
fs/smb/client/smbdirect.c
3337
ib_dma_unmap_single(rdma->device, sge->addr, sge->length,
fs/smb/client/smbdirect.c
3339
sge->addr = 0;
fs/smb/client/smbdirect.c
52
struct ib_sge *sge;
fs/smb/client/smbdirect.c
553
msg->sge[i].addr,
fs/smb/client/smbdirect.c
554
msg->sge[i].length,
fs/smb/client/smbdirect.c
793
response->sge.addr,
fs/smb/client/smbdirect.c
794
response->sge.length,
fs/smb/common/smbdirect/smbdirect_socket.h
474
struct ib_sge sge[SMBDIRECT_SEND_IO_MAX_SGE];
fs/smb/common/smbdirect/smbdirect_socket.h
516
struct ib_sge sge;
fs/smb/server/transport_rdma.c
1403
sendmsg->sge[0].addr = ib_dma_map_single(sc->ib.dev,
fs/smb/server/transport_rdma.c
1407
ret = ib_dma_mapping_error(sc->ib.dev, sendmsg->sge[0].addr);
fs/smb/server/transport_rdma.c
1414
sendmsg->sge[0].length = header_length;
fs/smb/server/transport_rdma.c
1415
sendmsg->sge[0].lkey = sc->ib.pd->local_dma_lkey;
fs/smb/server/transport_rdma.c
1471
msg->sge[i].addr, msg->sge[i].length,
fs/smb/server/transport_rdma.c
1476
msg->wr.sg_list = &msg->sge[0];
fs/smb/server/transport_rdma.c
1557
struct ib_sge *sge;
fs/smb/server/transport_rdma.c
1579
sge = &msg->sge[msg->num_sge];
fs/smb/server/transport_rdma.c
1580
sge->addr = sg_dma_address(&sg[j]);
fs/smb/server/transport_rdma.c
1581
sge->length = sg_dma_len(&sg[j]);
fs/smb/server/transport_rdma.c
1582
sge->lkey = sc->ib.pd->local_dma_lkey;
fs/smb/server/transport_rdma.c
186
if (likely(recvmsg->sge.length != 0)) {
fs/smb/server/transport_rdma.c
188
recvmsg->sge.addr,
fs/smb/server/transport_rdma.c
189
recvmsg->sge.length,
fs/smb/server/transport_rdma.c
191
recvmsg->sge.length = 0;
fs/smb/server/transport_rdma.c
2084
sendmsg->sge[0].addr = ib_dma_map_single(sc->ib.dev,
fs/smb/server/transport_rdma.c
2087
ret = ib_dma_mapping_error(sc->ib.dev, sendmsg->sge[0].addr);
fs/smb/server/transport_rdma.c
2094
sendmsg->sge[0].length = sizeof(*resp);
fs/smb/server/transport_rdma.c
2095
sendmsg->sge[0].lkey = sc->ib.pd->local_dma_lkey;
fs/smb/server/transport_rdma.c
2293
recvmsg->sge.length = 0;
fs/smb/server/transport_rdma.c
558
msg->sge[0].addr, msg->sge[0].length,
fs/smb/server/transport_rdma.c
562
msg->sge[i].addr, msg->sge[i].length,
fs/smb/server/transport_rdma.c
642
ib_dma_sync_single_for_cpu(wc->qp->device, recvmsg->sge.addr,
fs/smb/server/transport_rdma.c
643
recvmsg->sge.length, DMA_FROM_DEVICE);
fs/smb/server/transport_rdma.c
774
recv_io->sge.addr,
fs/smb/server/transport_rdma.c
775
recv_io->sge.length,
fs/smb/server/transport_rdma.c
871
recvmsg->sge.addr = ib_dma_map_single(sc->ib.dev,
fs/smb/server/transport_rdma.c
875
ret = ib_dma_mapping_error(sc->ib.dev, recvmsg->sge.addr);
fs/smb/server/transport_rdma.c
878
recvmsg->sge.length = sp->max_recv_size;
fs/smb/server/transport_rdma.c
879
recvmsg->sge.lkey = sc->ib.pd->local_dma_lkey;
fs/smb/server/transport_rdma.c
883
wr.sg_list = &recvmsg->sge;
fs/smb/server/transport_rdma.c
890
recvmsg->sge.addr, recvmsg->sge.length,
fs/smb/server/transport_rdma.c
892
recvmsg->sge.length = 0;
include/linux/qed/storage_common.h
70
struct scsi_sge sge[4];
include/linux/skmsg.h
247
struct scatterlist *sge = sk_msg_elem(msg, msg->sg.start);
include/linux/skmsg.h
253
msg->data = sg_virt(sge);
include/linux/skmsg.h
254
msg->data_end = msg->data + sge->length;
include/linux/skmsg.h
261
struct scatterlist *sge;
include/linux/skmsg.h
264
sge = sk_msg_elem(msg, msg->sg.end);
include/linux/skmsg.h
265
sg_set_page(sge, page, len, offset);
include/linux/skmsg.h
266
sg_unmark_end(sge);
include/net/mana/hw_channel.h
121
struct gdma_sge sge;
include/rdma/rdma_vt.h
524
int rvt_rkey_ok(struct rvt_qp *qp, struct rvt_sge *sge,
include/rdma/rdma_vt.h
528
struct ib_sge *sge, int acc);
include/rdma/rdmavt_mr.h
102
static inline u32 rvt_get_sge_length(struct rvt_sge *sge, u32 length)
include/rdma/rdmavt_mr.h
104
u32 len = sge->length;
include/rdma/rdmavt_mr.h
108
if (len > sge->sge_length)
include/rdma/rdmavt_mr.h
109
len = sge->sge_length;
include/rdma/rdmavt_mr.h
117
struct rvt_sge *sge = &ss->sge;
include/rdma/rdmavt_mr.h
119
sge->vaddr += length;
include/rdma/rdmavt_mr.h
120
sge->length -= length;
include/rdma/rdmavt_mr.h
121
sge->sge_length -= length;
include/rdma/rdmavt_mr.h
122
if (sge->sge_length == 0) {
include/rdma/rdmavt_mr.h
124
rvt_put_mr(sge->mr);
include/rdma/rdmavt_mr.h
126
*sge = *ss->sg_list++;
include/rdma/rdmavt_mr.h
127
} else if (sge->length == 0 && sge->mr->lkey) {
include/rdma/rdmavt_mr.h
128
if (++sge->n >= RVT_SEGSZ) {
include/rdma/rdmavt_mr.h
129
if (++sge->m >= sge->mr->mapsz)
include/rdma/rdmavt_mr.h
131
sge->n = 0;
include/rdma/rdmavt_mr.h
133
sge->vaddr = sge->mr->map[sge->m]->segs[sge->n].vaddr;
include/rdma/rdmavt_mr.h
134
sge->length = sge->mr->map[sge->m]->segs[sge->n].length;
include/rdma/rdmavt_mr.h
141
struct rvt_sge *sge = &ss->sge;
include/rdma/rdmavt_mr.h
144
u32 len = rvt_get_sge_length(sge, length);
include/rdma/rdmavt_mr.h
78
struct rvt_sge sge; /* progress state for the current SGE */
include/rdma/rdmavt_mr.h
96
rvt_put_mr(ss->sge.mr);
include/rdma/rdmavt_mr.h
98
ss->sge = *ss->sg_list++;
include/rdma/rdmavt_qp.h
592
struct rvt_sge *sge = &wqe->sg_list[i];
include/rdma/rdmavt_qp.h
594
rvt_put_mr(sge->mr);
include/rdma/rw.h
25
struct ib_sge sge;
include/rdma/rw.h
38
struct ib_sge sge;
include/rdma/rw.h
45
struct ib_sge sge;
include/uapi/rdma/rdma_user_rxe.h
157
__DECLARE_FLEX_ARRAY(struct rxe_sge, sge);
include/uapi/rdma/siw-abi.h
119
struct siw_sge sge[SIW_MAX_SGE];
include/uapi/rdma/siw-abi.h
135
struct siw_sge sge[SIW_MAX_SGE];
net/9p/trans_rdma.c
275
struct ib_sge sge;
net/9p/trans_rdma.c
286
sge.addr = c->busa;
net/9p/trans_rdma.c
287
sge.length = client->msize;
net/9p/trans_rdma.c
288
sge.lkey = rdma->pd->local_dma_lkey;
net/9p/trans_rdma.c
292
wr.sg_list = &sge;
net/9p/trans_rdma.c
310
struct ib_sge sge;
net/9p/trans_rdma.c
383
sge.addr = c->busa;
net/9p/trans_rdma.c
384
sge.length = c->req->tc.size;
net/9p/trans_rdma.c
385
sge.lkey = rdma->pd->local_dma_lkey;
net/9p/trans_rdma.c
391
wr.sg_list = &sge;
net/core/filter.c
2666
struct scatterlist *sge;
net/core/filter.c
2723
sge = sk_msg_elem(msg, i);
net/core/filter.c
2724
from = sg_virt(sge);
net/core/filter.c
2725
len = sge->length;
net/core/filter.c
2730
sge->length = 0;
net/core/filter.c
2731
put_page(sg_page(sge));
net/core/filter.c
2791
struct scatterlist sge, nsge, nnsge, rsge = {0}, *psge;
net/core/filter.c
2888
sge = sk_msg_elem_cpy(msg, new);
net/core/filter.c
2889
sg_unmark_end(&sge);
net/core/filter.c
2899
msg->sg.data[i] = sge;
net/core/filter.c
2900
sge = nsge;
net/core/filter.c
2939
struct scatterlist *sge = sk_msg_elem(msg, i);
net/core/filter.c
2942
put_page(sg_page(sge));
net/core/filter.c
2954
struct scatterlist tmp, sge;
net/core/filter.c
2957
sge = sk_msg_elem_cpy(msg, i);
net/core/filter.c
2962
msg->sg.data[i] = sge;
net/core/filter.c
2964
sge = tmp;
net/core/filter.c
3022
struct scatterlist *nsge, *sge = sk_msg_elem(msg, i);
net/core/filter.c
3024
int b = sge->length - pop - a;
net/core/filter.c
3030
sge->length = a;
net/core/filter.c
3033
get_page(sg_page(sge));
net/core/filter.c
3035
sg_page(sge),
net/core/filter.c
3036
b, sge->offset + pop + a);
net/core/filter.c
3047
orig = sg_page(sge);
net/core/filter.c
3048
from = sg_virt(sge);
net/core/filter.c
3052
sg_set_page(sge, page, a + b, 0);
net/core/filter.c
3057
pop -= (sge->length - a);
net/core/filter.c
3058
sge->length = a;
net/core/filter.c
3080
struct scatterlist *sge = sk_msg_elem(msg, i);
net/core/filter.c
3082
if (pop < sge->length) {
net/core/filter.c
3083
sge->length -= pop;
net/core/filter.c
3084
sge->offset += pop;
net/core/filter.c
3087
pop -= sge->length;
net/core/skmsg.c
102
sge = sk_msg_elem(src, i);
net/core/skmsg.c
106
sge_len = sge->length - off;
net/core/skmsg.c
114
(sg_page(sge) == sg_page(sgd)) &&
net/core/skmsg.c
115
(sg_virt(sge) + off == sg_virt(sgd) + sgd->length)) {
net/core/skmsg.c
119
sge_off = sge->offset + off;
net/core/skmsg.c
120
sk_msg_page_add(dst, sg_page(sge), sge_len, sge_off);
net/core/skmsg.c
131
sge = sk_msg_elem(src, i);
net/core/skmsg.c
143
struct scatterlist *sge = sk_msg_elem(msg, i);
net/core/skmsg.c
145
if (bytes < sge->length) {
net/core/skmsg.c
146
sge->length -= bytes;
net/core/skmsg.c
147
sge->offset += bytes;
net/core/skmsg.c
152
sk_mem_uncharge(sk, sge->length);
net/core/skmsg.c
153
bytes -= sge->length;
net/core/skmsg.c
154
sge->length = 0;
net/core/skmsg.c
155
sge->offset = 0;
net/core/skmsg.c
167
struct scatterlist *sge = &msg->sg.data[i];
net/core/skmsg.c
168
int uncharge = (bytes < sge->length) ? bytes : sge->length;
net/core/skmsg.c
180
struct scatterlist *sge = sk_msg_elem(msg, i);
net/core/skmsg.c
181
u32 len = sge->length;
net/core/skmsg.c
187
put_page(sg_page(sge));
net/core/skmsg.c
189
memset(sge, 0, sizeof(*sge));
net/core/skmsg.c
196
struct scatterlist *sge = sk_msg_elem(msg, i);
net/core/skmsg.c
200
msg->sg.size -= sge->length;
net/core/skmsg.c
204
sge = sk_msg_elem(msg, i);
net/core/skmsg.c
226
struct scatterlist *sge;
net/core/skmsg.c
230
sge = sk_msg_elem(msg, i);
net/core/skmsg.c
231
if (!sge->length)
net/core/skmsg.c
233
if (bytes < sge->length) {
net/core/skmsg.c
236
sge->length -= bytes;
net/core/skmsg.c
237
sge->offset += bytes;
net/core/skmsg.c
242
msg->sg.size -= sge->length;
net/core/skmsg.c
243
bytes -= sge->length;
net/core/skmsg.c
36
struct scatterlist *sge;
net/core/skmsg.c
373
struct scatterlist *sge;
net/core/skmsg.c
377
sge = sk_msg_elem(msg, i);
net/core/skmsg.c
379
if (msg->sg.copybreak >= sge->length) {
net/core/skmsg.c
384
sge = sk_msg_elem(msg, i);
net/core/skmsg.c
387
buf_size = sge->length - msg->sg.copybreak;
net/core/skmsg.c
389
to = sg_virt(sge) + msg->sg.copybreak;
net/core/skmsg.c
426
struct scatterlist *sge;
net/core/skmsg.c
437
sge = sk_msg_elem(msg_rx, i);
net/core/skmsg.c
438
copy = sge->length;
net/core/skmsg.c
439
page = sg_page(sge);
net/core/skmsg.c
443
copy = copy_page_to_iter(page, sge->offset, copy, iter);
net/core/skmsg.c
454
sge->offset += copy;
net/core/skmsg.c
455
sge->length -= copy;
net/core/skmsg.c
463
if (!sge->length) {
net/core/skmsg.c
472
if (copy != sge->length)
net/core/skmsg.c
479
} while ((i != msg_rx->sg.end) && !sg_is_last(sge));
net/core/skmsg.c
489
if (!sge->length && (i == msg_rx->sg.end || sg_is_last(sge))) {
net/core/skmsg.c
54
sge = &msg->sg.data[i];
net/core/skmsg.c
57
sg_page(sge) == pfrag->page &&
net/core/skmsg.c
58
sge->offset + sge->length == orig_offset) {
net/core/skmsg.c
59
sge->length += use;
net/core/skmsg.c
66
sge = &msg->sg.data[msg->sg.end];
net/core/skmsg.c
67
sg_unmark_end(sge);
net/core/skmsg.c
68
sg_set_page(sge, pfrag->page, use, orig_offset);
net/core/skmsg.c
91
struct scatterlist *sge = sk_msg_elem(src, i);
net/core/skmsg.c
96
if (sge->length > off)
net/core/skmsg.c
98
off -= sge->length;
net/ipv4/tcp_bpf.c
105
sge = sk_msg_elem(msg, msg->sg.start);
net/ipv4/tcp_bpf.c
106
size = (apply && apply_bytes < sge->length) ?
net/ipv4/tcp_bpf.c
107
apply_bytes : sge->length;
net/ipv4/tcp_bpf.c
108
off = sge->offset;
net/ipv4/tcp_bpf.c
109
page = sg_page(sge);
net/ipv4/tcp_bpf.c
118
if (size < sge->length && msg->sg.start != msg->sg.end)
net/ipv4/tcp_bpf.c
130
sge->offset += ret;
net/ipv4/tcp_bpf.c
131
sge->length -= ret;
net/ipv4/tcp_bpf.c
139
if (!sge->length) {
net/ipv4/tcp_bpf.c
142
sg_init_table(sge, 1);
net/ipv4/tcp_bpf.c
205
struct scatterlist *sge;
net/ipv4/tcp_bpf.c
211
sge = sk_msg_elem(msg_rx, i);
net/ipv4/tcp_bpf.c
212
if (!sge->length) {
net/ipv4/tcp_bpf.c
37
struct scatterlist *sge;
net/ipv4/tcp_bpf.c
50
sge = sk_msg_elem(msg, i);
net/ipv4/tcp_bpf.c
51
size = (apply && apply_bytes < sge->length) ?
net/ipv4/tcp_bpf.c
52
apply_bytes : sge->length;
net/ipv4/tcp_bpf.c
63
if (sge->length)
net/ipv4/tcp_bpf.c
70
if (sge->length)
net/ipv4/tcp_bpf.c
96
struct scatterlist *sge;
net/rds/ib_rdma.c
568
struct ib_sge sge = {};
net/rds/ib_rdma.c
597
sge.addr = virt_addr;
net/rds/ib_rdma.c
598
sge.length = length;
net/rds/ib_rdma.c
599
sge.lkey = ib_mr->lkey;
net/rds/ib_rdma.c
603
IB_UVERBS_ADVISE_MR_FLAG_FLUSH, &sge, 1);
net/rds/ib_recv.c
312
struct ib_sge *sge;
net/rds/ib_recv.c
346
sge = &recv->r_sge[0];
net/rds/ib_recv.c
347
sge->addr = ic->i_recv_hdrs_dma[recv - ic->i_recvs];
net/rds/ib_recv.c
348
sge->length = sizeof(struct rds_header);
net/rds/ib_recv.c
350
sge = &recv->r_sge[1];
net/rds/ib_recv.c
351
sge->addr = sg_dma_address(&recv->r_frag->f_sg);
net/rds/ib_recv.c
352
sge->length = sg_dma_len(&recv->r_frag->f_sg);
net/rds/ib_recv.c
54
struct ib_sge *sge;
net/rds/ib_recv.c
578
struct ib_sge *sge = &ic->i_ack_sge;
net/rds/ib_recv.c
580
sge->addr = ic->i_ack_dma;
net/rds/ib_recv.c
581
sge->length = sizeof(struct rds_header);
net/rds/ib_recv.c
582
sge->lkey = ic->i_pd->local_dma_lkey;
net/rds/ib_recv.c
584
wr->sg_list = sge;
net/rds/ib_recv.c
64
sge = &recv->r_sge[0];
net/rds/ib_recv.c
65
sge->addr = ic->i_recv_hdrs_dma[i];
net/rds/ib_recv.c
66
sge->length = sizeof(struct rds_header);
net/rds/ib_recv.c
67
sge->lkey = ic->i_pd->local_dma_lkey;
net/rds/ib_recv.c
69
sge = &recv->r_sge[1];
net/rds/ib_recv.c
70
sge->addr = 0;
net/rds/ib_recv.c
71
sge->length = RDS_FRAG_SIZE;
net/rds/ib_recv.c
72
sge->lkey = ic->i_pd->local_dma_lkey;
net/rds/ib_send.c
196
struct ib_sge *sge;
net/rds/ib_send.c
204
sge = &send->s_sge[0];
net/rds/ib_send.c
205
sge->addr = ic->i_send_hdrs_dma[i];
net/rds/ib_send.c
207
sge->length = sizeof(struct rds_header);
net/rds/ib_send.c
208
sge->lkey = ic->i_pd->local_dma_lkey;
net/smc/smc_tx.c
374
struct ib_sge *sge = wr->wr.sg_list;
net/smc/smc_tx.c
386
sge[srcchunk].addr = conn->sndbuf_desc->is_vm ?
net/smc/smc_tx.c
388
sge[srcchunk].length = src_len;
net/smc/smc_tx.c
390
sge[srcchunk].lkey =
net/sunrpc/xprtrdma/rpc_rdma.c
537
struct ib_sge *sge;
net/sunrpc/xprtrdma/rpc_rdma.c
546
for (sge = &sc->sc_sges[2]; sc->sc_unmap_count;
net/sunrpc/xprtrdma/rpc_rdma.c
547
++sge, --sc->sc_unmap_count)
net/sunrpc/xprtrdma/rpc_rdma.c
548
ib_dma_unmap_page(rdmab_device(rb), sge->addr, sge->length,
net/sunrpc/xprtrdma/rpc_rdma.c
561
struct ib_sge *sge = &sc->sc_sges[req->rl_wr.num_sge++];
net/sunrpc/xprtrdma/rpc_rdma.c
563
sge->addr = rdmab_addr(rb);
net/sunrpc/xprtrdma/rpc_rdma.c
564
sge->length = len;
net/sunrpc/xprtrdma/rpc_rdma.c
565
sge->lkey = rdmab_lkey(rb);
net/sunrpc/xprtrdma/rpc_rdma.c
567
ib_dma_sync_single_for_device(rdmab_device(rb), sge->addr, sge->length,
net/sunrpc/xprtrdma/rpc_rdma.c
578
struct ib_sge *sge = &sc->sc_sges[req->rl_wr.num_sge++];
net/sunrpc/xprtrdma/rpc_rdma.c
584
sge->addr = rdmab_addr(rb);
net/sunrpc/xprtrdma/rpc_rdma.c
585
sge->length = len;
net/sunrpc/xprtrdma/rpc_rdma.c
586
sge->lkey = rdmab_lkey(rb);
net/sunrpc/xprtrdma/rpc_rdma.c
588
ib_dma_sync_single_for_device(rdmab_device(rb), sge->addr, sge->length,
net/sunrpc/xprtrdma/rpc_rdma.c
603
struct ib_sge *sge;
net/sunrpc/xprtrdma/rpc_rdma.c
609
sge = &sc->sc_sges[req->rl_wr.num_sge++];
net/sunrpc/xprtrdma/rpc_rdma.c
611
sge->addr = ib_dma_map_page(rdmab_device(rb), *ppages,
net/sunrpc/xprtrdma/rpc_rdma.c
613
if (ib_dma_mapping_error(rdmab_device(rb), sge->addr))
net/sunrpc/xprtrdma/rpc_rdma.c
616
sge->length = len;
net/sunrpc/xprtrdma/rpc_rdma.c
617
sge->lkey = rdmab_lkey(rb);
net/sunrpc/xprtrdma/rpc_rdma.c
628
trace_xprtrdma_dma_maperr(sge->addr);
net/sunrpc/xprtrdma/rpc_rdma.c
641
struct ib_sge *sge = &sc->sc_sges[req->rl_wr.num_sge++];
net/sunrpc/xprtrdma/rpc_rdma.c
645
sge->addr = ib_dma_map_page(rdmab_device(rb), page, page_base, len,
net/sunrpc/xprtrdma/rpc_rdma.c
647
if (ib_dma_mapping_error(rdmab_device(rb), sge->addr))
net/sunrpc/xprtrdma/rpc_rdma.c
650
sge->length = len;
net/sunrpc/xprtrdma/rpc_rdma.c
651
sge->lkey = rdmab_lkey(rb);
net/sunrpc/xprtrdma/rpc_rdma.c
656
trace_xprtrdma_dma_maperr(sge->addr);
net/tls/tls_sw.c
473
struct scatterlist *sge;
net/tls/tls_sw.c
487
sge = sk_msg_elem(msg_en, msg_en->sg.curr);
net/tls/tls_sw.c
488
sge->offset -= prot->prepend_size;
net/tls/tls_sw.c
489
sge->length += prot->prepend_size;
net/tls/tls_sw.c
543
struct scatterlist *sge = sk_msg_elem(msg_en, start);
net/tls/tls_sw.c
564
sge->offset += prot->prepend_size;
net/tls/tls_sw.c
565
sge->length -= prot->prepend_size;
net/tls/tls_sw.c
600
sge->offset -= prot->prepend_size;
net/tls/tls_sw.c
601
sge->length += prot->prepend_size;
net/tls/tls_sw.c
623
struct scatterlist *sge, *osge, *nsge;
net/tls/tls_sw.c
642
sge = sk_msg_elem(msg_opl, i);
net/tls/tls_sw.c
643
while (apply && sge->length) {
net/tls/tls_sw.c
644
if (sge->length > apply) {
net/tls/tls_sw.c
645
u32 len = sge->length - apply;
net/tls/tls_sw.c
647
get_page(sg_page(sge));
net/tls/tls_sw.c
648
sg_set_page(&tmp, sg_page(sge), len,
net/tls/tls_sw.c
649
sge->offset + apply);
net/tls/tls_sw.c
650
sge->length = apply;
net/tls/tls_sw.c
654
apply -= sge->length;
net/tls/tls_sw.c
655
bytes += sge->length;
net/tls/tls_sw.c
661
sge = sk_msg_elem(msg_opl, i);