arch/mips/include/asm/octeon/cvmx-lmcx-defs.h
1401
uint64_t dq:5;
arch/mips/include/asm/octeon/cvmx-lmcx-defs.h
1407
uint64_t dq:5;
arch/mips/include/asm/octeon/cvmx-lmcx-defs.h
1414
uint64_t dq:4;
arch/mips/include/asm/octeon/cvmx-lmcx-defs.h
1424
uint64_t dq:4;
crypto/async_tx/async_raid6_recov.c
208
struct page *p, *q, *g, *dp, *dq;
crypto/async_tx/async_raid6_recov.c
244
dq = blocks[failb];
crypto/async_tx/async_raid6_recov.c
250
tx = async_mult(dq, dq_off, g, g_off,
crypto/async_tx/async_raid6_recov.c
263
srcs[0] = dq;
crypto/async_tx/async_raid6_recov.c
269
tx = async_xor_offs(dq, dq_off, srcs, src_offs, 2, bytes, submit);
crypto/async_tx/async_raid6_recov.c
274
srcs[1] = dq;
crypto/async_tx/async_raid6_recov.c
279
tx = async_sum_product(dq, dq_off, srcs, src_offs, coef, bytes, submit);
crypto/async_tx/async_raid6_recov.c
284
srcs[1] = dq;
crypto/async_tx/async_raid6_recov.c
299
struct page *p, *q, *dp, *dq;
crypto/async_tx/async_raid6_recov.c
323
dq = blocks[failb];
crypto/async_tx/async_raid6_recov.c
326
blocks[disks-1] = dq;
crypto/async_tx/async_raid6_recov.c
335
blocks[failb] = dq;
crypto/async_tx/async_raid6_recov.c
352
srcs[0] = dq;
crypto/async_tx/async_raid6_recov.c
358
tx = async_xor_offs(dq, dq_off, srcs, src_offs, 2, bytes, submit);
crypto/async_tx/async_raid6_recov.c
363
srcs[1] = dq;
crypto/async_tx/async_raid6_recov.c
368
tx = async_sum_product(dq, dq_off, srcs, src_offs, coef, bytes, submit);
crypto/async_tx/async_raid6_recov.c
373
srcs[1] = dq;
crypto/async_tx/async_raid6_recov.c
476
struct page *p, *q, *dq;
crypto/async_tx/async_raid6_recov.c
533
dq = blocks[faila];
crypto/async_tx/async_raid6_recov.c
536
blocks[disks-1] = dq;
crypto/async_tx/async_raid6_recov.c
552
tx = async_mult(dq, dq_off, g, g_off,
crypto/async_tx/async_raid6_recov.c
561
blocks[faila] = dq;
crypto/async_tx/async_raid6_recov.c
569
srcs[0] = dq;
crypto/async_tx/async_raid6_recov.c
575
tx = async_xor_offs(dq, dq_off, srcs, src_offs, 2, bytes, submit);
crypto/async_tx/async_raid6_recov.c
578
tx = async_mult(dq, dq_off, dq, dq_off, coef, bytes, submit);
crypto/async_tx/async_raid6_recov.c
582
srcs[1] = dq;
crypto/rsa.c
205
mpi_free(key->dq);
crypto/rsa.c
213
key->dq = NULL;
crypto/rsa.c
23
MPI dq;
crypto/rsa.c
346
mpi_key->dq = mpi_read_raw_data(raw_key.dq, raw_key.dq_sz);
crypto/rsa.c
347
if (!mpi_key->dq)
crypto/rsa.c
96
ret = mpi_powm(m2, c, key->dq, key->q);
crypto/rsa_helper.c
130
key->dq = value;
drivers/ata/sata_nv.c
1687
struct defer_queue *dq = &pp->defer_queue;
drivers/ata/sata_nv.c
1690
WARN_ON(dq->tail - dq->head == ATA_MAX_QUEUE);
drivers/ata/sata_nv.c
1691
dq->defer_bits |= (1 << qc->hw_tag);
drivers/ata/sata_nv.c
1692
dq->tag[dq->tail++ & (ATA_MAX_QUEUE - 1)] = qc->hw_tag;
drivers/ata/sata_nv.c
1698
struct defer_queue *dq = &pp->defer_queue;
drivers/ata/sata_nv.c
1701
if (dq->head == dq->tail) /* null queue */
drivers/ata/sata_nv.c
1704
tag = dq->tag[dq->head & (ATA_MAX_QUEUE - 1)];
drivers/ata/sata_nv.c
1705
dq->tag[dq->head++ & (ATA_MAX_QUEUE - 1)] = ATA_TAG_POISON;
drivers/ata/sata_nv.c
1706
WARN_ON(!(dq->defer_bits & (1 << tag)));
drivers/ata/sata_nv.c
1707
dq->defer_bits &= ~(1 << tag);
drivers/ata/sata_nv.c
1725
struct defer_queue *dq = &pp->defer_queue;
drivers/ata/sata_nv.c
1727
dq->head = 0;
drivers/ata/sata_nv.c
1728
dq->tail = 0;
drivers/ata/sata_nv.c
1729
dq->defer_bits = 0;
drivers/crypto/caam/caamalg_qi2.c
4880
struct dpaa2_dq *dq;
drivers/crypto/caam/caamalg_qi2.c
4884
dq = dpaa2_io_store_next(ppriv->store, &is_last);
drivers/crypto/caam/caamalg_qi2.c
4885
if (unlikely(!dq)) {
drivers/crypto/caam/caamalg_qi2.c
4901
dpaa2_caam_process_fd(ppriv->priv, dpaa2_dq_fd(dq));
drivers/crypto/caam/caampkc.c
1021
rsa_key->dq = caam_read_rsa_crt(raw_key->dq, raw_key->dq_sz, q_sz);
drivers/crypto/caam/caampkc.c
1022
if (!rsa_key->dq)
drivers/crypto/caam/caampkc.c
1035
kfree_sensitive(rsa_key->dq);
drivers/crypto/caam/caampkc.c
604
pdb->dq_dma = dma_map_single(dev, key->dq, q_sz, DMA_TO_DEVICE);
drivers/crypto/caam/caampkc.c
872
kfree_sensitive(key->dq);
drivers/crypto/caam/caampkc.h
76
u8 *dq;
drivers/crypto/hisilicon/hpre/hpre_crypto.c
947
rsa_key->dq, rsa_key->dq_sz);
drivers/crypto/intel/qat/qat_common/qat_asym_algs.c
1119
ptr = rsa_key->dq;
drivers/crypto/intel/qat/qat_common/qat_asym_algs.c
1124
ctx->dq = dma_alloc_coherent(dev, half_key_sz, &ctx->dma_dq,
drivers/crypto/intel/qat/qat_common/qat_asym_algs.c
1126
if (!ctx->dq)
drivers/crypto/intel/qat/qat_common/qat_asym_algs.c
1128
memcpy(ctx->dq + (half_key_sz - len), ptr, len);
drivers/crypto/intel/qat/qat_common/qat_asym_algs.c
1146
memset(ctx->dq, '\0', half_key_sz);
drivers/crypto/intel/qat/qat_common/qat_asym_algs.c
1147
dma_free_coherent(dev, half_key_sz, ctx->dq, ctx->dma_dq);
drivers/crypto/intel/qat/qat_common/qat_asym_algs.c
1148
ctx->dq = NULL;
drivers/crypto/intel/qat/qat_common/qat_asym_algs.c
1190
if (ctx->dq) {
drivers/crypto/intel/qat/qat_common/qat_asym_algs.c
1191
memset(ctx->dq, '\0', half_key_sz);
drivers/crypto/intel/qat/qat_common/qat_asym_algs.c
1192
dma_free_coherent(dev, half_key_sz, ctx->dq, ctx->dma_dq);
drivers/crypto/intel/qat/qat_common/qat_asym_algs.c
1205
ctx->dq = NULL;
drivers/crypto/intel/qat/qat_common/qat_asym_algs.c
40
dma_addr_t dq;
drivers/crypto/intel/qat/qat_common/qat_asym_algs.c
66
char *dq;
drivers/crypto/intel/qat/qat_common/qat_asym_algs.c
877
qat_req->in.rsa.dec_crt.dq = ctx->dma_dq;
drivers/crypto/virtio/virtio_crypto_core.c
97
struct data_queue *dq = &vcrypto->data_vq[vq->index];
drivers/crypto/virtio/virtio_crypto_core.c
99
tasklet_schedule(&dq->done_task);
drivers/dma/fsl-dpaa2-qdma/dpaa2-qdma.c
400
struct dpaa2_dq *dq;
drivers/dma/fsl-dpaa2-qdma/dpaa2-qdma.c
414
dq = dpaa2_io_store_next(ppriv->store, &is_last);
drivers/dma/fsl-dpaa2-qdma/dpaa2-qdma.c
415
} while (!is_last && !dq);
drivers/dma/fsl-dpaa2-qdma/dpaa2-qdma.c
416
if (!dq) {
drivers/dma/fsl-dpaa2-qdma/dpaa2-qdma.c
422
fd = dpaa2_dq_fd(dq);
drivers/media/common/saa7146/saa7146_vbi.c
293
struct saa7146_dmaqueue *dq = &dev->vv_data->vbi_dmaq;
drivers/media/common/saa7146/saa7146_vbi.c
296
if (dq->curr) {
drivers/media/common/saa7146/saa7146_vbi.c
297
buf = dq->curr;
drivers/media/common/saa7146/saa7146_vbi.c
298
dq->curr = NULL;
drivers/media/common/saa7146/saa7146_vbi.c
301
while (!list_empty(&dq->queue)) {
drivers/media/common/saa7146/saa7146_vbi.c
302
buf = list_entry(dq->queue.next, struct saa7146_buf, list);
drivers/media/common/saa7146/saa7146_video.c
638
struct saa7146_dmaqueue *dq = &dev->vv_data->video_dmaq;
drivers/media/common/saa7146/saa7146_video.c
641
if (dq->curr) {
drivers/media/common/saa7146/saa7146_video.c
642
buf = dq->curr;
drivers/media/common/saa7146/saa7146_video.c
643
dq->curr = NULL;
drivers/media/common/saa7146/saa7146_video.c
646
while (!list_empty(&dq->queue)) {
drivers/media/common/saa7146/saa7146_video.c
647
buf = list_entry(dq->queue.next, struct saa7146_buf, list);
drivers/media/common/saa7146/saa7146_video.c
669
struct saa7146_dmaqueue *dq = &dev->vv_data->video_dmaq;
drivers/media/common/saa7146/saa7146_video.c
671
timer_delete(&dq->timeout);
drivers/net/ethernet/cavium/liquidio/cn23xx_pf_regs.h
366
#define CN23XX_DMA_CNT(dq) \
drivers/net/ethernet/cavium/liquidio/cn23xx_pf_regs.h
367
(CN23XX_DMA_CNT_START + ((dq) * CN23XX_DMA_OFFSET))
drivers/net/ethernet/cavium/liquidio/cn23xx_pf_regs.h
369
#define CN23XX_DMA_INT_LEVEL(dq) \
drivers/net/ethernet/cavium/liquidio/cn23xx_pf_regs.h
370
(CN23XX_DMA_INT_LEVEL_START + ((dq) * CN23XX_DMA_OFFSET))
drivers/net/ethernet/cavium/liquidio/cn23xx_pf_regs.h
372
#define CN23XX_DMA_PKT_INT_LEVEL(dq) \
drivers/net/ethernet/cavium/liquidio/cn23xx_pf_regs.h
373
(CN23XX_DMA_INT_LEVEL_START + ((dq) * CN23XX_DMA_OFFSET))
drivers/net/ethernet/cavium/liquidio/cn23xx_pf_regs.h
375
#define CN23XX_DMA_TIME_INT_LEVEL(dq) \
drivers/net/ethernet/cavium/liquidio/cn23xx_pf_regs.h
376
(CN23XX_DMA_INT_LEVEL_START + 4 + ((dq) * CN23XX_DMA_OFFSET))
drivers/net/ethernet/cavium/liquidio/cn23xx_pf_regs.h
378
#define CN23XX_DMA_TIM(dq) \
drivers/net/ethernet/cavium/liquidio/cn23xx_pf_regs.h
379
(CN23XX_DMA_TIM_START + ((dq) * CN23XX_DMA_OFFSET))
drivers/net/ethernet/cavium/liquidio/cn66xx_regs.h
311
#define CN6XXX_DMA_CNT(dq) \
drivers/net/ethernet/cavium/liquidio/cn66xx_regs.h
312
(CN6XXX_DMA_CNT_START + ((dq) * CN6XXX_DMA_OFFSET))
drivers/net/ethernet/cavium/liquidio/cn66xx_regs.h
314
#define CN6XXX_DMA_INT_LEVEL(dq) \
drivers/net/ethernet/cavium/liquidio/cn66xx_regs.h
315
(CN6XXX_DMA_INT_LEVEL_START + ((dq) * CN6XXX_DMA_OFFSET))
drivers/net/ethernet/cavium/liquidio/cn66xx_regs.h
317
#define CN6XXX_DMA_PKT_INT_LEVEL(dq) \
drivers/net/ethernet/cavium/liquidio/cn66xx_regs.h
318
(CN6XXX_DMA_INT_LEVEL_START + ((dq) * CN6XXX_DMA_OFFSET))
drivers/net/ethernet/cavium/liquidio/cn66xx_regs.h
320
#define CN6XXX_DMA_TIME_INT_LEVEL(dq) \
drivers/net/ethernet/cavium/liquidio/cn66xx_regs.h
321
(CN6XXX_DMA_INT_LEVEL_START + 4 + ((dq) * CN6XXX_DMA_OFFSET))
drivers/net/ethernet/cavium/liquidio/cn66xx_regs.h
323
#define CN6XXX_DMA_TIM(dq) \
drivers/net/ethernet/cavium/liquidio/cn66xx_regs.h
324
(CN6XXX_DMA_TIM_START + ((dq) * CN6XXX_DMA_OFFSET))
drivers/net/ethernet/freescale/dpaa/dpaa_eth.c
2492
const struct qm_dqrr_entry *dq,
drivers/net/ethernet/freescale/dpaa/dpaa_eth.c
2503
dpaa_bp = dpaa_bpid2pool(dq->fd.bpid);
drivers/net/ethernet/freescale/dpaa/dpaa_eth.c
2513
dpaa_rx_error(net_dev, priv, percpu_priv, &dq->fd, fq->fqid);
drivers/net/ethernet/freescale/dpaa/dpaa_eth.c
2682
const struct qm_dqrr_entry *dq,
drivers/net/ethernet/freescale/dpaa/dpaa_eth.c
2690
const struct qm_fd *fd = &dq->fd;
drivers/net/ethernet/freescale/dpaa/dpaa_eth.c
2712
dpaa_bp = dpaa_bpid2pool(dq->fd.bpid);
drivers/net/ethernet/freescale/dpaa/dpaa_eth.c
2717
trace_dpaa_rx_fd(net_dev, fq, &dq->fd);
drivers/net/ethernet/freescale/dpaa/dpaa_eth.c
2732
dpaa_fd_release(net_dev, &dq->fd);
drivers/net/ethernet/freescale/dpaa/dpaa_eth.c
2841
const struct qm_dqrr_entry *dq,
drivers/net/ethernet/freescale/dpaa/dpaa_eth.c
2856
dpaa_tx_error(net_dev, priv, percpu_priv, &dq->fd, fq->fqid);
drivers/net/ethernet/freescale/dpaa/dpaa_eth.c
2863
const struct qm_dqrr_entry *dq,
drivers/net/ethernet/freescale/dpaa/dpaa_eth.c
2874
trace_dpaa_tx_conf_fd(net_dev, fq, &dq->fd);
drivers/net/ethernet/freescale/dpaa/dpaa_eth.c
2881
dpaa_tx_conf(net_dev, priv, percpu_priv, &dq->fd, fq->fqid);
drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c
715
struct dpaa2_dq *dq;
drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c
721
dq = dpaa2_io_store_next(ch->store, &is_last);
drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c
722
if (unlikely(!dq)) {
drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c
736
fd = dpaa2_dq_fd(dq);
drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c
737
fq = (struct dpaa2_eth_fq *)(uintptr_t)dpaa2_dq_fqd_ctx(dq);
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch.c
2815
struct dpaa2_dq *dq;
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch.c
2820
dq = dpaa2_io_store_next(fq->store, &is_last);
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch.c
2821
if (unlikely(!dq)) {
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch.c
2831
dpaa2_switch_rx(fq, dpaa2_dq_fd(dq));
drivers/net/ethernet/freescale/dpaa2/dpaa2-switch.c
2833
dpaa2_switch_tx_conf(fq, dpaa2_dq_fd(dq));
drivers/net/ethernet/mellanox/mlxsw/pci.c
330
cq->u.cq.dq = q;
drivers/net/ethernet/mellanox/mlxsw/pci.c
567
cq->u.cq.dq = q;
drivers/net/ethernet/mellanox/mlxsw/pci.c
596
cq->u.cq.dq = NULL;
drivers/net/ethernet/mellanox/mlxsw/pci.c
851
struct mlxsw_pci_queue *rdq = q->u.cq.dq;
drivers/net/ethernet/mellanox/mlxsw/pci.c
87
struct mlxsw_pci_queue *dq;
drivers/net/ethernet/mellanox/mlxsw/pci.c
913
struct mlxsw_pci_queue *sdq = q->u.cq.dq;
drivers/s390/block/dasd.c
2868
struct dasd_queue *dq;
drivers/s390/block/dasd.c
2880
dq = cqr->dq;
drivers/s390/block/dasd.c
2881
spin_lock_irq(&dq->lock);
drivers/s390/block/dasd.c
2884
spin_unlock_irq(&dq->lock);
drivers/s390/block/dasd.c
2916
spin_lock_irq(&cqr->dq->lock);
drivers/s390/block/dasd.c
2919
spin_unlock_irq(&cqr->dq->lock);
drivers/s390/block/dasd.c
2986
spin_lock_irqsave(&cqr->dq->lock, flags);
drivers/s390/block/dasd.c
2990
spin_unlock_irqrestore(&cqr->dq->lock, flags);
drivers/s390/block/dasd.c
3022
struct dasd_queue *dq = hctx->driver_data;
drivers/s390/block/dasd.c
3029
spin_lock_irq(&dq->lock);
drivers/s390/block/dasd.c
3089
cqr->dq = dq;
drivers/s390/block/dasd.c
3100
spin_unlock_irq(&dq->lock);
drivers/s390/block/dasd.c
3124
spin_lock_irqsave(&cqr->dq->lock, flags);
drivers/s390/block/dasd.c
3127
spin_unlock_irqrestore(&cqr->dq->lock, flags);
drivers/s390/block/dasd.c
3175
spin_unlock_irqrestore(&cqr->dq->lock, flags);
drivers/s390/block/dasd.c
3183
struct dasd_queue *dq = kzalloc_obj(*dq);
drivers/s390/block/dasd.c
3185
if (!dq)
drivers/s390/block/dasd.c
3188
spin_lock_init(&dq->lock);
drivers/s390/block/dasd.c
3189
hctx->driver_data = dq;
drivers/s390/block/dasd_int.h
144
struct dasd_queue *dq;
drivers/s390/crypto/zcrypt_msgtype50.c
104
unsigned char dq[64];
drivers/s390/crypto/zcrypt_msgtype50.c
117
unsigned char dq[128];
drivers/s390/crypto/zcrypt_msgtype50.c
130
unsigned char dq[256];
drivers/s390/crypto/zcrypt_msgtype50.c
262
unsigned char *p, *q, *dp, *dq, *u, *inp;
drivers/s390/crypto/zcrypt_msgtype50.c
284
dq = crb1->dq + sizeof(crb1->dq) - short_len;
drivers/s390/crypto/zcrypt_msgtype50.c
298
dq = crb2->dq + sizeof(crb2->dq) - short_len;
drivers/s390/crypto/zcrypt_msgtype50.c
313
dq = crb3->dq + sizeof(crb3->dq) - short_len;
drivers/s390/crypto/zcrypt_msgtype50.c
327
copy_from_user(dq, crt->bq_key, short_len) ||
drivers/scsi/be2iscsi/be_cmds.c
1041
dq->id = le16_to_cpu(resp->id);
drivers/scsi/be2iscsi/be_cmds.c
1042
dq->created = true;
drivers/scsi/be2iscsi/be_cmds.c
1049
defq_ring->id = dq->id;
drivers/scsi/be2iscsi/be_cmds.c
976
struct be_queue_info *dq, int length,
drivers/scsi/be2iscsi/be_cmds.c
982
struct be_dma_mem *q_mem = &dq->dma_mem;
drivers/scsi/be2iscsi/be_cmds.h
839
struct be_queue_info *dq, int length,
drivers/scsi/be2iscsi/be_main.c
3139
struct be_queue_info *dq, *cq;
drivers/scsi/be2iscsi/be_main.c
3145
dq = &phwi_context->be_def_hdrq[ulp_num];
drivers/scsi/be2iscsi/be_main.c
3147
mem = &dq->dma_mem;
drivers/scsi/be2iscsi/be_main.c
3152
ret = be_fill_queue(dq, mem_descr->mem_array[0].size /
drivers/scsi/be2iscsi/be_main.c
3164
ret = be_cmd_create_default_pdu_queue(&phba->ctrl, cq, dq,
drivers/scsi/hisi_sas/hisi_sas.h
235
struct hisi_sas_dq *dq;
drivers/scsi/hisi_sas/hisi_sas.h
328
void (*start_delivery)(struct hisi_sas_dq *dq);
drivers/scsi/hisi_sas/hisi_sas.h
371
struct hisi_sas_dq *dq;
drivers/scsi/hisi_sas/hisi_sas.h
440
struct hisi_sas_dq dq[HISI_SAS_MAX_QUEUES];
drivers/scsi/hisi_sas/hisi_sas_main.c
2192
struct hisi_sas_dq *dq = &hisi_hba->dq[i];
drivers/scsi/hisi_sas/hisi_sas_main.c
2199
dq->wr_point = 0;
drivers/scsi/hisi_sas/hisi_sas_main.c
2244
struct hisi_sas_dq *dq = &hisi_hba->dq[i];
drivers/scsi/hisi_sas/hisi_sas_main.c
2252
spin_lock_init(&dq->lock);
drivers/scsi/hisi_sas/hisi_sas_main.c
2253
INIT_LIST_HEAD(&dq->list);
drivers/scsi/hisi_sas/hisi_sas_main.c
2254
dq->id = i;
drivers/scsi/hisi_sas/hisi_sas_main.c
2255
dq->hisi_hba = hisi_hba;
drivers/scsi/hisi_sas/hisi_sas_main.c
432
struct hisi_sas_dq *dq,
drivers/scsi/hisi_sas/hisi_sas_main.c
440
spin_lock(&dq->lock);
drivers/scsi/hisi_sas/hisi_sas_main.c
441
wr_q_index = dq->wr_point;
drivers/scsi/hisi_sas/hisi_sas_main.c
442
dq->wr_point = (dq->wr_point + 1) % HISI_SAS_QUEUE_SLOTS;
drivers/scsi/hisi_sas/hisi_sas_main.c
443
list_add_tail(&slot->delivery, &dq->list);
drivers/scsi/hisi_sas/hisi_sas_main.c
444
spin_unlock(&dq->lock);
drivers/scsi/hisi_sas/hisi_sas_main.c
449
dlvry_queue = dq->id;
drivers/scsi/hisi_sas/hisi_sas_main.c
488
spin_lock(&dq->lock);
drivers/scsi/hisi_sas/hisi_sas_main.c
489
hisi_hba->hw->start_delivery(dq);
drivers/scsi/hisi_sas/hisi_sas_main.c
490
spin_unlock(&dq->lock);
drivers/scsi/hisi_sas/hisi_sas_main.c
500
struct hisi_sas_dq *dq = NULL;
drivers/scsi/hisi_sas/hisi_sas_main.c
566
dq = &hisi_hba->dq[dq_index];
drivers/scsi/hisi_sas/hisi_sas_main.c
583
dq = &hisi_hba->dq[queue];
drivers/scsi/hisi_sas/hisi_sas_main.c
599
dq = &hisi_hba->dq[task->abort_task.qid];
drivers/scsi/hisi_sas/hisi_sas_main.c
635
hisi_sas_task_deliver(hisi_hba, slot, dq, sas_dev);
drivers/scsi/hisi_sas/hisi_sas_main.c
701
struct hisi_sas_dq *dq = &hisi_hba->dq[queue];
drivers/scsi/hisi_sas/hisi_sas_main.c
709
sas_dev->dq = dq;
drivers/scsi/hisi_sas/hisi_sas_v1_hw.c
863
static void start_delivery_v1_hw(struct hisi_sas_dq *dq)
drivers/scsi/hisi_sas/hisi_sas_v1_hw.c
865
struct hisi_hba *hisi_hba = dq->hisi_hba;
drivers/scsi/hisi_sas/hisi_sas_v1_hw.c
867
int dlvry_queue = dq->id;
drivers/scsi/hisi_sas/hisi_sas_v1_hw.c
870
list_for_each_entry_safe(s, s1, &dq->list, delivery) {
drivers/scsi/hisi_sas/hisi_sas_v2_hw.c
1645
static void start_delivery_v2_hw(struct hisi_sas_dq *dq)
drivers/scsi/hisi_sas/hisi_sas_v2_hw.c
1647
struct hisi_hba *hisi_hba = dq->hisi_hba;
drivers/scsi/hisi_sas/hisi_sas_v2_hw.c
1649
int dlvry_queue = dq->id;
drivers/scsi/hisi_sas/hisi_sas_v2_hw.c
1652
list_for_each_entry_safe(s, s1, &dq->list, delivery) {
drivers/scsi/hisi_sas/hisi_sas_v2_hw.c
860
struct hisi_sas_dq *dq = &hisi_hba->dq[queue];
drivers/scsi/hisi_sas/hisi_sas_v2_hw.c
869
sas_dev->dq = dq;
drivers/scsi/hisi_sas/hisi_sas_v3_hw.c
1163
static void start_delivery_v3_hw(struct hisi_sas_dq *dq)
drivers/scsi/hisi_sas/hisi_sas_v3_hw.c
1165
struct hisi_hba *hisi_hba = dq->hisi_hba;
drivers/scsi/hisi_sas/hisi_sas_v3_hw.c
1167
int dlvry_queue = dq->id;
drivers/scsi/hisi_sas/hisi_sas_v3_hw.c
1170
list_for_each_entry_safe(s, s1, &dq->list, delivery) {
drivers/scsi/hisi_sas/hisi_sas_v3_hw.c
4722
struct hisi_sas_debugfs_dq *dq =
drivers/scsi/hisi_sas/hisi_sas_v3_hw.c
4725
dq->hdr = devm_kmalloc(dev, sz, GFP_KERNEL);
drivers/scsi/hisi_sas/hisi_sas_v3_hw.c
4726
if (!dq->hdr)
drivers/scsi/hisi_sas/hisi_sas_v3_hw.c
4728
dq->dq = &hisi_hba->dq[d];
drivers/scsi/lpfc/lpfc_crtn.h
299
int lpfc_sli4_rq_put(struct lpfc_queue *hq, struct lpfc_queue *dq,
drivers/scsi/lpfc/lpfc_sli.c
798
lpfc_sli4_rq_put(struct lpfc_queue *hq, struct lpfc_queue *dq,
drivers/scsi/lpfc/lpfc_sli.c
808
if (unlikely(!hq) || unlikely(!dq))
drivers/scsi/lpfc/lpfc_sli.c
811
dq_put_index = dq->host_index;
drivers/scsi/lpfc/lpfc_sli.c
813
temp_drqe = lpfc_sli4_qe(dq, dq_put_index);
drivers/scsi/lpfc/lpfc_sli.c
815
if (hq->type != LPFC_HRQ || dq->type != LPFC_DRQ)
drivers/scsi/lpfc/lpfc_sli.c
823
lpfc_sli4_pcimem_bcopy(drqe, temp_drqe, dq->entry_size);
drivers/scsi/lpfc/lpfc_sli.c
827
dq->host_index = ((dq_put_index + 1) % dq->entry_count);
drivers/scsi/lpfc/lpfc_sli.c
861
lpfc_sli4_rq_release(struct lpfc_queue *hq, struct lpfc_queue *dq)
drivers/scsi/lpfc/lpfc_sli.c
864
if (unlikely(!hq) || unlikely(!dq))
drivers/scsi/lpfc/lpfc_sli.c
867
if ((hq->type != LPFC_HRQ) || (dq->type != LPFC_DRQ))
drivers/scsi/lpfc/lpfc_sli.c
870
dq->hba_index = ((dq->hba_index + 1) % dq->entry_count);
drivers/scsi/qlogicpti.c
1208
struct scsi_cmnd *dq;
drivers/scsi/qlogicpti.c
1211
dq = qlogicpti_intr_handler(qpti);
drivers/scsi/qlogicpti.c
1213
if (dq != NULL) {
drivers/scsi/qlogicpti.c
1217
next = (struct scsi_cmnd *) dq->host_scribble;
drivers/scsi/qlogicpti.c
1218
scsi_done(dq);
drivers/scsi/qlogicpti.c
1219
dq = next;
drivers/scsi/qlogicpti.c
1220
} while (dq != NULL);
drivers/soc/fsl/dpio/dpio-service.c
228
const struct dpaa2_dq *dq;
drivers/soc/fsl/dpio/dpio-service.c
240
dq = qbman_swp_dqrr_next(swp);
drivers/soc/fsl/dpio/dpio-service.c
241
while (dq) {
drivers/soc/fsl/dpio/dpio-service.c
242
if (qbman_result_is_SCN(dq)) {
drivers/soc/fsl/dpio/dpio-service.c
246
q64 = qbman_result_SCN_ctx(dq);
drivers/soc/fsl/dpio/dpio-service.c
252
qbman_swp_dqrr_consume(swp, dq);
drivers/soc/fsl/dpio/dpio-service.c
256
dq = qbman_swp_dqrr_next(swp);
drivers/soc/fsl/dpio/qbman-portal.c
1218
verb = p->dq.verb;
drivers/soc/fsl/dpio/qbman-portal.c
1246
flags = p->dq.stat;
drivers/soc/fsl/dpio/qbman-portal.c
1310
verb = p->dq.verb;
drivers/soc/fsl/dpio/qbman-portal.c
1338
flags = p->dq.stat;
drivers/soc/fsl/dpio/qbman-portal.c
1356
void qbman_swp_dqrr_consume(struct qbman_swp *s, const struct dpaa2_dq *dq)
drivers/soc/fsl/dpio/qbman-portal.c
1358
qbman_write_register(s, QBMAN_CINH_SWP_DCAP, QBMAN_IDX_FROM_DQRR(dq));
drivers/soc/fsl/dpio/qbman-portal.c
1378
int qbman_result_has_new_result(struct qbman_swp *s, const struct dpaa2_dq *dq)
drivers/soc/fsl/dpio/qbman-portal.c
1380
if (dq->dq.tok != QMAN_DQ_TOKEN_VALID)
drivers/soc/fsl/dpio/qbman-portal.c
1388
((struct dpaa2_dq *)dq)->dq.tok = 0;
drivers/soc/fsl/dpio/qbman-portal.c
1395
if (s->vdq.storage == dq) {
drivers/soc/fsl/dpio/qbman-portal.h
219
void qbman_swp_dqrr_consume(struct qbman_swp *s, const struct dpaa2_dq *dq);
drivers/soc/fsl/dpio/qbman-portal.h
221
int qbman_result_has_new_result(struct qbman_swp *p, const struct dpaa2_dq *dq);
drivers/soc/fsl/dpio/qbman-portal.h
308
static inline int qbman_result_is_DQ(const struct dpaa2_dq *dq)
drivers/soc/fsl/dpio/qbman-portal.h
310
return ((dq->dq.verb & QBMAN_RESULT_MASK) == QBMAN_RESULT_DQ);
drivers/soc/fsl/dpio/qbman-portal.h
318
static inline int qbman_result_is_SCN(const struct dpaa2_dq *dq)
drivers/soc/fsl/dpio/qbman-portal.h
320
return !qbman_result_is_DQ(dq);
drivers/soc/fsl/dpio/qbman-portal.h
324
static inline int qbman_result_is_FQDAN(const struct dpaa2_dq *dq)
drivers/soc/fsl/dpio/qbman-portal.h
326
return ((dq->dq.verb & QBMAN_RESULT_MASK) == QBMAN_RESULT_FQDAN);
drivers/soc/fsl/dpio/qbman-portal.h
330
static inline int qbman_result_is_CDAN(const struct dpaa2_dq *dq)
drivers/soc/fsl/dpio/qbman-portal.h
332
return ((dq->dq.verb & QBMAN_RESULT_MASK) == QBMAN_RESULT_CDAN);
drivers/soc/fsl/dpio/qbman-portal.h
336
static inline int qbman_result_is_CSCN(const struct dpaa2_dq *dq)
drivers/soc/fsl/dpio/qbman-portal.h
338
return ((dq->dq.verb & QBMAN_RESULT_MASK) == QBMAN_RESULT_CSCN_WQ);
drivers/soc/fsl/dpio/qbman-portal.h
342
static inline int qbman_result_is_BPSCN(const struct dpaa2_dq *dq)
drivers/soc/fsl/dpio/qbman-portal.h
344
return ((dq->dq.verb & QBMAN_RESULT_MASK) == QBMAN_RESULT_BPSCN);
drivers/soc/fsl/dpio/qbman-portal.h
348
static inline int qbman_result_is_CGCU(const struct dpaa2_dq *dq)
drivers/soc/fsl/dpio/qbman-portal.h
350
return ((dq->dq.verb & QBMAN_RESULT_MASK) == QBMAN_RESULT_CGCU);
drivers/soc/fsl/dpio/qbman-portal.h
354
static inline int qbman_result_is_FQRN(const struct dpaa2_dq *dq)
drivers/soc/fsl/dpio/qbman-portal.h
356
return ((dq->dq.verb & QBMAN_RESULT_MASK) == QBMAN_RESULT_FQRN);
drivers/soc/fsl/dpio/qbman-portal.h
360
static inline int qbman_result_is_FQRNI(const struct dpaa2_dq *dq)
drivers/soc/fsl/dpio/qbman-portal.h
362
return ((dq->dq.verb & QBMAN_RESULT_MASK) == QBMAN_RESULT_FQRNI);
drivers/soc/fsl/dpio/qbman-portal.h
366
static inline int qbman_result_is_FQPN(const struct dpaa2_dq *dq)
drivers/soc/fsl/dpio/qbman-portal.h
368
return ((dq->dq.verb & QBMAN_RESULT_MASK) == QBMAN_RESULT_FQPN);
drivers/soc/fsl/qbman/qman.c
1610
const struct qm_dqrr_entry *dq;
drivers/soc/fsl/qbman/qman.c
1617
dq = qm_dqrr_current(&p->p);
drivers/soc/fsl/qbman/qman.c
1618
if (!dq)
drivers/soc/fsl/qbman/qman.c
1621
if (dq->stat & QM_DQRR_STAT_UNSCHEDULED) {
drivers/soc/fsl/qbman/qman.c
1634
if (dq->stat & QM_DQRR_STAT_FQ_EMPTY)
drivers/soc/fsl/qbman/qman.c
1642
res = fq->cb.dqrr(p, fq, dq, sched_napi);
drivers/soc/fsl/qbman/qman.c
1646
if (dq->stat & QM_DQRR_STAT_DQCR_EXPIRED)
drivers/soc/fsl/qbman/qman.c
1650
fq = tag_to_fq(be32_to_cpu(dq->context_b));
drivers/soc/fsl/qbman/qman.c
1652
res = fq->cb.dqrr(p, fq, dq, sched_napi);
drivers/soc/fsl/qbman/qman.c
1666
DPAA_ASSERT((dq->stat & QM_DQRR_STAT_FQ_HELDACTIVE) ||
drivers/soc/fsl/qbman/qman.c
1670
qm_dqrr_cdc_consume_1ptr(&p->p, dq,
drivers/soc/fsl/qbman/qman.c
694
const struct qm_dqrr_entry *dq,
drivers/soc/fsl/qbman/qman.c
698
int idx = dqrr_ptr2idx(dq);
drivers/soc/fsl/qbman/qman.c
701
DPAA_ASSERT((dqrr->ring + idx) == dq);
drivers/soc/fsl/qbman/qman_test_api.c
212
const struct qm_dqrr_entry *dq,
drivers/soc/fsl/qbman/qman_test_api.c
215
if (WARN_ON(fd_neq(&fd_dq, &dq->fd))) {
drivers/soc/fsl/qbman/qman_test_api.c
220
if (!(dq->stat & QM_DQRR_STAT_UNSCHEDULED) && !fd_neq(&fd_dq, &fd)) {
fs/xfs/scrub/dqiterate.c
150
struct xfs_dquot *dq;
fs/xfs/scrub/dqiterate.c
156
nr_found = radix_tree_gang_lookup(tree, (void **)&dq, cursor->id, 1);
fs/xfs/scrub/dqiterate.c
158
*next_incore_id = dq->q_id;
fs/xfs/scrub/dqiterate.c
176
struct xfs_dquot *dq = NULL;
fs/xfs/scrub/dqiterate.c
204
error = xfs_qm_dqget(mp, cursor->id, cursor->dqtype, false, &dq);
fs/xfs/scrub/dqiterate.c
208
cursor->id = dq->q_id + 1;
fs/xfs/scrub/dqiterate.c
209
*dqpp = dq;
fs/xfs/scrub/quota.c
113
if (XFS_FSB_TO_DADDR(mp, irec.br_startblock) != dq->q_blkno)
fs/xfs/scrub/quota.c
142
struct xfs_dquot *dq)
fs/xfs/scrub/quota.c
160
mutex_lock(&dq->q_qlock);
fs/xfs/scrub/quota.c
166
offset = dq->q_id / qi->qi_dqperchunk;
fs/xfs/scrub/quota.c
167
if (dq->q_id && dq->q_id <= sqi->last_id)
fs/xfs/scrub/quota.c
170
sqi->last_id = dq->q_id;
fs/xfs/scrub/quota.c
172
error = xchk_quota_item_bmap(sc, dq, offset);
fs/xfs/scrub/quota.c
175
mutex_unlock(&dq->q_qlock);
fs/xfs/scrub/quota.c
187
if (dq->q_blk.hardlimit > mp->m_sb.sb_dblocks)
fs/xfs/scrub/quota.c
189
if (dq->q_blk.softlimit > dq->q_blk.hardlimit)
fs/xfs/scrub/quota.c
192
if (dq->q_ino.hardlimit > M_IGEO(mp)->maxicount)
fs/xfs/scrub/quota.c
194
if (dq->q_ino.softlimit > dq->q_ino.hardlimit)
fs/xfs/scrub/quota.c
197
if (dq->q_rtb.hardlimit > mp->m_sb.sb_rblocks)
fs/xfs/scrub/quota.c
199
if (dq->q_rtb.softlimit > dq->q_rtb.hardlimit)
fs/xfs/scrub/quota.c
211
if (mp->m_sb.sb_dblocks < dq->q_blk.count)
fs/xfs/scrub/quota.c
214
if (mp->m_sb.sb_rblocks < dq->q_rtb.count)
fs/xfs/scrub/quota.c
218
if (mp->m_sb.sb_dblocks < dq->q_blk.count)
fs/xfs/scrub/quota.c
221
if (mp->m_sb.sb_rblocks < dq->q_rtb.count)
fs/xfs/scrub/quota.c
225
if (dq->q_ino.count > fs_icount)
fs/xfs/scrub/quota.c
233
if (dq->q_id == 0)
fs/xfs/scrub/quota.c
236
if (dq->q_blk.hardlimit != 0 &&
fs/xfs/scrub/quota.c
237
dq->q_blk.count > dq->q_blk.hardlimit)
fs/xfs/scrub/quota.c
240
if (dq->q_ino.hardlimit != 0 &&
fs/xfs/scrub/quota.c
241
dq->q_ino.count > dq->q_ino.hardlimit)
fs/xfs/scrub/quota.c
244
if (dq->q_rtb.hardlimit != 0 &&
fs/xfs/scrub/quota.c
245
dq->q_rtb.count > dq->q_rtb.hardlimit)
fs/xfs/scrub/quota.c
248
xchk_quota_item_timer(sc, offset, &dq->q_blk);
fs/xfs/scrub/quota.c
249
xchk_quota_item_timer(sc, offset, &dq->q_ino);
fs/xfs/scrub/quota.c
250
xchk_quota_item_timer(sc, offset, &dq->q_rtb);
fs/xfs/scrub/quota.c
253
mutex_unlock(&dq->q_qlock);
fs/xfs/scrub/quota.c
309
struct xfs_dquot *dq;
fs/xfs/scrub/quota.c
331
while ((error = xchk_dquot_iter(&cursor, &dq)) == 1) {
fs/xfs/scrub/quota.c
332
error = xchk_quota_item(&sqi, dq);
fs/xfs/scrub/quota.c
333
xfs_qm_dqrele(dq);
fs/xfs/scrub/quota.c
84
struct xfs_dquot *dq,
fs/xfs/scrub/quota.c
97
if (dq->q_fileoffset != offset) {
fs/xfs/scrub/quota_repair.c
108
struct xfs_dquot *dq,
fs/xfs/scrub/quota_repair.c
114
xfs_fileoff_t offset = dq->q_id / qi->qi_dqperchunk;
fs/xfs/scrub/quota_repair.c
123
dq->q_fileoffset = offset;
fs/xfs/scrub/quota_repair.c
131
error = xrep_quota_item_fill_bmap_hole(sc, dq, &irec);
fs/xfs/scrub/quota_repair.c
138
} else if (dq->q_blkno != XFS_FSB_TO_DADDR(mp, irec.br_startblock)) {
fs/xfs/scrub/quota_repair.c
144
dq->q_blkno = XFS_FSB_TO_DADDR(mp, irec.br_startblock);
fs/xfs/scrub/quota_repair.c
172
struct xfs_dquot *dq)
fs/xfs/scrub/quota_repair.c
189
mutex_lock(&dq->q_qlock);
fs/xfs/scrub/quota_repair.c
190
error = xrep_quota_item_bmap(sc, dq, &dirty);
fs/xfs/scrub/quota_repair.c
196
if (dq->q_blk.softlimit > dq->q_blk.hardlimit) {
fs/xfs/scrub/quota_repair.c
197
dq->q_blk.softlimit = dq->q_blk.hardlimit;
fs/xfs/scrub/quota_repair.c
201
if (dq->q_ino.softlimit > dq->q_ino.hardlimit) {
fs/xfs/scrub/quota_repair.c
202
dq->q_ino.softlimit = dq->q_ino.hardlimit;
fs/xfs/scrub/quota_repair.c
206
if (dq->q_rtb.softlimit > dq->q_rtb.hardlimit) {
fs/xfs/scrub/quota_repair.c
207
dq->q_rtb.softlimit = dq->q_rtb.hardlimit;
fs/xfs/scrub/quota_repair.c
217
if (!xfs_has_reflink(mp) && dq->q_blk.count > mp->m_sb.sb_dblocks) {
fs/xfs/scrub/quota_repair.c
218
dq->q_blk.reserved -= dq->q_blk.count;
fs/xfs/scrub/quota_repair.c
219
dq->q_blk.reserved += mp->m_sb.sb_dblocks;
fs/xfs/scrub/quota_repair.c
220
dq->q_blk.count = mp->m_sb.sb_dblocks;
fs/xfs/scrub/quota_repair.c
225
if (dq->q_ino.count > fs_icount) {
fs/xfs/scrub/quota_repair.c
226
dq->q_ino.reserved -= dq->q_ino.count;
fs/xfs/scrub/quota_repair.c
227
dq->q_ino.reserved += fs_icount;
fs/xfs/scrub/quota_repair.c
228
dq->q_ino.count = fs_icount;
fs/xfs/scrub/quota_repair.c
232
if (!xfs_has_reflink(mp) && dq->q_rtb.count > mp->m_sb.sb_rblocks) {
fs/xfs/scrub/quota_repair.c
233
dq->q_rtb.reserved -= dq->q_rtb.count;
fs/xfs/scrub/quota_repair.c
234
dq->q_rtb.reserved += mp->m_sb.sb_rblocks;
fs/xfs/scrub/quota_repair.c
235
dq->q_rtb.count = mp->m_sb.sb_rblocks;
fs/xfs/scrub/quota_repair.c
240
xrep_quota_item_timer(sc, &dq->q_blk, &dirty);
fs/xfs/scrub/quota_repair.c
241
xrep_quota_item_timer(sc, &dq->q_ino, &dirty);
fs/xfs/scrub/quota_repair.c
242
xrep_quota_item_timer(sc, &dq->q_rtb, &dirty);
fs/xfs/scrub/quota_repair.c
247
trace_xrep_dquot_item(sc->mp, dq->q_type, dq->q_id);
fs/xfs/scrub/quota_repair.c
249
dq->q_flags |= XFS_DQFLAG_DIRTY;
fs/xfs/scrub/quota_repair.c
250
xfs_trans_dqjoin(sc->tp, dq);
fs/xfs/scrub/quota_repair.c
251
if (dq->q_id) {
fs/xfs/scrub/quota_repair.c
252
xfs_qm_adjust_dqlimits(dq);
fs/xfs/scrub/quota_repair.c
253
xfs_qm_adjust_dqtimers(dq);
fs/xfs/scrub/quota_repair.c
255
xfs_trans_log_dquot(sc->tp, dq);
fs/xfs/scrub/quota_repair.c
259
mutex_unlock(&dq->q_qlock);
fs/xfs/scrub/quota_repair.c
508
struct xfs_dquot *dq;
fs/xfs/scrub/quota_repair.c
512
while ((error = xchk_dquot_iter(&cursor, &dq)) == 1) {
fs/xfs/scrub/quota_repair.c
513
error = xrep_quota_item(&rqi, dq);
fs/xfs/scrub/quota_repair.c
514
xfs_qm_dqrele(dq);
fs/xfs/scrub/quota_repair.c
58
struct xfs_dquot *dq,
fs/xfs/scrub/quota_repair.c
74
error = xfs_bmapi_write(sc->tp, sc->ip, dq->q_fileoffset,
fs/xfs/scrub/quota_repair.c
80
dq->q_blkno = XFS_FSB_TO_DADDR(mp, irec->br_startblock);
fs/xfs/scrub/quota_repair.c
82
trace_xrep_dquot_item_fill_bmap_hole(sc->mp, dq->q_type, dq->q_id);
fs/xfs/scrub/quota_repair.c
85
error = xfs_trans_get_buf(sc->tp, mp->m_ddev_targp, dq->q_blkno,
fs/xfs/scrub/quota_repair.c
91
xfs_qm_init_dquot_blk(sc->tp, dq->q_id, dq->q_type, bp);
fs/xfs/scrub/quotacheck.c
555
struct xfs_dquot *dq)
fs/xfs/scrub/quotacheck.c
566
mutex_lock(&dq->q_qlock);
fs/xfs/scrub/quotacheck.c
568
error = xfarray_load_sparse(counts, dq->q_id, &xcdq);
fs/xfs/scrub/quotacheck.c
572
if (xcdq.icount != dq->q_ino.count)
fs/xfs/scrub/quotacheck.c
573
xchk_qcheck_set_corrupt(xqc->sc, dqtype, dq->q_id);
fs/xfs/scrub/quotacheck.c
575
if (xcdq.bcount != dq->q_blk.count)
fs/xfs/scrub/quotacheck.c
576
xchk_qcheck_set_corrupt(xqc->sc, dqtype, dq->q_id);
fs/xfs/scrub/quotacheck.c
578
if (xcdq.rtbcount != dq->q_rtb.count)
fs/xfs/scrub/quotacheck.c
579
xchk_qcheck_set_corrupt(xqc->sc, dqtype, dq->q_id);
fs/xfs/scrub/quotacheck.c
582
error = xfarray_store(counts, dq->q_id, &xcdq);
fs/xfs/scrub/quotacheck.c
595
mutex_unlock(&dq->q_qlock);
fs/xfs/scrub/quotacheck.c
615
struct xfs_dquot *dq;
fs/xfs/scrub/quotacheck.c
629
error = xfs_qm_dqget(xqc->sc->mp, id, dqtype, false, &dq);
fs/xfs/scrub/quotacheck.c
637
error = xqcheck_compare_dquot(xqc, dqtype, dq);
fs/xfs/scrub/quotacheck.c
638
xfs_qm_dqrele(dq);
fs/xfs/scrub/quotacheck.c
660
struct xfs_dquot *dq;
fs/xfs/scrub/quotacheck.c
674
while ((error = xchk_dquot_iter(&cursor, &dq)) == 1) {
fs/xfs/scrub/quotacheck.c
675
error = xqcheck_compare_dquot(xqc, dqtype, dq);
fs/xfs/scrub/quotacheck.c
676
xfs_qm_dqrele(dq);
fs/xfs/scrub/quotacheck_repair.c
109
trace_xrep_quotacheck_dquot(xqc->sc->mp, dq->q_type, dq->q_id);
fs/xfs/scrub/quotacheck_repair.c
112
dq->q_flags |= XFS_DQFLAG_DIRTY;
fs/xfs/scrub/quotacheck_repair.c
113
if (dq->q_id)
fs/xfs/scrub/quotacheck_repair.c
114
xfs_qm_adjust_dqtimers(dq);
fs/xfs/scrub/quotacheck_repair.c
115
xfs_trans_log_dquot(xqc->sc->tp, dq);
fs/xfs/scrub/quotacheck_repair.c
136
struct xfs_dquot *dq;
fs/xfs/scrub/quotacheck_repair.c
144
while ((error = xchk_dquot_iter(&cursor, &dq)) == 1) {
fs/xfs/scrub/quotacheck_repair.c
145
error = xqcheck_commit_dquot(xqc, dqtype, dq);
fs/xfs/scrub/quotacheck_repair.c
146
xfs_qm_dqrele(dq);
fs/xfs/scrub/quotacheck_repair.c
172
error = xfs_qm_dqget(mp, id, dqtype, true, &dq);
fs/xfs/scrub/quotacheck_repair.c
176
error = xqcheck_commit_dquot(xqc, dqtype, dq);
fs/xfs/scrub/quotacheck_repair.c
177
xfs_qm_dqrele(dq);
fs/xfs/scrub/quotacheck_repair.c
47
struct xfs_dquot *dq)
fs/xfs/scrub/quotacheck_repair.c
59
mutex_lock(&dq->q_qlock);
fs/xfs/scrub/quotacheck_repair.c
60
xfs_trans_dqjoin(xqc->sc->tp, dq);
fs/xfs/scrub/quotacheck_repair.c
68
error = xfarray_load_sparse(counts, dq->q_id, &xcdq);
fs/xfs/scrub/quotacheck_repair.c
73
delta = (int64_t)xcdq.icount - dq->q_ino.count;
fs/xfs/scrub/quotacheck_repair.c
75
dq->q_ino.reserved += delta;
fs/xfs/scrub/quotacheck_repair.c
76
dq->q_ino.count += delta;
fs/xfs/scrub/quotacheck_repair.c
80
delta = (int64_t)xcdq.bcount - dq->q_blk.count;
fs/xfs/scrub/quotacheck_repair.c
82
dq->q_blk.reserved += delta;
fs/xfs/scrub/quotacheck_repair.c
83
dq->q_blk.count += delta;
fs/xfs/scrub/quotacheck_repair.c
87
delta = (int64_t)xcdq.rtbcount - dq->q_rtb.count;
fs/xfs/scrub/quotacheck_repair.c
89
dq->q_rtb.reserved += delta;
fs/xfs/scrub/quotacheck_repair.c
90
dq->q_rtb.count += delta;
fs/xfs/scrub/quotacheck_repair.c
95
error = xfarray_store(counts, dq->q_id, &xcdq);
fs/xfs/xfs_dquot.c
120
struct xfs_dquot *dq)
fs/xfs/xfs_dquot.c
122
struct xfs_mount *mp = dq->q_mount;
fs/xfs/xfs_dquot.c
127
ASSERT(dq->q_id);
fs/xfs/xfs_dquot.c
128
defq = xfs_get_defquota(q, xfs_dquot_type(dq));
fs/xfs/xfs_dquot.c
130
if (!dq->q_blk.softlimit) {
fs/xfs/xfs_dquot.c
131
dq->q_blk.softlimit = defq->blk.soft;
fs/xfs/xfs_dquot.c
134
if (!dq->q_blk.hardlimit) {
fs/xfs/xfs_dquot.c
135
dq->q_blk.hardlimit = defq->blk.hard;
fs/xfs/xfs_dquot.c
138
if (!dq->q_ino.softlimit)
fs/xfs/xfs_dquot.c
139
dq->q_ino.softlimit = defq->ino.soft;
fs/xfs/xfs_dquot.c
140
if (!dq->q_ino.hardlimit)
fs/xfs/xfs_dquot.c
141
dq->q_ino.hardlimit = defq->ino.hard;
fs/xfs/xfs_dquot.c
142
if (!dq->q_rtb.softlimit)
fs/xfs/xfs_dquot.c
143
dq->q_rtb.softlimit = defq->rtb.soft;
fs/xfs/xfs_dquot.c
144
if (!dq->q_rtb.hardlimit)
fs/xfs/xfs_dquot.c
145
dq->q_rtb.hardlimit = defq->rtb.hard;
fs/xfs/xfs_dquot.c
148
xfs_dquot_set_prealloc_limits(dq);
fs/xfs/xfs_dquot.c
208
struct xfs_dquot *dq)
fs/xfs/xfs_dquot.c
210
struct xfs_mount *mp = dq->q_mount;
fs/xfs/xfs_dquot.c
214
ASSERT(dq->q_id);
fs/xfs/xfs_dquot.c
215
defq = xfs_get_defquota(qi, xfs_dquot_type(dq));
fs/xfs/xfs_dquot.c
217
xfs_qm_adjust_res_timer(dq->q_mount, &dq->q_blk, &defq->blk);
fs/xfs/xfs_dquot.c
218
xfs_qm_adjust_res_timer(dq->q_mount, &dq->q_ino, &defq->ino);
fs/xfs/xfs_dquot.c
219
xfs_qm_adjust_res_timer(dq->q_mount, &dq->q_rtb, &defq->rtb);
fs/xfs/xfs_iomap.c
368
struct xfs_dquot *dq = xfs_inode_dquot(ip, type);
fs/xfs/xfs_iomap.c
372
if (!dq || !xfs_this_quota_on(ip->i_mount, type))
fs/xfs/xfs_iomap.c
376
res = &dq->q_rtb;
fs/xfs/xfs_iomap.c
377
pre = &dq->q_rtb_prealloc;
fs/xfs/xfs_iomap.c
379
res = &dq->q_blk;
fs/xfs/xfs_iomap.c
380
pre = &dq->q_blk_prealloc;
fs/xfs/xfs_iomap.c
402
struct xfs_dquot *dq = xfs_inode_dquot(ip, type);
fs/xfs/xfs_iomap.c
408
if (!dq) {
fs/xfs/xfs_iomap.c
412
res = &dq->q_rtb;
fs/xfs/xfs_iomap.c
413
pre = &dq->q_rtb_prealloc;
fs/xfs/xfs_iomap.c
415
res = &dq->q_blk;
fs/xfs/xfs_iomap.c
416
pre = &dq->q_blk_prealloc;
include/crypto/internal/rsa.h
39
const u8 *dq;
include/soc/fsl/dpaa2-global.h
104
static inline u16 dpaa2_dq_seqnum(const struct dpaa2_dq *dq)
include/soc/fsl/dpaa2-global.h
106
return le16_to_cpu(dq->dq.seqnum);
include/soc/fsl/dpaa2-global.h
117
static inline u16 dpaa2_dq_odpid(const struct dpaa2_dq *dq)
include/soc/fsl/dpaa2-global.h
119
return le16_to_cpu(dq->dq.oprid);
include/soc/fsl/dpaa2-global.h
128
static inline u32 dpaa2_dq_fqid(const struct dpaa2_dq *dq)
include/soc/fsl/dpaa2-global.h
130
return le32_to_cpu(dq->dq.fqid) & DQ_FQID_MASK;
include/soc/fsl/dpaa2-global.h
139
static inline u32 dpaa2_dq_byte_count(const struct dpaa2_dq *dq)
include/soc/fsl/dpaa2-global.h
141
return le32_to_cpu(dq->dq.fq_byte_cnt);
include/soc/fsl/dpaa2-global.h
150
static inline u32 dpaa2_dq_frame_count(const struct dpaa2_dq *dq)
include/soc/fsl/dpaa2-global.h
152
return le32_to_cpu(dq->dq.fq_frm_cnt) & DQ_FRAME_COUNT_MASK;
include/soc/fsl/dpaa2-global.h
161
static inline u64 dpaa2_dq_fqd_ctx(const struct dpaa2_dq *dq)
include/soc/fsl/dpaa2-global.h
163
return le64_to_cpu(dq->dq.fqd_ctx);
include/soc/fsl/dpaa2-global.h
172
static inline const struct dpaa2_fd *dpaa2_dq_fd(const struct dpaa2_dq *dq)
include/soc/fsl/dpaa2-global.h
174
return (const struct dpaa2_fd *)&dq->dq.fd[0];
include/soc/fsl/dpaa2-global.h
33
} dq;
include/soc/fsl/dpaa2-global.h
68
static inline u32 dpaa2_dq_flags(const struct dpaa2_dq *dq)
include/soc/fsl/dpaa2-global.h
70
return dq->dq.stat;
include/soc/fsl/dpaa2-global.h
80
static inline int dpaa2_dq_is_pull(const struct dpaa2_dq *dq)
include/soc/fsl/dpaa2-global.h
82
return (int)(dpaa2_dq_flags(dq) & DPAA2_DQ_STAT_VOLATILE);
include/soc/fsl/dpaa2-global.h
91
static inline bool dpaa2_dq_is_pull_complete(const struct dpaa2_dq *dq)
include/soc/fsl/dpaa2-global.h
93
return !!(dpaa2_dq_flags(dq) & DPAA2_DQ_STAT_EXPIRED);
lib/raid6/neon.h
16
uint8_t *dq, const uint8_t *pbmul,
lib/raid6/neon.h
19
void __raid6_datap_recov_neon(int bytes, uint8_t *p, uint8_t *q, uint8_t *dq,
lib/raid6/recov.c
22
u8 *p, *q, *dp, *dq;
lib/raid6/recov.c
36
dq = (u8 *)ptrs[failb];
lib/raid6/recov.c
38
ptrs[disks-1] = dq;
lib/raid6/recov.c
44
ptrs[failb] = dq;
lib/raid6/recov.c
55
qx = qmul[*q ^ *dq];
lib/raid6/recov.c
56
*dq++ = db = pbmul[px] ^ qx; /* Reconstructed B */
lib/raid6/recov.c
66
u8 *p, *q, *dq;
lib/raid6/recov.c
74
dq = (u8 *)ptrs[faila];
lib/raid6/recov.c
76
ptrs[disks-1] = dq;
lib/raid6/recov.c
81
ptrs[faila] = dq;
lib/raid6/recov.c
89
*p++ ^= *dq = qmul[*q ^ *dq];
lib/raid6/recov.c
90
q++; dq++;
lib/raid6/recov_avx2.c
120
asm volatile("vmovdqa %%ymm1, %0" : "=m" (dq[0]));
lib/raid6/recov_avx2.c
121
asm volatile("vmovdqa %%ymm13,%0" : "=m" (dq[32]));
lib/raid6/recov_avx2.c
132
dq += 64;
lib/raid6/recov_avx2.c
136
asm volatile("vpxor %0, %%ymm1, %%ymm1" : : "m" (*dq));
lib/raid6/recov_avx2.c
170
asm volatile("vmovdqa %%ymm1, %0" : "=m" (dq[0]));
lib/raid6/recov_avx2.c
179
dq += 32;
lib/raid6/recov_avx2.c
189
u8 *p, *q, *dq;
lib/raid6/recov_avx2.c
19
u8 *p, *q, *dp, *dq;
lib/raid6/recov_avx2.c
198
dq = (u8 *)ptrs[faila];
lib/raid6/recov_avx2.c
200
ptrs[disks-1] = dq;
lib/raid6/recov_avx2.c
205
ptrs[faila] = dq;
lib/raid6/recov_avx2.c
217
asm volatile("vmovdqa %0, %%ymm3" : : "m" (dq[0]));
lib/raid6/recov_avx2.c
218
asm volatile("vmovdqa %0, %%ymm8" : : "m" (dq[32]));
lib/raid6/recov_avx2.c
258
asm volatile("vmovdqa %%ymm1, %0" : "=m" (dq[0]));
lib/raid6/recov_avx2.c
259
asm volatile("vmovdqa %%ymm14, %0" : "=m" (dq[32]));
lib/raid6/recov_avx2.c
266
dq += 64;
lib/raid6/recov_avx2.c
268
asm volatile("vmovdqa %0, %%ymm3" : : "m" (dq[0]));
lib/raid6/recov_avx2.c
290
asm volatile("vmovdqa %%ymm1, %0" : "=m" (dq[0]));
lib/raid6/recov_avx2.c
296
dq += 32;
lib/raid6/recov_avx2.c
33
dq = (u8 *)ptrs[failb];
lib/raid6/recov_avx2.c
35
ptrs[disks-1] = dq;
lib/raid6/recov_avx2.c
41
ptrs[failb] = dq;
lib/raid6/recov_avx2.c
61
asm volatile("vpxor %0, %%ymm1, %%ymm1" : : "m" (dq[0]));
lib/raid6/recov_avx2.c
62
asm volatile("vpxor %0, %%ymm9, %%ymm9" : : "m" (dq[32]));
lib/raid6/recov_avx512.c
146
: "m" (dq[0]), "m" (dq[64]));
lib/raid6/recov_avx512.c
157
dq += 128;
lib/raid6/recov_avx512.c
164
: "m" (*q), "m" (*p), "m"(*dq), "m" (*dp));
lib/raid6/recov_avx512.c
207
: "m" (dq[0]));
lib/raid6/recov_avx512.c
218
dq += 64;
lib/raid6/recov_avx512.c
228
u8 *p, *q, *dq;
lib/raid6/recov_avx512.c
240
dq = (u8 *)ptrs[faila];
lib/raid6/recov_avx512.c
242
ptrs[disks-1] = dq;
lib/raid6/recov_avx512.c
247
ptrs[faila] = dq;
lib/raid6/recov_avx512.c
25
u8 *p, *q, *dp, *dq;
lib/raid6/recov_avx512.c
264
: "m" (dq[0]), "m" (dq[64]), "m" (q[0]),
lib/raid6/recov_avx512.c
314
: "m" (dq[0]), "m" (dq[64]), "m" (p[0]),
lib/raid6/recov_avx512.c
320
dq += 128;
lib/raid6/recov_avx512.c
325
: "m" (dq[0]), "m" (q[0]));
lib/raid6/recov_avx512.c
355
: "m" (dq[0]), "m" (p[0]));
lib/raid6/recov_avx512.c
360
dq += 64;
lib/raid6/recov_avx512.c
42
dq = (u8 *)ptrs[failb];
lib/raid6/recov_avx512.c
44
ptrs[disks-1] = dq;
lib/raid6/recov_avx512.c
50
ptrs[failb] = dq;
lib/raid6/recov_avx512.c
76
"m" (p[64]), "m" (dq[0]), "m" (dq[64]),
lib/raid6/recov_loongarch_simd.c
161
asm volatile("vst $vr4, %0" : "=m" (dq[0]));
lib/raid6/recov_loongarch_simd.c
162
asm volatile("vst $vr5, %0" : "=m" (dq[16]));
lib/raid6/recov_loongarch_simd.c
163
asm volatile("vst $vr6, %0" : "=m" (dq[32]));
lib/raid6/recov_loongarch_simd.c
164
asm volatile("vst $vr7, %0" : "=m" (dq[48]));
lib/raid6/recov_loongarch_simd.c
180
dq += 64;
lib/raid6/recov_loongarch_simd.c
189
u8 *p, *q, *dq;
lib/raid6/recov_loongarch_simd.c
199
dq = (u8 *)ptrs[faila];
lib/raid6/recov_loongarch_simd.c
201
ptrs[disks - 1] = dq;
lib/raid6/recov_loongarch_simd.c
206
ptrs[faila] = dq;
lib/raid6/recov_loongarch_simd.c
225
asm volatile("vld $vr4, %0" : : "m" (dq[0]));
lib/raid6/recov_loongarch_simd.c
226
asm volatile("vld $vr5, %0" : : "m" (dq[16]));
lib/raid6/recov_loongarch_simd.c
227
asm volatile("vld $vr6, %0" : : "m" (dq[32]));
lib/raid6/recov_loongarch_simd.c
228
asm volatile("vld $vr7, %0" : : "m" (dq[48]));
lib/raid6/recov_loongarch_simd.c
264
asm volatile("vst $vr4, %0" : "=m" (dq[0]));
lib/raid6/recov_loongarch_simd.c
265
asm volatile("vst $vr5, %0" : "=m" (dq[16]));
lib/raid6/recov_loongarch_simd.c
266
asm volatile("vst $vr6, %0" : "=m" (dq[32]));
lib/raid6/recov_loongarch_simd.c
267
asm volatile("vst $vr7, %0" : "=m" (dq[48]));
lib/raid6/recov_loongarch_simd.c
282
dq += 64;
lib/raid6/recov_loongarch_simd.c
306
u8 *p, *q, *dp, *dq;
lib/raid6/recov_loongarch_simd.c
32
u8 *p, *q, *dp, *dq;
lib/raid6/recov_loongarch_simd.c
321
dq = (u8 *)ptrs[failb];
lib/raid6/recov_loongarch_simd.c
323
ptrs[disks - 1] = dq;
lib/raid6/recov_loongarch_simd.c
329
ptrs[failb] = dq;
lib/raid6/recov_loongarch_simd.c
357
asm volatile("xvld $xr4, %0" : : "m" (dq[0]));
lib/raid6/recov_loongarch_simd.c
358
asm volatile("xvld $xr5, %0" : : "m" (dq[32]));
lib/raid6/recov_loongarch_simd.c
410
asm volatile("xvst $xr0, %0" : "=m" (dq[0]));
lib/raid6/recov_loongarch_simd.c
411
asm volatile("xvst $xr1, %0" : "=m" (dq[32]));
lib/raid6/recov_loongarch_simd.c
419
dq += 64;
lib/raid6/recov_loongarch_simd.c
428
u8 *p, *q, *dq;
lib/raid6/recov_loongarch_simd.c
438
dq = (u8 *)ptrs[faila];
lib/raid6/recov_loongarch_simd.c
440
ptrs[disks - 1] = dq;
lib/raid6/recov_loongarch_simd.c
445
ptrs[faila] = dq;
lib/raid6/recov_loongarch_simd.c
464
asm volatile("xvld $xr2, %0" : : "m" (dq[0]));
lib/raid6/recov_loongarch_simd.c
465
asm volatile("xvld $xr3, %0" : : "m" (dq[32]));
lib/raid6/recov_loongarch_simd.c
47
dq = (u8 *)ptrs[failb];
lib/raid6/recov_loongarch_simd.c
49
ptrs[disks - 1] = dq;
lib/raid6/recov_loongarch_simd.c
492
asm volatile("xvst $xr2, %0" : "=m" (dq[0]));
lib/raid6/recov_loongarch_simd.c
493
asm volatile("xvst $xr3, %0" : "=m" (dq[32]));
lib/raid6/recov_loongarch_simd.c
500
dq += 64;
lib/raid6/recov_loongarch_simd.c
55
ptrs[failb] = dq;
lib/raid6/recov_loongarch_simd.c
81
asm volatile("vld $vr8, %0" : : "m" (dq[0]));
lib/raid6/recov_loongarch_simd.c
82
asm volatile("vld $vr9, %0" : : "m" (dq[16]));
lib/raid6/recov_loongarch_simd.c
83
asm volatile("vld $vr10, %0" : : "m" (dq[32]));
lib/raid6/recov_loongarch_simd.c
84
asm volatile("vld $vr11, %0" : : "m" (dq[48]));
lib/raid6/recov_neon.c
25
u8 *p, *q, *dp, *dq;
lib/raid6/recov_neon.c
40
dq = (u8 *)ptrs[failb];
lib/raid6/recov_neon.c
42
ptrs[disks - 1] = dq;
lib/raid6/recov_neon.c
48
ptrs[failb] = dq;
lib/raid6/recov_neon.c
58
__raid6_2data_recov_neon(bytes, p, q, dp, dq, pbmul, qmul);
lib/raid6/recov_neon.c
64
u8 *p, *q, *dq;
lib/raid6/recov_neon.c
74
dq = (u8 *)ptrs[faila];
lib/raid6/recov_neon.c
76
ptrs[disks - 1] = dq;
lib/raid6/recov_neon.c
81
ptrs[faila] = dq;
lib/raid6/recov_neon.c
88
__raid6_datap_recov_neon(bytes, p, q, dq, qmul);
lib/raid6/recov_neon_inner.c
103
vst1q_u8(dq, vx);
lib/raid6/recov_neon_inner.c
109
dq += 16;
lib/raid6/recov_neon_inner.c
29
uint8_t *dq, const uint8_t *pbmul,
lib/raid6/recov_neon_inner.c
54
vx = veorq_u8(vld1q_u8(q), vld1q_u8(dq));
lib/raid6/recov_neon_inner.c
67
vst1q_u8(dq, db);
lib/raid6/recov_neon_inner.c
74
dq += 16;
lib/raid6/recov_neon_inner.c
78
void __raid6_datap_recov_neon(int bytes, uint8_t *p, uint8_t *q, uint8_t *dq,
lib/raid6/recov_neon_inner.c
95
vx = veorq_u8(vld1q_u8(q), vld1q_u8(dq));
lib/raid6/recov_rvv.c
11
u8 *dq, const u8 *pbmul,
lib/raid6/recov_rvv.c
133
[dq]"r"(dq),
lib/raid6/recov_rvv.c
141
dq += 16;
lib/raid6/recov_rvv.c
148
u8 *p, *q, *dp, *dq;
lib/raid6/recov_rvv.c
163
dq = (u8 *)ptrs[failb];
lib/raid6/recov_rvv.c
165
ptrs[disks - 1] = dq;
lib/raid6/recov_rvv.c
171
ptrs[failb] = dq;
lib/raid6/recov_rvv.c
181
__raid6_2data_recov_rvv(bytes, p, q, dp, dq, pbmul, qmul);
lib/raid6/recov_rvv.c
188
u8 *p, *q, *dq;
lib/raid6/recov_rvv.c
198
dq = (u8 *)ptrs[faila];
lib/raid6/recov_rvv.c
200
ptrs[disks - 1] = dq;
lib/raid6/recov_rvv.c
205
ptrs[faila] = dq;
lib/raid6/recov_rvv.c
212
__raid6_datap_recov_rvv(bytes, p, q, dq, qmul);
lib/raid6/recov_rvv.c
74
[dq]"r"(dq),
lib/raid6/recov_rvv.c
85
dq += 16;
lib/raid6/recov_rvv.c
90
u8 *dq, const u8 *qmul)
lib/raid6/recov_s390xc.c
100
dq[i] = qmul[dq[i]];
lib/raid6/recov_s390xc.c
101
xor_block(p, dq);
lib/raid6/recov_s390xc.c
104
dq += 256;
lib/raid6/recov_s390xc.c
25
u8 *p, *q, *dp, *dq;
lib/raid6/recov_s390xc.c
39
dq = (u8 *)ptrs[failb];
lib/raid6/recov_s390xc.c
41
ptrs[disks-1] = dq;
lib/raid6/recov_s390xc.c
47
ptrs[failb] = dq;
lib/raid6/recov_s390xc.c
58
xor_block(dq, q);
lib/raid6/recov_s390xc.c
60
dq[i] = pbmul[dp[i]] ^ qmul[dq[i]];
lib/raid6/recov_s390xc.c
61
xor_block(dp, dq);
lib/raid6/recov_s390xc.c
65
dq += 256;
lib/raid6/recov_s390xc.c
74
u8 *p, *q, *dq;
lib/raid6/recov_s390xc.c
83
dq = (u8 *)ptrs[faila];
lib/raid6/recov_s390xc.c
85
ptrs[disks-1] = dq;
lib/raid6/recov_s390xc.c
90
ptrs[faila] = dq;
lib/raid6/recov_s390xc.c
98
xor_block(dq, q);
lib/raid6/recov_ssse3.c
124
asm volatile("movdqa %%xmm1,%0" : "=m" (dq[0]));
lib/raid6/recov_ssse3.c
125
asm volatile("movdqa %%xmm9,%0" : "=m" (dq[16]));
lib/raid6/recov_ssse3.c
136
dq += 32;
lib/raid6/recov_ssse3.c
140
asm volatile("pxor %0,%%xmm1" : : "m" (*dq));
lib/raid6/recov_ssse3.c
174
asm volatile("movdqa %%xmm1,%0" : "=m" (*dq));
lib/raid6/recov_ssse3.c
183
dq += 16;
lib/raid6/recov_ssse3.c
19
u8 *p, *q, *dp, *dq;
lib/raid6/recov_ssse3.c
194
u8 *p, *q, *dq;
lib/raid6/recov_ssse3.c
205
dq = (u8 *)ptrs[faila];
lib/raid6/recov_ssse3.c
207
ptrs[disks-1] = dq;
lib/raid6/recov_ssse3.c
212
ptrs[faila] = dq;
lib/raid6/recov_ssse3.c
224
asm volatile("movdqa %0, %%xmm3" : : "m" (dq[0]));
lib/raid6/recov_ssse3.c
225
asm volatile("movdqa %0, %%xmm4" : : "m" (dq[16]));
lib/raid6/recov_ssse3.c
271
asm volatile("movdqa %%xmm1, %0" : "=m" (dq[0]));
lib/raid6/recov_ssse3.c
272
asm volatile("movdqa %%xmm11, %0" : "=m" (dq[16]));
lib/raid6/recov_ssse3.c
280
dq += 32;
lib/raid6/recov_ssse3.c
283
asm volatile("movdqa %0, %%xmm3" : : "m" (dq[0]));
lib/raid6/recov_ssse3.c
305
asm volatile("movdqa %%xmm1, %0" : "=m" (dq[0]));
lib/raid6/recov_ssse3.c
311
dq += 16;
lib/raid6/recov_ssse3.c
35
dq = (u8 *)ptrs[failb];
lib/raid6/recov_ssse3.c
37
ptrs[disks-1] = dq;
lib/raid6/recov_ssse3.c
43
ptrs[failb] = dq;
lib/raid6/recov_ssse3.c
71
asm volatile("pxor %0,%%xmm1" : : "m" (dq[0]));
lib/raid6/recov_ssse3.c
72
asm volatile("pxor %0,%%xmm9" : : "m" (dq[16]));