arch/mips/cavium-octeon/executive/cvmx-helper-rgmii.c
325
int queue = cvmx_pko_get_base_queue(ipd_port) + i;
arch/mips/cavium-octeon/executive/cvmx-helper-rgmii.c
326
cvmx_write_csr(CVMX_PKO_REG_READ_IDX, queue);
arch/mips/cavium-octeon/executive/cvmx-helper-rgmii.c
329
pko_mem_queue_qos.s.qid = queue;
arch/mips/cavium-octeon/executive/cvmx-helper-rgmii.c
437
int queue = cvmx_pko_get_base_queue(ipd_port) + i;
arch/mips/cavium-octeon/executive/cvmx-helper-rgmii.c
438
cvmx_write_csr(CVMX_PKO_REG_READ_IDX, queue);
arch/mips/cavium-octeon/executive/cvmx-helper-util.c
107
cvmx_write_csr(CVMX_IPD_QOSX_RED_MARKS(queue), red_marks.u64);
arch/mips/cavium-octeon/executive/cvmx-helper-util.c
116
cvmx_write_csr(CVMX_IPD_RED_QUEX_PARAM(queue), red_param.u64);
arch/mips/cavium-octeon/executive/cvmx-helper-util.c
136
int queue;
arch/mips/cavium-octeon/executive/cvmx-helper-util.c
151
for (queue = 0; queue < 8; queue++)
arch/mips/cavium-octeon/executive/cvmx-helper-util.c
152
cvmx_helper_setup_red_queue(queue, pass_thresh, drop_thresh);
arch/mips/cavium-octeon/executive/cvmx-helper-util.c
95
static int cvmx_helper_setup_red_queue(int queue, int pass_thresh,
arch/mips/cavium-octeon/executive/cvmx-pko.c
101
num_queues, queue);
arch/mips/cavium-octeon/executive/cvmx-pko.c
104
CVMX_CMD_QUEUE_PKO(base_queue + queue));
arch/mips/cavium-octeon/executive/cvmx-pko.c
284
int queue;
arch/mips/cavium-octeon/executive/cvmx-pko.c
288
for (queue = 0; queue < CVMX_PKO_MAX_OUTPUT_QUEUES; queue++) {
arch/mips/cavium-octeon/executive/cvmx-pko.c
293
config.s.queue = queue & 0x7f;
arch/mips/cavium-octeon/executive/cvmx-pko.c
299
config1.s.qid7 = queue >> 7;
arch/mips/cavium-octeon/executive/cvmx-pko.c
303
cvmx_cmd_queue_shutdown(CVMX_CMD_QUEUE_PKO(queue));
arch/mips/cavium-octeon/executive/cvmx-pko.c
331
uint64_t queue;
arch/mips/cavium-octeon/executive/cvmx-pko.c
360
for (queue = 0; queue < num_queues; queue++) {
arch/mips/cavium-octeon/executive/cvmx-pko.c
363
&& priority[queue] ==
arch/mips/cavium-octeon/executive/cvmx-pko.c
365
static_priority_base = queue;
arch/mips/cavium-octeon/executive/cvmx-pko.c
369
&& priority[queue] != CVMX_PKO_QUEUE_STATIC_PRIORITY
arch/mips/cavium-octeon/executive/cvmx-pko.c
370
&& queue)
arch/mips/cavium-octeon/executive/cvmx-pko.c
371
static_priority_end = queue - 1;
arch/mips/cavium-octeon/executive/cvmx-pko.c
374
&& queue == num_queues - 1)
arch/mips/cavium-octeon/executive/cvmx-pko.c
376
static_priority_end = queue;
arch/mips/cavium-octeon/executive/cvmx-pko.c
384
&& (int)queue > static_priority_end
arch/mips/cavium-octeon/executive/cvmx-pko.c
385
&& priority[queue] ==
arch/mips/cavium-octeon/executive/cvmx-pko.c
391
(int)queue, static_priority_end);
arch/mips/cavium-octeon/executive/cvmx-pko.c
422
for (queue = 0; queue < num_queues; queue++) {
arch/mips/cavium-octeon/executive/cvmx-pko.c
426
config1.s.idx3 = queue >> 3;
arch/mips/cavium-octeon/executive/cvmx-pko.c
427
config1.s.qid7 = (base_queue + queue) >> 7;
arch/mips/cavium-octeon/executive/cvmx-pko.c
430
config.s.tail = queue == (num_queues - 1);
arch/mips/cavium-octeon/executive/cvmx-pko.c
431
config.s.index = queue;
arch/mips/cavium-octeon/executive/cvmx-pko.c
433
config.s.queue = base_queue + queue;
arch/mips/cavium-octeon/executive/cvmx-pko.c
437
config.s.static_q = (int)queue <= static_priority_end;
arch/mips/cavium-octeon/executive/cvmx-pko.c
438
config.s.s_tail = (int)queue == static_priority_end;
arch/mips/cavium-octeon/executive/cvmx-pko.c
445
switch ((int)priority[queue]) {
arch/mips/cavium-octeon/executive/cvmx-pko.c
482
(unsigned long long)priority[queue]);
arch/mips/cavium-octeon/executive/cvmx-pko.c
491
(base_queue + queue),
arch/mips/cavium-octeon/executive/cvmx-pko.c
521
(base_queue + queue));
arch/mips/cavium-octeon/executive/cvmx-pko.c
70
int queue;
arch/mips/cavium-octeon/executive/cvmx-pko.c
76
for (queue = 0; queue < num_queues; queue++) {
arch/mips/cavium-octeon/executive/cvmx-pko.c
82
config.s.index = queue;
arch/mips/cavium-octeon/executive/cvmx-pko.c
83
config.s.qid = base_queue + queue;
arch/mips/cavium-octeon/executive/cvmx-pko.c
85
config.s.tail = (queue == (num_queues - 1));
arch/mips/cavium-octeon/executive/cvmx-pko.c
86
config.s.s_tail = (queue == static_priority_end);
arch/mips/cavium-octeon/executive/cvmx-pko.c
88
config.s.static_q = (queue <= static_priority_end);
arch/mips/cavium-octeon/executive/cvmx-pko.c
92
CVMX_CMD_QUEUE_PKO(base_queue + queue),
arch/mips/include/asm/octeon/cvmx-pko-defs.h
1176
uint64_t queue:7;
arch/mips/include/asm/octeon/cvmx-pko-defs.h
1178
uint64_t queue:7;
arch/mips/include/asm/octeon/cvmx-pko.h
151
uint64_t queue:9;
arch/mips/include/asm/octeon/cvmx-pko.h
156
uint64_t queue:9;
arch/mips/include/asm/octeon/cvmx-pko.h
324
static inline void cvmx_pko_doorbell(uint64_t port, uint64_t queue,
arch/mips/include/asm/octeon/cvmx-pko.h
334
ptr.s.queue = queue;
arch/mips/include/asm/octeon/cvmx-pko.h
376
static inline void cvmx_pko_send_packet_prepare(uint64_t port, uint64_t queue,
arch/mips/include/asm/octeon/cvmx-pko.h
394
(CVMX_TAG_SUBGROUP_MASK & queue);
arch/mips/include/asm/octeon/cvmx-pko.h
419
uint64_t queue,
arch/mips/include/asm/octeon/cvmx-pko.h
427
result = cvmx_cmd_queue_write2(CVMX_CMD_QUEUE_PKO(queue),
arch/mips/include/asm/octeon/cvmx-pko.h
431
cvmx_pko_doorbell(port, queue, 2);
arch/mips/include/asm/octeon/cvmx-pko.h
462
uint64_t queue,
arch/mips/include/asm/octeon/cvmx-pko.h
471
result = cvmx_cmd_queue_write3(CVMX_CMD_QUEUE_PKO(queue),
arch/mips/include/asm/octeon/cvmx-pko.h
475
cvmx_pko_doorbell(port, queue, 3);
arch/powerpc/include/asm/fsl_hcalls.h
492
static inline unsigned int fh_err_get_info(int queue, uint32_t *bufsize,
arch/powerpc/include/asm/fsl_hcalls.h
503
r3 = queue;
arch/powerpc/kvm/mpic.c
141
unsigned long queue[BITS_TO_LONGS((MAX_IRQ + 63) & ~63)];
arch/powerpc/kvm/mpic.c
283
set_bit(n_IRQ, q->queue);
arch/powerpc/kvm/mpic.c
288
clear_bit(n_IRQ, q->queue);
arch/powerpc/kvm/mpic.c
298
irq = find_next_bit(q->queue, opp->max_irq, irq + 1);
arch/powerpc/sysdev/xive/common.c
1475
if (xc->queue[xive_irq_priority].qpage)
arch/powerpc/sysdev/xive/common.c
1484
if (!xc->queue[xive_irq_priority].qpage)
arch/powerpc/sysdev/xive/common.c
164
irq = xive_read_eq(&xc->queue[prio], just_peek);
arch/powerpc/sysdev/xive/common.c
1822
xive_eq_debug_show_one(m, &xc->queue[xive_irq_priority],
arch/powerpc/sysdev/xive/common.c
189
q = &xc->queue[prio];
arch/powerpc/sysdev/xive/common.c
286
xive_dump_eq("EQ", &xc->queue[xive_irq_priority]);
arch/powerpc/sysdev/xive/common.c
506
struct xive_q *q = &xc->queue[xive_irq_priority];
arch/powerpc/sysdev/xive/common.c
530
struct xive_q *q = &xc->queue[xive_irq_priority];
arch/powerpc/sysdev/xive/native.c
218
struct xive_q *q = &xc->queue[prio];
arch/powerpc/sysdev/xive/native.c
231
struct xive_q *q = &xc->queue[prio];
arch/powerpc/sysdev/xive/spapr.c
534
struct xive_q *q = &xc->queue[prio];
arch/powerpc/sysdev/xive/spapr.c
548
struct xive_q *q = &xc->queue[prio];
arch/powerpc/sysdev/xive/xive-internal.h
27
struct xive_q queue[XIVE_MAX_QUEUES];
arch/um/drivers/ubd_kern.c
1290
struct ubd *ubd_dev = hctx->queue->queuedata;
arch/x86/kvm/x86.c
858
queue:
arch/x86/kvm/x86.c
896
goto queue;
block/bdev.c
491
bdev->bd_queue = disk->queue;
block/bfq-cgroup.c
1232
ret = blkcg_activate_policy(bfqd->queue->disk, &blkcg_policy_bfq);
block/bfq-cgroup.c
1236
return blkg_to_bfqg(bfqd->queue->root_blkg);
block/bfq-cgroup.c
940
list_for_each_entry(blkg, &bfqd->queue->blkg_list, q_node) {
block/bfq-iosched.c
234
(blk_queue_rot(bfqd->queue) || \
block/bfq-iosched.c
4168
if (!blk_queue_rot(bfqd->queue))
block/bfq-iosched.c
4490
blk_queue_rot(bfqd->queue) && !bfqd->hw_tag,
block/bfq-iosched.c
4524
((blk_queue_rot(bfqd->queue) || !bfqd->hw_tag) &&
block/bfq-iosched.c
4725
if (!blk_queue_rot(bfqd->queue) &&
block/bfq-iosched.c
473
blk_mq_run_hw_queues(bfqd->queue, true);
block/bfq-iosched.c
5115
bfq_dispatch_remove(bfqd->queue, rq);
block/bfq-iosched.c
5146
struct bfq_data *bfqd = hctx->queue->elevator->elevator_data;
block/bfq-iosched.c
5158
struct bfq_data *bfqd = hctx->queue->elevator->elevator_data;
block/bfq-iosched.c
5299
struct bfq_data *bfqd = hctx->queue->elevator->elevator_data;
block/bfq-iosched.c
5316
bfq_update_dispatch_stats(hctx->queue, rq,
block/bfq-iosched.c
5519
bdi_dev_name(bfqq->bfqd->queue->disk->bdi),
block/bfq-iosched.c
5851
bfqd->queue->node);
block/bfq-iosched.c
6239
struct request_queue *q = hctx->queue;
block/bfq-iosched.c
6343
!blk_queue_rot(bfqd->queue) && bfqd->hw_tag;
block/bfq-iosched.c
7164
blkcg_deactivate_policy(bfqd->queue->disk, &blkcg_policy_bfq);
block/bfq-iosched.c
7172
blk_stat_disable_accounting(bfqd->queue);
block/bfq-iosched.c
7173
blk_queue_flag_clear(QUEUE_FLAG_DISABLE_WBT_DEF, bfqd->queue);
block/bfq-iosched.c
7174
wbt_enable_default(bfqd->queue->disk);
block/bfq-iosched.c
7235
bfqd->queue = q;
block/bfq-iosched.c
7286
bfqd->nonrot_with_queueing = !blk_queue_rot(bfqd->queue);
block/bfq-iosched.c
7321
bfqd->rate_dur_prod = ref_rate[!blk_queue_rot(bfqd->queue)] *
block/bfq-iosched.c
7322
ref_wr_duration[!blk_queue_rot(bfqd->queue)];
block/bfq-iosched.c
7323
bfqd->peak_rate = ref_rate[!blk_queue_rot(bfqd->queue)] * 2 / 3;
block/bfq-iosched.h
1179
if (likely(!blk_trace_note_message_enabled((bfqd)->queue))) \
block/bfq-iosched.h
1182
blk_add_cgroup_trace_msg((bfqd)->queue, \
block/bfq-iosched.h
1191
if (likely(!blk_trace_note_message_enabled((bfqd)->queue))) \
block/bfq-iosched.h
1194
blk_add_trace_msg((bfqd)->queue, "%s " fmt, pid_str, ##args); \
block/bfq-iosched.h
1200
blk_add_trace_msg((bfqd)->queue, "bfq " fmt, ##args)
block/bfq-iosched.h
514
struct request_queue *queue;
block/bio.c
1281
if (bio->bi_bdev && blk_queue_pci_p2pdma(bio->bi_bdev->bd_disk->queue))
block/blk-cgroup.c
1146
struct blkcg_gq *blkg = bdev->bd_disk->queue->root_blkg;
block/blk-cgroup.c
1497
struct request_queue *q = disk->queue;
block/blk-cgroup.c
1580
struct request_queue *q = disk->queue;
block/blk-cgroup.c
1704
struct request_queue *q = disk->queue;
block/blk-cgroup.c
2012
blkg = blkg_lookup(blkcg, disk->queue);
block/blk-cgroup.c
305
blkg = kzalloc_node(sizeof(*blkg), gfp_mask, disk->queue->node);
block/blk-cgroup.c
313
if (!blk_get_queue(disk->queue))
block/blk-cgroup.c
316
blkg->q = disk->queue;
block/blk-cgroup.c
336
if (!blkcg_policy_enabled(disk->queue, pol))
block/blk-cgroup.c
355
blk_put_queue(disk->queue);
block/blk-cgroup.c
375
lockdep_assert_held(&disk->queue->queue_lock);
block/blk-cgroup.c
378
if (blk_queue_dying(disk->queue)) {
block/blk-cgroup.c
401
blkg->parent = blkg_lookup(blkcg_parent(blkcg), disk->queue);
block/blk-cgroup.c
419
ret = radix_tree_insert(&blkcg->blkg_tree, disk->queue->id, blkg);
block/blk-cgroup.c
422
list_add(&blkg->q_node, &disk->queue->blkg_list);
block/blk-cgroup.c
468
struct request_queue *q = disk->queue;
block/blk-cgroup.c
571
struct request_queue *q = disk->queue;
block/blk-cgroup.c
875
q = disk->queue;
block/blk-core.c
640
!(disk->queue->limits.features & BLK_FEAT_POLL)) {
block/blk-core.c
646
blk_queue_exit(disk->queue);
block/blk-crypto-sysfs.c
166
struct request_queue *q = disk->queue;
block/blk-crypto-sysfs.c
190
kobject_put(disk->queue->crypto_kobject);
block/blk-ia-ranges.c
111
struct request_queue *q = disk->queue;
block/blk-ia-ranges.c
153
struct request_queue *q = disk->queue;
block/blk-ia-ranges.c
267
GFP_KERNEL, disk->queue->node);
block/blk-ia-ranges.c
288
struct request_queue *q = disk->queue;
block/blk-integrity.c
200
return &dev_to_disk(dev)->queue->limits.integrity;
block/blk-integrity.c
229
struct request_queue *q = dev_to_disk(dev)->queue;
block/blk-iocost.c
3236
if (!queue_is_mq(disk->queue)) {
block/blk-iocost.c
3241
ioc = q_to_ioc(disk->queue);
block/blk-iocost.c
3246
ioc = q_to_ioc(disk->queue);
block/blk-iocost.c
3249
blk_mq_quiesce_queue(disk->queue);
block/blk-iocost.c
3323
blk_stat_enable_accounting(disk->queue);
block/blk-iocost.c
3324
blk_queue_flag_set(QUEUE_FLAG_RQ_ALLOC_TIME, disk->queue);
block/blk-iocost.c
3327
blk_stat_disable_accounting(disk->queue);
block/blk-iocost.c
3328
blk_queue_flag_clear(QUEUE_FLAG_RQ_ALLOC_TIME, disk->queue);
block/blk-iocost.c
3347
blk_mq_unquiesce_queue(disk->queue);
block/blk-iocost.c
3353
blk_mq_unquiesce_queue(disk->queue);
block/blk-iocost.c
815
if (blk_queue_rot(disk->queue))
block/blk-iocost.c
819
if (blk_queue_depth(disk->queue) == 1)
block/blk-iolatency.c
333
unsigned long qd = blkiolat->rqos.disk->queue->nr_requests;
block/blk-iolatency.c
375
unsigned long qd = iolat->blkiolat->rqos.disk->queue->nr_requests;
block/blk-iolatency.c
661
blkiolat->rqos.disk->queue->root_blkg) {
block/blk-iolatency.c
745
struct request_queue *q = blkiolat->rqos.disk->queue;
block/blk-iolatency.c
748
memflags = blk_mq_freeze_queue(blkiolat->rqos.disk->queue);
block/blk-iolatency.c
754
blk_mq_unfreeze_queue(blkiolat->rqos.disk->queue, memflags);
block/blk-mq-cpumap.c
113
unsigned int queue, cpu;
block/blk-mq-cpumap.c
118
for (queue = 0; queue < qmap->nr_queues; queue++) {
block/blk-mq-cpumap.c
119
mask = dev->bus->irq_get_affinity(dev, queue + offset);
block/blk-mq-cpumap.c
124
qmap->mq_map[cpu] = qmap->queue_offset + queue;
block/blk-mq-cpumap.c
62
unsigned int queue, cpu, nr_masks;
block/blk-mq-cpumap.c
71
for (queue = 0; queue < qmap->nr_queues; queue++) {
block/blk-mq-cpumap.c
72
for_each_cpu(cpu, &masks[queue % nr_masks])
block/blk-mq-cpumap.c
73
qmap->mq_map[cpu] = qmap->queue_offset + queue;
block/blk-mq-debugfs.c
356
res = mutex_lock_interruptible(&hctx->queue->elevator_lock);
block/blk-mq-debugfs.c
359
blk_mq_tagset_busy_iter(hctx->queue->tag_set, hctx_show_busy_rq,
block/blk-mq-debugfs.c
361
mutex_unlock(&hctx->queue->elevator_lock);
block/blk-mq-debugfs.c
409
struct request_queue *q = hctx->queue;
block/blk-mq-debugfs.c
425
struct request_queue *q = hctx->queue;
block/blk-mq-debugfs.c
441
struct request_queue *q = hctx->queue;
block/blk-mq-debugfs.c
457
struct request_queue *q = hctx->queue;
block/blk-mq-debugfs.c
655
debugfs_create_files(hctx->queue, ctx_dir, ctx,
block/blk-mq-debugfs.c
681
if (!hctx->queue->debugfs_dir)
block/blk-mq-debugfs.c
753
struct request_queue *q = rqos->disk->queue;
block/blk-mq-debugfs.c
810
lockdep_assert_held(&hctx->queue->debugfs_mutex);
block/blk-mq-debugfs.c
812
if (!hctx->queue->debugfs_dir)
block/blk-mq-sched.c
215
struct request_queue *q = hctx->queue;
block/blk-mq-sched.c
306
if (hctx->queue->elevator)
block/blk-mq-sched.c
319
struct request_queue *q = hctx->queue;
block/blk-mq-sched.c
87
struct request_queue *q = hctx->queue;
block/blk-mq-sched.c
98
max_dispatch = hctx->queue->nr_requests;
block/blk-mq-sched.h
116
struct elevator_queue *e = hctx->queue->elevator;
block/blk-mq-sysfs.c
161
struct request_queue *q = hctx->queue;
block/blk-mq-sysfs.c
222
struct request_queue *q = disk->queue;
block/blk-mq-sysfs.c
256
struct request_queue *q = disk->queue;
block/blk-mq-sysfs.c
58
q = hctx->queue;
block/blk-mq-tag.c
52
struct request_queue *q = hctx->queue;
block/blk-mq-tag.c
90
struct request_queue *q = hctx->queue;
block/blk-mq.c
1189
struct request_queue *q = hctx->queue;
block/blk-mq.c
2107
if (hctx->queue->mq_ops->commit_rqs && queued) {
block/blk-mq.c
2108
trace_block_unplug(hctx->queue, queued, !from_schedule);
block/blk-mq.c
2109
hctx->queue->mq_ops->commit_rqs(hctx);
block/blk-mq.c
2120
struct request_queue *q = hctx->queue;
block/blk-mq.c
2274
if (hctx->queue->nr_hw_queues == 1 || blk_mq_hctx_empty_cpumask(hctx))
block/blk-mq.c
2337
__blk_mq_run_dispatch_ops(hctx->queue, false,
block/blk-mq.c
2338
need_run = !blk_queue_quiesced(hctx->queue) &&
block/blk-mq.c
2373
spin_lock_irqsave(&hctx->queue->queue_lock, flags);
block/blk-mq.c
2375
spin_unlock_irqrestore(&hctx->queue->queue_lock, flags);
block/blk-mq.c
2386
blk_mq_run_dispatch_ops(hctx->queue,
block/blk-mq.c
2562
blk_mq_run_dispatch_ops(hctx->queue,
block/blk-mq.c
2598
blk_mq_run_dispatch_ops(hctx->queue,
block/blk-mq.c
2937
trace_block_unplug(this_hctx->queue, depth, !from_sched);
block/blk-mq.c
2939
percpu_ref_get(&this_hctx->queue->q_usage_counter);
block/blk-mq.c
2946
} else if (this_hctx->queue->elevator) {
block/blk-mq.c
2947
this_hctx->queue->elevator->type->ops.insert_requests(this_hctx,
block/blk-mq.c
2953
percpu_ref_put(&this_hctx->queue->q_usage_counter);
block/blk-mq.c
3695
srcu_idx = srcu_read_lock(&hctx->queue->tag_set->tags_srcu);
block/blk-mq.c
3697
srcu_read_unlock(&hctx->queue->tag_set->tags_srcu, srcu_idx);
block/blk-mq.c
3714
struct blk_mq_hw_ctx *h = blk_mq_map_queue_type(hctx->queue,
block/blk-mq.c
3752
if (percpu_ref_tryget(&hctx->queue->q_usage_counter)) {
block/blk-mq.c
3767
percpu_ref_put(&hctx->queue->q_usage_counter);
block/blk-mq.c
3783
struct blk_mq_hw_ctx *mapped_hctx = blk_mq_map_queue_type(hctx->queue,
block/blk-mq.c
3815
ctx = __blk_mq_get_ctx(hctx->queue, cpu);
block/blk-mq.c
4050
hctx->queue = q;
block/blk-mq.c
4103
__ctx->queue = q;
block/blk-mq.h
29
struct request_queue *queue;
block/blk-mq.h
298
atomic_add(val, &hctx->queue->nr_active_requests_shared_tags);
block/blk-mq.h
312
atomic_sub(val, &hctx->queue->nr_active_requests_shared_tags);
block/blk-mq.h
351
return atomic_read(&hctx->queue->nr_active_requests_shared_tags);
block/blk-mq.h
418
struct request_queue *q = hctx->queue;
block/blk-rq-qos.c
328
struct request_queue *q = disk->queue;
block/blk-rq-qos.c
358
struct request_queue *q = rqos->disk->queue;
block/blk-sysfs.c
138
guard(mutex)(&disk->queue->elevator_lock);
block/blk-sysfs.c
140
return queue_var_show(disk->queue->async_depth, page);
block/blk-sysfs.c
146
struct request_queue *q = disk->queue;
block/blk-sysfs.c
180
mutex_lock(&disk->queue->limits_lock);
block/blk-sysfs.c
182
mutex_unlock(&disk->queue->limits_lock);
block/blk-sysfs.c
192
struct request_queue *q = disk->queue;
block/blk-sysfs.c
215
return queue_var_show(disk->queue->limits._field, page); \
block/blk-sysfs.c
242
(unsigned long long)disk->queue->limits._field << \
block/blk-sysfs.c
258
return queue_var_show(disk->queue->limits._field >> 1, page); \
block/blk-sysfs.c
285
if (max_discard_bytes & (disk->queue->limits.discard_granularity - 1))
block/blk-sysfs.c
349
!!(disk->queue->limits.features & _feature)); \
block/blk-sysfs.c
366
!!(disk->queue->limits.features & _feature)); \
block/blk-sysfs.c
374
if (queue_is_mq(disk->queue))
block/blk-sysfs.c
375
return sysfs_emit(page, "%u\n", blk_mq_can_poll(disk->queue));
block/blk-sysfs.c
378
!!(disk->queue->limits.features & BLK_FEAT_POLL));
block/blk-sysfs.c
383
if (blk_queue_is_zoned(disk->queue))
block/blk-sysfs.c
395
return queue_var_show(!!blk_queue_passthrough_stat(disk->queue), page);
block/blk-sysfs.c
417
return queue_var_show((blk_queue_nomerges(disk->queue) << 1) |
block/blk-sysfs.c
418
blk_queue_noxmerges(disk->queue), page);
block/blk-sysfs.c
425
struct request_queue *q = disk->queue;
block/blk-sysfs.c
443
bool set = test_bit(QUEUE_FLAG_SAME_COMP, &disk->queue->queue_flags);
block/blk-sysfs.c
444
bool force = test_bit(QUEUE_FLAG_SAME_FORCE, &disk->queue->queue_flags);
block/blk-sysfs.c
454
struct request_queue *q = disk->queue;
block/blk-sysfs.c
491
struct request_queue *q = disk->queue;
block/blk-sysfs.c
507
jiffies_to_msecs(READ_ONCE(disk->queue->rq_timeout)));
block/blk-sysfs.c
515
struct request_queue *q = disk->queue;
block/blk-sysfs.c
528
if (blk_queue_write_cache(disk->queue))
block/blk-sysfs.c
58
mutex_lock(&disk->queue->elevator_lock);
block/blk-sysfs.c
59
ret = queue_var_show(disk->queue->nr_requests, page);
block/blk-sysfs.c
60
mutex_unlock(&disk->queue->elevator_lock);
block/blk-sysfs.c
664
struct request_queue *q = disk->queue;
block/blk-sysfs.c
67
struct request_queue *q = disk->queue;
block/blk-sysfs.c
786
struct request_queue *q = disk->queue;
block/blk-sysfs.c
800
struct request_queue *q = disk->queue;
block/blk-sysfs.c
835
mutex_lock(&disk->queue->limits_lock);
block/blk-sysfs.c
837
mutex_unlock(&disk->queue->limits_lock);
block/blk-sysfs.c
850
struct request_queue *q = disk->queue;
block/blk-sysfs.c
899
struct request_queue *q = disk->queue;
block/blk-sysfs.c
916
struct request_queue *q = disk->queue;
block/blk-sysfs.c
992
struct request_queue *q = disk->queue;
block/blk-throttle.c
1134
q = td->queue;
block/blk-throttle.c
118
if (likely(!blk_trace_note_message_enabled(__td->queue))) \
block/blk-throttle.c
1203
struct request_queue *q = td->queue;
block/blk-throttle.c
121
blk_add_cgroup_trace_msg(__td->queue, \
block/blk-throttle.c
124
blk_add_trace_msg(__td->queue, "throtl " fmt, ##args); \
block/blk-throttle.c
1281
global ? tg->td->queue->root_blkg : tg_to_blkg(tg)) {
block/blk-throttle.c
1311
struct request_queue *q = disk->queue;
block/blk-throttle.c
1323
memflags = blk_mq_freeze_queue(disk->queue);
block/blk-throttle.c
1324
blk_mq_quiesce_queue(disk->queue);
block/blk-throttle.c
1327
td->queue = q;
block/blk-throttle.c
1336
blk_mq_unquiesce_queue(disk->queue);
block/blk-throttle.c
1337
blk_mq_unfreeze_queue(disk->queue, memflags);
block/blk-throttle.c
1673
struct request_queue *q = disk->queue;
block/blk-throttle.c
1826
struct request_queue *q = disk->queue;
block/blk-throttle.c
36
struct request_queue *queue;
block/blk-timeout.c
43
int set = test_bit(QUEUE_FLAG_FAIL_IO, &disk->queue->queue_flags);
block/blk-timeout.c
55
struct request_queue *q = disk->queue;
block/blk-wbt.c
212
else if (blk_queue_write_cache(rwb->rqos.disk->queue) &&
block/blk-wbt.c
741
struct request_queue *q = disk->queue;
block/blk-wbt.c
777
struct request_queue *q = disk->queue;
block/blk-wbt.c
811
RQWB(rqos)->rq_depth.queue_depth = blk_queue_depth(rqos->disk->queue);
block/blk-wbt.c
819
blk_stat_remove_callback(rqos->disk->queue, rwb->cb);
block/blk-wbt.c
828
struct rq_qos *rqos = wbt_rq_qos(disk->queue);
block/blk-wbt.c
946
struct request_queue *q = disk->queue;
block/blk-wbt.c
976
struct request_queue *q = disk->queue;
block/blk-zoned.c
1220
percpu_ref_get(&bio->bi_bdev->bd_disk->queue->q_usage_counter);
block/blk-zoned.c
1242
trace_disk_zone_wplug_add_bio(zwplug->disk->queue, zwplug->zone_no,
block/blk-zoned.c
1752
trace_blk_zone_wplug_bio(zwplug->disk->queue, zwplug->zone_no,
block/blk-zoned.c
1773
blk_queue_exit(bdev->bd_disk->queue);
block/blk-zoned.c
1911
struct queue_limits *lim = &disk->queue->limits;
block/blk-zoned.c
1948
struct request_queue *q = disk->queue;
block/blk-zoned.c
2128
sector_t zone_sectors = disk->queue->limits.chunk_sectors;
block/blk-zoned.c
2207
struct request_queue *q = disk->queue;
block/blk-zoned.c
728
struct request_queue *q = zwplug->disk->queue;
block/blk-zoned.c
86
return queue_is_mq(disk->queue) ||
block/blk-zoned.c
87
queue_emulates_zone_append(disk->queue);
block/bsg-lib.c
275
struct request_queue *q = hctx->queue;
block/bsg.c
107
struct request_queue *q = bd->queue;
block/bsg.c
180
struct gendisk *disk = bd->queue->disk;
block/bsg.c
200
bd->queue = q;
block/bsg.c
24
struct request_queue *queue;
block/bsg.c
67
ret = bd->sg_io_fn(bd->queue, &hdr, open_for_write,
block/bsg.c
76
if (!blk_get_queue(to_bsg_device(inode)->queue))
block/bsg.c
83
blk_put_queue(to_bsg_device(inode)->queue);
block/elevator.c
792
struct request_queue *q = disk->queue;
block/elevator.c
837
struct request_queue *q = disk->queue;
block/genhd.c
1283
blk_trace_remove(disk->queue);
block/genhd.c
1292
if (queue_is_mq(disk->queue) &&
block/genhd.c
1295
blk_mq_exit_queue(disk->queue);
block/genhd.c
1307
disk->queue->disk = NULL;
block/genhd.c
1308
blk_put_queue(disk->queue);
block/genhd.c
1463
disk->queue = q;
block/genhd.c
426
blk_apply_bdi_limits(disk->bdi, &disk->queue->limits);
block/genhd.c
442
if (queue_is_mq(disk->queue)) {
block/genhd.c
557
rq_qos_exit(disk->queue);
block/genhd.c
593
if (queue_is_mq(disk->queue)) {
block/genhd.c
594
set = disk->queue->tag_set;
block/genhd.c
666
blk_queue_flag_set(QUEUE_FLAG_DYING, disk->queue);
block/genhd.c
676
return blk_queue_start_drain(disk->queue);
block/genhd.c
695
struct request_queue *q = disk->queue;
block/genhd.c
814
if (!queue_is_mq(disk->queue)) {
block/genhd.c
817
set = disk->queue->tag_set;
block/genhd.c
819
disable_elv_switch(disk->queue);
block/kyber-iosched.c
565
merged = blk_bio_list_merge(hctx->queue, rq_list, bio, nr_segs);
block/kyber-iosched.c
791
struct kyber_queue_data *kqd = hctx->queue->elevator->elevator_data;
block/mq-deadline.c
454
struct deadline_data *dd = hctx->queue->elevator->elevator_data;
block/mq-deadline.c
626
struct request_queue *q = hctx->queue;
block/mq-deadline.c
674
struct request_queue *q = hctx->queue;
block/mq-deadline.c
721
struct deadline_data *dd = hctx->queue->elevator->elevator_data;
block/partitions/atari.c
50
if (queue_logical_block_size(state->disk->queue) != 512)
block/partitions/efi.c
137
queue_logical_block_size(disk->queue)) - 1ULL;
block/partitions/efi.c
239
(queue_logical_block_size(state->disk->queue) / 512);
block/partitions/efi.c
309
unsigned ssz = queue_logical_block_size(state->disk->queue);
block/partitions/efi.c
356
queue_logical_block_size(state->disk->queue)) {
block/partitions/efi.c
359
queue_logical_block_size(state->disk->queue));
block/partitions/efi.c
717
unsigned ssz = queue_logical_block_size(state->disk->queue) / 512;
block/partitions/msdos.c
143
sector_size = queue_logical_block_size(state->disk->queue) / 512;
block/partitions/msdos.c
591
sector_size = queue_logical_block_size(state->disk->queue) / 512;
crypto/algapi.c
939
void crypto_init_queue(struct crypto_queue *queue, unsigned int max_qlen)
crypto/algapi.c
941
INIT_LIST_HEAD(&queue->list);
crypto/algapi.c
942
queue->backlog = &queue->list;
crypto/algapi.c
943
queue->qlen = 0;
crypto/algapi.c
944
queue->max_qlen = max_qlen;
crypto/algapi.c
948
int crypto_enqueue_request(struct crypto_queue *queue,
crypto/algapi.c
953
if (unlikely(queue->qlen >= queue->max_qlen)) {
crypto/algapi.c
959
if (queue->backlog == &queue->list)
crypto/algapi.c
960
queue->backlog = &request->list;
crypto/algapi.c
963
queue->qlen++;
crypto/algapi.c
964
list_add_tail(&request->list, &queue->list);
crypto/algapi.c
971
void crypto_enqueue_request_head(struct crypto_queue *queue,
crypto/algapi.c
974
if (unlikely(queue->qlen >= queue->max_qlen))
crypto/algapi.c
975
queue->backlog = queue->backlog->prev;
crypto/algapi.c
977
queue->qlen++;
crypto/algapi.c
978
list_add(&request->list, &queue->list);
crypto/algapi.c
982
struct crypto_async_request *crypto_dequeue_request(struct crypto_queue *queue)
crypto/algapi.c
986
if (unlikely(!queue->qlen))
crypto/algapi.c
989
queue->qlen--;
crypto/algapi.c
991
if (queue->backlog != &queue->list)
crypto/algapi.c
992
queue->backlog = queue->backlog->next;
crypto/algapi.c
994
request = queue->list.next;
crypto/cryptd.c
101
static int cryptd_init_queue(struct cryptd_queue *queue,
crypto/cryptd.c
107
queue->cpu_queue = alloc_percpu(struct cryptd_cpu_queue);
crypto/cryptd.c
108
if (!queue->cpu_queue)
crypto/cryptd.c
111
cpu_queue = per_cpu_ptr(queue->cpu_queue, cpu);
crypto/cryptd.c
112
crypto_init_queue(&cpu_queue->queue, max_cpu_qlen);
crypto/cryptd.c
1124
err = cryptd_init_queue(&queue, cryptd_max_cpu_qlen);
crypto/cryptd.c
1135
cryptd_fini_queue(&queue);
crypto/cryptd.c
1144
cryptd_fini_queue(&queue);
crypto/cryptd.c
120
static void cryptd_fini_queue(struct cryptd_queue *queue)
crypto/cryptd.c
126
cpu_queue = per_cpu_ptr(queue->cpu_queue, cpu);
crypto/cryptd.c
127
BUG_ON(cpu_queue->queue.qlen);
crypto/cryptd.c
129
free_percpu(queue->cpu_queue);
crypto/cryptd.c
132
static int cryptd_enqueue_request(struct cryptd_queue *queue,
crypto/cryptd.c
140
local_lock_nested_bh(&queue->cpu_queue->bh_lock);
crypto/cryptd.c
141
cpu_queue = this_cpu_ptr(queue->cpu_queue);
crypto/cryptd.c
142
err = crypto_enqueue_request(&cpu_queue->queue, request);
crypto/cryptd.c
157
local_unlock_nested_bh(&queue->cpu_queue->bh_lock);
crypto/cryptd.c
177
backlog = crypto_get_backlog(&cpu_queue->queue);
crypto/cryptd.c
178
req = crypto_dequeue_request(&cpu_queue->queue);
crypto/cryptd.c
189
if (cpu_queue->queue.qlen)
crypto/cryptd.c
197
return ictx->queue;
crypto/cryptd.c
324
struct cryptd_queue *queue;
crypto/cryptd.c
326
queue = cryptd_get_queue(crypto_skcipher_tfm(tfm));
crypto/cryptd.c
332
return cryptd_enqueue_request(queue, &req->base);
crypto/cryptd.c
38
struct crypto_queue queue;
crypto/cryptd.c
382
struct cryptd_queue *queue)
crypto/cryptd.c
398
ctx->queue = queue;
crypto/cryptd.c
494
struct cryptd_queue *queue =
crypto/cryptd.c
502
return cryptd_enqueue_request(queue, &req->base);
crypto/cryptd.c
52
struct cryptd_queue *queue;
crypto/cryptd.c
57
struct cryptd_queue *queue;
crypto/cryptd.c
62
struct cryptd_queue *queue;
crypto/cryptd.c
666
struct cryptd_queue *queue)
crypto/cryptd.c
67
struct cryptd_queue *queue;
crypto/cryptd.c
682
ctx->queue = queue;
crypto/cryptd.c
819
struct cryptd_queue *queue = cryptd_get_queue(crypto_aead_tfm(tfm));
crypto/cryptd.c
826
return cryptd_enqueue_request(queue, &req->base);
crypto/cryptd.c
875
struct cryptd_queue *queue)
crypto/cryptd.c
891
ctx->queue = queue;
crypto/cryptd.c
927
static struct cryptd_queue queue;
crypto/cryptd.c
939
return cryptd_create_skcipher(tmpl, tb, algt, &queue);
crypto/cryptd.c
941
return cryptd_create_hash(tmpl, tb, algt, &queue);
crypto/cryptd.c
943
return cryptd_create_aead(tmpl, tb, algt, &queue);
crypto/crypto_engine.c
103
backlog = crypto_get_backlog(&engine->queue);
crypto/crypto_engine.c
104
async_req = crypto_dequeue_request(&engine->queue);
crypto/crypto_engine.c
146
crypto_enqueue_request_head(&engine->queue, async_req);
crypto/crypto_engine.c
202
ret = crypto_enqueue_request(&engine->queue, req);
crypto/crypto_engine.c
401
while ((crypto_queue_len(&engine->queue) || engine->busy) && limit--) {
crypto/crypto_engine.c
407
if (crypto_queue_len(&engine->queue) || engine->busy)
crypto/crypto_engine.c
457
crypto_init_queue(&engine->queue, qlen);
crypto/crypto_engine.c
86
if (!crypto_queue_len(&engine->queue) || !engine->running) {
drivers/accel/habanalabs/common/firmware_if.c
407
struct hl_hw_queue *queue = &hdev->kernel_queues[hw_queue_id];
drivers/accel/habanalabs/common/firmware_if.c
436
pi = queue->pi;
drivers/accel/habanalabs/common/firmware_if.c
446
hl_hw_queue_submit_bd(hdev, queue, hl_queue_inc_ptr(queue->pi), size, pkt_dma_addr);
drivers/accel/habanalabs/common/firmware_if.c
449
expected_ack_val = queue->pi;
drivers/accel/habanalabs/common/firmware_if.c
467
struct hl_bd *bd = queue->kernel_address;
drivers/accel/habanalabs/common/firmware_if.c
474
queue->dram_bd);
drivers/accel/habanalabs/common/firmware_if.c
530
sent_bd = queue->kernel_address;
drivers/accel/habanalabs/common/irq.c
102
struct hl_hw_queue *queue;
drivers/accel/habanalabs/common/irq.c
115
queue = &hdev->kernel_queues[job->hw_queue_id];
drivers/accel/habanalabs/common/irq.c
116
atomic_inc(&queue->ci);
drivers/accel/habanalabs/common/irq.c
80
struct hl_hw_queue *queue;
drivers/accel/habanalabs/common/irq.c
83
queue = &hdev->kernel_queues[cq->hw_queue_id];
drivers/accel/habanalabs/common/irq.c
84
job = queue->shadow_queue[hl_pi_2_offset(cs_seq)];
drivers/accel/habanalabs/common/irq.c
88
atomic_inc(&queue->ci);
drivers/atm/idt77252.c
100
static void add_rx_skb(struct idt77252_dev *, int queue,
drivers/atm/idt77252.c
1119
__skb_queue_tail(&rpp->queue, skb);
drivers/atm/idt77252.c
1147
if (skb_queue_len(&rpp->queue) > 1) {
drivers/atm/idt77252.c
1163
skb_queue_walk(&rpp->queue, sb)
drivers/atm/idt77252.c
1242
struct sk_buff *queue;
drivers/atm/idt77252.c
1253
queue = card->raw_cell_head;
drivers/atm/idt77252.c
1254
if (!queue)
drivers/atm/idt77252.c
1257
head = IDT77252_PRV_PADDR(queue) + (queue->data - queue->head - 16);
drivers/atm/idt77252.c
1260
dma_sync_single_for_cpu(&card->pcidev->dev, IDT77252_PRV_PADDR(queue),
drivers/atm/idt77252.c
1261
skb_end_offset(queue) - 16,
drivers/atm/idt77252.c
1268
header = le32_to_cpu(*(u32 *) &queue->data[0]);
drivers/atm/idt77252.c
1284
printk(" %02x", queue->data[i]);
drivers/atm/idt77252.c
1327
skb_put_data(sb, &(queue->data[16]), ATM_CELL_PAYLOAD);
drivers/atm/idt77252.c
1335
skb_pull(queue, 64);
drivers/atm/idt77252.c
1337
head = IDT77252_PRV_PADDR(queue)
drivers/atm/idt77252.c
1338
+ (queue->data - queue->head - 16);
drivers/atm/idt77252.c
1340
if (queue->len < 128) {
drivers/atm/idt77252.c
1344
head = le32_to_cpu(*(u32 *) &queue->data[0]);
drivers/atm/idt77252.c
1345
handle = le32_to_cpu(*(u32 *) &queue->data[4]);
drivers/atm/idt77252.c
1348
recycle_rx_skb(card, queue);
drivers/atm/idt77252.c
1352
queue = card->raw_cell_head;
drivers/atm/idt77252.c
1354
IDT77252_PRV_PADDR(queue),
drivers/atm/idt77252.c
1355
(skb_end_pointer(queue) -
drivers/atm/idt77252.c
1356
queue->data),
drivers/atm/idt77252.c
1790
idt77252_fbq_full(struct idt77252_dev *card, int queue)
drivers/atm/idt77252.c
1792
return (readl(SAR_REG_STAT) >> (16 + (queue << 2))) == 0x0f;
drivers/atm/idt77252.c
1796
push_rx_skb(struct idt77252_dev *card, struct sk_buff *skb, int queue)
drivers/atm/idt77252.c
1808
switch (queue) {
drivers/atm/idt77252.c
1825
if (idt77252_fbq_full(card, queue))
drivers/atm/idt77252.c
1834
writel(handle, card->fbq[queue]);
drivers/atm/idt77252.c
1835
writel(addr, card->fbq[queue]);
drivers/atm/idt77252.c
1842
add_rx_skb(struct idt77252_dev *card, int queue,
drivers/atm/idt77252.c
1853
if (sb_pool_add(card, skb, queue)) {
drivers/atm/idt77252.c
1865
if (push_rx_skb(card, skb, queue)) {
drivers/atm/idt77252.c
1908
skb_queue_head_init(&rpp->queue);
drivers/atm/idt77252.c
1917
skb_queue_walk_safe(&rpp->queue, skb, tmp)
drivers/atm/idt77252.c
2519
if (skb_queue_len(&vc->rcv.rx_pool.queue) != 0) {
drivers/atm/idt77252.c
2909
if (skb_queue_len(&vc->rcv.rx_pool.queue) != 0) {
drivers/atm/idt77252.c
582
sb_pool_add(struct idt77252_dev *card, struct sk_buff *skb, int queue)
drivers/atm/idt77252.c
584
struct sb_pool *pool = &card->sbpool[queue];
drivers/atm/idt77252.c
595
IDT77252_PRV_POOL(skb) = POOL_HANDLE(queue, index);
drivers/atm/idt77252.c
604
unsigned int queue, index;
drivers/atm/idt77252.c
609
queue = POOL_QUEUE(handle);
drivers/atm/idt77252.c
610
if (queue > 3)
drivers/atm/idt77252.c
617
card->sbpool[queue].skb[index] = NULL;
drivers/atm/idt77252.c
623
unsigned int queue, index;
drivers/atm/idt77252.c
625
queue = POOL_QUEUE(handle);
drivers/atm/idt77252.c
626
if (queue > 3)
drivers/atm/idt77252.c
633
return card->sbpool[queue].skb[index];
drivers/atm/idt77252.c
95
struct sk_buff *, int queue);
drivers/atm/idt77252.h
176
struct sk_buff_head queue;
drivers/atm/idt77252.h
343
#define POOL_HANDLE(queue, index) (((queue + 1) << 16) | (index))
drivers/atm/nicstar.c
1001
skb_unlink(handle1, &card->sbpool.queue);
drivers/atm/nicstar.c
1003
skb_unlink(handle2, &card->sbpool.queue);
drivers/atm/nicstar.c
1011
skb_unlink(handle1, &card->lbpool.queue);
drivers/atm/nicstar.c
1013
skb_unlink(handle2, &card->lbpool.queue);
drivers/atm/nicstar.c
1182
skb_queue_tail(&card->sbpool.queue, sb);
drivers/atm/nicstar.c
1207
skb_queue_tail(&card->lbpool.queue, lb);
drivers/atm/nicstar.c
2073
iovb = skb_dequeue(&(card->iovpool.queue));
drivers/atm/nicstar.c
2089
skb_queue_tail(&card->iovpool.queue, new_iovb);
drivers/atm/nicstar.c
2230
hb = skb_dequeue(&(card->hbpool.queue));
drivers/atm/nicstar.c
2252
queue, new_hb);
drivers/atm/nicstar.c
2262
skb_queue_tail(&card->hbpool.queue,
drivers/atm/nicstar.c
2273
queue, new_hb);
drivers/atm/nicstar.c
2285
skb_queue_tail(&card->hbpool.queue, hb);
drivers/atm/nicstar.c
229
while ((hb = skb_dequeue(&card->hbpool.queue)) != NULL) {
drivers/atm/nicstar.c
2352
skb_queue_tail(&card->iovpool.queue, iovb);
drivers/atm/nicstar.c
2360
skb_unlink(sb, &card->sbpool.queue);
drivers/atm/nicstar.c
2365
skb_queue_tail(&card->sbpool.queue, new_sb);
drivers/atm/nicstar.c
237
while ((iovb = skb_dequeue(&card->iovpool.queue)) != NULL) {
drivers/atm/nicstar.c
2375
skb_queue_tail(&card->sbpool.queue, new_sb);
drivers/atm/nicstar.c
2384
skb_unlink(lb, &card->lbpool.queue);
drivers/atm/nicstar.c
2389
skb_queue_tail(&card->lbpool.queue, new_lb);
drivers/atm/nicstar.c
2399
skb_queue_tail(&card->lbpool.queue, new_lb);
drivers/atm/nicstar.c
242
while ((lb = skb_dequeue(&card->lbpool.queue)) != NULL)
drivers/atm/nicstar.c
244
while ((sb = skb_dequeue(&card->sbpool.queue)) != NULL)
drivers/atm/nicstar.c
2590
skb_queue_tail(&card->sbpool.queue, sb);
drivers/atm/nicstar.c
2604
skb_queue_tail(&card->lbpool.queue, lb);
drivers/atm/nicstar.c
2615
hb = skb_dequeue(&card->hbpool.queue);
drivers/atm/nicstar.c
2634
skb_queue_tail(&card->hbpool.queue, hb);
drivers/atm/nicstar.c
2645
iovb = skb_dequeue(&card->iovpool.queue);
drivers/atm/nicstar.c
2664
skb_queue_tail(&card->iovpool.queue, iovb);
drivers/atm/nicstar.c
656
skb_queue_head_init(&card->hbpool.queue);
drivers/atm/nicstar.c
670
skb_queue_tail(&card->hbpool.queue, hb);
drivers/atm/nicstar.c
675
skb_queue_head_init(&card->lbpool.queue);
drivers/atm/nicstar.c
689
skb_queue_tail(&card->lbpool.queue, lb);
drivers/atm/nicstar.c
711
skb_queue_head_init(&card->sbpool.queue);
drivers/atm/nicstar.c
725
skb_queue_tail(&card->sbpool.queue, sb);
drivers/atm/nicstar.c
741
skb_queue_head_init(&card->iovpool.queue);
drivers/atm/nicstar.c
755
skb_queue_tail(&card->iovpool.queue, iovb);
drivers/atm/nicstar.c
824
while ((iovb = skb_dequeue(&card->iovpool.queue)) != NULL)
drivers/atm/nicstar.c
829
while ((sb = skb_dequeue(&card->sbpool.queue)) != NULL)
drivers/atm/nicstar.c
835
while ((lb = skb_dequeue(&card->lbpool.queue)) != NULL)
drivers/atm/nicstar.c
840
while ((hb = skb_dequeue(&card->hbpool.queue)) != NULL)
drivers/atm/nicstar.h
690
struct sk_buff_head queue;
drivers/block/aoe/aoeblk.c
253
struct aoedev *d = hctx->queue->queuedata;
drivers/block/aoe/aoeblk.c
392
d->blkq = gd->queue;
drivers/block/aoe/aoechr.c
127
struct sk_buff_head queue;
drivers/block/aoe/aoechr.c
128
__skb_queue_head_init(&queue);
drivers/block/aoe/aoechr.c
129
__skb_queue_tail(&queue, skb);
drivers/block/aoe/aoechr.c
130
aoenet_xmit(&queue);
drivers/block/aoe/aoecmd.c
1371
struct sk_buff_head queue;
drivers/block/aoe/aoecmd.c
1373
__skb_queue_head_init(&queue);
drivers/block/aoe/aoecmd.c
1374
aoecmd_cfg_pkts(aoemajor, aoeminor, &queue);
drivers/block/aoe/aoecmd.c
1375
aoenet_xmit(&queue);
drivers/block/aoe/aoecmd.c
1538
struct sk_buff_head queue;
drivers/block/aoe/aoecmd.c
1605
__skb_queue_head_init(&queue);
drivers/block/aoe/aoecmd.c
1606
__skb_queue_tail(&queue, sl);
drivers/block/aoe/aoecmd.c
1607
aoenet_xmit(&queue);
drivers/block/aoe/aoecmd.c
374
struct sk_buff_head queue;
drivers/block/aoe/aoecmd.c
402
__skb_queue_head_init(&queue);
drivers/block/aoe/aoecmd.c
403
__skb_queue_tail(&queue, skb);
drivers/block/aoe/aoecmd.c
404
aoenet_xmit(&queue);
drivers/block/aoe/aoecmd.c
415
aoecmd_cfg_pkts(ushort aoemajor, unsigned char aoeminor, struct sk_buff_head *queue)
drivers/block/aoe/aoecmd.c
438
__skb_queue_tail(queue, skb);
drivers/block/aoe/aoecmd.c
457
struct sk_buff_head queue;
drivers/block/aoe/aoecmd.c
497
__skb_queue_head_init(&queue);
drivers/block/aoe/aoecmd.c
498
__skb_queue_tail(&queue, skb);
drivers/block/aoe/aoecmd.c
499
aoenet_xmit(&queue);
drivers/block/aoe/aoecmd.c
590
struct sk_buff_head queue;
drivers/block/aoe/aoecmd.c
623
__skb_queue_head_init(&queue);
drivers/block/aoe/aoecmd.c
624
__skb_queue_tail(&queue, skb);
drivers/block/aoe/aoecmd.c
625
aoenet_xmit(&queue);
drivers/block/aoe/aoenet.c
113
aoenet_xmit(struct sk_buff_head *queue)
drivers/block/aoe/aoenet.c
118
skb_queue_walk_safe(queue, skb, tmp) {
drivers/block/aoe/aoenet.c
119
__skb_unlink(skb, queue);
drivers/block/ataflop.c
761
q = unit[drive].disk[type]->queue;
drivers/block/drbd/drbd_main.c
2692
device->rq_queue = disk->queue;
drivers/block/drbd/drbd_nl.c
1271
b = bdev->backing_bdev->bd_disk->queue;
drivers/block/drbd/drbd_worker.c
2059
static bool dequeue_work_batch(struct drbd_work_queue *queue, struct list_head *work_list)
drivers/block/drbd/drbd_worker.c
2061
spin_lock_irq(&queue->q_lock);
drivers/block/drbd/drbd_worker.c
2062
list_splice_tail_init(&queue->q, work_list);
drivers/block/drbd/drbd_worker.c
2063
spin_unlock_irq(&queue->q_lock);
drivers/block/loop.c
2054
lo->lo_queue = lo->lo_disk->queue;
drivers/block/loop.c
2087
disk->queue = lo->lo_queue;
drivers/block/mtip32xx/mtip32xx.c
2517
blk_mq_quiesce_queue(dd->queue);
drivers/block/mtip32xx/mtip32xx.c
2529
blk_mq_unquiesce_queue(dd->queue);
drivers/block/mtip32xx/mtip32xx.c
3247
struct driver_data *dd = hctx->queue->queuedata;
drivers/block/mtip32xx/mtip32xx.c
3271
struct driver_data *dd = hctx->queue->queuedata;
drivers/block/mtip32xx/mtip32xx.c
3308
struct driver_data *dd = hctx->queue->queuedata;
drivers/block/mtip32xx/mtip32xx.c
3443
dd->queue = dd->disk->queue;
drivers/block/mtip32xx/mtip32xx.c
894
blk_mq_quiesce_queue(port->dd->queue);
drivers/block/mtip32xx/mtip32xx.c
914
blk_mq_unquiesce_queue(port->dd->queue);
drivers/block/mtip32xx/mtip32xx.c
917
blk_mq_unquiesce_queue(port->dd->queue);
drivers/block/mtip32xx/mtip32xx.c
974
rq = blk_mq_alloc_request(dd->queue, REQ_OP_DRV_IN, BLK_MQ_REQ_RESERVED);
drivers/block/mtip32xx/mtip32xx.h
432
struct request_queue *queue; /* Our request queue. */
drivers/block/nbd.c
1052
blk_mq_quiesce_queue(nbd->disk->queue);
drivers/block/nbd.c
1054
blk_mq_unquiesce_queue(nbd->disk->queue);
drivers/block/nbd.c
1262
memflags = blk_mq_freeze_queue(nbd->disk->queue);
drivers/block/nbd.c
1303
blk_mq_unfreeze_queue(nbd->disk->queue, memflags);
drivers/block/nbd.c
1308
blk_mq_unfreeze_queue(nbd->disk->queue, memflags);
drivers/block/nbd.c
1587
blk_queue_rq_timeout(nbd->disk->queue, timeout * HZ);
drivers/block/nbd.c
1589
blk_queue_rq_timeout(nbd->disk->queue, 30 * HZ);
drivers/block/nbd.c
354
lim = queue_limits_start_update(nbd->disk->queue);
drivers/block/nbd.c
374
error = queue_limits_commit_update_frozen(nbd->disk->queue, &lim);
drivers/block/nbd.c
977
struct request_queue *q = nbd->disk->queue;
drivers/block/null_blk/main.c
1735
struct nullb *nullb = hctx->queue->queuedata;
drivers/block/null_blk/main.c
2013
nullb->q = nullb->disk->queue;
drivers/block/ps3disk.c
187
struct request_queue *q = hctx->queue;
drivers/block/ps3disk.c
259
blk_mq_run_hw_queues(priv->gendisk->queue, true);
drivers/block/rbd.c
4786
struct rbd_device *rbd_dev = hctx->queue->queuedata;
drivers/block/rbd.c
7293
unsigned int memflags = blk_mq_freeze_queue(rbd_dev->disk->queue);
drivers/block/rbd.c
7296
blk_mq_unfreeze_queue(rbd_dev->disk->queue, memflags);
drivers/block/rnbd/rnbd-clt.c
1326
queue_for_each_hw_ctx(dev->queue, hctx, i) {
drivers/block/rnbd/rnbd-clt.c
1342
dev->gd->queue = dev->queue;
drivers/block/rnbd/rnbd-clt.c
1398
dev->queue = dev->gd->queue;
drivers/block/rnbd/rnbd-clt.c
952
struct queue_limits *limit = &dev->queue->limits;
drivers/block/rnbd/rnbd-clt.h
111
struct request_queue *queue;
drivers/block/sunvdc.c
1070
blk_mq_stop_hw_queues(port->disk->queue);
drivers/block/sunvdc.c
1116
struct request_queue *q = port->disk->queue;
drivers/block/sunvdc.c
1174
blk_mq_stop_hw_queues(port->disk->queue);
drivers/block/sunvdc.c
187
blk_mq_start_stopped_hw_queues(port->disk->queue, true);
drivers/block/sunvdc.c
539
struct vdc_port *port = hctx->queue->queuedata;
drivers/block/sunvdc.c
844
q = g->queue;
drivers/block/sunvdc.c
852
g->queue = q;
drivers/block/swim.c
529
struct floppy_state *fs = hctx->queue->queuedata;
drivers/block/swim3.c
310
struct floppy_state *fs = hctx->queue->queuedata;
drivers/block/swim3.c
842
struct request_queue *q = disks[fs->index]->queue;
drivers/block/ublk_drv.c
2478
blk_mq_kick_requeue_list(disk->queue);
drivers/block/ublk_drv.c
2654
blk_mq_quiesce_queue(disk->queue);
drivers/block/ublk_drv.c
2656
blk_mq_unquiesce_queue(disk->queue);
drivers/block/ublk_drv.c
2844
WARN_ON_ONCE(!blk_queue_quiesced(ub->ub_disk->queue));
drivers/block/ublk_drv.c
2863
blk_mq_quiesce_queue(ub->ub_disk->queue);
drivers/block/ublk_drv.c
2869
blk_mq_unquiesce_queue(ub->ub_disk->queue);
drivers/block/ublk_drv.c
2871
blk_mq_kick_requeue_list(ub->ub_disk->queue);
drivers/block/ublk_drv.c
4472
unsigned long queue;
drivers/block/ublk_drv.c
4484
queue = header->data[0];
drivers/block/ublk_drv.c
4485
if (queue >= ub->dev_info.nr_hw_queues)
drivers/block/ublk_drv.c
4492
if (ub->tag_set.map[HCTX_TYPE_DEFAULT].mq_map[i] == queue)
drivers/block/ublk_drv.c
4990
blk_mq_kick_requeue_list(ub->ub_disk->queue);
drivers/block/ublk_drv.c
509
struct request_queue *q = ublk->ub_disk->queue;
drivers/block/ublk_drv.c
5124
blk_mq_quiesce_queue(disk->queue);
drivers/block/ublk_drv.c
5126
blk_mq_unquiesce_queue(disk->queue);
drivers/block/ublk_drv.c
537
unsigned int zone_size_sectors = disk->queue->limits.chunk_sectors;
drivers/block/ublk_drv.c
564
req = blk_mq_alloc_request(disk->queue, REQ_OP_DRV_IN, 0);
drivers/block/virtio_blk.c
1108
lim = queue_limits_start_update(disk->queue);
drivers/block/virtio_blk.c
1113
i = queue_limits_commit_update_frozen(disk->queue, &lim);
drivers/block/virtio_blk.c
1204
struct virtio_blk *vblk = hctx->queue->queuedata;
drivers/block/virtio_blk.c
1225
blk_mq_start_stopped_hw_queues(vblk->disk->queue, true);
drivers/block/virtio_blk.c
133
struct virtio_blk *vblk = hctx->queue->queuedata;
drivers/block/virtio_blk.c
1591
struct request_queue *q = vblk->disk->queue;
drivers/block/virtio_blk.c
1627
blk_mq_unquiesce_queue(vblk->disk->queue);
drivers/block/virtio_blk.c
338
struct virtio_blk *vblk = req->mq_hctx->queue->queuedata;
drivers/block/virtio_blk.c
373
blk_mq_start_stopped_hw_queues(vblk->disk->queue, true);
drivers/block/virtio_blk.c
379
struct virtio_blk *vblk = hctx->queue->queuedata;
drivers/block/virtio_blk.c
429
struct virtio_blk *vblk = hctx->queue->queuedata;
drivers/block/virtio_blk.c
467
struct virtio_blk *vblk = req->mq_hctx->queue->queuedata;
drivers/block/virtio_blk.c
531
struct request_queue *q = vblk->disk->queue;
drivers/block/virtio_blk.c
560
struct request_queue *q = vblk->disk->queue;
drivers/block/virtio_blk.c
807
struct request_queue *q = vblk->disk->queue;
drivers/block/virtio_blk.c
916
struct request_queue *q = vblk->disk->queue;
drivers/block/xen-blkfront.c
1164
info->rq = gd->queue;
drivers/block/xen-blkfront.c
898
struct blkfront_info *info = hctx->queue->queuedata;
drivers/bluetooth/btintel_pcie.c
219
u8 queue = entry->entry;
drivers/bluetooth/btintel_pcie.c
220
struct msix_entry *entries = entry - queue;
drivers/cdrom/cdrom.c
1488
int media_changed(struct cdrom_device_info *cdi, int queue)
drivers/cdrom/cdrom.c
1490
unsigned int mask = (1 << (queue & 1));
drivers/cdrom/cdrom.c
1498
BUG_ON(!queue); /* shouldn't be called from VFS path */
drivers/cdrom/cdrom.c
2145
int max_frames = (queue_max_sectors(cdi->disk->queue) << 9) /
drivers/cdrom/gdrom.c
727
gd.disk->queue = gd.gdrom_rq;
drivers/cdrom/gdrom.c
789
gd.gdrom_rq = gd.disk->queue;
drivers/char/apm-emulation.c
113
struct apm_queue queue;
drivers/char/apm-emulation.c
186
queue_add_event(&as->queue, event);
drivers/char/apm-emulation.c
201
if (queue_empty(&as->queue) && fp->f_flags & O_NONBLOCK)
drivers/char/apm-emulation.c
204
wait_event_interruptible(apm_waitqueue, !queue_empty(&as->queue));
drivers/char/apm-emulation.c
206
while ((i >= sizeof(event)) && !queue_empty(&as->queue)) {
drivers/char/apm-emulation.c
207
event = queue_get_event(&as->queue);
drivers/char/apm-emulation.c
234
return queue_empty(&as->queue) ? 0 : EPOLLIN | EPOLLRDNORM;
drivers/char/apm-emulation.c
529
queue_add_event(&as->queue, apm_event);
drivers/char/ipmi/bt-bmc.c
187
if (wait_event_interruptible(bt_bmc->queue,
drivers/char/ipmi/bt-bmc.c
261
if (wait_event_interruptible(bt_bmc->queue,
drivers/char/ipmi/bt-bmc.c
325
poll_wait(file, &bt_bmc->queue, wait);
drivers/char/ipmi/bt-bmc.c
353
wake_up(&bt_bmc->queue);
drivers/char/ipmi/bt-bmc.c
371
wake_up(&bt_bmc->queue);
drivers/char/ipmi/bt-bmc.c
427
init_waitqueue_head(&bt_bmc->queue);
drivers/char/ipmi/bt-bmc.c
62
wait_queue_head_t queue;
drivers/char/ipmi/kcs_bmc_cdev_ipmi.c
167
wake_up_interruptible(&priv->queue);
drivers/char/ipmi/kcs_bmc_cdev_ipmi.c
310
poll_wait(filp, &priv->queue, wait);
drivers/char/ipmi/kcs_bmc_cdev_ipmi.c
329
wait_event_interruptible(priv->queue,
drivers/char/ipmi/kcs_bmc_cdev_ipmi.c
484
init_waitqueue_head(&priv->queue);
drivers/char/ipmi/kcs_bmc_cdev_ipmi.c
79
wait_queue_head_t queue;
drivers/char/tpm/st33zp24/st33zp24.c
206
wait_queue_head_t *queue, bool check_cancel)
drivers/char/tpm/st33zp24/st33zp24.c
236
ret = wait_event_interruptible_timeout(*queue,
drivers/char/tpm/tpm_i2c_nuvoton.c
172
u32 timeout, wait_queue_head_t *queue)
drivers/char/tpm/tpm_i2c_nuvoton.c
174
if ((chip->flags & TPM_CHIP_FLAG_IRQ) && queue) {
drivers/char/tpm/tpm_i2c_nuvoton.c
180
rc = wait_event_interruptible_timeout(*queue,
drivers/char/tpm/tpm_i2c_nuvoton.c
221
wait_queue_head_t *queue)
drivers/char/tpm/tpm_i2c_nuvoton.c
226
timeout, queue);
drivers/char/tpm/tpm_tis_core.c
65
unsigned long timeout, wait_queue_head_t *queue,
drivers/char/tpm/tpm_tis_core.c
94
rc = wait_event_interruptible_timeout(*queue,
drivers/char/tpm/xen-tpmfront.c
56
unsigned long timeout, wait_queue_head_t *queue,
drivers/char/tpm/xen-tpmfront.c
76
rc = wait_event_interruptible_timeout(*queue,
drivers/crypto/atmel-aes.c
191
struct crypto_queue queue;
drivers/crypto/atmel-aes.c
2352
crypto_init_queue(&aes_dd->queue, ATMEL_AES_QUEUE_LENGTH);
drivers/crypto/atmel-aes.c
909
ret = crypto_enqueue_request(&dd->queue, new_areq);
drivers/crypto/atmel-aes.c
914
backlog = crypto_get_backlog(&dd->queue);
drivers/crypto/atmel-aes.c
915
areq = crypto_dequeue_request(&dd->queue);
drivers/crypto/atmel-sha.c
1065
ret = ahash_enqueue_request(&dd->queue, req);
drivers/crypto/atmel-sha.c
1072
backlog = crypto_get_backlog(&dd->queue);
drivers/crypto/atmel-sha.c
1073
async_req = crypto_dequeue_request(&dd->queue);
drivers/crypto/atmel-sha.c
143
struct crypto_queue queue;
drivers/crypto/atmel-sha.c
2598
crypto_init_queue(&sha_dd->queue, ATMEL_SHA_QUEUE_LENGTH);
drivers/crypto/atmel-tdes.c
566
ret = crypto_enqueue_request(&dd->queue, &req->base);
drivers/crypto/atmel-tdes.c
571
backlog = crypto_get_backlog(&dd->queue);
drivers/crypto/atmel-tdes.c
572
async_req = crypto_dequeue_request(&dd->queue);
drivers/crypto/atmel-tdes.c
965
crypto_init_queue(&tdes_dd->queue, ATMEL_TDES_QUEUE_LENGTH);
drivers/crypto/atmel-tdes.c
97
struct crypto_queue queue;
drivers/crypto/axis/artpec6_crypto.c
2050
while (!list_empty(&ac->queue) && !artpec6_crypto_busy()) {
drivers/crypto/axis/artpec6_crypto.c
2051
req = list_first_entry(&ac->queue,
drivers/crypto/axis/artpec6_crypto.c
262
struct list_head queue; /* waiting for pdma fifo space */
drivers/crypto/axis/artpec6_crypto.c
2876
INIT_LIST_HEAD(&ac->queue);
drivers/crypto/axis/artpec6_crypto.c
469
list_add_tail(&req->list, &ac->queue);
drivers/crypto/caam/dpseci.c
198
u8 queue, const struct dpseci_rx_queue_cfg *cfg)
drivers/crypto/caam/dpseci.c
209
cmd_params->queue = queue;
drivers/crypto/caam/dpseci.c
232
u8 queue, struct dpseci_rx_queue_attr *attr)
drivers/crypto/caam/dpseci.c
242
cmd_params->queue = queue;
drivers/crypto/caam/dpseci.c
272
u8 queue, struct dpseci_tx_queue_attr *attr)
drivers/crypto/caam/dpseci.c
283
cmd_params->queue = queue;
drivers/crypto/caam/dpseci.h
156
u8 queue, const struct dpseci_rx_queue_cfg *cfg);
drivers/crypto/caam/dpseci.h
175
u8 queue, struct dpseci_rx_queue_attr *attr);
drivers/crypto/caam/dpseci.h
188
u8 queue, struct dpseci_tx_queue_attr *attr);
drivers/crypto/caam/dpseci_cmd.h
86
u8 queue;
drivers/crypto/cavium/cpt/cptvf.h
64
struct command_queue queue[CPT_NUM_QS_PER_VF];
drivers/crypto/cavium/cpt/cptvf.h
87
struct pending_queue queue[CPT_NUM_QS_PER_VF];
drivers/crypto/cavium/cpt/cptvf.h
91
for (i = 0, q = &qinfo->queue[i]; i < qinfo->nr_queues; i++, \
drivers/crypto/cavium/cpt/cptvf.h
92
q = &qinfo->queue[i])
drivers/crypto/cavium/cpt/cptvf_main.c
108
struct pending_queue *queue = NULL;
drivers/crypto/cavium/cpt/cptvf_main.c
113
for_each_pending_queue(pqinfo, queue, i) {
drivers/crypto/cavium/cpt/cptvf_main.c
114
queue->head = kzalloc_objs(*queue->head, qlen);
drivers/crypto/cavium/cpt/cptvf_main.c
115
if (!queue->head) {
drivers/crypto/cavium/cpt/cptvf_main.c
120
queue->front = 0;
drivers/crypto/cavium/cpt/cptvf_main.c
121
queue->rear = 0;
drivers/crypto/cavium/cpt/cptvf_main.c
122
atomic64_set((&queue->pending_count), (0));
drivers/crypto/cavium/cpt/cptvf_main.c
125
spin_lock_init(&queue->lock);
drivers/crypto/cavium/cpt/cptvf_main.c
170
struct command_queue *queue = NULL;
drivers/crypto/cavium/cpt/cptvf_main.c
177
queue = &cqinfo->queue[i];
drivers/crypto/cavium/cpt/cptvf_main.c
178
if (hlist_empty(&cqinfo->queue[i].chead))
drivers/crypto/cavium/cpt/cptvf_main.c
181
hlist_for_each_entry_safe(chunk, node, &cqinfo->queue[i].chead,
drivers/crypto/cavium/cpt/cptvf_main.c
193
queue->nchunks = 0;
drivers/crypto/cavium/cpt/cptvf_main.c
194
queue->idx = 0;
drivers/crypto/cavium/cpt/cptvf_main.c
207
struct command_queue *queue = NULL;
drivers/crypto/cavium/cpt/cptvf_main.c
225
queue = &cqinfo->queue[i];
drivers/crypto/cavium/cpt/cptvf_main.c
226
INIT_HLIST_HEAD(&cqinfo->queue[i].chead);
drivers/crypto/cavium/cpt/cptvf_main.c
240
i, queue->nchunks);
drivers/crypto/cavium/cpt/cptvf_main.c
246
if (queue->nchunks == 0) {
drivers/crypto/cavium/cpt/cptvf_main.c
248
&cqinfo->queue[i].chead);
drivers/crypto/cavium/cpt/cptvf_main.c
255
queue->nchunks++;
drivers/crypto/cavium/cpt/cptvf_main.c
267
queue->qhead = curr;
drivers/crypto/cavium/cpt/cptvf_main.c
268
spin_lock_init(&queue->lock);
drivers/crypto/cavium/cpt/cptvf_main.c
651
base_addr = (u64)(cptvf->cqinfo.queue[0].qhead->dma_addr);
drivers/crypto/cavium/cpt/cptvf_main.c
84
struct pending_queue *queue;
drivers/crypto/cavium/cpt/cptvf_main.c
86
for_each_pending_queue(pqinfo, queue, i) {
drivers/crypto/cavium/cpt/cptvf_main.c
87
if (!queue->head)
drivers/crypto/cavium/cpt/cptvf_main.c
91
kfree_sensitive((queue->head));
drivers/crypto/cavium/cpt/cptvf_main.c
93
queue->front = 0;
drivers/crypto/cavium/cpt/cptvf_main.c
94
queue->rear = 0;
drivers/crypto/cavium/cpt/cptvf_reqmanager.c
228
struct command_queue *queue;
drivers/crypto/cavium/cpt/cptvf_reqmanager.c
240
queue = &qinfo->queue[qno];
drivers/crypto/cavium/cpt/cptvf_reqmanager.c
242
spin_lock(&queue->lock);
drivers/crypto/cavium/cpt/cptvf_reqmanager.c
243
ent = &queue->qhead->head[queue->idx * qinfo->cmd_size];
drivers/crypto/cavium/cpt/cptvf_reqmanager.c
246
if (++queue->idx >= queue->qhead->size / 64) {
drivers/crypto/cavium/cpt/cptvf_reqmanager.c
247
hlist_for_each_entry(chunk, &queue->chead, nextchunk) {
drivers/crypto/cavium/cpt/cptvf_reqmanager.c
248
if (chunk == queue->qhead) {
drivers/crypto/cavium/cpt/cptvf_reqmanager.c
251
queue->qhead = chunk;
drivers/crypto/cavium/cpt/cptvf_reqmanager.c
255
queue->idx = 0;
drivers/crypto/cavium/cpt/cptvf_reqmanager.c
261
spin_unlock(&queue->lock);
drivers/crypto/cavium/cpt/cptvf_reqmanager.c
329
struct pending_queue *pqueue = &pqinfo->queue[qno];
drivers/crypto/cavium/cpt/cptvf_reqmanager.c
37
struct pending_queue *queue = &pqinfo->queue[qno];
drivers/crypto/cavium/cpt/cptvf_reqmanager.c
39
queue->front++;
drivers/crypto/cavium/cpt/cptvf_reqmanager.c
40
if (unlikely(queue->front == pqinfo->qlen))
drivers/crypto/cavium/cpt/cptvf_reqmanager.c
408
int ret = 0, clear = 0, queue = 0;
drivers/crypto/cavium/cpt/cptvf_reqmanager.c
41
queue->front = 0;
drivers/crypto/cavium/cpt/cptvf_reqmanager.c
475
queue = 0;
drivers/crypto/cavium/cpt/cptvf_reqmanager.c
476
pqueue = &cptvf->pqinfo.queue[queue];
drivers/crypto/cavium/cpt/cptvf_reqmanager.c
480
process_pending_queue(cptvf, &cptvf->pqinfo, queue);
drivers/crypto/cavium/cpt/cptvf_reqmanager.c
489
process_pending_queue(cptvf, &cptvf->pqinfo, queue);
drivers/crypto/cavium/cpt/cptvf_reqmanager.c
495
queue, pqueue->rear, pqueue->front);
drivers/crypto/cavium/cpt/cptvf_reqmanager.c
524
ret = send_cpt_command(cptvf, &cptinst, queue);
drivers/crypto/hifn_795x.c
1806
dev->success, dev->queue.qlen, dev->queue.max_qlen,
drivers/crypto/hifn_795x.c
1913
while ((async_req = crypto_dequeue_request(&dev->queue))) {
drivers/crypto/hifn_795x.c
1975
err = crypto_enqueue_request(&dev->queue, &req->base);
drivers/crypto/hifn_795x.c
2031
backlog = crypto_get_backlog(&dev->queue);
drivers/crypto/hifn_795x.c
2032
async_req = crypto_dequeue_request(&dev->queue);
drivers/crypto/hifn_795x.c
2062
if (dev->started < HIFN_QUEUE_LENGTH && dev->queue.qlen)
drivers/crypto/hifn_795x.c
2327
if (dev->started < HIFN_QUEUE_LENGTH && dev->queue.qlen)
drivers/crypto/hifn_795x.c
2405
crypto_init_queue(&dev->queue, 1);
drivers/crypto/hifn_795x.c
428
struct crypto_queue queue;
drivers/crypto/hisilicon/qm.c
577
u64 base, u16 queue, bool op)
drivers/crypto/hisilicon/qm.c
582
mailbox->queue_num = cpu_to_le16(queue);
drivers/crypto/hisilicon/qm.c
693
int hisi_qm_mb(struct hisi_qm *qm, u8 cmd, dma_addr_t dma_addr, u16 queue,
drivers/crypto/hisilicon/qm.c
711
qm_mb_pre_init(&mailbox, cmd, dma_addr, queue, op);
drivers/crypto/hisilicon/qm.c
721
int hisi_qm_mb_read(struct hisi_qm *qm, u64 *base, u8 cmd, u16 queue)
drivers/crypto/hisilicon/qm.c
726
qm_mb_pre_init(&mailbox, cmd, 0, queue, 1);
drivers/crypto/hisilicon/sec/sec_algs.c
232
struct device *dev = ctx->queue->dev_info->dev;
drivers/crypto/hisilicon/sec/sec_algs.c
384
static int sec_send_request(struct sec_request *sec_req, struct sec_queue *queue)
drivers/crypto/hisilicon/sec/sec_algs.c
401
if (!queue->havesoftqueue ||
drivers/crypto/hisilicon/sec/sec_algs.c
402
(kfifo_is_empty(&queue->softqueue) &&
drivers/crypto/hisilicon/sec/sec_algs.c
403
sec_queue_empty(queue))) {
drivers/crypto/hisilicon/sec/sec_algs.c
404
ret = sec_queue_send(queue, &el->req, sec_req);
drivers/crypto/hisilicon/sec/sec_algs.c
412
kfifo_put(&queue->softqueue, el);
drivers/crypto/hisilicon/sec/sec_algs.c
432
struct device *dev = ctx->queue->dev_info->dev;
drivers/crypto/hisilicon/sec/sec_algs.c
452
spin_lock_bh(&ctx->queue->queuelock);
drivers/crypto/hisilicon/sec/sec_algs.c
484
if (ctx->queue->havesoftqueue &&
drivers/crypto/hisilicon/sec/sec_algs.c
485
!kfifo_is_empty(&ctx->queue->softqueue) &&
drivers/crypto/hisilicon/sec/sec_algs.c
486
sec_queue_empty(ctx->queue)) {
drivers/crypto/hisilicon/sec/sec_algs.c
487
ret = kfifo_get(&ctx->queue->softqueue, &nextrequest);
drivers/crypto/hisilicon/sec/sec_algs.c
494
sec_queue_send(ctx->queue, &nextrequest->req,
drivers/crypto/hisilicon/sec/sec_algs.c
501
if (sec_queue_can_enqueue(ctx->queue,
drivers/crypto/hisilicon/sec/sec_algs.c
503
(ctx->queue->havesoftqueue &&
drivers/crypto/hisilicon/sec/sec_algs.c
504
kfifo_avail(&ctx->queue->softqueue) >
drivers/crypto/hisilicon/sec/sec_algs.c
506
sec_send_request(backlog_req, ctx->queue);
drivers/crypto/hisilicon/sec/sec_algs.c
512
spin_unlock_bh(&ctx->queue->queuelock);
drivers/crypto/hisilicon/sec/sec_algs.c
517
sec_alg_free_el(sec_req_el, ctx->queue->dev_info);
drivers/crypto/hisilicon/sec/sec_algs.c
713
struct sec_queue *queue = ctx->queue;
drivers/crypto/hisilicon/sec/sec_algs.c
715
struct sec_dev_info *info = queue->dev_info;
drivers/crypto/hisilicon/sec/sec_algs.c
801
spin_lock_bh(&queue->queuelock);
drivers/crypto/hisilicon/sec/sec_algs.c
810
if ((!sec_queue_can_enqueue(queue, steps) &&
drivers/crypto/hisilicon/sec/sec_algs.c
811
(!queue->havesoftqueue ||
drivers/crypto/hisilicon/sec/sec_algs.c
812
kfifo_avail(&queue->softqueue) > steps)) ||
drivers/crypto/hisilicon/sec/sec_algs.c
817
spin_unlock_bh(&queue->queuelock);
drivers/crypto/hisilicon/sec/sec_algs.c
821
spin_unlock_bh(&queue->queuelock);
drivers/crypto/hisilicon/sec/sec_algs.c
824
ret = sec_send_request(sec_req, queue);
drivers/crypto/hisilicon/sec/sec_algs.c
825
spin_unlock_bh(&queue->queuelock);
drivers/crypto/hisilicon/sec/sec_algs.c
880
ctx->queue = sec_queue_alloc_start_safe();
drivers/crypto/hisilicon/sec/sec_algs.c
881
if (IS_ERR(ctx->queue))
drivers/crypto/hisilicon/sec/sec_algs.c
882
return PTR_ERR(ctx->queue);
drivers/crypto/hisilicon/sec/sec_algs.c
884
spin_lock_init(&ctx->queue->queuelock);
drivers/crypto/hisilicon/sec/sec_algs.c
885
ctx->queue->havesoftqueue = false;
drivers/crypto/hisilicon/sec/sec_algs.c
893
struct device *dev = ctx->queue->dev_info->dev;
drivers/crypto/hisilicon/sec/sec_algs.c
900
sec_queue_stop_release(ctx->queue);
drivers/crypto/hisilicon/sec/sec_algs.c
912
INIT_KFIFO(ctx->queue->softqueue);
drivers/crypto/hisilicon/sec/sec_algs.c
913
ret = kfifo_alloc(&ctx->queue->softqueue, 512, GFP_KERNEL);
drivers/crypto/hisilicon/sec/sec_algs.c
918
ctx->queue->havesoftqueue = true;
drivers/crypto/hisilicon/sec/sec_algs.c
927
kfifo_free(&ctx->queue->softqueue);
drivers/crypto/hisilicon/sec/sec_drv.c
1003
struct sec_queue *queue, int queue_id)
drivers/crypto/hisilicon/sec/sec_drv.c
1005
queue->dev_info = info;
drivers/crypto/hisilicon/sec/sec_drv.c
1006
queue->queue_id = queue_id;
drivers/crypto/hisilicon/sec/sec_drv.c
1007
snprintf(queue->name, sizeof(queue->name),
drivers/crypto/hisilicon/sec/sec_drv.c
1008
"%s_%d", dev_name(info->dev), queue->queue_id);
drivers/crypto/hisilicon/sec/sec_drv.c
1079
static int sec_queue_res_cfg(struct sec_queue *queue)
drivers/crypto/hisilicon/sec/sec_drv.c
1081
struct device *dev = queue->dev_info->dev;
drivers/crypto/hisilicon/sec/sec_drv.c
1082
struct sec_queue_ring_cmd *ring_cmd = &queue->ring_cmd;
drivers/crypto/hisilicon/sec/sec_drv.c
1083
struct sec_queue_ring_cq *ring_cq = &queue->ring_cq;
drivers/crypto/hisilicon/sec/sec_drv.c
1084
struct sec_queue_ring_db *ring_db = &queue->ring_db;
drivers/crypto/hisilicon/sec/sec_drv.c
1109
queue->task_irq = platform_get_irq(to_platform_device(dev),
drivers/crypto/hisilicon/sec/sec_drv.c
1110
queue->queue_id * 2 + 1);
drivers/crypto/hisilicon/sec/sec_drv.c
1111
if (queue->task_irq < 0) {
drivers/crypto/hisilicon/sec/sec_drv.c
1112
ret = queue->task_irq;
drivers/crypto/hisilicon/sec/sec_drv.c
1119
dma_free_coherent(dev, SEC_Q_DB_SIZE, queue->ring_db.vaddr,
drivers/crypto/hisilicon/sec/sec_drv.c
1120
queue->ring_db.paddr);
drivers/crypto/hisilicon/sec/sec_drv.c
1122
dma_free_coherent(dev, SEC_Q_CQ_SIZE, queue->ring_cq.vaddr,
drivers/crypto/hisilicon/sec/sec_drv.c
1123
queue->ring_cq.paddr);
drivers/crypto/hisilicon/sec/sec_drv.c
1125
dma_free_coherent(dev, SEC_Q_CMD_SIZE, queue->ring_cmd.vaddr,
drivers/crypto/hisilicon/sec/sec_drv.c
1126
queue->ring_cmd.paddr);
drivers/crypto/hisilicon/sec/sec_drv.c
1131
static void sec_queue_free_ring_pages(struct sec_queue *queue)
drivers/crypto/hisilicon/sec/sec_drv.c
1133
struct device *dev = queue->dev_info->dev;
drivers/crypto/hisilicon/sec/sec_drv.c
1135
dma_free_coherent(dev, SEC_Q_DB_SIZE, queue->ring_db.vaddr,
drivers/crypto/hisilicon/sec/sec_drv.c
1136
queue->ring_db.paddr);
drivers/crypto/hisilicon/sec/sec_drv.c
1137
dma_free_coherent(dev, SEC_Q_CQ_SIZE, queue->ring_cq.vaddr,
drivers/crypto/hisilicon/sec/sec_drv.c
1138
queue->ring_cq.paddr);
drivers/crypto/hisilicon/sec/sec_drv.c
1139
dma_free_coherent(dev, SEC_Q_CMD_SIZE, queue->ring_cmd.vaddr,
drivers/crypto/hisilicon/sec/sec_drv.c
1140
queue->ring_cmd.paddr);
drivers/crypto/hisilicon/sec/sec_drv.c
1143
static int sec_queue_config(struct sec_dev_info *info, struct sec_queue *queue,
drivers/crypto/hisilicon/sec/sec_drv.c
1148
sec_queue_base_init(info, queue, queue_id);
drivers/crypto/hisilicon/sec/sec_drv.c
1150
ret = sec_queue_res_cfg(queue);
drivers/crypto/hisilicon/sec/sec_drv.c
1154
ret = sec_queue_map_io(queue);
drivers/crypto/hisilicon/sec/sec_drv.c
1157
sec_queue_free_ring_pages(queue);
drivers/crypto/hisilicon/sec/sec_drv.c
1161
sec_queue_hw_init(queue);
drivers/crypto/hisilicon/sec/sec_drv.c
1167
struct sec_queue *queue)
drivers/crypto/hisilicon/sec/sec_drv.c
1169
sec_queue_unmap_io(queue);
drivers/crypto/hisilicon/sec/sec_drv.c
1170
sec_queue_free_ring_pages(queue);
drivers/crypto/hisilicon/sec/sec_drv.c
227
static int sec_queue_map_io(struct sec_queue *queue)
drivers/crypto/hisilicon/sec/sec_drv.c
229
struct device *dev = queue->dev_info->dev;
drivers/crypto/hisilicon/sec/sec_drv.c
234
2 + queue->queue_id);
drivers/crypto/hisilicon/sec/sec_drv.c
237
queue->queue_id);
drivers/crypto/hisilicon/sec/sec_drv.c
240
queue->regs = ioremap(res->start, resource_size(res));
drivers/crypto/hisilicon/sec/sec_drv.c
241
if (!queue->regs)
drivers/crypto/hisilicon/sec/sec_drv.c
247
static void sec_queue_unmap_io(struct sec_queue *queue)
drivers/crypto/hisilicon/sec/sec_drv.c
249
iounmap(queue->regs);
drivers/crypto/hisilicon/sec/sec_drv.c
252
static int sec_queue_ar_pkgattr(struct sec_queue *queue, u32 ar_pkg)
drivers/crypto/hisilicon/sec/sec_drv.c
254
void __iomem *addr = queue->regs + SEC_Q_ARUSER_CFG_REG;
drivers/crypto/hisilicon/sec/sec_drv.c
267
static int sec_queue_aw_pkgattr(struct sec_queue *queue, u32 aw_pkg)
drivers/crypto/hisilicon/sec/sec_drv.c
269
void __iomem *addr = queue->regs + SEC_Q_AWUSER_CFG_REG;
drivers/crypto/hisilicon/sec/sec_drv.c
519
static void sec_queue_ar_alloc(struct sec_queue *queue, u32 alloc)
drivers/crypto/hisilicon/sec/sec_drv.c
521
void __iomem *addr = queue->regs + SEC_Q_ARUSER_CFG_REG;
drivers/crypto/hisilicon/sec/sec_drv.c
536
static void sec_queue_aw_alloc(struct sec_queue *queue, u32 alloc)
drivers/crypto/hisilicon/sec/sec_drv.c
538
void __iomem *addr = queue->regs + SEC_Q_AWUSER_CFG_REG;
drivers/crypto/hisilicon/sec/sec_drv.c
553
static void sec_queue_reorder(struct sec_queue *queue, bool reorder)
drivers/crypto/hisilicon/sec/sec_drv.c
555
void __iomem *base = queue->regs;
drivers/crypto/hisilicon/sec/sec_drv.c
566
static void sec_queue_depth(struct sec_queue *queue, u32 depth)
drivers/crypto/hisilicon/sec/sec_drv.c
568
void __iomem *addr = queue->regs + SEC_Q_DEPTH_CFG_REG;
drivers/crypto/hisilicon/sec/sec_drv.c
578
static void sec_queue_cmdbase_addr(struct sec_queue *queue, u64 addr)
drivers/crypto/hisilicon/sec/sec_drv.c
580
writel_relaxed(upper_32_bits(addr), queue->regs + SEC_Q_BASE_HADDR_REG);
drivers/crypto/hisilicon/sec/sec_drv.c
581
writel_relaxed(lower_32_bits(addr), queue->regs + SEC_Q_BASE_LADDR_REG);
drivers/crypto/hisilicon/sec/sec_drv.c
584
static void sec_queue_outorder_addr(struct sec_queue *queue, u64 addr)
drivers/crypto/hisilicon/sec/sec_drv.c
587
queue->regs + SEC_Q_OUTORDER_BASE_HADDR_REG);
drivers/crypto/hisilicon/sec/sec_drv.c
589
queue->regs + SEC_Q_OUTORDER_BASE_LADDR_REG);
drivers/crypto/hisilicon/sec/sec_drv.c
592
static void sec_queue_errbase_addr(struct sec_queue *queue, u64 addr)
drivers/crypto/hisilicon/sec/sec_drv.c
595
queue->regs + SEC_Q_ERR_BASE_HADDR_REG);
drivers/crypto/hisilicon/sec/sec_drv.c
597
queue->regs + SEC_Q_ERR_BASE_LADDR_REG);
drivers/crypto/hisilicon/sec/sec_drv.c
600
static void sec_queue_irq_disable(struct sec_queue *queue)
drivers/crypto/hisilicon/sec/sec_drv.c
602
writel_relaxed((u32)~0, queue->regs + SEC_Q_FLOW_INT_MKS_REG);
drivers/crypto/hisilicon/sec/sec_drv.c
605
static void sec_queue_irq_enable(struct sec_queue *queue)
drivers/crypto/hisilicon/sec/sec_drv.c
607
writel_relaxed(0, queue->regs + SEC_Q_FLOW_INT_MKS_REG);
drivers/crypto/hisilicon/sec/sec_drv.c
610
static void sec_queue_abn_irq_disable(struct sec_queue *queue)
drivers/crypto/hisilicon/sec/sec_drv.c
612
writel_relaxed((u32)~0, queue->regs + SEC_Q_FAIL_INT_MSK_REG);
drivers/crypto/hisilicon/sec/sec_drv.c
615
static void sec_queue_stop(struct sec_queue *queue)
drivers/crypto/hisilicon/sec/sec_drv.c
617
disable_irq(queue->task_irq);
drivers/crypto/hisilicon/sec/sec_drv.c
618
sec_queue_irq_disable(queue);
drivers/crypto/hisilicon/sec/sec_drv.c
619
writel_relaxed(0x0, queue->regs + SEC_QUEUE_ENB_REG);
drivers/crypto/hisilicon/sec/sec_drv.c
622
static void sec_queue_start(struct sec_queue *queue)
drivers/crypto/hisilicon/sec/sec_drv.c
624
sec_queue_irq_enable(queue);
drivers/crypto/hisilicon/sec/sec_drv.c
625
enable_irq(queue->task_irq);
drivers/crypto/hisilicon/sec/sec_drv.c
626
queue->expected = 0;
drivers/crypto/hisilicon/sec/sec_drv.c
627
writel_relaxed(SEC_Q_INIT_AND_STAT_CLEAR, queue->regs + SEC_Q_INIT_REG);
drivers/crypto/hisilicon/sec/sec_drv.c
628
writel_relaxed(0x1, queue->regs + SEC_QUEUE_ENB_REG);
drivers/crypto/hisilicon/sec/sec_drv.c
651
static int sec_queue_free(struct sec_queue *queue)
drivers/crypto/hisilicon/sec/sec_drv.c
653
struct sec_dev_info *info = queue->dev_info;
drivers/crypto/hisilicon/sec/sec_drv.c
655
if (queue->queue_id >= SEC_Q_NUM) {
drivers/crypto/hisilicon/sec/sec_drv.c
656
dev_err(info->dev, "No queue %u\n", queue->queue_id);
drivers/crypto/hisilicon/sec/sec_drv.c
660
if (!queue->in_use) {
drivers/crypto/hisilicon/sec/sec_drv.c
661
dev_err(info->dev, "Queue %u is idle\n", queue->queue_id);
drivers/crypto/hisilicon/sec/sec_drv.c
666
queue->in_use = false;
drivers/crypto/hisilicon/sec/sec_drv.c
681
struct sec_queue *queue = q;
drivers/crypto/hisilicon/sec/sec_drv.c
682
struct sec_queue_ring_cmd *msg_ring = &queue->ring_cmd;
drivers/crypto/hisilicon/sec/sec_drv.c
683
struct sec_queue_ring_cq *cq_ring = &queue->ring_cq;
drivers/crypto/hisilicon/sec/sec_drv.c
687
void __iomem *base = queue->regs;
drivers/crypto/hisilicon/sec/sec_drv.c
701
set_bit(q_id, queue->unprocessed);
drivers/crypto/hisilicon/sec/sec_drv.c
702
if (q_id == queue->expected)
drivers/crypto/hisilicon/sec/sec_drv.c
703
while (test_bit(queue->expected, queue->unprocessed)) {
drivers/crypto/hisilicon/sec/sec_drv.c
704
clear_bit(queue->expected, queue->unprocessed);
drivers/crypto/hisilicon/sec/sec_drv.c
705
msg = msg_ring->vaddr + queue->expected;
drivers/crypto/hisilicon/sec/sec_drv.c
708
queue->shadow[queue->expected]);
drivers/crypto/hisilicon/sec/sec_drv.c
709
queue->shadow[queue->expected] = NULL;
drivers/crypto/hisilicon/sec/sec_drv.c
710
queue->expected = (queue->expected + 1) %
drivers/crypto/hisilicon/sec/sec_drv.c
723
sec_queue_irq_enable(queue);
drivers/crypto/hisilicon/sec/sec_drv.c
728
static int sec_queue_irq_init(struct sec_queue *queue)
drivers/crypto/hisilicon/sec/sec_drv.c
730
struct sec_dev_info *info = queue->dev_info;
drivers/crypto/hisilicon/sec/sec_drv.c
731
int irq = queue->task_irq;
drivers/crypto/hisilicon/sec/sec_drv.c
735
IRQF_TRIGGER_RISING, queue->name, queue);
drivers/crypto/hisilicon/sec/sec_drv.c
745
static int sec_queue_irq_uninit(struct sec_queue *queue)
drivers/crypto/hisilicon/sec/sec_drv.c
747
free_irq(queue->task_irq, queue);
drivers/crypto/hisilicon/sec/sec_drv.c
774
struct sec_queue *queue;
drivers/crypto/hisilicon/sec/sec_drv.c
776
queue = sec_alloc_queue(info);
drivers/crypto/hisilicon/sec/sec_drv.c
777
if (IS_ERR(queue)) {
drivers/crypto/hisilicon/sec/sec_drv.c
779
PTR_ERR(queue));
drivers/crypto/hisilicon/sec/sec_drv.c
780
return queue;
drivers/crypto/hisilicon/sec/sec_drv.c
783
sec_queue_start(queue);
drivers/crypto/hisilicon/sec/sec_drv.c
785
return queue;
drivers/crypto/hisilicon/sec/sec_drv.c
799
struct sec_queue *queue = ERR_PTR(-ENODEV);
drivers/crypto/hisilicon/sec/sec_drv.c
806
queue = sec_queue_alloc_start(info);
drivers/crypto/hisilicon/sec/sec_drv.c
811
return queue;
drivers/crypto/hisilicon/sec/sec_drv.c
821
int sec_queue_stop_release(struct sec_queue *queue)
drivers/crypto/hisilicon/sec/sec_drv.c
823
struct device *dev = queue->dev_info->dev;
drivers/crypto/hisilicon/sec/sec_drv.c
826
sec_queue_stop(queue);
drivers/crypto/hisilicon/sec/sec_drv.c
828
ret = sec_queue_free(queue);
drivers/crypto/hisilicon/sec/sec_drv.c
843
bool sec_queue_empty(struct sec_queue *queue)
drivers/crypto/hisilicon/sec/sec_drv.c
845
struct sec_queue_ring_cmd *msg_ring = &queue->ring_cmd;
drivers/crypto/hisilicon/sec/sec_drv.c
858
int sec_queue_send(struct sec_queue *queue, struct sec_bd_info *msg, void *ctx)
drivers/crypto/hisilicon/sec/sec_drv.c
860
struct sec_queue_ring_cmd *msg_ring = &queue->ring_cmd;
drivers/crypto/hisilicon/sec/sec_drv.c
861
void __iomem *base = queue->regs;
drivers/crypto/hisilicon/sec/sec_drv.c
872
queue->shadow[write] = ctx;
drivers/crypto/hisilicon/sec/sec_drv.c
885
bool sec_queue_can_enqueue(struct sec_queue *queue, int num)
drivers/crypto/hisilicon/sec/sec_drv.c
887
struct sec_queue_ring_cmd *msg_ring = &queue->ring_cmd;
drivers/crypto/hisilicon/sec/sec_drv.c
892
static void sec_queue_hw_init(struct sec_queue *queue)
drivers/crypto/hisilicon/sec/sec_drv.c
894
sec_queue_ar_alloc(queue, SEC_QUEUE_AR_FROCE_NOALLOC);
drivers/crypto/hisilicon/sec/sec_drv.c
895
sec_queue_aw_alloc(queue, SEC_QUEUE_AW_FROCE_NOALLOC);
drivers/crypto/hisilicon/sec/sec_drv.c
896
sec_queue_ar_pkgattr(queue, 1);
drivers/crypto/hisilicon/sec/sec_drv.c
897
sec_queue_aw_pkgattr(queue, 1);
drivers/crypto/hisilicon/sec/sec_drv.c
900
sec_queue_reorder(queue, true);
drivers/crypto/hisilicon/sec/sec_drv.c
903
writel_relaxed(1, queue->regs + SEC_Q_PROC_NUM_CFG_REG);
drivers/crypto/hisilicon/sec/sec_drv.c
905
sec_queue_depth(queue, SEC_QUEUE_LEN - 1);
drivers/crypto/hisilicon/sec/sec_drv.c
907
sec_queue_cmdbase_addr(queue, queue->ring_cmd.paddr);
drivers/crypto/hisilicon/sec/sec_drv.c
909
sec_queue_outorder_addr(queue, queue->ring_cq.paddr);
drivers/crypto/hisilicon/sec/sec_drv.c
911
sec_queue_errbase_addr(queue, queue->ring_db.paddr);
drivers/crypto/hisilicon/sec/sec_drv.c
913
writel_relaxed(0x100, queue->regs + SEC_Q_OT_TH_REG);
drivers/crypto/hisilicon/sec/sec_drv.c
915
sec_queue_abn_irq_disable(queue);
drivers/crypto/hisilicon/sec/sec_drv.c
916
sec_queue_irq_disable(queue);
drivers/crypto/hisilicon/sec/sec_drv.c
917
writel_relaxed(SEC_Q_INIT_AND_STAT_CLEAR, queue->regs + SEC_Q_INIT_REG);
drivers/crypto/hisilicon/sec/sec_drv.h
247
struct sec_queue *queue;
drivers/crypto/hisilicon/sec/sec_drv.h
417
int sec_queue_send(struct sec_queue *queue, struct sec_bd_info *msg, void *ctx);
drivers/crypto/hisilicon/sec/sec_drv.h
418
bool sec_queue_can_enqueue(struct sec_queue *queue, int num);
drivers/crypto/hisilicon/sec/sec_drv.h
419
int sec_queue_stop_release(struct sec_queue *queue);
drivers/crypto/hisilicon/sec/sec_drv.h
421
bool sec_queue_empty(struct sec_queue *queue);
drivers/crypto/img-hash.c
133
struct crypto_queue queue;
drivers/crypto/img-hash.c
512
res = ahash_enqueue_request(&hdev->queue, req);
drivers/crypto/img-hash.c
519
backlog = crypto_get_backlog(&hdev->queue);
drivers/crypto/img-hash.c
520
async_req = crypto_dequeue_request(&hdev->queue);
drivers/crypto/img-hash.c
957
crypto_init_queue(&hdev->queue, IMG_HASH_QUEUE_LENGTH);
drivers/crypto/inside-secure/safexcel.c
1677
crypto_init_queue(&priv->ring[i].queue,
drivers/crypto/inside-secure/safexcel.c
836
backlog = crypto_get_backlog(&priv->ring[ring].queue);
drivers/crypto/inside-secure/safexcel.c
837
req = crypto_dequeue_request(&priv->ring[ring].queue);
drivers/crypto/inside-secure/safexcel.h
708
struct crypto_queue queue;
drivers/crypto/inside-secure/safexcel_cipher.c
1109
crypto_enqueue_request(&priv->ring[ring].queue, base);
drivers/crypto/inside-secure/safexcel_cipher.c
1185
ret = crypto_enqueue_request(&priv->ring[ring].queue, base);
drivers/crypto/inside-secure/safexcel_cipher.c
962
enq_ret = crypto_enqueue_request(&priv->ring[ring].queue, base);
drivers/crypto/inside-secure/safexcel_hash.c
551
enq_ret = crypto_enqueue_request(&priv->ring[ring].queue, async);
drivers/crypto/inside-secure/safexcel_hash.c
643
crypto_enqueue_request(&priv->ring[ring].queue, &req->base);
drivers/crypto/inside-secure/safexcel_hash.c
731
ret = crypto_enqueue_request(&priv->ring[ring].queue, &areq->base);
drivers/crypto/marvell/cesa/cesa.c
170
ret = crypto_enqueue_request(&engine->queue, req);
drivers/crypto/marvell/cesa/cesa.c
41
*backlog = crypto_get_backlog(&engine->queue);
drivers/crypto/marvell/cesa/cesa.c
43
return crypto_dequeue_request(&engine->queue);
drivers/crypto/marvell/cesa/cesa.c
529
crypto_init_queue(&engine->queue, CESA_CRYPTO_DEFAULT_MAX_QLEN);
drivers/crypto/marvell/cesa/cesa.h
466
struct crypto_queue queue;
drivers/crypto/marvell/octeontx/otx_cptvf.h
48
struct otx_cpt_cmd_queue queue[OTX_CPT_NUM_QS_PER_VF];
drivers/crypto/marvell/octeontx/otx_cptvf.h
53
struct otx_cpt_pending_queue queue[OTX_CPT_NUM_QS_PER_VF];
drivers/crypto/marvell/octeontx/otx_cptvf.h
57
for (i = 0, q = &qinfo->queue[i]; i < qinfo->num_queues; i++, \
drivers/crypto/marvell/octeontx/otx_cptvf.h
58
q = &qinfo->queue[i])
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
102
for_each_pending_queue(pqinfo, queue, i) {
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
103
queue->head = kzalloc_objs(*queue->head, qlen);
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
104
if (!queue->head) {
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
109
queue->pending_count = 0;
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
110
queue->front = 0;
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
111
queue->rear = 0;
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
112
queue->qlen = qlen;
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
115
spin_lock_init(&queue->lock);
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
158
struct otx_cpt_cmd_queue *queue = NULL;
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
165
queue = &cqinfo->queue[i];
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
167
while (!list_empty(&cqinfo->queue[i].chead)) {
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
168
chunk = list_first_entry(&cqinfo->queue[i].chead,
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
180
queue->num_chunks = 0;
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
181
queue->idx = 0;
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
191
struct otx_cpt_cmd_queue *queue = NULL;
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
212
queue = &cqinfo->queue[i];
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
213
INIT_LIST_HEAD(&queue->chead);
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
227
i, queue->num_chunks);
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
232
if (queue->num_chunks == 0) {
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
234
queue->base = first;
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
237
&cqinfo->queue[i].chead);
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
239
queue->num_chunks++;
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
253
queue->qhead = curr;
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
634
base_addr = (u64)(cptvf->cqinfo.queue[0].qhead->dma_addr);
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
77
struct otx_cpt_pending_queue *queue;
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
80
for_each_pending_queue(pqinfo, queue, i) {
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
81
if (!queue->head)
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
85
kfree_sensitive((queue->head));
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
86
queue->front = 0;
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
87
queue->rear = 0;
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
88
queue->qlen = 0;
drivers/crypto/marvell/octeontx/otx_cptvf_main.c
96
struct otx_cpt_pending_queue *queue = NULL;
drivers/crypto/marvell/octeontx/otx_cptvf_reqmgr.c
271
struct otx_cpt_cmd_queue *queue;
drivers/crypto/marvell/octeontx/otx_cptvf_reqmgr.c
275
queue = &qinfo->queue[0];
drivers/crypto/marvell/octeontx/otx_cptvf_reqmgr.c
280
ent = &queue->qhead->head[queue->idx * OTX_CPT_INST_SIZE];
drivers/crypto/marvell/octeontx/otx_cptvf_reqmgr.c
283
if (++queue->idx >= queue->qhead->size / 64) {
drivers/crypto/marvell/octeontx/otx_cptvf_reqmgr.c
284
curr = queue->qhead;
drivers/crypto/marvell/octeontx/otx_cptvf_reqmgr.c
286
if (list_is_last(&curr->nextchunk, &queue->chead))
drivers/crypto/marvell/octeontx/otx_cptvf_reqmgr.c
287
queue->qhead = queue->base;
drivers/crypto/marvell/octeontx/otx_cptvf_reqmgr.c
289
queue->qhead = list_next_entry(queue->qhead, nextchunk);
drivers/crypto/marvell/octeontx/otx_cptvf_reqmgr.c
290
queue->idx = 0;
drivers/crypto/marvell/octeontx/otx_cptvf_reqmgr.c
425
return process_request(pdev, req, &cptvf->pqinfo.queue[0], cptvf);
drivers/crypto/marvell/octeontx/otx_cptvf_reqmgr.c
608
process_pending_queue(wqe->cptvf->pdev, &wqe->cptvf->pqinfo.queue[0]);
drivers/crypto/mxs-dcp.c
1149
crypto_init_queue(&sdcp->queue[i], 50);
drivers/crypto/mxs-dcp.c
424
backlog = crypto_get_backlog(&sdcp->queue[chan]);
drivers/crypto/mxs-dcp.c
425
arq = crypto_dequeue_request(&sdcp->queue[chan]);
drivers/crypto/mxs-dcp.c
484
ret = crypto_enqueue_request(&sdcp->queue[actx->chan], &req->base);
drivers/crypto/mxs-dcp.c
754
backlog = crypto_get_backlog(&sdcp->queue[chan]);
drivers/crypto/mxs-dcp.c
755
arq = crypto_dequeue_request(&sdcp->queue[chan]);
drivers/crypto/mxs-dcp.c
82
struct crypto_queue queue[DCP_MAX_CHANS];
drivers/crypto/mxs-dcp.c
831
ret = crypto_enqueue_request(&sdcp->queue[actx->chan], &req->base);
drivers/crypto/omap-aes.c
1046
return sysfs_emit(buf, "%d\n", dd->engine->queue.max_qlen);
drivers/crypto/omap-aes.c
1073
dd->engine->queue.max_qlen = value;
drivers/crypto/omap-sham.c
2007
return sysfs_emit(buf, "%d\n", dd->queue.max_qlen);
drivers/crypto/omap-sham.c
2030
dd->queue.max_qlen = value;
drivers/crypto/omap-sham.c
2065
crypto_init_queue(&dd->queue, OMAP_SHAM_QUEUE_LENGTH);
drivers/crypto/omap-sham.c
228
struct crypto_queue queue;
drivers/crypto/qce/core.c
101
backlog = crypto_get_backlog(&qce->queue);
drivers/crypto/qce/core.c
102
async_req = crypto_dequeue_request(&qce->queue);
drivers/crypto/qce/core.c
243
crypto_init_queue(&qce->queue, QCE_QUEUE_LENGTH);
drivers/crypto/qce/core.c
95
ret = crypto_enqueue_request(&qce->queue, req);
drivers/crypto/qce/core.h
33
struct crypto_queue queue;
drivers/crypto/s5p-sss.c
1958
backlog = crypto_get_backlog(&dev->queue);
drivers/crypto/s5p-sss.c
1959
async_req = crypto_dequeue_request(&dev->queue);
drivers/crypto/s5p-sss.c
1985
err = crypto_enqueue_request(&dev->queue, &req->base);
drivers/crypto/s5p-sss.c
2236
crypto_init_queue(&pdata->queue, CRYPTO_QUEUE_LEN);
drivers/crypto/s5p-sss.c
312
struct crypto_queue queue;
drivers/dax/super.c
88
if (!blk_queue_dax(bdev->bd_disk->queue))
drivers/dma/amd/qdma/qdma.c
284
static int qdma_clear_queue_context(const struct qdma_queue *queue)
drivers/dma/amd/qdma/qdma.c
298
struct qdma_device *qdev = queue->qdev;
drivers/dma/amd/qdma/qdma.c
302
if (queue->dir == DMA_MEM_TO_DEV) {
drivers/dma/amd/qdma/qdma.c
311
queue->qid, NULL);
drivers/dma/amd/qdma/qdma.c
408
static inline int qdma_update_pidx(const struct qdma_queue *queue, u16 pidx)
drivers/dma/amd/qdma/qdma.c
410
struct qdma_device *qdev = queue->qdev;
drivers/dma/amd/qdma/qdma.c
412
return regmap_write(qdev->regmap, queue->pidx_reg,
drivers/dma/amd/qdma/qdma.c
416
static inline int qdma_update_cidx(const struct qdma_queue *queue,
drivers/dma/amd/qdma/qdma.c
419
struct qdma_device *qdev = queue->qdev;
drivers/dma/amd/qdma/qdma.c
421
return regmap_write(qdev->regmap, queue->cidx_reg,
drivers/dma/amd/qdma/qdma.c
544
struct qdma_queue *queue = to_qdma_queue(chan);
drivers/dma/amd/qdma/qdma.c
545
struct qdma_device *qdev = queue->qdev;
drivers/dma/amd/qdma/qdma.c
548
qdma_clear_queue_context(queue);
drivers/dma/amd/qdma/qdma.c
549
vchan_free_chan_resources(&queue->vchan);
drivers/dma/amd/qdma/qdma.c
551
dma_free_coherent(pdata->dma_dev, queue->ring_size * QDMA_MM_DESC_SIZE,
drivers/dma/amd/qdma/qdma.c
552
queue->desc_base, queue->dma_desc_base);
drivers/dma/amd/qdma/qdma.c
561
struct qdma_queue *queue = to_qdma_queue(chan);
drivers/dma/amd/qdma/qdma.c
562
struct qdma_device *qdev = queue->qdev;
drivers/dma/amd/qdma/qdma.c
568
ret = qdma_clear_queue_context(queue);
drivers/dma/amd/qdma/qdma.c
573
size = queue->ring_size * QDMA_MM_DESC_SIZE;
drivers/dma/amd/qdma/qdma.c
574
queue->desc_base = dma_alloc_coherent(pdata->dma_dev, size,
drivers/dma/amd/qdma/qdma.c
575
&queue->dma_desc_base,
drivers/dma/amd/qdma/qdma.c
577
if (!queue->desc_base) {
drivers/dma/amd/qdma/qdma.c
584
desc.desc_base = queue->dma_desc_base;
drivers/dma/amd/qdma/qdma.c
585
ret = qdma_setup_queue_context(qdev, &desc, queue->dir, queue->qid);
drivers/dma/amd/qdma/qdma.c
589
dma_free_coherent(pdata->dma_dev, size, queue->desc_base,
drivers/dma/amd/qdma/qdma.c
590
queue->dma_desc_base);
drivers/dma/amd/qdma/qdma.c
594
queue->pidx = 0;
drivers/dma/amd/qdma/qdma.c
595
queue->cidx = 0;
drivers/dma/amd/qdma/qdma.c
602
struct qdma_queue *queue = to_qdma_queue(chan);
drivers/dma/amd/qdma/qdma.c
605
return info->dir == queue->dir;
drivers/dma/amd/qdma/qdma.c
608
static int qdma_xfer_start(struct qdma_queue *queue)
drivers/dma/amd/qdma/qdma.c
610
struct qdma_device *qdev = queue->qdev;
drivers/dma/amd/qdma/qdma.c
613
if (!vchan_next_desc(&queue->vchan))
drivers/dma/amd/qdma/qdma.c
617
queue->issued_vdesc->pidx, CHAN_STR(queue), queue->qid);
drivers/dma/amd/qdma/qdma.c
619
ret = qdma_update_pidx(queue, queue->issued_vdesc->pidx);
drivers/dma/amd/qdma/qdma.c
622
queue->pidx, CHAN_STR(queue), queue->qid);
drivers/dma/amd/qdma/qdma.c
630
struct qdma_queue *queue = to_qdma_queue(chan);
drivers/dma/amd/qdma/qdma.c
633
spin_lock_irqsave(&queue->vchan.lock, flags);
drivers/dma/amd/qdma/qdma.c
634
if (vchan_issue_pending(&queue->vchan)) {
drivers/dma/amd/qdma/qdma.c
635
if (queue->submitted_vdesc) {
drivers/dma/amd/qdma/qdma.c
636
queue->issued_vdesc = queue->submitted_vdesc;
drivers/dma/amd/qdma/qdma.c
637
queue->submitted_vdesc = NULL;
drivers/dma/amd/qdma/qdma.c
639
qdma_xfer_start(queue);
drivers/dma/amd/qdma/qdma.c
642
spin_unlock_irqrestore(&queue->vchan.lock, flags);
drivers/dma/amd/qdma/qdma.h
192
struct qdma_queue *queue;
drivers/dma/dw/core.c
1109
BUG_ON(!list_empty(&dwc->queue));
drivers/dma/dw/core.c
1258
INIT_LIST_HEAD(&dwc->queue);
drivers/dma/dw/core.c
221
if (list_empty(&dwc->queue))
drivers/dma/dw/core.c
224
list_move(dwc->queue.next, &dwc->active_list);
drivers/dma/dw/core.c
438
list_move(dwc->queue.next, dwc->active_list.prev);
drivers/dma/dw/core.c
70
list_add_tail(&desc->desc_node, &dwc->queue);
drivers/dma/dw/core.c
956
list_splice_init(&dwc->queue, &list);
drivers/dma/dw/regs.h
281
struct list_head queue;
drivers/dma/ep93xx_dma.c
1009
BUG_ON(!list_empty(&edmac->queue));
drivers/dma/ep93xx_dma.c
1268
list_splice_init(&edmac->queue, &list);
drivers/dma/ep93xx_dma.c
1422
INIT_LIST_HEAD(&edmac->queue);
drivers/dma/ep93xx_dma.c
209
struct list_head queue;
drivers/dma/ep93xx_dma.c
789
if (!list_empty(&edmac->active) || list_empty(&edmac->queue)) {
drivers/dma/ep93xx_dma.c
795
new = list_first_entry(&edmac->queue, struct ep93xx_dma_desc, node);
drivers/dma/ep93xx_dma.c
909
list_add_tail(&desc->node, &edmac->queue);
drivers/dma/fsl-dpaa2-qdma/dpdmai.c
26
u8 queue;
drivers/dma/fsl-dpaa2-qdma/dpdmai.c
304
cmd_params->queue = priority;
drivers/dma/fsl-dpaa2-qdma/dpdmai.c
348
cmd_params->queue = priority;
drivers/dma/fsl-qdma.c
1028
fsl_queue = fsl_comp->qchan->queue;
drivers/dma/fsl-qdma.c
1039
struct fsl_qdma_queue *fsl_queue = fsl_chan->queue;
drivers/dma/fsl-qdma.c
1074
struct fsl_qdma_queue *fsl_queue = fsl_chan->queue;
drivers/dma/fsl-qdma.c
1199
fsl_qdma->queue = fsl_qdma_alloc_queue_resources(pdev, fsl_qdma);
drivers/dma/fsl-qdma.c
1200
if (!fsl_qdma->queue)
drivers/dma/fsl-qdma.c
1214
fsl_chan->queue = fsl_qdma->queue + i % (fsl_qdma->n_queues *
drivers/dma/fsl-qdma.c
178
u8 queue;
drivers/dma/fsl-qdma.c
188
struct fsl_qdma_queue *queue;
drivers/dma/fsl-qdma.c
227
struct fsl_qdma_queue *queue;
drivers/dma/fsl-qdma.c
315
struct fsl_qdma_queue *fsl_queue = fsl_chan->queue;
drivers/dma/fsl-qdma.c
403
static int fsl_qdma_pre_request_enqueue_desc(struct fsl_qdma_queue *queue)
drivers/dma/fsl-qdma.c
408
for (i = 0; i < queue->n_cq + FSL_COMMAND_QUEUE_OVERFLLOW; i++) {
drivers/dma/fsl-qdma.c
413
dma_pool_alloc(queue->comp_pool, GFP_KERNEL,
drivers/dma/fsl-qdma.c
419
dma_pool_alloc(queue->desc_pool, GFP_KERNEL,
drivers/dma/fsl-qdma.c
424
list_add_tail(&comp_temp->list, &queue->comp_free);
drivers/dma/fsl-qdma.c
430
dma_pool_free(queue->comp_pool, comp_temp->virt_addr,
drivers/dma/fsl-qdma.c
438
&queue->comp_free, list) {
drivers/dma/fsl-qdma.c
440
dma_pool_free(queue->comp_pool,
drivers/dma/fsl-qdma.c
444
dma_pool_free(queue->desc_pool,
drivers/dma/fsl-qdma.c
464
struct fsl_qdma_queue *queue = fsl_chan->queue;
drivers/dma/fsl-qdma.c
467
spin_lock_irqsave(&queue->queue_lock, flags);
drivers/dma/fsl-qdma.c
468
if (!list_empty(&queue->comp_free)) {
drivers/dma/fsl-qdma.c
469
comp_temp = list_first_entry(&queue->comp_free,
drivers/dma/fsl-qdma.c
474
spin_unlock_irqrestore(&queue->queue_lock, flags);
drivers/dma/fsl-qdma.c
478
spin_unlock_irqrestore(&queue->queue_lock, flags);
drivers/dma/fsl-qdma.c
640
struct fsl_qdma_queue *fsl_queue = fsl_qdma->queue;
drivers/dma/fsl-qdma.c
654
__this_cpu_read(pre.queue) &&
drivers/dma/fsl-qdma.c
661
__this_cpu_write(pre.queue, qdma_ccdf_get_queue(status_addr));
drivers/dma/fsl-qdma.c
878
struct fsl_qdma_queue *fsl_queue = fsl_qdma->queue;
drivers/dma/fsl-qdma.c
996
struct fsl_qdma_queue *fsl_queue = fsl_chan->queue;
drivers/dma/mediatek/mtk-cqdma.c
101
struct list_head queue;
drivers/dma/mediatek/mtk-cqdma.c
277
if (list_empty(&pc->queue))
drivers/dma/mediatek/mtk-cqdma.c
283
list_add_tail(&cvd->node, &pc->queue);
drivers/dma/mediatek/mtk-cqdma.c
302
list_for_each_entry(cvd, &cvc->pc->queue, node)
drivers/dma/mediatek/mtk-cqdma.c
319
cvd = list_first_entry_or_null(&pc->queue,
drivers/dma/mediatek/mtk-cqdma.c
350
cvd = list_first_entry_or_null(&pc->queue,
drivers/dma/mediatek/mtk-cqdma.c
424
list_for_each_entry(vd, &cvc->pc->queue, node)
drivers/dma/mediatek/mtk-cqdma.c
815
INIT_LIST_HEAD(&cqdma->pc[i]->queue);
drivers/dma/pch_dma.c
100
struct list_head queue;
drivers/dma/pch_dma.c
168
return list_first_entry(&pd_chan->queue,
drivers/dma/pch_dma.c
363
if (!list_empty(&pd_chan->queue))
drivers/dma/pch_dma.c
367
list_splice_init(&pd_chan->queue, &pd_chan->active_list);
drivers/dma/pch_dma.c
380
list_splice_init(&pd_chan->queue, pd_chan->active_list.prev);
drivers/dma/pch_dma.c
414
list_add_tail(&desc->desc_node, &pd_chan->queue);
drivers/dma/pch_dma.c
531
BUG_ON(!list_empty(&pd_chan->queue));
drivers/dma/pch_dma.c
658
list_splice_init(&pd_chan->queue, &list);
drivers/dma/pch_dma.c
877
INIT_LIST_HEAD(&pd_chan->queue);
drivers/dma/ste_dma40.c
2586
list_splice_tail_init(&d40c->pending_queue, &d40c->queue);
drivers/dma/ste_dma40.c
2813
INIT_LIST_HEAD(&d40c->queue);
drivers/dma/ste_dma40.c
497
struct list_head queue;
drivers/dma/ste_dma40.c
993
return list_first_entry_or_null(&d40c->queue, struct d40_desc, node);
drivers/dma/timb_dma.c
281
BUG_ON(list_empty(&td_chan->queue));
drivers/dma/timb_dma.c
284
td_desc = list_entry(td_chan->queue.next, struct timb_dma_desc,
drivers/dma/timb_dma.c
314
list_add_tail(&td_desc->desc_node, &td_chan->queue);
drivers/dma/timb_dma.c
443
BUG_ON(!list_empty(&td_chan->queue));
drivers/dma/timb_dma.c
483
if (list_empty(&td_chan->active_list) && !list_empty(&td_chan->queue))
drivers/dma/timb_dma.c
555
list_for_each_entry_safe(td_desc, _td_desc, &td_chan->queue,
drivers/dma/timb_dma.c
585
if (!list_empty(&td_chan->queue))
drivers/dma/timb_dma.c
698
INIT_LIST_HEAD(&td_chan->queue);
drivers/dma/timb_dma.c
79
struct list_head queue;
drivers/dma/txx9dmac.c
1055
BUG_ON(!list_empty(&dc->queue));
drivers/dma/txx9dmac.c
1136
INIT_LIST_HEAD(&dc->queue);
drivers/dma/txx9dmac.c
177
return list_entry(dc->queue.next, struct txx9dmac_desc, desc_node);
drivers/dma/txx9dmac.c
445
} while (!list_empty(&dc->queue));
drivers/dma/txx9dmac.c
458
if (!list_empty(&dc->queue)) {
drivers/dma/txx9dmac.c
521
if (list_empty(&dc->active_list) && !list_empty(&dc->queue))
drivers/dma/txx9dmac.c
598
if (!list_empty(&dc->queue)) {
drivers/dma/txx9dmac.c
702
list_add_tail(&desc->desc_node, &dc->queue);
drivers/dma/txx9dmac.c
909
list_splice_init(&dc->queue, &list);
drivers/dma/txx9dmac.c
968
if (!list_empty(&dc->queue)) {
drivers/dma/txx9dmac.h
173
struct list_head queue;
drivers/gpib/common/gpib_os.c
1906
unsigned int num_gpib_events(const struct gpib_event_queue *queue)
drivers/gpib/common/gpib_os.c
1908
return queue->num_events;
drivers/gpib/common/gpib_os.c
1913
struct gpib_event_queue *queue = &board->event_queue;
drivers/gpib/common/gpib_os.c
1914
struct list_head *head = &queue->event_head;
drivers/gpib/common/gpib_os.c
1919
if (num_gpib_events(queue) >= max_num_events) {
drivers/gpib/common/gpib_os.c
1922
queue->dropped_event = 1;
drivers/gpib/common/gpib_os.c
1923
retval = pop_gpib_event_nolock(board, queue, &lost_event);
drivers/gpib/common/gpib_os.c
1930
queue->dropped_event = 1;
drivers/gpib/common/gpib_os.c
1940
queue->num_events++;
drivers/gpib/common/gpib_os.c
1943
(int)event_type, num_gpib_events(queue));
drivers/gpib/common/gpib_os.c
1968
struct gpib_event_queue *queue, short *event_type)
drivers/gpib/common/gpib_os.c
1970
struct list_head *head = &queue->event_head;
drivers/gpib/common/gpib_os.c
1974
if (num_gpib_events(queue) == 0) {
drivers/gpib/common/gpib_os.c
1982
if (queue->dropped_event) {
drivers/gpib/common/gpib_os.c
1983
queue->dropped_event = 0;
drivers/gpib/common/gpib_os.c
1993
queue->num_events--;
drivers/gpib/common/gpib_os.c
1996
(int)*event_type, num_gpib_events(queue));
drivers/gpib/common/gpib_os.c
2002
int pop_gpib_event(struct gpib_board *board, struct gpib_event_queue *queue, short *event_type)
drivers/gpib/common/gpib_os.c
2007
spin_lock_irqsave(&queue->lock, flags);
drivers/gpib/common/gpib_os.c
2008
retval = pop_gpib_event_nolock(board, queue, event_type);
drivers/gpib/common/gpib_os.c
2009
spin_unlock_irqrestore(&queue->lock, flags);
drivers/gpib/common/gpib_os.c
78
struct gpib_event_queue *queue, short *event_type);
drivers/gpib/include/gpibP.h
29
unsigned int num_gpib_events(const struct gpib_event_queue *queue);
drivers/gpib/include/gpibP.h
31
int pop_gpib_event(struct gpib_board *board, struct gpib_event_queue *queue, short *event_type);
drivers/gpib/include/gpib_types.h
206
static inline void init_event_queue(struct gpib_event_queue *queue)
drivers/gpib/include/gpib_types.h
208
INIT_LIST_HEAD(&queue->event_head);
drivers/gpib/include/gpib_types.h
209
queue->num_events = 0;
drivers/gpib/include/gpib_types.h
210
queue->dropped_event = 0;
drivers/gpib/include/gpib_types.h
211
spin_lock_init(&queue->lock);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gc_9_4_3.c
513
int engine, int queue)
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gc_9_4_3.c
515
uint32_t reg_offset = get_sdma_rlc_reg_offset(adev, engine, queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gfx_v10.c
1088
int engine, int queue)
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gfx_v10.c
45
uint32_t queue, uint32_t vmid)
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gfx_v10.c
48
nv_grbm_select(adev, mec, pipe, queue, vmid);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gfx_v10.h
70
int engine, int queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gfx_v10_3.c
45
uint32_t queue, uint32_t vmid)
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gfx_v10_3.c
48
nv_grbm_select(adev, mec, pipe, queue, vmid);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gfx_v11.c
43
uint32_t queue, uint32_t vmid)
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gfx_v11.c
46
soc21_grbm_select(adev, mec, pipe, queue, vmid);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gfx_v11.c
804
int engine, int queue)
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gfx_v12.c
31
uint32_t queue, uint32_t vmid)
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gfx_v12.c
34
soc24_grbm_select(adev, mec, pipe, queue, vmid);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gfx_v12.c
365
int engine, int queue)
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gfx_v12_1.c
31
uint32_t queue, uint32_t vmid, uint32_t inst)
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gfx_v12_1.c
34
amdgpu_gfx_select_me_pipe_q(adev, mec, pipe, queue, vmid, inst);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gfx_v12_1.c
368
int engine, int queue)
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gfx_v7.c
49
uint32_t queue, uint32_t vmid)
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gfx_v7.c
51
uint32_t value = PIPEID(pipe) | MEID(mec) | VMID(vmid) | QUEUEID(queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gfx_v7.c
565
int engine, int queue)
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gfx_v8.c
43
uint32_t queue, uint32_t vmid)
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gfx_v8.c
45
uint32_t value = PIPEID(pipe) | MEID(mec) | VMID(vmid) | QUEUEID(queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gfx_v8.c
586
int engine, int queue)
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gfx_v9.c
1227
int engine, int queue)
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gfx_v9.c
51
uint32_t queue, uint32_t vmid, uint32_t inst)
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gfx_v9.c
54
soc15_grbm_select(adev, mec, pipe, queue, vmid, GET_INST(GC, inst));
drivers/gpu/drm/amd/amdgpu/amdgpu_amdkfd_gfx_v9.h
116
int engine, int queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
107
queue = (*pos & GENMASK_ULL(53, 44)) >> 44;
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
141
amdgpu_gfx_select_me_pipe_q(adev, me, pipe, queue, vmid, 0);
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
270
rd->id.srbm.queue, rd->id.srbm.vmid, rd->id.xcc_id);
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
353
rd->id.srbm.queue = v1_data.srbm.queue;
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
79
unsigned int instance_bank, sh_bank, se_bank, me, pipe, queue, vmid;
drivers/gpu/drm/amd/amdgpu/amdgpu_debugfs.c
82
instance_bank = sh_bank = se_bank = me = pipe = queue = vmid = 0;
drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.c
174
int queue = ring->queue;
drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.c
181
adev->gfx.num_gfx_rings > 1 && pipe == 1 && queue == 0) {
drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.c
185
bit = amdgpu_gfx_me_queue_to_bit(adev, me, pipe, queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.c
208
int i, j, queue, pipe;
drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.c
221
queue = (i / adev->gfx.mec.num_pipe_per_mec) %
drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.c
224
set_bit(pipe * adev->gfx.mec.num_queue_per_pipe + queue,
drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.c
244
int i, queue, pipe;
drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.c
254
queue = (i / adev->gfx.me.num_pipe_per_me) %
drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.c
257
set_bit(pipe * num_queue_per_pipe + queue,
drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.c
275
int mec, pipe, queue;
drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.c
285
amdgpu_queue_mask_bit_to_mec_queue(adev, queue_bit, &mec, &pipe, &queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.c
292
if ((mec == 1 && pipe > 1) || queue != 0)
drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.c
297
ring->queue = queue;
drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.c
333
(unsigned char)ring->pipe, (unsigned char)ring->queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.c
49
int pipe, int queue)
drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.c
56
bit += queue;
drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.c
613
int mec, pipe, queue;
drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.c
616
amdgpu_queue_mask_bit_to_mec_queue(adev, queue_bit, &mec, &pipe, &queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.c
618
set_resource_bit = mec * 4 * 8 + pipe * 8 + queue;
drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.c
62
int *mec, int *pipe, int *queue)
drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.c
64
*queue = bit % adev->gfx.mec.num_queue_per_pipe;
drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.c
693
kiq_ring->pipe, kiq_ring->queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.c
73
int xcc_id, int mec, int pipe, int queue)
drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.c
75
return test_bit(amdgpu_gfx_mec_queue_to_bit(adev, mec, pipe, queue),
drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.c
80
int me, int pipe, int queue)
drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.c
88
bit += queue;
drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.c
94
int me, int pipe, int queue)
drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.c
96
return test_bit(amdgpu_gfx_me_queue_to_bit(adev, me, pipe, queue),
drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.h
349
u32 queue, u32 vmid, u32 xcc_id);
drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.h
595
int pipe, int queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.h
597
int *mec, int *pipe, int *queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.h
599
int mec, int pipe, int queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.h
605
int pipe, int queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_mes.c
344
queue_input.queue_id = ring->queue;
drivers/gpu/drm/amd/amdgpu/amdgpu_mes.c
370
queue_input.queue_id = ring->queue;
drivers/gpu/drm/amd/amdgpu/amdgpu_mes.c
399
queue_input.queue_id = ring->queue;
drivers/gpu/drm/amd/amdgpu/amdgpu_ring.h
380
u32 queue;
drivers/gpu/drm/amd/amdgpu/amdgpu_umr.h
34
__u32 me, pipe, queue, vmid;
drivers/gpu/drm/amd/amdgpu/amdgpu_umr.h
44
__u32 me, pipe, queue, vmid;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1013
struct amdgpu_usermode_queue *queue;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1031
queue = __xa_erase(&fpriv->userq_mgr.userq_xa, args->in.queue_id);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1033
if (!queue)
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1036
amdgpu_userq_put(queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1051
struct amdgpu_usermode_queue *queue;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1056
xa_for_each(&uq_mgr->userq_xa, queue_id, queue) {
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1057
queue = amdgpu_userq_get(uq_mgr, queue_id);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1058
if (!queue)
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1061
if (!amdgpu_userq_buffer_vas_mapped(queue)) {
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1064
queue->state = AMDGPU_USERQ_STATE_INVALID_VA;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1065
amdgpu_userq_put(queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1069
r = amdgpu_userq_restore_helper(queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1073
amdgpu_userq_put(queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1300
struct amdgpu_usermode_queue *queue;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1306
xa_for_each(&uq_mgr->userq_xa, queue_id, queue) {
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1307
queue = amdgpu_userq_get(uq_mgr, queue_id);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1308
if (!queue)
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1310
r = amdgpu_userq_preempt_helper(queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1313
amdgpu_userq_put(queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1341
struct amdgpu_usermode_queue *queue;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1345
xa_for_each(&uq_mgr->userq_xa, queue_id, queue) {
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1346
queue = amdgpu_userq_get(uq_mgr, queue_id);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1347
if (!queue)
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1350
struct dma_fence *f = queue->last_fence;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1353
amdgpu_userq_put(queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1360
amdgpu_userq_put(queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1363
amdgpu_userq_put(queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1413
struct amdgpu_usermode_queue *queue;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1418
queue = xa_find(&userq_mgr->userq_xa, &queue_id, ULONG_MAX,
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1420
if (queue)
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1424
if (!queue)
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1427
amdgpu_userq_put(queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1437
struct amdgpu_usermode_queue *queue;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1445
xa_for_each(&adev->userq_doorbell_xa, queue_id, queue) {
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1446
uqm = queue->userq_mgr;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1451
r = amdgpu_userq_preempt_helper(queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1453
r = amdgpu_userq_unmap_helper(queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1463
struct amdgpu_usermode_queue *queue;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1471
xa_for_each(&adev->userq_doorbell_xa, queue_id, queue) {
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1472
uqm = queue->userq_mgr;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1475
r = amdgpu_userq_restore_helper(queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1477
r = amdgpu_userq_map_helper(queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1489
struct amdgpu_usermode_queue *queue;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1501
xa_for_each(&adev->userq_doorbell_xa, queue_id, queue) {
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1502
uqm = queue->userq_mgr;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1505
if (((queue->queue_type == AMDGPU_HW_IP_GFX) ||
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1506
(queue->queue_type == AMDGPU_HW_IP_COMPUTE)) &&
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1507
(queue->xcp_id == idx)) {
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1509
r = amdgpu_userq_preempt_helper(queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1523
struct amdgpu_usermode_queue *queue;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
153
struct amdgpu_usermode_queue *queue = container_of(work,
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1535
xa_for_each(&adev->userq_doorbell_xa, queue_id, queue) {
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1536
uqm = queue->userq_mgr;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1538
if (((queue->queue_type == AMDGPU_HW_IP_GFX) ||
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1539
(queue->queue_type == AMDGPU_HW_IP_COMPUTE)) &&
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1540
(queue->xcp_id == idx)) {
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1541
r = amdgpu_userq_restore_helper(queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1585
struct amdgpu_usermode_queue *queue;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1589
xa_for_each(&adev->userq_doorbell_xa, queue_id, queue) {
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
159
if (!queue || !queue->userq_mgr)
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1590
uqm = queue->userq_mgr;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1592
if (queue->state == AMDGPU_USERQ_STATE_MAPPED) {
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1593
amdgpu_userq_wait_for_last_fence(queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1594
userq_funcs = adev->userq_funcs[queue->queue_type];
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1595
userq_funcs->unmap(queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1600
queue->state = AMDGPU_USERQ_STATE_HUNG;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1601
amdgpu_userq_fence_driver_force_completion(queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1612
struct amdgpu_usermode_queue *queue;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1617
xa_for_each(&adev->userq_doorbell_xa, queue_id, queue) {
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1618
if (queue->state == AMDGPU_USERQ_STATE_HUNG && !vram_lost) {
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1619
userq_funcs = adev->userq_funcs[queue->queue_type];
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
162
uq_mgr = queue->userq_mgr;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1621
r = userq_funcs->map(queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
1626
queue->state = AMDGPU_USERQ_STATE_MAPPED;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
163
fence = READ_ONCE(queue->hang_detect_fence);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
177
void amdgpu_userq_start_hang_detect_work(struct amdgpu_usermode_queue *queue)
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
182
if (!queue || !queue->userq_mgr || !queue->userq_mgr->adev)
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
185
adev = queue->userq_mgr->adev;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
187
switch (queue->queue_type) {
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
203
WRITE_ONCE(queue->hang_detect_fence, queue->last_fence);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
204
schedule_delayed_work(&queue->hang_detect_work,
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
208
static void amdgpu_userq_init_hang_detect_work(struct amdgpu_usermode_queue *queue)
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
210
INIT_DELAYED_WORK(&queue->hang_detect_work, amdgpu_userq_hang_detect_work);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
211
queue->hang_detect_fence = NULL;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
214
static int amdgpu_userq_buffer_va_list_add(struct amdgpu_usermode_queue *queue,
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
227
list_add(&va_cursor->list, &queue->userq_va_list);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
233
struct amdgpu_usermode_queue *queue,
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
237
struct amdgpu_vm *vm = queue->vm;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
257
amdgpu_userq_buffer_va_list_add(queue, va_map, user_addr);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
286
static bool amdgpu_userq_buffer_vas_mapped(struct amdgpu_usermode_queue *queue)
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
291
list_for_each_entry_safe(va_cursor, tmp, &queue->userq_va_list, list) {
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
292
r += amdgpu_userq_buffer_va_mapped(queue->vm, va_cursor->gpu_addr);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
293
dev_dbg(queue->userq_mgr->adev->dev,
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
295
queue, va_cursor->gpu_addr, r);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
313
struct amdgpu_usermode_queue *queue)
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
319
r = amdgpu_bo_reserve(queue->vm->root.bo, false);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
323
list_for_each_entry_safe(va_cursor, tmp, &queue->userq_va_list, list) {
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
324
mapping = amdgpu_vm_bo_lookup_mapping(queue->vm, va_cursor->gpu_addr);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
330
queue, va_cursor->gpu_addr);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
334
amdgpu_bo_unreserve(queue->vm->root.bo);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
338
static int amdgpu_userq_preempt_helper(struct amdgpu_usermode_queue *queue)
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
340
struct amdgpu_userq_mgr *uq_mgr = queue->userq_mgr;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
343
adev->userq_funcs[queue->queue_type];
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
347
if (queue->state == AMDGPU_USERQ_STATE_MAPPED) {
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
348
r = userq_funcs->preempt(queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
350
queue->state = AMDGPU_USERQ_STATE_HUNG;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
353
queue->state = AMDGPU_USERQ_STATE_PREEMPTED;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
363
static int amdgpu_userq_restore_helper(struct amdgpu_usermode_queue *queue)
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
365
struct amdgpu_userq_mgr *uq_mgr = queue->userq_mgr;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
368
adev->userq_funcs[queue->queue_type];
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
371
if (queue->state == AMDGPU_USERQ_STATE_PREEMPTED) {
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
372
r = userq_funcs->restore(queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
374
queue->state = AMDGPU_USERQ_STATE_HUNG;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
376
queue->state = AMDGPU_USERQ_STATE_MAPPED;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
383
static int amdgpu_userq_unmap_helper(struct amdgpu_usermode_queue *queue)
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
385
struct amdgpu_userq_mgr *uq_mgr = queue->userq_mgr;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
388
adev->userq_funcs[queue->queue_type];
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
392
if ((queue->state == AMDGPU_USERQ_STATE_MAPPED) ||
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
393
(queue->state == AMDGPU_USERQ_STATE_PREEMPTED)) {
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
394
r = userq_funcs->unmap(queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
396
queue->state = AMDGPU_USERQ_STATE_HUNG;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
399
queue->state = AMDGPU_USERQ_STATE_UNMAPPED;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
409
static int amdgpu_userq_map_helper(struct amdgpu_usermode_queue *queue)
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
411
struct amdgpu_userq_mgr *uq_mgr = queue->userq_mgr;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
414
adev->userq_funcs[queue->queue_type];
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
417
if (queue->state == AMDGPU_USERQ_STATE_UNMAPPED) {
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
418
r = userq_funcs->map(queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
420
queue->state = AMDGPU_USERQ_STATE_HUNG;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
423
queue->state = AMDGPU_USERQ_STATE_MAPPED;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
430
static int amdgpu_userq_wait_for_last_fence(struct amdgpu_usermode_queue *queue)
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
432
struct amdgpu_userq_mgr *uq_mgr = queue->userq_mgr;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
433
struct dma_fence *f = queue->last_fence;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
441
queue->state = AMDGPU_USERQ_STATE_HUNG;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
449
static void amdgpu_userq_cleanup(struct amdgpu_usermode_queue *queue)
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
451
struct amdgpu_userq_mgr *uq_mgr = queue->userq_mgr;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
453
const struct amdgpu_userq_funcs *uq_funcs = adev->userq_funcs[queue->queue_type];
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
459
amdgpu_userq_buffer_vas_list_cleanup(adev, queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
460
uq_funcs->mqd_destroy(queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
461
amdgpu_userq_fence_driver_free(queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
463
xa_erase_irq(&adev->userq_doorbell_xa, queue->doorbell_index);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
464
queue->userq_mgr = NULL;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
465
list_del(&queue->userq_va_list);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
466
kfree(queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
627
amdgpu_userq_destroy(struct amdgpu_userq_mgr *uq_mgr, struct amdgpu_usermode_queue *queue)
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
634
amdgpu_userq_wait_for_last_fence(queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
636
if (queue->hang_detect_fence) {
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
637
cancel_delayed_work_sync(&queue->hang_detect_work);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
638
queue->hang_detect_fence = NULL;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
640
r = amdgpu_bo_reserve(queue->db_obj.obj, true);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
642
amdgpu_bo_unpin(queue->db_obj.obj);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
643
amdgpu_bo_unreserve(queue->db_obj.obj);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
645
amdgpu_bo_unref(&queue->db_obj.obj);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
647
r = amdgpu_bo_reserve(queue->wptr_obj.obj, true);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
649
amdgpu_bo_unpin(queue->wptr_obj.obj);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
650
amdgpu_bo_unreserve(queue->wptr_obj.obj);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
652
amdgpu_bo_unref(&queue->wptr_obj.obj);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
654
atomic_dec(&uq_mgr->userq_count[queue->queue_type]);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
656
debugfs_remove_recursive(queue->debugfs_queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
659
r = amdgpu_userq_unmap_helper(queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
663
queue->state = AMDGPU_USERQ_STATE_HUNG;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
665
amdgpu_userq_cleanup(queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
676
struct amdgpu_usermode_queue *queue =
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
678
struct amdgpu_userq_mgr *uq_mgr = queue->userq_mgr;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
680
r = amdgpu_userq_destroy(uq_mgr, queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
687
struct amdgpu_usermode_queue *queue;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
690
queue = xa_load(&uq_mgr->userq_xa, qid);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
691
if (queue)
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
692
kref_get(&queue->refcount);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
695
return queue;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
698
void amdgpu_userq_put(struct amdgpu_usermode_queue *queue)
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
700
if (queue)
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
701
kref_put(&queue->refcount, amdgpu_userq_kref_destroy);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
722
struct amdgpu_usermode_queue *queue = m->private;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
726
if (!queue || !queue->mqd.obj)
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
729
bo = amdgpu_bo_ref(queue->mqd.obj);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
736
seq_printf(m, "queue_type: %d\n", queue->queue_type);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
737
seq_printf(m, "mqd_gpu_address: 0x%llx\n", amdgpu_bo_gpu_offset(queue->mqd.obj));
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
766
struct amdgpu_usermode_queue *queue;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
805
queue = kzalloc_obj(struct amdgpu_usermode_queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
806
if (!queue) {
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
812
INIT_LIST_HEAD(&queue->userq_va_list);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
813
queue->doorbell_handle = args->in.doorbell_handle;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
814
queue->queue_type = args->in.ip_type;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
815
queue->vm = &fpriv->vm;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
816
queue->priority = priority;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
818
db_info.queue_type = queue->queue_type;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
819
db_info.doorbell_handle = queue->doorbell_handle;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
820
db_info.db_obj = &queue->db_obj;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
823
queue->userq_mgr = uq_mgr;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
825
if (amdgpu_userq_input_va_validate(adev, queue, args->in.queue_va, args->in.queue_size) ||
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
826
amdgpu_userq_input_va_validate(adev, queue, args->in.rptr_va, AMDGPU_GPU_PAGE_SIZE) ||
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
827
amdgpu_userq_input_va_validate(adev, queue, args->in.wptr_va, AMDGPU_GPU_PAGE_SIZE)) {
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
829
kfree(queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
837
kfree(queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
842
queue->doorbell_index = index;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
843
xa_init_flags(&queue->fence_drv_xa, XA_FLAGS_ALLOC);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
844
r = amdgpu_userq_fence_driver_alloc(adev, queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
850
r = uq_funcs->mqd_create(queue, &args->in);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
853
amdgpu_userq_fence_driver_free(queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
854
kfree(queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
859
kref_init(&queue->refcount);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
863
r = xa_err(xa_store_irq(&adev->userq_doorbell_xa, index, queue, GFP_KERNEL));
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
865
kfree(queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
870
r = xa_alloc(&uq_mgr->userq_xa, &qid, queue,
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
874
amdgpu_userq_fence_driver_free(queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
875
uq_funcs->mqd_destroy(queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
876
kfree(queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
885
((queue->queue_type == AMDGPU_HW_IP_GFX) ||
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
886
(queue->queue_type == AMDGPU_HW_IP_COMPUTE)))
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
891
r = amdgpu_userq_map_helper(queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
895
amdgpu_userq_fence_driver_free(queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
896
uq_funcs->mqd_destroy(queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
897
kfree(queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
910
queue->debugfs_queue = debugfs_create_dir(queue_name, filp->debugfs_client);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
911
debugfs_create_file("mqd_info", 0444, queue->debugfs_queue, queue, &amdgpu_mqd_info_fops);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
913
amdgpu_userq_init_hang_detect_work(queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.c
917
atomic_inc(&uq_mgr->userq_count[queue->queue_type]);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.h
117
void amdgpu_userq_put(struct amdgpu_usermode_queue *queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.h
156
void amdgpu_userq_start_hang_detect_work(struct amdgpu_usermode_queue *queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.h
159
struct amdgpu_usermode_queue *queue,
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.h
83
int (*mqd_create)(struct amdgpu_usermode_queue *queue,
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.h
86
int (*unmap)(struct amdgpu_usermode_queue *queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.h
87
int (*map)(struct amdgpu_usermode_queue *queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.h
88
int (*preempt)(struct amdgpu_usermode_queue *queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq.h
89
int (*restore)(struct amdgpu_usermode_queue *queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq_fence.c
369
struct amdgpu_usermode_queue *queue,
drivers/gpu/drm/amd/amdgpu/amdgpu_userq_fence.c
377
r = amdgpu_bo_reserve(queue->vm->root.bo, false);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq_fence.c
381
addr = queue->userq_prop->wptr_gpu_addr;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq_fence.c
384
mapping = amdgpu_vm_bo_lookup_mapping(queue->vm, addr >> PAGE_SHIFT);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq_fence.c
386
amdgpu_bo_unreserve(queue->vm->root.bo);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq_fence.c
392
amdgpu_bo_unreserve(queue->vm->root.bo);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq_fence.c
469
struct amdgpu_usermode_queue *queue = NULL;
drivers/gpu/drm/amd/amdgpu/amdgpu_userq_fence.c
556
queue = amdgpu_userq_get(userq_mgr, args->queue_id);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq_fence.c
557
if (!queue) {
drivers/gpu/drm/amd/amdgpu/amdgpu_userq_fence.c
562
r = amdgpu_userq_fence_read_wptr(adev, queue, &wptr);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq_fence.c
574
r = amdgpu_userq_fence_create(queue, userq_fence, wptr, &fence);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq_fence.c
581
dma_fence_put(queue->last_fence);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq_fence.c
582
queue->last_fence = dma_fence_get(fence);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq_fence.c
583
amdgpu_userq_start_hang_detect_work(queue);
drivers/gpu/drm/amd/amdgpu/amdgpu_userq_fence.c
651
if (queue)
drivers/gpu/drm/amd/amdgpu/amdgpu_userq_fence.c
652
amdgpu_userq_put(queue);
drivers/gpu/drm/amd/amdgpu/cik.c
945
u32 me, u32 pipe, u32 queue, u32 vmid)
drivers/gpu/drm/amd/amdgpu/cik.c
951
((queue << SRBM_GFX_CNTL__QUEUEID__SHIFT) & SRBM_GFX_CNTL__QUEUEID_MASK));
drivers/gpu/drm/amd/amdgpu/cik.h
30
u32 me, u32 pipe, u32 queue, u32 vmid);
drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c
3752
PACKET3_MAP_QUEUES_QUEUE(ring->queue) |
drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c
4651
int me, int pipe, int queue)
drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c
4661
ring->queue = queue;
drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c
4671
sprintf(ring->name, "gfx_%d.%d.%d", ring->me, ring->pipe, ring->queue);
drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c
4681
int mec, int pipe, int queue)
drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c
4692
ring->queue = queue;
drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c
4700
sprintf(ring->name, "comp_%d.%d.%d", ring->me, ring->pipe, ring->queue);
drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c
6733
tmp |= (ring->me << 5) | (ring->pipe << 3) | (ring->queue);
drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c
6739
tmp |= (ring->me << 5) | (ring->pipe << 3) | (ring->queue);
drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c
6855
nv_grbm_select(adev, ring->me, ring->pipe, ring->queue, 0);
drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c
6872
nv_grbm_select(adev, ring->me, ring->pipe, ring->queue, 0);
drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c
7149
nv_grbm_select(adev, ring->me, ring->pipe, ring->queue, 0);
drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c
7158
nv_grbm_select(adev, ring->me, ring->pipe, ring->queue, 0);
drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c
7180
nv_grbm_select(adev, ring->me, ring->pipe, ring->queue, 0);
drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c
9209
(ring->queue == queue_id))
drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c
9357
ring->queue == queue_id)
drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c
9366
ring->queue == queue_id)
drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c
9530
tmp = REG_SET_FIELD(tmp, CP_VMID_RESET, PIPE0_QUEUES, 1 << ring->queue);
drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c
9532
tmp = REG_SET_FIELD(tmp, CP_VMID_RESET, PIPE1_QUEUES, 1 << ring->queue);
drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c
9602
nv_grbm_select(adev, ring->me, ring->pipe, ring->queue, 0);
drivers/gpu/drm/amd/amdgpu/gfx_v11_0.c
1136
int me, int pipe, int queue)
drivers/gpu/drm/amd/amdgpu/gfx_v11_0.c
1146
ring->queue = queue;
drivers/gpu/drm/amd/amdgpu/gfx_v11_0.c
1160
sprintf(ring->name, "gfx_%d.%d.%d", ring->me, ring->pipe, ring->queue);
drivers/gpu/drm/amd/amdgpu/gfx_v11_0.c
1170
int mec, int pipe, int queue)
drivers/gpu/drm/amd/amdgpu/gfx_v11_0.c
1182
ring->queue = queue;
drivers/gpu/drm/amd/amdgpu/gfx_v11_0.c
1190
sprintf(ring->name, "comp_%d.%d.%d", ring->me, ring->pipe, ring->queue);
drivers/gpu/drm/amd/amdgpu/gfx_v11_0.c
400
PACKET3_MAP_QUEUES_QUEUE(ring->queue) |
drivers/gpu/drm/amd/amdgpu/gfx_v11_0.c
4062
tmp |= (ring->me << 5) | (ring->pipe << 3) | (ring->queue);
drivers/gpu/drm/amd/amdgpu/gfx_v11_0.c
4205
soc21_grbm_select(adev, ring->me, ring->pipe, ring->queue, 0);
drivers/gpu/drm/amd/amdgpu/gfx_v11_0.c
4506
soc21_grbm_select(adev, ring->me, ring->pipe, ring->queue, 0);
drivers/gpu/drm/amd/amdgpu/gfx_v11_0.c
4515
soc21_grbm_select(adev, ring->me, ring->pipe, ring->queue, 0);
drivers/gpu/drm/amd/amdgpu/gfx_v11_0.c
4537
soc21_grbm_select(adev, ring->me, ring->pipe, ring->queue, 0);
drivers/gpu/drm/amd/amdgpu/gfx_v11_0.c
6489
(ring->queue == queue_id))
drivers/gpu/drm/amd/amdgpu/gfx_v11_0.c
6639
ring->queue == queue_id)
drivers/gpu/drm/amd/amdgpu/gfx_v11_0.c
6648
ring->queue == queue_id)
drivers/gpu/drm/amd/amdgpu/gfx_v11_0.c
6782
soc21_grbm_select(adev, ring->me, ring->pipe, ring->queue, 0);
drivers/gpu/drm/amd/amdgpu/gfx_v11_0.c
6874
soc21_grbm_select(adev, ring->me, ring->pipe, ring->queue, 0);
drivers/gpu/drm/amd/amdgpu/gfx_v12_0.c
1011
ring->queue = queue;
drivers/gpu/drm/amd/amdgpu/gfx_v12_0.c
1019
sprintf(ring->name, "comp_%d.%d.%d", ring->me, ring->pipe, ring->queue);
drivers/gpu/drm/amd/amdgpu/gfx_v12_0.c
2950
tmp |= (ring->me << 5) | (ring->pipe << 3) | (ring->queue);
drivers/gpu/drm/amd/amdgpu/gfx_v12_0.c
3076
soc24_grbm_select(adev, ring->me, ring->pipe, ring->queue, 0);
drivers/gpu/drm/amd/amdgpu/gfx_v12_0.c
335
PACKET3_MAP_QUEUES_QUEUE(ring->queue) |
drivers/gpu/drm/amd/amdgpu/gfx_v12_0.c
3377
soc24_grbm_select(adev, ring->me, ring->pipe, ring->queue, 0);
drivers/gpu/drm/amd/amdgpu/gfx_v12_0.c
3386
soc24_grbm_select(adev, ring->me, ring->pipe, ring->queue, 0);
drivers/gpu/drm/amd/amdgpu/gfx_v12_0.c
3408
soc24_grbm_select(adev, ring->me, ring->pipe, ring->queue, 0);
drivers/gpu/drm/amd/amdgpu/gfx_v12_0.c
4855
(ring->queue == queue_id))
drivers/gpu/drm/amd/amdgpu/gfx_v12_0.c
5005
ring->queue == queue_id)
drivers/gpu/drm/amd/amdgpu/gfx_v12_0.c
5014
ring->queue == queue_id)
drivers/gpu/drm/amd/amdgpu/gfx_v12_0.c
5245
soc24_grbm_select(adev, ring->me, ring->pipe, ring->queue, 0);
drivers/gpu/drm/amd/amdgpu/gfx_v12_0.c
5336
soc24_grbm_select(adev, ring->me, ring->pipe, ring->queue, 0);
drivers/gpu/drm/amd/amdgpu/gfx_v12_0.c
968
int me, int pipe, int queue)
drivers/gpu/drm/amd/amdgpu/gfx_v12_0.c
978
ring->queue = queue;
drivers/gpu/drm/amd/amdgpu/gfx_v12_0.c
988
sprintf(ring->name, "gfx_%d.%d.%d", ring->me, ring->pipe, ring->queue);
drivers/gpu/drm/amd/amdgpu/gfx_v12_0.c
999
int mec, int pipe, int queue)
drivers/gpu/drm/amd/amdgpu/gfx_v12_1.c
135
PACKET3_MAP_QUEUES_QUEUE(ring->queue) |
drivers/gpu/drm/amd/amdgpu/gfx_v12_1.c
2090
tmp |= (ring->me << 5) | (ring->pipe << 3) | (ring->queue);
drivers/gpu/drm/amd/amdgpu/gfx_v12_1.c
2367
soc_v1_0_grbm_select(adev, ring->me, ring->pipe, ring->queue, 0, GET_INST(GC, xcc_id));
drivers/gpu/drm/amd/amdgpu/gfx_v12_1.c
2376
soc_v1_0_grbm_select(adev, ring->me, ring->pipe, ring->queue, 0, GET_INST(GC, xcc_id));
drivers/gpu/drm/amd/amdgpu/gfx_v12_1.c
2399
soc_v1_0_grbm_select(adev, ring->me, ring->pipe, ring->queue, 0, GET_INST(GC, xcc_id));
drivers/gpu/drm/amd/amdgpu/gfx_v12_1.c
3651
(ring->queue == queue_id))
drivers/gpu/drm/amd/amdgpu/gfx_v12_1.c
3738
ring->queue == queue_id)
drivers/gpu/drm/amd/amdgpu/gfx_v12_1.c
767
int xcc_id, int mec, int pipe, int queue)
drivers/gpu/drm/amd/amdgpu/gfx_v12_1.c
782
ring->queue = queue;
drivers/gpu/drm/amd/amdgpu/gfx_v12_1.c
794
ring->xcc_id, ring->me, ring->pipe, ring->queue);
drivers/gpu/drm/amd/amdgpu/gfx_v6_0.c
3095
ring->queue = i;
drivers/gpu/drm/amd/amdgpu/gfx_v6_0.c
3096
sprintf(ring->name, "comp_%d.%d.%d", ring->me, ring->pipe, ring->queue);
drivers/gpu/drm/amd/amdgpu/gfx_v7_0.c
2976
cik_srbm_select(adev, ring->me, ring->pipe, ring->queue, 0);
drivers/gpu/drm/amd/amdgpu/gfx_v7_0.c
4308
int mec, int pipe, int queue)
drivers/gpu/drm/amd/amdgpu/gfx_v7_0.c
4317
ring->queue = queue;
drivers/gpu/drm/amd/amdgpu/gfx_v7_0.c
4322
sprintf(ring->name, "comp_%d.%d.%d", ring->me, ring->pipe, ring->queue);
drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c
1847
int mec, int pipe, int queue)
drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c
1859
ring->queue = queue;
drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c
1866
sprintf(ring->name, "comp_%d.%d.%d", ring->me, ring->pipe, ring->queue);
drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c
4300
tmp |= (ring->me << 5) | (ring->pipe << 3) | (ring->queue);
drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c
4351
PACKET3_MAP_QUEUES_QUEUE(ring->queue) |
drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c
4599
vi_srbm_select(adev, ring->me, ring->pipe, ring->queue, 0);
drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c
4610
vi_srbm_select(adev, ring->me, ring->pipe, ring->queue, 0);
drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c
4634
vi_srbm_select(adev, ring->me, ring->pipe, ring->queue, 0);
drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c
4981
vi_srbm_select(adev, ring->me, ring->pipe, ring->queue, 0);
drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c
5076
vi_srbm_select(adev, ring->me, ring->pipe, ring->queue, 0);
drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c
6599
if ((ring->me == me_id) && (ring->pipe == pipe_id) && (ring->queue == queue_id))
drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c
6627
ring->queue == queue_id)
drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
2155
int mec, int pipe, int queue)
drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
2166
ring->queue = queue;
drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
2174
sprintf(ring->name, "comp_%d.%d.%d", ring->me, ring->pipe, ring->queue);
drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
3532
tmp |= (ring->me << 5) | (ring->pipe << 3) | (ring->queue);
drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
3858
soc15_grbm_select(adev, ring->me, ring->pipe, ring->queue, 0, 0);
drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
3869
soc15_grbm_select(adev, ring->me, ring->pipe, ring->queue, 0, 0);
drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
3900
soc15_grbm_select(adev, ring->me, ring->pipe, ring->queue, 0, 0);
drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
4085
adev->gfx.kiq[0].ring.queue, 0, 0);
drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
6248
if ((ring->me == me_id) && (ring->pipe == pipe_id) && (ring->queue == queue_id))
drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
6276
ring->queue == queue_id)
drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
7297
soc15_grbm_select(adev, ring->me, ring->pipe, ring->queue, 0, 0);
drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c
963
PACKET3_MAP_QUEUES_QUEUE(ring->queue) |
drivers/gpu/drm/amd/amdgpu/gfx_v9_4_3.c
1809
tmp |= (ring->me << 5) | (ring->pipe << 3) | (ring->queue);
drivers/gpu/drm/amd/amdgpu/gfx_v9_4_3.c
212
PACKET3_MAP_QUEUES_QUEUE(ring->queue) |
drivers/gpu/drm/amd/amdgpu/gfx_v9_4_3.c
2134
soc15_grbm_select(adev, ring->me, ring->pipe, ring->queue, 0, GET_INST(GC, xcc_id));
drivers/gpu/drm/amd/amdgpu/gfx_v9_4_3.c
2145
soc15_grbm_select(adev, ring->me, ring->pipe, ring->queue, 0, GET_INST(GC, xcc_id));
drivers/gpu/drm/amd/amdgpu/gfx_v9_4_3.c
2177
soc15_grbm_select(adev, ring->me, ring->pipe, ring->queue, 0, GET_INST(GC, xcc_id));
drivers/gpu/drm/amd/amdgpu/gfx_v9_4_3.c
2206
ring->queue, 0, GET_INST(GC, xcc_id));
drivers/gpu/drm/amd/amdgpu/gfx_v9_4_3.c
2337
adev->gfx.kiq[xcc_id].ring.queue, 0,
drivers/gpu/drm/amd/amdgpu/gfx_v9_4_3.c
3312
if ((ring->me == me_id) && (ring->pipe == pipe_id) && (ring->queue == queue_id))
drivers/gpu/drm/amd/amdgpu/gfx_v9_4_3.c
3345
ring->queue == queue_id)
drivers/gpu/drm/amd/amdgpu/gfx_v9_4_3.c
3457
uint32_t pipe, uint32_t queue,
drivers/gpu/drm/amd/amdgpu/gfx_v9_4_3.c
3464
soc15_grbm_select(adev, me, pipe, queue, 0, GET_INST(GC, xcc_id));
drivers/gpu/drm/amd/amdgpu/gfx_v9_4_3.c
3542
r = gfx_v9_4_3_unmap_done(adev, ring->me, ring->pipe, ring->queue, ring->xcc_id);
drivers/gpu/drm/amd/amdgpu/gfx_v9_4_3.c
3582
r = gfx_v9_4_3_unmap_done(adev, ring->me, ring->pipe, ring->queue, ring->xcc_id);
drivers/gpu/drm/amd/amdgpu/gfx_v9_4_3.c
972
int xcc_id, int mec, int pipe, int queue)
drivers/gpu/drm/amd/amdgpu/gfx_v9_4_3.c
986
ring->queue = queue;
drivers/gpu/drm/amd/amdgpu/gfx_v9_4_3.c
998
ring->xcc_id, ring->me, ring->pipe, ring->queue);
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
110
queue->wptr_obj.gpu_addr = amdgpu_bo_gpu_offset(wptr_obj->obj);
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
136
static int mes_userq_map(struct amdgpu_usermode_queue *queue)
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
138
struct amdgpu_userq_mgr *uq_mgr = queue->userq_mgr;
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
140
struct amdgpu_userq_obj *ctx = &queue->fw_obj;
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
141
struct amdgpu_mqd_prop *userq_props = queue->userq_prop;
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
158
queue_input.gang_global_priority_level = convert_to_mes_priority(queue->priority);
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
160
queue_input.process_id = queue->vm->pasid;
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
161
queue_input.queue_type = queue->queue_type;
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
162
queue_input.mqd_addr = queue->mqd.gpu_addr;
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
166
queue_input.page_table_base_addr = amdgpu_gmc_pd_addr(queue->vm->root.bo);
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
167
queue_input.wptr_mc_addr = queue->wptr_obj.gpu_addr;
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
181
static int mes_userq_unmap(struct amdgpu_usermode_queue *queue)
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
183
struct amdgpu_userq_mgr *uq_mgr = queue->userq_mgr;
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
186
struct amdgpu_userq_obj *ctx = &queue->fw_obj;
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
190
queue_input.doorbell_offset = queue->doorbell_index;
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
202
struct amdgpu_usermode_queue *queue,
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
205
struct amdgpu_userq_obj *ctx = &queue->fw_obj;
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
228
struct amdgpu_usermode_queue *queue;
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
252
xa_for_each(&adev->userq_doorbell_xa, queue_id, queue) {
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
253
if (queue->queue_type == queue_type) {
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
255
if (queue->doorbell_index == db_array[i]) {
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
256
queue->state = AMDGPU_USERQ_STATE_HUNG;
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
259
amdgpu_userq_fence_driver_force_completion(queue);
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
275
static int mes_userq_mqd_create(struct amdgpu_usermode_queue *queue,
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
278
struct amdgpu_userq_mgr *uq_mgr = queue->userq_mgr;
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
280
struct amdgpu_mqd *mqd_hw_default = &adev->mqds[queue->queue_type];
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
292
r = amdgpu_userq_create_object(uq_mgr, &queue->mqd,
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
304
userq_props->mqd_gpu_addr = queue->mqd.gpu_addr;
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
306
userq_props->doorbell_index = queue->doorbell_index;
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
307
userq_props->fence_address = queue->fence_drv->gpu_addr;
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
309
if (queue->queue_type == AMDGPU_HW_IP_COMPUTE) {
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
325
r = amdgpu_userq_input_va_validate(adev, queue, compute_mqd->eop_va,
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
339
} else if (queue->queue_type == AMDGPU_HW_IP_GFX) {
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
368
r = amdgpu_userq_input_va_validate(adev, queue, mqd_gfx_v11->shadow_va,
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
374
r = amdgpu_userq_input_va_validate(adev, queue, mqd_gfx_v11->csa_va,
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
382
} else if (queue->queue_type == AMDGPU_HW_IP_DMA) {
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
397
r = amdgpu_userq_input_va_validate(adev, queue, mqd_sdma_v11->csa_va,
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
408
queue->userq_prop = userq_props;
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
410
r = mqd_hw_default->init_mqd(adev, (void *)queue->mqd.cpu_ptr, userq_props);
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
417
r = mes_userq_create_ctx_space(uq_mgr, queue, mqd_user);
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
424
r = mes_userq_create_wptr_mapping(adev, uq_mgr, queue, userq_props->wptr_gpu_addr);
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
433
amdgpu_userq_destroy_object(uq_mgr, &queue->fw_obj);
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
436
amdgpu_userq_destroy_object(uq_mgr, &queue->mqd);
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
444
static void mes_userq_mqd_destroy(struct amdgpu_usermode_queue *queue)
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
446
struct amdgpu_userq_mgr *uq_mgr = queue->userq_mgr;
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
448
amdgpu_userq_destroy_object(uq_mgr, &queue->fw_obj);
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
449
kfree(queue->userq_prop);
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
450
amdgpu_userq_destroy_object(uq_mgr, &queue->mqd);
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
453
static int mes_userq_preempt(struct amdgpu_usermode_queue *queue)
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
455
struct amdgpu_userq_mgr *uq_mgr = queue->userq_mgr;
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
458
struct amdgpu_userq_obj *ctx = &queue->fw_obj;
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
465
if (queue->state != AMDGPU_USERQ_STATE_MAPPED)
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
499
static int mes_userq_restore(struct amdgpu_usermode_queue *queue)
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
501
struct amdgpu_userq_mgr *uq_mgr = queue->userq_mgr;
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
504
struct amdgpu_userq_obj *ctx = &queue->fw_obj;
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
507
if (queue->state == AMDGPU_USERQ_STATE_HUNG)
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
509
if (queue->state != AMDGPU_USERQ_STATE_PREEMPTED)
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
64
struct amdgpu_usermode_queue *queue,
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
69
struct amdgpu_userq_obj *wptr_obj = &queue->wptr_obj;
drivers/gpu/drm/amd/amdgpu/mes_userqueue.c
72
wptr_vm = queue->vm;
drivers/gpu/drm/amd/amdgpu/mes_v11_0.c
1341
ring->queue = 0;
drivers/gpu/drm/amd/amdgpu/mes_v11_0.c
1348
sprintf(ring->name, "mes_%d.%d.%d", ring->me, ring->pipe, ring->queue);
drivers/gpu/drm/amd/amdgpu/mes_v11_0.c
1364
ring->queue = 0;
drivers/gpu/drm/amd/amdgpu/mes_v11_0.c
1373
ring->me, ring->pipe, ring->queue);
drivers/gpu/drm/amd/amdgpu/mes_v11_0.c
1556
tmp |= (ring->me << 5) | (ring->pipe << 3) | (ring->queue);
drivers/gpu/drm/amd/amdgpu/mes_v12_0.c
1527
ring->queue = 0;
drivers/gpu/drm/amd/amdgpu/mes_v12_0.c
1533
sprintf(ring->name, "mes_%d.%d.%d", ring->me, ring->pipe, ring->queue);
drivers/gpu/drm/amd/amdgpu/mes_v12_0.c
1554
ring->queue = 0;
drivers/gpu/drm/amd/amdgpu/mes_v12_0.c
1563
ring->me, ring->pipe, ring->queue);
drivers/gpu/drm/amd/amdgpu/mes_v12_0.c
1736
tmp |= (ring->me << 5) | (ring->pipe << 3) | (ring->queue);
drivers/gpu/drm/amd/amdgpu/mes_v12_1.c
1453
ring->queue = 0;
drivers/gpu/drm/amd/amdgpu/mes_v12_1.c
1463
(unsigned char)ring->pipe, (unsigned char)ring->queue);
drivers/gpu/drm/amd/amdgpu/mes_v12_1.c
1491
ring->queue = 0;
drivers/gpu/drm/amd/amdgpu/mes_v12_1.c
1507
(unsigned char)ring->pipe, (unsigned char)ring->queue);
drivers/gpu/drm/amd/amdgpu/mes_v12_1.c
1691
tmp |= (ring->me << 5) | (ring->pipe << 3) | (ring->queue);
drivers/gpu/drm/amd/amdgpu/nv.c
318
u32 me, u32 pipe, u32 queue, u32 vmid)
drivers/gpu/drm/amd/amdgpu/nv.c
324
grbm_gfx_cntl = REG_SET_FIELD(grbm_gfx_cntl, GRBM_GFX_CNTL, QUEUEID, queue);
drivers/gpu/drm/amd/amdgpu/nv.h
32
u32 me, u32 pipe, u32 queue, u32 vmid);
drivers/gpu/drm/amd/amdgpu/sdma_v6_0.c
1627
int instances, queue;
drivers/gpu/drm/amd/amdgpu/sdma_v6_0.c
1631
queue = entry->ring_id & 0xf;
drivers/gpu/drm/amd/amdgpu/sdma_v6_0.c
1640
switch (queue) {
drivers/gpu/drm/amd/amdgpu/sdma_v7_0.c
1559
int instances, queue;
drivers/gpu/drm/amd/amdgpu/sdma_v7_0.c
1563
queue = entry->ring_id & 0xf;
drivers/gpu/drm/amd/amdgpu/sdma_v7_0.c
1572
switch (queue) {
drivers/gpu/drm/amd/amdgpu/sdma_v7_1.c
1496
int inst, instances, queue, xcc_id = 0;
drivers/gpu/drm/amd/amdgpu/sdma_v7_1.c
1505
queue = entry->ring_id & 0xf;
drivers/gpu/drm/amd/amdgpu/sdma_v7_1.c
1523
switch (queue) {
drivers/gpu/drm/amd/amdgpu/si.h
30
u32 me, u32 pipe, u32 queue, u32 vmid);
drivers/gpu/drm/amd/amdgpu/soc15.c
364
u32 me, u32 pipe, u32 queue, u32 vmid, int xcc_id)
drivers/gpu/drm/amd/amdgpu/soc15.c
370
grbm_gfx_cntl = REG_SET_FIELD(grbm_gfx_cntl, GRBM_GFX_CNTL, QUEUEID, queue);
drivers/gpu/drm/amd/amdgpu/soc15.h
110
u32 me, u32 pipe, u32 queue, u32 vmid, int xcc_id);
drivers/gpu/drm/amd/amdgpu/soc21.c
270
u32 me, u32 pipe, u32 queue, u32 vmid)
drivers/gpu/drm/amd/amdgpu/soc21.c
276
grbm_gfx_cntl = REG_SET_FIELD(grbm_gfx_cntl, GRBM_GFX_CNTL, QUEUEID, queue);
drivers/gpu/drm/amd/amdgpu/soc21.h
29
u32 me, u32 pipe, u32 queue, u32 vmid);
drivers/gpu/drm/amd/amdgpu/soc24.c
102
u32 me, u32 pipe, u32 queue, u32 vmid)
drivers/gpu/drm/amd/amdgpu/soc24.c
108
grbm_gfx_cntl = REG_SET_FIELD(grbm_gfx_cntl, GRBM_GFX_CNTL, QUEUEID, queue);
drivers/gpu/drm/amd/amdgpu/soc24.h
29
u32 me, u32 pipe, u32 queue, u32 vmid);
drivers/gpu/drm/amd/amdgpu/soc_v1_0.c
118
u32 queue, u32 vmid,
drivers/gpu/drm/amd/amdgpu/soc_v1_0.c
125
grbm_gfx_cntl = REG_SET_FIELD(grbm_gfx_cntl, GRBM_GFX_CNTL, QUEUEID, queue);
drivers/gpu/drm/amd/amdgpu/soc_v1_0.h
30
u32 queue, u32 vmid,
drivers/gpu/drm/amd/amdgpu/vi.c
579
u32 me, u32 pipe, u32 queue, u32 vmid)
drivers/gpu/drm/amd/amdgpu/vi.c
585
srbm_gfx_cntl = REG_SET_FIELD(srbm_gfx_cntl, SRBM_GFX_CNTL, QUEUEID, queue);
drivers/gpu/drm/amd/amdgpu/vi.h
30
u32 me, u32 pipe, u32 queue, u32 vmid);
drivers/gpu/drm/amd/amdkfd/kfd_chardev.c
1483
struct queue *q;
drivers/gpu/drm/amd/amdkfd/kfd_debug.c
299
static int kfd_dbg_set_queue_workaround(struct queue *q, bool enable)
drivers/gpu/drm/amd/amdkfd/kfd_debug.c
966
struct queue *queue = NULL;
drivers/gpu/drm/amd/amdkfd/kfd_debug.c
973
list_for_each_entry(queue, &qpd->queues_list, list) {
drivers/gpu/drm/amd/amdkfd/kfd_debug.c
974
if (!found && queue->properties.queue_id == source_id) {
drivers/gpu/drm/amd/amdkfd/kfd_debug.c
988
if (!(queue->properties.exception_status & KFD_EC_MASK(exception_code))) {
drivers/gpu/drm/amd/amdkfd/kfd_debug.c
992
exception_status_ptr = &queue->properties.exception_status;
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
1045
q->pipe, q->queue,
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
1064
struct queue *q)
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
1110
struct queue *q)
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
1143
struct queue *q;
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
1177
KFD_UNMAP_LATENCY_MS, q->pipe, q->queue);
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
1193
struct queue *q;
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
1256
struct queue *q;
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
1316
q->queue, &q->properties, mm);
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
1336
struct queue *q;
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
1502
int pipe, queue;
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
1520
for (queue = 0; queue < get_queues_per_pipe(dqm); queue++)
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
1521
if (test_bit(pipe_offset + queue,
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
1523
dqm->allocated_queues[pipe] |= 1 << queue;
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
1577
struct queue *q, const uint32_t *restore_sdma_id)
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
163
struct queue *q;
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
1684
struct queue *q)
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
1939
increment_queue_count(dqm, qpd, kq->queue);
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
1954
decrement_queue_count(dqm, qpd, kq->queue);
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
1968
static int create_queue_cpsch(struct device_queue_manager *dqm, struct queue *q,
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
201
static int add_queue_mes(struct device_queue_manager *dqm, struct queue *q,
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
2131
static void set_queue_as_reset(struct device_queue_manager *dqm, struct queue *q,
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
2157
uint32_t mec, pipe, queue;
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
2166
amdgpu_queue_mask_bit_to_mec_queue(dqm->dev->adev, i, &mec, &pipe, &queue);
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
2170
dqm->dev->adev, pipe, queue, xcc_id);
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
2177
hang_info.queue_id = queue;
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
2189
static struct queue *find_queue_by_address(struct device_queue_manager *dqm, uint64_t queue_address)
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
2193
struct queue *q;
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
2219
struct queue *q = find_queue_by_address(dqm, hang_info.queue_address);
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
2283
struct queue *q;
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
2424
if (mqd_mgr->check_preemption_failed(mqd_mgr, dqm->packet_mgr.priv_queue->queue->mqd) &&
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
2484
struct queue *q)
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
2513
struct queue *q)
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
2648
struct queue *q;
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
2660
q = list_first_entry(&qpd->queues_list, struct queue, list);
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
2694
struct queue *q,
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
2724
const struct queue *q,
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
2744
const struct queue *q,
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
276
static int remove_queue_mes(struct device_queue_manager *dqm, struct queue *q,
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
2777
struct queue *q;
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
2791
decrement_queue_count(dqm, qpd, kq->queue);
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
2844
q = list_first_entry(&qpd->queues_list, struct queue, list);
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
3067
struct queue *q = NULL;
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
313
struct queue *q;
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
3251
struct queue *q;
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
3330
struct queue *q;
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
340
struct queue *q;
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
3433
struct queue *q;
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
3545
void set_queue_snapshot_entry(struct queue *q,
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
3620
struct queue *q;
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
3668
int pipe, queue;
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
3698
for (queue = 0; queue < get_queues_per_pipe(dqm); queue++) {
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
3699
if (!test_bit(pipe_offset + queue,
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
3704
pipe, queue,
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
3712
xcc_id, pipe, queue);
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
3724
for (queue = 0;
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
3725
queue < dqm->dev->kfd->device_info.num_sdma_queues_per_engine;
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
3726
queue++) {
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
3728
dqm->dev->adev, pipe, queue, &dump, &n_regs);
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
3733
pipe, queue);
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
404
struct queue *q)
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
418
struct queue *q)
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
435
struct queue *q,
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
506
struct queue *q)
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
535
struct queue *q)
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
604
struct queue *q)
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
624
struct queue *q,
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
65
struct queue *q);
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
664
q->pipe, q->queue);
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
68
struct queue *q);
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
69
static int allocate_hqd(struct device_queue_manager *dqm, struct queue *q);
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
706
q->queue, &q->properties, current->mm);
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
71
struct queue *q, const uint32_t *restore_sdma_id);
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
744
static int allocate_hqd(struct device_queue_manager *dqm, struct queue *q)
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
762
q->queue = bit;
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
771
pr_debug("hqd slot - pipe %d, queue %d\n", q->pipe, q->queue);
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
779
struct queue *q)
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
781
dqm->allocated_queues[q->pipe] |= (1 << q->queue);
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
857
struct queue *q)
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
887
q->pipe, q->queue);
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
915
struct queue *q)
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
945
static int update_queue(struct device_queue_manager *dqm, struct queue *q,
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c
998
KFD_UNMAP_LATENCY_MS, q->pipe, q->queue);
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.h
138
struct queue *q,
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.h
146
struct queue *q);
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.h
149
struct queue *q, struct mqd_update_info *minfo);
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.h
188
struct queue *q,
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.h
196
const struct queue *q, u32 *mqd_size,
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.h
200
const struct queue *q,
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.h
216
struct queue *q,
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.h
323
void set_queue_snapshot_entry(struct queue *q,
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager_cik.c
160
struct queue *q,
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager_cik.c
48
struct queue *q,
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager_v10.c
32
static void init_sdma_vm_v10(struct device_queue_manager *dqm, struct queue *q,
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager_v10.c
85
static void init_sdma_vm_v10(struct device_queue_manager *dqm, struct queue *q,
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager_v11.c
31
static void init_sdma_vm_v11(struct device_queue_manager *dqm, struct queue *q,
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager_v11.c
85
static void init_sdma_vm_v11(struct device_queue_manager *dqm, struct queue *q,
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager_v12.c
31
static void init_sdma_vm_v12(struct device_queue_manager *dqm, struct queue *q,
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager_v12.c
85
static void init_sdma_vm_v12(struct device_queue_manager *dqm, struct queue *q,
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager_v12_1.c
31
static void init_sdma_vm_v12_1(struct device_queue_manager *dqm, struct queue *q,
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager_v12_1.c
93
static void init_sdma_vm_v12_1(struct device_queue_manager *dqm, struct queue *q,
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager_v9.c
130
static void init_sdma_vm_v9(struct device_queue_manager *dqm, struct queue *q,
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager_v9.c
31
static void init_sdma_vm_v9(struct device_queue_manager *dqm, struct queue *q,
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager_vi.c
162
struct queue *q,
drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager_vi.c
48
struct queue *q,
drivers/gpu/drm/amd/amdkfd/kfd_kernel_queue.c
128
if (init_queue(&kq->queue, &prop) != 0)
drivers/gpu/drm/amd/amdkfd/kfd_kernel_queue.c
131
kq->queue->device = dev;
drivers/gpu/drm/amd/amdkfd/kfd_kernel_queue.c
133
kq->queue->mqd_mem_obj = kq->mqd_mgr->allocate_mqd(kq->mqd_mgr,
drivers/gpu/drm/amd/amdkfd/kfd_kernel_queue.c
134
&kq->queue->properties);
drivers/gpu/drm/amd/amdkfd/kfd_kernel_queue.c
135
if (!kq->queue->mqd_mem_obj)
drivers/gpu/drm/amd/amdkfd/kfd_kernel_queue.c
137
kq->mqd_mgr->init_mqd(kq->mqd_mgr, &kq->queue->mqd,
drivers/gpu/drm/amd/amdkfd/kfd_kernel_queue.c
138
kq->queue->mqd_mem_obj,
drivers/gpu/drm/amd/amdkfd/kfd_kernel_queue.c
139
&kq->queue->gart_mqd_addr,
drivers/gpu/drm/amd/amdkfd/kfd_kernel_queue.c
140
&kq->queue->properties);
drivers/gpu/drm/amd/amdkfd/kfd_kernel_queue.c
144
kq->queue->pipe = KFD_CIK_HIQ_PIPE;
drivers/gpu/drm/amd/amdkfd/kfd_kernel_queue.c
145
kq->queue->queue = KFD_CIK_HIQ_QUEUE;
drivers/gpu/drm/amd/amdkfd/kfd_kernel_queue.c
146
kq->mqd_mgr->load_mqd(kq->mqd_mgr, kq->queue->mqd,
drivers/gpu/drm/amd/amdkfd/kfd_kernel_queue.c
147
kq->queue->pipe, kq->queue->queue,
drivers/gpu/drm/amd/amdkfd/kfd_kernel_queue.c
148
&kq->queue->properties, NULL);
drivers/gpu/drm/amd/amdkfd/kfd_kernel_queue.c
151
print_queue(kq->queue);
drivers/gpu/drm/amd/amdkfd/kfd_kernel_queue.c
155
uninit_queue(kq->queue);
drivers/gpu/drm/amd/amdkfd/kfd_kernel_queue.c
174
if (kq->queue->properties.type == KFD_QUEUE_TYPE_HIQ && down_read_trylock(&kq->dev->adev->reset_domain->sem)) {
drivers/gpu/drm/amd/amdkfd/kfd_kernel_queue.c
176
kq->queue->mqd,
drivers/gpu/drm/amd/amdkfd/kfd_kernel_queue.c
179
kq->queue->pipe,
drivers/gpu/drm/amd/amdkfd/kfd_kernel_queue.c
180
kq->queue->queue);
drivers/gpu/drm/amd/amdkfd/kfd_kernel_queue.c
184
kq->mqd_mgr->free_mqd(kq->mqd_mgr, kq->queue->mqd,
drivers/gpu/drm/amd/amdkfd/kfd_kernel_queue.c
185
kq->queue->mqd_mem_obj);
drivers/gpu/drm/amd/amdkfd/kfd_kernel_queue.c
197
kq->queue->properties.doorbell_ptr);
drivers/gpu/drm/amd/amdkfd/kfd_kernel_queue.c
198
uninit_queue(kq->queue);
drivers/gpu/drm/amd/amdkfd/kfd_kernel_queue.c
219
queue_size_dwords = kq->queue->properties.queue_size / 4;
drivers/gpu/drm/amd/amdkfd/kfd_kernel_queue.c
284
write_kernel_doorbell64(kq->queue->properties.doorbell_ptr,
drivers/gpu/drm/amd/amdkfd/kfd_kernel_queue.c
289
write_kernel_doorbell(kq->queue->properties.doorbell_ptr,
drivers/gpu/drm/amd/amdkfd/kfd_kernel_queue.c
301
(kq->queue->properties.queue_size / 4);
drivers/gpu/drm/amd/amdkfd/kfd_kernel_queue.h
58
struct queue *queue;
drivers/gpu/drm/amd/amdkfd/kfd_packet_manager.c
148
struct queue *q;
drivers/gpu/drm/amd/amdkfd/kfd_packet_manager.c
204
if (!kq->queue->properties.is_active)
drivers/gpu/drm/amd/amdkfd/kfd_packet_manager.c
209
kq->queue->queue, qpd->is_debug);
drivers/gpu/drm/amd/amdkfd/kfd_packet_manager.c
213
kq->queue,
drivers/gpu/drm/amd/amdkfd/kfd_packet_manager.c
229
q->queue, qpd->is_debug);
drivers/gpu/drm/amd/amdkfd/kfd_packet_manager_v9.c
228
struct queue *q, bool is_static)
drivers/gpu/drm/amd/amdkfd/kfd_packet_manager_vi.c
144
struct queue *q, bool is_static)
drivers/gpu/drm/amd/amdkfd/kfd_pm4_headers_vi.h
252
uint32_t queue:6;
drivers/gpu/drm/amd/amdkfd/kfd_priv.h
1155
int kfd_procfs_add_queue(struct queue *q);
drivers/gpu/drm/amd/amdkfd/kfd_priv.h
1156
void kfd_procfs_del_queue(struct queue *q);
drivers/gpu/drm/amd/amdkfd/kfd_priv.h
1335
int init_queue(struct queue **q, const struct queue_properties *properties);
drivers/gpu/drm/amd/amdkfd/kfd_priv.h
1336
void uninit_queue(struct queue *q);
drivers/gpu/drm/amd/amdkfd/kfd_priv.h
1338
void print_queue(struct queue *q);
drivers/gpu/drm/amd/amdkfd/kfd_priv.h
1373
struct queue *q;
drivers/gpu/drm/amd/amdkfd/kfd_priv.h
1397
struct queue *pqm_get_user_queue(struct process_queue_manager *pqm,
drivers/gpu/drm/amd/amdkfd/kfd_priv.h
1462
struct queue *q, bool is_static);
drivers/gpu/drm/amd/amdkfd/kfd_priv.h
613
uint32_t queue;
drivers/gpu/drm/amd/amdkfd/kfd_process.c
105
struct queue *q;
drivers/gpu/drm/amd/amdkfd/kfd_process.c
2329
struct queue *q;
drivers/gpu/drm/amd/amdkfd/kfd_process.c
406
struct queue *q = container_of(kobj, struct queue, kobj);
drivers/gpu/drm/amd/amdkfd/kfd_process.c
520
int kfd_procfs_add_queue(struct queue *q)
drivers/gpu/drm/amd/amdkfd/kfd_process.c
680
void kfd_procfs_del_queue(struct queue *q)
drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c
1097
struct queue *q;
drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c
246
struct kfd_node *dev, struct queue **q,
drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c
320
struct queue *q;
drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c
40
(pqn->kq && pqn->kq->queue->properties.queue_id == qid))
drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c
581
struct queue *q = pqn->q;
drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c
663
struct queue *pqm_get_user_queue(struct process_queue_manager *pqm,
drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c
735
struct queue *q,
drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c
756
struct queue *q;
drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c
814
struct queue *q,
drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c
867
struct queue *q;
drivers/gpu/drm/amd/amdkfd/kfd_queue.c
48
void print_queue(struct queue *q)
drivers/gpu/drm/amd/amdkfd/kfd_queue.c
69
int init_queue(struct queue **q, const struct queue_properties *properties)
drivers/gpu/drm/amd/amdkfd/kfd_queue.c
71
struct queue *tmp_q;
drivers/gpu/drm/amd/amdkfd/kfd_queue.c
83
void uninit_queue(struct queue *q)
drivers/gpu/drm/amd/include/kgd_kfd_interface.h
335
int engine, int queue);
drivers/gpu/drm/drm_atomic_helper.c
1911
wait_queue_head_t *queue = drm_crtc_vblank_waitqueue(crtc);
drivers/gpu/drm/drm_atomic_helper.c
1916
ret = wait_event_timeout(*queue,
drivers/gpu/drm/drm_vblank.c
1316
ret = wait_event_timeout(vblank->queue,
drivers/gpu/drm/drm_vblank.c
1366
wake_up(&vblank->queue);
drivers/gpu/drm/drm_vblank.c
1841
wait = wait_event_interruptible_timeout(vblank->queue,
drivers/gpu/drm/drm_vblank.c
1949
wake_up(&vblank->queue);
drivers/gpu/drm/drm_vblank.c
558
init_waitqueue_head(&vblank->queue);
drivers/gpu/drm/drm_vblank.c
610
return &vblank->queue;
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
1488
while ((rb = rb_first_cached(&sched_engine->queue))) {
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
1567
rb_erase_cached(&p->node, &sched_engine->queue);
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
289
rb = rb_first_cached(&sched_engine->queue);
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
3179
while ((rb = rb_first_cached(&sched_engine->queue))) {
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
3189
rb_erase_cached(&p->node, &sched_engine->queue);
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
3223
sched_engine->queue = RB_ROOT_CACHED;
drivers/gpu/drm/i915/gt/intel_execlists_submission.c
4113
for (rb = rb_first_cached(&sched_engine->queue); rb; rb = rb_next(rb)) {
drivers/gpu/drm/i915/gt/intel_gt_debugfs.c
38
wait_event(gt->reset.queue,
drivers/gpu/drm/i915/gt/intel_reset.c
1529
wait_event(gt->reset.queue,
drivers/gpu/drm/i915/gt/intel_reset.c
1563
wake_up_all(>->reset.queue);
drivers/gpu/drm/i915/gt/intel_reset.c
1582
if (wait_event_interruptible(gt->reset.queue,
drivers/gpu/drm/i915/gt/intel_reset.c
1622
if (wait_event_interruptible(gt->reset.queue,
drivers/gpu/drm/i915/gt/intel_reset.c
1652
init_waitqueue_head(>->reset.queue);
drivers/gpu/drm/i915/gt/intel_reset_types.h
55
wait_queue_head_t queue;
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
2002
while ((rb = rb_first_cached(&sched_engine->queue))) {
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
2013
rb_erase_cached(&p->node, &sched_engine->queue);
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
2020
sched_engine->queue = RB_ROOT_CACHED;
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
5511
for (rb = rb_first_cached(&sched_engine->queue); rb; rb = rb_next(rb)) {
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
953
while ((rb = rb_first_cached(&sched_engine->queue))) {
drivers/gpu/drm/i915/gt/uc/intel_guc_submission.c
984
rb_erase_cached(&p->node, &sched_engine->queue);
drivers/gpu/drm/i915/i915_scheduler.c
114
rb_insert_color_cached(&p->node, &sched_engine->queue, first);
drivers/gpu/drm/i915/i915_scheduler.c
462
sched_engine->queue = RB_ROOT_CACHED;
drivers/gpu/drm/i915/i915_scheduler.c
47
GEM_BUG_ON(rb_first_cached(&sched_engine->queue) !=
drivers/gpu/drm/i915/i915_scheduler.c
48
rb_first(&sched_engine->queue.rb_root));
drivers/gpu/drm/i915/i915_scheduler.c
51
for (rb = rb_first_cached(&sched_engine->queue); rb; rb = rb_next(rb)) {
drivers/gpu/drm/i915/i915_scheduler.c
75
parent = &sched_engine->queue.rb_root.rb_node;
drivers/gpu/drm/i915/i915_scheduler.h
71
return RB_EMPTY_ROOT(&sched_engine->queue.rb_root);
drivers/gpu/drm/i915/i915_scheduler_types.h
154
struct rb_root_cached queue;
drivers/gpu/drm/i915/selftests/igt_reset.c
22
wait_event(gt->reset.queue,
drivers/gpu/drm/i915/selftests/igt_reset.c
42
wake_up_all(>->reset.queue);
drivers/gpu/drm/imagination/pvr_device.c
135
struct pvr_queue *queue, *tmp_queue;
drivers/gpu/drm/imagination/pvr_device.c
146
list_for_each_entry_safe(queue, tmp_queue, &active_queues, node)
drivers/gpu/drm/imagination/pvr_device.c
147
pvr_queue_process(queue);
drivers/gpu/drm/imagination/pvr_job.c
616
get_last_queued_job_scheduled_fence(struct pvr_queue *queue,
drivers/gpu/drm/imagination/pvr_job.c
626
if (job->ctx == queue->ctx && job->type == queue->type)
drivers/gpu/drm/imagination/pvr_job.c
633
return dma_fence_get(queue->last_queued_job_scheduled_fence);
drivers/gpu/drm/imagination/pvr_queue.c
1001
lockdep_assert_held(&queue->ctx->pvr_dev->queues.lock);
drivers/gpu/drm/imagination/pvr_queue.c
1003
pvr_queue_check_job_waiting_for_cccb_space(queue);
drivers/gpu/drm/imagination/pvr_queue.c
1004
pvr_queue_signal_done_fences(queue);
drivers/gpu/drm/imagination/pvr_queue.c
1005
pvr_queue_update_active_state_locked(queue);
drivers/gpu/drm/imagination/pvr_queue.c
1008
static u32 get_dm_type(struct pvr_queue *queue)
drivers/gpu/drm/imagination/pvr_queue.c
1010
switch (queue->type) {
drivers/gpu/drm/imagination/pvr_queue.c
1032
static void init_fw_context(struct pvr_queue *queue, void *fw_ctx_map)
drivers/gpu/drm/imagination/pvr_queue.c
1034
struct pvr_context *ctx = queue->ctx;
drivers/gpu/drm/imagination/pvr_queue.c
1037
struct pvr_cccb *cccb = &queue->cccb;
drivers/gpu/drm/imagination/pvr_queue.c
1039
cctx_fw = fw_ctx_map + queue->ctx_offset;
drivers/gpu/drm/imagination/pvr_queue.c
1043
cctx_fw->dm = get_dm_type(queue);
drivers/gpu/drm/imagination/pvr_queue.c
1052
pvr_fw_object_get_fw_addr(queue->reg_state_obj, &cctx_fw->context_state_addr);
drivers/gpu/drm/imagination/pvr_queue.c
1063
static int pvr_queue_cleanup_fw_context(struct pvr_queue *queue)
drivers/gpu/drm/imagination/pvr_queue.c
1065
if (!queue->ctx->fw_obj)
drivers/gpu/drm/imagination/pvr_queue.c
1068
return pvr_fw_structure_cleanup(queue->ctx->pvr_dev,
drivers/gpu/drm/imagination/pvr_queue.c
1070
queue->ctx->fw_obj, queue->ctx_offset);
drivers/gpu/drm/imagination/pvr_queue.c
1090
struct pvr_queue *queue;
drivers/gpu/drm/imagination/pvr_queue.c
1096
queue = pvr_context_get_queue_for_job(job->ctx, job->type);
drivers/gpu/drm/imagination/pvr_queue.c
1097
if (!queue)
drivers/gpu/drm/imagination/pvr_queue.c
1100
if (!pvr_cccb_cmdseq_can_fit(&queue->cccb, job_cmds_size(job, min_native_dep_count)))
drivers/gpu/drm/imagination/pvr_queue.c
1103
err = drm_sched_job_init(&job->base, &queue->entity, 1, THIS_MODULE, drm_client_id);
drivers/gpu/drm/imagination/pvr_queue.c
116
pvr_context_put(fence->queue->ctx);
drivers/gpu/drm/imagination/pvr_queue.c
1166
struct pvr_queue *queue = container_of(job->base.sched, struct pvr_queue, scheduler);
drivers/gpu/drm/imagination/pvr_queue.c
1169
dma_fence_put(queue->last_queued_job_scheduled_fence);
drivers/gpu/drm/imagination/pvr_queue.c
1170
queue->last_queued_job_scheduled_fence = dma_fence_get(&job->base.s_fence->scheduled);
drivers/gpu/drm/imagination/pvr_queue.c
1178
struct pvr_queue *queue = priv;
drivers/gpu/drm/imagination/pvr_queue.c
1180
if (queue->type == DRM_PVR_JOB_TYPE_GEOMETRY) {
drivers/gpu/drm/imagination/pvr_queue.c
1184
queue->callstack_addr;
drivers/gpu/drm/imagination/pvr_queue.c
123
struct pvr_device *pvr_dev = fence->queue->ctx->pvr_dev;
drivers/gpu/drm/imagination/pvr_queue.c
1240
struct pvr_queue *queue;
drivers/gpu/drm/imagination/pvr_queue.c
1269
queue = kzalloc_obj(*queue);
drivers/gpu/drm/imagination/pvr_queue.c
1270
if (!queue)
drivers/gpu/drm/imagination/pvr_queue.c
1273
queue->type = type;
drivers/gpu/drm/imagination/pvr_queue.c
1274
queue->ctx_offset = get_ctx_offset(type);
drivers/gpu/drm/imagination/pvr_queue.c
1275
queue->ctx = ctx;
drivers/gpu/drm/imagination/pvr_queue.c
1276
queue->callstack_addr = args->callstack_addr;
drivers/gpu/drm/imagination/pvr_queue.c
1277
sched = &queue->scheduler;
drivers/gpu/drm/imagination/pvr_queue.c
1278
INIT_LIST_HEAD(&queue->node);
drivers/gpu/drm/imagination/pvr_queue.c
1279
mutex_init(&queue->cccb_fence_ctx.job_lock);
drivers/gpu/drm/imagination/pvr_queue.c
1280
pvr_queue_fence_ctx_init(&queue->cccb_fence_ctx.base);
drivers/gpu/drm/imagination/pvr_queue.c
1281
pvr_queue_fence_ctx_init(&queue->job_fence_ctx);
drivers/gpu/drm/imagination/pvr_queue.c
1283
err = pvr_cccb_init(pvr_dev, &queue->cccb, props[type].cccb_size, props[type].name);
drivers/gpu/drm/imagination/pvr_queue.c
1289
reg_state_init, queue, &queue->reg_state_obj);
drivers/gpu/drm/imagination/pvr_queue.c
1293
init_fw_context(queue, fw_ctx_map);
drivers/gpu/drm/imagination/pvr_queue.c
1301
cpu_map = pvr_fw_object_create_and_map(pvr_dev, sizeof(*queue->timeline_ufo.value),
drivers/gpu/drm/imagination/pvr_queue.c
1303
NULL, NULL, &queue->timeline_ufo.fw_obj);
drivers/gpu/drm/imagination/pvr_queue.c
1309
queue->timeline_ufo.value = cpu_map;
drivers/gpu/drm/imagination/pvr_queue.c
1311
err = drm_sched_init(&queue->scheduler, &sched_args);
drivers/gpu/drm/imagination/pvr_queue.c
1315
err = drm_sched_entity_init(&queue->entity,
drivers/gpu/drm/imagination/pvr_queue.c
1322
list_add_tail(&queue->node, &pvr_dev->queues.idle);
drivers/gpu/drm/imagination/pvr_queue.c
1325
return queue;
drivers/gpu/drm/imagination/pvr_queue.c
1328
drm_sched_fini(&queue->scheduler);
drivers/gpu/drm/imagination/pvr_queue.c
133
switch (fence->queue->type) {
drivers/gpu/drm/imagination/pvr_queue.c
1331
pvr_fw_object_unmap_and_destroy(queue->timeline_ufo.fw_obj);
drivers/gpu/drm/imagination/pvr_queue.c
1334
pvr_fw_object_destroy(queue->reg_state_obj);
drivers/gpu/drm/imagination/pvr_queue.c
1337
pvr_cccb_fini(&queue->cccb);
drivers/gpu/drm/imagination/pvr_queue.c
1340
mutex_destroy(&queue->cccb_fence_ctx.job_lock);
drivers/gpu/drm/imagination/pvr_queue.c
1341
kfree(queue);
drivers/gpu/drm/imagination/pvr_queue.c
1348
struct pvr_queue *queue;
drivers/gpu/drm/imagination/pvr_queue.c
1351
list_for_each_entry(queue, &pvr_dev->queues.idle, node)
drivers/gpu/drm/imagination/pvr_queue.c
1352
pvr_queue_stop(queue, NULL);
drivers/gpu/drm/imagination/pvr_queue.c
1353
list_for_each_entry(queue, &pvr_dev->queues.active, node)
drivers/gpu/drm/imagination/pvr_queue.c
1354
pvr_queue_stop(queue, NULL);
drivers/gpu/drm/imagination/pvr_queue.c
1360
struct pvr_queue *queue;
drivers/gpu/drm/imagination/pvr_queue.c
1363
list_for_each_entry(queue, &pvr_dev->queues.active, node)
drivers/gpu/drm/imagination/pvr_queue.c
1364
pvr_queue_start(queue);
drivers/gpu/drm/imagination/pvr_queue.c
1365
list_for_each_entry(queue, &pvr_dev->queues.idle, node)
drivers/gpu/drm/imagination/pvr_queue.c
1366
pvr_queue_start(queue);
drivers/gpu/drm/imagination/pvr_queue.c
1379
void pvr_queue_kill(struct pvr_queue *queue)
drivers/gpu/drm/imagination/pvr_queue.c
1381
drm_sched_entity_destroy(&queue->entity);
drivers/gpu/drm/imagination/pvr_queue.c
1382
dma_fence_put(queue->last_queued_job_scheduled_fence);
drivers/gpu/drm/imagination/pvr_queue.c
1383
queue->last_queued_job_scheduled_fence = NULL;
drivers/gpu/drm/imagination/pvr_queue.c
1393
void pvr_queue_destroy(struct pvr_queue *queue)
drivers/gpu/drm/imagination/pvr_queue.c
1395
if (!queue)
drivers/gpu/drm/imagination/pvr_queue.c
1398
mutex_lock(&queue->ctx->pvr_dev->queues.lock);
drivers/gpu/drm/imagination/pvr_queue.c
1399
list_del_init(&queue->node);
drivers/gpu/drm/imagination/pvr_queue.c
1400
mutex_unlock(&queue->ctx->pvr_dev->queues.lock);
drivers/gpu/drm/imagination/pvr_queue.c
1402
drm_sched_fini(&queue->scheduler);
drivers/gpu/drm/imagination/pvr_queue.c
1403
drm_sched_entity_fini(&queue->entity);
drivers/gpu/drm/imagination/pvr_queue.c
1405
if (WARN_ON(queue->last_queued_job_scheduled_fence))
drivers/gpu/drm/imagination/pvr_queue.c
1406
dma_fence_put(queue->last_queued_job_scheduled_fence);
drivers/gpu/drm/imagination/pvr_queue.c
1408
pvr_queue_cleanup_fw_context(queue);
drivers/gpu/drm/imagination/pvr_queue.c
1410
pvr_fw_object_unmap_and_destroy(queue->timeline_ufo.fw_obj);
drivers/gpu/drm/imagination/pvr_queue.c
1411
pvr_fw_object_destroy(queue->reg_state_obj);
drivers/gpu/drm/imagination/pvr_queue.c
1412
pvr_cccb_fini(&queue->cccb);
drivers/gpu/drm/imagination/pvr_queue.c
1413
mutex_destroy(&queue->cccb_fence_ctx.job_lock);
drivers/gpu/drm/imagination/pvr_queue.c
1414
kfree(queue);
drivers/gpu/drm/imagination/pvr_queue.c
156
switch (fence->queue->type) {
drivers/gpu/drm/imagination/pvr_queue.c
271
struct pvr_queue *queue,
drivers/gpu/drm/imagination/pvr_queue.c
277
pvr_context_get(queue->ctx);
drivers/gpu/drm/imagination/pvr_queue.c
278
fence->queue = queue;
drivers/gpu/drm/imagination/pvr_queue.c
296
pvr_queue_cccb_fence_init(struct dma_fence *fence, struct pvr_queue *queue)
drivers/gpu/drm/imagination/pvr_queue.c
298
pvr_queue_fence_init(fence, queue, &pvr_queue_cccb_fence_ops,
drivers/gpu/drm/imagination/pvr_queue.c
299
&queue->cccb_fence_ctx.base);
drivers/gpu/drm/imagination/pvr_queue.c
314
pvr_queue_job_fence_init(struct dma_fence *fence, struct pvr_queue *queue)
drivers/gpu/drm/imagination/pvr_queue.c
317
pvr_queue_fence_init(fence, queue, &pvr_queue_job_fence_ops,
drivers/gpu/drm/imagination/pvr_queue.c
318
&queue->job_fence_ctx);
drivers/gpu/drm/imagination/pvr_queue.c
398
pvr_queue_get_job_cccb_fence(struct pvr_queue *queue, struct pvr_job *job)
drivers/gpu/drm/imagination/pvr_queue.c
409
mutex_lock(&queue->cccb_fence_ctx.job_lock);
drivers/gpu/drm/imagination/pvr_queue.c
413
if (pvr_cccb_cmdseq_fits(&queue->cccb, job_cmds_size(job, native_deps_remaining))) {
drivers/gpu/drm/imagination/pvr_queue.c
422
if (WARN_ON(queue->cccb_fence_ctx.job))
drivers/gpu/drm/imagination/pvr_queue.c
423
pvr_job_put(queue->cccb_fence_ctx.job);
drivers/gpu/drm/imagination/pvr_queue.c
425
queue->cccb_fence_ctx.job = pvr_job_get(job);
drivers/gpu/drm/imagination/pvr_queue.c
429
if (!WARN_ON(cccb_fence->queue))
drivers/gpu/drm/imagination/pvr_queue.c
430
pvr_queue_cccb_fence_init(job->cccb_fence, queue);
drivers/gpu/drm/imagination/pvr_queue.c
433
mutex_unlock(&queue->cccb_fence_ctx.job_lock);
drivers/gpu/drm/imagination/pvr_queue.c
451
pvr_queue_get_job_kccb_fence(struct pvr_queue *queue, struct pvr_job *job)
drivers/gpu/drm/imagination/pvr_queue.c
453
struct pvr_device *pvr_dev = queue->ctx->pvr_dev;
drivers/gpu/drm/imagination/pvr_queue.c
471
pvr_queue_get_paired_frag_job_dep(struct pvr_queue *queue, struct pvr_job *job)
drivers/gpu/drm/imagination/pvr_queue.c
493
return frag_job->base.sched->ops->prepare_job(&frag_job->base, &queue->entity);
drivers/gpu/drm/imagination/pvr_queue.c
509
struct pvr_queue *queue = container_of(s_entity, struct pvr_queue, entity);
drivers/gpu/drm/imagination/pvr_queue.c
535
pvr_queue_job_fence_init(job->done_fence, queue);
drivers/gpu/drm/imagination/pvr_queue.c
541
internal_dep = pvr_queue_get_job_cccb_fence(queue, job);
drivers/gpu/drm/imagination/pvr_queue.c
547
internal_dep = pvr_queue_get_job_kccb_fence(queue, job);
drivers/gpu/drm/imagination/pvr_queue.c
558
internal_dep = pvr_queue_get_paired_frag_job_dep(queue, job);
drivers/gpu/drm/imagination/pvr_queue.c
570
static void pvr_queue_update_active_state_locked(struct pvr_queue *queue)
drivers/gpu/drm/imagination/pvr_queue.c
572
struct pvr_device *pvr_dev = queue->ctx->pvr_dev;
drivers/gpu/drm/imagination/pvr_queue.c
580
if (list_empty(&queue->node))
drivers/gpu/drm/imagination/pvr_queue.c
583
if (!atomic_read(&queue->in_flight_job_count))
drivers/gpu/drm/imagination/pvr_queue.c
584
list_move_tail(&queue->node, &pvr_dev->queues.idle);
drivers/gpu/drm/imagination/pvr_queue.c
586
list_move_tail(&queue->node, &pvr_dev->queues.active);
drivers/gpu/drm/imagination/pvr_queue.c
602
static void pvr_queue_update_active_state(struct pvr_queue *queue)
drivers/gpu/drm/imagination/pvr_queue.c
604
struct pvr_device *pvr_dev = queue->ctx->pvr_dev;
drivers/gpu/drm/imagination/pvr_queue.c
607
pvr_queue_update_active_state_locked(queue);
drivers/gpu/drm/imagination/pvr_queue.c
613
struct pvr_queue *queue = container_of(job->base.sched, struct pvr_queue, scheduler);
drivers/gpu/drm/imagination/pvr_queue.c
615
struct pvr_cccb *cccb = &queue->cccb;
drivers/gpu/drm/imagination/pvr_queue.c
625
atomic_inc(&queue->in_flight_job_count);
drivers/gpu/drm/imagination/pvr_queue.c
626
pvr_queue_update_active_state(queue);
drivers/gpu/drm/imagination/pvr_queue.c
641
pvr_fw_object_get_fw_addr(jfence->queue->timeline_ufo.fw_obj,
drivers/gpu/drm/imagination/pvr_queue.c
656
pvr_fw_object_get_fw_addr(jfence->queue->timeline_ufo.fw_obj,
drivers/gpu/drm/imagination/pvr_queue.c
673
pvr_fw_object_get_fw_addr(queue->timeline_ufo.fw_obj,
drivers/gpu/drm/imagination/pvr_queue.c
683
pvr_fw_object_get_fw_addr(queue->timeline_ufo.fw_obj, &ufos[0].addr);
drivers/gpu/drm/imagination/pvr_queue.c
755
struct pvr_queue *queue = container_of(job->base.sched,
drivers/gpu/drm/imagination/pvr_queue.c
758
pvr_cccb_send_kccb_kick(pvr_dev, &queue->cccb,
drivers/gpu/drm/imagination/pvr_queue.c
759
pvr_context_get_fw_addr(job->ctx) + queue->ctx_offset,
drivers/gpu/drm/imagination/pvr_queue.c
766
static void pvr_queue_stop(struct pvr_queue *queue, struct pvr_job *bad_job)
drivers/gpu/drm/imagination/pvr_queue.c
768
drm_sched_stop(&queue->scheduler, bad_job ? &bad_job->base : NULL);
drivers/gpu/drm/imagination/pvr_queue.c
771
static void pvr_queue_start(struct pvr_queue *queue)
drivers/gpu/drm/imagination/pvr_queue.c
778
*queue->timeline_ufo.value = atomic_read(&queue->job_fence_ctx.seqno);
drivers/gpu/drm/imagination/pvr_queue.c
780
list_for_each_entry(job, &queue->scheduler.pending_list, base.list) {
drivers/gpu/drm/imagination/pvr_queue.c
791
atomic_set(&queue->ctx->faulty, 1);
drivers/gpu/drm/imagination/pvr_queue.c
795
drm_sched_start(&queue->scheduler, 0);
drivers/gpu/drm/imagination/pvr_queue.c
812
struct pvr_queue *queue = container_of(sched, struct pvr_queue, scheduler);
drivers/gpu/drm/imagination/pvr_queue.c
813
struct pvr_device *pvr_dev = queue->ctx->pvr_dev;
drivers/gpu/drm/imagination/pvr_queue.c
830
list_del_init(&queue->node);
drivers/gpu/drm/imagination/pvr_queue.c
840
WARN_ON(atomic_read(&queue->in_flight_job_count) != job_count);
drivers/gpu/drm/imagination/pvr_queue.c
847
list_move_tail(&queue->node, &pvr_dev->queues.idle);
drivers/gpu/drm/imagination/pvr_queue.c
849
atomic_set(&queue->in_flight_job_count, job_count);
drivers/gpu/drm/imagination/pvr_queue.c
850
list_move_tail(&queue->node, &pvr_dev->queues.active);
drivers/gpu/drm/imagination/pvr_queue.c
851
pvr_queue_process(queue);
drivers/gpu/drm/imagination/pvr_queue.c
916
pvr_queue_signal_done_fences(struct pvr_queue *queue)
drivers/gpu/drm/imagination/pvr_queue.c
921
spin_lock(&queue->scheduler.job_list_lock);
drivers/gpu/drm/imagination/pvr_queue.c
922
cur_seqno = *queue->timeline_ufo.value;
drivers/gpu/drm/imagination/pvr_queue.c
923
list_for_each_entry_safe(job, tmp_job, &queue->scheduler.pending_list, base.list) {
drivers/gpu/drm/imagination/pvr_queue.c
930
atomic_dec(&queue->in_flight_job_count);
drivers/gpu/drm/imagination/pvr_queue.c
933
spin_unlock(&queue->scheduler.job_list_lock);
drivers/gpu/drm/imagination/pvr_queue.c
946
pvr_queue_check_job_waiting_for_cccb_space(struct pvr_queue *queue)
drivers/gpu/drm/imagination/pvr_queue.c
952
mutex_lock(&queue->cccb_fence_ctx.job_lock);
drivers/gpu/drm/imagination/pvr_queue.c
953
job = queue->cccb_fence_ctx.job;
drivers/gpu/drm/imagination/pvr_queue.c
967
if (WARN_ON(!cccb_fence->queue)) {
drivers/gpu/drm/imagination/pvr_queue.c
977
if (!pvr_cccb_cmdseq_fits(&queue->cccb, job_cmds_size(job, native_deps_remaining))) {
drivers/gpu/drm/imagination/pvr_queue.c
985
queue->cccb_fence_ctx.job = NULL;
drivers/gpu/drm/imagination/pvr_queue.c
988
mutex_unlock(&queue->cccb_fence_ctx.job_lock);
drivers/gpu/drm/imagination/pvr_queue.c
999
void pvr_queue_process(struct pvr_queue *queue)
drivers/gpu/drm/imagination/pvr_queue.h
159
void pvr_queue_kill(struct pvr_queue *queue);
drivers/gpu/drm/imagination/pvr_queue.h
161
void pvr_queue_destroy(struct pvr_queue *queue);
drivers/gpu/drm/imagination/pvr_queue.h
163
void pvr_queue_process(struct pvr_queue *queue);
drivers/gpu/drm/imagination/pvr_queue.h
66
struct pvr_queue *queue;
drivers/gpu/drm/msm/adreno/a2xx_gpu.c
25
if (ring->cur_ctx_seqno == submit->queue->ctx->seqno)
drivers/gpu/drm/msm/adreno/a3xx_gpu.c
43
if (ring->cur_ctx_seqno == submit->queue->ctx->seqno)
drivers/gpu/drm/msm/adreno/a4xx_gpu.c
37
if (ring->cur_ctx_seqno == submit->queue->ctx->seqno)
drivers/gpu/drm/msm/adreno/a5xx_gpu.c
176
if (ring->cur_ctx_seqno == submit->queue->ctx->seqno)
drivers/gpu/drm/msm/adreno/a5xx_gpu.c
80
if (ring->cur_ctx_seqno == submit->queue->ctx->seqno)
drivers/gpu/drm/msm/adreno/a6xx_gpu.c
205
struct msm_context *ctx = submit->queue->ctx;
drivers/gpu/drm/msm/adreno/a6xx_gpu.c
364
if (ring->cur_ctx_seqno == submit->queue->ctx->seqno)
drivers/gpu/drm/msm/adreno/a6xx_gpu.c
413
struct a6xx_gpu *a6xx_gpu, struct msm_gpu_submitqueue *queue)
drivers/gpu/drm/msm/adreno/a6xx_gpu.c
477
a6xx_emit_set_pseudo_reg(ring, a6xx_gpu, submit->queue);
drivers/gpu/drm/msm/adreno/a6xx_gpu.c
496
if (submit->queue->flags & MSM_SUBMITQUEUE_ALLOW_PREEMPT) {
drivers/gpu/drm/msm/adreno/a6xx_gpu.c
507
if (ring->cur_ctx_seqno == submit->queue->ctx->seqno)
drivers/gpu/drm/msm/adreno/a6xx_gpu.c
530
if (submit->queue->flags & MSM_SUBMITQUEUE_ALLOW_PREEMPT) {
drivers/gpu/drm/msm/adreno/a6xx_gpu.h
282
struct msm_gpu_submitqueue *queue);
drivers/gpu/drm/msm/adreno/a6xx_gpu.h
284
struct msm_gpu_submitqueue *queue);
drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
1298
struct a6xx_hfi_queue *queue = &gmu->queues[i];
drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
1300
unsigned idx = (j + queue->history_idx) % HFI_HISTORY_SZ;
drivers/gpu/drm/msm/adreno/a6xx_gpu_state.c
1301
a6xx_state->hfi_queue_history[i][j] = queue->history[idx];
drivers/gpu/drm/msm/adreno/a6xx_hfi.c
1011
struct a6xx_hfi_queue *queue = &gmu->queues[i];
drivers/gpu/drm/msm/adreno/a6xx_hfi.c
1013
if (!queue->header)
drivers/gpu/drm/msm/adreno/a6xx_hfi.c
1016
if (queue->header->read_index != queue->header->write_index)
drivers/gpu/drm/msm/adreno/a6xx_hfi.c
1019
queue->header->read_index = 0;
drivers/gpu/drm/msm/adreno/a6xx_hfi.c
1020
queue->header->write_index = 0;
drivers/gpu/drm/msm/adreno/a6xx_hfi.c
1022
memset(&queue->history, 0xff, sizeof(queue->history));
drivers/gpu/drm/msm/adreno/a6xx_hfi.c
1023
queue->history_idx = 0;
drivers/gpu/drm/msm/adreno/a6xx_hfi.c
1027
static void a6xx_hfi_queue_init(struct a6xx_hfi_queue *queue,
drivers/gpu/drm/msm/adreno/a6xx_hfi.c
1031
spin_lock_init(&queue->lock);
drivers/gpu/drm/msm/adreno/a6xx_hfi.c
1032
queue->header = header;
drivers/gpu/drm/msm/adreno/a6xx_hfi.c
1033
queue->data = virt;
drivers/gpu/drm/msm/adreno/a6xx_hfi.c
1034
atomic_set(&queue->seqnum, 0);
drivers/gpu/drm/msm/adreno/a6xx_hfi.c
1036
memset(&queue->history, 0xff, sizeof(queue->history));
drivers/gpu/drm/msm/adreno/a6xx_hfi.c
1037
queue->history_idx = 0;
drivers/gpu/drm/msm/adreno/a6xx_hfi.c
146
struct a6xx_hfi_queue *queue = &gmu->queues[HFI_RESPONSE_QUEUE];
drivers/gpu/drm/msm/adreno/a6xx_hfi.c
157
ret = a6xx_hfi_queue_read(gmu, queue, (u32 *) &resp,
drivers/gpu/drm/msm/adreno/a6xx_hfi.c
209
struct a6xx_hfi_queue *queue = &gmu->queues[HFI_COMMAND_QUEUE];
drivers/gpu/drm/msm/adreno/a6xx_hfi.c
213
seqnum = atomic_inc_return(&queue->seqnum) % 0xfff;
drivers/gpu/drm/msm/adreno/a6xx_hfi.c
219
ret = a6xx_hfi_queue_write(gmu, queue, data, dwords);
drivers/gpu/drm/msm/adreno/a6xx_hfi.c
32
struct a6xx_hfi_queue *queue, u32 *data, u32 dwords)
drivers/gpu/drm/msm/adreno/a6xx_hfi.c
34
struct a6xx_hfi_queue_header *header = queue->header;
drivers/gpu/drm/msm/adreno/a6xx_hfi.c
42
hdr = queue->data[index];
drivers/gpu/drm/msm/adreno/a6xx_hfi.c
44
queue->history[(queue->history_idx++) % HFI_HISTORY_SZ] = index;
drivers/gpu/drm/msm/adreno/a6xx_hfi.c
58
data[i] = queue->data[index];
drivers/gpu/drm/msm/adreno/a6xx_hfi.c
70
struct a6xx_hfi_queue *queue, u32 *data, u32 dwords)
drivers/gpu/drm/msm/adreno/a6xx_hfi.c
72
struct a6xx_hfi_queue_header *header = queue->header;
drivers/gpu/drm/msm/adreno/a6xx_hfi.c
75
spin_lock(&queue->lock);
drivers/gpu/drm/msm/adreno/a6xx_hfi.c
81
spin_unlock(&queue->lock);
drivers/gpu/drm/msm/adreno/a6xx_hfi.c
85
queue->history[(queue->history_idx++) % HFI_HISTORY_SZ] = index;
drivers/gpu/drm/msm/adreno/a6xx_hfi.c
88
queue->data[index] = data[i];
drivers/gpu/drm/msm/adreno/a6xx_hfi.c
95
queue->data[index] = 0xfafafafa;
drivers/gpu/drm/msm/adreno/a6xx_hfi.c
99
spin_unlock(&queue->lock);
drivers/gpu/drm/msm/disp/mdp4/mdp4_crtc.c
530
wait_queue_head_t *queue = drm_crtc_vblank_waitqueue(crtc);
drivers/gpu/drm/msm/disp/mdp4/mdp4_crtc.c
537
ret = wait_event_timeout(*queue,
drivers/gpu/drm/msm/disp/mdp5/mdp5_crtc.c
1236
wait_queue_head_t *queue = drm_crtc_vblank_waitqueue(crtc);
drivers/gpu/drm/msm/disp/mdp5/mdp5_crtc.c
1247
ret = wait_event_timeout(*queue,
drivers/gpu/drm/msm/msm_drv.c
657
static int wait_fence(struct msm_gpu_submitqueue *queue, uint32_t fence_id,
drivers/gpu/drm/msm/msm_drv.c
663
if (fence_after(fence_id, queue->last_fence)) {
drivers/gpu/drm/msm/msm_drv.c
665
fence_id, queue->last_fence);
drivers/gpu/drm/msm/msm_drv.c
677
spin_lock(&queue->idr_lock);
drivers/gpu/drm/msm/msm_drv.c
678
fence = idr_find(&queue->fence_idr, fence_id);
drivers/gpu/drm/msm/msm_drv.c
681
spin_unlock(&queue->idr_lock);
drivers/gpu/drm/msm/msm_drv.c
706
struct msm_gpu_submitqueue *queue;
drivers/gpu/drm/msm/msm_drv.c
717
queue = msm_submitqueue_get(file->driver_priv, args->queueid);
drivers/gpu/drm/msm/msm_drv.c
718
if (!queue)
drivers/gpu/drm/msm/msm_drv.c
721
ret = wait_fence(queue, args->fence, to_ktime(args->timeout), args->flags);
drivers/gpu/drm/msm/msm_drv.c
723
msm_submitqueue_put(queue);
drivers/gpu/drm/msm/msm_gem.h
450
struct msm_gpu_submitqueue *queue;
drivers/gpu/drm/msm/msm_gem_submit.c
101
spin_lock(&submit->queue->idr_lock);
drivers/gpu/drm/msm/msm_gem_submit.c
102
idr_remove(&submit->queue->fence_idr, submit->fence_id);
drivers/gpu/drm/msm/msm_gem_submit.c
103
spin_unlock(&submit->queue->idr_lock);
drivers/gpu/drm/msm/msm_gem_submit.c
122
msm_submitqueue_put(submit->queue);
drivers/gpu/drm/msm/msm_gem_submit.c
303
if (msm_context_is_vmbind(submit->queue->ctx))
drivers/gpu/drm/msm/msm_gem_submit.c
34
struct msm_gpu_submitqueue *queue, uint32_t nr_bos,
drivers/gpu/drm/msm/msm_gem_submit.c
412
if (msm_context_is_vmbind(submit->queue->ctx)) {
drivers/gpu/drm/msm/msm_gem_submit.c
557
struct msm_gpu_submitqueue *queue;
drivers/gpu/drm/msm/msm_gem_submit.c
59
ret = drm_sched_job_init(&submit->base, queue->entity, 1, queue,
drivers/gpu/drm/msm/msm_gem_submit.c
591
queue = msm_submitqueue_get(ctx, args->queueid);
drivers/gpu/drm/msm/msm_gem_submit.c
592
if (!queue)
drivers/gpu/drm/msm/msm_gem_submit.c
595
if (queue->flags & MSM_SUBMITQUEUE_VM_BIND) {
drivers/gpu/drm/msm/msm_gem_submit.c
600
ring = gpu->rb[queue->ring_nr];
drivers/gpu/drm/msm/msm_gem_submit.c
610
submit = submit_create(dev, gpu, queue, args->nr_bos, args->nr_cmds,
drivers/gpu/drm/msm/msm_gem_submit.c
620
ret = mutex_lock_interruptible(&queue->lock);
drivers/gpu/drm/msm/msm_gem_submit.c
69
submit->vm = msm_context_vm(dev, queue->ctx);
drivers/gpu/drm/msm/msm_gem_submit.c
72
submit->queue = queue;
drivers/gpu/drm/msm/msm_gem_submit.c
724
spin_lock(&queue->idr_lock);
drivers/gpu/drm/msm/msm_gem_submit.c
733
(!args->fence || idr_find(&queue->fence_idr, args->fence))) {
drivers/gpu/drm/msm/msm_gem_submit.c
734
spin_unlock(&queue->idr_lock);
drivers/gpu/drm/msm/msm_gem_submit.c
74
submit->ring = gpu->rb[queue->ring_nr];
drivers/gpu/drm/msm/msm_gem_submit.c
751
ret = idr_alloc_u32(&queue->fence_idr, submit->user_fence,
drivers/gpu/drm/msm/msm_gem_submit.c
764
submit->fence_id = idr_alloc_cyclic(&queue->fence_idr,
drivers/gpu/drm/msm/msm_gem_submit.c
769
spin_unlock(&queue->idr_lock);
drivers/gpu/drm/msm/msm_gem_submit.c
808
queue->last_fence = submit->fence_id;
drivers/gpu/drm/msm/msm_gem_submit.c
816
mutex_unlock(&queue->lock);
drivers/gpu/drm/msm/msm_gem_submit.c
835
msm_submitqueue_put(queue);
drivers/gpu/drm/msm/msm_gem_vma.c
116
struct msm_gpu_submitqueue *queue;
drivers/gpu/drm/msm/msm_gem_vma.c
1427
struct msm_gpu_submitqueue *queue;
drivers/gpu/drm/msm/msm_gem_vma.c
1457
queue = msm_submitqueue_get(ctx, args->queue_id);
drivers/gpu/drm/msm/msm_gem_vma.c
1458
if (!queue)
drivers/gpu/drm/msm/msm_gem_vma.c
1461
if (!(queue->flags & MSM_SUBMITQUEUE_VM_BIND)) {
drivers/gpu/drm/msm/msm_gem_vma.c
1474
job = vm_bind_job_create(dev, file, queue, args->nr_ops);
drivers/gpu/drm/msm/msm_gem_vma.c
1480
ret = mutex_lock_interruptible(&queue->lock);
drivers/gpu/drm/msm/msm_gem_vma.c
1579
mutex_unlock(&queue->lock);
drivers/gpu/drm/msm/msm_gem_vma.c
1599
msm_submitqueue_put(queue);
drivers/gpu/drm/msm/msm_gem_vma.c
525
.queue_id = job->queue->id,
drivers/gpu/drm/msm/msm_gem_vma.c
564
.queue_id = job->queue->id,
drivers/gpu/drm/msm/msm_gem_vma.c
671
.queue_id = job->queue->id,
drivers/gpu/drm/msm/msm_gem_vma.c
767
msm_submitqueue_put(job->queue);
drivers/gpu/drm/msm/msm_gem_vma.c
950
struct msm_gpu_submitqueue *queue, uint32_t nr_ops)
drivers/gpu/drm/msm/msm_gem_vma.c
959
ret = drm_sched_job_init(&job->base, queue->entity, 1, queue,
drivers/gpu/drm/msm/msm_gem_vma.c
966
job->vm = msm_context_vm(dev, queue->ctx);
drivers/gpu/drm/msm/msm_gem_vma.c
967
job->queue = queue;
drivers/gpu/drm/msm/msm_gpu.c
261
if (msm_context_is_vmbind(submit->queue->ctx)) {
drivers/gpu/drm/msm/msm_gpu.c
441
struct msm_context *ctx = submit->queue->ctx;
drivers/gpu/drm/msm/msm_gpu.c
488
submit->queue->faults++;
drivers/gpu/drm/msm/msm_gpu.c
798
submit->queue->ctx->elapsed_ns += elapsed;
drivers/gpu/drm/msm/msm_gpu.c
799
submit->queue->ctx->cycles += cycles;
drivers/gpu/drm/msm/msm_gpu.c
914
submit->ring->cur_ctx_seqno = submit->queue->ctx->seqno;
drivers/gpu/drm/msm/msm_gpu.h
715
static inline void msm_submitqueue_put(struct msm_gpu_submitqueue *queue)
drivers/gpu/drm/msm/msm_gpu.h
717
if (queue)
drivers/gpu/drm/msm/msm_gpu.h
718
kref_put(&queue->ref, msm_submitqueue_destroy);
drivers/gpu/drm/msm/msm_rd.c
375
if (msm_context_is_vmbind(submit->queue->ctx)) {
drivers/gpu/drm/msm/msm_ringbuffer.c
40
if (submit->queue->ctx->closed)
drivers/gpu/drm/msm/msm_submitqueue.c
112
struct msm_gpu_submitqueue *queue, *tmp;
drivers/gpu/drm/msm/msm_submitqueue.c
121
list_for_each_entry_safe(queue, tmp, &ctx->submitqueues, node) {
drivers/gpu/drm/msm/msm_submitqueue.c
122
if (queue->entity == &queue->_vm_bind_entity[0])
drivers/gpu/drm/msm/msm_submitqueue.c
123
drm_sched_entity_flush(queue->entity, MAX_WAIT_SCHED_ENTITY_Q_EMPTY);
drivers/gpu/drm/msm/msm_submitqueue.c
124
list_del(&queue->node);
drivers/gpu/drm/msm/msm_submitqueue.c
125
msm_submitqueue_put(queue);
drivers/gpu/drm/msm/msm_submitqueue.c
175
struct msm_gpu_submitqueue *queue;
drivers/gpu/drm/msm/msm_submitqueue.c
196
sz = struct_size(queue, _vm_bind_entity, 1);
drivers/gpu/drm/msm/msm_submitqueue.c
197
queue = kzalloc(sz, GFP_KERNEL);
drivers/gpu/drm/msm/msm_submitqueue.c
210
queue = kzalloc_obj(*queue);
drivers/gpu/drm/msm/msm_submitqueue.c
213
if (!queue)
drivers/gpu/drm/msm/msm_submitqueue.c
216
kref_init(&queue->ref);
drivers/gpu/drm/msm/msm_submitqueue.c
217
queue->flags = flags;
drivers/gpu/drm/msm/msm_submitqueue.c
222
queue->entity = &queue->_vm_bind_entity[0];
drivers/gpu/drm/msm/msm_submitqueue.c
224
drm_sched_entity_init(queue->entity, DRM_SCHED_PRIORITY_KERNEL,
drivers/gpu/drm/msm/msm_submitqueue.c
227
queue->ring_nr = ring_nr;
drivers/gpu/drm/msm/msm_submitqueue.c
229
queue->entity = get_sched_entity(ctx, priv->gpu->rb[ring_nr],
drivers/gpu/drm/msm/msm_submitqueue.c
233
if (IS_ERR(queue->entity)) {
drivers/gpu/drm/msm/msm_submitqueue.c
234
ret = PTR_ERR(queue->entity);
drivers/gpu/drm/msm/msm_submitqueue.c
235
kfree(queue);
drivers/gpu/drm/msm/msm_submitqueue.c
241
queue->ctx = msm_context_get(ctx);
drivers/gpu/drm/msm/msm_submitqueue.c
242
queue->id = ctx->queueid++;
drivers/gpu/drm/msm/msm_submitqueue.c
245
*id = queue->id;
drivers/gpu/drm/msm/msm_submitqueue.c
247
idr_init(&queue->fence_idr);
drivers/gpu/drm/msm/msm_submitqueue.c
248
spin_lock_init(&queue->idr_lock);
drivers/gpu/drm/msm/msm_submitqueue.c
249
mutex_init(&queue->lock);
drivers/gpu/drm/msm/msm_submitqueue.c
251
list_add_tail(&queue->node, &ctx->submitqueues);
drivers/gpu/drm/msm/msm_submitqueue.c
282
static int msm_submitqueue_query_faults(struct msm_gpu_submitqueue *queue,
drivers/gpu/drm/msm/msm_submitqueue.c
285
size_t size = min_t(size_t, args->len, sizeof(queue->faults));
drivers/gpu/drm/msm/msm_submitqueue.c
290
args->len = sizeof(queue->faults);
drivers/gpu/drm/msm/msm_submitqueue.c
297
ret = copy_to_user(u64_to_user_ptr(args->data), &queue->faults, size);
drivers/gpu/drm/msm/msm_submitqueue.c
305
struct msm_gpu_submitqueue *queue;
drivers/gpu/drm/msm/msm_submitqueue.c
311
queue = msm_submitqueue_get(ctx, args->id);
drivers/gpu/drm/msm/msm_submitqueue.c
312
if (!queue)
drivers/gpu/drm/msm/msm_submitqueue.c
316
ret = msm_submitqueue_query_faults(queue, args);
drivers/gpu/drm/msm/msm_submitqueue.c
318
msm_submitqueue_put(queue);
drivers/gpu/drm/msm/msm_submitqueue.c
74
struct msm_gpu_submitqueue *queue = container_of(kref,
drivers/gpu/drm/msm/msm_submitqueue.c
77
idr_destroy(&queue->fence_idr);
drivers/gpu/drm/msm/msm_submitqueue.c
79
if (queue->entity == &queue->_vm_bind_entity[0])
drivers/gpu/drm/msm/msm_submitqueue.c
80
drm_sched_entity_destroy(queue->entity);
drivers/gpu/drm/msm/msm_submitqueue.c
82
msm_context_put(queue->ctx);
drivers/gpu/drm/msm/msm_submitqueue.c
84
kfree(queue);
drivers/gpu/drm/panfrost/panfrost_job.c
103
fence->queue = js_num;
drivers/gpu/drm/panfrost/panfrost_job.c
104
fence->seqno = ++js->queue[js_num].emit_seqno;
drivers/gpu/drm/panfrost/panfrost_job.c
106
js->queue[js_num].fence_context, fence->seqno);
drivers/gpu/drm/panfrost/panfrost_job.c
1067
struct drm_gpu_scheduler *sched = &pfdev->js->queue[i].sched;
drivers/gpu/drm/panfrost/panfrost_job.c
42
struct panfrost_queue_state queue[NUM_JOB_SLOTS];
drivers/gpu/drm/panfrost/panfrost_job.c
502
drm_sched_fault(&pfdev->js->queue[js].sched);
drivers/gpu/drm/panfrost/panfrost_job.c
58
int queue;
drivers/gpu/drm/panfrost/panfrost_job.c
682
drm_sched_stop(&pfdev->js->queue[i].sched, bad);
drivers/gpu/drm/panfrost/panfrost_job.c
747
drm_sched_resubmit_jobs(&pfdev->js->queue[i].sched);
drivers/gpu/drm/panfrost/panfrost_job.c
752
drm_sched_start(&pfdev->js->queue[i].sched, 0);
drivers/gpu/drm/panfrost/panfrost_job.c
76
switch (f->queue) {
drivers/gpu/drm/panfrost/panfrost_job.c
898
js->queue[j].fence_context = dma_fence_context_alloc(1);
drivers/gpu/drm/panfrost/panfrost_job.c
901
ret = drm_sched_init(&js->queue[j].sched, &args);
drivers/gpu/drm/panfrost/panfrost_job.c
915
drm_sched_fini(&js->queue[j].sched);
drivers/gpu/drm/panfrost/panfrost_job.c
929
drm_sched_fini(&js->queue[j].sched);
drivers/gpu/drm/panfrost/panfrost_job.c
977
if (atomic_read(&js->queue[i].sched.credit_count))
drivers/gpu/drm/panthor/panthor_sched.c
1089
queue_timeout_is_suspended(struct panthor_queue *queue)
drivers/gpu/drm/panthor/panthor_sched.c
1092
return queue->timeout.remaining != MAX_SCHEDULE_TIMEOUT;
drivers/gpu/drm/panthor/panthor_sched.c
1096
queue_reset_timeout_locked(struct panthor_queue *queue)
drivers/gpu/drm/panthor/panthor_sched.c
1098
lockdep_assert_held(&queue->fence_ctx.lock);
drivers/gpu/drm/panthor/panthor_sched.c
1100
if (!queue_timeout_is_suspended(queue)) {
drivers/gpu/drm/panthor/panthor_sched.c
1101
mod_delayed_work(queue->scheduler.timeout_wq,
drivers/gpu/drm/panthor/panthor_sched.c
1102
&queue->timeout.work,
drivers/gpu/drm/panthor/panthor_sched.c
1108
queue_suspend_timeout_locked(struct panthor_queue *queue)
drivers/gpu/drm/panthor/panthor_sched.c
1115
lockdep_assert_held(&queue->fence_ctx.lock);
drivers/gpu/drm/panthor/panthor_sched.c
1118
if (queue_timeout_is_suspended(queue))
drivers/gpu/drm/panthor/panthor_sched.c
1121
job = list_first_entry_or_null(&queue->fence_ctx.in_flight_jobs,
drivers/gpu/drm/panthor/panthor_sched.c
1136
qtimeout = queue->timeout.work.timer.expires;
drivers/gpu/drm/panthor/panthor_sched.c
1139
timer_was_active = cancel_delayed_work(&queue->timeout.work);
drivers/gpu/drm/panthor/panthor_sched.c
1141
queue->timeout.remaining = msecs_to_jiffies(JOB_TIMEOUT_MS);
drivers/gpu/drm/panthor/panthor_sched.c
1143
queue->timeout.remaining = qtimeout - now;
drivers/gpu/drm/panthor/panthor_sched.c
1145
queue->timeout.remaining = 0;
drivers/gpu/drm/panthor/panthor_sched.c
1147
if (WARN_ON_ONCE(queue->timeout.remaining > msecs_to_jiffies(JOB_TIMEOUT_MS)))
drivers/gpu/drm/panthor/panthor_sched.c
1148
queue->timeout.remaining = msecs_to_jiffies(JOB_TIMEOUT_MS);
drivers/gpu/drm/panthor/panthor_sched.c
1152
queue_suspend_timeout(struct panthor_queue *queue)
drivers/gpu/drm/panthor/panthor_sched.c
1154
spin_lock(&queue->fence_ctx.lock);
drivers/gpu/drm/panthor/panthor_sched.c
1155
queue_suspend_timeout_locked(queue);
drivers/gpu/drm/panthor/panthor_sched.c
1156
spin_unlock(&queue->fence_ctx.lock);
drivers/gpu/drm/panthor/panthor_sched.c
1160
queue_resume_timeout(struct panthor_queue *queue)
drivers/gpu/drm/panthor/panthor_sched.c
1162
spin_lock(&queue->fence_ctx.lock);
drivers/gpu/drm/panthor/panthor_sched.c
1164
if (queue_timeout_is_suspended(queue)) {
drivers/gpu/drm/panthor/panthor_sched.c
1165
mod_delayed_work(queue->scheduler.timeout_wq,
drivers/gpu/drm/panthor/panthor_sched.c
1166
&queue->timeout.work,
drivers/gpu/drm/panthor/panthor_sched.c
1167
queue->timeout.remaining);
drivers/gpu/drm/panthor/panthor_sched.c
1169
queue->timeout.remaining = MAX_SCHEDULE_TIMEOUT;
drivers/gpu/drm/panthor/panthor_sched.c
1172
spin_unlock(&queue->fence_ctx.lock);
drivers/gpu/drm/panthor/panthor_sched.c
1189
struct panthor_queue *queue = ptdev->scheduler->csg_slots[csg_id].group->queues[cs_id];
drivers/gpu/drm/panthor/panthor_sched.c
1194
queue->iface.input->extract = queue->iface.output->extract;
drivers/gpu/drm/panthor/panthor_sched.c
1195
drm_WARN_ON(&ptdev->base, queue->iface.input->insert < queue->iface.input->extract);
drivers/gpu/drm/panthor/panthor_sched.c
1197
cs_iface->input->ringbuf_base = panthor_kernel_bo_gpuva(queue->ringbuf);
drivers/gpu/drm/panthor/panthor_sched.c
1198
cs_iface->input->ringbuf_size = panthor_kernel_bo_size(queue->ringbuf);
drivers/gpu/drm/panthor/panthor_sched.c
1199
cs_iface->input->ringbuf_input = queue->iface.input_fw_va;
drivers/gpu/drm/panthor/panthor_sched.c
1200
cs_iface->input->ringbuf_output = queue->iface.output_fw_va;
drivers/gpu/drm/panthor/panthor_sched.c
1201
cs_iface->input->config = CS_CONFIG_PRIORITY(queue->priority) |
drivers/gpu/drm/panthor/panthor_sched.c
1202
CS_CONFIG_DOORBELL(queue->doorbell_id);
drivers/gpu/drm/panthor/panthor_sched.c
1211
if (queue->iface.input->insert != queue->iface.input->extract)
drivers/gpu/drm/panthor/panthor_sched.c
1212
queue_resume_timeout(queue);
drivers/gpu/drm/panthor/panthor_sched.c
1231
struct panthor_queue *queue = group->queues[cs_id];
drivers/gpu/drm/panthor/panthor_sched.c
1239
queue_suspend_timeout(queue);
drivers/gpu/drm/panthor/panthor_sched.c
1279
struct panthor_queue *queue = group->queues[cs_id];
drivers/gpu/drm/panthor/panthor_sched.c
1287
if (queue->iface.input->insert == queue->iface.output->extract &&
drivers/gpu/drm/panthor/panthor_sched.c
1304
queue->syncwait.gpu_va = cs_iface->output->status_wait_sync_ptr;
drivers/gpu/drm/panthor/panthor_sched.c
1305
queue->syncwait.ref = cs_iface->output->status_wait_sync_value;
drivers/gpu/drm/panthor/panthor_sched.c
1307
queue->syncwait.gt = status_wait_cond == CS_STATUS_WAIT_SYNC_COND_GT;
drivers/gpu/drm/panthor/panthor_sched.c
1311
queue->syncwait.sync64 = true;
drivers/gpu/drm/panthor/panthor_sched.c
1312
queue->syncwait.ref |= sync_val_hi << 32;
drivers/gpu/drm/panthor/panthor_sched.c
1314
queue->syncwait.sync64 = false;
drivers/gpu/drm/panthor/panthor_sched.c
1529
struct panthor_queue *queue = group && cs_id < group->queue_count ?
drivers/gpu/drm/panthor/panthor_sched.c
1541
if (queue) {
drivers/gpu/drm/panthor/panthor_sched.c
1542
u64 cs_extract = queue->iface.output->extract;
drivers/gpu/drm/panthor/panthor_sched.c
1545
spin_lock(&queue->fence_ctx.lock);
drivers/gpu/drm/panthor/panthor_sched.c
1546
list_for_each_entry(job, &queue->fence_ctx.in_flight_jobs, node) {
drivers/gpu/drm/panthor/panthor_sched.c
1555
spin_unlock(&queue->fence_ctx.lock);
drivers/gpu/drm/panthor/panthor_sched.c
2029
struct list_head *queue,
drivers/gpu/drm/panthor/panthor_sched.c
2038
list_for_each_entry_safe(group, tmp, queue, run_node) {
drivers/gpu/drm/panthor/panthor_sched.c
2170
struct panthor_queue *queue = group->queues[i];
drivers/gpu/drm/panthor/panthor_sched.c
2181
if (!queue)
drivers/gpu/drm/panthor/panthor_sched.c
2184
spin_lock(&queue->fence_ctx.lock);
drivers/gpu/drm/panthor/panthor_sched.c
2185
list_for_each_entry_safe(job, tmp, &queue->fence_ctx.in_flight_jobs, node) {
drivers/gpu/drm/panthor/panthor_sched.c
2190
spin_unlock(&queue->fence_ctx.lock);
drivers/gpu/drm/panthor/panthor_sched.c
2195
syncobj->seqno = atomic64_read(&queue->fence_ctx.seqno);
drivers/gpu/drm/panthor/panthor_sched.c
2579
struct panthor_queue *queue = group->queues[queue_idx];
drivers/gpu/drm/panthor/panthor_sched.c
2587
syncobj = panthor_queue_get_syncwait_obj(group, queue);
drivers/gpu/drm/panthor/panthor_sched.c
2591
value = queue->syncwait.sync64 ?
drivers/gpu/drm/panthor/panthor_sched.c
2595
if (queue->syncwait.gt)
drivers/gpu/drm/panthor/panthor_sched.c
2596
result = value > queue->syncwait.ref;
drivers/gpu/drm/panthor/panthor_sched.c
2598
result = value <= queue->syncwait.ref;
drivers/gpu/drm/panthor/panthor_sched.c
2601
panthor_queue_put_syncwait_obj(queue);
drivers/gpu/drm/panthor/panthor_sched.c
2676
struct list_head *queue = &sched->groups.runnable[group->priority];
drivers/gpu/drm/panthor/panthor_sched.c
2694
list_move_tail(&group->run_node, queue);
drivers/gpu/drm/panthor/panthor_sched.c
2725
static void queue_stop(struct panthor_queue *queue,
drivers/gpu/drm/panthor/panthor_sched.c
2728
disable_delayed_work_sync(&queue->timeout.work);
drivers/gpu/drm/panthor/panthor_sched.c
2729
drm_sched_stop(&queue->scheduler, bad_job ? &bad_job->base : NULL);
drivers/gpu/drm/panthor/panthor_sched.c
2732
static void queue_start(struct panthor_queue *queue)
drivers/gpu/drm/panthor/panthor_sched.c
2737
list_for_each_entry(job, &queue->scheduler.pending_list, base.list)
drivers/gpu/drm/panthor/panthor_sched.c
2740
enable_delayed_work(&queue->timeout.work);
drivers/gpu/drm/panthor/panthor_sched.c
2741
drm_sched_start(&queue->scheduler, 0);
drivers/gpu/drm/panthor/panthor_sched.c
3008
struct panthor_queue *queue = group->queues[job->queue_idx];
drivers/gpu/drm/panthor/panthor_sched.c
3010
struct panthor_job_profiling_data *slots = queue->profiling.slots->kmap;
drivers/gpu/drm/panthor/panthor_sched.c
3041
static bool queue_check_job_completion(struct panthor_queue *queue)
drivers/gpu/drm/panthor/panthor_sched.c
3049
spin_lock(&queue->fence_ctx.lock);
drivers/gpu/drm/panthor/panthor_sched.c
3050
list_for_each_entry_safe(job, job_tmp, &queue->fence_ctx.in_flight_jobs, node) {
drivers/gpu/drm/panthor/panthor_sched.c
3065
if (list_empty(&queue->fence_ctx.in_flight_jobs)) {
drivers/gpu/drm/panthor/panthor_sched.c
3070
queue_suspend_timeout_locked(queue);
drivers/gpu/drm/panthor/panthor_sched.c
3078
queue_reset_timeout_locked(queue);
drivers/gpu/drm/panthor/panthor_sched.c
3081
spin_unlock(&queue->fence_ctx.lock);
drivers/gpu/drm/panthor/panthor_sched.c
3103
struct panthor_queue *queue = group->queues[queue_idx];
drivers/gpu/drm/panthor/panthor_sched.c
3105
if (!queue)
drivers/gpu/drm/panthor/panthor_sched.c
3108
queue_check_job_completion(queue);
drivers/gpu/drm/panthor/panthor_sched.c
3132
copy_instrs_to_ringbuf(struct panthor_queue *queue,
drivers/gpu/drm/panthor/panthor_sched.c
3136
u64 ringbuf_size = panthor_kernel_bo_size(queue->ringbuf);
drivers/gpu/drm/panthor/panthor_sched.c
3150
memcpy(queue->ringbuf->kmap + start, instrs->buffer, written);
drivers/gpu/drm/panthor/panthor_sched.c
3153
memcpy(queue->ringbuf->kmap,
drivers/gpu/drm/panthor/panthor_sched.c
3171
struct panthor_queue *queue = group->queues[job->queue_idx];
drivers/gpu/drm/panthor/panthor_sched.c
3183
params->times_addr = panthor_kernel_bo_gpuva(queue->profiling.slots) +
drivers/gpu/drm/panthor/panthor_sched.c
3300
struct panthor_queue *queue = group->queues[job->queue_idx];
drivers/gpu/drm/panthor/panthor_sched.c
3313
job->done_fence = dma_fence_get(queue->fence_ctx.last_fence);
drivers/gpu/drm/panthor/panthor_sched.c
3329
&queue->fence_ctx.lock,
drivers/gpu/drm/panthor/panthor_sched.c
3330
queue->fence_ctx.id,
drivers/gpu/drm/panthor/panthor_sched.c
3331
atomic64_inc_return(&queue->fence_ctx.seqno));
drivers/gpu/drm/panthor/panthor_sched.c
3333
job->profiling.slot = queue->profiling.seqno++;
drivers/gpu/drm/panthor/panthor_sched.c
3334
if (queue->profiling.seqno == queue->profiling.slot_count)
drivers/gpu/drm/panthor/panthor_sched.c
3335
queue->profiling.seqno = 0;
drivers/gpu/drm/panthor/panthor_sched.c
3337
job->ringbuf.start = queue->iface.input->insert;
drivers/gpu/drm/panthor/panthor_sched.c
3341
copy_instrs_to_ringbuf(queue, job, &instrs);
drivers/gpu/drm/panthor/panthor_sched.c
3346
spin_lock(&queue->fence_ctx.lock);
drivers/gpu/drm/panthor/panthor_sched.c
3347
list_add_tail(&job->node, &queue->fence_ctx.in_flight_jobs);
drivers/gpu/drm/panthor/panthor_sched.c
3348
spin_unlock(&queue->fence_ctx.lock);
drivers/gpu/drm/panthor/panthor_sched.c
3355
queue->iface.input->extract = queue->iface.output->extract;
drivers/gpu/drm/panthor/panthor_sched.c
3356
queue->iface.input->insert = job->ringbuf.end;
drivers/gpu/drm/panthor/panthor_sched.c
3373
gpu_write(ptdev, CSF_DOORBELL(queue->doorbell_id), 1);
drivers/gpu/drm/panthor/panthor_sched.c
3379
queue_resume_timeout(queue);
drivers/gpu/drm/panthor/panthor_sched.c
3384
dma_fence_put(queue->fence_ctx.last_fence);
drivers/gpu/drm/panthor/panthor_sched.c
3385
queue->fence_ctx.last_fence = dma_fence_get(job->done_fence);
drivers/gpu/drm/panthor/panthor_sched.c
3404
struct panthor_queue *queue = group->queues[job->queue_idx];
drivers/gpu/drm/panthor/panthor_sched.c
3411
queue_stop(queue, job);
drivers/gpu/drm/panthor/panthor_sched.c
3428
queue_start(queue);
drivers/gpu/drm/panthor/panthor_sched.c
3473
struct panthor_queue *queue = container_of(work, struct panthor_queue,
drivers/gpu/drm/panthor/panthor_sched.c
3477
progress = queue_check_job_completion(queue);
drivers/gpu/drm/panthor/panthor_sched.c
3479
drm_sched_fault(&queue->scheduler);
drivers/gpu/drm/panthor/panthor_sched.c
3503
struct panthor_queue *queue;
drivers/gpu/drm/panthor/panthor_sched.c
3516
queue = kzalloc_obj(*queue);
drivers/gpu/drm/panthor/panthor_sched.c
3517
if (!queue)
drivers/gpu/drm/panthor/panthor_sched.c
3520
queue->timeout.remaining = msecs_to_jiffies(JOB_TIMEOUT_MS);
drivers/gpu/drm/panthor/panthor_sched.c
3521
INIT_DELAYED_WORK(&queue->timeout.work, queue_timeout_work);
drivers/gpu/drm/panthor/panthor_sched.c
3522
queue->fence_ctx.id = dma_fence_context_alloc(1);
drivers/gpu/drm/panthor/panthor_sched.c
3523
spin_lock_init(&queue->fence_ctx.lock);
drivers/gpu/drm/panthor/panthor_sched.c
3524
INIT_LIST_HEAD(&queue->fence_ctx.in_flight_jobs);
drivers/gpu/drm/panthor/panthor_sched.c
3526
queue->priority = args->priority;
drivers/gpu/drm/panthor/panthor_sched.c
3528
queue->ringbuf = panthor_kernel_bo_create(group->ptdev, group->vm,
drivers/gpu/drm/panthor/panthor_sched.c
3535
if (IS_ERR(queue->ringbuf)) {
drivers/gpu/drm/panthor/panthor_sched.c
3536
ret = PTR_ERR(queue->ringbuf);
drivers/gpu/drm/panthor/panthor_sched.c
3540
ret = panthor_kernel_bo_vmap(queue->ringbuf);
drivers/gpu/drm/panthor/panthor_sched.c
3544
queue->iface.mem = panthor_fw_alloc_queue_iface_mem(group->ptdev,
drivers/gpu/drm/panthor/panthor_sched.c
3545
&queue->iface.input,
drivers/gpu/drm/panthor/panthor_sched.c
3546
&queue->iface.output,
drivers/gpu/drm/panthor/panthor_sched.c
3547
&queue->iface.input_fw_va,
drivers/gpu/drm/panthor/panthor_sched.c
3548
&queue->iface.output_fw_va);
drivers/gpu/drm/panthor/panthor_sched.c
3549
if (IS_ERR(queue->iface.mem)) {
drivers/gpu/drm/panthor/panthor_sched.c
3550
ret = PTR_ERR(queue->iface.mem);
drivers/gpu/drm/panthor/panthor_sched.c
3554
queue->profiling.slot_count =
drivers/gpu/drm/panthor/panthor_sched.c
3557
queue->profiling.slots =
drivers/gpu/drm/panthor/panthor_sched.c
3559
queue->profiling.slot_count *
drivers/gpu/drm/panthor/panthor_sched.c
3567
if (IS_ERR(queue->profiling.slots)) {
drivers/gpu/drm/panthor/panthor_sched.c
3568
ret = PTR_ERR(queue->profiling.slots);
drivers/gpu/drm/panthor/panthor_sched.c
3572
ret = panthor_kernel_bo_vmap(queue->profiling.slots);
drivers/gpu/drm/panthor/panthor_sched.c
3577
queue->name = kasprintf(GFP_KERNEL, "panthor-queue-%llu-%u-%u", drm_client_id, gid, qid);
drivers/gpu/drm/panthor/panthor_sched.c
3578
if (!queue->name) {
drivers/gpu/drm/panthor/panthor_sched.c
3583
sched_args.name = queue->name;
drivers/gpu/drm/panthor/panthor_sched.c
3585
ret = drm_sched_init(&queue->scheduler, &sched_args);
drivers/gpu/drm/panthor/panthor_sched.c
3589
drm_sched = &queue->scheduler;
drivers/gpu/drm/panthor/panthor_sched.c
3590
ret = drm_sched_entity_init(&queue->entity, 0, &drm_sched, 1, NULL);
drivers/gpu/drm/panthor/panthor_sched.c
3594
return queue;
drivers/gpu/drm/panthor/panthor_sched.c
3597
group_free_queue(group, queue);
drivers/gpu/drm/panthor/panthor_sched.c
3612
struct panthor_queue *queue;
drivers/gpu/drm/panthor/panthor_sched.c
3625
queue = group->queues[i];
drivers/gpu/drm/panthor/panthor_sched.c
3626
group->fdinfo.kbo_sizes += queue->ringbuf->obj->size;
drivers/gpu/drm/panthor/panthor_sched.c
3627
group->fdinfo.kbo_sizes += queue->iface.mem->obj->size;
drivers/gpu/drm/panthor/panthor_sched.c
3628
group->fdinfo.kbo_sizes += queue->profiling.slots->obj->size;
drivers/gpu/drm/panthor/panthor_sched.c
852
panthor_queue_put_syncwait_obj(struct panthor_queue *queue)
drivers/gpu/drm/panthor/panthor_sched.c
854
if (queue->syncwait.kmap) {
drivers/gpu/drm/panthor/panthor_sched.c
855
struct iosys_map map = IOSYS_MAP_INIT_VADDR(queue->syncwait.kmap);
drivers/gpu/drm/panthor/panthor_sched.c
857
drm_gem_vunmap(queue->syncwait.obj, &map);
drivers/gpu/drm/panthor/panthor_sched.c
858
queue->syncwait.kmap = NULL;
drivers/gpu/drm/panthor/panthor_sched.c
861
drm_gem_object_put(queue->syncwait.obj);
drivers/gpu/drm/panthor/panthor_sched.c
862
queue->syncwait.obj = NULL;
drivers/gpu/drm/panthor/panthor_sched.c
866
panthor_queue_get_syncwait_obj(struct panthor_group *group, struct panthor_queue *queue)
drivers/gpu/drm/panthor/panthor_sched.c
873
if (queue->syncwait.kmap) {
drivers/gpu/drm/panthor/panthor_sched.c
874
bo = container_of(queue->syncwait.obj,
drivers/gpu/drm/panthor/panthor_sched.c
880
queue->syncwait.gpu_va,
drivers/gpu/drm/panthor/panthor_sched.c
881
&queue->syncwait.offset);
drivers/gpu/drm/panthor/panthor_sched.c
885
queue->syncwait.obj = &bo->base.base;
drivers/gpu/drm/panthor/panthor_sched.c
886
ret = drm_gem_vmap(queue->syncwait.obj, &map);
drivers/gpu/drm/panthor/panthor_sched.c
890
queue->syncwait.kmap = map.vaddr;
drivers/gpu/drm/panthor/panthor_sched.c
891
if (drm_WARN_ON(&ptdev->base, !queue->syncwait.kmap))
drivers/gpu/drm/panthor/panthor_sched.c
901
queue->syncwait.offset,
drivers/gpu/drm/panthor/panthor_sched.c
902
queue->syncwait.sync64 ?
drivers/gpu/drm/panthor/panthor_sched.c
906
return queue->syncwait.kmap + queue->syncwait.offset;
drivers/gpu/drm/panthor/panthor_sched.c
909
panthor_queue_put_syncwait_obj(queue);
drivers/gpu/drm/panthor/panthor_sched.c
913
static void group_free_queue(struct panthor_group *group, struct panthor_queue *queue)
drivers/gpu/drm/panthor/panthor_sched.c
915
if (IS_ERR_OR_NULL(queue))
drivers/gpu/drm/panthor/panthor_sched.c
919
disable_delayed_work_sync(&queue->timeout.work);
drivers/gpu/drm/panthor/panthor_sched.c
921
if (queue->entity.fence_context)
drivers/gpu/drm/panthor/panthor_sched.c
922
drm_sched_entity_destroy(&queue->entity);
drivers/gpu/drm/panthor/panthor_sched.c
924
if (queue->scheduler.ops)
drivers/gpu/drm/panthor/panthor_sched.c
925
drm_sched_fini(&queue->scheduler);
drivers/gpu/drm/panthor/panthor_sched.c
927
kfree(queue->name);
drivers/gpu/drm/panthor/panthor_sched.c
929
panthor_queue_put_syncwait_obj(queue);
drivers/gpu/drm/panthor/panthor_sched.c
931
panthor_kernel_bo_destroy(queue->ringbuf);
drivers/gpu/drm/panthor/panthor_sched.c
932
panthor_kernel_bo_destroy(queue->iface.mem);
drivers/gpu/drm/panthor/panthor_sched.c
933
panthor_kernel_bo_destroy(queue->profiling.slots);
drivers/gpu/drm/panthor/panthor_sched.c
936
dma_fence_put(queue->fence_ctx.last_fence);
drivers/gpu/drm/panthor/panthor_sched.c
938
kfree(queue);
drivers/gpu/drm/radeon/cik.c
1842
u32 me, u32 pipe, u32 queue, u32 vmid)
drivers/gpu/drm/radeon/cik.c
1847
QUEUEID(queue & 0x7));
drivers/gpu/drm/radeon/cik.c
4148
cik_srbm_select(rdev, ring->me, ring->pipe, ring->queue, 0);
drivers/gpu/drm/radeon/cik.c
4167
cik_srbm_select(rdev, ring->me, ring->pipe, ring->queue, 0);
drivers/gpu/drm/radeon/cik.c
4189
cik_srbm_select(rdev, ring->me, ring->pipe, ring->queue, 0);
drivers/gpu/drm/radeon/cik.c
4607
rdev->ring[idx].queue, 0);
drivers/gpu/drm/radeon/cik.c
8412
ring->queue = 0; /* first queue */
drivers/gpu/drm/radeon/cik.c
8424
ring->queue = 1; /* second queue */
drivers/gpu/drm/radeon/radeon.h
817
u32 queue;
drivers/gpu/drm/v3d/v3d_drv.c
143
sched = &v3d->queue[i].sched;
drivers/gpu/drm/v3d/v3d_drv.c
167
struct v3d_queue_state *queue = &v3d->queue[q];
drivers/gpu/drm/v3d/v3d_drv.c
168
struct v3d_job *job = queue->active_job;
drivers/gpu/drm/v3d/v3d_drv.c
173
spin_lock_irqsave(&queue->queue_lock, irqflags);
drivers/gpu/drm/v3d/v3d_drv.c
175
spin_unlock_irqrestore(&queue->queue_lock, irqflags);
drivers/gpu/drm/v3d/v3d_drv.c
201
enum v3d_queue queue;
drivers/gpu/drm/v3d/v3d_drv.c
203
for (queue = 0; queue < V3D_MAX_QUEUES; queue++) {
drivers/gpu/drm/v3d/v3d_drv.c
204
struct v3d_stats *stats = &file_priv->stats[queue];
drivers/gpu/drm/v3d/v3d_drv.c
213
v3d_queue_to_string(queue), active_runtime);
drivers/gpu/drm/v3d/v3d_drv.c
219
v3d_queue_to_string(queue), jobs_completed);
drivers/gpu/drm/v3d/v3d_drv.h
163
struct v3d_queue_state queue[V3D_MAX_QUEUES];
drivers/gpu/drm/v3d/v3d_drv.h
264
enum v3d_queue queue;
drivers/gpu/drm/v3d/v3d_drv.h
27
static inline char *v3d_queue_to_string(enum v3d_queue queue)
drivers/gpu/drm/v3d/v3d_drv.h
29
switch (queue) {
drivers/gpu/drm/v3d/v3d_fence.c
16
fence->queue = q;
drivers/gpu/drm/v3d/v3d_fence.c
17
fence->seqno = ++queue->emit_seqno;
drivers/gpu/drm/v3d/v3d_fence.c
18
dma_fence_init(&fence->base, &v3d_fence_ops, &queue->fence_lock,
drivers/gpu/drm/v3d/v3d_fence.c
19
queue->fence_context, fence->seqno);
drivers/gpu/drm/v3d/v3d_fence.c
33
switch (f->queue) {
drivers/gpu/drm/v3d/v3d_fence.c
8
struct v3d_queue_state *queue = &v3d->queue[q];
drivers/gpu/drm/v3d/v3d_gem.c
288
struct v3d_queue_state *queue = &v3d->queue[i];
drivers/gpu/drm/v3d/v3d_gem.c
290
queue->fence_context = dma_fence_context_alloc(1);
drivers/gpu/drm/v3d/v3d_gem.c
291
memset(&queue->stats, 0, sizeof(queue->stats));
drivers/gpu/drm/v3d/v3d_gem.c
292
seqcount_init(&queue->stats.lock);
drivers/gpu/drm/v3d/v3d_gem.c
294
spin_lock_init(&queue->queue_lock);
drivers/gpu/drm/v3d/v3d_gem.c
295
spin_lock_init(&queue->fence_lock);
drivers/gpu/drm/v3d/v3d_gem.c
356
WARN_ON(v3d->queue[q].active_job);
drivers/gpu/drm/v3d/v3d_irq.c
47
struct v3d_queue_state *queue = &v3d->queue[V3D_BIN];
drivers/gpu/drm/v3d/v3d_irq.c
67
spin_lock_irqsave(&queue->queue_lock, irqflags);
drivers/gpu/drm/v3d/v3d_irq.c
68
bin_job = (struct v3d_bin_job *)queue->active_job;
drivers/gpu/drm/v3d/v3d_irq.c
71
spin_unlock_irqrestore(&queue->queue_lock, irqflags);
drivers/gpu/drm/v3d/v3d_irq.c
77
spin_unlock_irqrestore(&queue->queue_lock, irqflags);
drivers/gpu/drm/v3d/v3d_irq.c
92
struct v3d_queue_state *queue = &v3d->queue[q];
drivers/gpu/drm/v3d/v3d_irq.c
93
struct v3d_fence *fence = to_v3d_fence(queue->active_job->irq_fence);
drivers/gpu/drm/v3d/v3d_irq.c
95
v3d_job_update_stats(queue->active_job, q);
drivers/gpu/drm/v3d/v3d_irq.c
98
queue->active_job = NULL;
drivers/gpu/drm/v3d/v3d_sched.c
140
v3d_job_start_stats(struct v3d_job *job, enum v3d_queue queue)
drivers/gpu/drm/v3d/v3d_sched.c
144
struct v3d_stats *global_stats = &v3d->queue[queue].stats;
drivers/gpu/drm/v3d/v3d_sched.c
145
struct v3d_stats *local_stats = &file->stats[queue];
drivers/gpu/drm/v3d/v3d_sched.c
201
struct v3d_queue_state *queue = &v3d->queue[q];
drivers/gpu/drm/v3d/v3d_sched.c
202
struct v3d_stats *global_stats = &queue->stats;
drivers/gpu/drm/v3d/v3d_sched.c
213
spin_lock(&queue->queue_lock);
drivers/gpu/drm/v3d/v3d_sched.c
216
spin_unlock(&queue->queue_lock);
drivers/gpu/drm/v3d/v3d_sched.c
230
struct v3d_queue_state *queue = &v3d->queue[V3D_BIN];
drivers/gpu/drm/v3d/v3d_sched.c
236
spin_lock_irqsave(&queue->queue_lock, irqflags);
drivers/gpu/drm/v3d/v3d_sched.c
237
queue->active_job = NULL;
drivers/gpu/drm/v3d/v3d_sched.c
238
spin_unlock_irqrestore(&queue->queue_lock, irqflags);
drivers/gpu/drm/v3d/v3d_sched.c
245
spin_lock_irqsave(&queue->queue_lock, irqflags);
drivers/gpu/drm/v3d/v3d_sched.c
246
queue->active_job = &job->base;
drivers/gpu/drm/v3d/v3d_sched.c
251
spin_unlock_irqrestore(&queue->queue_lock, irqflags);
drivers/gpu/drm/v3d/v3d_sched.c
295
v3d->queue[V3D_RENDER].active_job = NULL;
drivers/gpu/drm/v3d/v3d_sched.c
299
v3d->queue[V3D_RENDER].active_job = &job->base;
drivers/gpu/drm/v3d/v3d_sched.c
343
v3d->queue[V3D_TFU].active_job = NULL;
drivers/gpu/drm/v3d/v3d_sched.c
347
v3d->queue[V3D_TFU].active_job = &job->base;
drivers/gpu/drm/v3d/v3d_sched.c
391
v3d->queue[V3D_CSD].active_job = NULL;
drivers/gpu/drm/v3d/v3d_sched.c
395
v3d->queue[V3D_CSD].active_job = &job->base;
drivers/gpu/drm/v3d/v3d_sched.c
738
drm_sched_stop(&v3d->queue[i].sched, sched_job);
drivers/gpu/drm/v3d/v3d_sched.c
747
spin_lock_irqsave(&v3d->queue[q].queue_lock, irqflags);
drivers/gpu/drm/v3d/v3d_sched.c
750
spin_unlock_irqrestore(&v3d->queue[q].queue_lock, irqflags);
drivers/gpu/drm/v3d/v3d_sched.c
753
drm_sched_resubmit_jobs(&v3d->queue[i].sched);
drivers/gpu/drm/v3d/v3d_sched.c
757
drm_sched_start(&v3d->queue[i].sched, 0);
drivers/gpu/drm/v3d/v3d_sched.c
869
enum v3d_queue queue, const char *name)
drivers/gpu/drm/v3d/v3d_sched.c
881
return drm_sched_init(&v3d->queue[queue].sched, &args);
drivers/gpu/drm/v3d/v3d_sched.c
931
if (v3d->queue[q].sched.ready)
drivers/gpu/drm/v3d/v3d_sched.c
932
drm_sched_fini(&v3d->queue[q].sched);
drivers/gpu/drm/v3d/v3d_submit.c
162
u32 in_sync, struct v3d_submit_ext *se, enum v3d_queue queue)
drivers/gpu/drm/v3d/v3d_submit.c
172
ret = drm_sched_job_init(&job->base, &v3d_priv->sched_entity[queue],
drivers/gpu/drm/v3d/v3d_submit.c
178
if (se->in_sync_count && se->wait_stage == queue) {
drivers/gpu/drm/v3d/v3d_sysfs.c
16
enum v3d_queue queue;
drivers/gpu/drm/v3d/v3d_sysfs.c
22
for (queue = 0; queue < V3D_MAX_QUEUES; queue++) {
drivers/gpu/drm/v3d/v3d_sysfs.c
23
struct v3d_stats *stats = &v3d->queue[queue].stats;
drivers/gpu/drm/v3d/v3d_sysfs.c
39
v3d_queue_to_string(queue),
drivers/gpu/drm/xe/xe_gt_sriov_vf.c
1355
goto queue;
drivers/gpu/drm/xe/xe_gt_sriov_vf.c
1381
goto queue;
drivers/gpu/drm/xe/xe_gt_sriov_vf.c
1398
queue:
drivers/hid/hid-wiimote-core.c
100
wdata->queue.head = newhead;
drivers/hid/hid-wiimote-core.c
101
schedule_work(&wdata->queue.worker);
drivers/hid/hid-wiimote-core.c
102
} else if (newhead != wdata->queue.tail) {
drivers/hid/hid-wiimote-core.c
103
wdata->queue.head = newhead;
drivers/hid/hid-wiimote-core.c
114
spin_unlock_irqrestore(&wdata->queue.lock, flags);
drivers/hid/hid-wiimote-core.c
1747
spin_lock_init(&wdata->queue.lock);
drivers/hid/hid-wiimote-core.c
1748
INIT_WORK(&wdata->queue.worker, wiimote_queue_worker);
drivers/hid/hid-wiimote-core.c
1782
cancel_work_sync(&wdata->queue.worker);
drivers/hid/hid-wiimote-core.c
43
struct wiimote_queue *queue = container_of(work, struct wiimote_queue,
drivers/hid/hid-wiimote-core.c
45
struct wiimote_data *wdata = container_of(queue, struct wiimote_data,
drivers/hid/hid-wiimote-core.c
46
queue);
drivers/hid/hid-wiimote-core.c
50
spin_lock_irqsave(&wdata->queue.lock, flags);
drivers/hid/hid-wiimote-core.c
52
while (wdata->queue.head != wdata->queue.tail) {
drivers/hid/hid-wiimote-core.c
53
spin_unlock_irqrestore(&wdata->queue.lock, flags);
drivers/hid/hid-wiimote-core.c
55
wdata->queue.outq[wdata->queue.tail].data,
drivers/hid/hid-wiimote-core.c
56
wdata->queue.outq[wdata->queue.tail].size);
drivers/hid/hid-wiimote-core.c
62
spin_lock_irqsave(&wdata->queue.lock, flags);
drivers/hid/hid-wiimote-core.c
64
wdata->queue.tail = (wdata->queue.tail + 1) % WIIMOTE_BUFSIZE;
drivers/hid/hid-wiimote-core.c
67
spin_unlock_irqrestore(&wdata->queue.lock, flags);
drivers/hid/hid-wiimote-core.c
79
spin_lock_irqsave(&wdata->queue.lock, flags);
drivers/hid/hid-wiimote-core.c
93
spin_lock_irqsave(&wdata->queue.lock, flags);
drivers/hid/hid-wiimote-core.c
95
memcpy(wdata->queue.outq[wdata->queue.head].data, buffer, count);
drivers/hid/hid-wiimote-core.c
96
wdata->queue.outq[wdata->queue.head].size = count;
drivers/hid/hid-wiimote-core.c
97
newhead = (wdata->queue.head + 1) % WIIMOTE_BUFSIZE;
drivers/hid/hid-wiimote-core.c
99
if (wdata->queue.head == wdata->queue.tail) {
drivers/hid/hid-wiimote.h
161
struct wiimote_queue queue;
drivers/hsi/clients/hsi_char.c
117
struct list_head *queue)
drivers/hsi/clients/hsi_char.c
122
list_add_tail(&msg->link, queue);
drivers/hsi/clients/hsi_char.c
127
struct list_head *queue)
drivers/hsi/clients/hsi_char.c
134
if (list_empty(queue))
drivers/hsi/clients/hsi_char.c
137
msg = list_first_entry(queue, struct hsi_msg, link);
drivers/hsi/controllers/omap_ssi_port.c
314
static int ssi_start_transfer(struct list_head *queue)
drivers/hsi/controllers/omap_ssi_port.c
319
if (list_empty(queue))
drivers/hsi/controllers/omap_ssi_port.c
321
msg = list_first_entry(queue, struct hsi_msg, link);
drivers/hsi/controllers/omap_ssi_port.c
374
struct list_head *queue;
drivers/hsi/controllers/omap_ssi_port.c
387
queue = &omap_port->txqueue[msg->channel];
drivers/hsi/controllers/omap_ssi_port.c
390
queue = &omap_port->rxqueue[msg->channel];
drivers/hsi/controllers/omap_ssi_port.c
396
list_add_tail(&msg->link, queue);
drivers/hsi/controllers/omap_ssi_port.c
397
err = ssi_start_transfer(queue);
drivers/hsi/controllers/omap_ssi_port.c
426
static void ssi_flush_queue(struct list_head *queue, struct hsi_client *cl)
drivers/hsi/controllers/omap_ssi_port.c
431
list_for_each_safe(node, tmp, queue) {
drivers/hsi/controllers/omap_ssi_port.c
631
struct list_head *queue)
drivers/hsi/controllers/omap_ssi_port.c
639
err = ssi_start_transfer(queue);
drivers/hsi/controllers/omap_ssi_port.c
641
msg = list_first_entry(queue, struct hsi_msg, link);
drivers/hsi/controllers/omap_ssi_port.c
879
static void ssi_pio_complete(struct hsi_port *port, struct list_head *queue)
drivers/hsi/controllers/omap_ssi_port.c
890
msg = list_first_entry(queue, struct hsi_msg, link);
drivers/hsi/controllers/omap_ssi_port.c
939
ssi_transfer(omap_port, queue);
drivers/i2c/busses/i2c-imx.c
1068
wake_up(&i2c_imx->queue);
drivers/i2c/busses/i2c-imx.c
1085
wake_up(&i2c_imx->queue);
drivers/i2c/busses/i2c-imx.c
1097
wake_up(&i2c_imx->queue);
drivers/i2c/busses/i2c-imx.c
1116
wake_up(&i2c_imx->queue);
drivers/i2c/busses/i2c-imx.c
1123
wake_up(&i2c_imx->queue);
drivers/i2c/busses/i2c-imx.c
1396
wait_event_timeout(i2c_imx->queue,
drivers/i2c/busses/i2c-imx.c
1515
wait_event_timeout(i2c_imx->queue,
drivers/i2c/busses/i2c-imx.c
1762
init_waitqueue_head(&i2c_imx->queue);
drivers/i2c/busses/i2c-imx.c
243
wait_queue_head_t queue;
drivers/i2c/busses/i2c-imx.c
593
wait_event_timeout(i2c_imx->queue, i2c_imx->i2csr & I2SR_IIF, HZ / 10);
drivers/i2c/busses/i2c-imx.c
978
wake_up(&i2c_imx->queue);
drivers/i2c/busses/i2c-octeon-core.c
189
time_left = wait_event_timeout(i2c->queue,
drivers/i2c/busses/i2c-octeon-core.c
34
wake_up(&i2c->queue);
drivers/i2c/busses/i2c-octeon-core.c
69
time_left = wait_event_timeout(i2c->queue, octeon_i2c_test_iflg(i2c),
drivers/i2c/busses/i2c-octeon-core.h
125
wait_queue_head_t queue;
drivers/i2c/busses/i2c-octeon-platdrv.c
108
wake_up(&i2c->queue);
drivers/i2c/busses/i2c-octeon-platdrv.c
190
init_waitqueue_head(&i2c->queue);
drivers/i2c/busses/i2c-qcom-cci.c
294
static int cci_run_queue(struct cci *cci, u8 master, u8 queue)
drivers/i2c/busses/i2c-qcom-cci.c
298
val = readl(cci->base + CCI_I2C_Mm_Qn_CUR_WORD_CNT(master, queue));
drivers/i2c/busses/i2c-qcom-cci.c
299
writel(val, cci->base + CCI_I2C_Mm_Qn_EXEC_WORD_CNT(master, queue));
drivers/i2c/busses/i2c-qcom-cci.c
302
val = BIT(master * 2 + queue);
drivers/i2c/busses/i2c-qcom-cci.c
308
master, queue);
drivers/i2c/busses/i2c-qcom-cci.c
317
static int cci_validate_queue(struct cci *cci, u8 master, u8 queue)
drivers/i2c/busses/i2c-qcom-cci.c
321
val = readl(cci->base + CCI_I2C_Mm_Qn_CUR_WORD_CNT(master, queue));
drivers/i2c/busses/i2c-qcom-cci.c
322
if (val == cci->data->queue_size[queue])
drivers/i2c/busses/i2c-qcom-cci.c
329
writel(val, cci->base + CCI_I2C_Mm_Qn_LOAD_DATA(master, queue));
drivers/i2c/busses/i2c-qcom-cci.c
331
return cci_run_queue(cci, master, queue);
drivers/i2c/busses/i2c-qcom-cci.c
338
u8 queue = QUEUE_1;
drivers/i2c/busses/i2c-qcom-cci.c
346
ret = cci_validate_queue(cci, master, queue);
drivers/i2c/busses/i2c-qcom-cci.c
351
writel(val, cci->base + CCI_I2C_Mm_Qn_LOAD_DATA(master, queue));
drivers/i2c/busses/i2c-qcom-cci.c
354
writel(val, cci->base + CCI_I2C_Mm_Qn_LOAD_DATA(master, queue));
drivers/i2c/busses/i2c-qcom-cci.c
356
ret = cci_run_queue(cci, master, queue);
drivers/i2c/busses/i2c-qcom-cci.c
390
u8 queue = QUEUE_0;
drivers/i2c/busses/i2c-qcom-cci.c
399
ret = cci_validate_queue(cci, master, queue);
drivers/i2c/busses/i2c-qcom-cci.c
404
writel(val, cci->base + CCI_I2C_Mm_Qn_LOAD_DATA(master, queue));
drivers/i2c/busses/i2c-qcom-cci.c
416
writel(val, cci->base + CCI_I2C_Mm_Qn_LOAD_DATA(master, queue));
drivers/i2c/busses/i2c-qcom-cci.c
420
writel(val, cci->base + CCI_I2C_Mm_Qn_LOAD_DATA(master, queue));
drivers/i2c/busses/i2c-qcom-cci.c
422
return cci_run_queue(cci, master, queue);
drivers/i2c/busses/i2c-thunderx-pcidrv.c
194
init_waitqueue_head(&i2c->queue);
drivers/i3c/master/mipi-i3c-hci/pio.c
743
u32 queue = pio_reg_read(QUEUE_CUR_STATUS);
drivers/i3c/master/mipi-i3c-hci/pio.c
749
FIELD_GET(CUR_CMD_Q_EMPTY_LEVEL, queue),
drivers/i3c/master/mipi-i3c-hci/pio.c
750
FIELD_GET(CUR_RESP_Q_LEVEL, queue),
drivers/i3c/master/mipi-i3c-hci/pio.c
751
FIELD_GET(CUR_IBI_Q_LEVEL, queue),
drivers/iio/buffer/industrialio-buffer-dma.c
102
struct iio_dma_buffer_queue *queue = block->queue;
drivers/iio/buffer/industrialio-buffer-dma.c
107
dma_free_coherent(queue->dev, PAGE_ALIGN(block->size),
drivers/iio/buffer/industrialio-buffer-dma.c
110
atomic_dec(&queue->num_dmabufs);
drivers/iio/buffer/industrialio-buffer-dma.c
113
iio_buffer_put(&queue->buffer);
drivers/iio/buffer/industrialio-buffer-dma.c
173
iio_dma_buffer_alloc_block(struct iio_dma_buffer_queue *queue, size_t size,
drivers/iio/buffer/industrialio-buffer-dma.c
182
block->vaddr = dma_alloc_coherent(queue->dev, PAGE_ALIGN(size),
drivers/iio/buffer/industrialio-buffer-dma.c
191
block->queue = queue;
drivers/iio/buffer/industrialio-buffer-dma.c
195
iio_buffer_get(&queue->buffer);
drivers/iio/buffer/industrialio-buffer-dma.c
198
atomic_inc(&queue->num_dmabufs);
drivers/iio/buffer/industrialio-buffer-dma.c
209
static void iio_dma_buffer_queue_wake(struct iio_dma_buffer_queue *queue)
drivers/iio/buffer/industrialio-buffer-dma.c
213
if (queue->buffer.direction == IIO_BUFFER_DIRECTION_IN)
drivers/iio/buffer/industrialio-buffer-dma.c
218
wake_up_interruptible_poll(&queue->buffer.pollq, flags);
drivers/iio/buffer/industrialio-buffer-dma.c
230
struct iio_dma_buffer_queue *queue = block->queue;
drivers/iio/buffer/industrialio-buffer-dma.c
235
scoped_guard(spinlock_irqsave, &queue->list_lock)
drivers/iio/buffer/industrialio-buffer-dma.c
242
iio_dma_buffer_queue_wake(queue);
drivers/iio/buffer/industrialio-buffer-dma.c
257
void iio_dma_buffer_block_list_abort(struct iio_dma_buffer_queue *queue,
drivers/iio/buffer/industrialio-buffer-dma.c
265
scoped_guard(spinlock_irqsave, &queue->list_lock) {
drivers/iio/buffer/industrialio-buffer-dma.c
278
if (queue->fileio.enabled)
drivers/iio/buffer/industrialio-buffer-dma.c
279
queue->fileio.enabled = false;
drivers/iio/buffer/industrialio-buffer-dma.c
281
iio_dma_buffer_queue_wake(queue);
drivers/iio/buffer/industrialio-buffer-dma.c
302
static bool iio_dma_buffer_can_use_fileio(struct iio_dma_buffer_queue *queue)
drivers/iio/buffer/industrialio-buffer-dma.c
309
return queue->fileio.enabled || !atomic_read(&queue->num_dmabufs);
drivers/iio/buffer/industrialio-buffer-dma.c
321
struct iio_dma_buffer_queue *queue = iio_buffer_to_queue(buffer);
drivers/iio/buffer/industrialio-buffer-dma.c
332
size = DIV_ROUND_UP(queue->buffer.bytes_per_datum *
drivers/iio/buffer/industrialio-buffer-dma.c
333
queue->buffer.length, 2);
drivers/iio/buffer/industrialio-buffer-dma.c
335
guard(mutex)(&queue->lock);
drivers/iio/buffer/industrialio-buffer-dma.c
337
queue->fileio.enabled = iio_dma_buffer_can_use_fileio(queue);
drivers/iio/buffer/industrialio-buffer-dma.c
340
if (!queue->fileio.enabled)
drivers/iio/buffer/industrialio-buffer-dma.c
344
if (PAGE_ALIGN(queue->fileio.block_size) == PAGE_ALIGN(size))
drivers/iio/buffer/industrialio-buffer-dma.c
347
queue->fileio.block_size = size;
drivers/iio/buffer/industrialio-buffer-dma.c
348
queue->fileio.active_block = NULL;
drivers/iio/buffer/industrialio-buffer-dma.c
350
scoped_guard(spinlock_irq, &queue->list_lock) {
drivers/iio/buffer/industrialio-buffer-dma.c
351
for (i = 0; i < ARRAY_SIZE(queue->fileio.blocks); i++) {
drivers/iio/buffer/industrialio-buffer-dma.c
352
block = queue->fileio.blocks[i];
drivers/iio/buffer/industrialio-buffer-dma.c
366
INIT_LIST_HEAD(&queue->incoming);
drivers/iio/buffer/industrialio-buffer-dma.c
368
for (i = 0; i < ARRAY_SIZE(queue->fileio.blocks); i++) {
drivers/iio/buffer/industrialio-buffer-dma.c
369
if (queue->fileio.blocks[i]) {
drivers/iio/buffer/industrialio-buffer-dma.c
370
block = queue->fileio.blocks[i];
drivers/iio/buffer/industrialio-buffer-dma.c
383
block = iio_dma_buffer_alloc_block(queue, size, true);
drivers/iio/buffer/industrialio-buffer-dma.c
387
queue->fileio.blocks[i] = block;
drivers/iio/buffer/industrialio-buffer-dma.c
402
if (queue->buffer.direction == IIO_BUFFER_DIRECTION_IN) {
drivers/iio/buffer/industrialio-buffer-dma.c
404
list_add_tail(&block->head, &queue->incoming);
drivers/iio/buffer/industrialio-buffer-dma.c
414
static void iio_dma_buffer_fileio_free(struct iio_dma_buffer_queue *queue)
drivers/iio/buffer/industrialio-buffer-dma.c
418
scoped_guard(spinlock_irq, &queue->list_lock) {
drivers/iio/buffer/industrialio-buffer-dma.c
419
for (i = 0; i < ARRAY_SIZE(queue->fileio.blocks); i++) {
drivers/iio/buffer/industrialio-buffer-dma.c
420
if (!queue->fileio.blocks[i])
drivers/iio/buffer/industrialio-buffer-dma.c
422
queue->fileio.blocks[i]->state = IIO_BLOCK_STATE_DEAD;
drivers/iio/buffer/industrialio-buffer-dma.c
426
INIT_LIST_HEAD(&queue->incoming);
drivers/iio/buffer/industrialio-buffer-dma.c
428
for (i = 0; i < ARRAY_SIZE(queue->fileio.blocks); i++) {
drivers/iio/buffer/industrialio-buffer-dma.c
429
if (!queue->fileio.blocks[i])
drivers/iio/buffer/industrialio-buffer-dma.c
431
iio_buffer_block_put(queue->fileio.blocks[i]);
drivers/iio/buffer/industrialio-buffer-dma.c
432
queue->fileio.blocks[i] = NULL;
drivers/iio/buffer/industrialio-buffer-dma.c
434
queue->fileio.active_block = NULL;
drivers/iio/buffer/industrialio-buffer-dma.c
437
static void iio_dma_buffer_submit_block(struct iio_dma_buffer_queue *queue,
drivers/iio/buffer/industrialio-buffer-dma.c
447
if (!queue->ops)
drivers/iio/buffer/industrialio-buffer-dma.c
453
ret = queue->ops->submit(queue, block);
drivers/iio/buffer/industrialio-buffer-dma.c
484
struct iio_dma_buffer_queue *queue = iio_buffer_to_queue(buffer);
drivers/iio/buffer/industrialio-buffer-dma.c
487
guard(mutex)(&queue->lock);
drivers/iio/buffer/industrialio-buffer-dma.c
488
queue->active = true;
drivers/iio/buffer/industrialio-buffer-dma.c
489
list_for_each_entry_safe(block, _block, &queue->incoming, head) {
drivers/iio/buffer/industrialio-buffer-dma.c
491
iio_dma_buffer_submit_block(queue, block);
drivers/iio/buffer/industrialio-buffer-dma.c
508
struct iio_dma_buffer_queue *queue = iio_buffer_to_queue(buffer);
drivers/iio/buffer/industrialio-buffer-dma.c
510
guard(mutex)(&queue->lock);
drivers/iio/buffer/industrialio-buffer-dma.c
511
queue->active = false;
drivers/iio/buffer/industrialio-buffer-dma.c
513
if (queue->ops && queue->ops->abort)
drivers/iio/buffer/industrialio-buffer-dma.c
514
queue->ops->abort(queue);
drivers/iio/buffer/industrialio-buffer-dma.c
520
static void iio_dma_buffer_enqueue(struct iio_dma_buffer_queue *queue,
drivers/iio/buffer/industrialio-buffer-dma.c
525
} else if (queue->active) {
drivers/iio/buffer/industrialio-buffer-dma.c
526
iio_dma_buffer_submit_block(queue, block);
drivers/iio/buffer/industrialio-buffer-dma.c
529
list_add_tail(&block->head, &queue->incoming);
drivers/iio/buffer/industrialio-buffer-dma.c
534
iio_dma_buffer_dequeue(struct iio_dma_buffer_queue *queue)
drivers/iio/buffer/industrialio-buffer-dma.c
539
guard(spinlock_irq)(&queue->list_lock);
drivers/iio/buffer/industrialio-buffer-dma.c
541
idx = queue->fileio.next_dequeue;
drivers/iio/buffer/industrialio-buffer-dma.c
542
block = queue->fileio.blocks[idx];
drivers/iio/buffer/industrialio-buffer-dma.c
547
idx = (idx + 1) % ARRAY_SIZE(queue->fileio.blocks);
drivers/iio/buffer/industrialio-buffer-dma.c
548
queue->fileio.next_dequeue = idx;
drivers/iio/buffer/industrialio-buffer-dma.c
556
struct iio_dma_buffer_queue *queue = iio_buffer_to_queue(buffer);
drivers/iio/buffer/industrialio-buffer-dma.c
564
guard(mutex)(&queue->lock);
drivers/iio/buffer/industrialio-buffer-dma.c
566
if (!queue->fileio.active_block) {
drivers/iio/buffer/industrialio-buffer-dma.c
567
block = iio_dma_buffer_dequeue(queue);
drivers/iio/buffer/industrialio-buffer-dma.c
571
queue->fileio.pos = 0;
drivers/iio/buffer/industrialio-buffer-dma.c
572
queue->fileio.active_block = block;
drivers/iio/buffer/industrialio-buffer-dma.c
574
block = queue->fileio.active_block;
drivers/iio/buffer/industrialio-buffer-dma.c
578
if (n > block->bytes_used - queue->fileio.pos)
drivers/iio/buffer/industrialio-buffer-dma.c
579
n = block->bytes_used - queue->fileio.pos;
drivers/iio/buffer/industrialio-buffer-dma.c
580
addr = block->vaddr + queue->fileio.pos;
drivers/iio/buffer/industrialio-buffer-dma.c
589
queue->fileio.pos += n;
drivers/iio/buffer/industrialio-buffer-dma.c
591
if (queue->fileio.pos == block->bytes_used) {
drivers/iio/buffer/industrialio-buffer-dma.c
592
queue->fileio.active_block = NULL;
drivers/iio/buffer/industrialio-buffer-dma.c
593
iio_dma_buffer_enqueue(queue, block);
drivers/iio/buffer/industrialio-buffer-dma.c
642
struct iio_dma_buffer_queue *queue = iio_buffer_to_queue(buf);
drivers/iio/buffer/industrialio-buffer-dma.c
654
guard(mutex)(&queue->lock);
drivers/iio/buffer/industrialio-buffer-dma.c
655
if (queue->fileio.active_block)
drivers/iio/buffer/industrialio-buffer-dma.c
656
data_available += queue->fileio.active_block->size;
drivers/iio/buffer/industrialio-buffer-dma.c
658
guard(spinlock_irq)(&queue->list_lock);
drivers/iio/buffer/industrialio-buffer-dma.c
660
for (i = 0; i < ARRAY_SIZE(queue->fileio.blocks); i++) {
drivers/iio/buffer/industrialio-buffer-dma.c
661
block = queue->fileio.blocks[i];
drivers/iio/buffer/industrialio-buffer-dma.c
663
if (block != queue->fileio.active_block && block->state == IIO_BLOCK_STATE_DONE)
drivers/iio/buffer/industrialio-buffer-dma.c
675
struct iio_dma_buffer_queue *queue = iio_buffer_to_queue(buffer);
drivers/iio/buffer/industrialio-buffer-dma.c
678
guard(mutex)(&queue->lock);
drivers/iio/buffer/industrialio-buffer-dma.c
684
if (queue->fileio.enabled)
drivers/iio/buffer/industrialio-buffer-dma.c
687
block = iio_dma_buffer_alloc_block(queue, attach->dmabuf->size, false);
drivers/iio/buffer/industrialio-buffer-dma.c
692
iio_dma_buffer_fileio_free(queue);
drivers/iio/buffer/industrialio-buffer-dma.c
708
struct iio_dma_buffer_queue *queue = block->queue;
drivers/iio/buffer/industrialio-buffer-dma.c
711
if (queue->fileio.enabled)
drivers/iio/buffer/industrialio-buffer-dma.c
733
struct iio_dma_buffer_queue *queue = iio_buffer_to_queue(buffer);
drivers/iio/buffer/industrialio-buffer-dma.c
737
lockdep_assert_held(&queue->lock);
drivers/iio/buffer/industrialio-buffer-dma.c
750
iio_dma_buffer_enqueue(queue, block);
drivers/iio/buffer/industrialio-buffer-dma.c
767
struct iio_dma_buffer_queue *queue = iio_buffer_to_queue(buffer);
drivers/iio/buffer/industrialio-buffer-dma.c
769
mutex_lock(&queue->lock);
drivers/iio/buffer/industrialio-buffer-dma.c
775
struct iio_dma_buffer_queue *queue = iio_buffer_to_queue(buffer);
drivers/iio/buffer/industrialio-buffer-dma.c
777
mutex_unlock(&queue->lock);
drivers/iio/buffer/industrialio-buffer-dma.c
827
void iio_dma_buffer_init(struct iio_dma_buffer_queue *queue, struct device *dev,
drivers/iio/buffer/industrialio-buffer-dma.c
830
iio_buffer_init(&queue->buffer);
drivers/iio/buffer/industrialio-buffer-dma.c
831
queue->buffer.length = PAGE_SIZE;
drivers/iio/buffer/industrialio-buffer-dma.c
832
queue->buffer.watermark = queue->buffer.length / 2;
drivers/iio/buffer/industrialio-buffer-dma.c
833
queue->dev = dev;
drivers/iio/buffer/industrialio-buffer-dma.c
834
queue->ops = ops;
drivers/iio/buffer/industrialio-buffer-dma.c
836
INIT_LIST_HEAD(&queue->incoming);
drivers/iio/buffer/industrialio-buffer-dma.c
838
mutex_init(&queue->lock);
drivers/iio/buffer/industrialio-buffer-dma.c
839
spin_lock_init(&queue->list_lock);
drivers/iio/buffer/industrialio-buffer-dma.c
850
void iio_dma_buffer_exit(struct iio_dma_buffer_queue *queue)
drivers/iio/buffer/industrialio-buffer-dma.c
852
guard(mutex)(&queue->lock);
drivers/iio/buffer/industrialio-buffer-dma.c
854
iio_dma_buffer_fileio_free(queue);
drivers/iio/buffer/industrialio-buffer-dma.c
855
queue->ops = NULL;
drivers/iio/buffer/industrialio-buffer-dma.c
867
void iio_dma_buffer_release(struct iio_dma_buffer_queue *queue)
drivers/iio/buffer/industrialio-buffer-dma.c
869
mutex_destroy(&queue->lock);
drivers/iio/buffer/industrialio-buffer-dmaengine.c
110
if (queue->buffer.direction == IIO_BUFFER_DIRECTION_IN)
drivers/iio/buffer/industrialio-buffer-dmaengine.c
132
scoped_guard(spinlock_irq, &dmaengine_buffer->queue.list_lock)
drivers/iio/buffer/industrialio-buffer-dmaengine.c
140
static void iio_dmaengine_buffer_abort(struct iio_dma_buffer_queue *queue)
drivers/iio/buffer/industrialio-buffer-dmaengine.c
143
iio_buffer_to_dmaengine_buffer(&queue->buffer);
drivers/iio/buffer/industrialio-buffer-dmaengine.c
146
iio_dma_buffer_block_list_abort(queue, &dmaengine_buffer->active);
drivers/iio/buffer/industrialio-buffer-dmaengine.c
154
iio_dma_buffer_release(&dmaengine_buffer->queue);
drivers/iio/buffer/industrialio-buffer-dmaengine.c
247
iio_dma_buffer_init(&dmaengine_buffer->queue, chan->device->dev,
drivers/iio/buffer/industrialio-buffer-dmaengine.c
250
dmaengine_buffer->queue.buffer.attrs = iio_dmaengine_buffer_attrs;
drivers/iio/buffer/industrialio-buffer-dmaengine.c
251
dmaengine_buffer->queue.buffer.access = &iio_dmaengine_buffer_ops;
drivers/iio/buffer/industrialio-buffer-dmaengine.c
253
return &dmaengine_buffer->queue.buffer;
drivers/iio/buffer/industrialio-buffer-dmaengine.c
267
iio_dma_buffer_exit(&dmaengine_buffer->queue);
drivers/iio/buffer/industrialio-buffer-dmaengine.c
34
struct iio_dma_buffer_queue queue;
drivers/iio/buffer/industrialio-buffer-dmaengine.c
45
return container_of(buffer, struct dmaengine_buffer, queue.buffer);
drivers/iio/buffer/industrialio-buffer-dmaengine.c
53
scoped_guard(spinlock_irqsave, &block->queue->list_lock)
drivers/iio/buffer/industrialio-buffer-dmaengine.c
59
static int iio_dmaengine_buffer_submit_block(struct iio_dma_buffer_queue *queue,
drivers/iio/buffer/industrialio-buffer-dmaengine.c
63
iio_buffer_to_dmaengine_buffer(&queue->buffer);
drivers/iio/buffer/industrialio-buffer-dmaengine.c
77
if (queue->buffer.direction == IIO_BUFFER_DIRECTION_IN)
drivers/infiniband/hw/cxgb4/cq.c
100
if (!cq->queue) {
drivers/infiniband/hw/cxgb4/cq.c
1067
(sizeof(*chp->cq.queue) / 2) : sizeof(*chp->cq.queue));
drivers/infiniband/hw/cxgb4/cq.c
108
((u8 *)cq->queue + (cq->size - 1) *
drivers/infiniband/hw/cxgb4/cq.c
109
(sizeof(*cq->queue) / 2)))->qp_err;
drivers/infiniband/hw/cxgb4/cq.c
112
((u8 *)cq->queue + (cq->size - 1) *
drivers/infiniband/hw/cxgb4/cq.c
113
sizeof(*cq->queue)))->qp_err;
drivers/infiniband/hw/cxgb4/cq.c
1130
mm->vaddr = chp->cq.queue;
drivers/infiniband/hw/cxgb4/cq.c
174
dma_free_coherent(&rdev->lldi.pdev->dev, cq->memsize, cq->queue,
drivers/infiniband/hw/cxgb4/cq.c
486
srq->queue[srq->size].status.host_wq_pidx =
drivers/infiniband/hw/cxgb4/cq.c
65
cq->memsize, cq->queue,
drivers/infiniband/hw/cxgb4/cq.c
98
cq->queue = dma_alloc_coherent(&rdev->lldi.pdev->dev, cq->memsize,
drivers/infiniband/hw/cxgb4/qp.c
102
dma_free_coherent(&(rdev->lldi.pdev->dev), sq->memsize, sq->queue,
drivers/infiniband/hw/cxgb4/qp.c
1139
wqe = (union t4_wr *)((u8 *)qhp->wq.sq.queue +
drivers/infiniband/hw/cxgb4/qp.c
123
sq->queue = (__force union t4_wr *)(rdev->oc_mw_kva + sq->dma_addr -
drivers/infiniband/hw/cxgb4/qp.c
1293
wqe = (union t4_recv_wr *)((u8 *)qhp->wq.rq.queue +
drivers/infiniband/hw/cxgb4/qp.c
131
sq->queue = dma_alloc_coherent(&(rdev->lldi.pdev->dev), sq->memsize,
drivers/infiniband/hw/cxgb4/qp.c
133
if (!sq->queue)
drivers/infiniband/hw/cxgb4/qp.c
135
sq->phys_addr = virt_to_phys(sq->queue);
drivers/infiniband/hw/cxgb4/qp.c
163
wq->rq.memsize, wq->rq.queue,
drivers/infiniband/hw/cxgb4/qp.c
2165
sizeof(*qhp->wq.sq.queue) + 16 * sizeof(__be64);
drivers/infiniband/hw/cxgb4/qp.c
2171
sizeof(*qhp->wq.rq.queue);
drivers/infiniband/hw/cxgb4/qp.c
2288
sq_key_mm->vaddr = qhp->wq.sq.queue;
drivers/infiniband/hw/cxgb4/qp.c
2296
rq_key_mm->vaddr = qhp->wq.rq.queue;
drivers/infiniband/hw/cxgb4/qp.c
2339
&qhp->wq.rq.queue[qhp->wq.rq.size].status.qp_err;
drivers/infiniband/hw/cxgb4/qp.c
2342
&qhp->wq.sq.queue[qhp->wq.sq.size].status.qp_err;
drivers/infiniband/hw/cxgb4/qp.c
2344
&qhp->wq.sq.queue[qhp->wq.sq.size].status.srqidx;
drivers/infiniband/hw/cxgb4/qp.c
2525
wq->memsize, wq->queue,
drivers/infiniband/hw/cxgb4/qp.c
2566
wq->queue = dma_alloc_coherent(&rdev->lldi.pdev->dev, wq->memsize,
drivers/infiniband/hw/cxgb4/qp.c
2568
if (!wq->queue)
drivers/infiniband/hw/cxgb4/qp.c
257
memset(wq->sq.queue, 0, wq->sq.memsize);
drivers/infiniband/hw/cxgb4/qp.c
261
wq->rq.queue = dma_alloc_coherent(&rdev->lldi.pdev->dev,
drivers/infiniband/hw/cxgb4/qp.c
2642
__func__, srq->idx, wq->qid, srq->pdid, wq->queue,
drivers/infiniband/hw/cxgb4/qp.c
2643
(u64)virt_to_phys(wq->queue), wq->bar2_va,
drivers/infiniband/hw/cxgb4/qp.c
2649
wq->memsize, wq->queue,
drivers/infiniband/hw/cxgb4/qp.c
265
if (!wq->rq.queue) {
drivers/infiniband/hw/cxgb4/qp.c
2670
dst = (u64 *)((u8 *)srq->queue + srq->wq_pidx * T4_EQ_ENTRY_SIZE);
drivers/infiniband/hw/cxgb4/qp.c
2673
if (dst >= (u64 *)&srq->queue[srq->size])
drivers/infiniband/hw/cxgb4/qp.c
2674
dst = (u64 *)srq->queue;
drivers/infiniband/hw/cxgb4/qp.c
2676
if (dst >= (u64 *)&srq->queue[srq->size])
drivers/infiniband/hw/cxgb4/qp.c
2677
dst = (u64 *)srq->queue;
drivers/infiniband/hw/cxgb4/qp.c
270
wq->sq.queue,
drivers/infiniband/hw/cxgb4/qp.c
271
(unsigned long long)virt_to_phys(wq->sq.queue),
drivers/infiniband/hw/cxgb4/qp.c
272
wq->rq.queue,
drivers/infiniband/hw/cxgb4/qp.c
273
(unsigned long long)virt_to_phys(wq->rq.queue));
drivers/infiniband/hw/cxgb4/qp.c
2743
sizeof(*srq->wq.queue);
drivers/infiniband/hw/cxgb4/qp.c
2786
srq_key_mm->vaddr = srq->wq.queue;
drivers/infiniband/hw/cxgb4/qp.c
391
wq->rq.memsize, wq->rq.queue,
drivers/infiniband/hw/cxgb4/qp.c
427
if (dstp == (u8 *)&sq->queue[sq->size])
drivers/infiniband/hw/cxgb4/qp.c
428
dstp = (u8 *)sq->queue;
drivers/infiniband/hw/cxgb4/qp.c
429
if (rem <= (u8 *)&sq->queue[sq->size] - dstp)
drivers/infiniband/hw/cxgb4/qp.c
432
len = (u8 *)&sq->queue[sq->size] - dstp;
drivers/infiniband/hw/cxgb4/qp.c
531
ret = build_isgl((__be64 *)sq->queue,
drivers/infiniband/hw/cxgb4/qp.c
532
(__be64 *)&sq->queue[sq->size],
drivers/infiniband/hw/cxgb4/qp.c
582
ret = build_isgl((__be64 *)sq->queue,
drivers/infiniband/hw/cxgb4/qp.c
583
(__be64 *)&sq->queue[sq->size],
drivers/infiniband/hw/cxgb4/qp.c
644
build_isgl((__be64 *)sq->queue, (__be64 *)&sq->queue[sq->size],
drivers/infiniband/hw/cxgb4/qp.c
648
build_isgl((__be64 *)sq->queue, (__be64 *)&sq->queue[sq->size],
drivers/infiniband/hw/cxgb4/qp.c
703
wqe = (union t4_wr *)((u8 *)qhp->wq.sq.queue +
drivers/infiniband/hw/cxgb4/qp.c
761
ret = build_isgl((__be64 *)qhp->wq.rq.queue,
drivers/infiniband/hw/cxgb4/qp.c
762
(__be64 *)&qhp->wq.rq.queue[qhp->wq.rq.size],
drivers/infiniband/hw/cxgb4/qp.c
866
if (++p == (__be64 *)&sq->queue[sq->size])
drivers/infiniband/hw/cxgb4/qp.c
867
p = (__be64 *)sq->queue;
drivers/infiniband/hw/cxgb4/qp.c
872
if (++p == (__be64 *)&sq->queue[sq->size])
drivers/infiniband/hw/cxgb4/qp.c
873
p = (__be64 *)sq->queue;
drivers/infiniband/hw/cxgb4/restrack.c
397
hwcqes[0] = chp->cq.queue[idx];
drivers/infiniband/hw/cxgb4/restrack.c
400
hwcqes[1] = chp->cq.queue[idx];
drivers/infiniband/hw/cxgb4/t4.h
333
union t4_wr *queue;
drivers/infiniband/hw/cxgb4/t4.h
362
union t4_recv_wr *queue;
drivers/infiniband/hw/cxgb4/t4.h
399
union t4_recv_wr *queue;
drivers/infiniband/hw/cxgb4/t4.h
438
srq->queue[srq->size].status.host_pidx = srq->pidx;
drivers/infiniband/hw/cxgb4/t4.h
468
srq->queue[srq->size].status.host_cidx = srq->cidx;
drivers/infiniband/hw/cxgb4/t4.h
477
srq->queue[srq->size].status.host_cidx = srq->cidx;
drivers/infiniband/hw/cxgb4/t4.h
514
return wq->rq.queue[wq->rq.size].status.host_wq_pidx;
drivers/infiniband/hw/cxgb4/t4.h
558
return wq->sq.queue[wq->sq.size].status.host_wq_pidx;
drivers/infiniband/hw/cxgb4/t4.h
664
wq->rq.queue[wq->rq.size].status.db_off = 1;
drivers/infiniband/hw/cxgb4/t4.h
669
wq->rq.queue[wq->rq.size].status.db_off = 0;
drivers/infiniband/hw/cxgb4/t4.h
677
struct t4_cqe *queue;
drivers/infiniband/hw/cxgb4/t4.h
756
cq->bits_type_ts = cq->queue[cq->cidx].bits_type_ts;
drivers/infiniband/hw/cxgb4/t4.h
777
return cq->sw_in_use || t4_valid_cqe(cq, &cq->queue[cq->cidx]);
drivers/infiniband/hw/cxgb4/t4.h
790
if (cq->queue[prev_cidx].bits_type_ts != cq->bits_type_ts) {
drivers/infiniband/hw/cxgb4/t4.h
794
} else if (t4_valid_cqe(cq, &cq->queue[cq->cidx])) {
drivers/infiniband/hw/cxgb4/t4.h
798
*cqe = &cq->queue[cq->cidx];
drivers/infiniband/hw/hfi1/tid_rdma.c
1496
goto queue;
drivers/infiniband/hw/hfi1/tid_rdma.c
1504
goto queue;
drivers/infiniband/hw/hfi1/tid_rdma.c
1536
queue:
drivers/infiniband/hw/hfi1/tid_rdma.c
3416
struct tid_queue *queue)
drivers/infiniband/hw/hfi1/tid_rdma.c
3418
return qpriv->tid_enqueue - queue->dequeue;
drivers/infiniband/hw/hfi1/tid_rdma.c
470
struct tid_queue *queue)
drivers/infiniband/hw/hfi1/tid_rdma.c
476
priv = list_first_entry_or_null(&queue->queue_head,
drivers/infiniband/hw/hfi1/tid_rdma.c
505
struct tid_queue *queue, struct rvt_qp *qp)
drivers/infiniband/hw/hfi1/tid_rdma.c
513
fqp = first_qp(rcd, queue);
drivers/infiniband/hw/hfi1/tid_rdma.c
538
struct tid_queue *queue, struct rvt_qp *qp)
drivers/infiniband/hw/hfi1/tid_rdma.c
549
queue->dequeue++;
drivers/infiniband/hw/hfi1/tid_rdma.c
565
struct tid_queue *queue, struct rvt_qp *qp)
drivers/infiniband/hw/hfi1/tid_rdma.c
574
list_add_tail(&priv->tid_wait, &queue->queue_head);
drivers/infiniband/hw/hfi1/tid_rdma.c
575
priv->tid_enqueue = ++queue->enqueue;
drivers/infiniband/hw/hfi1/tid_rdma.c
669
static void _tid_rdma_flush_wait(struct rvt_qp *qp, struct tid_queue *queue)
drivers/infiniband/hw/hfi1/tid_rdma.c
683
queue->dequeue++;
drivers/infiniband/hw/hfi1/tid_rdma.c
794
goto queue;
drivers/infiniband/hw/hfi1/tid_rdma.c
798
goto queue;
drivers/infiniband/hw/hfi1/tid_rdma.c
814
queue:
drivers/infiniband/hw/mana/cq.c
105
mana_ib_destroy_queue(mdev, &cq->queue);
drivers/infiniband/hw/mana/cq.c
125
mana_ib_destroy_queue(mdev, &cq->queue);
drivers/infiniband/hw/mana/cq.c
143
if (cq->queue.id >= gc->max_num_cqs)
drivers/infiniband/hw/mana/cq.c
146
WARN_ON(gc->cq_table[cq->queue.id]);
drivers/infiniband/hw/mana/cq.c
147
if (cq->queue.kmem)
drivers/infiniband/hw/mana/cq.c
148
gdma_cq = cq->queue.kmem;
drivers/infiniband/hw/mana/cq.c
157
gdma_cq->id = cq->queue.id;
drivers/infiniband/hw/mana/cq.c
158
gc->cq_table[cq->queue.id] = gdma_cq;
drivers/infiniband/hw/mana/cq.c
166
if (cq->queue.id >= gc->max_num_cqs || cq->queue.id == INVALID_QUEUE_ID)
drivers/infiniband/hw/mana/cq.c
169
if (cq->queue.kmem)
drivers/infiniband/hw/mana/cq.c
173
kfree(gc->cq_table[cq->queue.id]);
drivers/infiniband/hw/mana/cq.c
174
gc->cq_table[cq->queue.id] = NULL;
drivers/infiniband/hw/mana/cq.c
180
struct gdma_queue *gdma_cq = cq->queue.kmem;
drivers/infiniband/hw/mana/cq.c
326
struct gdma_queue *queue = cq->queue.kmem;
drivers/infiniband/hw/mana/cq.c
334
comp_read = mana_gd_poll_cq(queue, &gdma_cqe, 1);
drivers/infiniband/hw/mana/cq.c
47
&cq->queue);
drivers/infiniband/hw/mana/cq.c
63
err = mana_ib_create_kernel_queue(mdev, buf_size, GDMA_CQ, &cq->queue);
drivers/infiniband/hw/mana/cq.c
86
resp.cqid = cq->queue.id;
drivers/infiniband/hw/mana/main.c
245
struct mana_ib_queue *queue)
drivers/infiniband/hw/mana/main.c
250
queue->id = INVALID_QUEUE_ID;
drivers/infiniband/hw/mana/main.c
251
queue->gdma_region = GDMA_INVALID_DMA_REGION;
drivers/infiniband/hw/mana/main.c
255
err = mana_gd_create_mana_wq_cq(mdev->gdma_dev, &spec, &queue->kmem);
drivers/infiniband/hw/mana/main.c
259
queue->gdma_region = queue->kmem->mem_info.dma_region_handle;
drivers/infiniband/hw/mana/main.c
260
queue->kmem->mem_info.dma_region_handle = GDMA_INVALID_DMA_REGION;
drivers/infiniband/hw/mana/main.c
265
struct mana_ib_queue *queue)
drivers/infiniband/hw/mana/main.c
270
queue->umem = NULL;
drivers/infiniband/hw/mana/main.c
271
queue->id = INVALID_QUEUE_ID;
drivers/infiniband/hw/mana/main.c
272
queue->gdma_region = GDMA_INVALID_DMA_REGION;
drivers/infiniband/hw/mana/main.c
280
err = mana_ib_create_zero_offset_dma_region(mdev, umem, &queue->gdma_region);
drivers/infiniband/hw/mana/main.c
285
queue->umem = umem;
drivers/infiniband/hw/mana/main.c
287
ibdev_dbg(&mdev->ib_dev, "created dma region 0x%llx\n", queue->gdma_region);
drivers/infiniband/hw/mana/main.c
295
void mana_ib_destroy_queue(struct mana_ib_dev *mdev, struct mana_ib_queue *queue)
drivers/infiniband/hw/mana/main.c
300
mana_ib_gd_destroy_dma_region(mdev, queue->gdma_region);
drivers/infiniband/hw/mana/main.c
301
ib_umem_release(queue->umem);
drivers/infiniband/hw/mana/main.c
302
if (queue->kmem)
drivers/infiniband/hw/mana/main.c
303
mana_gd_destroy_queue(mdev_to_gc(mdev), queue->kmem);
drivers/infiniband/hw/mana/main.c
975
req.gdma_region = cq->queue.gdma_region;
drivers/infiniband/hw/mana/main.c
986
cq->queue.id = resp.cq_id;
drivers/infiniband/hw/mana/main.c
989
cq->queue.gdma_region = GDMA_INVALID_DMA_REGION;
drivers/infiniband/hw/mana/mana_ib.h
141
struct mana_ib_queue queue;
drivers/infiniband/hw/mana/mana_ib.h
626
struct mana_ib_queue *queue);
drivers/infiniband/hw/mana/mana_ib.h
628
struct mana_ib_queue *queue);
drivers/infiniband/hw/mana/mana_ib.h
629
void mana_ib_destroy_queue(struct mana_ib_dev *mdev, struct mana_ib_queue *queue);
drivers/infiniband/hw/mana/mana_ib.h
90
struct mana_ib_queue queue;
drivers/infiniband/hw/mana/qp.c
185
wq_spec.gdma_region = wq->queue.gdma_region;
drivers/infiniband/hw/mana/qp.c
188
cq_spec.gdma_region = cq->queue.gdma_region;
drivers/infiniband/hw/mana/qp.c
203
wq->queue.gdma_region = GDMA_INVALID_DMA_REGION;
drivers/infiniband/hw/mana/qp.c
204
cq->queue.gdma_region = GDMA_INVALID_DMA_REGION;
drivers/infiniband/hw/mana/qp.c
206
wq->queue.id = wq_spec.queue_index;
drivers/infiniband/hw/mana/qp.c
207
cq->queue.id = cq_spec.queue_index;
drivers/infiniband/hw/mana/qp.c
211
wq->rx_object, wq->queue.id, cq->queue.id);
drivers/infiniband/hw/mana/qp.c
213
resp.entries[i].cqid = cq->queue.id;
drivers/infiniband/hw/mana/qp.c
214
resp.entries[i].wqid = wq->queue.id;
drivers/infiniband/hw/mana/qp.c
339
cq_spec.gdma_region = send_cq->queue.gdma_region;
drivers/infiniband/hw/mana/qp.c
357
send_cq->queue.gdma_region = GDMA_INVALID_DMA_REGION;
drivers/infiniband/hw/mana/qp.c
360
send_cq->queue.id = cq_spec.queue_index;
drivers/infiniband/hw/mana/qp.c
369
qp->qp_handle, qp->raw_sq.id, send_cq->queue.id);
drivers/infiniband/hw/mana/qp.c
372
resp.cqid = send_cq->queue.id;
drivers/infiniband/hw/mana/shadow_queue.h
100
static inline void shadow_queue_advance_producer(struct shadow_queue *queue)
drivers/infiniband/hw/mana/shadow_queue.h
102
queue->prod_idx++;
drivers/infiniband/hw/mana/shadow_queue.h
105
static inline void shadow_queue_advance_consumer(struct shadow_queue *queue)
drivers/infiniband/hw/mana/shadow_queue.h
107
queue->cons_idx++;
drivers/infiniband/hw/mana/shadow_queue.h
110
static inline void shadow_queue_advance_next_to_complete(struct shadow_queue *queue)
drivers/infiniband/hw/mana/shadow_queue.h
112
queue->next_to_complete_idx++;
drivers/infiniband/hw/mana/shadow_queue.h
41
static inline int create_shadow_queue(struct shadow_queue *queue, uint32_t length, uint32_t stride)
drivers/infiniband/hw/mana/shadow_queue.h
43
queue->buffer = kvmalloc_array(length, stride, GFP_KERNEL);
drivers/infiniband/hw/mana/shadow_queue.h
44
if (!queue->buffer)
drivers/infiniband/hw/mana/shadow_queue.h
47
queue->length = length;
drivers/infiniband/hw/mana/shadow_queue.h
48
queue->stride = stride;
drivers/infiniband/hw/mana/shadow_queue.h
53
static inline void destroy_shadow_queue(struct shadow_queue *queue)
drivers/infiniband/hw/mana/shadow_queue.h
55
kvfree(queue->buffer);
drivers/infiniband/hw/mana/shadow_queue.h
58
static inline bool shadow_queue_full(struct shadow_queue *queue)
drivers/infiniband/hw/mana/shadow_queue.h
60
return (queue->prod_idx - queue->cons_idx) >= queue->length;
drivers/infiniband/hw/mana/shadow_queue.h
63
static inline bool shadow_queue_empty(struct shadow_queue *queue)
drivers/infiniband/hw/mana/shadow_queue.h
65
return queue->prod_idx == queue->cons_idx;
drivers/infiniband/hw/mana/shadow_queue.h
69
shadow_queue_get_element(const struct shadow_queue *queue, u64 unmasked_index)
drivers/infiniband/hw/mana/shadow_queue.h
71
u32 index = unmasked_index % queue->length;
drivers/infiniband/hw/mana/shadow_queue.h
73
return ((u8 *)queue->buffer + index * queue->stride);
drivers/infiniband/hw/mana/shadow_queue.h
77
shadow_queue_producer_entry(struct shadow_queue *queue)
drivers/infiniband/hw/mana/shadow_queue.h
79
return shadow_queue_get_element(queue, queue->prod_idx);
drivers/infiniband/hw/mana/shadow_queue.h
83
shadow_queue_get_next_to_consume(const struct shadow_queue *queue)
drivers/infiniband/hw/mana/shadow_queue.h
85
if (queue->cons_idx == queue->next_to_complete_idx)
drivers/infiniband/hw/mana/shadow_queue.h
88
return shadow_queue_get_element(queue, queue->cons_idx);
drivers/infiniband/hw/mana/shadow_queue.h
92
shadow_queue_get_next_to_complete(struct shadow_queue *queue)
drivers/infiniband/hw/mana/shadow_queue.h
94
if (queue->next_to_complete_idx == queue->prod_idx)
drivers/infiniband/hw/mana/shadow_queue.h
97
return shadow_queue_get_element(queue, queue->next_to_complete_idx);
drivers/infiniband/hw/mana/wq.c
34
err = mana_ib_create_queue(mdev, ucmd.wq_buf_addr, ucmd.wq_buf_size, &wq->queue);
drivers/infiniband/hw/mana/wq.c
67
mana_ib_destroy_queue(mdev, &wq->queue);
drivers/infiniband/hw/mana/wr.c
128
err = mana_gd_post_work_request(queue, &wqe_req, &wqe_info);
drivers/infiniband/hw/mana/wr.c
13
struct gdma_queue *queue = qp->ud_qp.queues[MANA_UD_RECV_QUEUE].kmem;
drivers/infiniband/hw/mana/wr.c
140
mana_gd_wq_ring_doorbell(mdev_to_gc(mdev), queue);
drivers/infiniband/hw/mana/wr.c
34
err = mana_gd_post_work_request(queue, &wqe_req, &wqe_info);
drivers/infiniband/hw/mana/wr.c
45
mana_gd_wq_ring_doorbell(mdev_to_gc(mdev), queue);
drivers/infiniband/hw/mana/wr.c
80
struct gdma_queue *queue = qp->ud_qp.queues[MANA_UD_SEND_QUEUE].kmem;
drivers/infiniband/hw/mthca/mthca_cq.c
163
return buf->queue.direct.buf + (entry * MTHCA_CQ_ENTRY_SIZE);
drivers/infiniband/hw/mthca/mthca_cq.c
165
return buf->queue.page_list[entry * MTHCA_CQ_ENTRY_SIZE / PAGE_SIZE].buf
drivers/infiniband/hw/mthca/mthca_cq.c
355
&buf->queue, &buf->is_direct,
drivers/infiniband/hw/mthca/mthca_cq.c
368
mthca_buf_free(dev, (cqe + 1) * MTHCA_CQ_ENTRY_SIZE, &buf->queue,
drivers/infiniband/hw/mthca/mthca_provider.h
168
union mthca_buf queue;
drivers/infiniband/hw/mthca/mthca_provider.h
221
union mthca_buf queue;
drivers/infiniband/hw/mthca/mthca_provider.h
274
union mthca_buf queue;
drivers/infiniband/hw/mthca/mthca_qp.c
1072
&qp->queue, &qp->is_direct, pd, 0, &qp->mr);
drivers/infiniband/hw/mthca/mthca_qp.c
1088
&qp->queue, qp->is_direct, &qp->mr);
drivers/infiniband/hw/mthca/mthca_qp.c
211
return qp->queue.direct.buf + (n << qp->rq.wqe_shift);
drivers/infiniband/hw/mthca/mthca_qp.c
213
return qp->queue.page_list[(n << qp->rq.wqe_shift) >> PAGE_SHIFT].buf +
drivers/infiniband/hw/mthca/mthca_qp.c
220
return qp->queue.direct.buf + qp->send_wqe_offset +
drivers/infiniband/hw/mthca/mthca_qp.c
223
return qp->queue.page_list[(qp->send_wqe_offset +
drivers/infiniband/hw/mthca/mthca_srq.c
149
mthca_buf_free(dev, srq->max << srq->wqe_shift, &srq->queue,
drivers/infiniband/hw/mthca/mthca_srq.c
171
&srq->queue, &srq->is_direct, pd, 1, &srq->mr);
drivers/infiniband/hw/mthca/mthca_srq.c
77
return srq->queue.direct.buf + (n << srq->wqe_shift);
drivers/infiniband/hw/mthca/mthca_srq.c
79
return srq->queue.page_list[(n << srq->wqe_shift) >> PAGE_SHIFT].buf +
drivers/infiniband/sw/rdmavt/cq.c
256
cq->queue = u_wc;
drivers/infiniband/sw/rdmavt/cq.c
319
if (cq->queue) {
drivers/infiniband/sw/rdmavt/cq.c
320
if (RDMA_READ_UAPI_ATOMIC(cq->queue->head) !=
drivers/infiniband/sw/rdmavt/cq.c
321
RDMA_READ_UAPI_ATOMIC(cq->queue->tail))
drivers/infiniband/sw/rdmavt/cq.c
386
old_u_wc = cq->queue;
drivers/infiniband/sw/rdmavt/cq.c
40
u_wc = cq->queue;
drivers/infiniband/sw/rdmavt/cq.c
421
cq->queue = u_wc;
drivers/infiniband/sw/rxe/rxe_comp.c
149
wqe = queue_head(qp->sq.queue, QUEUE_TYPE_FROM_CLIENT);
drivers/infiniband/sw/rxe/rxe_comp.c
457
queue_advance_consumer(qp->sq.queue, QUEUE_TYPE_FROM_CLIENT);
drivers/infiniband/sw/rxe/rxe_comp.c
591
struct rxe_queue *q = qp->sq.queue;
drivers/infiniband/sw/rxe/rxe_comp.c
595
if (!qp->sq.queue)
drivers/infiniband/sw/rxe/rxe_cq.c
108
addr = queue_producer_addr(cq->queue, QUEUE_TYPE_TO_CLIENT);
drivers/infiniband/sw/rxe/rxe_cq.c
111
queue_advance_producer(cq->queue, QUEUE_TYPE_TO_CLIENT);
drivers/infiniband/sw/rxe/rxe_cq.c
128
if (cq->queue)
drivers/infiniband/sw/rxe/rxe_cq.c
129
rxe_queue_cleanup(cq->queue);
drivers/infiniband/sw/rxe/rxe_cq.c
28
count = queue_count(cq->queue, QUEUE_TYPE_TO_CLIENT);
drivers/infiniband/sw/rxe/rxe_cq.c
50
cq->queue = rxe_queue_init(rxe, &cqe,
drivers/infiniband/sw/rxe/rxe_cq.c
52
if (!cq->queue) {
drivers/infiniband/sw/rxe/rxe_cq.c
58
cq->queue->buf, cq->queue->buf_size, &cq->queue->ip);
drivers/infiniband/sw/rxe/rxe_cq.c
75
err = rxe_queue_resize(cq->queue, (unsigned int *)&cqe,
drivers/infiniband/sw/rxe/rxe_cq.c
94
full = queue_full(cq->queue, QUEUE_TYPE_TO_CLIENT);
drivers/infiniband/sw/rxe/rxe_qp.c
249
qp->sq.queue = rxe_queue_init(rxe, &qp->sq.max_wr, wqe_size,
drivers/infiniband/sw/rxe/rxe_qp.c
251
if (!qp->sq.queue) {
drivers/infiniband/sw/rxe/rxe_qp.c
259
qp->sq.queue->buf, qp->sq.queue->buf_size,
drivers/infiniband/sw/rxe/rxe_qp.c
260
&qp->sq.queue->ip);
drivers/infiniband/sw/rxe/rxe_qp.c
276
vfree(qp->sq.queue->buf);
drivers/infiniband/sw/rxe/rxe_qp.c
277
kfree(qp->sq.queue);
drivers/infiniband/sw/rxe/rxe_qp.c
278
qp->sq.queue = NULL;
drivers/infiniband/sw/rxe/rxe_qp.c
311
qp->req.wqe_index = queue_get_producer(qp->sq.queue,
drivers/infiniband/sw/rxe/rxe_qp.c
340
qp->rq.queue = rxe_queue_init(rxe, &qp->rq.max_wr, wqe_size,
drivers/infiniband/sw/rxe/rxe_qp.c
342
if (!qp->rq.queue) {
drivers/infiniband/sw/rxe/rxe_qp.c
350
qp->rq.queue->buf, qp->rq.queue->buf_size,
drivers/infiniband/sw/rxe/rxe_qp.c
351
&qp->rq.queue->ip);
drivers/infiniband/sw/rxe/rxe_qp.c
365
vfree(qp->rq.queue->buf);
drivers/infiniband/sw/rxe/rxe_qp.c
366
kfree(qp->rq.queue);
drivers/infiniband/sw/rxe/rxe_qp.c
367
qp->rq.queue = NULL;
drivers/infiniband/sw/rxe/rxe_qp.c
441
rxe_queue_cleanup(qp->sq.queue);
drivers/infiniband/sw/rxe/rxe_qp.c
442
qp->sq.queue = NULL;
drivers/infiniband/sw/rxe/rxe_qp.c
572
if (qp->rq.queue)
drivers/infiniband/sw/rxe/rxe_qp.c
573
rxe_queue_reset(qp->rq.queue);
drivers/infiniband/sw/rxe/rxe_qp.c
574
if (qp->sq.queue)
drivers/infiniband/sw/rxe/rxe_qp.c
575
rxe_queue_reset(qp->sq.queue);
drivers/infiniband/sw/rxe/rxe_qp.c
883
if (qp->sq.queue)
drivers/infiniband/sw/rxe/rxe_qp.c
884
rxe_queue_cleanup(qp->sq.queue);
drivers/infiniband/sw/rxe/rxe_qp.c
889
if (qp->rq.queue)
drivers/infiniband/sw/rxe/rxe_qp.c
890
rxe_queue_cleanup(qp->rq.queue);
drivers/infiniband/sw/rxe/rxe_queue.h
93
void rxe_queue_cleanup(struct rxe_queue *queue);
drivers/infiniband/sw/rxe/rxe_req.c
128
q = qp->sq.queue;
drivers/infiniband/sw/rxe/rxe_req.c
166
struct rxe_queue *q = qp->sq.queue;
drivers/infiniband/sw/rxe/rxe_req.c
215
return qp->req.wqe_index != queue_get_consumer(qp->sq.queue,
drivers/infiniband/sw/rxe/rxe_req.c
44
struct rxe_queue *q = qp->sq.queue;
drivers/infiniband/sw/rxe/rxe_req.c
57
wqe = queue_addr_from_index(qp->sq.queue, wqe_index);
drivers/infiniband/sw/rxe/rxe_req.c
586
qp->req.wqe_index = queue_next_index(qp->sq.queue,
drivers/infiniband/sw/rxe/rxe_req.c
637
qp->req.wqe_index = queue_next_index(qp->sq.queue, qp->req.wqe_index);
drivers/infiniband/sw/rxe/rxe_req.c
654
struct rxe_queue *q = qp->sq.queue;
drivers/infiniband/sw/rxe/rxe_req.c
760
qp->req.wqe_index = queue_next_index(qp->sq.queue,
drivers/infiniband/sw/rxe/rxe_req.c
827
qp->req.wqe_index = queue_next_index(qp->sq.queue, qp->req.wqe_index);
drivers/infiniband/sw/rxe/rxe_resp.c
1160
queue_advance_consumer(qp->rq.queue, QUEUE_TYPE_FROM_CLIENT);
drivers/infiniband/sw/rxe/rxe_resp.c
1474
struct rxe_queue *q = qp->rq.queue;
drivers/infiniband/sw/rxe/rxe_resp.c
1491
if (!qp->rq.queue)
drivers/infiniband/sw/rxe/rxe_resp.c
262
struct rxe_queue *q = srq->rq.queue;
drivers/infiniband/sw/rxe/rxe_resp.c
330
qp->resp.wqe = queue_head(qp->rq.queue,
drivers/infiniband/sw/rxe/rxe_srq.c
137
if (attr->srq_limit > srq->rq.queue->buf->index_mask) {
drivers/infiniband/sw/rxe/rxe_srq.c
140
srq->rq.queue->buf->index_mask);
drivers/infiniband/sw/rxe/rxe_srq.c
155
struct rxe_queue *q = srq->rq.queue;
drivers/infiniband/sw/rxe/rxe_srq.c
192
if (srq->rq.queue)
drivers/infiniband/sw/rxe/rxe_srq.c
193
rxe_queue_cleanup(srq->rq.queue);
drivers/infiniband/sw/rxe/rxe_srq.c
88
srq->rq.queue = q;
drivers/infiniband/sw/rxe/rxe_verbs.c
1002
recv_wqe = queue_producer_addr(rq->queue, QUEUE_TYPE_FROM_ULP);
drivers/infiniband/sw/rxe/rxe_verbs.c
1013
queue_advance_producer(rq->queue, QUEUE_TYPE_FROM_ULP);
drivers/infiniband/sw/rxe/rxe_verbs.c
1174
cqe = queue_head(cq->queue, QUEUE_TYPE_TO_ULP);
drivers/infiniband/sw/rxe/rxe_verbs.c
1179
queue_advance_consumer(cq->queue, QUEUE_TYPE_TO_ULP);
drivers/infiniband/sw/rxe/rxe_verbs.c
1191
count = queue_count(cq->queue, QUEUE_TYPE_TO_ULP);
drivers/infiniband/sw/rxe/rxe_verbs.c
1205
empty = queue_empty(cq->queue, QUEUE_TYPE_TO_ULP);
drivers/infiniband/sw/rxe/rxe_verbs.c
499
attr->max_wr = srq->rq.queue->buf->index_mask;
drivers/infiniband/sw/rxe/rxe_verbs.c
892
full = queue_full(sq->queue, QUEUE_TYPE_FROM_ULP);
drivers/infiniband/sw/rxe/rxe_verbs.c
898
send_wqe = queue_producer_addr(sq->queue, QUEUE_TYPE_FROM_ULP);
drivers/infiniband/sw/rxe/rxe_verbs.c
901
queue_advance_producer(sq->queue, QUEUE_TYPE_FROM_ULP);
drivers/infiniband/sw/rxe/rxe_verbs.c
979
full = queue_full(rq->queue, QUEUE_TYPE_FROM_ULP);
drivers/infiniband/sw/rxe/rxe_verbs.h
63
struct rxe_queue *queue;
drivers/infiniband/sw/rxe/rxe_verbs.h
83
struct rxe_queue *queue;
drivers/infiniband/sw/rxe/rxe_verbs.h
91
struct rxe_queue *queue;
drivers/infiniband/sw/siw/siw.h
203
struct siw_cqe *queue;
drivers/infiniband/sw/siw/siw_cq.c
55
cqe = &cq->queue[cq->cq_get % cq->num_cqe];
drivers/infiniband/sw/siw/siw_qp.c
1046
cqe = &cq->queue[idx];
drivers/infiniband/sw/siw/siw_qp.c
1103
cqe = &cq->queue[idx];
drivers/infiniband/sw/siw/siw_verbs.c
1128
vfree(cq->queue);
drivers/infiniband/sw/siw/siw_verbs.c
1168
cq->queue = vmalloc_user(size * sizeof(struct siw_cqe) +
drivers/infiniband/sw/siw/siw_verbs.c
1171
cq->queue = vzalloc(size * sizeof(struct siw_cqe) +
drivers/infiniband/sw/siw/siw_verbs.c
1174
if (cq->queue == NULL) {
drivers/infiniband/sw/siw/siw_verbs.c
1183
cq->notify = (struct siw_cq_ctrl *)&cq->queue[size];
drivers/infiniband/sw/siw/siw_verbs.c
1194
siw_mmap_entry_insert(ctx, cq->queue,
drivers/infiniband/sw/siw/siw_verbs.c
1217
if (cq->queue) {
drivers/infiniband/sw/siw/siw_verbs.c
1223
vfree(cq->queue);
drivers/infiniband/ulp/ipoib/ipoib.h
430
struct sk_buff_head queue;
drivers/infiniband/ulp/ipoib/ipoib.h
448
struct sk_buff_head queue;
drivers/infiniband/ulp/ipoib/ipoib_cm.c
1036
while ((skb = __skb_dequeue(&p->neigh->queue)))
drivers/infiniband/ulp/ipoib/ipoib_main.c
1037
if (skb_queue_len(&neigh->queue) <
drivers/infiniband/ulp/ipoib/ipoib_main.c
1040
__skb_queue_tail(&neigh->queue, skb);
drivers/infiniband/ulp/ipoib/ipoib_main.c
1043
skb_queue_len(&neigh->queue));
drivers/infiniband/ulp/ipoib/ipoib_main.c
1058
if (skb_queue_len(&neigh->queue) < IPOIB_MAX_PATH_REC_QUEUE) {
drivers/infiniband/ulp/ipoib/ipoib_main.c
1060
__skb_queue_tail(&neigh->queue, skb);
drivers/infiniband/ulp/ipoib/ipoib_main.c
1114
if (skb_queue_len(&path->queue) < IPOIB_MAX_PATH_REC_QUEUE) {
drivers/infiniband/ulp/ipoib/ipoib_main.c
1116
__skb_queue_tail(&path->queue, skb);
drivers/infiniband/ulp/ipoib/ipoib_main.c
1212
if (skb_queue_len(&neigh->queue) < IPOIB_MAX_PATH_REC_QUEUE) {
drivers/infiniband/ulp/ipoib/ipoib_main.c
1215
__skb_queue_tail(&neigh->queue, skb);
drivers/infiniband/ulp/ipoib/ipoib_main.c
1367
if (likely(skb_queue_len(&neigh->queue) < IPOIB_MAX_PATH_REC_QUEUE))
drivers/infiniband/ulp/ipoib/ipoib_main.c
1452
skb_queue_head_init(&neigh->queue);
drivers/infiniband/ulp/ipoib/ipoib_main.c
1525
while ((skb = __skb_dequeue(&neigh->queue))) {
drivers/infiniband/ulp/ipoib/ipoib_main.c
656
while ((skb = __skb_dequeue(&path->queue)))
drivers/infiniband/ulp/ipoib/ipoib_main.c
846
while ((skb = __skb_dequeue(&path->queue)))
drivers/infiniband/ulp/ipoib/ipoib_main.c
875
while ((skb = __skb_dequeue(&neigh->queue)))
drivers/infiniband/ulp/ipoib/ipoib_main.c
931
skb_queue_head_init(&path->queue);
drivers/input/rmi4/rmi_f54.c
123
struct vb2_queue queue;
drivers/input/rmi4/rmi_f54.c
706
f54->queue = rmi_f54_queue;
drivers/input/rmi4/rmi_f54.c
707
f54->queue.drv_priv = f54;
drivers/input/rmi4/rmi_f54.c
708
f54->queue.lock = &f54->lock;
drivers/input/rmi4/rmi_f54.c
709
f54->queue.dev = &fn->dev;
drivers/input/rmi4/rmi_f54.c
711
ret = vb2_queue_init(&f54->queue);
drivers/input/rmi4/rmi_f54.c
719
f54->vdev.queue = &f54->queue;
drivers/input/serio/serio_raw.c
136
*c = serio_raw->queue[serio_raw->tail];
drivers/input/serio/serio_raw.c
255
serio_raw->queue[head] = data;
drivers/input/serio/serio_raw.c
29
unsigned char queue[SERIO_RAW_QUEUE_LEN];
drivers/input/touchscreen/atmel_mxt_ts.c
255
struct vb2_queue queue;
drivers/input/touchscreen/atmel_mxt_ts.c
2752
dbg->queue = mxt_queue;
drivers/input/touchscreen/atmel_mxt_ts.c
2753
dbg->queue.drv_priv = data;
drivers/input/touchscreen/atmel_mxt_ts.c
2754
dbg->queue.lock = &dbg->lock;
drivers/input/touchscreen/atmel_mxt_ts.c
2755
dbg->queue.dev = &data->client->dev;
drivers/input/touchscreen/atmel_mxt_ts.c
2757
error = vb2_queue_init(&dbg->queue);
drivers/input/touchscreen/atmel_mxt_ts.c
2765
dbg->vdev.queue = &dbg->queue;
drivers/input/touchscreen/sur40.c
217
struct vb2_queue queue;
drivers/input/touchscreen/sur40.c
537
if (!vb2_start_streaming_called(&sur40->queue))
drivers/input/touchscreen/sur40.c
746
sur40->queue = sur40_queue;
drivers/input/touchscreen/sur40.c
747
sur40->queue.drv_priv = sur40;
drivers/input/touchscreen/sur40.c
748
sur40->queue.lock = &sur40->lock;
drivers/input/touchscreen/sur40.c
749
sur40->queue.dev = sur40->dev;
drivers/input/touchscreen/sur40.c
752
error = vb2_queue_init(&sur40->queue);
drivers/input/touchscreen/sur40.c
760
sur40->vdev.queue = &sur40->queue;
drivers/iommu/dma-iommu.c
303
struct iova_fq *queue;
drivers/iommu/dma-iommu.c
305
queue = vmalloc(struct_size(queue, entries, fq_size));
drivers/iommu/dma-iommu.c
306
if (!queue)
drivers/iommu/dma-iommu.c
308
iommu_dma_init_one_fq(queue, fq_size);
drivers/iommu/dma-iommu.c
309
cookie->single_fq = queue;
drivers/iommu/dma-iommu.c
317
struct iova_fq __percpu *queue;
drivers/iommu/dma-iommu.c
320
queue = __alloc_percpu(struct_size(queue, entries, fq_size),
drivers/iommu/dma-iommu.c
321
__alignof__(*queue));
drivers/iommu/dma-iommu.c
322
if (!queue)
drivers/iommu/dma-iommu.c
326
iommu_dma_init_one_fq(per_cpu_ptr(queue, cpu), fq_size);
drivers/iommu/dma-iommu.c
327
cookie->percpu_fq = queue;
drivers/iommu/io-pgfault.c
311
flush_workqueue(iopf_param->queue->wq);
drivers/iommu/io-pgfault.c
355
int iopf_queue_discard_partial(struct iopf_queue *queue)
drivers/iommu/io-pgfault.c
360
if (!queue)
drivers/iommu/io-pgfault.c
363
mutex_lock(&queue->lock);
drivers/iommu/io-pgfault.c
364
list_for_each_entry(iopf_param, &queue->devices, queue_list) {
drivers/iommu/io-pgfault.c
373
mutex_unlock(&queue->lock);
drivers/iommu/io-pgfault.c
385
int iopf_queue_add_device(struct iopf_queue *queue, struct device *dev)
drivers/iommu/io-pgfault.c
395
mutex_lock(&queue->lock);
drivers/iommu/io-pgfault.c
414
list_add(&fault_param->queue_list, &queue->devices);
drivers/iommu/io-pgfault.c
415
fault_param->queue = queue;
drivers/iommu/io-pgfault.c
421
mutex_unlock(&queue->lock);
drivers/iommu/io-pgfault.c
450
void iopf_queue_remove_device(struct iopf_queue *queue, struct device *dev)
drivers/iommu/io-pgfault.c
459
mutex_lock(&queue->lock);
drivers/iommu/io-pgfault.c
464
if (WARN_ON(!fault_param || fault_param->queue != queue))
drivers/iommu/io-pgfault.c
492
mutex_unlock(&queue->lock);
drivers/iommu/io-pgfault.c
504
struct iopf_queue *queue;
drivers/iommu/io-pgfault.c
506
queue = kzalloc_obj(*queue);
drivers/iommu/io-pgfault.c
507
if (!queue)
drivers/iommu/io-pgfault.c
516
queue->wq = alloc_workqueue("iopf_queue/%s", WQ_UNBOUND, 0, name);
drivers/iommu/io-pgfault.c
517
if (!queue->wq) {
drivers/iommu/io-pgfault.c
518
kfree(queue);
drivers/iommu/io-pgfault.c
522
INIT_LIST_HEAD(&queue->devices);
drivers/iommu/io-pgfault.c
523
mutex_init(&queue->lock);
drivers/iommu/io-pgfault.c
525
return queue;
drivers/iommu/io-pgfault.c
536
void iopf_queue_free(struct iopf_queue *queue)
drivers/iommu/io-pgfault.c
540
if (!queue)
drivers/iommu/io-pgfault.c
543
list_for_each_entry_safe(iopf_param, next, &queue->devices, queue_list)
drivers/iommu/io-pgfault.c
544
iopf_queue_remove_device(queue, iopf_param->dev);
drivers/iommu/io-pgfault.c
546
destroy_workqueue(queue->wq);
drivers/iommu/io-pgfault.c
547
kfree(queue);
drivers/iommu/iommu-sva.c
303
if (!queue_work(fault_param->queue->wq, &group->work))
drivers/iommu/riscv/iommu.c
132
struct riscv_iommu_queue *queue,
drivers/iommu/riscv/iommu.c
142
riscv_iommu_writeq(iommu, queue->qbr, RISCV_IOMMU_QUEUE_LOG2SZ_FIELD);
drivers/iommu/riscv/iommu.c
143
qb = riscv_iommu_readq(iommu, queue->qbr);
drivers/iommu/riscv/iommu.c
150
logsz = ilog2(queue->mask);
drivers/iommu/riscv/iommu.c
162
queue->phys = pfn_to_phys(FIELD_GET(RISCV_IOMMU_PPN_FIELD, qb));
drivers/iommu/riscv/iommu.c
163
queue->base = devm_ioremap(iommu->dev, queue->phys, queue_size);
drivers/iommu/riscv/iommu.c
168
queue->base = riscv_iommu_get_pages(
drivers/iommu/riscv/iommu.c
170
queue->phys = __pa(queue->base);
drivers/iommu/riscv/iommu.c
171
} while (!queue->base && logsz-- > 0);
drivers/iommu/riscv/iommu.c
174
if (!queue->base)
drivers/iommu/riscv/iommu.c
177
qb = phys_to_ppn(queue->phys) |
drivers/iommu/riscv/iommu.c
181
riscv_iommu_writeq(iommu, queue->qbr, qb);
drivers/iommu/riscv/iommu.c
182
rb = riscv_iommu_readq(iommu, queue->qbr);
drivers/iommu/riscv/iommu.c
184
dev_err(iommu->dev, "queue #%u allocation failed\n", queue->qid);
drivers/iommu/riscv/iommu.c
189
queue->mask = (2U << logsz) - 1;
drivers/iommu/riscv/iommu.c
192
queue->qid, logsz + 1);
drivers/iommu/riscv/iommu.c
200
struct riscv_iommu_queue *queue = (struct riscv_iommu_queue *)data;
drivers/iommu/riscv/iommu.c
202
if (riscv_iommu_readl(queue->iommu, RISCV_IOMMU_REG_IPSR) & Q_IPSR(queue))
drivers/iommu/riscv/iommu.c
221
struct riscv_iommu_queue *queue,
drivers/iommu/riscv/iommu.c
224
const unsigned int irq = iommu->irqs[riscv_iommu_queue_vec(iommu, queue->qid)];
drivers/iommu/riscv/iommu.c
228
if (queue->iommu)
drivers/iommu/riscv/iommu.c
235
queue->iommu = iommu;
drivers/iommu/riscv/iommu.c
238
dev_name(iommu->dev), queue);
drivers/iommu/riscv/iommu.c
240
queue->iommu = NULL;
drivers/iommu/riscv/iommu.c
245
if (queue->qid == RISCV_IOMMU_INTR_CQ)
drivers/iommu/riscv/iommu.c
246
riscv_iommu_writel(queue->iommu, Q_TAIL(queue), 0);
drivers/iommu/riscv/iommu.c
248
riscv_iommu_writel(queue->iommu, Q_HEAD(queue), 0);
drivers/iommu/riscv/iommu.c
256
riscv_iommu_writel(iommu, queue->qcr,
drivers/iommu/riscv/iommu.c
261
riscv_iommu_readl_timeout(iommu, queue->qcr,
drivers/iommu/riscv/iommu.c
269
riscv_iommu_writel(iommu, queue->qcr, 0);
drivers/iommu/riscv/iommu.c
270
free_irq(irq, queue);
drivers/iommu/riscv/iommu.c
271
queue->iommu = NULL;
drivers/iommu/riscv/iommu.c
272
dev_err(iommu->dev, "queue #%u failed to start\n", queue->qid);
drivers/iommu/riscv/iommu.c
277
riscv_iommu_writel(iommu, RISCV_IOMMU_REG_IPSR, Q_IPSR(queue));
drivers/iommu/riscv/iommu.c
286
static void riscv_iommu_queue_disable(struct riscv_iommu_queue *queue)
drivers/iommu/riscv/iommu.c
288
struct riscv_iommu_device *iommu = queue->iommu;
drivers/iommu/riscv/iommu.c
294
free_irq(iommu->irqs[riscv_iommu_queue_vec(iommu, queue->qid)], queue);
drivers/iommu/riscv/iommu.c
295
riscv_iommu_writel(iommu, queue->qcr, 0);
drivers/iommu/riscv/iommu.c
296
riscv_iommu_readl_timeout(iommu, queue->qcr,
drivers/iommu/riscv/iommu.c
302
queue->qid, csr);
drivers/iommu/riscv/iommu.c
304
queue->iommu = NULL;
drivers/iommu/riscv/iommu.c
311
static int riscv_iommu_queue_consume(struct riscv_iommu_queue *queue,
drivers/iommu/riscv/iommu.c
314
unsigned int head = atomic_read(&queue->head);
drivers/iommu/riscv/iommu.c
315
unsigned int tail = atomic_read(&queue->tail);
drivers/iommu/riscv/iommu.c
316
unsigned int last = Q_ITEM(queue, tail);
drivers/iommu/riscv/iommu.c
325
if (riscv_iommu_readl_timeout(queue->iommu, Q_TAIL(queue),
drivers/iommu/riscv/iommu.c
326
tail, (tail & ~queue->mask) == 0,
drivers/iommu/riscv/iommu.c
328
dev_err_once(queue->iommu->dev,
drivers/iommu/riscv/iommu.c
337
return (int)(atomic_add_return((tail - last) & queue->mask, &queue->tail) - head);
drivers/iommu/riscv/iommu.c
343
static void riscv_iommu_queue_release(struct riscv_iommu_queue *queue, int count)
drivers/iommu/riscv/iommu.c
345
const unsigned int head = atomic_add_return(count, &queue->head);
drivers/iommu/riscv/iommu.c
347
riscv_iommu_writel(queue->iommu, Q_HEAD(queue), Q_ITEM(queue, head));
drivers/iommu/riscv/iommu.c
351
static unsigned int riscv_iommu_queue_cons(struct riscv_iommu_queue *queue)
drivers/iommu/riscv/iommu.c
353
const unsigned int cons = atomic_read(&queue->head);
drivers/iommu/riscv/iommu.c
354
const unsigned int last = Q_ITEM(queue, cons);
drivers/iommu/riscv/iommu.c
357
if (riscv_iommu_readl_timeout(queue->iommu, Q_HEAD(queue), head,
drivers/iommu/riscv/iommu.c
358
!(head & ~queue->mask),
drivers/iommu/riscv/iommu.c
362
return cons + ((head - last) & queue->mask);
drivers/iommu/riscv/iommu.c
366
static int riscv_iommu_queue_wait(struct riscv_iommu_queue *queue,
drivers/iommu/riscv/iommu.c
370
unsigned int cons = atomic_read(&queue->head);
drivers/iommu/riscv/iommu.c
377
return readx_poll_timeout(riscv_iommu_queue_cons, queue, cons,
drivers/iommu/riscv/iommu.c
387
static unsigned int riscv_iommu_queue_send(struct riscv_iommu_queue *queue,
drivers/iommu/riscv/iommu.c
399
prod = atomic_inc_return(&queue->prod) - 1;
drivers/iommu/riscv/iommu.c
400
head = atomic_read(&queue->head);
drivers/iommu/riscv/iommu.c
403
if ((prod - head) > queue->mask) {
drivers/iommu/riscv/iommu.c
404
if (readx_poll_timeout(atomic_read, &queue->head,
drivers/iommu/riscv/iommu.c
405
head, (prod - head) < queue->mask,
drivers/iommu/riscv/iommu.c
408
} else if ((prod - head) == queue->mask) {
drivers/iommu/riscv/iommu.c
409
const unsigned int last = Q_ITEM(queue, head);
drivers/iommu/riscv/iommu.c
411
if (riscv_iommu_readl_timeout(queue->iommu, Q_HEAD(queue), head,
drivers/iommu/riscv/iommu.c
412
!(head & ~queue->mask) && head != last,
drivers/iommu/riscv/iommu.c
415
atomic_add((head - last) & queue->mask, &queue->head);
drivers/iommu/riscv/iommu.c
419
memcpy(queue->base + Q_ITEM(queue, prod) * entry_size, entry, entry_size);
drivers/iommu/riscv/iommu.c
422
if (readx_poll_timeout(atomic_read, &queue->tail, tail, prod == tail,
drivers/iommu/riscv/iommu.c
432
riscv_iommu_writel(queue->iommu, Q_TAIL(queue), Q_ITEM(queue, prod + 1));
drivers/iommu/riscv/iommu.c
439
atomic_inc(&queue->tail);
drivers/iommu/riscv/iommu.c
448
dev_err_once(queue->iommu->dev, "Hardware error: command enqueue failed\n");
drivers/iommu/riscv/iommu.c
460
const struct riscv_iommu_queue *queue = (struct riscv_iommu_queue *)data;
drivers/iommu/riscv/iommu.c
464
ctrl = riscv_iommu_readl(queue->iommu, queue->qcr);
drivers/iommu/riscv/iommu.c
467
riscv_iommu_writel(queue->iommu, queue->qcr, ctrl);
drivers/iommu/riscv/iommu.c
468
dev_warn(queue->iommu->dev,
drivers/iommu/riscv/iommu.c
470
queue->qid,
drivers/iommu/riscv/iommu.c
480
riscv_iommu_writel(queue->iommu, RISCV_IOMMU_REG_IPSR, Q_IPSR(queue));
drivers/iommu/riscv/iommu.c
530
struct riscv_iommu_queue *queue = (struct riscv_iommu_queue *)data;
drivers/iommu/riscv/iommu.c
531
struct riscv_iommu_device *iommu = queue->iommu;
drivers/iommu/riscv/iommu.c
536
events = (struct riscv_iommu_fq_record *)queue->base;
drivers/iommu/riscv/iommu.c
539
riscv_iommu_writel(iommu, RISCV_IOMMU_REG_IPSR, Q_IPSR(queue));
drivers/iommu/riscv/iommu.c
542
cnt = riscv_iommu_queue_consume(queue, &idx);
drivers/iommu/riscv/iommu.c
544
riscv_iommu_fault(iommu, &events[Q_ITEM(queue, idx)]);
drivers/iommu/riscv/iommu.c
545
riscv_iommu_queue_release(queue, cnt);
drivers/iommu/riscv/iommu.c
549
ctrl = riscv_iommu_readl(iommu, queue->qcr);
drivers/iommu/riscv/iommu.c
551
riscv_iommu_writel(iommu, queue->qcr, ctrl);
drivers/iommu/riscv/iommu.c
554
queue->qid,
drivers/ipack/devices/ipoctal.c
31
wait_queue_head_t queue;
drivers/mailbox/ti-msgmgr.c
24
#define Q_DATA_OFFSET(proxy, queue, reg) \
drivers/mailbox/ti-msgmgr.c
25
((0x10000 * (proxy)) + (0x80 * (queue)) + ((reg) * 4))
drivers/mailbox/ti-msgmgr.c
26
#define Q_STATE_OFFSET(queue) ((queue) * 0x4)
drivers/md/dm-cache-policy-smq.c
270
static void q_init(struct queue *q, struct entry_space *es, unsigned int nr_levels)
drivers/md/dm-cache-policy-smq.c
288
static unsigned int q_size(struct queue *q)
drivers/md/dm-cache-policy-smq.c
296
static void q_push(struct queue *q, struct entry *e)
drivers/md/dm-cache-policy-smq.c
306
static void q_push_front(struct queue *q, struct entry *e)
drivers/md/dm-cache-policy-smq.c
316
static void q_push_before(struct queue *q, struct entry *old, struct entry *e)
drivers/md/dm-cache-policy-smq.c
326
static void q_del(struct queue *q, struct entry *e)
drivers/md/dm-cache-policy-smq.c
336
static struct entry *q_peek(struct queue *q, unsigned int max_level, bool can_cross_sentinel)
drivers/md/dm-cache-policy-smq.c
358
static struct entry *q_pop(struct queue *q)
drivers/md/dm-cache-policy-smq.c
373
static struct entry *__redist_pop_from(struct queue *q, unsigned int level)
drivers/md/dm-cache-policy-smq.c
387
static void q_set_targets_subrange_(struct queue *q, unsigned int nr_elts,
drivers/md/dm-cache-policy-smq.c
407
static void q_set_targets(struct queue *q)
drivers/md/dm-cache-policy-smq.c
429
static void q_redistribute(struct queue *q)
drivers/md/dm-cache-policy-smq.c
472
static void q_requeue(struct queue *q, struct entry *e, unsigned int extra_levels,
drivers/md/dm-cache-policy-smq.c
826
struct queue hotspot;
drivers/md/dm-cache-policy-smq.c
827
struct queue clean;
drivers/md/dm-cache-policy-smq.c
828
struct queue dirty;
drivers/md/dm-cache-policy-smq.c
889
struct queue *q = &mq->dirty;
drivers/md/dm-cache-policy-smq.c
902
struct queue *q = &mq->clean;
drivers/md/dm-core.h
182
if (blk_queue_is_zoned(md->queue))
drivers/md/dm-core.h
69
struct request_queue *queue;
drivers/md/dm-rq.c
185
__dm_mq_kick_requeue_list(md->queue, 0);
drivers/md/dm-rq.c
227
blk_queue_disable_discard(tio->md->queue);
drivers/md/dm-rq.c
230
blk_queue_disable_write_zeroes(tio->md->queue);
drivers/md/dm-rq.c
573
err = blk_mq_init_allocated_queue(md->tag_set, md->queue);
drivers/md/dm-rq.c
61
return queue_is_mq(md->queue);
drivers/md/dm-table.c
1060
if (md->queue->limits.features & BLK_FEAT_POLL)
drivers/md/dm-table.c
1428
if (t->md->queue &&
drivers/md/dm-table.c
1430
t->md->queue->crypto_profile)) {
drivers/md/dm-table.c
2227
if (t->md->queue)
drivers/md/dm-table.c
2228
blk_mq_run_hw_queues(t->md->queue, true);
drivers/md/dm-vdo/completion.c
126
vdo_enqueue_work_queue(vdo->threads[thread_id].queue, completion);
drivers/md/dm-vdo/data-vio.c
1265
vdo_funnel_queue_put(pool->queue, &completion->work_queue_entry_link);
drivers/md/dm-vdo/data-vio.c
175
struct funnel_queue *queue;
drivers/md/dm-vdo/data-vio.c
716
struct funnel_queue_entry *entry = vdo_funnel_queue_poll(pool->queue);
drivers/md/dm-vdo/data-vio.c
746
reschedule = !vdo_is_funnel_queue_empty(pool->queue);
drivers/md/dm-vdo/data-vio.c
865
result = vdo_make_funnel_queue(&pool->queue);
drivers/md/dm-vdo/data-vio.c
925
vdo_free_funnel_queue(vdo_forget(pool->queue));
drivers/md/dm-vdo/dump.c
73
vdo_dump_work_queue(vdo->threads[id].queue);
drivers/md/dm-vdo/funnel-queue.c
103
struct funnel_queue_entry *vdo_funnel_queue_poll(struct funnel_queue *queue)
drivers/md/dm-vdo/funnel-queue.c
105
struct funnel_queue_entry *oldest = get_oldest(queue);
drivers/md/dm-vdo/funnel-queue.c
116
queue->oldest = READ_ONCE(oldest->next);
drivers/md/dm-vdo/funnel-queue.c
127
uds_prefetch_address(queue->oldest, true);
drivers/md/dm-vdo/funnel-queue.c
137
bool vdo_is_funnel_queue_empty(struct funnel_queue *queue)
drivers/md/dm-vdo/funnel-queue.c
139
return get_oldest(queue) == NULL;
drivers/md/dm-vdo/funnel-queue.c
148
bool vdo_is_funnel_queue_idle(struct funnel_queue *queue)
drivers/md/dm-vdo/funnel-queue.c
15
struct funnel_queue *queue;
drivers/md/dm-vdo/funnel-queue.c
154
if (queue->oldest != &queue->stub)
drivers/md/dm-vdo/funnel-queue.c
166
if (READ_ONCE(queue->newest) != &queue->stub)
drivers/md/dm-vdo/funnel-queue.c
17
result = vdo_allocate(1, struct funnel_queue, "funnel queue", &queue);
drivers/md/dm-vdo/funnel-queue.c
25
queue->stub.next = NULL;
drivers/md/dm-vdo/funnel-queue.c
26
queue->newest = &queue->stub;
drivers/md/dm-vdo/funnel-queue.c
27
queue->oldest = &queue->stub;
drivers/md/dm-vdo/funnel-queue.c
29
*queue_ptr = queue;
drivers/md/dm-vdo/funnel-queue.c
33
void vdo_free_funnel_queue(struct funnel_queue *queue)
drivers/md/dm-vdo/funnel-queue.c
35
vdo_free(queue);
drivers/md/dm-vdo/funnel-queue.c
38
static struct funnel_queue_entry *get_oldest(struct funnel_queue *queue)
drivers/md/dm-vdo/funnel-queue.c
45
struct funnel_queue_entry *oldest = queue->oldest;
drivers/md/dm-vdo/funnel-queue.c
48
if (oldest == &queue->stub) {
drivers/md/dm-vdo/funnel-queue.c
60
queue->oldest = oldest;
drivers/md/dm-vdo/funnel-queue.c
69
struct funnel_queue_entry *newest = READ_ONCE(queue->newest);
drivers/md/dm-vdo/funnel-queue.c
83
vdo_funnel_queue_put(queue, &queue->stub);
drivers/md/dm-vdo/funnel-queue.h
104
struct funnel_queue_entry *__must_check vdo_funnel_queue_poll(struct funnel_queue *queue);
drivers/md/dm-vdo/funnel-queue.h
106
bool __must_check vdo_is_funnel_queue_empty(struct funnel_queue *queue);
drivers/md/dm-vdo/funnel-queue.h
108
bool __must_check vdo_is_funnel_queue_idle(struct funnel_queue *queue);
drivers/md/dm-vdo/funnel-queue.h
72
void vdo_free_funnel_queue(struct funnel_queue *queue);
drivers/md/dm-vdo/funnel-queue.h
82
static inline void vdo_funnel_queue_put(struct funnel_queue *queue,
drivers/md/dm-vdo/funnel-queue.h
96
previous = xchg(&queue->newest, entry);
drivers/md/dm-vdo/funnel-workqueue.c
100
for (i = queue->common.type->max_priority; i >= 0; i--) {
drivers/md/dm-vdo/funnel-workqueue.c
101
struct funnel_queue_entry *link = vdo_funnel_queue_poll(queue->priority_lists[i]);
drivers/md/dm-vdo/funnel-workqueue.c
110
static void enqueue_work_queue_completion(struct simple_work_queue *queue,
drivers/md/dm-vdo/funnel-workqueue.c
115
completion, completion->callback, queue, completion->my_queue);
drivers/md/dm-vdo/funnel-workqueue.c
117
completion->priority = queue->common.type->default_priority;
drivers/md/dm-vdo/funnel-workqueue.c
119
if (VDO_ASSERT(completion->priority <= queue->common.type->max_priority,
drivers/md/dm-vdo/funnel-workqueue.c
123
completion->my_queue = &queue->common;
drivers/md/dm-vdo/funnel-workqueue.c
126
vdo_funnel_queue_put(queue->priority_lists[completion->priority],
drivers/md/dm-vdo/funnel-workqueue.c
146
if ((atomic_read(&queue->idle) != 1) || (atomic_cmpxchg(&queue->idle, 1, 0) != 1))
drivers/md/dm-vdo/funnel-workqueue.c
150
wake_up(&queue->waiting_worker_threads);
drivers/md/dm-vdo/funnel-workqueue.c
153
static void run_start_hook(struct simple_work_queue *queue)
drivers/md/dm-vdo/funnel-workqueue.c
155
if (queue->common.type->start != NULL)
drivers/md/dm-vdo/funnel-workqueue.c
156
queue->common.type->start(queue->private);
drivers/md/dm-vdo/funnel-workqueue.c
159
static void run_finish_hook(struct simple_work_queue *queue)
drivers/md/dm-vdo/funnel-workqueue.c
161
if (queue->common.type->finish != NULL)
drivers/md/dm-vdo/funnel-workqueue.c
162
queue->common.type->finish(queue->private);
drivers/md/dm-vdo/funnel-workqueue.c
174
static struct vdo_completion *wait_for_next_completion(struct simple_work_queue *queue)
drivers/md/dm-vdo/funnel-workqueue.c
180
prepare_to_wait(&queue->waiting_worker_threads, &wait,
drivers/md/dm-vdo/funnel-workqueue.c
190
atomic_set(&queue->idle, 1);
drivers/md/dm-vdo/funnel-workqueue.c
193
completion = poll_for_completion(queue);
drivers/md/dm-vdo/funnel-workqueue.c
211
completion = poll_for_completion(queue);
drivers/md/dm-vdo/funnel-workqueue.c
216
finish_wait(&queue->waiting_worker_threads, &wait);
drivers/md/dm-vdo/funnel-workqueue.c
217
atomic_set(&queue->idle, 0);
drivers/md/dm-vdo/funnel-workqueue.c
222
static void process_completion(struct simple_work_queue *queue,
drivers/md/dm-vdo/funnel-workqueue.c
225
if (VDO_ASSERT(completion->my_queue == &queue->common,
drivers/md/dm-vdo/funnel-workqueue.c
227
completion, queue, completion->my_queue) == VDO_SUCCESS)
drivers/md/dm-vdo/funnel-workqueue.c
233
static void service_work_queue(struct simple_work_queue *queue)
drivers/md/dm-vdo/funnel-workqueue.c
235
run_start_hook(queue);
drivers/md/dm-vdo/funnel-workqueue.c
238
struct vdo_completion *completion = poll_for_completion(queue);
drivers/md/dm-vdo/funnel-workqueue.c
241
completion = wait_for_next_completion(queue);
drivers/md/dm-vdo/funnel-workqueue.c
248
process_completion(queue, completion);
drivers/md/dm-vdo/funnel-workqueue.c
258
run_finish_hook(queue);
drivers/md/dm-vdo/funnel-workqueue.c
263
struct simple_work_queue *queue = ptr;
drivers/md/dm-vdo/funnel-workqueue.c
265
complete(queue->started);
drivers/md/dm-vdo/funnel-workqueue.c
266
service_work_queue(queue);
drivers/md/dm-vdo/funnel-workqueue.c
272
static void free_simple_work_queue(struct simple_work_queue *queue)
drivers/md/dm-vdo/funnel-workqueue.c
277
vdo_free_funnel_queue(queue->priority_lists[i]);
drivers/md/dm-vdo/funnel-workqueue.c
278
vdo_free(queue->common.name);
drivers/md/dm-vdo/funnel-workqueue.c
279
vdo_free(queue);
drivers/md/dm-vdo/funnel-workqueue.c
282
static void free_round_robin_work_queue(struct round_robin_work_queue *queue)
drivers/md/dm-vdo/funnel-workqueue.c
284
struct simple_work_queue **queue_table = queue->service_queues;
drivers/md/dm-vdo/funnel-workqueue.c
285
unsigned int count = queue->num_service_queues;
drivers/md/dm-vdo/funnel-workqueue.c
288
queue->service_queues = NULL;
drivers/md/dm-vdo/funnel-workqueue.c
293
vdo_free(queue->common.name);
drivers/md/dm-vdo/funnel-workqueue.c
294
vdo_free(queue);
drivers/md/dm-vdo/funnel-workqueue.c
297
void vdo_free_work_queue(struct vdo_work_queue *queue)
drivers/md/dm-vdo/funnel-workqueue.c
299
if (queue == NULL)
drivers/md/dm-vdo/funnel-workqueue.c
302
vdo_finish_work_queue(queue);
drivers/md/dm-vdo/funnel-workqueue.c
304
if (queue->round_robin_mode)
drivers/md/dm-vdo/funnel-workqueue.c
305
free_round_robin_work_queue(as_round_robin_work_queue(queue));
drivers/md/dm-vdo/funnel-workqueue.c
307
free_simple_work_queue(as_simple_work_queue(queue));
drivers/md/dm-vdo/funnel-workqueue.c
316
struct simple_work_queue *queue;
drivers/md/dm-vdo/funnel-workqueue.c
325
result = vdo_allocate(1, struct simple_work_queue, "simple work queue", &queue);
drivers/md/dm-vdo/funnel-workqueue.c
329
queue->private = private;
drivers/md/dm-vdo/funnel-workqueue.c
330
queue->started = &started;
drivers/md/dm-vdo/funnel-workqueue.c
331
queue->common.type = type;
drivers/md/dm-vdo/funnel-workqueue.c
332
queue->common.owner = owner;
drivers/md/dm-vdo/funnel-workqueue.c
333
init_waitqueue_head(&queue->waiting_worker_threads);
drivers/md/dm-vdo/funnel-workqueue.c
335
result = vdo_duplicate_string(name, "queue name", &queue->common.name);
drivers/md/dm-vdo/funnel-workqueue.c
337
vdo_free(queue);
drivers/md/dm-vdo/funnel-workqueue.c
342
result = vdo_make_funnel_queue(&queue->priority_lists[i]);
drivers/md/dm-vdo/funnel-workqueue.c
344
free_simple_work_queue(queue);
drivers/md/dm-vdo/funnel-workqueue.c
349
thread = kthread_run(work_queue_runner, queue, "%s:%s", thread_name_prefix,
drivers/md/dm-vdo/funnel-workqueue.c
350
queue->common.name);
drivers/md/dm-vdo/funnel-workqueue.c
352
free_simple_work_queue(queue);
drivers/md/dm-vdo/funnel-workqueue.c
356
queue->thread = thread;
drivers/md/dm-vdo/funnel-workqueue.c
368
*queue_ptr = queue;
drivers/md/dm-vdo/funnel-workqueue.c
392
struct round_robin_work_queue *queue;
drivers/md/dm-vdo/funnel-workqueue.c
409
&queue);
drivers/md/dm-vdo/funnel-workqueue.c
414
"subordinate work queues", &queue->service_queues);
drivers/md/dm-vdo/funnel-workqueue.c
416
vdo_free(queue);
drivers/md/dm-vdo/funnel-workqueue.c
420
queue->num_service_queues = thread_count;
drivers/md/dm-vdo/funnel-workqueue.c
421
queue->common.round_robin_mode = true;
drivers/md/dm-vdo/funnel-workqueue.c
422
queue->common.owner = owner;
drivers/md/dm-vdo/funnel-workqueue.c
424
result = vdo_duplicate_string(name, "queue name", &queue->common.name);
drivers/md/dm-vdo/funnel-workqueue.c
426
vdo_free(queue->service_queues);
drivers/md/dm-vdo/funnel-workqueue.c
427
vdo_free(queue);
drivers/md/dm-vdo/funnel-workqueue.c
431
*queue_ptr = &queue->common;
drivers/md/dm-vdo/funnel-workqueue.c
438
context, type, &queue->service_queues[i]);
drivers/md/dm-vdo/funnel-workqueue.c
440
queue->num_service_queues = i;
drivers/md/dm-vdo/funnel-workqueue.c
450
static void finish_simple_work_queue(struct simple_work_queue *queue)
drivers/md/dm-vdo/funnel-workqueue.c
452
if (queue->thread == NULL)
drivers/md/dm-vdo/funnel-workqueue.c
456
kthread_stop(queue->thread);
drivers/md/dm-vdo/funnel-workqueue.c
457
queue->thread = NULL;
drivers/md/dm-vdo/funnel-workqueue.c
460
static void finish_round_robin_work_queue(struct round_robin_work_queue *queue)
drivers/md/dm-vdo/funnel-workqueue.c
462
struct simple_work_queue **queue_table = queue->service_queues;
drivers/md/dm-vdo/funnel-workqueue.c
463
unsigned int count = queue->num_service_queues;
drivers/md/dm-vdo/funnel-workqueue.c
471
void vdo_finish_work_queue(struct vdo_work_queue *queue)
drivers/md/dm-vdo/funnel-workqueue.c
473
if (queue == NULL)
drivers/md/dm-vdo/funnel-workqueue.c
476
if (queue->round_robin_mode)
drivers/md/dm-vdo/funnel-workqueue.c
477
finish_round_robin_work_queue(as_round_robin_work_queue(queue));
drivers/md/dm-vdo/funnel-workqueue.c
479
finish_simple_work_queue(as_simple_work_queue(queue));
drivers/md/dm-vdo/funnel-workqueue.c
484
static void dump_simple_work_queue(struct simple_work_queue *queue)
drivers/md/dm-vdo/funnel-workqueue.c
489
if (queue->thread != NULL) {
drivers/md/dm-vdo/funnel-workqueue.c
490
task_state_report = task_state_to_char(queue->thread);
drivers/md/dm-vdo/funnel-workqueue.c
491
thread_status = atomic_read(&queue->idle) ? "idle" : "running";
drivers/md/dm-vdo/funnel-workqueue.c
494
vdo_log_info("workQ %px (%s) %s (%c)", &queue->common, queue->common.name,
drivers/md/dm-vdo/funnel-workqueue.c
505
void vdo_dump_work_queue(struct vdo_work_queue *queue)
drivers/md/dm-vdo/funnel-workqueue.c
507
if (queue->round_robin_mode) {
drivers/md/dm-vdo/funnel-workqueue.c
508
struct round_robin_work_queue *round_robin = as_round_robin_work_queue(queue);
drivers/md/dm-vdo/funnel-workqueue.c
514
dump_simple_work_queue(as_simple_work_queue(queue));
drivers/md/dm-vdo/funnel-workqueue.c
562
void vdo_enqueue_work_queue(struct vdo_work_queue *queue,
drivers/md/dm-vdo/funnel-workqueue.c
571
if (!queue->round_robin_mode) {
drivers/md/dm-vdo/funnel-workqueue.c
572
simple_queue = as_simple_work_queue(queue);
drivers/md/dm-vdo/funnel-workqueue.c
574
struct round_robin_work_queue *round_robin = as_round_robin_work_queue(queue);
drivers/md/dm-vdo/funnel-workqueue.c
618
struct simple_work_queue *queue = get_current_thread_work_queue();
drivers/md/dm-vdo/funnel-workqueue.c
620
return (queue == NULL) ? NULL : &queue->common;
drivers/md/dm-vdo/funnel-workqueue.c
623
struct vdo_thread *vdo_get_work_queue_owner(struct vdo_work_queue *queue)
drivers/md/dm-vdo/funnel-workqueue.c
625
return queue->owner;
drivers/md/dm-vdo/funnel-workqueue.c
635
struct simple_work_queue *queue = get_current_thread_work_queue();
drivers/md/dm-vdo/funnel-workqueue.c
637
return (queue != NULL) ? queue->private : NULL;
drivers/md/dm-vdo/funnel-workqueue.c
640
bool vdo_work_queue_type_is(struct vdo_work_queue *queue,
drivers/md/dm-vdo/funnel-workqueue.c
643
return (queue->type == type);
drivers/md/dm-vdo/funnel-workqueue.c
73
static inline struct simple_work_queue *as_simple_work_queue(struct vdo_work_queue *queue)
drivers/md/dm-vdo/funnel-workqueue.c
75
return ((queue == NULL) ?
drivers/md/dm-vdo/funnel-workqueue.c
76
NULL : container_of(queue, struct simple_work_queue, common));
drivers/md/dm-vdo/funnel-workqueue.c
79
static inline struct round_robin_work_queue *as_round_robin_work_queue(struct vdo_work_queue *queue)
drivers/md/dm-vdo/funnel-workqueue.c
81
return ((queue == NULL) ?
drivers/md/dm-vdo/funnel-workqueue.c
83
container_of(queue, struct round_robin_work_queue, common));
drivers/md/dm-vdo/funnel-workqueue.c
96
static struct vdo_completion *poll_for_completion(struct simple_work_queue *queue)
drivers/md/dm-vdo/funnel-workqueue.h
33
void vdo_enqueue_work_queue(struct vdo_work_queue *queue, struct vdo_completion *completion);
drivers/md/dm-vdo/funnel-workqueue.h
35
void vdo_finish_work_queue(struct vdo_work_queue *queue);
drivers/md/dm-vdo/funnel-workqueue.h
37
void vdo_free_work_queue(struct vdo_work_queue *queue);
drivers/md/dm-vdo/funnel-workqueue.h
39
void vdo_dump_work_queue(struct vdo_work_queue *queue);
drivers/md/dm-vdo/funnel-workqueue.h
46
struct vdo_thread *vdo_get_work_queue_owner(struct vdo_work_queue *queue);
drivers/md/dm-vdo/funnel-workqueue.h
48
bool __must_check vdo_work_queue_type_is(struct vdo_work_queue *queue,
drivers/md/dm-vdo/indexer/funnel-requestqueue.c
104
if (!READ_ONCE(queue->running)) {
drivers/md/dm-vdo/indexer/funnel-requestqueue.c
115
static void wait_for_request(struct uds_request_queue *queue, bool dormant,
drivers/md/dm-vdo/indexer/funnel-requestqueue.c
120
wait_event_interruptible(queue->wait_head,
drivers/md/dm-vdo/indexer/funnel-requestqueue.c
121
(dequeue_request(queue, request, waited) ||
drivers/md/dm-vdo/indexer/funnel-requestqueue.c
122
!are_queues_idle(queue)));
drivers/md/dm-vdo/indexer/funnel-requestqueue.c
126
wait_event_interruptible_hrtimeout(queue->wait_head,
drivers/md/dm-vdo/indexer/funnel-requestqueue.c
127
dequeue_request(queue, request, waited),
drivers/md/dm-vdo/indexer/funnel-requestqueue.c
133
struct uds_request_queue *queue = arg;
drivers/md/dm-vdo/indexer/funnel-requestqueue.c
136
bool dormant = atomic_read(&queue->dormant);
drivers/md/dm-vdo/indexer/funnel-requestqueue.c
141
wait_for_request(queue, dormant, time_batch, &request, &waited);
drivers/md/dm-vdo/indexer/funnel-requestqueue.c
144
queue->processor(request);
drivers/md/dm-vdo/indexer/funnel-requestqueue.c
145
} else if (!READ_ONCE(queue->running)) {
drivers/md/dm-vdo/indexer/funnel-requestqueue.c
154
atomic_set(&queue->dormant, false);
drivers/md/dm-vdo/indexer/funnel-requestqueue.c
169
atomic_set(&queue->dormant, true);
drivers/md/dm-vdo/indexer/funnel-requestqueue.c
190
while ((request = poll_queues(queue)) != NULL)
drivers/md/dm-vdo/indexer/funnel-requestqueue.c
191
queue->processor(request);
drivers/md/dm-vdo/indexer/funnel-requestqueue.c
199
struct uds_request_queue *queue;
drivers/md/dm-vdo/indexer/funnel-requestqueue.c
201
result = vdo_allocate(1, struct uds_request_queue, __func__, &queue);
drivers/md/dm-vdo/indexer/funnel-requestqueue.c
205
queue->processor = processor;
drivers/md/dm-vdo/indexer/funnel-requestqueue.c
206
queue->running = true;
drivers/md/dm-vdo/indexer/funnel-requestqueue.c
207
atomic_set(&queue->dormant, false);
drivers/md/dm-vdo/indexer/funnel-requestqueue.c
208
init_waitqueue_head(&queue->wait_head);
drivers/md/dm-vdo/indexer/funnel-requestqueue.c
210
result = vdo_make_funnel_queue(&queue->main_queue);
drivers/md/dm-vdo/indexer/funnel-requestqueue.c
212
uds_request_queue_finish(queue);
drivers/md/dm-vdo/indexer/funnel-requestqueue.c
216
result = vdo_make_funnel_queue(&queue->retry_queue);
drivers/md/dm-vdo/indexer/funnel-requestqueue.c
218
uds_request_queue_finish(queue);
drivers/md/dm-vdo/indexer/funnel-requestqueue.c
222
result = vdo_create_thread(request_queue_worker, queue, queue_name,
drivers/md/dm-vdo/indexer/funnel-requestqueue.c
223
&queue->thread);
drivers/md/dm-vdo/indexer/funnel-requestqueue.c
225
uds_request_queue_finish(queue);
drivers/md/dm-vdo/indexer/funnel-requestqueue.c
229
queue->started = true;
drivers/md/dm-vdo/indexer/funnel-requestqueue.c
230
*queue_ptr = queue;
drivers/md/dm-vdo/indexer/funnel-requestqueue.c
234
static inline void wake_up_worker(struct uds_request_queue *queue)
drivers/md/dm-vdo/indexer/funnel-requestqueue.c
236
if (wq_has_sleeper(&queue->wait_head))
drivers/md/dm-vdo/indexer/funnel-requestqueue.c
237
wake_up(&queue->wait_head);
drivers/md/dm-vdo/indexer/funnel-requestqueue.c
240
void uds_request_queue_enqueue(struct uds_request_queue *queue,
drivers/md/dm-vdo/indexer/funnel-requestqueue.c
246
sub_queue = request->requeued ? queue->retry_queue : queue->main_queue;
drivers/md/dm-vdo/indexer/funnel-requestqueue.c
253
if (atomic_read(&queue->dormant) || unbatched)
drivers/md/dm-vdo/indexer/funnel-requestqueue.c
254
wake_up_worker(queue);
drivers/md/dm-vdo/indexer/funnel-requestqueue.c
257
void uds_request_queue_finish(struct uds_request_queue *queue)
drivers/md/dm-vdo/indexer/funnel-requestqueue.c
259
if (queue == NULL)
drivers/md/dm-vdo/indexer/funnel-requestqueue.c
269
WRITE_ONCE(queue->running, false);
drivers/md/dm-vdo/indexer/funnel-requestqueue.c
271
if (queue->started) {
drivers/md/dm-vdo/indexer/funnel-requestqueue.c
272
wake_up_worker(queue);
drivers/md/dm-vdo/indexer/funnel-requestqueue.c
273
vdo_join_threads(queue->thread);
drivers/md/dm-vdo/indexer/funnel-requestqueue.c
276
vdo_free_funnel_queue(queue->main_queue);
drivers/md/dm-vdo/indexer/funnel-requestqueue.c
277
vdo_free_funnel_queue(queue->retry_queue);
drivers/md/dm-vdo/indexer/funnel-requestqueue.c
278
vdo_free(queue);
drivers/md/dm-vdo/indexer/funnel-requestqueue.c
68
static inline struct uds_request *poll_queues(struct uds_request_queue *queue)
drivers/md/dm-vdo/indexer/funnel-requestqueue.c
72
entry = vdo_funnel_queue_poll(queue->retry_queue);
drivers/md/dm-vdo/indexer/funnel-requestqueue.c
76
entry = vdo_funnel_queue_poll(queue->main_queue);
drivers/md/dm-vdo/indexer/funnel-requestqueue.c
83
static inline bool are_queues_idle(struct uds_request_queue *queue)
drivers/md/dm-vdo/indexer/funnel-requestqueue.c
85
return vdo_is_funnel_queue_idle(queue->retry_queue) &&
drivers/md/dm-vdo/indexer/funnel-requestqueue.c
86
vdo_is_funnel_queue_idle(queue->main_queue);
drivers/md/dm-vdo/indexer/funnel-requestqueue.c
94
static inline bool dequeue_request(struct uds_request_queue *queue,
drivers/md/dm-vdo/indexer/funnel-requestqueue.c
97
struct uds_request *request = poll_queues(queue);
drivers/md/dm-vdo/indexer/funnel-requestqueue.h
26
void uds_request_queue_enqueue(struct uds_request_queue *queue,
drivers/md/dm-vdo/indexer/funnel-requestqueue.h
29
void uds_request_queue_finish(struct uds_request_queue *queue);
drivers/md/dm-vdo/indexer/index.c
1359
struct uds_request_queue *queue;
drivers/md/dm-vdo/indexer/index.c
1364
queue = index->triage_queue;
drivers/md/dm-vdo/indexer/index.c
1376
queue = index->zone_queues[request->zone_number];
drivers/md/dm-vdo/indexer/index.c
1384
uds_request_queue_enqueue(queue, request);
drivers/md/dm-vdo/io-submitter.c
35
struct vdo_work_queue *queue;
drivers/md/dm-vdo/io-submitter.c
435
bio_queue_data->queue = vdo->threads[vdo->thread_config.bio_threads[i]].queue;
drivers/md/dm-vdo/io-submitter.c
456
vdo_finish_work_queue(io_submitter->bio_queue_data[i].queue);
drivers/md/dm-vdo/io-submitter.c
476
vdo_forget(io_submitter->bio_queue_data[i].queue);
drivers/md/dm-vdo/priority-table.c
116
list_del_init(&table->buckets[priority].queue);
drivers/md/dm-vdo/priority-table.c
134
list_move_tail(entry, &table->buckets[priority].queue);
drivers/md/dm-vdo/priority-table.c
174
entry = bucket->queue.next;
drivers/md/dm-vdo/priority-table.c
178
if (list_empty(&bucket->queue))
drivers/md/dm-vdo/priority-table.c
212
mark_bucket_empty(table, list_entry(next_entry, struct bucket, queue));
drivers/md/dm-vdo/priority-table.c
27
struct list_head queue;
drivers/md/dm-vdo/priority-table.c
72
INIT_LIST_HEAD(&bucket->queue);
drivers/md/dm-vdo/slab-depot.c
250
struct vdo_wait_queue *queue)
drivers/md/dm-vdo/slab-depot.c
255
vdo_waitq_notify_all_waiters(queue, NULL, &result);
drivers/md/dm-vdo/vdo.c
1602
struct vdo_work_queue *queue = vdo_get_current_work_queue();
drivers/md/dm-vdo/vdo.c
1606
if (queue == NULL)
drivers/md/dm-vdo/vdo.c
1609
thread = vdo_get_work_queue_owner(queue);
drivers/md/dm-vdo/vdo.c
420
if (thread->queue != NULL) {
drivers/md/dm-vdo/vdo.c
421
return VDO_ASSERT(vdo_work_queue_type_is(thread->queue, type),
drivers/md/dm-vdo/vdo.c
430
type, queue_count, contexts, &thread->queue);
drivers/md/dm-vdo/vdo.c
634
vdo_finish_work_queue(vdo->threads[i].queue);
drivers/md/dm-vdo/vdo.c
707
vdo_free_work_queue(vdo_forget(vdo->threads[i].queue));
drivers/md/dm-vdo/vdo.h
59
struct vdo_work_queue *queue;
drivers/md/dm-zone.c
152
struct request_queue *q = md->queue;
drivers/md/dm-zone.c
327
max_active_zones = disk->queue->limits.max_active_zones;
drivers/md/dm-zone.c
333
max_open_zones = disk->queue->limits.max_open_zones;
drivers/md/dm.c
1075
return &md->queue->limits;
drivers/md/dm.c
1095
blk_queue_disable_discard(md->queue);
drivers/md/dm.c
1098
blk_queue_disable_write_zeroes(md->queue);
drivers/md/dm.c
1168
min(max_sectors ? : queue_max_sectors(ti->table->md->queue),
drivers/md/dm.c
1822
unsigned int zone_sectors = md->disk->queue->limits.chunk_sectors;
drivers/md/dm.c
2261
dm_queue_destroy_crypto_profile(md->queue);
drivers/md/dm.c
2336
md->queue = md->disk->queue;
drivers/md/dm.c
2465
int ret = dm_table_set_restrictions(t, md->queue, limits);
drivers/md/dm.c
2605
r = dm_table_set_restrictions(t, md->queue, &limits);
drivers/md/dm.c
2807
if (!queue_is_mq(md->queue))
drivers/md/dm.c
2811
if (!blk_mq_queue_inflight(md->queue))
drivers/md/dm.c
2890
limits = md->queue->limits;
drivers/md/dm.c
3001
dm_stop_queue(md->queue);
drivers/md/dm.c
3025
dm_start_queue(md->queue);
drivers/md/dm.c
3109
dm_start_queue(md->queue);
drivers/md/dm.c
598
if (blk_queue_io_stat(md->queue))
drivers/md/md-linear.c
84
return queue_limits_set(mddev->gendisk->queue, &lim);
drivers/md/md.c
1873
bmask = queue_logical_block_size(rdev->bdev->bd_disk->queue)-1;
drivers/md/md.c
226
rdev->bdev->bd_disk->queue->nr_hw_queues != 1 &&
drivers/md/md.c
2315
bmask = queue_logical_block_size(rdev->bdev->bd_disk->queue)-1;
drivers/md/md.c
5950
if (queue_logical_block_size(mddev->gendisk->queue) >= lbs) {
drivers/md/md.c
5956
lim = queue_limits_start_update(mddev->gendisk->queue);
drivers/md/md.c
5960
err = queue_limits_commit_update(mddev->gendisk->queue, &lim);
drivers/md/md.c
6002
curr_lbs = queue_logical_block_size(mddev->gendisk->queue);
drivers/md/md.c
6222
if (queue_logical_block_size(rdev->bdev->bd_disk->queue) >
drivers/md/md.c
6223
queue_logical_block_size(mddev->gendisk->queue)) {
drivers/md/md.c
6229
lim = queue_limits_start_update(mddev->gendisk->queue);
drivers/md/md.c
6236
queue_limits_cancel_update(mddev->gendisk->queue);
drivers/md/md.c
6240
return queue_limits_commit_update(mddev->gendisk->queue, &lim);
drivers/md/md.c
6255
lim = queue_limits_start_update(mddev->gendisk->queue);
drivers/md/md.c
6257
queue_limits_commit_update(mddev->gendisk->queue, &lim);
drivers/md/md.c
6899
blk_sync_queue(mddev->gendisk->queue);
drivers/md/md.c
9246
if (blk_queue_io_stat(bdev->bd_disk->queue))
drivers/md/md.h
1088
blk_add_trace_msg((mddev)->gendisk->queue, fmt, ##args); \
drivers/md/md.h
993
!bio->bi_bdev->bd_disk->queue->limits.max_write_zeroes_sectors)
drivers/md/md.h
994
mddev->gendisk->queue->limits.max_write_zeroes_sectors = 0;
drivers/md/raid0.c
398
return queue_limits_set(mddev->gendisk->queue, &lim);
drivers/md/raid0.c
76
blksize = queue_logical_block_size(mddev->gendisk->queue);
drivers/md/raid0.c
94
rdev1->bdev->bd_disk->queue));
drivers/md/raid1.c
3198
return queue_limits_set(mddev->gendisk->queue, &lim);
drivers/md/raid10.c
3945
return queue_limits_set(mddev->gendisk->queue, &lim);
drivers/md/raid5-ppl.c
1396
queue_logical_block_size(mddev->gendisk->queue);
drivers/md/raid5.c
4243
blk_add_trace_msg(mddev->gendisk->queue,
drivers/md/raid5.c
5657
trace_block_unplug(mddev->gendisk->queue, cnt, !from_schedule);
drivers/md/raid5.c
7133
struct request_queue *q = mddev->gendisk->queue;
drivers/md/raid5.c
7750
queue_max_hw_sectors(conf->mddev->gendisk->queue) >>
drivers/md/raid5.c
7822
return queue_limits_set(mddev->gendisk->queue, &lim);
drivers/media/common/saa7146/saa7146_fops.c
112
if (!list_empty(&q->queue)) {
drivers/media/common/saa7146/saa7146_fops.c
114
buf = list_entry(q->queue.next, struct saa7146_buf, list);
drivers/media/common/saa7146/saa7146_fops.c
116
if (!list_empty(&q->queue))
drivers/media/common/saa7146/saa7146_fops.c
117
next = list_entry(q->queue.next, struct saa7146_buf, list);
drivers/media/common/saa7146/saa7146_fops.c
120
buf, q->queue.prev, q->queue.next);
drivers/media/common/saa7146/saa7146_fops.c
395
vfd->queue = q;
drivers/media/common/saa7146/saa7146_fops.c
63
list_add_tail(&buf->list, &q->queue);
drivers/media/common/saa7146/saa7146_vbi.c
301
while (!list_empty(&dq->queue)) {
drivers/media/common/saa7146/saa7146_vbi.c
302
buf = list_entry(dq->queue.next, struct saa7146_buf, list);
drivers/media/common/saa7146/saa7146_vbi.c
418
INIT_LIST_HEAD(&vv->vbi_dmaq.queue);
drivers/media/common/saa7146/saa7146_video.c
646
while (!list_empty(&dq->queue)) {
drivers/media/common/saa7146/saa7146_video.c
647
buf = list_entry(dq->queue.next, struct saa7146_buf, list);
drivers/media/common/saa7146/saa7146_video.c
691
INIT_LIST_HEAD(&vv->video_dmaq.queue);
drivers/media/common/videobuf2/videobuf2-core.c
1666
.queue = vb2_req_queue,
drivers/media/common/videobuf2/videobuf2-v4l2.c
1004
if (vdev->queue->type != d->type)
drivers/media/common/videobuf2/videobuf2-v4l2.c
1010
if (vb2_queue_is_busy(vdev->queue, file))
drivers/media/common/videobuf2/videobuf2-v4l2.c
1013
if (vb2_fileio_is_active(vdev->queue)) {
drivers/media/common/videobuf2/videobuf2-v4l2.c
1014
dprintk(vdev->queue, 1, "file io in progress\n");
drivers/media/common/videobuf2/videobuf2-v4l2.c
1018
return vb2_core_remove_bufs(vdev->queue, d->index, d->count);
drivers/media/common/videobuf2/videobuf2-v4l2.c
1026
int res = vb2_verify_memory_type(vdev->queue, p->memory, p->type);
drivers/media/common/videobuf2/videobuf2-v4l2.c
1029
vb2_set_flags_and_caps(vdev->queue, p->memory, &flags,
drivers/media/common/videobuf2/videobuf2-v4l2.c
1034
if (vb2_queue_is_busy(vdev->queue, file))
drivers/media/common/videobuf2/videobuf2-v4l2.c
1036
res = vb2_core_reqbufs(vdev->queue, p->memory, p->flags, &p->count);
drivers/media/common/videobuf2/videobuf2-v4l2.c
1040
vdev->queue->owner = p->count ? file->private_data : NULL;
drivers/media/common/videobuf2/videobuf2-v4l2.c
1049
int res = vb2_verify_memory_type(vdev->queue, p->memory, p->format.type);
drivers/media/common/videobuf2/videobuf2-v4l2.c
1051
p->index = vb2_get_num_buffers(vdev->queue);
drivers/media/common/videobuf2/videobuf2-v4l2.c
1052
vb2_set_flags_and_caps(vdev->queue, p->memory, &p->flags,
drivers/media/common/videobuf2/videobuf2-v4l2.c
1062
if (vb2_queue_is_busy(vdev->queue, file))
drivers/media/common/videobuf2/videobuf2-v4l2.c
1065
res = vb2_create_bufs(vdev->queue, p);
drivers/media/common/videobuf2/videobuf2-v4l2.c
1067
vdev->queue->owner = file->private_data;
drivers/media/common/videobuf2/videobuf2-v4l2.c
1077
if (vb2_queue_is_busy(vdev->queue, file))
drivers/media/common/videobuf2/videobuf2-v4l2.c
1079
return vb2_prepare_buf(vdev->queue, vdev->v4l2_dev->mdev, p);
drivers/media/common/videobuf2/videobuf2-v4l2.c
1088
return vb2_querybuf(vdev->queue, p);
drivers/media/common/videobuf2/videobuf2-v4l2.c
1096
if (vb2_queue_is_busy(vdev->queue, file))
drivers/media/common/videobuf2/videobuf2-v4l2.c
1098
return vb2_qbuf(vdev->queue, vdev->v4l2_dev->mdev, p);
drivers/media/common/videobuf2/videobuf2-v4l2.c
1106
if (vb2_queue_is_busy(vdev->queue, file))
drivers/media/common/videobuf2/videobuf2-v4l2.c
1108
return vb2_dqbuf(vdev->queue, p, file->f_flags & O_NONBLOCK);
drivers/media/common/videobuf2/videobuf2-v4l2.c
1116
if (vb2_queue_is_busy(vdev->queue, file))
drivers/media/common/videobuf2/videobuf2-v4l2.c
1118
return vb2_streamon(vdev->queue, i);
drivers/media/common/videobuf2/videobuf2-v4l2.c
1126
if (vb2_queue_is_busy(vdev->queue, file))
drivers/media/common/videobuf2/videobuf2-v4l2.c
1128
return vb2_streamoff(vdev->queue, i);
drivers/media/common/videobuf2/videobuf2-v4l2.c
1136
if (vb2_queue_is_busy(vdev->queue, file))
drivers/media/common/videobuf2/videobuf2-v4l2.c
1138
return vb2_expbuf(vdev->queue, p);
drivers/media/common/videobuf2/videobuf2-v4l2.c
1148
return vb2_mmap(vdev->queue, vma);
drivers/media/common/videobuf2/videobuf2-v4l2.c
1158
if (!vdev->queue->owner || file->private_data == vdev->queue->owner) {
drivers/media/common/videobuf2/videobuf2-v4l2.c
1159
vb2_queue_release(vdev->queue);
drivers/media/common/videobuf2/videobuf2-v4l2.c
1160
vdev->queue->owner = NULL;
drivers/media/common/videobuf2/videobuf2-v4l2.c
1171
struct mutex *lock = vdev->queue->lock ? vdev->queue->lock : vdev->lock;
drivers/media/common/videobuf2/videobuf2-v4l2.c
1181
struct mutex *lock = vdev->queue->lock ? vdev->queue->lock : vdev->lock;
drivers/media/common/videobuf2/videobuf2-v4l2.c
1184
if (!(vdev->queue->io_modes & VB2_WRITE))
drivers/media/common/videobuf2/videobuf2-v4l2.c
1188
if (vb2_queue_is_busy(vdev->queue, file))
drivers/media/common/videobuf2/videobuf2-v4l2.c
1190
err = vb2_write(vdev->queue, buf, count, ppos,
drivers/media/common/videobuf2/videobuf2-v4l2.c
1192
if (vdev->queue->fileio)
drivers/media/common/videobuf2/videobuf2-v4l2.c
1193
vdev->queue->owner = file->private_data;
drivers/media/common/videobuf2/videobuf2-v4l2.c
1205
struct mutex *lock = vdev->queue->lock ? vdev->queue->lock : vdev->lock;
drivers/media/common/videobuf2/videobuf2-v4l2.c
1208
if (!(vdev->queue->io_modes & VB2_READ))
drivers/media/common/videobuf2/videobuf2-v4l2.c
1212
if (vb2_queue_is_busy(vdev->queue, file))
drivers/media/common/videobuf2/videobuf2-v4l2.c
1214
vdev->queue->owner = file->private_data;
drivers/media/common/videobuf2/videobuf2-v4l2.c
1215
err = vb2_read(vdev->queue, buf, count, ppos,
drivers/media/common/videobuf2/videobuf2-v4l2.c
1217
if (!vdev->queue->fileio)
drivers/media/common/videobuf2/videobuf2-v4l2.c
1218
vdev->queue->owner = NULL;
drivers/media/common/videobuf2/videobuf2-v4l2.c
1229
struct vb2_queue *q = vdev->queue;
drivers/media/common/videobuf2/videobuf2-v4l2.c
1245
res = vb2_poll(vdev->queue, file, wait);
drivers/media/common/videobuf2/videobuf2-v4l2.c
1262
return vb2_get_unmapped_area(vdev->queue, addr, len, pgoff, flags);
drivers/media/common/videobuf2/videobuf2-v4l2.c
1277
WARN_ON(!vdev->queue);
drivers/media/common/videobuf2/videobuf2-v4l2.c
1286
if (vdev->queue) {
drivers/media/common/videobuf2/videobuf2-v4l2.c
1287
struct mutex *lock = vdev->queue->lock ?
drivers/media/common/videobuf2/videobuf2-v4l2.c
1288
vdev->queue->lock : vdev->lock;
drivers/media/common/videobuf2/videobuf2-v4l2.c
1292
vb2_queue_release(vdev->queue);
drivers/media/common/videobuf2/videobuf2-v4l2.c
1293
vdev->queue->owner = NULL;
drivers/media/common/videobuf2/videobuf2-v4l2.c
1351
if (obj->ops->queue)
drivers/media/common/videobuf2/videobuf2-v4l2.c
1352
obj->ops->queue(obj);
drivers/media/dvb-core/dmxdev.c
1209
poll_wait(file, &dmxdevfilter->buffer.queue, wait);
drivers/media/dvb-core/dmxdev.c
1347
poll_wait(file, &dmxdev->dvr_buffer.queue, wait);
drivers/media/dvb-core/dmxdev.c
365
wake_up(&dmxdevfilter->buffer.queue);
drivers/media/dvb-core/dmxdev.c
390
wake_up(&dmxdevfilter->buffer.queue);
drivers/media/dvb-core/dmxdev.c
421
wake_up(&dmxdevfilter->buffer.queue);
drivers/media/dvb-core/dmxdev.c
465
wake_up(&buffer->queue);
drivers/media/dvb-core/dmxdev.c
476
wake_up(&buffer->queue);
drivers/media/dvb-core/dmxdev.c
76
ret = wait_event_interruptible(src->queue,
drivers/media/dvb-core/dmxdev.c
869
wake_up(&dmxdevfilter->buffer.queue);
drivers/media/dvb-core/dvb_ringbuffer.c
131
wake_up(&rbuf->queue);
drivers/media/dvb-core/dvb_ringbuffer.c
46
init_waitqueue_head(&rbuf->queue);
drivers/media/dvb-frontends/rtl2832_sdr.c
1424
dev->vdev.queue = &dev->vb_queue;
drivers/media/i2c/video-i2c.c
750
struct vb2_queue *queue;
drivers/media/i2c/video-i2c.c
777
queue = &data->vb_vidq;
drivers/media/i2c/video-i2c.c
778
queue->type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
drivers/media/i2c/video-i2c.c
779
queue->io_modes = VB2_DMABUF | VB2_MMAP | VB2_USERPTR | VB2_READ;
drivers/media/i2c/video-i2c.c
780
queue->timestamp_flags = V4L2_BUF_FLAG_TIMESTAMP_MONOTONIC;
drivers/media/i2c/video-i2c.c
781
queue->drv_priv = data;
drivers/media/i2c/video-i2c.c
782
queue->buf_struct_size = sizeof(struct video_i2c_buffer);
drivers/media/i2c/video-i2c.c
783
queue->min_queued_buffers = 1;
drivers/media/i2c/video-i2c.c
784
queue->ops = &video_i2c_video_qops;
drivers/media/i2c/video-i2c.c
785
queue->mem_ops = &vb2_vmalloc_memops;
drivers/media/i2c/video-i2c.c
786
queue->lock = &data->queue_lock;
drivers/media/i2c/video-i2c.c
788
ret = vb2_queue_init(queue);
drivers/media/i2c/video-i2c.c
792
data->vdev.queue = queue;
drivers/media/pci/bt8xx/bttv-driver.c
3119
vfd->queue = q;
drivers/media/pci/cobalt/cobalt-v4l2.c
1264
vdev->queue = q;
drivers/media/pci/cx18/cx18-fileops.c
712
file_to_v4l2_fh(filp) == vdev->queue->owner) {
drivers/media/pci/cx18/cx18-fileops.c
713
vb2_queue_release(vdev->queue);
drivers/media/pci/cx18/cx18-fileops.c
714
vdev->queue->owner = NULL;
drivers/media/pci/cx18/cx18-streams.c
296
s->video_dev.queue = &s->vidq;
drivers/media/pci/cx23885/cx23885-417.c
1177
struct cx23885_buffer, queue);
drivers/media/pci/cx23885/cx23885-417.c
1185
struct cx23885_buffer, queue);
drivers/media/pci/cx23885/cx23885-417.c
1187
list_del(&buf->queue);
drivers/media/pci/cx23885/cx23885-417.c
1540
dev->v4l_device->queue = q;
drivers/media/pci/cx23885/cx23885-core.c
1630
list_add_tail(&buf->queue, &cx88q->active);
drivers/media/pci/cx23885/cx23885-core.c
1636
queue);
drivers/media/pci/cx23885/cx23885-core.c
1637
list_add_tail(&buf->queue, &cx88q->active);
drivers/media/pci/cx23885/cx23885-core.c
1656
queue);
drivers/media/pci/cx23885/cx23885-core.c
1657
list_del(&buf->queue);
drivers/media/pci/cx23885/cx23885-core.c
435
struct cx23885_buffer, queue);
drivers/media/pci/cx23885/cx23885-core.c
446
list_del(&buf->queue);
drivers/media/pci/cx23885/cx23885-dvb.c
155
struct cx23885_buffer, queue);
drivers/media/pci/cx23885/cx23885-vbi.c
199
list_add_tail(&buf->queue, &q->active);
drivers/media/pci/cx23885/cx23885-vbi.c
207
queue);
drivers/media/pci/cx23885/cx23885-vbi.c
209
list_add_tail(&buf->queue, &q->active);
drivers/media/pci/cx23885/cx23885-vbi.c
222
struct cx23885_buffer, queue);
drivers/media/pci/cx23885/cx23885-vbi.c
238
struct cx23885_buffer, queue);
drivers/media/pci/cx23885/cx23885-vbi.c
240
list_del(&buf->queue);
drivers/media/pci/cx23885/cx23885-video.c
102
list_del(&buf->queue);
drivers/media/pci/cx23885/cx23885-video.c
1360
dev->video_dev->queue = &dev->vb2_vidq;
drivers/media/pci/cx23885/cx23885-video.c
1392
dev->vbi_dev->queue = &dev->vb2_vbiq;
drivers/media/pci/cx23885/cx23885-video.c
474
list_add_tail(&buf->queue, &q->active);
drivers/media/pci/cx23885/cx23885-video.c
480
queue);
drivers/media/pci/cx23885/cx23885-video.c
481
list_add_tail(&buf->queue, &q->active);
drivers/media/pci/cx23885/cx23885-video.c
494
struct cx23885_buffer, queue);
drivers/media/pci/cx23885/cx23885-video.c
510
struct cx23885_buffer, queue);
drivers/media/pci/cx23885/cx23885-video.c
512
list_del(&buf->queue);
drivers/media/pci/cx23885/cx23885-video.c
96
struct cx23885_buffer, queue);
drivers/media/pci/cx23885/cx23885.h
175
struct list_head queue;
drivers/media/pci/cx25821/cx25821-video.c
114
struct cx25821_buffer, queue);
drivers/media/pci/cx25821/cx25821-video.c
118
list_del(&buf->queue);
drivers/media/pci/cx25821/cx25821-video.c
251
list_add_tail(&buf->queue, &q->active);
drivers/media/pci/cx25821/cx25821-video.c
255
queue);
drivers/media/pci/cx25821/cx25821-video.c
256
list_add_tail(&buf->queue, &q->active);
drivers/media/pci/cx25821/cx25821-video.c
267
struct cx25821_buffer, queue);
drivers/media/pci/cx25821/cx25821-video.c
285
struct cx25821_buffer, queue);
drivers/media/pci/cx25821/cx25821-video.c
287
list_del(&buf->queue);
drivers/media/pci/cx25821/cx25821-video.c
754
vdev->queue = q;
drivers/media/pci/cx25821/cx25821.h
114
struct list_head queue;
drivers/media/pci/cx88/cx88-blackbird.c
1134
dev->mpeg_dev.queue = &dev->vb2_mpegq;
drivers/media/pci/cx88/cx88-video.c
1447
dev->video_dev.queue = &dev->vb2_vidq;
drivers/media/pci/cx88/cx88-video.c
1464
dev->vbi_dev.queue = &dev->vb2_vbiq;
drivers/media/pci/dt3155/dt3155.c
522
pd->vdev.queue = &pd->vidq;
drivers/media/pci/intel/ipu3/ipu3-cio2.c
1370
if (cio2->queue[s_asd->csi2.port].sensor)
drivers/media/pci/intel/ipu3/ipu3-cio2.c
1377
q = &cio2->queue[s_asd->csi2.port];
drivers/media/pci/intel/ipu3/ipu3-cio2.c
1394
cio2->queue[s_asd->csi2.port].sensor = NULL;
drivers/media/pci/intel/ipu3/ipu3-cio2.c
1408
q = &cio2->queue[s_asd->csi2.port];
drivers/media/pci/intel/ipu3/ipu3-cio2.c
1544
CIO2_ENTITY_NAME " %td", q - cio2->queue);
drivers/media/pci/intel/ipu3/ipu3-cio2.c
1578
"%s %td", CIO2_NAME, q - cio2->queue);
drivers/media/pci/intel/ipu3/ipu3-cio2.c
1584
vdev->queue = &q->vbq;
drivers/media/pci/intel/ipu3/ipu3-cio2.c
1633
r = cio2_queue_init(cio2, &cio2->queue[i]);
drivers/media/pci/intel/ipu3/ipu3-cio2.c
1642
cio2_queue_exit(cio2, &cio2->queue[i]);
drivers/media/pci/intel/ipu3/ipu3-cio2.c
1652
cio2_queue_exit(cio2, &cio2->queue[i]);
drivers/media/pci/intel/ipu3/ipu3-cio2.h
376
struct cio2_queue queue[CIO2_QUEUES];
drivers/media/pci/intel/ipu6/ipu6-fw-isys.c
367
ipu6_fw_isys_get_resp(void *context, unsigned int queue)
drivers/media/pci/intel/ipu6/ipu6-fw-isys.c
369
return ipu6_recv_get_token(context, queue);
drivers/media/pci/intel/ipu6/ipu6-fw-isys.c
372
void ipu6_fw_isys_put_resp(void *context, unsigned int queue)
drivers/media/pci/intel/ipu6/ipu6-fw-isys.c
374
ipu6_recv_put_token(context, queue);
drivers/media/pci/intel/ipu6/ipu6-fw-isys.h
594
ipu6_fw_isys_get_resp(void *context, unsigned int queue);
drivers/media/pci/intel/ipu6/ipu6-fw-isys.h
595
void ipu6_fw_isys_put_resp(void *context, unsigned int queue);
drivers/media/pci/intel/ipu6/ipu6-isys-video.c
1279
av->vdev.queue = &av->aq.vbq;
drivers/media/pci/mgb4/mgb4_sysfs_in.c
427
if (vb2_is_busy(vindev->vdev.queue)) {
drivers/media/pci/mgb4/mgb4_sysfs_in.c
624
if (vb2_is_busy(vindev->vdev.queue)) {
drivers/media/pci/mgb4/mgb4_sysfs_out.c
173
if (vb2_is_busy(voutdev->vdev.queue)) {
drivers/media/pci/mgb4/mgb4_sysfs_out.c
213
if (vb2_is_busy(voutdev->vdev.queue)) {
drivers/media/pci/mgb4/mgb4_sysfs_out.c
44
ret = vb2_is_busy(dev->queue);
drivers/media/pci/mgb4/mgb4_sysfs_out.c
678
if (vb2_is_busy(voutdev->vdev.queue)) {
drivers/media/pci/mgb4/mgb4_vin.c
1000
vindev->queue.gfp_flags = GFP_DMA32;
drivers/media/pci/mgb4/mgb4_vin.c
1001
vindev->queue.timestamp_flags = V4L2_BUF_FLAG_TIMESTAMP_MONOTONIC;
drivers/media/pci/mgb4/mgb4_vin.c
1002
vindev->queue.min_queued_buffers = 2;
drivers/media/pci/mgb4/mgb4_vin.c
1003
vindev->queue.drv_priv = vindev;
drivers/media/pci/mgb4/mgb4_vin.c
1004
vindev->queue.lock = &vindev->lock;
drivers/media/pci/mgb4/mgb4_vin.c
1005
vindev->queue.dev = dev;
drivers/media/pci/mgb4/mgb4_vin.c
1006
rv = vb2_queue_init(&vindev->queue);
drivers/media/pci/mgb4/mgb4_vin.c
1021
vindev->vdev.queue = &vindev->queue;
drivers/media/pci/mgb4/mgb4_vin.c
491
if (vb2_is_busy(&vindev->queue))
drivers/media/pci/mgb4/mgb4_vin.c
636
if (vb2_is_busy(&vindev->queue))
drivers/media/pci/mgb4/mgb4_vin.c
787
if (vb2_is_streaming(&vindev->queue))
drivers/media/pci/mgb4/mgb4_vin.c
788
vb2_queue_error(&vindev->queue);
drivers/media/pci/mgb4/mgb4_vin.c
995
vindev->queue.type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
drivers/media/pci/mgb4/mgb4_vin.c
996
vindev->queue.io_modes = VB2_MMAP | VB2_DMABUF | VB2_READ;
drivers/media/pci/mgb4/mgb4_vin.c
997
vindev->queue.buf_struct_size = sizeof(struct mgb4_frame_buffer);
drivers/media/pci/mgb4/mgb4_vin.c
998
vindev->queue.ops = &queue_ops;
drivers/media/pci/mgb4/mgb4_vin.c
999
vindev->queue.mem_ops = &vb2_dma_sg_memops;
drivers/media/pci/mgb4/mgb4_vin.h
43
struct vb2_queue queue;
drivers/media/pci/mgb4/mgb4_vout.c
341
if (vb2_is_busy(&voutdev->queue))
drivers/media/pci/mgb4/mgb4_vout.c
803
voutdev->queue.type = V4L2_BUF_TYPE_VIDEO_OUTPUT;
drivers/media/pci/mgb4/mgb4_vout.c
804
voutdev->queue.io_modes = VB2_MMAP | VB2_DMABUF | VB2_WRITE;
drivers/media/pci/mgb4/mgb4_vout.c
805
voutdev->queue.buf_struct_size = sizeof(struct mgb4_frame_buffer);
drivers/media/pci/mgb4/mgb4_vout.c
806
voutdev->queue.ops = &queue_ops;
drivers/media/pci/mgb4/mgb4_vout.c
807
voutdev->queue.mem_ops = &vb2_dma_sg_memops;
drivers/media/pci/mgb4/mgb4_vout.c
808
voutdev->queue.gfp_flags = GFP_DMA32;
drivers/media/pci/mgb4/mgb4_vout.c
809
voutdev->queue.timestamp_flags = V4L2_BUF_FLAG_TIMESTAMP_MONOTONIC;
drivers/media/pci/mgb4/mgb4_vout.c
810
voutdev->queue.min_queued_buffers = 2;
drivers/media/pci/mgb4/mgb4_vout.c
811
voutdev->queue.drv_priv = voutdev;
drivers/media/pci/mgb4/mgb4_vout.c
812
voutdev->queue.lock = &voutdev->lock;
drivers/media/pci/mgb4/mgb4_vout.c
813
voutdev->queue.dev = dev;
drivers/media/pci/mgb4/mgb4_vout.c
814
rv = vb2_queue_init(&voutdev->queue);
drivers/media/pci/mgb4/mgb4_vout.c
830
voutdev->vdev.queue = &voutdev->queue;
drivers/media/pci/mgb4/mgb4_vout.h
40
struct vb2_queue queue;
drivers/media/pci/ngene/ngene-dvb.c
155
wake_up(&dev->tsin_rbuf.queue);
drivers/media/pci/ngene/ngene-dvb.c
269
wake_up_interruptible(&dev->tsout_rbuf.queue);
drivers/media/pci/ngene/ngene-dvb.c
43
if (wait_event_interruptible(dev->tsout_rbuf.queue,
drivers/media/pci/ngene/ngene-dvb.c
64
dev->tsin_rbuf.queue,
drivers/media/pci/ngene/ngene-dvb.c
86
poll_wait(file, &rbuf->queue, wait);
drivers/media/pci/ngene/ngene-dvb.c
87
poll_wait(file, &wbuf->queue, wait);
drivers/media/pci/saa7134/saa7134-core.c
1181
dev->video_dev->queue = &dev->video_vbq;
drivers/media/pci/saa7134/saa7134-core.c
1200
dev->vbi_dev->queue = &dev->vbi_vbq;
drivers/media/pci/saa7134/saa7134-core.c
1363
if (!list_empty(&q->queue))
drivers/media/pci/saa7134/saa7134-core.c
1364
next = list_entry(q->queue.next, struct saa7134_buf,
drivers/media/pci/saa7134/saa7134-core.c
273
} else if (list_empty(&q->queue)) {
drivers/media/pci/saa7134/saa7134-core.c
274
list_add_tail(&buf->entry, &q->queue);
drivers/media/pci/saa7134/saa7134-core.c
276
next = list_entry(q->queue.next, struct saa7134_buf,
drivers/media/pci/saa7134/saa7134-core.c
282
list_add_tail(&buf->entry, &q->queue);
drivers/media/pci/saa7134/saa7134-core.c
309
if (!list_empty(&q->queue)) {
drivers/media/pci/saa7134/saa7134-core.c
311
buf = list_entry(q->queue.next, struct saa7134_buf, entry);
drivers/media/pci/saa7134/saa7134-core.c
313
buf, q->queue.prev, q->queue.next);
drivers/media/pci/saa7134/saa7134-core.c
315
if (!list_empty(&q->queue))
drivers/media/pci/saa7134/saa7134-core.c
316
next = list_entry(q->queue.next, struct saa7134_buf, entry);
drivers/media/pci/saa7134/saa7134-core.c
320
q->queue.prev, q->queue.next);
drivers/media/pci/saa7134/saa7134-core.c
359
list_for_each_safe(pos, n, &q->queue) {
drivers/media/pci/saa7134/saa7134-empress.c
292
dev->empress_dev->queue = q;
drivers/media/pci/saa7134/saa7134-ts.c
137
list_for_each_entry_safe(buf, tmp, &dmaq->queue, entry) {
drivers/media/pci/saa7134/saa7134-ts.c
214
INIT_LIST_HEAD(&dev->ts_q.queue);
drivers/media/pci/saa7134/saa7134-vbi.c
172
INIT_LIST_HEAD(&dev->vbi_q.queue);
drivers/media/pci/saa7134/saa7134-video.c
1684
INIT_LIST_HEAD(&dev->video_q.queue);
drivers/media/pci/saa7134/saa7134-video.c
802
list_for_each_entry_safe(buf, tmp, &dmaq->queue, entry) {
drivers/media/pci/saa7134/saa7134.h
479
struct list_head queue;
drivers/media/pci/saa7134/saa7134.h
728
return vdev->queue == &dev->empress_vbq;
drivers/media/pci/solo6x10/solo6x10-v4l2-enc.c
1300
solo_enc->vfd->queue = &solo_enc->vidq;
drivers/media/pci/solo6x10/solo6x10-v4l2.c
651
solo_dev->vfd->queue = &solo_dev->vidq;
drivers/media/pci/tw5864/tw5864-video.c
1125
input->vdev.queue = &input->vidq;
drivers/media/pci/tw68/tw68-video.c
961
dev->vdev.queue = &dev->vidq;
drivers/media/pci/tw686x/tw686x-video.c
1271
vdev->queue = &vc->vidq;
drivers/media/pci/zoran/zoran.h
37
struct list_head queue;
drivers/media/pci/zoran/zoran_device.c
747
buf = list_first_entry_or_null(&zr->queued_bufs, struct zr_buffer, queue);
drivers/media/pci/zoran/zoran_device.c
753
list_del(&buf->queue);
drivers/media/pci/zoran/zoran_driver.c
774
list_add_tail(&buf->queue, &zr->queued_bufs);
drivers/media/pci/zoran/zoran_driver.c
823
vb2_queue_error(zr->video_dev->queue);
drivers/media/pci/zoran/zoran_driver.c
827
buf = list_first_entry_or_null(&zr->queued_bufs, struct zr_buffer, queue);
drivers/media/pci/zoran/zoran_driver.c
830
vb2_queue_error(zr->video_dev->queue);
drivers/media/pci/zoran/zoran_driver.c
834
list_del(&buf->queue);
drivers/media/pci/zoran/zoran_driver.c
935
buf = list_entry(zr->queued_bufs.next, struct zr_buffer, queue);
drivers/media/pci/zoran/zoran_driver.c
936
list_del(&buf->queue);
drivers/media/pci/zoran/zoran_driver.c
976
zr->video_dev->queue = vq;
drivers/media/pci/zoran/zoran_driver.c
982
vb2_queue_release(zr->video_dev->queue);
drivers/media/platform/amlogic/c3/isp/c3-isp-capture.c
699
vdev->queue = vb2_q;
drivers/media/platform/amlogic/c3/isp/c3-isp-params.c
856
vdev->queue = vb2_q;
drivers/media/platform/amlogic/c3/isp/c3-isp-stats.c
258
vdev->queue = vb2_q;
drivers/media/platform/arm/mali-c55/mali-c55-capture.c
370
list_add_tail(&buf->queue, &cap_dev->buffers.input);
drivers/media/platform/arm/mali-c55/mali-c55-capture.c
395
struct mali_c55_buffer, queue);
drivers/media/platform/arm/mali-c55/mali-c55-capture.c
397
list_del(&buf->queue);
drivers/media/platform/arm/mali-c55/mali-c55-capture.c
443
list_add_tail(&buf->queue, &cap_dev->buffers.processing);
drivers/media/platform/arm/mali-c55/mali-c55-capture.c
465
struct mali_c55_buffer, queue);
drivers/media/platform/arm/mali-c55/mali-c55-capture.c
474
list_del(&buf->queue);
drivers/media/platform/arm/mali-c55/mali-c55-capture.c
519
queue) {
drivers/media/platform/arm/mali-c55/mali-c55-capture.c
520
list_del(&buf->queue);
drivers/media/platform/arm/mali-c55/mali-c55-capture.c
526
list_for_each_entry_safe(buf, tmp, &cap_dev->buffers.processing, queue) {
drivers/media/platform/arm/mali-c55/mali-c55-capture.c
527
list_del(&buf->queue);
drivers/media/platform/arm/mali-c55/mali-c55-capture.c
752
if (vb2_is_busy(&cap_dev->queue))
drivers/media/platform/arm/mali-c55/mali-c55-capture.c
835
vb2q = &cap_dev->queue;
drivers/media/platform/arm/mali-c55/mali-c55-capture.c
890
vdev->queue = &cap_dev->queue;
drivers/media/platform/arm/mali-c55/mali-c55-common.h
140
struct list_head queue;
drivers/media/platform/arm/mali-c55/mali-c55-common.h
149
struct vb2_queue queue;
drivers/media/platform/arm/mali-c55/mali-c55-common.h
172
struct list_head queue;
drivers/media/platform/arm/mali-c55/mali-c55-common.h
178
struct list_head queue;
drivers/media/platform/arm/mali-c55/mali-c55-common.h
185
struct vb2_queue queue;
drivers/media/platform/arm/mali-c55/mali-c55-common.h
193
struct list_head queue;
drivers/media/platform/arm/mali-c55/mali-c55-common.h
200
struct vb2_queue queue;
drivers/media/platform/arm/mali-c55/mali-c55-common.h
208
struct list_head queue;
drivers/media/platform/arm/mali-c55/mali-c55-core.c
514
return vb2_start_streaming_called(&fr->queue) &&
drivers/media/platform/arm/mali-c55/mali-c55-core.c
516
vb2_start_streaming_called(&ds->queue)) &&
drivers/media/platform/arm/mali-c55/mali-c55-core.c
517
vb2_start_streaming_called(¶ms->queue) &&
drivers/media/platform/arm/mali-c55/mali-c55-core.c
518
vb2_start_streaming_called(&stats->queue);
drivers/media/platform/arm/mali-c55/mali-c55-params.c
610
list_add_tail(&buf->queue, ¶ms->buffers.queue);
drivers/media/platform/arm/mali-c55/mali-c55-params.c
621
list_for_each_entry_safe(buf, tmp, ¶ms->buffers.queue, queue) {
drivers/media/platform/arm/mali-c55/mali-c55-params.c
622
list_del(&buf->queue);
drivers/media/platform/arm/mali-c55/mali-c55-params.c
700
buf = list_first_entry_or_null(¶ms->buffers.queue,
drivers/media/platform/arm/mali-c55/mali-c55-params.c
701
struct mali_c55_params_buf, queue);
drivers/media/platform/arm/mali-c55/mali-c55-params.c
703
list_del(&buf->queue);
drivers/media/platform/arm/mali-c55/mali-c55-params.c
751
struct vb2_queue *vb2q = ¶ms->queue;
drivers/media/platform/arm/mali-c55/mali-c55-params.c
755
INIT_LIST_HEAD(¶ms->buffers.queue);
drivers/media/platform/arm/mali-c55/mali-c55-params.c
787
vdev->queue = ¶ms->queue;
drivers/media/platform/arm/mali-c55/mali-c55-stats.c
125
list_add_tail(&buf->queue, &stats->buffers.queue);
drivers/media/platform/arm/mali-c55/mali-c55-stats.c
136
list_for_each_entry_safe(buf, tmp, &stats->buffers.queue, queue) {
drivers/media/platform/arm/mali-c55/mali-c55-stats.c
137
list_del(&buf->queue);
drivers/media/platform/arm/mali-c55/mali-c55-stats.c
229
if (!list_empty(&stats->buffers.queue)) {
drivers/media/platform/arm/mali-c55/mali-c55-stats.c
230
buf = list_first_entry(&stats->buffers.queue,
drivers/media/platform/arm/mali-c55/mali-c55-stats.c
231
struct mali_c55_stats_buf, queue);
drivers/media/platform/arm/mali-c55/mali-c55-stats.c
232
list_del(&buf->queue);
drivers/media/platform/arm/mali-c55/mali-c55-stats.c
263
struct vb2_queue *vb2q = &stats->queue;
drivers/media/platform/arm/mali-c55/mali-c55-stats.c
267
INIT_LIST_HEAD(&stats->buffers.queue);
drivers/media/platform/arm/mali-c55/mali-c55-stats.c
298
vdev->queue = &stats->queue;
drivers/media/platform/aspeed/aspeed-video.c
1481
if (vb2_is_busy(&video->queue))
drivers/media/platform/aspeed/aspeed-video.c
1534
if (vb2_is_busy(&video->queue))
drivers/media/platform/aspeed/aspeed-video.c
1668
if (vb2_is_busy(&video->queue))
drivers/media/platform/aspeed/aspeed-video.c
2122
struct vb2_queue *vbq = &video->queue;
drivers/media/platform/aspeed/aspeed-video.c
2181
vdev->queue = vbq;
drivers/media/platform/aspeed/aspeed-video.c
323
struct vb2_queue queue;
drivers/media/platform/atmel/atmel-isi.c
1208
q = &isi->queue;
drivers/media/platform/atmel/atmel-isi.c
1224
isi->vdev->queue = &isi->queue;
drivers/media/platform/atmel/atmel-isi.c
130
struct vb2_queue queue;
drivers/media/platform/atmel/atmel-isi.c
659
if (vb2_is_streaming(&isi->queue))
drivers/media/platform/broadcom/bcm2835-unicam.c
2278
vdev->queue = q;
drivers/media/platform/imagination/e5010-jpeg-enc.c
1159
struct e5010_q_data *queue;
drivers/media/platform/imagination/e5010-jpeg-enc.c
1162
queue = get_queue(ctx, vq->type);
drivers/media/platform/imagination/e5010-jpeg-enc.c
1165
if (*nplanes != queue->fmt->num_planes)
drivers/media/platform/imagination/e5010-jpeg-enc.c
1168
if (sizes[i] < queue->sizeimage[i])
drivers/media/platform/imagination/e5010-jpeg-enc.c
1174
*nplanes = queue->fmt->num_planes;
drivers/media/platform/imagination/e5010-jpeg-enc.c
1176
sizes[i] = queue->sizeimage[i];
drivers/media/platform/imagination/e5010-jpeg-enc.c
1214
struct e5010_q_data *queue;
drivers/media/platform/imagination/e5010-jpeg-enc.c
1219
queue = get_queue(ctx, vb->vb2_queue->type);
drivers/media/platform/imagination/e5010-jpeg-enc.c
1221
for (i = 0; i < queue->fmt->num_planes; i++) {
drivers/media/platform/imagination/e5010-jpeg-enc.c
1222
if (vb2_plane_size(vb, i) < (unsigned long)queue->sizeimage[i]) {
drivers/media/platform/imagination/e5010-jpeg-enc.c
1224
vb2_plane_size(vb, i), (unsigned long)queue->sizeimage[i]);
drivers/media/platform/imagination/e5010-jpeg-enc.c
1246
struct e5010_q_data *queue = get_queue(ctx, V4L2_BUF_TYPE_VIDEO_CAPTURE_MPLANE);
drivers/media/platform/imagination/e5010-jpeg-enc.c
1248
vbuf->sequence = queue->sequence++;
drivers/media/platform/imagination/e5010-jpeg-enc.c
1294
struct e5010_q_data *queue = get_queue(ctx, q->type);
drivers/media/platform/imagination/e5010-jpeg-enc.c
1297
queue->sequence = 0;
drivers/media/platform/imagination/e5010-jpeg-enc.c
283
struct e5010_q_data *queue;
drivers/media/platform/imagination/e5010-jpeg-enc.c
288
queue = get_queue(ctx, f->type);
drivers/media/platform/imagination/e5010-jpeg-enc.c
292
pix_mp->pixelformat = queue->fmt->fourcc;
drivers/media/platform/imagination/e5010-jpeg-enc.c
293
pix_mp->width = queue->width_adjusted;
drivers/media/platform/imagination/e5010-jpeg-enc.c
294
pix_mp->height = queue->height_adjusted;
drivers/media/platform/imagination/e5010-jpeg-enc.c
295
pix_mp->num_planes = queue->fmt->num_planes;
drivers/media/platform/imagination/e5010-jpeg-enc.c
301
for (i = 0; i < queue->fmt->num_planes; i++) {
drivers/media/platform/imagination/e5010-jpeg-enc.c
302
plane_fmt[i].sizeimage = queue->sizeimage[i];
drivers/media/platform/imagination/e5010-jpeg-enc.c
303
plane_fmt[i].bytesperline = queue->bytesperline[i];
drivers/media/platform/imagination/e5010-jpeg-enc.c
309
plane_fmt[0].sizeimage = queue->sizeimage[0];
drivers/media/platform/imagination/e5010-jpeg-enc.c
395
struct e5010_q_data *queue;
drivers/media/platform/imagination/e5010-jpeg-enc.c
410
queue = get_queue(ctx, f->type);
drivers/media/platform/imagination/e5010-jpeg-enc.c
412
queue->fmt = fmt;
drivers/media/platform/imagination/e5010-jpeg-enc.c
413
queue->width = pix_mp->width;
drivers/media/platform/imagination/e5010-jpeg-enc.c
414
queue->height = pix_mp->height;
drivers/media/platform/imagination/e5010-jpeg-enc.c
418
queue->bytesperline[i] = plane_fmt[i].bytesperline;
drivers/media/platform/imagination/e5010-jpeg-enc.c
419
queue->sizeimage[i] = plane_fmt[i].sizeimage;
drivers/media/platform/imagination/e5010-jpeg-enc.c
421
queue->crop.left = 0;
drivers/media/platform/imagination/e5010-jpeg-enc.c
422
queue->crop.top = 0;
drivers/media/platform/imagination/e5010-jpeg-enc.c
423
queue->crop.width = queue->width;
drivers/media/platform/imagination/e5010-jpeg-enc.c
424
queue->crop.height = queue->height;
drivers/media/platform/imagination/e5010-jpeg-enc.c
426
queue->sizeimage[0] = plane_fmt[0].sizeimage;
drivers/media/platform/imagination/e5010-jpeg-enc.c
427
queue->sizeimage[1] = 0;
drivers/media/platform/imagination/e5010-jpeg-enc.c
428
queue->bytesperline[0] = 0;
drivers/media/platform/imagination/e5010-jpeg-enc.c
429
queue->bytesperline[1] = 0;
drivers/media/platform/imagination/e5010-jpeg-enc.c
464
struct e5010_q_data *queue;
drivers/media/platform/imagination/e5010-jpeg-enc.c
469
queue = get_queue(ctx, V4L2_BUF_TYPE_VIDEO_OUTPUT_MPLANE);
drivers/media/platform/imagination/e5010-jpeg-enc.c
476
s->r.width = queue->width;
drivers/media/platform/imagination/e5010-jpeg-enc.c
477
s->r.height = queue->height;
drivers/media/platform/imagination/e5010-jpeg-enc.c
480
memcpy(&s->r, &queue->crop, sizeof(s->r));
drivers/media/platform/imagination/e5010-jpeg-enc.c
492
struct e5010_q_data *queue;
drivers/media/platform/imagination/e5010-jpeg-enc.c
505
queue = get_queue(ctx, V4L2_BUF_TYPE_VIDEO_OUTPUT_MPLANE);
drivers/media/platform/imagination/e5010-jpeg-enc.c
508
base_rect.width = queue->width;
drivers/media/platform/imagination/e5010-jpeg-enc.c
509
base_rect.height = queue->height;
drivers/media/platform/imagination/e5010-jpeg-enc.c
513
s->r.width = round_down(s->r.width, queue->fmt->frmsize.step_width);
drivers/media/platform/imagination/e5010-jpeg-enc.c
514
s->r.height = round_down(s->r.height, queue->fmt->frmsize.step_height);
drivers/media/platform/imagination/e5010-jpeg-enc.c
515
s->r.left = round_down(s->r.left, queue->fmt->frmsize.step_width);
drivers/media/platform/imagination/e5010-jpeg-enc.c
518
if (s->r.left + s->r.width > queue->width)
drivers/media/platform/imagination/e5010-jpeg-enc.c
519
s->r.width = round_down(s->r.width + s->r.left - queue->width,
drivers/media/platform/imagination/e5010-jpeg-enc.c
520
queue->fmt->frmsize.step_width);
drivers/media/platform/imagination/e5010-jpeg-enc.c
521
if (s->r.top + s->r.height > queue->height)
drivers/media/platform/imagination/e5010-jpeg-enc.c
522
s->r.top = round_down(s->r.top + s->r.height - queue->height, 2);
drivers/media/platform/imagination/e5010-jpeg-enc.c
525
s->r.width = round_up(s->r.width, queue->fmt->frmsize.step_width);
drivers/media/platform/imagination/e5010-jpeg-enc.c
526
s->r.height = round_up(s->r.height, queue->fmt->frmsize.step_height);
drivers/media/platform/imagination/e5010-jpeg-enc.c
527
s->r.left = round_up(s->r.left, queue->fmt->frmsize.step_width);
drivers/media/platform/imagination/e5010-jpeg-enc.c
531
s->r.width = round_down(s->r.width, queue->fmt->frmsize.step_width);
drivers/media/platform/imagination/e5010-jpeg-enc.c
532
s->r.height = round_down(s->r.height, queue->fmt->frmsize.step_height);
drivers/media/platform/imagination/e5010-jpeg-enc.c
533
s->r.left = round_down(s->r.left, queue->fmt->frmsize.step_width);
drivers/media/platform/imagination/e5010-jpeg-enc.c
537
if (!IS_ALIGNED(s->r.width, queue->fmt->frmsize.step_width) ||
drivers/media/platform/imagination/e5010-jpeg-enc.c
538
!IS_ALIGNED(s->r.height, queue->fmt->frmsize.step_height) ||
drivers/media/platform/imagination/e5010-jpeg-enc.c
539
!IS_ALIGNED(s->r.left, queue->fmt->frmsize.step_width) ||
drivers/media/platform/imagination/e5010-jpeg-enc.c
550
memcpy(&queue->crop, &s->r, sizeof(s->r));
drivers/media/platform/imagination/e5010-jpeg-enc.c
553
queue->crop_set = true;
drivers/media/platform/imagination/e5010-jpeg-enc.c
556
ctx, queue->crop.width, queue->crop.height, queue->crop.left, queue->crop.top);
drivers/media/platform/imagination/e5010-jpeg-enc.c
671
struct e5010_q_data *queue;
drivers/media/platform/imagination/e5010-jpeg-enc.c
681
queue = &ctx->out_queue;
drivers/media/platform/imagination/e5010-jpeg-enc.c
682
queue->fmt = fmt;
drivers/media/platform/imagination/e5010-jpeg-enc.c
683
queue->width = DEFAULT_WIDTH;
drivers/media/platform/imagination/e5010-jpeg-enc.c
684
queue->height = DEFAULT_HEIGHT;
drivers/media/platform/imagination/e5010-jpeg-enc.c
685
pix_mp->width = queue->width;
drivers/media/platform/imagination/e5010-jpeg-enc.c
686
pix_mp->height = queue->height;
drivers/media/platform/imagination/e5010-jpeg-enc.c
687
queue->crop.left = 0;
drivers/media/platform/imagination/e5010-jpeg-enc.c
688
queue->crop.top = 0;
drivers/media/platform/imagination/e5010-jpeg-enc.c
689
queue->crop.width = queue->width;
drivers/media/platform/imagination/e5010-jpeg-enc.c
690
queue->crop.height = queue->height;
drivers/media/platform/imagination/e5010-jpeg-enc.c
697
queue->bytesperline[i] = plane_fmt[i].bytesperline;
drivers/media/platform/imagination/e5010-jpeg-enc.c
698
queue->sizeimage[i] = plane_fmt[i].sizeimage;
drivers/media/platform/imagination/e5010-jpeg-enc.c
700
queue->width_adjusted = pix_mp->width;
drivers/media/platform/imagination/e5010-jpeg-enc.c
701
queue->height_adjusted = pix_mp->height;
drivers/media/platform/imagination/e5010-jpeg-enc.c
706
queue = &ctx->cap_queue;
drivers/media/platform/imagination/e5010-jpeg-enc.c
707
queue->fmt = fmt;
drivers/media/platform/imagination/e5010-jpeg-enc.c
708
queue->width = DEFAULT_WIDTH;
drivers/media/platform/imagination/e5010-jpeg-enc.c
709
queue->height = DEFAULT_HEIGHT;
drivers/media/platform/imagination/e5010-jpeg-enc.c
710
pix_mp->width = queue->width;
drivers/media/platform/imagination/e5010-jpeg-enc.c
711
pix_mp->height = queue->height;
drivers/media/platform/imagination/e5010-jpeg-enc.c
715
queue->sizeimage[0] = pix_mp->width * pix_mp->height * JPEG_MAX_BYTES_PER_PIXEL;
drivers/media/platform/imagination/e5010-jpeg-enc.c
716
queue->sizeimage[0] += HEADER_SIZE;
drivers/media/platform/imagination/e5010-jpeg-enc.c
717
queue->sizeimage[1] = 0;
drivers/media/platform/imagination/e5010-jpeg-enc.c
718
queue->bytesperline[0] = 0;
drivers/media/platform/imagination/e5010-jpeg-enc.c
719
queue->bytesperline[1] = 0;
drivers/media/platform/imagination/e5010-jpeg-enc.c
720
queue->width_adjusted = pix_mp->width;
drivers/media/platform/imagination/e5010-jpeg-enc.c
721
queue->height_adjusted = pix_mp->height;
drivers/media/platform/intel/pxa_camera.c
1033
WARN_ON(buf->inwork || list_empty(&buf->queue));
drivers/media/platform/intel/pxa_camera.c
1051
struct pxa_buffer, queue);
drivers/media/platform/intel/pxa_camera.c
1060
list_for_each_entry(buf, &pcdev->capture, queue)
drivers/media/platform/intel/pxa_camera.c
1165
struct pxa_buffer, queue);
drivers/media/platform/intel/pxa_camera.c
1345
INIT_LIST_HEAD(&buf->queue);
drivers/media/platform/intel/pxa_camera.c
1370
list_add_tail(&buf->queue, &pcdev->capture);
drivers/media/platform/intel/pxa_camera.c
1495
list_for_each_entry_safe(buf, tmp, &pcdev->capture, queue)
drivers/media/platform/intel/pxa_camera.c
2065
pcdev->vdev.queue = &pcdev->vb2_vq;
drivers/media/platform/intel/pxa_camera.c
636
struct list_head queue;
drivers/media/platform/intel/pxa_camera.c
949
list_del_init(&buf->queue);
drivers/media/platform/intel/pxa_camera.c
963
struct pxa_buffer, queue);
drivers/media/platform/marvell/mcam-core.c
1109
list_add(&mvb->queue, &cam->buffers);
drivers/media/platform/marvell/mcam-core.c
1126
list_for_each_entry_safe(buf, node, &cam->buffers, queue) {
drivers/media/platform/marvell/mcam-core.c
1128
list_del(&buf->queue);
drivers/media/platform/marvell/mcam-core.c
1790
cam->vdev.queue = &cam->vb_queue;
drivers/media/platform/marvell/mcam-core.c
197
struct list_head queue;
drivers/media/platform/marvell/mcam-core.c
1978
list_add(&cam->vb_bufs[0]->queue, &cam->buffers);
drivers/media/platform/marvell/mcam-core.c
466
queue);
drivers/media/platform/marvell/mcam-core.c
467
list_del_init(&buf->queue);
drivers/media/platform/marvell/mcam-core.c
553
queue);
drivers/media/platform/marvell/mcam-core.c
554
list_del_init(&buf->queue);
drivers/media/platform/marvell/mcam-core.c
608
buf = list_first_entry(&cam->buffers, struct mcam_vb_buffer, queue);
drivers/media/platform/marvell/mcam-core.c
609
list_del_init(&buf->queue);
drivers/media/platform/mediatek/vcodec/common/mtk_vcodec_intr.c
32
ctx_queue = ctx->queue;
drivers/media/platform/mediatek/vcodec/common/mtk_vcodec_intr.c
42
ctx_queue = ctx->queue;
drivers/media/platform/mediatek/vcodec/decoder/mtk_vcodec_dec_drv.c
225
init_waitqueue_head(&ctx->queue[i]);
drivers/media/platform/mediatek/vcodec/decoder/mtk_vcodec_dec_drv.c
227
init_waitqueue_head(&ctx->queue[0]);
drivers/media/platform/mediatek/vcodec/decoder/mtk_vcodec_dec_drv.h
211
wait_queue_head_t queue[MTK_VDEC_HW_MAX];
drivers/media/platform/mediatek/vcodec/decoder/mtk_vcodec_dec_drv.h
350
wake_up_interruptible(&ctx->queue[hw_id]);
drivers/media/platform/mediatek/vcodec/encoder/mtk_vcodec_enc_drv.c
136
init_waitqueue_head(&ctx->queue[0]);
drivers/media/platform/mediatek/vcodec/encoder/mtk_vcodec_enc_drv.h
146
wait_queue_head_t queue[MTK_VDEC_HW_MAX];
drivers/media/platform/mediatek/vcodec/encoder/mtk_vcodec_enc_drv.h
236
wake_up_interruptible(&ctx->queue[hw_id]);
drivers/media/platform/microchip/microchip-isc-base.c
1817
vdev->queue = q;
drivers/media/platform/nuvoton/npcm-video.c
1118
vb2_queue_error(&video->queue);
drivers/media/platform/nuvoton/npcm-video.c
117
struct vb2_queue queue;
drivers/media/platform/nuvoton/npcm-video.c
1200
if (vb2_is_busy(&video->queue)) {
drivers/media/platform/nuvoton/npcm-video.c
1250
if (vb2_is_busy(&video->queue)) {
drivers/media/platform/nuvoton/npcm-video.c
1575
struct vb2_queue *vbq = &video->queue;
drivers/media/platform/nuvoton/npcm-video.c
1623
vdev->queue = vbq;
drivers/media/platform/nuvoton/npcm-video.c
1796
vb2_queue_release(&video->queue);
drivers/media/platform/nuvoton/npcm-video.c
395
for (i = 0; i < vb2_get_num_buffers(&video->queue); i++) {
drivers/media/platform/nxp/imx7-media-csi.c
1667
vdev->queue = &csi->q;
drivers/media/platform/nxp/imx8-isi/imx8-isi-video.c
1394
vdev->queue = q;
drivers/media/platform/qcom/camss/camss-vfe.c
761
queue);
drivers/media/platform/qcom/camss/camss-vfe.c
762
list_del(&buffer->queue);
drivers/media/platform/qcom/camss/camss-vfe.c
771
INIT_LIST_HEAD(&buffer->queue);
drivers/media/platform/qcom/camss/camss-vfe.c
772
list_add_tail(&buffer->queue, &output->pending_bufs);
drivers/media/platform/qcom/camss/camss-vfe.c
786
list_for_each_entry_safe(buf, t, &output->pending_bufs, queue) {
drivers/media/platform/qcom/camss/camss-vfe.c
788
list_del(&buf->queue);
drivers/media/platform/qcom/camss/camss-video.c
726
vdev->queue = &video->vb2_q;
drivers/media/platform/qcom/camss/camss-video.h
25
struct list_head queue;
drivers/media/platform/qcom/iris/iris_hfi_queue.c
103
queue->rx_req = receive_request;
drivers/media/platform/qcom/iris/iris_hfi_queue.c
105
queue->read_idx = new_read_idx / sizeof(u32);
drivers/media/platform/qcom/iris/iris_hfi_queue.c
14
struct iris_hfi_queue_header *queue = qinfo->qhdr;
drivers/media/platform/qcom/iris/iris_hfi_queue.c
15
u32 write_idx = queue->write_idx * sizeof(u32);
drivers/media/platform/qcom/iris/iris_hfi_queue.c
16
u32 read_idx = queue->read_idx * sizeof(u32);
drivers/media/platform/qcom/iris/iris_hfi_queue.c
27
queue->tx_req = 0;
drivers/media/platform/qcom/iris/iris_hfi_queue.c
49
queue->write_idx = new_write_idx / sizeof(u32);
drivers/media/platform/qcom/iris/iris_hfi_queue.c
59
struct iris_hfi_queue_header *queue = qinfo->qhdr;
drivers/media/platform/qcom/iris/iris_hfi_queue.c
60
u32 write_idx = queue->write_idx * sizeof(u32);
drivers/media/platform/qcom/iris/iris_hfi_queue.c
61
u32 read_idx = queue->read_idx * sizeof(u32);
drivers/media/platform/qcom/iris/iris_hfi_queue.c
67
if (queue->queue_type == IFACEQ_MSGQ_ID)
drivers/media/platform/qcom/iris/iris_hfi_queue.c
71
queue->rx_req = receive_request;
drivers/media/platform/qcom/venus/hfi_venus.c
163
struct iface_queue *queue,
drivers/media/platform/qcom/venus/hfi_venus.c
171
if (!queue->qmem.kva)
drivers/media/platform/qcom/venus/hfi_venus.c
1733
struct iface_queue *queue;
drivers/media/platform/qcom/venus/hfi_venus.c
174
qhdr = queue->qhdr;
drivers/media/platform/qcom/venus/hfi_venus.c
1740
queue = &hdev->queues[i];
drivers/media/platform/qcom/venus/hfi_venus.c
1741
queue->qhdr =
drivers/media/platform/qcom/venus/hfi_venus.c
1744
venus_set_qhdr_defaults(queue->qhdr);
drivers/media/platform/qcom/venus/hfi_venus.c
1746
queue->qhdr->start_addr = queue->qmem.da;
drivers/media/platform/qcom/venus/hfi_venus.c
1749
queue->qhdr->type |= HFI_HOST_TO_CTRL_CMD_Q;
drivers/media/platform/qcom/venus/hfi_venus.c
1751
queue->qhdr->type |= HFI_CTRL_TO_HOST_MSG_Q;
drivers/media/platform/qcom/venus/hfi_venus.c
1753
queue->qhdr->type |= HFI_CTRL_TO_HOST_DBG_Q;
drivers/media/platform/qcom/venus/hfi_venus.c
1768
queue = &hdev->queues[IFACEQ_DBG_IDX];
drivers/media/platform/qcom/venus/hfi_venus.c
1769
queue->qhdr->rx_req = 0;
drivers/media/platform/qcom/venus/hfi_venus.c
210
wr_ptr = (u32 *)(queue->qmem.kva + (wr_idx << 2));
drivers/media/platform/qcom/venus/hfi_venus.c
212
if (wr_ptr < (u32 *)queue->qmem.kva ||
drivers/media/platform/qcom/venus/hfi_venus.c
213
wr_ptr > (u32 *)(queue->qmem.kva + queue->qmem.size - sizeof(*wr_ptr)))
drivers/media/platform/qcom/venus/hfi_venus.c
224
memcpy(queue->qmem.kva, packet + len, new_wr_idx << 2);
drivers/media/platform/qcom/venus/hfi_venus.c
240
struct iface_queue *queue, void *pkt, u32 *tx_req)
drivers/media/platform/qcom/venus/hfi_venus.c
250
if (!queue->qmem.kva)
drivers/media/platform/qcom/venus/hfi_venus.c
253
qhdr = queue->qhdr;
drivers/media/platform/qcom/venus/hfi_venus.c
286
rd_ptr = (u32 *)(queue->qmem.kva + (rd_idx << 2));
drivers/media/platform/qcom/venus/hfi_venus.c
288
if (rd_ptr < (u32 *)queue->qmem.kva ||
drivers/media/platform/qcom/venus/hfi_venus.c
289
rd_ptr > (u32 *)(queue->qmem.kva + queue->qmem.size - sizeof(*rd_ptr)))
drivers/media/platform/qcom/venus/hfi_venus.c
306
memcpy(pkt + len, queue->qmem.kva, new_rd_idx << 2);
drivers/media/platform/qcom/venus/hfi_venus.c
396
struct iface_queue *queue;
drivers/media/platform/qcom/venus/hfi_venus.c
406
queue = &hdev->queues[IFACEQ_CMD_IDX];
drivers/media/platform/qcom/venus/hfi_venus.c
408
ret = venus_write_queue(hdev, queue, pkt, &rx_req);
drivers/media/platform/qcom/venus/hfi_venus.c
419
queue = &hdev->queues[IFACEQ_MSG_IDX];
drivers/media/platform/qcom/venus/hfi_venus.c
420
queue->qhdr->rx_req = 1;
drivers/media/platform/qcom/venus/hfi_venus.c
695
struct iface_queue *queue;
drivers/media/platform/qcom/venus/hfi_venus.c
702
queue = &hdev->queues[IFACEQ_MSG_IDX];
drivers/media/platform/qcom/venus/hfi_venus.c
704
ret = venus_read_queue(hdev, queue, pkt, &tx_req);
drivers/media/platform/qcom/venus/hfi_venus.c
728
struct iface_queue *queue;
drivers/media/platform/qcom/venus/hfi_venus.c
736
queue = &hdev->queues[IFACEQ_DBG_IDX];
drivers/media/platform/qcom/venus/hfi_venus.c
738
ret = venus_read_queue(hdev, queue, pkt, &tx_req);
drivers/media/platform/qcom/venus/hfi_venus.c
795
struct iface_queue *queue;
drivers/media/platform/qcom/venus/hfi_venus.c
810
queue = &hdev->queues[i];
drivers/media/platform/qcom/venus/hfi_venus.c
811
queue->qmem.da = desc.da + offset;
drivers/media/platform/qcom/venus/hfi_venus.c
812
queue->qmem.kva = desc.kva + offset;
drivers/media/platform/qcom/venus/hfi_venus.c
813
queue->qmem.size = IFACEQ_QUEUE_SIZE;
drivers/media/platform/qcom/venus/hfi_venus.c
814
offset += queue->qmem.size;
drivers/media/platform/qcom/venus/hfi_venus.c
815
queue->qhdr =
drivers/media/platform/qcom/venus/hfi_venus.c
818
venus_set_qhdr_defaults(queue->qhdr);
drivers/media/platform/qcom/venus/hfi_venus.c
820
queue->qhdr->start_addr = queue->qmem.da;
drivers/media/platform/qcom/venus/hfi_venus.c
823
queue->qhdr->type |= HFI_HOST_TO_CTRL_CMD_Q;
drivers/media/platform/qcom/venus/hfi_venus.c
825
queue->qhdr->type |= HFI_CTRL_TO_HOST_MSG_Q;
drivers/media/platform/qcom/venus/hfi_venus.c
827
queue->qhdr->type |= HFI_CTRL_TO_HOST_DBG_Q;
drivers/media/platform/qcom/venus/hfi_venus.c
842
queue = &hdev->queues[IFACEQ_DBG_IDX];
drivers/media/platform/qcom/venus/hfi_venus.c
843
queue->qhdr->rx_req = 0;
drivers/media/platform/raspberrypi/pisp_be/pisp_be.c
1221
if (vb2_is_busy(&node->queue))
drivers/media/platform/raspberrypi/pisp_be/pisp_be.c
1244
if (vb2_is_busy(&node->queue))
drivers/media/platform/raspberrypi/pisp_be/pisp_be.c
1267
if (vb2_is_busy(&node->queue))
drivers/media/platform/raspberrypi/pisp_be/pisp_be.c
1284
if (f->type != node->queue.type)
drivers/media/platform/raspberrypi/pisp_be/pisp_be.c
1400
struct vb2_queue *q = &node->queue;
drivers/media/platform/raspberrypi/pisp_be/pisp_be.c
1437
vdev->queue = &node->queue;
drivers/media/platform/raspberrypi/pisp_be/pisp_be.c
1477
vb2_queue_release(&node->queue);
drivers/media/platform/raspberrypi/pisp_be/pisp_be.c
1580
vb2_queue_release(&pispbe->node[num_regist].queue);
drivers/media/platform/raspberrypi/pisp_be/pisp_be.c
1609
vb2_queue_release(&pispbe->node[i].queue);
drivers/media/platform/raspberrypi/pisp_be/pisp_be.c
167
struct vb2_queue queue;
drivers/media/platform/raspberrypi/pisp_be/pisp_be.c
193
struct list_head queue;
drivers/media/platform/raspberrypi/pisp_be/pisp_be.c
551
list_add_tail(&job->queue, &pispbe->job_queue);
drivers/media/platform/raspberrypi/pisp_be/pisp_be.c
586
queue);
drivers/media/platform/raspberrypi/pisp_be/pisp_be.c
590
list_del(&job->queue);
drivers/media/platform/raspberrypi/pisp_be/pisp_be.c
934
vb2_wait_for_all_buffers(&node->queue);
drivers/media/platform/raspberrypi/pisp_be/pisp_be.c
948
list_for_each_entry_safe(job, temp, &tmp_list, queue) {
drivers/media/platform/raspberrypi/pisp_be/pisp_be.c
949
list_del(&job->queue);
drivers/media/platform/raspberrypi/rp1-cfe/cfe.c
1650
ret = vb2_queue_change_type(vdev->queue, p->type);
drivers/media/platform/raspberrypi/rp1-cfe/cfe.c
1673
ret = vb2_queue_change_type(vdev->queue, p->format.type);
drivers/media/platform/raspberrypi/rp1-cfe/cfe.c
2019
vdev->queue = q;
drivers/media/platform/renesas/rcar-vin/rcar-dma.c
1450
struct vb2_queue *q = &vin->queue;
drivers/media/platform/renesas/rcar-vin/rcar-v4l2.c
533
if (vb2_is_busy(&vin->queue))
drivers/media/platform/renesas/rcar-vin/rcar-v4l2.c
703
vdev->queue = &vin->queue;
drivers/media/platform/renesas/rcar-vin/rcar-vin.h
192
struct vb2_queue queue;
drivers/media/platform/renesas/rcar_drif.c
1071
sdr->vdev->queue = &sdr->vb_queue;
drivers/media/platform/renesas/renesas-ceu.c
142
struct list_head queue;
drivers/media/platform/renesas/renesas-ceu.c
1431
vdev->queue = &ceudev->vb2_vq;
drivers/media/platform/renesas/renesas-ceu.c
531
queue);
drivers/media/platform/renesas/renesas-ceu.c
532
list_del(&buf->queue);
drivers/media/platform/renesas/renesas-ceu.c
549
list_for_each_entry(buf, &ceudev->capture, queue)
drivers/media/platform/renesas/renesas-ceu.c
653
list_add_tail(&buf->queue, &ceudev->capture);
drivers/media/platform/renesas/renesas-ceu.c
703
queue);
drivers/media/platform/renesas/renesas-ceu.c
705
list_del(&buf->queue);
drivers/media/platform/renesas/renesas-ceu.c
720
list_for_each_entry(buf, &ceudev->capture, queue)
drivers/media/platform/renesas/renesas-ceu.c
751
list_for_each_entry(buf, &ceudev->capture, queue)
drivers/media/platform/renesas/rzg2l-cru/rzg2l-cru.h
161
struct vb2_queue queue;
drivers/media/platform/renesas/rzg2l-cru/rzg2l-video.c
1150
vdev->queue = &cru->queue;
drivers/media/platform/renesas/rzg2l-cru/rzg2l-video.c
847
vb2_queue_release(&cru->queue);
drivers/media/platform/renesas/rzg2l-cru/rzg2l-video.c
852
struct vb2_queue *q = &cru->queue;
drivers/media/platform/renesas/rzg2l-cru/rzg2l-video.c
978
if (vb2_is_busy(&cru->queue))
drivers/media/platform/renesas/rzv2h-ivc/rzv2h-ivc-video.c
153
buf = list_first_entry_or_null(&ivc->buffers.queue,
drivers/media/platform/renesas/rzv2h-ivc/rzv2h-ivc-video.c
154
struct rzv2h_ivc_buf, queue);
drivers/media/platform/renesas/rzv2h-ivc/rzv2h-ivc-video.c
160
list_del(&buf->queue);
drivers/media/platform/renesas/rzv2h-ivc/rzv2h-ivc-video.c
199
list_add_tail(&buf->queue, &ivc->buffers.queue);
drivers/media/platform/renesas/rzv2h-ivc/rzv2h-ivc-video.c
254
list_for_each_entry_safe(buf, tmp, &ivc->buffers.queue, queue) {
drivers/media/platform/renesas/rzv2h-ivc/rzv2h-ivc-video.c
255
list_del(&buf->queue);
drivers/media/platform/renesas/rzv2h-ivc/rzv2h-ivc-video.c
31
struct list_head queue;
drivers/media/platform/renesas/rzv2h-ivc/rzv2h-ivc-video.c
445
INIT_LIST_HEAD(&ivc->buffers.queue);
drivers/media/platform/renesas/rzv2h-ivc/rzv2h-ivc-video.c
479
vdev->queue = vb2q;
drivers/media/platform/renesas/rzv2h-ivc/rzv2h-ivc.h
107
struct list_head queue;
drivers/media/platform/renesas/sh_vou.c
1290
q = &vou_dev->queue;
drivers/media/platform/renesas/sh_vou.c
1305
vdev->queue = q;
drivers/media/platform/renesas/sh_vou.c
701
if (vb2_is_busy(&vou_dev->queue))
drivers/media/platform/renesas/sh_vou.c
823
if (vb2_is_busy(&vou_dev->queue))
drivers/media/platform/renesas/sh_vou.c
85
struct vb2_queue queue;
drivers/media/platform/renesas/sh_vou.c
946
if (vb2_is_busy(&vou_dev->queue))
drivers/media/platform/renesas/vsp1/vsp1_histo.c
127
list_add_tail(&buf->queue, &histo->irqqueue);
drivers/media/platform/renesas/vsp1/vsp1_histo.c
144
list_for_each_entry(buffer, &histo->irqqueue, queue)
drivers/media/platform/renesas/vsp1/vsp1_histo.c
415
if (f->index > 0 || f->type != histo->queue.type)
drivers/media/platform/renesas/vsp1/vsp1_histo.c
430
if (format->type != histo->queue.type)
drivers/media/platform/renesas/vsp1/vsp1_histo.c
46
queue);
drivers/media/platform/renesas/vsp1/vsp1_histo.c
47
list_del(&buf->queue);
drivers/media/platform/renesas/vsp1/vsp1_histo.c
533
histo->queue.type = V4L2_BUF_TYPE_META_CAPTURE;
drivers/media/platform/renesas/vsp1/vsp1_histo.c
534
histo->queue.io_modes = VB2_MMAP | VB2_USERPTR | VB2_DMABUF;
drivers/media/platform/renesas/vsp1/vsp1_histo.c
535
histo->queue.lock = &histo->lock;
drivers/media/platform/renesas/vsp1/vsp1_histo.c
536
histo->queue.drv_priv = histo;
drivers/media/platform/renesas/vsp1/vsp1_histo.c
537
histo->queue.buf_struct_size = sizeof(struct vsp1_histogram_buffer);
drivers/media/platform/renesas/vsp1/vsp1_histo.c
538
histo->queue.ops = &histo_video_queue_qops;
drivers/media/platform/renesas/vsp1/vsp1_histo.c
539
histo->queue.mem_ops = &vb2_vmalloc_memops;
drivers/media/platform/renesas/vsp1/vsp1_histo.c
540
histo->queue.timestamp_flags = V4L2_BUF_FLAG_TIMESTAMP_COPY;
drivers/media/platform/renesas/vsp1/vsp1_histo.c
541
histo->queue.dev = vsp1->dev;
drivers/media/platform/renesas/vsp1/vsp1_histo.c
542
ret = vb2_queue_init(&histo->queue);
drivers/media/platform/renesas/vsp1/vsp1_histo.c
549
histo->video.queue = &histo->queue;
drivers/media/platform/renesas/vsp1/vsp1_histo.h
30
struct list_head queue;
drivers/media/platform/renesas/vsp1/vsp1_histo.h
45
struct vb2_queue queue;
drivers/media/platform/renesas/vsp1/vsp1_video.c
1032
ret = vb2_streamon(&video->queue, type);
drivers/media/platform/renesas/vsp1/vsp1_video.c
1278
video->queue.type = video->type;
drivers/media/platform/renesas/vsp1/vsp1_video.c
1279
video->queue.io_modes = VB2_MMAP | VB2_USERPTR | VB2_DMABUF;
drivers/media/platform/renesas/vsp1/vsp1_video.c
1280
video->queue.lock = &video->lock;
drivers/media/platform/renesas/vsp1/vsp1_video.c
1281
video->queue.drv_priv = video;
drivers/media/platform/renesas/vsp1/vsp1_video.c
1282
video->queue.buf_struct_size = sizeof(struct vsp1_vb2_buffer);
drivers/media/platform/renesas/vsp1/vsp1_video.c
1283
video->queue.ops = &vsp1_video_queue_qops;
drivers/media/platform/renesas/vsp1/vsp1_video.c
1284
video->queue.mem_ops = &vb2_dma_contig_memops;
drivers/media/platform/renesas/vsp1/vsp1_video.c
1285
video->queue.timestamp_flags = V4L2_BUF_FLAG_TIMESTAMP_COPY;
drivers/media/platform/renesas/vsp1/vsp1_video.c
1286
video->queue.dev = video->vsp1->bus_master;
drivers/media/platform/renesas/vsp1/vsp1_video.c
1287
ret = vb2_queue_init(&video->queue);
drivers/media/platform/renesas/vsp1/vsp1_video.c
1294
video->video.queue = &video->queue;
drivers/media/platform/renesas/vsp1/vsp1_video.c
223
struct vsp1_vb2_buffer, queue);
drivers/media/platform/renesas/vsp1/vsp1_video.c
225
list_del(&done->queue);
drivers/media/platform/renesas/vsp1/vsp1_video.c
229
struct vsp1_vb2_buffer, queue);
drivers/media/platform/renesas/vsp1/vsp1_video.c
669
list_add_tail(&buf->queue, &video->irqqueue);
drivers/media/platform/renesas/vsp1/vsp1_video.c
680
if (vb2_start_streaming_called(&video->queue) &&
drivers/media/platform/renesas/vsp1/vsp1_video.c
790
list_for_each_entry(buffer, &video->irqqueue, queue)
drivers/media/platform/renesas/vsp1/vsp1_video.c
938
if (format->type != video->queue.type)
drivers/media/platform/renesas/vsp1/vsp1_video.c
954
if (format->type != video->queue.type)
drivers/media/platform/renesas/vsp1/vsp1_video.c
968
if (format->type != video->queue.type)
drivers/media/platform/renesas/vsp1/vsp1_video.c
977
if (vb2_is_busy(&video->queue)) {
drivers/media/platform/renesas/vsp1/vsp1_video.c
999
if (vb2_queue_is_busy(&video->queue, file))
drivers/media/platform/renesas/vsp1/vsp1_video.h
21
struct list_head queue;
drivers/media/platform/renesas/vsp1/vsp1_video.h
44
struct vb2_queue queue;
drivers/media/platform/renesas/vsp1/vsp1_wpf.c
63
if (vb2_is_busy(&video->queue)) {
drivers/media/platform/rockchip/rga/rga.h
107
struct list_head queue;
drivers/media/platform/rockchip/rkcif/rkcif-common.h
89
struct list_head queue;
drivers/media/platform/rockchip/rkcif/rkcif-stream.c
175
static int rkcif_stream_setup_queue(struct vb2_queue *queue,
drivers/media/platform/rockchip/rkcif/rkcif-stream.c
181
struct rkcif_stream *stream = queue->drv_priv;
drivers/media/platform/rockchip/rkcif/rkcif-stream.c
251
static int rkcif_stream_start_streaming(struct vb2_queue *queue,
drivers/media/platform/rockchip/rkcif/rkcif-stream.c
254
struct rkcif_stream *stream = queue->drv_priv;
drivers/media/platform/rockchip/rkcif/rkcif-stream.c
304
static void rkcif_stream_stop_streaming(struct vb2_queue *queue)
drivers/media/platform/rockchip/rkcif/rkcif-stream.c
306
struct rkcif_stream *stream = queue->drv_priv;
drivers/media/platform/rockchip/rkcif/rkcif-stream.c
47
queue);
drivers/media/platform/rockchip/rkcif/rkcif-stream.c
48
list_del(&buffer->queue);
drivers/media/platform/rockchip/rkcif/rkcif-stream.c
566
vdev->queue = &stream->buf_queue;
drivers/media/platform/rockchip/rkcif/rkcif-stream.c
58
list_add_tail(&buffer->queue, &stream->driver_queue);
drivers/media/platform/rockchip/rkisp1/rkisp1-capture.c
1121
static void rkisp1_vb2_stop_streaming(struct vb2_queue *queue)
drivers/media/platform/rockchip/rkisp1/rkisp1-capture.c
1123
struct rkisp1_capture *cap = queue->drv_priv;
drivers/media/platform/rockchip/rkisp1/rkisp1-capture.c
1144
rkisp1_vb2_start_streaming(struct vb2_queue *queue, unsigned int count)
drivers/media/platform/rockchip/rkisp1/rkisp1-capture.c
1146
struct rkisp1_capture *cap = queue->drv_priv;
drivers/media/platform/rockchip/rkisp1/rkisp1-capture.c
1570
vdev->queue = q;
drivers/media/platform/rockchip/rkisp1/rkisp1-capture.c
1604
INIT_LIST_HEAD(&cap->buf.queue);
drivers/media/platform/rockchip/rkisp1/rkisp1-capture.c
740
if (!list_empty(&cap->buf.queue)) {
drivers/media/platform/rockchip/rkisp1/rkisp1-capture.c
743
cap->buf.next = list_first_entry(&cap->buf.queue, struct rkisp1_buffer, queue);
drivers/media/platform/rockchip/rkisp1/rkisp1-capture.c
744
list_del(&cap->buf.next->queue);
drivers/media/platform/rockchip/rkisp1/rkisp1-capture.c
865
static int rkisp1_vb2_queue_setup(struct vb2_queue *queue,
drivers/media/platform/rockchip/rkisp1/rkisp1-capture.c
871
struct rkisp1_capture *cap = queue->drv_priv;
drivers/media/platform/rockchip/rkisp1/rkisp1-capture.c
932
list_add_tail(&ispbuf->queue, &cap->buf.queue);
drivers/media/platform/rockchip/rkisp1/rkisp1-capture.c
970
while (!list_empty(&cap->buf.queue)) {
drivers/media/platform/rockchip/rkisp1/rkisp1-capture.c
971
buf = list_first_entry(&cap->buf.queue,
drivers/media/platform/rockchip/rkisp1/rkisp1-capture.c
972
struct rkisp1_buffer, queue);
drivers/media/platform/rockchip/rkisp1/rkisp1-capture.c
973
list_del(&buf->queue);
drivers/media/platform/rockchip/rkisp1/rkisp1-common.h
258
struct list_head queue;
drivers/media/platform/rockchip/rkisp1/rkisp1-common.h
272
struct list_head queue;
drivers/media/platform/rockchip/rkisp1/rkisp1-common.h
338
struct list_head queue;
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2266
list_del(&buf->queue);
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2280
struct rkisp1_params_buffer, queue);
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2388
struct rkisp1_params_buffer, queue);
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2423
struct rkisp1_params_buffer, queue);
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2509
f->type != video->queue->type)
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2524
if (f->type != video->queue->type)
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2538
if (f->type != video->queue->type)
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2553
if (f->type != video->queue->type)
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2556
if (vb2_is_busy(video->queue))
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2647
list_add_tail(¶ms_buf->queue, ¶ms->params);
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2721
list_for_each_entry(buf, &tmp_list, queue)
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2825
vdev->queue = &node->buf_queue;
drivers/media/platform/rockchip/rkisp1/rkisp1-params.c
2828
ret = rkisp1_params_init_vb2_queue(vdev->queue, params);
drivers/media/platform/rockchip/rkisp1/rkisp1-stats.c
117
list_add_tail(&stats_buf->queue, &stats_dev->stat);
drivers/media/platform/rockchip/rkisp1/rkisp1-stats.c
142
struct rkisp1_buffer, queue);
drivers/media/platform/rockchip/rkisp1/rkisp1-stats.c
143
list_del(&buf->queue);
drivers/media/platform/rockchip/rkisp1/rkisp1-stats.c
27
if (f->index > 0 || f->type != video->queue->type)
drivers/media/platform/rockchip/rkisp1/rkisp1-stats.c
349
struct rkisp1_buffer, queue);
drivers/media/platform/rockchip/rkisp1/rkisp1-stats.c
350
list_del(&cur_buf->queue);
drivers/media/platform/rockchip/rkisp1/rkisp1-stats.c
41
if (f->type != video->queue->type)
drivers/media/platform/rockchip/rkisp1/rkisp1-stats.c
431
vdev->queue = &node->buf_queue;
drivers/media/platform/rockchip/rkisp1/rkisp1-stats.c
434
rkisp1_stats_init_vb2_queue(vdev->queue, stats);
drivers/media/platform/samsung/exynos4-is/fimc-capture.c
1740
vfd->queue = q;
drivers/media/platform/samsung/exynos4-is/fimc-isp-video.c
610
vdev->queue = q;
drivers/media/platform/samsung/exynos4-is/fimc-lite.c
1256
vfd->queue = q;
drivers/media/platform/samsung/s5p-mfc/s5p_mfc.c
117
wake_up(&ctx->queue);
drivers/media/platform/samsung/s5p-mfc/s5p_mfc.c
127
wake_up(&dev->queue);
drivers/media/platform/samsung/s5p-mfc/s5p_mfc.c
1360
init_waitqueue_head(&dev->queue);
drivers/media/platform/samsung/s5p-mfc/s5p_mfc.c
1507
ret = wait_event_interruptible_timeout(m_dev->queue,
drivers/media/platform/samsung/s5p-mfc/s5p_mfc.c
606
wake_up(&ctx->queue);
drivers/media/platform/samsung/s5p-mfc/s5p_mfc.c
615
wake_up(&ctx->queue);
drivers/media/platform/samsung/s5p-mfc/s5p_mfc.c
642
wake_up(&ctx->queue);
drivers/media/platform/samsung/s5p-mfc/s5p_mfc.c
773
wake_up(&ctx->queue);
drivers/media/platform/samsung/s5p-mfc/s5p_mfc.c
802
init_waitqueue_head(&ctx->queue);
drivers/media/platform/samsung/s5p-mfc/s5p_mfc_common.h
325
wait_queue_head_t queue;
drivers/media/platform/samsung/s5p-mfc/s5p_mfc_common.h
640
wait_queue_head_t queue;
drivers/media/platform/samsung/s5p-mfc/s5p_mfc_intr.c
25
ret = wait_event_interruptible_timeout(dev->queue,
drivers/media/platform/samsung/s5p-mfc/s5p_mfc_intr.c
57
ret = wait_event_interruptible_timeout(ctx->queue,
drivers/media/platform/samsung/s5p-mfc/s5p_mfc_intr.c
62
ret = wait_event_timeout(ctx->queue,
drivers/media/platform/st/stm32/stm32-dcmi.c
1094
if (vb2_is_streaming(&dcmi->queue))
drivers/media/platform/st/stm32/stm32-dcmi.c
151
struct vb2_queue queue;
drivers/media/platform/st/stm32/stm32-dcmi.c
1968
q = &dcmi->queue;
drivers/media/platform/st/stm32/stm32-dcmi.c
1991
dcmi->vdev->queue = &dcmi->queue;
drivers/media/platform/st/stm32/stm32-dcmipp/dcmipp-bytecap.c
124
struct vb2_queue queue;
drivers/media/platform/st/stm32/stm32-dcmipp/dcmipp-bytecap.c
244
if (vb2_is_busy(&vcap->queue))
drivers/media/platform/st/stm32/stm32-dcmipp/dcmipp-bytecap.c
893
q = &vcap->queue;
drivers/media/platform/st/stm32/stm32-dcmipp/dcmipp-bytecap.c
941
vdev->queue = q;
drivers/media/platform/sunxi/sun4i-csi/sun4i_csi.h
151
struct vb2_queue queue;
drivers/media/platform/sunxi/sun4i-csi/sun4i_dma.c
401
struct vb2_queue *q = &csi->queue;
drivers/media/platform/sunxi/sun4i-csi/sun4i_v4l2.c
357
vdev->queue = &csi->queue;
drivers/media/platform/sunxi/sun6i-csi/sun6i_csi_capture.c
1006
queue->type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
drivers/media/platform/sunxi/sun6i-csi/sun6i_csi_capture.c
1007
queue->io_modes = VB2_MMAP | VB2_DMABUF;
drivers/media/platform/sunxi/sun6i-csi/sun6i_csi_capture.c
1008
queue->buf_struct_size = sizeof(struct sun6i_csi_buffer);
drivers/media/platform/sunxi/sun6i-csi/sun6i_csi_capture.c
1009
queue->ops = &sun6i_csi_capture_queue_ops;
drivers/media/platform/sunxi/sun6i-csi/sun6i_csi_capture.c
1010
queue->mem_ops = &vb2_dma_contig_memops;
drivers/media/platform/sunxi/sun6i-csi/sun6i_csi_capture.c
1011
queue->min_queued_buffers = 2;
drivers/media/platform/sunxi/sun6i-csi/sun6i_csi_capture.c
1012
queue->timestamp_flags = V4L2_BUF_FLAG_TIMESTAMP_MONOTONIC;
drivers/media/platform/sunxi/sun6i-csi/sun6i_csi_capture.c
1013
queue->lock = &capture->lock;
drivers/media/platform/sunxi/sun6i-csi/sun6i_csi_capture.c
1014
queue->dev = csi_dev->dev;
drivers/media/platform/sunxi/sun6i-csi/sun6i_csi_capture.c
1015
queue->drv_priv = csi_dev;
drivers/media/platform/sunxi/sun6i-csi/sun6i_csi_capture.c
1017
ret = vb2_queue_init(queue);
drivers/media/platform/sunxi/sun6i-csi/sun6i_csi_capture.c
1025
format->type = queue->type;
drivers/media/platform/sunxi/sun6i-csi/sun6i_csi_capture.c
1043
video_dev->queue = queue;
drivers/media/platform/sunxi/sun6i-csi/sun6i_csi_capture.c
464
list_for_each_entry(csi_buffer, &state->queue, list) {
drivers/media/platform/sunxi/sun6i-csi/sun6i_csi_capture.c
470
INIT_LIST_HEAD(&state->queue);
drivers/media/platform/sunxi/sun6i-csi/sun6i_csi_capture.c
483
if (list_empty(&state->queue))
drivers/media/platform/sunxi/sun6i-csi/sun6i_csi_capture.c
489
csi_buffer = list_first_entry(&state->queue, struct sun6i_csi_buffer,
drivers/media/platform/sunxi/sun6i-csi/sun6i_csi_capture.c
551
static int sun6i_csi_capture_queue_setup(struct vb2_queue *queue,
drivers/media/platform/sunxi/sun6i-csi/sun6i_csi_capture.c
557
struct sun6i_csi_device *csi_dev = vb2_get_drv_priv(queue);
drivers/media/platform/sunxi/sun6i-csi/sun6i_csi_capture.c
600
list_add_tail(&csi_buffer->list, &state->queue);
drivers/media/platform/sunxi/sun6i-csi/sun6i_csi_capture.c
604
static int sun6i_csi_capture_start_streaming(struct vb2_queue *queue,
drivers/media/platform/sunxi/sun6i-csi/sun6i_csi_capture.c
607
struct sun6i_csi_device *csi_dev = vb2_get_drv_priv(queue);
drivers/media/platform/sunxi/sun6i-csi/sun6i_csi_capture.c
638
static void sun6i_csi_capture_stop_streaming(struct vb2_queue *queue)
drivers/media/platform/sunxi/sun6i-csi/sun6i_csi_capture.c
640
struct sun6i_csi_device *csi_dev = vb2_get_drv_priv(queue);
drivers/media/platform/sunxi/sun6i-csi/sun6i_csi_capture.c
757
if (vb2_is_busy(&capture->queue))
drivers/media/platform/sunxi/sun6i-csi/sun6i_csi_capture.c
975
struct vb2_queue *queue = &capture->queue;
drivers/media/platform/sunxi/sun6i-csi/sun6i_csi_capture.c
987
INIT_LIST_HEAD(&state->queue);
drivers/media/platform/sunxi/sun6i-csi/sun6i_csi_capture.h
39
struct list_head queue;
drivers/media/platform/sunxi/sun6i-csi/sun6i_csi_capture.h
55
struct vb2_queue queue;
drivers/media/platform/synopsys/hdmirx/snps_hdmirx.c
100
struct list_head queue;
drivers/media/platform/synopsys/hdmirx/snps_hdmirx.c
1437
static int hdmirx_queue_setup(struct vb2_queue *queue,
drivers/media/platform/synopsys/hdmirx/snps_hdmirx.c
1443
struct hdmirx_stream *stream = vb2_get_drv_priv(queue);
drivers/media/platform/synopsys/hdmirx/snps_hdmirx.c
1474
v4l2_type_names[queue->type], *num_buffers, sizes[0]);
drivers/media/platform/synopsys/hdmirx/snps_hdmirx.c
1490
struct vb2_queue *queue;
drivers/media/platform/synopsys/hdmirx/snps_hdmirx.c
1496
queue = vb->vb2_queue;
drivers/media/platform/synopsys/hdmirx/snps_hdmirx.c
1497
stream = vb2_get_drv_priv(queue);
drivers/media/platform/synopsys/hdmirx/snps_hdmirx.c
1524
list_add_tail(&hdmirx_buf->queue, &stream->buf_head);
drivers/media/platform/synopsys/hdmirx/snps_hdmirx.c
1536
list_add_tail(&stream->curr_buf->queue, &stream->buf_head);
drivers/media/platform/synopsys/hdmirx/snps_hdmirx.c
1538
list_add_tail(&stream->next_buf->queue, &stream->buf_head);
drivers/media/platform/synopsys/hdmirx/snps_hdmirx.c
1542
list_for_each_entry_safe(buf, tmp, &stream->buf_head, queue) {
drivers/media/platform/synopsys/hdmirx/snps_hdmirx.c
1543
list_del(&buf->queue);
drivers/media/platform/synopsys/hdmirx/snps_hdmirx.c
1549
static void hdmirx_stop_streaming(struct vb2_queue *queue)
drivers/media/platform/synopsys/hdmirx/snps_hdmirx.c
1551
struct hdmirx_stream *stream = vb2_get_drv_priv(queue);
drivers/media/platform/synopsys/hdmirx/snps_hdmirx.c
1573
static int hdmirx_start_streaming(struct vb2_queue *queue, unsigned int count)
drivers/media/platform/synopsys/hdmirx/snps_hdmirx.c
1575
struct hdmirx_stream *stream = vb2_get_drv_priv(queue);
drivers/media/platform/synopsys/hdmirx/snps_hdmirx.c
1597
queue);
drivers/media/platform/synopsys/hdmirx/snps_hdmirx.c
1598
list_del(&stream->curr_buf->queue);
drivers/media/platform/synopsys/hdmirx/snps_hdmirx.c
1756
vdev->queue = &stream->buf_queue;
drivers/media/platform/synopsys/hdmirx/snps_hdmirx.c
2057
queue);
drivers/media/platform/synopsys/hdmirx/snps_hdmirx.c
2058
list_del(&stream->next_buf->queue);
drivers/media/platform/ti/am437x/am437x-vpfe.c
2253
vdev->queue = q;
drivers/media/platform/ti/cal/cal-video.c
1048
INIT_LIST_HEAD(&ctx->dma.queue);
drivers/media/platform/ti/cal/cal-video.c
1074
vfd->queue = q;
drivers/media/platform/ti/cal/cal-video.c
647
list_add_tail(&buf->list, &ctx->dma.queue);
drivers/media/platform/ti/cal/cal-video.c
659
list_for_each_entry_safe(buf, tmp, &ctx->dma.queue, list) {
drivers/media/platform/ti/cal/cal-video.c
758
buf = list_first_entry(&ctx->dma.queue, struct cal_buffer, list);
drivers/media/platform/ti/cal/cal.c
637
} else if (!list_empty(&ctx->dma.queue) && !ctx->dma.pending) {
drivers/media/platform/ti/cal/cal.c
645
buf = list_first_entry(&ctx->dma.queue, struct cal_buffer,
drivers/media/platform/ti/cal/cal.h
115
struct list_head queue;
drivers/media/platform/ti/davinci/vpif_capture.c
1449
vdev->queue = q;
drivers/media/platform/ti/davinci/vpif_display.c
1193
vdev->queue = q;
drivers/media/platform/ti/j721e-csi2rx/j721e-csi2rx.c
1008
INIT_LIST_HEAD(&csi->dma.queue);
drivers/media/platform/ti/j721e-csi2rx/j721e-csi2rx.c
650
while (!list_empty(&dma->queue)) {
drivers/media/platform/ti/j721e-csi2rx/j721e-csi2rx.c
651
buf = list_entry(dma->queue.next, struct ti_csi2rx_buffer, list);
drivers/media/platform/ti/j721e-csi2rx/j721e-csi2rx.c
735
list_for_each_entry_safe(buf, tmp, &csi->dma.queue, list) {
drivers/media/platform/ti/j721e-csi2rx/j721e-csi2rx.c
806
list_add_tail(&buf->list, &dma->queue);
drivers/media/platform/ti/j721e-csi2rx/j721e-csi2rx.c
846
if (list_empty(&dma->queue))
drivers/media/platform/ti/j721e-csi2rx/j721e-csi2rx.c
861
buf = list_entry(dma->queue.next, struct ti_csi2rx_buffer, list);
drivers/media/platform/ti/j721e-csi2rx/j721e-csi2rx.c
87
struct list_head queue;
drivers/media/platform/ti/j721e-csi2rx/j721e-csi2rx.c
938
csi->vdev.queue = q;
drivers/media/platform/ti/omap/omap_vout.c
1006
struct omap_vout_buffer, queue);
drivers/media/platform/ti/omap/omap_vout.c
1008
list_del(&vout->cur_frm->queue);
drivers/media/platform/ti/omap/omap_vout.c
1077
list_for_each_entry_safe(buf, tmp, &vout->dma_queue, queue) {
drivers/media/platform/ti/omap/omap_vout.c
1078
list_del(&buf->queue);
drivers/media/platform/ti/omap/omap_vout.c
1112
list_for_each_entry_safe(buf, tmp, &vout->dma_queue, queue) {
drivers/media/platform/ti/omap/omap_vout.c
1113
list_del(&buf->queue);
drivers/media/platform/ti/omap/omap_vout.c
1406
vfd->queue = vq;
drivers/media/platform/ti/omap/omap_vout.c
544
struct omap_vout_buffer, queue);
drivers/media/platform/ti/omap/omap_vout.c
545
list_del(&vout->next_frm->queue);
drivers/media/platform/ti/omap/omap_vout.c
992
list_add_tail(&voutbuf->queue, &vout->dma_queue);
drivers/media/platform/ti/omap/omap_voutdef.h
121
struct list_head queue;
drivers/media/platform/ti/omap3isp/ispccdc.c
1829
.queue = ccdc_video_queue,
drivers/media/platform/ti/omap3isp/ispccdc.c
368
struct list_head *queue)
drivers/media/platform/ti/omap3isp/ispccdc.c
374
list_for_each_entry_safe(req, n, queue, list) {
drivers/media/platform/ti/omap3isp/ispccp2.c
947
.queue = ccp2_video_queue,
drivers/media/platform/ti/omap3isp/ispcsi2.c
824
.queue = csi2_queue,
drivers/media/platform/ti/omap3isp/isppreview.c
1570
.queue = preview_video_queue,
drivers/media/platform/ti/omap3isp/ispresizer.c
1115
.queue = resizer_video_queue,
drivers/media/platform/ti/omap3isp/ispvideo.c
1004
ret = vb2_qbuf(&vfh->queue, video->video.v4l2_dev->mdev, b);
drivers/media/platform/ti/omap3isp/ispvideo.c
1018
ret = vb2_dqbuf(&vfh->queue, b, file->f_flags & O_NONBLOCK);
drivers/media/platform/ti/omap3isp/ispvideo.c
1217
video->queue = &vfh->queue;
drivers/media/platform/ti/omap3isp/ispvideo.c
1223
ret = vb2_streamon(&vfh->queue, type);
drivers/media/platform/ti/omap3isp/ispvideo.c
1244
video->queue = NULL;
drivers/media/platform/ti/omap3isp/ispvideo.c
1271
streaming = vb2_is_streaming(&vfh->queue);
drivers/media/platform/ti/omap3isp/ispvideo.c
1294
vb2_streamoff(&vfh->queue, type);
drivers/media/platform/ti/omap3isp/ispvideo.c
1296
video->queue = NULL;
drivers/media/platform/ti/omap3isp/ispvideo.c
1371
struct vb2_queue *queue;
drivers/media/platform/ti/omap3isp/ispvideo.c
1393
queue = &handle->queue;
drivers/media/platform/ti/omap3isp/ispvideo.c
1394
queue->type = video->type;
drivers/media/platform/ti/omap3isp/ispvideo.c
1395
queue->io_modes = VB2_MMAP | VB2_USERPTR;
drivers/media/platform/ti/omap3isp/ispvideo.c
1396
queue->drv_priv = handle;
drivers/media/platform/ti/omap3isp/ispvideo.c
1397
queue->ops = &isp_video_queue_ops;
drivers/media/platform/ti/omap3isp/ispvideo.c
1398
queue->mem_ops = &vb2_dma_contig_memops;
drivers/media/platform/ti/omap3isp/ispvideo.c
1399
queue->buf_struct_size = sizeof(struct isp_buffer);
drivers/media/platform/ti/omap3isp/ispvideo.c
1400
queue->timestamp_flags = V4L2_BUF_FLAG_TIMESTAMP_MONOTONIC;
drivers/media/platform/ti/omap3isp/ispvideo.c
1401
queue->dev = video->isp->dev;
drivers/media/platform/ti/omap3isp/ispvideo.c
1402
queue->lock = &video->queue_lock;
drivers/media/platform/ti/omap3isp/ispvideo.c
1404
ret = vb2_queue_init(&handle->queue);
drivers/media/platform/ti/omap3isp/ispvideo.c
1444
vb2_queue_release(&handle->queue);
drivers/media/platform/ti/omap3isp/ispvideo.c
1466
ret = vb2_poll(&vfh->queue, file, wait);
drivers/media/platform/ti/omap3isp/ispvideo.c
1476
return vb2_mmap(&vfh->queue, vma);
drivers/media/platform/ti/omap3isp/ispvideo.c
325
static int isp_video_queue_setup(struct vb2_queue *queue,
drivers/media/platform/ti/omap3isp/ispvideo.c
329
struct isp_video_fh *vfh = vb2_get_drv_priv(queue);
drivers/media/platform/ti/omap3isp/ispvideo.c
429
video->ops->queue(video, buffer);
drivers/media/platform/ti/omap3isp/ispvideo.c
467
static int isp_video_start_streaming(struct vb2_queue *queue,
drivers/media/platform/ti/omap3isp/ispvideo.c
470
struct isp_video_fh *vfh = vb2_get_drv_priv(queue);
drivers/media/platform/ti/omap3isp/ispvideo.c
639
vb2_discard_done(video->queue);
drivers/media/platform/ti/omap3isp/ispvideo.c
646
video->ops->queue(video, buf);
drivers/media/platform/ti/omap3isp/ispvideo.c
948
ret = vb2_reqbufs(&vfh->queue, rb);
drivers/media/platform/ti/omap3isp/ispvideo.c
962
ret = vb2_create_bufs(&vfh->queue, p);
drivers/media/platform/ti/omap3isp/ispvideo.c
976
ret = vb2_querybuf(&vfh->queue, b);
drivers/media/platform/ti/omap3isp/ispvideo.c
990
ret = vb2_prepare_buf(&vfh->queue, video->video.v4l2_dev->mdev, b);
drivers/media/platform/ti/omap3isp/ispvideo.h
152
int(*queue)(struct isp_video *video, struct isp_buffer *buffer);
drivers/media/platform/ti/omap3isp/ispvideo.h
178
struct vb2_queue *queue;
drivers/media/platform/ti/omap3isp/ispvideo.h
192
struct vb2_queue queue;
drivers/media/platform/ti/omap3isp/ispvideo.h
203
container_of(q, struct isp_video_fh, queue)
drivers/media/platform/ti/vpe/vip.c
3098
vfd->queue = q;
drivers/media/platform/verisilicon/hantro_postproc.c
185
struct vb2_queue *queue = &m2m_ctx->cap_q_ctx.q;
drivers/media/platform/verisilicon/hantro_postproc.c
188
for (i = 0; i < queue->max_num_buffers; ++i) {
drivers/media/platform/via/via-camera.c
108
struct list_head queue;
drivers/media/platform/via/via-camera.c
1255
cam->vdev.queue = vq;
drivers/media/platform/via/via-camera.c
325
return list_entry(cam->buffer_queue.next, struct via_buffer, queue);
drivers/media/platform/via/via-camera.c
361
list_del(&vb->queue);
drivers/media/platform/via/via-camera.c
574
list_add_tail(&via->queue, &cam->buffer_queue);
drivers/media/platform/via/via-camera.c
642
list_for_each_entry_safe(buf, tmp, &cam->buffer_queue, queue) {
drivers/media/platform/via/via-camera.c
643
list_del(&buf->queue);
drivers/media/platform/via/via-camera.c
657
list_for_each_entry_safe(buf, tmp, &cam->buffer_queue, queue) {
drivers/media/platform/via/via-camera.c
658
list_del(&buf->queue);
drivers/media/platform/xilinx/xilinx-dma.c
277
struct list_head queue;
drivers/media/platform/xilinx/xilinx-dma.c
289
list_del(&buf->queue);
drivers/media/platform/xilinx/xilinx-dma.c
336
if (dma->queue.type == V4L2_BUF_TYPE_VIDEO_CAPTURE) {
drivers/media/platform/xilinx/xilinx-dma.c
365
list_add_tail(&buf->queue, &dma->queued_bufs);
drivers/media/platform/xilinx/xilinx-dma.c
370
if (vb2_is_streaming(&dma->queue))
drivers/media/platform/xilinx/xilinx-dma.c
423
list_for_each_entry_safe(buf, nbuf, &dma->queued_bufs, queue) {
drivers/media/platform/xilinx/xilinx-dma.c
425
list_del(&buf->queue);
drivers/media/platform/xilinx/xilinx-dma.c
450
list_for_each_entry_safe(buf, nbuf, &dma->queued_bufs, queue) {
drivers/media/platform/xilinx/xilinx-dma.c
452
list_del(&buf->queue);
drivers/media/platform/xilinx/xilinx-dma.c
584
if (vb2_is_busy(&dma->queue))
drivers/media/platform/xilinx/xilinx-dma.c
662
dma->video.queue = &dma->queue;
drivers/media/platform/xilinx/xilinx-dma.c
689
dma->queue.type = type;
drivers/media/platform/xilinx/xilinx-dma.c
690
dma->queue.io_modes = VB2_MMAP | VB2_USERPTR | VB2_DMABUF;
drivers/media/platform/xilinx/xilinx-dma.c
691
dma->queue.lock = &dma->lock;
drivers/media/platform/xilinx/xilinx-dma.c
692
dma->queue.drv_priv = dma;
drivers/media/platform/xilinx/xilinx-dma.c
693
dma->queue.buf_struct_size = sizeof(struct xvip_dma_buffer);
drivers/media/platform/xilinx/xilinx-dma.c
694
dma->queue.ops = &xvip_dma_queue_qops;
drivers/media/platform/xilinx/xilinx-dma.c
695
dma->queue.mem_ops = &vb2_dma_contig_memops;
drivers/media/platform/xilinx/xilinx-dma.c
696
dma->queue.timestamp_flags = V4L2_BUF_FLAG_TIMESTAMP_MONOTONIC
drivers/media/platform/xilinx/xilinx-dma.c
698
dma->queue.dev = dma->xdev->dev;
drivers/media/platform/xilinx/xilinx-dma.c
699
ret = vb2_queue_init(&dma->queue);
drivers/media/platform/xilinx/xilinx-dma.h
91
struct vb2_queue queue;
drivers/media/rc/fintek-cir.h
61
wait_queue_head_t queue;
drivers/media/test-drivers/vimc/vimc-capture.c
122
if (vb2_is_busy(&vcapture->queue))
drivers/media/test-drivers/vimc/vimc-capture.c
20
struct vb2_queue queue;
drivers/media/test-drivers/vimc/vimc-capture.c
417
q = &vcapture->queue;
drivers/media/test-drivers/vimc/vimc-capture.c
465
vdev->queue = q;
drivers/media/test-drivers/vivid/vivid-core.c
1440
vfd->queue = &dev->vb_vid_cap_q;
drivers/media/test-drivers/vivid/vivid-core.c
1491
vfd->queue = &dev->vb_vid_out_q;
drivers/media/test-drivers/vivid/vivid-core.c
1539
vfd->queue = &dev->vb_vbi_cap_q;
drivers/media/test-drivers/vivid/vivid-core.c
1571
vfd->queue = &dev->vb_vbi_out_q;
drivers/media/test-drivers/vivid/vivid-core.c
1602
vfd->queue = &dev->vb_sdr_cap_q;
drivers/media/test-drivers/vivid/vivid-core.c
1668
vfd->queue = &dev->vb_meta_cap_q;
drivers/media/test-drivers/vivid/vivid-core.c
1698
vfd->queue = &dev->vb_meta_out_q;
drivers/media/test-drivers/vivid/vivid-core.c
1727
vfd->queue = &dev->vb_touch_cap_q;
drivers/media/test-drivers/vivid/vivid-core.c
666
if (vdev->queue)
drivers/media/test-drivers/vivid/vivid-core.c
704
r = vb2_queue_change_type(vdev->queue, p->type);
drivers/media/test-drivers/vivid/vivid-core.c
724
r = vb2_queue_change_type(vdev->queue, p->format.type);
drivers/media/test-drivers/vivid/vivid-vbi-out.c
167
dev->vbi_out_dev.queue->type = V4L2_BUF_TYPE_VBI_OUTPUT;
drivers/media/test-drivers/vivid/vivid-vbi-out.c
212
dev->vbi_out_dev.queue->type = V4L2_BUF_TYPE_SLICED_VBI_OUTPUT;
drivers/media/usb/airspy/airspy.c
1027
s->vdev.queue = &s->vb_queue;
drivers/media/usb/au0828/au0828-video.c
1973
dev->vdev.queue = &dev->vb_vidq;
drivers/media/usb/au0828/au0828-video.c
1974
dev->vdev.queue->lock = &dev->vb_queue_lock;
drivers/media/usb/au0828/au0828-video.c
1984
dev->vbi_dev.queue = &dev->vb_vbiq;
drivers/media/usb/au0828/au0828-video.c
1985
dev->vbi_dev.queue->lock = &dev->vb_vbi_queue_lock;
drivers/media/usb/cx231xx/cx231xx-417.c
1775
dev->v4l_device.queue = q;
drivers/media/usb/cx231xx/cx231xx-video.c
1817
dev->vdev.queue = q;
drivers/media/usb/cx231xx/cx231xx-video.c
1877
dev->vbi_dev.queue = q;
drivers/media/usb/dvb-usb/cxusb-analog.c
1652
cxdev->videodev->queue = &cxdev->videoqueue;
drivers/media/usb/em28xx/em28xx-video.c
2752
v4l2->vdev.queue = &v4l2->vb_vidq;
drivers/media/usb/em28xx/em28xx-video.c
2753
v4l2->vdev.queue->lock = &v4l2->vb_queue_lock;
drivers/media/usb/em28xx/em28xx-video.c
2795
v4l2->vbi_dev.queue = &v4l2->vb_vbiq;
drivers/media/usb/em28xx/em28xx-video.c
2796
v4l2->vbi_dev.queue->lock = &v4l2->vb_vbi_queue_lock;
drivers/media/usb/go7007/go7007-v4l2.c
1090
vdev->queue = &go->vidq;
drivers/media/usb/gspca/gspca.c
1098
if (vb2_is_busy(&gspca_dev->queue))
drivers/media/usb/gspca/gspca.c
1260
parm->parm.capture.readbuffers = gspca_dev->queue.min_queued_buffers;
drivers/media/usb/gspca/gspca.c
1276
parm->parm.capture.readbuffers = gspca_dev->queue.min_queued_buffers;
drivers/media/usb/gspca/gspca.c
1510
q = &gspca_dev->queue;
drivers/media/usb/gspca/gspca.c
1523
gspca_dev->vdev.queue = q;
drivers/media/usb/gspca/gspca.c
1633
vb2_queue_error(&gspca_dev->queue);
drivers/media/usb/gspca/gspca.c
1660
if (!vb2_start_streaming_called(&gspca_dev->queue))
drivers/media/usb/gspca/gspca.c
1692
streaming = vb2_start_streaming_called(&gspca_dev->queue);
drivers/media/usb/gspca/gspca.h
195
struct vb2_queue queue;
drivers/media/usb/hackrf/hackrf.c
1476
dev->rx_vdev.queue = &dev->rx_vb2_queue;
drivers/media/usb/hackrf/hackrf.c
1495
dev->tx_vdev.queue = &dev->tx_vb2_queue;
drivers/media/usb/msi2500/msi2500.c
1209
dev->vdev.queue = &dev->vb_queue;
drivers/media/usb/pwc/pwc-if.c
1064
pdev->vdev.queue = &pdev->vb_queue;
drivers/media/usb/s2255/s2255drv.c
1639
vc->vdev.queue = q;
drivers/media/usb/stk1160/stk1160-v4l.c
805
dev->vdev.queue = &dev->vb_vidq;
drivers/media/usb/usbtv/usbtv-video.c
944
usbtv->vdev.queue = &usbtv->vb2q;
drivers/media/usb/uvc/uvc_driver.c
2001
if (!video_is_registered(&stream->queue.vdev))
drivers/media/usb/uvc/uvc_driver.c
2004
vb2_video_unregister_device(&stream->queue.vdev);
drivers/media/usb/uvc/uvc_driver.c
2005
vb2_video_unregister_device(&stream->meta.queue.vdev);
drivers/media/usb/uvc/uvc_driver.c
2027
struct uvc_video_queue *queue,
drivers/media/usb/uvc/uvc_driver.c
2032
struct video_device *vdev = &queue->vdev;
drivers/media/usb/uvc/uvc_driver.c
2036
ret = uvc_queue_init(queue, type);
drivers/media/usb/uvc/uvc_driver.c
2052
vdev->queue = &queue->queue;
drivers/media/usb/uvc/uvc_driver.c
2053
vdev->lock = &queue->mutex;
drivers/media/usb/uvc/uvc_driver.c
2114
return uvc_register_video_device(dev, stream, &stream->queue,
drivers/media/usb/uvc/uvc_driver.c
2152
term->vdev = &stream->queue.vdev;
drivers/media/usb/uvc/uvc_driver.c
2426
mutex_lock(&stream->queue.mutex);
drivers/media/usb/uvc/uvc_driver.c
2427
vb2_streamoff(&stream->queue.queue,
drivers/media/usb/uvc/uvc_driver.c
2428
stream->queue.queue.type);
drivers/media/usb/uvc/uvc_driver.c
2429
mutex_unlock(&stream->queue.mutex);
drivers/media/usb/uvc/uvc_isight.c
126
ret = isight_decode(&stream->queue, buf,
drivers/media/usb/uvc/uvc_isight.c
136
buf = uvc_queue_next_buffer(&stream->queue,
drivers/media/usb/uvc/uvc_isight.c
35
static int isight_decode(struct uvc_video_queue *queue, struct uvc_buffer *buf,
drivers/media/usb/uvc/uvc_isight.c
44
struct uvc_streaming *stream = uvc_queue_to_stream(queue);
drivers/media/usb/uvc/uvc_metadata.c
102
if (vb2_is_busy(&stream->meta.queue.queue))
drivers/media/usb/uvc/uvc_metadata.c
117
if (fdesc->type != vfh->vdev->queue->type)
drivers/media/usb/uvc/uvc_metadata.c
229
struct uvc_video_queue *queue = &stream->meta.queue;
drivers/media/usb/uvc/uvc_metadata.c
233
return uvc_register_video_device(dev, stream, queue,
drivers/media/usb/uvc/uvc_metadata.c
49
if (format->type != vfh->vdev->queue->type)
drivers/media/usb/uvc/uvc_metadata.c
67
if (format->type != vfh->vdev->queue->type)
drivers/media/usb/uvc/uvc_queue.c
111
struct uvc_video_queue *queue = vb2_get_drv_priv(vb->vb2_queue);
drivers/media/usb/uvc/uvc_queue.c
116
uvc_dbg(uvc_queue_to_stream(queue)->dev, CAPTURE,
drivers/media/usb/uvc/uvc_queue.c
121
if (unlikely(queue->flags & UVC_QUEUE_DISCONNECTED))
drivers/media/usb/uvc/uvc_queue.c
139
struct uvc_video_queue *queue = vb2_get_drv_priv(vb->vb2_queue);
drivers/media/usb/uvc/uvc_queue.c
143
spin_lock_irqsave(&queue->irqlock, flags);
drivers/media/usb/uvc/uvc_queue.c
144
if (likely(!(queue->flags & UVC_QUEUE_DISCONNECTED))) {
drivers/media/usb/uvc/uvc_queue.c
146
list_add_tail(&buf->queue, &queue->irqqueue);
drivers/media/usb/uvc/uvc_queue.c
156
spin_unlock_irqrestore(&queue->irqlock, flags);
drivers/media/usb/uvc/uvc_queue.c
162
struct uvc_video_queue *queue = vb2_get_drv_priv(vb->vb2_queue);
drivers/media/usb/uvc/uvc_queue.c
163
struct uvc_streaming *stream = uvc_queue_to_stream(queue);
drivers/media/usb/uvc/uvc_queue.c
172
struct uvc_video_queue *queue = vb2_get_drv_priv(vq);
drivers/media/usb/uvc/uvc_queue.c
173
struct uvc_streaming *stream = uvc_queue_to_stream(queue);
drivers/media/usb/uvc/uvc_queue.c
182
queue->buf_used = 0;
drivers/media/usb/uvc/uvc_queue.c
193
uvc_queue_return_buffers(queue, UVC_BUF_STATE_QUEUED);
drivers/media/usb/uvc/uvc_queue.c
199
struct uvc_video_queue *queue = vb2_get_drv_priv(vq);
drivers/media/usb/uvc/uvc_queue.c
200
struct uvc_streaming *stream = uvc_queue_to_stream(queue);
drivers/media/usb/uvc/uvc_queue.c
204
uvc_video_stop_streaming(uvc_queue_to_stream(queue));
drivers/media/usb/uvc/uvc_queue.c
208
uvc_queue_return_buffers(queue, UVC_BUF_STATE_ERROR);
drivers/media/usb/uvc/uvc_queue.c
213
struct uvc_video_queue *queue = vb2_get_drv_priv(vq);
drivers/media/usb/uvc/uvc_queue.c
217
uvc_queue_return_buffers(queue, UVC_BUF_STATE_ERROR);
drivers/media/usb/uvc/uvc_queue.c
241
int uvc_queue_init(struct uvc_video_queue *queue, enum v4l2_buf_type type)
drivers/media/usb/uvc/uvc_queue.c
245
queue->queue.type = type;
drivers/media/usb/uvc/uvc_queue.c
246
queue->queue.io_modes = VB2_MMAP | VB2_USERPTR;
drivers/media/usb/uvc/uvc_queue.c
247
queue->queue.drv_priv = queue;
drivers/media/usb/uvc/uvc_queue.c
248
queue->queue.buf_struct_size = sizeof(struct uvc_buffer);
drivers/media/usb/uvc/uvc_queue.c
249
queue->queue.mem_ops = &vb2_vmalloc_memops;
drivers/media/usb/uvc/uvc_queue.c
250
queue->queue.timestamp_flags = V4L2_BUF_FLAG_TIMESTAMP_MONOTONIC
drivers/media/usb/uvc/uvc_queue.c
252
queue->queue.lock = &queue->mutex;
drivers/media/usb/uvc/uvc_queue.c
256
queue->queue.ops = &uvc_meta_queue_qops;
drivers/media/usb/uvc/uvc_queue.c
259
queue->queue.io_modes |= VB2_DMABUF;
drivers/media/usb/uvc/uvc_queue.c
260
queue->queue.ops = &uvc_queue_qops;
drivers/media/usb/uvc/uvc_queue.c
264
ret = vb2_queue_init(&queue->queue);
drivers/media/usb/uvc/uvc_queue.c
268
mutex_init(&queue->mutex);
drivers/media/usb/uvc/uvc_queue.c
269
spin_lock_init(&queue->irqlock);
drivers/media/usb/uvc/uvc_queue.c
270
INIT_LIST_HEAD(&queue->irqqueue);
drivers/media/usb/uvc/uvc_queue.c
291
void uvc_queue_cancel(struct uvc_video_queue *queue, int disconnect)
drivers/media/usb/uvc/uvc_queue.c
295
spin_lock_irqsave(&queue->irqlock, flags);
drivers/media/usb/uvc/uvc_queue.c
296
__uvc_queue_return_buffers(queue, UVC_BUF_STATE_ERROR);
drivers/media/usb/uvc/uvc_queue.c
305
queue->flags |= UVC_QUEUE_DISCONNECTED;
drivers/media/usb/uvc/uvc_queue.c
306
spin_unlock_irqrestore(&queue->irqlock, flags);
drivers/media/usb/uvc/uvc_queue.c
316
__uvc_queue_get_current_buffer(struct uvc_video_queue *queue)
drivers/media/usb/uvc/uvc_queue.c
318
if (list_empty(&queue->irqqueue))
drivers/media/usb/uvc/uvc_queue.c
321
return list_first_entry(&queue->irqqueue, struct uvc_buffer, queue);
drivers/media/usb/uvc/uvc_queue.c
324
struct uvc_buffer *uvc_queue_get_current_buffer(struct uvc_video_queue *queue)
drivers/media/usb/uvc/uvc_queue.c
329
spin_lock_irqsave(&queue->irqlock, flags);
drivers/media/usb/uvc/uvc_queue.c
330
nextbuf = __uvc_queue_get_current_buffer(queue);
drivers/media/usb/uvc/uvc_queue.c
331
spin_unlock_irqrestore(&queue->irqlock, flags);
drivers/media/usb/uvc/uvc_queue.c
343
static void uvc_queue_buffer_requeue(struct uvc_video_queue *queue,
drivers/media/usb/uvc/uvc_queue.c
358
struct uvc_video_queue *queue = vb2_get_drv_priv(vb->vb2_queue);
drivers/media/usb/uvc/uvc_queue.c
361
uvc_queue_buffer_requeue(queue, buf);
drivers/media/usb/uvc/uvc_queue.c
385
struct uvc_buffer *uvc_queue_next_buffer(struct uvc_video_queue *queue,
drivers/media/usb/uvc/uvc_queue.c
391
spin_lock_irqsave(&queue->irqlock, flags);
drivers/media/usb/uvc/uvc_queue.c
392
list_del(&buf->queue);
drivers/media/usb/uvc/uvc_queue.c
393
nextbuf = __uvc_queue_get_current_buffer(queue);
drivers/media/usb/uvc/uvc_queue.c
394
spin_unlock_irqrestore(&queue->irqlock, flags);
drivers/media/usb/uvc/uvc_queue.c
45
static void __uvc_queue_return_buffers(struct uvc_video_queue *queue,
drivers/media/usb/uvc/uvc_queue.c
52
lockdep_assert_held(&queue->irqlock);
drivers/media/usb/uvc/uvc_queue.c
54
while (!list_empty(&queue->irqqueue)) {
drivers/media/usb/uvc/uvc_queue.c
55
struct uvc_buffer *buf = list_first_entry(&queue->irqqueue,
drivers/media/usb/uvc/uvc_queue.c
57
queue);
drivers/media/usb/uvc/uvc_queue.c
58
list_del(&buf->queue);
drivers/media/usb/uvc/uvc_queue.c
64
static void uvc_queue_return_buffers(struct uvc_video_queue *queue,
drivers/media/usb/uvc/uvc_queue.c
67
spin_lock_irq(&queue->irqlock);
drivers/media/usb/uvc/uvc_queue.c
68
__uvc_queue_return_buffers(queue, state);
drivers/media/usb/uvc/uvc_queue.c
69
spin_unlock_irq(&queue->irqlock);
drivers/media/usb/uvc/uvc_queue.c
80
struct uvc_video_queue *queue = vb2_get_drv_priv(vq);
drivers/media/usb/uvc/uvc_queue.c
90
stream = uvc_queue_to_stream(queue);
drivers/media/usb/uvc/uvc_v4l2.c
436
if (vb2_is_busy(&stream->queue.queue))
drivers/media/usb/uvc/uvc_v4l2.c
506
if (uvc_queue_streaming(&stream->queue))
drivers/media/usb/uvc/uvc_v4l2.c
579
v4l2_fh_init(&handle->vfh, &stream->queue.vdev);
drivers/media/usb/uvc/uvc_v4l2.c
731
if (vb2_is_busy(&stream->queue.queue))
drivers/media/usb/uvc/uvc_video.c
1384
struct uvc_video_queue *queue = &stream->queue;
drivers/media/usb/uvc/uvc_video.c
1389
mem = buf->mem + queue->buf_used;
drivers/media/usb/uvc/uvc_video.c
1390
nbytes = min((unsigned int)len, buf->bytesused - queue->buf_used);
drivers/media/usb/uvc/uvc_video.c
1395
queue->buf_used += nbytes;
drivers/media/usb/uvc/uvc_video.c
1516
*meta_buf = uvc_queue_next_buffer(&stream->meta.queue,
drivers/media/usb/uvc/uvc_video.c
1519
*video_buf = uvc_queue_next_buffer(&stream->queue, *video_buf);
drivers/media/usb/uvc/uvc_video.c
1670
if (buf->bytesused == stream->queue.buf_used ||
drivers/media/usb/uvc/uvc_video.c
1672
if (buf->bytesused == stream->queue.buf_used) {
drivers/media/usb/uvc/uvc_video.c
1673
stream->queue.buf_used = 0;
drivers/media/usb/uvc/uvc_video.c
1676
uvc_queue_next_buffer(&stream->queue, buf);
drivers/media/usb/uvc/uvc_video.c
1691
struct uvc_video_queue *queue = &stream->queue;
drivers/media/usb/uvc/uvc_video.c
1692
struct uvc_video_queue *qmeta = &stream->meta.queue;
drivers/media/usb/uvc/uvc_video.c
1693
struct vb2_queue *vb2_qmeta = stream->meta.queue.vdev.queue;
drivers/media/usb/uvc/uvc_video.c
1714
uvc_queue_cancel(queue, urb->status == -ESHUTDOWN);
drivers/media/usb/uvc/uvc_video.c
1720
buf = uvc_queue_get_current_buffer(queue);
drivers/media/usb/uvc/uvc_video.c
1726
struct uvc_buffer, queue);
drivers/media/usb/uvc/uvc_video.c
2117
if (!uvc_queue_streaming(&stream->queue))
drivers/media/usb/uvc/uvc_video.c
2151
if (!uvc_queue_streaming(&stream->queue))
drivers/media/usb/uvc/uvcvideo.h
316
struct list_head queue;
drivers/media/usb/uvc/uvcvideo.h
335
struct vb2_queue queue;
drivers/media/usb/uvc/uvcvideo.h
477
struct uvc_video_queue queue;
drivers/media/usb/uvc/uvcvideo.h
483
struct uvc_video_queue queue;
drivers/media/usb/uvc/uvcvideo.h
694
int uvc_queue_init(struct uvc_video_queue *queue, enum v4l2_buf_type type);
drivers/media/usb/uvc/uvcvideo.h
695
void uvc_queue_cancel(struct uvc_video_queue *queue, int disconnect);
drivers/media/usb/uvc/uvcvideo.h
696
struct uvc_buffer *uvc_queue_next_buffer(struct uvc_video_queue *queue,
drivers/media/usb/uvc/uvcvideo.h
698
struct uvc_buffer *uvc_queue_get_current_buffer(struct uvc_video_queue *queue);
drivers/media/usb/uvc/uvcvideo.h
700
static inline int uvc_queue_streaming(struct uvc_video_queue *queue)
drivers/media/usb/uvc/uvcvideo.h
702
return vb2_is_streaming(&queue->queue);
drivers/media/usb/uvc/uvcvideo.h
706
uvc_queue_to_stream(struct uvc_video_queue *queue)
drivers/media/usb/uvc/uvcvideo.h
708
return container_of(queue, struct uvc_streaming, queue);
drivers/media/usb/uvc/uvcvideo.h
737
struct uvc_video_queue *queue,
drivers/media/v4l2-core/v4l2-ctrls-request.c
120
.queue = v4l2_ctrl_request_queue,
drivers/media/v4l2-core/v4l2-ioctl.c
3021
if (vdev->queue && vdev->queue->lock &&
drivers/media/v4l2-core/v4l2-ioctl.c
3023
return vdev->queue->lock;
drivers/media/v4l2-core/v4l2-mem2mem.c
1257
INIT_LIST_HEAD(&m2m_ctx->queue);
drivers/media/v4l2-core/v4l2-mem2mem.c
1342
if (!obj->ops->queue)
drivers/media/v4l2-core/v4l2-mem2mem.c
1361
obj->ops->queue(obj);
drivers/media/v4l2-core/v4l2-mem2mem.c
276
struct v4l2_m2m_ctx, queue);
drivers/media/v4l2-core/v4l2-mem2mem.c
367
list_add_tail(&m2m_ctx->queue, &m2m_dev->job_queue);
drivers/media/v4l2-core/v4l2-mem2mem.c
433
list_del(&m2m_ctx->queue);
drivers/media/v4l2-core/v4l2-mem2mem.c
477
list_del(&m2m_dev->curr_ctx->queue);
drivers/media/v4l2-core/v4l2-mem2mem.c
889
list_del(&m2m_ctx->queue);
drivers/memstick/core/ms_block.c
1977
struct memstick_dev *card = hctx->queue->queuedata;
drivers/memstick/core/ms_block.c
2024
blk_mq_stop_hw_queues(msb->queue);
drivers/memstick/core/ms_block.c
2060
blk_mq_start_hw_queues(msb->queue);
drivers/memstick/core/ms_block.c
2105
msb->queue = msb->disk->queue;
drivers/memstick/core/ms_block.c
2192
blk_mq_start_hw_queues(msb->queue);
drivers/memstick/core/ms_block.c
2197
msb->queue = NULL;
drivers/memstick/core/ms_block.h
148
struct request_queue *queue;
drivers/memstick/core/mspro_block.c
1147
msb->queue = msb->disk->queue;
drivers/memstick/core/mspro_block.c
1253
blk_mq_start_hw_queues(msb->queue);
drivers/memstick/core/mspro_block.c
1259
msb->queue = NULL;
drivers/memstick/core/mspro_block.c
1278
blk_mq_stop_hw_queues(msb->queue);
drivers/memstick/core/mspro_block.c
1336
blk_mq_start_hw_queues(msb->queue);
drivers/memstick/core/mspro_block.c
138
struct request_queue *queue;
drivers/memstick/core/mspro_block.c
731
blk_mq_stop_hw_queues(msb->queue);
drivers/memstick/core/mspro_block.c
747
blk_mq_start_hw_queues(msb->queue);
drivers/memstick/core/mspro_block.c
753
struct memstick_dev *card = hctx->queue->queuedata;
drivers/mfd/ipaq-micro.c
100
if (!list_empty(µ->queue)) {
drivers/mfd/ipaq-micro.c
101
micro->msg = list_entry(micro->queue.next,
drivers/mfd/ipaq-micro.c
413
INIT_LIST_HEAD(µ->queue);
drivers/mfd/ipaq-micro.c
69
list_add_tail(&msg->node, µ->queue);
drivers/misc/genwqe/card_base.h
281
struct ddcb_queue queue; /* genwqe DDCB queue */
drivers/misc/genwqe/card_base.h
384
struct ddcb_queue *queue; /* associated queue */
drivers/misc/genwqe/card_ddcb.c
1012
struct ddcb_queue *queue = &cd->queue;
drivers/misc/genwqe/card_ddcb.c
1014
spin_lock_irqsave(&queue->ddcb_lock, flags);
drivers/misc/genwqe/card_ddcb.c
1015
ddcbs_in_flight += queue->ddcbs_in_flight;
drivers/misc/genwqe/card_ddcb.c
1016
spin_unlock_irqrestore(&queue->ddcb_lock, flags);
drivers/misc/genwqe/card_ddcb.c
1021
static int setup_ddcb_queue(struct genwqe_dev *cd, struct ddcb_queue *queue)
drivers/misc/genwqe/card_ddcb.c
1034
queue->ddcbs_in_flight = 0; /* statistics */
drivers/misc/genwqe/card_ddcb.c
1035
queue->ddcbs_max_in_flight = 0;
drivers/misc/genwqe/card_ddcb.c
1036
queue->ddcbs_completed = 0;
drivers/misc/genwqe/card_ddcb.c
1037
queue->return_on_busy = 0;
drivers/misc/genwqe/card_ddcb.c
1038
queue->wait_on_busy = 0;
drivers/misc/genwqe/card_ddcb.c
1040
queue->ddcb_seq = 0x100; /* start sequence number */
drivers/misc/genwqe/card_ddcb.c
1041
queue->ddcb_max = GENWQE_DDCB_MAX;
drivers/misc/genwqe/card_ddcb.c
1042
queue->ddcb_vaddr = __genwqe_alloc_consistent(cd, queue_size,
drivers/misc/genwqe/card_ddcb.c
1043
&queue->ddcb_daddr);
drivers/misc/genwqe/card_ddcb.c
1044
if (queue->ddcb_vaddr == NULL) {
drivers/misc/genwqe/card_ddcb.c
1049
queue->ddcb_req = kzalloc_objs(struct ddcb_requ *, queue->ddcb_max);
drivers/misc/genwqe/card_ddcb.c
1050
if (!queue->ddcb_req) {
drivers/misc/genwqe/card_ddcb.c
1055
queue->ddcb_waitqs = kzalloc_objs(wait_queue_head_t, queue->ddcb_max);
drivers/misc/genwqe/card_ddcb.c
1056
if (!queue->ddcb_waitqs) {
drivers/misc/genwqe/card_ddcb.c
1061
for (i = 0; i < queue->ddcb_max; i++) {
drivers/misc/genwqe/card_ddcb.c
1062
pddcb = &queue->ddcb_vaddr[i]; /* DDCBs */
drivers/misc/genwqe/card_ddcb.c
1066
queue->ddcb_req[i] = NULL; /* requests */
drivers/misc/genwqe/card_ddcb.c
1067
init_waitqueue_head(&queue->ddcb_waitqs[i]); /* waitqueues */
drivers/misc/genwqe/card_ddcb.c
1070
queue->ddcb_act = 0;
drivers/misc/genwqe/card_ddcb.c
1071
queue->ddcb_next = 0; /* queue is empty */
drivers/misc/genwqe/card_ddcb.c
1073
spin_lock_init(&queue->ddcb_lock);
drivers/misc/genwqe/card_ddcb.c
1074
init_waitqueue_head(&queue->busy_waitq);
drivers/misc/genwqe/card_ddcb.c
1076
val64 = ((u64)(queue->ddcb_max - 1) << 8); /* lastptr */
drivers/misc/genwqe/card_ddcb.c
1077
__genwqe_writeq(cd, queue->IO_QUEUE_CONFIG, 0x07); /* iCRC/vCRC */
drivers/misc/genwqe/card_ddcb.c
1078
__genwqe_writeq(cd, queue->IO_QUEUE_SEGMENT, queue->ddcb_daddr);
drivers/misc/genwqe/card_ddcb.c
1079
__genwqe_writeq(cd, queue->IO_QUEUE_INITSQN, queue->ddcb_seq);
drivers/misc/genwqe/card_ddcb.c
1080
__genwqe_writeq(cd, queue->IO_QUEUE_WRAP, val64);
drivers/misc/genwqe/card_ddcb.c
1084
kfree(queue->ddcb_req);
drivers/misc/genwqe/card_ddcb.c
1085
queue->ddcb_req = NULL;
drivers/misc/genwqe/card_ddcb.c
1087
__genwqe_free_consistent(cd, queue_size, queue->ddcb_vaddr,
drivers/misc/genwqe/card_ddcb.c
1088
queue->ddcb_daddr);
drivers/misc/genwqe/card_ddcb.c
1089
queue->ddcb_vaddr = NULL;
drivers/misc/genwqe/card_ddcb.c
1090
queue->ddcb_daddr = 0ull;
drivers/misc/genwqe/card_ddcb.c
1095
static int ddcb_queue_initialized(struct ddcb_queue *queue)
drivers/misc/genwqe/card_ddcb.c
1097
return queue->ddcb_vaddr != NULL;
drivers/misc/genwqe/card_ddcb.c
1100
static void free_ddcb_queue(struct genwqe_dev *cd, struct ddcb_queue *queue)
drivers/misc/genwqe/card_ddcb.c
1104
queue_size = roundup(queue->ddcb_max * sizeof(struct ddcb), PAGE_SIZE);
drivers/misc/genwqe/card_ddcb.c
1106
kfree(queue->ddcb_req);
drivers/misc/genwqe/card_ddcb.c
1107
queue->ddcb_req = NULL;
drivers/misc/genwqe/card_ddcb.c
1109
if (queue->ddcb_vaddr) {
drivers/misc/genwqe/card_ddcb.c
1110
__genwqe_free_consistent(cd, queue_size, queue->ddcb_vaddr,
drivers/misc/genwqe/card_ddcb.c
1111
queue->ddcb_daddr);
drivers/misc/genwqe/card_ddcb.c
1112
queue->ddcb_vaddr = NULL;
drivers/misc/genwqe/card_ddcb.c
1113
queue->ddcb_daddr = 0ull;
drivers/misc/genwqe/card_ddcb.c
1192
genwqe_check_ddcb_queue(cd, &cd->queue);
drivers/misc/genwqe/card_ddcb.c
1228
struct ddcb_queue *queue;
drivers/misc/genwqe/card_ddcb.c
1241
queue = &cd->queue;
drivers/misc/genwqe/card_ddcb.c
1242
queue->IO_QUEUE_CONFIG = IO_SLC_QUEUE_CONFIG;
drivers/misc/genwqe/card_ddcb.c
1243
queue->IO_QUEUE_STATUS = IO_SLC_QUEUE_STATUS;
drivers/misc/genwqe/card_ddcb.c
1244
queue->IO_QUEUE_SEGMENT = IO_SLC_QUEUE_SEGMENT;
drivers/misc/genwqe/card_ddcb.c
1245
queue->IO_QUEUE_INITSQN = IO_SLC_QUEUE_INITSQN;
drivers/misc/genwqe/card_ddcb.c
1246
queue->IO_QUEUE_OFFSET = IO_SLC_QUEUE_OFFSET;
drivers/misc/genwqe/card_ddcb.c
1247
queue->IO_QUEUE_WRAP = IO_SLC_QUEUE_WRAP;
drivers/misc/genwqe/card_ddcb.c
1248
queue->IO_QUEUE_WTIME = IO_SLC_QUEUE_WTIME;
drivers/misc/genwqe/card_ddcb.c
1249
queue->IO_QUEUE_ERRCNTS = IO_SLC_QUEUE_ERRCNTS;
drivers/misc/genwqe/card_ddcb.c
1250
queue->IO_QUEUE_LRW = IO_SLC_QUEUE_LRW;
drivers/misc/genwqe/card_ddcb.c
1252
rc = setup_ddcb_queue(cd, queue);
drivers/misc/genwqe/card_ddcb.c
1300
free_ddcb_queue(cd, queue);
drivers/misc/genwqe/card_ddcb.c
1317
struct ddcb_queue *queue = &cd->queue;
drivers/misc/genwqe/card_ddcb.c
1319
spin_lock_irqsave(&queue->ddcb_lock, flags);
drivers/misc/genwqe/card_ddcb.c
1321
for (i = 0; i < queue->ddcb_max; i++)
drivers/misc/genwqe/card_ddcb.c
1322
wake_up_interruptible(&queue->ddcb_waitqs[queue->ddcb_act]);
drivers/misc/genwqe/card_ddcb.c
1324
wake_up_interruptible(&queue->busy_waitq);
drivers/misc/genwqe/card_ddcb.c
1325
spin_unlock_irqrestore(&queue->ddcb_lock, flags);
drivers/misc/genwqe/card_ddcb.c
1344
struct ddcb_queue *queue = &cd->queue;
drivers/misc/genwqe/card_ddcb.c
1346
if (!ddcb_queue_initialized(queue))
drivers/misc/genwqe/card_ddcb.c
1395
if (!ddcb_queue_initialized(&cd->queue))
drivers/misc/genwqe/card_ddcb.c
1406
free_ddcb_queue(cd, &cd->queue);
drivers/misc/genwqe/card_ddcb.c
163
static void print_ddcb_info(struct genwqe_dev *cd, struct ddcb_queue *queue)
drivers/misc/genwqe/card_ddcb.c
174
cd->card_idx, queue->ddcb_act, queue->ddcb_next);
drivers/misc/genwqe/card_ddcb.c
176
pddcb = queue->ddcb_vaddr;
drivers/misc/genwqe/card_ddcb.c
177
for (i = 0; i < queue->ddcb_max; i++) {
drivers/misc/genwqe/card_ddcb.c
180
i == queue->ddcb_act ? '>' : ' ',
drivers/misc/genwqe/card_ddcb.c
265
static int enqueue_ddcb(struct genwqe_dev *cd, struct ddcb_queue *queue,
drivers/misc/genwqe/card_ddcb.c
282
prev_no = (ddcb_no == 0) ? queue->ddcb_max - 1 : ddcb_no - 1;
drivers/misc/genwqe/card_ddcb.c
283
prev_ddcb = &queue->ddcb_vaddr[prev_no];
drivers/misc/genwqe/card_ddcb.c
312
__genwqe_writeq(cd, queue->IO_QUEUE_OFFSET, num); /* start queue */
drivers/misc/genwqe/card_ddcb.c
332
struct ddcb_queue *queue = req->queue;
drivers/misc/genwqe/card_ddcb.c
333
struct ddcb *pddcb = &queue->ddcb_vaddr[req->num];
drivers/misc/genwqe/card_ddcb.c
348
queue->ddcb_max - 1 : ddcb_no - 1;
drivers/misc/genwqe/card_ddcb.c
349
struct ddcb *prev_pddcb = &queue->ddcb_vaddr[prev_no];
drivers/misc/genwqe/card_ddcb.c
366
struct ddcb_queue *queue)
drivers/misc/genwqe/card_ddcb.c
372
spin_lock_irqsave(&queue->ddcb_lock, flags);
drivers/misc/genwqe/card_ddcb.c
375
while (!queue_empty(queue) && (ddcbs_finished < queue->ddcb_max)) {
drivers/misc/genwqe/card_ddcb.c
381
pddcb = &queue->ddcb_vaddr[queue->ddcb_act];
drivers/misc/genwqe/card_ddcb.c
390
req = queue->ddcb_req[queue->ddcb_act];
drivers/misc/genwqe/card_ddcb.c
408
u64 ddcb_offs = (u64)pddcb - (u64)queue->ddcb_vaddr;
drivers/misc/genwqe/card_ddcb.c
410
errcnts = __genwqe_readq(cd, queue->IO_QUEUE_ERRCNTS);
drivers/misc/genwqe/card_ddcb.c
411
status = __genwqe_readq(cd, queue->IO_QUEUE_STATUS);
drivers/misc/genwqe/card_ddcb.c
417
queue->ddcb_daddr + ddcb_offs);
drivers/misc/genwqe/card_ddcb.c
420
copy_ddcb_results(req, queue->ddcb_act);
drivers/misc/genwqe/card_ddcb.c
421
queue->ddcb_req[queue->ddcb_act] = NULL; /* take from queue */
drivers/misc/genwqe/card_ddcb.c
442
queue->ddcbs_completed++;
drivers/misc/genwqe/card_ddcb.c
443
queue->ddcbs_in_flight--;
drivers/misc/genwqe/card_ddcb.c
447
wake_up_interruptible(&queue->ddcb_waitqs[queue->ddcb_act]);
drivers/misc/genwqe/card_ddcb.c
448
wake_up_interruptible(&queue->busy_waitq);
drivers/misc/genwqe/card_ddcb.c
451
queue->ddcb_act = (queue->ddcb_act + 1) % queue->ddcb_max;
drivers/misc/genwqe/card_ddcb.c
456
spin_unlock_irqrestore(&queue->ddcb_lock, flags);
drivers/misc/genwqe/card_ddcb.c
481
struct ddcb_queue *queue;
drivers/misc/genwqe/card_ddcb.c
487
queue = req->queue;
drivers/misc/genwqe/card_ddcb.c
488
if (queue == NULL)
drivers/misc/genwqe/card_ddcb.c
492
if (ddcb_no >= queue->ddcb_max)
drivers/misc/genwqe/card_ddcb.c
495
rc = wait_event_interruptible_timeout(queue->ddcb_waitqs[ddcb_no],
drivers/misc/genwqe/card_ddcb.c
506
struct ddcb_queue *queue = req->queue;
drivers/misc/genwqe/card_ddcb.c
514
genwqe_check_ddcb_queue(cd, req->queue);
drivers/misc/genwqe/card_ddcb.c
524
__genwqe_readq(cd, queue->IO_QUEUE_STATUS));
drivers/misc/genwqe/card_ddcb.c
526
pddcb = &queue->ddcb_vaddr[req->num];
drivers/misc/genwqe/card_ddcb.c
529
print_ddcb_info(cd, req->queue);
drivers/misc/genwqe/card_ddcb.c
568
struct ddcb_queue *queue,
drivers/misc/genwqe/card_ddcb.c
574
if (queue_free_ddcbs(queue) == 0) /* queue is full */
drivers/misc/genwqe/card_ddcb.c
578
pddcb = &queue->ddcb_vaddr[queue->ddcb_next];
drivers/misc/genwqe/card_ddcb.c
585
*num = queue->ddcb_next; /* internal DDCB number */
drivers/misc/genwqe/card_ddcb.c
586
queue->ddcb_next = (queue->ddcb_next + 1) % queue->ddcb_max;
drivers/misc/genwqe/card_ddcb.c
601
pddcb->seqnum_16 = cpu_to_be16(queue->ddcb_seq++);
drivers/misc/genwqe/card_ddcb.c
625
struct ddcb_queue *queue = req->queue;
drivers/misc/genwqe/card_ddcb.c
638
pddcb = &queue->ddcb_vaddr[req->num];
drivers/misc/genwqe/card_ddcb.c
642
spin_lock_irqsave(&queue->ddcb_lock, flags);
drivers/misc/genwqe/card_ddcb.c
665
spin_unlock_irqrestore(&queue->ddcb_lock, flags);
drivers/misc/genwqe/card_ddcb.c
681
queue->ddcbs_in_flight--;
drivers/misc/genwqe/card_ddcb.c
682
queue->ddcb_req[req->num] = NULL; /* delete from array */
drivers/misc/genwqe/card_ddcb.c
697
(queue->ddcb_act == req->num)) {
drivers/misc/genwqe/card_ddcb.c
698
queue->ddcb_act = ((queue->ddcb_act + 1) %
drivers/misc/genwqe/card_ddcb.c
699
queue->ddcb_max);
drivers/misc/genwqe/card_ddcb.c
702
spin_unlock_irqrestore(&queue->ddcb_lock, flags);
drivers/misc/genwqe/card_ddcb.c
710
queue_status = __genwqe_readq(cd, queue->IO_QUEUE_STATUS);
drivers/misc/genwqe/card_ddcb.c
720
print_ddcb_info(cd, req->queue);
drivers/misc/genwqe/card_ddcb.c
759
struct ddcb_queue *queue;
drivers/misc/genwqe/card_ddcb.c
772
queue = req->queue = &cd->queue;
drivers/misc/genwqe/card_ddcb.c
778
genwqe_check_ddcb_queue(cd, queue);
drivers/misc/genwqe/card_ddcb.c
785
spin_lock_irqsave(&queue->ddcb_lock, flags);
drivers/misc/genwqe/card_ddcb.c
787
pddcb = get_next_ddcb(cd, queue, &req->num); /* get ptr and num */
drivers/misc/genwqe/card_ddcb.c
791
spin_unlock_irqrestore(&queue->ddcb_lock, flags);
drivers/misc/genwqe/card_ddcb.c
794
queue->return_on_busy++;
drivers/misc/genwqe/card_ddcb.c
798
queue->wait_on_busy++;
drivers/misc/genwqe/card_ddcb.c
799
rc = wait_event_interruptible(queue->busy_waitq,
drivers/misc/genwqe/card_ddcb.c
800
queue_free_ddcbs(queue) != 0);
drivers/misc/genwqe/card_ddcb.c
809
if (queue->ddcb_req[req->num] != NULL) {
drivers/misc/genwqe/card_ddcb.c
810
spin_unlock_irqrestore(&queue->ddcb_lock, flags);
drivers/misc/genwqe/card_ddcb.c
818
queue->ddcb_req[req->num] = req;
drivers/misc/genwqe/card_ddcb.c
82
static int queue_empty(struct ddcb_queue *queue)
drivers/misc/genwqe/card_ddcb.c
84
return queue->ddcb_next == queue->ddcb_act;
drivers/misc/genwqe/card_ddcb.c
87
static int queue_enqueued_ddcbs(struct ddcb_queue *queue)
drivers/misc/genwqe/card_ddcb.c
89
if (queue->ddcb_next >= queue->ddcb_act)
drivers/misc/genwqe/card_ddcb.c
892
enqueue_ddcb(cd, queue, pddcb, req->num);
drivers/misc/genwqe/card_ddcb.c
893
queue->ddcbs_in_flight++;
drivers/misc/genwqe/card_ddcb.c
895
if (queue->ddcbs_in_flight > queue->ddcbs_max_in_flight)
drivers/misc/genwqe/card_ddcb.c
896
queue->ddcbs_max_in_flight = queue->ddcbs_in_flight;
drivers/misc/genwqe/card_ddcb.c
899
spin_unlock_irqrestore(&queue->ddcb_lock, flags);
drivers/misc/genwqe/card_ddcb.c
90
return queue->ddcb_next - queue->ddcb_act;
drivers/misc/genwqe/card_ddcb.c
92
return queue->ddcb_max - (queue->ddcb_act - queue->ddcb_next);
drivers/misc/genwqe/card_ddcb.c
95
static int queue_free_ddcbs(struct ddcb_queue *queue)
drivers/misc/genwqe/card_ddcb.c
97
int free_ddcbs = queue->ddcb_max - queue_enqueued_ddcbs(queue) - 1;
drivers/misc/genwqe/card_ddcb.c
981
struct ddcb_queue *queue = &cd->queue;
drivers/misc/genwqe/card_ddcb.c
983
spin_lock_irqsave(&queue->ddcb_lock, flags);
drivers/misc/genwqe/card_ddcb.c
985
if (queue_empty(queue)) { /* empty queue */
drivers/misc/genwqe/card_ddcb.c
986
spin_unlock_irqrestore(&queue->ddcb_lock, flags);
drivers/misc/genwqe/card_ddcb.c
990
pddcb = &queue->ddcb_vaddr[queue->ddcb_act];
drivers/misc/genwqe/card_ddcb.c
992
spin_unlock_irqrestore(&queue->ddcb_lock, flags);
drivers/misc/genwqe/card_ddcb.c
996
spin_unlock_irqrestore(&queue->ddcb_lock, flags);
drivers/misc/genwqe/card_debugfs.c
215
struct ddcb_queue *queue;
drivers/misc/genwqe/card_debugfs.c
218
queue = &cd->queue;
drivers/misc/genwqe/card_debugfs.c
229
queue->ddcb_max, (long long)queue->ddcb_daddr,
drivers/misc/genwqe/card_debugfs.c
230
(long long)queue->ddcb_daddr +
drivers/misc/genwqe/card_debugfs.c
231
(queue->ddcb_max * DDCB_LENGTH),
drivers/misc/genwqe/card_debugfs.c
232
queue->ddcb_vaddr, queue->ddcbs_in_flight,
drivers/misc/genwqe/card_debugfs.c
233
queue->ddcbs_max_in_flight, queue->ddcbs_completed,
drivers/misc/genwqe/card_debugfs.c
234
queue->return_on_busy, queue->wait_on_busy,
drivers/misc/genwqe/card_debugfs.c
247
queue->IO_QUEUE_CONFIG,
drivers/misc/genwqe/card_debugfs.c
248
__genwqe_readq(cd, queue->IO_QUEUE_CONFIG),
drivers/misc/genwqe/card_debugfs.c
249
queue->IO_QUEUE_STATUS,
drivers/misc/genwqe/card_debugfs.c
250
__genwqe_readq(cd, queue->IO_QUEUE_STATUS),
drivers/misc/genwqe/card_debugfs.c
251
queue->IO_QUEUE_SEGMENT,
drivers/misc/genwqe/card_debugfs.c
252
__genwqe_readq(cd, queue->IO_QUEUE_SEGMENT),
drivers/misc/genwqe/card_debugfs.c
253
queue->IO_QUEUE_INITSQN,
drivers/misc/genwqe/card_debugfs.c
254
__genwqe_readq(cd, queue->IO_QUEUE_INITSQN),
drivers/misc/genwqe/card_debugfs.c
255
queue->IO_QUEUE_WRAP,
drivers/misc/genwqe/card_debugfs.c
256
__genwqe_readq(cd, queue->IO_QUEUE_WRAP),
drivers/misc/genwqe/card_debugfs.c
257
queue->IO_QUEUE_OFFSET,
drivers/misc/genwqe/card_debugfs.c
258
__genwqe_readq(cd, queue->IO_QUEUE_OFFSET),
drivers/misc/genwqe/card_debugfs.c
259
queue->IO_QUEUE_WTIME,
drivers/misc/genwqe/card_debugfs.c
260
__genwqe_readq(cd, queue->IO_QUEUE_WTIME),
drivers/misc/genwqe/card_debugfs.c
261
queue->IO_QUEUE_ERRCNTS,
drivers/misc/genwqe/card_debugfs.c
262
__genwqe_readq(cd, queue->IO_QUEUE_ERRCNTS),
drivers/misc/genwqe/card_debugfs.c
263
queue->IO_QUEUE_LRW,
drivers/misc/genwqe/card_debugfs.c
264
__genwqe_readq(cd, queue->IO_QUEUE_LRW));
drivers/misc/genwqe/card_debugfs.c
267
queue->ddcb_act, queue->ddcb_next);
drivers/misc/genwqe/card_debugfs.c
269
pddcb = queue->ddcb_vaddr;
drivers/misc/genwqe/card_debugfs.c
270
for (i = 0; i < queue->ddcb_max; i++) {
drivers/misc/ibmvmc.c
145
struct crq_queue *queue = &adapter->queue;
drivers/misc/ibmvmc.c
155
queue->msg_token,
drivers/misc/ibmvmc.c
156
queue->size * sizeof(*queue->msgs), DMA_BIDIRECTIONAL);
drivers/misc/ibmvmc.c
157
free_page((unsigned long)queue->msgs);
drivers/misc/ibmvmc.c
175
struct crq_queue *queue = &adapter->queue;
drivers/misc/ibmvmc.c
182
memset(queue->msgs, 0x00, PAGE_SIZE);
drivers/misc/ibmvmc.c
183
queue->cur = 0;
drivers/misc/ibmvmc.c
188
queue->msg_token, PAGE_SIZE);
drivers/misc/ibmvmc.c
205
static struct ibmvmc_crq_msg *crq_queue_next_crq(struct crq_queue *queue)
drivers/misc/ibmvmc.c
2077
while ((crq = crq_queue_next_crq(&adapter->queue)) != NULL) {
drivers/misc/ibmvmc.c
2088
crq = crq_queue_next_crq(&adapter->queue);
drivers/misc/ibmvmc.c
210
spin_lock_irqsave(&queue->lock, flags);
drivers/misc/ibmvmc.c
211
crq = &queue->msgs[queue->cur];
drivers/misc/ibmvmc.c
2116
struct crq_queue *queue = &adapter->queue;
drivers/misc/ibmvmc.c
2120
queue->msgs = (struct ibmvmc_crq_msg *)get_zeroed_page(GFP_KERNEL);
drivers/misc/ibmvmc.c
2122
if (!queue->msgs)
drivers/misc/ibmvmc.c
2125
queue->size = PAGE_SIZE / sizeof(*queue->msgs);
drivers/misc/ibmvmc.c
2127
queue->msg_token = dma_map_single(adapter->dev, queue->msgs,
drivers/misc/ibmvmc.c
2128
queue->size * sizeof(*queue->msgs),
drivers/misc/ibmvmc.c
213
if (++queue->cur == queue->size)
drivers/misc/ibmvmc.c
2131
if (dma_mapping_error(adapter->dev, queue->msg_token))
drivers/misc/ibmvmc.c
2136
queue->msg_token, PAGE_SIZE);
drivers/misc/ibmvmc.c
214
queue->cur = 0;
drivers/misc/ibmvmc.c
2150
queue->cur = 0;
drivers/misc/ibmvmc.c
2151
spin_lock_init(&queue->lock);
drivers/misc/ibmvmc.c
2179
queue->msg_token,
drivers/misc/ibmvmc.c
2180
queue->size * sizeof(*queue->msgs), DMA_BIDIRECTIONAL);
drivers/misc/ibmvmc.c
2182
free_page((unsigned long)queue->msgs);
drivers/misc/ibmvmc.c
224
spin_unlock_irqrestore(&queue->lock, flags);
drivers/misc/ibmvmc.c
665
crq_msg.crq_size = cpu_to_be16(adapter->queue.size);
drivers/misc/ibmvmc.h
162
struct crq_queue queue;
drivers/misc/sgi-gru/gru_instructions.h
560
static inline void gru_mesq(void *cb, unsigned long queue,
drivers/misc/sgi-gru/gru_instructions.h
566
ins->baddr0 = (long)queue;
drivers/misc/vmw_vmci/vmci_queue_pair.c
250
struct vmci_queue *queue = q;
drivers/misc/vmw_vmci/vmci_queue_pair.c
252
if (queue) {
drivers/misc/vmw_vmci/vmci_queue_pair.c
258
queue->kernel_if->u.g.vas[i],
drivers/misc/vmw_vmci/vmci_queue_pair.c
259
queue->kernel_if->u.g.pas[i]);
drivers/misc/vmw_vmci/vmci_queue_pair.c
262
vfree(queue);
drivers/misc/vmw_vmci/vmci_queue_pair.c
274
struct vmci_queue *queue;
drivers/misc/vmw_vmci/vmci_queue_pair.c
277
size_t queue_size = sizeof(*queue) + sizeof(*queue->kernel_if);
drivers/misc/vmw_vmci/vmci_queue_pair.c
285
(sizeof(*queue->kernel_if->u.g.pas) +
drivers/misc/vmw_vmci/vmci_queue_pair.c
286
sizeof(*queue->kernel_if->u.g.vas)))
drivers/misc/vmw_vmci/vmci_queue_pair.c
289
pas_size = num_pages * sizeof(*queue->kernel_if->u.g.pas);
drivers/misc/vmw_vmci/vmci_queue_pair.c
290
vas_size = num_pages * sizeof(*queue->kernel_if->u.g.vas);
drivers/misc/vmw_vmci/vmci_queue_pair.c
293
queue = vmalloc(queue_size);
drivers/misc/vmw_vmci/vmci_queue_pair.c
294
if (!queue)
drivers/misc/vmw_vmci/vmci_queue_pair.c
297
queue->q_header = NULL;
drivers/misc/vmw_vmci/vmci_queue_pair.c
298
queue->saved_header = NULL;
drivers/misc/vmw_vmci/vmci_queue_pair.c
299
queue->kernel_if = (struct vmci_queue_kern_if *)(queue + 1);
drivers/misc/vmw_vmci/vmci_queue_pair.c
300
queue->kernel_if->mutex = NULL;
drivers/misc/vmw_vmci/vmci_queue_pair.c
301
queue->kernel_if->num_pages = num_pages;
drivers/misc/vmw_vmci/vmci_queue_pair.c
302
queue->kernel_if->u.g.pas = (dma_addr_t *)(queue->kernel_if + 1);
drivers/misc/vmw_vmci/vmci_queue_pair.c
303
queue->kernel_if->u.g.vas =
drivers/misc/vmw_vmci/vmci_queue_pair.c
304
(void **)((u8 *)queue->kernel_if->u.g.pas + pas_size);
drivers/misc/vmw_vmci/vmci_queue_pair.c
305
queue->kernel_if->host = false;
drivers/misc/vmw_vmci/vmci_queue_pair.c
308
queue->kernel_if->u.g.vas[i] =
drivers/misc/vmw_vmci/vmci_queue_pair.c
310
&queue->kernel_if->u.g.pas[i],
drivers/misc/vmw_vmci/vmci_queue_pair.c
312
if (!queue->kernel_if->u.g.vas[i]) {
drivers/misc/vmw_vmci/vmci_queue_pair.c
314
qp_free_queue(queue, i * PAGE_SIZE);
drivers/misc/vmw_vmci/vmci_queue_pair.c
320
queue->q_header = queue->kernel_if->u.g.vas[0];
drivers/misc/vmw_vmci/vmci_queue_pair.c
322
return queue;
drivers/misc/vmw_vmci/vmci_queue_pair.c
331
static int qp_memcpy_to_queue_iter(struct vmci_queue *queue,
drivers/misc/vmw_vmci/vmci_queue_pair.c
336
struct vmci_queue_kern_if *kernel_if = queue->kernel_if;
drivers/misc/vmw_vmci/vmci_queue_pair.c
380
const struct vmci_queue *queue,
drivers/misc/vmw_vmci/vmci_queue_pair.c
383
struct vmci_queue_kern_if *kernel_if = queue->kernel_if;
drivers/misc/vmw_vmci/vmci_queue_pair.c
528
struct vmci_queue *queue;
drivers/misc/vmw_vmci/vmci_queue_pair.c
531
const size_t queue_size = sizeof(*queue) + sizeof(*(queue->kernel_if));
drivers/misc/vmw_vmci/vmci_queue_pair.c
537
sizeof(*queue->kernel_if->u.h.page))
drivers/misc/vmw_vmci/vmci_queue_pair.c
540
queue_page_size = num_pages * sizeof(*queue->kernel_if->u.h.page);
drivers/misc/vmw_vmci/vmci_queue_pair.c
545
queue = kzalloc(queue_size + queue_page_size, GFP_KERNEL);
drivers/misc/vmw_vmci/vmci_queue_pair.c
546
if (queue) {
drivers/misc/vmw_vmci/vmci_queue_pair.c
547
queue->q_header = NULL;
drivers/misc/vmw_vmci/vmci_queue_pair.c
548
queue->saved_header = NULL;
drivers/misc/vmw_vmci/vmci_queue_pair.c
549
queue->kernel_if = (struct vmci_queue_kern_if *)(queue + 1);
drivers/misc/vmw_vmci/vmci_queue_pair.c
550
queue->kernel_if->host = true;
drivers/misc/vmw_vmci/vmci_queue_pair.c
551
queue->kernel_if->mutex = NULL;
drivers/misc/vmw_vmci/vmci_queue_pair.c
552
queue->kernel_if->num_pages = num_pages;
drivers/misc/vmw_vmci/vmci_queue_pair.c
553
queue->kernel_if->u.h.header_page =
drivers/misc/vmw_vmci/vmci_queue_pair.c
554
(struct page **)((u8 *)queue + queue_size);
drivers/misc/vmw_vmci/vmci_queue_pair.c
555
queue->kernel_if->u.h.page =
drivers/misc/vmw_vmci/vmci_queue_pair.c
556
&queue->kernel_if->u.h.header_page[1];
drivers/misc/vmw_vmci/vmci_queue_pair.c
559
return queue;
drivers/misc/vmw_vmci/vmci_queue_pair.c
566
static void qp_host_free_queue(struct vmci_queue *queue, u64 queue_size)
drivers/misc/vmw_vmci/vmci_queue_pair.c
568
kfree(queue);
drivers/misc/vmw_vmci/vmci_queue_pair.c
610
static void qp_acquire_queue_mutex(struct vmci_queue *queue)
drivers/misc/vmw_vmci/vmci_queue_pair.c
612
if (queue->kernel_if->host)
drivers/misc/vmw_vmci/vmci_queue_pair.c
613
mutex_lock(queue->kernel_if->mutex);
drivers/misc/vmw_vmci/vmci_queue_pair.c
621
static void qp_release_queue_mutex(struct vmci_queue *queue)
drivers/misc/vmw_vmci/vmci_queue_pair.c
623
if (queue->kernel_if->host)
drivers/misc/vmw_vmci/vmci_queue_pair.c
624
mutex_unlock(queue->kernel_if->mutex);
drivers/mmc/core/block.c
117
struct mmc_queue queue;
drivers/mmc/core/block.c
1176
struct mmc_card *card = md->queue.card;
drivers/mmc/core/block.c
1219
struct mmc_card *card = md->queue.card;
drivers/mmc/core/block.c
1232
struct mmc_card *card = md->queue.card;
drivers/mmc/core/block.c
1302
struct mmc_card *card = md->queue.card;
drivers/mmc/core/block.c
1391
struct mmc_card *card = md->queue.card;
drivers/mmc/core/block.c
1458
brq->data.blocks = queue_physical_block_size(mq->queue) >> 9;
drivers/mmc/core/block.c
1789
size_t bytes_per_read = queue_physical_block_size(mq->queue);
drivers/mmc/core/block.c
1973
queue_physical_block_size(mq->queue) >> 9) {
drivers/mmc/core/block.c
238
struct mmc_card *card = md->queue.card;
drivers/mmc/core/block.c
2412
struct mmc_card *card = md->queue.card;
drivers/mmc/core/block.c
2591
md->disk = mmc_init_queue(&md->queue, card, features);
drivers/mmc/core/block.c
2601
md->queue.blkdata = md;
drivers/mmc/core/block.c
2649
blk_mq_free_tag_set(&md->queue.tag_set);
drivers/mmc/core/block.c
269
mq = &md->queue;
drivers/mmc/core/block.c
272
req = blk_mq_alloc_request(mq->queue, REQ_OP_DRV_OUT, 0);
drivers/mmc/core/block.c
2846
if (IS_ERR(md->queue.card))
drivers/mmc/core/block.c
2847
return PTR_ERR(md->queue.card);
drivers/mmc/core/block.c
2908
rq = blk_mq_alloc_request(md->queue.queue, REQ_OP_DRV_OUT, 0);
drivers/mmc/core/block.c
3046
mmc_cleanup_queue(&md->queue);
drivers/mmc/core/block.c
3077
struct mmc_queue *mq = &md->queue;
drivers/mmc/core/block.c
3082
req = blk_mq_alloc_request(mq->queue, REQ_OP_DRV_IN, 0);
drivers/mmc/core/block.c
3107
struct mmc_queue *mq = &md->queue;
drivers/mmc/core/block.c
3119
req = blk_mq_alloc_request(mq->queue, REQ_OP_DRV_IN, 0);
drivers/mmc/core/block.c
3344
mmc_queue_suspend(&md->queue);
drivers/mmc/core/block.c
3346
mmc_queue_suspend(&part_md->queue);
drivers/mmc/core/block.c
3376
mmc_queue_resume(&md->queue);
drivers/mmc/core/block.c
3378
mmc_queue_resume(&part_md->queue);
drivers/mmc/core/block.c
351
md->queue.card->ext_csd.boot_ro_lockable) {
drivers/mmc/core/block.c
353
if (!(md->queue.card->ext_csd.boot_ro_lock &
drivers/mmc/core/block.c
683
card = md->queue.card;
drivers/mmc/core/block.c
692
mq = &md->queue;
drivers/mmc/core/block.c
693
req = blk_mq_alloc_request(mq->queue,
drivers/mmc/core/block.c
755
card = md->queue.card;
drivers/mmc/core/block.c
765
mq = &md->queue;
drivers/mmc/core/block.c
766
req = blk_mq_alloc_request(mq->queue,
drivers/mmc/core/block.c
861
if (md->queue.card)
drivers/mmc/core/block.c
862
ret = mmc_card_alternative_gpt_sector(md->queue.card, sector);
drivers/mmc/core/queue.c
140
struct request_queue *q = mq->queue;
drivers/mmc/core/queue.c
387
mq->queue = disk->queue;
drivers/mmc/core/queue.c
389
blk_queue_rq_timeout(mq->queue, 60 * HZ);
drivers/mmc/core/queue.c
392
dma_set_max_seg_size(mmc_dev(host), queue_max_segment_size(mq->queue));
drivers/mmc/core/queue.c
401
mmc_crypto_setup_queue(mq->queue, host);
drivers/mmc/core/queue.c
474
blk_mq_quiesce_queue(mq->queue);
drivers/mmc/core/queue.c
486
blk_mq_unquiesce_queue(mq->queue);
drivers/mmc/core/queue.c
491
struct request_queue *q = mq->queue;
drivers/mmc/core/queue.h
79
struct request_queue *queue;
drivers/mmc/host/atmel-mci.c
1368
list_add_tail(&slot->queue_node, &host->queue);
drivers/mmc/host/atmel-mci.c
1492
if (list_empty(&host->queue)) {
drivers/mmc/host/atmel-mci.c
1610
if (!list_empty(&host->queue)) {
drivers/mmc/host/atmel-mci.c
1611
slot = list_entry(host->queue.next,
drivers/mmc/host/atmel-mci.c
2466
INIT_LIST_HEAD(&host->queue);
drivers/mmc/host/atmel-mci.c
361
struct list_head queue;
drivers/mmc/host/dw_mmc.c
1411
list_add_tail(&slot->queue_node, &host->queue);
drivers/mmc/host/dw_mmc.c
1905
if (!list_empty(&host->queue)) {
drivers/mmc/host/dw_mmc.c
1906
slot = list_entry(host->queue.next,
drivers/mmc/host/dw_mmc.c
3445
INIT_LIST_HEAD(&host->queue);
drivers/mmc/host/dw_mmc.h
202
struct list_head queue;
drivers/mtd/mtd_blkdevs.c
171
dev = hctx->queue->queuedata;
drivers/mtd/mtd_blkdevs.c
350
new->rq = new->disk->queue;
drivers/mtd/mtd_blkdevs.c
377
gd->queue = new->rq;
drivers/mtd/ubi/block.c
422
dev->rq = gd->queue;
drivers/net/dsa/microchip/ksz_common.c
3394
int queue;
drivers/net/dsa/microchip/ksz_common.c
3400
queue = ieee8021q_tt_to_tc(ipm, dev->info->num_tx_queues);
drivers/net/dsa/microchip/ksz_common.c
3401
if (queue < 0)
drivers/net/dsa/microchip/ksz_common.c
3402
return queue;
drivers/net/dsa/microchip/ksz_common.c
3404
queue_map |= queue << (ipm * KSZ9477_PORT_TC_MAP_S);
drivers/net/dsa/microchip/ksz_common.c
4187
if (qopt->queue > dev->info->num_tx_queues)
drivers/net/dsa/microchip/ksz_common.c
4191
ret = ksz_pwrite32(dev, port, REG_PORT_MTI_QUEUE_INDEX__4, qopt->queue);
drivers/net/dsa/microchip/ksz_common.c
4228
int queue, ret;
drivers/net/dsa/microchip/ksz_common.c
4233
for (queue = 0; queue < dev->info->num_tx_queues; queue++) {
drivers/net/dsa/microchip/ksz_common.c
4234
ret = ksz_pwrite8(dev, port, KSZ9477_REG_PORT_OUT_RATE_0 + queue,
drivers/net/dsa/microchip/ksz_common.c
4253
static u8 ksz8463_tc_ctrl(int port, int queue)
drivers/net/dsa/microchip/ksz_common.c
4258
reg += ((3 - queue) / 2) * 2;
drivers/net/dsa/microchip/ksz_common.c
4260
reg -= (queue & 1);
drivers/net/dsa/microchip/ksz_common.c
4292
int queue = ksz_ets_band_to_queue(p, band);
drivers/net/dsa/microchip/ksz_common.c
4298
reg = KSZ8873_TXQ_SPLIT_CTRL_REG(port, queue);
drivers/net/dsa/microchip/ksz_common.c
4300
reg = ksz8463_tc_ctrl(port, queue);
drivers/net/dsa/microchip/ksz_common.c
4326
int ret, queue;
drivers/net/dsa/microchip/ksz_common.c
4329
for (queue = 0; queue < dev->info->num_tx_queues; queue++) {
drivers/net/dsa/microchip/ksz_common.c
4335
reg = KSZ8873_TXQ_SPLIT_CTRL_REG(port, queue);
drivers/net/dsa/microchip/ksz_common.c
4337
reg = ksz8463_tc_ctrl(port, queue);
drivers/net/dsa/microchip/ksz_common.c
4351
static int ksz_queue_set_strict(struct ksz_device *dev, int port, int queue)
drivers/net/dsa/microchip/ksz_common.c
4355
ret = ksz_pwrite32(dev, port, REG_PORT_MTI_QUEUE_INDEX__4, queue);
drivers/net/dsa/microchip/ksz_common.c
4363
static int ksz_queue_set_wrr(struct ksz_device *dev, int port, int queue,
drivers/net/dsa/microchip/ksz_common.c
4368
ret = ksz_pwrite32(dev, port, REG_PORT_MTI_QUEUE_INDEX__4, queue);
drivers/net/dsa/microchip/ksz_common.c
4398
int queue = ksz_ets_band_to_queue(p, band);
drivers/net/dsa/microchip/ksz_common.c
4400
ret = ksz_queue_set_strict(dev, port, queue);
drivers/net/dsa/microchip/ksz_common.c
4410
int queue;
drivers/net/dsa/microchip/ksz_common.c
4415
queue = ksz_ets_band_to_queue(p, p->priomap[tc_prio]);
drivers/net/dsa/microchip/ksz_common.c
4416
queue_map |= queue << (tc_prio * KSZ9477_PORT_TC_MAP_S);
drivers/net/dsa/microchip/ksz_common.c
4424
int ret, queue;
drivers/net/dsa/microchip/ksz_common.c
4429
for (queue = 0; queue < dev->info->num_tx_queues; queue++) {
drivers/net/dsa/microchip/ksz_common.c
4430
ret = ksz_queue_set_wrr(dev, port, queue,
drivers/net/dsa/microchip/ksz_common.h
882
#define KSZ8873_TXQ_SPLIT_CTRL_REG(port, queue) \
drivers/net/dsa/microchip/ksz_common.h
883
(0xAF + ((port) * 4) + (3 - (queue)))
drivers/net/dsa/ocelot/felix_vsc9959.c
1598
int port_ix = port * 8 + cbs_qopt->queue;
drivers/net/dsa/ocelot/felix_vsc9959.c
1601
if (cbs_qopt->queue >= ds->num_tx_queues)
drivers/net/dsa/sja1105/sja1105_main.c
2178
return sja1105_delete_cbs_shaper(priv, port, offload->queue);
drivers/net/dsa/sja1105/sja1105_main.c
2181
index = sja1105_find_cbs_shaper(priv, port, offload->queue);
drivers/net/dsa/sja1105/sja1105_main.c
2191
cbs->prio = offload->queue;
drivers/net/eql.c
142
static void eql_kill_one_slave(slave_queue_t *queue, slave_t *slave);
drivers/net/eql.c
149
spin_lock(&eql->queue.lock);
drivers/net/eql.c
150
head = &eql->queue.all_slaves;
drivers/net/eql.c
159
eql_kill_one_slave(&eql->queue, slave);
drivers/net/eql.c
163
spin_unlock(&eql->queue.lock);
drivers/net/eql.c
186
spin_lock_init(&eql->queue.lock);
drivers/net/eql.c
187
INIT_LIST_HEAD(&eql->queue.all_slaves);
drivers/net/eql.c
188
eql->queue.master_dev = dev;
drivers/net/eql.c
213
BUG_ON(!list_empty(&eql->queue.all_slaves));
drivers/net/eql.c
223
static void eql_kill_one_slave(slave_queue_t *queue, slave_t *slave)
drivers/net/eql.c
226
queue->num_slaves--;
drivers/net/eql.c
232
static void eql_kill_slave_queue(slave_queue_t *queue)
drivers/net/eql.c
236
spin_lock_bh(&queue->lock);
drivers/net/eql.c
238
head = &queue->all_slaves;
drivers/net/eql.c
242
eql_kill_one_slave(queue, s);
drivers/net/eql.c
245
spin_unlock_bh(&queue->lock);
drivers/net/eql.c
259
eql_kill_slave_queue(&eql->queue);
drivers/net/eql.c
302
static slave_t *__eql_schedule_slaves(slave_queue_t *queue)
drivers/net/eql.c
311
head = &queue->all_slaves;
drivers/net/eql.c
331
eql_kill_one_slave(queue, slave);
drivers/net/eql.c
342
spin_lock(&eql->queue.lock);
drivers/net/eql.c
344
slave = __eql_schedule_slaves(&eql->queue);
drivers/net/eql.c
358
spin_unlock(&eql->queue.lock);
drivers/net/eql.c
368
static slave_t *__eql_find_slave_dev(slave_queue_t *queue, struct net_device *dev)
drivers/net/eql.c
372
head = &queue->all_slaves;
drivers/net/eql.c
383
static inline int eql_is_full(slave_queue_t *queue)
drivers/net/eql.c
385
equalizer_t *eql = netdev_priv(queue->master_dev);
drivers/net/eql.c
387
if (queue->num_slaves >= eql->max_slaves)
drivers/net/eql.c
393
static int __eql_insert_slave(slave_queue_t *queue, slave_t *slave)
drivers/net/eql.c
395
if (!eql_is_full(queue)) {
drivers/net/eql.c
398
duplicate_slave = __eql_find_slave_dev(queue, slave->dev);
drivers/net/eql.c
400
eql_kill_one_slave(queue, duplicate_slave);
drivers/net/eql.c
403
list_add(&slave->list, &queue->all_slaves);
drivers/net/eql.c
404
queue->num_slaves++;
drivers/net/eql.c
440
spin_lock_bh(&eql->queue.lock);
drivers/net/eql.c
441
ret = __eql_insert_slave(&eql->queue, s);
drivers/net/eql.c
445
spin_unlock_bh(&eql->queue.lock);
drivers/net/eql.c
469
spin_lock_bh(&eql->queue.lock);
drivers/net/eql.c
471
slave_t *slave = __eql_find_slave_dev(&eql->queue, slave_dev);
drivers/net/eql.c
473
eql_kill_one_slave(&eql->queue, slave);
drivers/net/eql.c
477
spin_unlock_bh(&eql->queue.lock);
drivers/net/eql.c
499
spin_lock_bh(&eql->queue.lock);
drivers/net/eql.c
501
slave = __eql_find_slave_dev(&eql->queue, slave_dev);
drivers/net/eql.c
507
spin_unlock_bh(&eql->queue.lock);
drivers/net/eql.c
533
spin_lock_bh(&eql->queue.lock);
drivers/net/eql.c
535
slave = __eql_find_slave_dev(&eql->queue, slave_dev);
drivers/net/eql.c
543
spin_unlock_bh(&eql->queue.lock);
drivers/net/ethernet/3com/3c509.c
167
struct sk_buff *queue[SKB_QUEUE_SIZE];
drivers/net/ethernet/airoha/airoha_eth.c
1844
int queue, channel;
drivers/net/ethernet/airoha/airoha_eth.c
1852
queue = (skb->priority - 1) % AIROHA_NUM_QOS_QUEUES; /* QoS queue */
drivers/net/ethernet/airoha/airoha_eth.c
1853
queue = channel * AIROHA_NUM_QOS_QUEUES + queue;
drivers/net/ethernet/airoha/airoha_eth.c
1855
return queue < dev->num_tx_queues ? queue : 0;
drivers/net/ethernet/airoha/airoha_eth.c
214
u32 port, u32 queue)
drivers/net/ethernet/airoha/airoha_eth.c
221
FIELD_PREP(PSE_CFG_QUEUE_ID_MASK, queue));
drivers/net/ethernet/airoha/airoha_eth.c
228
u32 port, u32 queue, u32 val)
drivers/net/ethernet/airoha/airoha_eth.c
236
FIELD_PREP(PSE_CFG_QUEUE_ID_MASK, queue) |
drivers/net/ethernet/airoha/airoha_eth.c
248
u32 port, u32 queue, u32 val)
drivers/net/ethernet/airoha/airoha_eth.c
250
u32 orig_val = airoha_fe_get_pse_queue_rsv_pages(eth, port, queue);
drivers/net/ethernet/airoha/airoha_eth.c
253
airoha_fe_set_pse_queue_rsv_pages(eth, port, queue, val);
drivers/net/ethernet/airoha/airoha_eth.c
2716
static void airoha_tc_remove_htb_queue(struct airoha_gdm_port *port, int queue)
drivers/net/ethernet/airoha/airoha_eth.c
2721
airoha_qdma_set_tx_rate_limit(port, queue + 1, 0, 0);
drivers/net/ethernet/airoha/airoha_eth.c
2722
clear_bit(queue, port->qos_sq_bmap);
drivers/net/ethernet/airoha/airoha_eth.c
918
u16 queue = skb_get_queue_mapping(skb);
drivers/net/ethernet/airoha/airoha_eth.c
921
txq = netdev_get_tx_queue(skb->dev, queue);
drivers/net/ethernet/airoha/airoha_eth.h
496
u8 queue;
drivers/net/ethernet/amazon/ena/ena_com.c
176
static void comp_ctxt_release(struct ena_com_admin_queue *queue,
drivers/net/ethernet/amazon/ena/ena_com.c
180
atomic_dec(&queue->outstanding_cmds);
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
2196
unsigned int queue,
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
2204
if (pdata->pfcq[queue] && (q_fifo_size > pdata->pfc_rfa)) {
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
2220
pdata->rx_rfa[queue] = 0;
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
2221
pdata->rx_rfd[queue] = 0;
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
2227
pdata->rx_rfa[queue] = 0; /* Full - 1024 bytes */
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
2228
pdata->rx_rfd[queue] = 1; /* Full - 1536 bytes */
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
2234
pdata->rx_rfa[queue] = 2; /* Full - 2048 bytes */
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
2235
pdata->rx_rfd[queue] = 5; /* Full - 3584 bytes */
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
2256
pdata->rx_rfa[queue] = XGMAC_FLOW_CONTROL_VALUE(rfa);
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
2257
pdata->rx_rfd[queue] = XGMAC_FLOW_CONTROL_VALUE(rfd);
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
2524
unsigned int qptc, qptc_extra, queue;
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
2536
for (i = 0, queue = 0; i < pdata->hw_feat.tc_cnt; i++) {
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
2539
"TXq%u mapped to TC%u\n", queue, i);
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
2540
XGMAC_MTL_IOWRITE_BITS(pdata, queue, MTL_Q_TQOMR,
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
2542
pdata->q2tc_map[queue++] = i;
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
2547
"TXq%u mapped to TC%u\n", queue, i);
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
2548
XGMAC_MTL_IOWRITE_BITS(pdata, queue, MTL_Q_TQOMR,
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
2550
pdata->q2tc_map[queue++] = i;
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
2605
unsigned int offset, queue, prio;
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
2614
for (i = 0, queue = 0, offset = 0; i < pdata->num_tcs; i++) {
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
2615
while ((queue < pdata->tx_q_count) &&
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
2616
(pdata->q2tc_map[queue] == i))
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
2617
queue++;
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
2620
i, offset, queue - 1);
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
2621
netdev_set_tc_queue(pdata->netdev, i, queue - offset, offset);
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
2622
offset = queue;
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
3154
unsigned int queue)
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
3165
tx_status = XGMAC_MTL_IOREAD(pdata, queue, MTL_Q_TQDR);
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
3176
queue);
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
3180
unsigned int queue)
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
3187
return xgbe_txq_prepare_tx_stop(pdata, queue);
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
3190
if (queue < DMA_DSRX_FIRST_QUEUE) {
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
3192
tx_pos = (queue * DMA_DSR_Q_WIDTH) + DMA_DSR0_TPS_START;
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
3194
tx_qidx = queue - DMA_DSRX_FIRST_QUEUE;
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
3219
queue);
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
3268
unsigned int queue)
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
3279
rx_status = XGMAC_MTL_IOREAD(pdata, queue, MTL_Q_RQDR);
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
3290
queue);
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
370
unsigned int queue)
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
376
if (pdata->prio2q_map[prio] != queue)
drivers/net/ethernet/apple/bmac.c
1259
bp->queue = (struct sk_buff_head *)(bp->rx_cmds + N_RX_RING + 1);
drivers/net/ethernet/apple/bmac.c
1260
skb_queue_head_init(bp->queue);
drivers/net/ethernet/apple/bmac.c
1394
skb = skb_dequeue(bp->queue);
drivers/net/ethernet/apple/bmac.c
1406
skb_queue_tail(bp->queue, skb);
drivers/net/ethernet/apple/bmac.c
59
struct sk_buff_head *queue;
drivers/net/ethernet/aquantia/atlantic/aq_filters.c
319
data->queue = fsp->ring_cookie;
drivers/net/ethernet/aquantia/atlantic/aq_filters.c
321
data->queue = -1;
drivers/net/ethernet/aquantia/atlantic/aq_filters.c
356
aq_vlans[i].queue != AQ_RX_QUEUE_NOT_ASSIGNED &&
drivers/net/ethernet/aquantia/atlantic/aq_filters.c
378
aq_vlans[i].queue != AQ_RX_QUEUE_NOT_ASSIGNED)
drivers/net/ethernet/aquantia/atlantic/aq_filters.c
386
aq_vlans[i].queue = AQ_RX_QUEUE_NOT_ASSIGNED;
drivers/net/ethernet/aquantia/atlantic/aq_filters.c
394
aq_vlans[i].queue = AQ_RX_QUEUE_NOT_ASSIGNED;
drivers/net/ethernet/aquantia/atlantic/aq_filters.c
425
aq_vlans[location].queue = fsp->ring_cookie & 0x1FU;
drivers/net/ethernet/aquantia/atlantic/aq_ptp.c
587
aq_ptp->eth_type_filter.queue = aq_ptp->ptp_rx.idx;
drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c
1507
if (data->queue < 0) {
drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c
1513
hw_atl_rpf_etht_rx_queue_set(self, data->queue, data->location);
drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c
1553
if (aq_vlans[i].queue != 0xFF) {
drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c
1555
aq_vlans[i].queue,
drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_llh.c
602
u32 queue)
drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_llh.c
604
aq_hw_write_reg(aq_hw, HW_ATL_RX_INTR_MODERATION_CTL_ADR(queue),
drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_llh.c
640
u32 queue)
drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_llh.c
642
aq_hw_write_reg(aq_hw, HW_ATL_TX_INTR_MODERATION_CTL_ADR(queue),
drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_llh.h
295
u32 queue);
drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_llh.h
319
u32 queue);
drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_llh_internal.h
2749
#define HW_ATL_TX_INTR_MODERATION_CTL_ADR(queue) (0x00008980u + (queue) * 0x4)
drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_utils.h
360
u8 queue;
drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_utils.h
366
s8 queue;
drivers/net/ethernet/aquantia/atlantic/hw_atl2/hw_atl2.c
774
u32 queue;
drivers/net/ethernet/aquantia/atlantic/hw_atl2/hw_atl2.c
781
queue = HW_ATL2_ACTION_ASSIGN_QUEUE(aq_vlans[i].queue);
drivers/net/ethernet/aquantia/atlantic/hw_atl2/hw_atl2.c
794
if (aq_vlans[i].queue != 0xFF) {
drivers/net/ethernet/aquantia/atlantic/hw_atl2/hw_atl2.c
796
aq_vlans[i].queue,
drivers/net/ethernet/aquantia/atlantic/hw_atl2/hw_atl2.c
806
HW_ATL2_RPF_TAG_VLAN_MASK, queue);
drivers/net/ethernet/aquantia/atlantic/hw_atl2/hw_atl2_llh.c
53
u32 queue)
drivers/net/ethernet/aquantia/atlantic/hw_atl2/hw_atl2_llh.c
58
queue);
drivers/net/ethernet/aquantia/atlantic/hw_atl2/hw_atl2_llh.c
90
u32 queue)
drivers/net/ethernet/aquantia/atlantic/hw_atl2/hw_atl2_llh.c
92
aq_hw_write_reg(aq_hw, HW_ATL2_TX_INTR_MODERATION_CTL_ADR(queue),
drivers/net/ethernet/aquantia/atlantic/hw_atl2/hw_atl2_llh.h
16
u32 queue);
drivers/net/ethernet/aquantia/atlantic/hw_atl2/hw_atl2_llh.h
36
u32 queue);
drivers/net/ethernet/aquantia/atlantic/hw_atl2/hw_atl2_llh_internal.h
125
#define HW_ATL2_RX_Q_TC_MAP_ADR(queue) \
drivers/net/ethernet/aquantia/atlantic/hw_atl2/hw_atl2_llh_internal.h
126
(((queue) < 32) ? 0x00005900 + ((queue) / 8) * 4 : 0)
drivers/net/ethernet/aquantia/atlantic/hw_atl2/hw_atl2_llh_internal.h
128
#define HW_ATL2_RX_Q_TC_MAP_SHIFT(queue) \
drivers/net/ethernet/aquantia/atlantic/hw_atl2/hw_atl2_llh_internal.h
129
(((queue) < 32) ? ((queue) * 4) % 32 : 0)
drivers/net/ethernet/aquantia/atlantic/hw_atl2/hw_atl2_llh_internal.h
178
#define HW_ATL2_TX_Q_TC_MAP_ADR(queue) \
drivers/net/ethernet/aquantia/atlantic/hw_atl2/hw_atl2_llh_internal.h
179
(((queue) < 32) ? 0x0000799C + ((queue) / 4) * 4 : 0)
drivers/net/ethernet/aquantia/atlantic/hw_atl2/hw_atl2_llh_internal.h
181
#define HW_ATL2_TX_Q_TC_MAP_SHIFT(queue) \
drivers/net/ethernet/aquantia/atlantic/hw_atl2/hw_atl2_llh_internal.h
182
(((queue) < 32) ? ((queue) * 8) % 32 : 0)
drivers/net/ethernet/aquantia/atlantic/hw_atl2/hw_atl2_llh_internal.h
250
#define HW_ATL2_TX_INTR_MODERATION_CTL_ADR(queue) (0x00007c28u + (queue) * 0x40)
drivers/net/ethernet/atheros/atl1c/atl1c_main.c
1746
static int atl1c_alloc_rx_buffer(struct atl1c_adapter *adapter, u32 queue,
drivers/net/ethernet/atheros/atl1c/atl1c_main.c
1749
struct atl1c_rfd_ring *rfd_ring = &adapter->rfd_ring[queue];
drivers/net/ethernet/atheros/atl1c/atl1c_main.c
1750
struct atl1c_rrd_ring *rrd_ring = &adapter->rrd_ring[queue];
drivers/net/ethernet/atheros/atl1c/atl1c_main.c
1825
AT_WRITE_REG(&adapter->hw, atl1c_qregs[queue].rfd_prod,
drivers/net/ethernet/atheros/atl1c/atl1c_main.c
1979
static inline u16 atl1c_tpd_avail(struct atl1c_adapter *adapter, u32 queue)
drivers/net/ethernet/atheros/atl1c/atl1c_main.c
1981
struct atl1c_tpd_ring *tpd_ring = &adapter->tpd_ring[queue];
drivers/net/ethernet/atheros/atl1c/atl1c_main.c
1999
u32 queue)
drivers/net/ethernet/atheros/atl1c/atl1c_main.c
2001
struct atl1c_tpd_ring *tpd_ring = &adapter->tpd_ring[queue];
drivers/net/ethernet/atheros/atl1c/atl1c_main.c
2043
u32 queue)
drivers/net/ethernet/atheros/atl1c/atl1c_main.c
2091
*tpd = atl1c_get_tpd(adapter, queue);
drivers/net/ethernet/atheros/atl1c/atl1c_main.c
2143
u32 queue)
drivers/net/ethernet/atheros/atl1c/atl1c_main.c
2145
struct atl1c_tpd_ring *tpd_ring = &adpt->tpd_ring[queue];
drivers/net/ethernet/atheros/atl1c/atl1c_main.c
2165
u32 queue)
drivers/net/ethernet/atheros/atl1c/atl1c_main.c
2206
use_tpd = atl1c_get_tpd(adapter, queue);
drivers/net/ethernet/atheros/atl1c/atl1c_main.c
2228
use_tpd = atl1c_get_tpd(adapter, queue);
drivers/net/ethernet/atheros/atl1c/atl1c_main.c
2261
static void atl1c_tx_queue(struct atl1c_adapter *adapter, u32 queue)
drivers/net/ethernet/atheros/atl1c/atl1c_main.c
2263
struct atl1c_tpd_ring *tpd_ring = &adapter->tpd_ring[queue];
drivers/net/ethernet/atheros/atl1c/atl1c_main.c
2265
AT_WRITE_REGW(&adapter->hw, atl1c_qregs[queue].tpd_prod,
drivers/net/ethernet/atheros/atl1c/atl1c_main.c
2273
u32 queue = skb_get_queue_mapping(skb);
drivers/net/ethernet/atheros/atl1c/atl1c_main.c
2274
struct netdev_queue *txq = netdev_get_tx_queue(netdev, queue);
drivers/net/ethernet/atheros/atl1c/atl1c_main.c
2285
if (atl1c_tpd_avail(adapter, queue) < tpd_req) {
drivers/net/ethernet/atheros/atl1c/atl1c_main.c
2287
atl1c_tx_queue(adapter, queue);
drivers/net/ethernet/atheros/atl1c/atl1c_main.c
2292
tpd = atl1c_get_tpd(adapter, queue);
drivers/net/ethernet/atheros/atl1c/atl1c_main.c
2295
if (atl1c_tso_csum(adapter, skb, &tpd, queue) != 0) {
drivers/net/ethernet/atheros/atl1c/atl1c_main.c
2296
atl1c_tx_queue(adapter, queue);
drivers/net/ethernet/atheros/atl1c/atl1c_main.c
2314
if (atl1c_tx_map(adapter, skb, tpd, queue) < 0) {
drivers/net/ethernet/atheros/atl1c/atl1c_main.c
2318
atl1c_tx_rollback(adapter, tpd, queue);
drivers/net/ethernet/atheros/atl1c/atl1c_main.c
2324
atl1c_tx_queue(adapter, queue);
drivers/net/ethernet/atheros/atl1c/atl1c_main.c
82
static int atl1c_alloc_rx_buffer(struct atl1c_adapter *adapter, u32 queue,
drivers/net/ethernet/atheros/atl1c/atl1c_main.c
876
u32 queue)
drivers/net/ethernet/atheros/atl1c/atl1c_main.c
878
struct atl1c_tpd_ring *tpd_ring = &adapter->tpd_ring[queue];
drivers/net/ethernet/atheros/atl1c/atl1c_main.c
889
netdev_tx_reset_queue(netdev_get_tx_queue(adapter->netdev, queue));
drivers/net/ethernet/atheros/atl1c/atl1c_main.c
903
static void atl1c_clean_rx_ring(struct atl1c_adapter *adapter, u32 queue)
drivers/net/ethernet/atheros/atl1c/atl1c_main.c
905
struct atl1c_rfd_ring *rfd_ring = &adapter->rfd_ring[queue];
drivers/net/ethernet/atheros/atl1c/atl1c_main.c
906
struct atl1c_rrd_ring *rrd_ring = &adapter->rrd_ring[queue];
drivers/net/ethernet/broadcom/bcmsysport.c
1295
u16 queue;
drivers/net/ethernet/broadcom/bcmsysport.c
1298
queue = skb_get_queue_mapping(skb);
drivers/net/ethernet/broadcom/bcmsysport.c
1299
txq = netdev_get_tx_queue(dev, queue);
drivers/net/ethernet/broadcom/bcmsysport.c
1300
ring = &priv->tx_rings[queue];
drivers/net/ethernet/broadcom/bcmsysport.c
1306
netdev_err(dev, "queue %d awake and ring full!\n", queue);
drivers/net/ethernet/broadcom/bcmsysport.c
2268
u16 queue = skb_get_queue_mapping(skb);
drivers/net/ethernet/broadcom/bcmsysport.c
2276
q = BRCM_TAG_GET_QUEUE(queue);
drivers/net/ethernet/broadcom/bcmsysport.c
2277
port = BRCM_TAG_GET_PORT(queue);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
1001
queue);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
430
static void bnx2x_tpa_start(struct bnx2x_fastpath *fp, u16 queue,
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
439
struct bnx2x_agg_info *tpa_info = &fp->tpa_info[queue];
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
444
BNX2X_ERR("start of bin not in stop [%d]\n", queue);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
488
fp->tpa_queue_used |= (1 << queue);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
928
u16 len, pad, queue;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
986
u16 queue = cqe_fp->queue_index;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
989
queue);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
991
bnx2x_tpa_start(fp, queue,
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
997
queue = cqe->end_agg_cqe.queue_index;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
998
tpa_info = &fp->tpa_info[queue];
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.h
496
int bnx2x_set_vf_mac(struct net_device *dev, int queue, u8 *mac);
drivers/net/ethernet/cadence/macb.h
1184
int (*mog_rx)(struct macb_queue *queue, struct napi_struct *napi,
drivers/net/ethernet/cadence/macb.h
819
#define queue_readl(queue, reg) (queue)->bp->macb_reg_readl((queue)->bp, (queue)->reg)
drivers/net/ethernet/cadence/macb.h
820
#define queue_writel(queue, reg, value) (queue)->bp->macb_reg_writel((queue)->bp, (queue)->reg, (value))
drivers/net/ethernet/cadence/macb_main.c
1114
struct macb_queue *queue = container_of(work, struct macb_queue,
drivers/net/ethernet/cadence/macb_main.c
1117
struct macb *bp = queue->bp;
drivers/net/ethernet/cadence/macb_main.c
1127
queue_index = queue - bp->queues;
drivers/net/ethernet/cadence/macb_main.c
1129
queue_index, queue->tx_tail, queue->tx_head);
drivers/net/ethernet/cadence/macb_main.c
1137
napi_disable(&queue->napi_tx);
drivers/net/ethernet/cadence/macb_main.c
1156
for (tail = queue->tx_tail; tail != queue->tx_head; tail++) {
drivers/net/ethernet/cadence/macb_main.c
1159
desc = macb_tx_desc(queue, tail);
drivers/net/ethernet/cadence/macb_main.c
1161
tx_skb = macb_tx_skb(queue, tail);
drivers/net/ethernet/cadence/macb_main.c
1169
tx_skb = macb_tx_skb(queue, tail);
drivers/net/ethernet/cadence/macb_main.c
1181
queue->stats.tx_packets++;
drivers/net/ethernet/cadence/macb_main.c
1184
queue->stats.tx_bytes += skb->len;
drivers/net/ethernet/cadence/macb_main.c
1206
desc = macb_tx_desc(queue, 0);
drivers/net/ethernet/cadence/macb_main.c
1214
queue_writel(queue, TBQP, lower_32_bits(queue->tx_ring_dma));
drivers/net/ethernet/cadence/macb_main.c
1216
queue->tx_head = 0;
drivers/net/ethernet/cadence/macb_main.c
1217
queue->tx_tail = 0;
drivers/net/ethernet/cadence/macb_main.c
1221
queue_writel(queue, IER, MACB_TX_INT_FLAGS);
drivers/net/ethernet/cadence/macb_main.c
1231
napi_enable(&queue->napi_tx);
drivers/net/ethernet/cadence/macb_main.c
1264
static int macb_tx_complete(struct macb_queue *queue, int budget)
drivers/net/ethernet/cadence/macb_main.c
1266
struct macb *bp = queue->bp;
drivers/net/ethernet/cadence/macb_main.c
1267
u16 queue_index = queue - bp->queues;
drivers/net/ethernet/cadence/macb_main.c
1274
spin_lock_irqsave(&queue->tx_ptr_lock, flags);
drivers/net/ethernet/cadence/macb_main.c
1275
head = queue->tx_head;
drivers/net/ethernet/cadence/macb_main.c
1276
for (tail = queue->tx_tail; tail != head && packets < budget; tail++) {
drivers/net/ethernet/cadence/macb_main.c
1282
desc = macb_tx_desc(queue, tail);
drivers/net/ethernet/cadence/macb_main.c
1297
tx_skb = macb_tx_skb(queue, tail);
drivers/net/ethernet/cadence/macb_main.c
1310
queue->stats.tx_packets++;
drivers/net/ethernet/cadence/macb_main.c
1312
queue->stats.tx_bytes += skb->len;
drivers/net/ethernet/cadence/macb_main.c
1332
queue->tx_tail = tail;
drivers/net/ethernet/cadence/macb_main.c
1334
CIRC_CNT(queue->tx_head, queue->tx_tail,
drivers/net/ethernet/cadence/macb_main.c
1337
spin_unlock_irqrestore(&queue->tx_ptr_lock, flags);
drivers/net/ethernet/cadence/macb_main.c
1342
static void gem_rx_refill(struct macb_queue *queue)
drivers/net/ethernet/cadence/macb_main.c
1347
struct macb *bp = queue->bp;
drivers/net/ethernet/cadence/macb_main.c
1350
while (CIRC_SPACE(queue->rx_prepared_head, queue->rx_tail,
drivers/net/ethernet/cadence/macb_main.c
1352
entry = macb_rx_ring_wrap(bp, queue->rx_prepared_head);
drivers/net/ethernet/cadence/macb_main.c
1357
desc = macb_rx_desc(queue, entry);
drivers/net/ethernet/cadence/macb_main.c
1359
if (!queue->rx_skbuff[entry]) {
drivers/net/ethernet/cadence/macb_main.c
1377
queue->rx_skbuff[entry] = skb;
drivers/net/ethernet/cadence/macb_main.c
1406
queue->rx_prepared_head++;
drivers/net/ethernet/cadence/macb_main.c
1413
queue, queue->rx_prepared_head, queue->rx_tail);
drivers/net/ethernet/cadence/macb_main.c
1417
static void discard_partial_frame(struct macb_queue *queue, unsigned int begin,
drivers/net/ethernet/cadence/macb_main.c
1423
struct macb_dma_desc *desc = macb_rx_desc(queue, frag);
drivers/net/ethernet/cadence/macb_main.c
1437
static int gem_rx(struct macb_queue *queue, struct napi_struct *napi,
drivers/net/ethernet/cadence/macb_main.c
1440
struct macb *bp = queue->bp;
drivers/net/ethernet/cadence/macb_main.c
1452
entry = macb_rx_ring_wrap(bp, queue->rx_tail);
drivers/net/ethernet/cadence/macb_main.c
1453
desc = macb_rx_desc(queue, entry);
drivers/net/ethernet/cadence/macb_main.c
1469
queue->rx_tail++;
drivers/net/ethernet/cadence/macb_main.c
1476
queue->stats.rx_dropped++;
drivers/net/ethernet/cadence/macb_main.c
1479
skb = queue->rx_skbuff[entry];
drivers/net/ethernet/cadence/macb_main.c
1484
queue->stats.rx_dropped++;
drivers/net/ethernet/cadence/macb_main.c
1488
queue->rx_skbuff[entry] = NULL;
drivers/net/ethernet/cadence/macb_main.c
1505
queue->stats.rx_packets++;
drivers/net/ethernet/cadence/macb_main.c
1507
queue->stats.rx_bytes += skb->len;
drivers/net/ethernet/cadence/macb_main.c
152
static struct macb_dma_desc *macb_tx_desc(struct macb_queue *queue,
drivers/net/ethernet/cadence/macb_main.c
1523
gem_rx_refill(queue);
drivers/net/ethernet/cadence/macb_main.c
1528
static int macb_rx_frame(struct macb_queue *queue, struct napi_struct *napi,
drivers/net/ethernet/cadence/macb_main.c
1536
struct macb *bp = queue->bp;
drivers/net/ethernet/cadence/macb_main.c
1538
desc = macb_rx_desc(queue, last_frag);
drivers/net/ethernet/cadence/macb_main.c
155
index = macb_tx_ring_wrap(queue->bp, index);
drivers/net/ethernet/cadence/macb_main.c
1557
desc = macb_rx_desc(queue, frag);
drivers/net/ethernet/cadence/macb_main.c
156
index = macb_adj_dma_desc_idx(queue->bp, index);
drivers/net/ethernet/cadence/macb_main.c
157
return &queue->tx_ring[index];
drivers/net/ethernet/cadence/macb_main.c
1585
macb_rx_buffer(queue, frag),
drivers/net/ethernet/cadence/macb_main.c
1588
desc = macb_rx_desc(queue, frag);
drivers/net/ethernet/cadence/macb_main.c
160
static struct macb_tx_skb *macb_tx_skb(struct macb_queue *queue,
drivers/net/ethernet/cadence/macb_main.c
1610
static inline void macb_init_rx_ring(struct macb_queue *queue)
drivers/net/ethernet/cadence/macb_main.c
1612
struct macb *bp = queue->bp;
drivers/net/ethernet/cadence/macb_main.c
1617
addr = queue->rx_buffers_dma;
drivers/net/ethernet/cadence/macb_main.c
1619
desc = macb_rx_desc(queue, i);
drivers/net/ethernet/cadence/macb_main.c
1625
queue->rx_tail = 0;
drivers/net/ethernet/cadence/macb_main.c
1628
static int macb_rx(struct macb_queue *queue, struct napi_struct *napi,
drivers/net/ethernet/cadence/macb_main.c
163
return &queue->tx_skb[macb_tx_ring_wrap(queue->bp, index)];
drivers/net/ethernet/cadence/macb_main.c
1631
struct macb *bp = queue->bp;
drivers/net/ethernet/cadence/macb_main.c
1637
for (tail = queue->rx_tail; budget > 0; tail++) {
drivers/net/ethernet/cadence/macb_main.c
1638
struct macb_dma_desc *desc = macb_rx_desc(queue, tail);
drivers/net/ethernet/cadence/macb_main.c
1654
discard_partial_frame(queue, first_frag, tail);
drivers/net/ethernet/cadence/macb_main.c
166
static dma_addr_t macb_tx_dma(struct macb_queue *queue, unsigned int index)
drivers/net/ethernet/cadence/macb_main.c
1666
dropped = macb_rx_frame(queue, napi, first_frag, tail);
drivers/net/ethernet/cadence/macb_main.c
1690
macb_init_rx_ring(queue);
drivers/net/ethernet/cadence/macb_main.c
1691
queue_writel(queue, RBQP, queue->rx_ring_dma);
drivers/net/ethernet/cadence/macb_main.c
170
offset = macb_tx_ring_wrap(queue->bp, index) *
drivers/net/ethernet/cadence/macb_main.c
1700
queue->rx_tail = first_frag;
drivers/net/ethernet/cadence/macb_main.c
1702
queue->rx_tail = tail;
drivers/net/ethernet/cadence/macb_main.c
1707
static bool macb_rx_pending(struct macb_queue *queue)
drivers/net/ethernet/cadence/macb_main.c
1709
struct macb *bp = queue->bp;
drivers/net/ethernet/cadence/macb_main.c
171
macb_dma_desc_get_size(queue->bp);
drivers/net/ethernet/cadence/macb_main.c
1713
entry = macb_rx_ring_wrap(bp, queue->rx_tail);
drivers/net/ethernet/cadence/macb_main.c
1714
desc = macb_rx_desc(queue, entry);
drivers/net/ethernet/cadence/macb_main.c
1724
struct macb_queue *queue = container_of(napi, struct macb_queue, napi_rx);
drivers/net/ethernet/cadence/macb_main.c
1725
struct macb *bp = queue->bp;
drivers/net/ethernet/cadence/macb_main.c
1728
work_done = bp->macbgem_ops.mog_rx(queue, napi, budget);
drivers/net/ethernet/cadence/macb_main.c
173
return queue->tx_ring_dma + offset;
drivers/net/ethernet/cadence/macb_main.c
1731
(unsigned int)(queue - bp->queues), work_done, budget);
drivers/net/ethernet/cadence/macb_main.c
1734
queue_writel(queue, IER, bp->rx_intr_mask);
drivers/net/ethernet/cadence/macb_main.c
1746
if (macb_rx_pending(queue)) {
drivers/net/ethernet/cadence/macb_main.c
1747
queue_writel(queue, IDR, bp->rx_intr_mask);
drivers/net/ethernet/cadence/macb_main.c
1749
queue_writel(queue, ISR, MACB_BIT(RCOMP));
drivers/net/ethernet/cadence/macb_main.c
1760
static void macb_tx_restart(struct macb_queue *queue)
drivers/net/ethernet/cadence/macb_main.c
1762
struct macb *bp = queue->bp;
drivers/net/ethernet/cadence/macb_main.c
1766
spin_lock_irqsave(&queue->tx_ptr_lock, flags);
drivers/net/ethernet/cadence/macb_main.c
1768
if (queue->tx_head == queue->tx_tail)
drivers/net/ethernet/cadence/macb_main.c
1771
tbqp = queue_readl(queue, TBQP) / macb_dma_desc_get_size(bp);
drivers/net/ethernet/cadence/macb_main.c
1773
head_idx = macb_adj_dma_desc_idx(bp, macb_tx_ring_wrap(bp, queue->tx_head));
drivers/net/ethernet/cadence/macb_main.c
1783
spin_unlock_irqrestore(&queue->tx_ptr_lock, flags);
drivers/net/ethernet/cadence/macb_main.c
1786
static bool macb_tx_complete_pending(struct macb_queue *queue)
drivers/net/ethernet/cadence/macb_main.c
1791
spin_lock_irqsave(&queue->tx_ptr_lock, flags);
drivers/net/ethernet/cadence/macb_main.c
1792
if (queue->tx_head != queue->tx_tail) {
drivers/net/ethernet/cadence/macb_main.c
1796
if (macb_tx_desc(queue, queue->tx_tail)->ctrl & MACB_BIT(TX_USED))
drivers/net/ethernet/cadence/macb_main.c
1799
spin_unlock_irqrestore(&queue->tx_ptr_lock, flags);
drivers/net/ethernet/cadence/macb_main.c
1805
struct macb_queue *queue = container_of(napi, struct macb_queue, napi_tx);
drivers/net/ethernet/cadence/macb_main.c
1806
struct macb *bp = queue->bp;
drivers/net/ethernet/cadence/macb_main.c
1809
work_done = macb_tx_complete(queue, budget);
drivers/net/ethernet/cadence/macb_main.c
181
static struct macb_dma_desc *macb_rx_desc(struct macb_queue *queue, unsigned int index)
drivers/net/ethernet/cadence/macb_main.c
1812
if (queue->txubr_pending) {
drivers/net/ethernet/cadence/macb_main.c
1813
queue->txubr_pending = false;
drivers/net/ethernet/cadence/macb_main.c
1815
macb_tx_restart(queue);
drivers/net/ethernet/cadence/macb_main.c
1819
(unsigned int)(queue - bp->queues), work_done, budget);
drivers/net/ethernet/cadence/macb_main.c
1822
queue_writel(queue, IER, MACB_BIT(TCOMP));
drivers/net/ethernet/cadence/macb_main.c
183
index = macb_rx_ring_wrap(queue->bp, index);
drivers/net/ethernet/cadence/macb_main.c
1834
if (macb_tx_complete_pending(queue)) {
drivers/net/ethernet/cadence/macb_main.c
1835
queue_writel(queue, IDR, MACB_BIT(TCOMP));
drivers/net/ethernet/cadence/macb_main.c
1837
queue_writel(queue, ISR, MACB_BIT(TCOMP));
drivers/net/ethernet/cadence/macb_main.c
184
index = macb_adj_dma_desc_idx(queue->bp, index);
drivers/net/ethernet/cadence/macb_main.c
185
return &queue->rx_ring[index];
drivers/net/ethernet/cadence/macb_main.c
1850
struct macb_queue *queue;
drivers/net/ethernet/cadence/macb_main.c
1854
for (q = 0, queue = bp->queues; q < bp->num_queues; ++q, ++queue) {
drivers/net/ethernet/cadence/macb_main.c
1855
queue_writel(queue, IDR, bp->rx_intr_mask |
drivers/net/ethernet/cadence/macb_main.c
1872
for (q = 0, queue = bp->queues; q < bp->num_queues; ++q, ++queue)
drivers/net/ethernet/cadence/macb_main.c
1873
queue_writel(queue, IER,
drivers/net/ethernet/cadence/macb_main.c
188
static void *macb_rx_buffer(struct macb_queue *queue, unsigned int index)
drivers/net/ethernet/cadence/macb_main.c
1887
struct macb_queue *queue = dev_id;
drivers/net/ethernet/cadence/macb_main.c
1888
struct macb *bp = queue->bp;
drivers/net/ethernet/cadence/macb_main.c
1891
status = queue_readl(queue, ISR);
drivers/net/ethernet/cadence/macb_main.c
1899
queue_writel(queue, IDR, MACB_BIT(WOL));
drivers/net/ethernet/cadence/macb_main.c
190
return queue->rx_buffers + queue->bp->rx_buffer_size *
drivers/net/ethernet/cadence/macb_main.c
1902
(unsigned int)(queue - bp->queues),
drivers/net/ethernet/cadence/macb_main.c
1905
queue_writel(queue, ISR, MACB_BIT(WOL));
drivers/net/ethernet/cadence/macb_main.c
191
macb_rx_ring_wrap(queue->bp, index);
drivers/net/ethernet/cadence/macb_main.c
1916
struct macb_queue *queue = dev_id;
drivers/net/ethernet/cadence/macb_main.c
1917
struct macb *bp = queue->bp;
drivers/net/ethernet/cadence/macb_main.c
1920
status = queue_readl(queue, ISR);
drivers/net/ethernet/cadence/macb_main.c
1928
queue_writel(queue, IDR, GEM_BIT(WOL));
drivers/net/ethernet/cadence/macb_main.c
1931
(unsigned int)(queue - bp->queues),
drivers/net/ethernet/cadence/macb_main.c
1934
queue_writel(queue, ISR, GEM_BIT(WOL));
drivers/net/ethernet/cadence/macb_main.c
1945
struct macb_queue *queue = dev_id;
drivers/net/ethernet/cadence/macb_main.c
1946
struct macb *bp = queue->bp;
drivers/net/ethernet/cadence/macb_main.c
1950
status = queue_readl(queue, ISR);
drivers/net/ethernet/cadence/macb_main.c
1960
queue_writel(queue, IDR, -1);
drivers/net/ethernet/cadence/macb_main.c
1962
queue_writel(queue, ISR, -1);
drivers/net/ethernet/cadence/macb_main.c
1967
(unsigned int)(queue - bp->queues),
drivers/net/ethernet/cadence/macb_main.c
1977
queue_writel(queue, IDR, bp->rx_intr_mask);
drivers/net/ethernet/cadence/macb_main.c
1979
queue_writel(queue, ISR, MACB_BIT(RCOMP));
drivers/net/ethernet/cadence/macb_main.c
1981
if (napi_schedule_prep(&queue->napi_rx)) {
drivers/net/ethernet/cadence/macb_main.c
1983
__napi_schedule(&queue->napi_rx);
drivers/net/ethernet/cadence/macb_main.c
1989
queue_writel(queue, IDR, MACB_BIT(TCOMP));
drivers/net/ethernet/cadence/macb_main.c
1991
queue_writel(queue, ISR, MACB_BIT(TCOMP) |
drivers/net/ethernet/cadence/macb_main.c
1995
queue->txubr_pending = true;
drivers/net/ethernet/cadence/macb_main.c
1999
if (napi_schedule_prep(&queue->napi_tx)) {
drivers/net/ethernet/cadence/macb_main.c
2001
__napi_schedule(&queue->napi_tx);
drivers/net/ethernet/cadence/macb_main.c
2006
queue_writel(queue, IDR, MACB_TX_INT_FLAGS);
drivers/net/ethernet/cadence/macb_main.c
2007
schedule_work(&queue->tx_error_task);
drivers/net/ethernet/cadence/macb_main.c
2010
queue_writel(queue, ISR, MACB_TX_ERR_FLAGS);
drivers/net/ethernet/cadence/macb_main.c
2033
queue_writel(queue, ISR, MACB_BIT(RXUBR));
drivers/net/ethernet/cadence/macb_main.c
2046
queue_writel(queue, ISR, MACB_BIT(ISR_ROVR));
drivers/net/ethernet/cadence/macb_main.c
2054
queue_writel(queue, ISR, MACB_BIT(HRESP));
drivers/net/ethernet/cadence/macb_main.c
2056
status = queue_readl(queue, ISR);
drivers/net/ethernet/cadence/macb_main.c
2071
struct macb_queue *queue;
drivers/net/ethernet/cadence/macb_main.c
2076
for (q = 0, queue = bp->queues; q < bp->num_queues; ++q, ++queue)
drivers/net/ethernet/cadence/macb_main.c
2077
macb_interrupt(dev->irq, queue);
drivers/net/ethernet/cadence/macb_main.c
2083
struct macb_queue *queue,
drivers/net/ethernet/cadence/macb_main.c
2088
unsigned int len, i, tx_head = queue->tx_head;
drivers/net/ethernet/cadence/macb_main.c
2114
tx_skb = macb_tx_skb(queue, tx_head);
drivers/net/ethernet/cadence/macb_main.c
2143
tx_skb = macb_tx_skb(queue, tx_head);
drivers/net/ethernet/cadence/macb_main.c
2180
desc = macb_tx_desc(queue, i);
drivers/net/ethernet/cadence/macb_main.c
2200
tx_skb = macb_tx_skb(queue, i);
drivers/net/ethernet/cadence/macb_main.c
2201
desc = macb_tx_desc(queue, i);
drivers/net/ethernet/cadence/macb_main.c
2212
if (i == queue->tx_head) {
drivers/net/ethernet/cadence/macb_main.c
2232
} while (i != queue->tx_head);
drivers/net/ethernet/cadence/macb_main.c
2234
queue->tx_head = tx_head;
drivers/net/ethernet/cadence/macb_main.c
2241
for (i = queue->tx_head; i != tx_head; i++) {
drivers/net/ethernet/cadence/macb_main.c
2242
tx_skb = macb_tx_skb(queue, i);
drivers/net/ethernet/cadence/macb_main.c
2358
struct macb_queue *queue = &bp->queues[queue_index];
drivers/net/ethernet/cadence/macb_main.c
2420
spin_lock_irqsave(&queue->tx_ptr_lock, flags);
drivers/net/ethernet/cadence/macb_main.c
2423
if (CIRC_SPACE(queue->tx_head, queue->tx_tail,
drivers/net/ethernet/cadence/macb_main.c
2427
queue->tx_head, queue->tx_tail);
drivers/net/ethernet/cadence/macb_main.c
2433
if (macb_tx_map(bp, queue, skb, hdrlen)) {
drivers/net/ethernet/cadence/macb_main.c
2448
if (CIRC_SPACE(queue->tx_head, queue->tx_tail, bp->tx_ring_size) < 1)
drivers/net/ethernet/cadence/macb_main.c
2452
spin_unlock_irqrestore(&queue->tx_ptr_lock, flags);
drivers/net/ethernet/cadence/macb_main.c
2481
struct macb_queue *queue;
drivers/net/ethernet/cadence/macb_main.c
2486
for (q = 0, queue = bp->queues; q < bp->num_queues; ++q, ++queue) {
drivers/net/ethernet/cadence/macb_main.c
2487
if (!queue->rx_skbuff)
drivers/net/ethernet/cadence/macb_main.c
2491
skb = queue->rx_skbuff[i];
drivers/net/ethernet/cadence/macb_main.c
2496
desc = macb_rx_desc(queue, i);
drivers/net/ethernet/cadence/macb_main.c
2505
kfree(queue->rx_skbuff);
drivers/net/ethernet/cadence/macb_main.c
2506
queue->rx_skbuff = NULL;
drivers/net/ethernet/cadence/macb_main.c
2512
struct macb_queue *queue = &bp->queues[0];
drivers/net/ethernet/cadence/macb_main.c
2514
if (queue->rx_buffers) {
drivers/net/ethernet/cadence/macb_main.c
2517
queue->rx_buffers, queue->rx_buffers_dma);
drivers/net/ethernet/cadence/macb_main.c
2518
queue->rx_buffers = NULL;
drivers/net/ethernet/cadence/macb_main.c
2535
struct macb_queue *queue;
drivers/net/ethernet/cadence/macb_main.c
2553
for (q = 0, queue = bp->queues; q < bp->num_queues; ++q, ++queue) {
drivers/net/ethernet/cadence/macb_main.c
2554
kfree(queue->tx_skb);
drivers/net/ethernet/cadence/macb_main.c
2555
queue->tx_skb = NULL;
drivers/net/ethernet/cadence/macb_main.c
2556
queue->tx_ring = NULL;
drivers/net/ethernet/cadence/macb_main.c
2557
queue->rx_ring = NULL;
drivers/net/ethernet/cadence/macb_main.c
2563
struct macb_queue *queue;
drivers/net/ethernet/cadence/macb_main.c
2567
for (q = 0, queue = bp->queues; q < bp->num_queues; ++q, ++queue) {
drivers/net/ethernet/cadence/macb_main.c
2569
queue->rx_skbuff = kzalloc(size, GFP_KERNEL);
drivers/net/ethernet/cadence/macb_main.c
2570
if (!queue->rx_skbuff)
drivers/net/ethernet/cadence/macb_main.c
2575
bp->rx_ring_size, queue->rx_skbuff);
drivers/net/ethernet/cadence/macb_main.c
2582
struct macb_queue *queue = &bp->queues[0];
drivers/net/ethernet/cadence/macb_main.c
2586
queue->rx_buffers = dma_alloc_coherent(&bp->pdev->dev, size,
drivers/net/ethernet/cadence/macb_main.c
2587
&queue->rx_buffers_dma, GFP_KERNEL);
drivers/net/ethernet/cadence/macb_main.c
2588
if (!queue->rx_buffers)
drivers/net/ethernet/cadence/macb_main.c
2593
size, (unsigned long)queue->rx_buffers_dma, queue->rx_buffers);
drivers/net/ethernet/cadence/macb_main.c
2601
struct macb_queue *queue;
drivers/net/ethernet/cadence/macb_main.c
2627
for (q = 0, queue = bp->queues; q < bp->num_queues; ++q, ++queue) {
drivers/net/ethernet/cadence/macb_main.c
2628
queue->tx_ring = tx + macb_tx_ring_size_per_queue(bp) * q;
drivers/net/ethernet/cadence/macb_main.c
2629
queue->tx_ring_dma = tx_dma + macb_tx_ring_size_per_queue(bp) * q;
drivers/net/ethernet/cadence/macb_main.c
2631
queue->rx_ring = rx + macb_rx_ring_size_per_queue(bp) * q;
drivers/net/ethernet/cadence/macb_main.c
2632
queue->rx_ring_dma = rx_dma + macb_rx_ring_size_per_queue(bp) * q;
drivers/net/ethernet/cadence/macb_main.c
2635
queue->tx_skb = kmalloc(size, GFP_KERNEL);
drivers/net/ethernet/cadence/macb_main.c
2636
if (!queue->tx_skb)
drivers/net/ethernet/cadence/macb_main.c
2672
static void gem_init_rx_ring(struct macb_queue *queue)
drivers/net/ethernet/cadence/macb_main.c
2674
queue->rx_tail = 0;
drivers/net/ethernet/cadence/macb_main.c
2675
queue->rx_prepared_head = 0;
drivers/net/ethernet/cadence/macb_main.c
2677
gem_rx_refill(queue);
drivers/net/ethernet/cadence/macb_main.c
2682
struct macb_queue *queue;
drivers/net/ethernet/cadence/macb_main.c
2687
for (q = 0, queue = bp->queues; q < bp->num_queues; ++q, ++queue) {
drivers/net/ethernet/cadence/macb_main.c
2689
desc = macb_tx_desc(queue, i);
drivers/net/ethernet/cadence/macb_main.c
2694
queue->tx_head = 0;
drivers/net/ethernet/cadence/macb_main.c
2695
queue->tx_tail = 0;
drivers/net/ethernet/cadence/macb_main.c
2697
gem_init_rx_ring(queue);
drivers/net/ethernet/cadence/macb_main.c
2724
struct macb_queue *queue;
drivers/net/ethernet/cadence/macb_main.c
2746
for (q = 0, queue = bp->queues; q < bp->num_queues; ++q, ++queue) {
drivers/net/ethernet/cadence/macb_main.c
2747
queue_writel(queue, IDR, -1);
drivers/net/ethernet/cadence/macb_main.c
2748
queue_readl(queue, ISR);
drivers/net/ethernet/cadence/macb_main.c
2750
queue_writel(queue, ISR, -1);
drivers/net/ethernet/cadence/macb_main.c
2829
struct macb_queue *queue;
drivers/net/ethernet/cadence/macb_main.c
2837
for (q = 0, queue = bp->queues; q < bp->num_queues; ++q, ++queue) {
drivers/net/ethernet/cadence/macb_main.c
2839
queue_writel(queue, RBQS, buffer_size);
drivers/net/ethernet/cadence/macb_main.c
3031
struct macb_queue *queue;
drivers/net/ethernet/cadence/macb_main.c
3054
for (q = 0, queue = bp->queues; q < bp->num_queues; ++q, ++queue) {
drivers/net/ethernet/cadence/macb_main.c
3055
napi_enable(&queue->napi_rx);
drivers/net/ethernet/cadence/macb_main.c
3056
napi_enable(&queue->napi_tx);
drivers/net/ethernet/cadence/macb_main.c
3085
for (q = 0, queue = bp->queues; q < bp->num_queues; ++q, ++queue) {
drivers/net/ethernet/cadence/macb_main.c
3086
napi_disable(&queue->napi_rx);
drivers/net/ethernet/cadence/macb_main.c
3087
napi_disable(&queue->napi_tx);
drivers/net/ethernet/cadence/macb_main.c
3098
struct macb_queue *queue;
drivers/net/ethernet/cadence/macb_main.c
3104
for (q = 0, queue = bp->queues; q < bp->num_queues; ++q, ++queue) {
drivers/net/ethernet/cadence/macb_main.c
3105
napi_disable(&queue->napi_rx);
drivers/net/ethernet/cadence/macb_main.c
3106
napi_disable(&queue->napi_tx);
drivers/net/ethernet/cadence/macb_main.c
3154
struct macb_queue *queue;
drivers/net/ethernet/cadence/macb_main.c
3176
for (q = 0, queue = bp->queues; q < bp->num_queues; ++q, ++queue)
drivers/net/ethernet/cadence/macb_main.c
3177
for (i = 0, stat = &queue->stats.first; i < QUEUE_STATS_LEN; ++i, ++stat)
drivers/net/ethernet/cadence/macb_main.c
3247
struct macb_queue *queue;
drivers/net/ethernet/cadence/macb_main.c
3257
for (q = 0, queue = bp->queues; q < bp->num_queues; ++q, ++queue) {
drivers/net/ethernet/cadence/macb_main.c
4182
struct macb_queue *queue;
drivers/net/ethernet/cadence/macb_main.c
4305
queue = &bp->queues[enst_queue[i].queue_id];
drivers/net/ethernet/cadence/macb_main.c
4307
queue_writel(queue, ENST_START_TIME,
drivers/net/ethernet/cadence/macb_main.c
4309
queue_writel(queue, ENST_ON_TIME,
drivers/net/ethernet/cadence/macb_main.c
4311
queue_writel(queue, ENST_OFF_TIME,
drivers/net/ethernet/cadence/macb_main.c
4330
struct macb_queue *queue;
drivers/net/ethernet/cadence/macb_main.c
4343
for (q = 0, queue = bp->queues; q < bp->num_queues; ++q, ++queue) {
drivers/net/ethernet/cadence/macb_main.c
4344
queue_writel(queue, ENST_START_TIME, 0);
drivers/net/ethernet/cadence/macb_main.c
4345
queue_writel(queue, ENST_ON_TIME, 0);
drivers/net/ethernet/cadence/macb_main.c
4346
queue_writel(queue, ENST_OFF_TIME, 0);
drivers/net/ethernet/cadence/macb_main.c
459
struct macb_queue *queue;
drivers/net/ethernet/cadence/macb_main.c
4590
struct macb_queue *queue;
drivers/net/ethernet/cadence/macb_main.c
4602
queue = &bp->queues[q];
drivers/net/ethernet/cadence/macb_main.c
4603
queue->bp = bp;
drivers/net/ethernet/cadence/macb_main.c
4604
spin_lock_init(&queue->tx_ptr_lock);
drivers/net/ethernet/cadence/macb_main.c
4605
netif_napi_add(dev, &queue->napi_rx, macb_rx_poll);
drivers/net/ethernet/cadence/macb_main.c
4606
netif_napi_add(dev, &queue->napi_tx, macb_tx_poll);
drivers/net/ethernet/cadence/macb_main.c
4608
queue->ISR = GEM_ISR(hw_q - 1);
drivers/net/ethernet/cadence/macb_main.c
4609
queue->IER = GEM_IER(hw_q - 1);
drivers/net/ethernet/cadence/macb_main.c
4610
queue->IDR = GEM_IDR(hw_q - 1);
drivers/net/ethernet/cadence/macb_main.c
4611
queue->IMR = GEM_IMR(hw_q - 1);
drivers/net/ethernet/cadence/macb_main.c
4612
queue->TBQP = GEM_TBQP(hw_q - 1);
drivers/net/ethernet/cadence/macb_main.c
4613
queue->RBQP = GEM_RBQP(hw_q - 1);
drivers/net/ethernet/cadence/macb_main.c
4614
queue->RBQS = GEM_RBQS(hw_q - 1);
drivers/net/ethernet/cadence/macb_main.c
4617
queue->ISR = MACB_ISR;
drivers/net/ethernet/cadence/macb_main.c
4618
queue->IER = MACB_IER;
drivers/net/ethernet/cadence/macb_main.c
4619
queue->IDR = MACB_IDR;
drivers/net/ethernet/cadence/macb_main.c
4620
queue->IMR = MACB_IMR;
drivers/net/ethernet/cadence/macb_main.c
4621
queue->TBQP = MACB_TBQP;
drivers/net/ethernet/cadence/macb_main.c
4622
queue->RBQP = MACB_RBQP;
drivers/net/ethernet/cadence/macb_main.c
4625
queue->ENST_START_TIME = GEM_ENST_START_TIME(hw_q);
drivers/net/ethernet/cadence/macb_main.c
4626
queue->ENST_ON_TIME = GEM_ENST_ON_TIME(hw_q);
drivers/net/ethernet/cadence/macb_main.c
4627
queue->ENST_OFF_TIME = GEM_ENST_OFF_TIME(hw_q);
drivers/net/ethernet/cadence/macb_main.c
4634
queue->irq = platform_get_irq(pdev, q);
drivers/net/ethernet/cadence/macb_main.c
4635
err = devm_request_irq(&pdev->dev, queue->irq, macb_interrupt,
drivers/net/ethernet/cadence/macb_main.c
4636
IRQF_SHARED, dev->name, queue);
drivers/net/ethernet/cadence/macb_main.c
4640
queue->irq, err);
drivers/net/ethernet/cadence/macb_main.c
4644
INIT_WORK(&queue->tx_error_task, macb_tx_error_task);
drivers/net/ethernet/cadence/macb_main.c
470
for (q = 0, queue = bp->queues; q < bp->num_queues; ++q, ++queue) {
drivers/net/ethernet/cadence/macb_main.c
471
queue_writel(queue, RBQP, lower_32_bits(queue->rx_ring_dma));
drivers/net/ethernet/cadence/macb_main.c
472
queue_writel(queue, TBQP, lower_32_bits(queue->tx_ring_dma));
drivers/net/ethernet/cadence/macb_main.c
5776
struct macb_queue *queue;
drivers/net/ethernet/cadence/macb_main.c
5817
for (q = 0, queue = bp->queues; q < bp->num_queues;
drivers/net/ethernet/cadence/macb_main.c
5818
++q, ++queue) {
drivers/net/ethernet/cadence/macb_main.c
5821
queue_writel(queue, RBQP, MACB_BIT(QUEUE_DISABLE));
drivers/net/ethernet/cadence/macb_main.c
5824
queue_writel(queue, RBQP,
drivers/net/ethernet/cadence/macb_main.c
5828
queue_writel(queue, IDR, -1);
drivers/net/ethernet/cadence/macb_main.c
5829
queue_readl(queue, ISR);
drivers/net/ethernet/cadence/macb_main.c
5831
queue_writel(queue, ISR, -1);
drivers/net/ethernet/cadence/macb_main.c
5883
for (q = 0, queue = bp->queues; q < bp->num_queues;
drivers/net/ethernet/cadence/macb_main.c
5884
++q, ++queue) {
drivers/net/ethernet/cadence/macb_main.c
5885
napi_disable(&queue->napi_rx);
drivers/net/ethernet/cadence/macb_main.c
5886
napi_disable(&queue->napi_tx);
drivers/net/ethernet/cadence/macb_main.c
5916
struct macb_queue *queue;
drivers/net/ethernet/cadence/macb_main.c
5970
for (q = 0, queue = bp->queues; q < bp->num_queues;
drivers/net/ethernet/cadence/macb_main.c
5971
++q, ++queue) {
drivers/net/ethernet/cadence/macb_main.c
5974
gem_init_rx_ring(queue);
drivers/net/ethernet/cadence/macb_main.c
5976
macb_init_rx_ring(queue);
drivers/net/ethernet/cadence/macb_main.c
5979
napi_enable(&queue->napi_rx);
drivers/net/ethernet/cadence/macb_main.c
5980
napi_enable(&queue->napi_tx);
drivers/net/ethernet/cadence/macb_main.c
656
struct macb_queue *queue;
drivers/net/ethernet/cadence/macb_main.c
661
for (q = 0, queue = bp->queues; q < bp->num_queues; ++q, ++queue)
drivers/net/ethernet/cadence/macb_main.c
662
queue_writel(queue, IDR,
drivers/net/ethernet/cadence/macb_main.c
673
static void gem_shuffle_tx_one_ring(struct macb_queue *queue)
drivers/net/ethernet/cadence/macb_main.c
679
struct macb *bp = queue->bp;
drivers/net/ethernet/cadence/macb_main.c
688
spin_lock_irqsave(&queue->tx_ptr_lock, flags);
drivers/net/ethernet/cadence/macb_main.c
689
head = queue->tx_head;
drivers/net/ethernet/cadence/macb_main.c
690
tail = queue->tx_tail;
drivers/net/ethernet/cadence/macb_main.c
698
queue->tx_head = 0;
drivers/net/ethernet/cadence/macb_main.c
699
queue->tx_tail = 0;
drivers/net/ethernet/cadence/macb_main.c
707
memcpy(&desc, macb_tx_desc(queue, i), desc_size);
drivers/net/ethernet/cadence/macb_main.c
708
memcpy(&tx_skb, macb_tx_skb(queue, i),
drivers/net/ethernet/cadence/macb_main.c
715
desc_curr = macb_tx_desc(queue, curr);
drivers/net/ethernet/cadence/macb_main.c
716
desc_next = macb_tx_desc(queue, next);
drivers/net/ethernet/cadence/macb_main.c
725
skb_curr = macb_tx_skb(queue, curr);
drivers/net/ethernet/cadence/macb_main.c
726
skb_next = macb_tx_skb(queue, next);
drivers/net/ethernet/cadence/macb_main.c
733
desc_curr = macb_tx_desc(queue, curr);
drivers/net/ethernet/cadence/macb_main.c
739
memcpy(macb_tx_skb(queue, curr), &tx_skb,
drivers/net/ethernet/cadence/macb_main.c
743
queue->tx_head = count;
drivers/net/ethernet/cadence/macb_main.c
744
queue->tx_tail = 0;
drivers/net/ethernet/cadence/macb_main.c
750
spin_unlock_irqrestore(&queue->tx_ptr_lock, flags);
drivers/net/ethernet/cadence/macb_main.c
756
struct macb_queue *queue;
drivers/net/ethernet/cadence/macb_main.c
759
for (q = 0, queue = bp->queues; q < bp->num_queues; q++, queue++)
drivers/net/ethernet/cadence/macb_main.c
760
gem_shuffle_tx_one_ring(queue);
drivers/net/ethernet/cadence/macb_main.c
771
struct macb_queue *queue;
drivers/net/ethernet/cadence/macb_main.c
800
for (q = 0, queue = bp->queues; q < bp->num_queues; ++q, ++queue) {
drivers/net/ethernet/cadence/macb_main.c
801
queue_writel(queue, IER,
drivers/net/ethernet/chelsio/cxgb3/sge.c
1905
struct sk_buff_head queue;
drivers/net/ethernet/chelsio/cxgb3/sge.c
1909
__skb_queue_head_init(&queue);
drivers/net/ethernet/chelsio/cxgb3/sge.c
1910
skb_queue_splice_init(&q->rx_queue, &queue);
drivers/net/ethernet/chelsio/cxgb3/sge.c
1911
if (skb_queue_empty(&queue)) {
drivers/net/ethernet/chelsio/cxgb3/sge.c
1919
skb_queue_walk_safe(&queue, skb, tmp) {
drivers/net/ethernet/chelsio/cxgb3/sge.c
1924
__skb_unlink(skb, &queue);
drivers/net/ethernet/chelsio/cxgb3/sge.c
1934
if (!skb_queue_empty(&queue)) {
drivers/net/ethernet/chelsio/cxgb3/sge.c
1937
skb_queue_splice(&queue, &q->rx_queue);
drivers/net/ethernet/chelsio/cxgb4/cxgb4.h
1287
s8 queue; /* queue index */
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1927
unsigned int queue)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1947
chan = rxq_to_chan(&adap->sge, queue);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1950
SYN_RSS_ENABLE_F | SYN_RSS_QUEUE_V(queue));
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1968
unsigned int queue)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1990
chan = rxq_to_chan(&adap->sge, queue);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
1993
SYN_RSS_ENABLE_F | SYN_RSS_QUEUE_V(queue));
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
2000
unsigned int queue, bool ipv6)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
2017
LISTSVR_IPV6_V(0)) | QUEUENO_V(queue));
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
2907
unsigned int queue, unsigned char port, unsigned char mask)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
2956
f->fs.iq = queue;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
2976
unsigned int queue, bool ipv6)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
3505
qe.queue = index;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
3527
qe.queue = index;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
3561
qe.queue = index;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_flower.c
547
fs->iq = act->queue.index;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_matchall.c
111
qe.queue = i;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_matchall.c
132
qe.queue = i;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_matchall.c
143
qe.queue = i;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_tc_matchall.c
158
qe.queue = i;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h
255
unsigned int queue);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h
258
unsigned int queue);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h
260
unsigned int queue, bool ipv6);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h
263
unsigned int queue,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h
266
unsigned int queue, bool ipv6);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h
299
static inline void set_wr_txq(struct sk_buff *skb, int prio, int queue)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h
301
skb_set_queue_mapping(skb, (queue << 1) | prio);
drivers/net/ethernet/chelsio/cxgb4/sched.c
177
if (p->queue < 0 || p->queue >= pi->nqsets)
drivers/net/ethernet/chelsio/cxgb4/sched.c
180
txq = &adap->sge.ethtxq[pi->first_qset + p->queue];
drivers/net/ethernet/chelsio/cxgb4/sched.c
193
if (p->queue < 0 || p->queue >= pi->nqsets)
drivers/net/ethernet/chelsio/cxgb4/sched.c
196
txq = &adap->sge.ethtxq[pi->first_qset + p->queue];
drivers/net/ethernet/chelsio/cxgb4/sched.c
225
if (p->queue < 0 || p->queue >= pi->nqsets)
drivers/net/ethernet/chelsio/cxgb4/sched.c
232
txq = &adap->sge.ethtxq[pi->first_qset + p->queue];
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_cm.c
2020
int status, int queue)
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_cm.c
2032
req->status = (queue << 1) | status;
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_cm.c
2039
set_wr_txq(reply_skb, CPL_PRIORITY_DATA, queue);
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_cm.c
2059
int queue;
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_cm.c
2063
queue = csk->txq_idx;
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_cm.c
2067
CPL_ABORT_NO_RST, queue);
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_cm.c
2096
int queue = csk->txq_idx;
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_cm.c
2098
chtls_send_abort_rpl(sk, skb, cdev, CPL_ABORT_NO_RST, queue);
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_cm.c
2114
int queue = csk->txq_idx;
drivers/net/ethernet/chelsio/inline_crypto/chtls/chtls_cm.c
2146
rst_status, queue);
drivers/net/ethernet/engleder/tsnep.h
199
struct tsnep_queue queue[TSNEP_MAX_QUEUES];
drivers/net/ethernet/engleder/tsnep.h
259
int tsnep_set_irq_coalesce(struct tsnep_queue *queue, u32 usecs);
drivers/net/ethernet/engleder/tsnep.h
260
u32 tsnep_get_irq_coalesce(struct tsnep_queue *queue);
drivers/net/ethernet/engleder/tsnep.h
261
int tsnep_enable_xsk(struct tsnep_queue *queue, struct xsk_buff_pool *pool);
drivers/net/ethernet/engleder/tsnep.h
262
void tsnep_disable_xsk(struct tsnep_queue *queue);
drivers/net/ethernet/engleder/tsnep_ethtool.c
338
if (adapter->queue[i].tx) {
drivers/net/ethernet/engleder/tsnep_ethtool.c
340
return &adapter->queue[i];
drivers/net/ethernet/engleder/tsnep_ethtool.c
355
if (adapter->queue[i].rx) {
drivers/net/ethernet/engleder/tsnep_ethtool.c
357
return &adapter->queue[i];
drivers/net/ethernet/engleder/tsnep_ethtool.c
372
struct tsnep_queue *queue;
drivers/net/ethernet/engleder/tsnep_ethtool.c
374
queue = tsnep_get_queue_with_rx(adapter, 0);
drivers/net/ethernet/engleder/tsnep_ethtool.c
375
if (queue)
drivers/net/ethernet/engleder/tsnep_ethtool.c
376
ec->rx_coalesce_usecs = tsnep_get_irq_coalesce(queue);
drivers/net/ethernet/engleder/tsnep_ethtool.c
378
queue = tsnep_get_queue_with_tx(adapter, 0);
drivers/net/ethernet/engleder/tsnep_ethtool.c
379
if (queue)
drivers/net/ethernet/engleder/tsnep_ethtool.c
380
ec->tx_coalesce_usecs = tsnep_get_irq_coalesce(queue);
drivers/net/ethernet/engleder/tsnep_ethtool.c
396
if (adapter->queue[i].rx)
drivers/net/ethernet/engleder/tsnep_ethtool.c
397
retval = tsnep_set_irq_coalesce(&adapter->queue[i],
drivers/net/ethernet/engleder/tsnep_ethtool.c
400
retval = tsnep_set_irq_coalesce(&adapter->queue[i],
drivers/net/ethernet/engleder/tsnep_ethtool.c
410
u32 queue,
drivers/net/ethernet/engleder/tsnep_ethtool.c
417
if (queue >= max(adapter->num_tx_queues, adapter->num_rx_queues))
drivers/net/ethernet/engleder/tsnep_ethtool.c
420
queue_with_rx = tsnep_get_queue_with_rx(adapter, queue);
drivers/net/ethernet/engleder/tsnep_ethtool.c
424
queue_with_tx = tsnep_get_queue_with_tx(adapter, queue);
drivers/net/ethernet/engleder/tsnep_ethtool.c
432
u32 queue,
drivers/net/ethernet/engleder/tsnep_ethtool.c
440
if (queue >= max(adapter->num_tx_queues, adapter->num_rx_queues))
drivers/net/ethernet/engleder/tsnep_ethtool.c
443
queue_with_rx = tsnep_get_queue_with_rx(adapter, queue);
drivers/net/ethernet/engleder/tsnep_ethtool.c
451
queue_with_tx = tsnep_get_queue_with_tx(adapter, queue);
drivers/net/ethernet/engleder/tsnep_main.c
101
if ((active & adapter->queue[0].irq_mask) != 0) {
drivers/net/ethernet/engleder/tsnep_main.c
102
if (napi_schedule_prep(&adapter->queue[0].napi)) {
drivers/net/ethernet/engleder/tsnep_main.c
103
tsnep_disable_irq(adapter, adapter->queue[0].irq_mask);
drivers/net/ethernet/engleder/tsnep_main.c
105
__napi_schedule(&adapter->queue[0].napi);
drivers/net/ethernet/engleder/tsnep_main.c
114
struct tsnep_queue *queue = arg;
drivers/net/ethernet/engleder/tsnep_main.c
117
if (napi_schedule_prep(&queue->napi)) {
drivers/net/ethernet/engleder/tsnep_main.c
118
tsnep_disable_irq(queue->adapter, queue->irq_mask);
drivers/net/ethernet/engleder/tsnep_main.c
120
__napi_schedule(&queue->napi);
drivers/net/ethernet/engleder/tsnep_main.c
126
int tsnep_set_irq_coalesce(struct tsnep_queue *queue, u32 usecs)
drivers/net/ethernet/engleder/tsnep_main.c
135
queue->irq_delay &= ~ECM_INT_DELAY_MASK;
drivers/net/ethernet/engleder/tsnep_main.c
136
queue->irq_delay |= usecs;
drivers/net/ethernet/engleder/tsnep_main.c
137
iowrite8(queue->irq_delay, queue->irq_delay_addr);
drivers/net/ethernet/engleder/tsnep_main.c
142
u32 tsnep_get_irq_coalesce(struct tsnep_queue *queue)
drivers/net/ethernet/engleder/tsnep_main.c
146
usecs = (queue->irq_delay & ECM_INT_DELAY_MASK);
drivers/net/ethernet/engleder/tsnep_main.c
1796
static bool tsnep_pending(struct tsnep_queue *queue)
drivers/net/ethernet/engleder/tsnep_main.c
1798
if (queue->tx && tsnep_tx_pending(queue->tx))
drivers/net/ethernet/engleder/tsnep_main.c
1801
if (queue->rx && tsnep_rx_pending(queue->rx))
drivers/net/ethernet/engleder/tsnep_main.c
1809
struct tsnep_queue *queue = container_of(napi, struct tsnep_queue,
drivers/net/ethernet/engleder/tsnep_main.c
1814
if (queue->tx)
drivers/net/ethernet/engleder/tsnep_main.c
1815
complete = tsnep_tx_poll(queue->tx, budget);
drivers/net/ethernet/engleder/tsnep_main.c
1821
if (queue->rx) {
drivers/net/ethernet/engleder/tsnep_main.c
1822
done = queue->rx->xsk_pool ?
drivers/net/ethernet/engleder/tsnep_main.c
1823
tsnep_rx_poll_zc(queue->rx, napi, budget) :
drivers/net/ethernet/engleder/tsnep_main.c
1824
tsnep_rx_poll(queue->rx, napi, budget);
drivers/net/ethernet/engleder/tsnep_main.c
1834
tsnep_enable_irq(queue->adapter, queue->irq_mask);
drivers/net/ethernet/engleder/tsnep_main.c
1840
if (tsnep_pending(queue)) {
drivers/net/ethernet/engleder/tsnep_main.c
1841
tsnep_disable_irq(queue->adapter, queue->irq_mask);
drivers/net/ethernet/engleder/tsnep_main.c
1849
static int tsnep_request_irq(struct tsnep_queue *queue, bool first)
drivers/net/ethernet/engleder/tsnep_main.c
1851
const char *name = netdev_name(queue->adapter->netdev);
drivers/net/ethernet/engleder/tsnep_main.c
1857
sprintf(queue->name, "%s-mac", name);
drivers/net/ethernet/engleder/tsnep_main.c
1859
dev = queue->adapter;
drivers/net/ethernet/engleder/tsnep_main.c
1861
if (queue->tx && queue->rx)
drivers/net/ethernet/engleder/tsnep_main.c
1862
snprintf(queue->name, sizeof(queue->name), "%s-txrx-%d",
drivers/net/ethernet/engleder/tsnep_main.c
1863
name, queue->rx->queue_index);
drivers/net/ethernet/engleder/tsnep_main.c
1864
else if (queue->tx)
drivers/net/ethernet/engleder/tsnep_main.c
1865
snprintf(queue->name, sizeof(queue->name), "%s-tx-%d",
drivers/net/ethernet/engleder/tsnep_main.c
1866
name, queue->tx->queue_index);
drivers/net/ethernet/engleder/tsnep_main.c
1868
snprintf(queue->name, sizeof(queue->name), "%s-rx-%d",
drivers/net/ethernet/engleder/tsnep_main.c
1869
name, queue->rx->queue_index);
drivers/net/ethernet/engleder/tsnep_main.c
1871
dev = queue;
drivers/net/ethernet/engleder/tsnep_main.c
1874
retval = request_irq(queue->irq, handler, 0, queue->name, dev);
drivers/net/ethernet/engleder/tsnep_main.c
1877
memset(queue->name, 0, sizeof(queue->name));
drivers/net/ethernet/engleder/tsnep_main.c
1883
static void tsnep_free_irq(struct tsnep_queue *queue, bool first)
drivers/net/ethernet/engleder/tsnep_main.c
1887
if (!strlen(queue->name))
drivers/net/ethernet/engleder/tsnep_main.c
1891
dev = queue->adapter;
drivers/net/ethernet/engleder/tsnep_main.c
1893
dev = queue;
drivers/net/ethernet/engleder/tsnep_main.c
1895
free_irq(queue->irq, dev);
drivers/net/ethernet/engleder/tsnep_main.c
1896
memset(queue->name, 0, sizeof(queue->name));
drivers/net/ethernet/engleder/tsnep_main.c
1899
static void tsnep_queue_close(struct tsnep_queue *queue, bool first)
drivers/net/ethernet/engleder/tsnep_main.c
1901
struct tsnep_rx *rx = queue->rx;
drivers/net/ethernet/engleder/tsnep_main.c
1903
tsnep_free_irq(queue, first);
drivers/net/ethernet/engleder/tsnep_main.c
1912
netif_napi_del(&queue->napi);
drivers/net/ethernet/engleder/tsnep_main.c
1916
struct tsnep_queue *queue, bool first)
drivers/net/ethernet/engleder/tsnep_main.c
1918
struct tsnep_rx *rx = queue->rx;
drivers/net/ethernet/engleder/tsnep_main.c
1919
struct tsnep_tx *tx = queue->tx;
drivers/net/ethernet/engleder/tsnep_main.c
1922
netif_napi_add(adapter->netdev, &queue->napi, tsnep_poll);
drivers/net/ethernet/engleder/tsnep_main.c
1938
rx->queue_index, queue->napi.napi_id);
drivers/net/ethernet/engleder/tsnep_main.c
1947
rx->queue_index, queue->napi.napi_id);
drivers/net/ethernet/engleder/tsnep_main.c
1959
retval = tsnep_request_irq(queue, first);
drivers/net/ethernet/engleder/tsnep_main.c
1962
"can't get assigned irq %d.\n", queue->irq);
drivers/net/ethernet/engleder/tsnep_main.c
1969
tsnep_queue_close(queue, first);
drivers/net/ethernet/engleder/tsnep_main.c
1974
static void tsnep_queue_enable(struct tsnep_queue *queue)
drivers/net/ethernet/engleder/tsnep_main.c
1976
struct tsnep_adapter *adapter = queue->adapter;
drivers/net/ethernet/engleder/tsnep_main.c
1978
netif_napi_set_irq(&queue->napi, queue->irq);
drivers/net/ethernet/engleder/tsnep_main.c
1979
napi_enable(&queue->napi);
drivers/net/ethernet/engleder/tsnep_main.c
1980
tsnep_enable_irq(adapter, queue->irq_mask);
drivers/net/ethernet/engleder/tsnep_main.c
1982
if (queue->tx) {
drivers/net/ethernet/engleder/tsnep_main.c
1983
netif_queue_set_napi(adapter->netdev, queue->tx->queue_index,
drivers/net/ethernet/engleder/tsnep_main.c
1984
NETDEV_QUEUE_TYPE_TX, &queue->napi);
drivers/net/ethernet/engleder/tsnep_main.c
1985
tsnep_tx_enable(queue->tx);
drivers/net/ethernet/engleder/tsnep_main.c
1988
if (queue->rx) {
drivers/net/ethernet/engleder/tsnep_main.c
1989
netif_queue_set_napi(adapter->netdev, queue->rx->queue_index,
drivers/net/ethernet/engleder/tsnep_main.c
1990
NETDEV_QUEUE_TYPE_RX, &queue->napi);
drivers/net/ethernet/engleder/tsnep_main.c
1991
tsnep_rx_enable(queue->rx);
drivers/net/ethernet/engleder/tsnep_main.c
1995
static void tsnep_queue_disable(struct tsnep_queue *queue)
drivers/net/ethernet/engleder/tsnep_main.c
1997
struct tsnep_adapter *adapter = queue->adapter;
drivers/net/ethernet/engleder/tsnep_main.c
1999
if (queue->rx)
drivers/net/ethernet/engleder/tsnep_main.c
2000
netif_queue_set_napi(adapter->netdev, queue->rx->queue_index,
drivers/net/ethernet/engleder/tsnep_main.c
2003
if (queue->tx) {
drivers/net/ethernet/engleder/tsnep_main.c
2004
tsnep_tx_disable(queue->tx, &queue->napi);
drivers/net/ethernet/engleder/tsnep_main.c
2005
netif_queue_set_napi(adapter->netdev, queue->tx->queue_index,
drivers/net/ethernet/engleder/tsnep_main.c
2009
napi_disable(&queue->napi);
drivers/net/ethernet/engleder/tsnep_main.c
2010
tsnep_disable_irq(adapter, queue->irq_mask);
drivers/net/ethernet/engleder/tsnep_main.c
2015
if (queue->rx)
drivers/net/ethernet/engleder/tsnep_main.c
2016
tsnep_rx_disable(queue->rx);
drivers/net/ethernet/engleder/tsnep_main.c
2025
if (adapter->queue[i].tx) {
drivers/net/ethernet/engleder/tsnep_main.c
2026
retval = tsnep_tx_open(adapter->queue[i].tx);
drivers/net/ethernet/engleder/tsnep_main.c
2030
if (adapter->queue[i].rx) {
drivers/net/ethernet/engleder/tsnep_main.c
2031
retval = tsnep_rx_open(adapter->queue[i].rx);
drivers/net/ethernet/engleder/tsnep_main.c
2036
retval = tsnep_queue_open(adapter, &adapter->queue[i], i == 0);
drivers/net/ethernet/engleder/tsnep_main.c
2056
tsnep_queue_enable(&adapter->queue[i]);
drivers/net/ethernet/engleder/tsnep_main.c
2064
tsnep_queue_close(&adapter->queue[i], i == 0);
drivers/net/ethernet/engleder/tsnep_main.c
2066
if (adapter->queue[i].rx)
drivers/net/ethernet/engleder/tsnep_main.c
2067
tsnep_rx_close(adapter->queue[i].rx);
drivers/net/ethernet/engleder/tsnep_main.c
2068
if (adapter->queue[i].tx)
drivers/net/ethernet/engleder/tsnep_main.c
2069
tsnep_tx_close(adapter->queue[i].tx);
drivers/net/ethernet/engleder/tsnep_main.c
2083
tsnep_queue_disable(&adapter->queue[i]);
drivers/net/ethernet/engleder/tsnep_main.c
2085
tsnep_queue_close(&adapter->queue[i], i == 0);
drivers/net/ethernet/engleder/tsnep_main.c
2087
if (adapter->queue[i].rx)
drivers/net/ethernet/engleder/tsnep_main.c
2088
tsnep_rx_close(adapter->queue[i].rx);
drivers/net/ethernet/engleder/tsnep_main.c
2089
if (adapter->queue[i].tx)
drivers/net/ethernet/engleder/tsnep_main.c
2090
tsnep_tx_close(adapter->queue[i].tx);
drivers/net/ethernet/engleder/tsnep_main.c
2096
int tsnep_enable_xsk(struct tsnep_queue *queue, struct xsk_buff_pool *pool)
drivers/net/ethernet/engleder/tsnep_main.c
2098
bool running = netif_running(queue->adapter->netdev);
drivers/net/ethernet/engleder/tsnep_main.c
2105
queue->rx->page_buffer = kzalloc_objs(*queue->rx->page_buffer,
drivers/net/ethernet/engleder/tsnep_main.c
2107
if (!queue->rx->page_buffer)
drivers/net/ethernet/engleder/tsnep_main.c
2109
queue->rx->xdp_batch = kzalloc_objs(*queue->rx->xdp_batch,
drivers/net/ethernet/engleder/tsnep_main.c
2111
if (!queue->rx->xdp_batch) {
drivers/net/ethernet/engleder/tsnep_main.c
2112
kfree(queue->rx->page_buffer);
drivers/net/ethernet/engleder/tsnep_main.c
2113
queue->rx->page_buffer = NULL;
drivers/net/ethernet/engleder/tsnep_main.c
2118
xsk_pool_set_rxq_info(pool, &queue->rx->xdp_rxq_zc);
drivers/net/ethernet/engleder/tsnep_main.c
2121
tsnep_queue_disable(queue);
drivers/net/ethernet/engleder/tsnep_main.c
2123
queue->tx->xsk_pool = pool;
drivers/net/ethernet/engleder/tsnep_main.c
2124
queue->rx->xsk_pool = pool;
drivers/net/ethernet/engleder/tsnep_main.c
2127
tsnep_rx_reopen_xsk(queue->rx);
drivers/net/ethernet/engleder/tsnep_main.c
2128
tsnep_queue_enable(queue);
drivers/net/ethernet/engleder/tsnep_main.c
2134
void tsnep_disable_xsk(struct tsnep_queue *queue)
drivers/net/ethernet/engleder/tsnep_main.c
2136
bool running = netif_running(queue->adapter->netdev);
drivers/net/ethernet/engleder/tsnep_main.c
2139
tsnep_queue_disable(queue);
drivers/net/ethernet/engleder/tsnep_main.c
2141
tsnep_rx_free_zc(queue->rx);
drivers/net/ethernet/engleder/tsnep_main.c
2143
queue->rx->xsk_pool = NULL;
drivers/net/ethernet/engleder/tsnep_main.c
2144
queue->tx->xsk_pool = NULL;
drivers/net/ethernet/engleder/tsnep_main.c
2147
tsnep_rx_reopen(queue->rx);
drivers/net/ethernet/engleder/tsnep_main.c
2148
tsnep_queue_enable(queue);
drivers/net/ethernet/engleder/tsnep_main.c
2151
kfree(queue->rx->xdp_batch);
drivers/net/ethernet/engleder/tsnep_main.c
2152
queue->rx->xdp_batch = NULL;
drivers/net/ethernet/engleder/tsnep_main.c
2153
kfree(queue->rx->page_buffer);
drivers/net/ethernet/engleder/tsnep_main.c
2154
queue->rx->page_buffer = NULL;
drivers/net/ethernet/engleder/tsnep_main.c
2357
struct tsnep_queue *queue;
drivers/net/ethernet/engleder/tsnep_main.c
2363
queue = &adapter->queue[queue_id];
drivers/net/ethernet/engleder/tsnep_main.c
2365
if (!napi_if_scheduled_mark_missed(&queue->napi))
drivers/net/ethernet/engleder/tsnep_main.c
2366
napi_schedule(&queue->napi);
drivers/net/ethernet/engleder/tsnep_main.c
2502
adapter->queue[0].adapter = adapter;
drivers/net/ethernet/engleder/tsnep_main.c
2503
adapter->queue[0].irq = retval;
drivers/net/ethernet/engleder/tsnep_main.c
2504
adapter->queue[0].tx = &adapter->tx[0];
drivers/net/ethernet/engleder/tsnep_main.c
2505
adapter->queue[0].tx->adapter = adapter;
drivers/net/ethernet/engleder/tsnep_main.c
2506
adapter->queue[0].tx->addr = adapter->addr + TSNEP_QUEUE(0);
drivers/net/ethernet/engleder/tsnep_main.c
2507
adapter->queue[0].tx->queue_index = 0;
drivers/net/ethernet/engleder/tsnep_main.c
2508
adapter->queue[0].rx = &adapter->rx[0];
drivers/net/ethernet/engleder/tsnep_main.c
2509
adapter->queue[0].rx->adapter = adapter;
drivers/net/ethernet/engleder/tsnep_main.c
2510
adapter->queue[0].rx->addr = adapter->addr + TSNEP_QUEUE(0);
drivers/net/ethernet/engleder/tsnep_main.c
2511
adapter->queue[0].rx->queue_index = 0;
drivers/net/ethernet/engleder/tsnep_main.c
2512
adapter->queue[0].irq_mask = irq_mask;
drivers/net/ethernet/engleder/tsnep_main.c
2513
adapter->queue[0].irq_delay_addr = adapter->addr + ECM_INT_DELAY;
drivers/net/ethernet/engleder/tsnep_main.c
2514
retval = tsnep_set_irq_coalesce(&adapter->queue[0],
drivers/net/ethernet/engleder/tsnep_main.c
2519
adapter->netdev->irq = adapter->queue[0].irq;
drivers/net/ethernet/engleder/tsnep_main.c
2533
adapter->queue[i].adapter = adapter;
drivers/net/ethernet/engleder/tsnep_main.c
2534
adapter->queue[i].irq = retval;
drivers/net/ethernet/engleder/tsnep_main.c
2535
adapter->queue[i].tx = &adapter->tx[i];
drivers/net/ethernet/engleder/tsnep_main.c
2536
adapter->queue[i].tx->adapter = adapter;
drivers/net/ethernet/engleder/tsnep_main.c
2537
adapter->queue[i].tx->addr = adapter->addr + TSNEP_QUEUE(i);
drivers/net/ethernet/engleder/tsnep_main.c
2538
adapter->queue[i].tx->queue_index = i;
drivers/net/ethernet/engleder/tsnep_main.c
2539
adapter->queue[i].rx = &adapter->rx[i];
drivers/net/ethernet/engleder/tsnep_main.c
2540
adapter->queue[i].rx->adapter = adapter;
drivers/net/ethernet/engleder/tsnep_main.c
2541
adapter->queue[i].rx->addr = adapter->addr + TSNEP_QUEUE(i);
drivers/net/ethernet/engleder/tsnep_main.c
2542
adapter->queue[i].rx->queue_index = i;
drivers/net/ethernet/engleder/tsnep_main.c
2543
adapter->queue[i].irq_mask =
drivers/net/ethernet/engleder/tsnep_main.c
2545
adapter->queue[i].irq_delay_addr =
drivers/net/ethernet/engleder/tsnep_main.c
2547
retval = tsnep_set_irq_coalesce(&adapter->queue[i],
drivers/net/ethernet/engleder/tsnep_xdp.c
24
struct tsnep_queue *queue;
drivers/net/ethernet/engleder/tsnep_xdp.c
31
queue = &adapter->queue[queue_id];
drivers/net/ethernet/engleder/tsnep_xdp.c
32
if (queue->rx->queue_index != queue_id ||
drivers/net/ethernet/engleder/tsnep_xdp.c
33
queue->tx->queue_index != queue_id) {
drivers/net/ethernet/engleder/tsnep_xdp.c
48
retval = tsnep_enable_xsk(queue, pool);
drivers/net/ethernet/engleder/tsnep_xdp.c
61
struct tsnep_queue *queue;
drivers/net/ethernet/engleder/tsnep_xdp.c
71
queue = &adapter->queue[queue_id];
drivers/net/ethernet/engleder/tsnep_xdp.c
73
tsnep_disable_xsk(queue);
drivers/net/ethernet/freescale/dpaa/dpaa_eth.c
2106
int queue,
drivers/net/ethernet/freescale/dpaa/dpaa_eth.c
2112
egress_fq = priv->egress_fqs[queue];
drivers/net/ethernet/freescale/dpaa/dpaa_eth.c
2114
fd->cmd |= cpu_to_be32(qman_fq_fqid(priv->conf_fqs[queue]));
drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c
3652
struct dpni_queue queue;
drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c
3669
&queue, &qid);
drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c
3899
struct dpni_queue queue;
drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c
3904
DPNI_QUEUE_RX, fq->tc, fq->flowid, &queue, &qid);
drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c
3912
queue.destination.id = fq->channel->dpcon_id;
drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c
3913
queue.destination.type = DPNI_DEST_DPCON;
drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c
3914
queue.destination.priority = 1;
drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c
3915
queue.user_context = (u64)(uintptr_t)fq;
drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c
3919
&queue);
drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c
3952
struct dpni_queue queue;
drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c
3959
&queue, &qid);
drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c
3972
&queue, &qid);
drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c
3980
queue.destination.id = fq->channel->dpcon_id;
drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c
3981
queue.destination.type = DPNI_DEST_DPCON;
drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c
3982
queue.destination.priority = 0;
drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c
3983
queue.user_context = (u64)(uintptr_t)fq;
drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c
3987
&queue);
drivers/net/ethernet/freescale/dpaa2/dpni.c
1526
const struct dpni_queue *queue)
drivers/net/ethernet/freescale/dpaa2/dpni.c
1540
cmd_params->dest_id = cpu_to_le32(queue->destination.id);
drivers/net/ethernet/freescale/dpaa2/dpni.c
1541
cmd_params->dest_prio = queue->destination.priority;
drivers/net/ethernet/freescale/dpaa2/dpni.c
1542
dpni_set_field(cmd_params->flags, DEST_TYPE, queue->destination.type);
drivers/net/ethernet/freescale/dpaa2/dpni.c
1543
dpni_set_field(cmd_params->flags, STASH_CTRL, queue->flc.stash_control);
drivers/net/ethernet/freescale/dpaa2/dpni.c
1545
queue->destination.hold_active);
drivers/net/ethernet/freescale/dpaa2/dpni.c
1546
cmd_params->flc = cpu_to_le64(queue->flc.value);
drivers/net/ethernet/freescale/dpaa2/dpni.c
1547
cmd_params->user_context = cpu_to_le64(queue->user_context);
drivers/net/ethernet/freescale/dpaa2/dpni.c
1573
struct dpni_queue *queue,
drivers/net/ethernet/freescale/dpaa2/dpni.c
1597
queue->destination.id = le32_to_cpu(rsp_params->dest_id);
drivers/net/ethernet/freescale/dpaa2/dpni.c
1598
queue->destination.priority = rsp_params->dest_prio;
drivers/net/ethernet/freescale/dpaa2/dpni.c
1599
queue->destination.type = dpni_get_field(rsp_params->flags,
drivers/net/ethernet/freescale/dpaa2/dpni.c
1601
queue->flc.stash_control = dpni_get_field(rsp_params->flags,
drivers/net/ethernet/freescale/dpaa2/dpni.c
1603
queue->destination.hold_active = dpni_get_field(rsp_params->flags,
drivers/net/ethernet/freescale/dpaa2/dpni.c
1605
queue->flc.value = le64_to_cpu(rsp_params->flc);
drivers/net/ethernet/freescale/dpaa2/dpni.c
1606
queue->user_context = le64_to_cpu(rsp_params->user_context);
drivers/net/ethernet/freescale/dpaa2/dpni.h
841
const struct dpni_queue *queue);
drivers/net/ethernet/freescale/dpaa2/dpni.h
849
struct dpni_queue *queue,
drivers/net/ethernet/freescale/enetc/enetc_qos.c
177
int queue = queue_stats->queue;
drivers/net/ethernet/freescale/enetc/enetc_qos.c
179
stats->window_drops = priv->tx_ring[queue]->stats.win_drop;
drivers/net/ethernet/freescale/enetc/enetc_qos.c
244
u8 tc = cbs->queue;
drivers/net/ethernet/freescale/enetc/enetc_qos.c
361
tc = qopt->queue;
drivers/net/ethernet/freescale/fec_main.c
1484
u32 queue)
drivers/net/ethernet/freescale/fec_main.c
1486
struct fec_enet_priv_tx_q *txq = fep->tx_queue[queue];
drivers/net/ethernet/freescale/fec_main.c
1497
nq = netdev_get_tx_queue(fep->netdev, queue);
drivers/net/ethernet/freescale/fec_main.c
1566
u16 queue, int budget)
drivers/net/ethernet/freescale/fec_main.c
1568
struct netdev_queue *nq = netdev_get_tx_queue(fep->netdev, queue);
drivers/net/ethernet/freescale/fec_main.c
1569
struct fec_enet_priv_tx_q *txq = fep->tx_queue[queue];
drivers/net/ethernet/freescale/fec_main.c
1732
if (!fec_enet_xsk_xmit(fep, pool, queue))
drivers/net/ethernet/freescale/fec_main.c
1895
u16 queue, int budget)
drivers/net/ethernet/freescale/fec_main.c
1897
struct fec_enet_priv_rx_q *rxq = fep->rx_queue[queue];
drivers/net/ethernet/freescale/fec_main.c
1926
writel(FEC_ENET_RXF_GET(queue), fep->hwp + FEC_IEVENT);
drivers/net/ethernet/freescale/fec_main.c
2018
static int fec_enet_rx_queue_xdp(struct fec_enet_private *fep, int queue,
drivers/net/ethernet/freescale/fec_main.c
2022
struct fec_enet_priv_rx_q *rxq = fep->rx_queue[queue];
drivers/net/ethernet/freescale/fec_main.c
2031
int tx_qid = queue;
drivers/net/ethernet/freescale/fec_main.c
2056
writel(FEC_ENET_RXF_GET(queue), fep->hwp + FEC_IEVENT);
drivers/net/ethernet/freescale/fec_main.c
2204
int queue)
drivers/net/ethernet/freescale/fec_main.c
2206
struct netdev_queue *nq = netdev_get_tx_queue(fep->netdev, queue);
drivers/net/ethernet/freescale/fec_main.c
2207
struct fec_enet_priv_tx_q *txq = fep->tx_queue[queue];
drivers/net/ethernet/freescale/fec_main.c
2269
static int fec_enet_rx_queue_xsk(struct fec_enet_private *fep, int queue,
drivers/net/ethernet/freescale/fec_main.c
2273
struct fec_enet_priv_rx_q *rxq = fep->rx_queue[queue];
drivers/net/ethernet/freescale/fec_main.c
2299
writel(FEC_ENET_RXF_GET(queue), fep->hwp + FEC_IEVENT);
drivers/net/ethernet/freescale/fec_main.c
2369
err = fec_enet_xsk_tx_xmit(fep, xsk, cpu, queue);
drivers/net/ethernet/freescale/fec_main.c
2440
fec_txq_trigger_xmit(fep, fep->tx_queue[queue]);
drivers/net/ethernet/freescale/fec_main.c
4071
fec_enet_alloc_rxq_buffers(struct net_device *ndev, unsigned int queue)
drivers/net/ethernet/freescale/fec_main.c
4077
rxq = fep->rx_queue[queue];
drivers/net/ethernet/freescale/fec_main.c
4102
fec_enet_alloc_txq_buffers(struct net_device *ndev, unsigned int queue)
drivers/net/ethernet/freescale/fec_main.c
4109
txq = fep->tx_queue[queue];
drivers/net/ethernet/freescale/fec_main.c
4420
fec_alloc_new_rxq_xsk(struct fec_enet_private *fep, int queue,
drivers/net/ethernet/freescale/fec_main.c
4423
struct fec_enet_priv_rx_q *old_rxq = fep->rx_queue[queue];
drivers/net/ethernet/freescale/fec_main.c
4434
rxq->id = queue;
drivers/net/ethernet/freescale/fec_main.c
4462
fec_alloc_new_rxq_pp(struct fec_enet_private *fep, int queue)
drivers/net/ethernet/freescale/fec_main.c
4464
struct fec_enet_priv_rx_q *old_rxq = fep->rx_queue[queue];
drivers/net/ethernet/freescale/fec_main.c
4474
rxq->id = queue;
drivers/net/ethernet/freescale/fec_main.c
4524
u16 queue)
drivers/net/ethernet/freescale/fec_main.c
4526
struct fec_enet_priv_tx_q *txq = fep->tx_queue[queue];
drivers/net/ethernet/freescale/fec_main.c
4535
rxq = pool ? fec_alloc_new_rxq_xsk(fep, queue, pool) :
drivers/net/ethernet/freescale/fec_main.c
4536
fec_alloc_new_rxq_pp(fep, queue);
drivers/net/ethernet/freescale/fec_main.c
4543
fec_free_rxq(fep->rx_queue[queue]);
drivers/net/ethernet/freescale/fec_main.c
4544
fep->rx_queue[queue] = rxq;
drivers/net/ethernet/freescale/fec_main.c
4563
u16 queue)
drivers/net/ethernet/freescale/fec_main.c
4574
struct fec_enet_priv_rx_q *rxq = fep->rx_queue[queue];
drivers/net/ethernet/freescale/fec_main.c
4575
struct fec_enet_priv_tx_q *txq = fep->tx_queue[queue];
drivers/net/ethernet/freescale/fec_main.c
4583
err = fec_xsk_restart_napi(fep, pool, queue);
drivers/net/ethernet/freescale/fec_main.c
4593
u16 queue)
drivers/net/ethernet/freescale/fec_main.c
4595
struct fec_enet_priv_tx_q *txq = fep->tx_queue[queue];
drivers/net/ethernet/freescale/fec_main.c
4600
struct fec_enet_priv_rx_q *rxq = fep->rx_queue[queue];
drivers/net/ethernet/freescale/fec_main.c
4609
err = fec_xsk_restart_napi(fep, NULL, queue);
drivers/net/ethernet/freescale/fec_main.c
4620
u16 queue)
drivers/net/ethernet/freescale/fec_main.c
4622
if (queue >= fep->num_rx_queues || queue >= fep->num_tx_queues)
drivers/net/ethernet/freescale/fec_main.c
4625
return pool ? fec_enable_xsk_pool(fep, pool, queue) :
drivers/net/ethernet/freescale/fec_main.c
4626
fec_disable_xsk_pool(fep, queue);
drivers/net/ethernet/freescale/fec_main.c
4764
u32 dma_sync_len, int queue)
drivers/net/ethernet/freescale/fec_main.c
4766
struct netdev_queue *nq = netdev_get_tx_queue(fep->netdev, queue);
drivers/net/ethernet/freescale/fec_main.c
4767
struct fec_enet_priv_tx_q *txq = fep->tx_queue[queue];
drivers/net/ethernet/freescale/fec_main.c
4791
unsigned int queue;
drivers/net/ethernet/freescale/fec_main.c
4794
queue = fec_enet_xdp_get_tx_queue(fep, cpu);
drivers/net/ethernet/freescale/fec_main.c
4795
txq = fep->tx_queue[queue];
drivers/net/ethernet/freescale/fec_main.c
4796
nq = netdev_get_tx_queue(fep->netdev, queue);
drivers/net/ethernet/freescale/fec_main.c
4816
static int fec_enet_xsk_wakeup(struct net_device *ndev, u32 queue, u32 flags)
drivers/net/ethernet/freescale/fec_main.c
4824
if (queue >= fep->num_rx_queues || queue >= fep->num_tx_queues)
drivers/net/ethernet/freescale/fec_main.c
4827
rxq = fep->rx_queue[queue];
drivers/net/ethernet/freescale/fec_main.c
83
u32 dma_sync_len, int queue);
drivers/net/ethernet/freescale/fec_main.c
948
unsigned short queue;
drivers/net/ethernet/freescale/fec_main.c
953
queue = skb_get_queue_mapping(skb);
drivers/net/ethernet/freescale/fec_main.c
954
txq = fep->tx_queue[queue];
drivers/net/ethernet/freescale/fec_main.c
955
nq = netdev_get_tx_queue(ndev, queue);
drivers/net/ethernet/freescale/ucc_geth_ethtool.c
182
int queue = 0;
drivers/net/ethernet/freescale/ucc_geth_ethtool.c
189
ring->rx_pending = ug_info->bdRingLenRx[queue];
drivers/net/ethernet/freescale/ucc_geth_ethtool.c
190
ring->rx_mini_pending = ug_info->bdRingLenRx[queue];
drivers/net/ethernet/freescale/ucc_geth_ethtool.c
191
ring->rx_jumbo_pending = ug_info->bdRingLenRx[queue];
drivers/net/ethernet/freescale/ucc_geth_ethtool.c
192
ring->tx_pending = ug_info->bdRingLenTx[queue];
drivers/net/ethernet/freescale/ucc_geth_ethtool.c
203
int queue = 0, ret = 0;
drivers/net/ethernet/freescale/ucc_geth_ethtool.c
224
ug_info->bdRingLenRx[queue] = ring->rx_pending;
drivers/net/ethernet/freescale/ucc_geth_ethtool.c
225
ug_info->bdRingLenTx[queue] = ring->tx_pending;
drivers/net/ethernet/hisilicon/hibmcge/hbg_common.h
71
struct hbg_buffer *queue;
drivers/net/ethernet/hisilicon/hibmcge/hbg_txrx.c
154
buffer = &ring->queue[ntu];
drivers/net/ethernet/hisilicon/hibmcge/hbg_txrx.c
217
buffer = &ring->queue[ntc];
drivers/net/ethernet/hisilicon/hibmcge/hbg_txrx.c
416
buffer = &ring->queue[ring->ntu];
drivers/net/ethernet/hisilicon/hibmcge/hbg_txrx.c
495
buffer = &ring->queue[ring->ntc];
drivers/net/ethernet/hisilicon/hibmcge/hbg_txrx.c
578
if (!ring->queue)
drivers/net/ethernet/hisilicon/hibmcge/hbg_txrx.c
585
buffer = &ring->queue[i];
drivers/net/ethernet/hisilicon/hibmcge/hbg_txrx.c
593
ring->len * sizeof(*ring->queue),
drivers/net/ethernet/hisilicon/hibmcge/hbg_txrx.c
594
ring->queue, ring->queue_dma);
drivers/net/ethernet/hisilicon/hibmcge/hbg_txrx.c
595
ring->queue = NULL;
drivers/net/ethernet/hisilicon/hibmcge/hbg_txrx.c
620
ring->queue = dma_alloc_coherent(&priv->pdev->dev,
drivers/net/ethernet/hisilicon/hibmcge/hbg_txrx.c
621
len * sizeof(*ring->queue),
drivers/net/ethernet/hisilicon/hibmcge/hbg_txrx.c
623
if (!ring->queue)
drivers/net/ethernet/hisilicon/hibmcge/hbg_txrx.c
627
buffer = &ring->queue[i];
drivers/net/ethernet/hisilicon/hibmcge/hbg_txrx.c
650
ring->len * sizeof(*ring->queue),
drivers/net/ethernet/hisilicon/hibmcge/hbg_txrx.c
651
ring->queue, ring->queue_dma);
drivers/net/ethernet/hisilicon/hibmcge/hbg_txrx.c
652
ring->queue = NULL;
drivers/net/ethernet/hisilicon/hisi_femac.c
348
struct hisi_femac_queue *queue,
drivers/net/ethernet/hisilicon/hisi_femac.c
351
queue->skb = devm_kcalloc(dev, num, sizeof(struct sk_buff *),
drivers/net/ethernet/hisilicon/hisi_femac.c
353
if (!queue->skb)
drivers/net/ethernet/hisilicon/hisi_femac.c
356
queue->dma_phys = devm_kcalloc(dev, num, sizeof(dma_addr_t),
drivers/net/ethernet/hisilicon/hisi_femac.c
358
if (!queue->dma_phys)
drivers/net/ethernet/hisilicon/hisi_femac.c
361
queue->num = num;
drivers/net/ethernet/hisilicon/hisi_femac.c
362
queue->head = 0;
drivers/net/ethernet/hisilicon/hisi_femac.c
363
queue->tail = 0;
drivers/net/ethernet/hisilicon/hns/hnae.h
482
void (*get_ring_bdnum_limit)(struct hnae_queue *queue,
drivers/net/ethernet/hisilicon/hns/hns_ae_adapt.c
465
static void hns_ae_get_ring_bdnum_limit(struct hnae_queue *queue,
drivers/net/ethernet/hisilicon/hns/hns_ae_adapt.c
625
struct hnae_queue *queue;
drivers/net/ethernet/hisilicon/hns/hns_ae_adapt.c
639
queue = handle->qs[idx];
drivers/net/ethernet/hisilicon/hns/hns_ae_adapt.c
640
hns_rcb_update_stats(queue);
drivers/net/ethernet/hisilicon/hns/hns_ae_adapt.c
642
tx_bytes += queue->tx_ring.stats.tx_bytes;
drivers/net/ethernet/hisilicon/hns/hns_ae_adapt.c
643
tx_packets += queue->tx_ring.stats.tx_pkts;
drivers/net/ethernet/hisilicon/hns/hns_ae_adapt.c
644
rx_bytes += queue->rx_ring.stats.rx_bytes;
drivers/net/ethernet/hisilicon/hns/hns_ae_adapt.c
645
rx_packets += queue->rx_ring.stats.rx_pkts;
drivers/net/ethernet/hisilicon/hns/hns_ae_adapt.c
647
rx_errors += queue->rx_ring.stats.err_pkt_len
drivers/net/ethernet/hisilicon/hns/hns_ae_adapt.c
648
+ queue->rx_ring.stats.l2_err
drivers/net/ethernet/hisilicon/hns/hns_ae_adapt.c
649
+ queue->rx_ring.stats.l3l4_csum_err;
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
1032
void hns_rcb_get_ring_regs(struct hnae_queue *queue, void *data)
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
1036
= container_of(queue, struct ring_pair_cb, q);
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
1040
regs[0] = dsaf_read_dev(queue, RCB_RING_RX_RING_BASEADDR_L_REG);
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
1041
regs[1] = dsaf_read_dev(queue, RCB_RING_RX_RING_BASEADDR_H_REG);
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
1042
regs[2] = dsaf_read_dev(queue, RCB_RING_RX_RING_BD_NUM_REG);
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
1043
regs[3] = dsaf_read_dev(queue, RCB_RING_RX_RING_BD_LEN_REG);
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
1044
regs[4] = dsaf_read_dev(queue, RCB_RING_RX_RING_PKTLINE_REG);
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
1045
regs[5] = dsaf_read_dev(queue, RCB_RING_RX_RING_TAIL_REG);
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
1046
regs[6] = dsaf_read_dev(queue, RCB_RING_RX_RING_HEAD_REG);
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
1047
regs[7] = dsaf_read_dev(queue, RCB_RING_RX_RING_FBDNUM_REG);
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
1048
regs[8] = dsaf_read_dev(queue, RCB_RING_RX_RING_PKTNUM_RECORD_REG);
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
1050
regs[9] = dsaf_read_dev(queue, RCB_RING_TX_RING_BASEADDR_L_REG);
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
1051
regs[10] = dsaf_read_dev(queue, RCB_RING_TX_RING_BASEADDR_H_REG);
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
1052
regs[11] = dsaf_read_dev(queue, RCB_RING_TX_RING_BD_NUM_REG);
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
1053
regs[12] = dsaf_read_dev(queue, RCB_RING_TX_RING_BD_LEN_REG);
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
1054
regs[13] = dsaf_read_dev(queue, RCB_RING_TX_RING_PKTLINE_REG);
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
1055
regs[15] = dsaf_read_dev(queue, RCB_RING_TX_RING_TAIL_REG);
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
1056
regs[16] = dsaf_read_dev(queue, RCB_RING_TX_RING_HEAD_REG);
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
1057
regs[17] = dsaf_read_dev(queue, RCB_RING_TX_RING_FBDNUM_REG);
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
1058
regs[18] = dsaf_read_dev(queue, RCB_RING_TX_RING_OFFSET_REG);
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
1059
regs[19] = dsaf_read_dev(queue, RCB_RING_TX_RING_PKTNUM_RECORD_REG);
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
1061
regs[20] = dsaf_read_dev(queue, RCB_RING_PREFETCH_EN_REG);
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
1062
regs[21] = dsaf_read_dev(queue, RCB_RING_CFG_VF_NUM_REG);
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
1063
regs[22] = dsaf_read_dev(queue, RCB_RING_ASID_REG);
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
1064
regs[23] = dsaf_read_dev(queue, RCB_RING_RX_VM_REG);
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
1065
regs[24] = dsaf_read_dev(queue, RCB_RING_T0_BE_RST);
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
1066
regs[25] = dsaf_read_dev(queue, RCB_RING_COULD_BE_RST);
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
1067
regs[26] = dsaf_read_dev(queue, RCB_RING_WRR_WEIGHT_REG);
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
1069
regs[27] = dsaf_read_dev(queue, RCB_RING_INTMSK_RXWL_REG);
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
1070
regs[28] = dsaf_read_dev(queue, RCB_RING_INTSTS_RX_RING_REG);
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
1071
regs[29] = dsaf_read_dev(queue, RCB_RING_INTMSK_TXWL_REG);
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
1072
regs[30] = dsaf_read_dev(queue, RCB_RING_INTSTS_TX_RING_REG);
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
1073
regs[31] = dsaf_read_dev(queue, RCB_RING_INTMSK_RX_OVERTIME_REG);
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
1074
regs[32] = dsaf_read_dev(queue, RCB_RING_INTSTS_RX_OVERTIME_REG);
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
1075
regs[33] = dsaf_read_dev(queue, RCB_RING_INTMSK_TX_OVERTIME_REG);
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
1076
regs[34] = dsaf_read_dev(queue, RCB_RING_INTSTS_TX_OVERTIME_REG);
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
809
void hns_rcb_update_stats(struct hnae_queue *queue)
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
812
container_of(queue, struct ring_pair_cb, q);
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
818
hw_stats->rx_pkts += dsaf_read_dev(queue,
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
820
dsaf_write_dev(queue, RCB_RING_RX_RING_PKTNUM_RECORD_REG, 0x1);
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
827
hw_stats->tx_pkts += dsaf_read_dev(queue,
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
829
dsaf_write_dev(queue, RCB_RING_TX_RING_PKTNUM_RECORD_REG, 0x1);
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
842
void hns_rcb_get_stats(struct hnae_queue *queue, u64 *data)
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
846
container_of(queue, struct ring_pair_cb, q);
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
853
dsaf_read_dev(queue, RCB_RING_TX_RING_FBDNUM_REG);
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
855
regs_buff[4] = queue->tx_ring.stats.tx_pkts;
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
856
regs_buff[5] = queue->tx_ring.stats.tx_bytes;
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
857
regs_buff[6] = queue->tx_ring.stats.tx_err_cnt;
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
858
regs_buff[7] = queue->tx_ring.stats.io_err_cnt;
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
859
regs_buff[8] = queue->tx_ring.stats.sw_err_cnt;
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
860
regs_buff[9] = queue->tx_ring.stats.seg_pkt_cnt;
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
861
regs_buff[10] = queue->tx_ring.stats.restart_queue;
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
862
regs_buff[11] = queue->tx_ring.stats.tx_busy;
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
868
dsaf_read_dev(queue, RCB_RING_RX_RING_FBDNUM_REG);
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
870
regs_buff[16] = queue->rx_ring.stats.rx_pkts;
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
871
regs_buff[17] = queue->rx_ring.stats.rx_bytes;
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
872
regs_buff[18] = queue->rx_ring.stats.rx_err_cnt;
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
873
regs_buff[19] = queue->rx_ring.stats.io_err_cnt;
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
874
regs_buff[20] = queue->rx_ring.stats.sw_err_cnt;
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
875
regs_buff[21] = queue->rx_ring.stats.seg_pkt_cnt;
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
876
regs_buff[22] = queue->rx_ring.stats.reuse_pg_cnt;
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
877
regs_buff[23] = queue->rx_ring.stats.err_pkt_len;
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
878
regs_buff[24] = queue->rx_ring.stats.non_vld_descs;
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
879
regs_buff[25] = queue->rx_ring.stats.err_bd_num;
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
880
regs_buff[26] = queue->rx_ring.stats.l2_err;
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.c
881
regs_buff[27] = queue->rx_ring.stats.l3l4_csum_err;
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.h
147
void hns_rcb_update_stats(struct hnae_queue *queue);
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.h
149
void hns_rcb_get_stats(struct hnae_queue *queue, u64 *data);
drivers/net/ethernet/hisilicon/hns/hns_dsaf_rcb.h
157
void hns_rcb_get_ring_regs(struct hnae_queue *queue, void *data);
drivers/net/ethernet/hisilicon/hns/hns_ethtool.c
673
struct hnae_queue *queue;
drivers/net/ethernet/hisilicon/hns/hns_ethtool.c
676
queue = priv->ae_handle->qs[0];
drivers/net/ethernet/hisilicon/hns/hns_ethtool.c
680
ops->get_ring_bdnum_limit(queue, &uplimit);
drivers/net/ethernet/hisilicon/hns/hns_ethtool.c
684
param->rx_pending = queue->rx_ring.desc_num;
drivers/net/ethernet/hisilicon/hns/hns_ethtool.c
685
param->tx_pending = queue->tx_ring.desc_num;
drivers/net/ethernet/hisilicon/hns3/hns3_common/hclge_comm_tqp_stats.c
106
struct hnae3_queue *queue;
drivers/net/ethernet/hisilicon/hns3/hns3_common/hclge_comm_tqp_stats.c
110
queue = kinfo->tqp[i];
drivers/net/ethernet/hisilicon/hns3/hns3_common/hclge_comm_tqp_stats.c
111
tqp = container_of(queue, struct hclge_comm_tqp, q);
drivers/net/ethernet/hisilicon/hns3/hns3_ethtool.c
1534
u32 queue)
drivers/net/ethernet/hisilicon/hns3/hns3_ethtool.c
1541
tx_vector = priv->ring[queue].tqp_vector;
drivers/net/ethernet/hisilicon/hns3/hns3_ethtool.c
1542
rx_vector = priv->ring[queue_num + queue].tqp_vector;
drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c
10815
struct hnae3_queue *queue;
drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c
10817
queue = handle->kinfo.tqp[queue_id];
drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.c
10818
tqp = container_of(queue, struct hclge_comm_tqp, q);
drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.h
1121
static inline int hclge_get_queue_id(struct hnae3_queue *queue)
drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_main.h
1124
container_of(queue, struct hclge_comm_tqp, q);
drivers/net/ethernet/huawei/hinic/hinic_ethtool.c
632
struct ethtool_coalesce *coal, u16 queue)
drivers/net/ethernet/huawei/hinic/hinic_ethtool.c
638
if (queue == COALESCE_ALL_QUEUE) {
drivers/net/ethernet/huawei/hinic/hinic_ethtool.c
643
if (queue >= nic_dev->num_qps) {
drivers/net/ethernet/huawei/hinic/hinic_ethtool.c
645
"Invalid queue_id: %d\n", queue);
drivers/net/ethernet/huawei/hinic/hinic_ethtool.c
648
rx_intr_coal_info = &nic_dev->rx_intr_coalesce[queue];
drivers/net/ethernet/huawei/hinic/hinic_ethtool.c
649
tx_intr_coal_info = &nic_dev->tx_intr_coalesce[queue];
drivers/net/ethernet/huawei/hinic/hinic_ethtool.c
718
u16 queue, bool set_rx_coal)
drivers/net/ethernet/huawei/hinic/hinic_ethtool.c
723
if (queue == COALESCE_ALL_QUEUE) {
drivers/net/ethernet/huawei/hinic/hinic_ethtool.c
731
if (queue >= nic_dev->num_qps) {
drivers/net/ethernet/huawei/hinic/hinic_ethtool.c
733
"Invalid queue_id: %d\n", queue);
drivers/net/ethernet/huawei/hinic/hinic_ethtool.c
736
err = set_queue_coalesce(nic_dev, queue, intr_coal,
drivers/net/ethernet/huawei/hinic/hinic_ethtool.c
746
struct ethtool_coalesce *coal, u16 queue)
drivers/net/ethernet/huawei/hinic/hinic_ethtool.c
786
err = __set_hw_coal_param(nic_dev, &rx_intr_coal, queue, true);
drivers/net/ethernet/huawei/hinic/hinic_ethtool.c
791
err = __set_hw_coal_param(nic_dev, &tx_intr_coal, queue, false);
drivers/net/ethernet/huawei/hinic/hinic_ethtool.c
814
static int hinic_get_per_queue_coalesce(struct net_device *netdev, u32 queue,
drivers/net/ethernet/huawei/hinic/hinic_ethtool.c
817
return __hinic_get_coalesce(netdev, coal, queue);
drivers/net/ethernet/huawei/hinic/hinic_ethtool.c
820
static int hinic_set_per_queue_coalesce(struct net_device *netdev, u32 queue,
drivers/net/ethernet/huawei/hinic/hinic_ethtool.c
823
return __hinic_set_coalesce(netdev, coal, queue);
drivers/net/ethernet/huawei/hinic3/hinic3_nic_io.h
82
static inline void hinic3_write_db(struct hinic3_io_queue *queue, int cos,
drivers/net/ethernet/huawei/hinic3/hinic3_nic_io.h
91
DB_INFO_SET(queue->q_id, QID));
drivers/net/ethernet/huawei/hinic3/hinic3_nic_io.h
94
writeq(*((u64 *)&db), DB_ADDR(queue, pi));
drivers/net/ethernet/ibm/ehea/ehea_qmr.c
100
nr_pages = queue->queue_length / queue->pagesize;
drivers/net/ethernet/ibm/ehea/ehea_qmr.c
103
free_page((unsigned long)(queue->queue_pages)[i]);
drivers/net/ethernet/ibm/ehea/ehea_qmr.c
105
kfree(queue->queue_pages);
drivers/net/ethernet/ibm/ehea/ehea_qmr.c
25
static void *hw_qpageit_get_inc(struct hw_queue *queue)
drivers/net/ethernet/ibm/ehea/ehea_qmr.c
27
void *retvalue = hw_qeit_get(queue);
drivers/net/ethernet/ibm/ehea/ehea_qmr.c
29
queue->current_q_offset += queue->pagesize;
drivers/net/ethernet/ibm/ehea/ehea_qmr.c
30
if (queue->current_q_offset > queue->queue_length) {
drivers/net/ethernet/ibm/ehea/ehea_qmr.c
31
queue->current_q_offset -= queue->pagesize;
drivers/net/ethernet/ibm/ehea/ehea_qmr.c
40
static int hw_queue_ctor(struct hw_queue *queue, const u32 nr_of_pages,
drivers/net/ethernet/ibm/ehea/ehea_qmr.c
52
queue->queue_length = nr_of_pages * pagesize;
drivers/net/ethernet/ibm/ehea/ehea_qmr.c
53
queue->queue_pages = kmalloc_array(nr_of_pages, sizeof(void *),
drivers/net/ethernet/ibm/ehea/ehea_qmr.c
55
if (!queue->queue_pages)
drivers/net/ethernet/ibm/ehea/ehea_qmr.c
69
(queue->queue_pages)[i] = (struct ehea_page *)kpage;
drivers/net/ethernet/ibm/ehea/ehea_qmr.c
75
queue->current_q_offset = 0;
drivers/net/ethernet/ibm/ehea/ehea_qmr.c
76
queue->qe_size = qe_size;
drivers/net/ethernet/ibm/ehea/ehea_qmr.c
77
queue->pagesize = pagesize;
drivers/net/ethernet/ibm/ehea/ehea_qmr.c
78
queue->toggle_state = 1;
drivers/net/ethernet/ibm/ehea/ehea_qmr.c
83
if (!(queue->queue_pages)[i])
drivers/net/ethernet/ibm/ehea/ehea_qmr.c
85
free_page((unsigned long)(queue->queue_pages)[i]);
drivers/net/ethernet/ibm/ehea/ehea_qmr.c
90
static void hw_queue_dtor(struct hw_queue *queue)
drivers/net/ethernet/ibm/ehea/ehea_qmr.c
95
if (!queue || !queue->queue_pages)
drivers/net/ethernet/ibm/ehea/ehea_qmr.c
98
pages_per_kpage = PAGE_SIZE / queue->pagesize;
drivers/net/ethernet/ibm/ehea/ehea_qmr.h
196
static inline void *hw_qeit_calc(struct hw_queue *queue, u64 q_offset)
drivers/net/ethernet/ibm/ehea/ehea_qmr.h
200
if (q_offset >= queue->queue_length)
drivers/net/ethernet/ibm/ehea/ehea_qmr.h
201
q_offset -= queue->queue_length;
drivers/net/ethernet/ibm/ehea/ehea_qmr.h
202
current_page = (queue->queue_pages)[q_offset >> EHEA_PAGESHIFT];
drivers/net/ethernet/ibm/ehea/ehea_qmr.h
206
static inline void *hw_qeit_get(struct hw_queue *queue)
drivers/net/ethernet/ibm/ehea/ehea_qmr.h
208
return hw_qeit_calc(queue, queue->current_q_offset);
drivers/net/ethernet/ibm/ehea/ehea_qmr.h
211
static inline void hw_qeit_inc(struct hw_queue *queue)
drivers/net/ethernet/ibm/ehea/ehea_qmr.h
213
queue->current_q_offset += queue->qe_size;
drivers/net/ethernet/ibm/ehea/ehea_qmr.h
214
if (queue->current_q_offset >= queue->queue_length) {
drivers/net/ethernet/ibm/ehea/ehea_qmr.h
215
queue->current_q_offset = 0;
drivers/net/ethernet/ibm/ehea/ehea_qmr.h
217
queue->toggle_state = (~queue->toggle_state) & 1;
drivers/net/ethernet/ibm/ehea/ehea_qmr.h
221
static inline void *hw_qeit_get_inc(struct hw_queue *queue)
drivers/net/ethernet/ibm/ehea/ehea_qmr.h
223
void *retvalue = hw_qeit_get(queue);
drivers/net/ethernet/ibm/ehea/ehea_qmr.h
224
hw_qeit_inc(queue);
drivers/net/ethernet/ibm/ehea/ehea_qmr.h
228
static inline void *hw_qeit_get_inc_valid(struct hw_queue *queue)
drivers/net/ethernet/ibm/ehea/ehea_qmr.h
230
struct ehea_cqe *retvalue = hw_qeit_get(queue);
drivers/net/ethernet/ibm/ehea/ehea_qmr.h
234
if ((valid >> 7) == (queue->toggle_state & 1)) {
drivers/net/ethernet/ibm/ehea/ehea_qmr.h
236
hw_qeit_inc(queue);
drivers/net/ethernet/ibm/ehea/ehea_qmr.h
237
pref = hw_qeit_calc(queue, queue->current_q_offset);
drivers/net/ethernet/ibm/ehea/ehea_qmr.h
245
static inline void *hw_qeit_get_valid(struct hw_queue *queue)
drivers/net/ethernet/ibm/ehea/ehea_qmr.h
247
struct ehea_cqe *retvalue = hw_qeit_get(queue);
drivers/net/ethernet/ibm/ehea/ehea_qmr.h
251
pref = hw_qeit_calc(queue, queue->current_q_offset);
drivers/net/ethernet/ibm/ehea/ehea_qmr.h
256
if (!((valid >> 7) == (queue->toggle_state & 1)))
drivers/net/ethernet/ibm/ehea/ehea_qmr.h
261
static inline void *hw_qeit_reset(struct hw_queue *queue)
drivers/net/ethernet/ibm/ehea/ehea_qmr.h
263
queue->current_q_offset = 0;
drivers/net/ethernet/ibm/ehea/ehea_qmr.h
264
return hw_qeit_get(queue);
drivers/net/ethernet/ibm/ehea/ehea_qmr.h
267
static inline void *hw_qeit_eq_get_inc(struct hw_queue *queue)
drivers/net/ethernet/ibm/ehea/ehea_qmr.h
269
u64 last_entry_in_q = queue->queue_length - queue->qe_size;
drivers/net/ethernet/ibm/ehea/ehea_qmr.h
272
retvalue = hw_qeit_get(queue);
drivers/net/ethernet/ibm/ehea/ehea_qmr.h
273
queue->current_q_offset += queue->qe_size;
drivers/net/ethernet/ibm/ehea/ehea_qmr.h
274
if (queue->current_q_offset > last_entry_in_q) {
drivers/net/ethernet/ibm/ehea/ehea_qmr.h
275
queue->current_q_offset = 0;
drivers/net/ethernet/ibm/ehea/ehea_qmr.h
276
queue->toggle_state = (~queue->toggle_state) & 1;
drivers/net/ethernet/ibm/ehea/ehea_qmr.h
281
static inline void *hw_eqit_eq_get_inc_valid(struct hw_queue *queue)
drivers/net/ethernet/ibm/ehea/ehea_qmr.h
283
void *retvalue = hw_qeit_get(queue);
drivers/net/ethernet/ibm/ehea/ehea_qmr.h
285
if ((qe >> 7) == (queue->toggle_state & 1))
drivers/net/ethernet/ibm/ehea/ehea_qmr.h
286
hw_qeit_eq_get_inc(queue);
drivers/net/ethernet/ibm/ehea/ehea_qmr.h
295
struct hw_queue *queue;
drivers/net/ethernet/ibm/ehea/ehea_qmr.h
298
queue = &qp->hw_rqueue1;
drivers/net/ethernet/ibm/ehea/ehea_qmr.h
300
queue = &qp->hw_rqueue2;
drivers/net/ethernet/ibm/ehea/ehea_qmr.h
302
queue = &qp->hw_rqueue3;
drivers/net/ethernet/ibm/ehea/ehea_qmr.h
304
return hw_qeit_get_inc(queue);
drivers/net/ethernet/ibm/ehea/ehea_qmr.h
310
struct hw_queue *queue = &my_qp->hw_squeue;
drivers/net/ethernet/ibm/ehea/ehea_qmr.h
313
*wqe_index = (queue->current_q_offset) >> (7 + EHEA_SG_SQ);
drivers/net/ethernet/ibm/ehea/ehea_qmr.h
327
struct hw_queue *queue = &qp->hw_rqueue1;
drivers/net/ethernet/ibm/ehea/ehea_qmr.h
329
*wqe_index = (queue->current_q_offset) >> (7 + EHEA_SG_RQ1);
drivers/net/ethernet/ibm/ehea/ehea_qmr.h
330
return hw_qeit_get_valid(queue);
drivers/net/ethernet/ibm/ibmvnic.c
180
struct ibmvnic_sub_crq_queue *queue)
drivers/net/ethernet/ibm/ibmvnic.c
182
if (!(queue && queue->irq))
drivers/net/ethernet/ibm/ibmvnic.c
185
cpumask_clear(queue->affinity_mask);
drivers/net/ethernet/ibm/ibmvnic.c
187
if (irq_set_affinity_and_hint(queue->irq, NULL))
drivers/net/ethernet/ibm/ibmvnic.c
190
__func__, queue, queue->irq);
drivers/net/ethernet/ibm/ibmvnic.c
216
static int ibmvnic_set_queue_affinity(struct ibmvnic_sub_crq_queue *queue,
drivers/net/ethernet/ibm/ibmvnic.c
224
if (!(queue && queue->irq))
drivers/net/ethernet/ibm/ibmvnic.c
249
cpumask_copy(queue->affinity_mask, mask);
drivers/net/ethernet/ibm/ibmvnic.c
250
rc = irq_set_affinity_and_hint(queue->irq, queue->affinity_mask);
drivers/net/ethernet/ibm/ibmvnic.c
261
struct ibmvnic_sub_crq_queue *queue;
drivers/net/ethernet/ibm/ibmvnic.c
290
queue = rxqs[i_rxqs++];
drivers/net/ethernet/ibm/ibmvnic.c
293
queue = txqs[i_txqs++];
drivers/net/ethernet/ibm/ibmvnic.c
296
rc = ibmvnic_set_queue_affinity(queue, &cpu, &stragglers,
drivers/net/ethernet/ibm/ibmvnic.c
301
if (!queue || is_rx_queue)
drivers/net/ethernet/ibm/ibmvnic.c
305
cpumask_bits(queue->affinity_mask),
drivers/net/ethernet/ibm/ibmvnic.c
316
__func__, queue, queue->irq, rc);
drivers/net/ethernet/ibm/ibmvnic.c
4702
struct ibmvnic_crq_queue *queue = &adapter->crq;
drivers/net/ethernet/ibm/ibmvnic.c
4705
crq = &queue->msgs[queue->cur];
drivers/net/ethernet/ibm/ibmvnic.c
4707
if (++queue->cur == queue->size)
drivers/net/ethernet/ibm/ibmvnic.c
4708
queue->cur = 0;
drivers/net/ethernet/ibm/ibmvnic.c
6168
struct ibmvnic_crq_queue *queue = &adapter->crq;
drivers/net/ethernet/ibm/ibmvnic.c
6172
spin_lock_irqsave(&queue->lock, flags);
drivers/net/ethernet/ibm/ibmvnic.c
6186
spin_unlock_irqrestore(&queue->lock, flags);
drivers/net/ethernet/intel/fm10k/fm10k_pf.c
503
u16 vsi, queue, pc, q_idx;
drivers/net/ethernet/intel/fm10k/fm10k_pf.c
524
for (queue = 0; queue < queue_count; queue++, q_idx++) {
drivers/net/ethernet/intel/fm10k/fm10k_pf.c
540
for (queue = 0; queue < queue_count; queue++) {
drivers/net/ethernet/intel/i40e/i40e_adminq_cmd.h
937
__le16 queue;
drivers/net/ethernet/intel/i40e/i40e_common.c
3752
u16 vsi_seid, u16 queue, bool is_add,
drivers/net/ethernet/intel/i40e/i40e_common.c
3769
cmd->queue = cpu_to_le16(queue);
drivers/net/ethernet/intel/i40e/i40e_common.c
482
void i40e_pre_tx_queue_cfg(struct i40e_hw *hw, u32 queue, bool enable)
drivers/net/ethernet/intel/i40e/i40e_common.c
484
u32 abs_queue_idx = hw->func_caps.base_queue + queue;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2880
int queue)
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2891
if (queue < 0)
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2892
queue = 0;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2893
else if (queue >= vsi->num_queue_pairs)
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2896
rx_ring = vsi->rx_rings[queue];
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2897
tx_ring = vsi->tx_rings[queue];
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2947
static int i40e_get_per_queue_coalesce(struct net_device *netdev, u32 queue,
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2950
return __i40e_get_coalesce(netdev, ec, queue);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2963
int queue)
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2965
struct i40e_ring *rx_ring = vsi->rx_rings[queue];
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2966
struct i40e_ring *tx_ring = vsi->tx_rings[queue];
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3012
int queue)
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3023
if (queue < 0) {
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3026
} else if (queue < vsi->num_queue_pairs) {
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3027
cur_rx_itr = vsi->rx_rings[queue]->itr_setting;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3028
cur_tx_itr = vsi->tx_rings[queue]->itr_setting;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3088
if (queue < 0) {
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3092
i40e_set_itr_per_queue(vsi, ec, queue);
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3123
static int i40e_set_per_queue_coalesce(struct net_device *netdev, u32 queue,
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
3126
return __i40e_set_coalesce(netdev, ec, queue);
drivers/net/ethernet/intel/i40e/i40e_lan_hmc.c
1043
u16 queue)
drivers/net/ethernet/intel/i40e/i40e_lan_hmc.c
1049
I40E_HMC_LAN_TX, queue);
drivers/net/ethernet/intel/i40e/i40e_lan_hmc.c
1063
u16 queue,
drivers/net/ethernet/intel/i40e/i40e_lan_hmc.c
1070
I40E_HMC_LAN_TX, queue);
drivers/net/ethernet/intel/i40e/i40e_lan_hmc.c
1084
u16 queue)
drivers/net/ethernet/intel/i40e/i40e_lan_hmc.c
1090
I40E_HMC_LAN_RX, queue);
drivers/net/ethernet/intel/i40e/i40e_lan_hmc.c
1104
u16 queue,
drivers/net/ethernet/intel/i40e/i40e_lan_hmc.c
1111
I40E_HMC_LAN_RX, queue);
drivers/net/ethernet/intel/i40e/i40e_lan_hmc.h
150
u16 queue);
drivers/net/ethernet/intel/i40e/i40e_lan_hmc.h
152
u16 queue,
drivers/net/ethernet/intel/i40e/i40e_lan_hmc.h
155
u16 queue);
drivers/net/ethernet/intel/i40e/i40e_lan_hmc.h
157
u16 queue,
drivers/net/ethernet/intel/i40e/i40e_main.c
11248
u16 queue = FIELD_GET(I40E_GL_MDET_TX_QUEUE_MASK, reg) -
drivers/net/ethernet/intel/i40e/i40e_main.c
11252
event, queue, pf_num, vf_num);
drivers/net/ethernet/intel/i40e/i40e_main.c
11260
u16 queue = FIELD_GET(I40E_GL_MDET_RX_QUEUE_MASK, reg) -
drivers/net/ethernet/intel/i40e/i40e_main.c
11264
event, queue, func);
drivers/net/ethernet/intel/i40e/i40e_main.c
9569
u32 queue = le32_to_cpu(data->prtdcb_rupto);
drivers/net/ethernet/intel/i40e/i40e_main.c
9576
queue, qtx_ctl);
drivers/net/ethernet/intel/i40e/i40e_prototype.h
320
void i40e_pre_tx_queue_cfg(struct i40e_hw *hw, u32 queue, bool enable);
drivers/net/ethernet/intel/i40e/i40e_prototype.h
391
u16 vsi_seid, u16 queue, bool is_add,
drivers/net/ethernet/intel/iavf/iavf_ethtool.c
556
struct ethtool_coalesce *ec, int queue)
drivers/net/ethernet/intel/iavf/iavf_ethtool.c
564
if (queue < 0)
drivers/net/ethernet/intel/iavf/iavf_ethtool.c
565
queue = 0;
drivers/net/ethernet/intel/iavf/iavf_ethtool.c
566
else if (queue >= adapter->num_active_queues)
drivers/net/ethernet/intel/iavf/iavf_ethtool.c
569
rx_ring = &adapter->rx_rings[queue];
drivers/net/ethernet/intel/iavf/iavf_ethtool.c
570
tx_ring = &adapter->tx_rings[queue];
drivers/net/ethernet/intel/iavf/iavf_ethtool.c
612
static int iavf_get_per_queue_coalesce(struct net_device *netdev, u32 queue,
drivers/net/ethernet/intel/iavf/iavf_ethtool.c
615
return __iavf_get_coalesce(netdev, ec, queue);
drivers/net/ethernet/intel/iavf/iavf_ethtool.c
627
struct ethtool_coalesce *ec, int queue)
drivers/net/ethernet/intel/iavf/iavf_ethtool.c
629
struct iavf_ring *rx_ring = &adapter->rx_rings[queue];
drivers/net/ethernet/intel/iavf/iavf_ethtool.c
630
struct iavf_ring *tx_ring = &adapter->tx_rings[queue];
drivers/net/ethernet/intel/iavf/iavf_ethtool.c
685
struct ethtool_coalesce *ec, int queue)
drivers/net/ethernet/intel/iavf/iavf_ethtool.c
701
if (queue < 0) {
drivers/net/ethernet/intel/iavf/iavf_ethtool.c
705
} else if (queue < adapter->num_active_queues) {
drivers/net/ethernet/intel/iavf/iavf_ethtool.c
706
if (iavf_set_itr_per_queue(adapter, ec, queue))
drivers/net/ethernet/intel/iavf/iavf_ethtool.c
742
static int iavf_set_per_queue_coalesce(struct net_device *netdev, u32 queue,
drivers/net/ethernet/intel/iavf/iavf_ethtool.c
745
return __iavf_set_coalesce(netdev, ec, queue);
drivers/net/ethernet/intel/iavf/iavf_fdir.c
667
vc_msg->rule_cfg.action_set.actions[0].act_conf.queue.index = fltr->q_index;
drivers/net/ethernet/intel/iavf/iavf_main.c
4293
vact->act_conf.queue.index = q_index;
drivers/net/ethernet/intel/ice/devlink/health.c
292
u16 queue;
drivers/net/ethernet/intel/ice/devlink/health.c
331
ICE_DEVLINK_FMSG_PUT_FIELD(fmsg, mdd_event, queue);
drivers/net/ethernet/intel/ice/devlink/health.c
349
u16 vf_num, u8 event, u16 queue)
drivers/net/ethernet/intel/ice/devlink/health.c
356
.queue = queue,
drivers/net/ethernet/intel/ice/devlink/health.c
382
u16 queue;
drivers/net/ethernet/intel/ice/devlink/health.c
403
ICE_DEVLINK_FMSG_PUT_FIELD(fmsg, event, queue);
drivers/net/ethernet/intel/ice/devlink/health.c
438
.queue = tx_ring->q_index,
drivers/net/ethernet/intel/ice/devlink/health.h
68
u16 vf_num, u8 event, u16 queue);
drivers/net/ethernet/intel/ice/ice_main.c
1799
u16 queue = FIELD_GET(GL_MDET_TX_PQM_QNUM_M, reg);
drivers/net/ethernet/intel/ice/ice_main.c
1803
event, queue, pf_num, vf_num);
drivers/net/ethernet/intel/ice/ice_main.c
1805
event, queue);
drivers/net/ethernet/intel/ice/ice_main.c
1814
u16 queue = FIELD_GET(GL_MDET_TX_TCLAN_QNUM_M, reg);
drivers/net/ethernet/intel/ice/ice_main.c
1818
event, queue, pf_num, vf_num);
drivers/net/ethernet/intel/ice/ice_main.c
1820
event, queue);
drivers/net/ethernet/intel/ice/ice_main.c
1829
u16 queue = FIELD_GET(GL_MDET_RX_QNUM_M, reg);
drivers/net/ethernet/intel/ice/ice_main.c
1833
event, queue, pf_num, vf_num);
drivers/net/ethernet/intel/ice/ice_main.c
1835
queue);
drivers/net/ethernet/intel/ice/ice_main.c
9377
u32 queue;
drivers/net/ethernet/intel/ice/ice_main.c
9391
queue = tx_ring->q_index;
drivers/net/ethernet/intel/ice/ice_main.c
9395
ice_qp_dis(vsi, queue);
drivers/net/ethernet/intel/ice/ice_main.c
9397
err = ice_qp_ena(vsi, queue);
drivers/net/ethernet/intel/ice/ice_main.c
9400
queue);
drivers/net/ethernet/intel/ice/ice_main.c
9427
if (!qopt_off || qopt->queue < 0 || qopt->queue >= vsi->num_txq)
drivers/net/ethernet/intel/ice/ice_main.c
9431
set_bit(qopt->queue, pf->txtime_txqs);
drivers/net/ethernet/intel/ice/ice_main.c
9433
clear_bit(qopt->queue, pf->txtime_txqs);
drivers/net/ethernet/intel/ice/ice_main.c
9436
tx_ring = vsi->tx_rings[qopt->queue];
drivers/net/ethernet/intel/ice/ice_main.c
9443
str_enable_disable(qopt->enable), qopt->queue);
drivers/net/ethernet/intel/ice/ice_main.c
9448
str_enable_disable(qopt->enable), qopt->queue);
drivers/net/ethernet/intel/ice/ice_main.c
9451
clear_bit(qopt->queue, pf->txtime_txqs);
drivers/net/ethernet/intel/ice/ice_sriov.c
1165
u32 gldcb_rtctq, queue;
drivers/net/ethernet/intel/ice/ice_sriov.c
1173
queue = FIELD_GET(GLDCB_RTCTQ_RXQNUM_M, gldcb_rtctq);
drivers/net/ethernet/intel/ice/ice_sriov.c
1175
vf = ice_get_vf_from_pfq(pf, ice_globalq_to_pfq(pf, queue));
drivers/net/ethernet/intel/ice/ice_tc_lib.c
1032
ice_locate_vsi_using_queue(struct ice_vsi *vsi, int queue)
drivers/net/ethernet/intel/ice/ice_tc_lib.c
1049
if (queue >= offset && queue < offset + qcount) {
drivers/net/ethernet/intel/ice/ice_tc_lib.c
1064
u16 queue = tc_fltr->action.fwd.q.queue;
drivers/net/ethernet/intel/ice/ice_tc_lib.c
1066
return queue < vsi->num_rxq ? vsi->rx_rings[queue] : NULL;
drivers/net/ethernet/intel/ice/ice_tc_lib.c
1108
tc_fltr->action.fwd.q.queue);
drivers/net/ethernet/intel/ice/ice_tc_lib.c
1114
q = tc_fltr->action.fwd.q.queue;
drivers/net/ethernet/intel/ice/ice_tc_lib.c
1210
tc_fltr->action.fwd.q.queue,
drivers/net/ethernet/intel/ice/ice_tc_lib.c
1273
lkups_cnt, flags, tc_fltr->action.fwd.q.queue,
drivers/net/ethernet/intel/ice/ice_tc_lib.c
2041
u16 queue = act->rx_queue;
drivers/net/ethernet/intel/ice/ice_tc_lib.c
2043
if (queue >= vsi->num_rxq) {
drivers/net/ethernet/intel/ice/ice_tc_lib.c
2049
fltr->action.fwd.q.queue = queue;
drivers/net/ethernet/intel/ice/ice_tc_lib.c
2051
fltr->action.fwd.q.hw_queue = vsi->rxq_map[queue];
drivers/net/ethernet/intel/ice/ice_tc_lib.c
2056
ch_vsi = ice_locate_vsi_using_queue(vsi, fltr->action.fwd.q.queue);
drivers/net/ethernet/intel/ice/ice_tc_lib.h
213
struct ice_vsi *ice_locate_vsi_using_queue(struct ice_vsi *vsi, int queue);
drivers/net/ethernet/intel/ice/ice_tc_lib.h
60
u16 queue; /* forward to queue */
drivers/net/ethernet/intel/ice/virt/fdir.c
1180
input->q_index = action->act_conf.queue.index;
drivers/net/ethernet/intel/ice/virt/fdir.c
1185
input->q_index = action->act_conf.queue.index;
drivers/net/ethernet/intel/ice/virt/fdir.c
1186
input->q_region = action->act_conf.queue.region;
drivers/net/ethernet/intel/igb/igb.h
533
u8 queue;
drivers/net/ethernet/intel/igb/igb.h
855
const u8 *addr, u8 queue, u8 flags);
drivers/net/ethernet/intel/igb/igb.h
857
const u8 *addr, u8 queue, u8 flags);
drivers/net/ethernet/intel/igb/igb_main.c
1600
static void set_tx_desc_fetch_prio(struct e1000_hw *hw, int queue,
drivers/net/ethernet/intel/igb/igb_main.c
1606
WARN_ON(queue < 0 || queue > 4);
drivers/net/ethernet/intel/igb/igb_main.c
1608
val = rd32(E1000_I210_TXDCTL(queue));
drivers/net/ethernet/intel/igb/igb_main.c
1615
wr32(E1000_I210_TXDCTL(queue), val);
drivers/net/ethernet/intel/igb/igb_main.c
1618
static void set_queue_mode(struct e1000_hw *hw, int queue, enum queue_mode mode)
drivers/net/ethernet/intel/igb/igb_main.c
1623
WARN_ON(queue < 0 || queue > 1);
drivers/net/ethernet/intel/igb/igb_main.c
1625
val = rd32(E1000_I210_TQAVCC(queue));
drivers/net/ethernet/intel/igb/igb_main.c
1632
wr32(E1000_I210_TQAVCC(queue), val);
drivers/net/ethernet/intel/igb/igb_main.c
1669
static void igb_config_tx_modes(struct igb_adapter *adapter, int queue)
drivers/net/ethernet/intel/igb/igb_main.c
1678
WARN_ON(queue < 0 || queue > 1);
drivers/net/ethernet/intel/igb/igb_main.c
1679
ring = adapter->tx_ring[queue];
drivers/net/ethernet/intel/igb/igb_main.c
1686
set_tx_desc_fetch_prio(hw, queue, TX_QUEUE_PRIO_HIGH);
drivers/net/ethernet/intel/igb/igb_main.c
1687
set_queue_mode(hw, queue, QUEUE_MODE_STREAM_RESERVATION);
drivers/net/ethernet/intel/igb/igb_main.c
1689
set_tx_desc_fetch_prio(hw, queue, TX_QUEUE_PRIO_LOW);
drivers/net/ethernet/intel/igb/igb_main.c
1690
set_queue_mode(hw, queue, QUEUE_MODE_STRICT_PRIORITY);
drivers/net/ethernet/intel/igb/igb_main.c
1694
if (ring->cbs_enable || queue == 0) {
drivers/net/ethernet/intel/igb/igb_main.c
1704
if (queue == 0 && !ring->cbs_enable) {
drivers/net/ethernet/intel/igb/igb_main.c
1777
tqavcc = rd32(E1000_I210_TQAVCC(queue));
drivers/net/ethernet/intel/igb/igb_main.c
1780
wr32(E1000_I210_TQAVCC(queue), tqavcc);
drivers/net/ethernet/intel/igb/igb_main.c
1782
wr32(E1000_I210_TQAVHC(queue),
drivers/net/ethernet/intel/igb/igb_main.c
1787
tqavcc = rd32(E1000_I210_TQAVCC(queue));
drivers/net/ethernet/intel/igb/igb_main.c
1789
wr32(E1000_I210_TQAVCC(queue), tqavcc);
drivers/net/ethernet/intel/igb/igb_main.c
1792
wr32(E1000_I210_TQAVHC(queue), 0);
drivers/net/ethernet/intel/igb/igb_main.c
1840
queue,
drivers/net/ethernet/intel/igb/igb_main.c
1845
static int igb_save_txtime_params(struct igb_adapter *adapter, int queue,
drivers/net/ethernet/intel/igb/igb_main.c
1850
if (queue < 0 || queue > adapter->num_tx_queues)
drivers/net/ethernet/intel/igb/igb_main.c
1853
ring = adapter->tx_ring[queue];
drivers/net/ethernet/intel/igb/igb_main.c
1859
static int igb_save_cbs_params(struct igb_adapter *adapter, int queue,
drivers/net/ethernet/intel/igb/igb_main.c
1865
if (queue < 0 || queue > adapter->num_tx_queues)
drivers/net/ethernet/intel/igb/igb_main.c
1868
ring = adapter->tx_ring[queue];
drivers/net/ethernet/intel/igb/igb_main.c
2573
static void igb_offload_apply(struct igb_adapter *adapter, s32 queue)
drivers/net/ethernet/intel/igb/igb_main.c
2580
igb_config_tx_modes(adapter, queue);
drivers/net/ethernet/intel/igb/igb_main.c
2597
if (qopt->queue < 0 || qopt->queue > 1)
drivers/net/ethernet/intel/igb/igb_main.c
2600
err = igb_save_cbs_params(adapter, qopt->queue, qopt->enable,
drivers/net/ethernet/intel/igb/igb_main.c
2606
igb_offload_apply(adapter, qopt->queue);
drivers/net/ethernet/intel/igb/igb_main.c
2835
if (qopt->queue < 0 || qopt->queue > 1)
drivers/net/ethernet/intel/igb/igb_main.c
2838
err = igb_save_txtime_params(adapter, qopt->queue, qopt->enable);
drivers/net/ethernet/intel/igb/igb_main.c
2842
igb_offload_apply(adapter, qopt->queue);
drivers/net/ethernet/intel/igb/igb_main.c
7716
adapter->mac_table[i].queue = 0;
drivers/net/ethernet/intel/igb/igb_main.c
7721
static int igb_available_rars(struct igb_adapter *adapter, u8 queue)
drivers/net/ethernet/intel/igb/igb_main.c
7736
(adapter->mac_table[i].queue != queue))
drivers/net/ethernet/intel/igb/igb_main.c
7751
mac_table->queue = adapter->vfs_allocated_count;
drivers/net/ethernet/intel/igb/igb_main.c
7784
const u8 *addr, const u8 queue,
drivers/net/ethernet/intel/igb/igb_main.c
7805
adapter->mac_table[i].queue = queue;
drivers/net/ethernet/intel/igb/igb_main.c
7816
const u8 queue)
drivers/net/ethernet/intel/igb/igb_main.c
7818
return igb_add_mac_filter_flags(adapter, addr, queue, 0);
drivers/net/ethernet/intel/igb/igb_main.c
7828
const u8 *addr, const u8 queue,
drivers/net/ethernet/intel/igb/igb_main.c
7848
if (adapter->mac_table[i].queue != queue)
drivers/net/ethernet/intel/igb/igb_main.c
7859
adapter->mac_table[i].queue =
drivers/net/ethernet/intel/igb/igb_main.c
7863
adapter->mac_table[i].queue = 0;
drivers/net/ethernet/intel/igb/igb_main.c
7875
const u8 queue)
drivers/net/ethernet/intel/igb/igb_main.c
7877
return igb_del_mac_filter_flags(adapter, addr, queue, 0);
drivers/net/ethernet/intel/igb/igb_main.c
7881
const u8 *addr, u8 queue, u8 flags)
drivers/net/ethernet/intel/igb/igb_main.c
7891
return igb_add_mac_filter_flags(adapter, addr, queue,
drivers/net/ethernet/intel/igb/igb_main.c
7896
const u8 *addr, u8 queue, u8 flags)
drivers/net/ethernet/intel/igb/igb_main.c
7898
return igb_del_mac_filter_flags(adapter, addr, queue,
drivers/net/ethernet/intel/igb/igb_main.c
9842
adapter->mac_table[index].queue;
drivers/net/ethernet/intel/igb/igb_main.c
9846
adapter->mac_table[index].queue;
drivers/net/ethernet/intel/igb/igb_main.c
9869
adapter->mac_table[rar_entry].queue = vf;
drivers/net/ethernet/intel/igc/igc_main.c
3375
int queue)
drivers/net/ethernet/intel/igc/igc_main.c
3390
addr, queue);
drivers/net/ethernet/intel/igc/igc_main.c
3393
igc_set_mac_filter_hw(adapter, index, type, addr, queue);
drivers/net/ethernet/intel/igc/igc_main.c
3440
int queue)
drivers/net/ethernet/intel/igc/igc_main.c
3453
vlanpqf |= IGC_VLANPQF_QSEL(prio, queue);
drivers/net/ethernet/intel/igc/igc_main.c
3459
prio, queue);
drivers/net/ethernet/intel/igc/igc_main.c
3510
int queue)
drivers/net/ethernet/intel/igc/igc_main.c
3525
if (queue >= 0) {
drivers/net/ethernet/intel/igc/igc_main.c
3527
etqf |= (queue << IGC_ETQF_QUEUE_SHIFT);
drivers/net/ethernet/intel/igc/igc_main.c
3536
etype, queue);
drivers/net/ethernet/intel/igc/igc_main.c
3881
static void igc_set_default_queue_filter(struct igc_adapter *adapter, u32 queue)
drivers/net/ethernet/intel/igc/igc_main.c
3887
mrqc |= FIELD_PREP(IGC_MRQC_DEFAULT_QUEUE_MASK, queue);
drivers/net/ethernet/intel/igc/igc_main.c
6330
static int igc_save_launchtime_params(struct igc_adapter *adapter, int queue,
drivers/net/ethernet/intel/igc/igc_main.c
6335
if (queue < 0 || queue >= adapter->num_tx_queues)
drivers/net/ethernet/intel/igc/igc_main.c
6338
ring = adapter->tx_ring[queue];
drivers/net/ethernet/intel/igc/igc_main.c
6415
err = igc_save_launchtime_params(adapter, qopt->queue, qopt->enable);
drivers/net/ethernet/intel/igc/igc_main.c
6640
static int igc_save_cbs_params(struct igc_adapter *adapter, int queue,
drivers/net/ethernet/intel/igc/igc_main.c
6652
if (queue < 0 || queue > 1)
drivers/net/ethernet/intel/igc/igc_main.c
6655
ring = adapter->tx_ring[queue];
drivers/net/ethernet/intel/igc/igc_main.c
6665
if (queue == 1 && !cbs_status[0]) {
drivers/net/ethernet/intel/igc/igc_main.c
6671
if (queue == 0 && cbs_status[1]) {
drivers/net/ethernet/intel/igc/igc_main.c
6696
if (qopt->queue < 0 || qopt->queue > 1)
drivers/net/ethernet/intel/igc/igc_main.c
6699
err = igc_save_cbs_params(adapter, qopt->queue, qopt->enable,
drivers/net/ethernet/intel/igc/igc_main.c
919
const u8 *addr, int queue)
drivers/net/ethernet/intel/igc/igc_main.c
936
if (queue >= 0) {
drivers/net/ethernet/intel/igc/igc_main.c
938
rah |= (queue << IGC_RAH_QSEL_SHIFT);
drivers/net/ethernet/intel/igc/igc_tsn.c
176
u32 i, queue = 0;
drivers/net/ethernet/intel/igc/igc_tsn.c
186
queue |= GENMASK(offset + count - 1, offset);
drivers/net/ethernet/intel/igc/igc_tsn.c
189
return queue;
drivers/net/ethernet/intel/ixgbe/ixgbe.h
1003
u16 soft_id, u8 queue);
drivers/net/ethernet/intel/ixgbe/ixgbe.h
981
const u8 *addr, u16 queue);
drivers/net/ethernet/intel/ixgbe/ixgbe.h
983
const u8 *addr, u16 queue);
drivers/net/ethernet/intel/ixgbe/ixgbe.h
998
u8 queue);
drivers/net/ethernet/intel/ixgbe/ixgbe_82599.c
1375
u8 queue)
drivers/net/ethernet/intel/ixgbe/ixgbe_82599.c
1406
fdircmd |= (u32)queue << IXGBE_FDIRCMD_RX_QUEUE_SHIFT;
drivers/net/ethernet/intel/ixgbe/ixgbe_82599.c
1418
hw_dbg(hw, "Tx Queue=%x hash=%x\n", queue, (u32)fdirhashcmd);
drivers/net/ethernet/intel/ixgbe/ixgbe_82599.c
1643
u16 soft_id, u8 queue)
drivers/net/ethernet/intel/ixgbe/ixgbe_82599.c
1688
if (queue == IXGBE_FDIR_DROP_QUEUE)
drivers/net/ethernet/intel/ixgbe/ixgbe_82599.c
1691
fdircmd |= (u32)queue << IXGBE_FDIRCMD_RX_QUEUE_SHIFT;
drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c
2949
u8 queue;
drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c
2959
queue = IXGBE_FDIR_DROP_QUEUE;
drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c
2973
queue = adapter->rx_ring[ring]->reg_idx;
drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c
2975
queue = ((vf - 1) *
drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c
3055
&input->filter, input->sw_idx, queue);
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
10044
u8 queue;
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
10060
data->queue = adapter->rx_ring[vadapter->rx_base_queue]->reg_idx;
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
10061
data->action = data->queue;
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
10070
u8 *queue, u64 *action)
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
10082
*queue = vf * __ALIGN_MASK(1, ~vmdq->mask);
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
10093
data.queue = 0;
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
10098
*queue = data.queue;
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
10107
struct tcf_exts *exts, u64 *action, u8 *queue)
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
10119
*queue = IXGBE_FDIR_DROP_QUEUE;
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
10130
queue, action);
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
10140
struct tcf_exts *exts, u64 *action, u8 *queue)
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
10209
u8 queue;
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
10352
&queue);
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
10372
input->sw_idx, queue);
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
1051
static int ixgbe_get_vf_idx(struct ixgbe_adapter *adapter, u16 queue, u16 *vf)
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
1057
if (queue >= adapter->num_tx_queues)
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
1078
*vf = queue / queue_count;
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
1259
static u32 ixgbe_poll_tx_icache(struct ixgbe_hw *hw, u16 queue, u16 idx)
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
1261
IXGBE_WRITE_REG(hw, IXGBE_TXDESCIC, queue * idx);
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
1276
u16 queue)
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
1290
hw, queue, i);
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
5694
u8 queue;
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
5704
queue = IXGBE_FDIR_DROP_QUEUE;
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
5723
queue = adapter->rx_ring[ring]->reg_idx;
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
5725
queue = ((vf - 1) *
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
5730
&filter->filter, filter->sw_idx, queue);
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
874
u8 queue, u8 msix_vector)
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
883
index = (((direction * 64) + queue) >> 2) & 0x1F;
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
885
ivar &= ~(0xFF << (8 * (queue & 0x3)));
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
886
ivar |= (msix_vector << (8 * (queue & 0x3)));
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
898
index = ((queue & 1) * 8);
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
907
index = ((16 * (queue & 1)) + (8 * direction));
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
908
ivar = IXGBE_READ_REG(hw, IXGBE_IVAR(queue >> 1));
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
911
IXGBE_WRITE_REG(hw, IXGBE_IVAR(queue >> 1), ivar);
drivers/net/ethernet/intel/ixgbe/ixgbe_sriov.c
1703
u16 queue, queues_per_pool;
drivers/net/ethernet/intel/ixgbe/ixgbe_sriov.c
1742
for (queue = 0; queue < queues_per_pool; queue++) {
drivers/net/ethernet/intel/ixgbe/ixgbe_sriov.c
1743
unsigned int reg_idx = (vf * queues_per_pool) + queue;
drivers/net/ethernet/intel/ixgbe/ixgbe_sriov.c
707
u32 queue;
drivers/net/ethernet/intel/ixgbe/ixgbe_sriov.c
751
for (queue = 0; queue < q_per_pool; queue++) {
drivers/net/ethernet/intel/ixgbe/ixgbe_sriov.c
752
unsigned int reg_idx = (vf * q_per_pool) + queue;
drivers/net/ethernet/intel/ixgbe/ixgbe_sriov.h
23
int ixgbe_ndo_set_vf_mac(struct net_device *netdev, int queue, u8 *mac);
drivers/net/ethernet/intel/ixgbe/ixgbe_sriov.h
24
int ixgbe_ndo_set_vf_vlan(struct net_device *netdev, int queue, u16 vlan,
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
171
u8 queue, u8 msix_vector)
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
186
index = ((16 * (queue & 1)) + (8 * direction));
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
187
ivar = IXGBE_READ_REG(hw, IXGBE_VTIVAR(queue >> 1));
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
190
IXGBE_WRITE_REG(hw, IXGBE_VTIVAR(queue >> 1), ivar);
drivers/net/ethernet/lantiq_etop.c
474
int queue = skb_get_queue_mapping(skb);
drivers/net/ethernet/lantiq_etop.c
475
struct netdev_queue *txq = netdev_get_tx_queue(dev, queue);
drivers/net/ethernet/lantiq_etop.c
477
struct ltq_etop_chan *ch = &priv->ch[(queue << 1) | 1];
drivers/net/ethernet/marvell/mv643xx_eth.c
1000
queue = skb_get_queue_mapping(skb);
drivers/net/ethernet/marvell/mv643xx_eth.c
1001
txq = mp->txq + queue;
drivers/net/ethernet/marvell/mv643xx_eth.c
1002
nq = netdev_get_tx_queue(dev, queue);
drivers/net/ethernet/marvell/mv643xx_eth.c
2256
int queue;
drivers/net/ethernet/marvell/mv643xx_eth.c
2276
queue = fls(queue_mask) - 1;
drivers/net/ethernet/marvell/mv643xx_eth.c
2277
queue_mask = 1 << queue;
drivers/net/ethernet/marvell/mv643xx_eth.c
2284
txq_kick(mp->txq + queue);
drivers/net/ethernet/marvell/mv643xx_eth.c
2286
work_done += txq_reclaim(mp->txq + queue, work_tbd, 0);
drivers/net/ethernet/marvell/mv643xx_eth.c
2287
txq_maybe_wake(mp->txq + queue);
drivers/net/ethernet/marvell/mv643xx_eth.c
2289
work_done += rxq_process(mp->rxq + queue, work_tbd);
drivers/net/ethernet/marvell/mv643xx_eth.c
2291
work_done += rxq_refill(mp->rxq + queue, work_tbd);
drivers/net/ethernet/marvell/mv643xx_eth.c
996
int length, queue, ret;
drivers/net/ethernet/marvell/mvneta.c
1254
int queue;
drivers/net/ethernet/marvell/mvneta.c
1259
for (queue = 0; queue < txq_number; queue++) {
drivers/net/ethernet/marvell/mvneta.c
1260
struct mvneta_tx_queue *txq = &pp->txqs[queue];
drivers/net/ethernet/marvell/mvneta.c
1262
q_map |= (1 << queue);
drivers/net/ethernet/marvell/mvneta.c
1268
for (queue = 0; queue < rxq_number; queue++) {
drivers/net/ethernet/marvell/mvneta.c
1269
struct mvneta_rx_queue *rxq = &pp->rxqs[queue];
drivers/net/ethernet/marvell/mvneta.c
1272
q_map |= (1 << queue);
drivers/net/ethernet/marvell/mvneta.c
1375
static void mvneta_set_ucast_table(struct mvneta_port *pp, int queue)
drivers/net/ethernet/marvell/mvneta.c
1380
if (queue == -1) {
drivers/net/ethernet/marvell/mvneta.c
1383
val = 0x1 | (queue << 1);
drivers/net/ethernet/marvell/mvneta.c
1392
static void mvneta_set_special_mcast_table(struct mvneta_port *pp, int queue)
drivers/net/ethernet/marvell/mvneta.c
1397
if (queue == -1) {
drivers/net/ethernet/marvell/mvneta.c
1400
val = 0x1 | (queue << 1);
drivers/net/ethernet/marvell/mvneta.c
1410
static void mvneta_set_other_mcast_table(struct mvneta_port *pp, int queue)
drivers/net/ethernet/marvell/mvneta.c
1415
if (queue == -1) {
drivers/net/ethernet/marvell/mvneta.c
1420
val = 0x1 | (queue << 1);
drivers/net/ethernet/marvell/mvneta.c
1477
int queue;
drivers/net/ethernet/marvell/mvneta.c
1530
for (queue = 0; queue < txq_number; queue++) {
drivers/net/ethernet/marvell/mvneta.c
1531
mvreg_write(pp, MVETH_TXQ_TOKEN_COUNT_REG(queue), 0);
drivers/net/ethernet/marvell/mvneta.c
1532
mvreg_write(pp, MVETH_TXQ_TOKEN_CFG_REG(queue), 0);
drivers/net/ethernet/marvell/mvneta.c
1597
int queue;
drivers/net/ethernet/marvell/mvneta.c
1619
for (queue = 0; queue < txq_number; queue++) {
drivers/net/ethernet/marvell/mvneta.c
1620
val = mvreg_read(pp, MVNETA_TXQ_TOKEN_SIZE_REG(queue));
drivers/net/ethernet/marvell/mvneta.c
1627
mvreg_write(pp, MVNETA_TXQ_TOKEN_SIZE_REG(queue), val);
drivers/net/ethernet/marvell/mvneta.c
1634
int queue)
drivers/net/ethernet/marvell/mvneta.c
1651
if (queue == -1) {
drivers/net/ethernet/marvell/mvneta.c
1656
unicast_reg |= ((0x01 | (queue << 1)) << (8 * reg_offset));
drivers/net/ethernet/marvell/mvneta.c
1664
const unsigned char *addr, int queue)
drivers/net/ethernet/marvell/mvneta.c
1669
if (queue != -1) {
drivers/net/ethernet/marvell/mvneta.c
1679
mvneta_set_ucast_addr(pp, addr[5], queue);
drivers/net/ethernet/marvell/mvneta.c
1862
int queue = fls(cause) - 1;
drivers/net/ethernet/marvell/mvneta.c
1864
return &pp->txqs[queue];
drivers/net/ethernet/marvell/mvneta.c
3082
int queue)
drivers/net/ethernet/marvell/mvneta.c
3096
if (queue == -1)
drivers/net/ethernet/marvell/mvneta.c
3100
smc_table_reg |= ((0x01 | (queue << 1)) << (8 * reg_offset));
drivers/net/ethernet/marvell/mvneta.c
3117
int queue)
drivers/net/ethernet/marvell/mvneta.c
3128
if (queue == -1) {
drivers/net/ethernet/marvell/mvneta.c
3133
omc_table_reg |= ((0x01 | (queue << 1)) << (8 * reg_offset));
drivers/net/ethernet/marvell/mvneta.c
3149
int queue)
drivers/net/ethernet/marvell/mvneta.c
3154
mvneta_set_special_mcast_addr(pp, p_addr[5], queue);
drivers/net/ethernet/marvell/mvneta.c
3159
if (queue == -1) {
drivers/net/ethernet/marvell/mvneta.c
3176
mvneta_set_other_mcast_addr(pp, crc_result, queue);
drivers/net/ethernet/marvell/mvneta.c
3431
int queue;
drivers/net/ethernet/marvell/mvneta.c
3434
for (queue = 0; queue < txq_number; queue++)
drivers/net/ethernet/marvell/mvneta.c
3435
mvneta_txq_done_force(pp, &pp->txqs[queue]);
drivers/net/ethernet/marvell/mvneta.c
3653
int queue;
drivers/net/ethernet/marvell/mvneta.c
3655
for (queue = 0; queue < txq_number; queue++)
drivers/net/ethernet/marvell/mvneta.c
3656
mvneta_txq_deinit(pp, &pp->txqs[queue]);
drivers/net/ethernet/marvell/mvneta.c
3662
int queue;
drivers/net/ethernet/marvell/mvneta.c
3664
for (queue = 0; queue < rxq_number; queue++)
drivers/net/ethernet/marvell/mvneta.c
3665
mvneta_rxq_deinit(pp, &pp->rxqs[queue]);
drivers/net/ethernet/marvell/mvneta.c
3672
int queue;
drivers/net/ethernet/marvell/mvneta.c
3674
for (queue = 0; queue < rxq_number; queue++) {
drivers/net/ethernet/marvell/mvneta.c
3675
int err = mvneta_rxq_init(pp, &pp->rxqs[queue]);
drivers/net/ethernet/marvell/mvneta.c
3679
__func__, queue);
drivers/net/ethernet/marvell/mvneta.c
3691
int queue;
drivers/net/ethernet/marvell/mvneta.c
3693
for (queue = 0; queue < txq_number; queue++) {
drivers/net/ethernet/marvell/mvneta.c
3694
int err = mvneta_txq_init(pp, &pp->txqs[queue]);
drivers/net/ethernet/marvell/mvneta.c
3697
__func__, queue);
drivers/net/ethernet/marvell/mvneta.c
4733
int queue;
drivers/net/ethernet/marvell/mvneta.c
4735
for (queue = 0; queue < rxq_number; queue++) {
drivers/net/ethernet/marvell/mvneta.c
4736
struct mvneta_rx_queue *rxq = &pp->rxqs[queue];
drivers/net/ethernet/marvell/mvneta.c
4743
for (queue = 0; queue < txq_number; queue++) {
drivers/net/ethernet/marvell/mvneta.c
4744
struct mvneta_tx_queue *txq = &pp->txqs[queue];
drivers/net/ethernet/marvell/mvneta.c
5212
static int mvneta_setup_queue_rates(struct mvneta_port *pp, int queue,
drivers/net/ethernet/marvell/mvneta.c
5235
mvreg_write(pp, MVNETA_TXQ_BUCKET_REFILL_REG(queue), val);
drivers/net/ethernet/marvell/mvneta.c
5366
int queue;
drivers/net/ethernet/marvell/mvneta.c
5379
for (queue = 0; queue < txq_number; queue++) {
drivers/net/ethernet/marvell/mvneta.c
5380
struct mvneta_tx_queue *txq = &pp->txqs[queue];
drivers/net/ethernet/marvell/mvneta.c
5381
txq->id = queue;
drivers/net/ethernet/marvell/mvneta.c
5391
for (queue = 0; queue < rxq_number; queue++) {
drivers/net/ethernet/marvell/mvneta.c
5392
struct mvneta_rx_queue *rxq = &pp->rxqs[queue];
drivers/net/ethernet/marvell/mvneta.c
5393
rxq->id = queue;
drivers/net/ethernet/marvell/mvneta.c
5797
int queue;
drivers/net/ethernet/marvell/mvneta.c
5819
for (queue = 0; queue < rxq_number; queue++) {
drivers/net/ethernet/marvell/mvneta.c
5820
struct mvneta_rx_queue *rxq = &pp->rxqs[queue];
drivers/net/ethernet/marvell/mvneta.c
5825
for (queue = 0; queue < txq_number; queue++) {
drivers/net/ethernet/marvell/mvneta.c
5826
struct mvneta_tx_queue *txq = &pp->txqs[queue];
drivers/net/ethernet/marvell/mvneta.c
5844
int err, queue;
drivers/net/ethernet/marvell/mvneta.c
5871
for (queue = 0; queue < rxq_number; queue++) {
drivers/net/ethernet/marvell/mvneta.c
5872
struct mvneta_rx_queue *rxq = &pp->rxqs[queue];
drivers/net/ethernet/marvell/mvneta.c
5878
for (queue = 0; queue < txq_number; queue++) {
drivers/net/ethernet/marvell/mvneta.c
5879
struct mvneta_tx_queue *txq = &pp->txqs[queue];
drivers/net/ethernet/marvell/mvpp2/mvpp2_cls.c
1133
if (act->queue.ctx)
drivers/net/ethernet/marvell/mvpp2/mvpp2_cls.c
1148
if (act->queue.ctx) {
drivers/net/ethernet/marvell/mvpp2/mvpp2_cls.c
1150
ctx = mvpp22_rss_ctx(port, act->queue.ctx);
drivers/net/ethernet/marvell/mvpp2/mvpp2_cls.c
1157
qh = ((act->queue.index + port->first_rxq) >> 3) &
drivers/net/ethernet/marvell/mvpp2/mvpp2_cls.c
1159
ql = (act->queue.index + port->first_rxq) &
drivers/net/ethernet/marvell/mvpp2/mvpp2_cls.c
1328
if (act->queue.ctx && act->queue.index)
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
2260
int tx_port_num, val, queue, lrxq;
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
2281
for (queue = 0; queue < MVPP2_MAX_TXQ; queue++)
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
2283
MVPP2_TXQ_SCHED_TOKEN_CNTR_REG(queue), 0);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
2305
queue = port->rxqs[lrxq]->id;
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
2306
val = mvpp2_read(port->priv, MVPP2_RXQ_CONFIG_REG(queue));
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
2309
mvpp2_write(port->priv, MVPP2_RXQ_CONFIG_REG(queue), val);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
2320
int lrxq, queue;
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
2323
queue = port->rxqs[lrxq]->id;
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
2324
val = mvpp2_read(port->priv, MVPP2_RXQ_CONFIG_REG(queue));
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
2326
mvpp2_write(port->priv, MVPP2_RXQ_CONFIG_REG(queue), val);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
2333
int lrxq, queue;
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
2336
queue = port->rxqs[lrxq]->id;
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
2337
val = mvpp2_read(port->priv, MVPP2_RXQ_CONFIG_REG(queue));
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
2339
mvpp2_write(port->priv, MVPP2_RXQ_CONFIG_REG(queue), val);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
2349
int queue;
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
2354
for (queue = 0; queue < port->ntxqs; queue++) {
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
2355
struct mvpp2_tx_queue *txq = port->txqs[queue];
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
2358
qmap |= (1 << queue);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
2642
int queue;
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
2648
for (queue = 0; queue < port->ntxqs; queue++) {
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
2649
int id = port->txqs[queue]->id;
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
2844
int queue = fls(cause) - 1;
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
2846
return port->rxqs[queue];
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
2852
int queue = fls(cause) - 1;
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
2854
return port->txqs[queue];
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
3273
int queue;
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
3282
for (queue = 0; queue < port->ntxqs; queue++) {
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
3283
txq = port->txqs[queue];
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
3297
int queue;
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
3299
for (queue = 0; queue < port->nrxqs; queue++)
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
3300
mvpp2_rxq_deinit(port, port->rxqs[queue]);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
3309
int queue, err;
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
3311
for (queue = 0; queue < port->nrxqs; queue++) {
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
3312
err = mvpp2_rxq_init(port, port->rxqs[queue]);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
3331
int queue, err;
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
3333
for (queue = 0; queue < port->ntxqs; queue++) {
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
3334
txq = port->txqs[queue];
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
3340
if (queue < num_possible_cpus())
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
3341
netif_set_xps_queue(port->dev, cpumask_of(queue), queue);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
3346
for (queue = 0; queue < port->ntxqs; queue++) {
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
3347
txq = port->txqs[queue];
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
3375
struct mvpp2_hwtstamp_queue *queue;
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
3385
queue = &port->tx_hwtstamp_queue[nq];
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
3397
skb = queue->skb[id];
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
3398
queue->skb[id] = NULL;
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
4112
struct mvpp2_hwtstamp_queue *queue;
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
4133
queue = &port->tx_hwtstamp_queue[0];
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
4146
queue = &port->tx_hwtstamp_queue[1];
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
4152
i = queue->next;
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
4153
queue->next = (i + 1) & 31;
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
4154
if (queue->skb[i])
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
4155
dev_kfree_skb_any(queue->skb[i]);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
4156
queue->skb[i] = skb_get(skb);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
5405
int queue;
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
5407
for (queue = 0; queue < port->nrxqs; queue++) {
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
5408
struct mvpp2_rx_queue *rxq = port->rxqs[queue];
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
5421
for (queue = 0; queue < port->ntxqs; queue++) {
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
5422
struct mvpp2_tx_queue *txq = port->txqs[queue];
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
5989
int queue, err, val;
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
6025
for (queue = 0; queue < port->ntxqs; queue++) {
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
6026
int queue_phy_id = mvpp2_txq_phys(port->id, queue);
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
6042
txq->log_id = queue;
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
6049
port->txqs[queue] = txq;
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
6060
for (queue = 0; queue < port->nrxqs; queue++) {
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
6070
rxq->id = port->first_rxq + queue;
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
6072
rxq->logic_rxq = queue;
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
6074
port->rxqs[queue] = rxq;
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
6080
for (queue = 0; queue < port->nrxqs; queue++) {
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
6081
struct mvpp2_rx_queue *rxq = port->rxqs[queue];
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
6116
for (queue = 0; queue < port->ntxqs; queue++) {
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
6117
if (!port->txqs[queue])
drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
6119
free_percpu(port->txqs[queue]->pcpu);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_xsk.c
229
int queue, int budget)
drivers/net/ethernet/marvell/octeontx2/nic/otx2_xsk.c
234
budget = min(budget, otx2_read_free_sqe(pfvf, queue));
drivers/net/ethernet/marvell/octeontx2/nic/otx2_xsk.c
243
otx2_xsk_sq_append_pkt(pfvf, dma_addr, xdp_desc[i].len, queue);
drivers/net/ethernet/marvell/octeontx2/nic/otx2_xsk.h
21
int queue, int budget);
drivers/net/ethernet/marvell/skge.h
532
#define RB_ADDR(offs, queue) ((u16)B16_RAM_REGS + (u16)(queue) + (offs))
drivers/net/ethernet/marvell/sky2.h
818
#define RB_ADDR(offs, queue) ((u16) B16_RAM_REGS + (queue) + (offs))
drivers/net/ethernet/mediatek/mtk_eth_soc.c
1581
int queue = skb_get_queue_mapping(skb);
drivers/net/ethernet/mediatek/mtk_eth_soc.c
1584
txq = netdev_get_tx_queue(dev, queue);
drivers/net/ethernet/mediatek/mtk_eth_soc.c
1632
txd_info.qid = queue;
drivers/net/ethernet/mediatek/mtk_eth_soc.c
4726
unsigned int queue = 0;
drivers/net/ethernet/mediatek/mtk_eth_soc.c
4729
queue = skb_get_queue_mapping(skb) + 3;
drivers/net/ethernet/mediatek/mtk_eth_soc.c
4731
queue = mac->id;
drivers/net/ethernet/mediatek/mtk_eth_soc.c
4733
if (queue >= dev->num_tx_queues)
drivers/net/ethernet/mediatek/mtk_eth_soc.c
4734
queue = 0;
drivers/net/ethernet/mediatek/mtk_eth_soc.c
4736
return queue;
drivers/net/ethernet/mediatek/mtk_ppe.c
465
unsigned int queue)
drivers/net/ethernet/mediatek/mtk_ppe.c
471
*ib2 |= FIELD_PREP(MTK_FOE_IB2_QID_V2, queue);
drivers/net/ethernet/mediatek/mtk_ppe.c
475
*ib2 |= FIELD_PREP(MTK_FOE_IB2_QID, queue);
drivers/net/ethernet/mediatek/mtk_ppe.h
398
unsigned int queue);
drivers/net/ethernet/mediatek/mtk_ppe_offload.c
115
info->queue = path->mtk_wdma.queue;
drivers/net/ethernet/mediatek/mtk_ppe_offload.c
196
int pse_port, dsa_port, queue;
drivers/net/ethernet/mediatek/mtk_ppe_offload.c
199
mtk_foe_entry_set_wdma(eth, foe, info.wdma_idx, info.queue,
drivers/net/ethernet/mediatek/mtk_ppe_offload.c
235
queue = 3 + dsa_port;
drivers/net/ethernet/mediatek/mtk_ppe_offload.c
237
queue = pse_port - 1;
drivers/net/ethernet/mediatek/mtk_ppe_offload.c
239
mtk_foe_entry_set_queue(eth, foe, queue);
drivers/net/ethernet/mediatek/mtk_wed.h
58
u8 queue;
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
2558
static int mlx4_en_set_vf_mac(struct net_device *dev, int queue, u8 *mac)
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
2563
return mlx4_set_vf_mac(mdev->dev, en_priv->port, queue, mac);
drivers/net/ethernet/mellanox/mlx5/core/en.h
1232
int mlx5e_get_per_queue_coalesce(struct net_device *dev, u32 queue,
drivers/net/ethernet/mellanox/mlx5/core/en.h
1234
int mlx5e_set_per_queue_coalesce(struct net_device *dev, u32 queue,
drivers/net/ethernet/mellanox/mlx5/core/en_dim.c
68
void *queue)
drivers/net/ethernet/mellanox/mlx5/core/en_dim.c
80
dim->priv = queue;
drivers/net/ethernet/mellanox/mlx5/core/en_ethtool.c
632
static int mlx5e_ethtool_get_per_queue_coalesce(struct mlx5e_priv *priv, u32 queue,
drivers/net/ethernet/mellanox/mlx5/core/en_ethtool.c
645
if (chs->num <= queue) {
drivers/net/ethernet/mellanox/mlx5/core/en_ethtool.c
650
c = chs->c[queue];
drivers/net/ethernet/mellanox/mlx5/core/en_ethtool.c
686
int mlx5e_get_per_queue_coalesce(struct net_device *dev, u32 queue,
drivers/net/ethernet/mellanox/mlx5/core/en_ethtool.c
691
return mlx5e_ethtool_get_per_queue_coalesce(priv, queue, coal);
drivers/net/ethernet/mellanox/mlx5/core/en_ethtool.c
860
static int mlx5e_ethtool_set_per_queue_coalesce(struct mlx5e_priv *priv, u32 queue,
drivers/net/ethernet/mellanox/mlx5/core/en_ethtool.c
893
if (chs->num <= queue) {
drivers/net/ethernet/mellanox/mlx5/core/en_ethtool.c
898
c = chs->c[queue];
drivers/net/ethernet/mellanox/mlx5/core/en_ethtool.c
934
int mlx5e_set_per_queue_coalesce(struct net_device *dev, u32 queue,
drivers/net/ethernet/mellanox/mlx5/core/en_ethtool.c
939
return mlx5e_ethtool_set_per_queue_coalesce(priv, queue, coal);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/action.c
1705
struct mlx5hws_send_engine *queue;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/action.c
1711
queue = &ctx->send_queue[ctx->queues - 1];
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/action.c
1712
if (unlikely(mlx5hws_send_engine_err(queue))) {
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/action.c
1766
mlx5hws_send_stes_fw(ctx, queue, &ste_attr);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/action.c
2106
static void hws_action_modify_write(struct mlx5hws_send_engine *queue,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/action.c
2130
mlx5hws_arg_write(queue, NULL, arg_idx,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/action.c
2262
hws_action_modify_write(apply->queue,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/action.c
2295
mlx5hws_arg_write(apply->queue, NULL,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/action.c
2327
mlx5hws_arg_decapl3_write(apply->queue,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/action.h
82
struct mlx5hws_send_engine *queue;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/bwc.c
132
queue = &ctx->send_queue[rule_attr.queue_id];
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/bwc.c
133
mlx5hws_send_engine_flush_queue(queue);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/bwc.c
77
struct mlx5hws_send_engine *queue;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/pat_arg.c
300
void mlx5hws_arg_decapl3_write(struct mlx5hws_send_engine *queue,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/pat_arg.c
313
ctrl = mlx5hws_send_engine_post_start(queue);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/pat_arg.c
322
void mlx5hws_arg_write(struct mlx5hws_send_engine *queue,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/pat_arg.c
342
ctrl = mlx5hws_send_engine_post_start(queue);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/pat_arg.c
355
ctrl = mlx5hws_send_engine_post_start(queue);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/pat_arg.c
370
struct mlx5hws_send_engine *queue;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/pat_arg.c
376
queue = &ctx->send_queue[ctx->queues - 1];
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/pat_arg.c
378
mlx5hws_arg_write(queue, arg_data, arg_idx, arg_data, data_size);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/pat_arg.c
380
mlx5hws_send_engine_flush_queue(queue);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/pat_arg.h
83
void mlx5hws_arg_write(struct mlx5hws_send_engine *queue,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/pat_arg.h
89
void mlx5hws_arg_decapl3_write(struct mlx5hws_send_engine *queue,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/rule.c
110
static void hws_rule_gen_comp(struct mlx5hws_send_engine *queue,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/rule.c
126
mlx5hws_send_engine_inc_rule(queue);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/rule.c
127
mlx5hws_send_engine_gen_comp(queue, user_data, comp_status);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/rule.c
301
struct mlx5hws_send_engine *queue;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/rule.c
312
queue = &ctx->send_queue[attr->queue_id];
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/rule.c
313
if (unlikely(mlx5hws_send_engine_err(queue)))
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/rule.c
322
dep_wqe = mlx5hws_send_add_new_dep_wqe(queue);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/rule.c
330
apply.queue = queue;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/rule.c
339
mlx5hws_send_abort_new_dep_wqe(queue);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/rule.c
377
mlx5hws_send_abort_new_dep_wqe(queue);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/rule.c
393
mlx5hws_send_ste(queue, &ste_attr);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/rule.c
403
mlx5hws_send_engine_inc_rule(queue);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/rule.c
406
mlx5hws_send_all_dep_wqe(queue);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/rule.c
415
struct mlx5hws_send_engine *queue;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/rule.c
417
queue = &ctx->send_queue[attr->queue_id];
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/rule.c
419
hws_rule_gen_comp(queue, rule, false,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/rule.c
441
mlx5hws_send_all_dep_wqe(queue);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/rule.c
442
mlx5hws_send_engine_flush_queue(queue);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/rule.c
453
struct mlx5hws_send_engine *queue;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/rule.c
455
queue = &ctx->send_queue[attr->queue_id];
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/rule.c
457
if (unlikely(mlx5hws_send_engine_err(queue))) {
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/rule.c
478
hws_rule_gen_comp(queue, rule, false,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/rule.c
486
mlx5hws_send_engine_inc_rule(queue);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/rule.c
490
mlx5hws_send_all_dep_wqe(queue);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/rule.c
513
mlx5hws_send_ste(queue, &ste_attr);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/rule.c
577
struct mlx5hws_send_engine *queue = queue_ptr;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/rule.c
580
mlx5hws_send_all_dep_wqe(queue);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/rule.c
603
mlx5hws_send_ste(queue, &ste_attr);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/rule.c
615
struct mlx5hws_send_engine *queue;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/rule.c
622
queue = &ctx->send_queue[attr->queue_id];
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/rule.c
624
ret = mlx5hws_send_engine_err(queue);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/rule.c
649
mlx5hws_send_ste(queue, &ste_attr);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/rule.c
650
mlx5hws_send_engine_inc_rule(queue);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/rule.c
653
mlx5hws_send_all_dep_wqe(queue);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/rule.h
78
void *queue, void *user_data);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
10
mlx5hws_send_add_new_dep_wqe(struct mlx5hws_send_engine *queue)
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
1000
err = hws_send_ring_open_cq(ctx->mdev, queue, numa_node, &ring->send_cq);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
1004
err = hws_send_ring_open_sq(ctx, numa_node, queue, &ring->send_sq,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
1016
static void mlx5hws_send_queue_close(struct mlx5hws_send_engine *queue)
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
1018
if (!queue->num_entries)
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
1021
hws_send_ring_close(queue);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
1022
kfree(queue->completed.entries);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
1026
struct mlx5hws_send_engine *queue,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
1031
mutex_init(&queue->lock);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
1033
queue->num_entries = roundup_pow_of_two(queue_size);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
1034
queue->used_entries = 0;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
1036
queue->completed.entries = kzalloc_objs(queue->completed.entries[0],
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
1037
queue->num_entries);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
1038
if (!queue->completed.entries)
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
1041
queue->completed.pi = 0;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
1042
queue->completed.ci = 0;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
1043
queue->completed.mask = queue->num_entries - 1;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
1044
err = mlx5hws_send_ring_open(ctx, queue);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
1051
kfree(queue->completed.entries);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
1171
struct mlx5hws_send_engine *queue;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
1175
queue = &ctx->send_queue[queue_id];
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
1176
send_sq = &queue->send_ring.send_sq;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
1185
mlx5hws_send_all_dep_wqe(queue);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
1188
mlx5hws_send_engine_flush_queue(queue);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
1191
while (wait_comp && !mlx5hws_send_engine_empty(queue))
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
1192
hws_send_engine_poll_cq(queue, NULL, &polled, 0);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
12
struct mlx5hws_send_ring_sq *send_sq = &queue->send_ring.send_sq;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
1287
struct mlx5hws_send_engine *queue,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
1297
queue_id = queue - ctx->send_queue;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
13
unsigned int idx = send_sq->head_dep_idx++ & (queue->num_entries - 1);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
1345
mlx5hws_send_engine_gen_comp(queue, send_attr->user_data, MLX5HWS_FLOW_OP_SUCCESS);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
1354
mlx5hws_send_engine_gen_comp(queue, send_attr->user_data, MLX5HWS_FLOW_OP_ERROR);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
180
static void hws_send_wqe(struct mlx5hws_send_engine *queue,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
194
ctrl = mlx5hws_send_engine_post_start(queue);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
20
void mlx5hws_send_abort_new_dep_wqe(struct mlx5hws_send_engine *queue)
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
210
void mlx5hws_send_ste(struct mlx5hws_send_engine *queue,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
22
queue->send_ring.send_sq.head_dep_idx--;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
223
hws_send_wqe(queue, send_attr,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
238
hws_send_wqe(queue, send_attr,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
25
void mlx5hws_send_all_dep_wqe(struct mlx5hws_send_engine *queue)
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
252
static void hws_send_engine_retry_post_send(struct mlx5hws_send_engine *queue,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
27
struct mlx5hws_send_ring_sq *send_sq = &queue->send_ring.send_sq;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
275
ctrl = mlx5hws_send_engine_post_start(queue);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
296
void mlx5hws_send_engine_flush_queue(struct mlx5hws_send_engine *queue)
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
298
struct mlx5hws_send_ring_sq *sq = &queue->send_ring.send_sq;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
308
hws_send_engine_update_rule_resize(struct mlx5hws_send_engine *queue,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
328
mlx5hws_rule_move_hws_remove(priv->rule, queue, priv->user_data);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
347
static void hws_send_engine_dump_error_cqe(struct mlx5hws_send_engine *queue,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
360
if (queue->error_cqe_printed)
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
363
queue->error_cqe_printed = true;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
40
dep_wqe = &send_sq->dep_wqe[send_sq->tail_dep_idx++ & (queue->num_entries - 1)];
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
457
static void hws_send_engine_update_rule(struct mlx5hws_send_engine *queue,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
470
hws_send_engine_retry_post_send(queue, priv, wqe_cnt);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
473
hws_send_engine_dump_error_cqe(queue, priv, cqe);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
484
hws_send_engine_update_rule_resize(queue, priv, status);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
519
static void hws_send_engine_update(struct mlx5hws_send_engine *queue,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
538
hws_send_engine_update_rule(queue, priv, wqe_cnt,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
549
mlx5hws_send_engine_dec_rule(queue);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
551
mlx5hws_send_engine_gen_comp(queue, priv->user_data, status);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
57
mlx5hws_send_ste(queue, &ste_attr);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
597
static void hws_send_engine_poll_cq(struct mlx5hws_send_engine *queue,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
602
struct mlx5hws_send_ring *send_ring = &queue->send_ring;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
619
queue->err = true;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
625
hws_send_engine_update(queue, NULL, priv, res, polled, res_nb, 0);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
631
hws_send_engine_update(queue, cqe, priv, res, polled, res_nb, wqe_cnt);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
635
static void hws_send_engine_poll_list(struct mlx5hws_send_engine *queue,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
640
struct mlx5hws_completed_poll *comp = &queue->completed;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
65
mlx5hws_send_engine_post_start(struct mlx5hws_send_engine *queue)
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
650
mlx5hws_send_engine_dec_rule(queue);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
657
static int hws_send_engine_poll(struct mlx5hws_send_engine *queue,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
663
hws_send_engine_poll_list(queue, res, &polled, res_nb);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
668
hws_send_engine_poll_cq(queue, res, &polled, res_nb);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
683
struct mlx5hws_send_engine *queue,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
69
ctrl.queue = queue;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
703
buf_sz = queue->num_entries * MAX_WQES_PER_RULE;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
704
sq->dep_wqe = kzalloc_objs(*sq->dep_wqe, queue->num_entries);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
71
ctrl.send_ring = &queue->send_ring;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
716
sq->buf_mask = (queue->num_entries * MAX_WQES_PER_RULE) - 1;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
738
struct mlx5hws_send_engine *queue,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
818
struct mlx5hws_send_engine *queue,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
824
err = hws_send_ring_create_sq(mdev, pdn, sqc_data, queue, sq, cq);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
837
struct mlx5hws_send_engine *queue,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
849
buf_sz = queue->num_entries * MAX_WQES_PER_RULE;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
857
err = hws_send_ring_alloc_sq(ctx->mdev, numa_node, queue, sq, sqc_data);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
862
queue, sq, cq);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
878
struct mlx5hws_send_engine *queue,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
910
struct mlx5hws_send_engine *queue,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
948
struct mlx5hws_send_engine *queue,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
960
MLX5_SET(cqc, cqc_data, log_cq_size, ilog2(queue->num_entries));
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
962
err = hws_send_ring_alloc_cq(mdev, numa_node, queue, cqc_data, cq);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
966
err = hws_send_ring_create_cq(mdev, queue, cqc_data, cq);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
987
static void hws_send_ring_close(struct mlx5hws_send_engine *queue)
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
989
hws_send_ring_close_sq(&queue->send_ring.send_sq);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
990
hws_send_ring_close_cq(&queue->send_ring.send_cq);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
994
struct mlx5hws_send_engine *queue)
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.c
997
struct mlx5hws_send_ring *ring = &queue->send_ring;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.h
148
struct mlx5hws_send_engine *queue;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.h
187
mlx5hws_send_add_new_dep_wqe(struct mlx5hws_send_engine *queue);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.h
189
void mlx5hws_send_abort_new_dep_wqe(struct mlx5hws_send_engine *queue);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.h
191
void mlx5hws_send_all_dep_wqe(struct mlx5hws_send_engine *queue);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.h
208
mlx5hws_send_engine_post_start(struct mlx5hws_send_engine *queue);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.h
216
void mlx5hws_send_ste(struct mlx5hws_send_engine *queue,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.h
220
struct mlx5hws_send_engine *queue,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.h
223
void mlx5hws_send_engine_flush_queue(struct mlx5hws_send_engine *queue);
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.h
225
static inline bool mlx5hws_send_engine_empty(struct mlx5hws_send_engine *queue)
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.h
227
struct mlx5hws_send_ring_sq *send_sq = &queue->send_ring.send_sq;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.h
228
struct mlx5hws_send_ring_cq *send_cq = &queue->send_ring.send_cq;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.h
233
static inline bool mlx5hws_send_engine_full(struct mlx5hws_send_engine *queue)
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.h
235
return queue->used_entries >= queue->num_entries;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.h
238
static inline void mlx5hws_send_engine_inc_rule(struct mlx5hws_send_engine *queue)
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.h
240
queue->used_entries++;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.h
243
static inline void mlx5hws_send_engine_dec_rule(struct mlx5hws_send_engine *queue)
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.h
245
queue->used_entries--;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.h
248
static inline void mlx5hws_send_engine_gen_comp(struct mlx5hws_send_engine *queue,
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.h
252
struct mlx5hws_completed_poll *comp = &queue->completed;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.h
260
static inline bool mlx5hws_send_engine_err(struct mlx5hws_send_engine *queue)
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/send.h
262
return queue->err;
drivers/net/ethernet/mellanox/mlxsw/spectrum.c
1857
list_splice_init(&events->queue, &event_queue);
drivers/net/ethernet/mellanox/mlxsw/spectrum.c
1906
list_add_tail(&event->list, &events->queue);
drivers/net/ethernet/mellanox/mlxsw/spectrum.c
1921
list_for_each_entry_safe(event, next_event, &events->queue, list) {
drivers/net/ethernet/mellanox/mlxsw/spectrum.c
1974
INIT_LIST_HEAD(&events->queue);
drivers/net/ethernet/mellanox/mlxsw/spectrum.h
145
struct list_head queue;
drivers/net/ethernet/microchip/lan966x/lan966x_cbs.c
18
se_idx = SE_IDX_QUEUE + port->chip_port * NUM_PRIO_QUEUES + qopt->queue;
drivers/net/ethernet/microchip/lan966x/lan966x_cbs.c
57
se_idx = SE_IDX_QUEUE + port->chip_port * NUM_PRIO_QUEUES + qopt->queue;
drivers/net/ethernet/microchip/lan966x/lan966x_tbf.c
10
u32 queue = 0;
drivers/net/ethernet/microchip/lan966x/lan966x_tbf.c
15
queue = TC_H_MIN(qopt->parent) - 1;
drivers/net/ethernet/microchip/lan966x/lan966x_tbf.c
16
if (queue >= NUM_PRIO_QUEUES)
drivers/net/ethernet/microchip/lan966x/lan966x_tbf.c
23
se_idx = SE_IDX_QUEUE + port->chip_port * NUM_PRIO_QUEUES + queue;
drivers/net/ethernet/microchip/lan966x/lan966x_tbf.c
60
u32 queue = 0;
drivers/net/ethernet/microchip/lan966x/lan966x_tbf.c
64
queue = TC_H_MIN(qopt->parent) - 1;
drivers/net/ethernet/microchip/lan966x/lan966x_tbf.c
65
if (queue >= NUM_PRIO_QUEUES)
drivers/net/ethernet/microchip/lan966x/lan966x_tbf.c
72
se_idx = SE_IDX_QUEUE + port->chip_port * NUM_PRIO_QUEUES + queue;
drivers/net/ethernet/microchip/sparx5/sparx5_qos.h
21
#define SPX5_HSCH_L0_GET_IDX(port, queue) ((64 * (port)) + (8 * (queue)))
drivers/net/ethernet/microchip/sparx5/sparx5_tc.c
70
u32 queue = TC_H_MIN(parent) - 1;
drivers/net/ethernet/microchip/sparx5/sparx5_tc.c
72
*idx = SPX5_HSCH_L0_GET_IDX(portno, queue);
drivers/net/ethernet/microsoft/mana/gdma_main.c
1059
struct gdma_queue *queue;
drivers/net/ethernet/microsoft/mana/gdma_main.c
1065
queue = kzalloc_obj(*queue);
drivers/net/ethernet/microsoft/mana/gdma_main.c
1066
if (!queue)
drivers/net/ethernet/microsoft/mana/gdma_main.c
1069
gmi = &queue->mem_info;
drivers/net/ethernet/microsoft/mana/gdma_main.c
1081
queue->head = 0;
drivers/net/ethernet/microsoft/mana/gdma_main.c
1082
queue->tail = 0;
drivers/net/ethernet/microsoft/mana/gdma_main.c
1083
queue->queue_mem_ptr = gmi->virt_addr;
drivers/net/ethernet/microsoft/mana/gdma_main.c
1084
queue->queue_size = spec->queue_size;
drivers/net/ethernet/microsoft/mana/gdma_main.c
1085
queue->monitor_avl_buf = spec->monitor_avl_buf;
drivers/net/ethernet/microsoft/mana/gdma_main.c
1086
queue->type = spec->type;
drivers/net/ethernet/microsoft/mana/gdma_main.c
1087
queue->gdma_dev = gd;
drivers/net/ethernet/microsoft/mana/gdma_main.c
1089
err = mana_gd_create_eq(gd, spec, true, queue);
drivers/net/ethernet/microsoft/mana/gdma_main.c
1093
*queue_ptr = queue;
drivers/net/ethernet/microsoft/mana/gdma_main.c
1100
kfree(queue);
drivers/net/ethernet/microsoft/mana/gdma_main.c
1111
struct gdma_queue *queue;
drivers/net/ethernet/microsoft/mana/gdma_main.c
1118
queue = kzalloc_obj(*queue);
drivers/net/ethernet/microsoft/mana/gdma_main.c
1119
if (!queue)
drivers/net/ethernet/microsoft/mana/gdma_main.c
1122
gmi = &queue->mem_info;
drivers/net/ethernet/microsoft/mana/gdma_main.c
1134
queue->head = 0;
drivers/net/ethernet/microsoft/mana/gdma_main.c
1135
queue->tail = 0;
drivers/net/ethernet/microsoft/mana/gdma_main.c
1136
queue->queue_mem_ptr = gmi->virt_addr;
drivers/net/ethernet/microsoft/mana/gdma_main.c
1137
queue->queue_size = spec->queue_size;
drivers/net/ethernet/microsoft/mana/gdma_main.c
1138
queue->monitor_avl_buf = spec->monitor_avl_buf;
drivers/net/ethernet/microsoft/mana/gdma_main.c
1139
queue->type = spec->type;
drivers/net/ethernet/microsoft/mana/gdma_main.c
1140
queue->gdma_dev = gd;
drivers/net/ethernet/microsoft/mana/gdma_main.c
1143
mana_gd_create_cq(spec, queue);
drivers/net/ethernet/microsoft/mana/gdma_main.c
1145
*queue_ptr = queue;
drivers/net/ethernet/microsoft/mana/gdma_main.c
1152
kfree(queue);
drivers/net/ethernet/microsoft/mana/gdma_main.c
1157
void mana_gd_destroy_queue(struct gdma_context *gc, struct gdma_queue *queue)
drivers/net/ethernet/microsoft/mana/gdma_main.c
1159
struct gdma_mem_info *gmi = &queue->mem_info;
drivers/net/ethernet/microsoft/mana/gdma_main.c
1161
switch (queue->type) {
drivers/net/ethernet/microsoft/mana/gdma_main.c
1163
mana_gd_destroy_eq(gc, queue->eq.disable_needed, queue);
drivers/net/ethernet/microsoft/mana/gdma_main.c
1167
mana_gd_destroy_cq(gc, queue);
drivers/net/ethernet/microsoft/mana/gdma_main.c
1178
queue->type);
drivers/net/ethernet/microsoft/mana/gdma_main.c
1184
kfree(queue);
drivers/net/ethernet/microsoft/mana/gdma_main.c
1435
int mana_gd_post_and_ring(struct gdma_queue *queue,
drivers/net/ethernet/microsoft/mana/gdma_main.c
1439
struct gdma_context *gc = queue->gdma_dev->gdma_context;
drivers/net/ethernet/microsoft/mana/gdma_main.c
1442
err = mana_gd_post_work_request(queue, wqe_req, wqe_info);
drivers/net/ethernet/microsoft/mana/gdma_main.c
1445
queue->type, queue->queue_size, err);
drivers/net/ethernet/microsoft/mana/gdma_main.c
1449
mana_gd_wq_ring_doorbell(gc, queue);
drivers/net/ethernet/microsoft/mana/gdma_main.c
262
struct gdma_queue *queue)
drivers/net/ethernet/microsoft/mana/gdma_main.c
268
if (queue->type != GDMA_EQ)
drivers/net/ethernet/microsoft/mana/gdma_main.c
274
req.hdr.dev_id = queue->gdma_dev->dev_id;
drivers/net/ethernet/microsoft/mana/gdma_main.c
275
req.type = queue->type;
drivers/net/ethernet/microsoft/mana/gdma_main.c
276
req.pdid = queue->gdma_dev->pdid;
drivers/net/ethernet/microsoft/mana/gdma_main.c
277
req.doolbell_id = queue->gdma_dev->doorbell;
drivers/net/ethernet/microsoft/mana/gdma_main.c
278
req.gdma_region = queue->mem_info.dma_region_handle;
drivers/net/ethernet/microsoft/mana/gdma_main.c
279
req.queue_size = queue->queue_size;
drivers/net/ethernet/microsoft/mana/gdma_main.c
280
req.log2_throttle_limit = queue->eq.log2_throttle_limit;
drivers/net/ethernet/microsoft/mana/gdma_main.c
281
req.eq_pci_msix_index = queue->eq.msix_index;
drivers/net/ethernet/microsoft/mana/gdma_main.c
290
queue->id = resp.queue_index;
drivers/net/ethernet/microsoft/mana/gdma_main.c
291
queue->eq.disable_needed = true;
drivers/net/ethernet/microsoft/mana/gdma_main.c
292
queue->mem_info.dma_region_handle = GDMA_INVALID_DMA_REGION;
drivers/net/ethernet/microsoft/mana/gdma_main.c
296
static int mana_gd_disable_queue(struct gdma_queue *queue)
drivers/net/ethernet/microsoft/mana/gdma_main.c
298
struct gdma_context *gc = queue->gdma_dev->gdma_context;
drivers/net/ethernet/microsoft/mana/gdma_main.c
303
WARN_ON(queue->type != GDMA_EQ);
drivers/net/ethernet/microsoft/mana/gdma_main.c
308
req.hdr.dev_id = queue->gdma_dev->dev_id;
drivers/net/ethernet/microsoft/mana/gdma_main.c
309
req.type = queue->type;
drivers/net/ethernet/microsoft/mana/gdma_main.c
310
req.queue_index = queue->id;
drivers/net/ethernet/microsoft/mana/gdma_main.c
379
void mana_gd_wq_ring_doorbell(struct gdma_context *gc, struct gdma_queue *queue)
drivers/net/ethernet/microsoft/mana/gdma_main.c
384
mana_gd_ring_doorbell(gc, queue->gdma_dev->doorbell, queue->type,
drivers/net/ethernet/microsoft/mana/gdma_main.c
385
queue->id, queue->head * GDMA_WQE_BU_SIZE, 0);
drivers/net/ethernet/microsoft/mana/gdma_main.c
710
static int mana_gd_register_irq(struct gdma_queue *queue,
drivers/net/ethernet/microsoft/mana/gdma_main.c
713
struct gdma_dev *gd = queue->gdma_dev;
drivers/net/ethernet/microsoft/mana/gdma_main.c
733
queue->eq.msix_index = msi_index;
drivers/net/ethernet/microsoft/mana/gdma_main.c
739
list_add_rcu(&queue->entry, &gic->eq_list);
drivers/net/ethernet/microsoft/mana/gdma_main.c
745
static void mana_gd_deregister_irq(struct gdma_queue *queue)
drivers/net/ethernet/microsoft/mana/gdma_main.c
747
struct gdma_dev *gd = queue->gdma_dev;
drivers/net/ethernet/microsoft/mana/gdma_main.c
757
msix_index = queue->eq.msix_index;
drivers/net/ethernet/microsoft/mana/gdma_main.c
767
if (queue == eq) {
drivers/net/ethernet/microsoft/mana/gdma_main.c
774
queue->eq.msix_index = INVALID_PCI_MSIX_INDEX;
drivers/net/ethernet/microsoft/mana/gdma_main.c
828
struct gdma_queue *queue)
drivers/net/ethernet/microsoft/mana/gdma_main.c
833
err = mana_gd_test_eq(gc, queue);
drivers/net/ethernet/microsoft/mana/gdma_main.c
838
mana_gd_deregister_irq(queue);
drivers/net/ethernet/microsoft/mana/gdma_main.c
840
if (queue->eq.disable_needed)
drivers/net/ethernet/microsoft/mana/gdma_main.c
841
mana_gd_disable_queue(queue);
drivers/net/ethernet/microsoft/mana/gdma_main.c
846
bool create_hwq, struct gdma_queue *queue)
drivers/net/ethernet/microsoft/mana/gdma_main.c
853
queue->eq.msix_index = INVALID_PCI_MSIX_INDEX;
drivers/net/ethernet/microsoft/mana/gdma_main.c
854
queue->id = INVALID_QUEUE_ID;
drivers/net/ethernet/microsoft/mana/gdma_main.c
856
log2_num_entries = ilog2(queue->queue_size / GDMA_EQE_SIZE);
drivers/net/ethernet/microsoft/mana/gdma_main.c
864
err = mana_gd_register_irq(queue, spec);
drivers/net/ethernet/microsoft/mana/gdma_main.c
870
queue->eq.callback = spec->eq.callback;
drivers/net/ethernet/microsoft/mana/gdma_main.c
871
queue->eq.context = spec->eq.context;
drivers/net/ethernet/microsoft/mana/gdma_main.c
872
queue->head |= INITIALIZED_OWNER_BIT(log2_num_entries);
drivers/net/ethernet/microsoft/mana/gdma_main.c
873
queue->eq.log2_throttle_limit = spec->eq.log2_throttle_limit ?: 1;
drivers/net/ethernet/microsoft/mana/gdma_main.c
876
err = mana_gd_create_hw_eq(gc, queue);
drivers/net/ethernet/microsoft/mana/gdma_main.c
880
err = mana_gd_test_eq(gc, queue);
drivers/net/ethernet/microsoft/mana/gdma_main.c
888
mana_gd_destroy_eq(gc, false, queue);
drivers/net/ethernet/microsoft/mana/gdma_main.c
893
struct gdma_queue *queue)
drivers/net/ethernet/microsoft/mana/gdma_main.c
897
queue->head |= INITIALIZED_OWNER_BIT(log2_num_entries);
drivers/net/ethernet/microsoft/mana/gdma_main.c
898
queue->cq.parent = spec->cq.parent_eq;
drivers/net/ethernet/microsoft/mana/gdma_main.c
899
queue->cq.context = spec->cq.context;
drivers/net/ethernet/microsoft/mana/gdma_main.c
900
queue->cq.callback = spec->cq.callback;
drivers/net/ethernet/microsoft/mana/gdma_main.c
904
struct gdma_queue *queue)
drivers/net/ethernet/microsoft/mana/gdma_main.c
906
u32 id = queue->id;
drivers/net/ethernet/microsoft/mana/gdma_main.c
923
struct gdma_queue *queue;
drivers/net/ethernet/microsoft/mana/gdma_main.c
926
queue = kzalloc_obj(*queue);
drivers/net/ethernet/microsoft/mana/gdma_main.c
927
if (!queue)
drivers/net/ethernet/microsoft/mana/gdma_main.c
930
gmi = &queue->mem_info;
drivers/net/ethernet/microsoft/mana/gdma_main.c
938
queue->head = 0;
drivers/net/ethernet/microsoft/mana/gdma_main.c
939
queue->tail = 0;
drivers/net/ethernet/microsoft/mana/gdma_main.c
940
queue->queue_mem_ptr = gmi->virt_addr;
drivers/net/ethernet/microsoft/mana/gdma_main.c
941
queue->queue_size = spec->queue_size;
drivers/net/ethernet/microsoft/mana/gdma_main.c
942
queue->monitor_avl_buf = spec->monitor_avl_buf;
drivers/net/ethernet/microsoft/mana/gdma_main.c
943
queue->type = spec->type;
drivers/net/ethernet/microsoft/mana/gdma_main.c
944
queue->gdma_dev = gd;
drivers/net/ethernet/microsoft/mana/gdma_main.c
947
err = mana_gd_create_eq(gd, spec, false, queue);
drivers/net/ethernet/microsoft/mana/gdma_main.c
949
mana_gd_create_cq(spec, queue);
drivers/net/ethernet/microsoft/mana/gdma_main.c
954
*queue_ptr = queue;
drivers/net/ethernet/microsoft/mana/gdma_main.c
961
kfree(queue);
drivers/net/ethernet/microsoft/mana/hw_channel.c
297
struct gdma_queue **queue)
drivers/net/ethernet/microsoft/mana/hw_channel.c
308
return mana_gd_create_hwc_queue(hwc->gdma_dev, &spec, queue);
drivers/net/ethernet/microsoft/mana/hw_channel.c
315
struct gdma_queue **queue)
drivers/net/ethernet/microsoft/mana/hw_channel.c
326
return mana_gd_create_hwc_queue(hwc->gdma_dev, &spec, queue);
drivers/net/ethernet/microsoft/mana/hw_channel.c
332
struct gdma_queue **queue)
drivers/net/ethernet/microsoft/mana/hw_channel.c
344
return mana_gd_create_hwc_queue(hwc->gdma_dev, &spec, queue);
drivers/net/ethernet/microsoft/mana/hw_channel.c
527
struct gdma_queue *queue;
drivers/net/ethernet/microsoft/mana/hw_channel.c
546
err = mana_hwc_create_gdma_wq(hwc, q_type, queue_size, &queue);
drivers/net/ethernet/microsoft/mana/hw_channel.c
551
hwc_wq->gdma_wq = queue;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2658
u16 pseudo_hdr_offset, cksum_offset, queue;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2662
queue = skb_get_queue_mapping(skb);
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2663
ss = &mgp->ss[queue];
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2664
netdev_queue = netdev_get_tx_queue(mgp->dev, queue);
drivers/net/ethernet/netronome/nfp/abm/ctrl.c
103
threshold = band * NFP_NET_MAX_RX_RINGS + alink->queue_base + queue;
drivers/net/ethernet/netronome/nfp/abm/ctrl.c
132
unsigned int queue, enum nfp_abm_q_action act)
drivers/net/ethernet/netronome/nfp/abm/ctrl.c
136
qid = band * NFP_NET_MAX_RX_RINGS + alink->queue_base + queue;
drivers/net/ethernet/netronome/nfp/abm/ctrl.c
141
u64 nfp_abm_ctrl_stat_non_sto(struct nfp_abm_link *alink, unsigned int queue)
drivers/net/ethernet/netronome/nfp/abm/ctrl.c
149
band, queue, true, &val))
drivers/net/ethernet/netronome/nfp/abm/ctrl.c
157
u64 nfp_abm_ctrl_stat_sto(struct nfp_abm_link *alink, unsigned int queue)
drivers/net/ethernet/netronome/nfp/abm/ctrl.c
165
band, queue, true, &val))
drivers/net/ethernet/netronome/nfp/abm/ctrl.c
175
unsigned int queue, unsigned int off, u64 *val)
drivers/net/ethernet/netronome/nfp/abm/ctrl.c
179
unsigned int id = alink->queue_base + queue;
drivers/net/ethernet/netronome/nfp/abm/ctrl.c
190
NFP_Q_STAT_STRIDE, off, band, queue,
drivers/net/ethernet/netronome/nfp/abm/ctrl.c
196
unsigned int queue, struct nfp_alink_stats *stats)
drivers/net/ethernet/netronome/nfp/abm/ctrl.c
200
err = nfp_abm_ctrl_stat_basic(alink, band, queue, NFP_Q_STAT_PKTS,
drivers/net/ethernet/netronome/nfp/abm/ctrl.c
205
err = nfp_abm_ctrl_stat_basic(alink, band, queue, NFP_Q_STAT_BYTES,
drivers/net/ethernet/netronome/nfp/abm/ctrl.c
211
NFP_QLVL_BLOG_BYTES, band, queue, false,
drivers/net/ethernet/netronome/nfp/abm/ctrl.c
218
band, queue, false, &stats->backlog_pkts);
drivers/net/ethernet/netronome/nfp/abm/ctrl.c
224
band, queue, true, &stats->drops);
drivers/net/ethernet/netronome/nfp/abm/ctrl.c
230
band, queue, true, &stats->overlimits);
drivers/net/ethernet/netronome/nfp/abm/ctrl.c
234
unsigned int band, unsigned int queue,
drivers/net/ethernet/netronome/nfp/abm/ctrl.c
241
band, queue, true, &xstats->pdrop);
drivers/net/ethernet/netronome/nfp/abm/ctrl.c
247
band, queue, true, &xstats->ecn_marked);
drivers/net/ethernet/netronome/nfp/abm/ctrl.c
50
unsigned int queue, bool is_u64, u64 *res)
drivers/net/ethernet/netronome/nfp/abm/ctrl.c
58
qid = band * NFP_NET_MAX_RX_RINGS + alink->queue_base + queue;
drivers/net/ethernet/netronome/nfp/abm/ctrl.c
67
alink->id, band, queue, alink->queue_base);
drivers/net/ethernet/netronome/nfp/abm/ctrl.c
99
unsigned int queue, u32 val)
drivers/net/ethernet/netronome/nfp/abm/main.h
256
unsigned int queue, u32 val);
drivers/net/ethernet/netronome/nfp/abm/main.h
260
unsigned int queue, enum nfp_abm_q_action act);
drivers/net/ethernet/netronome/nfp/abm/main.h
262
unsigned int band, unsigned int queue,
drivers/net/ethernet/netronome/nfp/abm/main.h
265
unsigned int band, unsigned int queue,
drivers/net/ethernet/netronome/nfp/abm/qdisc.c
139
unsigned int queue, struct nfp_alink_stats *prev_stats,
drivers/net/ethernet/netronome/nfp/abm/qdisc.c
151
err = nfp_abm_ctrl_read_q_stats(alink, band, queue, prev_stats);
drivers/net/ethernet/netronome/nfp/abm/qdisc.c
155
band, queue, err);
drivers/net/ethernet/netronome/nfp/abm/qdisc.c
159
err = nfp_abm_ctrl_read_q_xstats(alink, band, queue, prev_xstats);
drivers/net/ethernet/netronome/nfp/abm/qdisc.c
163
band, queue, err);
drivers/net/ethernet/netronome/nfp/abm/qdisc.c
174
unsigned int queue)
drivers/net/ethernet/netronome/nfp/abm/qdisc.c
180
err = __nfp_abm_stats_init(alink, i, queue,
drivers/net/ethernet/netronome/nfp/abm/qdisc.c
192
unsigned int queue)
drivers/net/ethernet/netronome/nfp/abm/qdisc.c
209
if (nfp_abm_stats_init(alink, qdisc, queue))
drivers/net/ethernet/netronome/nfp/abm/qdisc.c
218
nfp_abm_ctrl_set_q_lvl(alink, i, queue,
drivers/net/ethernet/netronome/nfp/abm/qdisc.c
222
nfp_abm_ctrl_set_q_act(alink, i, queue, act);
drivers/net/ethernet/netronome/nfp/abm/qdisc.c
46
unsigned int queue)
drivers/net/ethernet/netronome/nfp/abm/qdisc.c
56
err = nfp_abm_ctrl_read_q_stats(alink, i, queue,
drivers/net/ethernet/netronome/nfp/abm/qdisc.c
60
i, queue, err);
drivers/net/ethernet/netronome/nfp/abm/qdisc.c
62
err = nfp_abm_ctrl_read_q_xstats(alink, i, queue,
drivers/net/ethernet/netronome/nfp/abm/qdisc.c
66
i, queue, err);
drivers/net/ethernet/netronome/nfp/abm/qdisc.c
830
opt->graft_params.queue);
drivers/net/ethernet/netronome/nfp/ccm_mbox.c
118
return skb_queue_is_first(&nn->mbox_cmsg.queue, skb);
drivers/net/ethernet/netronome/nfp/ccm_mbox.c
133
skb = skb_peek(&nn->mbox_cmsg.queue);
drivers/net/ethernet/netronome/nfp/ccm_mbox.c
159
skb = __skb_peek(&nn->mbox_cmsg.queue);
drivers/net/ethernet/netronome/nfp/ccm_mbox.c
192
skb = skb_queue_next(&nn->mbox_cmsg.queue, skb);
drivers/net/ethernet/netronome/nfp/ccm_mbox.c
203
skb = __skb_peek(&nn->mbox_cmsg.queue);
drivers/net/ethernet/netronome/nfp/ccm_mbox.c
210
skb = skb_queue_next(&nn->mbox_cmsg.queue, skb);
drivers/net/ethernet/netronome/nfp/ccm_mbox.c
330
spin_lock_bh(&nn->mbox_cmsg.queue.lock);
drivers/net/ethernet/netronome/nfp/ccm_mbox.c
332
skb = __skb_dequeue(&nn->mbox_cmsg.queue);
drivers/net/ethernet/netronome/nfp/ccm_mbox.c
351
spin_unlock_bh(&nn->mbox_cmsg.queue.lock);
drivers/net/ethernet/netronome/nfp/ccm_mbox.c
360
spin_lock_bh(&nn->mbox_cmsg.queue.lock);
drivers/net/ethernet/netronome/nfp/ccm_mbox.c
362
skb = __skb_dequeue(&nn->mbox_cmsg.queue);
drivers/net/ethernet/netronome/nfp/ccm_mbox.c
371
spin_unlock_bh(&nn->mbox_cmsg.queue.lock);
drivers/net/ethernet/netronome/nfp/ccm_mbox.c
375
__releases(&nn->mbox_cmsg.queue.lock)
drivers/net/ethernet/netronome/nfp/ccm_mbox.c
385
last = skb = __skb_peek(&nn->mbox_cmsg.queue);
drivers/net/ethernet/netronome/nfp/ccm_mbox.c
388
while (!skb_queue_is_last(&nn->mbox_cmsg.queue, last)) {
drivers/net/ethernet/netronome/nfp/ccm_mbox.c
389
skb = skb_queue_next(&nn->mbox_cmsg.queue, last);
drivers/net/ethernet/netronome/nfp/ccm_mbox.c
399
spin_unlock_bh(&nn->mbox_cmsg.queue.lock);
drivers/net/ethernet/netronome/nfp/ccm_mbox.c
436
__releases(&nn->mbox_cmsg.queue.lock)
drivers/net/ethernet/netronome/nfp/ccm_mbox.c
441
spin_unlock_bh(&nn->mbox_cmsg.queue.lock);
drivers/net/ethernet/netronome/nfp/ccm_mbox.c
449
__skb_unlink(skb, &nn->mbox_cmsg.queue);
drivers/net/ethernet/netronome/nfp/ccm_mbox.c
453
spin_unlock_bh(&nn->mbox_cmsg.queue.lock);
drivers/net/ethernet/netronome/nfp/ccm_mbox.c
522
assert_spin_locked(&nn->mbox_cmsg.queue.lock);
drivers/net/ethernet/netronome/nfp/ccm_mbox.c
524
if (!critical && nn->mbox_cmsg.queue.qlen >= NFP_CCM_MAX_QLEN) {
drivers/net/ethernet/netronome/nfp/ccm_mbox.c
534
__skb_queue_tail(&nn->mbox_cmsg.queue, skb);
drivers/net/ethernet/netronome/nfp/ccm_mbox.c
551
spin_lock_bh(&nn->mbox_cmsg.queue.lock);
drivers/net/ethernet/netronome/nfp/ccm_mbox.c
561
spin_unlock_bh(&nn->mbox_cmsg.queue.lock);
drivers/net/ethernet/netronome/nfp/ccm_mbox.c
574
spin_lock_bh(&nn->mbox_cmsg.queue.lock);
drivers/net/ethernet/netronome/nfp/ccm_mbox.c
591
spin_unlock_bh(&nn->mbox_cmsg.queue.lock);
drivers/net/ethernet/netronome/nfp/ccm_mbox.c
613
spin_lock_bh(&nn->mbox_cmsg.queue.lock);
drivers/net/ethernet/netronome/nfp/ccm_mbox.c
615
skb = __skb_peek(&nn->mbox_cmsg.queue);
drivers/net/ethernet/netronome/nfp/ccm_mbox.c
618
spin_unlock_bh(&nn->mbox_cmsg.queue.lock);
drivers/net/ethernet/netronome/nfp/ccm_mbox.c
633
skb = skb_peek(&nn->mbox_cmsg.queue);
drivers/net/ethernet/netronome/nfp/ccm_mbox.c
660
spin_lock_bh(&nn->mbox_cmsg.queue.lock);
drivers/net/ethernet/netronome/nfp/ccm_mbox.c
678
spin_unlock_bh(&nn->mbox_cmsg.queue.lock);
drivers/net/ethernet/netronome/nfp/ccm_mbox.c
683
spin_unlock_bh(&nn->mbox_cmsg.queue.lock);
drivers/net/ethernet/netronome/nfp/ccm_mbox.c
728
skb_queue_head_init(&nn->mbox_cmsg.queue);
drivers/net/ethernet/netronome/nfp/ccm_mbox.c
742
WARN_ON(!skb_queue_empty(&nn->mbox_cmsg.queue));
drivers/net/ethernet/netronome/nfp/flower/main.c
389
const u8 queue = 0;
drivers/net/ethernet/netronome/nfp/flower/main.c
443
i, queue);
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
1223
__skb_queue_tail(&r_vec->queue, skb);
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
1225
__skb_queue_head(&r_vec->queue, skb);
drivers/net/ethernet/netronome/nfp/nfd3/dp.c
1286
while ((skb = __skb_dequeue(&r_vec->queue)))
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
1336
__skb_queue_tail(&r_vec->queue, skb);
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
1338
__skb_queue_head(&r_vec->queue, skb);
drivers/net/ethernet/netronome/nfp/nfdk/dp.c
1448
while ((skb = __skb_dequeue(&r_vec->queue)))
drivers/net/ethernet/netronome/nfp/nfp_net.h
393
struct sk_buff_head queue;
drivers/net/ethernet/netronome/nfp/nfp_net.h
709
struct sk_buff_head queue;
drivers/net/ethernet/netronome/nfp/nfp_net.h
912
u32 nfp_qcp_queue_offset(const struct nfp_dev_info *dev_info, u16 queue);
drivers/net/ethernet/netronome/nfp/nfp_net_common.c
72
u32 nfp_qcp_queue_offset(const struct nfp_dev_info *dev_info, u16 queue)
drivers/net/ethernet/netronome/nfp/nfp_net_common.c
74
queue &= dev_info->qc_idx_mask;
drivers/net/ethernet/netronome/nfp/nfp_net_common.c
75
return dev_info->qc_addr_offset + NFP_QCP_QUEUE_ADDR_SZ * queue;
drivers/net/ethernet/netronome/nfp/nfp_net_common.c
762
__skb_queue_head_init(&r_vec->queue);
drivers/net/ethernet/qlogic/qed/qed_sp.h
137
struct list_head *queue;
drivers/net/ethernet/qlogic/qed/qed_sp_commands.c
31
if (p_ent->queue == &p_hwfn->p_spq->unlimited_pending)
drivers/net/ethernet/qlogic/qed/qed_spq.c
643
p_ent->queue = &p_spq->unlimited_pending;
drivers/net/ethernet/qlogic/qed/qed_spq.c
648
p_ent->queue = &p_spq->pending;
drivers/net/ethernet/qlogic/qed/qed_spq.c
692
if (p_ent->queue == &p_spq->unlimited_pending) {
drivers/net/ethernet/qlogic/qed/qed_spq.c
899
p_ent->queue == &p_spq->unlimited_pending);
drivers/net/ethernet/qlogic/qed/qed_spq.c
901
if (p_ent->queue == &p_spq->unlimited_pending) {
drivers/net/ethernet/qlogic/qede/qede.h
585
int qede_set_per_coalesce(struct net_device *dev, u32 queue,
drivers/net/ethernet/qlogic/qede/qede_ethtool.c
2145
int qede_set_per_coalesce(struct net_device *dev, u32 queue,
drivers/net/ethernet/qlogic/qede/qede_ethtool.c
2167
if (queue >= edev->num_queues) {
drivers/net/ethernet/qlogic/qede/qede_ethtool.c
2178
fp = &edev->fp_array[queue];
drivers/net/ethernet/qlogic/qede/qede_ethtool.c
2180
if (edev->fp_array[queue].type & QEDE_FASTPATH_RX) {
drivers/net/ethernet/qlogic/qede/qede_ethtool.c
2189
edev->coal_entry[queue].rxc = rxc;
drivers/net/ethernet/qlogic/qede/qede_ethtool.c
2190
edev->coal_entry[queue].isvalid = true;
drivers/net/ethernet/qlogic/qede/qede_ethtool.c
2193
if (edev->fp_array[queue].type & QEDE_FASTPATH_TX) {
drivers/net/ethernet/qlogic/qede/qede_ethtool.c
2202
edev->coal_entry[queue].txc = txc;
drivers/net/ethernet/qlogic/qede/qede_ethtool.c
2203
edev->coal_entry[queue].isvalid = true;
drivers/net/ethernet/qlogic/qede/qede_ethtool.c
2212
u32 queue,
drivers/net/ethernet/qlogic/qede/qede_ethtool.c
2227
if (queue >= edev->num_queues) {
drivers/net/ethernet/qlogic/qede/qede_ethtool.c
2238
fp = &edev->fp_array[queue];
drivers/net/ethernet/qlogic/qede/qede_ethtool.c
2250
fp = &edev->fp_array[queue];
drivers/net/ethernet/qlogic/qede/qede_filter.c
1683
if (act->queue.vf)
drivers/net/ethernet/qlogic/qede/qede_filter.c
1686
if (act->queue.index >= QEDE_RSS_COUNT(edev)) {
drivers/net/ethernet/realtek/rtase/rtase_main.c
1679
static void rtase_set_hw_cbs(const struct rtase_private *tp, u32 queue)
drivers/net/ethernet/realtek/rtase/rtase_main.c
1681
u32 idle = tp->tx_qos[queue].idleslope * RTASE_1T_CLOCK;
drivers/net/ethernet/realtek/rtase/rtase_main.c
1695
rtase_w32(tp, RTASE_TXQCRDT_0 + queue * 4, val);
drivers/net/ethernet/realtek/rtase/rtase_main.c
1701
int queue = qopt->queue;
drivers/net/ethernet/realtek/rtase/rtase_main.c
1703
if (queue < 0 || queue >= tp->func_tx_queue_num)
drivers/net/ethernet/realtek/rtase/rtase_main.c
1707
tp->tx_qos[queue].hicredit = 0;
drivers/net/ethernet/realtek/rtase/rtase_main.c
1708
tp->tx_qos[queue].locredit = 0;
drivers/net/ethernet/realtek/rtase/rtase_main.c
1709
tp->tx_qos[queue].idleslope = 0;
drivers/net/ethernet/realtek/rtase/rtase_main.c
1710
tp->tx_qos[queue].sendslope = 0;
drivers/net/ethernet/realtek/rtase/rtase_main.c
1712
rtase_w32(tp, RTASE_TXQCRDT_0 + queue * 4, 0);
drivers/net/ethernet/realtek/rtase/rtase_main.c
1714
tp->tx_qos[queue].hicredit = qopt->hicredit;
drivers/net/ethernet/realtek/rtase/rtase_main.c
1715
tp->tx_qos[queue].locredit = qopt->locredit;
drivers/net/ethernet/realtek/rtase/rtase_main.c
1716
tp->tx_qos[queue].idleslope = qopt->idleslope;
drivers/net/ethernet/realtek/rtase/rtase_main.c
1717
tp->tx_qos[queue].sendslope = qopt->sendslope;
drivers/net/ethernet/realtek/rtase/rtase_main.c
1719
rtase_set_hw_cbs(tp, queue);
drivers/net/ethernet/renesas/rswitch.h
788
#define GWTRC(queue) (GWTRC0 + (queue) / 32 * 4)
drivers/net/ethernet/renesas/rswitch.h
790
#define GWDCC_OFFS(queue) (GWDCC0 + (queue) * 4)
drivers/net/ethernet/renesas/rswitch.h
855
#define TPEMIMC7(queue) (TPEMIMC70 + (queue) * 4)
drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c
192
static inline u32 sxgbe_tx_avail(struct sxgbe_tx_queue *queue, int tx_qsize)
drivers/net/ethernet/samsung/sxgbe/sxgbe_main.c
194
return queue->dirty_tx + tx_qsize - queue->cur_tx - 1;
drivers/net/ethernet/sfc/ef10.c
2229
ER_DZ_TX_DESC_UPD, tx_queue->queue);
drivers/net/ethernet/sfc/ef10.c
2415
tx_queue->queue);
drivers/net/ethernet/sfc/ef10.c
2427
ER_DZ_TX_DESC_UPD_DWORD, tx_queue->queue);
drivers/net/ethernet/sfc/ef10.c
832
if (tx_queue->queue == nic_data->pio_write_vi_base) {
drivers/net/ethernet/sfc/ef10.c
841
tx_queue->queue);
drivers/net/ethernet/sfc/ef10.c
853
tx_queue->queue, index, rc);
drivers/net/ethernet/sfc/ef10.c
862
tx_queue->queue, index,
drivers/net/ethernet/sfc/ef100_tx.c
124
ER_GZ_TX_RING_DOORBELL, tx_queue->queue);
drivers/net/ethernet/sfc/ef100_tx.c
49
"failed to initialise TXQ %d\n", tx_queue->queue);
drivers/net/ethernet/sfc/efx_channels.c
548
tx_queue->queue = -1;
drivers/net/ethernet/sfc/efx_channels.c
774
xdp_queue_number, tx_queue->queue);
drivers/net/ethernet/sfc/efx_channels.c
797
tx_queue->queue = next_queue++;
drivers/net/ethernet/sfc/efx_channels.c
805
tx_queue->queue = next_queue++;
drivers/net/ethernet/sfc/efx_channels.c
809
tx_queue->queue);
drivers/net/ethernet/sfc/ethtool_common.c
410
channel->tx_queue[0].queue /
drivers/net/ethernet/sfc/falcon/efx.c
1602
tx_queue->queue -= (efx->tx_channel_offset *
drivers/net/ethernet/sfc/falcon/efx.c
437
tx_queue->queue = i * EF4_TXQ_TYPES + j;
drivers/net/ethernet/sfc/falcon/ethtool.c
234
#define EF4_TX_QUEUE_NAME(_tx_queue) "txq%d", _tx_queue->queue
drivers/net/ethernet/sfc/falcon/ethtool.c
235
#define EF4_RX_QUEUE_NAME(_rx_queue) "rxq%d", _rx_queue->queue
drivers/net/ethernet/sfc/falcon/ethtool.c
263
&lb_tests->tx_sent[tx_queue->queue],
drivers/net/ethernet/sfc/falcon/ethtool.c
267
&lb_tests->tx_done[tx_queue->queue],
drivers/net/ethernet/sfc/falcon/ethtool.c
368
channel->tx_queue[0].queue /
drivers/net/ethernet/sfc/falcon/farch.c
280
FR_AZ_TX_DESC_UPD_DWORD_P0, tx_queue->queue);
drivers/net/ethernet/sfc/falcon/farch.c
298
FR_BZ_TX_DESC_UPD_P0, tx_queue->queue);
drivers/net/ethernet/sfc/falcon/farch.c
390
FRF_AZ_TX_DESCQ_LABEL, tx_queue->queue,
drivers/net/ethernet/sfc/falcon/farch.c
397
int csum = tx_queue->queue & EF4_TXQ_TYPE_OFFLOAD;
drivers/net/ethernet/sfc/falcon/farch.c
404
tx_queue->queue);
drivers/net/ethernet/sfc/falcon/farch.c
411
if (tx_queue->queue & EF4_TXQ_TYPE_OFFLOAD)
drivers/net/ethernet/sfc/falcon/farch.c
412
__clear_bit_le(tx_queue->queue, ®);
drivers/net/ethernet/sfc/falcon/farch.c
414
__set_bit_le(tx_queue->queue, ®);
drivers/net/ethernet/sfc/falcon/farch.c
421
(tx_queue->queue & EF4_TXQ_TYPE_HIGHPRI) ?
drivers/net/ethernet/sfc/falcon/farch.c
425
tx_queue->queue);
drivers/net/ethernet/sfc/falcon/farch.c
439
FRF_AZ_TX_FLUSH_DESCQ, tx_queue->queue);
drivers/net/ethernet/sfc/falcon/farch.c
451
tx_queue->queue);
drivers/net/ethernet/sfc/falcon/farch.c
624
FR_BZ_TX_DESC_PTR_TBL, tx_queue->queue);
drivers/net/ethernet/sfc/falcon/farch.c
631
tx_queue->queue);
drivers/net/ethernet/sfc/falcon/farch.c
640
"the queue\n", tx_queue->queue);
drivers/net/ethernet/sfc/falcon/farch.c
74
(_tx_queue)->queue)
drivers/net/ethernet/sfc/falcon/net_driver.h
1212
tx_queue->queue & EF4_TXQ_TYPE_HIGHPRI);
drivers/net/ethernet/sfc/falcon/net_driver.h
215
unsigned queue;
drivers/net/ethernet/sfc/falcon/nic.h
71
if (tx_queue->queue & EF4_TXQ_TYPE_OFFLOAD)
drivers/net/ethernet/sfc/falcon/selftest.c
458
"%d in %s loopback test\n", tx_queue->queue,
drivers/net/ethernet/sfc/falcon/selftest.c
510
tx_queue->queue, tx_done, state->packet_count,
drivers/net/ethernet/sfc/falcon/selftest.c
521
tx_queue->queue, rx_good, state->packet_count,
drivers/net/ethernet/sfc/falcon/selftest.c
528
lb_tests->tx_sent[tx_queue->queue] += state->packet_count;
drivers/net/ethernet/sfc/falcon/selftest.c
529
lb_tests->tx_done[tx_queue->queue] += tx_done;
drivers/net/ethernet/sfc/falcon/selftest.c
555
tx_queue->queue, LOOPBACK_MODE(efx),
drivers/net/ethernet/sfc/falcon/selftest.c
582
"of %d packets\n", tx_queue->queue, LOOPBACK_MODE(efx),
drivers/net/ethernet/sfc/falcon/selftest.c
672
state->offload_csum = (tx_queue->queue &
drivers/net/ethernet/sfc/falcon/tx.c
365
tx_queue->queue, read_ptr);
drivers/net/ethernet/sfc/falcon/tx.c
413
tx_queue->queue / EF4_TXQ_TYPES +
drivers/net/ethernet/sfc/falcon/tx.c
414
((tx_queue->queue & EF4_TXQ_TYPE_HIGHPRI) ?
drivers/net/ethernet/sfc/falcon/tx.c
451
if (!(tx_queue->queue & EF4_TXQ_TYPE_HIGHPRI))
drivers/net/ethernet/sfc/falcon/tx.c
544
tx_queue->queue, efx->txq_entries, tx_queue->ptr_mask);
drivers/net/ethernet/sfc/falcon/tx.c
579
"initialising TX queue %d\n", tx_queue->queue);
drivers/net/ethernet/sfc/falcon/tx.c
603
"shutting down TX queue %d\n", tx_queue->queue);
drivers/net/ethernet/sfc/falcon/tx.c
628
"destroying TX queue %d\n", tx_queue->queue);
drivers/net/ethernet/sfc/falcon/tx.c
66
tx_queue->queue, tx_queue->read_count);
drivers/net/ethernet/sfc/mcdi_functions.c
181
MCDI_SET_DWORD(inbuf, INIT_TXQ_IN_INSTANCE, tx_queue->queue);
drivers/net/ethernet/sfc/mcdi_functions.c
188
tx_queue->queue, entries, (u64)dma_addr);
drivers/net/ethernet/sfc/mcdi_functions.c
255
tx_queue->queue);
drivers/net/ethernet/sfc/net_driver.h
263
unsigned int queue;
drivers/net/ethernet/sfc/ptp.c
1245
struct efx_rx_queue *queue = efx_channel_get_rx_queue(channel);
drivers/net/ethernet/sfc/ptp.c
1248
efx_rx_queue_index(queue));
drivers/net/ethernet/sfc/selftest.c
552
tx_queue->label, tx_queue->queue, LOOPBACK_MODE(efx),
drivers/net/ethernet/sfc/siena/efx_channels.c
550
tx_queue->queue = -1;
drivers/net/ethernet/sfc/siena/efx_channels.c
780
xdp_queue_number, tx_queue->queue);
drivers/net/ethernet/sfc/siena/efx_channels.c
803
tx_queue->queue = next_queue++;
drivers/net/ethernet/sfc/siena/efx_channels.c
811
tx_queue->queue = next_queue++;
drivers/net/ethernet/sfc/siena/efx_channels.c
815
tx_queue->queue);
drivers/net/ethernet/sfc/siena/ethtool_common.c
410
channel->tx_queue[0].queue /
drivers/net/ethernet/sfc/siena/farch.c
291
FR_AZ_TX_DESC_UPD_DWORD_P0, tx_queue->queue);
drivers/net/ethernet/sfc/siena/farch.c
309
FR_BZ_TX_DESC_UPD_P0, tx_queue->queue);
drivers/net/ethernet/sfc/siena/farch.c
411
tx_queue->queue);
drivers/net/ethernet/sfc/siena/farch.c
418
efx_writeo_table(efx, ®, FR_BZ_TX_PACE_TBL, tx_queue->queue);
drivers/net/ethernet/sfc/siena/farch.c
433
FRF_AZ_TX_FLUSH_DESCQ, tx_queue->queue);
drivers/net/ethernet/sfc/siena/farch.c
445
tx_queue->queue);
drivers/net/ethernet/sfc/siena/farch.c
612
FR_BZ_TX_DESC_PTR_TBL, tx_queue->queue);
drivers/net/ethernet/sfc/siena/farch.c
619
tx_queue->queue);
drivers/net/ethernet/sfc/siena/farch.c
628
"the queue\n", tx_queue->queue);
drivers/net/ethernet/sfc/siena/farch.c
78
(_tx_queue)->queue)
drivers/net/ethernet/sfc/siena/net_driver.h
260
unsigned int queue;
drivers/net/ethernet/sfc/siena/selftest.c
553
tx_queue->label, tx_queue->queue, LOOPBACK_MODE(efx),
drivers/net/ethernet/sfc/siena/siena_sriov.c
1494
unsigned queue, qid;
drivers/net/ethernet/sfc/siena/siena_sriov.c
1496
queue = EFX_QWORD_FIELD(*event, FSF_AZ_DRIVER_EV_SUBDATA);
drivers/net/ethernet/sfc/siena/siena_sriov.c
1497
if (map_vi_index(efx, queue, &vf, &qid))
drivers/net/ethernet/sfc/siena/siena_sriov.c
1513
unsigned ev_failed, queue, qid;
drivers/net/ethernet/sfc/siena/siena_sriov.c
1515
queue = EFX_QWORD_FIELD(*event, FSF_AZ_DRIVER_EV_RX_DESCQ_ID);
drivers/net/ethernet/sfc/siena/siena_sriov.c
1518
if (map_vi_index(efx, queue, &vf, &qid))
drivers/net/ethernet/sfc/siena/tx_common.c
105
"destroying TX queue %d\n", tx_queue->queue);
drivers/net/ethernet/sfc/siena/tx_common.c
160
tx_queue->queue, tx_queue->read_count);
drivers/net/ethernet/sfc/siena/tx_common.c
174
"shutting down TX queue %d\n", tx_queue->queue);
drivers/net/ethernet/sfc/siena/tx_common.c
214
tx_queue->queue, read_ptr);
drivers/net/ethernet/sfc/siena/tx_common.c
36
tx_queue->queue, efx->txq_entries, tx_queue->ptr_mask);
drivers/net/ethernet/sfc/siena/tx_common.c
72
"initialising TX queue %d\n", tx_queue->queue);
drivers/net/ethernet/sfc/tx.c
564
tx_queue->queue);
drivers/net/ethernet/sfc/tx_common.c
107
"shutting down TX queue %d\n", tx_queue->queue);
drivers/net/ethernet/sfc/tx_common.c
139
"destroying TX queue %d\n", tx_queue->queue);
drivers/net/ethernet/sfc/tx_common.c
203
tx_queue->queue, tx_queue->read_count);
drivers/net/ethernet/sfc/tx_common.c
241
tx_queue->queue, read_ptr);
drivers/net/ethernet/sfc/tx_common.c
36
tx_queue->queue, efx->txq_entries, tx_queue->ptr_mask);
drivers/net/ethernet/sfc/tx_common.c
72
"initialising TX queue %d\n", tx_queue->queue);
drivers/net/ethernet/stmicro/stmmac/dwmac1000_dma.c
248
void __iomem *ioaddr, u32 riwt, u32 queue)
drivers/net/ethernet/stmicro/stmmac/dwmac1000_dma.c
250
writel(riwt, ioaddr + DMA_CHAN_RX_WATCHDOG(queue));
drivers/net/ethernet/stmicro/stmmac/dwmac4.h
81
#define GMAC_RX_QUEUE_CLEAR(queue) ~(GENMASK(1, 0) << ((queue) * 2))
drivers/net/ethernet/stmicro/stmmac/dwmac4.h
82
#define GMAC_RX_AV_QUEUE_ENABLE(queue) BIT((queue) * 2)
drivers/net/ethernet/stmicro/stmmac/dwmac4.h
83
#define GMAC_RX_DCB_QUEUE_ENABLE(queue) BIT(((queue) * 2) + 1)
drivers/net/ethernet/stmicro/stmmac/dwmac4_core.c
117
if (queue < 4) {
drivers/net/ethernet/stmicro/stmmac/dwmac4_core.c
118
ctrl2 |= (prio << GMAC_RXQCTRL_PSRQX_SHIFT(queue)) &
drivers/net/ethernet/stmicro/stmmac/dwmac4_core.c
119
GMAC_RXQCTRL_PSRQX_MASK(queue);
drivers/net/ethernet/stmicro/stmmac/dwmac4_core.c
124
queue -= 4;
drivers/net/ethernet/stmicro/stmmac/dwmac4_core.c
126
ctrl3 |= (prio << GMAC_RXQCTRL_PSRQX_SHIFT(queue)) &
drivers/net/ethernet/stmicro/stmmac/dwmac4_core.c
127
GMAC_RXQCTRL_PSRQX_MASK(queue);
drivers/net/ethernet/stmicro/stmmac/dwmac4_core.c
135
u32 prio, u32 queue)
drivers/net/ethernet/stmicro/stmmac/dwmac4_core.c
141
base_register = (queue < 4) ? GMAC_TXQ_PRTY_MAP0 : GMAC_TXQ_PRTY_MAP1;
drivers/net/ethernet/stmicro/stmmac/dwmac4_core.c
142
if (queue >= 4)
drivers/net/ethernet/stmicro/stmmac/dwmac4_core.c
143
queue -= 4;
drivers/net/ethernet/stmicro/stmmac/dwmac4_core.c
147
value &= ~GMAC_TXQCTRL_PSTQX_MASK(queue);
drivers/net/ethernet/stmicro/stmmac/dwmac4_core.c
148
value |= (prio << GMAC_TXQCTRL_PSTQX_SHIFT(queue)) &
drivers/net/ethernet/stmicro/stmmac/dwmac4_core.c
149
GMAC_TXQCTRL_PSTQX_MASK(queue);
drivers/net/ethernet/stmicro/stmmac/dwmac4_core.c
155
u8 packet, u32 queue)
drivers/net/ethernet/stmicro/stmmac/dwmac4_core.c
172
value |= (queue << route_possibilities[packet-1].reg_shift) &
drivers/net/ethernet/stmicro/stmmac/dwmac4_core.c
237
u32 weight, u32 queue)
drivers/net/ethernet/stmicro/stmmac/dwmac4_core.c
242
queue));
drivers/net/ethernet/stmicro/stmmac/dwmac4_core.c
246
writel(value, ioaddr + mtl_txqx_weight_base_addr(dwmac4_addrs, queue));
drivers/net/ethernet/stmicro/stmmac/dwmac4_core.c
249
static void dwmac4_map_mtl_dma(struct mac_device_info *hw, u32 queue, u32 chan)
drivers/net/ethernet/stmicro/stmmac/dwmac4_core.c
254
if (queue < 4) {
drivers/net/ethernet/stmicro/stmmac/dwmac4_core.c
256
value &= ~MTL_RXQ_DMA_QXMDMACH_MASK(queue);
drivers/net/ethernet/stmicro/stmmac/dwmac4_core.c
257
value |= MTL_RXQ_DMA_QXMDMACH(chan, queue);
drivers/net/ethernet/stmicro/stmmac/dwmac4_core.c
261
value &= ~MTL_RXQ_DMA_QXMDMACH_MASK(queue - 4);
drivers/net/ethernet/stmicro/stmmac/dwmac4_core.c
262
value |= MTL_RXQ_DMA_QXMDMACH(chan, queue - 4);
drivers/net/ethernet/stmicro/stmmac/dwmac4_core.c
270
u32 high_credit, u32 low_credit, u32 queue)
drivers/net/ethernet/stmicro/stmmac/dwmac4_core.c
276
pr_debug("Queue %d configured as AVB. Parameters:\n", queue);
drivers/net/ethernet/stmicro/stmmac/dwmac4_core.c
283
value = readl(ioaddr + mtl_etsx_ctrl_base_addr(dwmac4_addrs, queue));
drivers/net/ethernet/stmicro/stmmac/dwmac4_core.c
286
writel(value, ioaddr + mtl_etsx_ctrl_base_addr(dwmac4_addrs, queue));
drivers/net/ethernet/stmicro/stmmac/dwmac4_core.c
290
queue));
drivers/net/ethernet/stmicro/stmmac/dwmac4_core.c
294
queue));
drivers/net/ethernet/stmicro/stmmac/dwmac4_core.c
297
dwmac4_set_mtl_tx_queue_weight(priv, hw, idle_slope, queue);
drivers/net/ethernet/stmicro/stmmac/dwmac4_core.c
300
value = readl(ioaddr + mtl_high_credx_base_addr(dwmac4_addrs, queue));
drivers/net/ethernet/stmicro/stmmac/dwmac4_core.c
303
writel(value, ioaddr + mtl_high_credx_base_addr(dwmac4_addrs, queue));
drivers/net/ethernet/stmicro/stmmac/dwmac4_core.c
306
value = readl(ioaddr + mtl_low_credx_base_addr(dwmac4_addrs, queue));
drivers/net/ethernet/stmicro/stmmac/dwmac4_core.c
309
writel(value, ioaddr + mtl_low_credx_base_addr(dwmac4_addrs, queue));
drivers/net/ethernet/stmicro/stmmac/dwmac4_core.c
554
u32 queue = 0;
drivers/net/ethernet/stmicro/stmmac/dwmac4_core.c
571
for (queue = 0; queue < tx_cnt; queue++) {
drivers/net/ethernet/stmicro/stmmac/dwmac4_core.c
578
writel(flow, ioaddr + GMAC_QX_TX_FLOW_CTRL(queue));
drivers/net/ethernet/stmicro/stmmac/dwmac4_core.c
581
for (queue = 0; queue < tx_cnt; queue++)
drivers/net/ethernet/stmicro/stmmac/dwmac4_core.c
582
writel(0, ioaddr + GMAC_QX_TX_FLOW_CTRL(queue));
drivers/net/ethernet/stmicro/stmmac/dwmac4_core.c
673
u32 queue;
drivers/net/ethernet/stmicro/stmmac/dwmac4_core.c
675
for (queue = 0; queue < tx_queues; queue++) {
drivers/net/ethernet/stmicro/stmmac/dwmac4_core.c
676
value = readl(ioaddr + MTL_CHAN_TX_DEBUG(dwmac4_addrs, queue));
drivers/net/ethernet/stmicro/stmmac/dwmac4_core.c
700
for (queue = 0; queue < rx_queues; queue++) {
drivers/net/ethernet/stmicro/stmmac/dwmac4_core.c
701
value = readl(ioaddr + MTL_CHAN_RX_DEBUG(dwmac4_addrs, queue));
drivers/net/ethernet/stmicro/stmmac/dwmac4_core.c
79
u8 mode, u32 queue)
drivers/net/ethernet/stmicro/stmmac/dwmac4_core.c
84
value &= GMAC_RX_QUEUE_CLEAR(queue);
drivers/net/ethernet/stmicro/stmmac/dwmac4_core.c
86
value |= GMAC_RX_AV_QUEUE_ENABLE(queue);
drivers/net/ethernet/stmicro/stmmac/dwmac4_core.c
88
value |= GMAC_RX_DCB_QUEUE_ENABLE(queue);
drivers/net/ethernet/stmicro/stmmac/dwmac4_core.c
94
u32 prio, u32 queue)
drivers/net/ethernet/stmicro/stmmac/dwmac4_dma.c
226
u32 riwt, u32 queue)
drivers/net/ethernet/stmicro/stmmac/dwmac4_dma.c
230
writel(riwt, ioaddr + DMA_CHAN_RX_WATCHDOG(dwmac4_addrs, queue));
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c
121
if (queue < 4) {
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c
122
ctrl2 |= (prio << XGMAC_PSRQ_SHIFT(queue)) &
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c
123
XGMAC_PSRQ(queue);
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c
128
queue -= 4;
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c
130
ctrl3 |= (prio << XGMAC_PSRQ_SHIFT(queue)) &
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c
131
XGMAC_PSRQ(queue);
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c
139
u32 queue)
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c
144
reg = (queue < 4) ? XGMAC_TC_PRTY_MAP0 : XGMAC_TC_PRTY_MAP1;
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c
145
if (queue >= 4)
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c
146
queue -= 4;
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c
1470
u32 queue)
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c
1475
value = readl(ioaddr + XLGMAC_RXQ_ENABLE_CTRL0) & ~XGMAC_RXQEN(queue);
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c
1477
value |= 0x1 << XGMAC_RXQEN_SHIFT(queue);
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c
1479
value |= 0x2 << XGMAC_RXQEN_SHIFT(queue);
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c
149
value &= ~XGMAC_PSTC(queue);
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c
150
value |= (prio << XGMAC_PSTC_SHIFT(queue)) & XGMAC_PSTC(queue);
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c
156
u8 packet, u32 queue)
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c
173
value |= (queue << dwxgmac2_route_possibilities[packet - 1].reg_shift) &
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c
247
u32 weight, u32 queue)
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c
251
writel(weight, ioaddr + XGMAC_MTL_TCx_QUANTUM_WEIGHT(queue));
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c
254
static void dwxgmac2_map_mtl_to_dma(struct mac_device_info *hw, u32 queue,
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c
260
reg = (queue < 4) ? XGMAC_MTL_RXQ_DMA_MAP0 : XGMAC_MTL_RXQ_DMA_MAP1;
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c
261
if (queue >= 4)
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c
262
queue -= 4;
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c
265
value &= ~XGMAC_QxMDMACH(queue);
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c
266
value |= (chan << XGMAC_QxMDMACH_SHIFT(queue)) & XGMAC_QxMDMACH(queue);
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c
274
u32 high_credit, u32 low_credit, u32 queue)
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c
279
writel(send_slope, ioaddr + XGMAC_MTL_TCx_SENDSLOPE(queue));
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c
280
writel(idle_slope, ioaddr + XGMAC_MTL_TCx_QUANTUM_WEIGHT(queue));
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c
281
writel(high_credit, ioaddr + XGMAC_MTL_TCx_HICREDIT(queue));
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c
282
writel(low_credit, ioaddr + XGMAC_MTL_TCx_LOCREDIT(queue));
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c
284
value = readl(ioaddr + XGMAC_MTL_TCx_ETS_CONTROL(queue));
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c
287
writel(value, ioaddr + XGMAC_MTL_TCx_ETS_CONTROL(queue));
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c
84
u32 queue)
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c
89
value = readl(ioaddr + XGMAC_RXQ_CTRL0) & ~XGMAC_RXQEN(queue);
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c
91
value |= 0x1 << XGMAC_RXQEN_SHIFT(queue);
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c
93
value |= 0x2 << XGMAC_RXQEN_SHIFT(queue);
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c
98
u32 queue)
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_dma.c
474
u32 riwt, u32 queue)
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_dma.c
476
writel(riwt & XGMAC_RWT, ioaddr + XGMAC_DMA_CH_Rx_WATCHDOG(queue));
drivers/net/ethernet/stmicro/stmmac/hwif.h
224
u32 riwt, u32 queue);
drivers/net/ethernet/stmicro/stmmac/hwif.h
332
void (*rx_queue_enable)(struct mac_device_info *hw, u8 mode, u32 queue);
drivers/net/ethernet/stmicro/stmmac/hwif.h
334
void (*rx_queue_prio)(struct mac_device_info *hw, u32 prio, u32 queue);
drivers/net/ethernet/stmicro/stmmac/hwif.h
336
void (*tx_queue_prio)(struct mac_device_info *hw, u32 prio, u32 queue);
drivers/net/ethernet/stmicro/stmmac/hwif.h
339
u32 queue);
drivers/net/ethernet/stmicro/stmmac/hwif.h
347
u32 weight, u32 queue);
drivers/net/ethernet/stmicro/stmmac/hwif.h
349
void (*map_mtl_to_dma)(struct mac_device_info *hw, u32 queue, u32 chan);
drivers/net/ethernet/stmicro/stmmac/hwif.h
353
u32 low_credit, u32 queue);
drivers/net/ethernet/stmicro/stmmac/stmmac.h
418
void stmmac_disable_rx_queue(struct stmmac_priv *priv, u32 queue);
drivers/net/ethernet/stmicro/stmmac/stmmac.h
419
void stmmac_enable_rx_queue(struct stmmac_priv *priv, u32 queue);
drivers/net/ethernet/stmicro/stmmac/stmmac.h
420
void stmmac_disable_tx_queue(struct stmmac_priv *priv, u32 queue);
drivers/net/ethernet/stmicro/stmmac/stmmac.h
421
void stmmac_enable_tx_queue(struct stmmac_priv *priv, u32 queue);
drivers/net/ethernet/stmicro/stmmac/stmmac.h
422
int stmmac_xsk_wakeup(struct net_device *dev, u32 queue, u32 flags);
drivers/net/ethernet/stmicro/stmmac/stmmac_ethtool.c
781
int queue)
drivers/net/ethernet/stmicro/stmmac/stmmac_ethtool.c
792
if (queue < 0)
drivers/net/ethernet/stmicro/stmmac/stmmac_ethtool.c
793
queue = 0;
drivers/net/ethernet/stmicro/stmmac/stmmac_ethtool.c
794
else if (queue >= max_cnt)
drivers/net/ethernet/stmicro/stmmac/stmmac_ethtool.c
797
if (queue < tx_cnt) {
drivers/net/ethernet/stmicro/stmmac/stmmac_ethtool.c
798
ec->tx_coalesce_usecs = priv->tx_coal_timer[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_ethtool.c
799
ec->tx_max_coalesced_frames = priv->tx_coal_frames[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_ethtool.c
805
if (priv->use_riwt && queue < rx_cnt) {
drivers/net/ethernet/stmicro/stmmac/stmmac_ethtool.c
806
ec->rx_max_coalesced_frames = priv->rx_coal_frames[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_ethtool.c
807
ec->rx_coalesce_usecs = stmmac_riwt2usec(priv->rx_riwt[queue],
drivers/net/ethernet/stmicro/stmmac/stmmac_ethtool.c
825
static int stmmac_get_per_queue_coalesce(struct net_device *dev, u32 queue,
drivers/net/ethernet/stmicro/stmmac/stmmac_ethtool.c
828
return __stmmac_get_coalesce(dev, ec, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_ethtool.c
833
int queue)
drivers/net/ethernet/stmicro/stmmac/stmmac_ethtool.c
846
if (queue < 0)
drivers/net/ethernet/stmicro/stmmac/stmmac_ethtool.c
848
else if (queue >= max_cnt)
drivers/net/ethernet/stmicro/stmmac/stmmac_ethtool.c
867
} else if (queue < rx_cnt) {
drivers/net/ethernet/stmicro/stmmac/stmmac_ethtool.c
868
priv->rx_riwt[queue] = rx_riwt;
drivers/net/ethernet/stmicro/stmmac/stmmac_ethtool.c
870
rx_riwt, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_ethtool.c
871
priv->rx_coal_frames[queue] =
drivers/net/ethernet/stmicro/stmmac/stmmac_ethtool.c
893
} else if (queue < tx_cnt) {
drivers/net/ethernet/stmicro/stmmac/stmmac_ethtool.c
894
priv->tx_coal_frames[queue] =
drivers/net/ethernet/stmicro/stmmac/stmmac_ethtool.c
896
priv->tx_coal_timer[queue] =
drivers/net/ethernet/stmicro/stmmac/stmmac_ethtool.c
911
static int stmmac_set_per_queue_coalesce(struct net_device *dev, u32 queue,
drivers/net/ethernet/stmicro/stmmac/stmmac_ethtool.c
914
return __stmmac_set_coalesce(dev, ec, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1418
u32 queue;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1421
for (queue = 0; queue < rx_cnt; queue++) {
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1422
struct stmmac_rx_queue *rx_q = &dma_conf->rx_queue[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1424
pr_info("\tRX Queue %u rings\n", queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1446
u32 queue;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1449
for (queue = 0; queue < tx_cnt; queue++) {
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1450
struct stmmac_tx_queue *tx_q = &dma_conf->tx_queue[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1452
pr_info("\tTX Queue %d rings\n", queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1516
u32 queue)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1518
struct stmmac_rx_queue *rx_q = &dma_conf->rx_queue[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
152
static void stmmac_reset_rx_queue(struct stmmac_priv *priv, u32 queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
153
static void stmmac_reset_tx_queue(struct stmmac_priv *priv, u32 queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1545
u32 queue)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1547
struct stmmac_tx_queue *tx_q = &dma_conf->tx_queue[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
155
static void stmmac_tx_timer_arm(struct stmmac_priv *priv, u32 queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
156
static void stmmac_flush_tx_descriptors(struct stmmac_priv *priv, int queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1578
u32 queue;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1581
for (queue = 0; queue < rx_queue_cnt; queue++)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1582
stmmac_clear_rx_descriptors(priv, dma_conf, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1585
for (queue = 0; queue < tx_queue_cnt; queue++)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1586
stmmac_clear_tx_descriptors(priv, dma_conf, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1603
int i, gfp_t flags, u32 queue)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1605
struct stmmac_rx_queue *rx_q = &dma_conf->rx_queue[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1670
u32 queue, int i)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1672
struct stmmac_tx_queue *tx_q = &dma_conf->tx_queue[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1716
u32 queue)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1718
struct stmmac_rx_queue *rx_q = &dma_conf->rx_queue[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1727
u32 queue, gfp_t flags)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1729
struct stmmac_rx_queue *rx_q = &dma_conf->rx_queue[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1742
queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1760
u32 queue)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1762
struct stmmac_rx_queue *rx_q = &dma_conf->rx_queue[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1778
u32 queue)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1780
struct stmmac_rx_queue *rx_q = &dma_conf->rx_queue[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1813
static struct xsk_buff_pool *stmmac_get_xsk_pool(struct stmmac_priv *priv, u32 queue)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1815
if (!stmmac_xdp_is_enabled(priv) || !test_bit(queue, priv->af_xdp_zc_qps))
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1818
return xsk_get_pool_from_qid(priv->dev, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1833
u32 queue, gfp_t flags)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1835
struct stmmac_rx_queue *rx_q = &dma_conf->rx_queue[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1842
stmmac_clear_rx_descriptors(priv, dma_conf, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1846
rx_q->xsk_pool = stmmac_get_xsk_pool(priv, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1869
stmmac_alloc_rx_buffers_zc(priv, dma_conf, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1871
ret = stmmac_alloc_rx_buffers(priv, dma_conf, queue, flags);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1897
int queue;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1904
for (queue = 0; queue < rx_count; queue++) {
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1905
ret = __init_dma_rx_desc_rings(priv, dma_conf, queue, flags);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1913
while (queue >= 0) {
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1914
struct stmmac_rx_queue *rx_q = &dma_conf->rx_queue[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1917
dma_free_rx_xskbufs(priv, dma_conf, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1919
dma_free_rx_skbufs(priv, dma_conf, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1924
queue--;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1941
u32 queue)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1943
struct stmmac_tx_queue *tx_q = &dma_conf->tx_queue[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1962
tx_q->xsk_pool = stmmac_get_xsk_pool(priv, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1991
u32 queue;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1995
for (queue = 0; queue < tx_queue_cnt; queue++)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
1996
__init_dma_tx_desc_rings(priv, dma_conf, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2039
u32 queue)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2041
struct stmmac_tx_queue *tx_q = &dma_conf->tx_queue[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2047
stmmac_free_tx_buffer(priv, dma_conf, queue, i);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2063
u32 queue;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2065
for (queue = 0; queue < tx_queue_cnt; queue++)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2066
dma_free_tx_skbufs(priv, &priv->dma_conf, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2077
u32 queue)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2079
struct stmmac_rx_queue *rx_q = &dma_conf->rx_queue[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2083
dma_free_rx_xskbufs(priv, dma_conf, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2085
dma_free_rx_skbufs(priv, dma_conf, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2112
u32 queue;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2115
for (queue = 0; queue < rx_count; queue++)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2116
__free_dma_rx_desc_resources(priv, dma_conf, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2127
u32 queue)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2129
struct stmmac_tx_queue *tx_q = &dma_conf->tx_queue[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2134
dma_free_tx_skbufs(priv, dma_conf, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2159
u32 queue;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2162
for (queue = 0; queue < tx_count; queue++)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2163
__free_dma_tx_desc_resources(priv, dma_conf, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2178
u32 queue)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2180
struct stmmac_rx_queue *rx_q = &dma_conf->rx_queue[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2181
struct stmmac_channel *ch = &priv->channel[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2192
rx_q->queue_index = queue;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2241
test_bit(queue, priv->af_xdp_zc_qps))
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2261
u32 queue;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2265
for (queue = 0; queue < rx_count; queue++) {
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2266
ret = __alloc_dma_rx_desc_resources(priv, dma_conf, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2291
u32 queue)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2293
struct stmmac_tx_queue *tx_q = &dma_conf->tx_queue[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2297
tx_q->queue_index = queue;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2337
u32 queue;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2341
for (queue = 0; queue < tx_count; queue++) {
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2342
ret = __alloc_dma_tx_desc_resources(priv, dma_conf, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2402
int queue;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2405
for (queue = 0; queue < rx_queues_count; queue++) {
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2406
mode = priv->plat->rx_queues_cfg[queue].mode_to_use;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2407
stmmac_rx_queue_enable(priv, priv->hw, mode, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2646
static bool stmmac_xdp_xmit_zc(struct stmmac_priv *priv, u32 queue, u32 budget)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2648
struct netdev_queue *nq = netdev_get_tx_queue(priv->dev, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2649
struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2650
struct stmmac_txq_stats *txq_stats = &priv->xstats.txq_stats[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2651
bool csum = !priv->plat->tx_queues_cfg[queue].coe_unsupported;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2662
budget = min(budget, stmmac_tx_avail(priv, queue));
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2673
if (unlikely(stmmac_tx_avail(priv, queue) < STMMAC_TX_XSK_AVAIL) ||
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2683
priv->est->max_sdu[queue] &&
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2684
xdp_desc.len > priv->est->max_sdu[queue]) {
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2685
priv->xstats.max_sdu_txq_drop[queue]++;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
269
u32 queue;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
271
for (queue = 0; queue < maxq; queue++) {
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2718
if (!priv->tx_coal_frames[queue])
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
272
struct stmmac_channel *ch = &priv->channel[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2720
else if (tx_q->tx_count_frames % priv->tx_coal_frames[queue] == 0)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2742
stmmac_enable_dma_transmission(priv, priv->ioaddr, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
275
test_bit(queue, priv->af_xdp_zc_qps)) {
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2755
stmmac_flush_tx_descriptors(priv, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2792
static int stmmac_tx_clean(struct stmmac_priv *priv, int budget, u32 queue,
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2795
struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2796
struct stmmac_txq_stats *txq_stats = &priv->xstats.txq_stats[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
280
if (queue < rx_queues_cnt)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2801
__netif_tx_lock_bh(netdev_get_tx_queue(priv->dev, queue));
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
282
if (queue < tx_queues_cnt)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2851
stmmac_bump_dma_threshold(priv, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2922
netdev_tx_completed_queue(netdev_get_tx_queue(priv->dev, queue),
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2926
queue))) &&
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2927
stmmac_tx_avail(priv, queue) > STMMAC_TX_THRESH(priv)) {
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2931
netif_tx_wake_queue(netdev_get_tx_queue(priv->dev, queue));
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2948
work_done = stmmac_xdp_xmit_zc(priv, queue,
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
295
u32 queue;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
2971
__netif_tx_unlock_bh(netdev_get_tx_queue(priv->dev, queue));
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
298
for (queue = 0; queue < rx_queues_cnt; queue++) {
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
299
rx_q = &priv->dma_conf.rx_queue[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
318
u32 queue;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
320
for (queue = 0; queue < maxq; queue++) {
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
321
struct stmmac_channel *ch = &priv->channel[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
324
test_bit(queue, priv->af_xdp_zc_qps)) {
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
329
if (queue < rx_queues_cnt)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
3309
static void stmmac_tx_timer_arm(struct stmmac_priv *priv, u32 queue)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
331
if (queue < tx_queues_cnt)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
3311
struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
3312
u32 tx_coal_timer = priv->tx_coal_timer[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
3415
u32 queue;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
3417
for (queue = 0; queue < tx_queues_count; queue++) {
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
3418
weight = priv->plat->tx_queues_cfg[queue].weight;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
3419
stmmac_set_mtl_tx_queue_weight(priv, priv->hw, weight, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
3432
u32 queue;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
3435
for (queue = 1; queue < tx_queues_count; queue++) {
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
3436
mode_to_use = priv->plat->tx_queues_cfg[queue].mode_to_use;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
3441
priv->plat->tx_queues_cfg[queue].send_slope,
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
3442
priv->plat->tx_queues_cfg[queue].idle_slope,
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
3443
priv->plat->tx_queues_cfg[queue].high_credit,
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
3444
priv->plat->tx_queues_cfg[queue].low_credit,
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
3445
queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
3457
u32 queue;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
3460
for (queue = 0; queue < rx_queues_count; queue++) {
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
3461
chan = priv->plat->rx_queues_cfg[queue].chan;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
3462
stmmac_map_mtl_to_dma(priv, priv->hw, queue, chan);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
3474
u32 queue;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
3477
for (queue = 0; queue < rx_queues_count; queue++) {
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
3478
if (!priv->plat->rx_queues_cfg[queue].use_prio)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
3481
prio = priv->plat->rx_queues_cfg[queue].prio;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
3482
stmmac_rx_queue_prio(priv, priv->hw, prio, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
3494
u32 queue;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
3497
for (queue = 0; queue < tx_queues_count; queue++) {
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
3498
if (!priv->plat->tx_queues_cfg[queue].use_prio)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
3501
prio = priv->plat->tx_queues_cfg[queue].prio;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
3502
stmmac_tx_queue_prio(priv, priv->hw, prio, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
3514
u32 queue;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
3517
for (queue = 0; queue < rx_queues_count; queue++) {
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
3519
if (priv->plat->rx_queues_cfg[queue].pkt_route == 0x0)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
3522
packet = priv->plat->rx_queues_cfg[queue].pkt_route;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
3523
stmmac_rx_queue_routing(priv, priv->hw, packet, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
356
static inline u32 stmmac_tx_avail(struct stmmac_priv *priv, u32 queue)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
358
struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
3675
u32 queue;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
3677
for (queue = 0; queue < rx_cnt; queue++) {
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
3678
if (!priv->rx_riwt[queue])
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
3679
priv->rx_riwt[queue] = DEF_DMA_RIWT;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
3682
priv->rx_riwt[queue], queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
374
static inline u32 stmmac_rx_dirty(struct stmmac_priv *priv, u32 queue)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
376
struct stmmac_rx_queue *rx_q = &priv->dma_conf.rx_queue[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
390
u32 queue;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
393
for (queue = 0; queue < tx_cnt; queue++) {
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
394
struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4277
int total_len, bool last_segment, u32 queue)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4279
struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4312
static void stmmac_flush_tx_descriptors(struct stmmac_priv *priv, int queue)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4314
struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4331
stmmac_set_tx_tail_ptr(priv, priv->ioaddr, tx_q->tx_tail_addr, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4379
u32 pay_len, mss, queue;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4398
queue = skb_get_queue_mapping(skb);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4400
tx_q = &priv->dma_conf.tx_queue[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4401
txq_stats = &priv->xstats.txq_stats[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4414
if (unlikely(stmmac_tx_avail(priv, queue) <
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4416
if (!netif_tx_queue_stopped(netdev_get_tx_queue(dev, queue))) {
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4418
queue));
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4469
(nfrags == 0), queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4499
(i == nfrags - 1), queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4519
else if (!priv->tx_coal_frames[queue])
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4521
else if (tx_packets > priv->tx_coal_frames[queue])
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4524
priv->tx_coal_frames[queue]) < tx_packets)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4546
if (unlikely(stmmac_tx_avail(priv, queue) <= (MAX_SKB_FRAGS + 1))) {
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4549
netif_tx_stop_queue(netdev_get_tx_queue(priv->dev, queue));
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4600
netdev_tx_sent_queue(netdev_get_tx_queue(dev, queue), skb->len);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4603
stmmac_flush_tx_descriptors(priv, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4604
stmmac_tx_timer_arm(priv, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4650
u32 queue = skb_get_queue_mapping(skb);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4663
tx_q = &priv->dma_conf.tx_queue[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4664
txq_stats = &priv->xstats.txq_stats[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4680
priv->est->max_sdu[queue]) {
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4685
if (sdu_len > priv->est->max_sdu[queue]) {
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4686
priv->xstats.max_sdu_txq_drop[queue]++;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4691
if (unlikely(stmmac_tx_avail(priv, queue) < nfrags + 1)) {
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4692
if (!netif_tx_queue_stopped(netdev_get_tx_queue(dev, queue))) {
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4694
queue));
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4719
(priv->plat->tx_queues_cfg[queue].coe_unsupported ||
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4797
else if (!priv->tx_coal_frames[queue])
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4799
else if (tx_packets > priv->tx_coal_frames[queue])
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4802
priv->tx_coal_frames[queue]) < tx_packets)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4837
if (unlikely(stmmac_tx_avail(priv, queue) <= (MAX_SKB_FRAGS + 1))) {
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4840
netif_tx_stop_queue(netdev_get_tx_queue(priv->dev, queue));
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4895
netdev_tx_sent_queue(netdev_get_tx_queue(dev, queue), skb->len);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4897
stmmac_enable_dma_transmission(priv, priv->ioaddr, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4899
stmmac_flush_tx_descriptors(priv, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4900
stmmac_tx_timer_arm(priv, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4937
static inline void stmmac_rx_refill(struct stmmac_priv *priv, u32 queue)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4939
struct stmmac_rx_queue *rx_q = &priv->dma_conf.rx_queue[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4940
int dirty = stmmac_rx_dirty(priv, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4981
rx_q->rx_count_frames += priv->rx_coal_frames[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4982
if (rx_q->rx_count_frames > priv->rx_coal_frames[queue])
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4985
use_rx_wd = !priv->rx_coal_frames[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
4998
stmmac_set_rx_tail_ptr(priv, priv->ioaddr, rx_q->rx_tail_addr, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5000
stmmac_enable_dma_reception(priv, priv->ioaddr, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5066
static int stmmac_xdp_xmit_xdpf(struct stmmac_priv *priv, int queue,
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5069
struct stmmac_txq_stats *txq_stats = &priv->xstats.txq_stats[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5070
struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5071
bool csum = !priv->plat->tx_queues_cfg[queue].coe_unsupported;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5077
if (stmmac_tx_avail(priv, queue) < STMMAC_TX_THRESH(priv))
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5081
priv->est->max_sdu[queue] &&
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5082
xdpf->len > priv->est->max_sdu[queue]) {
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5083
priv->xstats.max_sdu_txq_drop[queue]++;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5128
if (tx_q->tx_count_frames % priv->tx_coal_frames[queue] == 0)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5141
stmmac_enable_dma_transmission(priv, priv->ioaddr, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5170
int queue;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5176
queue = stmmac_xdp_get_tx_queue(priv, cpu);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5177
nq = netdev_get_tx_queue(priv->dev, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5184
res = stmmac_xdp_xmit_xdpf(priv, queue, xdpf, zc);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5186
stmmac_flush_tx_descriptors(priv, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5257
int queue;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5259
queue = stmmac_xdp_get_tx_queue(priv, cpu);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5262
stmmac_tx_timer_arm(priv, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5288
static void stmmac_dispatch_skb_zc(struct stmmac_priv *priv, u32 queue,
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5292
struct stmmac_rxq_stats *rxq_stats = &priv->xstats.rxq_stats[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5293
struct stmmac_channel *ch = &priv->channel[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5323
skb_record_rx_queue(skb, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5332
static bool stmmac_rx_refill_zc(struct stmmac_priv *priv, u32 queue, u32 budget)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5334
struct stmmac_rx_queue *rx_q = &priv->dma_conf.rx_queue[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5339
budget = min(budget, stmmac_rx_dirty(priv, queue));
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5365
rx_q->rx_count_frames += priv->rx_coal_frames[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5366
if (rx_q->rx_count_frames > priv->rx_coal_frames[queue])
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5369
use_rx_wd = !priv->rx_coal_frames[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5384
stmmac_set_rx_tail_ptr(priv, priv->ioaddr, rx_q->rx_tail_addr, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5400
static int stmmac_rx_zc(struct stmmac_priv *priv, int limit, u32 queue)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5402
struct stmmac_rxq_stats *rxq_stats = &priv->xstats.rxq_stats[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5403
struct stmmac_rx_queue *rx_q = &priv->dma_conf.rx_queue[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5405
int dirty = stmmac_rx_dirty(priv, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5456
!stmmac_rx_refill_zc(priv, queue, dirty);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5539
stmmac_dispatch_skb_zc(priv, queue, p, np, buf->xdp);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5575
if (failure || stmmac_rx_dirty(priv, queue) > 0)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5594
static int stmmac_rx(struct stmmac_priv *priv, int limit, u32 queue)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5597
struct stmmac_rxq_stats *rxq_stats = &priv->xstats.rxq_stats[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5598
struct stmmac_rx_queue *rx_q = &priv->dma_conf.rx_queue[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5599
struct stmmac_channel *ch = &priv->channel[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5843
skb_record_rx_queue(skb, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
5861
stmmac_rx_refill(priv, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
6158
u32 queue;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
6186
for (queue = 0; queue < queues_count; queue++)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
6187
stmmac_host_mtl_irq_status(priv, priv->hw, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
6459
u32 queue;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
6464
for (queue = 0; queue < rx_count; queue++) {
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
6465
struct stmmac_rx_queue *rx_q = &priv->dma_conf.rx_queue[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
6467
seq_printf(seq, "RX Queue %d:\n", queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
6480
for (queue = 0; queue < tx_count; queue++) {
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
6481
struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
6483
seq_printf(seq, "TX Queue %d:\n", queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
6895
int queue;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
6903
queue = stmmac_xdp_get_tx_queue(priv, cpu);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
6904
nq = netdev_get_tx_queue(priv->dev, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
6913
res = stmmac_xdp_xmit_xdpf(priv, queue, frames[i], true);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
6921
stmmac_flush_tx_descriptors(priv, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
6922
stmmac_tx_timer_arm(priv, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
6930
void stmmac_disable_rx_queue(struct stmmac_priv *priv, u32 queue)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
6932
struct stmmac_channel *ch = &priv->channel[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
6936
stmmac_disable_dma_irq(priv, priv->ioaddr, queue, 1, 0);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
6939
stmmac_stop_rx_dma(priv, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
6940
__free_dma_rx_desc_resources(priv, &priv->dma_conf, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
6943
void stmmac_enable_rx_queue(struct stmmac_priv *priv, u32 queue)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
6945
struct stmmac_rx_queue *rx_q = &priv->dma_conf.rx_queue[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
6946
struct stmmac_channel *ch = &priv->channel[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
6951
ret = __alloc_dma_rx_desc_resources(priv, &priv->dma_conf, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
6957
ret = __init_dma_rx_desc_rings(priv, &priv->dma_conf, queue, GFP_KERNEL);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
6959
__free_dma_rx_desc_resources(priv, &priv->dma_conf, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
6964
stmmac_reset_rx_queue(priv, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
6965
stmmac_clear_rx_descriptors(priv, &priv->dma_conf, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
6986
stmmac_start_rx_dma(priv, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
6989
stmmac_enable_dma_irq(priv, priv->ioaddr, queue, 1, 0);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
6993
void stmmac_disable_tx_queue(struct stmmac_priv *priv, u32 queue)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
6995
struct stmmac_channel *ch = &priv->channel[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
6999
stmmac_disable_dma_irq(priv, priv->ioaddr, queue, 0, 1);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
7002
stmmac_stop_tx_dma(priv, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
7003
__free_dma_tx_desc_resources(priv, &priv->dma_conf, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
7006
void stmmac_enable_tx_queue(struct stmmac_priv *priv, u32 queue)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
7008
struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
7009
struct stmmac_channel *ch = &priv->channel[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
7013
ret = __alloc_dma_tx_desc_resources(priv, &priv->dma_conf, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
7019
ret = __init_dma_tx_desc_rings(priv, &priv->dma_conf, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
7021
__free_dma_tx_desc_resources(priv, &priv->dma_conf, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
7026
stmmac_reset_tx_queue(priv, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
7027
stmmac_clear_tx_descriptors(priv, &priv->dma_conf, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
7039
stmmac_start_tx_dma(priv, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
7042
stmmac_enable_dma_irq(priv, priv->ioaddr, queue, 0, 1);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
7186
int stmmac_xsk_wakeup(struct net_device *dev, u32 queue, u32 flags)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
7200
if (queue >= priv->plat->rx_queues_to_use ||
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
7201
queue >= priv->plat->tx_queues_to_use)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
7204
rx_q = &priv->dma_conf.rx_queue[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
7205
tx_q = &priv->dma_conf.tx_queue[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
7206
ch = &priv->channel[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
7509
u32 queue, maxq;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
7513
for (queue = 0; queue < maxq; queue++) {
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
7514
struct stmmac_channel *ch = &priv->channel[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
7517
ch->index = queue;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
7520
if (queue < priv->plat->rx_queues_to_use) {
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
7523
if (queue < priv->plat->tx_queues_to_use) {
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
7527
if (queue < priv->plat->rx_queues_to_use &&
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
7528
queue < priv->plat->tx_queues_to_use) {
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
7538
u32 queue, maxq;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
7542
for (queue = 0; queue < maxq; queue++) {
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
7543
struct stmmac_channel *ch = &priv->channel[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
7545
if (queue < priv->plat->rx_queues_to_use)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
7547
if (queue < priv->plat->tx_queues_to_use)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
7549
if (queue < priv->plat->rx_queues_to_use &&
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
7550
queue < priv->plat->tx_queues_to_use) {
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
8188
static void stmmac_reset_rx_queue(struct stmmac_priv *priv, u32 queue)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
8190
struct stmmac_rx_queue *rx_q = &priv->dma_conf.rx_queue[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
8196
static void stmmac_reset_tx_queue(struct stmmac_priv *priv, u32 queue)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
8198
struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
8204
netdev_tx_reset_queue(netdev_get_tx_queue(priv->dev, queue));
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
8215
u32 queue;
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
8217
for (queue = 0; queue < rx_cnt; queue++)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
8218
stmmac_reset_rx_queue(priv, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
8220
for (queue = 0; queue < tx_cnt; queue++)
drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
8221
stmmac_reset_tx_queue(priv, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_platform.c
139
u8 queue = 0;
drivers/net/ethernet/stmicro/stmmac/stmmac_platform.c
172
if (queue >= plat->rx_queues_to_use)
drivers/net/ethernet/stmicro/stmmac/stmmac_platform.c
176
plat->rx_queues_cfg[queue].mode_to_use = MTL_QUEUE_DCB;
drivers/net/ethernet/stmicro/stmmac/stmmac_platform.c
178
plat->rx_queues_cfg[queue].mode_to_use = MTL_QUEUE_AVB;
drivers/net/ethernet/stmicro/stmmac/stmmac_platform.c
180
plat->rx_queues_cfg[queue].mode_to_use = MTL_QUEUE_DCB;
drivers/net/ethernet/stmicro/stmmac/stmmac_platform.c
183
&plat->rx_queues_cfg[queue].chan);
drivers/net/ethernet/stmicro/stmmac/stmmac_platform.c
187
&plat->rx_queues_cfg[queue].prio))
drivers/net/ethernet/stmicro/stmmac/stmmac_platform.c
188
plat->rx_queues_cfg[queue].use_prio = true;
drivers/net/ethernet/stmicro/stmmac/stmmac_platform.c
192
plat->rx_queues_cfg[queue].pkt_route = PACKET_AVCPQ;
drivers/net/ethernet/stmicro/stmmac/stmmac_platform.c
194
plat->rx_queues_cfg[queue].pkt_route = PACKET_PTPQ;
drivers/net/ethernet/stmicro/stmmac/stmmac_platform.c
196
plat->rx_queues_cfg[queue].pkt_route = PACKET_DCBCPQ;
drivers/net/ethernet/stmicro/stmmac/stmmac_platform.c
198
plat->rx_queues_cfg[queue].pkt_route = PACKET_UPQ;
drivers/net/ethernet/stmicro/stmmac/stmmac_platform.c
200
plat->rx_queues_cfg[queue].pkt_route = PACKET_MCBCQ;
drivers/net/ethernet/stmicro/stmmac/stmmac_platform.c
202
queue++;
drivers/net/ethernet/stmicro/stmmac/stmmac_platform.c
204
if (queue != plat->rx_queues_to_use) {
drivers/net/ethernet/stmicro/stmmac/stmmac_platform.c
223
queue = 0;
drivers/net/ethernet/stmicro/stmmac/stmmac_platform.c
227
if (queue >= plat->tx_queues_to_use)
drivers/net/ethernet/stmicro/stmmac/stmmac_platform.c
231
&plat->tx_queues_cfg[queue].weight))
drivers/net/ethernet/stmicro/stmmac/stmmac_platform.c
232
plat->tx_queues_cfg[queue].weight = 0x10 + queue;
drivers/net/ethernet/stmicro/stmmac/stmmac_platform.c
235
plat->tx_queues_cfg[queue].mode_to_use = MTL_QUEUE_DCB;
drivers/net/ethernet/stmicro/stmmac/stmmac_platform.c
238
plat->tx_queues_cfg[queue].mode_to_use = MTL_QUEUE_AVB;
drivers/net/ethernet/stmicro/stmmac/stmmac_platform.c
242
&plat->tx_queues_cfg[queue].send_slope))
drivers/net/ethernet/stmicro/stmmac/stmmac_platform.c
243
plat->tx_queues_cfg[queue].send_slope = 0x0;
drivers/net/ethernet/stmicro/stmmac/stmmac_platform.c
245
&plat->tx_queues_cfg[queue].idle_slope))
drivers/net/ethernet/stmicro/stmmac/stmmac_platform.c
246
plat->tx_queues_cfg[queue].idle_slope = 0x0;
drivers/net/ethernet/stmicro/stmmac/stmmac_platform.c
248
&plat->tx_queues_cfg[queue].high_credit))
drivers/net/ethernet/stmicro/stmmac/stmmac_platform.c
249
plat->tx_queues_cfg[queue].high_credit = 0x0;
drivers/net/ethernet/stmicro/stmmac/stmmac_platform.c
251
&plat->tx_queues_cfg[queue].low_credit))
drivers/net/ethernet/stmicro/stmmac/stmmac_platform.c
252
plat->tx_queues_cfg[queue].low_credit = 0x0;
drivers/net/ethernet/stmicro/stmmac/stmmac_platform.c
254
plat->tx_queues_cfg[queue].mode_to_use = MTL_QUEUE_DCB;
drivers/net/ethernet/stmicro/stmmac/stmmac_platform.c
258
&plat->tx_queues_cfg[queue].prio))
drivers/net/ethernet/stmicro/stmmac/stmmac_platform.c
259
plat->tx_queues_cfg[queue].use_prio = true;
drivers/net/ethernet/stmicro/stmmac/stmmac_platform.c
261
plat->tx_queues_cfg[queue].coe_unsupported =
drivers/net/ethernet/stmicro/stmmac/stmmac_platform.c
264
queue++;
drivers/net/ethernet/stmicro/stmmac/stmmac_platform.c
266
if (queue != plat->tx_queues_to_use) {
drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c
1684
static int __stmmac_test_jumbo(struct stmmac_priv *priv, u16 queue)
drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c
1691
attr.queue_mapping = queue;
drivers/net/ethernet/stmicro/stmmac/stmmac_selftests.c
1777
qopt.queue = i;
drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c
1113
int queue = qopt->queue_stats.queue;
drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c
1115
q_stats->stats.window_drops = priv->xstats.max_sdu_txq_drop[queue] +
drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c
1116
priv->xstats.mtl_est_txq_hlbf[queue] +
drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c
1117
priv->xstats.mtl_est_txq_hlbs[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c
1163
if (qopt->queue >= priv->plat->tx_queues_to_use)
drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c
1165
if (!(priv->dma_conf.tx_queue[qopt->queue].tbs & STMMAC_TBS_AVAIL))
drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c
1169
priv->dma_conf.tx_queue[qopt->queue].tbs |= STMMAC_TBS_EN;
drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c
1171
priv->dma_conf.tx_queue[qopt->queue].tbs &= ~STMMAC_TBS_EN;
drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c
1174
qopt->enable ? "enabled" : "disabled", qopt->queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c
337
u32 queue = qopt->queue;
drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c
344
if (queue <= 0 || queue >= tx_queues_count)
drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c
375
mode_to_use = priv->plat->tx_queues_cfg[queue].mode_to_use;
drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c
377
ret = stmmac_dma_qmode(priv, priv->ioaddr, queue, MTL_QUEUE_AVB);
drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c
381
priv->plat->tx_queues_cfg[queue].mode_to_use = MTL_QUEUE_AVB;
drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c
383
ret = stmmac_dma_qmode(priv, priv->ioaddr, queue,
drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c
388
priv->plat->tx_queues_cfg[queue].mode_to_use = MTL_QUEUE_DCB;
drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c
394
priv->plat->tx_queues_cfg[queue].idle_slope = value & GENMASK(31, 0);
drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c
397
priv->plat->tx_queues_cfg[queue].send_slope = value & GENMASK(31, 0);
drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c
400
priv->plat->tx_queues_cfg[queue].high_credit = value & GENMASK(31, 0);
drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c
403
priv->plat->tx_queues_cfg[queue].low_credit = value & GENMASK(31, 0);
drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c
406
priv->plat->tx_queues_cfg[queue].send_slope,
drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c
407
priv->plat->tx_queues_cfg[queue].idle_slope,
drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c
408
priv->plat->tx_queues_cfg[queue].high_credit,
drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c
409
priv->plat->tx_queues_cfg[queue].low_credit,
drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c
410
queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_tc.c
415
queue, qopt->sendslope, qopt->idleslope,
drivers/net/ethernet/stmicro/stmmac/stmmac_xdp.c
10
struct xsk_buff_pool *pool, u16 queue)
drivers/net/ethernet/stmicro/stmmac/stmmac_xdp.c
12
struct stmmac_channel *ch = &priv->channel[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_xdp.c
17
if (queue >= priv->plat->rx_queues_to_use ||
drivers/net/ethernet/stmicro/stmmac/stmmac_xdp.c
18
queue >= priv->plat->tx_queues_to_use)
drivers/net/ethernet/stmicro/stmmac/stmmac_xdp.c
39
stmmac_disable_rx_queue(priv, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_xdp.c
40
stmmac_disable_tx_queue(priv, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_xdp.c
43
set_bit(queue, priv->af_xdp_zc_qps);
drivers/net/ethernet/stmicro/stmmac/stmmac_xdp.c
46
stmmac_enable_rx_queue(priv, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_xdp.c
47
stmmac_enable_tx_queue(priv, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_xdp.c
50
err = stmmac_xsk_wakeup(priv->dev, queue, XDP_WAKEUP_RX);
drivers/net/ethernet/stmicro/stmmac/stmmac_xdp.c
58
static int stmmac_xdp_disable_pool(struct stmmac_priv *priv, u16 queue)
drivers/net/ethernet/stmicro/stmmac/stmmac_xdp.c
60
struct stmmac_channel *ch = &priv->channel[queue];
drivers/net/ethernet/stmicro/stmmac/stmmac_xdp.c
64
if (queue >= priv->plat->rx_queues_to_use ||
drivers/net/ethernet/stmicro/stmmac/stmmac_xdp.c
65
queue >= priv->plat->tx_queues_to_use)
drivers/net/ethernet/stmicro/stmmac/stmmac_xdp.c
68
pool = xsk_get_pool_from_qid(priv->dev, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_xdp.c
76
stmmac_disable_rx_queue(priv, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_xdp.c
77
stmmac_disable_tx_queue(priv, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_xdp.c
83
clear_bit(queue, priv->af_xdp_zc_qps);
drivers/net/ethernet/stmicro/stmmac/stmmac_xdp.c
86
stmmac_enable_rx_queue(priv, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_xdp.c
87
stmmac_enable_tx_queue(priv, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_xdp.c
96
u16 queue)
drivers/net/ethernet/stmicro/stmmac/stmmac_xdp.c
98
return pool ? stmmac_xdp_enable_pool(priv, pool, queue) :
drivers/net/ethernet/stmicro/stmmac/stmmac_xdp.c
99
stmmac_xdp_disable_pool(priv, queue);
drivers/net/ethernet/stmicro/stmmac/stmmac_xdp.h
10
u16 queue);
drivers/net/ethernet/sunplus/spl2sw_int.c
101
desc->cmd2 = (rx_pos == comm->rx_desc_num[queue] - 1) ?
drivers/net/ethernet/sunplus/spl2sw_int.c
110
rx_pos = ((rx_pos + 1) == comm->rx_desc_num[queue]) ? 0 : rx_pos + 1;
drivers/net/ethernet/sunplus/spl2sw_int.c
117
if (queue == 1 && !(h_desc->cmd1 & RXD_OWN))
drivers/net/ethernet/sunplus/spl2sw_int.c
121
comm->rx_pos[queue] = rx_pos;
drivers/net/ethernet/sunplus/spl2sw_int.c
124
if (queue == 0)
drivers/net/ethernet/sunplus/spl2sw_int.c
125
h_desc = comm->rx_desc[queue] + rx_pos;
drivers/net/ethernet/sunplus/spl2sw_int.c
28
s32 queue;
drivers/net/ethernet/sunplus/spl2sw_int.c
35
for (queue = 0; queue < RX_DESC_QUEUE_NUM; queue++) {
drivers/net/ethernet/sunplus/spl2sw_int.c
36
rx_pos = comm->rx_pos[queue];
drivers/net/ethernet/sunplus/spl2sw_int.c
37
rx_count = comm->rx_desc_num[queue];
drivers/net/ethernet/sunplus/spl2sw_int.c
40
sinfo = comm->rx_skb_info[queue] + rx_pos;
drivers/net/ethernet/sunplus/spl2sw_int.c
41
desc = comm->rx_desc[queue] + rx_pos;
drivers/net/ethernet/sunplus/spl2sw_int.c
76
desc->cmd2 = (rx_pos == comm->rx_desc_num[queue] - 1) ?
drivers/net/ethernet/sunplus/spl2sw_int.c
89
desc->cmd2 = (rx_pos == comm->rx_desc_num[queue] - 1) ?
drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c
1507
unsigned int qptc, qptc_extra, queue;
drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c
1518
for (i = 0, queue = 0; i < pdata->hw_feat.tc_cnt; i++) {
drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c
1521
"TXq%u mapped to TC%u\n", queue, i);
drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c
1522
regval = readl(XLGMAC_MTL_REG(pdata, queue,
drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c
1528
writel(regval, XLGMAC_MTL_REG(pdata, queue,
drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c
1530
queue++;
drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c
1535
"TXq%u mapped to TC%u\n", queue, i);
drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c
1536
regval = readl(XLGMAC_MTL_REG(pdata, queue,
drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c
1542
writel(regval, XLGMAC_MTL_REG(pdata, queue,
drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c
1544
queue++;
drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c
574
unsigned int queue)
drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c
585
rx_status = readl(XLGMAC_MTL_REG(pdata, queue, MTL_Q_RQDR));
drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c
599
queue);
drivers/net/ethernet/ti/am65-cpsw-ethtool.c
915
static int am65_cpsw_get_per_queue_coalesce(struct net_device *ndev, u32 queue,
drivers/net/ethernet/ti/am65-cpsw-ethtool.c
922
if (queue >= AM65_CPSW_MAX_QUEUES)
drivers/net/ethernet/ti/am65-cpsw-ethtool.c
925
tx_chn = &common->tx_chns[queue];
drivers/net/ethernet/ti/am65-cpsw-ethtool.c
928
rx_flow = &common->rx_chns.flows[queue];
drivers/net/ethernet/ti/am65-cpsw-ethtool.c
941
static int am65_cpsw_set_per_queue_coalesce(struct net_device *ndev, u32 queue,
drivers/net/ethernet/ti/am65-cpsw-ethtool.c
948
if (queue >= AM65_CPSW_MAX_QUEUES)
drivers/net/ethernet/ti/am65-cpsw-ethtool.c
951
tx_chn = &common->tx_chns[queue];
drivers/net/ethernet/ti/am65-cpsw-ethtool.c
957
rx_flow = &common->rx_chns.flows[queue];
drivers/net/ethernet/ti/am65-cpsw-qos.c
1224
int queue, u32 rate_mbps)
drivers/net/ethernet/ti/am65-cpsw-qos.c
1234
queue, rate_mbps, common->tx_ch_rate_msk);
drivers/net/ethernet/ti/am65-cpsw-qos.c
1241
ch_rate = netdev_get_tx_queue(ndev, queue)->tx_maxrate;
drivers/net/ethernet/ti/am65-cpsw-qos.c
1253
if (rate_mbps && !(tx_ch_rate_msk_new & BIT(queue))) {
drivers/net/ethernet/ti/am65-cpsw-qos.c
1254
tx_ch_rate_msk_new |= BIT(queue);
drivers/net/ethernet/ti/am65-cpsw-qos.c
1255
ch_msk = GENMASK(common->tx_ch_num - 1, queue);
drivers/net/ethernet/ti/am65-cpsw-qos.c
1258
tx_ch_rate_msk_new &= ~BIT(queue);
drivers/net/ethernet/ti/am65-cpsw-qos.c
1259
ch_msk = queue ? GENMASK(queue - 1, 0) : 0;
drivers/net/ethernet/ti/am65-cpsw-qos.c
1270
tx_chn = &common->tx_chns[queue];
drivers/net/ethernet/ti/am65-cpsw-qos.c
1278
am65_cpsw_qos_tx_p0_rate_apply(common, queue, rate_mbps);
drivers/net/ethernet/ti/am65-cpsw-qos.h
101
int am65_cpsw_qos_ndo_tx_p0_set_maxrate(struct net_device *ndev, int queue, u32 rate_mbps);
drivers/net/ethernet/ti/am65-cpsw-qos.h
121
int queue,
drivers/net/ethernet/ti/cpsw_ethtool.c
524
struct netdev_queue *queue;
drivers/net/ethernet/ti/cpsw_ethtool.c
541
queue = netdev_get_tx_queue(priv->ndev, *ch);
drivers/net/ethernet/ti/cpsw_ethtool.c
542
queue->tx_maxrate = 0;
drivers/net/ethernet/ti/cpsw_priv.c
713
int cpsw_ndo_set_tx_maxrate(struct net_device *ndev, int queue, u32 rate)
drivers/net/ethernet/ti/cpsw_priv.c
722
ch_rate = netdev_get_tx_queue(ndev, queue)->tx_maxrate;
drivers/net/ethernet/ti/cpsw_priv.c
743
ret = cpdma_chan_set_rate(cpsw->txv[queue].ch, ch_rate);
drivers/net/ethernet/ti/cpsw_priv.c
755
netdev_get_tx_queue(slave->ndev, queue)->tx_maxrate = rate;
drivers/net/ethernet/ti/cpsw_priv.c
922
tc = netdev_txq_to_tc(priv->ndev, qopt->queue);
drivers/net/ethernet/ti/cpsw_priv.h
465
int cpsw_ndo_set_tx_maxrate(struct net_device *ndev, int queue, u32 rate);
drivers/net/ethernet/ti/icssg/icssg_config.c
173
int queue = 0, i, j;
drivers/net/ethernet/ti/icssg/icssg_config.c
178
queue = ICSSG_NUM_TX_QUEUES;
drivers/net/ethernet/ti/icssg/icssg_config.c
181
regmap_write(miig_rt, ICSSG_QUEUE_RESET_OFFSET, queue);
drivers/net/ethernet/ti/icssg/icssg_config.c
182
queue++;
drivers/net/ethernet/ti/icssg/icssg_config.c
185
queue = slice ? RECYCLE_Q_SLICE1 : RECYCLE_Q_SLICE0;
drivers/net/ethernet/ti/icssg/icssg_config.c
186
regmap_write(miig_rt, ICSSG_QUEUE_RESET_OFFSET, queue);
drivers/net/ethernet/ti/icssg/icssg_config.c
190
hwq_map[slice][i].queue);
drivers/net/ethernet/ti/icssg/icssg_config.c
219
queue = mp->queue;
drivers/net/ethernet/ti/icssg/icssg_config.c
220
regmap_write(miig_rt, ICSSG_QUEUE_OFFSET + 4 * queue,
drivers/net/ethernet/ti/icssg/icssg_config.c
89
int queue;
drivers/net/ethernet/ti/icssg/icssg_ethtool.c
226
static int emac_get_per_queue_coalesce(struct net_device *ndev, u32 queue,
drivers/net/ethernet/ti/icssg/icssg_ethtool.c
232
if (queue >= PRUETH_MAX_TX_QUEUES)
drivers/net/ethernet/ti/icssg/icssg_ethtool.c
235
tx_chn = &emac->tx_chns[queue];
drivers/net/ethernet/ti/icssg/icssg_ethtool.c
273
static int emac_set_per_queue_coalesce(struct net_device *ndev, u32 queue,
drivers/net/ethernet/ti/icssg/icssg_ethtool.c
280
if (queue >= PRUETH_MAX_TX_QUEUES)
drivers/net/ethernet/ti/icssg/icssg_ethtool.c
283
tx_chn = &emac->tx_chns[queue];
drivers/net/ethernet/ti/icssg/icssg_ethtool.c
288
ICSSG_MIN_COALESCE_USECS, queue);
drivers/net/ethernet/ti/icssg/icssg_prueth.h
439
int icssg_queue_pop(struct prueth *prueth, u8 queue);
drivers/net/ethernet/ti/icssg/icssg_prueth.h
440
void icssg_queue_push(struct prueth *prueth, int queue, u16 addr);
drivers/net/ethernet/ti/icssg/icssg_prueth.h
441
u32 icssg_queue_level(struct prueth *prueth, int queue);
drivers/net/ethernet/ti/icssg/icssg_queues.c
16
int icssg_queue_pop(struct prueth *prueth, u8 queue)
drivers/net/ethernet/ti/icssg/icssg_queues.c
20
if (queue >= ICSSG_QUEUES_MAX)
drivers/net/ethernet/ti/icssg/icssg_queues.c
23
regmap_read(prueth->miig_rt, ICSSG_QUEUE_CNT_OFFSET + 4 * queue, &cnt);
drivers/net/ethernet/ti/icssg/icssg_queues.c
27
regmap_read(prueth->miig_rt, ICSSG_QUEUE_OFFSET + 4 * queue, &val);
drivers/net/ethernet/ti/icssg/icssg_queues.c
33
void icssg_queue_push(struct prueth *prueth, int queue, u16 addr)
drivers/net/ethernet/ti/icssg/icssg_queues.c
35
if (queue >= ICSSG_QUEUES_MAX)
drivers/net/ethernet/ti/icssg/icssg_queues.c
38
regmap_write(prueth->miig_rt, ICSSG_QUEUE_OFFSET + 4 * queue, addr);
drivers/net/ethernet/ti/icssg/icssg_queues.c
42
u32 icssg_queue_level(struct prueth *prueth, int queue)
drivers/net/ethernet/ti/icssg/icssg_queues.c
46
if (queue >= ICSSG_QUEUES_MAX)
drivers/net/ethernet/ti/icssg/icssg_queues.c
49
regmap_read(prueth->miig_rt, ICSSG_QUEUE_CNT_OFFSET + 4 * queue, ®);
drivers/net/ethernet/ti/icssm/icssm_prueth.c
451
icssm_get_buff_desc_count(const struct prueth_queue_info *queue)
drivers/net/ethernet/ti/icssm/icssm_prueth.c
455
buffer_desc_count = queue->buffer_desc_end -
drivers/net/ethernet/ti/icssm/icssm_prueth.c
456
queue->buffer_desc_offset;
drivers/net/ethernet/ti/icssm/icssm_prueth.c
464
const struct prueth_queue_info *queue,
drivers/net/ethernet/ti/icssm/icssm_prueth.c
468
queue->buffer_desc_offset) / BD_SIZE;
drivers/net/ethernet/ti/icssm/icssm_prueth.c
470
queue->buffer_desc_offset) / BD_SIZE;
drivers/net/ethernet/wangxun/libwx/wx_lib.c
2451
u16 queue, u16 msix_vector)
drivers/net/ethernet/wangxun/libwx/wx_lib.c
2468
index = ((16 * (queue & 1)) + (8 * direction));
drivers/net/ethernet/wangxun/libwx/wx_lib.c
2469
ivar = rd32(wx, WX_PX_IVAR(queue >> 1));
drivers/net/ethernet/wangxun/libwx/wx_lib.c
2472
wr32(wx, WX_PX_IVAR(queue >> 1), ivar);
drivers/net/ethernet/wangxun/libwx/wx_vf_lib.c
29
static void wx_set_ivar_vf(struct wx *wx, s8 direction, u8 queue,
drivers/net/ethernet/wangxun/libwx/wx_vf_lib.c
44
index = ((16 * (queue & 1)) + (8 * direction));
drivers/net/ethernet/wangxun/libwx/wx_vf_lib.c
45
ivar = rd32(wx, WX_VXIVAR(queue >> 1));
drivers/net/ethernet/wangxun/libwx/wx_vf_lib.c
48
wr32(wx, WX_VXIVAR(queue >> 1), ivar);
drivers/net/ethernet/wangxun/txgbe/txgbe_ethtool.c
360
u8 queue;
drivers/net/ethernet/wangxun/txgbe/txgbe_ethtool.c
369
queue = TXGBE_RDB_FDIR_DROP_QUEUE;
drivers/net/ethernet/wangxun/txgbe/txgbe_ethtool.c
382
queue = wx->rx_ring[ring]->reg_idx;
drivers/net/ethernet/wangxun/txgbe/txgbe_ethtool.c
384
queue = ((vf - 1) * wx->num_rx_queues_per_pool) + ring;
drivers/net/ethernet/wangxun/txgbe/txgbe_ethtool.c
490
input->sw_idx, queue);
drivers/net/ethernet/wangxun/txgbe/txgbe_fdir.c
180
u8 queue)
drivers/net/ethernet/wangxun/txgbe/txgbe_fdir.c
209
fdircmd |= TXGBE_RDB_FDIR_CMD_RX_QUEUE(queue);
drivers/net/ethernet/wangxun/txgbe/txgbe_fdir.c
216
wx_dbg(wx, "Tx Queue=%x hash=%x\n", queue, (u32)fdirhashcmd);
drivers/net/ethernet/wangxun/txgbe/txgbe_fdir.c
397
u16 soft_id, u8 queue)
drivers/net/ethernet/wangxun/txgbe/txgbe_fdir.c
442
if (queue == TXGBE_RDB_FDIR_DROP_QUEUE)
drivers/net/ethernet/wangxun/txgbe/txgbe_fdir.c
445
fdircmd |= TXGBE_RDB_FDIR_CMD_RX_QUEUE(queue);
drivers/net/ethernet/wangxun/txgbe/txgbe_fdir.c
577
u8 queue = 0;
drivers/net/ethernet/wangxun/txgbe/txgbe_fdir.c
591
queue = TXGBE_RDB_FDIR_DROP_QUEUE;
drivers/net/ethernet/wangxun/txgbe/txgbe_fdir.c
602
queue = wx->rx_ring[ring]->reg_idx;
drivers/net/ethernet/wangxun/txgbe/txgbe_fdir.c
608
queue);
drivers/net/ethernet/wangxun/txgbe/txgbe_fdir.h
13
u16 soft_id, u8 queue);
drivers/net/ethernet/wangxun/txgbevf/txgbevf_main.c
52
u16 rss, queue;
drivers/net/ethernet/wangxun/txgbevf/txgbevf_main.c
72
queue = min_t(u16, wx->mac.max_rx_queues, wx->mac.max_tx_queues);
drivers/net/ethernet/wangxun/txgbevf/txgbevf_main.c
73
rss = min_t(u16, queue, rss);
drivers/net/ethernet/xscale/ixp4xx_eth.c
650
static inline int queue_get_desc(unsigned int queue, struct port *port,
drivers/net/ethernet/xscale/ixp4xx_eth.c
656
if (!(phys = qmgr_get_entry(queue)))
drivers/net/ethernet/xscale/ixp4xx_eth.c
669
static inline void queue_put_desc(unsigned int queue, u32 phys,
drivers/net/ethernet/xscale/ixp4xx_eth.c
674
qmgr_put_entry(queue, phys);
drivers/net/fddi/skfp/fplustm.c
270
struct s_smt_rx_queue *queue ;
drivers/net/fddi/skfp/fplustm.c
275
smc->hw.fp.rx[QUEUE_R1] = queue = &smc->hw.fp.rx_q[QUEUE_R1] ;
drivers/net/fddi/skfp/fplustm.c
276
queue->rx_bmu_ctl = (HW_PTR) ADDR(B0_R1_CSR) ;
drivers/net/fddi/skfp/fplustm.c
277
queue->rx_bmu_dsc = (HW_PTR) ADDR(B4_R1_DA) ;
drivers/net/fddi/skfp/fplustm.c
282
smc->hw.fp.rx[QUEUE_R2] = queue = &smc->hw.fp.rx_q[QUEUE_R2] ;
drivers/net/fddi/skfp/fplustm.c
283
queue->rx_bmu_ctl = (HW_PTR) ADDR(B0_R2_CSR) ;
drivers/net/fddi/skfp/fplustm.c
284
queue->rx_bmu_dsc = (HW_PTR) ADDR(B4_R2_DA) ;
drivers/net/fddi/skfp/fplustm.c
300
struct s_smt_tx_queue *queue ;
drivers/net/fddi/skfp/fplustm.c
305
smc->hw.fp.tx[QUEUE_S] = queue = &smc->hw.fp.tx_q[QUEUE_S] ;
drivers/net/fddi/skfp/fplustm.c
306
queue->tx_bmu_ctl = (HW_PTR) ADDR(B0_XS_CSR) ;
drivers/net/fddi/skfp/fplustm.c
307
queue->tx_bmu_dsc = (HW_PTR) ADDR(B5_XS_DA) ;
drivers/net/fddi/skfp/fplustm.c
316
smc->hw.fp.tx[QUEUE_A0] = queue = &smc->hw.fp.tx_q[QUEUE_A0] ;
drivers/net/fddi/skfp/fplustm.c
317
queue->tx_bmu_ctl = (HW_PTR) ADDR(B0_XA_CSR) ;
drivers/net/fddi/skfp/fplustm.c
318
queue->tx_bmu_dsc = (HW_PTR) ADDR(B5_XA_DA) ;
drivers/net/fddi/skfp/fplustm.c
582
void enable_tx_irq(struct s_smc *smc, u_short queue)
drivers/net/fddi/skfp/fplustm.c
589
if (queue == 0) {
drivers/net/fddi/skfp/fplustm.c
592
if (queue == 1) {
drivers/net/fddi/skfp/fplustm.c
617
void disable_tx_irq(struct s_smc *smc, u_short queue)
drivers/net/fddi/skfp/fplustm.c
624
if (queue == 0) {
drivers/net/fddi/skfp/fplustm.c
627
if (queue == 1) {
drivers/net/fddi/skfp/h/hwmtm.h
242
#define HWM_GET_TX_USED(smc,queue) (int) (smc)->hw.fp.tx_q[queue].tx_used
drivers/net/fddi/skfp/h/hwmtm.h
260
#define HWM_GET_CURR_TXD(smc,queue) (struct s_smt_fp_txd volatile *)\
drivers/net/fddi/skfp/h/hwmtm.h
261
(smc)->hw.fp.tx_q[queue].tx_curr_put
drivers/net/fddi/skfp/hwmtm.c
1039
struct s_smt_rx_queue *queue ; /* points to the queue ctl struct */
drivers/net/fddi/skfp/hwmtm.c
1052
queue = smc->hw.fp.rx[QUEUE_R1] ;
drivers/net/fddi/skfp/hwmtm.c
1055
r = queue->rx_curr_get ;
drivers/net/fddi/skfp/hwmtm.c
1056
rx_used = queue->rx_used ;
drivers/net/fddi/skfp/hwmtm.c
1148
rxd = queue->rx_curr_get ;
drivers/net/fddi/skfp/hwmtm.c
1149
queue->rx_curr_get = r ;
drivers/net/fddi/skfp/hwmtm.c
1150
queue->rx_free += frag_count ;
drivers/net/fddi/skfp/hwmtm.c
1151
queue->rx_used = rx_used ;
drivers/net/fddi/skfp/hwmtm.c
1353
DB_RX(3, "next RxD is %p", queue->rx_curr_get);
drivers/net/fddi/skfp/hwmtm.c
1354
NDD_TRACE("RHx1",queue->rx_curr_get,0,0) ;
drivers/net/fddi/skfp/hwmtm.c
1362
DB_RX(3, "next RxD is %p", queue->rx_curr_get);
drivers/net/fddi/skfp/hwmtm.c
1363
NDD_TRACE("RHx2",queue->rx_curr_get,0,0) ;
drivers/net/fddi/skfp/hwmtm.c
1459
struct s_smt_rx_queue *queue ;
drivers/net/fddi/skfp/hwmtm.c
1469
queue = smc->hw.fp.rx[QUEUE_R1] ;
drivers/net/fddi/skfp/hwmtm.c
1475
r = queue->rx_curr_get ;
drivers/net/fddi/skfp/hwmtm.c
1476
while (queue->rx_used) {
drivers/net/fddi/skfp/hwmtm.c
1484
while (r != queue->rx_curr_put &&
drivers/net/fddi/skfp/hwmtm.c
1496
for (r=queue->rx_curr_get,i=frag_count; i ; r=r->rxd_next,i--){
drivers/net/fddi/skfp/hwmtm.c
1502
queue->rx_curr_get, frag_count);
drivers/net/fddi/skfp/hwmtm.c
1503
mac_drv_clear_rxd(smc,queue->rx_curr_get,frag_count) ;
drivers/net/fddi/skfp/hwmtm.c
1505
queue->rx_curr_get = next_rxd ;
drivers/net/fddi/skfp/hwmtm.c
1506
queue->rx_used -= frag_count ;
drivers/net/fddi/skfp/hwmtm.c
1507
queue->rx_free += frag_count ;
drivers/net/fddi/skfp/hwmtm.c
1624
struct s_smt_tx_queue *queue ;
drivers/net/fddi/skfp/hwmtm.c
1627
queue = smc->os.hwm.tx_p ;
drivers/net/fddi/skfp/hwmtm.c
1635
t = queue->tx_curr_put ;
drivers/net/fddi/skfp/hwmtm.c
1651
outpd(queue->tx_bmu_ctl,CSR_START) ;
drivers/net/fddi/skfp/hwmtm.c
1661
queue->tx_free-- ;
drivers/net/fddi/skfp/hwmtm.c
1662
queue->tx_used++ ;
drivers/net/fddi/skfp/hwmtm.c
1663
queue->tx_curr_put = t->txd_next ;
drivers/net/fddi/skfp/hwmtm.c
1719
NDD_TRACE("THfE",t,queue->tx_free,0) ;
drivers/net/fddi/skfp/hwmtm.c
1808
struct s_smt_tx_queue *queue ;
drivers/net/fddi/skfp/hwmtm.c
1844
queue = smc->hw.fp.tx[QUEUE_A0] ;
drivers/net/fddi/skfp/hwmtm.c
1855
if (!smc->hw.mac_ring_is_up || frag_count > queue->tx_free) {
drivers/net/fddi/skfp/hwmtm.c
1873
t = queue->tx_curr_put ;
drivers/net/fddi/skfp/hwmtm.c
1892
outpd(queue->tx_bmu_ctl,CSR_START) ;
drivers/net/fddi/skfp/hwmtm.c
1898
queue->tx_curr_put = t = t->txd_next ;
drivers/net/fddi/skfp/hwmtm.c
1899
queue->tx_free-- ;
drivers/net/fddi/skfp/hwmtm.c
1900
queue->tx_used++ ;
drivers/net/fddi/skfp/hwmtm.c
1916
NDD_TRACE("THSE",t,queue->tx_free,frag_count) ;
drivers/net/fddi/skfp/hwmtm.c
1935
struct s_smt_tx_queue *queue ;
drivers/net/fddi/skfp/hwmtm.c
1946
queue = smc->hw.fp.tx[i] ;
drivers/net/fddi/skfp/hwmtm.c
1947
t1 = queue->tx_curr_get ;
drivers/net/fddi/skfp/hwmtm.c
1958
if (tbctrl & BMU_OWN || !queue->tx_used){
drivers/net/fddi/skfp/hwmtm.c
1966
t1 = queue->tx_curr_get ;
drivers/net/fddi/skfp/hwmtm.c
1987
queue->tx_curr_get);
drivers/net/fddi/skfp/hwmtm.c
1988
mac_drv_tx_complete(smc,queue->tx_curr_get) ;
drivers/net/fddi/skfp/hwmtm.c
1991
queue->tx_curr_get = t1 ;
drivers/net/fddi/skfp/hwmtm.c
1992
queue->tx_free += frag_count ;
drivers/net/fddi/skfp/hwmtm.c
1993
queue->tx_used -= frag_count ;
drivers/net/fddi/skfp/hwmtm.c
2026
struct s_smt_tx_queue *queue ;
drivers/net/fddi/skfp/hwmtm.c
2037
queue = smc->hw.fp.tx[i] ;
drivers/net/fddi/skfp/hwmtm.c
2043
t = queue->tx_curr_get ;
drivers/net/fddi/skfp/hwmtm.c
2044
tx_used = queue->tx_used ;
drivers/net/fddi/skfp/hwmtm.c
2061
queue = smc->hw.fp.tx[i] ;
drivers/net/fddi/skfp/hwmtm.c
2062
t = queue->tx_curr_get ;
drivers/net/fddi/skfp/hwmtm.c
2076
queue->tx_curr_put = queue->tx_curr_get->txd_next ;
drivers/net/fddi/skfp/hwmtm.c
2077
queue->tx_curr_get = queue->tx_curr_put ;
drivers/net/fddi/skfp/hwmtm.c
350
struct s_smt_tx_queue *queue ;
drivers/net/fddi/skfp/hwmtm.c
358
queue = smc->hw.fp.tx[QUEUE_A0] ;
drivers/net/fddi/skfp/hwmtm.c
364
queue->tx_curr_put = queue->tx_curr_get = ds ;
drivers/net/fddi/skfp/hwmtm.c
366
queue->tx_free = HWM_ASYNC_TXD_COUNT ;
drivers/net/fddi/skfp/hwmtm.c
367
queue->tx_used = 0 ;
drivers/net/fddi/skfp/hwmtm.c
372
queue = smc->hw.fp.tx[QUEUE_S] ;
drivers/net/fddi/skfp/hwmtm.c
378
queue->tx_curr_put = queue->tx_curr_get = ds ;
drivers/net/fddi/skfp/hwmtm.c
379
queue->tx_free = HWM_SYNC_TXD_COUNT ;
drivers/net/fddi/skfp/hwmtm.c
380
queue->tx_used = 0 ;
drivers/net/fddi/skfp/hwmtm.c
387
struct s_smt_rx_queue *queue ;
drivers/net/fddi/skfp/hwmtm.c
394
queue = smc->hw.fp.rx[QUEUE_R1] ;
drivers/net/fddi/skfp/hwmtm.c
400
queue->rx_curr_put = queue->rx_curr_get = ds ;
drivers/net/fddi/skfp/hwmtm.c
401
queue->rx_free = SMT_R1_RXD_COUNT ;
drivers/net/fddi/skfp/hwmtm.c
402
queue->rx_used = 0 ;
drivers/net/fddi/skfp/hwmtm.c
578
static u_long repair_txd_ring(struct s_smc *smc, struct s_smt_tx_queue *queue)
drivers/net/fddi/skfp/hwmtm.c
588
t = queue->tx_curr_get ;
drivers/net/fddi/skfp/hwmtm.c
589
tx_used = queue->tx_used ;
drivers/net/fddi/skfp/hwmtm.c
590
for (i = tx_used+queue->tx_free-1 ; i ; i-- ) {
drivers/net/fddi/skfp/hwmtm.c
595
t = queue->tx_curr_get ;
drivers/net/fddi/skfp/hwmtm.c
630
static u_long repair_rxd_ring(struct s_smc *smc, struct s_smt_rx_queue *queue)
drivers/net/fddi/skfp/hwmtm.c
640
r = queue->rx_curr_get ;
drivers/net/fddi/skfp/hwmtm.c
641
rx_used = queue->rx_used ;
drivers/net/fddi/skfp/hwmtm.c
647
r = queue->rx_curr_get ;
drivers/net/fddi/skfp/hwmtm.c
77
static u_long repair_txd_ring(struct s_smc *smc, struct s_smt_tx_queue *queue);
drivers/net/fddi/skfp/hwmtm.c
78
static u_long repair_rxd_ring(struct s_smc *smc, struct s_smt_rx_queue *queue);
drivers/net/fddi/skfp/skfddi.c
1112
int queue;
drivers/net/fddi/skfp/skfddi.c
1131
queue = (fc & FC_SYNC_BIT) ? QUEUE_S : QUEUE_A0;
drivers/net/fddi/skfp/skfddi.c
1149
frame_status = hwm_tx_init(smc, fc, 1, skb->len, queue);
drivers/net/fddi/skfp/skfddi.c
1177
txd = (struct s_smt_fp_txd *) HWM_GET_CURR_TXD(smc, queue);
drivers/net/fddi/skfp/skfddi.c
149
extern void enable_tx_irq(struct s_smc *smc, u_short queue);
drivers/net/fjes/fjes_main.c
839
struct netdev_queue *queue = netdev_get_tx_queue(netdev, 0);
drivers/net/fjes/fjes_main.c
841
netif_tx_wake_queue(queue);
drivers/net/hyperv/netvsc_trace.h
48
__field( u16, queue )
drivers/net/hyperv/netvsc_trace.h
55
__entry->queue = q;
drivers/net/hyperv/netvsc_trace.h
61
__get_str(name), __entry->queue, __entry->req_id,
drivers/net/netdevsim/netdev.c
880
unsigned int queue, mode;
drivers/net/netdevsim/netdev.c
890
ret = sscanf(buf, "%u %u", &queue, &mode);
drivers/net/netdevsim/netdev.c
895
if (queue >= ns->netdev->real_num_rx_queues) {
drivers/net/netdevsim/netdev.c
901
ret = netdev_rx_queue_restart(ns->netdev, queue);
drivers/net/ovpn/tcp.c
370
struct sk_buff_head queue;
drivers/net/ovpn/tcp.c
393
__skb_queue_head_init(&queue);
drivers/net/ovpn/tcp.c
394
skb_queue_splice_init(&peer->tcp.out_queue, &queue);
drivers/net/ovpn/tcp.c
396
while ((skb = __skb_dequeue(&queue)))
drivers/net/ovpn/tcp.c
599
struct sk_buff_head *queue = &sock->sk->sk_receive_queue;
drivers/net/ovpn/tcp.c
618
queue = &peer->tcp.user_queue;
drivers/net/ovpn/tcp.c
622
mask = datagram_poll_queue(file, sock, wait, queue);
drivers/net/tap.c
192
struct tap_queue *queue = NULL;
drivers/net/tap.c
210
queue = rcu_dereference(tap->taps[rxq % numvtaps]);
drivers/net/tap.c
220
queue = rcu_dereference(tap->taps[rxq]);
drivers/net/tap.c
225
queue = rcu_dereference(tap->taps[0]);
drivers/net/tap.c
227
return queue;
drivers/net/tun.c
1008
struct netdev_queue *queue;
drivers/net/tun.c
1069
queue = netdev_get_tx_queue(dev, txq);
drivers/net/tun.c
1070
txq_trans_cond_update(queue);
drivers/net/tun.c
1477
struct sk_buff_head *queue = &tfile->sk.sk_write_queue;
drivers/net/tun.c
1482
if (!rx_batched || (!more && skb_queue_empty(queue))) {
drivers/net/tun.c
1490
spin_lock(&queue->lock);
drivers/net/tun.c
1491
if (!more || skb_queue_len(queue) == rx_batched) {
drivers/net/tun.c
1493
skb_queue_splice_tail_init(queue, &process_queue);
drivers/net/tun.c
1496
__skb_queue_tail(queue, skb);
drivers/net/tun.c
1498
spin_unlock(&queue->lock);
drivers/net/tun.c
1932
struct sk_buff_head *queue = &tfile->sk.sk_write_queue;
drivers/net/tun.c
1935
spin_lock_bh(&queue->lock);
drivers/net/tun.c
1938
spin_unlock_bh(&queue->lock);
drivers/net/tun.c
1944
__skb_queue_tail(queue, skb);
drivers/net/tun.c
1945
queue_len = skb_queue_len(queue);
drivers/net/tun.c
1946
spin_unlock(&queue->lock);
drivers/net/tun.c
230
struct sk_buff_head *queue = &tfile->sk.sk_write_queue;
drivers/net/tun.c
237
spin_lock(&queue->lock);
drivers/net/tun.c
238
skb_queue_splice_tail_init(queue, &process_queue);
drivers/net/tun.c
2386
struct sk_buff_head *queue;
drivers/net/tun.c
239
spin_unlock(&queue->lock);
drivers/net/tun.c
247
spin_lock(&queue->lock);
drivers/net/tun.c
248
skb_queue_splice(&process_queue, queue);
drivers/net/tun.c
2480
queue = &tfile->sk.sk_write_queue;
drivers/net/tun.c
2481
spin_lock(&queue->lock);
drivers/net/tun.c
2484
spin_unlock(&queue->lock);
drivers/net/tun.c
2489
__skb_queue_tail(queue, skb);
drivers/net/tun.c
249
spin_unlock(&queue->lock);
drivers/net/tun.c
2490
spin_unlock(&queue->lock);
drivers/net/virtio_net.c
4052
u16 queue, u32 max_usecs,
drivers/net/virtio_net.c
4060
err = virtnet_send_ctrl_coal_vq_cmd(vi, rxq2vq(queue),
drivers/net/virtio_net.c
4065
vi->rq[queue].intr_coal.max_usecs = max_usecs;
drivers/net/virtio_net.c
4066
vi->rq[queue].intr_coal.max_packets = max_packets;
drivers/net/virtio_net.c
4072
u16 queue, u32 max_usecs,
drivers/net/virtio_net.c
4080
err = virtnet_send_ctrl_coal_vq_cmd(vi, txq2vq(queue),
drivers/net/virtio_net.c
4085
vi->sq[queue].intr_coal.max_usecs = max_usecs;
drivers/net/virtio_net.c
4086
vi->sq[queue].intr_coal.max_packets = max_packets;
drivers/net/virtio_net.c
5237
u16 queue)
drivers/net/virtio_net.c
5244
mutex_lock(&vi->rq[queue].dim_lock);
drivers/net/virtio_net.c
5245
cur_rx_dim = vi->rq[queue].dim_enabled;
drivers/net/virtio_net.c
5246
max_usecs = vi->rq[queue].intr_coal.max_usecs;
drivers/net/virtio_net.c
5247
max_packets = vi->rq[queue].intr_coal.max_packets;
drivers/net/virtio_net.c
5251
mutex_unlock(&vi->rq[queue].dim_lock);
drivers/net/virtio_net.c
5256
vi->rq[queue].dim_enabled = true;
drivers/net/virtio_net.c
5257
mutex_unlock(&vi->rq[queue].dim_lock);
drivers/net/virtio_net.c
5262
vi->rq[queue].dim_enabled = false;
drivers/net/virtio_net.c
5267
err = virtnet_send_rx_ctrl_coal_vq_cmd(vi, queue,
drivers/net/virtio_net.c
5270
mutex_unlock(&vi->rq[queue].dim_lock);
drivers/net/virtio_net.c
5276
u16 queue)
drivers/net/virtio_net.c
5280
err = virtnet_send_rx_notf_coal_vq_cmds(vi, ec, queue);
drivers/net/virtio_net.c
5284
err = virtnet_send_tx_ctrl_coal_vq_cmd(vi, queue,
drivers/net/virtio_net.c
5425
u32 queue,
drivers/net/virtio_net.c
5432
if (queue >= vi->max_queue_pairs)
drivers/net/virtio_net.c
5438
vi->sq[queue].napi.weight,
drivers/net/virtio_net.c
5444
ret = virtnet_send_notf_coal_vq_cmds(vi, ec, queue);
drivers/net/virtio_net.c
5452
vi->sq[queue].napi.weight = napi_weight;
drivers/net/virtio_net.c
5458
u32 queue,
drivers/net/virtio_net.c
5463
if (queue >= vi->max_queue_pairs)
drivers/net/virtio_net.c
5467
mutex_lock(&vi->rq[queue].dim_lock);
drivers/net/virtio_net.c
5468
ec->rx_coalesce_usecs = vi->rq[queue].intr_coal.max_usecs;
drivers/net/virtio_net.c
5469
ec->tx_coalesce_usecs = vi->sq[queue].intr_coal.max_usecs;
drivers/net/virtio_net.c
5470
ec->tx_max_coalesced_frames = vi->sq[queue].intr_coal.max_packets;
drivers/net/virtio_net.c
5471
ec->rx_max_coalesced_frames = vi->rq[queue].intr_coal.max_packets;
drivers/net/virtio_net.c
5472
ec->use_adaptive_rx_coalesce = vi->rq[queue].dim_enabled;
drivers/net/virtio_net.c
5473
mutex_unlock(&vi->rq[queue].dim_lock);
drivers/net/virtio_net.c
5477
if (vi->sq[queue].napi.weight)
drivers/net/virtio_net.c
6533
static ssize_t mergeable_rx_buffer_size_show(struct netdev_rx_queue *queue,
drivers/net/virtio_net.c
6536
struct virtnet_info *vi = netdev_priv(queue->dev);
drivers/net/virtio_net.c
6537
unsigned int queue_index = get_netdev_rx_queue_index(queue);
drivers/net/wan/farsync.c
576
fst_q_work_item(u64 *queue, int card_index)
drivers/net/wan/farsync.c
591
*queue |= mask;
drivers/net/wan/ixp4xx_hss.c
588
static inline int queue_get_desc(unsigned int queue, struct port *port,
drivers/net/wan/ixp4xx_hss.c
594
phys = qmgr_get_entry(queue);
drivers/net/wan/ixp4xx_hss.c
608
static inline void queue_put_desc(unsigned int queue, u32 phys,
drivers/net/wan/ixp4xx_hss.c
613
qmgr_put_entry(queue, phys);
drivers/net/wireguard/queueing.c
101
queue->tail = next;
drivers/net/wireguard/queueing.c
102
atomic_dec(&queue->count);
drivers/net/wireguard/queueing.c
25
int wg_packet_queue_init(struct crypt_queue *queue, work_func_t function,
drivers/net/wireguard/queueing.c
30
memset(queue, 0, sizeof(*queue));
drivers/net/wireguard/queueing.c
31
queue->last_cpu = -1;
drivers/net/wireguard/queueing.c
32
ret = ptr_ring_init(&queue->ring, len, GFP_KERNEL);
drivers/net/wireguard/queueing.c
35
queue->worker = wg_packet_percpu_multicore_worker_alloc(function, queue);
drivers/net/wireguard/queueing.c
36
if (!queue->worker) {
drivers/net/wireguard/queueing.c
37
ptr_ring_cleanup(&queue->ring, NULL);
drivers/net/wireguard/queueing.c
43
void wg_packet_queue_free(struct crypt_queue *queue, bool purge)
drivers/net/wireguard/queueing.c
45
free_percpu(queue->worker);
drivers/net/wireguard/queueing.c
46
WARN_ON(!purge && !__ptr_ring_empty(&queue->ring));
drivers/net/wireguard/queueing.c
47
ptr_ring_cleanup(&queue->ring, purge ? __skb_array_destroy_skb : NULL);
drivers/net/wireguard/queueing.c
51
#define STUB(queue) ((struct sk_buff *)&queue->empty)
drivers/net/wireguard/queueing.c
53
void wg_prev_queue_init(struct prev_queue *queue)
drivers/net/wireguard/queueing.c
55
NEXT(STUB(queue)) = NULL;
drivers/net/wireguard/queueing.c
56
queue->head = queue->tail = STUB(queue);
drivers/net/wireguard/queueing.c
57
queue->peeked = NULL;
drivers/net/wireguard/queueing.c
58
atomic_set(&queue->count, 0);
drivers/net/wireguard/queueing.c
66
static void __wg_prev_queue_enqueue(struct prev_queue *queue, struct sk_buff *skb)
drivers/net/wireguard/queueing.c
69
WRITE_ONCE(NEXT(xchg_release(&queue->head, skb)), skb);
drivers/net/wireguard/queueing.c
72
bool wg_prev_queue_enqueue(struct prev_queue *queue, struct sk_buff *skb)
drivers/net/wireguard/queueing.c
74
if (!atomic_add_unless(&queue->count, 1, MAX_QUEUED_PACKETS))
drivers/net/wireguard/queueing.c
76
__wg_prev_queue_enqueue(queue, skb);
drivers/net/wireguard/queueing.c
80
struct sk_buff *wg_prev_queue_dequeue(struct prev_queue *queue)
drivers/net/wireguard/queueing.c
82
struct sk_buff *tail = queue->tail, *next = smp_load_acquire(&NEXT(tail));
drivers/net/wireguard/queueing.c
84
if (tail == STUB(queue)) {
drivers/net/wireguard/queueing.c
87
queue->tail = next;
drivers/net/wireguard/queueing.c
92
queue->tail = next;
drivers/net/wireguard/queueing.c
93
atomic_dec(&queue->count);
drivers/net/wireguard/queueing.c
96
if (tail != READ_ONCE(queue->head))
drivers/net/wireguard/queueing.c
98
__wg_prev_queue_enqueue(queue, STUB(queue));
drivers/net/wireguard/queueing.h
129
void wg_prev_queue_init(struct prev_queue *queue);
drivers/net/wireguard/queueing.h
132
bool wg_prev_queue_enqueue(struct prev_queue *queue, struct sk_buff *skb);
drivers/net/wireguard/queueing.h
135
struct sk_buff *wg_prev_queue_dequeue(struct prev_queue *queue);
drivers/net/wireguard/queueing.h
138
static inline struct sk_buff *wg_prev_queue_peek(struct prev_queue *queue)
drivers/net/wireguard/queueing.h
140
if (queue->peeked)
drivers/net/wireguard/queueing.h
141
return queue->peeked;
drivers/net/wireguard/queueing.h
142
queue->peeked = wg_prev_queue_dequeue(queue);
drivers/net/wireguard/queueing.h
143
return queue->peeked;
drivers/net/wireguard/queueing.h
147
static inline void wg_prev_queue_drop_peeked(struct prev_queue *queue)
drivers/net/wireguard/queueing.h
149
queue->peeked = NULL;
drivers/net/wireguard/queueing.h
24
int wg_packet_queue_init(struct crypt_queue *queue, work_func_t function,
drivers/net/wireguard/queueing.h
26
void wg_packet_queue_free(struct crypt_queue *queue, bool purge);
drivers/net/wireguard/receive.c
208
struct crypt_queue *queue = container_of(work, struct multicore_worker, work)->ptr;
drivers/net/wireguard/receive.c
209
struct wg_device *wg = container_of(queue, struct wg_device, handshake_queue);
drivers/net/wireguard/receive.c
212
while ((skb = ptr_ring_consume_bh(&queue->ring)) != NULL) {
drivers/net/wireguard/receive.c
495
struct crypt_queue *queue = container_of(work, struct multicore_worker,
drivers/net/wireguard/receive.c
499
while ((skb = ptr_ring_consume_bh(&queue->ring)) != NULL) {
drivers/net/wireguard/send.c
289
struct crypt_queue *queue = container_of(work, struct multicore_worker,
drivers/net/wireguard/send.c
293
while ((first = ptr_ring_consume_bh(&queue->ring)) != NULL) {
drivers/net/wireless/ath/ath5k/ath5k.h
1508
int ath5k_hw_start_tx_dma(struct ath5k_hw *ah, unsigned int queue);
drivers/net/wireless/ath/ath5k/ath5k.h
1509
int ath5k_hw_stop_beacon_queue(struct ath5k_hw *ah, unsigned int queue);
drivers/net/wireless/ath/ath5k/ath5k.h
1510
u32 ath5k_hw_get_txdp(struct ath5k_hw *ah, unsigned int queue);
drivers/net/wireless/ath/ath5k/ath5k.h
1511
int ath5k_hw_set_txdp(struct ath5k_hw *ah, unsigned int queue,
drivers/net/wireless/ath/ath5k/ath5k.h
1558
int ath5k_hw_get_tx_queueprops(struct ath5k_hw *ah, int queue,
drivers/net/wireless/ath/ath5k/ath5k.h
1560
int ath5k_hw_set_tx_queueprops(struct ath5k_hw *ah, int queue,
drivers/net/wireless/ath/ath5k/ath5k.h
1566
unsigned int queue);
drivers/net/wireless/ath/ath5k/ath5k.h
1567
u32 ath5k_hw_num_tx_pending(struct ath5k_hw *ah, unsigned int queue);
drivers/net/wireless/ath/ath5k/ath5k.h
1568
void ath5k_hw_release_tx_queue(struct ath5k_hw *ah, unsigned int queue);
drivers/net/wireless/ath/ath5k/ath5k.h
1569
int ath5k_hw_reset_tx_queue(struct ath5k_hw *ah, unsigned int queue);
drivers/net/wireless/ath/ath5k/dma.c
130
ath5k_hw_start_tx_dma(struct ath5k_hw *ah, unsigned int queue)
drivers/net/wireless/ath/ath5k/dma.c
134
AR5K_ASSERT_ENTRY(queue, ah->ah_capabilities.cap_queues.q_tx_num);
drivers/net/wireless/ath/ath5k/dma.c
137
if (ah->ah_txq[queue].tqi_type == AR5K_TX_QUEUE_INACTIVE)
drivers/net/wireless/ath/ath5k/dma.c
146
switch (ah->ah_txq[queue].tqi_type) {
drivers/net/wireless/ath/ath5k/dma.c
168
if (AR5K_REG_READ_Q(ah, AR5K_QCU_TXD, queue))
drivers/net/wireless/ath/ath5k/dma.c
172
AR5K_REG_WRITE_Q(ah, AR5K_QCU_TXE, queue);
drivers/net/wireless/ath/ath5k/dma.c
188
ath5k_hw_stop_tx_dma(struct ath5k_hw *ah, unsigned int queue)
drivers/net/wireless/ath/ath5k/dma.c
193
AR5K_ASSERT_ENTRY(queue, ah->ah_capabilities.cap_queues.q_tx_num);
drivers/net/wireless/ath/ath5k/dma.c
196
if (ah->ah_txq[queue].tqi_type == AR5K_TX_QUEUE_INACTIVE)
drivers/net/wireless/ath/ath5k/dma.c
205
switch (ah->ah_txq[queue].tqi_type) {
drivers/net/wireless/ath/ath5k/dma.c
228
AR5K_REG_ENABLE_BITS(ah, AR5K_QUEUE_MISC(queue),
drivers/net/wireless/ath/ath5k/dma.c
234
AR5K_REG_WRITE_Q(ah, AR5K_QCU_TXD, queue);
drivers/net/wireless/ath/ath5k/dma.c
238
(AR5K_REG_READ_Q(ah, AR5K_QCU_TXE, queue) != 0);
drivers/net/wireless/ath/ath5k/dma.c
242
if (AR5K_REG_READ_Q(ah, AR5K_QCU_TXE, queue))
drivers/net/wireless/ath/ath5k/dma.c
244
"queue %i didn't stop !\n", queue);
drivers/net/wireless/ath/ath5k/dma.c
250
AR5K_QUEUE_STATUS(queue)) &
drivers/net/wireless/ath/ath5k/dma.c
286
AR5K_QUEUE_STATUS(queue)) &
drivers/net/wireless/ath/ath5k/dma.c
297
queue);
drivers/net/wireless/ath/ath5k/dma.c
303
AR5K_REG_DISABLE_BITS(ah, AR5K_QUEUE_MISC(queue),
drivers/net/wireless/ath/ath5k/dma.c
311
queue, pending);
drivers/net/wireless/ath/ath5k/dma.c
328
ath5k_hw_stop_beacon_queue(struct ath5k_hw *ah, unsigned int queue)
drivers/net/wireless/ath/ath5k/dma.c
331
ret = ath5k_hw_stop_tx_dma(ah, queue);
drivers/net/wireless/ath/ath5k/dma.c
353
ath5k_hw_get_txdp(struct ath5k_hw *ah, unsigned int queue)
drivers/net/wireless/ath/ath5k/dma.c
357
AR5K_ASSERT_ENTRY(queue, ah->ah_capabilities.cap_queues.q_tx_num);
drivers/net/wireless/ath/ath5k/dma.c
364
switch (ah->ah_txq[queue].tqi_type) {
drivers/net/wireless/ath/ath5k/dma.c
376
tx_reg = AR5K_QUEUE_TXDP(queue);
drivers/net/wireless/ath/ath5k/dma.c
396
ath5k_hw_set_txdp(struct ath5k_hw *ah, unsigned int queue, u32 phys_addr)
drivers/net/wireless/ath/ath5k/dma.c
400
AR5K_ASSERT_ENTRY(queue, ah->ah_capabilities.cap_queues.q_tx_num);
drivers/net/wireless/ath/ath5k/dma.c
407
switch (ah->ah_txq[queue].tqi_type) {
drivers/net/wireless/ath/ath5k/dma.c
424
if (AR5K_REG_READ_Q(ah, AR5K_QCU_TXE, queue))
drivers/net/wireless/ath/ath5k/dma.c
427
tx_reg = AR5K_QUEUE_TXDP(queue);
drivers/net/wireless/ath/ath5k/mac80211-ops.c
575
unsigned int link_id, u16 queue,
drivers/net/wireless/ath/ath5k/mac80211-ops.c
582
if (queue >= ah->ah_capabilities.cap_queues.q_tx_num)
drivers/net/wireless/ath/ath5k/mac80211-ops.c
587
ath5k_hw_get_tx_queueprops(ah, queue, &qi);
drivers/net/wireless/ath/ath5k/mac80211-ops.c
597
queue, params->aifs, params->cw_min,
drivers/net/wireless/ath/ath5k/mac80211-ops.c
600
if (ath5k_hw_set_tx_queueprops(ah, queue, &qi)) {
drivers/net/wireless/ath/ath5k/mac80211-ops.c
602
"Unable to update hardware queue %u!\n", queue);
drivers/net/wireless/ath/ath5k/mac80211-ops.c
605
ath5k_hw_reset_tx_queue(ah, queue);
drivers/net/wireless/ath/ath5k/qcu.c
100
ah->ah_txq[queue].tqi_type = AR5K_TX_QUEUE_INACTIVE;
drivers/net/wireless/ath/ath5k/qcu.c
102
AR5K_Q_DISABLE_BITS(ah->ah_txq_status, queue);
drivers/net/wireless/ath/ath5k/qcu.c
138
ath5k_hw_get_tx_queueprops(struct ath5k_hw *ah, int queue,
drivers/net/wireless/ath/ath5k/qcu.c
141
memcpy(queue_info, &ah->ah_txq[queue], sizeof(struct ath5k_txq_info));
drivers/net/wireless/ath/ath5k/qcu.c
154
ath5k_hw_set_tx_queueprops(struct ath5k_hw *ah, int queue,
drivers/net/wireless/ath/ath5k/qcu.c
159
AR5K_ASSERT_ENTRY(queue, ah->ah_capabilities.cap_queues.q_tx_num);
drivers/net/wireless/ath/ath5k/qcu.c
161
qi = &ah->ah_txq[queue];
drivers/net/wireless/ath/ath5k/qcu.c
206
unsigned int queue;
drivers/net/wireless/ath/ath5k/qcu.c
216
queue = AR5K_TX_QUEUE_ID_NOQCU_DATA;
drivers/net/wireless/ath/ath5k/qcu.c
220
queue = AR5K_TX_QUEUE_ID_NOQCU_BEACON;
drivers/net/wireless/ath/ath5k/qcu.c
228
queue = queue_info->tqi_subtype;
drivers/net/wireless/ath/ath5k/qcu.c
231
queue = AR5K_TX_QUEUE_ID_UAPSD;
drivers/net/wireless/ath/ath5k/qcu.c
234
queue = AR5K_TX_QUEUE_ID_BEACON;
drivers/net/wireless/ath/ath5k/qcu.c
237
queue = AR5K_TX_QUEUE_ID_CAB;
drivers/net/wireless/ath/ath5k/qcu.c
247
memset(&ah->ah_txq[queue], 0, sizeof(struct ath5k_txq_info));
drivers/net/wireless/ath/ath5k/qcu.c
248
ah->ah_txq[queue].tqi_type = queue_type;
drivers/net/wireless/ath/ath5k/qcu.c
252
ret = ath5k_hw_set_tx_queueprops(ah, queue, queue_info);
drivers/net/wireless/ath/ath5k/qcu.c
262
AR5K_Q_ENABLE_BITS(ah->ah_txq_status, queue);
drivers/net/wireless/ath/ath5k/qcu.c
264
return queue;
drivers/net/wireless/ath/ath5k/qcu.c
282
unsigned int queue)
drivers/net/wireless/ath/ath5k/qcu.c
286
struct ath5k_txq_info *tq = &ah->ah_txq[queue];
drivers/net/wireless/ath/ath5k/qcu.c
288
if (queue > 0)
drivers/net/wireless/ath/ath5k/qcu.c
311
AR5K_QUEUE_DFS_RETRY_LIMIT(queue));
drivers/net/wireless/ath/ath5k/qcu.c
324
ath5k_hw_reset_tx_queue(struct ath5k_hw *ah, unsigned int queue)
drivers/net/wireless/ath/ath5k/qcu.c
326
struct ath5k_txq_info *tq = &ah->ah_txq[queue];
drivers/net/wireless/ath/ath5k/qcu.c
328
AR5K_ASSERT_ENTRY(queue, ah->ah_capabilities.cap_queues.q_tx_num);
drivers/net/wireless/ath/ath5k/qcu.c
344
AR5K_QUEUE_DFS_LOCAL_IFS(queue));
drivers/net/wireless/ath/ath5k/qcu.c
349
ath5k_hw_set_tx_retry_limits(ah, queue);
drivers/net/wireless/ath/ath5k/qcu.c
357
AR5K_REG_ENABLE_BITS(ah, AR5K_QUEUE_DFS_MISC(queue),
drivers/net/wireless/ath/ath5k/qcu.c
362
AR5K_REG_ENABLE_BITS(ah, AR5K_QUEUE_DFS_MISC(queue),
drivers/net/wireless/ath/ath5k/qcu.c
371
AR5K_QUEUE_CBRCFG(queue));
drivers/net/wireless/ath/ath5k/qcu.c
373
AR5K_REG_ENABLE_BITS(ah, AR5K_QUEUE_MISC(queue),
drivers/net/wireless/ath/ath5k/qcu.c
377
AR5K_REG_ENABLE_BITS(ah, AR5K_QUEUE_MISC(queue),
drivers/net/wireless/ath/ath5k/qcu.c
386
AR5K_QUEUE_RDYTIMECFG(queue));
drivers/net/wireless/ath/ath5k/qcu.c
392
AR5K_QUEUE_DFS_CHANNEL_TIME(queue));
drivers/net/wireless/ath/ath5k/qcu.c
395
AR5K_REG_ENABLE_BITS(ah, AR5K_QUEUE_MISC(queue),
drivers/net/wireless/ath/ath5k/qcu.c
402
AR5K_QUEUE_DFS_MISC(queue));
drivers/net/wireless/ath/ath5k/qcu.c
407
AR5K_QUEUE_DFS_MISC(queue));
drivers/net/wireless/ath/ath5k/qcu.c
414
AR5K_REG_ENABLE_BITS(ah, AR5K_QUEUE_MISC(queue),
drivers/net/wireless/ath/ath5k/qcu.c
419
AR5K_REG_ENABLE_BITS(ah, AR5K_QUEUE_DFS_MISC(queue),
drivers/net/wireless/ath/ath5k/qcu.c
429
AR5K_REG_ENABLE_BITS(ah, AR5K_QUEUE_MISC(queue),
drivers/net/wireless/ath/ath5k/qcu.c
439
AR5K_QUEUE_RDYTIMECFG(queue));
drivers/net/wireless/ath/ath5k/qcu.c
441
AR5K_REG_ENABLE_BITS(ah, AR5K_QUEUE_DFS_MISC(queue),
drivers/net/wireless/ath/ath5k/qcu.c
447
AR5K_REG_ENABLE_BITS(ah, AR5K_QUEUE_MISC(queue),
drivers/net/wireless/ath/ath5k/qcu.c
463
AR5K_Q_ENABLE_BITS(ah->ah_txq_imr_txok, queue);
drivers/net/wireless/ath/ath5k/qcu.c
466
AR5K_Q_ENABLE_BITS(ah->ah_txq_imr_txerr, queue);
drivers/net/wireless/ath/ath5k/qcu.c
469
AR5K_Q_ENABLE_BITS(ah->ah_txq_imr_txurn, queue);
drivers/net/wireless/ath/ath5k/qcu.c
472
AR5K_Q_ENABLE_BITS(ah->ah_txq_imr_txdesc, queue);
drivers/net/wireless/ath/ath5k/qcu.c
475
AR5K_Q_ENABLE_BITS(ah->ah_txq_imr_txeol, queue);
drivers/net/wireless/ath/ath5k/qcu.c
478
AR5K_Q_ENABLE_BITS(ah->ah_txq_imr_cbrorn, queue);
drivers/net/wireless/ath/ath5k/qcu.c
481
AR5K_Q_ENABLE_BITS(ah->ah_txq_imr_cbrurn, queue);
drivers/net/wireless/ath/ath5k/qcu.c
484
AR5K_Q_ENABLE_BITS(ah->ah_txq_imr_qtrig, queue);
drivers/net/wireless/ath/ath5k/qcu.c
487
AR5K_Q_ENABLE_BITS(ah->ah_txq_imr_nofrm, queue);
drivers/net/wireless/ath/ath5k/qcu.c
539
AR5K_REG_WRITE_Q(ah, AR5K_QUEUE_QCUMASK(queue), queue);
drivers/net/wireless/ath/ath5k/qcu.c
63
ath5k_hw_num_tx_pending(struct ath5k_hw *ah, unsigned int queue)
drivers/net/wireless/ath/ath5k/qcu.c
66
AR5K_ASSERT_ENTRY(queue, ah->ah_capabilities.cap_queues.q_tx_num);
drivers/net/wireless/ath/ath5k/qcu.c
69
if (ah->ah_txq[queue].tqi_type == AR5K_TX_QUEUE_INACTIVE)
drivers/net/wireless/ath/ath5k/qcu.c
76
pending = ath5k_hw_reg_read(ah, AR5K_QUEUE_STATUS(queue));
drivers/net/wireless/ath/ath5k/qcu.c
82
if (!pending && AR5K_REG_READ_Q(ah, AR5K_QCU_TXE, queue))
drivers/net/wireless/ath/ath5k/qcu.c
94
ath5k_hw_release_tx_queue(struct ath5k_hw *ah, unsigned int queue)
drivers/net/wireless/ath/ath5k/qcu.c
96
if (WARN_ON(queue >= ah->ah_capabilities.cap_queues.q_tx_num))
drivers/net/wireless/ath/ath6kl/htc.h
670
static inline int get_queue_depth(struct list_head *queue)
drivers/net/wireless/ath/ath6kl/htc.h
675
list_for_each(tmp_list, queue)
drivers/net/wireless/ath/ath6kl/htc_mbox.c
1124
struct list_head queue;
drivers/net/wireless/ath/ath6kl/htc_mbox.c
1140
INIT_LIST_HEAD(&queue);
drivers/net/wireless/ath/ath6kl/htc_mbox.c
1141
list_add(&packet->list, &queue);
drivers/net/wireless/ath/ath6kl/htc_mbox.c
1142
htc_tx_complete(endpoint, &queue);
drivers/net/wireless/ath/ath6kl/htc_mbox.c
1270
struct list_head queue;
drivers/net/wireless/ath/ath6kl/htc_mbox.c
1272
INIT_LIST_HEAD(&queue);
drivers/net/wireless/ath/ath6kl/htc_mbox.c
1273
list_add_tail(&packet->list, &queue);
drivers/net/wireless/ath/ath6kl/htc_mbox.c
1274
return ath6kl_htc_mbox_add_rxbuf_multiple(target, &queue);
drivers/net/wireless/ath/ath6kl/htc_mbox.c
1367
u32 *lk_ahds, struct list_head *queue, int n_msg)
drivers/net/wireless/ath/ath6kl/htc_mbox.c
1447
list_add_tail(&packet->list, queue);
drivers/net/wireless/ath/ath6kl/htc_mbox.c
1471
struct list_head *queue)
drivers/net/wireless/ath/ath6kl/htc_mbox.c
1537
queue, n_msg);
drivers/net/wireless/ath/ath6kl/htc_mbox.c
1556
list_for_each_entry_safe(packet, tmp_pkt, queue, list) {
drivers/net/wireless/ath/ath6kl/htc_mbox.c
600
struct list_head *queue)
drivers/net/wireless/ath/ath6kl/htc_mbox.c
629
list_move_tail(&packet->list, queue);
drivers/net/wireless/ath/ath6kl/htc_mbox.c
682
struct list_head *queue)
drivers/net/wireless/ath/ath6kl/htc_mbox.c
694
if (list_empty(queue))
drivers/net/wireless/ath/ath6kl/htc_mbox.c
697
packet = list_first_entry(queue, struct htc_packet, list);
drivers/net/wireless/ath/ath6kl/htc_mbox.c
736
list_add(&packet->list, queue);
drivers/net/wireless/ath/ath6kl/htc_mbox.c
755
struct list_head *queue,
drivers/net/wireless/ath/ath6kl/htc_mbox.c
772
n_scat = get_queue_depth(queue);
drivers/net/wireless/ath/ath6kl/htc_mbox.c
822
queue);
drivers/net/wireless/ath/ath6kl/htc_pipe.c
1541
struct list_head queue;
drivers/net/wireless/ath/ath6kl/htc_pipe.c
1548
INIT_LIST_HEAD(&queue);
drivers/net/wireless/ath/ath6kl/htc_pipe.c
1549
list_add_tail(&packet->list, &queue);
drivers/net/wireless/ath/ath6kl/htc_pipe.c
1551
return htc_send_packets_multiple(target, &queue);
drivers/net/wireless/ath/ath6kl/htc_pipe.c
169
list_add_tail(&packet->list, queue);
drivers/net/wireless/ath/ath6kl/htc_pipe.c
175
struct list_head *queue, int resources)
drivers/net/wireless/ath/ath6kl/htc_pipe.c
198
list_add_tail(&packet->list, queue);
drivers/net/wireless/ath/ath6kl/htc_pipe.c
93
struct list_head *queue)
drivers/net/wireless/ath/ath6kl/txrx.c
881
struct list_head queue;
drivers/net/wireless/ath/ath6kl/txrx.c
889
INIT_LIST_HEAD(&queue);
drivers/net/wireless/ath/ath6kl/txrx.c
909
list_add_tail(&packet->list, &queue);
drivers/net/wireless/ath/ath6kl/txrx.c
912
if (!list_empty(&queue))
drivers/net/wireless/ath/ath6kl/txrx.c
913
ath6kl_htc_add_rxbuf_multiple(ar->htc_target, &queue);
drivers/net/wireless/ath/ath9k/ar9003_hw.c
1098
static bool ath9k_hw_verify_hang(struct ath_hw *ah, unsigned int queue)
drivers/net/wireless/ath/ath9k/ar9003_hw.c
1105
if (queue < 6) {
drivers/net/wireless/ath/ath9k/ar9003_hw.c
1107
reg_offset = queue * 5;
drivers/net/wireless/ath/ath9k/ar9003_hw.c
1110
reg_offset = (queue - 6) * 5;
drivers/net/wireless/ath/ath9k/ar9003_hw.c
1125
"MAC Hang signature found for queue: %d\n", queue);
drivers/net/wireless/ath/ath9k/ar9003_mci.c
1081
u32 *payload, bool queue)
drivers/net/wireless/ath/ath9k/ar9003_mci.c
1102
mci->update_2g5g = queue;
drivers/net/wireless/ath/ath9k/ar9003_mci.c
1106
mci->wlan_channels_update = queue;
drivers/net/wireless/ath/ath9k/ar9003_mci.c
1111
mci->unhalt_bt_gpm = queue;
drivers/net/wireless/ath/ath9k/ar9003_mci.c
1113
if (!queue)
drivers/net/wireless/ath/ath9k/ar9003_mci.c
1120
mci->halted_bt_gpm = !queue;
drivers/net/wireless/ath/ath9k/ath9k.h
596
void ath_update_max_aggr_framelen(struct ath_softc *sc, int queue, int txop);
drivers/net/wireless/ath/ath9k/ath9k.h
616
void ath9k_wake_tx_queue(struct ieee80211_hw *hw, struct ieee80211_txq *queue);
drivers/net/wireless/ath/ath9k/hif_usb.c
233
struct sk_buff_head *queue,
drivers/net/wireless/ath/ath9k/hif_usb.c
238
while ((skb = __skb_dequeue(queue)) != NULL) {
drivers/net/wireless/ath/ath9k/htc.h
583
int get_hw_qnum(u16 queue, int *hwq_map);
drivers/net/wireless/ath/ath9k/htc_drv_main.c
1374
unsigned int link_id, u16 queue,
drivers/net/wireless/ath/ath9k/htc_drv_main.c
1382
if (queue >= IEEE80211_NUM_ACS)
drivers/net/wireless/ath/ath9k/htc_drv_main.c
1395
qnum = get_hw_qnum(queue, priv->hwq_map);
drivers/net/wireless/ath/ath9k/htc_drv_main.c
1399
queue, qnum, params->aifs, params->cw_min,
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
40
int get_hw_qnum(u16 queue, int *hwq_map)
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
42
switch (queue) {
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
530
struct sk_buff_head *queue)
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
534
while ((skb = skb_dequeue(queue)) != NULL) {
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
739
struct sk_buff_head queue;
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
741
skb_queue_head_init(&queue);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
747
__skb_queue_tail(&queue, skb);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
754
skb_queue_walk_safe(&queue, skb, tmp) {
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
755
__skb_unlink(skb, &queue);
drivers/net/wireless/ath/ath9k/main.c
1715
unsigned int link_id, u16 queue,
drivers/net/wireless/ath/ath9k/main.c
1724
if (queue >= IEEE80211_NUM_ACS)
drivers/net/wireless/ath/ath9k/main.c
1727
txq = sc->tx.txq_map[queue];
drivers/net/wireless/ath/ath9k/main.c
1741
queue, txq->axq_qnum, params->aifs, params->cw_min,
drivers/net/wireless/ath/ath9k/main.c
1744
ath_update_max_aggr_framelen(sc, queue, qi.tqi_burstTime);
drivers/net/wireless/ath/ath9k/xmit.c
1153
void ath_update_max_aggr_framelen(struct ath_softc *sc, int queue, int txop)
drivers/net/wireless/ath/ath9k/xmit.c
1162
cur_ht20 = sc->tx.max_aggr_framelen[queue][MCS_HT20];
drivers/net/wireless/ath/ath9k/xmit.c
1163
cur_ht20_sgi = sc->tx.max_aggr_framelen[queue][MCS_HT20_SGI];
drivers/net/wireless/ath/ath9k/xmit.c
1164
cur_ht40 = sc->tx.max_aggr_framelen[queue][MCS_HT40];
drivers/net/wireless/ath/ath9k/xmit.c
1165
cur_ht40_sgi = sc->tx.max_aggr_framelen[queue][MCS_HT40_SGI];
drivers/net/wireless/ath/ath9k/xmit.c
123
struct ieee80211_txq *queue =
drivers/net/wireless/ath/ath9k/xmit.c
126
ieee80211_schedule_txq(sc->hw, queue);
drivers/net/wireless/ath/ath9k/xmit.c
129
void ath9k_wake_tx_queue(struct ieee80211_hw *hw, struct ieee80211_txq *queue)
drivers/net/wireless/ath/ath9k/xmit.c
133
struct ath_atx_tid *tid = (struct ath_atx_tid *) queue->drv_priv;
drivers/net/wireless/ath/ath9k/xmit.c
137
queue->sta ? queue->sta->addr : queue->vif->addr,
drivers/net/wireless/ath/ath9k/xmit.c
1984
struct ieee80211_txq *queue;
drivers/net/wireless/ath/ath9k/xmit.c
2001
while ((queue = ieee80211_next_txq(hw, txq->mac80211_qnum))) {
drivers/net/wireless/ath/ath9k/xmit.c
2004
tid = (struct ath_atx_tid *)queue->drv_priv;
drivers/net/wireless/ath/ath9k/xmit.c
2010
ieee80211_return_txq(hw, queue, force);
drivers/net/wireless/ath/carl9170/carl9170.h
138
struct sk_buff_head queue;
drivers/net/wireless/ath/carl9170/debug.c
331
" currently queued:%d\n", skb_queue_len(&iter->queue));
drivers/net/wireless/ath/carl9170/debug.c
334
skb_queue_walk(&iter->queue, skb) {
drivers/net/wireless/ath/carl9170/debug.c
353
ssize_t *len, size_t bufsize, struct sk_buff_head *queue)
drivers/net/wireless/ath/carl9170/debug.c
359
spin_lock_bh(&queue->lock);
drivers/net/wireless/ath/carl9170/debug.c
360
skb_queue_walk(queue, skb) {
drivers/net/wireless/ath/carl9170/debug.c
366
spin_unlock_bh(&queue->lock);
drivers/net/wireless/ath/carl9170/fwcmd.h
266
u8 queue:2;
drivers/net/wireless/ath/carl9170/main.c
1369
unsigned int link_id, u16 queue,
drivers/net/wireless/ath/carl9170/main.c
1376
memcpy(&ar->edcf[ar9170_qmap(queue)], param, sizeof(*param));
drivers/net/wireless/ath/carl9170/main.c
1428
skb_queue_head_init(&tid_info->queue);
drivers/net/wireless/ath/carl9170/main.c
211
while ((skb = __skb_dequeue(&tid_info->queue)))
drivers/net/wireless/ath/carl9170/main.c
264
while ((skb = __skb_dequeue(&tid_info->queue)))
drivers/net/wireless/ath/carl9170/main.c
328
#define CARL9170_FILL_QUEUE(queue, ai_fs, cwmin, cwmax, _txop) \
drivers/net/wireless/ath/carl9170/main.c
330
queue.aifs = ai_fs; \
drivers/net/wireless/ath/carl9170/main.c
331
queue.cw_min = cwmin; \
drivers/net/wireless/ath/carl9170/main.c
332
queue.cw_max = cwmax; \
drivers/net/wireless/ath/carl9170/main.c
333
queue.txop = _txop; \
drivers/net/wireless/ath/carl9170/rx.c
577
unsigned int queue;
drivers/net/wireless/ath/carl9170/rx.c
585
queue = TID_TO_WME_AC(((le16_to_cpu(bar->control) &
drivers/net/wireless/ath/carl9170/rx.c
590
list_for_each_entry_rcu(entry, &ar->bar_list[queue], list) {
drivers/net/wireless/ath/carl9170/rx.c
608
spin_lock_bh(&ar->bar_list_lock[queue]);
drivers/net/wireless/ath/carl9170/rx.c
610
spin_unlock_bh(&ar->bar_list_lock[queue]);
drivers/net/wireless/ath/carl9170/tx.c
1136
u16 seq, queue, tmpssn;
drivers/net/wireless/ath/carl9170/tx.c
1163
queue = TID_TO_WME_AC(tid_info->tid);
drivers/net/wireless/ath/carl9170/tx.c
1170
first = skb_peek(&tid_info->queue);
drivers/net/wireless/ath/carl9170/tx.c
1181
while ((skb = skb_peek(&tid_info->queue))) {
drivers/net/wireless/ath/carl9170/tx.c
1201
__skb_unlink(skb, &tid_info->queue);
drivers/net/wireless/ath/carl9170/tx.c
1209
if (skb_queue_empty(&tid_info->queue) ||
drivers/net/wireless/ath/carl9170/tx.c
1210
carl9170_get_seq(skb_peek(&tid_info->queue)) !=
drivers/net/wireless/ath/carl9170/tx.c
1232
spin_lock_bh(&ar->tx_pending[queue].lock);
drivers/net/wireless/ath/carl9170/tx.c
1233
skb_queue_splice_tail_init(&agg, &ar->tx_pending[queue]);
drivers/net/wireless/ath/carl9170/tx.c
1234
spin_unlock_bh(&ar->tx_pending[queue].lock);
drivers/net/wireless/ath/carl9170/tx.c
1245
struct sk_buff_head *queue)
drivers/net/wireless/ath/carl9170/tx.c
1253
spin_lock_bh(&queue->lock);
drivers/net/wireless/ath/carl9170/tx.c
1254
skb = skb_peek(queue);
drivers/net/wireless/ath/carl9170/tx.c
1261
__skb_unlink(skb, queue);
drivers/net/wireless/ath/carl9170/tx.c
1262
spin_unlock_bh(&queue->lock);
drivers/net/wireless/ath/carl9170/tx.c
1271
spin_unlock_bh(&queue->lock);
drivers/net/wireless/ath/carl9170/tx.c
1329
unsigned int queue = skb_get_queue_mapping(skb);
drivers/net/wireless/ath/carl9170/tx.c
1334
spin_lock_bh(&ar->bar_list_lock[queue]);
drivers/net/wireless/ath/carl9170/tx.c
1335
list_add_tail_rcu(&entry->list, &ar->bar_list[queue]);
drivers/net/wireless/ath/carl9170/tx.c
1336
spin_unlock_bh(&ar->bar_list_lock[queue]);
drivers/net/wireless/ath/carl9170/tx.c
1432
__skb_queue_tail(&agg->queue, skb);
drivers/net/wireless/ath/carl9170/tx.c
1437
skb_queue_reverse_walk(&agg->queue, iter) {
drivers/net/wireless/ath/carl9170/tx.c
1441
__skb_queue_after(&agg->queue, iter, skb);
drivers/net/wireless/ath/carl9170/tx.c
1446
__skb_queue_head(&agg->queue, skb);
drivers/net/wireless/ath/carl9170/tx.c
1450
if (agg->snx == carl9170_get_seq(skb_peek(&agg->queue))) {
drivers/net/wireless/ath/carl9170/tx.c
1515
unsigned int queue = skb_get_queue_mapping(skb);
drivers/net/wireless/ath/carl9170/tx.c
1519
skb_queue_tail(&ar->tx_pending[queue], skb);
drivers/net/wireless/ath/carl9170/tx.c
158
int queue;
drivers/net/wireless/ath/carl9170/tx.c
160
queue = skb_get_queue_mapping(skb);
drivers/net/wireless/ath/carl9170/tx.c
164
ar->tx_stats[queue].len--;
drivers/net/wireless/ath/carl9170/tx.c
455
int queue = skb_get_queue_mapping(skb);
drivers/net/wireless/ath/carl9170/tx.c
458
list_for_each_entry_rcu(entry, &ar->bar_list[queue], list) {
drivers/net/wireless/ath/carl9170/tx.c
460
spin_lock_bh(&ar->bar_list_lock[queue]);
drivers/net/wireless/ath/carl9170/tx.c
462
spin_unlock_bh(&ar->bar_list_lock[queue]);
drivers/net/wireless/ath/carl9170/tx.c
469
queue, bar->ra, bar->ta, bar->control,
drivers/net/wireless/ath/carl9170/tx.c
49
unsigned int queue)
drivers/net/wireless/ath/carl9170/tx.c
514
struct sk_buff_head *queue)
drivers/net/wireless/ath/carl9170/tx.c
518
spin_lock_bh(&queue->lock);
drivers/net/wireless/ath/carl9170/tx.c
519
skb_queue_walk(queue, skb) {
drivers/net/wireless/ath/carl9170/tx.c
52
return queue;
drivers/net/wireless/ath/carl9170/tx.c
525
__skb_unlink(skb, queue);
drivers/net/wireless/ath/carl9170/tx.c
526
spin_unlock_bh(&queue->lock);
drivers/net/wireless/ath/carl9170/tx.c
531
spin_unlock_bh(&queue->lock);
drivers/net/wireless/ath/carl9170/tx.c
617
skb = skb_peek(&iter->queue);
drivers/net/wireless/ath/carl9170/tx.c
78
int queue, i;
drivers/net/wireless/ath/carl9170/tx.c
83
queue = skb_get_queue_mapping(skb);
drivers/net/wireless/ath/carl9170/tx.c
92
ar->tx_stats[queue].len++;
drivers/net/wireless/ath/carl9170/tx.c
93
ar->tx_stats[queue].count++;
drivers/net/wireless/ath/carl9170/wlan.h
253
u8 queue:2;
drivers/net/wireless/broadcom/b43/b43.h
708
#define B43_QOS_PARAMS(queue) (B43_SHM_SH_EDCFQ + \
drivers/net/wireless/broadcom/b43/b43.h
709
(B43_NR_QOSPARAMS * sizeof(u16) * (queue)))
drivers/net/wireless/broadcom/b43/main.c
3794
unsigned int queue = (unsigned int)_queue;
drivers/net/wireless/broadcom/b43/main.c
3797
if (queue >= ARRAY_SIZE(wl->qos_params)) {
drivers/net/wireless/broadcom/b43/main.c
3811
memcpy(&(wl->qos_params[queue].p), params, sizeof(*params));
drivers/net/wireless/broadcom/b43/main.c
3813
b43_qos_params_upload(dev, &(wl->qos_params[queue].p),
drivers/net/wireless/broadcom/b43/main.c
3814
b43_qos_shm_offsets[queue]);
drivers/net/wireless/broadcom/b43/pio.c
152
p->queue = q;
drivers/net/wireless/broadcom/b43/pio.c
209
#define destroy_queue_tx(pio, queue) do { \
drivers/net/wireless/broadcom/b43/pio.c
210
b43_destroy_pioqueue_tx((pio)->queue, __stringify(queue)); \
drivers/net/wireless/broadcom/b43/pio.c
211
(pio)->queue = NULL; \
drivers/net/wireless/broadcom/b43/pio.c
214
#define destroy_queue_rx(pio, queue) do { \
drivers/net/wireless/broadcom/b43/pio.c
215
b43_destroy_pioqueue_rx((pio)->queue, __stringify(queue)); \
drivers/net/wireless/broadcom/b43/pio.c
216
(pio)->queue = NULL; \
drivers/net/wireless/broadcom/b43/pio.c
352
struct b43_pio_txqueue *q = pack->queue;
drivers/net/wireless/broadcom/b43/pio.c
422
struct b43_pio_txqueue *q = pack->queue;
drivers/net/wireless/broadcom/b43/pio.h
61
struct b43_pio_txqueue *queue;
drivers/net/wireless/broadcom/b43legacy/main.c
2509
unsigned int link_id, u16 queue,
drivers/net/wireless/broadcom/b43legacy/pio.c
103
static u16 generate_cookie(struct b43legacy_pioqueue *queue,
drivers/net/wireless/broadcom/b43legacy/pio.c
113
switch (queue->mmio_base) {
drivers/net/wireless/broadcom/b43legacy/pio.c
141
struct b43legacy_pioqueue *queue = NULL;
drivers/net/wireless/broadcom/b43legacy/pio.c
146
queue = pio->queue0;
drivers/net/wireless/broadcom/b43legacy/pio.c
149
queue = pio->queue1;
drivers/net/wireless/broadcom/b43legacy/pio.c
152
queue = pio->queue2;
drivers/net/wireless/broadcom/b43legacy/pio.c
155
queue = pio->queue3;
drivers/net/wireless/broadcom/b43legacy/pio.c
163
*packet = &(queue->tx_packets_cache[packetindex]);
drivers/net/wireless/broadcom/b43legacy/pio.c
165
return queue;
drivers/net/wireless/broadcom/b43legacy/pio.c
172
static int pio_tx_write_fragment(struct b43legacy_pioqueue *queue,
drivers/net/wireless/broadcom/b43legacy/pio.c
185
err = b43legacy_generate_txhdr(queue->dev,
drivers/net/wireless/broadcom/b43legacy/pio.c
188
generate_cookie(queue, packet));
drivers/net/wireless/broadcom/b43legacy/pio.c
192
tx_start(queue);
drivers/net/wireless/broadcom/b43legacy/pio.c
194
if (queue->need_workarounds)
drivers/net/wireless/broadcom/b43legacy/pio.c
196
tx_data(queue, txhdr, (u8 *)skb->data, octets);
drivers/net/wireless/broadcom/b43legacy/pio.c
197
tx_complete(queue, skb);
drivers/net/wireless/broadcom/b43legacy/pio.c
205
struct b43legacy_pioqueue *queue = packet->queue;
drivers/net/wireless/broadcom/b43legacy/pio.c
213
list_move(&packet->list, &queue->txfree);
drivers/net/wireless/broadcom/b43legacy/pio.c
214
queue->nr_txfree++;
drivers/net/wireless/broadcom/b43legacy/pio.c
219
struct b43legacy_pioqueue *queue = packet->queue;
drivers/net/wireless/broadcom/b43legacy/pio.c
22
static void tx_start(struct b43legacy_pioqueue *queue)
drivers/net/wireless/broadcom/b43legacy/pio.c
225
if (queue->tx_devq_size < octets) {
drivers/net/wireless/broadcom/b43legacy/pio.c
226
b43legacywarn(queue->dev->wl, "PIO queue too small. "
drivers/net/wireless/broadcom/b43legacy/pio.c
232
B43legacy_WARN_ON(queue->tx_devq_packets >
drivers/net/wireless/broadcom/b43legacy/pio.c
234
B43legacy_WARN_ON(queue->tx_devq_used > queue->tx_devq_size);
drivers/net/wireless/broadcom/b43legacy/pio.c
239
if (queue->tx_devq_packets == B43legacy_PIO_MAXTXDEVQPACKETS)
drivers/net/wireless/broadcom/b43legacy/pio.c
24
b43legacy_pio_write(queue, B43legacy_PIO_TXCTL,
drivers/net/wireless/broadcom/b43legacy/pio.c
241
if (queue->tx_devq_used + octets > queue->tx_devq_size)
drivers/net/wireless/broadcom/b43legacy/pio.c
244
err = pio_tx_write_fragment(queue, skb, packet,
drivers/net/wireless/broadcom/b43legacy/pio.c
256
queue->tx_devq_packets++;
drivers/net/wireless/broadcom/b43legacy/pio.c
257
queue->tx_devq_used += octets;
drivers/net/wireless/broadcom/b43legacy/pio.c
262
list_move_tail(&packet->list, &queue->txrunning);
drivers/net/wireless/broadcom/b43legacy/pio.c
269
struct b43legacy_pioqueue *queue = from_tasklet(queue, t, txtask);
drivers/net/wireless/broadcom/b43legacy/pio.c
270
struct b43legacy_wldev *dev = queue->dev;
drivers/net/wireless/broadcom/b43legacy/pio.c
277
if (queue->tx_frozen)
drivers/net/wireless/broadcom/b43legacy/pio.c
279
txctl = b43legacy_pio_read(queue, B43legacy_PIO_TXCTL);
drivers/net/wireless/broadcom/b43legacy/pio.c
28
static void tx_octet(struct b43legacy_pioqueue *queue,
drivers/net/wireless/broadcom/b43legacy/pio.c
283
list_for_each_entry_safe(packet, tmp_packet, &queue->txqueue, list) {
drivers/net/wireless/broadcom/b43legacy/pio.c
299
static void setup_txqueues(struct b43legacy_pioqueue *queue)
drivers/net/wireless/broadcom/b43legacy/pio.c
304
queue->nr_txfree = B43legacy_PIO_MAXTXPACKETS;
drivers/net/wireless/broadcom/b43legacy/pio.c
306
packet = &(queue->tx_packets_cache[i]);
drivers/net/wireless/broadcom/b43legacy/pio.c
308
packet->queue = queue;
drivers/net/wireless/broadcom/b43legacy/pio.c
31
if (queue->need_workarounds) {
drivers/net/wireless/broadcom/b43legacy/pio.c
311
list_add(&packet->list, &queue->txfree);
drivers/net/wireless/broadcom/b43legacy/pio.c
319
struct b43legacy_pioqueue *queue;
drivers/net/wireless/broadcom/b43legacy/pio.c
32
b43legacy_pio_write(queue, B43legacy_PIO_TXDATA, octet);
drivers/net/wireless/broadcom/b43legacy/pio.c
323
queue = kzalloc_obj(*queue);
drivers/net/wireless/broadcom/b43legacy/pio.c
324
if (!queue)
drivers/net/wireless/broadcom/b43legacy/pio.c
327
queue->dev = dev;
drivers/net/wireless/broadcom/b43legacy/pio.c
328
queue->mmio_base = pio_mmio_base;
drivers/net/wireless/broadcom/b43legacy/pio.c
329
queue->need_workarounds = (dev->dev->id.revision < 3);
drivers/net/wireless/broadcom/b43legacy/pio.c
33
b43legacy_pio_write(queue, B43legacy_PIO_TXCTL,
drivers/net/wireless/broadcom/b43legacy/pio.c
331
INIT_LIST_HEAD(&queue->txfree);
drivers/net/wireless/broadcom/b43legacy/pio.c
332
INIT_LIST_HEAD(&queue->txqueue);
drivers/net/wireless/broadcom/b43legacy/pio.c
333
INIT_LIST_HEAD(&queue->txrunning);
drivers/net/wireless/broadcom/b43legacy/pio.c
334
tasklet_setup(&queue->txtask, tx_tasklet);
drivers/net/wireless/broadcom/b43legacy/pio.c
340
qsize = b43legacy_read16(dev, queue->mmio_base
drivers/net/wireless/broadcom/b43legacy/pio.c
354
queue->tx_devq_size = qsize;
drivers/net/wireless/broadcom/b43legacy/pio.c
356
setup_txqueues(queue);
drivers/net/wireless/broadcom/b43legacy/pio.c
359
return queue;
drivers/net/wireless/broadcom/b43legacy/pio.c
36
b43legacy_pio_write(queue, B43legacy_PIO_TXCTL,
drivers/net/wireless/broadcom/b43legacy/pio.c
362
kfree(queue);
drivers/net/wireless/broadcom/b43legacy/pio.c
363
queue = NULL;
drivers/net/wireless/broadcom/b43legacy/pio.c
367
static void cancel_transfers(struct b43legacy_pioqueue *queue)
drivers/net/wireless/broadcom/b43legacy/pio.c
371
tasklet_kill(&queue->txtask);
drivers/net/wireless/broadcom/b43legacy/pio.c
373
list_for_each_entry_safe(packet, tmp_packet, &queue->txrunning, list)
drivers/net/wireless/broadcom/b43legacy/pio.c
375
list_for_each_entry_safe(packet, tmp_packet, &queue->txqueue, list)
drivers/net/wireless/broadcom/b43legacy/pio.c
379
static void b43legacy_destroy_pioqueue(struct b43legacy_pioqueue *queue)
drivers/net/wireless/broadcom/b43legacy/pio.c
38
b43legacy_pio_write(queue, B43legacy_PIO_TXDATA, octet);
drivers/net/wireless/broadcom/b43legacy/pio.c
381
if (!queue)
drivers/net/wireless/broadcom/b43legacy/pio.c
384
cancel_transfers(queue);
drivers/net/wireless/broadcom/b43legacy/pio.c
385
kfree(queue);
drivers/net/wireless/broadcom/b43legacy/pio.c
409
struct b43legacy_pioqueue *queue;
drivers/net/wireless/broadcom/b43legacy/pio.c
412
queue = b43legacy_setup_pioqueue(dev, B43legacy_MMIO_PIO1_BASE);
drivers/net/wireless/broadcom/b43legacy/pio.c
413
if (!queue)
drivers/net/wireless/broadcom/b43legacy/pio.c
415
pio->queue0 = queue;
drivers/net/wireless/broadcom/b43legacy/pio.c
417
queue = b43legacy_setup_pioqueue(dev, B43legacy_MMIO_PIO2_BASE);
drivers/net/wireless/broadcom/b43legacy/pio.c
418
if (!queue)
drivers/net/wireless/broadcom/b43legacy/pio.c
420
pio->queue1 = queue;
drivers/net/wireless/broadcom/b43legacy/pio.c
422
queue = b43legacy_setup_pioqueue(dev, B43legacy_MMIO_PIO3_BASE);
drivers/net/wireless/broadcom/b43legacy/pio.c
423
if (!queue)
drivers/net/wireless/broadcom/b43legacy/pio.c
425
pio->queue2 = queue;
drivers/net/wireless/broadcom/b43legacy/pio.c
427
queue = b43legacy_setup_pioqueue(dev, B43legacy_MMIO_PIO4_BASE);
drivers/net/wireless/broadcom/b43legacy/pio.c
428
if (!queue)
drivers/net/wireless/broadcom/b43legacy/pio.c
430
pio->queue3 = queue;
drivers/net/wireless/broadcom/b43legacy/pio.c
455
struct b43legacy_pioqueue *queue = dev->pio.queue1;
drivers/net/wireless/broadcom/b43legacy/pio.c
458
B43legacy_WARN_ON(queue->tx_suspended);
drivers/net/wireless/broadcom/b43legacy/pio.c
459
B43legacy_WARN_ON(list_empty(&queue->txfree));
drivers/net/wireless/broadcom/b43legacy/pio.c
461
packet = list_entry(queue->txfree.next, struct b43legacy_pio_txpacket,
drivers/net/wireless/broadcom/b43legacy/pio.c
465
list_move_tail(&packet->list, &queue->txqueue);
drivers/net/wireless/broadcom/b43legacy/pio.c
466
queue->nr_txfree--;
drivers/net/wireless/broadcom/b43legacy/pio.c
467
B43legacy_WARN_ON(queue->nr_txfree >= B43legacy_PIO_MAXTXPACKETS);
drivers/net/wireless/broadcom/b43legacy/pio.c
469
tasklet_schedule(&queue->txtask);
drivers/net/wireless/broadcom/b43legacy/pio.c
477
struct b43legacy_pioqueue *queue;
drivers/net/wireless/broadcom/b43legacy/pio.c
482
queue = parse_cookie(dev, status->cookie, &packet);
drivers/net/wireless/broadcom/b43legacy/pio.c
483
B43legacy_WARN_ON(!queue);
drivers/net/wireless/broadcom/b43legacy/pio.c
488
queue->tx_devq_packets--;
drivers/net/wireless/broadcom/b43legacy/pio.c
489
queue->tx_devq_used -= (packet->skb->len +
drivers/net/wireless/broadcom/b43legacy/pio.c
531
if (!list_empty(&queue->txqueue))
drivers/net/wireless/broadcom/b43legacy/pio.c
532
tasklet_schedule(&queue->txtask);
drivers/net/wireless/broadcom/b43legacy/pio.c
535
static void pio_rx_error(struct b43legacy_pioqueue *queue,
drivers/net/wireless/broadcom/b43legacy/pio.c
541
b43legacyerr(queue->dev->wl, "PIO RX error: %s\n", error);
drivers/net/wireless/broadcom/b43legacy/pio.c
542
b43legacy_pio_write(queue, B43legacy_PIO_RXCTL,
drivers/net/wireless/broadcom/b43legacy/pio.c
545
B43legacy_WARN_ON(queue->mmio_base != B43legacy_MMIO_PIO1_BASE);
drivers/net/wireless/broadcom/b43legacy/pio.c
548
b43legacy_pio_read(queue, B43legacy_PIO_RXDATA);
drivers/net/wireless/broadcom/b43legacy/pio.c
553
void b43legacy_pio_rx(struct b43legacy_pioqueue *queue)
drivers/net/wireless/broadcom/b43legacy/pio.c
564
tmp = b43legacy_pio_read(queue, B43legacy_PIO_RXCTL);
drivers/net/wireless/broadcom/b43legacy/pio.c
567
b43legacy_pio_write(queue, B43legacy_PIO_RXCTL,
drivers/net/wireless/broadcom/b43legacy/pio.c
571
tmp = b43legacy_pio_read(queue, B43legacy_PIO_RXCTL);
drivers/net/wireless/broadcom/b43legacy/pio.c
576
b43legacydbg(queue->dev->wl, "PIO RX timed out\n");
drivers/net/wireless/broadcom/b43legacy/pio.c
580
len = b43legacy_pio_read(queue, B43legacy_PIO_RXDATA);
drivers/net/wireless/broadcom/b43legacy/pio.c
582
pio_rx_error(queue, 0, "len > 0x700");
drivers/net/wireless/broadcom/b43legacy/pio.c
585
if (unlikely(len == 0 && queue->mmio_base !=
drivers/net/wireless/broadcom/b43legacy/pio.c
587
pio_rx_error(queue, 0, "len == 0");
drivers/net/wireless/broadcom/b43legacy/pio.c
591
if (queue->mmio_base == B43legacy_MMIO_PIO4_BASE)
drivers/net/wireless/broadcom/b43legacy/pio.c
596
tmp = b43legacy_pio_read(queue, B43legacy_PIO_RXDATA);
drivers/net/wireless/broadcom/b43legacy/pio.c
602
pio_rx_error(queue,
drivers/net/wireless/broadcom/b43legacy/pio.c
603
(queue->mmio_base == B43legacy_MMIO_PIO1_BASE),
drivers/net/wireless/broadcom/b43legacy/pio.c
607
if (queue->mmio_base == B43legacy_MMIO_PIO4_BASE) {
drivers/net/wireless/broadcom/b43legacy/pio.c
612
b43legacy_handle_hwtxstatus(queue->dev, hw);
drivers/net/wireless/broadcom/b43legacy/pio.c
619
pio_rx_error(queue, 1, "OOM");
drivers/net/wireless/broadcom/b43legacy/pio.c
624
tmp = b43legacy_pio_read(queue, B43legacy_PIO_RXDATA);
drivers/net/wireless/broadcom/b43legacy/pio.c
628
tmp = b43legacy_pio_read(queue, B43legacy_PIO_RXDATA);
drivers/net/wireless/broadcom/b43legacy/pio.c
63
static void tx_data(struct b43legacy_pioqueue *queue,
drivers/net/wireless/broadcom/b43legacy/pio.c
631
b43legacy_rx(queue->dev, skb, rxhdr);
drivers/net/wireless/broadcom/b43legacy/pio.c
634
void b43legacy_pio_tx_suspend(struct b43legacy_pioqueue *queue)
drivers/net/wireless/broadcom/b43legacy/pio.c
636
b43legacy_power_saving_ctl_bits(queue->dev, -1, 1);
drivers/net/wireless/broadcom/b43legacy/pio.c
637
b43legacy_pio_write(queue, B43legacy_PIO_TXCTL,
drivers/net/wireless/broadcom/b43legacy/pio.c
638
b43legacy_pio_read(queue, B43legacy_PIO_TXCTL)
drivers/net/wireless/broadcom/b43legacy/pio.c
642
void b43legacy_pio_tx_resume(struct b43legacy_pioqueue *queue)
drivers/net/wireless/broadcom/b43legacy/pio.c
644
b43legacy_pio_write(queue, B43legacy_PIO_TXCTL,
drivers/net/wireless/broadcom/b43legacy/pio.c
645
b43legacy_pio_read(queue, B43legacy_PIO_TXCTL)
drivers/net/wireless/broadcom/b43legacy/pio.c
647
b43legacy_power_saving_ctl_bits(queue->dev, -1, -1);
drivers/net/wireless/broadcom/b43legacy/pio.c
648
tasklet_schedule(&queue->txtask);
drivers/net/wireless/broadcom/b43legacy/pio.c
71
if (queue->need_workarounds) {
drivers/net/wireless/broadcom/b43legacy/pio.c
74
b43legacy_pio_write(queue, B43legacy_PIO_TXDATA, data);
drivers/net/wireless/broadcom/b43legacy/pio.c
76
b43legacy_pio_write(queue, B43legacy_PIO_TXCTL,
drivers/net/wireless/broadcom/b43legacy/pio.c
82
b43legacy_pio_write(queue, B43legacy_PIO_TXDATA, data);
drivers/net/wireless/broadcom/b43legacy/pio.c
85
tx_octet(queue, packet[octets -
drivers/net/wireless/broadcom/b43legacy/pio.c
89
static void tx_complete(struct b43legacy_pioqueue *queue,
drivers/net/wireless/broadcom/b43legacy/pio.c
92
if (queue->need_workarounds) {
drivers/net/wireless/broadcom/b43legacy/pio.c
93
b43legacy_pio_write(queue, B43legacy_PIO_TXDATA,
drivers/net/wireless/broadcom/b43legacy/pio.c
95
b43legacy_pio_write(queue, B43legacy_PIO_TXCTL,
drivers/net/wireless/broadcom/b43legacy/pio.c
99
b43legacy_pio_write(queue, B43legacy_PIO_TXCTL,
drivers/net/wireless/broadcom/b43legacy/pio.h
105
void b43legacy_pio_rx(struct b43legacy_pioqueue *queue);
drivers/net/wireless/broadcom/b43legacy/pio.h
108
void b43legacy_pio_tx_suspend(struct b43legacy_pioqueue *queue);
drivers/net/wireless/broadcom/b43legacy/pio.h
109
void b43legacy_pio_tx_resume(struct b43legacy_pioqueue *queue);
drivers/net/wireless/broadcom/b43legacy/pio.h
137
void b43legacy_pio_rx(struct b43legacy_pioqueue *queue)
drivers/net/wireless/broadcom/b43legacy/pio.h
141
void b43legacy_pio_tx_suspend(struct b43legacy_pioqueue *queue)
drivers/net/wireless/broadcom/b43legacy/pio.h
145
void b43legacy_pio_tx_resume(struct b43legacy_pioqueue *queue)
drivers/net/wireless/broadcom/b43legacy/pio.h
43
struct b43legacy_pioqueue *queue;
drivers/net/wireless/broadcom/b43legacy/pio.h
49
(packet)->queue->tx_packets_cache))
drivers/net/wireless/broadcom/b43legacy/pio.h
84
u16 b43legacy_pio_read(struct b43legacy_pioqueue *queue,
drivers/net/wireless/broadcom/b43legacy/pio.h
87
return b43legacy_read16(queue->dev, queue->mmio_base + offset);
drivers/net/wireless/broadcom/b43legacy/pio.h
91
void b43legacy_pio_write(struct b43legacy_pioqueue *queue,
drivers/net/wireless/broadcom/b43legacy/pio.h
94
b43legacy_write16(queue->dev, queue->mmio_base + offset, value);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c
1241
struct sk_buff_head *queue;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c
1265
queue = &pq->q[prec].skblist;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c
1267
p_head = skb_peek(queue);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c
1268
p_tail = skb_peek_tail(queue);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c
1280
p_tail = skb_queue_prev(queue, p_tail);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c
1285
__skb_queue_tail(queue, p);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c
1294
__skb_queue_after(queue, p_tail, p);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/fwsignal.c
1297
__skb_insert(p, p_tail->prev, p_tail, queue);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c
1669
queue);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c
1670
list_del(&work->queue);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c
222
struct list_head queue;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c
573
struct brcmf_msgbuf_work_item, queue);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c
574
list_del(&work->queue);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/msgbuf.c
690
list_add_tail(&create->queue, &msgbuf->work_queue);
drivers/net/wireless/broadcom/brcm80211/brcmsmac/ampdu.c
1022
u8 queue = txs->frameid & TXFID_QUEUE_MASK;
drivers/net/wireless/broadcom/brcm80211/brcmsmac/ampdu.c
1035
p = dma_getnexttxp(wlc->hw->di[queue],
drivers/net/wireless/broadcom/brcm80211/brcmsmac/ampdu.c
809
u8 bitmap[8], queue, tid;
drivers/net/wireless/broadcom/brcm80211/brcmsmac/ampdu.c
834
queue = txs->frameid & TXFID_QUEUE_MASK;
drivers/net/wireless/broadcom/brcm80211/brcmsmac/ampdu.c
885
brcms_c_ffpld_check_txfunfl(wlc, queue);
drivers/net/wireless/broadcom/brcm80211/brcmsmac/ampdu.c
953
ret = brcms_c_txfifo(wlc, queue, p);
drivers/net/wireless/broadcom/brcm80211/brcmsmac/ampdu.c
959
WARN_ONCE(ret, "queue %d out of txds\n", queue);
drivers/net/wireless/broadcom/brcm80211/brcmsmac/ampdu.c
983
p = dma_getnexttxp(wlc->hw->di[queue], DMA_RANGE_TRANSMITTED);
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
793
unsigned int link_id, u16 queue,
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
799
brcms_c_wme_setparams(wl->wlc, queue, params, true);
drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c
1000
u16 ac_queue = brcms_fifo_to_ac(queue);
drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c
1001
if (dma->txavail > TX_HEADROOM && queue < TX_BCMC_FIFO &&
drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c
6154
uint nfrags, uint queue, uint next_frag_len)
drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c
6208
if (queue == TX_BCMC_FIFO) {
drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c
6224
(queue & TXFID_QUEUE_MASK);
drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c
6227
frameid |= queue & TXFID_QUEUE_MASK;
drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c
6765
if (wlc->fragthresh[queue] !=
drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c
6767
wlc->fragthresh[queue] =
drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c
6772
wlc->pub->unit, fifo_names[queue],
drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c
6780
fifo_names[queue],
drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c
6781
phylen, wlc->fragthresh[queue],
drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c
6864
u16 queue;
drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c
6875
queue = skb_get_queue_mapping(p);
drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c
6877
!ieee80211_queue_stopped(wlc->pub->ieee_hw, queue))
drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c
6878
ieee80211_stop_queue(wlc->pub->ieee_hw, queue);
drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c
829
uint queue = NFIFO;
drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c
858
queue = txs->frameid & TXFID_QUEUE_MASK;
drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c
859
if (queue >= NFIFO) {
drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c
860
brcms_err(wlc->hw->d11core, "queue %u >= NFIFO\n", queue);
drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c
864
dma = wlc->hw->di[queue];
drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c
866
p = dma_getnexttxp(wlc->hw->di[queue], DMA_RANGE_TRANSMITTED);
drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c
929
if (queue < IEEE80211_NUM_ACS) {
drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c
930
sfbl = GFIELD(wlc->wme_retries[wme_fifo2ac[queue]],
drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c
932
lfbl = GFIELD(wlc->wme_retries[wme_fifo2ac[queue]],
drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c
999
if (dma && queue < NFIFO) {
drivers/net/wireless/intel/ipw2x00/ipw2200.c
5108
rxq->queue[rxq->write] = rxb;
drivers/net/wireless/intel/ipw2x00/ipw2200.c
8223
rxb = priv->rxq->queue[i];
drivers/net/wireless/intel/ipw2x00/ipw2200.c
8228
priv->rxq->queue[i] = NULL;
drivers/net/wireless/intel/ipw2x00/ipw2200.h
706
struct ipw_rx_mem_buffer *queue[RX_QUEUE_SIZE];
drivers/net/wireless/intel/iwlegacy/3945-mac.c
1167
rxb = rxq->queue[i];
drivers/net/wireless/intel/iwlegacy/3945-mac.c
1174
rxq->queue[i] = NULL;
drivers/net/wireless/intel/iwlegacy/3945-mac.c
944
rxq->queue[rxq->write] = rxb;
drivers/net/wireless/intel/iwlegacy/4965-mac.c
108
rxq->queue[i] = NULL;
drivers/net/wireless/intel/iwlegacy/4965-mac.c
261
rxb = rxq->queue[rxq->write];
drivers/net/wireless/intel/iwlegacy/4965-mac.c
272
rxq->queue[rxq->write] = rxb;
drivers/net/wireless/intel/iwlegacy/4965-mac.c
4237
rxb = rxq->queue[i];
drivers/net/wireless/intel/iwlegacy/4965-mac.c
4244
rxq->queue[i] = NULL;
drivers/net/wireless/intel/iwlegacy/common.c
4461
unsigned int link_id, u16 queue,
drivers/net/wireless/intel/iwlegacy/common.c
4475
if (queue >= AC_NUM) {
drivers/net/wireless/intel/iwlegacy/common.c
4476
D_MAC80211("leave - queue >= AC_NUM %d\n", queue);
drivers/net/wireless/intel/iwlegacy/common.c
4480
q = AC_NUM - 1 - queue;
drivers/net/wireless/intel/iwlegacy/common.h
1698
unsigned int link_id, u16 queue,
drivers/net/wireless/intel/iwlegacy/common.h
2281
u8 queue = txq->swq_id;
drivers/net/wireless/intel/iwlegacy/common.h
2282
u8 ac = queue & 3;
drivers/net/wireless/intel/iwlegacy/common.h
2283
u8 hwq = (queue >> 2) & 0x1f;
drivers/net/wireless/intel/iwlegacy/common.h
2292
u8 queue = txq->swq_id;
drivers/net/wireless/intel/iwlegacy/common.h
2293
u8 ac = queue & 3;
drivers/net/wireless/intel/iwlegacy/common.h
2294
u8 hwq = (queue >> 2) & 0x1f;
drivers/net/wireless/intel/iwlegacy/common.h
608
struct il_rx_buf *queue[RX_QUEUE_SIZE];
drivers/net/wireless/intel/iwlwifi/dvm/mac80211.c
1147
unsigned int link_id, u16 queue,
drivers/net/wireless/intel/iwlwifi/dvm/mac80211.c
1165
if (queue >= AC_NUM) {
drivers/net/wireless/intel/iwlwifi/dvm/mac80211.c
1166
IWL_DEBUG_MAC80211(priv, "leave - queue >= AC_NUM %d\n", queue);
drivers/net/wireless/intel/iwlwifi/dvm/mac80211.c
1170
q = AC_NUM - 1 - queue;
drivers/net/wireless/intel/iwlwifi/dvm/main.c
2055
static void iwl_stop_sw_queue(struct iwl_op_mode *op_mode, int queue)
drivers/net/wireless/intel/iwlwifi/dvm/main.c
2058
int mq = priv->queue_to_mac80211[queue];
drivers/net/wireless/intel/iwlwifi/dvm/main.c
2066
queue, mq);
drivers/net/wireless/intel/iwlwifi/dvm/main.c
2074
static void iwl_wake_sw_queue(struct iwl_op_mode *op_mode, int queue)
drivers/net/wireless/intel/iwlwifi/dvm/main.c
2077
int mq = priv->queue_to_mac80211[queue];
drivers/net/wireless/intel/iwlwifi/dvm/main.c
2085
queue, mq);
drivers/net/wireless/intel/iwlwifi/iwl-op-mode.h
162
struct iwl_rx_cmd_buffer *rxb, unsigned int queue);
drivers/net/wireless/intel/iwlwifi/iwl-op-mode.h
163
void (*queue_full)(struct iwl_op_mode *op_mode, int queue);
drivers/net/wireless/intel/iwlwifi/iwl-op-mode.h
164
void (*queue_not_full)(struct iwl_op_mode *op_mode, int queue);
drivers/net/wireless/intel/iwlwifi/iwl-op-mode.h
214
unsigned int queue)
drivers/net/wireless/intel/iwlwifi/iwl-op-mode.h
216
op_mode->ops->rx_rss(op_mode, napi, rxb, queue);
drivers/net/wireless/intel/iwlwifi/iwl-op-mode.h
220
int queue)
drivers/net/wireless/intel/iwlwifi/iwl-op-mode.h
222
op_mode->ops->queue_full(op_mode, queue);
drivers/net/wireless/intel/iwlwifi/iwl-op-mode.h
226
int queue)
drivers/net/wireless/intel/iwlwifi/iwl-op-mode.h
228
op_mode->ops->queue_not_full(op_mode, queue);
drivers/net/wireless/intel/iwlwifi/iwl-trans.c
647
struct iwl_device_tx_cmd *dev_cmd, int queue)
drivers/net/wireless/intel/iwlwifi/iwl-trans.c
657
return iwl_txq_gen2_tx(trans, skb, dev_cmd, queue);
drivers/net/wireless/intel/iwlwifi/iwl-trans.c
659
return iwl_trans_pcie_tx(trans, skb, dev_cmd, queue);
drivers/net/wireless/intel/iwlwifi/iwl-trans.c
663
void iwl_trans_reclaim(struct iwl_trans *trans, int queue, int ssn,
drivers/net/wireless/intel/iwlwifi/iwl-trans.c
673
iwl_pcie_reclaim(trans, queue, ssn, skbs, is_flush);
drivers/net/wireless/intel/iwlwifi/iwl-trans.c
677
void iwl_trans_txq_disable(struct iwl_trans *trans, int queue,
drivers/net/wireless/intel/iwlwifi/iwl-trans.c
680
iwl_trans_pcie_txq_disable(trans, queue, configure_scd);
drivers/net/wireless/intel/iwlwifi/iwl-trans.c
684
bool iwl_trans_txq_enable_cfg(struct iwl_trans *trans, int queue, u16 ssn,
drivers/net/wireless/intel/iwlwifi/iwl-trans.c
694
return iwl_trans_pcie_txq_enable(trans, queue, ssn,
drivers/net/wireless/intel/iwlwifi/iwl-trans.c
699
int iwl_trans_wait_txq_empty(struct iwl_trans *trans, int queue)
drivers/net/wireless/intel/iwlwifi/iwl-trans.c
708
return iwl_trans_pcie_wait_txq_empty(trans, queue);
drivers/net/wireless/intel/iwlwifi/iwl-trans.c
747
void iwl_trans_set_q_ptrs(struct iwl_trans *trans, int queue, int ptr)
drivers/net/wireless/intel/iwlwifi/iwl-trans.c
753
iwl_pcie_set_q_ptrs(trans, queue, ptr);
drivers/net/wireless/intel/iwlwifi/iwl-trans.c
771
void iwl_trans_txq_free(struct iwl_trans *trans, int queue)
drivers/net/wireless/intel/iwlwifi/iwl-trans.c
773
iwl_txq_dyn_free(trans, queue);
drivers/net/wireless/intel/iwlwifi/iwl-trans.c
777
int iwl_trans_get_rxq_dma_data(struct iwl_trans *trans, int queue,
drivers/net/wireless/intel/iwlwifi/iwl-trans.c
780
return iwl_trans_pcie_rxq_dma_data(trans, queue, data);
drivers/net/wireless/intel/iwlwifi/iwl-trans.h
1007
int iwl_trans_wait_txq_empty(struct iwl_trans *trans, int queue);
drivers/net/wireless/intel/iwlwifi/iwl-trans.h
946
struct iwl_device_tx_cmd *dev_cmd, int queue);
drivers/net/wireless/intel/iwlwifi/iwl-trans.h
948
void iwl_trans_reclaim(struct iwl_trans *trans, int queue, int ssn,
drivers/net/wireless/intel/iwlwifi/iwl-trans.h
951
void iwl_trans_set_q_ptrs(struct iwl_trans *trans, int queue, int ptr);
drivers/net/wireless/intel/iwlwifi/iwl-trans.h
953
void iwl_trans_txq_disable(struct iwl_trans *trans, int queue,
drivers/net/wireless/intel/iwlwifi/iwl-trans.h
956
bool iwl_trans_txq_enable_cfg(struct iwl_trans *trans, int queue, u16 ssn,
drivers/net/wireless/intel/iwlwifi/iwl-trans.h
960
int iwl_trans_get_rxq_dma_data(struct iwl_trans *trans, int queue,
drivers/net/wireless/intel/iwlwifi/iwl-trans.h
963
void iwl_trans_txq_free(struct iwl_trans *trans, int queue);
drivers/net/wireless/intel/iwlwifi/iwl-trans.h
971
static inline void iwl_trans_txq_enable(struct iwl_trans *trans, int queue,
drivers/net/wireless/intel/iwlwifi/iwl-trans.h
984
iwl_trans_txq_enable_cfg(trans, queue, ssn, &cfg, queue_wdg_timeout);
drivers/net/wireless/intel/iwlwifi/iwl-trans.h
988
void iwl_trans_ac_txq_enable(struct iwl_trans *trans, int queue, int fifo,
drivers/net/wireless/intel/iwlwifi/iwl-trans.h
999
iwl_trans_txq_enable_cfg(trans, queue, 0, &cfg, queue_wdg_timeout);
drivers/net/wireless/intel/iwlwifi/mei/main.c
293
int dir, queue;
drivers/net/wireless/intel/iwlwifi/mei/main.c
319
for (queue = 0; queue < SAP_QUEUE_IDX_MAX; queue++) {
drivers/net/wireless/intel/iwlwifi/mei/main.c
320
mem->q_head[dir][queue] = q_head;
drivers/net/wireless/intel/iwlwifi/mei/main.c
322
le32_to_cpu(mem->ctrl->dir[dir].q_ctrl_blk[queue].size);
drivers/net/wireless/intel/iwlwifi/mei/main.c
323
mem->q_size[dir][queue] =
drivers/net/wireless/intel/iwlwifi/mei/main.c
324
le32_to_cpu(mem->ctrl->dir[dir].q_ctrl_blk[queue].size);
drivers/net/wireless/intel/iwlwifi/mld/agg.c
100
int queue)
drivers/net/wireless/intel/iwlwifi/mld/agg.c
145
iwl_mld_release_frames_from_notif(mld, napi, baid, nssn, queue);
drivers/net/wireless/intel/iwlwifi/mld/agg.c
150
void iwl_mld_del_ba(struct iwl_mld *mld, int queue,
drivers/net/wireless/intel/iwlwifi/mld/agg.c
17
&baid_data->entries[reorder_buf->queue *
drivers/net/wireless/intel/iwlwifi/mld/agg.c
174
reorder_buf = &ba_data->reorder_buf[queue];
drivers/net/wireless/intel/iwlwifi/mld/agg.c
190
int queue, struct ieee80211_sta *sta,
drivers/net/wireless/intel/iwlwifi/mld/agg.c
252
buffer = &baid_data->reorder_buf[queue];
drivers/net/wireless/intel/iwlwifi/mld/agg.c
253
entries = &baid_data->entries[queue * baid_data->entries_per_queue];
drivers/net/wireless/intel/iwlwifi/mld/agg.c
34
reorder_buf->queue,
drivers/net/wireless/intel/iwlwifi/mld/agg.c
44
u8 baid, u16 nssn, int queue)
drivers/net/wireless/intel/iwlwifi/mld/agg.c
461
reorder_buf->queue = i;
drivers/net/wireless/intel/iwlwifi/mld/agg.c
72
reorder_buf = &ba_data->reorder_buf[queue];
drivers/net/wireless/intel/iwlwifi/mld/agg.c
82
struct iwl_rx_packet *pkt, int queue)
drivers/net/wireless/intel/iwlwifi/mld/agg.c
94
queue);
drivers/net/wireless/intel/iwlwifi/mld/agg.h
109
int queue, struct ieee80211_sta *sta,
drivers/net/wireless/intel/iwlwifi/mld/agg.h
114
struct iwl_rx_packet *pkt, int queue);
drivers/net/wireless/intel/iwlwifi/mld/agg.h
118
int queue);
drivers/net/wireless/intel/iwlwifi/mld/agg.h
120
void iwl_mld_del_ba(struct iwl_mld *mld, int queue,
drivers/net/wireless/intel/iwlwifi/mld/agg.h
21
int queue;
drivers/net/wireless/intel/iwlwifi/mld/low_latency.c
263
u8 queue)
drivers/net/wireless/intel/iwlwifi/mld/low_latency.c
276
queue >= mld->trans->info.num_rxqs))
drivers/net/wireless/intel/iwlwifi/mld/low_latency.c
285
counters = &mld->low_latency.pkts_counters[queue];
drivers/net/wireless/intel/iwlwifi/mld/low_latency.h
64
u8 queue);
drivers/net/wireless/intel/iwlwifi/mld/mld.h
523
struct ieee80211_rx_status *rx_status, int queue);
drivers/net/wireless/intel/iwlwifi/mld/notif.c
617
struct iwl_rx_cmd_buffer *rxb, unsigned int queue)
drivers/net/wireless/intel/iwlwifi/mld/notif.c
623
if (unlikely(queue >= mld->trans->info.num_rxqs))
drivers/net/wireless/intel/iwlwifi/mld/notif.c
627
iwl_mld_rx_mpdu(mld, napi, rxb, queue);
drivers/net/wireless/intel/iwlwifi/mld/notif.c
630
iwl_mld_handle_rx_queues_sync_notif(mld, napi, pkt, queue);
drivers/net/wireless/intel/iwlwifi/mld/notif.c
632
iwl_mld_handle_frame_release_notif(mld, napi, pkt, queue);
drivers/net/wireless/intel/iwlwifi/mld/notif.h
14
struct iwl_rx_cmd_buffer *rxb, unsigned int queue);
drivers/net/wireless/intel/iwlwifi/mld/rx.c
107
res = memcmp(pn, ptk_pn->q[queue].pn[tid], IEEE80211_CCMP_PN_LEN);
drivers/net/wireless/intel/iwlwifi/mld/rx.c
113
memcpy(ptk_pn->q[queue].pn[tid], pn, IEEE80211_CCMP_PN_LEN);
drivers/net/wireless/intel/iwlwifi/mld/rx.c
122
struct sk_buff *skb, int queue,
drivers/net/wireless/intel/iwlwifi/mld/rx.c
126
mld, napi, skb, queue, sta);
drivers/net/wireless/intel/iwlwifi/mld/rx.c
128
if (unlikely(iwl_mld_check_pn(mld, skb, queue, sta))) {
drivers/net/wireless/intel/iwlwifi/mld/rx.c
1638
struct ieee80211_rx_status *rx_status, int queue)
drivers/net/wireless/intel/iwlwifi/mld/rx.c
1652
dup_data = &mld_sta->dup_data[queue];
drivers/net/wireless/intel/iwlwifi/mld/rx.c
1735
const struct iwl_rx_packet *pkt, int queue, bool *drop)
drivers/net/wireless/intel/iwlwifi/mld/rx.c
1781
if (iwl_mld_is_dup(mld, sta, hdr, mpdu_desc, rx_status, queue)) {
drivers/net/wireless/intel/iwlwifi/mld/rx.c
1799
iwl_mld_count_mpdu_rx(link_sta, queue, 1);
drivers/net/wireless/intel/iwlwifi/mld/rx.c
1803
queue);
drivers/net/wireless/intel/iwlwifi/mld/rx.c
1911
struct iwl_rx_mpdu_desc *desc, int queue,
drivers/net/wireless/intel/iwlwifi/mld/rx.c
2021
struct iwl_rx_cmd_buffer *rxb, int queue)
drivers/net/wireless/intel/iwlwifi/mld/rx.c
2089
sta = iwl_mld_rx_with_sta(mld, hdr, skb, mpdu_desc, pkt, queue, &drop);
drivers/net/wireless/intel/iwlwifi/mld/rx.c
2115
if (!queue && unlikely(ieee80211_is_beacon(hdr->frame_control) ||
drivers/net/wireless/intel/iwlwifi/mld/rx.c
2130
if (iwl_mld_rx_crypto(mld, sta, hdr, rx_status, mpdu_desc, queue,
drivers/net/wireless/intel/iwlwifi/mld/rx.c
2143
reorder_res = iwl_mld_reorder(mld, napi, queue, sta, skb, mpdu_desc);
drivers/net/wireless/intel/iwlwifi/mld/rx.c
2156
iwl_mld_pass_packet_to_mac80211(mld, napi, skb, queue, sta);
drivers/net/wireless/intel/iwlwifi/mld/rx.c
2217
struct iwl_rx_packet *pkt, int queue)
drivers/net/wireless/intel/iwlwifi/mld/rx.c
2236
internal_notif->cookie, mld->rxq_sync.cookie, queue))
drivers/net/wireless/intel/iwlwifi/mld/rx.c
2249
iwl_mld_del_ba(mld, queue, (void *)internal_notif->payload);
drivers/net/wireless/intel/iwlwifi/mld/rx.c
2255
IWL_FW_CHECK(mld, !test_and_clear_bit(queue, &mld->rxq_sync.state),
drivers/net/wireless/intel/iwlwifi/mld/rx.c
2256
"RXQ sync: queue %d responded a second time!\n", queue);
drivers/net/wireless/intel/iwlwifi/mld/rx.c
51
int queue, struct ieee80211_sta *sta)
drivers/net/wireless/intel/iwlwifi/mld/rx.c
67
if (queue == 0 || !ieee80211_is_data(hdr->frame_control) ||
drivers/net/wireless/intel/iwlwifi/mld/rx.h
54
struct iwl_rx_cmd_buffer *rxb, int queue);
drivers/net/wireless/intel/iwlwifi/mld/rx.h
62
struct iwl_rx_packet *pkt, int queue);
drivers/net/wireless/intel/iwlwifi/mld/rx.h
66
struct sk_buff *skb, int queue,
drivers/net/wireless/intel/iwlwifi/mld/sta.c
853
static void iwl_mld_count_mpdu(struct ieee80211_link_sta *link_sta, int queue,
drivers/net/wireless/intel/iwlwifi/mld/sta.c
877
queue_counter = &mld_sta->mpdu_counters[queue];
drivers/net/wireless/intel/iwlwifi/mld/sta.c
927
void iwl_mld_count_mpdu_rx(struct ieee80211_link_sta *link_sta, int queue,
drivers/net/wireless/intel/iwlwifi/mld/sta.c
930
iwl_mld_count_mpdu(link_sta, queue, count, false);
drivers/net/wireless/intel/iwlwifi/mld/sta.c
945
int queue, size;
drivers/net/wireless/intel/iwlwifi/mld/sta.c
950
queue = iwl_trans_txq_alloc(mld->trans, 0, sta_mask, tid, size,
drivers/net/wireless/intel/iwlwifi/mld/sta.c
953
if (queue >= 0)
drivers/net/wireless/intel/iwlwifi/mld/sta.c
956
queue, sta_mask, tid);
drivers/net/wireless/intel/iwlwifi/mld/sta.c
957
return queue;
drivers/net/wireless/intel/iwlwifi/mld/sta.h
203
void iwl_mld_count_mpdu_rx(struct ieee80211_link_sta *link_sta, int queue,
drivers/net/wireless/intel/iwlwifi/mld/tests/agg.c
443
struct sk_buff *skb, int queue,
drivers/net/wireless/intel/iwlwifi/mld/tests/agg.c
542
buffer->queue = QUEUE;
drivers/net/wireless/intel/iwlwifi/mld/tx.c
743
int queue = iwl_mld_get_tx_queue_id(mld, txq, skb);
drivers/net/wireless/intel/iwlwifi/mld/tx.c
746
if (WARN_ONCE(queue == IWL_MLD_INVALID_QUEUE, "Invalid TX Queue id") ||
drivers/net/wireless/intel/iwlwifi/mld/tx.c
747
queue == IWL_MLD_INVALID_DROP_TX)
drivers/net/wireless/intel/iwlwifi/mld/tx.c
774
tid, queue, skb->len);
drivers/net/wireless/intel/iwlwifi/mld/tx.c
782
if (iwl_trans_tx(mld->trans, skb, dev_tx_cmd, queue))
drivers/net/wireless/intel/iwlwifi/mld/tx.c
795
IWL_DEBUG_TX(mld, "TX from Q:%d dropped\n", queue);
drivers/net/wireless/intel/iwlwifi/mld/tx.c
80
int queue, size;
drivers/net/wireless/intel/iwlwifi/mld/tx.c
90
queue = iwl_trans_txq_alloc(mld->trans, 0, fw_sta_mask, tid, size,
drivers/net/wireless/intel/iwlwifi/mld/tx.c
93
if (queue >= 0)
drivers/net/wireless/intel/iwlwifi/mld/tx.c
96
queue, fw_sta_mask, tid);
drivers/net/wireless/intel/iwlwifi/mld/tx.c
97
return queue;
drivers/net/wireless/intel/iwlwifi/mvm/mld-sta.c
120
u16 *queue, u8 tid,
drivers/net/wireless/intel/iwlwifi/mvm/mld-sta.c
147
*queue = txq;
drivers/net/wireless/intel/iwlwifi/mvm/mld-sta.c
157
struct iwl_mvm_int_sta *int_sta, u16 *queue,
drivers/net/wireless/intel/iwlwifi/mvm/mld-sta.c
174
queue, tid, wdg_timeout);
drivers/net/wireless/intel/iwlwifi/mvm/mld-sta.c
198
u16 *queue;
drivers/net/wireless/intel/iwlwifi/mvm/mld-sta.c
207
queue = &mvm_link->mgmt_queue;
drivers/net/wireless/intel/iwlwifi/mvm/mld-sta.c
209
queue = &mvm->p2p_dev_queue;
drivers/net/wireless/intel/iwlwifi/mvm/mld-sta.c
215
return iwl_mvm_mld_add_int_sta(mvm, bsta, queue,
drivers/net/wireless/intel/iwlwifi/mvm/mld-sta.c
291
int queue = *queueptr;
drivers/net/wireless/intel/iwlwifi/mvm/mld-sta.c
311
iwl_trans_txq_free(mvm->trans, queue);
drivers/net/wireless/intel/iwlwifi/mvm/mvm.h
1420
static inline bool iwl_mvm_is_dqa_data_queue(struct iwl_mvm *mvm, u8 queue)
drivers/net/wireless/intel/iwlwifi/mvm/mvm.h
1422
return (queue >= IWL_MVM_DQA_MIN_DATA_QUEUE) &&
drivers/net/wireless/intel/iwlwifi/mvm/mvm.h
1423
(queue <= IWL_MVM_DQA_MAX_DATA_QUEUE);
drivers/net/wireless/intel/iwlwifi/mvm/mvm.h
1426
static inline bool iwl_mvm_is_dqa_mgmt_queue(struct iwl_mvm *mvm, u8 queue)
drivers/net/wireless/intel/iwlwifi/mvm/mvm.h
1428
return (queue >= IWL_MVM_DQA_MIN_MGMT_QUEUE) &&
drivers/net/wireless/intel/iwlwifi/mvm/mvm.h
1429
(queue <= IWL_MVM_DQA_MAX_MGMT_QUEUE);
drivers/net/wireless/intel/iwlwifi/mvm/mvm.h
1820
struct iwl_rx_cmd_buffer *rxb, int queue);
drivers/net/wireless/intel/iwlwifi/mvm/mvm.h
1822
struct iwl_rx_cmd_buffer *rxb, int queue);
drivers/net/wireless/intel/iwlwifi/mvm/mvm.h
1824
struct iwl_rx_cmd_buffer *rxb, int queue);
drivers/net/wireless/intel/iwlwifi/mvm/mvm.h
1826
struct iwl_rx_cmd_buffer *rxb, int queue);
drivers/net/wireless/intel/iwlwifi/mvm/mvm.h
1828
struct iwl_rx_cmd_buffer *rxb, int queue);
drivers/net/wireless/intel/iwlwifi/mvm/mvm.h
701
int queue;
drivers/net/wireless/intel/iwlwifi/mvm/mvm.h
757
sizeof(*buf) * buf->queue);
drivers/net/wireless/intel/iwlwifi/mvm/ops.c
1811
static int iwl_mvm_is_static_queue(struct iwl_mvm *mvm, int queue)
drivers/net/wireless/intel/iwlwifi/mvm/ops.c
1813
return queue == mvm->aux_queue || queue == mvm->probe_queue ||
drivers/net/wireless/intel/iwlwifi/mvm/ops.c
1814
queue == mvm->p2p_dev_queue || queue == mvm->snif_queue;
drivers/net/wireless/intel/iwlwifi/mvm/ops.c
2120
unsigned int queue)
drivers/net/wireless/intel/iwlwifi/mvm/ops.c
2126
if (unlikely(queue >= mvm->trans->info.num_rxqs))
drivers/net/wireless/intel/iwlwifi/mvm/ops.c
2130
iwl_mvm_rx_frame_release(mvm, napi, rxb, queue);
drivers/net/wireless/intel/iwlwifi/mvm/ops.c
2133
iwl_mvm_rx_queue_notif(mvm, napi, rxb, queue);
drivers/net/wireless/intel/iwlwifi/mvm/ops.c
2135
iwl_mvm_rx_mpdu_mq(mvm, napi, rxb, queue);
drivers/net/wireless/intel/iwlwifi/mvm/rxmq.c
1130
int queue)
drivers/net/wireless/intel/iwlwifi/mvm/rxmq.c
15
int queue, struct ieee80211_sta *sta)
drivers/net/wireless/intel/iwlwifi/mvm/rxmq.c
1651
int queue)
drivers/net/wireless/intel/iwlwifi/mvm/rxmq.c
1688
if (!queue && !(phy_info & IWL_RX_MPDU_PHY_AMPDU)) {
drivers/net/wireless/intel/iwlwifi/mvm/rxmq.c
1696
if (!queue && (phy_info & IWL_RX_MPDU_PHY_TSF_OVERLOAD) &&
drivers/net/wireless/intel/iwlwifi/mvm/rxmq.c
1794
int queue)
drivers/net/wireless/intel/iwlwifi/mvm/rxmq.c
1830
if (!queue && !(phy_info & IWL_RX_MPDU_PHY_AMPDU)) {
drivers/net/wireless/intel/iwlwifi/mvm/rxmq.c
1839
queue);
drivers/net/wireless/intel/iwlwifi/mvm/rxmq.c
1842
if (!queue && (phy_info & IWL_RX_MPDU_PHY_TSF_OVERLOAD) &&
drivers/net/wireless/intel/iwlwifi/mvm/rxmq.c
1977
int queue)
drivers/net/wireless/intel/iwlwifi/mvm/rxmq.c
2012
iwl_mvm_rx_he(mvm, skb, phy_data, queue);
drivers/net/wireless/intel/iwlwifi/mvm/rxmq.c
2034
iwl_mvm_rx_eht(mvm, skb, phy_data, queue);
drivers/net/wireless/intel/iwlwifi/mvm/rxmq.c
2096
struct iwl_rx_cmd_buffer *rxb, int queue)
drivers/net/wireless/intel/iwlwifi/mvm/rxmq.c
2230
if (!queue && (phy_data.phy_info & IWL_RX_MPDU_PHY_AMPDU)) {
drivers/net/wireless/intel/iwlwifi/mvm/rxmq.c
2270
le32_to_cpu(pkt->len_n_flags), queue,
drivers/net/wireless/intel/iwlwifi/mvm/rxmq.c
2276
iwl_mvm_rx_fill_status(mvm, desc, hdr, skb, &phy_data, queue);
drivers/net/wireless/intel/iwlwifi/mvm/rxmq.c
2343
if (iwl_mvm_is_dup(sta, queue, rx_status, hdr, desc)) {
drivers/net/wireless/intel/iwlwifi/mvm/rxmq.c
2378
if (!queue) {
drivers/net/wireless/intel/iwlwifi/mvm/rxmq.c
238
struct sk_buff *skb, int queue,
drivers/net/wireless/intel/iwlwifi/mvm/rxmq.c
2395
if (!iwl_mvm_reorder(mvm, napi, queue, sta, skb, desc) &&
drivers/net/wireless/intel/iwlwifi/mvm/rxmq.c
2403
iwl_mvm_pass_packet_to_mac80211(mvm, napi, skb, queue, sta);
drivers/net/wireless/intel/iwlwifi/mvm/rxmq.c
241
if (unlikely(iwl_mvm_check_pn(mvm, skb, queue, sta))) {
drivers/net/wireless/intel/iwlwifi/mvm/rxmq.c
2410
struct iwl_rx_cmd_buffer *rxb, int queue)
drivers/net/wireless/intel/iwlwifi/mvm/rxmq.c
2496
iwl_mvm_rx_fill_status(mvm, NULL, NULL, skb, &phy_data, queue);
drivers/net/wireless/intel/iwlwifi/mvm/rxmq.c
2536
struct iwl_rx_cmd_buffer *rxb, int queue)
drivers/net/wireless/intel/iwlwifi/mvm/rxmq.c
2546
queue);
drivers/net/wireless/intel/iwlwifi/mvm/rxmq.c
2550
struct iwl_rx_cmd_buffer *rxb, int queue)
drivers/net/wireless/intel/iwlwifi/mvm/rxmq.c
2595
iwl_mvm_release_frames_from_notif(mvm, napi, baid, nssn, queue);
drivers/net/wireless/intel/iwlwifi/mvm/rxmq.c
418
u32 pkt_flags, int queue, u8 *crypt_len)
drivers/net/wireless/intel/iwlwifi/mvm/rxmq.c
42
if (queue == 0)
drivers/net/wireless/intel/iwlwifi/mvm/rxmq.c
550
static bool iwl_mvm_is_dup(struct ieee80211_sta *sta, int queue,
drivers/net/wireless/intel/iwlwifi/mvm/rxmq.c
567
dup_data = &mvm_sta->dup_data[queue];
drivers/net/wireless/intel/iwlwifi/mvm/rxmq.c
618
&baid_data->entries[reorder_buf->queue *
drivers/net/wireless/intel/iwlwifi/mvm/rxmq.c
638
reorder_buf->queue,
drivers/net/wireless/intel/iwlwifi/mvm/rxmq.c
646
static void iwl_mvm_del_ba(struct iwl_mvm *mvm, int queue,
drivers/net/wireless/intel/iwlwifi/mvm/rxmq.c
670
reorder_buf = &ba_data->reorder_buf[queue];
drivers/net/wireless/intel/iwlwifi/mvm/rxmq.c
685
u8 baid, u16 nssn, int queue)
drivers/net/wireless/intel/iwlwifi/mvm/rxmq.c
717
reorder_buf = &ba_data->reorder_buf[queue];
drivers/net/wireless/intel/iwlwifi/mvm/rxmq.c
729
struct iwl_rx_cmd_buffer *rxb, int queue)
drivers/net/wireless/intel/iwlwifi/mvm/rxmq.c
748
internal_notif->cookie, mvm->queue_sync_cookie, queue))
drivers/net/wireless/intel/iwlwifi/mvm/rxmq.c
760
iwl_mvm_del_ba(mvm, queue, (void *)internal_notif->data);
drivers/net/wireless/intel/iwlwifi/mvm/rxmq.c
767
WARN_ONCE(!test_and_clear_bit(queue, &mvm->queue_sync_state),
drivers/net/wireless/intel/iwlwifi/mvm/rxmq.c
769
queue);
drivers/net/wireless/intel/iwlwifi/mvm/rxmq.c
78
res = memcmp(pn, ptk_pn->q[queue].pn[tid], IEEE80211_CCMP_PN_LEN);
drivers/net/wireless/intel/iwlwifi/mvm/rxmq.c
781
int queue,
drivers/net/wireless/intel/iwlwifi/mvm/rxmq.c
84
memcpy(ptk_pn->q[queue].pn[tid], pn, IEEE80211_CCMP_PN_LEN);
drivers/net/wireless/intel/iwlwifi/mvm/rxmq.c
852
buffer = &baid_data->reorder_buf[queue];
drivers/net/wireless/intel/iwlwifi/mvm/rxmq.c
853
entries = &baid_data->entries[queue * baid_data->entries_per_queue];
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1003
"Failed to configure queue %d on FIFO %d\n", queue, cfg->fifo);
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1008
static void iwl_mvm_change_queue_tid(struct iwl_mvm *mvm, int queue)
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1011
.scd_queue = queue,
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1023
tid_bitmap = mvm->queue_info[queue].tid_bitmap;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1025
if (WARN(!tid_bitmap, "TXQ %d has no tids assigned to it\n", queue))
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1036
queue, ret);
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1040
mvm->queue_info[queue].txq_tid = tid;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1042
queue, tid);
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1045
static void iwl_mvm_unshare_queue(struct iwl_mvm *mvm, int queue)
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1062
sta_id = mvm->queue_info[queue].ra_sta_id;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1063
tid_bitmap = mvm->queue_info[queue].tid_bitmap;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1069
queue, tid_bitmap);
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1073
IWL_DEBUG_TX_QUEUES(mvm, "Unsharing TXQ %d, keeping tid %d\n", queue,
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1087
ret = iwl_mvm_redirect_queue(mvm, queue, tid,
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1092
IWL_ERR(mvm, "Failed to redirect TXQ %d\n", queue);
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1114
queue);
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1117
iwl_trans_txq_set_shared_mode(mvm->trans, queue, false);
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1121
mvm->queue_info[queue].status = IWL_MVM_QUEUE_READY;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1132
struct iwl_mvm_sta *mvmsta, int queue,
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1157
if (tid_bitmap == mvm->queue_info[queue].tid_bitmap) {
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1158
IWL_DEBUG_TX_QUEUES(mvm, "Queue %d is inactive\n", queue);
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1170
mvm->queue_info[queue].tid_bitmap &= ~BIT(tid);
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1172
q_tid_bitmap = mvm->queue_info[queue].tid_bitmap;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1185
if (!(q_tid_bitmap & BIT(mvm->queue_info[queue].txq_tid)))
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1186
set_bit(queue, changetid_queues);
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1190
tid, queue);
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1194
"TXQ #%d left with tid bitmap 0x%x\n", queue,
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1195
mvm->queue_info[queue].tid_bitmap);
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1201
tid_bitmap = mvm->queue_info[queue].tid_bitmap;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1204
if (hweight16(mvm->queue_info[queue].tid_bitmap) == 1 &&
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1205
mvm->queue_info[queue].status == IWL_MVM_QUEUE_SHARED) {
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1207
queue);
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1208
set_bit(queue, unshare_queues);
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1334
int queue = -1;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1354
queue = iwl_mvm_find_free_queue(mvm, mvmsta->deflink.sta_id,
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1357
if (queue >= IWL_MVM_DQA_MIN_MGMT_QUEUE)
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1359
queue);
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1364
if ((queue < 0 && mvmsta->reserved_queue != IEEE80211_INVAL_HW_QUEUE) &&
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1367
queue = mvmsta->reserved_queue;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1368
mvm->queue_info[queue].reserved = true;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1369
IWL_DEBUG_TX_QUEUES(mvm, "Using reserved queue #%d\n", queue);
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1372
if (queue < 0)
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1373
queue = iwl_mvm_find_free_queue(mvm, mvmsta->deflink.sta_id,
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1376
if (queue < 0) {
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1378
queue = iwl_mvm_inactivity_check(mvm, mvmsta->deflink.sta_id);
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1382
if (queue <= 0) {
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1383
queue = iwl_mvm_get_shared_queue(mvm, tfd_queue_mask, ac);
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1384
if (queue > 0) {
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1386
mvm->queue_info[queue].status = IWL_MVM_QUEUE_SHARED;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1396
if (queue > 0 && !shared_queue)
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1397
mvm->queue_info[queue].status = IWL_MVM_QUEUE_READY;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1400
if (WARN_ON(queue <= 0)) {
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1403
return queue;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1412
cfg.aggregate = (queue >= IWL_MVM_DQA_MIN_DATA_QUEUE ||
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1413
queue == IWL_MVM_DQA_BSS_CLIENT_QUEUE);
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1417
shared_queue ? "shared " : "", queue,
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1422
disable_agg_tids = iwl_mvm_get_queue_agg_tids(mvm, queue);
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1426
queue);
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1427
iwl_mvm_invalidate_sta_queue(mvm, queue,
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1432
inc_ssn = iwl_mvm_enable_txq(mvm, sta, queue, ssn, &cfg, wdg_timeout);
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1441
iwl_trans_txq_set_shared_mode(mvm->trans, queue, true);
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1453
mvmsta->tid_data[tid].txq_id = queue;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1454
mvmsta->tfd_queue_msk |= BIT(queue);
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1457
if (mvmsta->reserved_queue == queue)
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1468
ret = iwl_mvm_sta_tx_agg(mvm, sta, tid, queue, true);
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1474
ret = iwl_mvm_redirect_queue(mvm, queue, tid, ac, ssn,
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1484
queue_tmp = queue;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1579
int queue;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1593
queue = IWL_MVM_DQA_BSS_CLIENT_QUEUE;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1595
queue = iwl_mvm_find_free_queue(mvm, mvmsta->deflink.sta_id,
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1598
if (queue < 0) {
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1600
queue = iwl_mvm_inactivity_check(mvm, mvmsta->deflink.sta_id);
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1601
if (queue < 0) {
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1606
mvm->queue_info[queue].status = IWL_MVM_QUEUE_RESERVED;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1608
mvmsta->reserved_queue = queue;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
1611
queue, mvmsta->deflink.sta_id);
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
2175
static void iwl_mvm_enable_aux_snif_queue(struct iwl_mvm *mvm, u16 queue,
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
2190
iwl_mvm_enable_txq(mvm, NULL, queue, 0, &cfg, wdg_timeout);
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
2207
u16 *queue, int fifo)
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
2213
iwl_mvm_enable_aux_snif_queue(mvm, *queue, sta->sta_id, fifo);
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
2218
iwl_mvm_disable_txq(mvm, NULL, sta->sta_id, queue,
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
2236
*queue = txq;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
2340
int queue;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
2357
queue = mvm->probe_queue;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
2359
queue = mvm->p2p_dev_queue;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
2365
bsta->tfd_queue_msk |= BIT(queue);
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
2367
iwl_mvm_enable_txq(mvm, NULL, queue, 0, &cfg, wdg_timeout);
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
2386
queue = iwl_mvm_tvqm_enable_txq(mvm, NULL, bsta->sta_id,
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
2389
if (queue < 0) {
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
2391
return queue;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
2397
mvm->probe_queue = queue;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
2399
mvmvif->deflink.mgmt_queue = queue;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
2401
mvm->p2p_dev_queue = queue;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
2416
u16 *queueptr, queue;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
2437
queue = *queueptr;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
2447
WARN_ON(!(mvmvif->deflink.bcast_sta.tfd_queue_msk & BIT(queue)));
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
2448
mvmvif->deflink.bcast_sta.tfd_queue_msk &= ~BIT(queue);
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
2594
int queue = iwl_mvm_tvqm_enable_txq(mvm, NULL, msta->sta_id,
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
2596
if (queue < 0) {
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
2597
ret = queue;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
2600
mvmvif->deflink.cab_queue = queue;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
2750
reorder_buf->queue = i;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
3010
int tid, u8 queue, bool start)
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
3020
mvm_sta->tfd_queue_msk |= BIT(queue);
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
303
static int iwl_mvm_invalidate_sta_queue(struct iwl_mvm *mvm, int queue,
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
316
sta_id = mvm->queue_info[queue].ra_sta_id;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
3192
int queue, ret;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
3216
queue = tid_data->txq_id;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
3238
ret = iwl_mvm_sta_tx_agg(mvm, sta, tid, queue, true);
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
3246
queue_status = mvm->queue_info[queue].status;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
3249
if (mvm->queue_info[queue].status == IWL_MVM_QUEUE_READY)
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
3262
BIT(queue));
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
3269
ret = iwl_mvm_reconfig_scd(mvm, queue, cfg.fifo,
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
3274
"Error reconfiguring TXQ #%d\n", queue);
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
3280
iwl_mvm_enable_txq(mvm, sta, queue, ssn,
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
3285
ret = iwl_mvm_sta_tx_agg(mvm, sta, tid, queue, true);
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
3291
mvm->queue_info[queue].status = IWL_MVM_QUEUE_READY;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
354
int queue = *queueptr;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
356
.scd_queue = queue,
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
384
iwl_trans_txq_free(mvm->trans, queue);
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
390
if (WARN_ON(mvm->queue_info[queue].tid_bitmap == 0))
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
393
mvm->queue_info[queue].tid_bitmap &= ~BIT(tid);
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
395
cmd.action = mvm->queue_info[queue].tid_bitmap ?
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
398
mvm->queue_info[queue].status = IWL_MVM_QUEUE_FREE;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
402
queue,
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
403
mvm->queue_info[queue].tid_bitmap);
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
409
cmd.sta_id = mvm->queue_info[queue].ra_sta_id;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
410
cmd.tid = mvm->queue_info[queue].txq_tid;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
413
WARN(mvm->queue_info[queue].tid_bitmap,
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
415
queue, mvm->queue_info[queue].tid_bitmap);
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
418
mvm->queue_info[queue].tid_bitmap = 0;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
432
mvm->queue_info[queue].reserved = false;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
434
iwl_trans_txq_disable(mvm->trans, queue, false);
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
440
queue, ret);
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
444
static int iwl_mvm_get_queue_agg_tids(struct iwl_mvm *mvm, int queue)
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
458
sta_id = mvm->queue_info[queue].ra_sta_id;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
459
tid_bitmap = mvm->queue_info[queue].tid_bitmap;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
484
static int iwl_mvm_remove_sta_queue_marking(struct iwl_mvm *mvm, int queue)
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
498
sta_id = mvm->queue_info[queue].ra_sta_id;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
499
tid_bitmap = mvm->queue_info[queue].tid_bitmap;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
529
mvmsta->tfd_queue_msk &= ~BIT(queue); /* Don't use this queue anymore */
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
547
static int iwl_mvm_free_inactive_queue(struct iwl_mvm *mvm, int queue,
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
555
u16 queue_tmp = queue;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
563
sta_id = mvm->queue_info[queue].ra_sta_id;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
564
tid = mvm->queue_info[queue].txq_tid;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
572
disable_agg_tids = iwl_mvm_remove_sta_queue_marking(mvm, queue);
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
575
iwl_mvm_invalidate_sta_queue(mvm, queue,
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
582
queue, ret);
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
589
iwl_mvm_invalidate_sta_queue(mvm, queue, 0, true);
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
597
int queue = 0;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
633
queue = ac_to_queue[IEEE80211_AC_BE];
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
636
queue = ac_to_queue[ac];
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
640
queue = ac_to_queue[IEEE80211_AC_VI];
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
643
queue = ac_to_queue[IEEE80211_AC_BK];
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
646
queue = ac_to_queue[IEEE80211_AC_VI];
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
649
queue = ac_to_queue[IEEE80211_AC_VO];
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
652
if (!iwl_mvm_is_dqa_data_queue(mvm, queue) &&
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
653
!iwl_mvm_is_dqa_mgmt_queue(mvm, queue) &&
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
654
(queue != IWL_MVM_DQA_BSS_CLIENT_QUEUE)) {
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
659
return queue;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
663
static int iwl_mvm_reconfig_scd(struct iwl_mvm *mvm, int queue, int fifo,
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
667
.scd_queue = queue,
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
673
.aggregate = (queue >= IWL_MVM_DQA_MIN_DATA_QUEUE ||
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
674
queue == IWL_MVM_DQA_BSS_CLIENT_QUEUE),
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
682
if (WARN(mvm->queue_info[queue].tid_bitmap == 0,
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
683
"Trying to reconfig unallocated queue %d\n", queue))
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
686
IWL_DEBUG_TX_QUEUES(mvm, "Reconfig SCD for TXQ #%d\n", queue);
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
690
queue, fifo, ret);
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
701
static int iwl_mvm_redirect_queue(struct iwl_mvm *mvm, int queue, int tid,
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
706
.scd_queue = queue,
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
723
if (ac <= mvm->queue_info[queue].mac80211_ac && !force) {
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
726
queue);
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
730
cmd.sta_id = mvm->queue_info[queue].ra_sta_id;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
731
cmd.tx_fifo = iwl_mvm_ac_to_tx_fifo[mvm->queue_info[queue].mac80211_ac];
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
732
cmd.tid = mvm->queue_info[queue].txq_tid;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
733
shared_queue = hweight16(mvm->queue_info[queue].tid_bitmap) > 1;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
736
queue, iwl_mvm_ac_to_tx_fifo[ac]);
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
741
ret = iwl_trans_wait_tx_queues_empty(mvm->trans, BIT(queue));
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
744
queue);
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
750
iwl_trans_txq_disable(mvm->trans, queue, false);
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
753
IWL_ERR(mvm, "Failed SCD disable TXQ %d (ret=%d)\n", queue,
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
757
iwl_trans_txq_enable_cfg(mvm->trans, queue, ssn, NULL, wdg_timeout);
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
760
mvm->queue_info[queue].txq_tid = tid;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
765
iwl_mvm_reconfig_scd(mvm, queue, iwl_mvm_ac_to_tx_fifo[ac],
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
769
mvm->queue_info[queue].mac80211_ac = ac;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
778
iwl_trans_txq_set_shared_mode(mvm->trans, queue, true);
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
849
int queue, size;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
884
queue = iwl_trans_txq_alloc(mvm->trans, 0, sta_mask,
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
887
if (queue >= 0)
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
890
queue, sta_mask, tid);
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
892
return queue;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
904
int queue = -1;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
911
queue = iwl_mvm_tvqm_enable_txq(mvm, sta, mvmsta->deflink.sta_id,
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
913
if (queue < 0)
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
914
return queue;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
916
mvmtxq->txq_id = queue;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
917
mvm->tvqm_info[queue].txq_tid = tid;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
918
mvm->tvqm_info[queue].sta_id = mvmsta->deflink.sta_id;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
920
IWL_DEBUG_TX_QUEUES(mvm, "Allocated queue is %d\n", queue);
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
923
mvmsta->tid_data[tid].txq_id = queue;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
931
int queue, u8 sta_id, u8 tid)
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
936
if (mvm->queue_info[queue].tid_bitmap & BIT(tid)) {
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
938
queue, tid);
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
943
if (mvm->queue_info[queue].tid_bitmap)
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
946
mvm->queue_info[queue].tid_bitmap |= BIT(tid);
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
947
mvm->queue_info[queue].ra_sta_id = sta_id;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
951
mvm->queue_info[queue].mac80211_ac =
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
954
mvm->queue_info[queue].mac80211_ac = IEEE80211_AC_VO;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
956
mvm->queue_info[queue].txq_tid = tid;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
963
mvmtxq->txq_id = queue;
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
968
queue, mvm->queue_info[queue].tid_bitmap);
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
974
int queue, u16 ssn,
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
979
.scd_queue = queue,
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
994
if (!iwl_mvm_update_txq_mapping(mvm, sta, queue, cfg->sta_id, cfg->tid))
drivers/net/wireless/intel/iwlwifi/mvm/sta.c
997
inc_ssn = iwl_trans_txq_enable_cfg(mvm->trans, queue, ssn,
drivers/net/wireless/intel/iwlwifi/mvm/sta.h
529
int tid, u8 queue, bool start);
drivers/net/wireless/intel/iwlwifi/mvm/sta.h
649
u16 *queue, u8 tid,
drivers/net/wireless/intel/iwlwifi/mvm/tx.c
1847
int queue = SEQ_TO_QUEUE(sequence);
drivers/net/wireless/intel/iwlwifi/mvm/tx.c
1850
if (WARN_ON_ONCE(queue < IWL_MVM_DQA_MIN_DATA_QUEUE &&
drivers/net/wireless/intel/iwlwifi/mvm/tx.c
1851
(queue != IWL_MVM_DQA_BSS_CLIENT_QUEUE)))
drivers/net/wireless/intel/iwlwifi/mvm/tx.c
780
int queue = -1;
drivers/net/wireless/intel/iwlwifi/mvm/tx.c
813
queue = mvm->aux_queue;
drivers/net/wireless/intel/iwlwifi/mvm/tx.c
814
if (WARN_ON(queue == IWL_MVM_INVALID_QUEUE))
drivers/net/wireless/intel/iwlwifi/mvm/tx.c
826
queue = iwl_mvm_get_ctrl_vif_queue(mvm,
drivers/net/wireless/intel/iwlwifi/mvm/tx.c
830
queue = mvm->snif_queue;
drivers/net/wireless/intel/iwlwifi/mvm/tx.c
835
if (queue < 0) {
drivers/net/wireless/intel/iwlwifi/mvm/tx.c
843
IWL_DEBUG_TX(mvm, "station Id %d, queue=%d\n", sta_id, queue);
drivers/net/wireless/intel/iwlwifi/mvm/tx.c
853
if (iwl_trans_tx(mvm->trans, skb, dev_cmd, queue)) {
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/internal.h
1153
int iwl_trans_pcie_rxq_dma_data(struct iwl_trans *trans, int queue,
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/internal.h
164
struct iwl_rx_mem_buffer *queue[RX_QUEUE_SIZE];
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/internal.h
528
static inline void iwl_pcie_clear_irq(struct iwl_trans *trans, int queue)
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/internal.h
538
iwl_write32(trans, CSR_MSIX_AUTOMASK_ST_AD, BIT(queue));
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/internal.h
621
bool iwl_trans_pcie_txq_enable(struct iwl_trans *trans, int queue, u16 ssn,
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/internal.h
624
void iwl_trans_pcie_txq_disable(struct iwl_trans *trans, int queue,
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/internal.h
768
void iwl_txq_dyn_free(struct iwl_trans *trans, int queue);
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/rx.c
1107
def_rxq->queue[i] = NULL;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/rx.c
1455
rxb = rxq->queue[i];
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/rx.c
1456
rxq->queue[i] = NULL;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/rx.c
1498
static int iwl_pcie_rx_handle(struct iwl_trans *trans, int queue, int budget)
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/rx.c
1505
if (WARN_ON_ONCE(!trans_pcie->rxq || !trans_pcie->rxq[queue].bd))
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/rx.c
1508
rxq = &trans_pcie->rxq[queue];
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/rx.c
1629
u8 queue = entry->entry;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/rx.c
1630
struct msix_entry *entries = entry - queue;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/rx.c
312
rxb = rxq->queue[rxq->write];
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/rx.c
323
rxq->queue[rxq->write] = rxb;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/rx.c
998
static int iwl_pcie_rx_handle(struct iwl_trans *trans, int queue, int budget);
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/trans.c
2518
int iwl_trans_pcie_rxq_dma_data(struct iwl_trans *trans, int queue,
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/trans.c
2523
if (queue >= trans->info.num_rxqs || !trans_pcie->rxq)
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/trans.c
2526
data->fr_bd_cb = trans_pcie->rxq[queue].bd_dma;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/trans.c
2527
data->urbd_stts_wrptr = trans_pcie->rxq[queue].rb_stts_dma;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/trans.c
2528
data->ur_bd_cb = trans_pcie->rxq[queue].used_bd_dma;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/trans.c
3292
struct iwl_rx_mem_buffer *rxb = rxq->queue[i];
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx-gen2.c
1105
void iwl_txq_dyn_free(struct iwl_trans *trans, int queue)
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx-gen2.c
1109
if (WARN(queue >= IWL_MAX_TVQM_QUEUES,
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx-gen2.c
1110
"queue %d out of range", queue))
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx-gen2.c
1119
if (!test_and_clear_bit(queue, trans_pcie->txqs.queue_used)) {
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx-gen2.c
1121
"queue %d not used", queue);
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx-gen2.c
1125
iwl_txq_gen2_free(trans, queue);
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx-gen2.c
1127
IWL_DEBUG_TX_QUEUES(trans, "Deactivate queue %d\n", queue);
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx-gen2.c
1150
struct iwl_txq *queue;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx-gen2.c
1155
queue = kzalloc_obj(*queue);
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx-gen2.c
1156
if (!queue) {
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx-gen2.c
1160
trans_pcie->txqs.txq[txq_id] = queue;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx-gen2.c
1161
ret = iwl_pcie_txq_alloc(trans, queue, queue_size, true);
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx-gen2.c
1167
queue = trans_pcie->txqs.txq[txq_id];
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx-gen2.c
1170
ret = iwl_txq_init(trans, queue, queue_size,
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
2499
int queue;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
2501
for_each_set_bit(queue, &txqs, BITS_PER_LONG) {
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
2502
struct iwl_txq *txq = trans_pcie->txqs.txq[queue];
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/tx.c
2513
freeze ? "Freezing" : "Waking", queue);
drivers/net/wireless/intersil/p54/fwio.c
590
memcpy(edcf->queue, priv->qos_params, sizeof(edcf->queue));
drivers/net/wireless/intersil/p54/lmac.h
404
struct p54_edcf_queue_param queue[8];
drivers/net/wireless/intersil/p54/lmac.h
448
u8 queue;
drivers/net/wireless/intersil/p54/main.c
410
unsigned int link_id, u16 queue,
drivers/net/wireless/intersil/p54/main.c
417
P54_SET_QUEUE(priv->qos_params[queue], params->aifs,
drivers/net/wireless/intersil/p54/p54.h
101
#define P54_SET_QUEUE(queue, ai_fs, cw_min, cw_max, _txop) \
drivers/net/wireless/intersil/p54/p54.h
103
queue.aifs = cpu_to_le16(ai_fs); \
drivers/net/wireless/intersil/p54/p54.h
104
queue.cwmin = cpu_to_le16(cw_min); \
drivers/net/wireless/intersil/p54/p54.h
105
queue.cwmax = cpu_to_le16(cw_max); \
drivers/net/wireless/intersil/p54/p54.h
106
queue.txop = cpu_to_le16(_txop); \
drivers/net/wireless/intersil/p54/txrx.c
189
struct p54_tx_queue_stats *queue;
drivers/net/wireless/intersil/p54/txrx.c
195
queue = &priv->tx_stats[p54_queue];
drivers/net/wireless/intersil/p54/txrx.c
198
if (unlikely(queue->len >= queue->limit && IS_QOS_QUEUE(p54_queue))) {
drivers/net/wireless/intersil/p54/txrx.c
203
queue->len++;
drivers/net/wireless/intersil/p54/txrx.c
204
queue->count++;
drivers/net/wireless/intersil/p54/txrx.c
206
if (unlikely(queue->len == queue->limit && IS_QOS_QUEUE(p54_queue))) {
drivers/net/wireless/intersil/p54/txrx.c
691
u8 *queue, u32 *extra_len, u16 *flags, u16 *aid,
drivers/net/wireless/intersil/p54/txrx.c
710
*queue = skb_get_queue_mapping(skb) + P54_QUEUE_DATA;
drivers/net/wireless/intersil/p54/txrx.c
730
*queue = P54_QUEUE_CAB;
drivers/net/wireless/intersil/p54/txrx.c
754
*queue = P54_QUEUE_BEACON;
drivers/net/wireless/intersil/p54/txrx.c
793
u8 rate, queue = 0, crypt_offset = 0;
drivers/net/wireless/intersil/p54/txrx.c
800
p54_tx_80211_header(priv, skb, info, control->sta, &queue, &extra_len,
drivers/net/wireless/intersil/p54/txrx.c
803
if (p54_tx_qos_accounting_alloc(priv, skb, queue)) {
drivers/net/wireless/intersil/p54/txrx.c
927
txhdr->hw_queue = queue;
drivers/net/wireless/intersil/p54/txrx.c
928
txhdr->backlog = priv->tx_stats[queue].len - 1;
drivers/net/wireless/marvell/mwl8k.c
5441
unsigned int link_id, u16 queue,
drivers/net/wireless/marvell/mwl8k.c
5449
BUG_ON(queue > MWL8K_TX_WMM_QUEUES - 1);
drivers/net/wireless/marvell/mwl8k.c
5450
memcpy(&priv->wmm_params[queue], params, sizeof(*params));
drivers/net/wireless/marvell/mwl8k.c
5456
int q = MWL8K_TX_WMM_QUEUES - 1 - queue;
drivers/net/wireless/mediatek/mt76/mt7603/mac.c
184
int i, port, queue;
drivers/net/wireless/mediatek/mt76/mt7603/mac.c
188
queue = 8; /* free queue */
drivers/net/wireless/mediatek/mt76/mt7603/mac.c
191
queue = 1; /* MCU queue */
drivers/net/wireless/mediatek/mt76/mt7603/mac.c
216
FIELD_PREP(MT_DMA_FQCR0_DEST_QUEUE_ID, queue));
drivers/net/wireless/mediatek/mt76/mt7603/main.c
525
unsigned int link_id, u16 queue,
drivers/net/wireless/mediatek/mt76/mt7603/main.c
533
queue = dev->mphy.q_tx[queue]->hw_idx;
drivers/net/wireless/mediatek/mt76/mt7603/main.c
543
val = mt76_rr(dev, MT_WMM_TXOP(queue));
drivers/net/wireless/mediatek/mt76/mt7603/main.c
544
val &= ~(MT_WMM_TXOP_MASK << MT_WMM_TXOP_SHIFT(queue));
drivers/net/wireless/mediatek/mt76/mt7603/main.c
545
val |= params->txop << MT_WMM_TXOP_SHIFT(queue);
drivers/net/wireless/mediatek/mt76/mt7603/main.c
546
mt76_wr(dev, MT_WMM_TXOP(queue), val);
drivers/net/wireless/mediatek/mt76/mt7603/main.c
549
val &= ~(MT_WMM_AIFSN_MASK << MT_WMM_AIFSN_SHIFT(queue));
drivers/net/wireless/mediatek/mt76/mt7603/main.c
550
val |= params->aifs << MT_WMM_AIFSN_SHIFT(queue);
drivers/net/wireless/mediatek/mt76/mt7603/main.c
554
val &= ~(MT_WMM_CWMIN_MASK << MT_WMM_CWMIN_SHIFT(queue));
drivers/net/wireless/mediatek/mt76/mt7603/main.c
555
val |= cw_min << MT_WMM_CWMIN_SHIFT(queue);
drivers/net/wireless/mediatek/mt76/mt7603/main.c
558
val = mt76_rr(dev, MT_WMM_CWMAX(queue));
drivers/net/wireless/mediatek/mt76/mt7603/main.c
559
val &= ~(MT_WMM_CWMAX_MASK << MT_WMM_CWMAX_SHIFT(queue));
drivers/net/wireless/mediatek/mt76/mt7603/main.c
560
val |= cw_max << MT_WMM_CWMAX_SHIFT(queue);
drivers/net/wireless/mediatek/mt76/mt7603/main.c
561
mt76_wr(dev, MT_WMM_CWMAX(queue), val);
drivers/net/wireless/mediatek/mt76/mt7615/debugfs.c
396
char *queue;
drivers/net/wireless/mediatek/mt76/mt7615/debugfs.c
409
queue_map[i].queue, q->queued, q->head,
drivers/net/wireless/mediatek/mt76/mt7615/main.c
462
unsigned int link_id, u16 queue,
drivers/net/wireless/mediatek/mt76/mt7615/main.c
471
queue = mt7615_lmac_mapping(dev, queue);
drivers/net/wireless/mediatek/mt76/mt7615/main.c
472
queue += mvif->wmm_idx * MT7615_MAX_WMM_SETS;
drivers/net/wireless/mediatek/mt76/mt7615/main.c
473
err = mt7615_mcu_set_wmm(dev, queue, params);
drivers/net/wireless/mediatek/mt76/mt7615/mcu.c
1789
int mt7615_mcu_set_wmm(struct mt7615_dev *dev, u8 queue,
drivers/net/wireless/mediatek/mt76/mt7615/mcu.c
1801
u8 queue;
drivers/net/wireless/mediatek/mt76/mt7615/mcu.c
1809
.queue = queue,
drivers/net/wireless/mediatek/mt76/mt7615/mt7615.h
404
int mt7615_mcu_set_wmm(struct mt7615_dev *dev, u8 queue,
drivers/net/wireless/mediatek/mt76/mt76x02.h
173
unsigned int link_id, u16 queue,
drivers/net/wireless/mediatek/mt76/mt76x02_util.c
488
unsigned int link_id, u16 queue,
drivers/net/wireless/mediatek/mt76/mt76x02_util.c
495
qid = dev->mphy.q_tx[queue]->hw_idx;
drivers/net/wireless/mediatek/mt76/mt7915/debugfs.c
949
char *queue;
drivers/net/wireless/mediatek/mt76/mt7915/debugfs.c
966
queue_map[i].queue, q->queued, q->head,
drivers/net/wireless/mediatek/mt76/mt7915/mac.c
168
u8 queue = mt76_connac_lmac_mapping(i);
drivers/net/wireless/mediatek/mt76/mt7915/mac.c
169
u32 tx_cur = tx_time[queue];
drivers/net/wireless/mediatek/mt76/mt7915/mac.c
170
u32 rx_cur = rx_time[queue];
drivers/net/wireless/mediatek/mt76/mt7915/main.c
1740
path->mtk_wdma.queue = phy != &dev->phy;
drivers/net/wireless/mediatek/mt76/mt7915/main.c
509
unsigned int link_id, u16 queue,
drivers/net/wireless/mediatek/mt76/mt7915/main.c
515
queue = mt76_connac_lmac_mapping(queue);
drivers/net/wireless/mediatek/mt76/mt7915/main.c
516
mvif->queue_params[queue] = *params;
drivers/net/wireless/mediatek/mt76/mt7915/mcu.c
2553
e->queue = ac + mvif->mt76.wmm_idx * MT76_CONNAC_MAX_WMM_SETS;
drivers/net/wireless/mediatek/mt76/mt7915/mcu.h
195
u8 queue;
drivers/net/wireless/mediatek/mt76/mt7915/testmode.c
178
e->queue = qid + mvif->mt76.wmm_idx * MT76_CONNAC_MAX_WMM_SETS;
drivers/net/wireless/mediatek/mt76/mt7925/main.c
1682
unsigned int link_id, u16 queue,
drivers/net/wireless/mediatek/mt76/mt7925/main.c
1695
mconf->queue_params[mq_to_aci[queue]] = *params;
drivers/net/wireless/mediatek/mt76/mt7925/mcu.c
1171
e->queue = ac;
drivers/net/wireless/mediatek/mt76/mt7925/mcu.h
338
u8 queue;
drivers/net/wireless/mediatek/mt76/mt792x.h
385
unsigned int link_id, u16 queue,
drivers/net/wireless/mediatek/mt76/mt792x_core.c
210
unsigned int link_id, u16 queue,
drivers/net/wireless/mediatek/mt76/mt792x_core.c
216
queue = mt76_connac_lmac_mapping(queue);
drivers/net/wireless/mediatek/mt76/mt792x_core.c
217
mvif->bss_conf.queue_params[queue] = *params;
drivers/net/wireless/mediatek/mt76/mt792x_debugfs.c
103
char *queue;
drivers/net/wireless/mediatek/mt76/mt792x_debugfs.c
119
queue_map[i].queue, q->queued, q->head,
drivers/net/wireless/mediatek/mt76/mt7996/debugfs.c
761
char *queue;
drivers/net/wireless/mediatek/mt76/mt7996/debugfs.c
788
queue_map[i].queue, q->queued, q->head,
drivers/net/wireless/mediatek/mt76/mt7996/dma.c
179
u8 queue, val;
drivers/net/wireless/mediatek/mt76/mt7996/dma.c
204
queue = is_mt7996(&dev->mt76) ? MT_RXQ_BAND2_WA : MT_RXQ_BAND1_WA;
drivers/net/wireless/mediatek/mt76/mt7996/dma.c
205
mt76_wr(dev, MT_RXQ_EXT_CTRL(queue) + ofs, PREFETCH(val));
drivers/net/wireless/mediatek/mt76/mt7996/dma.c
220
queue = is_mt7996(&dev->mt76) ? MT_RXQ_BAND2 : MT_RXQ_BAND1;
drivers/net/wireless/mediatek/mt76/mt7996/dma.c
221
mt76_wr(dev, MT_RXQ_EXT_CTRL(queue) + ofs, PREFETCH(0x10));
drivers/net/wireless/mediatek/mt76/mt7996/dma.c
226
queue = is_mt7996(&dev->mt76) ? MT_RXQ_RRO_BAND2 : MT_RXQ_RRO_BAND1;
drivers/net/wireless/mediatek/mt76/mt7996/dma.c
227
mt76_wr(dev, MT_RXQ_EXT_CTRL(queue) + ofs, PREFETCH(0x10));
drivers/net/wireless/mediatek/mt76/mt7996/main.c
2209
path->mtk_wdma.queue = 0;
drivers/net/wireless/mediatek/mt76/mt7996/main.c
60
u8 queue = mt76_connac_lmac_mapping(IEEE80211_AC_VI);
drivers/net/wireless/mediatek/mt76/mt7996/main.c
62
ret = mt7996_mcu_cp_support(dev, queue);
drivers/net/wireless/mediatek/mt76/mt7996/main.c
657
unsigned int link_id, u16 queue,
drivers/net/wireless/mediatek/mt76/mt7996/main.c
670
link_info->queue_params[mq_to_aci[queue]] = *params;
drivers/net/wireless/mediatek/mt76/mt7996/mcu.c
3462
e->queue = ac;
drivers/net/wireless/mediatek/mt76/mt7996/mcu.h
288
u8 queue;
drivers/net/wireless/mediatek/mt7601u/mt7601u.h
371
unsigned int link_id, u16 queue,
drivers/net/wireless/mediatek/mt7601u/tx.c
261
unsigned int link_id, u16 queue,
drivers/net/wireless/mediatek/mt7601u/tx.c
265
u8 cw_min = 5, cw_max = 10, hw_q = q2hwq(queue);
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
1168
struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
1210
struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
1254
entry->queue->rt2x00dev->rssi_offset;
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
1268
struct data_queue *queue = rt2x00queue_get_tx_queue(rt2x00dev, queue_idx);
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
1274
while (!rt2x00queue_empty(queue)) {
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
1275
entry = rt2x00queue_get_entry(queue, Q_INDEX_DONE);
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
1658
unsigned int link_id, u16 queue,
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
1668
if (queue != 0)
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
1671
if (rt2x00mac_conf_tx(hw, vif, link_id, queue, params))
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
1764
static void rt2400pci_queue_init(struct data_queue *queue)
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
1766
switch (queue->qid) {
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
1768
queue->limit = 24;
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
1769
queue->data_size = DATA_FRAME_SIZE;
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
1770
queue->desc_size = RXD_DESC_SIZE;
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
1771
queue->priv_size = sizeof(struct queue_entry_priv_mmio);
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
1778
queue->limit = 24;
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
1779
queue->data_size = DATA_FRAME_SIZE;
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
1780
queue->desc_size = TXD_DESC_SIZE;
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
1781
queue->priv_size = sizeof(struct queue_entry_priv_mmio);
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
1785
queue->limit = 1;
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
1786
queue->data_size = MGMT_FRAME_SIZE;
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
1787
queue->desc_size = TXD_DESC_SIZE;
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
1788
queue->priv_size = sizeof(struct queue_entry_priv_mmio);
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
1792
queue->limit = 8;
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
1793
queue->data_size = DATA_FRAME_SIZE;
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
1794
queue->desc_size = TXD_DESC_SIZE;
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
1795
queue->priv_size = sizeof(struct queue_entry_priv_mmio);
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
627
static void rt2400pci_start_queue(struct data_queue *queue)
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
629
struct rt2x00_dev *rt2x00dev = queue->rt2x00dev;
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
632
switch (queue->qid) {
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
650
static void rt2400pci_kick_queue(struct data_queue *queue)
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
652
struct rt2x00_dev *rt2x00dev = queue->rt2x00dev;
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
655
switch (queue->qid) {
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
676
static void rt2400pci_stop_queue(struct data_queue *queue)
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
678
struct rt2x00_dev *rt2x00dev = queue->rt2x00dev;
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
681
switch (queue->qid) {
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
719
if (entry->queue->qid == QID_RX) {
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
737
if (entry->queue->qid == QID_RX) {
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
1264
rt2x00_set_field32(&word, TXD_W2_AIFS, entry->queue->aifs);
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
1265
rt2x00_set_field32(&word, TXD_W2_CWMIN, entry->queue->cw_min);
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
1266
rt2x00_set_field32(&word, TXD_W2_CWMAX, entry->queue->cw_max);
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
1320
struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
1379
entry->queue->rt2x00dev->rssi_offset;
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
1396
struct data_queue *queue = rt2x00queue_get_tx_queue(rt2x00dev, queue_idx);
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
1402
while (!rt2x00queue_empty(queue)) {
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
1403
entry = rt2x00queue_get_entry(queue, Q_INDEX_DONE);
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
2062
static void rt2500pci_queue_init(struct data_queue *queue)
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
2064
switch (queue->qid) {
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
2066
queue->limit = 32;
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
2067
queue->data_size = DATA_FRAME_SIZE;
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
2068
queue->desc_size = RXD_DESC_SIZE;
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
2069
queue->priv_size = sizeof(struct queue_entry_priv_mmio);
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
2076
queue->limit = 32;
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
2077
queue->data_size = DATA_FRAME_SIZE;
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
2078
queue->desc_size = TXD_DESC_SIZE;
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
2079
queue->priv_size = sizeof(struct queue_entry_priv_mmio);
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
2083
queue->limit = 1;
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
2084
queue->data_size = MGMT_FRAME_SIZE;
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
2085
queue->desc_size = TXD_DESC_SIZE;
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
2086
queue->priv_size = sizeof(struct queue_entry_priv_mmio);
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
2090
queue->limit = 8;
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
2091
queue->data_size = DATA_FRAME_SIZE;
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
2092
queue->desc_size = TXD_DESC_SIZE;
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
2093
queue->priv_size = sizeof(struct queue_entry_priv_mmio);
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
286
struct data_queue *queue = rt2x00dev->bcn;
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
297
rt2x00_set_field32(®, BCNCSR1_BEACON_CWMIN, queue->cw_min);
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
716
static void rt2500pci_start_queue(struct data_queue *queue)
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
718
struct rt2x00_dev *rt2x00dev = queue->rt2x00dev;
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
721
switch (queue->qid) {
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
739
static void rt2500pci_kick_queue(struct data_queue *queue)
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
741
struct rt2x00_dev *rt2x00dev = queue->rt2x00dev;
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
744
switch (queue->qid) {
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
765
static void rt2500pci_stop_queue(struct data_queue *queue)
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
767
struct rt2x00_dev *rt2x00dev = queue->rt2x00dev;
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
770
switch (queue->qid) {
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
808
if (entry->queue->qid == QID_RX) {
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
826
if (entry->queue->qid == QID_RX) {
drivers/net/wireless/ralink/rt2x00/rt2500usb.c
1086
rt2x00_set_field32(&word, TXD_W1_AIFS, entry->queue->aifs);
drivers/net/wireless/ralink/rt2x00/rt2500usb.c
1087
rt2x00_set_field32(&word, TXD_W1_CWMIN, entry->queue->cw_min);
drivers/net/wireless/ralink/rt2x00/rt2500usb.c
1088
rt2x00_set_field32(&word, TXD_W1_CWMAX, entry->queue->cw_max);
drivers/net/wireless/ralink/rt2x00/rt2500usb.c
1121
struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
drivers/net/wireless/ralink/rt2x00/rt2500usb.c
1124
int pipe = usb_sndbulkpipe(usb_dev, entry->queue->usb_endpoint);
drivers/net/wireless/ralink/rt2x00/rt2500usb.c
1208
length += (2 * !(length % entry->queue->usb_maxpacket));
drivers/net/wireless/ralink/rt2x00/rt2500usb.c
1219
struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
drivers/net/wireless/ralink/rt2x00/rt2500usb.c
1225
entry->queue->desc_size));
drivers/net/wireless/ralink/rt2x00/rt2500usb.c
1297
if (!test_bit(DEVICE_STATE_ENABLED_RADIO, &entry->queue->rt2x00dev->flags))
drivers/net/wireless/ralink/rt2x00/rt2500usb.c
1851
static void rt2500usb_queue_init(struct data_queue *queue)
drivers/net/wireless/ralink/rt2x00/rt2500usb.c
1853
switch (queue->qid) {
drivers/net/wireless/ralink/rt2x00/rt2500usb.c
1855
queue->limit = 32;
drivers/net/wireless/ralink/rt2x00/rt2500usb.c
1856
queue->data_size = DATA_FRAME_SIZE;
drivers/net/wireless/ralink/rt2x00/rt2500usb.c
1857
queue->desc_size = RXD_DESC_SIZE;
drivers/net/wireless/ralink/rt2x00/rt2500usb.c
1858
queue->priv_size = sizeof(struct queue_entry_priv_usb);
drivers/net/wireless/ralink/rt2x00/rt2500usb.c
1865
queue->limit = 32;
drivers/net/wireless/ralink/rt2x00/rt2500usb.c
1866
queue->data_size = DATA_FRAME_SIZE;
drivers/net/wireless/ralink/rt2x00/rt2500usb.c
1867
queue->desc_size = TXD_DESC_SIZE;
drivers/net/wireless/ralink/rt2x00/rt2500usb.c
1868
queue->priv_size = sizeof(struct queue_entry_priv_usb);
drivers/net/wireless/ralink/rt2x00/rt2500usb.c
1872
queue->limit = 1;
drivers/net/wireless/ralink/rt2x00/rt2500usb.c
1873
queue->data_size = MGMT_FRAME_SIZE;
drivers/net/wireless/ralink/rt2x00/rt2500usb.c
1874
queue->desc_size = TXD_DESC_SIZE;
drivers/net/wireless/ralink/rt2x00/rt2500usb.c
1875
queue->priv_size = sizeof(struct queue_entry_priv_usb_bcn);
drivers/net/wireless/ralink/rt2x00/rt2500usb.c
1879
queue->limit = 8;
drivers/net/wireless/ralink/rt2x00/rt2500usb.c
1880
queue->data_size = DATA_FRAME_SIZE;
drivers/net/wireless/ralink/rt2x00/rt2500usb.c
1881
queue->desc_size = TXD_DESC_SIZE;
drivers/net/wireless/ralink/rt2x00/rt2500usb.c
1882
queue->priv_size = sizeof(struct queue_entry_priv_usb);
drivers/net/wireless/ralink/rt2x00/rt2500usb.c
717
static void rt2500usb_start_queue(struct data_queue *queue)
drivers/net/wireless/ralink/rt2x00/rt2500usb.c
719
struct rt2x00_dev *rt2x00dev = queue->rt2x00dev;
drivers/net/wireless/ralink/rt2x00/rt2500usb.c
722
switch (queue->qid) {
drivers/net/wireless/ralink/rt2x00/rt2500usb.c
740
static void rt2500usb_stop_queue(struct data_queue *queue)
drivers/net/wireless/ralink/rt2x00/rt2500usb.c
742
struct rt2x00_dev *rt2x00dev = queue->rt2x00dev;
drivers/net/wireless/ralink/rt2x00/rt2500usb.c
745
switch (queue->qid) {
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
1010
rt2x00_dbg(entry->queue->rt2x00dev,
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
1012
entry->queue->qid, entry->entry_idx);
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
1022
struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
1123
struct data_queue *queue;
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
1135
queue = rt2x00queue_get_tx_queue(rt2x00dev, qid);
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
1137
if (unlikely(rt2x00queue_empty(queue))) {
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
1143
entry = rt2x00queue_get_entry(queue, Q_INDEX_DONE);
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
1174
rt2x00_dbg(entry->queue->rt2x00dev,
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
1176
entry->entry_idx, entry->queue->qid);
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
1182
struct data_queue *queue;
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
1185
tx_queue_for_each(rt2x00dev, queue) {
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
1186
entry = rt2x00queue_get_entry(queue, Q_INDEX_DONE);
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
1201
struct data_queue *queue;
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
1203
tx_queue_for_each(rt2x00dev, queue) {
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
1204
if (rt2x00queue_get_entry(queue, Q_INDEX_DMA_DONE) !=
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
1205
rt2x00queue_get_entry(queue, Q_INDEX_DONE))
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
1214
struct data_queue *queue;
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
12179
struct data_queue *queue;
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
12202
queue = rt2x00queue_get_tx_queue(rt2x00dev, queue_idx);
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
12210
rt2x00_set_field32(®, field, queue->txop);
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
12218
rt2x00_set_field32(®, field, queue->aifs);
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
12222
rt2x00_set_field32(®, field, queue->cw_min);
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
12226
rt2x00_set_field32(®, field, queue->cw_max);
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
12233
rt2x00_set_field32(®, EDCA_AC0_CFG_TX_OP, queue->txop);
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
12234
rt2x00_set_field32(®, EDCA_AC0_CFG_AIFSN, queue->aifs);
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
12235
rt2x00_set_field32(®, EDCA_AC0_CFG_CWMIN, queue->cw_min);
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
12236
rt2x00_set_field32(®, EDCA_AC0_CFG_CWMAX, queue->cw_max);
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
1224
tx_queue_for_each(rt2x00dev, queue) {
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
1225
while (!rt2x00queue_empty(queue)) {
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
1226
entry = rt2x00queue_get_entry(queue, Q_INDEX_DONE);
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
1242
static bool rt2800_check_hung(struct data_queue *queue)
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
1244
unsigned int cur_idx = rt2800_drv_get_dma_done(queue);
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
1246
if (queue->wd_idx != cur_idx) {
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
1247
queue->wd_idx = cur_idx;
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
1248
queue->wd_count = 0;
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
1250
queue->wd_count++;
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
1252
return queue->wd_count > 16;
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
1268
struct data_queue *queue;
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
1274
queue_for_each(rt2x00dev, queue) {
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
1275
switch (queue->qid) {
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
1281
if (rt2x00queue_empty(queue))
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
1283
hung_tx = hung_tx || rt2800_check_hung(queue);
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
1292
hung_rx = hung_rx || rt2800_check_hung(queue);
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
1308
queue_for_each(rt2x00dev, queue)
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
1309
queue->wd_count = 0;
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
1382
struct data_queue *queue = rt2x00dev->bcn;
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
1391
for (i = 0; i < queue->limit; i++) {
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
1392
entry = &queue->entries[i];
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
1414
struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
1419
const int txwi_desc_size = entry->queue->winfo_size;
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
1508
struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
1951
struct data_queue *queue = rt2x00dev->bcn;
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
1960
for (i = 0; i < queue->limit; i++) {
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
1961
entry = &queue->entries[i];
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
838
rt2x00_set_field32(&word, TXWI_W1_PACKETID_QUEUE, entry->queue->qid);
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
851
for (i = 2; i < entry->queue->winfo_size / sizeof(__le32); i++)
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
937
rxdesc->rssi = rt2800_agc_to_rssi(entry->queue->rt2x00dev, word);
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
941
skb_pull(entry->skb, entry->queue->winfo_size);
drivers/net/wireless/ralink/rt2x00/rt2800lib.h
175
const struct rt2800_ops *rt2800ops = entry->queue->rt2x00dev->ops->drv;
drivers/net/wireless/ralink/rt2x00/rt2800lib.h
180
static inline unsigned int rt2800_drv_get_dma_done(struct data_queue *queue)
drivers/net/wireless/ralink/rt2x00/rt2800lib.h
182
const struct rt2800_ops *rt2800ops = queue->rt2x00dev->ops->drv;
drivers/net/wireless/ralink/rt2x00/rt2800lib.h
184
return rt2800ops->drv_get_dma_done(queue);
drivers/net/wireless/ralink/rt2x00/rt2800lib.h
78
unsigned int (*drv_get_dma_done)(struct data_queue *queue);
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
27
unsigned int rt2800mmio_get_dma_done(struct data_queue *queue)
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
29
struct rt2x00_dev *rt2x00dev = queue->rt2x00dev;
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
33
switch (queue->qid) {
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
38
qid = queue->qid;
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
424
void rt2800mmio_start_queue(struct data_queue *queue)
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
426
struct rt2x00_dev *rt2x00dev = queue->rt2x00dev;
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
429
switch (queue->qid) {
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
45
entry = rt2x00queue_get_entry(queue, Q_INDEX_DMA_DONE);
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
455
void rt2800mmio_kick_queue(struct data_queue *queue)
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
457
struct rt2x00_dev *rt2x00dev = queue->rt2x00dev;
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
460
switch (queue->qid) {
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
465
WARN_ON_ONCE(rt2x00queue_empty(queue));
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
466
entry = rt2x00queue_get_entry(queue, Q_INDEX);
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
467
rt2x00mmio_register_write(rt2x00dev, TX_CTX_IDX(queue->qid),
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
473
entry = rt2x00queue_get_entry(queue, Q_INDEX);
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
483
void rt2800mmio_flush_queue(struct data_queue *queue, bool drop)
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
485
struct rt2x00_dev *rt2x00dev = queue->rt2x00dev;
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
489
switch (queue->qid) {
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
508
if (rt2x00queue_empty(queue))
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
527
void rt2800mmio_stop_queue(struct data_queue *queue)
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
529
struct rt2x00_dev *rt2x00dev = queue->rt2x00dev;
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
532
switch (queue->qid) {
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
564
void rt2800mmio_queue_init(struct data_queue *queue)
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
566
struct rt2x00_dev *rt2x00dev = queue->rt2x00dev;
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
571
switch (queue->qid) {
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
573
queue->limit = 128;
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
574
queue->data_size = AGGREGATION_SIZE;
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
575
queue->desc_size = RXD_DESC_SIZE;
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
576
queue->winfo_size = rxwi_size;
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
577
queue->priv_size = sizeof(struct queue_entry_priv_mmio);
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
584
queue->limit = 64;
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
585
queue->data_size = AGGREGATION_SIZE;
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
586
queue->desc_size = TXD_DESC_SIZE;
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
587
queue->winfo_size = txwi_size;
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
588
queue->priv_size = sizeof(struct queue_entry_priv_mmio);
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
592
queue->limit = 8;
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
593
queue->data_size = 0; /* No DMA required for beacons */
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
594
queue->desc_size = TXD_DESC_SIZE;
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
595
queue->winfo_size = txwi_size;
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
596
queue->priv_size = sizeof(struct queue_entry_priv_mmio);
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
615
if (entry->queue->qid == QID_RX) {
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
631
struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
634
if (entry->queue->qid == QID_RX) {
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
655
if (entry->queue->length == 1)
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
74
const unsigned int txwi_size = entry->queue->winfo_size;
drivers/net/wireless/ralink/rt2x00/rt2800mmio.h
117
unsigned int rt2800mmio_get_dma_done(struct data_queue *queue);
drivers/net/wireless/ralink/rt2x00/rt2800mmio.h
139
void rt2800mmio_start_queue(struct data_queue *queue);
drivers/net/wireless/ralink/rt2x00/rt2800mmio.h
140
void rt2800mmio_kick_queue(struct data_queue *queue);
drivers/net/wireless/ralink/rt2x00/rt2800mmio.h
141
void rt2800mmio_flush_queue(struct data_queue *queue, bool drop);
drivers/net/wireless/ralink/rt2x00/rt2800mmio.h
142
void rt2800mmio_stop_queue(struct data_queue *queue);
drivers/net/wireless/ralink/rt2x00/rt2800mmio.h
143
void rt2800mmio_queue_init(struct data_queue *queue);
drivers/net/wireless/ralink/rt2x00/rt2800usb.c
157
struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
drivers/net/wireless/ralink/rt2x00/rt2800usb.c
382
static unsigned int rt2800usb_get_dma_done(struct data_queue *queue)
drivers/net/wireless/ralink/rt2x00/rt2800usb.c
386
entry = rt2x00queue_get_entry(queue, Q_INDEX_DMA_DONE);
drivers/net/wireless/ralink/rt2x00/rt2800usb.c
395
if (entry->queue->qid == QID_BEACON)
drivers/net/wireless/ralink/rt2x00/rt2800usb.c
434
skbdesc->desc_len = TXINFO_DESC_SIZE + entry->queue->winfo_size;
drivers/net/wireless/ralink/rt2x00/rt2800usb.c
46
static void rt2800usb_start_queue(struct data_queue *queue)
drivers/net/wireless/ralink/rt2x00/rt2800usb.c
48
struct rt2x00_dev *rt2x00dev = queue->rt2x00dev;
drivers/net/wireless/ralink/rt2x00/rt2800usb.c
51
switch (queue->qid) {
drivers/net/wireless/ralink/rt2x00/rt2800usb.c
514
rx_pkt_len > entry->queue->data_size)) {
drivers/net/wireless/ralink/rt2x00/rt2800usb.c
515
rt2x00_err(entry->queue->rt2x00dev,
drivers/net/wireless/ralink/rt2x00/rt2800usb.c
69
static void rt2800usb_stop_queue(struct data_queue *queue)
drivers/net/wireless/ralink/rt2x00/rt2800usb.c
71
struct rt2x00_dev *rt2x00dev = queue->rt2x00dev;
drivers/net/wireless/ralink/rt2x00/rt2800usb.c
718
static void rt2800usb_queue_init(struct data_queue *queue)
drivers/net/wireless/ralink/rt2x00/rt2800usb.c
720
struct rt2x00_dev *rt2x00dev = queue->rt2x00dev;
drivers/net/wireless/ralink/rt2x00/rt2800usb.c
725
switch (queue->qid) {
drivers/net/wireless/ralink/rt2x00/rt2800usb.c
727
queue->limit = 128;
drivers/net/wireless/ralink/rt2x00/rt2800usb.c
728
queue->data_size = AGGREGATION_SIZE;
drivers/net/wireless/ralink/rt2x00/rt2800usb.c
729
queue->desc_size = RXINFO_DESC_SIZE;
drivers/net/wireless/ralink/rt2x00/rt2800usb.c
730
queue->winfo_size = rxwi_size;
drivers/net/wireless/ralink/rt2x00/rt2800usb.c
731
queue->priv_size = sizeof(struct queue_entry_priv_usb);
drivers/net/wireless/ralink/rt2x00/rt2800usb.c
738
queue->limit = 16;
drivers/net/wireless/ralink/rt2x00/rt2800usb.c
739
queue->data_size = AGGREGATION_SIZE;
drivers/net/wireless/ralink/rt2x00/rt2800usb.c
74
switch (queue->qid) {
drivers/net/wireless/ralink/rt2x00/rt2800usb.c
740
queue->desc_size = TXINFO_DESC_SIZE;
drivers/net/wireless/ralink/rt2x00/rt2800usb.c
741
queue->winfo_size = txwi_size;
drivers/net/wireless/ralink/rt2x00/rt2800usb.c
742
queue->priv_size = sizeof(struct queue_entry_priv_usb);
drivers/net/wireless/ralink/rt2x00/rt2800usb.c
746
queue->limit = 8;
drivers/net/wireless/ralink/rt2x00/rt2800usb.c
747
queue->data_size = MGMT_FRAME_SIZE;
drivers/net/wireless/ralink/rt2x00/rt2800usb.c
748
queue->desc_size = TXINFO_DESC_SIZE;
drivers/net/wireless/ralink/rt2x00/rt2800usb.c
749
queue->winfo_size = txwi_size;
drivers/net/wireless/ralink/rt2x00/rt2800usb.c
750
queue->priv_size = sizeof(struct queue_entry_priv_usb);
drivers/net/wireless/ralink/rt2x00/rt2x00.h
1311
enum data_queue_qid queue)
drivers/net/wireless/ralink/rt2x00/rt2x00.h
1313
if (queue >= rt2x00dev->ops->tx_queues && queue < IEEE80211_NUM_ACS)
drivers/net/wireless/ralink/rt2x00/rt2x00.h
1314
queue = rt2x00dev->ops->tx_queues - 1;
drivers/net/wireless/ralink/rt2x00/rt2x00.h
1316
if (queue < rt2x00dev->ops->tx_queues && rt2x00dev->tx)
drivers/net/wireless/ralink/rt2x00/rt2x00.h
1317
return &rt2x00dev->tx[queue];
drivers/net/wireless/ralink/rt2x00/rt2x00.h
1319
if (queue == QID_ATIM)
drivers/net/wireless/ralink/rt2x00/rt2x00.h
1330
struct queue_entry *rt2x00queue_get_entry(struct data_queue *queue,
drivers/net/wireless/ralink/rt2x00/rt2x00.h
1341
void rt2x00queue_pause_queue(struct data_queue *queue);
drivers/net/wireless/ralink/rt2x00/rt2x00.h
1350
void rt2x00queue_unpause_queue(struct data_queue *queue);
drivers/net/wireless/ralink/rt2x00/rt2x00.h
1358
void rt2x00queue_start_queue(struct data_queue *queue);
drivers/net/wireless/ralink/rt2x00/rt2x00.h
1366
void rt2x00queue_stop_queue(struct data_queue *queue);
drivers/net/wireless/ralink/rt2x00/rt2x00.h
1376
void rt2x00queue_flush_queue(struct data_queue *queue, bool drop);
drivers/net/wireless/ralink/rt2x00/rt2x00.h
1487
unsigned int link_id, u16 queue,
drivers/net/wireless/ralink/rt2x00/rt2x00.h
567
void (*start_queue) (struct data_queue *queue);
drivers/net/wireless/ralink/rt2x00/rt2x00.h
568
void (*kick_queue) (struct data_queue *queue);
drivers/net/wireless/ralink/rt2x00/rt2x00.h
569
void (*stop_queue) (struct data_queue *queue);
drivers/net/wireless/ralink/rt2x00/rt2x00.h
570
void (*flush_queue) (struct data_queue *queue, bool drop);
drivers/net/wireless/ralink/rt2x00/rt2x00.h
636
void (*queue_init)(struct data_queue *queue);
drivers/net/wireless/ralink/rt2x00/rt2x00debug.c
171
dump_hdr->queue_index = entry->queue->qid;
drivers/net/wireless/ralink/rt2x00/rt2x00debug.c
303
struct data_queue *queue;
drivers/net/wireless/ralink/rt2x00/rt2x00debug.c
320
queue_for_each(intf->rt2x00dev, queue) {
drivers/net/wireless/ralink/rt2x00/rt2x00debug.c
321
spin_lock_irqsave(&queue->index_lock, irqflags);
drivers/net/wireless/ralink/rt2x00/rt2x00debug.c
324
queue->qid, (unsigned int)queue->flags,
drivers/net/wireless/ralink/rt2x00/rt2x00debug.c
325
queue->count, queue->limit, queue->length,
drivers/net/wireless/ralink/rt2x00/rt2x00debug.c
326
queue->index[Q_INDEX],
drivers/net/wireless/ralink/rt2x00/rt2x00debug.c
327
queue->index[Q_INDEX_DMA_DONE],
drivers/net/wireless/ralink/rt2x00/rt2x00debug.c
328
queue->index[Q_INDEX_DONE]);
drivers/net/wireless/ralink/rt2x00/rt2x00debug.c
330
spin_unlock_irqrestore(&queue->index_lock, irqflags);
drivers/net/wireless/ralink/rt2x00/rt2x00dev.c
263
struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
drivers/net/wireless/ralink/rt2x00/rt2x00dev.c
404
spin_lock_bh(&entry->queue->tx_lock);
drivers/net/wireless/ralink/rt2x00/rt2x00dev.c
405
if (!rt2x00queue_threshold(entry->queue))
drivers/net/wireless/ralink/rt2x00/rt2x00dev.c
406
rt2x00queue_unpause_queue(entry->queue);
drivers/net/wireless/ralink/rt2x00/rt2x00dev.c
407
spin_unlock_bh(&entry->queue->tx_lock);
drivers/net/wireless/ralink/rt2x00/rt2x00dev.c
413
struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
drivers/net/wireless/ralink/rt2x00/rt2x00dev.c
463
struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
drivers/net/wireless/ralink/rt2x00/rt2x00dev.c
739
struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
drivers/net/wireless/ralink/rt2x00/rt2x00dev.c
777
rxdesc.size > entry->queue->data_size)) {
drivers/net/wireless/ralink/rt2x00/rt2x00dev.c
779
rxdesc.size, entry->queue->data_size);
drivers/net/wireless/ralink/rt2x00/rt2x00lib.h
136
int rt2x00queue_write_tx_frame(struct data_queue *queue, struct sk_buff *skb,
drivers/net/wireless/ralink/rt2x00/rt2x00mac.c
114
queue = rt2x00queue_get_tx_queue(rt2x00dev, qid);
drivers/net/wireless/ralink/rt2x00/rt2x00mac.c
115
if (unlikely(!queue)) {
drivers/net/wireless/ralink/rt2x00/rt2x00mac.c
134
if (rt2x00queue_available(queue) <= 1) {
drivers/net/wireless/ralink/rt2x00/rt2x00mac.c
139
spin_lock(&queue->tx_lock);
drivers/net/wireless/ralink/rt2x00/rt2x00mac.c
140
if (rt2x00queue_threshold(queue))
drivers/net/wireless/ralink/rt2x00/rt2x00mac.c
141
rt2x00queue_pause_queue(queue);
drivers/net/wireless/ralink/rt2x00/rt2x00mac.c
142
spin_unlock(&queue->tx_lock);
drivers/net/wireless/ralink/rt2x00/rt2x00mac.c
147
if (rt2x00mac_tx_rts_cts(rt2x00dev, queue, skb))
drivers/net/wireless/ralink/rt2x00/rt2x00mac.c
151
if (unlikely(rt2x00queue_write_tx_frame(queue, skb, control->sta, false)))
drivers/net/wireless/ralink/rt2x00/rt2x00mac.c
20
struct data_queue *queue,
drivers/net/wireless/ralink/rt2x00/rt2x00mac.c
208
struct data_queue *queue = rt2x00dev->bcn;
drivers/net/wireless/ralink/rt2x00/rt2x00mac.c
226
for (i = 0; i < queue->limit; i++) {
drivers/net/wireless/ralink/rt2x00/rt2x00mac.c
227
entry = &queue->entries[i];
drivers/net/wireless/ralink/rt2x00/rt2x00mac.c
232
if (unlikely(i == queue->limit))
drivers/net/wireless/ralink/rt2x00/rt2x00mac.c
684
struct data_queue *queue;
drivers/net/wireless/ralink/rt2x00/rt2x00mac.c
686
queue = rt2x00queue_get_tx_queue(rt2x00dev, queue_idx);
drivers/net/wireless/ralink/rt2x00/rt2x00mac.c
687
if (unlikely(!queue))
drivers/net/wireless/ralink/rt2x00/rt2x00mac.c
695
queue->cw_min = fls(params->cw_min);
drivers/net/wireless/ralink/rt2x00/rt2x00mac.c
697
queue->cw_min = 5; /* cw_min: 2^5 = 32. */
drivers/net/wireless/ralink/rt2x00/rt2x00mac.c
700
queue->cw_max = fls(params->cw_max);
drivers/net/wireless/ralink/rt2x00/rt2x00mac.c
702
queue->cw_max = 10; /* cw_min: 2^10 = 1024. */
drivers/net/wireless/ralink/rt2x00/rt2x00mac.c
704
queue->aifs = params->aifs;
drivers/net/wireless/ralink/rt2x00/rt2x00mac.c
705
queue->txop = params->txop;
drivers/net/wireless/ralink/rt2x00/rt2x00mac.c
709
queue_idx, queue->cw_min, queue->cw_max, queue->aifs,
drivers/net/wireless/ralink/rt2x00/rt2x00mac.c
710
queue->txop);
drivers/net/wireless/ralink/rt2x00/rt2x00mac.c
729
struct data_queue *queue;
drivers/net/wireless/ralink/rt2x00/rt2x00mac.c
736
tx_queue_for_each(rt2x00dev, queue)
drivers/net/wireless/ralink/rt2x00/rt2x00mac.c
737
rt2x00queue_flush_queue(queue, drop);
drivers/net/wireless/ralink/rt2x00/rt2x00mac.c
80
retval = rt2x00queue_write_tx_frame(queue, skb, NULL, true);
drivers/net/wireless/ralink/rt2x00/rt2x00mac.c
816
struct data_queue *queue;
drivers/net/wireless/ralink/rt2x00/rt2x00mac.c
818
tx_queue_for_each(rt2x00dev, queue) {
drivers/net/wireless/ralink/rt2x00/rt2x00mac.c
819
*tx += queue->length;
drivers/net/wireless/ralink/rt2x00/rt2x00mac.c
820
*tx_max += queue->limit;
drivers/net/wireless/ralink/rt2x00/rt2x00mac.c
831
struct data_queue *queue;
drivers/net/wireless/ralink/rt2x00/rt2x00mac.c
833
tx_queue_for_each(rt2x00dev, queue) {
drivers/net/wireless/ralink/rt2x00/rt2x00mac.c
834
if (!rt2x00queue_empty(queue))
drivers/net/wireless/ralink/rt2x00/rt2x00mac.c
96
struct data_queue *queue = NULL;
drivers/net/wireless/ralink/rt2x00/rt2x00mmio.c
101
struct data_queue *queue)
drivers/net/wireless/ralink/rt2x00/rt2x00mmio.c
112
queue->limit * queue->desc_size, &dma,
drivers/net/wireless/ralink/rt2x00/rt2x00mmio.c
120
for (i = 0; i < queue->limit; i++) {
drivers/net/wireless/ralink/rt2x00/rt2x00mmio.c
121
entry_priv = queue->entries[i].priv_data;
drivers/net/wireless/ralink/rt2x00/rt2x00mmio.c
122
entry_priv->desc = addr + i * queue->desc_size;
drivers/net/wireless/ralink/rt2x00/rt2x00mmio.c
123
entry_priv->desc_dma = dma + i * queue->desc_size;
drivers/net/wireless/ralink/rt2x00/rt2x00mmio.c
130
struct data_queue *queue)
drivers/net/wireless/ralink/rt2x00/rt2x00mmio.c
133
queue->entries[0].priv_data;
drivers/net/wireless/ralink/rt2x00/rt2x00mmio.c
137
queue->limit * queue->desc_size,
drivers/net/wireless/ralink/rt2x00/rt2x00mmio.c
144
struct data_queue *queue;
drivers/net/wireless/ralink/rt2x00/rt2x00mmio.c
150
queue_for_each(rt2x00dev, queue) {
drivers/net/wireless/ralink/rt2x00/rt2x00mmio.c
151
status = rt2x00mmio_alloc_queue_dma(rt2x00dev, queue);
drivers/net/wireless/ralink/rt2x00/rt2x00mmio.c
171
queue_for_each(rt2x00dev, queue)
drivers/net/wireless/ralink/rt2x00/rt2x00mmio.c
172
rt2x00mmio_free_queue_dma(rt2x00dev, queue);
drivers/net/wireless/ralink/rt2x00/rt2x00mmio.c
180
struct data_queue *queue;
drivers/net/wireless/ralink/rt2x00/rt2x00mmio.c
190
queue_for_each(rt2x00dev, queue)
drivers/net/wireless/ralink/rt2x00/rt2x00mmio.c
191
rt2x00mmio_free_queue_dma(rt2x00dev, queue);
drivers/net/wireless/ralink/rt2x00/rt2x00mmio.c
51
struct data_queue *queue = rt2x00dev->rx;
drivers/net/wireless/ralink/rt2x00/rt2x00mmio.c
58
entry = rt2x00queue_get_entry(queue, Q_INDEX);
drivers/net/wireless/ralink/rt2x00/rt2x00mmio.c
69
skbdesc->desc_len = entry->queue->desc_size;
drivers/net/wireless/ralink/rt2x00/rt2x00mmio.c
88
void rt2x00mmio_flush_queue(struct data_queue *queue, bool drop)
drivers/net/wireless/ralink/rt2x00/rt2x00mmio.c
92
for (i = 0; !rt2x00queue_empty(queue) && i < 10; i++)
drivers/net/wireless/ralink/rt2x00/rt2x00mmio.h
95
void rt2x00mmio_flush_queue(struct data_queue *queue, bool drop);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1004
queue->rt2x00dev->ops->lib->kick_queue(queue);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1011
if (likely(queue->rt2x00dev->ops->lib->flush_queue))
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1012
queue->rt2x00dev->ops->lib->flush_queue(queue, drop);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1017
if (unlikely(!rt2x00queue_empty(queue)))
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1018
rt2x00_warn(queue->rt2x00dev, "Queue %d failed to flush\n",
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1019
queue->qid);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1025
struct data_queue *queue;
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1031
tx_queue_for_each(rt2x00dev, queue)
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1032
rt2x00queue_start_queue(queue);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1040
struct data_queue *queue;
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1050
tx_queue_for_each(rt2x00dev, queue)
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1051
rt2x00queue_stop_queue(queue);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1059
struct data_queue *queue;
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1061
tx_queue_for_each(rt2x00dev, queue)
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1062
rt2x00queue_flush_queue(queue, drop);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1068
static void rt2x00queue_reset(struct data_queue *queue)
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1073
spin_lock_irqsave(&queue->index_lock, irqflags);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1075
queue->count = 0;
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1076
queue->length = 0;
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1079
queue->index[i] = 0;
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1081
spin_unlock_irqrestore(&queue->index_lock, irqflags);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1086
struct data_queue *queue;
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1089
queue_for_each(rt2x00dev, queue) {
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1090
rt2x00queue_reset(queue);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1092
for (i = 0; i < queue->limit; i++)
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1093
rt2x00dev->ops->lib->clear_entry(&queue->entries[i]);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1097
static int rt2x00queue_alloc_entries(struct data_queue *queue)
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1103
rt2x00queue_reset(queue);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1108
entry_size = sizeof(*entries) + queue->priv_size;
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1109
entries = kcalloc(queue->limit, entry_size, GFP_KERNEL);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1117
for (i = 0; i < queue->limit; i++) {
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1119
entries[i].queue = queue;
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
112
struct device *dev = entry->queue->rt2x00dev->dev;
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1123
QUEUE_ENTRY_PRIV_OFFSET(entries, i, queue->limit,
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1124
sizeof(*entries), queue->priv_size);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1129
queue->entries = entries;
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1134
static void rt2x00queue_free_skbs(struct data_queue *queue)
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1138
if (!queue->entries)
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1141
for (i = 0; i < queue->limit; i++) {
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1142
rt2x00queue_free_skb(&queue->entries[i]);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1146
static int rt2x00queue_alloc_rxskbs(struct data_queue *queue)
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1151
for (i = 0; i < queue->limit; i++) {
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1152
skb = rt2x00queue_alloc_rxskb(&queue->entries[i], GFP_KERNEL);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1155
queue->entries[i].skb = skb;
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1163
struct data_queue *queue;
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1170
tx_queue_for_each(rt2x00dev, queue) {
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1171
status = rt2x00queue_alloc_entries(queue);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1202
struct data_queue *queue;
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1206
queue_for_each(rt2x00dev, queue) {
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1207
kfree(queue->entries);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1208
queue->entries = NULL;
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1213
struct data_queue *queue, enum data_queue_qid qid)
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1215
mutex_init(&queue->status_lock);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1216
spin_lock_init(&queue->tx_lock);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1217
spin_lock_init(&queue->index_lock);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1219
queue->rt2x00dev = rt2x00dev;
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1220
queue->qid = qid;
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1221
queue->txop = 0;
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1222
queue->aifs = 2;
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1223
queue->cw_min = 5;
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1224
queue->cw_max = 10;
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1226
rt2x00dev->ops->queue_init(queue);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1228
queue->threshold = DIV_ROUND_UP(queue->limit, 10);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1233
struct data_queue *queue;
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1247
queue = kzalloc_objs(*queue, rt2x00dev->data_queues);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1248
if (!queue)
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1254
rt2x00dev->rx = queue;
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1255
rt2x00dev->tx = &queue[1];
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1256
rt2x00dev->bcn = &queue[1 + rt2x00dev->ops->tx_queues];
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1257
rt2x00dev->atim = req_atim ? &queue[2 + rt2x00dev->ops->tx_queues] : NULL;
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1271
tx_queue_for_each(rt2x00dev, queue)
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1272
rt2x00queue_init(rt2x00dev, queue, qid++);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
25
struct data_queue *queue = entry->queue;
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
26
struct rt2x00_dev *rt2x00dev = queue->rt2x00dev;
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
37
frame_size = queue->data_size + queue->desc_size + queue->winfo_size;
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
488
struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
500
entry->queue->qid, DRV_PROJECT);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
529
struct data_queue *queue = entry->queue;
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
531
queue->rt2x00dev->ops->lib->write_tx_desc(entry, txdesc);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
537
rt2x00debug_dump_frame(queue->rt2x00dev, DUMP_FRAME_TX, entry);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
540
static void rt2x00queue_kick_tx_queue(struct data_queue *queue,
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
552
if (rt2x00queue_threshold(queue) ||
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
554
queue->rt2x00dev->ops->lib->kick_queue(queue);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
559
struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
599
int rt2x00queue_write_tx_frame(struct data_queue *queue, struct sk_buff *skb,
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
614
rt2x00queue_create_tx_descriptor(queue->rt2x00dev, skb, &txdesc, sta);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
639
if (rt2x00_has_cap_flag(queue->rt2x00dev, REQUIRE_COPY_IV))
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
653
if (rt2x00_has_cap_flag(queue->rt2x00dev, REQUIRE_L2PAD))
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
655
else if (rt2x00_has_cap_flag(queue->rt2x00dev, REQUIRE_DMA))
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
661
spin_lock(&queue->tx_lock);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
663
if (unlikely(rt2x00queue_full(queue))) {
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
664
rt2x00_dbg(queue->rt2x00dev, "Dropping frame due to full tx queue %d\n",
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
665
queue->qid);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
670
entry = rt2x00queue_get_entry(queue, Q_INDEX);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
674
rt2x00_err(queue->rt2x00dev,
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
677
queue->qid, DRV_PROJECT);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
705
rt2x00queue_kick_tx_queue(queue, &txdesc);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
713
if (rt2x00queue_threshold(queue))
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
714
rt2x00queue_pause_queue(queue);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
716
spin_unlock(&queue->tx_lock);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
784
bool rt2x00queue_for_each_entry(struct data_queue *queue,
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
797
rt2x00_err(queue->rt2x00dev,
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
809
spin_lock_irqsave(&queue->index_lock, irqflags);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
810
index_start = queue->index[start];
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
811
index_end = queue->index[end];
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
812
spin_unlock_irqrestore(&queue->index_lock, irqflags);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
820
if (fn(&queue->entries[i], data))
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
824
for (i = index_start; i < queue->limit; i++) {
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
825
if (fn(&queue->entries[i], data))
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
830
if (fn(&queue->entries[i], data))
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
839
struct queue_entry *rt2x00queue_get_entry(struct data_queue *queue,
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
846
rt2x00_err(queue->rt2x00dev, "Entry requested from invalid index type (%d)\n",
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
851
spin_lock_irqsave(&queue->index_lock, irqflags);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
853
entry = &queue->entries[queue->index[index]];
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
855
spin_unlock_irqrestore(&queue->index_lock, irqflags);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
863
struct data_queue *queue = entry->queue;
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
867
rt2x00_err(queue->rt2x00dev,
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
872
spin_lock_irqsave(&queue->index_lock, irqflags);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
874
queue->index[index]++;
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
875
if (queue->index[index] >= queue->limit)
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
876
queue->index[index] = 0;
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
881
queue->length++;
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
883
queue->length--;
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
884
queue->count++;
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
887
spin_unlock_irqrestore(&queue->index_lock, irqflags);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
890
static void rt2x00queue_pause_queue_nocheck(struct data_queue *queue)
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
892
switch (queue->qid) {
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
901
ieee80211_stop_queue(queue->rt2x00dev->hw, queue->qid);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
907
void rt2x00queue_pause_queue(struct data_queue *queue)
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
909
if (!test_bit(DEVICE_STATE_PRESENT, &queue->rt2x00dev->flags) ||
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
910
!test_bit(QUEUE_STARTED, &queue->flags) ||
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
911
test_and_set_bit(QUEUE_PAUSED, &queue->flags))
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
914
rt2x00queue_pause_queue_nocheck(queue);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
918
void rt2x00queue_unpause_queue(struct data_queue *queue)
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
920
if (!test_bit(DEVICE_STATE_PRESENT, &queue->rt2x00dev->flags) ||
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
921
!test_bit(QUEUE_STARTED, &queue->flags) ||
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
922
!test_and_clear_bit(QUEUE_PAUSED, &queue->flags))
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
925
switch (queue->qid) {
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
934
ieee80211_wake_queue(queue->rt2x00dev->hw, queue->qid);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
941
queue->rt2x00dev->ops->lib->kick_queue(queue);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
949
void rt2x00queue_start_queue(struct data_queue *queue)
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
95
struct device *dev = entry->queue->rt2x00dev->dev;
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
951
mutex_lock(&queue->status_lock);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
953
if (!test_bit(DEVICE_STATE_PRESENT, &queue->rt2x00dev->flags) ||
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
954
test_and_set_bit(QUEUE_STARTED, &queue->flags)) {
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
955
mutex_unlock(&queue->status_lock);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
959
set_bit(QUEUE_PAUSED, &queue->flags);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
961
queue->rt2x00dev->ops->lib->start_queue(queue);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
963
rt2x00queue_unpause_queue(queue);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
965
mutex_unlock(&queue->status_lock);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
969
void rt2x00queue_stop_queue(struct data_queue *queue)
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
971
mutex_lock(&queue->status_lock);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
973
if (!test_and_clear_bit(QUEUE_STARTED, &queue->flags)) {
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
974
mutex_unlock(&queue->status_lock);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
978
rt2x00queue_pause_queue_nocheck(queue);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
980
queue->rt2x00dev->ops->lib->stop_queue(queue);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
982
mutex_unlock(&queue->status_lock);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
986
void rt2x00queue_flush_queue(struct data_queue *queue, bool drop)
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
989
(queue->qid == QID_AC_VO) ||
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
990
(queue->qid == QID_AC_VI) ||
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
991
(queue->qid == QID_AC_BE) ||
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
992
(queue->qid == QID_AC_BK);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
994
if (rt2x00queue_empty(queue))
drivers/net/wireless/ralink/rt2x00/rt2x00queue.h
372
struct data_queue *queue;
drivers/net/wireless/ralink/rt2x00/rt2x00queue.h
580
bool rt2x00queue_for_each_entry(struct data_queue *queue,
drivers/net/wireless/ralink/rt2x00/rt2x00queue.h
591
static inline int rt2x00queue_empty(struct data_queue *queue)
drivers/net/wireless/ralink/rt2x00/rt2x00queue.h
593
return queue->length == 0;
drivers/net/wireless/ralink/rt2x00/rt2x00queue.h
600
static inline int rt2x00queue_full(struct data_queue *queue)
drivers/net/wireless/ralink/rt2x00/rt2x00queue.h
602
return queue->length == queue->limit;
drivers/net/wireless/ralink/rt2x00/rt2x00queue.h
609
static inline int rt2x00queue_available(struct data_queue *queue)
drivers/net/wireless/ralink/rt2x00/rt2x00queue.h
611
return queue->limit - queue->length;
drivers/net/wireless/ralink/rt2x00/rt2x00queue.h
618
static inline int rt2x00queue_threshold(struct data_queue *queue)
drivers/net/wireless/ralink/rt2x00/rt2x00queue.h
620
return rt2x00queue_available(queue) < queue->threshold;
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
252
struct data_queue *queue;
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
255
tx_queue_for_each(rt2x00dev, queue) {
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
256
while (!rt2x00queue_empty(queue)) {
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
257
entry = rt2x00queue_get_entry(queue, Q_INDEX_DONE);
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
271
struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
298
struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
326
usb_sndbulkpipe(usb_dev, entry->queue->usb_endpoint),
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
363
skbdesc->desc_len = entry->queue->desc_size;
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
375
struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
385
if (urb->actual_length < entry->queue->desc_size || urb->status)
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
401
struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
412
usb_rcvbulkpipe(usb_dev, entry->queue->usb_endpoint),
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
427
void rt2x00usb_kick_queue(struct data_queue *queue)
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
429
switch (queue->qid) {
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
434
if (!rt2x00queue_empty(queue))
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
435
rt2x00queue_for_each_entry(queue,
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
442
if (!rt2x00queue_full(queue))
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
443
rt2x00queue_for_each_entry(queue,
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
457
struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
469
if ((entry->queue->qid == QID_BEACON) &&
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
476
void rt2x00usb_flush_queue(struct data_queue *queue, bool drop)
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
482
rt2x00queue_for_each_entry(queue, Q_INDEX_DONE, Q_INDEX, NULL,
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
488
switch (queue->qid) {
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
493
completion = &queue->rt2x00dev->txdone_work;
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
496
completion = &queue->rt2x00dev->rxdone_work;
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
508
if (rt2x00queue_empty(queue))
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
515
queue_work(queue->rt2x00dev->workqueue, completion);
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
526
static void rt2x00usb_watchdog_tx_dma(struct data_queue *queue)
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
528
rt2x00_warn(queue->rt2x00dev, "TX queue %d DMA timed out, invoke forced reset\n",
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
529
queue->qid);
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
531
rt2x00queue_stop_queue(queue);
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
532
rt2x00queue_flush_queue(queue, true);
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
533
rt2x00queue_start_queue(queue);
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
536
static int rt2x00usb_dma_timeout(struct data_queue *queue)
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
540
entry = rt2x00queue_get_entry(queue, Q_INDEX_DMA_DONE);
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
546
struct data_queue *queue;
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
548
tx_queue_for_each(rt2x00dev, queue) {
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
549
if (!rt2x00queue_empty(queue)) {
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
550
if (rt2x00usb_dma_timeout(queue))
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
551
rt2x00usb_watchdog_tx_dma(queue);
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
574
if (entry->queue->qid == QID_RX)
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
579
static void rt2x00usb_assign_endpoint(struct data_queue *queue,
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
582
struct usb_device *usb_dev = to_usb_device_intf(queue->rt2x00dev->dev);
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
585
queue->usb_endpoint = usb_endpoint_num(ep_desc);
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
587
if (queue->qid == QID_RX) {
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
588
pipe = usb_rcvbulkpipe(usb_dev, queue->usb_endpoint);
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
589
queue->usb_maxpacket = usb_maxpacket(usb_dev, pipe);
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
591
pipe = usb_sndbulkpipe(usb_dev, queue->usb_endpoint);
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
592
queue->usb_maxpacket = usb_maxpacket(usb_dev, pipe);
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
595
if (!queue->usb_maxpacket)
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
596
queue->usb_maxpacket = 1;
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
604
struct data_queue *queue = rt2x00dev->tx;
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
620
(queue != queue_end(rt2x00dev))) {
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
621
rt2x00usb_assign_endpoint(queue, ep_desc);
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
622
queue = queue_next(queue);
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
641
txall_queue_for_each(rt2x00dev, queue) {
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
642
if (!queue->usb_endpoint)
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
643
rt2x00usb_assign_endpoint(queue, tx_ep_desc);
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
649
static int rt2x00usb_alloc_entries(struct data_queue *queue)
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
651
struct rt2x00_dev *rt2x00dev = queue->rt2x00dev;
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
656
for (i = 0; i < queue->limit; i++) {
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
657
entry_priv = queue->entries[i].priv_data;
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
668
if (queue->qid != QID_BEACON ||
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
672
for (i = 0; i < queue->limit; i++) {
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
673
bcn_priv = queue->entries[i].priv_data;
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
682
static void rt2x00usb_free_entries(struct data_queue *queue)
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
684
struct rt2x00_dev *rt2x00dev = queue->rt2x00dev;
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
689
if (!queue->entries)
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
692
for (i = 0; i < queue->limit; i++) {
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
693
entry_priv = queue->entries[i].priv_data;
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
703
if (queue->qid != QID_BEACON ||
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
707
for (i = 0; i < queue->limit; i++) {
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
708
bcn_priv = queue->entries[i].priv_data;
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
716
struct data_queue *queue;
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
729
queue_for_each(rt2x00dev, queue) {
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
730
status = rt2x00usb_alloc_entries(queue);
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
746
struct data_queue *queue;
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
753
queue_for_each(rt2x00dev, queue)
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
754
rt2x00usb_free_entries(queue);
drivers/net/wireless/ralink/rt2x00/rt2x00usb.h
365
void rt2x00usb_kick_queue(struct data_queue *queue);
drivers/net/wireless/ralink/rt2x00/rt2x00usb.h
376
void rt2x00usb_flush_queue(struct data_queue *queue, bool drop);
drivers/net/wireless/ralink/rt2x00/rt61pci.c
1041
static void rt61pci_start_queue(struct data_queue *queue)
drivers/net/wireless/ralink/rt2x00/rt61pci.c
1043
struct rt2x00_dev *rt2x00dev = queue->rt2x00dev;
drivers/net/wireless/ralink/rt2x00/rt61pci.c
1046
switch (queue->qid) {
drivers/net/wireless/ralink/rt2x00/rt61pci.c
1064
static void rt61pci_kick_queue(struct data_queue *queue)
drivers/net/wireless/ralink/rt2x00/rt61pci.c
1066
struct rt2x00_dev *rt2x00dev = queue->rt2x00dev;
drivers/net/wireless/ralink/rt2x00/rt61pci.c
1069
switch (queue->qid) {
drivers/net/wireless/ralink/rt2x00/rt61pci.c
1095
static void rt61pci_stop_queue(struct data_queue *queue)
drivers/net/wireless/ralink/rt2x00/rt61pci.c
1097
struct rt2x00_dev *rt2x00dev = queue->rt2x00dev;
drivers/net/wireless/ralink/rt2x00/rt61pci.c
1100
switch (queue->qid) {
drivers/net/wireless/ralink/rt2x00/rt61pci.c
1292
if (entry->queue->qid == QID_RX) {
drivers/net/wireless/ralink/rt2x00/rt61pci.c
1310
if (entry->queue->qid == QID_RX) {
drivers/net/wireless/ralink/rt2x00/rt61pci.c
1787
rt2x00_set_field32(&word, TXD_W1_HOST_Q_ID, entry->queue->qid);
drivers/net/wireless/ralink/rt2x00/rt61pci.c
1788
rt2x00_set_field32(&word, TXD_W1_AIFSN, entry->queue->aifs);
drivers/net/wireless/ralink/rt2x00/rt61pci.c
1789
rt2x00_set_field32(&word, TXD_W1_CWMIN, entry->queue->cw_min);
drivers/net/wireless/ralink/rt2x00/rt61pci.c
1790
rt2x00_set_field32(&word, TXD_W1_CWMAX, entry->queue->cw_max);
drivers/net/wireless/ralink/rt2x00/rt61pci.c
1812
rt2x00_set_field32(&word, TXD_W5_PID_TYPE, entry->queue->qid);
drivers/net/wireless/ralink/rt2x00/rt61pci.c
1815
TXPOWER_TO_DEV(entry->queue->rt2x00dev->tx_power));
drivers/net/wireless/ralink/rt2x00/rt61pci.c
1819
if (entry->queue->qid != QID_BEACON) {
drivers/net/wireless/ralink/rt2x00/rt61pci.c
1865
skbdesc->desc_len = (entry->queue->qid == QID_BEACON) ? TXINFO_SIZE :
drivers/net/wireless/ralink/rt2x00/rt61pci.c
1875
struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
drivers/net/wireless/ralink/rt2x00/rt61pci.c
1939
struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
drivers/net/wireless/ralink/rt2x00/rt61pci.c
1997
struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
drivers/net/wireless/ralink/rt2x00/rt61pci.c
2061
struct data_queue *queue;
drivers/net/wireless/ralink/rt2x00/rt61pci.c
2091
queue = rt2x00queue_get_tx_queue(rt2x00dev, type);
drivers/net/wireless/ralink/rt2x00/rt61pci.c
2092
if (unlikely(!queue))
drivers/net/wireless/ralink/rt2x00/rt61pci.c
2100
if (unlikely(index >= queue->limit))
drivers/net/wireless/ralink/rt2x00/rt61pci.c
2103
entry = &queue->entries[index];
drivers/net/wireless/ralink/rt2x00/rt61pci.c
2111
entry_done = rt2x00queue_get_entry(queue, Q_INDEX_DONE);
drivers/net/wireless/ralink/rt2x00/rt61pci.c
2120
entry_done = rt2x00queue_get_entry(queue, Q_INDEX_DONE);
drivers/net/wireless/ralink/rt2x00/rt61pci.c
2807
struct data_queue *queue;
drivers/net/wireless/ralink/rt2x00/rt61pci.c
2830
queue = rt2x00queue_get_tx_queue(rt2x00dev, queue_idx);
drivers/net/wireless/ralink/rt2x00/rt61pci.c
2838
rt2x00_set_field32(®, field, queue->txop);
drivers/net/wireless/ralink/rt2x00/rt61pci.c
2846
rt2x00_set_field32(®, field, queue->aifs);
drivers/net/wireless/ralink/rt2x00/rt61pci.c
2850
rt2x00_set_field32(®, field, queue->cw_min);
drivers/net/wireless/ralink/rt2x00/rt61pci.c
2854
rt2x00_set_field32(®, field, queue->cw_max);
drivers/net/wireless/ralink/rt2x00/rt61pci.c
2938
static void rt61pci_queue_init(struct data_queue *queue)
drivers/net/wireless/ralink/rt2x00/rt61pci.c
2940
switch (queue->qid) {
drivers/net/wireless/ralink/rt2x00/rt61pci.c
2942
queue->limit = 32;
drivers/net/wireless/ralink/rt2x00/rt61pci.c
2943
queue->data_size = DATA_FRAME_SIZE;
drivers/net/wireless/ralink/rt2x00/rt61pci.c
2944
queue->desc_size = RXD_DESC_SIZE;
drivers/net/wireless/ralink/rt2x00/rt61pci.c
2945
queue->priv_size = sizeof(struct queue_entry_priv_mmio);
drivers/net/wireless/ralink/rt2x00/rt61pci.c
2952
queue->limit = 32;
drivers/net/wireless/ralink/rt2x00/rt61pci.c
2953
queue->data_size = DATA_FRAME_SIZE;
drivers/net/wireless/ralink/rt2x00/rt61pci.c
2954
queue->desc_size = TXD_DESC_SIZE;
drivers/net/wireless/ralink/rt2x00/rt61pci.c
2955
queue->priv_size = sizeof(struct queue_entry_priv_mmio);
drivers/net/wireless/ralink/rt2x00/rt61pci.c
2959
queue->limit = 4;
drivers/net/wireless/ralink/rt2x00/rt61pci.c
2960
queue->data_size = 0; /* No DMA required for beacons */
drivers/net/wireless/ralink/rt2x00/rt61pci.c
2961
queue->desc_size = TXINFO_SIZE;
drivers/net/wireless/ralink/rt2x00/rt61pci.c
2962
queue->priv_size = sizeof(struct queue_entry_priv_mmio);
drivers/net/wireless/ralink/rt2x00/rt73usb.c
1013
static void rt73usb_start_queue(struct data_queue *queue)
drivers/net/wireless/ralink/rt2x00/rt73usb.c
1015
struct rt2x00_dev *rt2x00dev = queue->rt2x00dev;
drivers/net/wireless/ralink/rt2x00/rt73usb.c
1018
switch (queue->qid) {
drivers/net/wireless/ralink/rt2x00/rt73usb.c
1036
static void rt73usb_stop_queue(struct data_queue *queue)
drivers/net/wireless/ralink/rt2x00/rt73usb.c
1038
struct rt2x00_dev *rt2x00dev = queue->rt2x00dev;
drivers/net/wireless/ralink/rt2x00/rt73usb.c
1041
switch (queue->qid) {
drivers/net/wireless/ralink/rt2x00/rt73usb.c
1481
rt2x00_set_field32(&word, TXD_W1_HOST_Q_ID, entry->queue->qid);
drivers/net/wireless/ralink/rt2x00/rt73usb.c
1482
rt2x00_set_field32(&word, TXD_W1_AIFSN, entry->queue->aifs);
drivers/net/wireless/ralink/rt2x00/rt73usb.c
1483
rt2x00_set_field32(&word, TXD_W1_CWMIN, entry->queue->cw_min);
drivers/net/wireless/ralink/rt2x00/rt73usb.c
1484
rt2x00_set_field32(&word, TXD_W1_CWMAX, entry->queue->cw_max);
drivers/net/wireless/ralink/rt2x00/rt73usb.c
1506
TXPOWER_TO_DEV(entry->queue->rt2x00dev->tx_power));
drivers/net/wireless/ralink/rt2x00/rt73usb.c
1524
struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
drivers/net/wireless/ralink/rt2x00/rt73usb.c
1590
struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
drivers/net/wireless/ralink/rt2x00/rt73usb.c
1624
length += (4 * !(length % entry->queue->usb_maxpacket));
drivers/net/wireless/ralink/rt2x00/rt73usb.c
1670
struct rt2x00_dev *rt2x00dev = entry->queue->rt2x00dev;
drivers/net/wireless/ralink/rt2x00/rt73usb.c
1742
skb_pull(entry->skb, entry->queue->desc_size);
drivers/net/wireless/ralink/rt2x00/rt73usb.c
2226
struct data_queue *queue;
drivers/net/wireless/ralink/rt2x00/rt73usb.c
2249
queue = rt2x00queue_get_tx_queue(rt2x00dev, queue_idx);
drivers/net/wireless/ralink/rt2x00/rt73usb.c
2257
rt2x00_set_field32(®, field, queue->txop);
drivers/net/wireless/ralink/rt2x00/rt73usb.c
2265
rt2x00_set_field32(®, field, queue->aifs);
drivers/net/wireless/ralink/rt2x00/rt73usb.c
2269
rt2x00_set_field32(®, field, queue->cw_min);
drivers/net/wireless/ralink/rt2x00/rt73usb.c
2273
rt2x00_set_field32(®, field, queue->cw_max);
drivers/net/wireless/ralink/rt2x00/rt73usb.c
2354
static void rt73usb_queue_init(struct data_queue *queue)
drivers/net/wireless/ralink/rt2x00/rt73usb.c
2356
switch (queue->qid) {
drivers/net/wireless/ralink/rt2x00/rt73usb.c
2358
queue->limit = 32;
drivers/net/wireless/ralink/rt2x00/rt73usb.c
2359
queue->data_size = DATA_FRAME_SIZE;
drivers/net/wireless/ralink/rt2x00/rt73usb.c
2360
queue->desc_size = RXD_DESC_SIZE;
drivers/net/wireless/ralink/rt2x00/rt73usb.c
2361
queue->priv_size = sizeof(struct queue_entry_priv_usb);
drivers/net/wireless/ralink/rt2x00/rt73usb.c
2368
queue->limit = 32;
drivers/net/wireless/ralink/rt2x00/rt73usb.c
2369
queue->data_size = DATA_FRAME_SIZE;
drivers/net/wireless/ralink/rt2x00/rt73usb.c
2370
queue->desc_size = TXD_DESC_SIZE;
drivers/net/wireless/ralink/rt2x00/rt73usb.c
2371
queue->priv_size = sizeof(struct queue_entry_priv_usb);
drivers/net/wireless/ralink/rt2x00/rt73usb.c
2375
queue->limit = 4;
drivers/net/wireless/ralink/rt2x00/rt73usb.c
2376
queue->data_size = MGMT_FRAME_SIZE;
drivers/net/wireless/ralink/rt2x00/rt73usb.c
2377
queue->desc_size = TXINFO_SIZE;
drivers/net/wireless/ralink/rt2x00/rt73usb.c
2378
queue->priv_size = sizeof(struct queue_entry_priv_usb);
drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c
1090
skb_queue_head_init(&priv->tx_ring[prio].queue);
drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c
1104
while (skb_queue_len(&ring->queue)) {
drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c
1106
struct sk_buff *skb = __skb_dequeue(&ring->queue);
drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c
1378
static void rtl8187se_conf_ac_parm(struct ieee80211_hw *dev, u8 queue)
drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c
1390
params = &priv->queue_param[queue];
drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c
1405
switch (queue) {
drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c
1423
unsigned int link_id, u16 queue,
drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c
1437
priv->queue_param[queue] = *params;
drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c
1438
rtl8187se_conf_ac_parm(dev, queue);
drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c
347
while (skb_queue_len(&ring->queue)) {
drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c
357
skb = __skb_dequeue(&ring->queue);
drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c
371
if (ring->entries - skb_queue_len(&ring->queue) == 2)
drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c
548
idx = (ring->idx + skb_queue_len(&ring->queue)) % ring->entries;
drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c
577
__skb_queue_tail(&ring->queue, skb);
drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c
578
if (ring->entries - skb_queue_len(&ring->queue) < 2)
drivers/net/wireless/realtek/rtl818x/rtl8180/rtl8180.h
93
struct sk_buff_head queue;
drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c
1057
while ((skb = skb_dequeue(&priv->b_tx_status.queue)))
drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c
1211
int queue;
drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c
1245
for (queue = 0; queue < 4; queue++)
drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c
1246
rtl818x_iowrite8(priv, (u8 *) rtl8187b_ac_addr[queue],
drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c
1247
priv->aifsn[queue] * priv->slot_time +
drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c
1354
unsigned int link_id, u16 queue,
drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c
1360
if (queue > 3)
drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c
1367
priv->aifsn[queue] = params->aifs;
drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c
1376
rtl818x_iowrite32(priv, rtl8187b_ac_addr[queue],
drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c
1381
if (queue != 0)
drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c
1648
skb_queue_head_init(&priv->b_tx_status.queue);
drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c
199
skb_queue_tail(&priv->b_tx_status.queue, skb);
drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c
202
while (skb_queue_len(&priv->b_tx_status.queue) > 5) {
drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c
208
old_skb = skb_dequeue(&priv->b_tx_status.queue);
drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c
222
skb_queue_tail(&priv->b_tx_status.queue, skb);
drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c
519
spin_lock_irqsave(&priv->b_tx_status.queue.lock, flags);
drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c
521
skb_queue_reverse_walk(&priv->b_tx_status.queue, iter) {
drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c
543
__skb_unlink(skb, &priv->b_tx_status.queue);
drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c
550
spin_unlock_irqrestore(&priv->b_tx_status.queue.lock, flags);
drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c
914
length = skb_queue_len(&priv->b_tx_status.queue);
drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c
920
while (skb_queue_len(&priv->b_tx_status.queue) > 0) {
drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c
923
old_skb = skb_dequeue(&priv->b_tx_status.queue);
drivers/net/wireless/realtek/rtl818x/rtl8187/rtl8187.h
145
struct sk_buff_head queue;
drivers/net/wireless/realtek/rtl8xxxu/core.c
4994
static u32 rtl8xxxu_80211_to_rtl_queue(u32 queue)
drivers/net/wireless/realtek/rtl8xxxu/core.c
4998
switch (queue) {
drivers/net/wireless/realtek/rtl8xxxu/core.c
5020
u32 queue;
drivers/net/wireless/realtek/rtl8xxxu/core.c
5023
queue = TXDESC_QUEUE_BEACON;
drivers/net/wireless/realtek/rtl8xxxu/core.c
5025
queue = TXDESC_QUEUE_MGNT;
drivers/net/wireless/realtek/rtl8xxxu/core.c
5027
queue = rtl8xxxu_80211_to_rtl_queue(skb_get_queue_mapping(skb));
drivers/net/wireless/realtek/rtl8xxxu/core.c
5029
return queue;
drivers/net/wireless/realtek/rtl8xxxu/core.c
5425
u32 queue, rts_rate;
drivers/net/wireless/realtek/rtl8xxxu/core.c
5459
queue = rtl8xxxu_queue_select(hdr, skb);
drivers/net/wireless/realtek/rtl8xxxu/core.c
5478
tx_desc->txdw1 = cpu_to_le32(queue << TXDESC_QUEUE_SHIFT);
drivers/net/wireless/realtek/rtl8xxxu/core.c
5550
usb_fill_bulk_urb(&tx_urb->urb, priv->udev, priv->pipe_out[queue],
drivers/net/wireless/realtek/rtl8xxxu/core.c
6802
unsigned int link_id, u16 queue,
drivers/net/wireless/realtek/rtl8xxxu/core.c
6820
__func__, queue, val32, param->acm, acm_ctrl);
drivers/net/wireless/realtek/rtl8xxxu/core.c
6822
switch (queue) {
drivers/net/wireless/realtek/rtlwifi/base.c
1569
__skb_queue_tail(&tx_report->queue, skb);
drivers/net/wireless/realtek/rtlwifi/base.c
1648
struct sk_buff_head *queue = &tx_report->queue;
drivers/net/wireless/realtek/rtlwifi/base.c
1665
skb_queue_walk(queue, skb) {
drivers/net/wireless/realtek/rtlwifi/base.c
1668
skb_unlink(skb, queue);
drivers/net/wireless/realtek/rtlwifi/base.c
1941
struct sk_buff_head *queue = &tx_report->queue;
drivers/net/wireless/realtek/rtlwifi/base.c
1945
skb_queue_walk_safe(queue, skb, tmp) {
drivers/net/wireless/realtek/rtlwifi/base.c
1950
skb_unlink(skb, queue);
drivers/net/wireless/realtek/rtlwifi/base.c
562
skb_queue_head_init(&rtlpriv->tx_report.queue);
drivers/net/wireless/realtek/rtlwifi/core.c
1851
pskb = __skb_dequeue(&ring->queue);
drivers/net/wireless/realtek/rtlwifi/core.c
1860
__skb_queue_tail(&ring->queue, skb);
drivers/net/wireless/realtek/rtlwifi/core.c
923
static int _rtl_get_hal_qnum(u16 queue)
drivers/net/wireless/realtek/rtlwifi/core.c
927
switch (queue) {
drivers/net/wireless/realtek/rtlwifi/core.c
953
unsigned int link_id, u16 queue,
drivers/net/wireless/realtek/rtlwifi/core.c
960
if (queue >= AC_MAX) {
drivers/net/wireless/realtek/rtlwifi/core.c
962
"queue number %d is incorrect!\n", queue);
drivers/net/wireless/realtek/rtlwifi/core.c
966
aci = _rtl_get_hal_qnum(queue);
drivers/net/wireless/realtek/rtlwifi/pci.c
1023
pskb = __skb_dequeue(&ring->queue);
drivers/net/wireless/realtek/rtlwifi/pci.c
1050
__skb_queue_tail(&ring->queue, pskb);
drivers/net/wireless/realtek/rtlwifi/pci.c
1183
skb_queue_head_init(&rtlpci->tx_ring[prio].queue);
drivers/net/wireless/realtek/rtlwifi/pci.c
1271
while (skb_queue_len(&ring->queue)) {
drivers/net/wireless/realtek/rtlwifi/pci.c
1273
struct sk_buff *skb = __skb_dequeue(&ring->queue);
drivers/net/wireless/realtek/rtlwifi/pci.c
1449
while (skb_queue_len(&ring->queue)) {
drivers/net/wireless/realtek/rtlwifi/pci.c
1452
__skb_dequeue(&ring->queue);
drivers/net/wireless/realtek/rtlwifi/pci.c
1564
idx = (ring->idx + skb_queue_len(&ring->queue)) %
drivers/net/wireless/realtek/rtlwifi/pci.c
1581
skb_queue_len(&ring->queue));
drivers/net/wireless/realtek/rtlwifi/pci.c
1603
__skb_queue_tail(&ring->queue, skb);
drivers/net/wireless/realtek/rtlwifi/pci.c
1613
if ((ring->entries - skb_queue_len(&ring->queue)) < 2 &&
drivers/net/wireless/realtek/rtlwifi/pci.c
1618
skb_queue_len(&ring->queue));
drivers/net/wireless/realtek/rtlwifi/pci.c
1651
queue_len = skb_queue_len(&ring->queue);
drivers/net/wireless/realtek/rtlwifi/pci.c
428
(ring->entries - skb_queue_len(&ring->queue) >
drivers/net/wireless/realtek/rtlwifi/pci.c
457
while (skb_queue_len(&ring->queue)) {
drivers/net/wireless/realtek/rtlwifi/pci.c
473
skb = __skb_dequeue(&ring->queue);
drivers/net/wireless/realtek/rtlwifi/pci.c
486
skb_queue_len(&ring->queue),
drivers/net/wireless/realtek/rtlwifi/pci.c
533
if ((ring->entries - skb_queue_len(&ring->queue)) <= 4) {
drivers/net/wireless/realtek/rtlwifi/pci.c
537
skb_queue_len(&ring->queue));
drivers/net/wireless/realtek/rtlwifi/pci.h
133
struct sk_buff_head queue;
drivers/net/wireless/realtek/rtlwifi/rtl8188ee/hw.c
176
u8 queue;
drivers/net/wireless/realtek/rtlwifi/rtl8188ee/hw.c
188
for (queue = 0; queue < RTL_PCI_MAX_TX_QUEUE_COUNT; queue++) {
drivers/net/wireless/realtek/rtlwifi/rtl8188ee/hw.c
189
ring = &rtlpci->tx_ring[queue];
drivers/net/wireless/realtek/rtlwifi/rtl8188ee/hw.c
190
if (skb_queue_len(&ring->queue)) {
drivers/net/wireless/realtek/rtlwifi/rtl8188ee/hw.c
76
while (skb_queue_len(&ring->queue)) {
drivers/net/wireless/realtek/rtlwifi/rtl8188ee/hw.c
78
struct sk_buff *skb = __skb_dequeue(&ring->queue);
drivers/net/wireless/realtek/rtlwifi/rtl8188ee/phy.c
2199
skb_queue_len(&ring->queue) == 0) {
drivers/net/wireless/realtek/rtlwifi/rtl8188ee/phy.c
2206
skb_queue_len(&ring->queue));
drivers/net/wireless/realtek/rtlwifi/rtl8188ee/phy.c
2216
skb_queue_len(&ring->queue));
drivers/net/wireless/realtek/rtlwifi/rtl8188ee/phy.c
2242
if (skb_queue_len(&ring->queue) == 0) {
drivers/net/wireless/realtek/rtlwifi/rtl8188ee/phy.c
2249
skb_queue_len(&ring->queue));
drivers/net/wireless/realtek/rtlwifi/rtl8188ee/phy.c
2259
skb_queue_len(&ring->queue));
drivers/net/wireless/realtek/rtlwifi/rtl8192ce/phy.c
466
skb_queue_len(&ring->queue) == 0) {
drivers/net/wireless/realtek/rtlwifi/rtl8192ce/phy.c
473
skb_queue_len(&ring->queue));
drivers/net/wireless/realtek/rtlwifi/rtl8192ce/phy.c
483
skb_queue_len(&ring->queue));
drivers/net/wireless/realtek/rtlwifi/rtl8192cu/mac.c
467
u16 queue, u16 txop, u8 cw_min, u8 cw_max, u8 aifs)
drivers/net/wireless/realtek/rtlwifi/rtl8192cu/mac.c
480
rtl_write_dword(rtlpriv, (REG_EDCA_VO_PARAM + (queue * 4)), value);
drivers/net/wireless/realtek/rtlwifi/rtl8192cu/mac.h
33
u16 queue,
drivers/net/wireless/realtek/rtlwifi/rtl8192cu/phy.c
419
if (skb_queue_len(&ring->queue) == 0 ||
drivers/net/wireless/realtek/rtlwifi/rtl8192cu/phy.c
428
skb_queue_len(&ring->queue));
drivers/net/wireless/realtek/rtlwifi/rtl8192cu/phy.c
437
skb_queue_len(&ring->queue));
drivers/net/wireless/realtek/rtlwifi/rtl8192cu/phy.c
462
if (skb_queue_len(&ring->queue) == 0) {
drivers/net/wireless/realtek/rtlwifi/rtl8192cu/phy.c
469
skb_queue_len(&ring->queue));
drivers/net/wireless/realtek/rtlwifi/rtl8192cu/phy.c
478
skb_queue_len(&ring->queue));
drivers/net/wireless/realtek/rtlwifi/rtl8192de/fw.c
124
pskb = __skb_dequeue(&ring->queue);
drivers/net/wireless/realtek/rtlwifi/rtl8192de/fw.c
131
__skb_queue_tail(&ring->queue, skb);
drivers/net/wireless/realtek/rtlwifi/rtl8192de/phy.c
2406
if (skb_queue_len(&ring->queue) == 0 ||
drivers/net/wireless/realtek/rtlwifi/rtl8192de/phy.c
2419
skb_queue_len(&ring->queue));
drivers/net/wireless/realtek/rtlwifi/rtl8192de/phy.c
2428
skb_queue_len(&ring->queue));
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c
150
u8 queue;
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c
163
for (queue = 0; queue < RTL_PCI_MAX_TX_QUEUE_COUNT; queue++) {
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c
164
ring = &rtlpci->tx_ring[queue];
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/hw.c
165
if (skb_queue_len(&ring->queue)) {
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/phy.c
3072
skb_queue_len(&ring->queue) == 0) {
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/phy.c
3079
skb_queue_len(&ring->queue));
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/phy.c
3089
skb_queue_len(&ring->queue));
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/phy.c
3115
if (skb_queue_len(&ring->queue) == 0) {
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/phy.c
3122
skb_queue_len(&ring->queue));
drivers/net/wireless/realtek/rtlwifi/rtl8192ee/phy.c
3131
skb_queue_len(&ring->queue));
drivers/net/wireless/realtek/rtlwifi/rtl8192se/fw.c
123
idx = (ring->idx + skb_queue_len(&ring->queue)) % ring->entries;
drivers/net/wireless/realtek/rtlwifi/rtl8192se/fw.c
126
__skb_queue_tail(&ring->queue, skb);
drivers/net/wireless/realtek/rtlwifi/rtl8192se/phy.c
568
if (skb_queue_len(&ring->queue) == 0 ||
drivers/net/wireless/realtek/rtlwifi/rtl8192se/phy.c
576
skb_queue_len(&ring->queue));
drivers/net/wireless/realtek/rtlwifi/rtl8192se/phy.c
587
skb_queue_len(&ring->queue));
drivers/net/wireless/realtek/rtlwifi/rtl8723ae/phy.c
1615
skb_queue_len(&ring->queue) == 0) {
drivers/net/wireless/realtek/rtlwifi/rtl8723ae/phy.c
1622
skb_queue_len(&ring->queue));
drivers/net/wireless/realtek/rtlwifi/rtl8723ae/phy.c
1632
skb_queue_len(&ring->queue));
drivers/net/wireless/realtek/rtlwifi/rtl8723be/hw.c
179
u8 queue;
drivers/net/wireless/realtek/rtlwifi/rtl8723be/hw.c
191
for (queue = 0; queue < RTL_PCI_MAX_TX_QUEUE_COUNT; queue++) {
drivers/net/wireless/realtek/rtlwifi/rtl8723be/hw.c
192
ring = &rtlpci->tx_ring[queue];
drivers/net/wireless/realtek/rtlwifi/rtl8723be/hw.c
193
if (skb_queue_len(&ring->queue)) {
drivers/net/wireless/realtek/rtlwifi/rtl8723be/hw.c
38
while (skb_queue_len(&ring->queue)) {
drivers/net/wireless/realtek/rtlwifi/rtl8723be/hw.c
40
struct sk_buff *skb = __skb_dequeue(&ring->queue);
drivers/net/wireless/realtek/rtlwifi/rtl8723be/phy.c
2561
skb_queue_len(&ring->queue) == 0) {
drivers/net/wireless/realtek/rtlwifi/rtl8723be/phy.c
2568
skb_queue_len(&ring->queue));
drivers/net/wireless/realtek/rtlwifi/rtl8723be/phy.c
2578
skb_queue_len(&ring->queue));
drivers/net/wireless/realtek/rtlwifi/rtl8723be/phy.c
2605
if (skb_queue_len(&ring->queue) == 0) {
drivers/net/wireless/realtek/rtlwifi/rtl8723be/phy.c
2612
skb_queue_len(&ring->queue));
drivers/net/wireless/realtek/rtlwifi/rtl8723be/phy.c
2622
skb_queue_len(&ring->queue));
drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c
177
u8 queue;
drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c
189
for (queue = 0; queue < RTL_PCI_MAX_TX_QUEUE_COUNT; queue++) {
drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c
190
ring = &rtlpci->tx_ring[queue];
drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c
191
if (skb_queue_len(&ring->queue)) {
drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c
34
while (skb_queue_len(&ring->queue)) {
drivers/net/wireless/realtek/rtlwifi/rtl8821ae/hw.c
36
struct sk_buff *skb = __skb_dequeue(&ring->queue);
drivers/net/wireless/realtek/rtlwifi/rtl8821ae/phy.c
4728
skb_queue_len(&ring->queue) == 0) {
drivers/net/wireless/realtek/rtlwifi/rtl8821ae/phy.c
4735
skb_queue_len(&ring->queue));
drivers/net/wireless/realtek/rtlwifi/rtl8821ae/phy.c
4745
skb_queue_len(&ring->queue));
drivers/net/wireless/realtek/rtlwifi/wifi.h
1949
struct sk_buff_head queue;
drivers/net/wireless/realtek/rtw88/coex.c
644
skb_queue_tail(&coex->queue, skb);
drivers/net/wireless/realtek/rtw88/coex.c
658
if (!wait_event_timeout(coex->wait, !skb_queue_empty(&coex->queue),
drivers/net/wireless/realtek/rtw88/coex.c
664
skb_resp = skb_dequeue(&coex->queue);
drivers/net/wireless/realtek/rtw88/main.c
2166
skb_queue_head_init(&rtwdev->coex.queue);
drivers/net/wireless/realtek/rtw88/main.c
2167
skb_queue_head_init(&rtwdev->tx_report.queue);
drivers/net/wireless/realtek/rtw88/main.c
2233
skb_queue_purge(&rtwdev->tx_report.queue);
drivers/net/wireless/realtek/rtw88/main.c
2235
skb_queue_purge(&rtwdev->coex.queue);
drivers/net/wireless/realtek/rtw88/main.h
1562
struct sk_buff_head queue;
drivers/net/wireless/realtek/rtw88/main.h
744
struct sk_buff_head queue;
drivers/net/wireless/realtek/rtw88/pci.c
102
skb_queue_walk_safe(&tx_ring->queue, skb, tmp) {
drivers/net/wireless/realtek/rtw88/pci.c
103
__skb_unlink(skb, &tx_ring->queue);
drivers/net/wireless/realtek/rtw88/pci.c
198
skb_queue_head_init(&tx_ring->queue);
drivers/net/wireless/realtek/rtw88/pci.c
35
enum rtw_tx_queue_type queue)
drivers/net/wireless/realtek/rtw88/pci.c
37
switch (queue) {
drivers/net/wireless/realtek/rtw88/pci.c
540
enum rtw_tx_queue_type queue;
drivers/net/wireless/realtek/rtw88/pci.c
543
for (queue = 0; queue < RTK_MAX_TX_QUEUE_NUM; queue++) {
drivers/net/wireless/realtek/rtw88/pci.c
544
tx_ring = &rtwpci->tx_rings[queue];
drivers/net/wireless/realtek/rtw88/pci.c
606
enum rtw_tx_queue_type queue;
drivers/net/wireless/realtek/rtw88/pci.c
615
for (queue = 0; queue < RTK_MAX_TX_QUEUE_NUM; queue++) {
drivers/net/wireless/realtek/rtw88/pci.c
619
if (queue == RTW_TX_QUEUE_BCN ||
drivers/net/wireless/realtek/rtw88/pci.c
620
queue == RTW_TX_QUEUE_H2C)
drivers/net/wireless/realtek/rtw88/pci.c
623
tx_ring = &rtwpci->tx_rings[queue];
drivers/net/wireless/realtek/rtw88/pci.c
626
if (skb_queue_len(&tx_ring->queue)) {
drivers/net/wireless/realtek/rtw88/pci.c
670
struct sk_buff *prev = skb_dequeue(&ring->queue);
drivers/net/wireless/realtek/rtw88/pci.c
773
enum rtw_tx_queue_type queue)
drivers/net/wireless/realtek/rtw88/pci.c
779
ring = &rtwpci->tx_rings[queue];
drivers/net/wireless/realtek/rtw88/pci.c
780
bd_idx = rtw_pci_tx_queue_idx_addr[queue];
drivers/net/wireless/realtek/rtw88/pci.c
792
enum rtw_tx_queue_type queue;
drivers/net/wireless/realtek/rtw88/pci.c
794
for (queue = 0; queue < RTK_MAX_TX_QUEUE_NUM; queue++)
drivers/net/wireless/realtek/rtw88/pci.c
795
if (test_and_clear_bit(queue, rtwpci->tx_queued))
drivers/net/wireless/realtek/rtw88/pci.c
796
rtw_pci_tx_kick_off_queue(rtwdev, queue);
drivers/net/wireless/realtek/rtw88/pci.c
802
enum rtw_tx_queue_type queue)
drivers/net/wireless/realtek/rtw88/pci.c
816
ring = &rtwpci->tx_rings[queue];
drivers/net/wireless/realtek/rtw88/pci.c
820
if (queue == RTW_TX_QUEUE_BCN)
drivers/net/wireless/realtek/rtw88/pci.c
827
pkt_info->qsel = rtw_pci_get_tx_qsel(skb, queue);
drivers/net/wireless/realtek/rtw88/pci.c
838
if (queue == RTW_TX_QUEUE_BCN)
drivers/net/wireless/realtek/rtw88/pci.c
853
skb_queue_tail(&ring->queue, skb);
drivers/net/wireless/realtek/rtw88/pci.c
855
if (queue == RTW_TX_QUEUE_BCN)
drivers/net/wireless/realtek/rtw88/pci.c
859
set_bit(queue, rtwpci->tx_queued);
drivers/net/wireless/realtek/rtw88/pci.c
920
enum rtw_tx_queue_type queue = rtw_tx_queue_mapping(skb);
drivers/net/wireless/realtek/rtw88/pci.c
925
ret = rtw_pci_tx_write_data(rtwdev, pkt_info, skb, queue);
drivers/net/wireless/realtek/rtw88/pci.c
929
ring = &rtwpci->tx_rings[queue];
drivers/net/wireless/realtek/rtw88/pci.c
966
skb = skb_dequeue(&ring->queue);
drivers/net/wireless/realtek/rtw88/pci.h
189
struct sk_buff_head queue;
drivers/net/wireless/realtek/rtw88/pci.h
240
static inline u32 max_num_of_tx_queue(u8 queue)
drivers/net/wireless/realtek/rtw88/pci.h
244
switch (queue) {
drivers/net/wireless/realtek/rtw88/sdio.c
1229
enum rtw_tx_queue_type queue)
drivers/net/wireless/realtek/rtw88/sdio.c
1235
skb = skb_dequeue(&rtwsdio->tx_queue[queue]);
drivers/net/wireless/realtek/rtw88/sdio.c
1239
ret = rtw_sdio_write_port(rtwdev, skb, queue);
drivers/net/wireless/realtek/rtw88/sdio.c
1241
skb_queue_head(&rtwsdio->tx_queue[queue], skb);
drivers/net/wireless/realtek/rtw88/sdio.c
1254
int limit, queue;
drivers/net/wireless/realtek/rtw88/sdio.c
1262
for (queue = RTK_MAX_TX_QUEUE_NUM - 1; queue >= 0; queue--) {
drivers/net/wireless/realtek/rtw88/sdio.c
1264
rtw_sdio_process_tx_queue(rtwdev, queue);
drivers/net/wireless/realtek/rtw88/sdio.c
1266
if (skb_queue_empty(&rtwsdio->tx_queue[queue]))
drivers/net/wireless/realtek/rtw88/sdio.c
471
enum rtw_tx_queue_type queue)
drivers/net/wireless/realtek/rtw88/sdio.c
475
switch (queue) {
drivers/net/wireless/realtek/rtw88/sdio.c
498
queue);
drivers/net/wireless/realtek/rtw88/sdio.c
551
static int rtw_sdio_check_free_txpg(struct rtw_dev *rtwdev, u8 queue,
drivers/net/wireless/realtek/rtw88/sdio.c
561
switch (queue) {
drivers/net/wireless/realtek/rtw88/sdio.c
580
rtw_warn(rtwdev, "Unknown mapping for queue %u\n", queue);
drivers/net/wireless/realtek/rtw88/sdio.c
593
switch (queue) {
drivers/net/wireless/realtek/rtw88/sdio.c
615
rtw_warn(rtwdev, "Unknown mapping for queue %u\n", queue);
drivers/net/wireless/realtek/rtw88/sdio.c
628
pages_needed, pages_free, queue, count);
drivers/net/wireless/realtek/rtw88/sdio.c
636
enum rtw_tx_queue_type queue)
drivers/net/wireless/realtek/rtw88/sdio.c
644
txaddr = rtw_sdio_get_tx_addr(rtwdev, skb->len, queue);
drivers/net/wireless/realtek/rtw88/sdio.c
650
ret = rtw_sdio_check_free_txpg(rtwdev, queue, txsize);
drivers/net/wireless/realtek/rtw88/sdio.c
656
__func__, queue);
drivers/net/wireless/realtek/rtw88/sdio.c
730
u8 queue)
drivers/net/wireless/realtek/rtw88/sdio.c
732
switch (queue) {
drivers/net/wireless/realtek/rtw88/sdio.c
769
u8 queue;
drivers/net/wireless/realtek/rtw88/sdio.c
773
for (queue = 0; queue < RTK_MAX_TX_QUEUE_NUM; queue++) {
drivers/net/wireless/realtek/rtw88/sdio.c
777
if (queue == RTW_TX_QUEUE_BCN ||
drivers/net/wireless/realtek/rtw88/sdio.c
778
queue == RTW_TX_QUEUE_H2C)
drivers/net/wireless/realtek/rtw88/sdio.c
782
if (skb_queue_len(&rtwsdio->tx_queue[queue])) {
drivers/net/wireless/realtek/rtw88/sdio.c
850
enum rtw_tx_queue_type queue)
drivers/net/wireless/realtek/rtw88/sdio.c
879
pkt_info->qsel = rtw_sdio_get_tx_qsel(rtwdev, skb, queue);
drivers/net/wireless/realtek/rtw88/sdio.c
888
enum rtw_tx_queue_type queue)
drivers/net/wireless/realtek/rtw88/sdio.c
892
rtw_sdio_tx_skb_prepare(rtwdev, pkt_info, skb, queue);
drivers/net/wireless/realtek/rtw88/sdio.c
894
ret = rtw_sdio_write_port(rtwdev, skb, queue);
drivers/net/wireless/realtek/rtw88/sdio.c
930
enum rtw_tx_queue_type queue = rtw_tx_queue_mapping(skb);
drivers/net/wireless/realtek/rtw88/sdio.c
933
rtw_sdio_tx_skb_prepare(rtwdev, pkt_info, skb, queue);
drivers/net/wireless/realtek/rtw88/sdio.c
938
skb_queue_tail(&rtwsdio->tx_queue[queue], skb);
drivers/net/wireless/realtek/rtw88/tx.c
186
if (skb_queue_len(&tx_report->queue) == 0)
drivers/net/wireless/realtek/rtw88/tx.c
192
skb_queue_purge(&tx_report->queue);
drivers/net/wireless/realtek/rtw88/tx.c
207
__skb_queue_tail(&tx_report->queue, skb);
drivers/net/wireless/realtek/rtw88/tx.c
249
skb_queue_walk_safe(&tx_report->queue, cur, tmp) {
drivers/net/wireless/realtek/rtw88/tx.c
252
__skb_unlink(cur, &tx_report->queue);
drivers/net/wireless/realtek/rtw88/tx.c
736
enum rtw_tx_queue_type queue;
drivers/net/wireless/realtek/rtw88/tx.c
739
queue = RTW_TX_QUEUE_BCN;
drivers/net/wireless/realtek/rtw88/tx.c
741
queue = RTW_TX_QUEUE_MGMT;
drivers/net/wireless/realtek/rtw88/tx.c
744
queue = RTW_TX_QUEUE_HI0;
drivers/net/wireless/realtek/rtw88/tx.c
746
queue = ac_to_hwq[IEEE80211_AC_BE];
drivers/net/wireless/realtek/rtw88/tx.c
748
queue = ac_to_hwq[q_mapping];
drivers/net/wireless/realtek/rtw88/tx.c
750
return queue;
drivers/net/wireless/realtek/rtw88/usb.c
417
goto queue;
drivers/net/wireless/realtek/rtw88/usb.c
423
goto queue;
drivers/net/wireless/realtek/rtw88/usb.c
454
queue:
drivers/net/wireless/realtek/rtw89/pci.c
1543
skb_queue_tail(&txwd->queue, skb);
drivers/net/wireless/realtek/rtw89/pci.c
3558
skb_queue_head_init(&txwd->queue);
drivers/net/wireless/realtek/rtw89/pci.c
517
if (skb_queue_len(&txwd->queue) == 0)
drivers/net/wireless/realtek/rtw89/pci.c
558
skb_queue_walk_safe(&txwd->queue, skb, tmp) {
drivers/net/wireless/realtek/rtw89/pci.c
559
skb_unlink(skb, &txwd->queue);
drivers/net/wireless/realtek/rtw89/pci.h
1535
struct sk_buff_head queue;
drivers/net/wireless/rsi/rsi_91x_mac80211.c
903
unsigned int link_id, u16 queue,
drivers/net/wireless/rsi/rsi_91x_mac80211.c
910
if (queue >= IEEE80211_NUM_ACS)
drivers/net/wireless/rsi/rsi_91x_mac80211.c
915
__func__, queue, params->aifs,
drivers/net/wireless/rsi/rsi_91x_mac80211.c
920
switch (queue) {
drivers/net/wireless/silabs/wfx/data_tx.c
548
struct wfx_queue *queue;
drivers/net/wireless/silabs/wfx/data_tx.c
554
queue = &wvif->tx_queue[i];
drivers/net/wireless/silabs/wfx/data_tx.c
556
wfx_tx_queue_drop(wvif, queue, dropped);
drivers/net/wireless/silabs/wfx/data_tx.c
563
queue = &wvif->tx_queue[i];
drivers/net/wireless/silabs/wfx/data_tx.c
564
if (wait_event_timeout(wvif->wdev->tx_dequeue, wfx_tx_queue_empty(wvif, queue),
drivers/net/wireless/silabs/wfx/hif_tx.c
409
int wfx_hif_set_edca_queue_params(struct wfx_vif *wvif, u16 queue,
drivers/net/wireless/silabs/wfx/hif_tx.c
426
body->queue_id = 3 - queue;
drivers/net/wireless/silabs/wfx/hif_tx.c
428
if (wfx_api_older_than(wvif->wdev, 2, 0) && queue == IEEE80211_AC_BE)
drivers/net/wireless/silabs/wfx/hif_tx.c
430
if (wfx_api_older_than(wvif->wdev, 2, 0) && queue == IEEE80211_AC_BK)
drivers/net/wireless/silabs/wfx/hif_tx.h
51
int wfx_hif_set_edca_queue_params(struct wfx_vif *wvif, u16 queue,
drivers/net/wireless/silabs/wfx/queue.c
106
void wfx_tx_queue_drop(struct wfx_vif *wvif, struct wfx_queue *queue,
drivers/net/wireless/silabs/wfx/queue.c
109
__wfx_tx_queue_drop(wvif, &queue->normal, dropped);
drivers/net/wireless/silabs/wfx/queue.c
110
__wfx_tx_queue_drop(wvif, &queue->cab, dropped);
drivers/net/wireless/silabs/wfx/queue.c
111
__wfx_tx_queue_drop(wvif, &queue->offchan, dropped);
drivers/net/wireless/silabs/wfx/queue.c
117
struct wfx_queue *queue = &wvif->tx_queue[skb_get_queue_mapping(skb)];
drivers/net/wireless/silabs/wfx/queue.c
121
skb_queue_tail(&queue->offchan, skb);
drivers/net/wireless/silabs/wfx/queue.c
123
skb_queue_tail(&queue->cab, skb);
drivers/net/wireless/silabs/wfx/queue.c
125
skb_queue_tail(&queue->normal, skb);
drivers/net/wireless/silabs/wfx/queue.c
130
struct wfx_queue *queue;
drivers/net/wireless/silabs/wfx/queue.c
138
queue = &wvif->tx_queue[skb_get_queue_mapping(skb)];
drivers/net/wireless/silabs/wfx/queue.c
140
WARN_ON(!atomic_read(&queue->pending_frames));
drivers/net/wireless/silabs/wfx/queue.c
141
atomic_dec(&queue->pending_frames);
drivers/net/wireless/silabs/wfx/queue.c
149
struct wfx_queue *queue;
drivers/net/wireless/silabs/wfx/queue.c
164
queue = &wvif->tx_queue[skb_get_queue_mapping(skb)];
drivers/net/wireless/silabs/wfx/queue.c
166
WARN_ON(!atomic_read(&queue->pending_frames));
drivers/net/wireless/silabs/wfx/queue.c
167
atomic_dec(&queue->pending_frames);
drivers/net/wireless/silabs/wfx/queue.c
226
static int wfx_tx_queue_get_weight(struct wfx_queue *queue)
drivers/net/wireless/silabs/wfx/queue.c
228
return atomic_read(&queue->pending_frames) * queue->priority;
drivers/net/wireless/silabs/wfx/queue.c
76
bool wfx_tx_queue_empty(struct wfx_vif *wvif, struct wfx_queue *queue)
drivers/net/wireless/silabs/wfx/queue.c
78
return skb_queue_empty_lockless(&queue->normal) &&
drivers/net/wireless/silabs/wfx/queue.c
79
skb_queue_empty_lockless(&queue->cab) &&
drivers/net/wireless/silabs/wfx/queue.c
80
skb_queue_empty_lockless(&queue->offchan);
drivers/net/wireless/silabs/wfx/queue.h
36
bool wfx_tx_queue_empty(struct wfx_vif *wvif, struct wfx_queue *queue);
drivers/net/wireless/silabs/wfx/queue.h
37
void wfx_tx_queue_drop(struct wfx_vif *wvif, struct wfx_queue *queue,
drivers/net/wireless/silabs/wfx/sta.c
202
unsigned int link_id, u16 queue,
drivers/net/wireless/silabs/wfx/sta.c
209
WARN_ON(queue >= hw->queues);
drivers/net/wireless/silabs/wfx/sta.c
212
assign_bit(queue, &wvif->uapsd_mask, params->uapsd);
drivers/net/wireless/silabs/wfx/sta.c
213
wfx_hif_set_edca_queue_params(wvif, queue, params);
drivers/net/wireless/silabs/wfx/sta.h
39
unsigned int link_id, u16 queue,
drivers/net/wireless/silabs/wfx/traces.h
448
const struct wfx_queue *queue;
drivers/net/wireless/silabs/wfx/traces.h
464
queue = &wvif->tx_queue[i];
drivers/net/wireless/silabs/wfx/traces.h
465
__entry->hw[j] = atomic_read(&queue->pending_frames);
drivers/net/wireless/silabs/wfx/traces.h
466
__entry->drv[j] = skb_queue_len(&queue->normal);
drivers/net/wireless/silabs/wfx/traces.h
467
__entry->cab[j] = skb_queue_len(&queue->cab);
drivers/net/wireless/silabs/wfx/traces.h
468
if (queue == elected_queue) {
drivers/net/wireless/st/cw1200/queue.c
102
--queue->num_queued;
drivers/net/wireless/st/cw1200/queue.c
103
--queue->link_map_cache[iter->txpriv.link_id];
drivers/net/wireless/st/cw1200/queue.c
112
list_move_tail(&iter->head, &queue->free_pool);
drivers/net/wireless/st/cw1200/queue.c
118
if (queue->overfull) {
drivers/net/wireless/st/cw1200/queue.c
119
if (queue->num_queued <= (queue->capacity >> 1)) {
drivers/net/wireless/st/cw1200/queue.c
120
queue->overfull = false;
drivers/net/wireless/st/cw1200/queue.c
122
__cw1200_queue_unlock(queue);
drivers/net/wireless/st/cw1200/queue.c
124
unsigned long tmo = item->queue_timestamp + queue->ttl;
drivers/net/wireless/st/cw1200/queue.c
125
mod_timer(&queue->gc, tmo);
drivers/net/wireless/st/cw1200/queue.c
135
struct cw1200_queue *queue =
drivers/net/wireless/st/cw1200/queue.c
136
timer_container_of(queue, t, gc);
drivers/net/wireless/st/cw1200/queue.c
138
spin_lock_bh(&queue->lock);
drivers/net/wireless/st/cw1200/queue.c
139
__cw1200_queue_gc(queue, &list, true);
drivers/net/wireless/st/cw1200/queue.c
140
spin_unlock_bh(&queue->lock);
drivers/net/wireless/st/cw1200/queue.c
141
cw1200_queue_post_gc(queue->stats, &list);
drivers/net/wireless/st/cw1200/queue.c
163
int cw1200_queue_init(struct cw1200_queue *queue,
drivers/net/wireless/st/cw1200/queue.c
171
memset(queue, 0, sizeof(*queue));
drivers/net/wireless/st/cw1200/queue.c
172
queue->stats = stats;
drivers/net/wireless/st/cw1200/queue.c
173
queue->capacity = capacity;
drivers/net/wireless/st/cw1200/queue.c
174
queue->queue_id = queue_id;
drivers/net/wireless/st/cw1200/queue.c
175
queue->ttl = ttl;
drivers/net/wireless/st/cw1200/queue.c
176
INIT_LIST_HEAD(&queue->queue);
drivers/net/wireless/st/cw1200/queue.c
177
INIT_LIST_HEAD(&queue->pending);
drivers/net/wireless/st/cw1200/queue.c
178
INIT_LIST_HEAD(&queue->free_pool);
drivers/net/wireless/st/cw1200/queue.c
179
spin_lock_init(&queue->lock);
drivers/net/wireless/st/cw1200/queue.c
180
timer_setup(&queue->gc, cw1200_queue_gc, 0);
drivers/net/wireless/st/cw1200/queue.c
182
queue->pool = kzalloc_objs(struct cw1200_queue_item, capacity);
drivers/net/wireless/st/cw1200/queue.c
183
if (!queue->pool)
drivers/net/wireless/st/cw1200/queue.c
186
queue->link_map_cache = kzalloc_objs(int, stats->map_capacity);
drivers/net/wireless/st/cw1200/queue.c
187
if (!queue->link_map_cache) {
drivers/net/wireless/st/cw1200/queue.c
188
kfree(queue->pool);
drivers/net/wireless/st/cw1200/queue.c
189
queue->pool = NULL;
drivers/net/wireless/st/cw1200/queue.c
194
list_add_tail(&queue->pool[i].head, &queue->free_pool);
drivers/net/wireless/st/cw1200/queue.c
199
int cw1200_queue_clear(struct cw1200_queue *queue)
drivers/net/wireless/st/cw1200/queue.c
203
struct cw1200_queue_stats *stats = queue->stats;
drivers/net/wireless/st/cw1200/queue.c
206
spin_lock_bh(&queue->lock);
drivers/net/wireless/st/cw1200/queue.c
207
queue->generation++;
drivers/net/wireless/st/cw1200/queue.c
208
list_splice_tail_init(&queue->queue, &queue->pending);
drivers/net/wireless/st/cw1200/queue.c
209
list_for_each_entry_safe(item, tmp, &queue->pending, head) {
drivers/net/wireless/st/cw1200/queue.c
213
list_move_tail(&item->head, &queue->free_pool);
drivers/net/wireless/st/cw1200/queue.c
215
queue->num_queued = 0;
drivers/net/wireless/st/cw1200/queue.c
216
queue->num_pending = 0;
drivers/net/wireless/st/cw1200/queue.c
220
stats->num_queued -= queue->link_map_cache[i];
drivers/net/wireless/st/cw1200/queue.c
221
stats->link_map_cache[i] -= queue->link_map_cache[i];
drivers/net/wireless/st/cw1200/queue.c
222
queue->link_map_cache[i] = 0;
drivers/net/wireless/st/cw1200/queue.c
225
if (queue->overfull) {
drivers/net/wireless/st/cw1200/queue.c
226
queue->overfull = false;
drivers/net/wireless/st/cw1200/queue.c
227
__cw1200_queue_unlock(queue);
drivers/net/wireless/st/cw1200/queue.c
229
spin_unlock_bh(&queue->lock);
drivers/net/wireless/st/cw1200/queue.c
241
void cw1200_queue_deinit(struct cw1200_queue *queue)
drivers/net/wireless/st/cw1200/queue.c
243
cw1200_queue_clear(queue);
drivers/net/wireless/st/cw1200/queue.c
244
timer_delete_sync(&queue->gc);
drivers/net/wireless/st/cw1200/queue.c
245
INIT_LIST_HEAD(&queue->free_pool);
drivers/net/wireless/st/cw1200/queue.c
246
kfree(queue->pool);
drivers/net/wireless/st/cw1200/queue.c
247
kfree(queue->link_map_cache);
drivers/net/wireless/st/cw1200/queue.c
248
queue->pool = NULL;
drivers/net/wireless/st/cw1200/queue.c
249
queue->link_map_cache = NULL;
drivers/net/wireless/st/cw1200/queue.c
250
queue->capacity = 0;
drivers/net/wireless/st/cw1200/queue.c
253
size_t cw1200_queue_get_num_queued(struct cw1200_queue *queue,
drivers/net/wireless/st/cw1200/queue.c
258
size_t map_capacity = queue->stats->map_capacity;
drivers/net/wireless/st/cw1200/queue.c
263
spin_lock_bh(&queue->lock);
drivers/net/wireless/st/cw1200/queue.c
265
ret = queue->num_queued - queue->num_pending;
drivers/net/wireless/st/cw1200/queue.c
27
static inline void __cw1200_queue_lock(struct cw1200_queue *queue)
drivers/net/wireless/st/cw1200/queue.c
270
ret += queue->link_map_cache[i];
drivers/net/wireless/st/cw1200/queue.c
273
spin_unlock_bh(&queue->lock);
drivers/net/wireless/st/cw1200/queue.c
277
int cw1200_queue_put(struct cw1200_queue *queue,
drivers/net/wireless/st/cw1200/queue.c
282
struct cw1200_queue_stats *stats = queue->stats;
drivers/net/wireless/st/cw1200/queue.c
284
if (txpriv->link_id >= queue->stats->map_capacity)
drivers/net/wireless/st/cw1200/queue.c
287
spin_lock_bh(&queue->lock);
drivers/net/wireless/st/cw1200/queue.c
288
if (!WARN_ON(list_empty(&queue->free_pool))) {
drivers/net/wireless/st/cw1200/queue.c
29
struct cw1200_queue_stats *stats = queue->stats;
drivers/net/wireless/st/cw1200/queue.c
290
&queue->free_pool, struct cw1200_queue_item, head);
drivers/net/wireless/st/cw1200/queue.c
293
list_move_tail(&item->head, &queue->queue);
drivers/net/wireless/st/cw1200/queue.c
297
item->packet_id = cw1200_queue_mk_packet_id(queue->generation,
drivers/net/wireless/st/cw1200/queue.c
298
queue->queue_id,
drivers/net/wireless/st/cw1200/queue.c
30
if (queue->tx_locked_cnt++ == 0) {
drivers/net/wireless/st/cw1200/queue.c
300
item - queue->pool);
drivers/net/wireless/st/cw1200/queue.c
303
++queue->num_queued;
drivers/net/wireless/st/cw1200/queue.c
304
++queue->link_map_cache[txpriv->link_id];
drivers/net/wireless/st/cw1200/queue.c
314
if (queue->overfull == false &&
drivers/net/wireless/st/cw1200/queue.c
315
queue->num_queued >=
drivers/net/wireless/st/cw1200/queue.c
316
(queue->capacity - (num_present_cpus() - 1))) {
drivers/net/wireless/st/cw1200/queue.c
317
queue->overfull = true;
drivers/net/wireless/st/cw1200/queue.c
318
__cw1200_queue_lock(queue);
drivers/net/wireless/st/cw1200/queue.c
319
mod_timer(&queue->gc, jiffies);
drivers/net/wireless/st/cw1200/queue.c
32
queue->queue_id);
drivers/net/wireless/st/cw1200/queue.c
324
spin_unlock_bh(&queue->lock);
drivers/net/wireless/st/cw1200/queue.c
328
int cw1200_queue_get(struct cw1200_queue *queue,
drivers/net/wireless/st/cw1200/queue.c
33
ieee80211_stop_queue(stats->priv->hw, queue->queue_id);
drivers/net/wireless/st/cw1200/queue.c
336
struct cw1200_queue_stats *stats = queue->stats;
drivers/net/wireless/st/cw1200/queue.c
339
spin_lock_bh(&queue->lock);
drivers/net/wireless/st/cw1200/queue.c
340
list_for_each_entry(item, &queue->queue, head) {
drivers/net/wireless/st/cw1200/queue.c
352
list_move_tail(&item->head, &queue->pending);
drivers/net/wireless/st/cw1200/queue.c
353
++queue->num_pending;
drivers/net/wireless/st/cw1200/queue.c
354
--queue->link_map_cache[item->txpriv.link_id];
drivers/net/wireless/st/cw1200/queue.c
363
spin_unlock_bh(&queue->lock);
drivers/net/wireless/st/cw1200/queue.c
369
int cw1200_queue_requeue(struct cw1200_queue *queue, u32 packet_id)
drivers/net/wireless/st/cw1200/queue.c
37
static inline void __cw1200_queue_unlock(struct cw1200_queue *queue)
drivers/net/wireless/st/cw1200/queue.c
374
struct cw1200_queue_stats *stats = queue->stats;
drivers/net/wireless/st/cw1200/queue.c
379
item = &queue->pool[item_id];
drivers/net/wireless/st/cw1200/queue.c
381
spin_lock_bh(&queue->lock);
drivers/net/wireless/st/cw1200/queue.c
382
BUG_ON(queue_id != queue->queue_id);
drivers/net/wireless/st/cw1200/queue.c
383
if (queue_generation != queue->generation) {
drivers/net/wireless/st/cw1200/queue.c
385
} else if (item_id >= (unsigned) queue->capacity) {
drivers/net/wireless/st/cw1200/queue.c
39
struct cw1200_queue_stats *stats = queue->stats;
drivers/net/wireless/st/cw1200/queue.c
392
--queue->num_pending;
drivers/net/wireless/st/cw1200/queue.c
393
++queue->link_map_cache[item->txpriv.link_id];
drivers/net/wireless/st/cw1200/queue.c
40
BUG_ON(!queue->tx_locked_cnt);
drivers/net/wireless/st/cw1200/queue.c
405
list_move(&item->head, &queue->queue);
drivers/net/wireless/st/cw1200/queue.c
407
spin_unlock_bh(&queue->lock);
drivers/net/wireless/st/cw1200/queue.c
41
if (--queue->tx_locked_cnt == 0) {
drivers/net/wireless/st/cw1200/queue.c
411
int cw1200_queue_remove(struct cw1200_queue *queue, u32 packet_id)
drivers/net/wireless/st/cw1200/queue.c
416
struct cw1200_queue_stats *stats = queue->stats;
drivers/net/wireless/st/cw1200/queue.c
423
item = &queue->pool[item_id];
drivers/net/wireless/st/cw1200/queue.c
425
spin_lock_bh(&queue->lock);
drivers/net/wireless/st/cw1200/queue.c
426
BUG_ON(queue_id != queue->queue_id);
drivers/net/wireless/st/cw1200/queue.c
427
if (queue_generation != queue->generation) {
drivers/net/wireless/st/cw1200/queue.c
429
} else if (item_id >= (unsigned) queue->capacity) {
drivers/net/wireless/st/cw1200/queue.c
43
queue->queue_id);
drivers/net/wireless/st/cw1200/queue.c
439
--queue->num_pending;
drivers/net/wireless/st/cw1200/queue.c
44
ieee80211_wake_queue(stats->priv->hw, queue->queue_id);
drivers/net/wireless/st/cw1200/queue.c
440
--queue->num_queued;
drivers/net/wireless/st/cw1200/queue.c
441
++queue->num_sent;
drivers/net/wireless/st/cw1200/queue.c
446
list_move(&item->head, &queue->free_pool);
drivers/net/wireless/st/cw1200/queue.c
448
if (queue->overfull &&
drivers/net/wireless/st/cw1200/queue.c
449
(queue->num_queued <= (queue->capacity >> 1))) {
drivers/net/wireless/st/cw1200/queue.c
450
queue->overfull = false;
drivers/net/wireless/st/cw1200/queue.c
451
__cw1200_queue_unlock(queue);
drivers/net/wireless/st/cw1200/queue.c
454
spin_unlock_bh(&queue->lock);
drivers/net/wireless/st/cw1200/queue.c
462
int cw1200_queue_get_skb(struct cw1200_queue *queue, u32 packet_id,
drivers/net/wireless/st/cw1200/queue.c
472
item = &queue->pool[item_id];
drivers/net/wireless/st/cw1200/queue.c
474
spin_lock_bh(&queue->lock);
drivers/net/wireless/st/cw1200/queue.c
475
BUG_ON(queue_id != queue->queue_id);
drivers/net/wireless/st/cw1200/queue.c
476
if (queue_generation != queue->generation) {
drivers/net/wireless/st/cw1200/queue.c
478
} else if (item_id >= (unsigned) queue->capacity) {
drivers/net/wireless/st/cw1200/queue.c
488
spin_unlock_bh(&queue->lock);
drivers/net/wireless/st/cw1200/queue.c
492
void cw1200_queue_lock(struct cw1200_queue *queue)
drivers/net/wireless/st/cw1200/queue.c
494
spin_lock_bh(&queue->lock);
drivers/net/wireless/st/cw1200/queue.c
495
__cw1200_queue_lock(queue);
drivers/net/wireless/st/cw1200/queue.c
496
spin_unlock_bh(&queue->lock);
drivers/net/wireless/st/cw1200/queue.c
499
void cw1200_queue_unlock(struct cw1200_queue *queue)
drivers/net/wireless/st/cw1200/queue.c
501
spin_lock_bh(&queue->lock);
drivers/net/wireless/st/cw1200/queue.c
502
__cw1200_queue_unlock(queue);
drivers/net/wireless/st/cw1200/queue.c
503
spin_unlock_bh(&queue->lock);
drivers/net/wireless/st/cw1200/queue.c
506
bool cw1200_queue_get_xmit_timestamp(struct cw1200_queue *queue,
drivers/net/wireless/st/cw1200/queue.c
513
spin_lock_bh(&queue->lock);
drivers/net/wireless/st/cw1200/queue.c
514
ret = !list_empty(&queue->pending);
drivers/net/wireless/st/cw1200/queue.c
516
list_for_each_entry(item, &queue->pending, head) {
drivers/net/wireless/st/cw1200/queue.c
523
spin_unlock_bh(&queue->lock);
drivers/net/wireless/st/cw1200/queue.c
89
static void __cw1200_queue_gc(struct cw1200_queue *queue,
drivers/net/wireless/st/cw1200/queue.c
93
struct cw1200_queue_stats *stats = queue->stats;
drivers/net/wireless/st/cw1200/queue.c
97
list_for_each_entry_safe(iter, tmp, &queue->queue, head) {
drivers/net/wireless/st/cw1200/queue.c
98
if (time_is_after_jiffies(iter->queue_timestamp + queue->ttl)) {
drivers/net/wireless/st/cw1200/queue.h
33
struct list_head queue;
drivers/net/wireless/st/cw1200/queue.h
68
int cw1200_queue_init(struct cw1200_queue *queue,
drivers/net/wireless/st/cw1200/queue.h
73
int cw1200_queue_clear(struct cw1200_queue *queue);
drivers/net/wireless/st/cw1200/queue.h
75
void cw1200_queue_deinit(struct cw1200_queue *queue);
drivers/net/wireless/st/cw1200/queue.h
77
size_t cw1200_queue_get_num_queued(struct cw1200_queue *queue,
drivers/net/wireless/st/cw1200/queue.h
79
int cw1200_queue_put(struct cw1200_queue *queue,
drivers/net/wireless/st/cw1200/queue.h
82
int cw1200_queue_get(struct cw1200_queue *queue,
drivers/net/wireless/st/cw1200/queue.h
87
int cw1200_queue_requeue(struct cw1200_queue *queue, u32 packet_id);
drivers/net/wireless/st/cw1200/queue.h
88
int cw1200_queue_remove(struct cw1200_queue *queue,
drivers/net/wireless/st/cw1200/queue.h
90
int cw1200_queue_get_skb(struct cw1200_queue *queue, u32 packet_id,
drivers/net/wireless/st/cw1200/queue.h
93
void cw1200_queue_lock(struct cw1200_queue *queue);
drivers/net/wireless/st/cw1200/queue.h
94
void cw1200_queue_unlock(struct cw1200_queue *queue);
drivers/net/wireless/st/cw1200/queue.h
95
bool cw1200_queue_get_xmit_timestamp(struct cw1200_queue *queue,
drivers/net/wireless/st/cw1200/scan.c
355
struct cw1200_queue *queue = &priv->tx_queue[queue_id];
drivers/net/wireless/st/cw1200/scan.c
393
if (cw1200_queue_get_skb(queue, priv->pending_frame_id,
drivers/net/wireless/st/cw1200/scan.c
451
BUG_ON(cw1200_queue_remove(queue, priv->pending_frame_id));
drivers/net/wireless/st/cw1200/sta.c
609
unsigned int link_id, u16 queue,
drivers/net/wireless/st/cw1200/sta.c
619
if (queue < dev->queues) {
drivers/net/wireless/st/cw1200/sta.c
622
WSM_TX_QUEUE_SET(&priv->tx_queue_params, queue, 0, 0, 0);
drivers/net/wireless/st/cw1200/sta.c
624
&priv->tx_queue_params.params[queue], queue);
drivers/net/wireless/st/cw1200/sta.c
630
WSM_EDCA_SET(&priv->edca, queue, params->aifs,
drivers/net/wireless/st/cw1200/sta.c
847
struct cw1200_queue *queue = &priv->tx_queue[queue_id];
drivers/net/wireless/st/cw1200/sta.c
856
cw1200_queue_requeue(queue, priv->pending_frame_id);
drivers/net/wireless/st/cw1200/sta.h
31
unsigned int link_id, u16 queue,
drivers/net/wireless/st/cw1200/txrx.c
402
unsigned queue;
drivers/net/wireless/st/cw1200/txrx.c
469
if (t->sta && (t->sta->uapsd_queues & BIT(t->queue)))
drivers/net/wireless/st/cw1200/txrx.c
579
wsm->queue_id = wsm_queue_id_to_wsm(t->queue);
drivers/net/wireless/st/cw1200/txrx.c
710
.queue = skb_get_queue_mapping(skb),
drivers/net/wireless/st/cw1200/txrx.c
732
if (WARN_ON(t.queue >= 4))
drivers/net/wireless/st/cw1200/txrx.c
740
skb->len, t.queue, t.txpriv.link_id,
drivers/net/wireless/st/cw1200/txrx.c
770
BUG_ON(cw1200_queue_put(&priv->tx_queue[t.queue],
drivers/net/wireless/st/cw1200/txrx.c
852
struct cw1200_queue *queue = &priv->tx_queue[queue_id];
drivers/net/wireless/st/cw1200/txrx.c
883
cw1200_queue_requeue(queue, arg->packet_id);
drivers/net/wireless/st/cw1200/txrx.c
893
} else if (!cw1200_queue_get_skb(queue, arg->packet_id,
drivers/net/wireless/st/cw1200/txrx.c
953
cw1200_queue_remove(queue, arg->packet_id);
drivers/net/wireless/st/cw1200/wsm.c
1058
arg.queue = (flags >> 1) & 3;
drivers/net/wireless/st/cw1200/wsm.c
1457
struct cw1200_queue *queue)
drivers/net/wireless/st/cw1200/wsm.c
1554
BUG_ON(cw1200_queue_remove(queue, wsm->packet_id));
drivers/net/wireless/st/cw1200/wsm.c
1663
struct cw1200_queue *queue = NULL;
drivers/net/wireless/st/cw1200/wsm.c
1689
ret = wsm_get_tx_queue_and_mask(priv, &queue,
drivers/net/wireless/st/cw1200/wsm.c
1691
queue_num = queue - priv->tx_queue;
drivers/net/wireless/st/cw1200/wsm.c
1709
if (cw1200_queue_get(queue,
drivers/net/wireless/st/cw1200/wsm.c
1715
tx_info, txpriv, queue))
drivers/net/wireless/st/cw1200/wsm.c
1730
(int)cw1200_queue_get_num_queued(queue, tx_allowed_mask) + 1);
drivers/net/wireless/st/cw1200/wsm.h
1094
#define WSM_TX_QUEUE_SET(queue_params, queue, ack_policy, allowed_time,\
drivers/net/wireless/st/cw1200/wsm.h
1097
struct wsm_set_tx_queue_params *p = &(queue_params)->params[queue]; \
drivers/net/wireless/st/cw1200/wsm.h
1247
/* [out] */ int queue;
drivers/net/wireless/ti/wl1251/acx.c
981
int wl1251_acx_tid_cfg(struct wl1251 *wl, u8 queue,
drivers/net/wireless/ti/wl1251/acx.c
990
"ps_scheme %d ack_policy %d", queue, type, tsid,
drivers/net/wireless/ti/wl1251/acx.c
997
acx->queue = queue;
drivers/net/wireless/ti/wl1251/acx.h
1296
u8 queue;
drivers/net/wireless/ti/wl1251/acx.h
1479
int wl1251_acx_tid_cfg(struct wl1251 *wl, u8 queue,
drivers/net/wireless/ti/wl1251/main.c
1287
unsigned int link_id, u16 queue,
drivers/net/wireless/ti/wl1251/main.c
1296
wl1251_debug(DEBUG_MAC80211, "mac80211 conf tx %d", queue);
drivers/net/wireless/ti/wl1251/main.c
1303
ret = wl1251_acx_ac_cfg(wl, wl1251_tx_get_queue(queue),
drivers/net/wireless/ti/wl1251/main.c
1314
ret = wl1251_acx_tid_cfg(wl, wl1251_tx_get_queue(queue),
drivers/net/wireless/ti/wl1251/main.c
1316
wl1251_tx_get_queue(queue), ps_scheme,
drivers/net/wireless/ti/wl1251/tx.h
197
static inline int wl1251_tx_get_queue(int queue)
drivers/net/wireless/ti/wl1251/tx.h
199
switch (queue) {
drivers/net/wireless/ti/wlcore/main.c
4936
unsigned int link_id, u16 queue,
drivers/net/wireless/ti/wlcore/main.c
4949
wl1271_debug(DEBUG_MAC80211, "mac80211 conf tx %d", queue);
drivers/net/wireless/ti/wlcore/main.c
4967
ret = wl1271_acx_ac_cfg(wl, wlvif, wl1271_tx_get_queue(queue),
drivers/net/wireless/ti/wlcore/main.c
4973
ret = wl1271_acx_tid_cfg(wl, wlvif, wl1271_tx_get_queue(queue),
drivers/net/wireless/ti/wlcore/main.c
4975
wl1271_tx_get_queue(queue),
drivers/net/wireless/ti/wlcore/tx.c
1198
u8 queue, enum wlcore_queue_stop_reason reason)
drivers/net/wireless/ti/wlcore/tx.c
1200
int hwq = wlcore_tx_get_mac80211_queue(wlvif, queue);
drivers/net/wireless/ti/wlcore/tx.c
1212
void wlcore_stop_queue(struct wl1271 *wl, struct wl12xx_vif *wlvif, u8 queue,
drivers/net/wireless/ti/wlcore/tx.c
1218
wlcore_stop_queue_locked(wl, wlvif, queue, reason);
drivers/net/wireless/ti/wlcore/tx.c
1222
void wlcore_wake_queue(struct wl1271 *wl, struct wl12xx_vif *wlvif, u8 queue,
drivers/net/wireless/ti/wlcore/tx.c
1226
int hwq = wlcore_tx_get_mac80211_queue(wlvif, queue);
drivers/net/wireless/ti/wlcore/tx.c
1285
struct wl12xx_vif *wlvif, u8 queue,
drivers/net/wireless/ti/wlcore/tx.c
1292
stopped = wlcore_is_queue_stopped_by_reason_locked(wl, wlvif, queue,
drivers/net/wireless/ti/wlcore/tx.c
1300
struct wl12xx_vif *wlvif, u8 queue,
drivers/net/wireless/ti/wlcore/tx.c
1303
int hwq = wlcore_tx_get_mac80211_queue(wlvif, queue);
drivers/net/wireless/ti/wlcore/tx.c
1310
u8 queue)
drivers/net/wireless/ti/wlcore/tx.c
1312
int hwq = wlcore_tx_get_mac80211_queue(wlvif, queue);
drivers/net/wireless/ti/wlcore/tx.h
184
static inline int wl1271_tx_get_queue(int queue)
drivers/net/wireless/ti/wlcore/tx.h
186
switch (queue) {
drivers/net/wireless/ti/wlcore/tx.h
201
int wlcore_tx_get_mac80211_queue(struct wl12xx_vif *wlvif, int queue)
drivers/net/wireless/ti/wlcore/tx.h
205
switch (queue) {
drivers/net/wireless/ti/wlcore/tx.h
249
u8 queue, enum wlcore_queue_stop_reason reason);
drivers/net/wireless/ti/wlcore/tx.h
250
void wlcore_stop_queue(struct wl1271 *wl, struct wl12xx_vif *wlvif, u8 queue,
drivers/net/wireless/ti/wlcore/tx.h
252
void wlcore_wake_queue(struct wl1271 *wl, struct wl12xx_vif *wlvif, u8 queue,
drivers/net/wireless/ti/wlcore/tx.h
259
struct wl12xx_vif *wlvif, u8 queue,
drivers/net/wireless/ti/wlcore/tx.h
264
u8 queue,
drivers/net/wireless/ti/wlcore/tx.h
267
u8 queue);
drivers/net/wireless/virtual/mac80211_hwsim.c
2788
unsigned int link_id, u16 queue,
drivers/net/wireless/virtual/mac80211_hwsim.c
2793
__func__, queue,
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
1098
static int t7xx_cldma_default_recv_skb(struct cldma_queue *queue, struct sk_buff *skb)
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
110
static int t7xx_cldma_gpd_rx_from_q(struct cldma_queue *queue, int budget, bool *over_budget)
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
112
struct cldma_ctrl *md_ctrl = queue->md_ctrl;
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
127
req = queue->tr_done;
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
142
queue->index * sizeof(u64));
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
155
queue->tr_ring->pkt_size, DMA_FROM_DEVICE);
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
163
ret = queue->recv_skb(queue, skb);
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
171
spin_lock_irqsave(&queue->ring_lock, flags);
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
172
queue->tr_done = list_next_entry_circular(req, &queue->tr_ring->gpd_ring, entry);
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
173
spin_unlock_irqrestore(&queue->ring_lock, flags);
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
174
req = queue->rx_refill;
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
176
ret = t7xx_cldma_alloc_and_map_skb(md_ctrl, req, queue->tr_ring->pkt_size, GFP_KERNEL);
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
185
spin_lock_irqsave(&queue->ring_lock, flags);
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
186
queue->rx_refill = list_next_entry_circular(req, &queue->tr_ring->gpd_ring, entry);
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
187
spin_unlock_irqrestore(&queue->ring_lock, flags);
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
196
static int t7xx_cldma_gpd_rx_collect(struct cldma_queue *queue, int budget)
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
198
struct cldma_ctrl *md_ctrl = queue->md_ctrl;
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
208
ret = t7xx_cldma_gpd_rx_from_q(queue, budget, &over_budget);
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
217
if (md_ctrl->rxq_active & BIT(queue->index)) {
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
218
if (!t7xx_cldma_hw_queue_status(hw_info, queue->index, MTK_RX))
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
219
t7xx_cldma_hw_resume_queue(hw_info, queue->index, MTK_RX);
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
221
pending_rx_int = t7xx_cldma_hw_int_status(hw_info, BIT(queue->index),
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
240
struct cldma_queue *queue = container_of(work, struct cldma_queue, cldma_work);
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
241
struct cldma_ctrl *md_ctrl = queue->md_ctrl;
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
244
value = t7xx_cldma_gpd_rx_collect(queue, queue->budget);
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
245
if (value && md_ctrl->rxq_active & BIT(queue->index)) {
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
246
queue_work(queue->worker, &queue->cldma_work);
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
251
t7xx_cldma_hw_irq_en_txrx(&md_ctrl->hw_info, queue->index, MTK_RX);
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
252
t7xx_cldma_hw_irq_en_eq(&md_ctrl->hw_info, queue->index, MTK_RX);
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
256
static int t7xx_cldma_gpd_tx_collect(struct cldma_queue *queue)
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
258
struct cldma_ctrl *md_ctrl = queue->md_ctrl;
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
267
spin_lock_irqsave(&queue->ring_lock, flags);
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
268
req = queue->tr_done;
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
270
spin_unlock_irqrestore(&queue->ring_lock, flags);
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
275
spin_unlock_irqrestore(&queue->ring_lock, flags);
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
278
queue->budget++;
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
283
queue->tr_done = list_next_entry_circular(req, &queue->tr_ring->gpd_ring, entry);
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
284
spin_unlock_irqrestore(&queue->ring_lock, flags);
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
292
wake_up_nr(&queue->req_wq, count);
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
297
static void t7xx_cldma_txq_empty_hndl(struct cldma_queue *queue)
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
299
struct cldma_ctrl *md_ctrl = queue->md_ctrl;
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
305
if (!(md_ctrl->txq_active & BIT(queue->index)))
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
308
spin_lock_irqsave(&queue->ring_lock, flags);
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
309
req = list_prev_entry_circular(queue->tx_next, &queue->tr_ring->gpd_ring, entry);
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
310
spin_unlock_irqrestore(&queue->ring_lock, flags);
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
320
queue->index * sizeof(u64));
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
324
md_ctrl->hif_id, queue->index);
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
328
t7xx_cldma_hw_resume_queue(hw_info, queue->index, MTK_TX);
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
335
struct cldma_queue *queue = container_of(work, struct cldma_queue, cldma_work);
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
336
struct cldma_ctrl *md_ctrl = queue->md_ctrl;
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
342
t7xx_cldma_gpd_tx_collect(queue);
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
343
l2_tx_int = t7xx_cldma_hw_int_status(hw_info, BIT(queue->index) | EQ_STA_BIT(queue->index),
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
345
if (l2_tx_int & EQ_STA_BIT(queue->index)) {
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
346
t7xx_cldma_hw_tx_done(hw_info, EQ_STA_BIT(queue->index));
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
347
t7xx_cldma_txq_empty_hndl(queue);
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
350
if (l2_tx_int & BIT(queue->index)) {
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
351
t7xx_cldma_hw_tx_done(hw_info, BIT(queue->index));
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
352
queue_work(queue->worker, &queue->cldma_work);
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
357
if (md_ctrl->txq_active & BIT(queue->index)) {
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
359
t7xx_cldma_hw_irq_en_eq(hw_info, queue->index, MTK_TX);
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
360
t7xx_cldma_hw_irq_en_txrx(hw_info, queue->index, MTK_TX);
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
504
static void t7xx_cldma_q_reset(struct cldma_queue *queue)
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
508
req = list_first_entry(&queue->tr_ring->gpd_ring, struct cldma_request, entry);
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
509
queue->tr_done = req;
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
510
queue->budget = queue->tr_ring->length;
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
512
if (queue->dir == MTK_TX)
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
513
queue->tx_next = req;
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
515
queue->rx_refill = req;
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
518
static void t7xx_cldma_rxq_init(struct cldma_queue *queue)
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
520
struct cldma_ctrl *md_ctrl = queue->md_ctrl;
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
522
queue->dir = MTK_RX;
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
523
queue->tr_ring = &md_ctrl->rx_ring[queue->index];
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
524
t7xx_cldma_q_reset(queue);
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
527
static void t7xx_cldma_txq_init(struct cldma_queue *queue)
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
529
struct cldma_ctrl *md_ctrl = queue->md_ctrl;
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
531
queue->dir = MTK_TX;
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
532
queue->tr_ring = &md_ctrl->tx_ring[queue->index];
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
533
t7xx_cldma_q_reset(queue);
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
60
static void md_cd_queue_struct_reset(struct cldma_queue *queue, struct cldma_ctrl *md_ctrl,
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
63
queue->dir = tx_rx;
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
64
queue->index = index;
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
65
queue->md_ctrl = md_ctrl;
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
66
queue->tr_ring = NULL;
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
67
queue->tr_done = NULL;
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
68
queue->tx_next = NULL;
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
71
static void md_cd_queue_struct_init(struct cldma_queue *queue, struct cldma_ctrl *md_ctrl,
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
74
md_cd_queue_struct_reset(queue, md_ctrl, tx_rx, index);
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
75
init_waitqueue_head(&queue->req_wq);
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
76
spin_lock_init(&queue->ring_lock);
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
841
static int t7xx_cldma_gpd_handle_tx_request(struct cldma_queue *queue, struct cldma_request *tx_req,
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
844
struct cldma_ctrl *md_ctrl = queue->md_ctrl;
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
863
if (md_ctrl->txq_active & BIT(queue->index))
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
900
static void t7xx_cldma_set_recv_skb(struct cldma_queue *queue,
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
901
int (*recv_skb)(struct cldma_queue *queue, struct sk_buff *skb))
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
903
queue->recv_skb = recv_skb;
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
922
struct cldma_queue *queue;
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
934
queue = &md_ctrl->txq[qno];
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
945
spin_lock_irqsave(&queue->ring_lock, flags);
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
946
tx_req = queue->tx_next;
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
947
if (queue->budget > 0 && !tx_req->skb) {
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
948
struct list_head *gpd_ring = &queue->tr_ring->gpd_ring;
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
950
queue->budget--;
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
951
t7xx_cldma_gpd_handle_tx_request(queue, tx_req, skb);
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
952
queue->tx_next = list_next_entry_circular(tx_req, gpd_ring, entry);
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
953
spin_unlock_irqrestore(&queue->ring_lock, flags);
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
970
spin_unlock_irqrestore(&queue->ring_lock, flags);
drivers/net/wwan/t7xx/t7xx_hif_cldma.c
983
ret = wait_event_interruptible_exclusive(queue->req_wq, queue->budget > 0);
drivers/net/wwan/t7xx/t7xx_hif_cldma.h
94
int (*recv_skb)(struct cldma_queue *queue, struct sk_buff *skb);
drivers/net/wwan/t7xx/t7xx_hif_dpmaif_rx.c
1049
int t7xx_dpmaif_rxq_init(struct dpmaif_rx_queue *queue)
drivers/net/wwan/t7xx/t7xx_hif_dpmaif_rx.c
1053
ret = t7xx_dpmaif_rx_alloc(queue);
drivers/net/wwan/t7xx/t7xx_hif_dpmaif_rx.c
1055
dev_err(queue->dpmaif_ctrl->dev, "Failed to allocate RX buffers: %d\n", ret);
drivers/net/wwan/t7xx/t7xx_hif_dpmaif_rx.c
1060
void t7xx_dpmaif_rxq_free(struct dpmaif_rx_queue *queue)
drivers/net/wwan/t7xx/t7xx_hif_dpmaif_rx.c
1062
t7xx_dpmaif_rx_buf_free(queue);
drivers/net/wwan/t7xx/t7xx_hif_dpmaif_rx.h
109
void t7xx_dpmaif_rxq_free(struct dpmaif_rx_queue *queue);
drivers/net/wwan/t7xx/t7xx_hif_dpmaif_rx.h
98
int t7xx_dpmaif_rxq_init(struct dpmaif_rx_queue *queue);
drivers/net/wwan/t7xx/t7xx_port_proxy.c
390
int t7xx_port_proxy_recv_skb_from_dedicated_queue(struct cldma_queue *queue, struct sk_buff *skb)
drivers/net/wwan/t7xx/t7xx_port_proxy.c
392
struct t7xx_pci_dev *t7xx_dev = queue->md_ctrl->t7xx_dev;
drivers/net/wwan/t7xx/t7xx_port_proxy.c
399
if (WARN_ON_ONCE(port->port_conf->rxq_index != queue->index)) {
drivers/net/wwan/t7xx/t7xx_port_proxy.c
415
struct cldma_queue *queue, u16 channel)
drivers/net/wwan/t7xx/t7xx_port_proxy.c
427
if (queue->md_ctrl->hif_id == port_conf->path_id &&
drivers/net/wwan/t7xx/t7xx_port_proxy.c
444
int t7xx_port_proxy_recv_skb(struct cldma_queue *queue, struct sk_buff *skb)
drivers/net/wwan/t7xx/t7xx_port_proxy.c
447
struct t7xx_pci_dev *t7xx_dev = queue->md_ctrl->t7xx_dev;
drivers/net/wwan/t7xx/t7xx_port_proxy.c
449
struct device *dev = queue->md_ctrl->dev;
drivers/net/wwan/t7xx/t7xx_port_proxy.c
461
port = t7xx_port_proxy_find_port(t7xx_dev, queue, channel);
drivers/net/wwan/t7xx/t7xx_port_proxy.h
110
int t7xx_port_proxy_recv_skb(struct cldma_queue *queue, struct sk_buff *skb);
drivers/net/wwan/t7xx/t7xx_port_proxy.h
111
int t7xx_port_proxy_recv_skb_from_dedicated_queue(struct cldma_queue *queue, struct sk_buff *skb);
drivers/net/xen-netback/common.h
350
int xenvif_init_queue(struct xenvif_queue *queue);
drivers/net/xen-netback/common.h
351
void xenvif_deinit_queue(struct xenvif_queue *queue);
drivers/net/xen-netback/common.h
353
int xenvif_connect_data(struct xenvif_queue *queue,
drivers/net/xen-netback/common.h
368
void xenvif_unmap_frontend_data_rings(struct xenvif_queue *queue);
drivers/net/xen-netback/common.h
369
int xenvif_map_frontend_data_rings(struct xenvif_queue *queue,
drivers/net/xen-netback/common.h
374
void xenvif_napi_schedule_or_enable_events(struct xenvif_queue *queue);
drivers/net/xen-netback/common.h
379
int xenvif_tx_action(struct xenvif_queue *queue, int budget);
drivers/net/xen-netback/common.h
382
void xenvif_kick_thread(struct xenvif_queue *queue);
drivers/net/xen-netback/common.h
388
bool xenvif_have_rx_work(struct xenvif_queue *queue, bool test_kthread);
drivers/net/xen-netback/common.h
389
bool xenvif_rx_queue_tail(struct xenvif_queue *queue, struct sk_buff *skb);
drivers/net/xen-netback/common.h
396
static inline pending_ring_idx_t nr_pending_reqs(struct xenvif_queue *queue)
drivers/net/xen-netback/common.h
399
queue->pending_prod + queue->pending_cons;
drivers/net/xen-netback/common.h
416
void xenvif_skb_zerocopy_prepare(struct xenvif_queue *queue,
drivers/net/xen-netback/common.h
418
void xenvif_skb_zerocopy_complete(struct xenvif_queue *queue);
drivers/net/xen-netback/interface.c
105
struct xenvif_queue *queue =
drivers/net/xen-netback/interface.c
113
if (unlikely(queue->vif->disabled)) {
drivers/net/xen-netback/interface.c
118
work_done = xenvif_tx_action(queue, budget);
drivers/net/xen-netback/interface.c
125
if (likely(!queue->rate_limited))
drivers/net/xen-netback/interface.c
126
xenvif_napi_schedule_or_enable_events(queue);
drivers/net/xen-netback/interface.c
132
static bool xenvif_handle_rx_interrupt(struct xenvif_queue *queue)
drivers/net/xen-netback/interface.c
136
rc = xenvif_have_rx_work(queue, false);
drivers/net/xen-netback/interface.c
138
xenvif_kick_thread(queue);
drivers/net/xen-netback/interface.c
144
struct xenvif_queue *queue = dev_id;
drivers/net/xen-netback/interface.c
147
old = atomic_fetch_or(NETBK_RX_EOI, &queue->eoi_pending);
drivers/net/xen-netback/interface.c
150
if (!xenvif_handle_rx_interrupt(queue)) {
drivers/net/xen-netback/interface.c
151
atomic_andnot(NETBK_RX_EOI, &queue->eoi_pending);
drivers/net/xen-netback/interface.c
160
struct xenvif_queue *queue = dev_id;
drivers/net/xen-netback/interface.c
164
old = atomic_fetch_or(NETBK_COMMON_EOI, &queue->eoi_pending);
drivers/net/xen-netback/interface.c
167
has_tx = xenvif_handle_tx_interrupt(queue);
drivers/net/xen-netback/interface.c
168
has_rx = xenvif_handle_rx_interrupt(queue);
drivers/net/xen-netback/interface.c
171
atomic_andnot(NETBK_COMMON_EOI, &queue->eoi_pending);
drivers/net/xen-netback/interface.c
208
struct xenvif_queue *queue = NULL;
drivers/net/xen-netback/interface.c
230
queue = &vif->queues[index];
drivers/net/xen-netback/interface.c
233
if (queue->task == NULL ||
drivers/net/xen-netback/interface.c
234
queue->dealloc_task == NULL ||
drivers/net/xen-netback/interface.c
258
if (!xenvif_rx_queue_tail(queue, skb))
drivers/net/xen-netback/interface.c
261
xenvif_kick_thread(queue);
drivers/net/xen-netback/interface.c
274
struct xenvif_queue *queue = NULL;
drivers/net/xen-netback/interface.c
287
queue = &vif->queues[index];
drivers/net/xen-netback/interface.c
288
rx_bytes += queue->stats.rx_bytes;
drivers/net/xen-netback/interface.c
289
rx_packets += queue->stats.rx_packets;
drivers/net/xen-netback/interface.c
290
tx_bytes += queue->stats.tx_bytes;
drivers/net/xen-netback/interface.c
291
tx_packets += queue->stats.tx_packets;
drivers/net/xen-netback/interface.c
306
struct xenvif_queue *queue = NULL;
drivers/net/xen-netback/interface.c
311
queue = &vif->queues[queue_index];
drivers/net/xen-netback/interface.c
312
napi_enable(&queue->napi);
drivers/net/xen-netback/interface.c
313
enable_irq(queue->tx_irq);
drivers/net/xen-netback/interface.c
314
if (queue->tx_irq != queue->rx_irq)
drivers/net/xen-netback/interface.c
315
enable_irq(queue->rx_irq);
drivers/net/xen-netback/interface.c
316
xenvif_napi_schedule_or_enable_events(queue);
drivers/net/xen-netback/interface.c
322
struct xenvif_queue *queue = NULL;
drivers/net/xen-netback/interface.c
327
queue = &vif->queues[queue_index];
drivers/net/xen-netback/interface.c
328
disable_irq(queue->tx_irq);
drivers/net/xen-netback/interface.c
329
if (queue->tx_irq != queue->rx_irq)
drivers/net/xen-netback/interface.c
330
disable_irq(queue->rx_irq);
drivers/net/xen-netback/interface.c
331
napi_disable(&queue->napi);
drivers/net/xen-netback/interface.c
332
timer_delete_sync(&queue->credit_timeout);
drivers/net/xen-netback/interface.c
52
void xenvif_skb_zerocopy_prepare(struct xenvif_queue *queue,
drivers/net/xen-netback/interface.c
56
atomic_inc(&queue->inflight_packets);
drivers/net/xen-netback/interface.c
561
int xenvif_init_queue(struct xenvif_queue *queue)
drivers/net/xen-netback/interface.c
565
queue->credit_bytes = queue->remaining_credit = ~0UL;
drivers/net/xen-netback/interface.c
566
queue->credit_usec = 0UL;
drivers/net/xen-netback/interface.c
567
timer_setup(&queue->credit_timeout, xenvif_tx_credit_callback, 0);
drivers/net/xen-netback/interface.c
568
queue->credit_window_start = get_jiffies_64();
drivers/net/xen-netback/interface.c
570
queue->rx_queue_max = XENVIF_RX_QUEUE_BYTES;
drivers/net/xen-netback/interface.c
572
skb_queue_head_init(&queue->rx_queue);
drivers/net/xen-netback/interface.c
573
skb_queue_head_init(&queue->tx_queue);
drivers/net/xen-netback/interface.c
575
queue->pending_cons = 0;
drivers/net/xen-netback/interface.c
576
queue->pending_prod = MAX_PENDING_REQS;
drivers/net/xen-netback/interface.c
578
queue->pending_ring[i] = i;
drivers/net/xen-netback/interface.c
580
spin_lock_init(&queue->callback_lock);
drivers/net/xen-netback/interface.c
581
spin_lock_init(&queue->response_lock);
drivers/net/xen-netback/interface.c
588
queue->mmap_pages);
drivers/net/xen-netback/interface.c
59
void xenvif_skb_zerocopy_complete(struct xenvif_queue *queue)
drivers/net/xen-netback/interface.c
590
netdev_err(queue->vif->dev, "Could not reserve mmap_pages\n");
drivers/net/xen-netback/interface.c
595
queue->pending_tx_info[i].callback_struct = (struct ubuf_info_msgzc)
drivers/net/xen-netback/interface.c
599
queue->grant_tx_handle[i] = NETBACK_INVALID_HANDLE;
drivers/net/xen-netback/interface.c
61
atomic_dec(&queue->inflight_packets);
drivers/net/xen-netback/interface.c
67
wake_up(&queue->dealloc_wq);
drivers/net/xen-netback/interface.c
671
static void xenvif_disconnect_queue(struct xenvif_queue *queue)
drivers/net/xen-netback/interface.c
673
if (queue->task) {
drivers/net/xen-netback/interface.c
674
kthread_stop_put(queue->task);
drivers/net/xen-netback/interface.c
675
queue->task = NULL;
drivers/net/xen-netback/interface.c
678
if (queue->dealloc_task) {
drivers/net/xen-netback/interface.c
679
kthread_stop(queue->dealloc_task);
drivers/net/xen-netback/interface.c
680
queue->dealloc_task = NULL;
drivers/net/xen-netback/interface.c
683
if (queue->napi.poll) {
drivers/net/xen-netback/interface.c
684
netif_napi_del(&queue->napi);
drivers/net/xen-netback/interface.c
685
queue->napi.poll = NULL;
drivers/net/xen-netback/interface.c
688
if (queue->tx_irq) {
drivers/net/xen-netback/interface.c
689
unbind_from_irqhandler(queue->tx_irq, queue);
drivers/net/xen-netback/interface.c
690
if (queue->tx_irq == queue->rx_irq)
drivers/net/xen-netback/interface.c
691
queue->rx_irq = 0;
drivers/net/xen-netback/interface.c
692
queue->tx_irq = 0;
drivers/net/xen-netback/interface.c
695
if (queue->rx_irq) {
drivers/net/xen-netback/interface.c
696
unbind_from_irqhandler(queue->rx_irq, queue);
drivers/net/xen-netback/interface.c
697
queue->rx_irq = 0;
drivers/net/xen-netback/interface.c
700
xenvif_unmap_frontend_data_rings(queue);
drivers/net/xen-netback/interface.c
703
int xenvif_connect_data(struct xenvif_queue *queue,
drivers/net/xen-netback/interface.c
709
struct xenbus_device *dev = xenvif_to_xenbus_device(queue->vif);
drivers/net/xen-netback/interface.c
713
BUG_ON(queue->tx_irq);
drivers/net/xen-netback/interface.c
714
BUG_ON(queue->task);
drivers/net/xen-netback/interface.c
715
BUG_ON(queue->dealloc_task);
drivers/net/xen-netback/interface.c
717
err = xenvif_map_frontend_data_rings(queue, tx_ring_ref,
drivers/net/xen-netback/interface.c
722
init_waitqueue_head(&queue->wq);
drivers/net/xen-netback/interface.c
723
init_waitqueue_head(&queue->dealloc_wq);
drivers/net/xen-netback/interface.c
724
atomic_set(&queue->inflight_packets, 0);
drivers/net/xen-netback/interface.c
726
netif_napi_add(queue->vif->dev, &queue->napi, xenvif_poll);
drivers/net/xen-netback/interface.c
728
queue->stalled = true;
drivers/net/xen-netback/interface.c
730
task = kthread_run(xenvif_kthread_guest_rx, queue,
drivers/net/xen-netback/interface.c
731
"%s-guest-rx", queue->name);
drivers/net/xen-netback/interface.c
734
queue->task = task;
drivers/net/xen-netback/interface.c
741
task = kthread_run(xenvif_dealloc_kthread, queue,
drivers/net/xen-netback/interface.c
742
"%s-dealloc", queue->name);
drivers/net/xen-netback/interface.c
745
queue->dealloc_task = task;
drivers/net/xen-netback/interface.c
751
queue->name, queue);
drivers/net/xen-netback/interface.c
754
queue->tx_irq = queue->rx_irq = err;
drivers/net/xen-netback/interface.c
755
disable_irq(queue->tx_irq);
drivers/net/xen-netback/interface.c
758
snprintf(queue->tx_irq_name, sizeof(queue->tx_irq_name),
drivers/net/xen-netback/interface.c
759
"%s-tx", queue->name);
drivers/net/xen-netback/interface.c
762
queue->tx_irq_name, queue);
drivers/net/xen-netback/interface.c
765
queue->tx_irq = err;
drivers/net/xen-netback/interface.c
766
disable_irq(queue->tx_irq);
drivers/net/xen-netback/interface.c
768
snprintf(queue->rx_irq_name, sizeof(queue->rx_irq_name),
drivers/net/xen-netback/interface.c
769
"%s-rx", queue->name);
drivers/net/xen-netback/interface.c
77
static bool xenvif_handle_tx_interrupt(struct xenvif_queue *queue)
drivers/net/xen-netback/interface.c
772
queue->rx_irq_name, queue);
drivers/net/xen-netback/interface.c
775
queue->rx_irq = err;
drivers/net/xen-netback/interface.c
776
disable_irq(queue->rx_irq);
drivers/net/xen-netback/interface.c
782
pr_warn("Could not allocate kthread for %s\n", queue->name);
drivers/net/xen-netback/interface.c
785
xenvif_disconnect_queue(queue);
drivers/net/xen-netback/interface.c
804
struct xenvif_queue *queue = NULL;
drivers/net/xen-netback/interface.c
81
rc = RING_HAS_UNCONSUMED_REQUESTS(&queue->tx);
drivers/net/xen-netback/interface.c
811
queue = &vif->queues[queue_index];
drivers/net/xen-netback/interface.c
813
xenvif_disconnect_queue(queue);
drivers/net/xen-netback/interface.c
83
napi_schedule(&queue->napi);
drivers/net/xen-netback/interface.c
838
void xenvif_deinit_queue(struct xenvif_queue *queue)
drivers/net/xen-netback/interface.c
840
gnttab_free_pages(MAX_PENDING_REQS, queue->mmap_pages);
drivers/net/xen-netback/interface.c
89
struct xenvif_queue *queue = dev_id;
drivers/net/xen-netback/interface.c
92
old = atomic_fetch_or(NETBK_TX_EOI, &queue->eoi_pending);
drivers/net/xen-netback/interface.c
95
if (!xenvif_handle_tx_interrupt(queue)) {
drivers/net/xen-netback/interface.c
96
atomic_andnot(NETBK_TX_EOI, &queue->eoi_pending);
drivers/net/xen-netback/netback.c
1001
ret = xenvif_count_requests(queue, &txreq, extra_count,
drivers/net/xen-netback/netback.c
1010
netdev_dbg(queue->vif->dev,
drivers/net/xen-netback/netback.c
1012
xenvif_tx_err(queue, &txreq, extra_count, idx);
drivers/net/xen-netback/netback.c
1018
netdev_err(queue->vif->dev, "Cross page boundary, txreq.offset: %u, size: %u\n",
drivers/net/xen-netback/netback.c
1020
xenvif_fatal_tx_err(queue->vif);
drivers/net/xen-netback/netback.c
1029
netdev_dbg(queue->vif->dev,
drivers/net/xen-netback/netback.c
1031
xenvif_tx_err(queue, &txreq, extra_count, idx);
drivers/net/xen-netback/netback.c
1049
xenvif_tx_err(queue, &txreq, extra_count, idx);
drivers/net/xen-netback/netback.c
1051
netdev_err(queue->vif->dev,
drivers/net/xen-netback/netback.c
1061
if (xenvif_set_skb_gso(queue->vif, skb, gso)) {
drivers/net/xen-netback/netback.c
107
static void xenvif_idx_release(struct xenvif_queue *queue, u16 pending_idx,
drivers/net/xen-netback/netback.c
1097
xenvif_get_requests(queue, skb, &txreq, txfrags, copy_ops,
drivers/net/xen-netback/netback.c
110
static void make_tx_response(struct xenvif_queue *queue,
drivers/net/xen-netback/netback.c
1101
__skb_queue_tail(&queue->tx_queue, skb);
drivers/net/xen-netback/netback.c
1103
queue->tx.req_cons = idx;
drivers/net/xen-netback/netback.c
1112
static int xenvif_handle_frag_list(struct xenvif_queue *queue, struct sk_buff *skb)
drivers/net/xen-netback/netback.c
1120
queue->stats.tx_zerocopy_sent += 2;
drivers/net/xen-netback/netback.c
1121
queue->stats.tx_frag_overflow++;
drivers/net/xen-netback/netback.c
1123
xenvif_fill_frags(queue, nskb);
drivers/net/xen-netback/netback.c
115
static void xenvif_idx_unmap(struct xenvif_queue *queue, u16 pending_idx);
drivers/net/xen-netback/netback.c
1160
atomic_inc(&queue->inflight_packets);
drivers/net/xen-netback/netback.c
117
static inline int tx_work_todo(struct xenvif_queue *queue);
drivers/net/xen-netback/netback.c
1172
static int xenvif_tx_submit(struct xenvif_queue *queue)
drivers/net/xen-netback/netback.c
1174
struct gnttab_map_grant_ref *gop_map = queue->tx_map_ops;
drivers/net/xen-netback/netback.c
1175
struct gnttab_copy *gop_copy = queue->tx_copy_ops;
drivers/net/xen-netback/netback.c
1179
while ((skb = __skb_dequeue(&queue->tx_queue)) != NULL) {
drivers/net/xen-netback/netback.c
1184
txp = &queue->pending_tx_info[pending_idx].req;
drivers/net/xen-netback/netback.c
1187
if (unlikely(xenvif_tx_check_gop(queue, skb, &gop_map, &gop_copy))) {
drivers/net/xen-netback/netback.c
119
static inline unsigned long idx_to_pfn(struct xenvif_queue *queue,
drivers/net/xen-netback/netback.c
1207
xenvif_fill_frags(queue, skb);
drivers/net/xen-netback/netback.c
1211
xenvif_skb_zerocopy_prepare(queue, nskb);
drivers/net/xen-netback/netback.c
1212
if (xenvif_handle_frag_list(queue, skb)) {
drivers/net/xen-netback/netback.c
1214
netdev_err(queue->vif->dev,
drivers/net/xen-netback/netback.c
1216
xenvif_skb_zerocopy_prepare(queue, skb);
drivers/net/xen-netback/netback.c
122
return page_to_pfn(queue->mmap_pages[idx]);
drivers/net/xen-netback/netback.c
1225
skb->dev = queue->vif->dev;
drivers/net/xen-netback/netback.c
1229
if (checksum_setup(queue, skb)) {
drivers/net/xen-netback/netback.c
1230
netdev_dbg(queue->vif->dev,
drivers/net/xen-netback/netback.c
1234
xenvif_skb_zerocopy_prepare(queue, skb);
drivers/net/xen-netback/netback.c
125
static inline unsigned long idx_to_kaddr(struct xenvif_queue *queue,
drivers/net/xen-netback/netback.c
1262
queue->stats.rx_bytes += skb->len;
drivers/net/xen-netback/netback.c
1263
queue->stats.rx_packets++;
drivers/net/xen-netback/netback.c
1273
xenvif_skb_zerocopy_prepare(queue, skb);
drivers/net/xen-netback/netback.c
1274
queue->stats.tx_zerocopy_sent++;
drivers/net/xen-netback/netback.c
128
return (unsigned long)pfn_to_kaddr(idx_to_pfn(queue, idx));
drivers/net/xen-netback/netback.c
1290
struct xenvif_queue *queue = ubuf_to_queue(ubuf);
drivers/net/xen-netback/netback.c
1295
spin_lock_irqsave(&queue->callback_lock, flags);
drivers/net/xen-netback/netback.c
1299
BUG_ON(queue->dealloc_prod - queue->dealloc_cons >=
drivers/net/xen-netback/netback.c
1301
index = pending_index(queue->dealloc_prod);
drivers/net/xen-netback/netback.c
1302
queue->dealloc_ring[index] = pending_idx;
drivers/net/xen-netback/netback.c
1307
queue->dealloc_prod++;
drivers/net/xen-netback/netback.c
1309
spin_unlock_irqrestore(&queue->callback_lock, flags);
drivers/net/xen-netback/netback.c
1312
queue->stats.tx_zerocopy_success++;
drivers/net/xen-netback/netback.c
1314
queue->stats.tx_zerocopy_fail++;
drivers/net/xen-netback/netback.c
1315
xenvif_skb_zerocopy_complete(queue);
drivers/net/xen-netback/netback.c
1322
static inline void xenvif_tx_dealloc_action(struct xenvif_queue *queue)
drivers/net/xen-netback/netback.c
1329
dc = queue->dealloc_cons;
drivers/net/xen-netback/netback.c
1330
gop = queue->tx_unmap_ops;
drivers/net/xen-netback/netback.c
1334
dp = queue->dealloc_prod;
drivers/net/xen-netback/netback.c
1342
BUG_ON(gop - queue->tx_unmap_ops >= MAX_PENDING_REQS);
drivers/net/xen-netback/netback.c
1344
queue->dealloc_ring[pending_index(dc++)];
drivers/net/xen-netback/netback.c
1346
pending_idx_release[gop - queue->tx_unmap_ops] =
drivers/net/xen-netback/netback.c
1348
queue->pages_to_unmap[gop - queue->tx_unmap_ops] =
drivers/net/xen-netback/netback.c
1349
queue->mmap_pages[pending_idx];
drivers/net/xen-netback/netback.c
1351
idx_to_kaddr(queue, pending_idx),
drivers/net/xen-netback/netback.c
1353
queue->grant_tx_handle[pending_idx]);
drivers/net/xen-netback/netback.c
1354
xenvif_grant_handle_reset(queue, pending_idx);
drivers/net/xen-netback/netback.c
1358
} while (dp != queue->dealloc_prod);
drivers/net/xen-netback/netback.c
1360
queue->dealloc_cons = dc;
drivers/net/xen-netback/netback.c
1362
if (gop - queue->tx_unmap_ops > 0) {
drivers/net/xen-netback/netback.c
1364
ret = gnttab_unmap_refs(queue->tx_unmap_ops,
drivers/net/xen-netback/netback.c
1366
queue->pages_to_unmap,
drivers/net/xen-netback/netback.c
1367
gop - queue->tx_unmap_ops);
drivers/net/xen-netback/netback.c
1369
netdev_err(queue->vif->dev, "Unmap fail: nr_ops %tu ret %d\n",
drivers/net/xen-netback/netback.c
1370
gop - queue->tx_unmap_ops, ret);
drivers/net/xen-netback/netback.c
1371
for (i = 0; i < gop - queue->tx_unmap_ops; ++i) {
drivers/net/xen-netback/netback.c
1373
netdev_err(queue->vif->dev,
drivers/net/xen-netback/netback.c
1383
for (i = 0; i < gop - queue->tx_unmap_ops; ++i)
drivers/net/xen-netback/netback.c
1384
xenvif_idx_release(queue, pending_idx_release[i],
drivers/net/xen-netback/netback.c
1390
int xenvif_tx_action(struct xenvif_queue *queue, int budget)
drivers/net/xen-netback/netback.c
1395
if (unlikely(!tx_work_todo(queue)))
drivers/net/xen-netback/netback.c
1398
xenvif_tx_build_gops(queue, budget, &nr_cops, &nr_mops);
drivers/net/xen-netback/netback.c
1403
gnttab_batch_copy(queue->tx_copy_ops, nr_cops);
drivers/net/xen-netback/netback.c
1405
ret = gnttab_map_refs(queue->tx_map_ops,
drivers/net/xen-netback/netback.c
1407
queue->pages_to_map,
drivers/net/xen-netback/netback.c
1412
netdev_err(queue->vif->dev, "Map fail: nr %u ret %d\n",
drivers/net/xen-netback/netback.c
1415
WARN_ON_ONCE(queue->tx_map_ops[i].status ==
drivers/net/xen-netback/netback.c
1420
work_done = xenvif_tx_submit(queue);
drivers/net/xen-netback/netback.c
1425
static void _make_tx_response(struct xenvif_queue *queue,
drivers/net/xen-netback/netback.c
1430
RING_IDX i = queue->tx.rsp_prod_pvt;
drivers/net/xen-netback/netback.c
1433
resp = RING_GET_RESPONSE(&queue->tx, i);
drivers/net/xen-netback/netback.c
1438
RING_GET_RESPONSE(&queue->tx, ++i)->status = XEN_NETIF_RSP_NULL;
drivers/net/xen-netback/netback.c
1440
queue->tx.rsp_prod_pvt = ++i;
drivers/net/xen-netback/netback.c
1443
static void push_tx_responses(struct xenvif_queue *queue)
drivers/net/xen-netback/netback.c
1447
RING_PUSH_RESPONSES_AND_CHECK_NOTIFY(&queue->tx, notify);
drivers/net/xen-netback/netback.c
1449
notify_remote_via_irq(queue->tx_irq);
drivers/net/xen-netback/netback.c
1452
static void xenvif_idx_release(struct xenvif_queue *queue, u16 pending_idx,
drivers/net/xen-netback/netback.c
1459
pending_tx_info = &queue->pending_tx_info[pending_idx];
drivers/net/xen-netback/netback.c
1461
spin_lock_irqsave(&queue->response_lock, flags);
drivers/net/xen-netback/netback.c
1463
_make_tx_response(queue, &pending_tx_info->req,
drivers/net/xen-netback/netback.c
1470
index = pending_index(queue->pending_prod++);
drivers/net/xen-netback/netback.c
1471
queue->pending_ring[index] = pending_idx;
drivers/net/xen-netback/netback.c
1473
push_tx_responses(queue);
drivers/net/xen-netback/netback.c
1475
spin_unlock_irqrestore(&queue->response_lock, flags);
drivers/net/xen-netback/netback.c
1478
static void make_tx_response(struct xenvif_queue *queue,
drivers/net/xen-netback/netback.c
1485
spin_lock_irqsave(&queue->response_lock, flags);
drivers/net/xen-netback/netback.c
1487
_make_tx_response(queue, txp, extra_count, status);
drivers/net/xen-netback/netback.c
1488
push_tx_responses(queue);
drivers/net/xen-netback/netback.c
1490
spin_unlock_irqrestore(&queue->response_lock, flags);
drivers/net/xen-netback/netback.c
1493
static void xenvif_idx_unmap(struct xenvif_queue *queue, u16 pending_idx)
drivers/net/xen-netback/netback.c
1499
idx_to_kaddr(queue, pending_idx),
drivers/net/xen-netback/netback.c
1501
queue->grant_tx_handle[pending_idx]);
drivers/net/xen-netback/netback.c
1502
xenvif_grant_handle_reset(queue, pending_idx);
drivers/net/xen-netback/netback.c
1505
&queue->mmap_pages[pending_idx], 1);
drivers/net/xen-netback/netback.c
1507
netdev_err(queue->vif->dev,
drivers/net/xen-netback/netback.c
1518
static inline int tx_work_todo(struct xenvif_queue *queue)
drivers/net/xen-netback/netback.c
1520
if (likely(RING_HAS_UNCONSUMED_REQUESTS(&queue->tx)))
drivers/net/xen-netback/netback.c
1526
static inline bool tx_dealloc_work_todo(struct xenvif_queue *queue)
drivers/net/xen-netback/netback.c
1528
return queue->dealloc_cons != queue->dealloc_prod;
drivers/net/xen-netback/netback.c
1531
void xenvif_unmap_frontend_data_rings(struct xenvif_queue *queue)
drivers/net/xen-netback/netback.c
1533
if (queue->tx.sring)
drivers/net/xen-netback/netback.c
1534
xenbus_unmap_ring_vfree(xenvif_to_xenbus_device(queue->vif),
drivers/net/xen-netback/netback.c
1535
queue->tx.sring);
drivers/net/xen-netback/netback.c
1536
if (queue->rx.sring)
drivers/net/xen-netback/netback.c
1537
xenbus_unmap_ring_vfree(xenvif_to_xenbus_device(queue->vif),
drivers/net/xen-netback/netback.c
1538
queue->rx.sring);
drivers/net/xen-netback/netback.c
1541
int xenvif_map_frontend_data_rings(struct xenvif_queue *queue,
drivers/net/xen-netback/netback.c
1551
err = xenbus_map_ring_valloc(xenvif_to_xenbus_device(queue->vif),
drivers/net/xen-netback/netback.c
1560
BACK_RING_ATTACH(&queue->tx, txs, rsp_prod, XEN_PAGE_SIZE);
drivers/net/xen-netback/netback.c
1563
if (req_prod - rsp_prod > RING_SIZE(&queue->tx))
drivers/net/xen-netback/netback.c
1566
err = xenbus_map_ring_valloc(xenvif_to_xenbus_device(queue->vif),
drivers/net/xen-netback/netback.c
1575
BACK_RING_ATTACH(&queue->rx, rxs, rsp_prod, XEN_PAGE_SIZE);
drivers/net/xen-netback/netback.c
1578
if (req_prod - rsp_prod > RING_SIZE(&queue->rx))
drivers/net/xen-netback/netback.c
1584
xenvif_unmap_frontend_data_rings(queue);
drivers/net/xen-netback/netback.c
1588
static bool xenvif_dealloc_kthread_should_stop(struct xenvif_queue *queue)
drivers/net/xen-netback/netback.c
1594
!atomic_read(&queue->inflight_packets);
drivers/net/xen-netback/netback.c
1599
struct xenvif_queue *queue = data;
drivers/net/xen-netback/netback.c
1602
wait_event_interruptible(queue->dealloc_wq,
drivers/net/xen-netback/netback.c
1603
tx_dealloc_work_todo(queue) ||
drivers/net/xen-netback/netback.c
1604
xenvif_dealloc_kthread_should_stop(queue));
drivers/net/xen-netback/netback.c
1605
if (xenvif_dealloc_kthread_should_stop(queue))
drivers/net/xen-netback/netback.c
1608
xenvif_tx_dealloc_action(queue);
drivers/net/xen-netback/netback.c
161
void xenvif_kick_thread(struct xenvif_queue *queue)
drivers/net/xen-netback/netback.c
1613
if (tx_dealloc_work_todo(queue))
drivers/net/xen-netback/netback.c
1614
xenvif_tx_dealloc_action(queue);
drivers/net/xen-netback/netback.c
163
wake_up(&queue->wq);
drivers/net/xen-netback/netback.c
166
void xenvif_napi_schedule_or_enable_events(struct xenvif_queue *queue)
drivers/net/xen-netback/netback.c
170
RING_FINAL_CHECK_FOR_REQUESTS(&queue->tx, more_to_do);
drivers/net/xen-netback/netback.c
173
napi_schedule(&queue->napi);
drivers/net/xen-netback/netback.c
175
&queue->eoi_pending) &
drivers/net/xen-netback/netback.c
177
xen_irq_lateeoi(queue->tx_irq, 0);
drivers/net/xen-netback/netback.c
180
static void tx_add_credit(struct xenvif_queue *queue)
drivers/net/xen-netback/netback.c
188
max_burst = max(131072UL, queue->credit_bytes);
drivers/net/xen-netback/netback.c
191
max_credit = queue->remaining_credit + queue->credit_bytes;
drivers/net/xen-netback/netback.c
192
if (max_credit < queue->remaining_credit)
drivers/net/xen-netback/netback.c
195
queue->remaining_credit = min(max_credit, max_burst);
drivers/net/xen-netback/netback.c
196
queue->rate_limited = false;
drivers/net/xen-netback/netback.c
201
struct xenvif_queue *queue = timer_container_of(queue, t,
drivers/net/xen-netback/netback.c
203
tx_add_credit(queue);
drivers/net/xen-netback/netback.c
204
xenvif_napi_schedule_or_enable_events(queue);
drivers/net/xen-netback/netback.c
207
static void xenvif_tx_err(struct xenvif_queue *queue,
drivers/net/xen-netback/netback.c
211
RING_IDX cons = queue->tx.req_cons;
drivers/net/xen-netback/netback.c
214
make_tx_response(queue, txp, extra_count, XEN_NETIF_RSP_ERROR);
drivers/net/xen-netback/netback.c
217
RING_COPY_REQUEST(&queue->tx, cons++, txp);
drivers/net/xen-netback/netback.c
220
queue->tx.req_cons = cons;
drivers/net/xen-netback/netback.c
232
static int xenvif_count_requests(struct xenvif_queue *queue,
drivers/net/xen-netback/netback.c
238
RING_IDX cons = queue->tx.req_cons;
drivers/net/xen-netback/netback.c
250
netdev_err(queue->vif->dev,
drivers/net/xen-netback/netback.c
253
xenvif_fatal_tx_err(queue->vif);
drivers/net/xen-netback/netback.c
261
netdev_err(queue->vif->dev,
drivers/net/xen-netback/netback.c
264
xenvif_fatal_tx_err(queue->vif);
drivers/net/xen-netback/netback.c
277
netdev_dbg(queue->vif->dev,
drivers/net/xen-netback/netback.c
286
RING_COPY_REQUEST(&queue->tx, cons + slots, txp);
drivers/net/xen-netback/netback.c
299
netdev_dbg(queue->vif->dev,
drivers/net/xen-netback/netback.c
309
netdev_err(queue->vif->dev, "Cross page boundary, txp->offset: %u, size: %u\n",
drivers/net/xen-netback/netback.c
311
xenvif_fatal_tx_err(queue->vif);
drivers/net/xen-netback/netback.c
323
xenvif_tx_err(queue, first, extra_count, cons + slots);
drivers/net/xen-netback/netback.c
341
static inline void xenvif_tx_create_map_op(struct xenvif_queue *queue,
drivers/net/xen-netback/netback.c
347
queue->pages_to_map[mop-queue->tx_map_ops] = queue->mmap_pages[pending_idx];
drivers/net/xen-netback/netback.c
348
gnttab_set_map_op(mop, idx_to_kaddr(queue, pending_idx),
drivers/net/xen-netback/netback.c
350
txp->gref, queue->vif->domid);
drivers/net/xen-netback/netback.c
352
memcpy(&queue->pending_tx_info[pending_idx].req, txp,
drivers/net/xen-netback/netback.c
354
queue->pending_tx_info[pending_idx].extra_count = extra_count;
drivers/net/xen-netback/netback.c
376
static void xenvif_get_requests(struct xenvif_queue *queue,
drivers/net/xen-netback/netback.c
392
struct gnttab_copy *cop = queue->tx_copy_ops + *copy_ops;
drivers/net/xen-netback/netback.c
393
struct gnttab_map_grant_ref *gop = queue->tx_map_ops + *map_ops;
drivers/net/xen-netback/netback.c
408
cop->source.domid = queue->vif->domid;
drivers/net/xen-netback/netback.c
428
index = pending_index(queue->pending_cons);
drivers/net/xen-netback/netback.c
429
pending_idx = queue->pending_ring[index];
drivers/net/xen-netback/netback.c
430
callback_param(queue, pending_idx).ctx = NULL;
drivers/net/xen-netback/netback.c
441
memcpy(&queue->pending_tx_info[pending_idx].req,
drivers/net/xen-netback/netback.c
443
queue->pending_tx_info[pending_idx].extra_count =
drivers/net/xen-netback/netback.c
450
queue->pending_cons++;
drivers/net/xen-netback/netback.c
465
make_tx_response(queue, txp, 0, XEN_NETIF_RSP_OKAY);
drivers/net/xen-netback/netback.c
470
index = pending_index(queue->pending_cons++);
drivers/net/xen-netback/netback.c
471
pending_idx = queue->pending_ring[index];
drivers/net/xen-netback/netback.c
472
xenvif_tx_create_map_op(queue, pending_idx, txp,
drivers/net/xen-netback/netback.c
491
make_tx_response(queue, txp, 0,
drivers/net/xen-netback/netback.c
496
index = pending_index(queue->pending_cons++);
drivers/net/xen-netback/netback.c
497
pending_idx = queue->pending_ring[index];
drivers/net/xen-netback/netback.c
498
xenvif_tx_create_map_op(queue, pending_idx, txp, 0,
drivers/net/xen-netback/netback.c
520
(*copy_ops) = cop - queue->tx_copy_ops;
drivers/net/xen-netback/netback.c
521
(*map_ops) = gop - queue->tx_map_ops;
drivers/net/xen-netback/netback.c
524
static inline void xenvif_grant_handle_set(struct xenvif_queue *queue,
drivers/net/xen-netback/netback.c
528
if (unlikely(queue->grant_tx_handle[pending_idx] !=
drivers/net/xen-netback/netback.c
530
netdev_err(queue->vif->dev,
drivers/net/xen-netback/netback.c
535
queue->grant_tx_handle[pending_idx] = handle;
drivers/net/xen-netback/netback.c
538
static inline void xenvif_grant_handle_reset(struct xenvif_queue *queue,
drivers/net/xen-netback/netback.c
541
if (unlikely(queue->grant_tx_handle[pending_idx] ==
drivers/net/xen-netback/netback.c
543
netdev_err(queue->vif->dev,
drivers/net/xen-netback/netback.c
548
queue->grant_tx_handle[pending_idx] = NETBACK_INVALID_HANDLE;
drivers/net/xen-netback/netback.c
551
static int xenvif_tx_check_gop(struct xenvif_queue *queue,
drivers/net/xen-netback/netback.c
589
xenvif_idx_release(queue, pending_idx,
drivers/net/xen-netback/netback.c
594
netdev_dbg(queue->vif->dev,
drivers/net/xen-netback/netback.c
601
xenvif_idx_release(queue, pending_idx,
drivers/net/xen-netback/netback.c
617
xenvif_grant_handle_set(queue,
drivers/net/xen-netback/netback.c
622
xenvif_idx_unmap(queue, pending_idx);
drivers/net/xen-netback/netback.c
628
xenvif_idx_release(queue, pending_idx,
drivers/net/xen-netback/netback.c
631
xenvif_idx_release(queue, pending_idx,
drivers/net/xen-netback/netback.c
639
netdev_dbg(queue->vif->dev,
drivers/net/xen-netback/netback.c
646
xenvif_idx_release(queue, pending_idx, XEN_NETIF_RSP_ERROR);
drivers/net/xen-netback/netback.c
655
xenvif_idx_unmap(queue, pending_idx);
drivers/net/xen-netback/netback.c
656
xenvif_idx_release(queue, pending_idx,
drivers/net/xen-netback/netback.c
666
xenvif_idx_unmap(queue, pending_idx);
drivers/net/xen-netback/netback.c
667
xenvif_idx_release(queue, pending_idx,
drivers/net/xen-netback/netback.c
688
static void xenvif_fill_frags(struct xenvif_queue *queue, struct sk_buff *skb)
drivers/net/xen-netback/netback.c
706
&callback_param(queue, pending_idx);
drivers/net/xen-netback/netback.c
708
callback_param(queue, prev_pending_idx).ctx =
drivers/net/xen-netback/netback.c
709
&callback_param(queue, pending_idx);
drivers/net/xen-netback/netback.c
711
callback_param(queue, pending_idx).ctx = NULL;
drivers/net/xen-netback/netback.c
714
txp = &queue->pending_tx_info[pending_idx].req;
drivers/net/xen-netback/netback.c
715
page = virt_to_page((void *)idx_to_kaddr(queue, pending_idx));
drivers/net/xen-netback/netback.c
722
get_page(queue->mmap_pages[pending_idx]);
drivers/net/xen-netback/netback.c
726
static int xenvif_get_extras(struct xenvif_queue *queue,
drivers/net/xen-netback/netback.c
732
RING_IDX cons = queue->tx.req_cons;
drivers/net/xen-netback/netback.c
736
netdev_err(queue->vif->dev, "Missing extra info\n");
drivers/net/xen-netback/netback.c
737
xenvif_fatal_tx_err(queue->vif);
drivers/net/xen-netback/netback.c
741
RING_COPY_REQUEST(&queue->tx, cons, &extra);
drivers/net/xen-netback/netback.c
743
queue->tx.req_cons = ++cons;
drivers/net/xen-netback/netback.c
748
netdev_err(queue->vif->dev,
drivers/net/xen-netback/netback.c
750
xenvif_fatal_tx_err(queue->vif);
drivers/net/xen-netback/netback.c
789
static int checksum_setup(struct xenvif_queue *queue, struct sk_buff *skb)
drivers/net/xen-netback/netback.c
799
queue->stats.rx_gso_checksum_fixup++;
drivers/net/xen-netback/netback.c
811
static bool tx_credit_exceeded(struct xenvif_queue *queue, unsigned size)
drivers/net/xen-netback/netback.c
814
u64 next_credit = queue->credit_window_start +
drivers/net/xen-netback/netback.c
815
msecs_to_jiffies(queue->credit_usec / 1000);
drivers/net/xen-netback/netback.c
818
if (timer_pending(&queue->credit_timeout)) {
drivers/net/xen-netback/netback.c
819
queue->rate_limited = true;
drivers/net/xen-netback/netback.c
825
queue->credit_window_start = now;
drivers/net/xen-netback/netback.c
826
tx_add_credit(queue);
drivers/net/xen-netback/netback.c
830
if (size > queue->remaining_credit) {
drivers/net/xen-netback/netback.c
831
mod_timer(&queue->credit_timeout,
drivers/net/xen-netback/netback.c
833
queue->credit_window_start = next_credit;
drivers/net/xen-netback/netback.c
834
queue->rate_limited = true;
drivers/net/xen-netback/netback.c
916
static void xenvif_tx_build_gops(struct xenvif_queue *queue,
drivers/net/xen-netback/netback.c
925
while (skb_queue_len(&queue->tx_queue) < budget) {
drivers/net/xen-netback/netback.c
934
if (queue->tx.sring->req_prod - queue->tx.req_cons >
drivers/net/xen-netback/netback.c
936
netdev_err(queue->vif->dev,
drivers/net/xen-netback/netback.c
939
queue->tx.sring->req_prod, queue->tx.req_cons,
drivers/net/xen-netback/netback.c
941
xenvif_fatal_tx_err(queue->vif);
drivers/net/xen-netback/netback.c
945
work_to_do = XEN_RING_NR_UNCONSUMED_REQUESTS(&queue->tx);
drivers/net/xen-netback/netback.c
949
idx = queue->tx.req_cons;
drivers/net/xen-netback/netback.c
951
RING_COPY_REQUEST(&queue->tx, idx, &txreq);
drivers/net/xen-netback/netback.c
954
if (txreq.size > queue->remaining_credit &&
drivers/net/xen-netback/netback.c
955
tx_credit_exceeded(queue, txreq.size))
drivers/net/xen-netback/netback.c
958
queue->remaining_credit -= txreq.size;
drivers/net/xen-netback/netback.c
961
queue->tx.req_cons = ++idx;
drivers/net/xen-netback/netback.c
966
work_to_do = xenvif_get_extras(queue, extras,
drivers/net/xen-netback/netback.c
969
idx = queue->tx.req_cons;
drivers/net/xen-netback/netback.c
978
ret = xenvif_mcast_add(queue->vif, extra->u.mcast.addr);
drivers/net/xen-netback/netback.c
980
make_tx_response(queue, &txreq, extra_count,
drivers/net/xen-netback/netback.c
991
xenvif_mcast_del(queue->vif, extra->u.mcast.addr);
drivers/net/xen-netback/netback.c
993
make_tx_response(queue, &txreq, extra_count,
drivers/net/xen-netback/rx.c
101
__skb_queue_tail(&queue->rx_queue, skb);
drivers/net/xen-netback/rx.c
103
queue->rx_queue_len += skb->len;
drivers/net/xen-netback/rx.c
106
spin_unlock_irqrestore(&queue->rx_queue.lock, flags);
drivers/net/xen-netback/rx.c
111
static struct sk_buff *xenvif_rx_dequeue(struct xenvif_queue *queue)
drivers/net/xen-netback/rx.c
115
spin_lock_irq(&queue->rx_queue.lock);
drivers/net/xen-netback/rx.c
117
skb = __skb_dequeue(&queue->rx_queue);
drivers/net/xen-netback/rx.c
119
xenvif_update_needed_slots(queue, skb_peek(&queue->rx_queue));
drivers/net/xen-netback/rx.c
121
queue->rx_queue_len -= skb->len;
drivers/net/xen-netback/rx.c
122
if (queue->rx_queue_len < queue->rx_queue_max) {
drivers/net/xen-netback/rx.c
125
txq = netdev_get_tx_queue(queue->vif->dev, queue->id);
drivers/net/xen-netback/rx.c
130
spin_unlock_irq(&queue->rx_queue.lock);
drivers/net/xen-netback/rx.c
135
static void xenvif_rx_queue_purge(struct xenvif_queue *queue)
drivers/net/xen-netback/rx.c
139
while ((skb = xenvif_rx_dequeue(queue)) != NULL)
drivers/net/xen-netback/rx.c
143
static void xenvif_rx_queue_drop_expired(struct xenvif_queue *queue)
drivers/net/xen-netback/rx.c
148
skb = skb_peek(&queue->rx_queue);
drivers/net/xen-netback/rx.c
153
xenvif_rx_dequeue(queue);
drivers/net/xen-netback/rx.c
155
queue->vif->dev->stats.rx_dropped++;
drivers/net/xen-netback/rx.c
159
static void xenvif_rx_copy_flush(struct xenvif_queue *queue)
drivers/net/xen-netback/rx.c
164
gnttab_batch_copy(queue->rx_copy.op, queue->rx_copy.num);
drivers/net/xen-netback/rx.c
166
for (i = 0; i < queue->rx_copy.num; i++) {
drivers/net/xen-netback/rx.c
169
op = &queue->rx_copy.op[i];
drivers/net/xen-netback/rx.c
177
rsp = RING_GET_RESPONSE(&queue->rx,
drivers/net/xen-netback/rx.c
178
queue->rx_copy.idx[i]);
drivers/net/xen-netback/rx.c
183
queue->rx_copy.num = 0;
drivers/net/xen-netback/rx.c
186
RING_PUSH_RESPONSES_AND_CHECK_NOTIFY(&queue->rx, notify);
drivers/net/xen-netback/rx.c
188
notify_remote_via_irq(queue->rx_irq);
drivers/net/xen-netback/rx.c
190
__skb_queue_purge(queue->rx_copy.completed);
drivers/net/xen-netback/rx.c
193
static void xenvif_rx_copy_add(struct xenvif_queue *queue,
drivers/net/xen-netback/rx.c
201
if (queue->rx_copy.num == COPY_BATCH_SIZE)
drivers/net/xen-netback/rx.c
202
xenvif_rx_copy_flush(queue);
drivers/net/xen-netback/rx.c
204
op = &queue->rx_copy.op[queue->rx_copy.num];
drivers/net/xen-netback/rx.c
222
op->dest.domid = queue->vif->domid;
drivers/net/xen-netback/rx.c
226
queue->rx_copy.idx[queue->rx_copy.num] = queue->rx.req_cons;
drivers/net/xen-netback/rx.c
227
queue->rx_copy.num++;
drivers/net/xen-netback/rx.c
252
static void xenvif_rx_next_skb(struct xenvif_queue *queue,
drivers/net/xen-netback/rx.c
258
skb = xenvif_rx_dequeue(queue);
drivers/net/xen-netback/rx.c
260
queue->stats.tx_bytes += skb->len;
drivers/net/xen-netback/rx.c
261
queue->stats.tx_packets++;
drivers/net/xen-netback/rx.c
272
if ((1 << gso_type) & queue->vif->gso_mask) {
drivers/net/xen-netback/rx.c
287
if (queue->vif->xdp_headroom) {
drivers/net/xen-netback/rx.c
293
extra->u.xdp.headroom = queue->vif->xdp_headroom;
drivers/net/xen-netback/rx.c
328
static void xenvif_rx_complete(struct xenvif_queue *queue,
drivers/net/xen-netback/rx.c
332
queue->rx.rsp_prod_pvt = queue->rx.req_cons;
drivers/net/xen-netback/rx.c
334
__skb_queue_tail(queue->rx_copy.completed, pkt->skb);
drivers/net/xen-netback/rx.c
355
static void xenvif_rx_next_chunk(struct xenvif_queue *queue,
drivers/net/xen-netback/rx.c
393
static void xenvif_rx_data_slot(struct xenvif_queue *queue,
drivers/net/xen-netback/rx.c
398
unsigned int offset = queue->vif->xdp_headroom;
drivers/net/xen-netback/rx.c
405
xenvif_rx_next_chunk(queue, pkt, offset, &data, &len);
drivers/net/xen-netback/rx.c
406
xenvif_rx_copy_add(queue, req, offset, data, len);
drivers/net/xen-netback/rx.c
42
static void xenvif_update_needed_slots(struct xenvif_queue *queue,
drivers/net/xen-netback/rx.c
437
static void xenvif_rx_extra_slot(struct xenvif_queue *queue,
drivers/net/xen-netback/rx.c
461
static void xenvif_rx_skb(struct xenvif_queue *queue)
drivers/net/xen-netback/rx.c
465
xenvif_rx_next_skb(queue, &pkt);
drivers/net/xen-netback/rx.c
467
queue->last_rx_time = jiffies;
drivers/net/xen-netback/rx.c
473
req = RING_GET_REQUEST(&queue->rx, queue->rx.req_cons);
drivers/net/xen-netback/rx.c
474
rsp = RING_GET_RESPONSE(&queue->rx, queue->rx.req_cons);
drivers/net/xen-netback/rx.c
478
xenvif_rx_extra_slot(queue, &pkt, req, rsp);
drivers/net/xen-netback/rx.c
480
xenvif_rx_data_slot(queue, &pkt, req, rsp);
drivers/net/xen-netback/rx.c
482
queue->rx.req_cons++;
drivers/net/xen-netback/rx.c
486
xenvif_rx_complete(queue, &pkt);
drivers/net/xen-netback/rx.c
491
static void xenvif_rx_action(struct xenvif_queue *queue)
drivers/net/xen-netback/rx.c
497
queue->rx_copy.completed = &completed_skbs;
drivers/net/xen-netback/rx.c
499
while (xenvif_rx_ring_slots_available(queue) &&
drivers/net/xen-netback/rx.c
500
!skb_queue_empty(&queue->rx_queue) &&
drivers/net/xen-netback/rx.c
502
xenvif_rx_skb(queue);
drivers/net/xen-netback/rx.c
507
xenvif_rx_copy_flush(queue);
drivers/net/xen-netback/rx.c
510
static RING_IDX xenvif_rx_queue_slots(const struct xenvif_queue *queue)
drivers/net/xen-netback/rx.c
514
prod = queue->rx.sring->req_prod;
drivers/net/xen-netback/rx.c
515
cons = queue->rx.req_cons;
drivers/net/xen-netback/rx.c
520
static bool xenvif_rx_queue_stalled(const struct xenvif_queue *queue)
drivers/net/xen-netback/rx.c
522
unsigned int needed = READ_ONCE(queue->rx_slots_needed);
drivers/net/xen-netback/rx.c
524
return !queue->stalled &&
drivers/net/xen-netback/rx.c
525
xenvif_rx_queue_slots(queue) < needed &&
drivers/net/xen-netback/rx.c
527
queue->last_rx_time + queue->vif->stall_timeout);
drivers/net/xen-netback/rx.c
530
static bool xenvif_rx_queue_ready(struct xenvif_queue *queue)
drivers/net/xen-netback/rx.c
532
unsigned int needed = READ_ONCE(queue->rx_slots_needed);
drivers/net/xen-netback/rx.c
534
return queue->stalled && xenvif_rx_queue_slots(queue) >= needed;
drivers/net/xen-netback/rx.c
537
bool xenvif_have_rx_work(struct xenvif_queue *queue, bool test_kthread)
drivers/net/xen-netback/rx.c
539
return xenvif_rx_ring_slots_available(queue) ||
drivers/net/xen-netback/rx.c
540
(queue->vif->stall_timeout &&
drivers/net/xen-netback/rx.c
541
(xenvif_rx_queue_stalled(queue) ||
drivers/net/xen-netback/rx.c
542
xenvif_rx_queue_ready(queue))) ||
drivers/net/xen-netback/rx.c
544
queue->vif->disabled;
drivers/net/xen-netback/rx.c
547
static long xenvif_rx_queue_timeout(struct xenvif_queue *queue)
drivers/net/xen-netback/rx.c
55
WRITE_ONCE(queue->rx_slots_needed, needed);
drivers/net/xen-netback/rx.c
552
skb = skb_peek(&queue->rx_queue);
drivers/net/xen-netback/rx.c
570
static void xenvif_wait_for_rx_work(struct xenvif_queue *queue)
drivers/net/xen-netback/rx.c
574
if (xenvif_have_rx_work(queue, true))
drivers/net/xen-netback/rx.c
58
static bool xenvif_rx_ring_slots_available(struct xenvif_queue *queue)
drivers/net/xen-netback/rx.c
580
prepare_to_wait(&queue->wq, &wait, TASK_INTERRUPTIBLE);
drivers/net/xen-netback/rx.c
581
if (xenvif_have_rx_work(queue, true))
drivers/net/xen-netback/rx.c
584
&queue->eoi_pending) &
drivers/net/xen-netback/rx.c
586
xen_irq_lateeoi(queue->rx_irq, 0);
drivers/net/xen-netback/rx.c
588
ret = schedule_timeout(xenvif_rx_queue_timeout(queue));
drivers/net/xen-netback/rx.c
592
finish_wait(&queue->wq, &wait);
drivers/net/xen-netback/rx.c
595
static void xenvif_queue_carrier_off(struct xenvif_queue *queue)
drivers/net/xen-netback/rx.c
597
struct xenvif *vif = queue->vif;
drivers/net/xen-netback/rx.c
599
queue->stalled = true;
drivers/net/xen-netback/rx.c
610
static void xenvif_queue_carrier_on(struct xenvif_queue *queue)
drivers/net/xen-netback/rx.c
612
struct xenvif *vif = queue->vif;
drivers/net/xen-netback/rx.c
614
queue->last_rx_time = jiffies; /* Reset Rx stall detection. */
drivers/net/xen-netback/rx.c
615
queue->stalled = false;
drivers/net/xen-netback/rx.c
628
struct xenvif_queue *queue = data;
drivers/net/xen-netback/rx.c
629
struct xenvif *vif = queue->vif;
drivers/net/xen-netback/rx.c
63
needed = READ_ONCE(queue->rx_slots_needed);
drivers/net/xen-netback/rx.c
632
xenvif_queue_carrier_on(queue);
drivers/net/xen-netback/rx.c
635
xenvif_wait_for_rx_work(queue);
drivers/net/xen-netback/rx.c
647
if (unlikely(vif->disabled && queue->id == 0)) {
drivers/net/xen-netback/rx.c
652
if (!skb_queue_empty(&queue->rx_queue))
drivers/net/xen-netback/rx.c
653
xenvif_rx_action(queue);
drivers/net/xen-netback/rx.c
660
if (xenvif_rx_queue_stalled(queue))
drivers/net/xen-netback/rx.c
661
xenvif_queue_carrier_off(queue);
drivers/net/xen-netback/rx.c
662
else if (xenvif_rx_queue_ready(queue))
drivers/net/xen-netback/rx.c
663
xenvif_queue_carrier_on(queue);
drivers/net/xen-netback/rx.c
671
xenvif_rx_queue_drop_expired(queue);
drivers/net/xen-netback/rx.c
677
xenvif_rx_queue_purge(queue);
drivers/net/xen-netback/rx.c
68
prod = queue->rx.sring->req_prod;
drivers/net/xen-netback/rx.c
69
cons = queue->rx.req_cons;
drivers/net/xen-netback/rx.c
74
queue->rx.sring->req_event = prod + 1;
drivers/net/xen-netback/rx.c
80
} while (queue->rx.sring->req_prod != prod);
drivers/net/xen-netback/rx.c
85
bool xenvif_rx_queue_tail(struct xenvif_queue *queue, struct sk_buff *skb)
drivers/net/xen-netback/rx.c
90
spin_lock_irqsave(&queue->rx_queue.lock, flags);
drivers/net/xen-netback/rx.c
92
if (queue->rx_queue_len >= queue->rx_queue_max) {
drivers/net/xen-netback/rx.c
93
struct net_device *dev = queue->vif->dev;
drivers/net/xen-netback/rx.c
95
netif_tx_stop_queue(netdev_get_tx_queue(dev, queue->id));
drivers/net/xen-netback/rx.c
98
if (skb_queue_empty(&queue->rx_queue))
drivers/net/xen-netback/rx.c
99
xenvif_update_needed_slots(queue, skb);
drivers/net/xen-netback/xenbus.c
109
struct xenvif_queue *queue =
drivers/net/xen-netback/xenbus.c
131
xenvif_interrupt(0, (void *)queue);
drivers/net/xen-netback/xenbus.c
134
queue->id);
drivers/net/xen-netback/xenbus.c
14
struct xenvif_queue *queue);
drivers/net/xen-netback/xenbus.c
143
void *queue = NULL;
drivers/net/xen-netback/xenbus.c
146
queue = inode->i_private;
drivers/net/xen-netback/xenbus.c
147
ret = single_open(filp, xenvif_read_io_ring, queue);
drivers/net/xen-netback/xenbus.c
28
struct xenvif_queue *queue = m->private;
drivers/net/xen-netback/xenbus.c
29
struct xen_netif_tx_back_ring *tx_ring = &queue->tx;
drivers/net/xen-netback/xenbus.c
30
struct xen_netif_rx_back_ring *rx_ring = &queue->rx;
drivers/net/xen-netback/xenbus.c
36
seq_printf(m, "Queue %d\nTX: nr_ents %u\n", queue->id,
drivers/net/xen-netback/xenbus.c
52
queue->pending_prod,
drivers/net/xen-netback/xenbus.c
53
queue->pending_cons,
drivers/net/xen-netback/xenbus.c
534
struct xenvif_queue *queue = &vif->queues[queue_index];
drivers/net/xen-netback/xenbus.c
536
queue->credit_bytes = credit_bytes;
drivers/net/xen-netback/xenbus.c
537
queue->credit_usec = credit_usec;
drivers/net/xen-netback/xenbus.c
538
if (!mod_timer_pending(&queue->credit_timeout, jiffies) &&
drivers/net/xen-netback/xenbus.c
539
queue->remaining_credit > queue->credit_bytes) {
drivers/net/xen-netback/xenbus.c
54
nr_pending_reqs(queue));
drivers/net/xen-netback/xenbus.c
540
queue->remaining_credit = queue->credit_bytes;
drivers/net/xen-netback/xenbus.c
56
queue->dealloc_prod,
drivers/net/xen-netback/xenbus.c
57
queue->dealloc_cons,
drivers/net/xen-netback/xenbus.c
58
queue->dealloc_prod - queue->dealloc_cons);
drivers/net/xen-netback/xenbus.c
731
struct xenvif_queue *queue;
drivers/net/xen-netback/xenbus.c
777
queue = &be->vif->queues[queue_index];
drivers/net/xen-netback/xenbus.c
778
queue->vif = be->vif;
drivers/net/xen-netback/xenbus.c
779
queue->id = queue_index;
drivers/net/xen-netback/xenbus.c
780
snprintf(queue->name, sizeof(queue->name), "%s-q%u",
drivers/net/xen-netback/xenbus.c
781
be->vif->dev->name, queue->id);
drivers/net/xen-netback/xenbus.c
783
err = xenvif_init_queue(queue);
drivers/net/xen-netback/xenbus.c
795
queue->credit_bytes = credit_bytes;
drivers/net/xen-netback/xenbus.c
796
queue->remaining_credit = credit_bytes;
drivers/net/xen-netback/xenbus.c
797
queue->credit_usec = credit_usec;
drivers/net/xen-netback/xenbus.c
799
err = connect_data_rings(be, queue);
drivers/net/xen-netback/xenbus.c
806
xenvif_deinit_queue(queue);
drivers/net/xen-netback/xenbus.c
83
queue->napi.state, queue->napi.weight,
drivers/net/xen-netback/xenbus.c
84
skb_queue_len(&queue->tx_queue),
drivers/net/xen-netback/xenbus.c
85
timer_pending(&queue->credit_timeout),
drivers/net/xen-netback/xenbus.c
851
struct xenvif_queue *queue)
drivers/net/xen-netback/xenbus.c
854
unsigned int num_queues = queue->vif->num_queues;
drivers/net/xen-netback/xenbus.c
86
queue->credit_bytes,
drivers/net/xen-netback/xenbus.c
87
queue->credit_usec,
drivers/net/xen-netback/xenbus.c
88
queue->remaining_credit,
drivers/net/xen-netback/xenbus.c
884
queue->id);
drivers/net/xen-netback/xenbus.c
89
queue->credit_timeout.expires,
drivers/net/xen-netback/xenbus.c
914
err = xenvif_connect_data(queue, tx_ring_ref, rx_ring_ref,
drivers/net/xen-netback/xenbus.c
92
dev_queue = netdev_get_tx_queue(queue->vif->dev, queue->id);
drivers/net/xen-netback/xenbus.c
95
queue->rx_queue_len, queue->rx_queue_max,
drivers/net/xen-netback/xenbus.c
96
skb_queue_len(&queue->rx_queue),
drivers/net/xen-netfront.c
1002
trace_xdp_exception(queue->info->netdev, prog, act);
drivers/net/xen-netfront.c
1011
trace_xdp_exception(queue->info->netdev, prog, act);
drivers/net/xen-netfront.c
1015
bpf_warn_invalid_xdp_action(queue->info->netdev, prog, act);
drivers/net/xen-netfront.c
1021
static int xennet_get_responses(struct netfront_queue *queue,
drivers/net/xen-netfront.c
1028
RING_IDX cons = queue->rx.rsp_cons;
drivers/net/xen-netfront.c
1029
struct sk_buff *skb = xennet_get_rx_skb(queue, cons);
drivers/net/xen-netfront.c
1031
grant_ref_t ref = xennet_get_rx_ref(queue, cons);
drivers/net/xen-netfront.c
1032
struct device *dev = &queue->info->netdev->dev;
drivers/net/xen-netfront.c
1040
err = xennet_get_extras(queue, extras, rp);
drivers/net/xen-netfront.c
1049
cons = queue->rx.rsp_cons;
drivers/net/xen-netfront.c
1071
xennet_move_rx_slot(queue, skb, ref);
drivers/net/xen-netfront.c
1079
queue->info->broken = true;
drivers/net/xen-netfront.c
1084
gnttab_release_grant_reference(&queue->gref_rx_head, ref);
drivers/net/xen-netfront.c
1087
xdp_prog = rcu_dereference(queue->xdp_prog);
drivers/net/xen-netfront.c
1091
verdict = xennet_run_xdp(queue,
drivers/net/xen-netfront.c
1116
RING_COPY_RESPONSE(&queue->rx, cons + slots, &rx_local);
drivers/net/xen-netfront.c
1118
skb = xennet_get_rx_skb(queue, cons + slots);
drivers/net/xen-netfront.c
1119
ref = xennet_get_rx_ref(queue, cons + slots);
drivers/net/xen-netfront.c
1130
xennet_set_rx_rsp_cons(queue, cons + slots);
drivers/net/xen-netfront.c
1164
static int xennet_fill_frags(struct netfront_queue *queue,
drivers/net/xen-netfront.c
1168
RING_IDX cons = queue->rx.rsp_cons;
drivers/net/xen-netfront.c
1175
RING_COPY_RESPONSE(&queue->rx, ++cons, &rx);
drivers/net/xen-netfront.c
1184
xennet_set_rx_rsp_cons(queue,
drivers/net/xen-netfront.c
1198
xennet_set_rx_rsp_cons(queue, cons);
drivers/net/xen-netfront.c
1227
static int handle_incoming_queue(struct netfront_queue *queue,
drivers/net/xen-netfront.c
1230
struct netfront_stats *rx_stats = this_cpu_ptr(queue->info->rx_stats);
drivers/net/xen-netfront.c
1241
skb->protocol = eth_type_trans(skb, queue->info->netdev);
drivers/net/xen-netfront.c
1244
if (checksum_setup(queue->info->netdev, skb)) {
drivers/net/xen-netfront.c
1247
queue->info->netdev->stats.rx_errors++;
drivers/net/xen-netfront.c
1257
napi_gro_receive(&queue->napi, skb);
drivers/net/xen-netfront.c
1265
struct netfront_queue *queue = container_of(napi, struct netfront_queue, napi);
drivers/net/xen-netfront.c
1266
struct net_device *dev = queue->info->netdev;
drivers/net/xen-netfront.c
1279
spin_lock(&queue->rx_lock);
drivers/net/xen-netfront.c
1285
rp = queue->rx.sring->rsp_prod;
drivers/net/xen-netfront.c
1286
if (RING_RESPONSE_PROD_OVERFLOW(&queue->rx, rp)) {
drivers/net/xen-netfront.c
1288
rp - queue->rx.rsp_cons);
drivers/net/xen-netfront.c
1289
queue->info->broken = true;
drivers/net/xen-netfront.c
1290
spin_unlock(&queue->rx_lock);
drivers/net/xen-netfront.c
1295
i = queue->rx.rsp_cons;
drivers/net/xen-netfront.c
1298
RING_COPY_RESPONSE(&queue->rx, i, rx);
drivers/net/xen-netfront.c
1301
err = xennet_get_responses(queue, &rinfo, rp, &tmpq,
drivers/net/xen-netfront.c
1305
if (queue->info->broken) {
drivers/net/xen-netfront.c
1306
spin_unlock(&queue->rx_lock);
drivers/net/xen-netfront.c
1313
i = queue->rx.rsp_cons;
drivers/net/xen-netfront.c
1325
xennet_set_rx_rsp_cons(queue,
drivers/net/xen-netfront.c
1326
queue->rx.rsp_cons +
drivers/net/xen-netfront.c
1341
if (unlikely(xennet_fill_frags(queue, skb, &tmpq)))
drivers/net/xen-netfront.c
1351
i = queue->rx.rsp_cons + 1;
drivers/net/xen-netfront.c
1352
xennet_set_rx_rsp_cons(queue, i);
drivers/net/xen-netfront.c
1360
work_done -= handle_incoming_queue(queue, &rxq);
drivers/net/xen-netfront.c
1362
xennet_alloc_rx_buffers(queue);
drivers/net/xen-netfront.c
1369
RING_FINAL_CHECK_FOR_RESPONSES(&queue->rx, more_to_do);
drivers/net/xen-netfront.c
1374
spin_unlock(&queue->rx_lock);
drivers/net/xen-netfront.c
1423
static void xennet_release_tx_bufs(struct netfront_queue *queue)
drivers/net/xen-netfront.c
1430
if (!queue->tx_skbs[i])
drivers/net/xen-netfront.c
1433
skb = queue->tx_skbs[i];
drivers/net/xen-netfront.c
1434
queue->tx_skbs[i] = NULL;
drivers/net/xen-netfront.c
1435
get_page(queue->grant_tx_page[i]);
drivers/net/xen-netfront.c
1436
gnttab_end_foreign_access(queue->grant_tx_ref[i],
drivers/net/xen-netfront.c
1437
queue->grant_tx_page[i]);
drivers/net/xen-netfront.c
1438
queue->grant_tx_page[i] = NULL;
drivers/net/xen-netfront.c
1439
queue->grant_tx_ref[i] = INVALID_GRANT_REF;
drivers/net/xen-netfront.c
1440
add_id_to_list(&queue->tx_skb_freelist, queue->tx_link, i);
drivers/net/xen-netfront.c
1445
static void xennet_release_rx_bufs(struct netfront_queue *queue)
drivers/net/xen-netfront.c
1449
spin_lock_bh(&queue->rx_lock);
drivers/net/xen-netfront.c
1455
skb = queue->rx_skbs[id];
drivers/net/xen-netfront.c
1459
ref = queue->grant_rx_ref[id];
drivers/net/xen-netfront.c
1470
queue->grant_rx_ref[id] = INVALID_GRANT_REF;
drivers/net/xen-netfront.c
1475
spin_unlock_bh(&queue->rx_lock);
drivers/net/xen-netfront.c
1514
static bool xennet_handle_tx(struct netfront_queue *queue, unsigned int *eoi)
drivers/net/xen-netfront.c
1518
if (unlikely(queue->info->broken))
drivers/net/xen-netfront.c
1521
spin_lock_irqsave(&queue->tx_lock, flags);
drivers/net/xen-netfront.c
1522
if (xennet_tx_buf_gc(queue))
drivers/net/xen-netfront.c
1524
spin_unlock_irqrestore(&queue->tx_lock, flags);
drivers/net/xen-netfront.c
1539
static bool xennet_handle_rx(struct netfront_queue *queue, unsigned int *eoi)
drivers/net/xen-netfront.c
1544
if (unlikely(queue->info->broken))
drivers/net/xen-netfront.c
1547
spin_lock_irqsave(&queue->rx_cons_lock, flags);
drivers/net/xen-netfront.c
1548
work_queued = XEN_RING_NR_UNCONSUMED_RESPONSES(&queue->rx);
drivers/net/xen-netfront.c
1549
if (work_queued > queue->rx_rsp_unconsumed) {
drivers/net/xen-netfront.c
1550
queue->rx_rsp_unconsumed = work_queued;
drivers/net/xen-netfront.c
1552
} else if (unlikely(work_queued < queue->rx_rsp_unconsumed)) {
drivers/net/xen-netfront.c
1553
const struct device *dev = &queue->info->netdev->dev;
drivers/net/xen-netfront.c
1555
spin_unlock_irqrestore(&queue->rx_cons_lock, flags);
drivers/net/xen-netfront.c
1558
queue->info->broken = true;
drivers/net/xen-netfront.c
1561
spin_unlock_irqrestore(&queue->rx_cons_lock, flags);
drivers/net/xen-netfront.c
1563
if (likely(netif_carrier_ok(queue->info->netdev) && work_queued))
drivers/net/xen-netfront.c
1564
napi_schedule(&queue->napi);
drivers/net/xen-netfront.c
1823
struct netfront_queue *queue = &info->queues[i];
drivers/net/xen-netfront.c
1825
timer_delete_sync(&queue->rx_refill_timer);
drivers/net/xen-netfront.c
1827
if (queue->tx_irq && (queue->tx_irq == queue->rx_irq))
drivers/net/xen-netfront.c
1828
unbind_from_irqhandler(queue->tx_irq, queue);
drivers/net/xen-netfront.c
1829
if (queue->tx_irq && (queue->tx_irq != queue->rx_irq)) {
drivers/net/xen-netfront.c
1830
unbind_from_irqhandler(queue->tx_irq, queue);
drivers/net/xen-netfront.c
1831
unbind_from_irqhandler(queue->rx_irq, queue);
drivers/net/xen-netfront.c
1833
queue->tx_evtchn = queue->rx_evtchn = 0;
drivers/net/xen-netfront.c
1834
queue->tx_irq = queue->rx_irq = 0;
drivers/net/xen-netfront.c
1837
napi_synchronize(&queue->napi);
drivers/net/xen-netfront.c
1839
xennet_release_tx_bufs(queue);
drivers/net/xen-netfront.c
1840
xennet_release_rx_bufs(queue);
drivers/net/xen-netfront.c
1841
gnttab_free_grant_references(queue->gref_tx_head);
drivers/net/xen-netfront.c
1842
gnttab_free_grant_references(queue->gref_rx_head);
drivers/net/xen-netfront.c
1845
xennet_end_access(queue->tx_ring_ref, queue->tx.sring);
drivers/net/xen-netfront.c
1846
xennet_end_access(queue->rx_ring_ref, queue->rx.sring);
drivers/net/xen-netfront.c
1848
queue->tx_ring_ref = INVALID_GRANT_REF;
drivers/net/xen-netfront.c
1849
queue->rx_ring_ref = INVALID_GRANT_REF;
drivers/net/xen-netfront.c
1850
queue->tx.sring = NULL;
drivers/net/xen-netfront.c
1851
queue->rx.sring = NULL;
drivers/net/xen-netfront.c
1853
page_pool_destroy(queue->page_pool);
drivers/net/xen-netfront.c
1905
static int setup_netfront_single(struct netfront_queue *queue)
drivers/net/xen-netfront.c
1909
err = xenbus_alloc_evtchn(queue->info->xbdev, &queue->tx_evtchn);
drivers/net/xen-netfront.c
1913
err = bind_evtchn_to_irqhandler_lateeoi(queue->tx_evtchn,
drivers/net/xen-netfront.c
1915
queue->info->netdev->name,
drivers/net/xen-netfront.c
1916
queue);
drivers/net/xen-netfront.c
1919
queue->rx_evtchn = queue->tx_evtchn;
drivers/net/xen-netfront.c
1920
queue->rx_irq = queue->tx_irq = err;
drivers/net/xen-netfront.c
1925
xenbus_free_evtchn(queue->info->xbdev, queue->tx_evtchn);
drivers/net/xen-netfront.c
1926
queue->tx_evtchn = 0;
drivers/net/xen-netfront.c
1931
static int setup_netfront_split(struct netfront_queue *queue)
drivers/net/xen-netfront.c
1935
err = xenbus_alloc_evtchn(queue->info->xbdev, &queue->tx_evtchn);
drivers/net/xen-netfront.c
1938
err = xenbus_alloc_evtchn(queue->info->xbdev, &queue->rx_evtchn);
drivers/net/xen-netfront.c
1942
snprintf(queue->tx_irq_name, sizeof(queue->tx_irq_name),
drivers/net/xen-netfront.c
1943
"%s-tx", queue->name);
drivers/net/xen-netfront.c
1944
err = bind_evtchn_to_irqhandler_lateeoi(queue->tx_evtchn,
drivers/net/xen-netfront.c
1946
queue->tx_irq_name, queue);
drivers/net/xen-netfront.c
1949
queue->tx_irq = err;
drivers/net/xen-netfront.c
1951
snprintf(queue->rx_irq_name, sizeof(queue->rx_irq_name),
drivers/net/xen-netfront.c
1952
"%s-rx", queue->name);
drivers/net/xen-netfront.c
1953
err = bind_evtchn_to_irqhandler_lateeoi(queue->rx_evtchn,
drivers/net/xen-netfront.c
1955
queue->rx_irq_name, queue);
drivers/net/xen-netfront.c
1958
queue->rx_irq = err;
drivers/net/xen-netfront.c
1963
unbind_from_irqhandler(queue->tx_irq, queue);
drivers/net/xen-netfront.c
1964
queue->tx_irq = 0;
drivers/net/xen-netfront.c
1966
xenbus_free_evtchn(queue->info->xbdev, queue->rx_evtchn);
drivers/net/xen-netfront.c
1967
queue->rx_evtchn = 0;
drivers/net/xen-netfront.c
1969
xenbus_free_evtchn(queue->info->xbdev, queue->tx_evtchn);
drivers/net/xen-netfront.c
1970
queue->tx_evtchn = 0;
drivers/net/xen-netfront.c
1976
struct netfront_queue *queue, unsigned int feature_split_evtchn)
drivers/net/xen-netfront.c
1982
queue->tx_ring_ref = INVALID_GRANT_REF;
drivers/net/xen-netfront.c
1983
queue->rx_ring_ref = INVALID_GRANT_REF;
drivers/net/xen-netfront.c
1984
queue->rx.sring = NULL;
drivers/net/xen-netfront.c
1985
queue->tx.sring = NULL;
drivers/net/xen-netfront.c
1988
1, &queue->tx_ring_ref);
drivers/net/xen-netfront.c
1992
XEN_FRONT_RING_INIT(&queue->tx, txs, XEN_PAGE_SIZE);
drivers/net/xen-netfront.c
1995
1, &queue->rx_ring_ref);
drivers/net/xen-netfront.c
1999
XEN_FRONT_RING_INIT(&queue->rx, rxs, XEN_PAGE_SIZE);
drivers/net/xen-netfront.c
2002
err = setup_netfront_split(queue);
drivers/net/xen-netfront.c
2008
err = setup_netfront_single(queue);
drivers/net/xen-netfront.c
2016
xenbus_teardown_ring((void **)&queue->rx.sring, 1, &queue->rx_ring_ref);
drivers/net/xen-netfront.c
2017
xenbus_teardown_ring((void **)&queue->tx.sring, 1, &queue->tx_ring_ref);
drivers/net/xen-netfront.c
2026
static int xennet_init_queue(struct netfront_queue *queue)
drivers/net/xen-netfront.c
2032
spin_lock_init(&queue->tx_lock);
drivers/net/xen-netfront.c
2033
spin_lock_init(&queue->rx_lock);
drivers/net/xen-netfront.c
2034
spin_lock_init(&queue->rx_cons_lock);
drivers/net/xen-netfront.c
2036
timer_setup(&queue->rx_refill_timer, rx_refill_timeout, 0);
drivers/net/xen-netfront.c
2038
devid = strrchr(queue->info->xbdev->nodename, '/') + 1;
drivers/net/xen-netfront.c
2039
snprintf(queue->name, sizeof(queue->name), "vif%s-q%u",
drivers/net/xen-netfront.c
2040
devid, queue->id);
drivers/net/xen-netfront.c
2043
queue->tx_skb_freelist = 0;
drivers/net/xen-netfront.c
2044
queue->tx_pend_queue = TX_LINK_NONE;
drivers/net/xen-netfront.c
2046
queue->tx_link[i] = i + 1;
drivers/net/xen-netfront.c
2047
queue->grant_tx_ref[i] = INVALID_GRANT_REF;
drivers/net/xen-netfront.c
2048
queue->grant_tx_page[i] = NULL;
drivers/net/xen-netfront.c
2050
queue->tx_link[NET_TX_RING_SIZE - 1] = TX_LINK_NONE;
drivers/net/xen-netfront.c
2054
queue->rx_skbs[i] = NULL;
drivers/net/xen-netfront.c
2055
queue->grant_rx_ref[i] = INVALID_GRANT_REF;
drivers/net/xen-netfront.c
2060
&queue->gref_tx_head) < 0) {
drivers/net/xen-netfront.c
2068
&queue->gref_rx_head) < 0) {
drivers/net/xen-netfront.c
2077
gnttab_free_grant_references(queue->gref_tx_head);
drivers/net/xen-netfront.c
2082
static int write_queue_xenstore_keys(struct netfront_queue *queue,
drivers/net/xen-netfront.c
2089
struct xenbus_device *dev = queue->info->xbdev;
drivers/net/xen-netfront.c
2105
dev->nodename, queue->id);
drivers/net/xen-netfront.c
2112
queue->tx_ring_ref);
drivers/net/xen-netfront.c
2119
queue->rx_ring_ref);
drivers/net/xen-netfront.c
2128
if (queue->tx_evtchn == queue->rx_evtchn) {
drivers/net/xen-netfront.c
2131
"event-channel", "%u", queue->tx_evtchn);
drivers/net/xen-netfront.c
2139
"event-channel-tx", "%u", queue->tx_evtchn);
drivers/net/xen-netfront.c
2146
"event-channel-rx", "%u", queue->rx_evtchn);
drivers/net/xen-netfront.c
2166
static int xennet_create_page_pool(struct netfront_queue *queue)
drivers/net/xen-netfront.c
2174
.dev = &queue->info->netdev->dev,
drivers/net/xen-netfront.c
2179
queue->page_pool = page_pool_create(&pp_params);
drivers/net/xen-netfront.c
218
static struct sk_buff *xennet_get_rx_skb(struct netfront_queue *queue,
drivers/net/xen-netfront.c
2180
if (IS_ERR(queue->page_pool)) {
drivers/net/xen-netfront.c
2181
err = PTR_ERR(queue->page_pool);
drivers/net/xen-netfront.c
2182
queue->page_pool = NULL;
drivers/net/xen-netfront.c
2186
err = xdp_rxq_info_reg(&queue->xdp_rxq, queue->info->netdev,
drivers/net/xen-netfront.c
2187
queue->id, 0);
drivers/net/xen-netfront.c
2189
netdev_err(queue->info->netdev, "xdp_rxq_info_reg failed\n");
drivers/net/xen-netfront.c
2193
err = xdp_rxq_info_reg_mem_model(&queue->xdp_rxq,
drivers/net/xen-netfront.c
2194
MEM_TYPE_PAGE_POOL, queue->page_pool);
drivers/net/xen-netfront.c
2196
netdev_err(queue->info->netdev, "xdp_rxq_info_reg_mem_model failed\n");
drivers/net/xen-netfront.c
2202
xdp_rxq_info_unreg(&queue->xdp_rxq);
drivers/net/xen-netfront.c
2204
page_pool_destroy(queue->page_pool);
drivers/net/xen-netfront.c
2205
queue->page_pool = NULL;
drivers/net/xen-netfront.c
222
struct sk_buff *skb = queue->rx_skbs[i];
drivers/net/xen-netfront.c
2220
struct netfront_queue *queue = &info->queues[i];
drivers/net/xen-netfront.c
2222
queue->id = i;
drivers/net/xen-netfront.c
2223
queue->info = info;
drivers/net/xen-netfront.c
2225
ret = xennet_init_queue(queue);
drivers/net/xen-netfront.c
223
queue->rx_skbs[i] = NULL;
drivers/net/xen-netfront.c
2234
ret = xennet_create_page_pool(queue);
drivers/net/xen-netfront.c
2241
netif_napi_add(queue->info->netdev, &queue->napi, xennet_poll);
drivers/net/xen-netfront.c
2243
napi_enable(&queue->napi);
drivers/net/xen-netfront.c
2265
struct netfront_queue *queue = NULL;
drivers/net/xen-netfront.c
227
static grant_ref_t xennet_get_rx_ref(struct netfront_queue *queue,
drivers/net/xen-netfront.c
231
grant_ref_t ref = queue->grant_rx_ref[i];
drivers/net/xen-netfront.c
232
queue->grant_rx_ref[i] = INVALID_GRANT_REF;
drivers/net/xen-netfront.c
2321
queue = &info->queues[i];
drivers/net/xen-netfront.c
2322
err = setup_netfront(dev, queue, feature_split_evtchn);
drivers/net/xen-netfront.c
2352
queue = &info->queues[i];
drivers/net/xen-netfront.c
2353
err = write_queue_xenstore_keys(queue, &xbt, 1); /* hierarchical */
drivers/net/xen-netfront.c
2429
struct netfront_queue *queue = NULL;
drivers/net/xen-netfront.c
2474
queue = &np->queues[j];
drivers/net/xen-netfront.c
2476
notify_remote_via_irq(queue->tx_irq);
drivers/net/xen-netfront.c
2477
if (queue->tx_irq != queue->rx_irq)
drivers/net/xen-netfront.c
2478
notify_remote_via_irq(queue->rx_irq);
drivers/net/xen-netfront.c
248
struct netfront_queue *queue = timer_container_of(queue, t,
drivers/net/xen-netfront.c
2480
spin_lock_bh(&queue->rx_lock);
drivers/net/xen-netfront.c
2481
xennet_alloc_rx_buffers(queue);
drivers/net/xen-netfront.c
2482
spin_unlock_bh(&queue->rx_lock);
drivers/net/xen-netfront.c
250
napi_schedule(&queue->napi);
drivers/net/xen-netfront.c
253
static int netfront_tx_slot_available(struct netfront_queue *queue)
drivers/net/xen-netfront.c
255
return (queue->tx.req_prod_pvt - queue->tx.rsp_cons) <
drivers/net/xen-netfront.c
259
static void xennet_maybe_wake_tx(struct netfront_queue *queue)
drivers/net/xen-netfront.c
261
struct net_device *dev = queue->info->netdev;
drivers/net/xen-netfront.c
262
struct netdev_queue *dev_queue = netdev_get_tx_queue(dev, queue->id);
drivers/net/xen-netfront.c
265
netfront_tx_slot_available(queue) &&
drivers/net/xen-netfront.c
267
netif_tx_wake_queue(netdev_get_tx_queue(dev, queue->id));
drivers/net/xen-netfront.c
271
static struct sk_buff *xennet_alloc_one_rx_buffer(struct netfront_queue *queue)
drivers/net/xen-netfront.c
276
skb = __netdev_alloc_skb(queue->info->netdev,
drivers/net/xen-netfront.c
282
page = page_pool_alloc_pages(queue->page_pool,
drivers/net/xen-netfront.c
293
skb->dev = queue->info->netdev;
drivers/net/xen-netfront.c
299
static void xennet_alloc_rx_buffers(struct netfront_queue *queue)
drivers/net/xen-netfront.c
301
RING_IDX req_prod = queue->rx.req_prod_pvt;
drivers/net/xen-netfront.c
305
if (unlikely(!netif_carrier_ok(queue->info->netdev)))
drivers/net/xen-netfront.c
308
for (req_prod = queue->rx.req_prod_pvt;
drivers/net/xen-netfront.c
309
req_prod - queue->rx.rsp_cons < NET_RX_RING_SIZE;
drivers/net/xen-netfront.c
317
skb = xennet_alloc_one_rx_buffer(queue);
drivers/net/xen-netfront.c
325
BUG_ON(queue->rx_skbs[id]);
drivers/net/xen-netfront.c
326
queue->rx_skbs[id] = skb;
drivers/net/xen-netfront.c
328
ref = gnttab_claim_grant_reference(&queue->gref_rx_head);
drivers/net/xen-netfront.c
330
queue->grant_rx_ref[id] = ref;
drivers/net/xen-netfront.c
334
req = RING_GET_REQUEST(&queue->rx, req_prod);
drivers/net/xen-netfront.c
336
queue->info->xbdev->otherend_id,
drivers/net/xen-netfront.c
343
queue->rx.req_prod_pvt = req_prod;
drivers/net/xen-netfront.c
350
if (req_prod - queue->rx.rsp_cons < NET_RX_SLOTS_MIN ||
drivers/net/xen-netfront.c
352
mod_timer(&queue->rx_refill_timer, jiffies + (HZ/10));
drivers/net/xen-netfront.c
356
RING_PUSH_REQUESTS_AND_CHECK_NOTIFY(&queue->rx, notify);
drivers/net/xen-netfront.c
358
notify_remote_via_irq(queue->rx_irq);
drivers/net/xen-netfront.c
366
struct netfront_queue *queue = NULL;
drivers/net/xen-netfront.c
372
queue = &np->queues[i];
drivers/net/xen-netfront.c
373
napi_enable(&queue->napi);
drivers/net/xen-netfront.c
375
spin_lock_bh(&queue->rx_lock);
drivers/net/xen-netfront.c
377
xennet_alloc_rx_buffers(queue);
drivers/net/xen-netfront.c
378
queue->rx.sring->rsp_event = queue->rx.rsp_cons + 1;
drivers/net/xen-netfront.c
379
if (RING_HAS_UNCONSUMED_RESPONSES(&queue->rx))
drivers/net/xen-netfront.c
380
napi_schedule(&queue->napi);
drivers/net/xen-netfront.c
382
spin_unlock_bh(&queue->rx_lock);
drivers/net/xen-netfront.c
390
static bool xennet_tx_buf_gc(struct netfront_queue *queue)
drivers/net/xen-netfront.c
397
const struct device *dev = &queue->info->netdev->dev;
drivers/net/xen-netfront.c
399
BUG_ON(!netif_carrier_ok(queue->info->netdev));
drivers/net/xen-netfront.c
402
prod = queue->tx.sring->rsp_prod;
drivers/net/xen-netfront.c
403
if (RING_RESPONSE_PROD_OVERFLOW(&queue->tx, prod)) {
drivers/net/xen-netfront.c
405
prod - queue->tx.rsp_cons);
drivers/net/xen-netfront.c
410
for (cons = queue->tx.rsp_cons; cons != prod; cons++) {
drivers/net/xen-netfront.c
415
RING_COPY_RESPONSE(&queue->tx, cons, &txrsp);
drivers/net/xen-netfront.c
420
if (id >= RING_SIZE(&queue->tx)) {
drivers/net/xen-netfront.c
426
if (queue->tx_link[id] != TX_PENDING) {
drivers/net/xen-netfront.c
432
queue->tx_link[id] = TX_LINK_NONE;
drivers/net/xen-netfront.c
433
skb = queue->tx_skbs[id];
drivers/net/xen-netfront.c
434
queue->tx_skbs[id] = NULL;
drivers/net/xen-netfront.c
436
queue->grant_tx_ref[id]))) {
drivers/net/xen-netfront.c
442
&queue->gref_tx_head, queue->grant_tx_ref[id]);
drivers/net/xen-netfront.c
443
queue->grant_tx_ref[id] = INVALID_GRANT_REF;
drivers/net/xen-netfront.c
444
queue->grant_tx_page[id] = NULL;
drivers/net/xen-netfront.c
445
add_id_to_list(&queue->tx_skb_freelist, queue->tx_link, id);
drivers/net/xen-netfront.c
449
queue->tx.rsp_cons = prod;
drivers/net/xen-netfront.c
451
RING_FINAL_CHECK_FOR_RESPONSES(&queue->tx, more_to_do);
drivers/net/xen-netfront.c
454
xennet_maybe_wake_tx(queue);
drivers/net/xen-netfront.c
459
queue->info->broken = true;
drivers/net/xen-netfront.c
466
struct netfront_queue *queue;
drivers/net/xen-netfront.c
483
struct netfront_queue *queue = info->queue;
drivers/net/xen-netfront.c
486
id = get_id_from_list(&queue->tx_skb_freelist, queue->tx_link);
drivers/net/xen-netfront.c
487
tx = RING_GET_REQUEST(&queue->tx, queue->tx.req_prod_pvt++);
drivers/net/xen-netfront.c
488
ref = gnttab_claim_grant_reference(&queue->gref_tx_head);
drivers/net/xen-netfront.c
491
gnttab_grant_foreign_access_ref(ref, queue->info->xbdev->otherend_id,
drivers/net/xen-netfront.c
494
queue->tx_skbs[id] = skb;
drivers/net/xen-netfront.c
495
queue->grant_tx_page[id] = page;
drivers/net/xen-netfront.c
496
queue->grant_tx_ref[id] = ref;
drivers/net/xen-netfront.c
510
add_id_to_list(&queue->tx_pend_queue, queue->tx_link, id);
drivers/net/xen-netfront.c
604
static void xennet_mark_tx_pending(struct netfront_queue *queue)
drivers/net/xen-netfront.c
608
while ((i = get_id_from_list(&queue->tx_pend_queue, queue->tx_link)) !=
drivers/net/xen-netfront.c
610
queue->tx_link[i] = TX_PENDING;
drivers/net/xen-netfront.c
614
struct netfront_queue *queue,
drivers/net/xen-netfront.c
620
.queue = queue,
drivers/net/xen-netfront.c
630
xennet_mark_tx_pending(queue);
drivers/net/xen-netfront.c
632
RING_PUSH_REQUESTS_AND_CHECK_NOTIFY(&queue->tx, notify);
drivers/net/xen-netfront.c
634
notify_remote_via_irq(queue->tx_irq);
drivers/net/xen-netfront.c
649
struct netfront_queue *queue = NULL;
drivers/net/xen-netfront.c
659
queue = &np->queues[smp_processor_id() % num_queues];
drivers/net/xen-netfront.c
661
spin_lock_irqsave(&queue->tx_lock, irq_flags);
drivers/net/xen-netfront.c
667
if (xennet_xdp_xmit_one(dev, queue, xdpf))
drivers/net/xen-netfront.c
671
spin_unlock_irqrestore(&queue->tx_lock, irq_flags);
drivers/net/xen-netfront.c
718
struct netfront_queue *queue = NULL;
drivers/net/xen-netfront.c
731
queue = &np->queues[queue_index];
drivers/net/xen-netfront.c
773
spin_lock_irqsave(&queue->tx_lock, flags);
drivers/net/xen-netfront.c
778
spin_unlock_irqrestore(&queue->tx_lock, flags);
drivers/net/xen-netfront.c
783
info.queue = queue;
drivers/net/xen-netfront.c
807
RING_GET_REQUEST(&queue->tx, queue->tx.req_prod_pvt++);
drivers/net/xen-netfront.c
839
xennet_mark_tx_pending(queue);
drivers/net/xen-netfront.c
841
RING_PUSH_REQUESTS_AND_CHECK_NOTIFY(&queue->tx, notify);
drivers/net/xen-netfront.c
843
notify_remote_via_irq(queue->tx_irq);
drivers/net/xen-netfront.c
850
if (!netfront_tx_slot_available(queue))
drivers/net/xen-netfront.c
851
netif_tx_stop_queue(netdev_get_tx_queue(dev, queue->id));
drivers/net/xen-netfront.c
853
spin_unlock_irqrestore(&queue->tx_lock, flags);
drivers/net/xen-netfront.c
868
struct netfront_queue *queue;
drivers/net/xen-netfront.c
871
queue = &np->queues[i];
drivers/net/xen-netfront.c
872
napi_disable(&queue->napi);
drivers/net/xen-netfront.c
885
struct netfront_queue *queue = &info->queues[i];
drivers/net/xen-netfront.c
888
napi_disable(&queue->napi);
drivers/net/xen-netfront.c
889
netif_napi_del(&queue->napi);
drivers/net/xen-netfront.c
902
static void xennet_set_rx_rsp_cons(struct netfront_queue *queue, RING_IDX val)
drivers/net/xen-netfront.c
906
spin_lock_irqsave(&queue->rx_cons_lock, flags);
drivers/net/xen-netfront.c
907
queue->rx.rsp_cons = val;
drivers/net/xen-netfront.c
908
queue->rx_rsp_unconsumed = XEN_RING_NR_UNCONSUMED_RESPONSES(&queue->rx);
drivers/net/xen-netfront.c
909
spin_unlock_irqrestore(&queue->rx_cons_lock, flags);
drivers/net/xen-netfront.c
912
static void xennet_move_rx_slot(struct netfront_queue *queue, struct sk_buff *skb,
drivers/net/xen-netfront.c
915
int new = xennet_rxidx(queue->rx.req_prod_pvt);
drivers/net/xen-netfront.c
917
BUG_ON(queue->rx_skbs[new]);
drivers/net/xen-netfront.c
918
queue->rx_skbs[new] = skb;
drivers/net/xen-netfront.c
919
queue->grant_rx_ref[new] = ref;
drivers/net/xen-netfront.c
920
RING_GET_REQUEST(&queue->rx, queue->rx.req_prod_pvt)->id = new;
drivers/net/xen-netfront.c
921
RING_GET_REQUEST(&queue->rx, queue->rx.req_prod_pvt)->gref = ref;
drivers/net/xen-netfront.c
922
queue->rx.req_prod_pvt++;
drivers/net/xen-netfront.c
925
static int xennet_get_extras(struct netfront_queue *queue,
drivers/net/xen-netfront.c
931
struct device *dev = &queue->info->netdev->dev;
drivers/net/xen-netfront.c
932
RING_IDX cons = queue->rx.rsp_cons;
drivers/net/xen-netfront.c
946
RING_COPY_RESPONSE(&queue->rx, ++cons, &extra);
drivers/net/xen-netfront.c
958
skb = xennet_get_rx_skb(queue, cons);
drivers/net/xen-netfront.c
959
ref = xennet_get_rx_ref(queue, cons);
drivers/net/xen-netfront.c
960
xennet_move_rx_slot(queue, skb, ref);
drivers/net/xen-netfront.c
963
xennet_set_rx_rsp_cons(queue, cons);
drivers/net/xen-netfront.c
967
static u32 xennet_run_xdp(struct netfront_queue *queue, struct page *pdata,
drivers/net/xen-netfront.c
977
&queue->xdp_rxq);
drivers/net/xen-netfront.c
986
trace_xdp_exception(queue->info->netdev, prog, act);
drivers/net/xen-netfront.c
990
err = xennet_xdp_xmit(queue->info->netdev, 1, &xdpf, 0);
drivers/net/xen-netfront.c
993
trace_xdp_exception(queue->info->netdev, prog, act);
drivers/net/xen-netfront.c
999
err = xdp_do_redirect(queue->info->netdev, xdp, prog);
drivers/nfc/pn533/pn533.c
2803
list_for_each_entry_safe(cmd, n, &priv->cmd_queue, queue) {
drivers/nfc/pn533/pn533.c
2804
list_del(&cmd->queue);
drivers/nfc/pn533/pn533.c
475
INIT_LIST_HEAD(&cmd->queue);
drivers/nfc/pn533/pn533.c
476
list_add_tail(&cmd->queue, &dev->cmd_queue);
drivers/nfc/pn533/pn533.c
566
cmd = list_first_entry(&dev->cmd_queue, struct pn533_cmd, queue);
drivers/nfc/pn533/pn533.c
568
list_del(&cmd->queue);
drivers/nfc/pn533/pn533.h
187
struct list_head queue;
drivers/nvdimm/btt.c
1447
do_acct = blk_queue_io_stat(bio->bi_bdev->bd_disk->queue);
drivers/nvdimm/pmem.c
214
do_acct = blk_queue_io_stat(bio->bi_bdev->bd_disk->queue);
drivers/nvme/host/apple.c
768
struct nvme_ns *ns = hctx->queue->queuedata;
drivers/nvme/host/core.c
2159
if (blk_queue_is_zoned(ns->disk->queue)) {
drivers/nvme/host/core.c
2176
lim = queue_limits_start_update(ns->disk->queue);
drivers/nvme/host/core.c
2179
memflags = blk_mq_freeze_queue(ns->disk->queue);
drivers/nvme/host/core.c
2180
ret = queue_limits_commit_update(ns->disk->queue, &lim);
drivers/nvme/host/core.c
2182
blk_mq_unfreeze_queue(ns->disk->queue, memflags);
drivers/nvme/host/core.c
2304
ret = nvme_submit_sync_cmd(ns->queue, &c, ruhs, size);
drivers/nvme/host/core.c
2375
lim = queue_limits_start_update(ns->disk->queue);
drivers/nvme/host/core.c
2377
memflags = blk_mq_freeze_queue(ns->disk->queue);
drivers/nvme/host/core.c
2426
ret = queue_limits_commit_update(ns->disk->queue, &lim);
drivers/nvme/host/core.c
2428
blk_mq_unfreeze_queue(ns->disk->queue, memflags);
drivers/nvme/host/core.c
2435
blk_mq_unfreeze_queue(ns->disk->queue, memflags);
drivers/nvme/host/core.c
2437
if (blk_queue_is_zoned(ns->queue)) {
drivers/nvme/host/core.c
2489
struct queue_limits *ns_lim = &ns->disk->queue->limits;
drivers/nvme/host/core.c
2493
lim = queue_limits_start_update(ns->head->disk->queue);
drivers/nvme/host/core.c
2494
memflags = blk_mq_freeze_queue(ns->head->disk->queue);
drivers/nvme/host/core.c
2522
ret = queue_limits_commit_update(ns->head->disk->queue, &lim);
drivers/nvme/host/core.c
2528
blk_mq_unfreeze_queue(ns->head->disk->queue, memflags);
drivers/nvme/host/core.c
4129
ns->queue = disk->queue;
drivers/nvme/host/core.c
5208
blk_mq_unfreeze_queue_non_owner(ns->queue);
drivers/nvme/host/core.c
5222
timeout = blk_mq_freeze_queue_wait_timeout(ns->queue, timeout);
drivers/nvme/host/core.c
5239
blk_mq_freeze_queue_wait(ns->queue);
drivers/nvme/host/core.c
5258
blk_freeze_queue_start_non_owner(ns->queue);
drivers/nvme/host/core.c
5307
blk_sync_queue(ns->queue);
drivers/nvme/host/fc.c
100
struct nvme_fc_queue *queue;
drivers/nvme/host/fc.c
1131
struct nvme_fc_queue *queue, u16 qsize, u16 ersp_ratio)
drivers/nvme/host/fc.c
1179
lsop->queue = queue;
drivers/nvme/host/fc.c
1225
queue->qnum, validation_errors[fcret]);
drivers/nvme/host/fc.c
1230
queue->connection_id =
drivers/nvme/host/fc.c
1232
set_bit(NVME_FC_Q_CONNECTED, &queue->flags);
drivers/nvme/host/fc.c
1242
queue->qnum, ret);
drivers/nvme/host/fc.c
1247
nvme_fc_connect_queue(struct nvme_fc_ctrl *ctrl, struct nvme_fc_queue *queue,
drivers/nvme/host/fc.c
1291
conn_rqst->connect_cmd.qid = cpu_to_be16(queue->qnum);
drivers/nvme/host/fc.c
1294
lsop->queue = queue;
drivers/nvme/host/fc.c
1331
queue->qnum, validation_errors[fcret]);
drivers/nvme/host/fc.c
1333
queue->connection_id =
drivers/nvme/host/fc.c
1335
set_bit(NVME_FC_Q_CONNECTED, &queue->flags);
drivers/nvme/host/fc.c
1344
queue->qnum, ret);
drivers/nvme/host/fc.c
1831
op->queue->lldd_handle,
drivers/nvme/host/fc.c
1902
struct nvme_fc_queue *queue = op->queue;
drivers/nvme/host/fc.c
2039
nvme_complete_async_event(&queue->ctrl->ctrl, status, &result);
drivers/nvme/host/fc.c
2059
struct nvme_fc_queue *queue, struct nvme_fc_fcp_op *op,
drivers/nvme/host/fc.c
2074
op->queue = queue;
drivers/nvme/host/fc.c
2081
if (queue->qnum)
drivers/nvme/host/fc.c
2117
struct nvme_fc_queue *queue = &ctrl->queues[queue_idx];
drivers/nvme/host/fc.c
2120
res = __nvme_fc_init_request(ctrl, queue, &op->op, rq, queue->rqcnt++);
drivers/nvme/host/fc.c
2189
struct nvme_fc_queue *queue = &ctrl->queues[qidx];
drivers/nvme/host/fc.c
2191
hctx->driver_data = queue;
drivers/nvme/host/fc.c
2192
queue->hctx = hctx;
drivers/nvme/host/fc.c
2212
struct nvme_fc_queue *queue;
drivers/nvme/host/fc.c
2214
queue = &ctrl->queues[idx];
drivers/nvme/host/fc.c
2215
memset(queue, 0, sizeof(*queue));
drivers/nvme/host/fc.c
2216
queue->ctrl = ctrl;
drivers/nvme/host/fc.c
2217
queue->qnum = idx;
drivers/nvme/host/fc.c
2218
atomic_set(&queue->csn, 0);
drivers/nvme/host/fc.c
2219
queue->dev = ctrl->dev;
drivers/nvme/host/fc.c
2222
queue->cmnd_capsule_len = ctrl->ctrl.ioccsz * 16;
drivers/nvme/host/fc.c
2224
queue->cmnd_capsule_len = sizeof(struct nvme_command);
drivers/nvme/host/fc.c
2247
nvme_fc_free_queue(struct nvme_fc_queue *queue)
drivers/nvme/host/fc.c
2249
if (!test_and_clear_bit(NVME_FC_Q_CONNECTED, &queue->flags))
drivers/nvme/host/fc.c
2252
clear_bit(NVME_FC_Q_LIVE, &queue->flags);
drivers/nvme/host/fc.c
2259
queue->connection_id = 0;
drivers/nvme/host/fc.c
2260
atomic_set(&queue->csn, 0);
drivers/nvme/host/fc.c
2265
struct nvme_fc_queue *queue, unsigned int qidx)
drivers/nvme/host/fc.c
2269
queue->lldd_handle);
drivers/nvme/host/fc.c
2270
queue->lldd_handle = NULL;
drivers/nvme/host/fc.c
2284
struct nvme_fc_queue *queue, unsigned int qidx, u16 qsize)
drivers/nvme/host/fc.c
2288
queue->lldd_handle = NULL;
drivers/nvme/host/fc.c
2291
qidx, qsize, &queue->lldd_handle);
drivers/nvme/host/fc.c
2299
struct nvme_fc_queue *queue = &ctrl->queues[ctrl->ctrl.queue_count - 1];
drivers/nvme/host/fc.c
2302
for (i = ctrl->ctrl.queue_count - 1; i >= 1; i--, queue--)
drivers/nvme/host/fc.c
2303
__nvme_fc_delete_hw_queue(ctrl, queue, i);
drivers/nvme/host/fc.c
2309
struct nvme_fc_queue *queue = &ctrl->queues[1];
drivers/nvme/host/fc.c
2312
for (i = 1; i < ctrl->ctrl.queue_count; i++, queue++) {
drivers/nvme/host/fc.c
2313
ret = __nvme_fc_create_hw_queue(ctrl, queue, i, qsize);
drivers/nvme/host/fc.c
2535
u16 qnum = op->queue->qnum;
drivers/nvme/host/fc.c
2636
nvme_fc_start_fcp_op(struct nvme_fc_ctrl *ctrl, struct nvme_fc_queue *queue,
drivers/nvme/host/fc.c
2655
cmdiu->connection_id = cpu_to_be64(queue->connection_id);
drivers/nvme/host/fc.c
2673
op->fcp_req.sqid = cpu_to_le16(queue->qnum);
drivers/nvme/host/fc.c
2713
cmdiu->csn = cpu_to_be32(atomic_inc_return(&queue->csn));
drivers/nvme/host/fc.c
2716
queue->lldd_handle, &op->fcp_req);
drivers/nvme/host/fc.c
2755
struct nvme_ns *ns = hctx->queue->queuedata;
drivers/nvme/host/fc.c
2756
struct nvme_fc_queue *queue = hctx->driver_data;
drivers/nvme/host/fc.c
2757
struct nvme_fc_ctrl *ctrl = queue->ctrl;
drivers/nvme/host/fc.c
2761
bool queue_ready = test_bit(NVME_FC_Q_LIVE, &queue->flags);
drivers/nvme/host/fc.c
2766
!nvme_check_ready(&queue->ctrl->ctrl, rq, queue_ready))
drivers/nvme/host/fc.c
2767
return nvme_fail_nonready_command(&queue->ctrl->ctrl, rq);
drivers/nvme/host/fc.c
2791
return nvme_fc_start_fcp_op(ctrl, queue, op, data_len, io_dir);
drivers/nvme/host/fc.c
2806
ret = nvme_fc_start_fcp_op(ctrl, aen_op->queue, aen_op, 0,
drivers/nvme/host/fc.c
59
struct nvme_fc_queue *queue;
drivers/nvme/host/ioctl.c
264
return nvme_submit_user_cmd(ns->queue, &c, io.addr, length, metadata,
drivers/nvme/host/ioctl.c
317
status = nvme_submit_user_cmd(ns ? ns->queue : ctrl->admin_q, &c,
drivers/nvme/host/ioctl.c
364
status = nvme_submit_user_cmd(ns ? ns->queue : ctrl->admin_q, &c,
drivers/nvme/host/ioctl.c
452
struct request_queue *q = ns ? ns->queue : ctrl->admin_q;
drivers/nvme/host/multipath.c
114
blk_mq_unfreeze_queue_nomemrestore(h->disk->queue);
drivers/nvme/host/multipath.c
124
blk_mq_freeze_queue_wait(h->disk->queue);
drivers/nvme/host/multipath.c
1288
if (blk_queue_is_zoned(ns->queue) && ns->head->disk)
drivers/nvme/host/multipath.c
134
blk_freeze_queue_start(h->disk->queue);
drivers/nvme/host/multipath.c
191
if (!blk_queue_io_stat(disk->queue) || blk_rq_is_passthrough(rq) ||
drivers/nvme/host/nvme.h
1049
trace_block_bio_complete(ns->head->disk->queue, req->bio);
drivers/nvme/host/nvme.h
587
struct request_queue *queue;
drivers/nvme/host/pr.c
62
ret = nvme_submit_sync_cmd(ns->queue, c, data, data_len);
drivers/nvme/host/pr.c
72
return nvme_submit_sync_cmd(ns->queue, c, data, data_len);
drivers/nvme/host/rdma.c
1170
struct nvme_rdma_queue *queue = wc->qp->qp_context;
drivers/nvme/host/rdma.c
1171
struct nvme_rdma_ctrl *ctrl = queue->ctrl;
drivers/nvme/host/rdma.c
1198
static int nvme_rdma_inv_rkey(struct nvme_rdma_queue *queue,
drivers/nvme/host/rdma.c
1212
return ib_post_send(queue->qp, &wr, NULL);
drivers/nvme/host/rdma.c
1231
static void nvme_rdma_unmap_data(struct nvme_rdma_queue *queue,
drivers/nvme/host/rdma.c
1235
struct nvme_rdma_device *dev = queue->device;
drivers/nvme/host/rdma.c
1237
struct list_head *pool = &queue->qp->rdma_mrs;
drivers/nvme/host/rdma.c
1243
pool = &queue->qp->sig_mrs;
drivers/nvme/host/rdma.c
1246
ib_mr_pool_put(queue->qp, pool, req->mr);
drivers/nvme/host/rdma.c
1264
static int nvme_rdma_map_sg_inline(struct nvme_rdma_queue *queue,
drivers/nvme/host/rdma.c
1277
sge->lkey = queue->device->pd->local_dma_lkey;
drivers/nvme/host/rdma.c
1282
sg->addr = cpu_to_le64(queue->ctrl->ctrl.icdoff);
drivers/nvme/host/rdma.c
1290
static int nvme_rdma_map_sg_single(struct nvme_rdma_queue *queue,
drivers/nvme/host/rdma.c
1297
put_unaligned_le32(queue->device->pd->unsafe_global_rkey, sg->key);
drivers/nvme/host/rdma.c
1302
static int nvme_rdma_map_sg_fr(struct nvme_rdma_queue *queue,
drivers/nvme/host/rdma.c
1309
req->mr = ib_mr_pool_get(queue->qp, &queue->qp->rdma_mrs);
drivers/nvme/host/rdma.c
1320
ib_mr_pool_put(queue->qp, &queue->qp->rdma_mrs, req->mr);
drivers/nvme/host/rdma.c
1406
static int nvme_rdma_map_sg_pi(struct nvme_rdma_queue *queue,
drivers/nvme/host/rdma.c
1420
req->mr = ib_mr_pool_get(queue->qp, &queue->qp->sig_mrs);
drivers/nvme/host/rdma.c
1459
ib_mr_pool_put(queue->qp, &queue->qp->sig_mrs, req->mr);
drivers/nvme/host/rdma.c
1525
static int nvme_rdma_map_data(struct nvme_rdma_queue *queue,
drivers/nvme/host/rdma.c
1529
struct nvme_rdma_device *dev = queue->device;
drivers/nvme/host/rdma.c
1547
ret = nvme_rdma_map_sg_pi(queue, req, c, count, pi_count);
drivers/nvme/host/rdma.c
1552
if (rq_data_dir(rq) == WRITE && nvme_rdma_queue_idx(queue) &&
drivers/nvme/host/rdma.c
1553
queue->ctrl->use_inline_data &&
drivers/nvme/host/rdma.c
1555
nvme_rdma_inline_data_size(queue)) {
drivers/nvme/host/rdma.c
1556
ret = nvme_rdma_map_sg_inline(queue, req, c, count);
drivers/nvme/host/rdma.c
1561
ret = nvme_rdma_map_sg_single(queue, req, c);
drivers/nvme/host/rdma.c
1566
ret = nvme_rdma_map_sg_fr(queue, req, c, count);
drivers/nvme/host/rdma.c
159
static inline int nvme_rdma_queue_idx(struct nvme_rdma_queue *queue)
drivers/nvme/host/rdma.c
1591
static int nvme_rdma_post_send(struct nvme_rdma_queue *queue,
drivers/nvme/host/rdma.c
1600
sge->lkey = queue->device->pd->local_dma_lkey;
drivers/nvme/host/rdma.c
161
return queue - queue->ctrl->queues;
drivers/nvme/host/rdma.c
1614
ret = ib_post_send(queue->qp, first, NULL);
drivers/nvme/host/rdma.c
1616
dev_err(queue->ctrl->ctrl.device,
drivers/nvme/host/rdma.c
1622
static int nvme_rdma_post_recv(struct nvme_rdma_queue *queue,
drivers/nvme/host/rdma.c
1631
list.lkey = queue->device->pd->local_dma_lkey;
drivers/nvme/host/rdma.c
164
static bool nvme_rdma_poll_queue(struct nvme_rdma_queue *queue)
drivers/nvme/host/rdma.c
1640
ret = ib_post_recv(queue->qp, &wr, NULL);
drivers/nvme/host/rdma.c
1642
dev_err(queue->ctrl->ctrl.device,
drivers/nvme/host/rdma.c
1648
static struct blk_mq_tags *nvme_rdma_tagset(struct nvme_rdma_queue *queue)
drivers/nvme/host/rdma.c
1650
u32 queue_idx = nvme_rdma_queue_idx(queue);
drivers/nvme/host/rdma.c
1653
return queue->ctrl->admin_tag_set.tags[queue_idx];
drivers/nvme/host/rdma.c
1654
return queue->ctrl->tag_set.tags[queue_idx - 1];
drivers/nvme/host/rdma.c
166
return nvme_rdma_queue_idx(queue) >
drivers/nvme/host/rdma.c
1666
struct nvme_rdma_queue *queue = &ctrl->queues[0];
drivers/nvme/host/rdma.c
1667
struct ib_device *dev = queue->device->dev;
drivers/nvme/host/rdma.c
167
queue->ctrl->io_queues[HCTX_TYPE_DEFAULT] +
drivers/nvme/host/rdma.c
168
queue->ctrl->io_queues[HCTX_TYPE_READ];
drivers/nvme/host/rdma.c
1686
ret = nvme_rdma_post_send(queue, sqe, &sge, 1, NULL);
drivers/nvme/host/rdma.c
1690
static void nvme_rdma_process_nvme_rsp(struct nvme_rdma_queue *queue,
drivers/nvme/host/rdma.c
1696
rq = nvme_find_rq(nvme_rdma_tagset(queue), cqe->command_id);
drivers/nvme/host/rdma.c
1698
dev_err(queue->ctrl->ctrl.device,
drivers/nvme/host/rdma.c
1700
cqe->command_id, queue->qp->qp_num);
drivers/nvme/host/rdma.c
1701
nvme_rdma_error_recovery(queue->ctrl);
drivers/nvme/host/rdma.c
171
static inline size_t nvme_rdma_inline_data_size(struct nvme_rdma_queue *queue)
drivers/nvme/host/rdma.c
1712
dev_err(queue->ctrl->ctrl.device,
drivers/nvme/host/rdma.c
1715
nvme_rdma_error_recovery(queue->ctrl);
drivers/nvme/host/rdma.c
1720
ret = nvme_rdma_inv_rkey(queue, req);
drivers/nvme/host/rdma.c
1722
dev_err(queue->ctrl->ctrl.device,
drivers/nvme/host/rdma.c
1725
nvme_rdma_error_recovery(queue->ctrl);
drivers/nvme/host/rdma.c
173
return queue->cmnd_capsule_len - sizeof(struct nvme_command);
drivers/nvme/host/rdma.c
1738
struct nvme_rdma_queue *queue = wc->qp->qp_context;
drivers/nvme/host/rdma.c
1739
struct ib_device *ibdev = queue->device->dev;
drivers/nvme/host/rdma.c
1750
dev_err(queue->ctrl->ctrl.device,
drivers/nvme/host/rdma.c
1752
nvme_rdma_error_recovery(queue->ctrl);
drivers/nvme/host/rdma.c
1763
if (unlikely(nvme_is_aen_req(nvme_rdma_queue_idx(queue),
drivers/nvme/host/rdma.c
1765
nvme_complete_async_event(&queue->ctrl->ctrl, cqe->status,
drivers/nvme/host/rdma.c
1768
nvme_rdma_process_nvme_rsp(queue, cqe, wc);
drivers/nvme/host/rdma.c
1771
nvme_rdma_post_recv(queue, qe);
drivers/nvme/host/rdma.c
1774
static int nvme_rdma_conn_established(struct nvme_rdma_queue *queue)
drivers/nvme/host/rdma.c
1778
for (i = 0; i < queue->queue_size; i++) {
drivers/nvme/host/rdma.c
1779
ret = nvme_rdma_post_recv(queue, &queue->rsp_ring[i]);
drivers/nvme/host/rdma.c
1787
static int nvme_rdma_conn_rejected(struct nvme_rdma_queue *queue,
drivers/nvme/host/rdma.c
1790
struct rdma_cm_id *cm_id = queue->cm_id;
drivers/nvme/host/rdma.c
1802
dev_err(queue->ctrl->ctrl.device,
drivers/nvme/host/rdma.c
1806
dev_err(queue->ctrl->ctrl.device,
drivers/nvme/host/rdma.c
1813
static int nvme_rdma_addr_resolved(struct nvme_rdma_queue *queue)
drivers/nvme/host/rdma.c
1815
struct nvme_ctrl *ctrl = &queue->ctrl->ctrl;
drivers/nvme/host/rdma.c
1818
ret = nvme_rdma_create_queue_ib(queue);
drivers/nvme/host/rdma.c
1823
rdma_set_service_type(queue->cm_id, ctrl->opts->tos);
drivers/nvme/host/rdma.c
1824
ret = rdma_resolve_route(queue->cm_id, NVME_RDMA_CM_TIMEOUT_MS);
drivers/nvme/host/rdma.c
1827
queue->cm_error);
drivers/nvme/host/rdma.c
1834
nvme_rdma_destroy_queue_ib(queue);
drivers/nvme/host/rdma.c
1838
static int nvme_rdma_route_resolved(struct nvme_rdma_queue *queue)
drivers/nvme/host/rdma.c
1840
struct nvme_rdma_ctrl *ctrl = queue->ctrl;
drivers/nvme/host/rdma.c
1845
param.qp_num = queue->qp->qp_num;
drivers/nvme/host/rdma.c
1848
param.responder_resources = queue->device->dev->attrs.max_qp_rd_atom;
drivers/nvme/host/rdma.c
1856
priv.qid = cpu_to_le16(nvme_rdma_queue_idx(queue));
drivers/nvme/host/rdma.c
1870
priv.hrqsize = cpu_to_le16(queue->queue_size);
drivers/nvme/host/rdma.c
1871
priv.hsqsize = cpu_to_le16(queue->ctrl->ctrl.sqsize);
drivers/nvme/host/rdma.c
1876
ret = rdma_connect_locked(queue->cm_id, ¶m);
drivers/nvme/host/rdma.c
1889
struct nvme_rdma_queue *queue = cm_id->context;
drivers/nvme/host/rdma.c
1892
dev_dbg(queue->ctrl->ctrl.device, "%s (%d): status %d id %p\n",
drivers/nvme/host/rdma.c
1898
cm_error = nvme_rdma_addr_resolved(queue);
drivers/nvme/host/rdma.c
1901
cm_error = nvme_rdma_route_resolved(queue);
drivers/nvme/host/rdma.c
1904
queue->cm_error = nvme_rdma_conn_established(queue);
drivers/nvme/host/rdma.c
1906
complete(&queue->cm_done);
drivers/nvme/host/rdma.c
1909
cm_error = nvme_rdma_conn_rejected(queue, ev);
drivers/nvme/host/rdma.c
1915
dev_dbg(queue->ctrl->ctrl.device,
drivers/nvme/host/rdma.c
1922
dev_dbg(queue->ctrl->ctrl.device,
drivers/nvme/host/rdma.c
1924
nvme_rdma_error_recovery(queue->ctrl);
drivers/nvme/host/rdma.c
1930
dev_err(queue->ctrl->ctrl.device,
drivers/nvme/host/rdma.c
1932
nvme_rdma_error_recovery(queue->ctrl);
drivers/nvme/host/rdma.c
1937
queue->cm_error = cm_error;
drivers/nvme/host/rdma.c
1938
complete(&queue->cm_done);
drivers/nvme/host/rdma.c
1947
struct nvme_rdma_queue *queue = req->queue;
drivers/nvme/host/rdma.c
1949
nvme_rdma_stop_queue(queue);
drivers/nvme/host/rdma.c
1956
struct nvme_rdma_queue *queue = req->queue;
drivers/nvme/host/rdma.c
1957
struct nvme_rdma_ctrl *ctrl = queue->ctrl;
drivers/nvme/host/rdma.c
1959
int qid = nvme_rdma_queue_idx(queue);
drivers/nvme/host/rdma.c
1995
struct nvme_ns *ns = hctx->queue->queuedata;
drivers/nvme/host/rdma.c
1996
struct nvme_rdma_queue *queue = hctx->driver_data;
drivers/nvme/host/rdma.c
2002
bool queue_ready = test_bit(NVME_RDMA_Q_LIVE, &queue->flags);
drivers/nvme/host/rdma.c
2008
if (!nvme_check_ready(&queue->ctrl->ctrl, rq, queue_ready))
drivers/nvme/host/rdma.c
2009
return nvme_fail_nonready_command(&queue->ctrl->ctrl, rq);
drivers/nvme/host/rdma.c
2011
dev = queue->device->dev;
drivers/nvme/host/rdma.c
2030
queue->pi_support &&
drivers/nvme/host/rdma.c
2038
err = nvme_rdma_map_data(queue, rq, c);
drivers/nvme/host/rdma.c
2040
dev_err(queue->ctrl->ctrl.device,
drivers/nvme/host/rdma.c
2050
err = nvme_rdma_post_send(queue, sqe, req->sge, req->num_sge,
drivers/nvme/host/rdma.c
2058
nvme_rdma_unmap_data(queue, rq);
drivers/nvme/host/rdma.c
2075
struct nvme_rdma_queue *queue = hctx->driver_data;
drivers/nvme/host/rdma.c
2077
return ib_process_cq_direct(queue->ib_cq, -1);
drivers/nvme/host/rdma.c
2114
struct nvme_rdma_queue *queue = req->queue;
drivers/nvme/host/rdma.c
2115
struct ib_device *ibdev = queue->device->dev;
drivers/nvme/host/rdma.c
2120
nvme_rdma_unmap_data(queue, rq);
drivers/nvme/host/rdma.c
246
static int nvme_rdma_wait_for_cm(struct nvme_rdma_queue *queue)
drivers/nvme/host/rdma.c
250
ret = wait_for_completion_interruptible(&queue->cm_done);
drivers/nvme/host/rdma.c
253
WARN_ON_ONCE(queue->cm_error > 0);
drivers/nvme/host/rdma.c
254
return queue->cm_error;
drivers/nvme/host/rdma.c
257
static int nvme_rdma_create_qp(struct nvme_rdma_queue *queue, const int factor)
drivers/nvme/host/rdma.c
259
struct nvme_rdma_device *dev = queue->device;
drivers/nvme/host/rdma.c
266
init_attr.cap.max_send_wr = factor * queue->queue_size + 1;
drivers/nvme/host/rdma.c
268
init_attr.cap.max_recv_wr = queue->queue_size + 1;
drivers/nvme/host/rdma.c
273
init_attr.send_cq = queue->ib_cq;
drivers/nvme/host/rdma.c
274
init_attr.recv_cq = queue->ib_cq;
drivers/nvme/host/rdma.c
275
if (queue->pi_support)
drivers/nvme/host/rdma.c
277
init_attr.qp_context = queue;
drivers/nvme/host/rdma.c
279
ret = rdma_create_qp(queue->cm_id, dev->pd, &init_attr);
drivers/nvme/host/rdma.c
281
queue->qp = queue->cm_id->qp;
drivers/nvme/host/rdma.c
300
struct nvme_rdma_queue *queue = &ctrl->queues[queue_idx];
drivers/nvme/host/rdma.c
308
if (queue->pi_support)
drivers/nvme/host/rdma.c
313
req->queue = queue;
drivers/nvme/host/rdma.c
323
struct nvme_rdma_queue *queue = &ctrl->queues[hctx_idx + 1];
drivers/nvme/host/rdma.c
327
hctx->driver_data = queue;
drivers/nvme/host/rdma.c
335
struct nvme_rdma_queue *queue = &ctrl->queues[0];
drivers/nvme/host/rdma.c
339
hctx->driver_data = queue;
drivers/nvme/host/rdma.c
413
static void nvme_rdma_free_cq(struct nvme_rdma_queue *queue)
drivers/nvme/host/rdma.c
415
if (nvme_rdma_poll_queue(queue))
drivers/nvme/host/rdma.c
416
ib_free_cq(queue->ib_cq);
drivers/nvme/host/rdma.c
418
ib_cq_pool_put(queue->ib_cq, queue->cq_size);
drivers/nvme/host/rdma.c
421
static void nvme_rdma_destroy_queue_ib(struct nvme_rdma_queue *queue)
drivers/nvme/host/rdma.c
426
if (!test_and_clear_bit(NVME_RDMA_Q_TR_READY, &queue->flags))
drivers/nvme/host/rdma.c
429
dev = queue->device;
drivers/nvme/host/rdma.c
432
if (queue->pi_support)
drivers/nvme/host/rdma.c
433
ib_mr_pool_destroy(queue->qp, &queue->qp->sig_mrs);
drivers/nvme/host/rdma.c
434
ib_mr_pool_destroy(queue->qp, &queue->qp->rdma_mrs);
drivers/nvme/host/rdma.c
441
ib_destroy_qp(queue->qp);
drivers/nvme/host/rdma.c
442
nvme_rdma_free_cq(queue);
drivers/nvme/host/rdma.c
444
nvme_rdma_free_ring(ibdev, queue->rsp_ring, queue->queue_size,
drivers/nvme/host/rdma.c
463
struct nvme_rdma_queue *queue)
drivers/nvme/host/rdma.c
465
int ret, comp_vector, idx = nvme_rdma_queue_idx(queue);
drivers/nvme/host/rdma.c
474
if (nvme_rdma_poll_queue(queue))
drivers/nvme/host/rdma.c
475
queue->ib_cq = ib_alloc_cq(ibdev, queue, queue->cq_size,
drivers/nvme/host/rdma.c
478
queue->ib_cq = ib_cq_pool_get(ibdev, queue->cq_size,
drivers/nvme/host/rdma.c
481
if (IS_ERR(queue->ib_cq)) {
drivers/nvme/host/rdma.c
482
ret = PTR_ERR(queue->ib_cq);
drivers/nvme/host/rdma.c
489
static int nvme_rdma_create_queue_ib(struct nvme_rdma_queue *queue)
drivers/nvme/host/rdma.c
496
queue->device = nvme_rdma_find_get_device(queue->cm_id);
drivers/nvme/host/rdma.c
497
if (!queue->device) {
drivers/nvme/host/rdma.c
498
dev_err(queue->cm_id->device->dev.parent,
drivers/nvme/host/rdma.c
502
ibdev = queue->device->dev;
drivers/nvme/host/rdma.c
505
queue->cq_size = cq_factor * queue->queue_size + 1;
drivers/nvme/host/rdma.c
507
ret = nvme_rdma_create_cq(ibdev, queue);
drivers/nvme/host/rdma.c
511
ret = nvme_rdma_create_qp(queue, send_wr_factor);
drivers/nvme/host/rdma.c
515
queue->rsp_ring = nvme_rdma_alloc_ring(ibdev, queue->queue_size,
drivers/nvme/host/rdma.c
517
if (!queue->rsp_ring) {
drivers/nvme/host/rdma.c
527
pages_per_mr = nvme_rdma_get_max_fr_pages(ibdev, queue->pi_support) + 1;
drivers/nvme/host/rdma.c
528
ret = ib_mr_pool_init(queue->qp, &queue->qp->rdma_mrs,
drivers/nvme/host/rdma.c
529
queue->queue_size,
drivers/nvme/host/rdma.c
533
dev_err(queue->ctrl->ctrl.device,
drivers/nvme/host/rdma.c
535
queue->queue_size, nvme_rdma_queue_idx(queue));
drivers/nvme/host/rdma.c
539
if (queue->pi_support) {
drivers/nvme/host/rdma.c
540
ret = ib_mr_pool_init(queue->qp, &queue->qp->sig_mrs,
drivers/nvme/host/rdma.c
541
queue->queue_size, IB_MR_TYPE_INTEGRITY,
drivers/nvme/host/rdma.c
544
dev_err(queue->ctrl->ctrl.device,
drivers/nvme/host/rdma.c
546
queue->queue_size, nvme_rdma_queue_idx(queue));
drivers/nvme/host/rdma.c
551
set_bit(NVME_RDMA_Q_TR_READY, &queue->flags);
drivers/nvme/host/rdma.c
556
ib_mr_pool_destroy(queue->qp, &queue->qp->rdma_mrs);
drivers/nvme/host/rdma.c
558
nvme_rdma_free_ring(ibdev, queue->rsp_ring, queue->queue_size,
drivers/nvme/host/rdma.c
561
rdma_destroy_qp(queue->cm_id);
drivers/nvme/host/rdma.c
563
nvme_rdma_free_cq(queue);
drivers/nvme/host/rdma.c
565
nvme_rdma_dev_put(queue->device);
drivers/nvme/host/rdma.c
572
struct nvme_rdma_queue *queue;
drivers/nvme/host/rdma.c
576
queue = &ctrl->queues[idx];
drivers/nvme/host/rdma.c
577
mutex_init(&queue->queue_lock);
drivers/nvme/host/rdma.c
578
queue->ctrl = ctrl;
drivers/nvme/host/rdma.c
580
queue->pi_support = true;
drivers/nvme/host/rdma.c
582
queue->pi_support = false;
drivers/nvme/host/rdma.c
583
init_completion(&queue->cm_done);
drivers/nvme/host/rdma.c
586
queue->cmnd_capsule_len = ctrl->ctrl.ioccsz * 16;
drivers/nvme/host/rdma.c
588
queue->cmnd_capsule_len = sizeof(struct nvme_command);
drivers/nvme/host/rdma.c
590
queue->queue_size = queue_size;
drivers/nvme/host/rdma.c
592
queue->cm_id = rdma_create_id(&init_net, nvme_rdma_cm_handler, queue,
drivers/nvme/host/rdma.c
594
if (IS_ERR(queue->cm_id)) {
drivers/nvme/host/rdma.c
596
"failed to create CM ID: %ld\n", PTR_ERR(queue->cm_id));
drivers/nvme/host/rdma.c
597
ret = PTR_ERR(queue->cm_id);
drivers/nvme/host/rdma.c
604
queue->cm_error = -ETIMEDOUT;
drivers/nvme/host/rdma.c
605
ret = rdma_resolve_addr(queue->cm_id, src_addr,
drivers/nvme/host/rdma.c
614
ret = nvme_rdma_wait_for_cm(queue);
drivers/nvme/host/rdma.c
621
set_bit(NVME_RDMA_Q_ALLOCATED, &queue->flags);
drivers/nvme/host/rdma.c
626
rdma_destroy_id(queue->cm_id);
drivers/nvme/host/rdma.c
627
nvme_rdma_destroy_queue_ib(queue);
drivers/nvme/host/rdma.c
629
mutex_destroy(&queue->queue_lock);
drivers/nvme/host/rdma.c
633
static void __nvme_rdma_stop_queue(struct nvme_rdma_queue *queue)
drivers/nvme/host/rdma.c
635
rdma_disconnect(queue->cm_id);
drivers/nvme/host/rdma.c
636
ib_drain_qp(queue->qp);
drivers/nvme/host/rdma.c
639
static void nvme_rdma_stop_queue(struct nvme_rdma_queue *queue)
drivers/nvme/host/rdma.c
641
if (!test_bit(NVME_RDMA_Q_ALLOCATED, &queue->flags))
drivers/nvme/host/rdma.c
644
mutex_lock(&queue->queue_lock);
drivers/nvme/host/rdma.c
645
if (test_and_clear_bit(NVME_RDMA_Q_LIVE, &queue->flags))
drivers/nvme/host/rdma.c
646
__nvme_rdma_stop_queue(queue);
drivers/nvme/host/rdma.c
647
mutex_unlock(&queue->queue_lock);
drivers/nvme/host/rdma.c
650
static void nvme_rdma_free_queue(struct nvme_rdma_queue *queue)
drivers/nvme/host/rdma.c
652
if (!test_and_clear_bit(NVME_RDMA_Q_ALLOCATED, &queue->flags))
drivers/nvme/host/rdma.c
655
rdma_destroy_id(queue->cm_id);
drivers/nvme/host/rdma.c
656
nvme_rdma_destroy_queue_ib(queue);
drivers/nvme/host/rdma.c
657
mutex_destroy(&queue->queue_lock);
drivers/nvme/host/rdma.c
678
struct nvme_rdma_queue *queue = &ctrl->queues[idx];
drivers/nvme/host/rdma.c
687
set_bit(NVME_RDMA_Q_LIVE, &queue->flags);
drivers/nvme/host/rdma.c
689
if (test_bit(NVME_RDMA_Q_ALLOCATED, &queue->flags))
drivers/nvme/host/rdma.c
690
__nvme_rdma_stop_queue(queue);
drivers/nvme/host/rdma.c
73
struct nvme_rdma_queue *queue;
drivers/nvme/host/tcp.c
1001
if (queue->recv_ddgst != queue->exp_ddgst) {
drivers/nvme/host/tcp.c
1002
struct request *rq = nvme_cid_to_rq(nvme_tcp_tagset(queue),
drivers/nvme/host/tcp.c
1008
dev_err(queue->ctrl->ctrl.device,
drivers/nvme/host/tcp.c
1010
le32_to_cpu(queue->recv_ddgst),
drivers/nvme/host/tcp.c
1011
le32_to_cpu(queue->exp_ddgst));
drivers/nvme/host/tcp.c
1015
struct request *rq = nvme_cid_to_rq(nvme_tcp_tagset(queue),
drivers/nvme/host/tcp.c
1020
queue->nr_cqe++;
drivers/nvme/host/tcp.c
1023
nvme_tcp_init_recv_ctx(queue);
drivers/nvme/host/tcp.c
1030
struct nvme_tcp_queue *queue = desc->arg.data;
drivers/nvme/host/tcp.c
1034
if (unlikely(!queue->rd_enabled))
drivers/nvme/host/tcp.c
1038
switch (nvme_tcp_recv_state(queue)) {
drivers/nvme/host/tcp.c
1040
result = nvme_tcp_recv_pdu(queue, skb, &offset, &len);
drivers/nvme/host/tcp.c
1043
result = nvme_tcp_recv_data(queue, skb, &offset, &len);
drivers/nvme/host/tcp.c
1046
result = nvme_tcp_recv_ddgst(queue, skb, &offset, &len);
drivers/nvme/host/tcp.c
1052
dev_err(queue->ctrl->ctrl.device,
drivers/nvme/host/tcp.c
1054
queue->rd_enabled = false;
drivers/nvme/host/tcp.c
1055
nvme_tcp_error_recovery(&queue->ctrl->ctrl);
drivers/nvme/host/tcp.c
1065
struct nvme_tcp_queue *queue;
drivers/nvme/host/tcp.c
107
struct nvme_tcp_queue *queue;
drivers/nvme/host/tcp.c
1070
queue = sk->sk_user_data;
drivers/nvme/host/tcp.c
1071
if (likely(queue && queue->rd_enabled) &&
drivers/nvme/host/tcp.c
1072
!test_bit(NVME_TCP_Q_POLLING, &queue->flags))
drivers/nvme/host/tcp.c
1073
queue_work_on(queue->io_cpu, nvme_tcp_wq, &queue->io_work);
drivers/nvme/host/tcp.c
1079
struct nvme_tcp_queue *queue;
drivers/nvme/host/tcp.c
1082
queue = sk->sk_user_data;
drivers/nvme/host/tcp.c
1083
if (likely(queue && sk_stream_is_writeable(sk))) {
drivers/nvme/host/tcp.c
1086
if (nvme_tcp_queue_tls(queue))
drivers/nvme/host/tcp.c
1087
queue->write_space(sk);
drivers/nvme/host/tcp.c
1088
queue_work_on(queue->io_cpu, nvme_tcp_wq, &queue->io_work);
drivers/nvme/host/tcp.c
1095
struct nvme_tcp_queue *queue;
drivers/nvme/host/tcp.c
1098
queue = sk->sk_user_data;
drivers/nvme/host/tcp.c
1099
if (!queue)
drivers/nvme/host/tcp.c
1108
nvme_tcp_error_recovery(&queue->ctrl->ctrl);
drivers/nvme/host/tcp.c
1111
dev_info(queue->ctrl->ctrl.device,
drivers/nvme/host/tcp.c
1113
nvme_tcp_queue_id(queue), sk->sk_state);
drivers/nvme/host/tcp.c
1116
queue->state_change(sk);
drivers/nvme/host/tcp.c
1121
static inline void nvme_tcp_done_send_req(struct nvme_tcp_queue *queue)
drivers/nvme/host/tcp.c
1123
queue->request = NULL;
drivers/nvme/host/tcp.c
1131
nvme_complete_async_event(&req->queue->ctrl->ctrl,
drivers/nvme/host/tcp.c
1141
struct nvme_tcp_queue *queue = req->queue;
drivers/nvme/host/tcp.c
1157
if (last && !queue->data_digest && !nvme_tcp_queue_more(queue))
drivers/nvme/host/tcp.c
1167
ret = sock_sendmsg(queue->sock, &msg);
drivers/nvme/host/tcp.c
1171
if (queue->data_digest)
drivers/nvme/host/tcp.c
1172
nvme_tcp_ddgst_update(&queue->snd_crc, page,
drivers/nvme/host/tcp.c
1185
if (queue->data_digest) {
drivers/nvme/host/tcp.c
1187
nvme_tcp_ddgst_final(queue->snd_crc);
drivers/nvme/host/tcp.c
1194
nvme_tcp_done_send_req(queue);
drivers/nvme/host/tcp.c
1204
struct nvme_tcp_queue *queue = req->queue;
drivers/nvme/host/tcp.c
1209
u8 hdgst = nvme_tcp_hdgst_len(queue);
drivers/nvme/host/tcp.c
1213
if (inline_data || nvme_tcp_queue_more(queue))
drivers/nvme/host/tcp.c
1218
if (queue->hdr_digest && !req->offset)
drivers/nvme/host/tcp.c
1223
ret = sock_sendmsg(queue->sock, &msg);
drivers/nvme/host/tcp.c
1231
if (queue->data_digest)
drivers/nvme/host/tcp.c
1232
queue->snd_crc = NVME_TCP_CRC_SEED;
drivers/nvme/host/tcp.c
1234
nvme_tcp_done_send_req(queue);
drivers/nvme/host/tcp.c
1245
struct nvme_tcp_queue *queue = req->queue;
drivers/nvme/host/tcp.c
1249
u8 hdgst = nvme_tcp_hdgst_len(queue);
drivers/nvme/host/tcp.c
1253
if (queue->hdr_digest && !req->offset)
drivers/nvme/host/tcp.c
1261
ret = sock_sendmsg(queue->sock, &msg);
drivers/nvme/host/tcp.c
1268
if (queue->data_digest)
drivers/nvme/host/tcp.c
1269
queue->snd_crc = NVME_TCP_CRC_SEED;
drivers/nvme/host/tcp.c
1279
struct nvme_tcp_queue *queue = req->queue;
drivers/nvme/host/tcp.c
1289
if (nvme_tcp_queue_more(queue))
drivers/nvme/host/tcp.c
1294
ret = kernel_sendmsg(queue->sock, &msg, &iov, 1, iov.iov_len);
drivers/nvme/host/tcp.c
1302
nvme_tcp_done_send_req(queue);
drivers/nvme/host/tcp.c
1310
static int nvme_tcp_try_send(struct nvme_tcp_queue *queue)
drivers/nvme/host/tcp.c
1316
if (!queue->request) {
drivers/nvme/host/tcp.c
1317
queue->request = nvme_tcp_fetch_request(queue);
drivers/nvme/host/tcp.c
1318
if (!queue->request)
drivers/nvme/host/tcp.c
1321
req = queue->request;
drivers/nvme/host/tcp.c
1350
dev_err(queue->ctrl->ctrl.device,
drivers/nvme/host/tcp.c
1352
nvme_tcp_fail_request(queue->request);
drivers/nvme/host/tcp.c
1353
nvme_tcp_done_send_req(queue);
drivers/nvme/host/tcp.c
1360
static int nvme_tcp_try_recv(struct nvme_tcp_queue *queue)
drivers/nvme/host/tcp.c
1362
struct socket *sock = queue->sock;
drivers/nvme/host/tcp.c
1367
rd_desc.arg.data = queue;
drivers/nvme/host/tcp.c
1370
queue->nr_cqe = 0;
drivers/nvme/host/tcp.c
1378
struct nvme_tcp_queue *queue =
drivers/nvme/host/tcp.c
1386
if (mutex_trylock(&queue->send_mutex)) {
drivers/nvme/host/tcp.c
1387
result = nvme_tcp_try_send(queue);
drivers/nvme/host/tcp.c
1388
mutex_unlock(&queue->send_mutex);
drivers/nvme/host/tcp.c
1395
result = nvme_tcp_try_recv(queue);
drivers/nvme/host/tcp.c
1402
if (nvme_tcp_queue_has_pending(queue) &&
drivers/nvme/host/tcp.c
1403
sk_stream_is_writeable(queue->sock->sk))
drivers/nvme/host/tcp.c
1406
if (!pending || !queue->rd_enabled)
drivers/nvme/host/tcp.c
1411
queue_work_on(queue->io_cpu, nvme_tcp_wq, &queue->io_work);
drivers/nvme/host/tcp.c
1423
struct nvme_tcp_queue *queue = &ctrl->queues[0];
drivers/nvme/host/tcp.c
1425
u8 hdgst = nvme_tcp_hdgst_len(queue);
drivers/nvme/host/tcp.c
1427
async->pdu = page_frag_alloc(&queue->pf_cache,
drivers/nvme/host/tcp.c
1433
async->queue = &ctrl->queues[0];
drivers/nvme/host/tcp.c
1440
struct nvme_tcp_queue *queue = &ctrl->queues[qid];
drivers/nvme/host/tcp.c
1443
if (!test_and_clear_bit(NVME_TCP_Q_ALLOCATED, &queue->flags))
drivers/nvme/host/tcp.c
1446
page_frag_cache_drain(&queue->pf_cache);
drivers/nvme/host/tcp.c
1450
fput(queue->sock->file);
drivers/nvme/host/tcp.c
1451
queue->sock = NULL;
drivers/nvme/host/tcp.c
1454
kfree(queue->pdu);
drivers/nvme/host/tcp.c
1455
mutex_destroy(&queue->send_mutex);
drivers/nvme/host/tcp.c
1456
mutex_destroy(&queue->queue_lock);
drivers/nvme/host/tcp.c
1459
static int nvme_tcp_init_connection(struct nvme_tcp_queue *queue)
drivers/nvme/host/tcp.c
1488
if (queue->hdr_digest)
drivers/nvme/host/tcp.c
1490
if (queue->data_digest)
drivers/nvme/host/tcp.c
1495
ret = kernel_sendmsg(queue->sock, &msg, &iov, 1, iov.iov_len);
drivers/nvme/host/tcp.c
1498
nvme_tcp_queue_id(queue), ret);
drivers/nvme/host/tcp.c
1505
if (nvme_tcp_queue_tls(queue)) {
drivers/nvme/host/tcp.c
1510
ret = kernel_recvmsg(queue->sock, &msg, &iov, 1,
drivers/nvme/host/tcp.c
1516
nvme_tcp_queue_id(queue), ret);
drivers/nvme/host/tcp.c
1520
if (nvme_tcp_queue_tls(queue)) {
drivers/nvme/host/tcp.c
1521
ctype = tls_get_record_type(queue->sock->sk,
drivers/nvme/host/tcp.c
1525
nvme_tcp_queue_id(queue), ctype);
drivers/nvme/host/tcp.c
1532
nvme_tcp_queue_id(queue), icresp->hdr.type);
drivers/nvme/host/tcp.c
1538
nvme_tcp_queue_id(queue), icresp->hdr.plen);
drivers/nvme/host/tcp.c
1544
nvme_tcp_queue_id(queue), icresp->pfv);
drivers/nvme/host/tcp.c
1549
if ((queue->data_digest && !ctrl_ddgst) ||
drivers/nvme/host/tcp.c
1550
(!queue->data_digest && ctrl_ddgst)) {
drivers/nvme/host/tcp.c
1552
nvme_tcp_queue_id(queue),
drivers/nvme/host/tcp.c
1553
queue->data_digest ? "enabled" : "disabled",
drivers/nvme/host/tcp.c
1559
if ((queue->hdr_digest && !ctrl_hdgst) ||
drivers/nvme/host/tcp.c
1560
(!queue->hdr_digest && ctrl_hdgst)) {
drivers/nvme/host/tcp.c
1562
nvme_tcp_queue_id(queue),
drivers/nvme/host/tcp.c
1563
queue->hdr_digest ? "enabled" : "disabled",
drivers/nvme/host/tcp.c
1570
nvme_tcp_queue_id(queue), icresp->cpda);
drivers/nvme/host/tcp.c
1577
nvme_tcp_queue_id(queue), maxh2cdata);
drivers/nvme/host/tcp.c
1580
queue->maxh2cdata = maxh2cdata;
drivers/nvme/host/tcp.c
1590
static bool nvme_tcp_admin_queue(struct nvme_tcp_queue *queue)
drivers/nvme/host/tcp.c
1592
return nvme_tcp_queue_id(queue) == 0;
drivers/nvme/host/tcp.c
1595
static bool nvme_tcp_default_queue(struct nvme_tcp_queue *queue)
drivers/nvme/host/tcp.c
1597
struct nvme_tcp_ctrl *ctrl = queue->ctrl;
drivers/nvme/host/tcp.c
1598
int qid = nvme_tcp_queue_id(queue);
drivers/nvme/host/tcp.c
1600
return !nvme_tcp_admin_queue(queue) &&
drivers/nvme/host/tcp.c
1604
static bool nvme_tcp_read_queue(struct nvme_tcp_queue *queue)
drivers/nvme/host/tcp.c
1606
struct nvme_tcp_ctrl *ctrl = queue->ctrl;
drivers/nvme/host/tcp.c
1607
int qid = nvme_tcp_queue_id(queue);
drivers/nvme/host/tcp.c
1609
return !nvme_tcp_admin_queue(queue) &&
drivers/nvme/host/tcp.c
1610
!nvme_tcp_default_queue(queue) &&
drivers/nvme/host/tcp.c
1615
static bool nvme_tcp_poll_queue(struct nvme_tcp_queue *queue)
drivers/nvme/host/tcp.c
1617
struct nvme_tcp_ctrl *ctrl = queue->ctrl;
drivers/nvme/host/tcp.c
1618
int qid = nvme_tcp_queue_id(queue);
drivers/nvme/host/tcp.c
1620
return !nvme_tcp_admin_queue(queue) &&
drivers/nvme/host/tcp.c
1621
!nvme_tcp_default_queue(queue) &&
drivers/nvme/host/tcp.c
1622
!nvme_tcp_read_queue(queue) &&
drivers/nvme/host/tcp.c
1637
static void nvme_tcp_set_queue_io_cpu(struct nvme_tcp_queue *queue)
drivers/nvme/host/tcp.c
1639
struct nvme_tcp_ctrl *ctrl = queue->ctrl;
drivers/nvme/host/tcp.c
1641
int qid = nvme_tcp_queue_id(queue) - 1;
drivers/nvme/host/tcp.c
1648
if (nvme_tcp_default_queue(queue))
drivers/nvme/host/tcp.c
1650
else if (nvme_tcp_read_queue(queue))
drivers/nvme/host/tcp.c
1652
else if (nvme_tcp_poll_queue(queue))
drivers/nvme/host/tcp.c
1671
queue->io_cpu = io_cpu;
drivers/nvme/host/tcp.c
1673
set_bit(NVME_TCP_Q_IO_CPU_SET, &queue->flags);
drivers/nvme/host/tcp.c
1677
qid, queue->io_cpu);
drivers/nvme/host/tcp.c
1682
struct nvme_tcp_queue *queue = data;
drivers/nvme/host/tcp.c
1683
struct nvme_tcp_ctrl *ctrl = queue->ctrl;
drivers/nvme/host/tcp.c
1684
int qid = nvme_tcp_queue_id(queue);
drivers/nvme/host/tcp.c
1691
queue->tls_err = -status;
drivers/nvme/host/tcp.c
1699
queue->tls_err = -ENOKEY;
drivers/nvme/host/tcp.c
1701
queue->tls_enabled = true;
drivers/nvme/host/tcp.c
1705
queue->tls_err = 0;
drivers/nvme/host/tcp.c
1709
complete(&queue->tls_complete);
drivers/nvme/host/tcp.c
1713
struct nvme_tcp_queue *queue,
drivers/nvme/host/tcp.c
1716
int qid = nvme_tcp_queue_id(queue);
drivers/nvme/host/tcp.c
1725
args.ta_sock = queue->sock;
drivers/nvme/host/tcp.c
1727
args.ta_data = queue;
drivers/nvme/host/tcp.c
1734
queue->tls_err = -EOPNOTSUPP;
drivers/nvme/host/tcp.c
1735
init_completion(&queue->tls_complete);
drivers/nvme/host/tcp.c
1742
ret = wait_for_completion_interruptible_timeout(&queue->tls_complete, tmo);
drivers/nvme/host/tcp.c
1750
tls_handshake_cancel(queue->sock->sk);
drivers/nvme/host/tcp.c
1752
if (queue->tls_err) {
drivers/nvme/host/tcp.c
1755
qid, queue->tls_err);
drivers/nvme/host/tcp.c
1760
ret = queue->tls_err;
drivers/nvme/host/tcp.c
1769
struct nvme_tcp_queue *queue = &ctrl->queues[qid];
drivers/nvme/host/tcp.c
1773
mutex_init(&queue->queue_lock);
drivers/nvme/host/tcp.c
1774
queue->ctrl = ctrl;
drivers/nvme/host/tcp.c
1775
init_llist_head(&queue->req_list);
drivers/nvme/host/tcp.c
1776
INIT_LIST_HEAD(&queue->send_list);
drivers/nvme/host/tcp.c
1777
mutex_init(&queue->send_mutex);
drivers/nvme/host/tcp.c
1778
INIT_WORK(&queue->io_work, nvme_tcp_io_work);
drivers/nvme/host/tcp.c
1781
queue->cmnd_capsule_len = nctrl->ioccsz * 16;
drivers/nvme/host/tcp.c
1783
queue->cmnd_capsule_len = sizeof(struct nvme_command) +
drivers/nvme/host/tcp.c
1788
IPPROTO_TCP, &queue->sock);
drivers/nvme/host/tcp.c
1795
sock_file = sock_alloc_file(queue->sock, O_CLOEXEC, NULL);
drivers/nvme/host/tcp.c
1801
sk_net_refcnt_upgrade(queue->sock->sk);
drivers/nvme/host/tcp.c
1802
nvme_tcp_reclassify_socket(queue->sock);
drivers/nvme/host/tcp.c
1805
tcp_sock_set_syncnt(queue->sock->sk, 1);
drivers/nvme/host/tcp.c
1808
tcp_sock_set_nodelay(queue->sock->sk);
drivers/nvme/host/tcp.c
1815
sock_no_linger(queue->sock->sk);
drivers/nvme/host/tcp.c
1818
sock_set_priority(queue->sock->sk, so_priority);
drivers/nvme/host/tcp.c
1822
ip_sock_set_tos(queue->sock->sk, nctrl->opts->tos);
drivers/nvme/host/tcp.c
1825
queue->sock->sk->sk_rcvtimeo = 10 * HZ;
drivers/nvme/host/tcp.c
1827
queue->sock->sk->sk_allocation = GFP_ATOMIC;
drivers/nvme/host/tcp.c
1828
queue->sock->sk->sk_use_task_frag = false;
drivers/nvme/host/tcp.c
1829
queue->io_cpu = WORK_CPU_UNBOUND;
drivers/nvme/host/tcp.c
1830
queue->request = NULL;
drivers/nvme/host/tcp.c
1831
queue->data_remaining = 0;
drivers/nvme/host/tcp.c
1832
queue->ddgst_remaining = 0;
drivers/nvme/host/tcp.c
1833
queue->pdu_remaining = 0;
drivers/nvme/host/tcp.c
1834
queue->pdu_offset = 0;
drivers/nvme/host/tcp.c
1835
sk_set_memalloc(queue->sock->sk);
drivers/nvme/host/tcp.c
1838
ret = kernel_bind(queue->sock, (struct sockaddr_unsized *)&ctrl->src_addr,
drivers/nvme/host/tcp.c
1852
ret = sock_setsockopt(queue->sock, SOL_SOCKET, SO_BINDTODEVICE,
drivers/nvme/host/tcp.c
1862
queue->hdr_digest = nctrl->opts->hdr_digest;
drivers/nvme/host/tcp.c
1863
queue->data_digest = nctrl->opts->data_digest;
drivers/nvme/host/tcp.c
1866
nvme_tcp_hdgst_len(queue);
drivers/nvme/host/tcp.c
1867
queue->pdu = kmalloc(rcv_pdu_size, GFP_KERNEL);
drivers/nvme/host/tcp.c
1868
if (!queue->pdu) {
drivers/nvme/host/tcp.c
1874
nvme_tcp_queue_id(queue));
drivers/nvme/host/tcp.c
1876
ret = kernel_connect(queue->sock, (struct sockaddr_unsized *)&ctrl->addr,
drivers/nvme/host/tcp.c
1886
ret = nvme_tcp_start_tls(nctrl, queue, pskid);
drivers/nvme/host/tcp.c
1891
ret = nvme_tcp_init_connection(queue);
drivers/nvme/host/tcp.c
1895
set_bit(NVME_TCP_Q_ALLOCATED, &queue->flags);
drivers/nvme/host/tcp.c
1900
kernel_sock_shutdown(queue->sock, SHUT_RDWR);
drivers/nvme/host/tcp.c
1902
kfree(queue->pdu);
drivers/nvme/host/tcp.c
1905
fput(queue->sock->file);
drivers/nvme/host/tcp.c
1906
queue->sock = NULL;
drivers/nvme/host/tcp.c
1908
mutex_destroy(&queue->send_mutex);
drivers/nvme/host/tcp.c
1909
mutex_destroy(&queue->queue_lock);
drivers/nvme/host/tcp.c
1913
static void nvme_tcp_restore_sock_ops(struct nvme_tcp_queue *queue)
drivers/nvme/host/tcp.c
1915
struct socket *sock = queue->sock;
drivers/nvme/host/tcp.c
1919
sock->sk->sk_data_ready = queue->data_ready;
drivers/nvme/host/tcp.c
1920
sock->sk->sk_state_change = queue->state_change;
drivers/nvme/host/tcp.c
1921
sock->sk->sk_write_space = queue->write_space;
drivers/nvme/host/tcp.c
1925
static void __nvme_tcp_stop_queue(struct nvme_tcp_queue *queue)
drivers/nvme/host/tcp.c
1927
kernel_sock_shutdown(queue->sock, SHUT_RDWR);
drivers/nvme/host/tcp.c
1928
nvme_tcp_restore_sock_ops(queue);
drivers/nvme/host/tcp.c
1929
cancel_work_sync(&queue->io_work);
drivers/nvme/host/tcp.c
1935
struct nvme_tcp_queue *queue = &ctrl->queues[qid];
drivers/nvme/host/tcp.c
1937
if (!test_bit(NVME_TCP_Q_ALLOCATED, &queue->flags))
drivers/nvme/host/tcp.c
1940
if (test_and_clear_bit(NVME_TCP_Q_IO_CPU_SET, &queue->flags))
drivers/nvme/host/tcp.c
1941
atomic_dec(&nvme_tcp_cpu_queues[queue->io_cpu]);
drivers/nvme/host/tcp.c
1943
mutex_lock(&queue->queue_lock);
drivers/nvme/host/tcp.c
1944
if (test_and_clear_bit(NVME_TCP_Q_LIVE, &queue->flags))
drivers/nvme/host/tcp.c
1945
__nvme_tcp_stop_queue(queue);
drivers/nvme/host/tcp.c
1947
queue->tls_enabled = false;
drivers/nvme/host/tcp.c
1948
mutex_unlock(&queue->queue_lock);
drivers/nvme/host/tcp.c
1954
struct nvme_tcp_queue *queue = &ctrl->queues[qid];
drivers/nvme/host/tcp.c
1958
if (!test_bit(NVME_TCP_Q_ALLOCATED, &queue->flags) ||
drivers/nvme/host/tcp.c
1959
!sk_wmem_alloc_get(queue->sock->sk))
drivers/nvme/host/tcp.c
1976
static void nvme_tcp_setup_sock_ops(struct nvme_tcp_queue *queue)
drivers/nvme/host/tcp.c
1978
write_lock_bh(&queue->sock->sk->sk_callback_lock);
drivers/nvme/host/tcp.c
1979
queue->sock->sk->sk_user_data = queue;
drivers/nvme/host/tcp.c
1980
queue->state_change = queue->sock->sk->sk_state_change;
drivers/nvme/host/tcp.c
1981
queue->data_ready = queue->sock->sk->sk_data_ready;
drivers/nvme/host/tcp.c
1982
queue->write_space = queue->sock->sk->sk_write_space;
drivers/nvme/host/tcp.c
1983
queue->sock->sk->sk_data_ready = nvme_tcp_data_ready;
drivers/nvme/host/tcp.c
1984
queue->sock->sk->sk_state_change = nvme_tcp_state_change;
drivers/nvme/host/tcp.c
1985
queue->sock->sk->sk_write_space = nvme_tcp_write_space;
drivers/nvme/host/tcp.c
1987
queue->sock->sk->sk_ll_usec = 1;
drivers/nvme/host/tcp.c
1989
write_unlock_bh(&queue->sock->sk->sk_callback_lock);
drivers/nvme/host/tcp.c
1995
struct nvme_tcp_queue *queue = &ctrl->queues[idx];
drivers/nvme/host/tcp.c
1998
queue->rd_enabled = true;
drivers/nvme/host/tcp.c
1999
nvme_tcp_init_recv_ctx(queue);
drivers/nvme/host/tcp.c
2000
nvme_tcp_setup_sock_ops(queue);
drivers/nvme/host/tcp.c
2003
nvme_tcp_set_queue_io_cpu(queue);
drivers/nvme/host/tcp.c
2009
set_bit(NVME_TCP_Q_LIVE, &queue->flags);
drivers/nvme/host/tcp.c
2011
if (test_bit(NVME_TCP_Q_ALLOCATED, &queue->flags))
drivers/nvme/host/tcp.c
2012
__nvme_tcp_stop_queue(queue);
drivers/nvme/host/tcp.c
208
static int nvme_tcp_try_send(struct nvme_tcp_queue *queue);
drivers/nvme/host/tcp.c
215
static inline int nvme_tcp_queue_id(struct nvme_tcp_queue *queue)
drivers/nvme/host/tcp.c
217
return queue - queue->ctrl->queues;
drivers/nvme/host/tcp.c
236
static inline bool nvme_tcp_queue_tls(struct nvme_tcp_queue *queue)
drivers/nvme/host/tcp.c
241
return queue->tls_enabled;
drivers/nvme/host/tcp.c
255
static inline struct blk_mq_tags *nvme_tcp_tagset(struct nvme_tcp_queue *queue)
drivers/nvme/host/tcp.c
257
u32 queue_idx = nvme_tcp_queue_id(queue);
drivers/nvme/host/tcp.c
2580
static void nvme_tcp_set_sg_inline(struct nvme_tcp_queue *queue,
drivers/nvme/host/tcp.c
2585
sg->addr = cpu_to_le64(queue->ctrl->ctrl.icdoff);
drivers/nvme/host/tcp.c
260
return queue->ctrl->admin_tag_set.tags[queue_idx];
drivers/nvme/host/tcp.c
2604
struct nvme_tcp_queue *queue = &ctrl->queues[0];
drivers/nvme/host/tcp.c
2607
u8 hdgst = nvme_tcp_hdgst_len(queue);
drivers/nvme/host/tcp.c
261
return queue->ctrl->tag_set.tags[queue_idx - 1];
drivers/nvme/host/tcp.c
2611
if (queue->hdr_digest)
drivers/nvme/host/tcp.c
2634
struct nvme_ctrl *ctrl = &req->queue->ctrl->ctrl;
drivers/nvme/host/tcp.c
2636
nvme_tcp_stop_queue(ctrl, nvme_tcp_queue_id(req->queue));
drivers/nvme/host/tcp.c
264
static inline u8 nvme_tcp_hdgst_len(struct nvme_tcp_queue *queue)
drivers/nvme/host/tcp.c
2643
struct nvme_ctrl *ctrl = &req->queue->ctrl->ctrl;
drivers/nvme/host/tcp.c
2646
int qid = nvme_tcp_queue_id(req->queue);
drivers/nvme/host/tcp.c
266
return queue->hdr_digest ? NVME_TCP_DIGEST_LENGTH : 0;
drivers/nvme/host/tcp.c
2679
static blk_status_t nvme_tcp_map_data(struct nvme_tcp_queue *queue,
drivers/nvme/host/tcp.c
269
static inline u8 nvme_tcp_ddgst_len(struct nvme_tcp_queue *queue)
drivers/nvme/host/tcp.c
2692
nvme_tcp_set_sg_inline(queue, c, req->data_len);
drivers/nvme/host/tcp.c
2704
struct nvme_tcp_queue *queue = req->queue;
drivers/nvme/host/tcp.c
2705
u8 hdgst = nvme_tcp_hdgst_len(queue), ddgst = 0;
drivers/nvme/host/tcp.c
271
return queue->data_digest ? NVME_TCP_DIGEST_LENGTH : 0;
drivers/nvme/host/tcp.c
2731
if (queue->hdr_digest)
drivers/nvme/host/tcp.c
2733
if (queue->data_digest && req->pdu_len) {
drivers/nvme/host/tcp.c
2735
ddgst = nvme_tcp_ddgst_len(queue);
drivers/nvme/host/tcp.c
2742
ret = nvme_tcp_map_data(queue, rq);
drivers/nvme/host/tcp.c
2745
dev_err(queue->ctrl->ctrl.device,
drivers/nvme/host/tcp.c
2755
struct nvme_tcp_queue *queue = hctx->driver_data;
drivers/nvme/host/tcp.c
2757
if (!llist_empty(&queue->req_list))
drivers/nvme/host/tcp.c
2758
queue_work_on(queue->io_cpu, nvme_tcp_wq, &queue->io_work);
drivers/nvme/host/tcp.c
2764
struct nvme_ns *ns = hctx->queue->queuedata;
drivers/nvme/host/tcp.c
2765
struct nvme_tcp_queue *queue = hctx->driver_data;
drivers/nvme/host/tcp.c
2768
bool queue_ready = test_bit(NVME_TCP_Q_LIVE, &queue->flags);
drivers/nvme/host/tcp.c
2771
if (!nvme_check_ready(&queue->ctrl->ctrl, rq, queue_ready))
drivers/nvme/host/tcp.c
2772
return nvme_fail_nonready_command(&queue->ctrl->ctrl, rq);
drivers/nvme/host/tcp.c
2794
struct nvme_tcp_queue *queue = hctx->driver_data;
drivers/nvme/host/tcp.c
2795
struct sock *sk = queue->sock->sk;
drivers/nvme/host/tcp.c
2798
if (!test_bit(NVME_TCP_Q_LIVE, &queue->flags))
drivers/nvme/host/tcp.c
2801
set_bit(NVME_TCP_Q_POLLING, &queue->flags);
drivers/nvme/host/tcp.c
2804
ret = nvme_tcp_try_recv(queue);
drivers/nvme/host/tcp.c
2805
clear_bit(NVME_TCP_Q_POLLING, &queue->flags);
drivers/nvme/host/tcp.c
2806
return ret < 0 ? ret : queue->nr_cqe;
drivers/nvme/host/tcp.c
2811
struct nvme_tcp_queue *queue = &to_tcp_ctrl(ctrl)->queues[0];
drivers/nvme/host/tcp.c
2817
if (!test_bit(NVME_TCP_Q_LIVE, &queue->flags))
drivers/nvme/host/tcp.c
2820
mutex_lock(&queue->queue_lock);
drivers/nvme/host/tcp.c
2822
ret = kernel_getsockname(queue->sock, (struct sockaddr *)&src_addr);
drivers/nvme/host/tcp.c
2830
mutex_unlock(&queue->queue_lock);
drivers/nvme/host/tcp.c
290
return req->queue->cmnd_capsule_len - sizeof(struct nvme_command);
drivers/nvme/host/tcp.c
295
return req == &req->queue->ctrl->async_req;
drivers/nvme/host/tcp.c
384
static inline void nvme_tcp_send_all(struct nvme_tcp_queue *queue)
drivers/nvme/host/tcp.c
390
ret = nvme_tcp_try_send(queue);
drivers/nvme/host/tcp.c
394
static inline bool nvme_tcp_queue_has_pending(struct nvme_tcp_queue *queue)
drivers/nvme/host/tcp.c
396
return !list_empty(&queue->send_list) ||
drivers/nvme/host/tcp.c
397
!llist_empty(&queue->req_list);
drivers/nvme/host/tcp.c
400
static inline bool nvme_tcp_queue_more(struct nvme_tcp_queue *queue)
drivers/nvme/host/tcp.c
402
return !nvme_tcp_queue_tls(queue) &&
drivers/nvme/host/tcp.c
403
nvme_tcp_queue_has_pending(queue);
drivers/nvme/host/tcp.c
409
struct nvme_tcp_queue *queue = req->queue;
drivers/nvme/host/tcp.c
412
empty = llist_add(&req->lentry, &queue->req_list) &&
drivers/nvme/host/tcp.c
413
list_empty(&queue->send_list) && !queue->request;
drivers/nvme/host/tcp.c
420
if (queue->io_cpu == raw_smp_processor_id() &&
drivers/nvme/host/tcp.c
421
empty && mutex_trylock(&queue->send_mutex)) {
drivers/nvme/host/tcp.c
422
nvme_tcp_send_all(queue);
drivers/nvme/host/tcp.c
423
mutex_unlock(&queue->send_mutex);
drivers/nvme/host/tcp.c
426
if (last && nvme_tcp_queue_has_pending(queue))
drivers/nvme/host/tcp.c
427
queue_work_on(queue->io_cpu, nvme_tcp_wq, &queue->io_work);
drivers/nvme/host/tcp.c
430
static void nvme_tcp_process_req_list(struct nvme_tcp_queue *queue)
drivers/nvme/host/tcp.c
435
for (node = llist_del_all(&queue->req_list); node; node = node->next) {
drivers/nvme/host/tcp.c
437
list_add(&req->entry, &queue->send_list);
drivers/nvme/host/tcp.c
442
nvme_tcp_fetch_request(struct nvme_tcp_queue *queue)
drivers/nvme/host/tcp.c
446
req = list_first_entry_or_null(&queue->send_list,
drivers/nvme/host/tcp.c
449
nvme_tcp_process_req_list(queue);
drivers/nvme/host/tcp.c
450
req = list_first_entry_or_null(&queue->send_list,
drivers/nvme/host/tcp.c
495
static int nvme_tcp_verify_hdgst(struct nvme_tcp_queue *queue,
drivers/nvme/host/tcp.c
503
dev_err(queue->ctrl->ctrl.device,
drivers/nvme/host/tcp.c
505
nvme_tcp_queue_id(queue));
drivers/nvme/host/tcp.c
512
dev_err(queue->ctrl->ctrl.device,
drivers/nvme/host/tcp.c
521
static int nvme_tcp_check_ddgst(struct nvme_tcp_queue *queue, void *pdu)
drivers/nvme/host/tcp.c
524
u8 digest_len = nvme_tcp_hdgst_len(queue);
drivers/nvme/host/tcp.c
531
dev_err(queue->ctrl->ctrl.device,
drivers/nvme/host/tcp.c
533
nvme_tcp_queue_id(queue));
drivers/nvme/host/tcp.c
536
queue->rcv_crc = NVME_TCP_CRC_SEED;
drivers/nvme/host/tcp.c
557
struct nvme_tcp_queue *queue = &ctrl->queues[queue_idx];
drivers/nvme/host/tcp.c
558
u8 hdgst = nvme_tcp_hdgst_len(queue);
drivers/nvme/host/tcp.c
560
req->pdu = page_frag_alloc(&queue->pf_cache,
drivers/nvme/host/tcp.c
567
req->queue = queue;
drivers/nvme/host/tcp.c
580
struct nvme_tcp_queue *queue = &ctrl->queues[hctx_idx + 1];
drivers/nvme/host/tcp.c
582
hctx->driver_data = queue;
drivers/nvme/host/tcp.c
590
struct nvme_tcp_queue *queue = &ctrl->queues[0];
drivers/nvme/host/tcp.c
592
hctx->driver_data = queue;
drivers/nvme/host/tcp.c
597
nvme_tcp_recv_state(struct nvme_tcp_queue *queue)
drivers/nvme/host/tcp.c
599
return (queue->pdu_remaining) ? NVME_TCP_RECV_PDU :
drivers/nvme/host/tcp.c
600
(queue->ddgst_remaining) ? NVME_TCP_RECV_DDGST :
drivers/nvme/host/tcp.c
604
static void nvme_tcp_init_recv_ctx(struct nvme_tcp_queue *queue)
drivers/nvme/host/tcp.c
606
queue->pdu_remaining = sizeof(struct nvme_tcp_rsp_pdu) +
drivers/nvme/host/tcp.c
607
nvme_tcp_hdgst_len(queue);
drivers/nvme/host/tcp.c
608
queue->pdu_offset = 0;
drivers/nvme/host/tcp.c
609
queue->data_remaining = -1;
drivers/nvme/host/tcp.c
610
queue->ddgst_remaining = 0;
drivers/nvme/host/tcp.c
622
static int nvme_tcp_process_nvme_cqe(struct nvme_tcp_queue *queue,
drivers/nvme/host/tcp.c
628
rq = nvme_find_rq(nvme_tcp_tagset(queue), cqe->command_id);
drivers/nvme/host/tcp.c
630
dev_err(queue->ctrl->ctrl.device,
drivers/nvme/host/tcp.c
632
cqe->command_id, nvme_tcp_queue_id(queue));
drivers/nvme/host/tcp.c
633
nvme_tcp_error_recovery(&queue->ctrl->ctrl);
drivers/nvme/host/tcp.c
643
queue->nr_cqe++;
drivers/nvme/host/tcp.c
648
static int nvme_tcp_handle_c2h_data(struct nvme_tcp_queue *queue,
drivers/nvme/host/tcp.c
653
rq = nvme_find_rq(nvme_tcp_tagset(queue), pdu->command_id);
drivers/nvme/host/tcp.c
655
dev_err(queue->ctrl->ctrl.device,
drivers/nvme/host/tcp.c
657
pdu->command_id, nvme_tcp_queue_id(queue));
drivers/nvme/host/tcp.c
662
dev_err(queue->ctrl->ctrl.device,
drivers/nvme/host/tcp.c
664
nvme_tcp_queue_id(queue), rq->tag);
drivers/nvme/host/tcp.c
668
queue->data_remaining = le32_to_cpu(pdu->data_length);
drivers/nvme/host/tcp.c
672
dev_err(queue->ctrl->ctrl.device,
drivers/nvme/host/tcp.c
674
nvme_tcp_queue_id(queue), rq->tag);
drivers/nvme/host/tcp.c
675
nvme_tcp_error_recovery(&queue->ctrl->ctrl);
drivers/nvme/host/tcp.c
682
static int nvme_tcp_handle_comp(struct nvme_tcp_queue *queue,
drivers/nvme/host/tcp.c
694
if (unlikely(nvme_is_aen_req(nvme_tcp_queue_id(queue),
drivers/nvme/host/tcp.c
696
nvme_complete_async_event(&queue->ctrl->ctrl, cqe->status,
drivers/nvme/host/tcp.c
699
ret = nvme_tcp_process_nvme_cqe(queue, cqe);
drivers/nvme/host/tcp.c
707
struct nvme_tcp_queue *queue = req->queue;
drivers/nvme/host/tcp.c
710
u8 hdgst = nvme_tcp_hdgst_len(queue);
drivers/nvme/host/tcp.c
711
u8 ddgst = nvme_tcp_ddgst_len(queue);
drivers/nvme/host/tcp.c
715
req->pdu_len = min(req->h2cdata_left, queue->maxh2cdata);
drivers/nvme/host/tcp.c
724
if (queue->hdr_digest)
drivers/nvme/host/tcp.c
726
if (queue->data_digest)
drivers/nvme/host/tcp.c
738
static int nvme_tcp_handle_r2t(struct nvme_tcp_queue *queue,
drivers/nvme/host/tcp.c
746
rq = nvme_find_rq(nvme_tcp_tagset(queue), pdu->command_id);
drivers/nvme/host/tcp.c
748
dev_err(queue->ctrl->ctrl.device,
drivers/nvme/host/tcp.c
750
pdu->command_id, nvme_tcp_queue_id(queue));
drivers/nvme/host/tcp.c
756
dev_err(queue->ctrl->ctrl.device,
drivers/nvme/host/tcp.c
763
dev_err(queue->ctrl->ctrl.device,
drivers/nvme/host/tcp.c
770
dev_err(queue->ctrl->ctrl.device,
drivers/nvme/host/tcp.c
778
dev_err(queue->ctrl->ctrl.device,
drivers/nvme/host/tcp.c
791
llist_add(&req->lentry, &queue->req_list);
drivers/nvme/host/tcp.c
792
queue_work_on(queue->io_cpu, nvme_tcp_wq, &queue->io_work);
drivers/nvme/host/tcp.c
797
static void nvme_tcp_handle_c2h_term(struct nvme_tcp_queue *queue,
drivers/nvme/host/tcp.c
815
dev_err(queue->ctrl->ctrl.device,
drivers/nvme/host/tcp.c
827
dev_err(queue->ctrl->ctrl.device,
drivers/nvme/host/tcp.c
831
static int nvme_tcp_recv_pdu(struct nvme_tcp_queue *queue, struct sk_buff *skb,
drivers/nvme/host/tcp.c
835
char *pdu = queue->pdu;
drivers/nvme/host/tcp.c
836
size_t rcv_len = min_t(size_t, *len, queue->pdu_remaining);
drivers/nvme/host/tcp.c
840
&pdu[queue->pdu_offset], rcv_len);
drivers/nvme/host/tcp.c
844
queue->pdu_remaining -= rcv_len;
drivers/nvme/host/tcp.c
845
queue->pdu_offset += rcv_len;
drivers/nvme/host/tcp.c
848
if (queue->pdu_remaining)
drivers/nvme/host/tcp.c
851
hdr = queue->pdu;
drivers/nvme/host/tcp.c
856
dev_err(queue->ctrl->ctrl.device,
drivers/nvme/host/tcp.c
867
nvme_tcp_handle_c2h_term(queue, (void *)queue->pdu);
drivers/nvme/host/tcp.c
871
if (queue->hdr_digest) {
drivers/nvme/host/tcp.c
872
ret = nvme_tcp_verify_hdgst(queue, queue->pdu, hdr->hlen);
drivers/nvme/host/tcp.c
878
if (queue->data_digest) {
drivers/nvme/host/tcp.c
879
ret = nvme_tcp_check_ddgst(queue, queue->pdu);
drivers/nvme/host/tcp.c
886
return nvme_tcp_handle_c2h_data(queue, (void *)queue->pdu);
drivers/nvme/host/tcp.c
888
nvme_tcp_init_recv_ctx(queue);
drivers/nvme/host/tcp.c
889
return nvme_tcp_handle_comp(queue, (void *)queue->pdu);
drivers/nvme/host/tcp.c
891
nvme_tcp_init_recv_ctx(queue);
drivers/nvme/host/tcp.c
892
return nvme_tcp_handle_r2t(queue, (void *)queue->pdu);
drivers/nvme/host/tcp.c
898
dev_err(queue->ctrl->ctrl.device,
drivers/nvme/host/tcp.c
911
static int nvme_tcp_recv_data(struct nvme_tcp_queue *queue, struct sk_buff *skb,
drivers/nvme/host/tcp.c
914
struct nvme_tcp_data_pdu *pdu = (void *)queue->pdu;
drivers/nvme/host/tcp.c
916
nvme_cid_to_rq(nvme_tcp_tagset(queue), pdu->command_id);
drivers/nvme/host/tcp.c
922
recv_len = min_t(size_t, *len, queue->data_remaining);
drivers/nvme/host/tcp.c
934
dev_err(queue->ctrl->ctrl.device,
drivers/nvme/host/tcp.c
936
nvme_tcp_queue_id(queue), rq->tag);
drivers/nvme/host/tcp.c
937
nvme_tcp_init_recv_ctx(queue);
drivers/nvme/host/tcp.c
947
if (queue->data_digest)
drivers/nvme/host/tcp.c
949
&req->iter, recv_len, &queue->rcv_crc);
drivers/nvme/host/tcp.c
954
dev_err(queue->ctrl->ctrl.device,
drivers/nvme/host/tcp.c
956
nvme_tcp_queue_id(queue), rq->tag);
drivers/nvme/host/tcp.c
962
queue->data_remaining -= recv_len;
drivers/nvme/host/tcp.c
965
if (!queue->data_remaining) {
drivers/nvme/host/tcp.c
966
if (queue->data_digest) {
drivers/nvme/host/tcp.c
967
queue->exp_ddgst = nvme_tcp_ddgst_final(queue->rcv_crc);
drivers/nvme/host/tcp.c
968
queue->ddgst_remaining = NVME_TCP_DIGEST_LENGTH;
drivers/nvme/host/tcp.c
973
queue->nr_cqe++;
drivers/nvme/host/tcp.c
975
nvme_tcp_init_recv_ctx(queue);
drivers/nvme/host/tcp.c
982
static int nvme_tcp_recv_ddgst(struct nvme_tcp_queue *queue,
drivers/nvme/host/tcp.c
985
struct nvme_tcp_data_pdu *pdu = (void *)queue->pdu;
drivers/nvme/host/tcp.c
986
char *ddgst = (char *)&queue->recv_ddgst;
drivers/nvme/host/tcp.c
987
size_t recv_len = min_t(size_t, *len, queue->ddgst_remaining);
drivers/nvme/host/tcp.c
988
off_t off = NVME_TCP_DIGEST_LENGTH - queue->ddgst_remaining;
drivers/nvme/host/tcp.c
995
queue->ddgst_remaining -= recv_len;
drivers/nvme/host/tcp.c
998
if (queue->ddgst_remaining)
drivers/nvme/host/zns.c
122
struct request_queue *q = ns->disk->queue;
drivers/nvme/host/zns.c
203
ret = nvme_submit_sync_cmd(ns->queue, &c, report, buflen);
drivers/nvme/target/core.c
488
if (!blk_queue_pci_p2pdma(ns->bdev->bd_disk->queue)) {
drivers/nvme/target/fc.c
1554
struct nvmet_fc_tgt_queue *queue;
drivers/nvme/target/fc.c
1568
queue = assoc->queues[0];
drivers/nvme/target/fc.c
1569
if (queue && queue->nvme_sq.ctrl == ctrl) {
drivers/nvme/target/fc.c
1668
struct nvmet_fc_tgt_queue *queue;
drivers/nvme/target/fc.c
1704
queue = nvmet_fc_alloc_target_queue(iod->assoc, 0,
drivers/nvme/target/fc.c
1706
if (!queue) {
drivers/nvme/target/fc.c
1724
queue->ersp_ratio = be16_to_cpu(rqst->assoc_cmd.ersp_ratio);
drivers/nvme/target/fc.c
1725
atomic_set(&queue->connected, 1);
drivers/nvme/target/fc.c
1726
queue->sqhd = 0; /* best place to init value */
drivers/nvme/target/fc.c
1758
struct nvmet_fc_tgt_queue *queue;
drivers/nvme/target/fc.c
1794
queue = nvmet_fc_alloc_target_queue(iod->assoc,
drivers/nvme/target/fc.c
1797
if (!queue)
drivers/nvme/target/fc.c
1818
queue->ersp_ratio = be16_to_cpu(rqst->connect_cmd.ersp_ratio);
drivers/nvme/target/fc.c
1819
atomic_set(&queue->connected, 1);
drivers/nvme/target/fc.c
1820
queue->sqhd = 0; /* best place to init value */
drivers/nvme/target/fc.c
2182
rspcnt = atomic_inc_return(&fod->queue->zrspcnt);
drivers/nvme/target/fc.c
2183
if (!(rspcnt % fod->queue->ersp_ratio) ||
drivers/nvme/target/fc.c
2188
queue_90percent_full(fod->queue, le16_to_cpu(cqe->sq_head)))
drivers/nvme/target/fc.c
2200
rsn = atomic_inc_return(&fod->queue->rsn);
drivers/nvme/target/fc.c
2229
nvmet_fc_free_fcp_iod(fod->queue, fod);
drivers/nvme/target/fc.c
235
static void nvmet_fc_tgt_q_put(struct nvmet_fc_tgt_queue *queue);
drivers/nvme/target/fc.c
236
static int nvmet_fc_tgt_q_get(struct nvmet_fc_tgt_queue *queue);
drivers/nvme/target/fc.c
2411
nvmet_fc_free_fcp_iod(fod->queue, fod);
drivers/nvme/target/fc.c
2435
nvmet_fc_free_fcp_iod(fod->queue, fod);
drivers/nvme/target/fc.c
2469
fod->queue->sqhd = cqe->sq_head;
drivers/nvme/target/fc.c
2480
cqe->sq_head = fod->queue->sqhd; /* echo last cqe sqhd */
drivers/nvme/target/fc.c
2481
cqe->sq_id = cpu_to_le16(fod->queue->qid);
drivers/nvme/target/fc.c
2566
ret = nvmet_req_init(&fod->req, &fod->queue->nvme_sq,
drivers/nvme/target/fc.c
2577
atomic_inc(&fod->queue->sqtail);
drivers/nvme/target/fc.c
2663
struct nvmet_fc_tgt_queue *queue;
drivers/nvme/target/fc.c
2675
queue = nvmet_fc_find_target_queue(tgtport,
drivers/nvme/target/fc.c
2677
if (!queue)
drivers/nvme/target/fc.c
2687
spin_lock_irqsave(&queue->qlock, flags);
drivers/nvme/target/fc.c
2689
fod = nvmet_fc_alloc_fcp_iod(queue);
drivers/nvme/target/fc.c
2691
spin_unlock_irqrestore(&queue->qlock, flags);
drivers/nvme/target/fc.c
2698
nvmet_fc_queue_fcp_req(tgtport, queue, fcpreq);
drivers/nvme/target/fc.c
2704
spin_unlock_irqrestore(&queue->qlock, flags);
drivers/nvme/target/fc.c
2706
nvmet_fc_tgt_q_put(queue);
drivers/nvme/target/fc.c
2710
deferfcp = list_first_entry_or_null(&queue->avail_defer_list,
drivers/nvme/target/fc.c
2716
spin_unlock_irqrestore(&queue->qlock, flags);
drivers/nvme/target/fc.c
2722
nvmet_fc_tgt_q_put(queue);
drivers/nvme/target/fc.c
2725
spin_lock_irqsave(&queue->qlock, flags);
drivers/nvme/target/fc.c
2734
list_add_tail(&deferfcp->req_list, &queue->pending_cmd_list);
drivers/nvme/target/fc.c
2738
spin_unlock_irqrestore(&queue->qlock, flags);
drivers/nvme/target/fc.c
2772
struct nvmet_fc_tgt_queue *queue;
drivers/nvme/target/fc.c
2779
queue = fod->queue;
drivers/nvme/target/fc.c
2781
spin_lock_irqsave(&queue->qlock, flags);
drivers/nvme/target/fc.c
2793
spin_unlock_irqrestore(&queue->qlock, flags);
drivers/nvme/target/fc.c
2964
struct nvmet_fc_tgt_queue *queue =
drivers/nvme/target/fc.c
2966
struct nvmet_fc_tgtport *tgtport = queue->assoc ? queue->assoc->tgtport : NULL;
drivers/nvme/target/fc.c
2967
struct nvmet_fc_hostport *hostport = queue->assoc ? queue->assoc->hostport : NULL;
drivers/nvme/target/fc.c
619
struct nvmet_fc_tgt_queue *queue)
drivers/nvme/target/fc.c
621
struct nvmet_fc_fcp_iod *fod = queue->fod;
drivers/nvme/target/fc.c
624
for (i = 0; i < queue->sqsize; fod++, i++) {
drivers/nvme/target/fc.c
627
fod->queue = queue;
drivers/nvme/target/fc.c
632
list_add_tail(&fod->fcp_list, &queue->fod_list);
drivers/nvme/target/fc.c
654
struct nvmet_fc_tgt_queue *queue)
drivers/nvme/target/fc.c
656
struct nvmet_fc_fcp_iod *fod = queue->fod;
drivers/nvme/target/fc.c
659
for (i = 0; i < queue->sqsize; fod++, i++) {
drivers/nvme/target/fc.c
667
nvmet_fc_alloc_fcp_iod(struct nvmet_fc_tgt_queue *queue)
drivers/nvme/target/fc.c
671
lockdep_assert_held(&queue->qlock);
drivers/nvme/target/fc.c
673
fod = list_first_entry_or_null(&queue->fod_list,
drivers/nvme/target/fc.c
690
struct nvmet_fc_tgt_queue *queue,
drivers/nvme/target/fc.c
699
fcpreq->hwqid = queue->qid ?
drivers/nvme/target/fc.c
700
((queue->qid - 1) % tgtport->ops->max_hw_queues) : 0;
drivers/nvme/target/fc.c
712
nvmet_fc_queue_fcp_req(fod->tgtport, fod->queue, fod->fcpreq);
drivers/nvme/target/fc.c
717
nvmet_fc_free_fcp_iod(struct nvmet_fc_tgt_queue *queue,
drivers/nvme/target/fc.c
739
nvmet_fc_tgt_q_put(queue);
drivers/nvme/target/fc.c
741
spin_lock_irqsave(&queue->qlock, flags);
drivers/nvme/target/fc.c
742
deferfcp = list_first_entry_or_null(&queue->pending_cmd_list,
drivers/nvme/target/fc.c
745
list_add_tail(&fod->fcp_list, &fod->queue->fod_list);
drivers/nvme/target/fc.c
746
spin_unlock_irqrestore(&queue->qlock, flags);
drivers/nvme/target/fc.c
756
list_add_tail(&deferfcp->req_list, &queue->avail_defer_list);
drivers/nvme/target/fc.c
758
spin_unlock_irqrestore(&queue->qlock, flags);
drivers/nvme/target/fc.c
778
queue_work(queue->work_q, &fod->defer_work);
drivers/nvme/target/fc.c
785
struct nvmet_fc_tgt_queue *queue;
drivers/nvme/target/fc.c
791
queue = kzalloc_flex(*queue, fod, sqsize);
drivers/nvme/target/fc.c
792
if (!queue)
drivers/nvme/target/fc.c
795
queue->work_q = alloc_workqueue("ntfc%d.%d.%d", 0, 0,
drivers/nvme/target/fc.c
798
if (!queue->work_q)
drivers/nvme/target/fc.c
801
queue->qid = qid;
drivers/nvme/target/fc.c
802
queue->sqsize = sqsize;
drivers/nvme/target/fc.c
803
queue->assoc = assoc;
drivers/nvme/target/fc.c
804
INIT_LIST_HEAD(&queue->fod_list);
drivers/nvme/target/fc.c
805
INIT_LIST_HEAD(&queue->avail_defer_list);
drivers/nvme/target/fc.c
806
INIT_LIST_HEAD(&queue->pending_cmd_list);
drivers/nvme/target/fc.c
807
atomic_set(&queue->connected, 0);
drivers/nvme/target/fc.c
808
atomic_set(&queue->sqtail, 0);
drivers/nvme/target/fc.c
809
atomic_set(&queue->rsn, 1);
drivers/nvme/target/fc.c
810
atomic_set(&queue->zrspcnt, 0);
drivers/nvme/target/fc.c
811
spin_lock_init(&queue->qlock);
drivers/nvme/target/fc.c
812
kref_init(&queue->ref);
drivers/nvme/target/fc.c
814
nvmet_fc_prep_fcp_iodlist(assoc->tgtport, queue);
drivers/nvme/target/fc.c
816
nvmet_cq_init(&queue->nvme_cq);
drivers/nvme/target/fc.c
817
ret = nvmet_sq_init(&queue->nvme_sq, &queue->nvme_cq);
drivers/nvme/target/fc.c
822
assoc->queues[qid] = queue;
drivers/nvme/target/fc.c
824
return queue;
drivers/nvme/target/fc.c
827
nvmet_cq_put(&queue->nvme_cq);
drivers/nvme/target/fc.c
828
nvmet_fc_destroy_fcp_iodlist(assoc->tgtport, queue);
drivers/nvme/target/fc.c
829
destroy_workqueue(queue->work_q);
drivers/nvme/target/fc.c
831
kfree(queue);
drivers/nvme/target/fc.c
839
struct nvmet_fc_tgt_queue *queue =
drivers/nvme/target/fc.c
842
nvmet_fc_destroy_fcp_iodlist(queue->assoc->tgtport, queue);
drivers/nvme/target/fc.c
844
destroy_workqueue(queue->work_q);
drivers/nvme/target/fc.c
846
kfree(queue);
drivers/nvme/target/fc.c
850
nvmet_fc_tgt_q_put(struct nvmet_fc_tgt_queue *queue)
drivers/nvme/target/fc.c
852
kref_put(&queue->ref, nvmet_fc_tgt_queue_free);
drivers/nvme/target/fc.c
856
nvmet_fc_tgt_q_get(struct nvmet_fc_tgt_queue *queue)
drivers/nvme/target/fc.c
858
return kref_get_unless_zero(&queue->ref);
drivers/nvme/target/fc.c
863
nvmet_fc_delete_target_queue(struct nvmet_fc_tgt_queue *queue)
drivers/nvme/target/fc.c
865
struct nvmet_fc_tgtport *tgtport = queue->assoc->tgtport;
drivers/nvme/target/fc.c
866
struct nvmet_fc_fcp_iod *fod = queue->fod;
drivers/nvme/target/fc.c
872
disconnect = atomic_xchg(&queue->connected, 0);
drivers/nvme/target/fc.c
878
spin_lock_irqsave(&queue->qlock, flags);
drivers/nvme/target/fc.c
880
for (i = 0; i < queue->sqsize; fod++, i++) {
drivers/nvme/target/fc.c
900
list_for_each_entry_safe(deferfcp, tempptr, &queue->avail_defer_list,
drivers/nvme/target/fc.c
907
deferfcp = list_first_entry_or_null(&queue->pending_cmd_list,
drivers/nvme/target/fc.c
913
spin_unlock_irqrestore(&queue->qlock, flags);
drivers/nvme/target/fc.c
925
nvmet_fc_tgt_q_put(queue);
drivers/nvme/target/fc.c
929
spin_lock_irqsave(&queue->qlock, flags);
drivers/nvme/target/fc.c
93
struct nvmet_fc_tgt_queue *queue;
drivers/nvme/target/fc.c
931
spin_unlock_irqrestore(&queue->qlock, flags);
drivers/nvme/target/fc.c
933
flush_workqueue(queue->work_q);
drivers/nvme/target/fc.c
935
nvmet_sq_destroy(&queue->nvme_sq);
drivers/nvme/target/fc.c
936
nvmet_cq_put(&queue->nvme_cq);
drivers/nvme/target/fc.c
938
nvmet_fc_tgt_q_put(queue);
drivers/nvme/target/fc.c
946
struct nvmet_fc_tgt_queue *queue;
drivers/nvme/target/fc.c
956
queue = assoc->queues[qid];
drivers/nvme/target/fc.c
957
if (queue &&
drivers/nvme/target/fc.c
958
(!atomic_read(&queue->connected) ||
drivers/nvme/target/fc.c
959
!nvmet_fc_tgt_q_get(queue)))
drivers/nvme/target/fc.c
960
queue = NULL;
drivers/nvme/target/fc.c
962
return queue;
drivers/nvme/target/loop.c
105
if (unlikely(nvme_is_aen_req(nvme_loop_queue_idx(queue),
drivers/nvme/target/loop.c
107
nvme_complete_async_event(&queue->ctrl->ctrl, cqe->status,
drivers/nvme/target/loop.c
112
rq = nvme_find_rq(nvme_loop_tagset(queue), cqe->command_id);
drivers/nvme/target/loop.c
114
dev_err(queue->ctrl->ctrl.device,
drivers/nvme/target/loop.c
116
cqe->command_id, nvme_loop_queue_idx(queue));
drivers/nvme/target/loop.c
136
struct nvme_ns *ns = hctx->queue->queuedata;
drivers/nvme/target/loop.c
137
struct nvme_loop_queue *queue = hctx->driver_data;
drivers/nvme/target/loop.c
140
bool queue_ready = test_bit(NVME_LOOP_Q_LIVE, &queue->flags);
drivers/nvme/target/loop.c
143
if (!nvme_check_ready(&queue->ctrl->ctrl, req, queue_ready))
drivers/nvme/target/loop.c
144
return nvme_fail_nonready_command(&queue->ctrl->ctrl, req);
drivers/nvme/target/loop.c
152
iod->req.port = queue->ctrl->port;
drivers/nvme/target/loop.c
153
if (!nvmet_req_init(&iod->req, &queue->nvme_sq, &nvme_loop_ops))
drivers/nvme/target/loop.c
177
struct nvme_loop_queue *queue = &ctrl->queues[0];
drivers/nvme/target/loop.c
185
if (!nvmet_req_init(&iod->req, &queue->nvme_sq, &nvme_loop_ops)) {
drivers/nvme/target/loop.c
198
iod->queue = &ctrl->queues[queue_idx];
drivers/nvme/target/loop.c
222
struct nvme_loop_queue *queue = &ctrl->queues[hctx_idx + 1];
drivers/nvme/target/loop.c
23
struct nvme_loop_queue *queue;
drivers/nvme/target/loop.c
234
hctx->driver_data = queue;
drivers/nvme/target/loop.c
242
struct nvme_loop_queue *queue = &ctrl->queues[0];
drivers/nvme/target/loop.c
246
hctx->driver_data = queue;
drivers/nvme/target/loop.c
71
static inline int nvme_loop_queue_idx(struct nvme_loop_queue *queue)
drivers/nvme/target/loop.c
73
return queue - queue->ctrl->queues;
drivers/nvme/target/loop.c
84
static struct blk_mq_tags *nvme_loop_tagset(struct nvme_loop_queue *queue)
drivers/nvme/target/loop.c
86
u32 queue_idx = nvme_loop_queue_idx(queue);
drivers/nvme/target/loop.c
89
return queue->ctrl->admin_tag_set.tags[queue_idx];
drivers/nvme/target/loop.c
90
return queue->ctrl->tag_set.tags[queue_idx - 1];
drivers/nvme/target/loop.c
95
struct nvme_loop_queue *queue =
drivers/nvme/target/passthru.c
318
q = ns->queue;
drivers/nvme/target/pci-epf.c
1192
static void nvmet_pci_epf_drain_queue(struct nvmet_pci_epf_queue *queue)
drivers/nvme/target/pci-epf.c
1197
spin_lock_irqsave(&queue->lock, flags);
drivers/nvme/target/pci-epf.c
1198
while (!list_empty(&queue->list)) {
drivers/nvme/target/pci-epf.c
1199
iod = list_first_entry(&queue->list, struct nvmet_pci_epf_iod,
drivers/nvme/target/pci-epf.c
1204
spin_unlock_irqrestore(&queue->lock, flags);
drivers/nvme/target/pci-epf.c
1547
struct nvmet_pci_epf_queue *queue;
drivers/nvme/target/pci-epf.c
1550
queue = &ctrl->sq[qid];
drivers/nvme/target/pci-epf.c
1552
queue = &ctrl->cq[qid];
drivers/nvme/target/pci-epf.c
1553
INIT_DELAYED_WORK(&queue->work, nvmet_pci_epf_cq_work);
drivers/nvme/target/pci-epf.c
1555
queue->ctrl = ctrl;
drivers/nvme/target/pci-epf.c
1556
queue->qid = qid;
drivers/nvme/target/pci-epf.c
1557
spin_lock_init(&queue->lock);
drivers/nvme/target/pci-epf.c
1558
INIT_LIST_HEAD(&queue->list);
drivers/nvme/target/rdma.c
1003
spin_lock_irqsave(&queue->state_lock, flags);
drivers/nvme/target/rdma.c
1008
if (queue->state == NVMET_RDMA_Q_LIVE)
drivers/nvme/target/rdma.c
1010
else if (queue->state == NVMET_RDMA_Q_CONNECTING)
drivers/nvme/target/rdma.c
1011
list_add_tail(&rsp->wait_list, &queue->rsp_wait_list);
drivers/nvme/target/rdma.c
1014
spin_unlock_irqrestore(&queue->state_lock, flags);
drivers/nvme/target/rdma.c
1022
struct nvmet_rdma_queue *queue = wc->qp->qp_context;
drivers/nvme/target/rdma.c
1030
nvmet_rdma_error_comp(queue);
drivers/nvme/target/rdma.c
1037
nvmet_rdma_error_comp(queue);
drivers/nvme/target/rdma.c
1041
cmd->queue = queue;
drivers/nvme/target/rdma.c
1042
rsp = nvmet_rdma_get_rsp(queue);
drivers/nvme/target/rdma.c
1049
nvmet_rdma_post_recv(queue->dev, cmd);
drivers/nvme/target/rdma.c
1052
rsp->queue = queue;
drivers/nvme/target/rdma.c
1056
rsp->req.port = queue->port;
drivers/nvme/target/rdma.c
1060
if (unlikely(queue->state != NVMET_RDMA_Q_LIVE) &&
drivers/nvme/target/rdma.c
1061
nvmet_rdma_recv_not_live(queue, rsp))
drivers/nvme/target/rdma.c
1064
nvmet_rdma_handle_command(queue, rsp);
drivers/nvme/target/rdma.c
1262
static int nvmet_rdma_create_queue_ib(struct nvmet_rdma_queue *queue)
drivers/nvme/target/rdma.c
1265
struct nvmet_rdma_device *ndev = queue->dev;
drivers/nvme/target/rdma.c
1271
nr_cqe = queue->recv_queue_size + 2 * queue->send_queue_size;
drivers/nvme/target/rdma.c
1273
queue->cq = ib_cq_pool_get(ndev->device, nr_cqe + 1,
drivers/nvme/target/rdma.c
1274
queue->comp_vector, IB_POLL_WORKQUEUE);
drivers/nvme/target/rdma.c
1275
if (IS_ERR(queue->cq)) {
drivers/nvme/target/rdma.c
1276
ret = PTR_ERR(queue->cq);
drivers/nvme/target/rdma.c
1282
qp_attr.qp_context = queue;
drivers/nvme/target/rdma.c
1284
qp_attr.send_cq = queue->cq;
drivers/nvme/target/rdma.c
1285
qp_attr.recv_cq = queue->cq;
drivers/nvme/target/rdma.c
1289
qp_attr.cap.max_send_wr = queue->send_queue_size + 1;
drivers/nvme/target/rdma.c
1290
factor = rdma_rw_mr_factor(ndev->device, queue->cm_id->port_num,
drivers/nvme/target/rdma.c
1292
qp_attr.cap.max_rdma_ctxs = queue->send_queue_size * factor;
drivers/nvme/target/rdma.c
1296
if (queue->nsrq) {
drivers/nvme/target/rdma.c
1297
qp_attr.srq = queue->nsrq->srq;
drivers/nvme/target/rdma.c
1300
qp_attr.cap.max_recv_wr = 1 + queue->recv_queue_size;
drivers/nvme/target/rdma.c
1304
if (queue->port->pi_enable && queue->host_qid)
drivers/nvme/target/rdma.c
1307
ret = rdma_create_qp(queue->cm_id, ndev->pd, &qp_attr);
drivers/nvme/target/rdma.c
1312
queue->qp = queue->cm_id->qp;
drivers/nvme/target/rdma.c
1314
atomic_set(&queue->sq_wr_avail, qp_attr.cap.max_send_wr);
drivers/nvme/target/rdma.c
1317
__func__, queue->cq->cqe, qp_attr.cap.max_send_sge,
drivers/nvme/target/rdma.c
1318
qp_attr.cap.max_send_wr, queue->cm_id);
drivers/nvme/target/rdma.c
1320
if (!queue->nsrq) {
drivers/nvme/target/rdma.c
1321
for (i = 0; i < queue->recv_queue_size; i++) {
drivers/nvme/target/rdma.c
1322
queue->cmds[i].queue = queue;
drivers/nvme/target/rdma.c
1323
ret = nvmet_rdma_post_recv(ndev, &queue->cmds[i]);
drivers/nvme/target/rdma.c
1333
rdma_destroy_qp(queue->cm_id);
drivers/nvme/target/rdma.c
1335
ib_cq_pool_put(queue->cq, nr_cqe + 1);
drivers/nvme/target/rdma.c
1339
static void nvmet_rdma_destroy_queue_ib(struct nvmet_rdma_queue *queue)
drivers/nvme/target/rdma.c
1341
ib_drain_qp(queue->qp);
drivers/nvme/target/rdma.c
1342
if (queue->cm_id)
drivers/nvme/target/rdma.c
1343
rdma_destroy_id(queue->cm_id);
drivers/nvme/target/rdma.c
1344
ib_destroy_qp(queue->qp);
drivers/nvme/target/rdma.c
1345
ib_cq_pool_put(queue->cq, queue->recv_queue_size + 2 *
drivers/nvme/target/rdma.c
1346
queue->send_queue_size + 1);
drivers/nvme/target/rdma.c
1349
static void nvmet_rdma_free_queue(struct nvmet_rdma_queue *queue)
drivers/nvme/target/rdma.c
1351
pr_debug("freeing queue %d\n", queue->idx);
drivers/nvme/target/rdma.c
1353
nvmet_sq_destroy(&queue->nvme_sq);
drivers/nvme/target/rdma.c
1354
nvmet_cq_put(&queue->nvme_cq);
drivers/nvme/target/rdma.c
1356
nvmet_rdma_destroy_queue_ib(queue);
drivers/nvme/target/rdma.c
1357
if (!queue->nsrq) {
drivers/nvme/target/rdma.c
1358
nvmet_rdma_free_cmds(queue->dev, queue->cmds,
drivers/nvme/target/rdma.c
1359
queue->recv_queue_size,
drivers/nvme/target/rdma.c
1360
!queue->host_qid);
drivers/nvme/target/rdma.c
1362
nvmet_rdma_free_rsps(queue);
drivers/nvme/target/rdma.c
1363
ida_free(&nvmet_rdma_queue_ida, queue->idx);
drivers/nvme/target/rdma.c
1364
kfree(queue);
drivers/nvme/target/rdma.c
1369
struct nvmet_rdma_queue *queue =
drivers/nvme/target/rdma.c
1371
struct nvmet_rdma_device *dev = queue->dev;
drivers/nvme/target/rdma.c
1373
nvmet_rdma_free_queue(queue);
drivers/nvme/target/rdma.c
1380
struct nvmet_rdma_queue *queue)
drivers/nvme/target/rdma.c
1391
queue->host_qid = le16_to_cpu(req->qid);
drivers/nvme/target/rdma.c
1397
queue->recv_queue_size = le16_to_cpu(req->hsqsize) + 1;
drivers/nvme/target/rdma.c
1398
queue->send_queue_size = le16_to_cpu(req->hrqsize);
drivers/nvme/target/rdma.c
1400
if (!queue->host_qid && queue->recv_queue_size > NVME_AQ_DEPTH)
drivers/nvme/target/rdma.c
1429
struct nvmet_rdma_queue *queue;
drivers/nvme/target/rdma.c
1432
queue = kzalloc_obj(*queue);
drivers/nvme/target/rdma.c
1433
if (!queue) {
drivers/nvme/target/rdma.c
1438
nvmet_cq_init(&queue->nvme_cq);
drivers/nvme/target/rdma.c
1439
ret = nvmet_sq_init(&queue->nvme_sq, &queue->nvme_cq);
drivers/nvme/target/rdma.c
1445
ret = nvmet_rdma_parse_cm_connect_req(&event->param.conn, queue);
drivers/nvme/target/rdma.c
1453
INIT_WORK(&queue->release_work, nvmet_rdma_release_queue_work);
drivers/nvme/target/rdma.c
1454
queue->dev = ndev;
drivers/nvme/target/rdma.c
1455
queue->cm_id = cm_id;
drivers/nvme/target/rdma.c
1456
queue->port = port->nport;
drivers/nvme/target/rdma.c
1458
spin_lock_init(&queue->state_lock);
drivers/nvme/target/rdma.c
1459
queue->state = NVMET_RDMA_Q_CONNECTING;
drivers/nvme/target/rdma.c
1460
INIT_LIST_HEAD(&queue->rsp_wait_list);
drivers/nvme/target/rdma.c
1461
INIT_LIST_HEAD(&queue->rsp_wr_wait_list);
drivers/nvme/target/rdma.c
1462
spin_lock_init(&queue->rsp_wr_wait_lock);
drivers/nvme/target/rdma.c
1463
INIT_LIST_HEAD(&queue->queue_list);
drivers/nvme/target/rdma.c
1465
queue->idx = ida_alloc(&nvmet_rdma_queue_ida, GFP_KERNEL);
drivers/nvme/target/rdma.c
1466
if (queue->idx < 0) {
drivers/nvme/target/rdma.c
1475
queue->comp_vector = !queue->host_qid ? 0 :
drivers/nvme/target/rdma.c
1476
queue->idx % ndev->device->num_comp_vectors;
drivers/nvme/target/rdma.c
1479
ret = nvmet_rdma_alloc_rsps(queue);
drivers/nvme/target/rdma.c
1486
queue->nsrq = ndev->srqs[queue->comp_vector % ndev->srq_count];
drivers/nvme/target/rdma.c
1488
queue->cmds = nvmet_rdma_alloc_cmds(ndev,
drivers/nvme/target/rdma.c
1489
queue->recv_queue_size,
drivers/nvme/target/rdma.c
1490
!queue->host_qid);
drivers/nvme/target/rdma.c
1491
if (IS_ERR(queue->cmds)) {
drivers/nvme/target/rdma.c
1497
ret = nvmet_rdma_create_queue_ib(queue);
drivers/nvme/target/rdma.c
1505
return queue;
drivers/nvme/target/rdma.c
1508
if (!queue->nsrq) {
drivers/nvme/target/rdma.c
1509
nvmet_rdma_free_cmds(queue->dev, queue->cmds,
drivers/nvme/target/rdma.c
1510
queue->recv_queue_size,
drivers/nvme/target/rdma.c
1511
!queue->host_qid);
drivers/nvme/target/rdma.c
1514
nvmet_rdma_free_rsps(queue);
drivers/nvme/target/rdma.c
1516
ida_free(&nvmet_rdma_queue_ida, queue->idx);
drivers/nvme/target/rdma.c
1518
nvmet_sq_destroy(&queue->nvme_sq);
drivers/nvme/target/rdma.c
1520
nvmet_cq_put(&queue->nvme_cq);
drivers/nvme/target/rdma.c
1521
kfree(queue);
drivers/nvme/target/rdma.c
1529
struct nvmet_rdma_queue *queue = priv;
drivers/nvme/target/rdma.c
1533
rdma_notify(queue->cm_id, event->event);
drivers/nvme/target/rdma.c
1537
queue);
drivers/nvme/target/rdma.c
1547
struct nvmet_rdma_queue *queue,
drivers/nvme/target/rdma.c
1557
queue->dev->device->attrs.max_qp_init_rd_atom);
drivers/nvme/target/rdma.c
1561
priv.crqsize = cpu_to_le16(queue->recv_queue_size);
drivers/nvme/target/rdma.c
1574
struct nvmet_rdma_queue *queue;
drivers/nvme/target/rdma.c
1583
queue = nvmet_rdma_alloc_queue(ndev, cm_id, event);
drivers/nvme/target/rdma.c
1584
if (!queue) {
drivers/nvme/target/rdma.c
1589
if (queue->host_qid == 0) {
drivers/nvme/target/rdma.c
1596
if (q->nvme_sq.ctrl == queue->nvme_sq.ctrl &&
drivers/nvme/target/rdma.c
1605
ret = nvmet_rdma_cm_accept(cm_id, queue, &event->param.conn);
drivers/nvme/target/rdma.c
1611
queue->cm_id = NULL;
drivers/nvme/target/rdma.c
1616
list_add_tail(&queue->queue_list, &nvmet_rdma_queue_list);
drivers/nvme/target/rdma.c
1622
nvmet_rdma_free_queue(queue);
drivers/nvme/target/rdma.c
1629
static void nvmet_rdma_queue_established(struct nvmet_rdma_queue *queue)
drivers/nvme/target/rdma.c
1633
spin_lock_irqsave(&queue->state_lock, flags);
drivers/nvme/target/rdma.c
1634
if (queue->state != NVMET_RDMA_Q_CONNECTING) {
drivers/nvme/target/rdma.c
1638
queue->state = NVMET_RDMA_Q_LIVE;
drivers/nvme/target/rdma.c
1640
while (!list_empty(&queue->rsp_wait_list)) {
drivers/nvme/target/rdma.c
1643
cmd = list_first_entry(&queue->rsp_wait_list,
drivers/nvme/target/rdma.c
1647
spin_unlock_irqrestore(&queue->state_lock, flags);
drivers/nvme/target/rdma.c
1648
nvmet_rdma_handle_command(queue, cmd);
drivers/nvme/target/rdma.c
1649
spin_lock_irqsave(&queue->state_lock, flags);
drivers/nvme/target/rdma.c
1653
spin_unlock_irqrestore(&queue->state_lock, flags);
drivers/nvme/target/rdma.c
1656
static void __nvmet_rdma_queue_disconnect(struct nvmet_rdma_queue *queue)
drivers/nvme/target/rdma.c
1661
pr_debug("cm_id= %p queue->state= %d\n", queue->cm_id, queue->state);
drivers/nvme/target/rdma.c
1663
spin_lock_irqsave(&queue->state_lock, flags);
drivers/nvme/target/rdma.c
1664
switch (queue->state) {
drivers/nvme/target/rdma.c
1666
while (!list_empty(&queue->rsp_wait_list)) {
drivers/nvme/target/rdma.c
1669
rsp = list_first_entry(&queue->rsp_wait_list,
drivers/nvme/target/rdma.c
1677
queue->state = NVMET_RDMA_Q_DISCONNECTING;
drivers/nvme/target/rdma.c
1683
spin_unlock_irqrestore(&queue->state_lock, flags);
drivers/nvme/target/rdma.c
1686
rdma_disconnect(queue->cm_id);
drivers/nvme/target/rdma.c
1687
queue_work(nvmet_wq, &queue->release_work);
drivers/nvme/target/rdma.c
1691
static void nvmet_rdma_queue_disconnect(struct nvmet_rdma_queue *queue)
drivers/nvme/target/rdma.c
1696
if (!list_empty(&queue->queue_list)) {
drivers/nvme/target/rdma.c
1697
list_del_init(&queue->queue_list);
drivers/nvme/target/rdma.c
1703
__nvmet_rdma_queue_disconnect(queue);
drivers/nvme/target/rdma.c
1707
struct nvmet_rdma_queue *queue)
drivers/nvme/target/rdma.c
1709
WARN_ON_ONCE(queue->state != NVMET_RDMA_Q_CONNECTING);
drivers/nvme/target/rdma.c
1712
if (!list_empty(&queue->queue_list))
drivers/nvme/target/rdma.c
1713
list_del_init(&queue->queue_list);
drivers/nvme/target/rdma.c
1716
pr_err("failed to connect queue %d\n", queue->idx);
drivers/nvme/target/rdma.c
1717
queue_work(nvmet_wq, &queue->release_work);
drivers/nvme/target/rdma.c
172
static void nvmet_rdma_queue_disconnect(struct nvmet_rdma_queue *queue);
drivers/nvme/target/rdma.c
1736
struct nvmet_rdma_queue *queue)
drivers/nvme/target/rdma.c
1740
if (queue) {
drivers/nvme/target/rdma.c
1770
struct nvmet_rdma_queue *queue = NULL;
drivers/nvme/target/rdma.c
1774
queue = cm_id->qp->qp_context;
drivers/nvme/target/rdma.c
1785
nvmet_rdma_queue_established(queue);
drivers/nvme/target/rdma.c
1788
if (!queue) {
drivers/nvme/target/rdma.c
1797
nvmet_rdma_queue_disconnect(queue);
drivers/nvme/target/rdma.c
1800
ret = nvmet_rdma_device_removal(cm_id, queue);
drivers/nvme/target/rdma.c
1808
nvmet_rdma_queue_connect_fail(cm_id, queue);
drivers/nvme/target/rdma.c
1821
struct nvmet_rdma_queue *queue, *n;
drivers/nvme/target/rdma.c
1824
list_for_each_entry_safe(queue, n, &nvmet_rdma_queue_list, queue_list) {
drivers/nvme/target/rdma.c
1825
if (queue->nvme_sq.ctrl != ctrl)
drivers/nvme/target/rdma.c
1827
list_del_init(&queue->queue_list);
drivers/nvme/target/rdma.c
1828
__nvmet_rdma_queue_disconnect(queue);
drivers/nvme/target/rdma.c
1835
struct nvmet_rdma_queue *queue, *tmp;
drivers/nvme/target/rdma.c
1839
list_for_each_entry_safe(queue, tmp, &nvmet_rdma_queue_list,
drivers/nvme/target/rdma.c
1841
if (queue->port != nport)
drivers/nvme/target/rdma.c
1844
list_del_init(&queue->queue_list);
drivers/nvme/target/rdma.c
1845
__nvmet_rdma_queue_disconnect(queue);
drivers/nvme/target/rdma.c
2006
struct rdma_cm_id *req_cm_id = rsp->queue->cm_id;
drivers/nvme/target/rdma.c
2019
struct nvmet_rdma_queue *queue =
drivers/nvme/target/rdma.c
2023
(struct sockaddr *)&queue->cm_id->route.addr.dst_addr);
drivers/nvme/target/rdma.c
2057
struct nvmet_rdma_queue *queue, *tmp;
drivers/nvme/target/rdma.c
2078
list_for_each_entry_safe(queue, tmp, &nvmet_rdma_queue_list,
drivers/nvme/target/rdma.c
2080
if (queue->dev->device != ib_device)
drivers/nvme/target/rdma.c
2083
pr_info("Removing queue %d\n", queue->idx);
drivers/nvme/target/rdma.c
2084
list_del_init(&queue->queue_list);
drivers/nvme/target/rdma.c
2085
__nvmet_rdma_queue_disconnect(queue);
drivers/nvme/target/rdma.c
213
nvmet_rdma_get_rsp(struct nvmet_rdma_queue *queue)
drivers/nvme/target/rdma.c
218
tag = sbitmap_get(&queue->rsp_tags);
drivers/nvme/target/rdma.c
220
rsp = &queue->rsps[tag];
drivers/nvme/target/rdma.c
228
ret = nvmet_rdma_alloc_rsp(queue->dev, rsp,
drivers/nvme/target/rdma.c
243
nvmet_rdma_free_rsp(rsp->queue->dev, rsp);
drivers/nvme/target/rdma.c
248
sbitmap_clear_bit(&rsp->queue->rsp_tags, rsp->tag);
drivers/nvme/target/rdma.c
448
nvmet_rdma_alloc_rsps(struct nvmet_rdma_queue *queue)
drivers/nvme/target/rdma.c
450
struct nvmet_rdma_device *ndev = queue->dev;
drivers/nvme/target/rdma.c
451
int nr_rsps = queue->recv_queue_size * 2;
drivers/nvme/target/rdma.c
454
if (sbitmap_init_node(&queue->rsp_tags, nr_rsps, -1, GFP_KERNEL,
drivers/nvme/target/rdma.c
458
queue->rsps = kvzalloc_objs(struct nvmet_rdma_rsp, nr_rsps);
drivers/nvme/target/rdma.c
459
if (!queue->rsps)
drivers/nvme/target/rdma.c
463
struct nvmet_rdma_rsp *rsp = &queue->rsps[i];
drivers/nvme/target/rdma.c
474
nvmet_rdma_free_rsp(ndev, &queue->rsps[i]);
drivers/nvme/target/rdma.c
475
kvfree(queue->rsps);
drivers/nvme/target/rdma.c
477
sbitmap_free(&queue->rsp_tags);
drivers/nvme/target/rdma.c
482
static void nvmet_rdma_free_rsps(struct nvmet_rdma_queue *queue)
drivers/nvme/target/rdma.c
484
struct nvmet_rdma_device *ndev = queue->dev;
drivers/nvme/target/rdma.c
485
int i, nr_rsps = queue->recv_queue_size * 2;
drivers/nvme/target/rdma.c
488
nvmet_rdma_free_rsp(ndev, &queue->rsps[i]);
drivers/nvme/target/rdma.c
489
kvfree(queue->rsps);
drivers/nvme/target/rdma.c
490
sbitmap_free(&queue->rsp_tags);
drivers/nvme/target/rdma.c
505
ret = ib_post_recv(cmd->queue->qp, &cmd->wr, NULL);
drivers/nvme/target/rdma.c
513
static void nvmet_rdma_process_wr_wait_list(struct nvmet_rdma_queue *queue)
drivers/nvme/target/rdma.c
515
spin_lock(&queue->rsp_wr_wait_lock);
drivers/nvme/target/rdma.c
516
while (!list_empty(&queue->rsp_wr_wait_list)) {
drivers/nvme/target/rdma.c
52
struct nvmet_rdma_queue *queue;
drivers/nvme/target/rdma.c
520
rsp = list_entry(queue->rsp_wr_wait_list.next,
drivers/nvme/target/rdma.c
524
spin_unlock(&queue->rsp_wr_wait_lock);
drivers/nvme/target/rdma.c
526
spin_lock(&queue->rsp_wr_wait_lock);
drivers/nvme/target/rdma.c
529
list_add(&rsp->wait_list, &queue->rsp_wr_wait_list);
drivers/nvme/target/rdma.c
533
spin_unlock(&queue->rsp_wr_wait_lock);
drivers/nvme/target/rdma.c
628
struct rdma_cm_id *cm_id = rsp->queue->cm_id;
drivers/nvme/target/rdma.c
647
struct rdma_cm_id *cm_id = rsp->queue->cm_id;
drivers/nvme/target/rdma.c
66
struct nvmet_rdma_queue *queue;
drivers/nvme/target/rdma.c
662
struct nvmet_rdma_queue *queue = rsp->queue;
drivers/nvme/target/rdma.c
664
atomic_add(1 + rsp->n_rdma, &queue->sq_wr_avail);
drivers/nvme/target/rdma.c
672
if (unlikely(!list_empty_careful(&queue->rsp_wr_wait_list)))
drivers/nvme/target/rdma.c
673
nvmet_rdma_process_wr_wait_list(queue);
drivers/nvme/target/rdma.c
678
static void nvmet_rdma_error_comp(struct nvmet_rdma_queue *queue)
drivers/nvme/target/rdma.c
680
if (queue->nvme_sq.ctrl) {
drivers/nvme/target/rdma.c
681
nvmet_ctrl_fatal_error(queue->nvme_sq.ctrl);
drivers/nvme/target/rdma.c
688
nvmet_rdma_queue_disconnect(queue);
drivers/nvme/target/rdma.c
696
struct nvmet_rdma_queue *queue = wc->qp->qp_context;
drivers/nvme/target/rdma.c
704
nvmet_rdma_error_comp(queue);
drivers/nvme/target/rdma.c
712
struct rdma_cm_id *cm_id = rsp->queue->cm_id;
drivers/nvme/target/rdma.c
733
nvmet_rdma_post_recv(rsp->queue->dev, rsp->cmd);
drivers/nvme/target/rdma.c
735
ib_dma_sync_single_for_device(rsp->queue->dev->device,
drivers/nvme/target/rdma.c
749
struct nvmet_rdma_queue *queue = wc->qp->qp_context;
drivers/nvme/target/rdma.c
753
atomic_add(rsp->n_rdma, &queue->sq_wr_avail);
drivers/nvme/target/rdma.c
763
nvmet_rdma_error_comp(queue);
drivers/nvme/target/rdma.c
782
struct nvmet_rdma_queue *queue = wc->qp->qp_context;
drivers/nvme/target/rdma.c
783
struct rdma_cm_id *cm_id = rsp->queue->cm_id;
drivers/nvme/target/rdma.c
790
atomic_add(rsp->n_rdma, &queue->sq_wr_avail);
drivers/nvme/target/rdma.c
800
nvmet_rdma_error_comp(queue);
drivers/nvme/target/rdma.c
857
if (off + len > rsp->queue->dev->inline_data_size) {
drivers/nvme/target/rdma.c
944
struct nvmet_rdma_queue *queue = rsp->queue;
drivers/nvme/target/rdma.c
947
&queue->sq_wr_avail) < 0)) {
drivers/nvme/target/rdma.c
949
1 + rsp->n_rdma, queue->idx,
drivers/nvme/target/rdma.c
950
queue->nvme_sq.ctrl->cntlid);
drivers/nvme/target/rdma.c
951
atomic_add(1 + rsp->n_rdma, &queue->sq_wr_avail);
drivers/nvme/target/rdma.c
956
if (rdma_rw_ctx_post(&rsp->rw, queue->qp,
drivers/nvme/target/rdma.c
957
queue->cm_id->port_num, &rsp->read_cqe, NULL))
drivers/nvme/target/rdma.c
966
static void nvmet_rdma_handle_command(struct nvmet_rdma_queue *queue,
drivers/nvme/target/rdma.c
971
ib_dma_sync_single_for_cpu(queue->dev->device,
drivers/nvme/target/rdma.c
974
ib_dma_sync_single_for_cpu(queue->dev->device,
drivers/nvme/target/rdma.c
978
if (!nvmet_req_init(&cmd->req, &queue->nvme_sq, &nvmet_rdma_ops))
drivers/nvme/target/rdma.c
986
spin_lock(&queue->rsp_wr_wait_lock);
drivers/nvme/target/rdma.c
987
list_add_tail(&cmd->wait_list, &queue->rsp_wr_wait_list);
drivers/nvme/target/rdma.c
988
spin_unlock(&queue->rsp_wr_wait_lock);
drivers/nvme/target/rdma.c
997
static bool nvmet_rdma_recv_not_live(struct nvmet_rdma_queue *queue,
drivers/nvme/target/tcp.c
1010
queue->idx, data->ttag);
drivers/nvme/target/tcp.c
1015
queue->cmd = cmd;
drivers/nvme/target/tcp.c
1016
queue->rcv_state = NVMET_TCP_RECV_DATA;
drivers/nvme/target/tcp.c
1022
nvmet_tcp_fatal_error(queue);
drivers/nvme/target/tcp.c
1026
static int nvmet_tcp_done_recv_pdu(struct nvmet_tcp_queue *queue)
drivers/nvme/target/tcp.c
1028
struct nvme_tcp_hdr *hdr = &queue->pdu.cmd.hdr;
drivers/nvme/target/tcp.c
1029
struct nvme_command *nvme_cmd = &queue->pdu.cmd.cmd;
drivers/nvme/target/tcp.c
1033
if (unlikely(queue->state == NVMET_TCP_Q_CONNECTING)) {
drivers/nvme/target/tcp.c
1037
nvmet_tcp_fatal_error(queue);
drivers/nvme/target/tcp.c
1040
return nvmet_tcp_handle_icreq(queue);
drivers/nvme/target/tcp.c
1045
queue->idx, queue->state);
drivers/nvme/target/tcp.c
1046
nvmet_tcp_fatal_error(queue);
drivers/nvme/target/tcp.c
1051
ret = nvmet_tcp_handle_h2c_data_pdu(queue);
drivers/nvme/target/tcp.c
1057
queue->cmd = nvmet_tcp_get_cmd(queue);
drivers/nvme/target/tcp.c
1058
if (unlikely(!queue->cmd)) {
drivers/nvme/target/tcp.c
1061
queue->idx, queue->nr_cmds, queue->send_list_len,
drivers/nvme/target/tcp.c
1063
nvmet_tcp_fatal_error(queue);
drivers/nvme/target/tcp.c
1067
req = &queue->cmd->req;
drivers/nvme/target/tcp.c
1070
if (unlikely(!nvmet_req_init(req, &queue->nvme_sq, &nvmet_tcp_ops))) {
drivers/nvme/target/tcp.c
1077
nvmet_tcp_handle_req_failure(queue, queue->cmd, req);
drivers/nvme/target/tcp.c
1081
ret = nvmet_tcp_map_data(queue->cmd);
drivers/nvme/target/tcp.c
1083
pr_err("queue %d: failed to map data\n", queue->idx);
drivers/nvme/target/tcp.c
1084
if (nvmet_tcp_has_inline_data(queue->cmd))
drivers/nvme/target/tcp.c
1085
nvmet_tcp_fatal_error(queue);
drivers/nvme/target/tcp.c
109
struct nvmet_tcp_queue *queue;
drivers/nvme/target/tcp.c
1092
if (nvmet_tcp_need_data_in(queue->cmd)) {
drivers/nvme/target/tcp.c
1093
if (nvmet_tcp_has_inline_data(queue->cmd)) {
drivers/nvme/target/tcp.c
1094
queue->rcv_state = NVMET_TCP_RECV_DATA;
drivers/nvme/target/tcp.c
1095
nvmet_tcp_build_pdu_iovec(queue->cmd);
drivers/nvme/target/tcp.c
1099
nvmet_tcp_queue_response(&queue->cmd->req);
drivers/nvme/target/tcp.c
1103
queue->cmd->req.execute(&queue->cmd->req);
drivers/nvme/target/tcp.c
1105
nvmet_prepare_receive_pdu(queue);
drivers/nvme/target/tcp.c
1137
static int nvmet_tcp_tls_record_ok(struct nvmet_tcp_queue *queue,
drivers/nvme/target/tcp.c
1144
ctype = tls_get_record_type(queue->sock->sk, cmsg);
drivers/nvme/target/tcp.c
1151
tls_alert_recv(queue->sock->sk, msg, &level, &description);
drivers/nvme/target/tcp.c
1154
queue->idx, description);
drivers/nvme/target/tcp.c
1158
queue->idx, description);
drivers/nvme/target/tcp.c
1165
queue->idx, ctype);
drivers/nvme/target/tcp.c
1172
static int nvmet_tcp_try_recv_pdu(struct nvmet_tcp_queue *queue)
drivers/nvme/target/tcp.c
1174
struct nvme_tcp_hdr *hdr = &queue->pdu.cmd.hdr;
drivers/nvme/target/tcp.c
1181
iov.iov_base = (void *)&queue->pdu + queue->offset;
drivers/nvme/target/tcp.c
1182
iov.iov_len = queue->left;
drivers/nvme/target/tcp.c
1183
if (queue->tls_pskid) {
drivers/nvme/target/tcp.c
1187
len = kernel_recvmsg(queue->sock, &msg, &iov, 1,
drivers/nvme/target/tcp.c
1191
if (queue->tls_pskid) {
drivers/nvme/target/tcp.c
1192
ret = nvmet_tcp_tls_record_ok(queue, &msg, cbuf);
drivers/nvme/target/tcp.c
1197
queue->offset += len;
drivers/nvme/target/tcp.c
1198
queue->left -= len;
drivers/nvme/target/tcp.c
1199
if (queue->left)
drivers/nvme/target/tcp.c
1202
if (queue->offset == sizeof(struct nvme_tcp_hdr)) {
drivers/nvme/target/tcp.c
1203
u8 hdgst = nvmet_tcp_hdgst_len(queue);
drivers/nvme/target/tcp.c
1207
nvmet_tcp_fatal_error(queue);
drivers/nvme/target/tcp.c
1216
queue->left = hdr->hlen - queue->offset + hdgst;
drivers/nvme/target/tcp.c
1220
if (queue->hdr_digest &&
drivers/nvme/target/tcp.c
1221
nvmet_tcp_verify_hdgst(queue, &queue->pdu, hdr->hlen)) {
drivers/nvme/target/tcp.c
1222
nvmet_tcp_fatal_error(queue); /* fatal */
drivers/nvme/target/tcp.c
1226
if (queue->data_digest &&
drivers/nvme/target/tcp.c
1227
nvmet_tcp_check_ddgst(queue, &queue->pdu)) {
drivers/nvme/target/tcp.c
1228
nvmet_tcp_fatal_error(queue); /* fatal */
drivers/nvme/target/tcp.c
1232
return nvmet_tcp_done_recv_pdu(queue);
drivers/nvme/target/tcp.c
1237
struct nvmet_tcp_queue *queue = cmd->queue;
drivers/nvme/target/tcp.c
1240
queue->offset = 0;
drivers/nvme/target/tcp.c
1241
queue->left = NVME_TCP_DIGEST_LENGTH;
drivers/nvme/target/tcp.c
1242
queue->rcv_state = NVMET_TCP_RECV_DDGST;
drivers/nvme/target/tcp.c
1245
static int nvmet_tcp_try_recv_data(struct nvmet_tcp_queue *queue)
drivers/nvme/target/tcp.c
1247
struct nvmet_tcp_cmd *cmd = queue->cmd;
drivers/nvme/target/tcp.c
1251
len = sock_recvmsg(cmd->queue->sock, &cmd->recv_msg,
drivers/nvme/target/tcp.c
1255
if (queue->tls_pskid) {
drivers/nvme/target/tcp.c
1256
ret = nvmet_tcp_tls_record_ok(cmd->queue,
drivers/nvme/target/tcp.c
1266
if (queue->data_digest) {
drivers/nvme/target/tcp.c
1274
nvmet_prepare_receive_pdu(queue);
drivers/nvme/target/tcp.c
1278
static int nvmet_tcp_try_recv_ddgst(struct nvmet_tcp_queue *queue)
drivers/nvme/target/tcp.c
1280
struct nvmet_tcp_cmd *cmd = queue->cmd;
drivers/nvme/target/tcp.c
1285
.iov_base = (void *)&cmd->recv_ddgst + queue->offset,
drivers/nvme/target/tcp.c
1286
.iov_len = queue->left
drivers/nvme/target/tcp.c
1289
if (queue->tls_pskid) {
drivers/nvme/target/tcp.c
1293
len = kernel_recvmsg(queue->sock, &msg, &iov, 1,
drivers/nvme/target/tcp.c
1297
if (queue->tls_pskid) {
drivers/nvme/target/tcp.c
1298
ret = nvmet_tcp_tls_record_ok(queue, &msg, cbuf);
drivers/nvme/target/tcp.c
1303
queue->offset += len;
drivers/nvme/target/tcp.c
1304
queue->left -= len;
drivers/nvme/target/tcp.c
1305
if (queue->left)
drivers/nvme/target/tcp.c
1308
if (queue->data_digest && cmd->exp_ddgst != cmd->recv_ddgst) {
drivers/nvme/target/tcp.c
1310
queue->idx, cmd->req.cmd->common.command_id,
drivers/nvme/target/tcp.c
1311
queue->pdu.cmd.hdr.type, le32_to_cpu(cmd->recv_ddgst),
drivers/nvme/target/tcp.c
1315
nvmet_tcp_fatal_error(queue);
drivers/nvme/target/tcp.c
1325
nvmet_prepare_receive_pdu(queue);
drivers/nvme/target/tcp.c
1329
static int nvmet_tcp_try_recv_one(struct nvmet_tcp_queue *queue)
drivers/nvme/target/tcp.c
1333
if (unlikely(queue->rcv_state == NVMET_TCP_RECV_ERR))
drivers/nvme/target/tcp.c
1336
if (queue->rcv_state == NVMET_TCP_RECV_PDU) {
drivers/nvme/target/tcp.c
1337
result = nvmet_tcp_try_recv_pdu(queue);
drivers/nvme/target/tcp.c
1342
if (queue->rcv_state == NVMET_TCP_RECV_DATA) {
drivers/nvme/target/tcp.c
1343
result = nvmet_tcp_try_recv_data(queue);
drivers/nvme/target/tcp.c
1348
if (queue->rcv_state == NVMET_TCP_RECV_DDGST) {
drivers/nvme/target/tcp.c
1349
result = nvmet_tcp_try_recv_ddgst(queue);
drivers/nvme/target/tcp.c
1363
static int nvmet_tcp_try_recv(struct nvmet_tcp_queue *queue,
drivers/nvme/target/tcp.c
1369
ret = nvmet_tcp_try_recv_one(queue);
drivers/nvme/target/tcp.c
1371
nvmet_tcp_socket_error(queue, ret);
drivers/nvme/target/tcp.c
1384
struct nvmet_tcp_queue *queue =
drivers/nvme/target/tcp.c
1387
WARN_ON(queue->state != NVMET_TCP_Q_DISCONNECTING);
drivers/nvme/target/tcp.c
1388
queue_work(nvmet_wq, &queue->release_work);
drivers/nvme/target/tcp.c
1391
static void nvmet_tcp_schedule_release_queue(struct nvmet_tcp_queue *queue)
drivers/nvme/target/tcp.c
1393
spin_lock_bh(&queue->state_lock);
drivers/nvme/target/tcp.c
1394
if (queue->state == NVMET_TCP_Q_TLS_HANDSHAKE) {
drivers/nvme/target/tcp.c
1396
tls_handshake_cancel(queue->sock->sk);
drivers/nvme/target/tcp.c
1398
if (queue->state != NVMET_TCP_Q_DISCONNECTING) {
drivers/nvme/target/tcp.c
1399
queue->state = NVMET_TCP_Q_DISCONNECTING;
drivers/nvme/target/tcp.c
1400
kref_put(&queue->kref, nvmet_tcp_release_queue);
drivers/nvme/target/tcp.c
1402
spin_unlock_bh(&queue->state_lock);
drivers/nvme/target/tcp.c
1405
static inline void nvmet_tcp_arm_queue_deadline(struct nvmet_tcp_queue *queue)
drivers/nvme/target/tcp.c
1407
queue->poll_end = jiffies + usecs_to_jiffies(idle_poll_period_usecs);
drivers/nvme/target/tcp.c
1410
static bool nvmet_tcp_check_queue_deadline(struct nvmet_tcp_queue *queue,
drivers/nvme/target/tcp.c
1417
nvmet_tcp_arm_queue_deadline(queue);
drivers/nvme/target/tcp.c
1419
return !time_after(jiffies, queue->poll_end);
drivers/nvme/target/tcp.c
1424
struct nvmet_tcp_queue *queue =
drivers/nvme/target/tcp.c
1432
ret = nvmet_tcp_try_recv(queue, NVMET_TCP_RECV_BUDGET, &ops);
drivers/nvme/target/tcp.c
1438
ret = nvmet_tcp_try_send(queue, NVMET_TCP_SEND_BUDGET, &ops);
drivers/nvme/target/tcp.c
1450
if (nvmet_tcp_check_queue_deadline(queue, ops) || pending)
drivers/nvme/target/tcp.c
1451
queue_work_on(queue_cpu(queue), nvmet_tcp_wq, &queue->io_work);
drivers/nvme/target/tcp.c
1454
static int nvmet_tcp_alloc_cmd(struct nvmet_tcp_queue *queue,
drivers/nvme/target/tcp.c
1457
u8 hdgst = nvmet_tcp_hdgst_len(queue);
drivers/nvme/target/tcp.c
1459
c->queue = queue;
drivers/nvme/target/tcp.c
1460
c->req.port = queue->port->nport;
drivers/nvme/target/tcp.c
1462
c->cmd_pdu = page_frag_alloc(&queue->pf_cache,
drivers/nvme/target/tcp.c
1468
c->rsp_pdu = page_frag_alloc(&queue->pf_cache,
drivers/nvme/target/tcp.c
1474
c->data_pdu = page_frag_alloc(&queue->pf_cache,
drivers/nvme/target/tcp.c
1479
c->r2t_pdu = page_frag_alloc(&queue->pf_cache,
drivers/nvme/target/tcp.c
1484
if (queue->state == NVMET_TCP_Q_TLS_HANDSHAKE) {
drivers/nvme/target/tcp.c
1490
list_add_tail(&c->entry, &queue->free_list);
drivers/nvme/target/tcp.c
1510
static int nvmet_tcp_alloc_cmds(struct nvmet_tcp_queue *queue)
drivers/nvme/target/tcp.c
1513
int i, ret = -EINVAL, nr_cmds = queue->nr_cmds;
drivers/nvme/target/tcp.c
1520
ret = nvmet_tcp_alloc_cmd(queue, cmds + i);
drivers/nvme/target/tcp.c
1525
queue->cmds = cmds;
drivers/nvme/target/tcp.c
1536
static void nvmet_tcp_free_cmds(struct nvmet_tcp_queue *queue)
drivers/nvme/target/tcp.c
1538
struct nvmet_tcp_cmd *cmds = queue->cmds;
drivers/nvme/target/tcp.c
1541
for (i = 0; i < queue->nr_cmds; i++)
drivers/nvme/target/tcp.c
1544
nvmet_tcp_free_cmd(&queue->connect);
drivers/nvme/target/tcp.c
1548
static void nvmet_tcp_restore_socket_callbacks(struct nvmet_tcp_queue *queue)
drivers/nvme/target/tcp.c
1550
struct socket *sock = queue->sock;
drivers/nvme/target/tcp.c
1552
if (!queue->state_change)
drivers/nvme/target/tcp.c
1556
sock->sk->sk_data_ready = queue->data_ready;
drivers/nvme/target/tcp.c
1557
sock->sk->sk_state_change = queue->state_change;
drivers/nvme/target/tcp.c
1558
sock->sk->sk_write_space = queue->write_space;
drivers/nvme/target/tcp.c
1563
static void nvmet_tcp_uninit_data_in_cmds(struct nvmet_tcp_queue *queue)
drivers/nvme/target/tcp.c
1565
struct nvmet_tcp_cmd *cmd = queue->cmds;
drivers/nvme/target/tcp.c
1568
for (i = 0; i < queue->nr_cmds; i++, cmd++) {
drivers/nvme/target/tcp.c
1573
if (!queue->nr_cmds && nvmet_tcp_need_data_in(&queue->connect)) {
drivers/nvme/target/tcp.c
1575
nvmet_req_uninit(&queue->connect.req);
drivers/nvme/target/tcp.c
1579
static void nvmet_tcp_free_cmd_data_in_buffers(struct nvmet_tcp_queue *queue)
drivers/nvme/target/tcp.c
1581
struct nvmet_tcp_cmd *cmd = queue->cmds;
drivers/nvme/target/tcp.c
1584
for (i = 0; i < queue->nr_cmds; i++, cmd++)
drivers/nvme/target/tcp.c
1586
nvmet_tcp_free_cmd_buffers(&queue->connect);
drivers/nvme/target/tcp.c
1591
struct nvmet_tcp_queue *queue =
drivers/nvme/target/tcp.c
1595
list_del_init(&queue->queue_list);
drivers/nvme/target/tcp.c
1598
nvmet_tcp_restore_socket_callbacks(queue);
drivers/nvme/target/tcp.c
1599
cancel_delayed_work_sync(&queue->tls_handshake_tmo_work);
drivers/nvme/target/tcp.c
1600
cancel_work_sync(&queue->io_work);
drivers/nvme/target/tcp.c
1602
queue->rcv_state = NVMET_TCP_RECV_ERR;
drivers/nvme/target/tcp.c
1604
nvmet_sq_put_tls_key(&queue->nvme_sq);
drivers/nvme/target/tcp.c
1605
nvmet_tcp_uninit_data_in_cmds(queue);
drivers/nvme/target/tcp.c
1606
nvmet_sq_destroy(&queue->nvme_sq);
drivers/nvme/target/tcp.c
1607
nvmet_cq_put(&queue->nvme_cq);
drivers/nvme/target/tcp.c
1608
cancel_work_sync(&queue->io_work);
drivers/nvme/target/tcp.c
1609
nvmet_tcp_free_cmd_data_in_buffers(queue);
drivers/nvme/target/tcp.c
1611
fput(queue->sock->file);
drivers/nvme/target/tcp.c
1612
nvmet_tcp_free_cmds(queue);
drivers/nvme/target/tcp.c
1613
ida_free(&nvmet_tcp_queue_ida, queue->idx);
drivers/nvme/target/tcp.c
1614
page_frag_cache_drain(&queue->pf_cache);
drivers/nvme/target/tcp.c
1615
kfree(queue);
drivers/nvme/target/tcp.c
1620
struct nvmet_tcp_queue *queue;
drivers/nvme/target/tcp.c
1625
queue = sk->sk_user_data;
drivers/nvme/target/tcp.c
1626
if (likely(queue)) {
drivers/nvme/target/tcp.c
1627
if (queue->data_ready)
drivers/nvme/target/tcp.c
1628
queue->data_ready(sk);
drivers/nvme/target/tcp.c
1629
if (queue->state != NVMET_TCP_Q_TLS_HANDSHAKE)
drivers/nvme/target/tcp.c
1630
queue_work_on(queue_cpu(queue), nvmet_tcp_wq,
drivers/nvme/target/tcp.c
1631
&queue->io_work);
drivers/nvme/target/tcp.c
1638
struct nvmet_tcp_queue *queue;
drivers/nvme/target/tcp.c
1641
queue = sk->sk_user_data;
drivers/nvme/target/tcp.c
1642
if (unlikely(!queue))
drivers/nvme/target/tcp.c
1645
if (unlikely(queue->state == NVMET_TCP_Q_CONNECTING)) {
drivers/nvme/target/tcp.c
1646
queue->write_space(sk);
drivers/nvme/target/tcp.c
1652
queue_work_on(queue_cpu(queue), nvmet_tcp_wq, &queue->io_work);
drivers/nvme/target/tcp.c
1660
struct nvmet_tcp_queue *queue;
drivers/nvme/target/tcp.c
1663
queue = sk->sk_user_data;
drivers/nvme/target/tcp.c
1664
if (!queue)
drivers/nvme/target/tcp.c
1675
nvmet_tcp_schedule_release_queue(queue);
drivers/nvme/target/tcp.c
1679
queue->idx, sk->sk_state);
drivers/nvme/target/tcp.c
1685
static int nvmet_tcp_set_queue_sock(struct nvmet_tcp_queue *queue)
drivers/nvme/target/tcp.c
1687
struct socket *sock = queue->sock;
drivers/nvme/target/tcp.c
1692
(struct sockaddr *)&queue->sockaddr);
drivers/nvme/target/tcp.c
1697
(struct sockaddr *)&queue->sockaddr_peer);
drivers/nvme/target/tcp.c
1724
sock->sk->sk_user_data = queue;
drivers/nvme/target/tcp.c
1725
queue->data_ready = sock->sk->sk_data_ready;
drivers/nvme/target/tcp.c
1727
queue->state_change = sock->sk->sk_state_change;
drivers/nvme/target/tcp.c
1729
queue->write_space = sock->sk->sk_write_space;
drivers/nvme/target/tcp.c
1732
nvmet_tcp_arm_queue_deadline(queue);
drivers/nvme/target/tcp.c
1733
queue_work_on(queue_cpu(queue), nvmet_tcp_wq, &queue->io_work);
drivers/nvme/target/tcp.c
1741
static int nvmet_tcp_try_peek_pdu(struct nvmet_tcp_queue *queue)
drivers/nvme/target/tcp.c
1743
struct nvme_tcp_hdr *hdr = &queue->pdu.cmd.hdr;
drivers/nvme/target/tcp.c
1746
.iov_base = (u8 *)&queue->pdu + queue->offset,
drivers/nvme/target/tcp.c
1756
if (nvmet_port_secure_channel_required(queue->port->nport))
drivers/nvme/target/tcp.c
1759
len = kernel_recvmsg(queue->sock, &msg, &iov, 1,
drivers/nvme/target/tcp.c
1763
queue->idx, len);
drivers/nvme/target/tcp.c
1767
ret = nvmet_tcp_tls_record_ok(queue, &msg, cbuf);
drivers/nvme/target/tcp.c
1773
queue->idx, (int)iov.iov_len - len);
drivers/nvme/target/tcp.c
1777
queue->idx, hdr->type, hdr->hlen, hdr->plen,
drivers/nvme/target/tcp.c
1783
queue->idx);
drivers/nvme/target/tcp.c
1789
static int nvmet_tcp_tls_key_lookup(struct nvmet_tcp_queue *queue,
drivers/nvme/target/tcp.c
1797
__func__, queue->idx, peerid);
drivers/nvme/target/tcp.c
1798
spin_lock_bh(&queue->state_lock);
drivers/nvme/target/tcp.c
1799
queue->state = NVMET_TCP_Q_FAILED;
drivers/nvme/target/tcp.c
1800
spin_unlock_bh(&queue->state_lock);
drivers/nvme/target/tcp.c
1804
__func__, queue->idx, peerid);
drivers/nvme/target/tcp.c
1805
queue->nvme_sq.tls_key = tls_key;
drivers/nvme/target/tcp.c
1813
struct nvmet_tcp_queue *queue = data;
drivers/nvme/target/tcp.c
1816
queue->idx, peerid, status);
drivers/nvme/target/tcp.c
1817
spin_lock_bh(&queue->state_lock);
drivers/nvme/target/tcp.c
1818
if (WARN_ON(queue->state != NVMET_TCP_Q_TLS_HANDSHAKE)) {
drivers/nvme/target/tcp.c
1819
spin_unlock_bh(&queue->state_lock);
drivers/nvme/target/tcp.c
1823
queue->tls_pskid = peerid;
drivers/nvme/target/tcp.c
1824
queue->state = NVMET_TCP_Q_CONNECTING;
drivers/nvme/target/tcp.c
1826
queue->state = NVMET_TCP_Q_FAILED;
drivers/nvme/target/tcp.c
1827
spin_unlock_bh(&queue->state_lock);
drivers/nvme/target/tcp.c
1829
cancel_delayed_work_sync(&queue->tls_handshake_tmo_work);
drivers/nvme/target/tcp.c
1832
status = nvmet_tcp_tls_key_lookup(queue, peerid);
drivers/nvme/target/tcp.c
1835
nvmet_tcp_schedule_release_queue(queue);
drivers/nvme/target/tcp.c
1837
nvmet_tcp_set_queue_sock(queue);
drivers/nvme/target/tcp.c
1838
kref_put(&queue->kref, nvmet_tcp_release_queue);
drivers/nvme/target/tcp.c
1843
struct nvmet_tcp_queue *queue = container_of(to_delayed_work(w),
drivers/nvme/target/tcp.c
1846
pr_warn("queue %d: TLS handshake timeout\n", queue->idx);
drivers/nvme/target/tcp.c
1850
if (!tls_handshake_cancel(queue->sock->sk))
drivers/nvme/target/tcp.c
1852
spin_lock_bh(&queue->state_lock);
drivers/nvme/target/tcp.c
1853
if (WARN_ON(queue->state != NVMET_TCP_Q_TLS_HANDSHAKE)) {
drivers/nvme/target/tcp.c
1854
spin_unlock_bh(&queue->state_lock);
drivers/nvme/target/tcp.c
1857
queue->state = NVMET_TCP_Q_FAILED;
drivers/nvme/target/tcp.c
1858
spin_unlock_bh(&queue->state_lock);
drivers/nvme/target/tcp.c
1859
nvmet_tcp_schedule_release_queue(queue);
drivers/nvme/target/tcp.c
1860
kref_put(&queue->kref, nvmet_tcp_release_queue);
drivers/nvme/target/tcp.c
1863
static int nvmet_tcp_tls_handshake(struct nvmet_tcp_queue *queue)
drivers/nvme/target/tcp.c
1868
if (queue->state != NVMET_TCP_Q_TLS_HANDSHAKE) {
drivers/nvme/target/tcp.c
1869
pr_warn("cannot start TLS in state %d\n", queue->state);
drivers/nvme/target/tcp.c
1873
kref_get(&queue->kref);
drivers/nvme/target/tcp.c
1874
pr_debug("queue %d: TLS ServerHello\n", queue->idx);
drivers/nvme/target/tcp.c
1876
args.ta_sock = queue->sock;
drivers/nvme/target/tcp.c
1878
args.ta_data = queue;
drivers/nvme/target/tcp.c
1879
args.ta_keyring = key_serial(queue->port->nport->keyring);
drivers/nvme/target/tcp.c
1884
kref_put(&queue->kref, nvmet_tcp_release_queue);
drivers/nvme/target/tcp.c
1887
queue_delayed_work(nvmet_wq, &queue->tls_handshake_tmo_work,
drivers/nvme/target/tcp.c
1899
struct nvmet_tcp_queue *queue;
drivers/nvme/target/tcp.c
1903
queue = kzalloc_obj(*queue);
drivers/nvme/target/tcp.c
1904
if (!queue) {
drivers/nvme/target/tcp.c
1909
INIT_WORK(&queue->release_work, nvmet_tcp_release_queue_work);
drivers/nvme/target/tcp.c
1910
INIT_WORK(&queue->io_work, nvmet_tcp_io_work);
drivers/nvme/target/tcp.c
1911
kref_init(&queue->kref);
drivers/nvme/target/tcp.c
1912
queue->sock = newsock;
drivers/nvme/target/tcp.c
1913
queue->port = port;
drivers/nvme/target/tcp.c
1914
queue->nr_cmds = 0;
drivers/nvme/target/tcp.c
1915
spin_lock_init(&queue->state_lock);
drivers/nvme/target/tcp.c
1916
if (queue->port->nport->disc_addr.tsas.tcp.sectype ==
drivers/nvme/target/tcp.c
1918
queue->state = NVMET_TCP_Q_TLS_HANDSHAKE;
drivers/nvme/target/tcp.c
1920
queue->state = NVMET_TCP_Q_CONNECTING;
drivers/nvme/target/tcp.c
1921
INIT_LIST_HEAD(&queue->free_list);
drivers/nvme/target/tcp.c
1922
init_llist_head(&queue->resp_list);
drivers/nvme/target/tcp.c
1923
INIT_LIST_HEAD(&queue->resp_send_list);
drivers/nvme/target/tcp.c
1925
sock_file = sock_alloc_file(queue->sock, O_CLOEXEC, NULL);
drivers/nvme/target/tcp.c
1931
queue->idx = ida_alloc(&nvmet_tcp_queue_ida, GFP_KERNEL);
drivers/nvme/target/tcp.c
1932
if (queue->idx < 0) {
drivers/nvme/target/tcp.c
1933
ret = queue->idx;
drivers/nvme/target/tcp.c
1937
ret = nvmet_tcp_alloc_cmd(queue, &queue->connect);
drivers/nvme/target/tcp.c
1941
nvmet_cq_init(&queue->nvme_cq);
drivers/nvme/target/tcp.c
1942
ret = nvmet_sq_init(&queue->nvme_sq, &queue->nvme_cq);
drivers/nvme/target/tcp.c
1946
nvmet_prepare_receive_pdu(queue);
drivers/nvme/target/tcp.c
1949
list_add_tail(&queue->queue_list, &nvmet_tcp_queue_list);
drivers/nvme/target/tcp.c
1952
INIT_DELAYED_WORK(&queue->tls_handshake_tmo_work,
drivers/nvme/target/tcp.c
1955
if (queue->state == NVMET_TCP_Q_TLS_HANDSHAKE) {
drivers/nvme/target/tcp.c
1956
struct sock *sk = queue->sock->sk;
drivers/nvme/target/tcp.c
1963
if (!nvmet_tcp_try_peek_pdu(queue)) {
drivers/nvme/target/tcp.c
1964
if (!nvmet_tcp_tls_handshake(queue))
drivers/nvme/target/tcp.c
1970
queue->state = NVMET_TCP_Q_CONNECTING;
drivers/nvme/target/tcp.c
1974
ret = nvmet_tcp_set_queue_sock(queue);
drivers/nvme/target/tcp.c
1981
list_del_init(&queue->queue_list);
drivers/nvme/target/tcp.c
1983
nvmet_sq_destroy(&queue->nvme_sq);
drivers/nvme/target/tcp.c
1985
nvmet_cq_put(&queue->nvme_cq);
drivers/nvme/target/tcp.c
1986
nvmet_tcp_free_cmd(&queue->connect);
drivers/nvme/target/tcp.c
1988
ida_free(&nvmet_tcp_queue_ida, queue->idx);
drivers/nvme/target/tcp.c
1990
fput(queue->sock->file);
drivers/nvme/target/tcp.c
1992
kfree(queue);
drivers/nvme/target/tcp.c
2113
struct nvmet_tcp_queue *queue;
drivers/nvme/target/tcp.c
2116
list_for_each_entry(queue, &nvmet_tcp_queue_list, queue_list)
drivers/nvme/target/tcp.c
2117
if (queue->port == port)
drivers/nvme/target/tcp.c
2118
kernel_sock_shutdown(queue->sock, SHUT_RDWR);
drivers/nvme/target/tcp.c
2143
struct nvmet_tcp_queue *queue;
drivers/nvme/target/tcp.c
2146
list_for_each_entry(queue, &nvmet_tcp_queue_list, queue_list)
drivers/nvme/target/tcp.c
2147
if (queue->nvme_sq.ctrl == ctrl)
drivers/nvme/target/tcp.c
2148
kernel_sock_shutdown(queue->sock, SHUT_RDWR);
drivers/nvme/target/tcp.c
2154
struct nvmet_tcp_queue *queue =
drivers/nvme/target/tcp.c
2173
queue->nr_cmds = sq->size * 2;
drivers/nvme/target/tcp.c
2174
if (nvmet_tcp_alloc_cmds(queue)) {
drivers/nvme/target/tcp.c
2175
queue->nr_cmds = 0;
drivers/nvme/target/tcp.c
218
static inline u16 nvmet_tcp_cmd_tag(struct nvmet_tcp_queue *queue,
drivers/nvme/target/tcp.c
2189
struct nvmet_tcp_queue *queue = cmd->queue;
drivers/nvme/target/tcp.c
2191
sprintf(traddr, "%pISc", (struct sockaddr *)&queue->sockaddr);
drivers/nvme/target/tcp.c
2201
struct nvmet_tcp_queue *queue =
drivers/nvme/target/tcp.c
2204
if (queue->sockaddr_peer.ss_family == AF_UNSPEC)
drivers/nvme/target/tcp.c
2207
(struct sockaddr *)&queue->sockaddr_peer);
drivers/nvme/target/tcp.c
221
if (unlikely(!queue->nr_cmds)) {
drivers/nvme/target/tcp.c
2244
struct nvmet_tcp_queue *queue;
drivers/nvme/target/tcp.c
2250
list_for_each_entry(queue, &nvmet_tcp_queue_list, queue_list)
drivers/nvme/target/tcp.c
2251
kernel_sock_shutdown(queue->sock, SHUT_RDWR);
drivers/nvme/target/tcp.c
226
return cmd - queue->cmds;
drivers/nvme/target/tcp.c
254
nvmet_tcp_get_cmd(struct nvmet_tcp_queue *queue)
drivers/nvme/target/tcp.c
258
cmd = list_first_entry_or_null(&queue->free_list,
drivers/nvme/target/tcp.c
274
if (unlikely(cmd == &cmd->queue->connect))
drivers/nvme/target/tcp.c
277
list_add_tail(&cmd->entry, &cmd->queue->free_list);
drivers/nvme/target/tcp.c
280
static inline int queue_cpu(struct nvmet_tcp_queue *queue)
drivers/nvme/target/tcp.c
282
return queue->sock->sk->sk_incoming_cpu;
drivers/nvme/target/tcp.c
285
static inline u8 nvmet_tcp_hdgst_len(struct nvmet_tcp_queue *queue)
drivers/nvme/target/tcp.c
287
return queue->hdr_digest ? NVME_TCP_DIGEST_LENGTH : 0;
drivers/nvme/target/tcp.c
290
static inline u8 nvmet_tcp_ddgst_len(struct nvmet_tcp_queue *queue)
drivers/nvme/target/tcp.c
292
return queue->data_digest ? NVME_TCP_DIGEST_LENGTH : 0;
drivers/nvme/target/tcp.c
300
static int nvmet_tcp_verify_hdgst(struct nvmet_tcp_queue *queue,
drivers/nvme/target/tcp.c
309
queue->idx);
drivers/nvme/target/tcp.c
318
queue->idx, le32_to_cpu(recv_digest),
drivers/nvme/target/tcp.c
326
static int nvmet_tcp_check_ddgst(struct nvmet_tcp_queue *queue, void *pdu)
drivers/nvme/target/tcp.c
329
u8 digest_len = nvmet_tcp_hdgst_len(queue);
drivers/nvme/target/tcp.c
336
pr_err("queue %d: data digest flag is cleared\n", queue->idx);
drivers/nvme/target/tcp.c
352
static void nvmet_tcp_fatal_error(struct nvmet_tcp_queue *queue);
drivers/nvme/target/tcp.c
368
nvmet_tcp_fatal_error(cmd->queue);
drivers/nvme/target/tcp.c
376
nvmet_tcp_fatal_error(cmd->queue);
drivers/nvme/target/tcp.c
380
nvmet_tcp_fatal_error(cmd->queue);
drivers/nvme/target/tcp.c
399
static void nvmet_tcp_fatal_error(struct nvmet_tcp_queue *queue)
drivers/nvme/target/tcp.c
401
queue->rcv_state = NVMET_TCP_RECV_ERR;
drivers/nvme/target/tcp.c
402
if (queue->nvme_sq.ctrl)
drivers/nvme/target/tcp.c
403
nvmet_ctrl_fatal_error(queue->nvme_sq.ctrl);
drivers/nvme/target/tcp.c
405
kernel_sock_shutdown(queue->sock, SHUT_RDWR);
drivers/nvme/target/tcp.c
408
static void nvmet_tcp_socket_error(struct nvmet_tcp_queue *queue, int status)
drivers/nvme/target/tcp.c
410
queue->rcv_state = NVMET_TCP_RECV_ERR;
drivers/nvme/target/tcp.c
412
kernel_sock_shutdown(queue->sock, SHUT_RDWR);
drivers/nvme/target/tcp.c
414
nvmet_tcp_fatal_error(queue);
drivers/nvme/target/tcp.c
476
struct nvmet_tcp_queue *queue = cmd->queue;
drivers/nvme/target/tcp.c
477
u8 hdgst = nvmet_tcp_hdgst_len(cmd->queue);
drivers/nvme/target/tcp.c
478
u8 ddgst = nvmet_tcp_ddgst_len(cmd->queue);
drivers/nvme/target/tcp.c
484
pdu->hdr.flags = NVME_TCP_F_DATA_LAST | (queue->nvme_sq.sqhd_disabled ?
drivers/nvme/target/tcp.c
495
if (queue->data_digest) {
drivers/nvme/target/tcp.c
500
if (cmd->queue->hdr_digest) {
drivers/nvme/target/tcp.c
509
u8 hdgst = nvmet_tcp_hdgst_len(cmd->queue);
drivers/nvme/target/tcp.c
521
pdu->ttag = nvmet_tcp_cmd_tag(cmd->queue, cmd);
drivers/nvme/target/tcp.c
524
if (cmd->queue->hdr_digest) {
drivers/nvme/target/tcp.c
533
u8 hdgst = nvmet_tcp_hdgst_len(cmd->queue);
drivers/nvme/target/tcp.c
543
if (cmd->queue->hdr_digest) {
drivers/nvme/target/tcp.c
549
static void nvmet_tcp_process_resp_list(struct nvmet_tcp_queue *queue)
drivers/nvme/target/tcp.c
554
for (node = llist_del_all(&queue->resp_list); node; node = node->next) {
drivers/nvme/target/tcp.c
556
list_add(&cmd->entry, &queue->resp_send_list);
drivers/nvme/target/tcp.c
557
queue->send_list_len++;
drivers/nvme/target/tcp.c
561
static struct nvmet_tcp_cmd *nvmet_tcp_fetch_cmd(struct nvmet_tcp_queue *queue)
drivers/nvme/target/tcp.c
563
queue->snd_cmd = list_first_entry_or_null(&queue->resp_send_list,
drivers/nvme/target/tcp.c
565
if (!queue->snd_cmd) {
drivers/nvme/target/tcp.c
566
nvmet_tcp_process_resp_list(queue);
drivers/nvme/target/tcp.c
567
queue->snd_cmd =
drivers/nvme/target/tcp.c
568
list_first_entry_or_null(&queue->resp_send_list,
drivers/nvme/target/tcp.c
570
if (unlikely(!queue->snd_cmd))
drivers/nvme/target/tcp.c
574
list_del_init(&queue->snd_cmd->entry);
drivers/nvme/target/tcp.c
575
queue->send_list_len--;
drivers/nvme/target/tcp.c
577
if (nvmet_tcp_need_data_out(queue->snd_cmd))
drivers/nvme/target/tcp.c
578
nvmet_setup_c2h_data_pdu(queue->snd_cmd);
drivers/nvme/target/tcp.c
579
else if (nvmet_tcp_need_data_in(queue->snd_cmd))
drivers/nvme/target/tcp.c
580
nvmet_setup_r2t_pdu(queue->snd_cmd);
drivers/nvme/target/tcp.c
582
nvmet_setup_response_pdu(queue->snd_cmd);
drivers/nvme/target/tcp.c
584
return queue->snd_cmd;
drivers/nvme/target/tcp.c
591
struct nvmet_tcp_queue *queue = cmd->queue;
drivers/nvme/target/tcp.c
598
queue_state = smp_load_acquire(&queue->rcv_state);
drivers/nvme/target/tcp.c
599
queue_cmd = READ_ONCE(queue->cmd);
drivers/nvme/target/tcp.c
616
llist_add(&cmd->lentry, &queue->resp_list);
drivers/nvme/target/tcp.c
617
queue_work_on(queue_cpu(queue), nvmet_tcp_wq, &cmd->queue->io_work);
drivers/nvme/target/tcp.c
634
u8 hdgst = nvmet_tcp_hdgst_len(cmd->queue);
drivers/nvme/target/tcp.c
640
ret = sock_sendmsg(cmd->queue->sock, &msg);
drivers/nvme/target/tcp.c
657
struct nvmet_tcp_queue *queue = cmd->queue;
drivers/nvme/target/tcp.c
668
if ((!last_in_batch && cmd->queue->send_list_len) ||
drivers/nvme/target/tcp.c
670
queue->data_digest || !queue->nvme_sq.sqhd_disabled)
drivers/nvme/target/tcp.c
675
ret = sock_sendmsg(cmd->queue->sock, &msg);
drivers/nvme/target/tcp.c
689
if (queue->data_digest) {
drivers/nvme/target/tcp.c
693
if (queue->nvme_sq.sqhd_disabled) {
drivers/nvme/target/tcp.c
694
cmd->queue->snd_cmd = NULL;
drivers/nvme/target/tcp.c
701
if (queue->nvme_sq.sqhd_disabled)
drivers/nvme/target/tcp.c
713
u8 hdgst = nvmet_tcp_hdgst_len(cmd->queue);
drivers/nvme/target/tcp.c
717
if (!last_in_batch && cmd->queue->send_list_len)
drivers/nvme/target/tcp.c
724
ret = sock_sendmsg(cmd->queue->sock, &msg);
drivers/nvme/target/tcp.c
734
cmd->queue->snd_cmd = NULL;
drivers/nvme/target/tcp.c
743
u8 hdgst = nvmet_tcp_hdgst_len(cmd->queue);
drivers/nvme/target/tcp.c
747
if (!last_in_batch && cmd->queue->send_list_len)
drivers/nvme/target/tcp.c
754
ret = sock_sendmsg(cmd->queue->sock, &msg);
drivers/nvme/target/tcp.c
763
cmd->queue->snd_cmd = NULL;
drivers/nvme/target/tcp.c
769
struct nvmet_tcp_queue *queue = cmd->queue;
drivers/nvme/target/tcp.c
778
if (!last_in_batch && cmd->queue->send_list_len)
drivers/nvme/target/tcp.c
783
ret = kernel_sendmsg(queue->sock, &msg, &iov, 1, iov.iov_len);
drivers/nvme/target/tcp.c
793
if (queue->nvme_sq.sqhd_disabled) {
drivers/nvme/target/tcp.c
794
cmd->queue->snd_cmd = NULL;
drivers/nvme/target/tcp.c
802
static int nvmet_tcp_try_send_one(struct nvmet_tcp_queue *queue,
drivers/nvme/target/tcp.c
805
struct nvmet_tcp_cmd *cmd = queue->snd_cmd;
drivers/nvme/target/tcp.c
808
if (!cmd || queue->state == NVMET_TCP_Q_DISCONNECTING) {
drivers/nvme/target/tcp.c
809
cmd = nvmet_tcp_fetch_cmd(queue);
drivers/nvme/target/tcp.c
851
static int nvmet_tcp_try_send(struct nvmet_tcp_queue *queue,
drivers/nvme/target/tcp.c
857
ret = nvmet_tcp_try_send_one(queue, i == budget - 1);
drivers/nvme/target/tcp.c
859
nvmet_tcp_socket_error(queue, ret);
drivers/nvme/target/tcp.c
870
static void nvmet_prepare_receive_pdu(struct nvmet_tcp_queue *queue)
drivers/nvme/target/tcp.c
872
queue->offset = 0;
drivers/nvme/target/tcp.c
873
queue->left = sizeof(struct nvme_tcp_hdr);
drivers/nvme/target/tcp.c
874
WRITE_ONCE(queue->cmd, NULL);
drivers/nvme/target/tcp.c
876
smp_store_release(&queue->rcv_state, NVMET_TCP_RECV_PDU);
drivers/nvme/target/tcp.c
879
static int nvmet_tcp_handle_icreq(struct nvmet_tcp_queue *queue)
drivers/nvme/target/tcp.c
881
struct nvme_tcp_icreq_pdu *icreq = &queue->pdu.icreq;
drivers/nvme/target/tcp.c
882
struct nvme_tcp_icresp_pdu *icresp = &queue->pdu.icresp;
drivers/nvme/target/tcp.c
890
nvmet_tcp_fatal_error(queue);
drivers/nvme/target/tcp.c
895
pr_err("queue %d: bad pfv %d\n", queue->idx, icreq->pfv);
drivers/nvme/target/tcp.c
900
pr_err("queue %d: unsupported hpda %d\n", queue->idx,
drivers/nvme/target/tcp.c
905
queue->hdr_digest = !!(icreq->digest & NVME_TCP_HDR_DIGEST_ENABLE);
drivers/nvme/target/tcp.c
906
queue->data_digest = !!(icreq->digest & NVME_TCP_DATA_DIGEST_ENABLE);
drivers/nvme/target/tcp.c
916
if (queue->hdr_digest)
drivers/nvme/target/tcp.c
918
if (queue->data_digest)
drivers/nvme/target/tcp.c
923
ret = kernel_sendmsg(queue->sock, &msg, &iov, 1, iov.iov_len);
drivers/nvme/target/tcp.c
925
queue->state = NVMET_TCP_Q_FAILED;
drivers/nvme/target/tcp.c
929
queue->state = NVMET_TCP_Q_LIVE;
drivers/nvme/target/tcp.c
930
nvmet_prepare_receive_pdu(queue);
drivers/nvme/target/tcp.c
934
static void nvmet_tcp_handle_req_failure(struct nvmet_tcp_queue *queue,
drivers/nvme/target/tcp.c
949
nvmet_prepare_receive_pdu(queue);
drivers/nvme/target/tcp.c
955
pr_err("queue %d: failed to map data\n", queue->idx);
drivers/nvme/target/tcp.c
956
nvmet_tcp_fatal_error(queue);
drivers/nvme/target/tcp.c
960
queue->rcv_state = NVMET_TCP_RECV_DATA;
drivers/nvme/target/tcp.c
965
static int nvmet_tcp_handle_h2c_data_pdu(struct nvmet_tcp_queue *queue)
drivers/nvme/target/tcp.c
967
struct nvme_tcp_data_pdu *data = &queue->pdu.data;
drivers/nvme/target/tcp.c
971
if (likely(queue->nr_cmds)) {
drivers/nvme/target/tcp.c
972
if (unlikely(data->ttag >= queue->nr_cmds)) {
drivers/nvme/target/tcp.c
974
queue->idx, data->ttag, queue->nr_cmds);
drivers/nvme/target/tcp.c
977
cmd = &queue->cmds[data->ttag];
drivers/nvme/target/tcp.c
979
cmd = &queue->connect;
drivers/nvme/target/tcp.c
990
nvmet_tcp_hdgst_len(queue) -
drivers/nvme/target/tcp.c
991
nvmet_tcp_ddgst_len(queue) -
drivers/pci/hotplug/cpqphp.h
308
wait_queue_head_t queue; /* sleep & wake process */
drivers/pci/hotplug/cpqphp.h
710
add_wait_queue(&ctrl->queue, &wait);
drivers/pci/hotplug/cpqphp.h
713
remove_wait_queue(&ctrl->queue, &wait);
drivers/pci/hotplug/cpqphp_core.c
1035
init_waitqueue_head(&ctrl->queue);
drivers/pci/hotplug/cpqphp_ctrl.c
906
wake_up_interruptible(&ctrl->queue);
drivers/pci/hotplug/cpqphp_ctrl.c
936
wake_up_interruptible(&ctrl->queue);
drivers/pci/hotplug/pciehp.h
102
wait_queue_head_t queue;
drivers/pci/hotplug/pciehp_hpc.c
1037
init_waitqueue_head(&ctrl->queue);
drivers/pci/hotplug/pciehp_hpc.c
142
rc = wait_event_timeout(ctrl->queue, !ctrl->cmd_busy, timeout);
drivers/pci/hotplug/pciehp_hpc.c
710
wake_up(&ctrl->queue);
drivers/pci/hotplug/shpchp.h
80
wait_queue_head_t queue; /* sleep & wake process */
drivers/pci/hotplug/shpchp_hpc.c
274
rc = wait_event_interruptible_timeout(ctrl->queue,
drivers/pci/hotplug/shpchp_hpc.c
786
wake_up_interruptible(&ctrl->queue);
drivers/pci/hotplug/shpchp_hpc.c
956
init_waitqueue_head(&ctrl->queue);
drivers/perf/fsl_imx8_ddr_perf.c
273
IMX8_DDR_PMU_EVENT_ATTR(read-queue-depth, 0x08),
drivers/perf/fsl_imx8_ddr_perf.c
274
IMX8_DDR_PMU_EVENT_ATTR(write-queue-depth, 0x09),
drivers/perf/hisilicon/hns3_pmu.c
337
HNS3_PMU_FILTER_ATTR(queue, config1, 24, 39);
drivers/perf/hisilicon/hns3_pmu.c
658
HNS3_PMU_FORMAT_ATTR(queue, "config1:24-39"),
drivers/perf/hisilicon/hns3_pmu.c
836
u16 queue)
drivers/perf/hisilicon/hns3_pmu.c
841
val |= (u32)queue << HNS3_PMU_QID_PARA_QUEUE_S;
drivers/perf/hisilicon/hns3_pmu.c
873
u16 queue)
drivers/perf/hisilicon/hns3_pmu.c
875
hns3_pmu_set_qid_para(hns3_pmu, idx, bdf, queue);
drivers/perf/xgene_pmu.c
341
XGENE_PMU_EVENT_ATTR(collision-queue-not-empty, 0x10),
drivers/perf/xgene_pmu.c
342
XGENE_PMU_EVENT_ATTR(collision-queue-full, 0x11),
drivers/perf/xgene_pmu.c
551
XGENE_PMU_EVENT_ATTR(queue-fill-gt-thresh, 0x22),
drivers/perf/xgene_pmu.c
552
XGENE_PMU_EVENT_ATTR(queue-rds-gt-thresh, 0x23),
drivers/perf/xgene_pmu.c
553
XGENE_PMU_EVENT_ATTR(queue-wrs-gt-thresh, 0x24),
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
1361
notify_bulks(struct vchiq_service *service, struct vchiq_bulk_queue *queue,
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
1369
(queue == &service->bulk_tx) ? 't' : 'r',
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
1370
queue->process, queue->remote_notify, queue->remove);
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
1372
queue->remote_notify = queue->process;
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
1374
while (queue->remove != queue->remote_notify) {
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
1376
&queue->bulks[BULK_INDEX(queue->remove)];
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
1388
queue->remove++;
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
1395
request_poll(service->state, service, (queue == &service->bulk_tx) ?
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
1786
struct vchiq_bulk_queue *queue)
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
1788
int is_tx = (queue == &service->bulk_tx);
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
1793
is_tx ? 't' : 'r', queue->local_insert,
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
1794
queue->remote_insert, queue->process);
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
1796
WARN_ON((int)(queue->local_insert - queue->process) < 0);
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
1797
WARN_ON((int)(queue->remote_insert - queue->process) < 0);
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
1799
while ((queue->process != queue->local_insert) ||
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
1800
(queue->process != queue->remote_insert)) {
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
1801
struct vchiq_bulk *bulk = &queue->bulks[BULK_INDEX(queue->process)];
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
1803
if (queue->process == queue->remote_insert) {
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
1807
queue->remote_insert++;
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
1810
if (queue->process != queue->local_insert) {
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
1825
queue->local_insert++;
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
1828
queue->process++;
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
2091
struct vchiq_bulk_queue *queue;
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
2094
queue = (type == VCHIQ_MSG_BULK_RX_DONE) ?
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
2102
if ((int)(queue->remote_insert -
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
2103
queue->local_insert) >= 0) {
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
2107
localport, queue->remote_insert, queue->local_insert);
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
2111
if (queue->process != queue->remote_insert) {
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
2113
__func__, queue->process,
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
2114
queue->remote_insert);
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
2119
bulk = &queue->bulks[BULK_INDEX(queue->remote_insert)];
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
2121
queue->remote_insert++;
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
2130
queue->local_insert, queue->remote_insert, queue->process);
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
2133
WARN_ON(queue->process == queue->local_insert);
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
2135
queue->process++;
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
2138
notify_bulks(service, queue, RETRY_POLL);
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
3037
struct vchiq_bulk_queue *queue;
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
3054
queue = (bulk_params->dir == VCHIQ_BULK_TRANSMIT) ?
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
3060
if (queue->local_insert == queue->remove + VCHIQ_NUM_SERVICE_BULKS) {
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
3068
} while (queue->local_insert == queue->remove +
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
3072
bulk = &queue->bulks[BULK_INDEX(queue->local_insert)];
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
3126
queue->local_insert++;
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
3132
state->id, service->localport, dir_char, queue->local_insert,
drivers/platform/raspberrypi/vchiq-interface/vchiq_core.c
3133
queue->remote_insert, queue->process);
drivers/platform/surface/aggregator/controller.c
715
return &cplt->event.target[tidx].queue[event];
drivers/platform/surface/aggregator/controller.c
778
struct ssam_event_queue *queue;
drivers/platform/surface/aggregator/controller.c
784
queue = container_of(work, struct ssam_event_queue, work);
drivers/platform/surface/aggregator/controller.c
785
nf = &queue->cplt->event.notif;
drivers/platform/surface/aggregator/controller.c
786
dev = queue->cplt->dev;
drivers/platform/surface/aggregator/controller.c
790
item = ssam_event_queue_pop(queue);
drivers/platform/surface/aggregator/controller.c
798
if (!ssam_event_queue_is_empty(queue))
drivers/platform/surface/aggregator/controller.c
799
ssam_cplt_submit(queue->cplt, &queue->work);
drivers/platform/surface/aggregator/controller.c
835
for (i = 0; i < ARRAY_SIZE(target->queue); i++)
drivers/platform/surface/aggregator/controller.c
836
ssam_event_queue_init(cplt, &target->queue[i]);
drivers/platform/surface/aggregator/controller.h
118
struct ssam_event_queue queue[SSH_NUM_EVENTS];
drivers/platform/surface/aggregator/ssh_packet_layer.c
1372
spin_lock(&packet->ptl->queue.lock);
drivers/platform/surface/aggregator/ssh_packet_layer.c
1377
spin_unlock(&packet->ptl->queue.lock);
drivers/platform/surface/aggregator/ssh_packet_layer.c
1388
spin_unlock(&packet->ptl->queue.lock);
drivers/platform/surface/aggregator/ssh_packet_layer.c
1394
spin_unlock(&packet->ptl->queue.lock);
drivers/platform/surface/aggregator/ssh_packet_layer.c
1967
spin_lock(&ptl->queue.lock);
drivers/platform/surface/aggregator/ssh_packet_layer.c
1968
list_for_each_entry_safe(p, n, &ptl->queue.head, queue_node) {
drivers/platform/surface/aggregator/ssh_packet_layer.c
1976
spin_unlock(&ptl->queue.lock);
drivers/platform/surface/aggregator/ssh_packet_layer.c
2034
spin_lock_init(&ptl->queue.lock);
drivers/platform/surface/aggregator/ssh_packet_layer.c
2035
INIT_LIST_HEAD(&ptl->queue.head);
drivers/platform/surface/aggregator/ssh_packet_layer.c
686
lockdep_assert_held(&p->ptl->queue.lock);
drivers/platform/surface/aggregator/ssh_packet_layer.c
702
lockdep_assert_held(&p->ptl->queue.lock);
drivers/platform/surface/aggregator/ssh_packet_layer.c
718
list_for_each(head, &p->ptl->queue.head) {
drivers/platform/surface/aggregator/ssh_packet_layer.c
725
list_for_each_prev(head, &p->ptl->queue.head) {
drivers/platform/surface/aggregator/ssh_packet_layer.c
744
lockdep_assert_held(&ptl->queue.lock);
drivers/platform/surface/aggregator/ssh_packet_layer.c
767
spin_lock(&packet->ptl->queue.lock);
drivers/platform/surface/aggregator/ssh_packet_layer.c
769
spin_unlock(&packet->ptl->queue.lock);
drivers/platform/surface/aggregator/ssh_packet_layer.c
778
spin_lock(&ptl->queue.lock);
drivers/platform/surface/aggregator/ssh_packet_layer.c
781
spin_unlock(&ptl->queue.lock);
drivers/platform/surface/aggregator/ssh_packet_layer.c
787
spin_unlock(&ptl->queue.lock);
drivers/platform/surface/aggregator/ssh_packet_layer.c
907
spin_lock(&ptl->queue.lock);
drivers/platform/surface/aggregator/ssh_packet_layer.c
908
list_for_each_entry_safe(p, n, &ptl->queue.head, queue_node) {
drivers/platform/surface/aggregator/ssh_packet_layer.c
950
spin_unlock(&ptl->queue.lock);
drivers/platform/surface/aggregator/ssh_packet_layer.h
84
} queue;
drivers/platform/surface/aggregator/ssh_request_layer.c
1037
spin_lock_init(&rtl->queue.lock);
drivers/platform/surface/aggregator/ssh_request_layer.c
1038
INIT_LIST_HEAD(&rtl->queue.head);
drivers/platform/surface/aggregator/ssh_request_layer.c
104
spin_lock(&rtl->queue.lock);
drivers/platform/surface/aggregator/ssh_request_layer.c
107
spin_unlock(&rtl->queue.lock);
drivers/platform/surface/aggregator/ssh_request_layer.c
113
spin_unlock(&rtl->queue.lock);
drivers/platform/surface/aggregator/ssh_request_layer.c
121
spin_lock(&rtl->queue.lock);
drivers/platform/surface/aggregator/ssh_request_layer.c
1213
spin_lock(&rtl->queue.lock);
drivers/platform/surface/aggregator/ssh_request_layer.c
1214
list_for_each_entry_safe(r, n, &rtl->queue.head, node) {
drivers/platform/surface/aggregator/ssh_request_layer.c
122
empty = list_empty(&rtl->queue.head);
drivers/platform/surface/aggregator/ssh_request_layer.c
1222
spin_unlock(&rtl->queue.lock);
drivers/platform/surface/aggregator/ssh_request_layer.c
123
spin_unlock(&rtl->queue.lock);
drivers/platform/surface/aggregator/ssh_request_layer.c
212
spin_lock(&rtl->queue.lock);
drivers/platform/surface/aggregator/ssh_request_layer.c
215
list_for_each_entry_safe(p, n, &rtl->queue.head, node) {
drivers/platform/surface/aggregator/ssh_request_layer.c
236
spin_unlock(&rtl->queue.lock);
drivers/platform/surface/aggregator/ssh_request_layer.c
374
spin_lock(&rtl->queue.lock);
drivers/platform/surface/aggregator/ssh_request_layer.c
392
spin_unlock(&rtl->queue.lock);
drivers/platform/surface/aggregator/ssh_request_layer.c
408
spin_unlock(&rtl->queue.lock);
drivers/platform/surface/aggregator/ssh_request_layer.c
413
spin_unlock(&rtl->queue.lock);
drivers/platform/surface/aggregator/ssh_request_layer.c
418
list_add_tail(&ssh_request_get(rqst)->node, &rtl->queue.head);
drivers/platform/surface/aggregator/ssh_request_layer.c
420
spin_unlock(&rtl->queue.lock);
drivers/platform/surface/aggregator/ssh_request_layer.c
628
spin_lock(&rtl->queue.lock);
drivers/platform/surface/aggregator/ssh_request_layer.c
639
spin_unlock(&rtl->queue.lock);
drivers/platform/surface/aggregator/ssh_request_layer.c
646
spin_unlock(&rtl->queue.lock);
drivers/platform/surface/aggregator/ssh_request_layer.h
72
} queue;
drivers/pps/generators/pps_gen.c
241
init_waitqueue_head(&pps_gen->queue);
drivers/pps/generators/pps_gen.c
297
wake_up_interruptible_all(&pps_gen->queue);
drivers/pps/generators/pps_gen.c
41
poll_wait(file, &pps_gen->queue, wait);
drivers/pps/generators/pps_gen.c
91
ret = wait_event_interruptible(pps_gen->queue,
drivers/pps/kapi.c
104
init_waitqueue_head(&pps->queue);
drivers/pps/kapi.c
212
wake_up_interruptible_all(&pps->queue);
drivers/pps/pps.c
42
poll_wait(file, &pps->queue, wait);
drivers/pps/pps.c
63
err = wait_event_interruptible(pps->queue,
drivers/pps/pps.c
76
pps->queue,
drivers/ptp/ptp_chardev.c
136
struct timestamp_event_queue *queue;
drivers/ptp/ptp_chardev.c
139
queue = kzalloc(sizeof(*queue), GFP_KERNEL);
drivers/ptp/ptp_chardev.c
140
if (!queue)
drivers/ptp/ptp_chardev.c
142
queue->mask = bitmap_alloc(PTP_MAX_CHANNELS, GFP_KERNEL);
drivers/ptp/ptp_chardev.c
143
if (!queue->mask) {
drivers/ptp/ptp_chardev.c
144
kfree(queue);
drivers/ptp/ptp_chardev.c
147
bitmap_set(queue->mask, 0, PTP_MAX_CHANNELS);
drivers/ptp/ptp_chardev.c
148
spin_lock_init(&queue->lock);
drivers/ptp/ptp_chardev.c
150
list_add_tail(&queue->qlist, &ptp->tsevqs);
drivers/ptp/ptp_chardev.c
151
pccontext->private_clkdata = queue;
drivers/ptp/ptp_chardev.c
154
sprintf(debugfsname, "0x%p", queue);
drivers/ptp/ptp_chardev.c
155
queue->debugfs_instance =
drivers/ptp/ptp_chardev.c
157
queue->dfs_bitmap.array = (u32 *)queue->mask;
drivers/ptp/ptp_chardev.c
158
queue->dfs_bitmap.n_elements =
drivers/ptp/ptp_chardev.c
160
debugfs_create_u32_array("mask", 0444, queue->debugfs_instance,
drivers/ptp/ptp_chardev.c
161
&queue->dfs_bitmap);
drivers/ptp/ptp_chardev.c
168
struct timestamp_event_queue *queue = pccontext->private_clkdata;
drivers/ptp/ptp_chardev.c
172
debugfs_remove(queue->debugfs_instance);
drivers/ptp/ptp_chardev.c
175
list_del(&queue->qlist);
drivers/ptp/ptp_chardev.c
176
bitmap_free(queue->mask);
drivers/ptp/ptp_chardev.c
177
kfree(queue);
drivers/ptp/ptp_chardev.c
584
struct timestamp_event_queue *queue;
drivers/ptp/ptp_chardev.c
586
queue = pccontext->private_clkdata;
drivers/ptp/ptp_chardev.c
587
if (!queue)
drivers/ptp/ptp_chardev.c
592
return queue_cnt(queue) ? EPOLLIN : 0;
drivers/ptp/ptp_chardev.c
601
struct timestamp_event_queue *queue;
drivers/ptp/ptp_chardev.c
605
queue = pccontext->private_clkdata;
drivers/ptp/ptp_chardev.c
606
if (!queue)
drivers/ptp/ptp_chardev.c
615
if (wait_event_interruptible(ptp->tsev_wq, ptp->defunct || queue_cnt(queue)))
drivers/ptp/ptp_chardev.c
625
scoped_guard(spinlock_irq, &queue->lock) {
drivers/ptp/ptp_chardev.c
626
size_t qcnt = min((size_t)queue_cnt(queue), cnt / sizeof(*event));
drivers/ptp/ptp_chardev.c
629
event[i] = queue->buf[queue->head];
drivers/ptp/ptp_chardev.c
631
WRITE_ONCE(queue->head, (queue->head + 1) % PTP_MAX_TIMESTAMPS);
drivers/ptp/ptp_clock.c
320
struct timestamp_event_queue *queue = NULL;
drivers/ptp/ptp_clock.c
347
queue = kzalloc_obj(*queue);
drivers/ptp/ptp_clock.c
348
if (!queue) {
drivers/ptp/ptp_clock.c
352
list_add_tail(&queue->qlist, &ptp->tsevqs);
drivers/ptp/ptp_clock.c
354
queue->mask = bitmap_alloc(PTP_MAX_CHANNELS, GFP_KERNEL);
drivers/ptp/ptp_clock.c
355
if (!queue->mask) {
drivers/ptp/ptp_clock.c
359
bitmap_set(queue->mask, 0, PTP_MAX_CHANNELS);
drivers/ptp/ptp_clock.c
360
spin_lock_init(&queue->lock);
drivers/ptp/ptp_clock.c
47
static void enqueue_external_timestamp(struct timestamp_event_queue *queue,
drivers/ptp/ptp_clock.c
476
bitmap_free(queue->mask);
drivers/ptp/ptp_clock.c
478
list_del(&queue->qlist);
drivers/ptp/ptp_clock.c
479
kfree(queue);
drivers/ptp/ptp_clock.c
67
spin_lock_irqsave(&queue->lock, flags);
drivers/ptp/ptp_clock.c
69
dst = &queue->buf[queue->tail];
drivers/ptp/ptp_clock.c
78
if (!queue_free(queue))
drivers/ptp/ptp_clock.c
79
WRITE_ONCE(queue->head, (queue->head + 1) % PTP_MAX_TIMESTAMPS);
drivers/ptp/ptp_clock.c
81
WRITE_ONCE(queue->tail, (queue->tail + 1) % PTP_MAX_TIMESTAMPS);
drivers/ptp/ptp_clock.c
83
spin_unlock_irqrestore(&queue->lock, flags);
drivers/ptp/ptp_sysfs.c
77
struct timestamp_event_queue *queue;
drivers/ptp/ptp_sysfs.c
88
queue = list_first_entry(&ptp->tsevqs, struct timestamp_event_queue,
drivers/ptp/ptp_sysfs.c
92
spin_lock_irqsave(&queue->lock, flags);
drivers/ptp/ptp_sysfs.c
93
qcnt = queue_cnt(queue);
drivers/ptp/ptp_sysfs.c
95
event = queue->buf[queue->head];
drivers/ptp/ptp_sysfs.c
97
WRITE_ONCE(queue->head, (queue->head + 1) % PTP_MAX_TIMESTAMPS);
drivers/ptp/ptp_sysfs.c
99
spin_unlock_irqrestore(&queue->lock, flags);
drivers/rapidio/devices/tsi721.h
705
struct list_head queue;
drivers/rapidio/devices/tsi721_dma.c
548
if (!desc && !bdma_chan->active_tx && !list_empty(&bdma_chan->queue)) {
drivers/rapidio/devices/tsi721_dma.c
549
desc = list_first_entry(&bdma_chan->queue,
drivers/rapidio/devices/tsi721_dma.c
716
list_add_tail(&desc->desc_node, &bdma_chan->queue);
drivers/rapidio/devices/tsi721_dma.c
923
list_splice_init(&bdma_chan->queue, &list);
drivers/rapidio/devices/tsi721_dma.c
989
INIT_LIST_HEAD(&bdma_chan->queue);
drivers/rpmsg/rpmsg_char.c
114
skb_queue_tail(&eptdev->queue, skb);
drivers/rpmsg/rpmsg_char.c
195
skb_queue_purge(&eptdev->queue);
drivers/rpmsg/rpmsg_char.c
216
if (skb_queue_empty(&eptdev->queue)) {
drivers/rpmsg/rpmsg_char.c
224
!skb_queue_empty(&eptdev->queue) ||
drivers/rpmsg/rpmsg_char.c
235
skb = skb_dequeue(&eptdev->queue);
drivers/rpmsg/rpmsg_char.c
303
if (!skb_queue_empty(&eptdev->queue))
drivers/rpmsg/rpmsg_char.c
422
skb_queue_head_init(&eptdev->queue);
drivers/rpmsg/rpmsg_char.c
70
struct sk_buff_head queue;
drivers/s390/block/dasd.c
1533
blk_mq_run_hw_queues(device->block->gdp->queue, true);
drivers/s390/block/dasd.c
2643
blk_mq_run_hw_queues(block->gdp->queue, true);
drivers/s390/block/dasd.c
3021
struct dasd_block *block = hctx->queue->queuedata;
drivers/s390/block/dasd.c
320
lim = queue_limits_start_update(block->gdp->queue);
drivers/s390/block/dasd.c
342
rc = queue_limits_commit_update(block->gdp->queue, &lim);
drivers/s390/block/dasd.c
3684
blk_mq_run_hw_queues(device->block->gdp->queue, true);
drivers/s390/block/dasd.c
3839
blk_mq_run_hw_queues(device->block->gdp->queue, true);
drivers/s390/block/dasd_devmap.c
1710
blk_queue_rq_timeout(device->block->gdp->queue, val * HZ);
drivers/s390/block/scm_blk.c
284
struct scm_device *scmdev = hctx->queue->queuedata;
drivers/s390/cio/qdio.h
106
static inline int do_eqbs(u64 token, unsigned char *state, int queue,
drivers/s390/cio/qdio.h
109
unsigned long _queuestart = ((unsigned long)queue << 32) | *start;
drivers/s390/cio/qdio.h
88
static inline int do_sqbs(u64 token, unsigned char state, int queue,
drivers/s390/cio/qdio.h
91
unsigned long _queuestart = ((unsigned long)queue << 32) | *start;
drivers/s390/cio/qdio_setup.c
299
static void qdio_fill_qdr_desc(struct qdesfmt0 *desc, struct qdio_q *queue)
drivers/s390/cio/qdio_setup.c
301
desc->sliba = virt_to_dma64(queue->slib);
drivers/s390/cio/qdio_setup.c
302
desc->sla = virt_to_dma64(queue->sl);
drivers/s390/cio/qdio_setup.c
303
desc->slsba = virt_to_dma64(&queue->slsb);
drivers/s390/crypto/ap_bus.c
1001
test_bit_inv(queue, ap_perms.aqm);
drivers/s390/crypto/ap_bus.c
1037
card, queue, ap_drv->driver.name);
drivers/s390/crypto/ap_bus.c
862
int rc, card, queue, devres, drvres;
drivers/s390/crypto/ap_bus.c
870
queue = AP_QID_QUEUE(aq->qid);
drivers/s390/crypto/ap_bus.c
875
pr_debug("reprobing queue=%02x.%04x\n", card, queue);
drivers/s390/crypto/ap_bus.c
879
__func__, card, queue);
drivers/s390/crypto/ap_bus.c
885
test_bit_inv(queue, ap_perms.aqm);
drivers/s390/crypto/ap_bus.c
890
pr_debug("reprobing queue=%02x.%04x\n", card, queue);
drivers/s390/crypto/ap_bus.c
894
__func__, card, queue);
drivers/s390/crypto/ap_bus.c
919
int ap_owned_by_def_drv(int card, int queue)
drivers/s390/crypto/ap_bus.c
924
if (card < 0 || card >= AP_DEVICES || queue < 0 || queue >= AP_DOMAINS)
drivers/s390/crypto/ap_bus.c
927
aq = ap_get_qdev(AP_MKQID(card, queue));
drivers/s390/crypto/ap_bus.c
941
test_bit_inv(queue, ap_perms.aqm))
drivers/s390/crypto/ap_bus.c
964
int card, queue, rc = 0;
drivers/s390/crypto/ap_bus.c
968
for (queue = 0; !rc && queue < AP_DOMAINS; queue++)
drivers/s390/crypto/ap_bus.c
969
if (test_bit_inv(queue, aqm))
drivers/s390/crypto/ap_bus.c
970
rc = ap_owned_by_def_drv(card, queue);
drivers/s390/crypto/ap_bus.c
980
int card, queue, devres, drvres, rc = -ENODEV;
drivers/s390/crypto/ap_bus.c
993
queue = AP_QID_QUEUE(to_ap_queue(dev)->qid);
drivers/s390/crypto/vfio_ap_ops.c
1903
struct ap_queue *queue;
drivers/s390/crypto/vfio_ap_ops.c
1906
queue = ap_get_qdev(apqn);
drivers/s390/crypto/vfio_ap_ops.c
1907
if (!queue)
drivers/s390/crypto/vfio_ap_ops.c
1910
if (queue->ap_dev.device.driver == &matrix_dev->vfio_ap_drv->driver)
drivers/s390/crypto/vfio_ap_ops.c
1911
q = dev_get_drvdata(&queue->ap_dev.device);
drivers/s390/crypto/vfio_ap_ops.c
1913
put_device(&queue->ap_dev.device);
drivers/s390/crypto/vfio_ap_private.h
158
int vfio_ap_mdev_probe_queue(struct ap_device *queue);
drivers/s390/crypto/vfio_ap_private.h
159
void vfio_ap_mdev_remove_queue(struct ap_device *queue);
drivers/s390/crypto/zcrypt_api.c
1128
!ap_queue_usable(zq->queue) ||
drivers/s390/crypto/zcrypt_api.c
1130
!is_desired_ep11_queue(zq->queue->qid,
drivers/s390/crypto/zcrypt_api.c
1135
AP_QID_QUEUE(zq->queue->qid)))
drivers/s390/crypto/zcrypt_api.c
1139
tr->last_qid == zq->queue->qid) ?
drivers/s390/crypto/zcrypt_api.c
1166
qid = pref_zq->queue->qid;
drivers/s390/crypto/zcrypt_api.c
1251
!ap_queue_usable(zq->queue))
drivers/s390/crypto/zcrypt_api.c
1269
qid = pref_zq->queue->qid;
drivers/s390/crypto/zcrypt_api.c
1289
int card, queue;
drivers/s390/crypto/zcrypt_api.c
1297
card = AP_QID_CARD(zq->queue->qid);
drivers/s390/crypto/zcrypt_api.c
1300
queue = AP_QID_QUEUE(zq->queue->qid);
drivers/s390/crypto/zcrypt_api.c
1301
stat = &devstatus[card * AP_DOMAINS + queue];
drivers/s390/crypto/zcrypt_api.c
1304
stat->qid = zq->queue->qid;
drivers/s390/crypto/zcrypt_api.c
1317
int card, queue;
drivers/s390/crypto/zcrypt_api.c
1325
card = AP_QID_CARD(zq->queue->qid);
drivers/s390/crypto/zcrypt_api.c
1326
queue = AP_QID_QUEUE(zq->queue->qid);
drivers/s390/crypto/zcrypt_api.c
1327
if (card >= maxcard || queue >= maxqueue)
drivers/s390/crypto/zcrypt_api.c
1329
stat = &devstatus[card * maxqueue + queue];
drivers/s390/crypto/zcrypt_api.c
1332
stat->qid = zq->queue->qid;
drivers/s390/crypto/zcrypt_api.c
1340
int zcrypt_device_status_ext(int card, int queue,
drivers/s390/crypto/zcrypt_api.c
1351
if (card == AP_QID_CARD(zq->queue->qid) &&
drivers/s390/crypto/zcrypt_api.c
1352
queue == AP_QID_QUEUE(zq->queue->qid)) {
drivers/s390/crypto/zcrypt_api.c
1355
devstat->qid = zq->queue->qid;
drivers/s390/crypto/zcrypt_api.c
1378
card = AP_QID_CARD(zq->queue->qid);
drivers/s390/crypto/zcrypt_api.c
1379
if (AP_QID_QUEUE(zq->queue->qid) != ap_domain_index ||
drivers/s390/crypto/zcrypt_api.c
1399
card = AP_QID_CARD(zq->queue->qid);
drivers/s390/crypto/zcrypt_api.c
1400
if (AP_QID_QUEUE(zq->queue->qid) != ap_domain_index ||
drivers/s390/crypto/zcrypt_api.c
1403
spin_lock(&zq->queue->lock);
drivers/s390/crypto/zcrypt_api.c
1405
zq->queue->pendingq_count +
drivers/s390/crypto/zcrypt_api.c
1406
zq->queue->requestq_count;
drivers/s390/crypto/zcrypt_api.c
1407
spin_unlock(&zq->queue->lock);
drivers/s390/crypto/zcrypt_api.c
1426
card = AP_QID_CARD(zq->queue->qid);
drivers/s390/crypto/zcrypt_api.c
1427
if (AP_QID_QUEUE(zq->queue->qid) != ap_domain_index ||
drivers/s390/crypto/zcrypt_api.c
1430
spin_lock(&zq->queue->lock);
drivers/s390/crypto/zcrypt_api.c
1431
cnt = zq->queue->total_request_count;
drivers/s390/crypto/zcrypt_api.c
1432
spin_unlock(&zq->queue->lock);
drivers/s390/crypto/zcrypt_api.c
1451
if (AP_QID_QUEUE(zq->queue->qid) != ap_domain_index)
drivers/s390/crypto/zcrypt_api.c
1453
spin_lock(&zq->queue->lock);
drivers/s390/crypto/zcrypt_api.c
1454
pendingq_count += zq->queue->pendingq_count;
drivers/s390/crypto/zcrypt_api.c
1455
spin_unlock(&zq->queue->lock);
drivers/s390/crypto/zcrypt_api.c
1474
if (AP_QID_QUEUE(zq->queue->qid) != ap_domain_index)
drivers/s390/crypto/zcrypt_api.c
1476
spin_lock(&zq->queue->lock);
drivers/s390/crypto/zcrypt_api.c
1477
requestq_count += zq->queue->requestq_count;
drivers/s390/crypto/zcrypt_api.c
1478
spin_unlock(&zq->queue->lock);
drivers/s390/crypto/zcrypt_api.c
571
static inline bool zcrypt_check_queue(struct ap_perms *perms, int queue)
drivers/s390/crypto/zcrypt_api.c
573
return test_bit_inv(queue, perms->aqm) ? true : false;
drivers/s390/crypto/zcrypt_api.c
581
if (!zq || !try_module_get(zq->queue->ap_dev.device.driver->owner))
drivers/s390/crypto/zcrypt_api.c
585
get_device(&zq->queue->ap_dev.device);
drivers/s390/crypto/zcrypt_api.c
589
*pmod = zq->queue->ap_dev.device.driver->owner;
drivers/s390/crypto/zcrypt_api.c
601
put_device(&zq->queue->ap_dev.device);
drivers/s390/crypto/zcrypt_api.c
632
return zq->queue->total_request_count <
drivers/s390/crypto/zcrypt_api.c
633
pref_zq->queue->total_request_count;
drivers/s390/crypto/zcrypt_api.c
700
!ap_queue_usable(zq->queue))
drivers/s390/crypto/zcrypt_api.c
704
AP_QID_QUEUE(zq->queue->qid)))
drivers/s390/crypto/zcrypt_api.c
708
tr->last_qid == zq->queue->qid) ?
drivers/s390/crypto/zcrypt_api.c
727
qid = pref_zq->queue->qid;
drivers/s390/crypto/zcrypt_api.c
806
!ap_queue_usable(zq->queue))
drivers/s390/crypto/zcrypt_api.c
810
AP_QID_QUEUE(zq->queue->qid)))
drivers/s390/crypto/zcrypt_api.c
814
tr->last_qid == zq->queue->qid) ?
drivers/s390/crypto/zcrypt_api.c
833
qid = pref_zq->queue->qid;
drivers/s390/crypto/zcrypt_api.c
931
!ap_queue_usable(zq->queue) ||
drivers/s390/crypto/zcrypt_api.c
933
tdom != AP_QID_QUEUE(zq->queue->qid)))
drivers/s390/crypto/zcrypt_api.c
937
AP_QID_QUEUE(zq->queue->qid)))
drivers/s390/crypto/zcrypt_api.c
941
tr->last_qid == zq->queue->qid) ?
drivers/s390/crypto/zcrypt_api.c
962
qid = pref_zq->queue->qid;
drivers/s390/crypto/zcrypt_api.h
126
struct ap_queue *queue; /* The "real" ap queue device. */
drivers/s390/crypto/zcrypt_api.h
177
int zcrypt_device_status_ext(int card, int queue,
drivers/s390/crypto/zcrypt_card.c
104
ap_send_online_uevent(&zq->queue->ap_dev, online);
drivers/s390/crypto/zcrypt_cex4.c
120
cca_get_info(AP_QID_CARD(zq->queue->qid),
drivers/s390/crypto/zcrypt_cex4.c
121
AP_QID_QUEUE(zq->queue->qid),
drivers/s390/crypto/zcrypt_cex4.c
344
ep11_get_domain_info(AP_QID_CARD(zq->queue->qid),
drivers/s390/crypto/zcrypt_cex4.c
345
AP_QID_QUEUE(zq->queue->qid),
drivers/s390/crypto/zcrypt_cex4.c
391
ep11_get_domain_info(AP_QID_CARD(zq->queue->qid),
drivers/s390/crypto/zcrypt_cex4.c
392
AP_QID_QUEUE(zq->queue->qid),
drivers/s390/crypto/zcrypt_cex4.c
673
zq->queue = aq;
drivers/s390/crypto/zcrypt_error.h
110
__func__, card, queue,
drivers/s390/crypto/zcrypt_error.h
114
__func__, card, queue,
drivers/s390/crypto/zcrypt_error.h
136
__func__, card, queue, ehdr->reply_code, apfs);
drivers/s390/crypto/zcrypt_error.h
139
__func__, card, queue,
drivers/s390/crypto/zcrypt_error.h
146
__func__, card, queue, ehdr->reply_code);
drivers/s390/crypto/zcrypt_error.h
82
int card = AP_QID_CARD(zq->queue->qid);
drivers/s390/crypto/zcrypt_error.h
83
int queue = AP_QID_QUEUE(zq->queue->qid);
drivers/s390/crypto/zcrypt_msgtype50.c
357
AP_QID_CARD(zq->queue->qid),
drivers/s390/crypto/zcrypt_msgtype50.c
358
AP_QID_QUEUE(zq->queue->qid), t80h->code);
drivers/s390/crypto/zcrypt_msgtype50.c
360
__func__, AP_QID_CARD(zq->queue->qid),
drivers/s390/crypto/zcrypt_msgtype50.c
361
AP_QID_QUEUE(zq->queue->qid), t80h->code);
drivers/s390/crypto/zcrypt_msgtype50.c
362
ap_send_online_uevent(&zq->queue->ap_dev, zq->online);
drivers/s390/crypto/zcrypt_msgtype50.c
390
AP_QID_CARD(zq->queue->qid),
drivers/s390/crypto/zcrypt_msgtype50.c
391
AP_QID_QUEUE(zq->queue->qid),
drivers/s390/crypto/zcrypt_msgtype50.c
395
__func__, AP_QID_CARD(zq->queue->qid),
drivers/s390/crypto/zcrypt_msgtype50.c
396
AP_QID_QUEUE(zq->queue->qid), (int)rtype);
drivers/s390/crypto/zcrypt_msgtype50.c
397
ap_send_online_uevent(&zq->queue->ap_dev, zq->online);
drivers/s390/crypto/zcrypt_msgtype50.c
471
rc = ap_queue_message(zq->queue, ap_msg);
drivers/s390/crypto/zcrypt_msgtype50.c
483
ap_cancel_message(zq->queue, ap_msg);
drivers/s390/crypto/zcrypt_msgtype50.c
489
AP_QID_CARD(zq->queue->qid),
drivers/s390/crypto/zcrypt_msgtype50.c
490
AP_QID_QUEUE(zq->queue->qid), rc);
drivers/s390/crypto/zcrypt_msgtype50.c
520
rc = ap_queue_message(zq->queue, ap_msg);
drivers/s390/crypto/zcrypt_msgtype50.c
532
ap_cancel_message(zq->queue, ap_msg);
drivers/s390/crypto/zcrypt_msgtype50.c
538
AP_QID_CARD(zq->queue->qid),
drivers/s390/crypto/zcrypt_msgtype50.c
539
AP_QID_QUEUE(zq->queue->qid), rc);
drivers/s390/crypto/zcrypt_msgtype6.c
1024
rc = ap_queue_message(zq->queue, ap_msg);
drivers/s390/crypto/zcrypt_msgtype6.c
1036
ap_cancel_message(zq->queue, ap_msg);
drivers/s390/crypto/zcrypt_msgtype6.c
1109
rc = ap_queue_message(zq->queue, ap_msg);
drivers/s390/crypto/zcrypt_msgtype6.c
1119
ap_cancel_message(zq->queue, ap_msg);
drivers/s390/crypto/zcrypt_msgtype6.c
1128
AP_QID_CARD(zq->queue->qid),
drivers/s390/crypto/zcrypt_msgtype6.c
1129
AP_QID_QUEUE(zq->queue->qid), rc);
drivers/s390/crypto/zcrypt_msgtype6.c
1196
AP_QID_QUEUE(zq->queue->qid);
drivers/s390/crypto/zcrypt_msgtype6.c
1214
AP_QID_QUEUE(zq->queue->qid);
drivers/s390/crypto/zcrypt_msgtype6.c
1225
rc = ap_queue_message(zq->queue, ap_msg);
drivers/s390/crypto/zcrypt_msgtype6.c
1235
ap_cancel_message(zq->queue, ap_msg);
drivers/s390/crypto/zcrypt_msgtype6.c
1244
AP_QID_CARD(zq->queue->qid),
drivers/s390/crypto/zcrypt_msgtype6.c
1245
AP_QID_QUEUE(zq->queue->qid), rc);
drivers/s390/crypto/zcrypt_msgtype6.c
1297
msg->cprbx.domain = AP_QID_QUEUE(zq->queue->qid);
drivers/s390/crypto/zcrypt_msgtype6.c
1300
rc = ap_queue_message(zq->queue, ap_msg);
drivers/s390/crypto/zcrypt_msgtype6.c
1310
ap_cancel_message(zq->queue, ap_msg);
drivers/s390/crypto/zcrypt_msgtype6.c
234
msg->cprbx.domain = AP_QID_QUEUE(zq->queue->qid);
drivers/s390/crypto/zcrypt_msgtype6.c
304
msg->cprbx.domain = AP_QID_QUEUE(zq->queue->qid);
drivers/s390/crypto/zcrypt_msgtype6.c
580
__func__, AP_QID_CARD(zq->queue->qid),
drivers/s390/crypto/zcrypt_msgtype6.c
581
AP_QID_QUEUE(zq->queue->qid),
drivers/s390/crypto/zcrypt_msgtype6.c
587
AP_QID_CARD(zq->queue->qid),
drivers/s390/crypto/zcrypt_msgtype6.c
588
AP_QID_QUEUE(zq->queue->qid),
drivers/s390/crypto/zcrypt_msgtype6.c
591
__func__, AP_QID_CARD(zq->queue->qid),
drivers/s390/crypto/zcrypt_msgtype6.c
592
AP_QID_QUEUE(zq->queue->qid),
drivers/s390/crypto/zcrypt_msgtype6.c
594
ap_send_online_uevent(&zq->queue->ap_dev, zq->online);
drivers/s390/crypto/zcrypt_msgtype6.c
729
AP_QID_CARD(zq->queue->qid),
drivers/s390/crypto/zcrypt_msgtype6.c
730
AP_QID_QUEUE(zq->queue->qid),
drivers/s390/crypto/zcrypt_msgtype6.c
734
__func__, AP_QID_CARD(zq->queue->qid),
drivers/s390/crypto/zcrypt_msgtype6.c
735
AP_QID_QUEUE(zq->queue->qid), (int)msg->hdr.type);
drivers/s390/crypto/zcrypt_msgtype6.c
736
ap_send_online_uevent(&zq->queue->ap_dev, zq->online);
drivers/s390/crypto/zcrypt_msgtype6.c
764
AP_QID_CARD(zq->queue->qid),
drivers/s390/crypto/zcrypt_msgtype6.c
765
AP_QID_QUEUE(zq->queue->qid),
drivers/s390/crypto/zcrypt_msgtype6.c
769
__func__, AP_QID_CARD(zq->queue->qid),
drivers/s390/crypto/zcrypt_msgtype6.c
770
AP_QID_QUEUE(zq->queue->qid), (int)msg->hdr.type);
drivers/s390/crypto/zcrypt_msgtype6.c
771
ap_send_online_uevent(&zq->queue->ap_dev, zq->online);
drivers/s390/crypto/zcrypt_msgtype6.c
794
AP_QID_CARD(zq->queue->qid),
drivers/s390/crypto/zcrypt_msgtype6.c
795
AP_QID_QUEUE(zq->queue->qid),
drivers/s390/crypto/zcrypt_msgtype6.c
799
__func__, AP_QID_CARD(zq->queue->qid),
drivers/s390/crypto/zcrypt_msgtype6.c
800
AP_QID_QUEUE(zq->queue->qid), (int)msg->hdr.type);
drivers/s390/crypto/zcrypt_msgtype6.c
801
ap_send_online_uevent(&zq->queue->ap_dev, zq->online);
drivers/s390/crypto/zcrypt_msgtype6.c
825
AP_QID_CARD(zq->queue->qid),
drivers/s390/crypto/zcrypt_msgtype6.c
826
AP_QID_QUEUE(zq->queue->qid),
drivers/s390/crypto/zcrypt_msgtype6.c
830
__func__, AP_QID_CARD(zq->queue->qid),
drivers/s390/crypto/zcrypt_msgtype6.c
831
AP_QID_QUEUE(zq->queue->qid), (int)msg->hdr.type);
drivers/s390/crypto/zcrypt_msgtype6.c
832
ap_send_online_uevent(&zq->queue->ap_dev, zq->online);
drivers/s390/crypto/zcrypt_msgtype6.c
979
rc = ap_queue_message(zq->queue, ap_msg);
drivers/s390/crypto/zcrypt_msgtype6.c
991
ap_cancel_message(zq->queue, ap_msg);
drivers/s390/crypto/zcrypt_queue.c
108
ap_flush_queue(zq->queue);
drivers/s390/crypto/zcrypt_queue.c
173
zc = dev_get_drvdata(&zq->queue->card->ap_dev.device);
drivers/s390/crypto/zcrypt_queue.c
179
__func__, AP_QID_CARD(zq->queue->qid),
drivers/s390/crypto/zcrypt_queue.c
180
AP_QID_QUEUE(zq->queue->qid));
drivers/s390/crypto/zcrypt_queue.c
185
rc = sysfs_create_group(&zq->queue->ap_dev.device.kobj,
drivers/s390/crypto/zcrypt_queue.c
198
sysfs_remove_group(&zq->queue->ap_dev.device.kobj,
drivers/s390/crypto/zcrypt_queue.c
220
__func__, AP_QID_CARD(zq->queue->qid),
drivers/s390/crypto/zcrypt_queue.c
221
AP_QID_QUEUE(zq->queue->qid));
drivers/s390/crypto/zcrypt_queue.c
229
sysfs_remove_group(&zq->queue->ap_dev.device.kobj,
drivers/s390/crypto/zcrypt_queue.c
70
__func__, AP_QID_CARD(zq->queue->qid),
drivers/s390/crypto/zcrypt_queue.c
71
AP_QID_QUEUE(zq->queue->qid), online);
drivers/s390/crypto/zcrypt_queue.c
76
ap_flush_queue(zq->queue);
drivers/s390/net/qeth_core.h
1103
struct qeth_qdio_out_q *queue, __be16 proto,
drivers/s390/net/qeth_core.h
1104
void (*fill_header)(struct qeth_qdio_out_q *queue,
drivers/s390/net/qeth_core.h
529
static inline void qeth_tx_arm_timer(struct qeth_qdio_out_q *queue,
drivers/s390/net/qeth_core.h
532
timer_reduce(&queue->timer, usecs_to_jiffies(usecs) + jiffies);
drivers/s390/net/qeth_core.h
535
static inline bool qeth_out_queue_is_full(struct qeth_qdio_out_q *queue)
drivers/s390/net/qeth_core.h
537
return atomic_read(&queue->used_buffers) >= QDIO_MAX_BUFFERS_PER_Q;
drivers/s390/net/qeth_core.h
540
static inline bool qeth_out_queue_is_empty(struct qeth_qdio_out_q *queue)
drivers/s390/net/qeth_core.h
542
return atomic_read(&queue->used_buffers) == 0;
drivers/s390/net/qeth_core.h
913
struct qeth_qdio_out_q *queue)
drivers/s390/net/qeth_core.h
915
return qeth_iqd_translate_txq(card->dev, queue->queue_no) ==
drivers/s390/net/qeth_core_main.c
1343
static void qeth_tx_complete_buf(struct qeth_qdio_out_q *queue,
drivers/s390/net/qeth_core_main.c
1353
QETH_TXQ_STAT_INC(queue, bufs);
drivers/s390/net/qeth_core_main.c
1354
QETH_TXQ_STAT_ADD(queue, buf_elements, buf->next_element_to_fill);
drivers/s390/net/qeth_core_main.c
1356
QETH_TXQ_STAT_ADD(queue, tx_errors, buf->frames);
drivers/s390/net/qeth_core_main.c
1358
QETH_TXQ_STAT_ADD(queue, tx_packets, buf->frames);
drivers/s390/net/qeth_core_main.c
1359
QETH_TXQ_STAT_ADD(queue, tx_bytes, buf->bytes);
drivers/s390/net/qeth_core_main.c
1370
QETH_TXQ_STAT_ADD(queue, skbs_csum, packets);
drivers/s390/net/qeth_core_main.c
1372
QETH_TXQ_STAT_INC(queue, skbs_sg);
drivers/s390/net/qeth_core_main.c
1374
QETH_TXQ_STAT_INC(queue, skbs_tso);
drivers/s390/net/qeth_core_main.c
1375
QETH_TXQ_STAT_ADD(queue, tso_bytes, bytes);
drivers/s390/net/qeth_core_main.c
1383
static void qeth_clear_output_buffer(struct qeth_qdio_out_q *queue,
drivers/s390/net/qeth_core_main.c
1391
atomic_dec(&queue->set_pci_flags_count);
drivers/s390/net/qeth_core_main.c
1392
QETH_TXQ_STAT_INC(queue, completion_irq);
drivers/s390/net/qeth_core_main.c
1395
qeth_tx_complete_buf(queue, buf, error, budget);
drivers/s390/net/qeth_core_main.c
1397
for (i = 0; i < queue->max_elements; ++i) {
drivers/s390/net/qeth_core_main.c
1404
qeth_scrub_qdio_buffer(buf->buffer, queue->max_elements);
drivers/s390/net/qeth_core_main.c
1419
struct qeth_qdio_out_q *queue,
drivers/s390/net/qeth_core_main.c
1424
list_for_each_entry_safe(buf, tmp, &queue->pending_bufs, list_entry) {
drivers/s390/net/qeth_core_main.c
1437
qeth_notify_skbs(queue, buf, notify);
drivers/s390/net/qeth_core_main.c
1438
qeth_tx_complete_buf(queue, buf, drain, budget);
drivers/s390/net/qeth_core_main.c
1441
i < aob->sb_count && i < queue->max_elements;
drivers/s390/net/qeth_core_main.c
188
struct qeth_qdio_q *queue = card->qdio.in_q;
drivers/s390/net/qeth_core_main.c
196
for (i = 0; i < ARRAY_SIZE(queue->bufs); i++)
drivers/s390/net/qeth_core_main.c
197
queue->bufs[i].pool_entry = NULL;
drivers/s390/net/qeth_core_main.c
2621
struct qeth_qdio_out_q *queue = timer_container_of(queue, timer,
drivers/s390/net/qeth_core_main.c
2624
napi_schedule(&queue->napi);
drivers/s390/net/qeth_core_main.c
2625
QETH_TXQ_STAT_INC(queue, completion_timer);
drivers/s390/net/qeth_core_main.c
2648
struct qeth_qdio_out_q *queue;
drivers/s390/net/qeth_core_main.c
2650
queue = qeth_alloc_output_queue();
drivers/s390/net/qeth_core_main.c
2651
if (!queue)
drivers/s390/net/qeth_core_main.c
2654
QETH_CARD_HEX(card, 2, &queue, sizeof(void *));
drivers/s390/net/qeth_core_main.c
2655
card->qdio.out_qs[i] = queue;
drivers/s390/net/qeth_core_main.c
2656
queue->card = card;
drivers/s390/net/qeth_core_main.c
2657
queue->queue_no = i;
drivers/s390/net/qeth_core_main.c
2658
INIT_LIST_HEAD(&queue->pending_bufs);
drivers/s390/net/qeth_core_main.c
2659
spin_lock_init(&queue->lock);
drivers/s390/net/qeth_core_main.c
2660
timer_setup(&queue->timer, qeth_tx_completion_timer, 0);
drivers/s390/net/qeth_core_main.c
2662
queue->coalesce_usecs = QETH_TX_COALESCE_USECS;
drivers/s390/net/qeth_core_main.c
2663
queue->max_coalesced_frames = QETH_TX_MAX_COALESCED_FRAMES;
drivers/s390/net/qeth_core_main.c
2664
queue->rescan_usecs = QETH_TX_TIMER_USECS;
drivers/s390/net/qeth_core_main.c
2666
queue->coalesce_usecs = USEC_PER_SEC;
drivers/s390/net/qeth_core_main.c
2667
queue->max_coalesced_frames = 0;
drivers/s390/net/qeth_core_main.c
2668
queue->rescan_usecs = 10 * USEC_PER_SEC;
drivers/s390/net/qeth_core_main.c
2670
queue->priority = QETH_QIB_PQUE_PRIO_DEFAULT;
drivers/s390/net/qeth_core_main.c
2717
struct qeth_qdio_out_q *queue;
drivers/s390/net/qeth_core_main.c
2750
qeth_for_each_output_queue(card, queue, i)
drivers/s390/net/qeth_core_main.c
2751
parms->pque_priority[i] = queue->priority;
drivers/s390/net/qeth_core_main.c
2973
struct qeth_qdio_out_q *queue)
drivers/s390/net/qeth_core_main.c
2976
qeth_iqd_is_mcast_queue(card, queue) ||
drivers/s390/net/qeth_core_main.c
3019
struct qeth_qdio_out_q *queue = card->qdio.out_qs[i];
drivers/s390/net/qeth_core_main.c
3021
qdio_reset_buffers(queue->qdio_bufs, QDIO_MAX_BUFFERS_PER_Q);
drivers/s390/net/qeth_core_main.c
3022
queue->max_elements = QETH_MAX_BUFFER_ELEMENTS(card);
drivers/s390/net/qeth_core_main.c
3023
queue->next_buf_to_fill = 0;
drivers/s390/net/qeth_core_main.c
3024
queue->do_pack = 0;
drivers/s390/net/qeth_core_main.c
3025
queue->prev_hdr = NULL;
drivers/s390/net/qeth_core_main.c
3026
queue->coalesced_frames = 0;
drivers/s390/net/qeth_core_main.c
3027
queue->bulk_start = 0;
drivers/s390/net/qeth_core_main.c
3028
queue->bulk_count = 0;
drivers/s390/net/qeth_core_main.c
3029
queue->bulk_max = qeth_tx_select_bulk_max(card, queue);
drivers/s390/net/qeth_core_main.c
3030
atomic_set(&queue->used_buffers, 0);
drivers/s390/net/qeth_core_main.c
3031
atomic_set(&queue->set_pci_flags_count, 0);
drivers/s390/net/qeth_core_main.c
3453
struct qeth_qdio_q *queue = card->qdio.in_q;
drivers/s390/net/qeth_core_main.c
3461
for (i = queue->next_buf_to_init;
drivers/s390/net/qeth_core_main.c
3462
i < queue->next_buf_to_init + count; ++i) {
drivers/s390/net/qeth_core_main.c
3464
&queue->bufs[QDIO_BUFNR(i)])) {
drivers/s390/net/qeth_core_main.c
3494
queue->next_buf_to_init,
drivers/s390/net/qeth_core_main.c
3499
queue->next_buf_to_init = QDIO_BUFNR(queue->next_buf_to_init +
drivers/s390/net/qeth_core_main.c
3544
static int qeth_prep_flush_pack_buffer(struct qeth_qdio_out_q *queue)
drivers/s390/net/qeth_core_main.c
3548
buffer = queue->bufs[queue->next_buf_to_fill];
drivers/s390/net/qeth_core_main.c
3553
queue->next_buf_to_fill =
drivers/s390/net/qeth_core_main.c
3554
QDIO_BUFNR(queue->next_buf_to_fill + 1);
drivers/s390/net/qeth_core_main.c
3564
static void qeth_switch_to_packing_if_needed(struct qeth_qdio_out_q *queue)
drivers/s390/net/qeth_core_main.c
3566
if (!queue->do_pack) {
drivers/s390/net/qeth_core_main.c
3567
if (atomic_read(&queue->used_buffers)
drivers/s390/net/qeth_core_main.c
3570
QETH_CARD_TEXT(queue->card, 6, "np->pack");
drivers/s390/net/qeth_core_main.c
3571
QETH_TXQ_STAT_INC(queue, packing_mode_switch);
drivers/s390/net/qeth_core_main.c
3572
queue->do_pack = 1;
drivers/s390/net/qeth_core_main.c
3583
static int qeth_switch_to_nonpacking_if_needed(struct qeth_qdio_out_q *queue)
drivers/s390/net/qeth_core_main.c
3585
if (queue->do_pack) {
drivers/s390/net/qeth_core_main.c
3586
if (atomic_read(&queue->used_buffers)
drivers/s390/net/qeth_core_main.c
3589
QETH_CARD_TEXT(queue->card, 6, "pack->np");
drivers/s390/net/qeth_core_main.c
3590
QETH_TXQ_STAT_INC(queue, packing_mode_switch);
drivers/s390/net/qeth_core_main.c
3591
queue->do_pack = 0;
drivers/s390/net/qeth_core_main.c
3592
return qeth_prep_flush_pack_buffer(queue);
drivers/s390/net/qeth_core_main.c
3598
static void qeth_flush_buffers(struct qeth_qdio_out_q *queue, int index,
drivers/s390/net/qeth_core_main.c
3601
struct qeth_qdio_out_buffer *buf = queue->bufs[index];
drivers/s390/net/qeth_core_main.c
3602
struct qeth_card *card = queue->card;
drivers/s390/net/qeth_core_main.c
3612
buf = queue->bufs[bidx];
drivers/s390/net/qeth_core_main.c
3615
queue->coalesced_frames += buf->frames;
drivers/s390/net/qeth_core_main.c
3625
!qeth_iqd_is_mcast_queue(card, queue) &&
drivers/s390/net/qeth_core_main.c
3636
priv->queue_no = queue->queue_no;
drivers/s390/net/qeth_core_main.c
3640
if (!queue->do_pack) {
drivers/s390/net/qeth_core_main.c
3641
if ((atomic_read(&queue->used_buffers) >=
drivers/s390/net/qeth_core_main.c
3644
!atomic_read(&queue->set_pci_flags_count)) {
drivers/s390/net/qeth_core_main.c
3647
atomic_inc(&queue->set_pci_flags_count);
drivers/s390/net/qeth_core_main.c
3651
if (!atomic_read(&queue->set_pci_flags_count)) {
drivers/s390/net/qeth_core_main.c
3660
atomic_inc(&queue->set_pci_flags_count);
drivers/s390/net/qeth_core_main.c
3666
QETH_TXQ_STAT_INC(queue, doorbell);
drivers/s390/net/qeth_core_main.c
3667
rc = qdio_add_bufs_to_output_queue(CARD_DDEV(card), queue->queue_no,
drivers/s390/net/qeth_core_main.c
3676
frames = READ_ONCE(queue->max_coalesced_frames);
drivers/s390/net/qeth_core_main.c
3677
usecs = READ_ONCE(queue->coalesce_usecs);
drivers/s390/net/qeth_core_main.c
3679
if (frames && queue->coalesced_frames >= frames) {
drivers/s390/net/qeth_core_main.c
3680
napi_schedule(&queue->napi);
drivers/s390/net/qeth_core_main.c
3681
queue->coalesced_frames = 0;
drivers/s390/net/qeth_core_main.c
3682
QETH_TXQ_STAT_INC(queue, coal_frames);
drivers/s390/net/qeth_core_main.c
3684
atomic_read(&queue->used_buffers) >= 32) {
drivers/s390/net/qeth_core_main.c
3686
napi_schedule(&queue->napi);
drivers/s390/net/qeth_core_main.c
3687
QETH_TXQ_STAT_INC(queue, coal_frames);
drivers/s390/net/qeth_core_main.c
3689
qeth_tx_arm_timer(queue, usecs);
drivers/s390/net/qeth_core_main.c
3694
QETH_CARD_TEXT(queue->card, 2, "flushbuf");
drivers/s390/net/qeth_core_main.c
3695
QETH_CARD_TEXT_(queue->card, 2, " q%d", queue->queue_no);
drivers/s390/net/qeth_core_main.c
3696
QETH_CARD_TEXT_(queue->card, 2, " idx%d", index);
drivers/s390/net/qeth_core_main.c
3697
QETH_CARD_TEXT_(queue->card, 2, " c%d", count);
drivers/s390/net/qeth_core_main.c
3698
QETH_CARD_TEXT_(queue->card, 2, " err%d", rc);
drivers/s390/net/qeth_core_main.c
3702
qeth_schedule_recovery(queue->card);
drivers/s390/net/qeth_core_main.c
3706
static void qeth_flush_queue(struct qeth_qdio_out_q *queue)
drivers/s390/net/qeth_core_main.c
3708
qeth_flush_buffers(queue, queue->bulk_start, queue->bulk_count);
drivers/s390/net/qeth_core_main.c
3710
queue->bulk_start = QDIO_BUFNR(queue->bulk_start + queue->bulk_count);
drivers/s390/net/qeth_core_main.c
3711
queue->prev_hdr = NULL;
drivers/s390/net/qeth_core_main.c
3712
queue->bulk_count = 0;
drivers/s390/net/qeth_core_main.c
3715
static void qeth_check_outbound_queue(struct qeth_qdio_out_q *queue)
drivers/s390/net/qeth_core_main.c
3721
if ((atomic_read(&queue->used_buffers) <= QETH_LOW_WATERMARK_PACK) ||
drivers/s390/net/qeth_core_main.c
3722
!atomic_read(&queue->set_pci_flags_count)) {
drivers/s390/net/qeth_core_main.c
3725
spin_lock(&queue->lock);
drivers/s390/net/qeth_core_main.c
3727
index = queue->next_buf_to_fill;
drivers/s390/net/qeth_core_main.c
3729
flush_cnt = qeth_switch_to_nonpacking_if_needed(queue);
drivers/s390/net/qeth_core_main.c
3730
if (!flush_cnt && !atomic_read(&queue->set_pci_flags_count))
drivers/s390/net/qeth_core_main.c
3731
flush_cnt = qeth_prep_flush_pack_buffer(queue);
drivers/s390/net/qeth_core_main.c
3734
qeth_flush_buffers(queue, index, flush_cnt);
drivers/s390/net/qeth_core_main.c
3735
QETH_TXQ_STAT_ADD(queue, bufs_pack, flush_cnt);
drivers/s390/net/qeth_core_main.c
3738
spin_unlock(&queue->lock);
drivers/s390/net/qeth_core_main.c
3772
unsigned int queue, int first_element,
drivers/s390/net/qeth_core_main.c
3802
rc = qdio_add_bufs_to_input_queue(CARD_DDEV(card), queue,
drivers/s390/net/qeth_core_main.c
3814
unsigned int qdio_err, int queue,
drivers/s390/net/qeth_core_main.c
3820
QETH_CARD_TEXT_(card, 2, "qihq%d", queue);
drivers/s390/net/qeth_core_main.c
3950
static int qeth_add_hw_header(struct qeth_qdio_out_q *queue,
drivers/s390/net/qeth_core_main.c
3957
const unsigned int max_elements = queue->max_elements;
drivers/s390/net/qeth_core_main.c
3996
QETH_TXQ_STAT_INC(queue, skbs_linearized_fail);
drivers/s390/net/qeth_core_main.c
4000
QETH_TXQ_STAT_INC(queue, skbs_linearized);
drivers/s390/net/qeth_core_main.c
4023
static bool qeth_iqd_may_bulk(struct qeth_qdio_out_q *queue,
drivers/s390/net/qeth_core_main.c
4027
struct qeth_qdio_out_buffer *buffer = queue->bufs[queue->bulk_start];
drivers/s390/net/qeth_core_main.c
4028
struct qeth_hdr *prev_hdr = queue->prev_hdr;
drivers/s390/net/qeth_core_main.c
4140
static int __qeth_xmit(struct qeth_card *card, struct qeth_qdio_out_q *queue,
drivers/s390/net/qeth_core_main.c
4152
buffer = queue->bufs[QDIO_BUFNR(queue->bulk_start + queue->bulk_count)];
drivers/s390/net/qeth_core_main.c
4161
flush = !qeth_iqd_may_bulk(queue, skb, hdr);
drivers/s390/net/qeth_core_main.c
4164
(buffer->next_element_to_fill + elements > queue->max_elements)) {
drivers/s390/net/qeth_core_main.c
4167
queue->bulk_count++;
drivers/s390/net/qeth_core_main.c
4170
if (queue->bulk_count >= queue->bulk_max)
drivers/s390/net/qeth_core_main.c
4174
qeth_flush_queue(queue);
drivers/s390/net/qeth_core_main.c
4176
buffer = queue->bufs[QDIO_BUFNR(queue->bulk_start +
drivers/s390/net/qeth_core_main.c
4177
queue->bulk_count)];
drivers/s390/net/qeth_core_main.c
4185
atomic_inc_return(&queue->used_buffers) >= QDIO_MAX_BUFFERS_PER_Q) {
drivers/s390/net/qeth_core_main.c
4189
QETH_TXQ_STAT_INC(queue, stopped);
drivers/s390/net/qeth_core_main.c
4197
queue->prev_hdr = hdr;
drivers/s390/net/qeth_core_main.c
4202
if (flush || next_element >= queue->max_elements) {
drivers/s390/net/qeth_core_main.c
4204
queue->bulk_count++;
drivers/s390/net/qeth_core_main.c
4206
if (queue->bulk_count >= queue->bulk_max)
drivers/s390/net/qeth_core_main.c
4210
qeth_flush_queue(queue);
drivers/s390/net/qeth_core_main.c
4213
if (stopped && !qeth_out_queue_is_full(queue))
drivers/s390/net/qeth_core_main.c
4219
struct qeth_qdio_out_q *queue,
drivers/s390/net/qeth_core_main.c
4224
unsigned int start_index = queue->next_buf_to_fill;
drivers/s390/net/qeth_core_main.c
4233
buffer = queue->bufs[queue->next_buf_to_fill];
drivers/s390/net/qeth_core_main.c
4244
qeth_switch_to_packing_if_needed(queue);
drivers/s390/net/qeth_core_main.c
4245
if (queue->do_pack) {
drivers/s390/net/qeth_core_main.c
4249
queue->max_elements) {
drivers/s390/net/qeth_core_main.c
4253
queue->next_buf_to_fill =
drivers/s390/net/qeth_core_main.c
4254
QDIO_BUFNR(queue->next_buf_to_fill + 1);
drivers/s390/net/qeth_core_main.c
4255
buffer = queue->bufs[queue->next_buf_to_fill];
drivers/s390/net/qeth_core_main.c
4260
qeth_flush_buffers(queue, start_index,
drivers/s390/net/qeth_core_main.c
4269
atomic_inc_return(&queue->used_buffers) >= QDIO_MAX_BUFFERS_PER_Q) {
drivers/s390/net/qeth_core_main.c
4273
QETH_TXQ_STAT_INC(queue, stopped);
drivers/s390/net/qeth_core_main.c
4282
if (queue->do_pack)
drivers/s390/net/qeth_core_main.c
4283
QETH_TXQ_STAT_INC(queue, skbs_pack);
drivers/s390/net/qeth_core_main.c
4284
if (!queue->do_pack || stopped || next_element >= queue->max_elements) {
drivers/s390/net/qeth_core_main.c
4287
queue->next_buf_to_fill =
drivers/s390/net/qeth_core_main.c
4288
QDIO_BUFNR(queue->next_buf_to_fill + 1);
drivers/s390/net/qeth_core_main.c
4292
qeth_flush_buffers(queue, start_index, flush_count);
drivers/s390/net/qeth_core_main.c
4296
QETH_TXQ_STAT_ADD(queue, bufs_pack, flush_count);
drivers/s390/net/qeth_core_main.c
4298
if (stopped && !qeth_out_queue_is_full(queue))
drivers/s390/net/qeth_core_main.c
4320
struct qeth_qdio_out_q *queue, __be16 proto,
drivers/s390/net/qeth_core_main.c
4321
void (*fill_header)(struct qeth_qdio_out_q *queue,
drivers/s390/net/qeth_core_main.c
4346
push_len = qeth_add_hw_header(queue, skb, &hdr, hw_hdr_len, proto_len,
drivers/s390/net/qeth_core_main.c
4356
fill_header(queue, hdr, skb, proto, frame_len);
drivers/s390/net/qeth_core_main.c
4362
rc = __qeth_xmit(card, queue, skb, elements, hdr, data_offset,
drivers/s390/net/qeth_core_main.c
4367
spin_lock(&queue->lock);
drivers/s390/net/qeth_core_main.c
4368
rc = qeth_do_send_packet(card, queue, skb, hdr, data_offset,
drivers/s390/net/qeth_core_main.c
4370
spin_unlock(&queue->lock);
drivers/s390/net/qeth_core_main.c
5849
struct qeth_qdio_out_q *queue;
drivers/s390/net/qeth_core_main.c
5852
qeth_for_each_output_queue(card, queue, i) {
drivers/s390/net/qeth_core_main.c
5853
if (!qeth_out_queue_is_empty(queue))
drivers/s390/net/qeth_core_main.c
5854
napi_schedule(&queue->napi);
drivers/s390/net/qeth_core_main.c
5880
static void qeth_iqd_tx_complete(struct qeth_qdio_out_q *queue,
drivers/s390/net/qeth_core_main.c
5884
struct qeth_qdio_out_buffer *buffer = queue->bufs[bidx];
drivers/s390/net/qeth_core_main.c
5886
struct qeth_card *card = queue->card;
drivers/s390/net/qeth_core_main.c
5897
bidx, queue->queue_no);
drivers/s390/net/qeth_core_main.c
5907
qeth_notify_skbs(queue, buffer, TX_NOTIFY_PENDING);
drivers/s390/net/qeth_core_main.c
5910
qeth_scrub_qdio_buffer(buffer->buffer, queue->max_elements);
drivers/s390/net/qeth_core_main.c
5911
if (qeth_alloc_out_buf(queue, bidx, GFP_ATOMIC)) {
drivers/s390/net/qeth_core_main.c
5916
list_add(&buffer->list_entry, &queue->pending_bufs);
drivers/s390/net/qeth_core_main.c
5923
qeth_notify_skbs(queue, buffer, notify);
drivers/s390/net/qeth_core_main.c
5927
qeth_notify_skbs(queue, buffer,
drivers/s390/net/qeth_core_main.c
5931
qeth_clear_output_buffer(queue, buffer, error, budget);
drivers/s390/net/qeth_core_main.c
5936
struct qeth_qdio_out_q *queue = qeth_napi_to_out_queue(napi);
drivers/s390/net/qeth_core_main.c
5937
unsigned int queue_no = queue->queue_no;
drivers/s390/net/qeth_core_main.c
5938
struct qeth_card *card = queue->card;
drivers/s390/net/qeth_core_main.c
5954
qeth_tx_complete_pending_bufs(card, queue, false, budget);
drivers/s390/net/qeth_core_main.c
5956
if (qeth_out_queue_is_empty(queue)) {
drivers/s390/net/qeth_core_main.c
5963
QETH_TXQ_STAT_INC(queue, completion_yield);
drivers/s390/net/qeth_core_main.c
5974
!atomic_read(&queue->set_pci_flags_count))
drivers/s390/net/qeth_core_main.c
5975
qeth_tx_arm_timer(queue, queue->rescan_usecs);
drivers/s390/net/qeth_core_main.c
5983
buffer = queue->bufs[bidx];
drivers/s390/net/qeth_core_main.c
5989
qeth_iqd_tx_complete(queue, bidx, error, budget);
drivers/s390/net/qeth_core_main.c
5991
qeth_clear_output_buffer(queue, buffer, error,
drivers/s390/net/qeth_core_main.c
5995
atomic_sub(completed, &queue->used_buffers);
drivers/s390/net/qeth_core_main.c
6000
qeth_check_outbound_queue(queue);
drivers/s390/net/qeth_core_main.c
6008
!qeth_out_queue_is_full(queue))
drivers/s390/net/qeth_core_main.c
6943
struct qeth_qdio_out_q *queue;
drivers/s390/net/qeth_core_main.c
6962
queue = card->qdio.out_qs[i];
drivers/s390/net/qeth_core_main.c
6964
stats->tx_packets += queue->stats.tx_packets;
drivers/s390/net/qeth_core_main.c
6965
stats->tx_bytes += queue->stats.tx_bytes;
drivers/s390/net/qeth_core_main.c
6966
stats->tx_errors += queue->stats.tx_errors;
drivers/s390/net/qeth_core_main.c
6967
stats->tx_dropped += queue->stats.tx_dropped;
drivers/s390/net/qeth_core_main.c
7044
struct qeth_qdio_out_q *queue;
drivers/s390/net/qeth_core_main.c
7052
qeth_for_each_output_queue(card, queue, i) {
drivers/s390/net/qeth_core_main.c
7053
netif_napi_add_tx(dev, &queue->napi, qeth_tx_poll);
drivers/s390/net/qeth_core_main.c
7054
napi_enable(&queue->napi);
drivers/s390/net/qeth_core_main.c
7059
qeth_for_each_output_queue(card, queue, i) {
drivers/s390/net/qeth_core_main.c
7060
napi_schedule(&queue->napi);
drivers/s390/net/qeth_core_main.c
7073
struct qeth_qdio_out_q *queue;
drivers/s390/net/qeth_core_main.c
7083
qeth_for_each_output_queue(card, queue, i)
drivers/s390/net/qeth_core_main.c
7084
napi_disable(&queue->napi);
drivers/s390/net/qeth_core_main.c
7089
qeth_for_each_output_queue(card, queue, i) {
drivers/s390/net/qeth_core_main.c
7090
timer_delete_sync(&queue->timer);
drivers/s390/net/qeth_core_main.c
7092
netif_napi_del(&queue->napi);
drivers/s390/net/qeth_core_sys.c
307
struct qeth_qdio_out_q *queue;
drivers/s390/net/qeth_core_sys.c
319
queue = card->qdio.out_qs[i];
drivers/s390/net/qeth_core_sys.c
320
if (!queue)
drivers/s390/net/qeth_core_sys.c
322
memset(&queue->stats, 0, sizeof(queue->stats));
drivers/s390/net/qeth_ethtool.c
112
struct qeth_qdio_out_q *queue,
drivers/s390/net/qeth_ethtool.c
115
WRITE_ONCE(queue->coalesce_usecs, coal->tx_coalesce_usecs);
drivers/s390/net/qeth_ethtool.c
116
WRITE_ONCE(queue->max_coalesced_frames, coal->tx_max_coalesced_frames);
drivers/s390/net/qeth_ethtool.c
120
!qeth_out_queue_is_empty(queue))
drivers/s390/net/qeth_ethtool.c
121
qeth_tx_arm_timer(queue, coal->tx_coalesce_usecs);
drivers/s390/net/qeth_ethtool.c
130
struct qeth_qdio_out_q *queue;
drivers/s390/net/qeth_ethtool.c
139
qeth_for_each_output_queue(card, queue, i)
drivers/s390/net/qeth_ethtool.c
140
__qeth_set_coalesce(dev, queue, coal);
drivers/s390/net/qeth_ethtool.c
292
struct qeth_qdio_out_q *queue;
drivers/s390/net/qeth_ethtool.c
300
queue = card->qdio.out_qs[__queue];
drivers/s390/net/qeth_ethtool.c
302
coal->tx_coalesce_usecs = queue->coalesce_usecs;
drivers/s390/net/qeth_ethtool.c
303
coal->tx_max_coalesced_frames = queue->max_coalesced_frames;
drivers/s390/net/qeth_ethtool.c
307
static int qeth_set_per_queue_coalesce(struct net_device *dev, u32 queue,
drivers/s390/net/qeth_ethtool.c
315
if (queue >= card->qdio.no_out_queues)
drivers/s390/net/qeth_ethtool.c
321
__qeth_set_coalesce(dev, card->qdio.out_qs[queue], coal);
drivers/s390/net/qeth_l2_main.c
159
static void qeth_l2_fill_header(struct qeth_qdio_out_q *queue,
drivers/s390/net/qeth_l2_main.c
504
struct qeth_qdio_out_q *queue;
drivers/s390/net/qeth_l2_main.c
511
queue = card->qdio.out_qs[txq];
drivers/s390/net/qeth_l2_main.c
513
rc = qeth_xmit(card, skb, queue, vlan_get_protocol(skb),
drivers/s390/net/qeth_l2_main.c
518
QETH_TXQ_STAT_INC(queue, tx_dropped);
drivers/s390/net/qeth_l3_main.c
1626
static void qeth_l3_fill_header(struct qeth_qdio_out_q *queue,
drivers/s390/net/qeth_l3_main.c
1632
struct qeth_card *card = queue->card;
drivers/s390/net/qeth_l3_main.c
1711
struct qeth_qdio_out_q *queue, __be16 proto)
drivers/s390/net/qeth_l3_main.c
1725
return qeth_xmit(card, skb, queue, proto, qeth_l3_fill_header);
drivers/s390/net/qeth_l3_main.c
1734
struct qeth_qdio_out_q *queue;
drivers/s390/net/qeth_l3_main.c
1740
queue = card->qdio.out_qs[qeth_iqd_translate_txq(dev, txq)];
drivers/s390/net/qeth_l3_main.c
1759
queue = card->qdio.out_qs[txq];
drivers/s390/net/qeth_l3_main.c
1767
rc = qeth_l3_xmit(card, skb, queue, proto);
drivers/s390/net/qeth_l3_main.c
1769
rc = qeth_xmit(card, skb, queue, proto, qeth_l3_fill_header);
drivers/s390/net/qeth_l3_main.c
1775
QETH_TXQ_STAT_INC(queue, tx_dropped);
drivers/s390/virtio/virtio_ccw.c
122
dma64_t queue;
drivers/s390/virtio/virtio_ccw.c
499
info->info_block->l.queue = 0;
drivers/s390/virtio/virtio_ccw.c
559
u64 queue;
drivers/s390/virtio/virtio_ccw.c
606
queue = virtqueue_get_desc_addr(vq);
drivers/s390/virtio/virtio_ccw.c
608
info->info_block->l.queue = u64_to_dma64(queue);
drivers/s390/virtio/virtio_ccw.c
614
info->info_block->s.desc = u64_to_dma64(queue);
drivers/scsi/3w-9xxx.c
291
int first_reset = 0, queue = 0, retval = 1;
drivers/scsi/3w-9xxx.c
336
queue = 0;
drivers/scsi/3w-9xxx.c
350
queue = 1;
drivers/scsi/3w-9xxx.c
355
queue = 1;
drivers/scsi/3w-9xxx.c
359
if (queue)
drivers/scsi/3w-sas.c
564
int first_reset = 0, queue = 0, retval = 1;
drivers/scsi/3w-sas.c
604
queue = 0;
drivers/scsi/3w-sas.c
618
queue = 1;
drivers/scsi/3w-sas.c
623
queue = 1;
drivers/scsi/3w-sas.c
627
if (queue)
drivers/scsi/3w-xxxx.c
685
int queue = 0;
drivers/scsi/3w-xxxx.c
766
queue = 0;
drivers/scsi/3w-xxxx.c
782
queue = 1;
drivers/scsi/3w-xxxx.c
800
queue = 1;
drivers/scsi/3w-xxxx.c
804
if (queue == 1) {
drivers/scsi/aacraid/aacraid.h
1007
struct aac_queue queue[8];
drivers/scsi/aacraid/comminit.c
397
comm->queue[HostNormCmdQueue].base = queues;
drivers/scsi/aacraid/comminit.c
398
aac_queue_init(dev, &comm->queue[HostNormCmdQueue], headers, HOST_NORM_CMD_ENTRIES);
drivers/scsi/aacraid/comminit.c
403
comm->queue[HostHighCmdQueue].base = queues;
drivers/scsi/aacraid/comminit.c
404
aac_queue_init(dev, &comm->queue[HostHighCmdQueue], headers, HOST_HIGH_CMD_ENTRIES);
drivers/scsi/aacraid/comminit.c
410
comm->queue[AdapNormCmdQueue].base = queues;
drivers/scsi/aacraid/comminit.c
411
aac_queue_init(dev, &comm->queue[AdapNormCmdQueue], headers, ADAP_NORM_CMD_ENTRIES);
drivers/scsi/aacraid/comminit.c
417
comm->queue[AdapHighCmdQueue].base = queues;
drivers/scsi/aacraid/comminit.c
418
aac_queue_init(dev, &comm->queue[AdapHighCmdQueue], headers, ADAP_HIGH_CMD_ENTRIES);
drivers/scsi/aacraid/comminit.c
424
comm->queue[HostNormRespQueue].base = queues;
drivers/scsi/aacraid/comminit.c
425
aac_queue_init(dev, &comm->queue[HostNormRespQueue], headers, HOST_NORM_RESP_ENTRIES);
drivers/scsi/aacraid/comminit.c
430
comm->queue[HostHighRespQueue].base = queues;
drivers/scsi/aacraid/comminit.c
431
aac_queue_init(dev, &comm->queue[HostHighRespQueue], headers, HOST_HIGH_RESP_ENTRIES);
drivers/scsi/aacraid/comminit.c
437
comm->queue[AdapNormRespQueue].base = queues;
drivers/scsi/aacraid/comminit.c
438
aac_queue_init(dev, &comm->queue[AdapNormRespQueue], headers, ADAP_NORM_RESP_ENTRIES);
drivers/scsi/aacraid/comminit.c
444
comm->queue[AdapHighRespQueue].base = queues;
drivers/scsi/aacraid/comminit.c
445
aac_queue_init(dev, &comm->queue[AdapHighRespQueue], headers, ADAP_HIGH_RESP_ENTRIES);
drivers/scsi/aacraid/comminit.c
447
comm->queue[AdapNormCmdQueue].lock = comm->queue[HostNormRespQueue].lock;
drivers/scsi/aacraid/comminit.c
448
comm->queue[AdapHighCmdQueue].lock = comm->queue[HostHighRespQueue].lock;
drivers/scsi/aacraid/comminit.c
449
comm->queue[AdapNormRespQueue].lock = comm->queue[HostNormCmdQueue].lock;
drivers/scsi/aacraid/comminit.c
450
comm->queue[AdapHighRespQueue].lock = comm->queue[HostHighCmdQueue].lock;
drivers/scsi/aacraid/commsup.c
2036
t_lock = dev->queues->queue[HostNormCmdQueue].lock;
drivers/scsi/aacraid/commsup.c
2039
while (!list_empty(&(dev->queues->queue[HostNormCmdQueue].cmdq))) {
drivers/scsi/aacraid/commsup.c
2048
entry = dev->queues->queue[HostNormCmdQueue].cmdq.next;
drivers/scsi/aacraid/commsup.c
2051
t_lock = dev->queues->queue[HostNormCmdQueue].lock;
drivers/scsi/aacraid/commsup.c
2142
t_lock = dev->queues->queue[HostNormCmdQueue].lock;
drivers/scsi/aacraid/commsup.c
2148
t_lock = dev->queues->queue[HostNormCmdQueue].lock;
drivers/scsi/aacraid/commsup.c
2318
add_wait_queue(&dev->queues->queue[HostNormCmdQueue].cmdready, &wait);
drivers/scsi/aacraid/commsup.c
2390
remove_wait_queue(&dev->queues->queue[HostNormCmdQueue].cmdready, &wait);
drivers/scsi/aacraid/commsup.c
363
q = &dev->queues->queue[qid];
drivers/scsi/aacraid/commsup.c
644
struct aac_queue * q = &dev->queues->queue[AdapNormCmdQueue];
drivers/scsi/aacraid/commsup.c
912
q = &dev->queues->queue[AdapNormRespQueue];
drivers/scsi/aacraid/dpcsup.c
278
struct aac_queue *q = &dev->queues->queue[HostNormCmdQueue];
drivers/scsi/aacraid/dpcsup.c
344
atomic_dec(&dev->queues->queue[AdapNormCmdQueue].numpending);
drivers/scsi/aacraid/dpcsup.c
72
atomic_dec(&dev->queues->queue[AdapNormCmdQueue].numpending);
drivers/scsi/aacraid/rx.c
401
struct aac_queue *q = &dev->queues->queue[AdapNormCmdQueue];
drivers/scsi/aacraid/rx.c
424
struct aac_queue *q = &dev->queues->queue[AdapNormCmdQueue];
drivers/scsi/aacraid/rx.c
56
aac_command_normal(&dev->queues->queue[HostNormCmdQueue]);
drivers/scsi/aacraid/rx.c
60
aac_response_normal(&dev->queues->queue[HostNormRespQueue]);
drivers/scsi/aacraid/sa.c
55
aac_command_normal(&dev->queues->queue[HostNormCmdQueue]);
drivers/scsi/aacraid/sa.c
58
aac_response_normal(&dev->queues->queue[HostNormRespQueue]);
drivers/scsi/aacraid/src.c
486
struct aac_queue *q = &dev->queues->queue[AdapNormCmdQueue];
drivers/scsi/aic7xxx/aic7xxx_core.c
241
struct scb_tailq *queue);
drivers/scsi/aic7xxx/aic7xxx_core.c
932
ahc_run_untagged_queue(struct ahc_softc *ahc, struct scb_tailq *queue)
drivers/scsi/aic7xxx/aic7xxx_core.c
939
if ((scb = TAILQ_FIRST(queue)) != NULL
drivers/scsi/arm/queue.c
107
int __queue_add(Queue_t *queue, struct scsi_cmnd *SCpnt, int head)
drivers/scsi/arm/queue.c
114
spin_lock_irqsave(&queue->queue_lock, flags);
drivers/scsi/arm/queue.c
115
if (list_empty(&queue->free))
drivers/scsi/arm/queue.c
118
l = queue->free.next;
drivers/scsi/arm/queue.c
128
list_add(l, &queue->head);
drivers/scsi/arm/queue.c
130
list_add_tail(l, &queue->head);
drivers/scsi/arm/queue.c
134
spin_unlock_irqrestore(&queue->queue_lock, flags);
drivers/scsi/arm/queue.c
138
static struct scsi_cmnd *__queue_remove(Queue_t *queue, struct list_head *ent)
drivers/scsi/arm/queue.c
150
list_add(ent, &queue->free);
drivers/scsi/arm/queue.c
162
struct scsi_cmnd *queue_remove_exclude(Queue_t *queue, unsigned long *exclude)
drivers/scsi/arm/queue.c
168
spin_lock_irqsave(&queue->queue_lock, flags);
drivers/scsi/arm/queue.c
169
list_for_each(l, &queue->head) {
drivers/scsi/arm/queue.c
173
SCpnt = __queue_remove(queue, l);
drivers/scsi/arm/queue.c
177
spin_unlock_irqrestore(&queue->queue_lock, flags);
drivers/scsi/arm/queue.c
188
struct scsi_cmnd *queue_remove(Queue_t *queue)
drivers/scsi/arm/queue.c
193
spin_lock_irqsave(&queue->queue_lock, flags);
drivers/scsi/arm/queue.c
194
if (!list_empty(&queue->head))
drivers/scsi/arm/queue.c
195
SCpnt = __queue_remove(queue, queue->head.next);
drivers/scsi/arm/queue.c
196
spin_unlock_irqrestore(&queue->queue_lock, flags);
drivers/scsi/arm/queue.c
210
struct scsi_cmnd *queue_remove_tgtluntag(Queue_t *queue, int target, int lun,
drivers/scsi/arm/queue.c
217
spin_lock_irqsave(&queue->queue_lock, flags);
drivers/scsi/arm/queue.c
218
list_for_each(l, &queue->head) {
drivers/scsi/arm/queue.c
222
SCpnt = __queue_remove(queue, l);
drivers/scsi/arm/queue.c
226
spin_unlock_irqrestore(&queue->queue_lock, flags);
drivers/scsi/arm/queue.c
238
void queue_remove_all_target(Queue_t *queue, int target)
drivers/scsi/arm/queue.c
243
spin_lock_irqsave(&queue->queue_lock, flags);
drivers/scsi/arm/queue.c
244
list_for_each(l, &queue->head) {
drivers/scsi/arm/queue.c
247
__queue_remove(queue, l);
drivers/scsi/arm/queue.c
249
spin_unlock_irqrestore(&queue->queue_lock, flags);
drivers/scsi/arm/queue.c
261
int queue_probetgtlun (Queue_t *queue, int target, int lun)
drivers/scsi/arm/queue.c
267
spin_lock_irqsave(&queue->queue_lock, flags);
drivers/scsi/arm/queue.c
268
list_for_each(l, &queue->head) {
drivers/scsi/arm/queue.c
275
spin_unlock_irqrestore(&queue->queue_lock, flags);
drivers/scsi/arm/queue.c
287
int queue_remove_cmd(Queue_t *queue, struct scsi_cmnd *SCpnt)
drivers/scsi/arm/queue.c
293
spin_lock_irqsave(&queue->queue_lock, flags);
drivers/scsi/arm/queue.c
294
list_for_each(l, &queue->head) {
drivers/scsi/arm/queue.c
297
__queue_remove(queue, l);
drivers/scsi/arm/queue.c
302
spin_unlock_irqrestore(&queue->queue_lock, flags);
drivers/scsi/arm/queue.c
59
int queue_initialise (Queue_t *queue)
drivers/scsi/arm/queue.c
64
spin_lock_init(&queue->queue_lock);
drivers/scsi/arm/queue.c
65
INIT_LIST_HEAD(&queue->head);
drivers/scsi/arm/queue.c
66
INIT_LIST_HEAD(&queue->free);
drivers/scsi/arm/queue.c
74
queue->alloc = q = kmalloc_objs(QE_t, nqueues);
drivers/scsi/arm/queue.c
79
list_add(&q->list, &queue->free);
drivers/scsi/arm/queue.c
83
return queue->alloc != NULL;
drivers/scsi/arm/queue.c
91
void queue_free (Queue_t *queue)
drivers/scsi/arm/queue.c
93
if (!list_empty(&queue->head))
drivers/scsi/arm/queue.c
94
printk(KERN_WARNING "freeing non-empty queue %p\n", queue);
drivers/scsi/arm/queue.c
95
kfree(queue->alloc);
drivers/scsi/arm/queue.h
102
int queue_remove_cmd(Queue_t *queue, struct scsi_cmnd *SCpnt);
drivers/scsi/arm/queue.h
22
extern int queue_initialise (Queue_t *queue);
drivers/scsi/arm/queue.h
29
extern void queue_free (Queue_t *queue);
drivers/scsi/arm/queue.h
37
extern struct scsi_cmnd *queue_remove (Queue_t *queue);
drivers/scsi/arm/queue.h
46
extern struct scsi_cmnd *queue_remove_exclude(Queue_t *queue,
drivers/scsi/arm/queue.h
49
#define queue_add_cmd_ordered(queue,SCpnt) \
drivers/scsi/arm/queue.h
50
__queue_add(queue,SCpnt,(SCpnt)->cmnd[0] == REQUEST_SENSE)
drivers/scsi/arm/queue.h
51
#define queue_add_cmd_tail(queue,SCpnt) \
drivers/scsi/arm/queue.h
52
__queue_add(queue,SCpnt,0)
drivers/scsi/arm/queue.h
61
extern int __queue_add(Queue_t *queue, struct scsi_cmnd *SCpnt, int head);
drivers/scsi/arm/queue.h
72
extern struct scsi_cmnd *queue_remove_tgtluntag(Queue_t *queue, int target,
drivers/scsi/arm/queue.h
82
extern void queue_remove_all_target(Queue_t *queue, int target);
drivers/scsi/arm/queue.h
93
extern int queue_probetgtlun (Queue_t *queue, int target, int lun);
drivers/scsi/bfa/bfa_core.c
766
int queue;
drivers/scsi/bfa/bfa_core.c
777
for (queue = 0; queue < BFI_IOC_MAX_CQS; queue++)
drivers/scsi/bfa/bfa_core.c
778
bfa_isr_rspq(bfa, queue);
drivers/scsi/bfa/bfa_core.c
790
for (queue = 0; queue < BFI_IOC_MAX_CQS; queue++)
drivers/scsi/bfa/bfa_core.c
791
bfa_isr_reqq(bfa, queue);
drivers/scsi/bfa/bfa_core.c
804
int queue;
drivers/scsi/bfa/bfa_core.c
817
for (queue = 0; queue < BFI_IOC_MAX_CQS; queue++)
drivers/scsi/bfa/bfa_core.c
818
if (bfa_isr_rspq(bfa, queue))
drivers/scsi/bfa/bfa_core.c
830
for (queue = 0; queue < BFI_IOC_MAX_CQS; queue++)
drivers/scsi/bfa/bfa_core.c
831
bfa_isr_reqq(bfa, queue);
drivers/scsi/bfa/bfa_ioc.h
566
u8 queue;
drivers/scsi/bfa/bfa_svc.c
5653
res->queue = fcdiag->qtest.all;
drivers/scsi/bfa/bfa_svc.c
5667
req = bfa_reqq_next(fcdiag->bfa, fcdiag->qtest.queue);
drivers/scsi/bfa/bfa_svc.c
5678
bfa_trc(fcdiag, fcdiag->qtest.queue);
drivers/scsi/bfa/bfa_svc.c
5680
bfa_reqq_produce(fcdiag->bfa, fcdiag->qtest.queue, req->mh);
drivers/scsi/bfa/bfa_svc.c
5717
fcdiag->qtest.queue < (BFI_IOC_MAX_CQS - 1)) {
drivers/scsi/bfa/bfa_svc.c
5719
fcdiag->qtest.queue++;
drivers/scsi/bfa/bfa_svc.c
5733
res->queue = fcdiag->qtest.queue;
drivers/scsi/bfa/bfa_svc.c
5947
bfa_fcdiag_queuetest(struct bfa_s *bfa, u32 force, u32 queue,
drivers/scsi/bfa/bfa_svc.c
5954
bfa_trc(fcdiag, queue);
drivers/scsi/bfa/bfa_svc.c
5977
if (queue < BFI_IOC_MAX_CQS) {
drivers/scsi/bfa/bfa_svc.c
5978
fcdiag->qtest.result->queue = (u8)queue;
drivers/scsi/bfa/bfa_svc.c
5979
fcdiag->qtest.queue = (u8)queue;
drivers/scsi/bfa/bfa_svc.c
5982
fcdiag->qtest.result->queue = 0;
drivers/scsi/bfa/bfa_svc.c
5983
fcdiag->qtest.queue = 0;
drivers/scsi/bfa/bfa_svc.h
747
u8 queue;
drivers/scsi/bfa/bfa_svc.h
812
u32 queue, struct bfa_diag_qtest_result_s *result,
drivers/scsi/bfa/bfad_bsg.c
1713
iocmd->queue, &iocmd->result,
drivers/scsi/bfa/bfad_bsg.h
620
u32 queue;
drivers/scsi/elx/efct/efct_hw.c
2013
spin_lock_irqsave(&io_to_abort->wq->queue->lock, flags);
drivers/scsi/elx/efct/efct_hw.c
2016
spin_unlock_irqrestore(&io_to_abort->wq->queue->lock,
drivers/scsi/elx/efct/efct_hw.c
2020
spin_unlock_irqrestore(&io_to_abort->wq->queue->lock, flags);
drivers/scsi/elx/efct/efct_hw.c
2214
while (!done && !sli_eq_read(&hw->sli, eq->queue, eqe)) {
drivers/scsi/elx/efct/efct_hw.c
2244
if (eq->queue->n_posted > eq->queue->posted_limit)
drivers/scsi/elx/efct/efct_hw.c
2245
sli_queue_arm(&hw->sli, eq->queue, false);
drivers/scsi/elx/efct/efct_hw.c
2254
sli_queue_eq_arm(&hw->sli, eq->queue, true);
drivers/scsi/elx/efct/efct_hw.c
2278
queue_rc = sli_wq_write(&wq->hw->sli, wq->queue, wqe->wqebuf);
drivers/scsi/elx/efct/efct_hw.c
2289
spin_lock_irqsave(&wq->queue->lock, flags);
drivers/scsi/elx/efct/efct_hw.c
2309
spin_unlock_irqrestore(&wq->queue->lock, flags);
drivers/scsi/elx/efct/efct_hw.c
2324
while (!sli_cq_read(&hw->sli, cq->queue, cqe)) {
drivers/scsi/elx/efct/efct_hw.c
2327
status = sli_cq_parse(&hw->sli, cq->queue, cqe, &ctype, &rid);
drivers/scsi/elx/efct/efct_hw.c
2395
if (n_processed == cq->queue->proc_limit)
drivers/scsi/elx/efct/efct_hw.c
2398
if (cq->queue->n_posted >= cq->queue->posted_limit)
drivers/scsi/elx/efct/efct_hw.c
2399
sli_queue_arm(&hw->sli, cq->queue, false);
drivers/scsi/elx/efct/efct_hw.c
2402
sli_queue_arm(&hw->sli, cq->queue, true);
drivers/scsi/elx/efct/efct_hw.c
2404
if (n_processed > cq->queue->max_num_processed)
drivers/scsi/elx/efct/efct_hw.c
2405
cq->queue->max_num_processed = n_processed;
drivers/scsi/elx/efct/efct_hw.c
2407
if (telapsed > cq->queue->max_process_time)
drivers/scsi/elx/efct/efct_hw.c
2408
cq->queue->max_process_time = telapsed;
drivers/scsi/elx/efct/efct_hw.c
2516
spin_lock_irqsave(&wq->queue->lock, flags);
drivers/scsi/elx/efct/efct_hw.c
2526
spin_unlock_irqrestore(&wq->queue->lock, flags);
drivers/scsi/elx/efct/efct_hw.c
2554
spin_unlock_irqrestore(&wq->queue->lock, flags);
drivers/scsi/elx/efct/efct_hw.h
476
struct sli4_queue *queue;
drivers/scsi/elx/efct/efct_hw.h
488
struct sli4_queue *queue;
drivers/scsi/elx/efct/efct_hw.h
506
struct sli4_queue *queue;
drivers/scsi/elx/efct/efct_hw.h
520
struct sli4_queue *queue;
drivers/scsi/elx/efct/efct_hw_queues.c
139
eq->queue = &hw->eq[eq->instance];
drivers/scsi/elx/efct/efct_hw_queues.c
142
if (sli_queue_alloc(&hw->sli, SLI4_QTYPE_EQ, eq->queue, entry_count,
drivers/scsi/elx/efct/efct_hw_queues.c
149
sli_eq_modify_delay(&hw->sli, eq->queue, 1, 0, 8);
drivers/scsi/elx/efct/efct_hw_queues.c
154
eq->queue->id, eq->entry_count);
drivers/scsi/elx/efct/efct_hw_queues.c
171
cq->queue = &hw->cq[cq->instance];
drivers/scsi/elx/efct/efct_hw_queues.c
175
if (sli_queue_alloc(&hw->sli, SLI4_QTYPE_CQ, cq->queue,
drivers/scsi/elx/efct/efct_hw_queues.c
176
cq->entry_count, eq->queue)) {
drivers/scsi/elx/efct/efct_hw_queues.c
187
cq->queue->id, cq->entry_count);
drivers/scsi/elx/efct/efct_hw_queues.c
216
cq->queue = &hw->cq[cq->instance];
drivers/scsi/elx/efct/efct_hw_queues.c
217
qs[i] = cq->queue;
drivers/scsi/elx/efct/efct_hw_queues.c
218
assefct[i] = eqs[i]->queue;
drivers/scsi/elx/efct/efct_hw_queues.c
257
mq->queue = &hw->mq[mq->instance];
drivers/scsi/elx/efct/efct_hw_queues.c
259
if (sli_queue_alloc(&hw->sli, SLI4_QTYPE_MQ, mq->queue, mq->entry_size,
drivers/scsi/elx/efct/efct_hw_queues.c
260
cq->queue)) {
drivers/scsi/elx/efct/efct_hw_queues.c
270
mq->queue->id, mq->entry_count);
drivers/scsi/elx/efct/efct_hw_queues.c
288
wq->queue = &hw->wq[wq->instance];
drivers/scsi/elx/efct/efct_hw_queues.c
294
if (sli_queue_alloc(&hw->sli, SLI4_QTYPE_WQ, wq->queue,
drivers/scsi/elx/efct/efct_hw_queues.c
295
wq->entry_count, cq->queue)) {
drivers/scsi/elx/efct/efct_hw_queues.c
305
wq->instance, wq->queue->id, wq->entry_count, wq->class);
drivers/scsi/elx/efct/efct_hw_queues.c
355
cqs[0]->queue->id,
drivers/scsi/elx/efct/efct_hw_queues.c
360
cqs[0]->queue->id);
drivers/scsi/esas2r/esas2r_main.c
1006
queue = &a->defer_list;
drivers/scsi/esas2r/esas2r_main.c
1010
result = esas2r_check_active_queue(a, &abort_request, cmd, queue);
drivers/scsi/esas2r/esas2r_main.c
1015
} else if (result == 2 && (queue == &a->defer_list)) {
drivers/scsi/esas2r/esas2r_main.c
1016
queue = &a->active_list;
drivers/scsi/esas2r/esas2r_main.c
908
struct list_head *queue)
drivers/scsi/esas2r/esas2r_main.c
915
list_for_each_safe(element, next, queue) {
drivers/scsi/esas2r/esas2r_main.c
922
if (queue == &a->active_list) {
drivers/scsi/esas2r/esas2r_main.c
984
struct list_head *queue;
drivers/scsi/hisi_sas/hisi_sas_main.c
568
int queue;
drivers/scsi/hisi_sas/hisi_sas_main.c
576
queue = 0;
drivers/scsi/hisi_sas/hisi_sas_main.c
581
queue = qmap->mq_map[raw_smp_processor_id()];
drivers/scsi/hisi_sas/hisi_sas_main.c
583
dq = &hisi_hba->dq[queue];
drivers/scsi/hisi_sas/hisi_sas_main.c
700
int queue = i % hisi_hba->queue_count;
drivers/scsi/hisi_sas/hisi_sas_main.c
701
struct hisi_sas_dq *dq = &hisi_hba->dq[queue];
drivers/scsi/hisi_sas/hisi_sas_v1_hw.c
1484
int queue = cq->id;
drivers/scsi/hisi_sas/hisi_sas_v1_hw.c
1487
hisi_hba->complete_hdr[queue];
drivers/scsi/hisi_sas/hisi_sas_v1_hw.c
1491
hisi_sas_write32(hisi_hba, OQ_INT_SRC, 1 << queue);
drivers/scsi/hisi_sas/hisi_sas_v1_hw.c
1493
COMPL_Q_0_WR_PTR + (0x14 * queue));
drivers/scsi/hisi_sas/hisi_sas_v1_hw.c
1511
slot->cmplt_queue = queue;
drivers/scsi/hisi_sas/hisi_sas_v1_hw.c
1520
hisi_sas_write32(hisi_hba, COMPL_Q_0_RD_PTR + (0x14 * queue), rd_point);
drivers/scsi/hisi_sas/hisi_sas_v2_hw.c
3121
int queue = cq->id;
drivers/scsi/hisi_sas/hisi_sas_v2_hw.c
3126
complete_queue = hisi_hba->complete_hdr[queue];
drivers/scsi/hisi_sas/hisi_sas_v2_hw.c
3129
(0x14 * queue));
drivers/scsi/hisi_sas/hisi_sas_v2_hw.c
3160
slot->cmplt_queue = queue;
drivers/scsi/hisi_sas/hisi_sas_v2_hw.c
3172
slot->cmplt_queue = queue;
drivers/scsi/hisi_sas/hisi_sas_v2_hw.c
3182
hisi_sas_write32(hisi_hba, COMPL_Q_0_RD_PTR + (0x14 * queue), rd_point);
drivers/scsi/hisi_sas/hisi_sas_v2_hw.c
3191
int queue = cq->id;
drivers/scsi/hisi_sas/hisi_sas_v2_hw.c
3193
hisi_sas_write32(hisi_hba, OQ_INT_SRC, 1 << queue);
drivers/scsi/hisi_sas/hisi_sas_v2_hw.c
3559
unsigned int queue, cpu;
drivers/scsi/hisi_sas/hisi_sas_v2_hw.c
3561
for (queue = 0; queue < qmap->nr_queues; queue++) {
drivers/scsi/hisi_sas/hisi_sas_v2_hw.c
3562
mask = irq_get_affinity_mask(hisi_hba->irq_map[96 + queue]);
drivers/scsi/hisi_sas/hisi_sas_v2_hw.c
3567
qmap->mq_map[cpu] = qmap->queue_offset + queue;
drivers/scsi/hisi_sas/hisi_sas_v2_hw.c
859
int queue = i % hisi_hba->queue_count;
drivers/scsi/hisi_sas/hisi_sas_v2_hw.c
860
struct hisi_sas_dq *dq = &hisi_hba->dq[queue];
drivers/scsi/hisi_sas/hisi_sas_v3_hw.c
2498
int queue = cq->id;
drivers/scsi/hisi_sas/hisi_sas_v3_hw.c
2502
complete_queue = hisi_hba->complete_hdr[queue];
drivers/scsi/hisi_sas/hisi_sas_v3_hw.c
2505
(0x14 * queue));
drivers/scsi/hisi_sas/hisi_sas_v3_hw.c
2542
slot->cmplt_queue = queue;
drivers/scsi/hisi_sas/hisi_sas_v3_hw.c
2553
hisi_sas_write32(hisi_hba, COMPL_Q_0_RD_PTR + (0x14 * queue), rd_point);
drivers/scsi/hisi_sas/hisi_sas_v3_hw.c
2559
static int queue_complete_v3_hw(struct Scsi_Host *shost, unsigned int queue)
drivers/scsi/hisi_sas/hisi_sas_v3_hw.c
2562
struct hisi_sas_cq *cq = &hisi_hba->cq[queue];
drivers/scsi/hisi_sas/hisi_sas_v3_hw.c
2585
int queue = cq->id;
drivers/scsi/hisi_sas/hisi_sas_v3_hw.c
2587
hisi_sas_write32(hisi_hba, OQ_INT_SRC, 1 << queue);
drivers/scsi/hpsa.c
6994
static struct ctlr_info *queue_to_hba(u8 *queue)
drivers/scsi/hpsa.c
6996
return container_of((queue - *queue), struct ctlr_info, q[0]);
drivers/scsi/hpsa.c
6999
static irqreturn_t hpsa_intx_discard_completions(int irq, void *queue)
drivers/scsi/hpsa.c
7001
struct ctlr_info *h = queue_to_hba(queue);
drivers/scsi/hpsa.c
7002
u8 q = *(u8 *) queue;
drivers/scsi/hpsa.c
7019
static irqreturn_t hpsa_msix_discard_completions(int irq, void *queue)
drivers/scsi/hpsa.c
7021
struct ctlr_info *h = queue_to_hba(queue);
drivers/scsi/hpsa.c
7023
u8 q = *(u8 *) queue;
drivers/scsi/hpsa.c
7035
static irqreturn_t do_hpsa_intr_intx(int irq, void *queue)
drivers/scsi/hpsa.c
7037
struct ctlr_info *h = queue_to_hba((u8 *) queue);
drivers/scsi/hpsa.c
7039
u8 q = *(u8 *) queue;
drivers/scsi/hpsa.c
7054
static irqreturn_t do_hpsa_intr_msi(int irq, void *queue)
drivers/scsi/hpsa.c
7056
struct ctlr_info *h = queue_to_hba(queue);
drivers/scsi/hpsa.c
7058
u8 q = *(u8 *) queue;
drivers/scsi/hpsa.c
7474
unsigned int queue, cpu;
drivers/scsi/hpsa.c
7476
for (queue = 0; queue < h->msix_vectors; queue++) {
drivers/scsi/hpsa.c
7477
mask = pci_irq_get_affinity(h->pdev, queue);
drivers/scsi/hpsa.c
7482
h->reply_map[cpu] = queue;
drivers/scsi/ibmvscsi/ibmvfc.c
1034
struct ibmvfc_event_pool *pool = &evt->queue->evt_pool;
drivers/scsi/ibmvscsi/ibmvfc.c
1041
spin_lock_irqsave(&evt->queue->l_lock, flags);
drivers/scsi/ibmvscsi/ibmvfc.c
1042
list_add_tail(&evt->queue_list, &evt->queue->free);
drivers/scsi/ibmvscsi/ibmvfc.c
1045
evt->queue->reserved_free++;
drivers/scsi/ibmvscsi/ibmvfc.c
1047
evt->queue->evt_free++;
drivers/scsi/ibmvscsi/ibmvfc.c
1051
spin_unlock_irqrestore(&evt->queue->l_lock, flags);
drivers/scsi/ibmvscsi/ibmvfc.c
1235
list_for_each_entry(tgt, &vhost->targets, queue)
drivers/scsi/ibmvscsi/ibmvfc.c
1538
static struct ibmvfc_event *__ibmvfc_get_event(struct ibmvfc_queue *queue, int reserved)
drivers/scsi/ibmvscsi/ibmvfc.c
1543
spin_lock_irqsave(&queue->l_lock, flags);
drivers/scsi/ibmvscsi/ibmvfc.c
1544
if (reserved && queue->reserved_free) {
drivers/scsi/ibmvscsi/ibmvfc.c
1545
evt = list_entry(queue->free.next, struct ibmvfc_event, queue_list);
drivers/scsi/ibmvscsi/ibmvfc.c
1547
queue->reserved_free--;
drivers/scsi/ibmvscsi/ibmvfc.c
1548
} else if (queue->evt_free) {
drivers/scsi/ibmvscsi/ibmvfc.c
1549
evt = list_entry(queue->free.next, struct ibmvfc_event, queue_list);
drivers/scsi/ibmvscsi/ibmvfc.c
1550
queue->evt_free--;
drivers/scsi/ibmvscsi/ibmvfc.c
1558
spin_unlock_irqrestore(&queue->l_lock, flags);
drivers/scsi/ibmvscsi/ibmvfc.c
1562
#define ibmvfc_get_event(queue) __ibmvfc_get_event(queue, 0)
drivers/scsi/ibmvscsi/ibmvfc.c
1563
#define ibmvfc_get_reserved_event(queue) __ibmvfc_get_event(queue, 1)
drivers/scsi/ibmvscsi/ibmvfc.c
1737
spin_lock_irqsave(&evt->queue->l_lock, flags);
drivers/scsi/ibmvscsi/ibmvfc.c
1738
list_add_tail(&evt->queue_list, &evt->queue->sent);
drivers/scsi/ibmvscsi/ibmvfc.c
1743
if (evt->queue->fmt == IBMVFC_SUB_CRQ_FMT)
drivers/scsi/ibmvscsi/ibmvfc.c
1745
evt->queue->vios_cookie,
drivers/scsi/ibmvscsi/ibmvfc.c
1756
spin_unlock_irqrestore(&evt->queue->l_lock, flags);
drivers/scsi/ibmvscsi/ibmvfc.c
1782
spin_unlock_irqrestore(&evt->queue->l_lock, flags);
drivers/scsi/ibmvscsi/ibmvfc.c
1835
list_for_each_entry(tgt, &vhost->targets, queue) {
drivers/scsi/ibmvscsi/ibmvfc.c
2133
list_for_each_entry(tgt, &vhost->targets, queue) {
drivers/scsi/ibmvscsi/ibmvfc.c
2473
list_for_each_entry(loop_evt, &evt->queue->free, queue_list)
drivers/scsi/ibmvscsi/ibmvfc.c
2558
static struct ibmvfc_event *ibmvfc_init_tmf(struct ibmvfc_queue *queue,
drivers/scsi/ibmvscsi/ibmvfc.c
2568
evt = ibmvfc_get_reserved_event(queue);
drivers/scsi/ibmvscsi/ibmvfc.c
2651
status = be16_to_cpu(evt->queue->cancel_rsp.mad_common.status);
drivers/scsi/ibmvscsi/ibmvfc.c
3105
list_for_each_entry(tgt, &vhost->targets, queue) {
drivers/scsi/ibmvscsi/ibmvfc.c
3245
list_for_each_entry(tgt, &vhost->targets, queue) {
drivers/scsi/ibmvscsi/ibmvfc.c
3360
spin_lock(&evt->queue->l_lock);
drivers/scsi/ibmvscsi/ibmvfc.c
3362
spin_unlock(&evt->queue->l_lock);
drivers/scsi/ibmvscsi/ibmvfc.c
3748
struct ibmvfc_queue *queue = &vhost->crq;
drivers/scsi/ibmvscsi/ibmvfc.c
3751
crq = &queue->msgs.crq[queue->cur];
drivers/scsi/ibmvscsi/ibmvfc.c
3753
if (++queue->cur == queue->size)
drivers/scsi/ibmvscsi/ibmvfc.c
3754
queue->cur = 0;
drivers/scsi/ibmvscsi/ibmvfc.c
3882
if (unlikely(!ibmvfc_valid_event(&evt->queue->evt_pool, evt))) {
drivers/scsi/ibmvscsi/ibmvfc.c
3894
spin_lock(&evt->queue->l_lock);
drivers/scsi/ibmvscsi/ibmvfc.c
3896
spin_unlock(&evt->queue->l_lock);
drivers/scsi/ibmvscsi/ibmvfc.c
4866
list_for_each_entry(tgt, &vhost->targets, queue) {
drivers/scsi/ibmvscsi/ibmvfc.c
4873
list_for_each_entry(tgt, &vhost->targets, queue) {
drivers/scsi/ibmvscsi/ibmvfc.c
4930
list_add_tail(&tgt->queue, &vhost->targets);
drivers/scsi/ibmvscsi/ibmvfc.c
5386
list_for_each_entry(tgt, &vhost->targets, queue) {
drivers/scsi/ibmvscsi/ibmvfc.c
5406
list_for_each_entry(tgt, &vhost->targets, queue) {
drivers/scsi/ibmvscsi/ibmvfc.c
5436
list_for_each_entry(tgt, &vhost->targets, queue)
drivers/scsi/ibmvscsi/ibmvfc.c
5439
list_for_each_entry(tgt, &vhost->targets, queue)
drivers/scsi/ibmvscsi/ibmvfc.c
5447
list_for_each_entry(tgt, &vhost->targets, queue)
drivers/scsi/ibmvscsi/ibmvfc.c
5450
list_for_each_entry(tgt, &vhost->targets, queue)
drivers/scsi/ibmvscsi/ibmvfc.c
5520
list_del(&tgt->queue);
drivers/scsi/ibmvscsi/ibmvfc.c
5638
list_for_each_entry(tgt, &vhost->targets, queue)
drivers/scsi/ibmvscsi/ibmvfc.c
5643
list_for_each_entry(tgt, &vhost->targets, queue) {
drivers/scsi/ibmvscsi/ibmvfc.c
5655
list_for_each_entry(tgt, &vhost->targets, queue) {
drivers/scsi/ibmvscsi/ibmvfc.c
5667
list_for_each_entry(tgt, &vhost->targets, queue) {
drivers/scsi/ibmvscsi/ibmvfc.c
5672
list_del(&tgt->queue);
drivers/scsi/ibmvscsi/ibmvfc.c
5742
list_for_each_entry(tgt, &vhost->targets, queue) {
drivers/scsi/ibmvscsi/ibmvfc.c
5799
struct ibmvfc_queue *queue,
drivers/scsi/ibmvscsi/ibmvfc.c
5806
spin_lock_init(&queue->_lock);
drivers/scsi/ibmvscsi/ibmvfc.c
5807
queue->q_lock = &queue->_lock;
drivers/scsi/ibmvscsi/ibmvfc.c
5811
fmt_size = sizeof(*queue->msgs.crq);
drivers/scsi/ibmvscsi/ibmvfc.c
5812
queue->total_depth = scsi_qdepth + IBMVFC_NUM_INTERNAL_REQ;
drivers/scsi/ibmvscsi/ibmvfc.c
5813
queue->evt_depth = scsi_qdepth;
drivers/scsi/ibmvscsi/ibmvfc.c
5814
queue->reserved_depth = IBMVFC_NUM_INTERNAL_REQ;
drivers/scsi/ibmvscsi/ibmvfc.c
5817
fmt_size = sizeof(*queue->msgs.async);
drivers/scsi/ibmvscsi/ibmvfc.c
5820
fmt_size = sizeof(*queue->msgs.scrq);
drivers/scsi/ibmvscsi/ibmvfc.c
5822
queue->total_depth = scsi_qdepth + IBMVFC_NUM_INTERNAL_SUBQ_REQ;
drivers/scsi/ibmvscsi/ibmvfc.c
5823
queue->evt_depth = scsi_qdepth;
drivers/scsi/ibmvscsi/ibmvfc.c
5824
queue->reserved_depth = IBMVFC_NUM_INTERNAL_SUBQ_REQ;
drivers/scsi/ibmvscsi/ibmvfc.c
5831
queue->fmt = fmt;
drivers/scsi/ibmvscsi/ibmvfc.c
5832
if (ibmvfc_init_event_pool(vhost, queue)) {
drivers/scsi/ibmvscsi/ibmvfc.c
5837
queue->msgs.handle = (void *)get_zeroed_page(GFP_KERNEL);
drivers/scsi/ibmvscsi/ibmvfc.c
5838
if (!queue->msgs.handle)
drivers/scsi/ibmvscsi/ibmvfc.c
5841
queue->msg_token = dma_map_single(dev, queue->msgs.handle, PAGE_SIZE,
drivers/scsi/ibmvscsi/ibmvfc.c
5844
if (dma_mapping_error(dev, queue->msg_token)) {
drivers/scsi/ibmvscsi/ibmvfc.c
5845
free_page((unsigned long)queue->msgs.handle);
drivers/scsi/ibmvscsi/ibmvfc.c
5846
queue->msgs.handle = NULL;
drivers/scsi/ibmvscsi/ibmvfc.c
5850
queue->cur = 0;
drivers/scsi/ibmvscsi/ibmvfc.c
5851
queue->size = PAGE_SIZE / fmt_size;
drivers/scsi/ibmvscsi/ibmvfc.c
5853
queue->vhost = vhost;
drivers/scsi/ibmvscsi/ibmvfc.c
6279
list_for_each_entry(tgt, &vhost->targets, queue) {
drivers/scsi/ibmvscsi/ibmvfc.c
6400
list_add_tail(&vhost->queue, &ibmvfc_head);
drivers/scsi/ibmvscsi/ibmvfc.c
6457
list_del(&vhost->queue);
drivers/scsi/ibmvscsi/ibmvfc.c
685
list_for_each_entry(tgt, &vhost->targets, queue)
drivers/scsi/ibmvscsi/ibmvfc.c
718
list_for_each_entry(tgt, &vhost->targets, queue) {
drivers/scsi/ibmvscsi/ibmvfc.c
790
struct ibmvfc_queue *queue)
drivers/scsi/ibmvscsi/ibmvfc.c
793
struct ibmvfc_event_pool *pool = &queue->evt_pool;
drivers/scsi/ibmvscsi/ibmvfc.c
796
if (!queue->total_depth)
drivers/scsi/ibmvscsi/ibmvfc.c
799
pool->size = queue->total_depth;
drivers/scsi/ibmvscsi/ibmvfc.c
813
INIT_LIST_HEAD(&queue->sent);
drivers/scsi/ibmvscsi/ibmvfc.c
814
INIT_LIST_HEAD(&queue->free);
drivers/scsi/ibmvscsi/ibmvfc.c
815
queue->evt_free = queue->evt_depth;
drivers/scsi/ibmvscsi/ibmvfc.c
816
queue->reserved_free = queue->reserved_depth;
drivers/scsi/ibmvscsi/ibmvfc.c
817
spin_lock_init(&queue->l_lock);
drivers/scsi/ibmvscsi/ibmvfc.c
834
evt->queue = queue;
drivers/scsi/ibmvscsi/ibmvfc.c
836
list_add_tail(&evt->queue_list, &queue->free);
drivers/scsi/ibmvscsi/ibmvfc.c
850
struct ibmvfc_queue *queue)
drivers/scsi/ibmvscsi/ibmvfc.c
853
struct ibmvfc_event_pool *pool = &queue->evt_pool;
drivers/scsi/ibmvscsi/ibmvfc.c
880
struct ibmvfc_queue *queue)
drivers/scsi/ibmvscsi/ibmvfc.c
884
dma_unmap_single(dev, queue->msg_token, PAGE_SIZE, DMA_BIDIRECTIONAL);
drivers/scsi/ibmvscsi/ibmvfc.c
885
free_page((unsigned long)queue->msgs.handle);
drivers/scsi/ibmvscsi/ibmvfc.c
886
queue->msgs.handle = NULL;
drivers/scsi/ibmvscsi/ibmvfc.c
888
ibmvfc_free_event_pool(vhost, queue);
drivers/scsi/ibmvscsi/ibmvfc.h
725
struct list_head queue;
drivers/scsi/ibmvscsi/ibmvfc.h
753
struct ibmvfc_queue *queue;
drivers/scsi/ibmvscsi/ibmvfc.h
864
struct list_head queue;
drivers/scsi/ibmvscsi/ibmvscsi.c
141
static void ibmvscsi_release_crq_queue(struct crq_queue *queue,
drivers/scsi/ibmvscsi/ibmvscsi.c
155
queue->msg_token,
drivers/scsi/ibmvscsi/ibmvscsi.c
156
queue->size * sizeof(*queue->msgs), DMA_BIDIRECTIONAL);
drivers/scsi/ibmvscsi/ibmvscsi.c
157
free_page((unsigned long)queue->msgs);
drivers/scsi/ibmvscsi/ibmvscsi.c
167
static struct viosrp_crq *crq_queue_next_crq(struct crq_queue *queue)
drivers/scsi/ibmvscsi/ibmvscsi.c
172
spin_lock_irqsave(&queue->lock, flags);
drivers/scsi/ibmvscsi/ibmvscsi.c
173
crq = &queue->msgs[queue->cur];
drivers/scsi/ibmvscsi/ibmvscsi.c
175
if (++queue->cur == queue->size)
drivers/scsi/ibmvscsi/ibmvscsi.c
176
queue->cur = 0;
drivers/scsi/ibmvscsi/ibmvscsi.c
184
spin_unlock_irqrestore(&queue->lock, flags);
drivers/scsi/ibmvscsi/ibmvscsi.c
2141
rc = ibmvscsi_reset_crq_queue(&hostdata->queue, hostdata);
drivers/scsi/ibmvscsi/ibmvscsi.c
2150
rc = ibmvscsi_reenable_crq_queue(&hostdata->queue, hostdata);
drivers/scsi/ibmvscsi/ibmvscsi.c
221
while ((crq = crq_queue_next_crq(&hostdata->queue)) != NULL) {
drivers/scsi/ibmvscsi/ibmvscsi.c
2268
rc = ibmvscsi_init_crq_queue(&hostdata->queue, hostdata, max_events);
drivers/scsi/ibmvscsi/ibmvscsi.c
228
crq = crq_queue_next_crq(&hostdata->queue);
drivers/scsi/ibmvscsi/ibmvscsi.c
2333
ibmvscsi_release_crq_queue(&hostdata->queue, hostdata, max_events);
drivers/scsi/ibmvscsi/ibmvscsi.c
2354
ibmvscsi_release_crq_queue(&hostdata->queue, hostdata,
drivers/scsi/ibmvscsi/ibmvscsi.c
284
static int ibmvscsi_reset_crq_queue(struct crq_queue *queue,
drivers/scsi/ibmvscsi/ibmvscsi.c
298
memset(queue->msgs, 0x00, PAGE_SIZE);
drivers/scsi/ibmvscsi/ibmvscsi.c
299
queue->cur = 0;
drivers/scsi/ibmvscsi/ibmvscsi.c
306
queue->msg_token, PAGE_SIZE);
drivers/scsi/ibmvscsi/ibmvscsi.c
326
static int ibmvscsi_init_crq_queue(struct crq_queue *queue,
drivers/scsi/ibmvscsi/ibmvscsi.c
334
queue->msgs = (struct viosrp_crq *)get_zeroed_page(GFP_KERNEL);
drivers/scsi/ibmvscsi/ibmvscsi.c
336
if (!queue->msgs)
drivers/scsi/ibmvscsi/ibmvscsi.c
338
queue->size = PAGE_SIZE / sizeof(*queue->msgs);
drivers/scsi/ibmvscsi/ibmvscsi.c
340
queue->msg_token = dma_map_single(hostdata->dev, queue->msgs,
drivers/scsi/ibmvscsi/ibmvscsi.c
341
queue->size * sizeof(*queue->msgs),
drivers/scsi/ibmvscsi/ibmvscsi.c
344
if (dma_mapping_error(hostdata->dev, queue->msg_token))
drivers/scsi/ibmvscsi/ibmvscsi.c
352
queue->msg_token, PAGE_SIZE);
drivers/scsi/ibmvscsi/ibmvscsi.c
355
rc = ibmvscsi_reset_crq_queue(queue,
drivers/scsi/ibmvscsi/ibmvscsi.c
367
queue->cur = 0;
drivers/scsi/ibmvscsi/ibmvscsi.c
368
spin_lock_init(&queue->lock);
drivers/scsi/ibmvscsi/ibmvscsi.c
399
queue->msg_token,
drivers/scsi/ibmvscsi/ibmvscsi.c
400
queue->size * sizeof(*queue->msgs), DMA_BIDIRECTIONAL);
drivers/scsi/ibmvscsi/ibmvscsi.c
402
free_page((unsigned long)queue->msgs);
drivers/scsi/ibmvscsi/ibmvscsi.c
412
static int ibmvscsi_reenable_crq_queue(struct crq_queue *queue,
drivers/scsi/ibmvscsi/ibmvscsi.h
92
struct crq_queue queue;
drivers/scsi/ibmvscsi_tgt/libsrp.c
142
if (kfifo_out_locked(&target->iu_queue.queue, (void *)&iue,
drivers/scsi/ibmvscsi_tgt/libsrp.c
157
kfifo_in_locked(&iue->target->iu_queue.queue, (void *)&iue,
drivers/scsi/ibmvscsi_tgt/libsrp.c
38
kfifo_init(&q->queue, (void *)q->pool, max * sizeof(void *));
drivers/scsi/ibmvscsi_tgt/libsrp.c
41
kfifo_in(&q->queue, (void *)&iue, sizeof(void *));
drivers/scsi/ibmvscsi_tgt/libsrp.h
79
struct kfifo queue;
drivers/scsi/ipr.c
10009
list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q);
drivers/scsi/ipr.c
10032
list_for_each_entry(ioa_cfg, &ipr_ioa_head, queue) {
drivers/scsi/ipr.c
1064
list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_pending_q);
drivers/scsi/ipr.c
1065
list_add_tail(&hostrcb->queue, &ioa_cfg->hostrcb_pending_q);
drivers/scsi/ipr.c
1089
list_add_tail(&hostrcb->queue, &ioa_cfg->hostrcb_free_q);
drivers/scsi/ipr.c
1131
list_for_each_entry(gscsi_res, &ioa_cfg->used_res_q, queue) {
drivers/scsi/ipr.c
1323
list_for_each_entry(gscsi_res, &ioa_cfg->used_res_q, queue)
drivers/scsi/ipr.c
1357
list_for_each_entry(res, &ioa_cfg->used_res_q, queue) {
drivers/scsi/ipr.c
1373
struct ipr_resource_entry, queue);
drivers/scsi/ipr.c
1375
list_del(&res->queue);
drivers/scsi/ipr.c
1377
list_add_tail(&res->queue, &ioa_cfg->used_res_q);
drivers/scsi/ipr.c
1389
list_move_tail(&res->queue, &ioa_cfg->free_res_q);
drivers/scsi/ipr.c
1415
list_del_init(&hostrcb->queue);
drivers/scsi/ipr.c
1416
list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q);
drivers/scsi/ipr.c
2525
struct ipr_hostrcb, queue);
drivers/scsi/ipr.c
2530
struct ipr_hostrcb, queue);
drivers/scsi/ipr.c
2533
list_del_init(&hostrcb->queue);
drivers/scsi/ipr.c
2560
list_del_init(&hostrcb->queue);
drivers/scsi/ipr.c
2561
list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q);
drivers/scsi/ipr.c
2573
list_add_tail(&hostrcb->queue, &ioa_cfg->hostrcb_report_q);
drivers/scsi/ipr.c
2701
list_for_each_entry(res, &ioa_cfg->used_res_q, queue) {
drivers/scsi/ipr.c
3257
list_for_each_entry(res, &ioa_cfg->used_res_q, queue) {
drivers/scsi/ipr.c
3263
list_move_tail(&res->queue, &ioa_cfg->free_res_q);
drivers/scsi/ipr.c
3276
list_for_each_entry(res, &ioa_cfg->used_res_q, queue) {
drivers/scsi/ipr.c
4102
struct ipr_hostrcb, queue);
drivers/scsi/ipr.c
4125
struct ipr_hostrcb, queue);
drivers/scsi/ipr.c
4132
list_move_tail(&hostrcb->queue, &ioa_cfg->hostrcb_free_q);
drivers/scsi/ipr.c
4693
list_for_each_entry(res, &ioa_cfg->used_res_q, queue) {
drivers/scsi/ipr.c
4737
list_for_each_entry(res, &ioa_cfg->used_res_q, queue) {
drivers/scsi/ipr.c
4888
list_for_each_entry(loop_cmd, &ipr_cmd->hrrq->hrrq_free_q, queue) {
drivers/scsi/ipr.c
5035
list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q);
drivers/scsi/ipr.c
5121
list_for_each_entry(res, &ioa_cfg->used_res_q, queue) {
drivers/scsi/ipr.c
5137
list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q);
drivers/scsi/ipr.c
5262
list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q);
drivers/scsi/ipr.c
5349
list_del(&ioa_cfg->reset_cmd->queue);
drivers/scsi/ipr.c
5364
list_del(&ioa_cfg->reset_cmd->queue);
drivers/scsi/ipr.c
5449
list_move_tail(&ipr_cmd->queue, doneq);
drivers/scsi/ipr.c
5482
list_for_each_entry_safe(ipr_cmd, temp, &doneq, queue) {
drivers/scsi/ipr.c
5483
list_del(&ipr_cmd->queue);
drivers/scsi/ipr.c
5551
list_for_each_entry_safe(ipr_cmd, temp, &doneq, queue) {
drivers/scsi/ipr.c
5552
list_del(&ipr_cmd->queue);
drivers/scsi/ipr.c
5601
list_for_each_entry_safe(ipr_cmd, temp, &doneq, queue) {
drivers/scsi/ipr.c
5602
list_del(&ipr_cmd->queue);
drivers/scsi/ipr.c
5756
list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q);
drivers/scsi/ipr.c
6193
list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q);
drivers/scsi/ipr.c
6222
list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q);
drivers/scsi/ipr.c
6337
list_add_tail(&ipr_cmd->queue, &hrrq->hrrq_free_q);
drivers/scsi/ipr.c
6345
list_add_tail(&ipr_cmd->queue, &hrrq->hrrq_free_q);
drivers/scsi/ipr.c
6356
list_add_tail(&ipr_cmd->queue, &hrrq->hrrq_pending_q);
drivers/scsi/ipr.c
6449
list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q);
drivers/scsi/ipr.c
6484
list_for_each_entry(res, &ioa_cfg->used_res_q, queue) {
drivers/scsi/ipr.c
6493
list_del_init(&ioa_cfg->hostrcb[j]->queue);
drivers/scsi/ipr.c
6508
list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q);
drivers/scsi/ipr.c
6554
list_for_each_entry_continue(res, &ioa_cfg->used_res_q, queue) {
drivers/scsi/ipr.c
6791
struct ipr_resource_entry, queue);
drivers/scsi/ipr.c
6843
list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q);
drivers/scsi/ipr.c
6865
struct ipr_resource_entry, queue);
drivers/scsi/ipr.c
691
struct ipr_cmnd, queue);
drivers/scsi/ipr.c
692
list_del(&ipr_cmd->queue);
drivers/scsi/ipr.c
7018
list_for_each_entry_safe(res, temp, &ioa_cfg->used_res_q, queue)
drivers/scsi/ipr.c
7019
list_move_tail(&res->queue, &old_res);
drivers/scsi/ipr.c
7033
list_for_each_entry_safe(res, temp, &old_res, queue) {
drivers/scsi/ipr.c
7035
list_move_tail(&res->queue, &ioa_cfg->used_res_q);
drivers/scsi/ipr.c
7049
struct ipr_resource_entry, queue);
drivers/scsi/ipr.c
7050
list_move_tail(&res->queue, &ioa_cfg->used_res_q);
drivers/scsi/ipr.c
7060
list_for_each_entry_safe(res, temp, &old_res, queue) {
drivers/scsi/ipr.c
7064
list_move_tail(&res->queue, &ioa_cfg->used_res_q);
drivers/scsi/ipr.c
7068
list_for_each_entry_safe(res, temp, &old_res, queue) {
drivers/scsi/ipr.c
7070
list_move_tail(&res->queue, &ioa_cfg->free_res_q);
drivers/scsi/ipr.c
7486
list_del(&ipr_cmd->queue);
drivers/scsi/ipr.c
7512
list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_pending_q);
drivers/scsi/ipr.c
7605
list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_pending_q);
drivers/scsi/ipr.c
7674
list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_pending_q);
drivers/scsi/ipr.c
7766
struct ipr_hostrcb, queue);
drivers/scsi/ipr.c
7767
list_del_init(&hostrcb->queue);
drivers/scsi/ipr.c
7784
list_add_tail(&hostrcb->queue, &ioa_cfg->hostrcb_free_q);
drivers/scsi/ipr.c
822
list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q);
drivers/scsi/ipr.c
8229
list_for_each_entry(loop_cmd, &hrrq->hrrq_pending_q, queue) {
drivers/scsi/ipr.c
8232
list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q);
drivers/scsi/ipr.c
8268
list_for_each_entry(hcam_cmd, &hrrq->hrrq_pending_q, queue) {
drivers/scsi/ipr.c
8435
list_add_tail(&ipr_cmd->queue,
drivers/scsi/ipr.c
8574
list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_pending_q);
drivers/scsi/ipr.c
863
temp, &hrrq->hrrq_pending_q, queue) {
drivers/scsi/ipr.c
864
list_del(&ipr_cmd->queue);
drivers/scsi/ipr.c
8946
list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_free_q);
drivers/scsi/ipr.c
8974
list_add_tail(&ioa_cfg->res_entries[i].queue, &ioa_cfg->free_res_q);
drivers/scsi/ipr.c
9026
list_add_tail(&ioa_cfg->hostrcb[i]->queue, &ioa_cfg->hostrcb_free_q);
drivers/scsi/ipr.c
930
list_add_tail(&ipr_cmd->queue, &ipr_cmd->hrrq->hrrq_pending_q);
drivers/scsi/ipr.c
9615
list_add_tail(&ioa_cfg->queue, &ipr_ioa_head);
drivers/scsi/ipr.c
9707
list_del(&ioa_cfg->queue);
drivers/scsi/ipr.h
1182
struct list_head queue;
drivers/scsi/ipr.h
1266
struct list_head queue;
drivers/scsi/ipr.h
1409
struct list_head queue;
drivers/scsi/ipr.h
1550
struct list_head queue;
drivers/scsi/ips.c
2739
ips_putq_scb_head(ips_scb_queue_t * queue, ips_scb_t * item)
drivers/scsi/ips.c
2746
item->q_next = queue->head;
drivers/scsi/ips.c
2747
queue->head = item;
drivers/scsi/ips.c
2749
if (!queue->tail)
drivers/scsi/ips.c
2750
queue->tail = item;
drivers/scsi/ips.c
2752
queue->count++;
drivers/scsi/ips.c
2767
ips_removeq_scb_head(ips_scb_queue_t * queue)
drivers/scsi/ips.c
2773
item = queue->head;
drivers/scsi/ips.c
2779
queue->head = item->q_next;
drivers/scsi/ips.c
2782
if (queue->tail == item)
drivers/scsi/ips.c
2783
queue->tail = NULL;
drivers/scsi/ips.c
2785
queue->count--;
drivers/scsi/ips.c
2802
ips_removeq_scb(ips_scb_queue_t * queue, ips_scb_t * item)
drivers/scsi/ips.c
2811
if (item == queue->head) {
drivers/scsi/ips.c
2812
return (ips_removeq_scb_head(queue));
drivers/scsi/ips.c
2815
p = queue->head;
drivers/scsi/ips.c
2825
queue->tail = p;
drivers/scsi/ips.c
2828
queue->count--;
drivers/scsi/ips.c
2847
static void ips_putq_wait_tail(ips_wait_queue_entry_t *queue, struct scsi_cmnd *item)
drivers/scsi/ips.c
2856
if (queue->tail)
drivers/scsi/ips.c
2857
queue->tail->host_scribble = (char *) item;
drivers/scsi/ips.c
2859
queue->tail = item;
drivers/scsi/ips.c
2861
if (!queue->head)
drivers/scsi/ips.c
2862
queue->head = item;
drivers/scsi/ips.c
2864
queue->count++;
drivers/scsi/ips.c
2878
static struct scsi_cmnd *ips_removeq_wait_head(ips_wait_queue_entry_t *queue)
drivers/scsi/ips.c
2884
item = queue->head;
drivers/scsi/ips.c
2890
queue->head = (struct scsi_cmnd *) item->host_scribble;
drivers/scsi/ips.c
2893
if (queue->tail == item)
drivers/scsi/ips.c
2894
queue->tail = NULL;
drivers/scsi/ips.c
2896
queue->count--;
drivers/scsi/ips.c
2912
static struct scsi_cmnd *ips_removeq_wait(ips_wait_queue_entry_t *queue,
drivers/scsi/ips.c
2922
if (item == queue->head) {
drivers/scsi/ips.c
2923
return (ips_removeq_wait_head(queue));
drivers/scsi/ips.c
2926
p = queue->head;
drivers/scsi/ips.c
2936
queue->tail = p;
drivers/scsi/ips.c
2939
queue->count--;
drivers/scsi/ips.c
2959
ips_putq_copp_tail(ips_copp_queue_t * queue, ips_copp_wait_item_t * item)
drivers/scsi/ips.c
2968
if (queue->tail)
drivers/scsi/ips.c
2969
queue->tail->next = item;
drivers/scsi/ips.c
2971
queue->tail = item;
drivers/scsi/ips.c
2973
if (!queue->head)
drivers/scsi/ips.c
2974
queue->head = item;
drivers/scsi/ips.c
2976
queue->count++;
drivers/scsi/ips.c
2991
ips_removeq_copp_head(ips_copp_queue_t * queue)
drivers/scsi/ips.c
2997
item = queue->head;
drivers/scsi/ips.c
3003
queue->head = item->next;
drivers/scsi/ips.c
3006
if (queue->tail == item)
drivers/scsi/ips.c
3007
queue->tail = NULL;
drivers/scsi/ips.c
3009
queue->count--;
drivers/scsi/ips.c
3026
ips_removeq_copp(ips_copp_queue_t * queue, ips_copp_wait_item_t * item)
drivers/scsi/ips.c
3035
if (item == queue->head) {
drivers/scsi/ips.c
3036
return (ips_removeq_copp_head(queue));
drivers/scsi/ips.c
3039
p = queue->head;
drivers/scsi/ips.c
3049
queue->tail = p;
drivers/scsi/ips.c
3052
queue->count--;
drivers/scsi/libiscsi.c
1718
if (!kfifo_out(&conn->session->cmdpool.queue,
drivers/scsi/libiscsi.c
2787
kfifo_init(&q->queue, (void*)q->pool, max * sizeof(void*));
drivers/scsi/libiscsi.c
2795
kfifo_in(&q->queue, (void*)&q->pool[i], sizeof(void*));
drivers/scsi/libiscsi.c
3206
if (!kfifo_out(&session->cmdpool.queue,
drivers/scsi/libiscsi.c
3231
kfifo_in(&session->cmdpool.queue, (void*)&conn->login_task,
drivers/scsi/libiscsi.c
3277
kfifo_in(&session->cmdpool.queue, (void*)&conn->login_task,
drivers/scsi/libiscsi.c
473
kfifo_in(&session->cmdpool.queue, (void*)&task, sizeof(void*));
drivers/scsi/libiscsi.c
749
if (!kfifo_out(&session->cmdpool.queue,
drivers/scsi/libiscsi_tcp.c
1029
kfifo_in(&tcp_task->r2tpool.queue,
drivers/scsi/libiscsi_tcp.c
455
kfifo_in(&tcp_task->r2tpool.queue, (void*)&r2t,
drivers/scsi/libiscsi_tcp.c
462
kfifo_in(&tcp_task->r2tpool.queue, (void*)&r2t,
drivers/scsi/libiscsi_tcp.c
611
rc = kfifo_out(&tcp_task->r2tpool.queue, (void *)&r2t, sizeof(void *));
drivers/scsi/lpfc/lpfc_sli.c
15839
lpfc_sli4_queue_free(struct lpfc_queue *queue)
drivers/scsi/lpfc/lpfc_sli.c
15843
if (!queue)
drivers/scsi/lpfc/lpfc_sli.c
15846
if (!list_empty(&queue->wq_list))
drivers/scsi/lpfc/lpfc_sli.c
15847
list_del(&queue->wq_list);
drivers/scsi/lpfc/lpfc_sli.c
15849
while (!list_empty(&queue->page_list)) {
drivers/scsi/lpfc/lpfc_sli.c
15850
list_remove_head(&queue->page_list, dmabuf, struct lpfc_dmabuf,
drivers/scsi/lpfc/lpfc_sli.c
15852
dma_free_coherent(&queue->phba->pcidev->dev, queue->page_size,
drivers/scsi/lpfc/lpfc_sli.c
15856
if (queue->rqbp) {
drivers/scsi/lpfc/lpfc_sli.c
15857
lpfc_free_rq_buffer(queue->phba, queue);
drivers/scsi/lpfc/lpfc_sli.c
15858
kfree(queue->rqbp);
drivers/scsi/lpfc/lpfc_sli.c
15861
if (!list_empty(&queue->cpu_list))
drivers/scsi/lpfc/lpfc_sli.c
15862
list_del(&queue->cpu_list);
drivers/scsi/lpfc/lpfc_sli.c
15864
kfree(queue);
drivers/scsi/lpfc/lpfc_sli.c
15884
struct lpfc_queue *queue;
drivers/scsi/lpfc/lpfc_sli.c
15898
queue = kzalloc_node(sizeof(*queue) + (sizeof(void *) * pgcnt),
drivers/scsi/lpfc/lpfc_sli.c
15900
if (!queue)
drivers/scsi/lpfc/lpfc_sli.c
15903
INIT_LIST_HEAD(&queue->list);
drivers/scsi/lpfc/lpfc_sli.c
15904
INIT_LIST_HEAD(&queue->_poll_list);
drivers/scsi/lpfc/lpfc_sli.c
15905
INIT_LIST_HEAD(&queue->wq_list);
drivers/scsi/lpfc/lpfc_sli.c
15906
INIT_LIST_HEAD(&queue->wqfull_list);
drivers/scsi/lpfc/lpfc_sli.c
15907
INIT_LIST_HEAD(&queue->page_list);
drivers/scsi/lpfc/lpfc_sli.c
15908
INIT_LIST_HEAD(&queue->child_list);
drivers/scsi/lpfc/lpfc_sli.c
15909
INIT_LIST_HEAD(&queue->cpu_list);
drivers/scsi/lpfc/lpfc_sli.c
15914
queue->page_count = pgcnt;
drivers/scsi/lpfc/lpfc_sli.c
15915
queue->q_pgs = (void **)&queue[1];
drivers/scsi/lpfc/lpfc_sli.c
15916
queue->entry_cnt_per_pg = hw_page_size / entry_size;
drivers/scsi/lpfc/lpfc_sli.c
15917
queue->entry_size = entry_size;
drivers/scsi/lpfc/lpfc_sli.c
15918
queue->entry_count = entry_count;
drivers/scsi/lpfc/lpfc_sli.c
15919
queue->page_size = hw_page_size;
drivers/scsi/lpfc/lpfc_sli.c
15920
queue->phba = phba;
drivers/scsi/lpfc/lpfc_sli.c
15922
for (x = 0; x < queue->page_count; x++) {
drivers/scsi/lpfc/lpfc_sli.c
15935
list_add_tail(&dmabuf->list, &queue->page_list);
drivers/scsi/lpfc/lpfc_sli.c
15937
queue->q_pgs[x] = dmabuf->virt;
drivers/scsi/lpfc/lpfc_sli.c
15939
INIT_WORK(&queue->irqwork, lpfc_sli4_hba_process_cq);
drivers/scsi/lpfc/lpfc_sli.c
15940
INIT_WORK(&queue->spwork, lpfc_sli4_sp_process_cq);
drivers/scsi/lpfc/lpfc_sli.c
15941
INIT_DELAYED_WORK(&queue->sched_irqwork, lpfc_sli4_dly_hba_process_cq);
drivers/scsi/lpfc/lpfc_sli.c
15942
INIT_DELAYED_WORK(&queue->sched_spwork, lpfc_sli4_dly_sp_process_cq);
drivers/scsi/lpfc/lpfc_sli.c
15946
return queue;
drivers/scsi/lpfc/lpfc_sli.c
15948
lpfc_sli4_queue_free(queue);
drivers/scsi/megaraid/megaraid_sas_base.c
5841
unsigned int queue, cpu, low_latency_index_start;
drivers/scsi/megaraid/megaraid_sas_base.c
5845
for (queue = low_latency_index_start; queue < instance->msix_vectors; queue++) {
drivers/scsi/megaraid/megaraid_sas_base.c
5846
mask = pci_irq_get_affinity(instance->pdev, queue);
drivers/scsi/megaraid/megaraid_sas_base.c
5851
instance->reply_map[cpu] = queue;
drivers/scsi/megaraid/megaraid_sas_base.c
5856
queue = low_latency_index_start;
drivers/scsi/megaraid/megaraid_sas_base.c
5858
instance->reply_map[cpu] = queue;
drivers/scsi/megaraid/megaraid_sas_base.c
5859
if (queue == (instance->msix_vectors - 1))
drivers/scsi/megaraid/megaraid_sas_base.c
5860
queue = low_latency_index_start;
drivers/scsi/megaraid/megaraid_sas_base.c
5862
queue++;
drivers/scsi/pmcraid.c
1573
list_for_each_entry(res, &pinstance->used_res_q, queue) {
drivers/scsi/pmcraid.c
158
list_for_each_entry(temp, &pinstance->used_res_q, queue) {
drivers/scsi/pmcraid.c
1609
struct pmcraid_resource_entry, queue);
drivers/scsi/pmcraid.c
1611
list_del(&res->queue);
drivers/scsi/pmcraid.c
1614
list_add_tail(&res->queue, &pinstance->used_res_q);
drivers/scsi/pmcraid.c
1632
list_move_tail(&res->queue, &pinstance->free_res_q);
drivers/scsi/pmcraid.c
3044
list_for_each_entry(temp, &pinstance->used_res_q, queue) {
drivers/scsi/pmcraid.c
3854
list_for_each_entry_safe(res, temp, &pinstance->used_res_q, queue) {
drivers/scsi/pmcraid.c
3870
list_move_tail(&res->queue,
drivers/scsi/pmcraid.c
3888
list_for_each_entry(res, &pinstance->used_res_q, queue) {
drivers/scsi/pmcraid.c
4368
list_del(&pinstance->res_entries[i].queue);
drivers/scsi/pmcraid.c
4397
list_add_tail(&pinstance->res_entries[i].queue,
drivers/scsi/pmcraid.c
5023
list_for_each_entry_safe(res, temp, &pinstance->used_res_q, queue)
drivers/scsi/pmcraid.c
5024
list_move_tail(&res->queue, &old_res);
drivers/scsi/pmcraid.c
5040
list_for_each_entry_safe(res, temp, &old_res, queue) {
drivers/scsi/pmcraid.c
5046
list_move_tail(&res->queue,
drivers/scsi/pmcraid.c
5063
struct pmcraid_resource_entry, queue);
drivers/scsi/pmcraid.c
5068
list_move_tail(&res->queue, &pinstance->used_res_q);
drivers/scsi/pmcraid.c
5087
list_for_each_entry_safe(res, temp, &old_res, queue) {
drivers/scsi/pmcraid.c
5093
list_move_tail(&res->queue, &pinstance->used_res_q);
drivers/scsi/pmcraid.c
5095
list_move_tail(&res->queue, &pinstance->free_res_q);
drivers/scsi/pmcraid.h
795
struct list_head queue; /* link to "to be exposed" resources */
drivers/scsi/qla2xxx/qla_dbg.c
587
qh->queue = htonl(TYPE_ATIO_QUEUE);
drivers/scsi/qla2xxx/qla_dbg.c
631
qh->queue = htonl(TYPE_REQUEST_QUEUE);
drivers/scsi/qla2xxx/qla_dbg.c
659
qh->queue = htonl(TYPE_RESPONSE_QUEUE);
drivers/scsi/qla2xxx/qla_dbg.h
254
__be32 queue;
drivers/scsi/sd.c
1119
blk_queue_disable_write_zeroes(sdkp->disk->queue);
drivers/scsi/sd.c
122
blk_queue_disable_discard(sdkp->disk->queue);
drivers/scsi/sd.c
229
lim = queue_limits_start_update(sdkp->disk->queue);
drivers/scsi/sd.c
231
ret = queue_limits_commit_update_frozen(sdkp->disk->queue,
drivers/scsi/sd.c
3748
*lim = queue_limits_start_update(sdkp->disk->queue);
drivers/scsi/sd.c
3826
err = queue_limits_commit_update_frozen(sdkp->disk->queue, lim);
drivers/scsi/sd.c
563
lim = queue_limits_start_update(sdkp->disk->queue);
drivers/scsi/sd.c
565
err = queue_limits_commit_update_frozen(sdkp->disk->queue, &lim);
drivers/scsi/sd.c
672
lim = queue_limits_start_update(sdkp->disk->queue);
drivers/scsi/sd.c
674
err = queue_limits_commit_update_frozen(sdkp->disk->queue, &lim);
drivers/scsi/sd_zbc.c
165
struct request_queue *q = sdkp->disk->queue;
drivers/scsi/sd_zbc.c
544
struct request_queue *q = disk->queue;
drivers/scsi/sr.c
945
rq = scsi_alloc_request(disk->queue, REQ_OP_DRV_IN, 0);
drivers/scsi/sr.c
950
ret = blk_rq_map_user(disk->queue, rq, NULL, ubuf, len, GFP_KERNEL);
drivers/soc/fsl/qbman/qman_test_stash.c
196
static DECLARE_WAIT_QUEUE_HEAD(queue);
drivers/soc/fsl/qbman/qman_test_stash.c
309
wake_up(&queue);
drivers/soc/fsl/qbman/qman_test_stash.c
613
wait_event(queue, loop_counter == HP_LOOPS);
drivers/soc/ixp4xx/ixp4xx-qmgr.c
100
if (queue >= HALF_QUEUES)
drivers/soc/ixp4xx/ixp4xx-qmgr.c
102
(queue - HALF_QUEUES)) & 0x01;
drivers/soc/ixp4xx/ixp4xx-qmgr.c
103
return __qmgr_get_stat1(queue) & QUEUE_STAT1_FULL;
drivers/soc/ixp4xx/ixp4xx-qmgr.c
112
int qmgr_stat_overflow(unsigned int queue)
drivers/soc/ixp4xx/ixp4xx-qmgr.c
114
return __qmgr_get_stat2(queue) & QUEUE_STAT2_OVERFLOW;
drivers/soc/ixp4xx/ixp4xx-qmgr.c
117
void qmgr_set_irq(unsigned int queue, int src,
drivers/soc/ixp4xx/ixp4xx-qmgr.c
123
if (queue < HALF_QUEUES) {
drivers/soc/ixp4xx/ixp4xx-qmgr.c
127
reg = &qmgr_regs->irqsrc[queue >> 3]; /* 8 queues per u32 */
drivers/soc/ixp4xx/ixp4xx-qmgr.c
128
bit = (queue % 8) * 4; /* 3 bits + 1 reserved bit per queue */
drivers/soc/ixp4xx/ixp4xx-qmgr.c
135
irq_handlers[queue] = handler;
drivers/soc/ixp4xx/ixp4xx-qmgr.c
136
irq_pdevs[queue] = pdev;
drivers/soc/ixp4xx/ixp4xx-qmgr.c
205
void qmgr_enable_irq(unsigned int queue)
drivers/soc/ixp4xx/ixp4xx-qmgr.c
208
int half = queue / 32;
drivers/soc/ixp4xx/ixp4xx-qmgr.c
209
u32 mask = 1 << (queue & (HALF_QUEUES - 1));
drivers/soc/ixp4xx/ixp4xx-qmgr.c
217
void qmgr_disable_irq(unsigned int queue)
drivers/soc/ixp4xx/ixp4xx-qmgr.c
220
int half = queue / 32;
drivers/soc/ixp4xx/ixp4xx-qmgr.c
221
u32 mask = 1 << (queue & (HALF_QUEUES - 1));
drivers/soc/ixp4xx/ixp4xx-qmgr.c
239
int qmgr_request_queue(unsigned int queue, unsigned int len /* dwords */,
drivers/soc/ixp4xx/ixp4xx-qmgr.c
244
int __qmgr_request_queue(unsigned int queue, unsigned int len /* dwords */,
drivers/soc/ixp4xx/ixp4xx-qmgr.c
252
BUG_ON(queue >= QUEUES);
drivers/soc/ixp4xx/ixp4xx-qmgr.c
287
if (__raw_readl(&qmgr_regs->sram[queue])) {
drivers/soc/ixp4xx/ixp4xx-qmgr.c
29
void qmgr_put_entry(unsigned int queue, u32 val)
drivers/soc/ixp4xx/ixp4xx-qmgr.c
303
" queue %i\n", queue);
drivers/soc/ixp4xx/ixp4xx-qmgr.c
313
__raw_writel(cfg | (addr << 14), &qmgr_regs->sram[queue]);
drivers/soc/ixp4xx/ixp4xx-qmgr.c
315
snprintf(qmgr_queue_descs[queue], sizeof(qmgr_queue_descs[0]),
drivers/soc/ixp4xx/ixp4xx-qmgr.c
318
qmgr_queue_descs[queue], queue, addr);
drivers/soc/ixp4xx/ixp4xx-qmgr.c
32
BUG_ON(!qmgr_queue_descs[queue]); /* not yet requested */
drivers/soc/ixp4xx/ixp4xx-qmgr.c
329
void qmgr_release_queue(unsigned int queue)
drivers/soc/ixp4xx/ixp4xx-qmgr.c
333
BUG_ON(queue >= QUEUES); /* not in valid range */
drivers/soc/ixp4xx/ixp4xx-qmgr.c
336
cfg = __raw_readl(&qmgr_regs->sram[queue]);
drivers/soc/ixp4xx/ixp4xx-qmgr.c
35
qmgr_queue_descs[queue], queue, val);
drivers/soc/ixp4xx/ixp4xx-qmgr.c
355
qmgr_queue_descs[queue], queue);
drivers/soc/ixp4xx/ixp4xx-qmgr.c
356
qmgr_queue_descs[queue][0] = '\x0';
drivers/soc/ixp4xx/ixp4xx-qmgr.c
359
while ((addr = qmgr_get_entry(queue)))
drivers/soc/ixp4xx/ixp4xx-qmgr.c
361
queue, addr);
drivers/soc/ixp4xx/ixp4xx-qmgr.c
363
__raw_writel(0, &qmgr_regs->sram[queue]);
drivers/soc/ixp4xx/ixp4xx-qmgr.c
369
irq_handlers[queue] = NULL; /* catch IRQ bugs */
drivers/soc/ixp4xx/ixp4xx-qmgr.c
37
__raw_writel(val, &qmgr_regs->acc[queue][0]);
drivers/soc/ixp4xx/ixp4xx-qmgr.c
40
u32 qmgr_get_entry(unsigned int queue)
drivers/soc/ixp4xx/ixp4xx-qmgr.c
43
val = __raw_readl(&qmgr_regs->acc[queue][0]);
drivers/soc/ixp4xx/ixp4xx-qmgr.c
45
BUG_ON(!qmgr_queue_descs[queue]); /* not yet requested */
drivers/soc/ixp4xx/ixp4xx-qmgr.c
48
qmgr_queue_descs[queue], queue, val);
drivers/soc/ixp4xx/ixp4xx-qmgr.c
53
static int __qmgr_get_stat1(unsigned int queue)
drivers/soc/ixp4xx/ixp4xx-qmgr.c
55
return (__raw_readl(&qmgr_regs->stat1[queue >> 3])
drivers/soc/ixp4xx/ixp4xx-qmgr.c
56
>> ((queue & 7) << 2)) & 0xF;
drivers/soc/ixp4xx/ixp4xx-qmgr.c
59
static int __qmgr_get_stat2(unsigned int queue)
drivers/soc/ixp4xx/ixp4xx-qmgr.c
61
BUG_ON(queue >= HALF_QUEUES);
drivers/soc/ixp4xx/ixp4xx-qmgr.c
62
return (__raw_readl(&qmgr_regs->stat2[queue >> 4])
drivers/soc/ixp4xx/ixp4xx-qmgr.c
63
>> ((queue & 0xF) << 1)) & 0x3;
drivers/soc/ixp4xx/ixp4xx-qmgr.c
72
int qmgr_stat_empty(unsigned int queue)
drivers/soc/ixp4xx/ixp4xx-qmgr.c
74
BUG_ON(queue >= HALF_QUEUES);
drivers/soc/ixp4xx/ixp4xx-qmgr.c
75
return __qmgr_get_stat1(queue) & QUEUE_STAT1_EMPTY;
drivers/soc/ixp4xx/ixp4xx-qmgr.c
84
int qmgr_stat_below_low_watermark(unsigned int queue)
drivers/soc/ixp4xx/ixp4xx-qmgr.c
86
if (queue >= HALF_QUEUES)
drivers/soc/ixp4xx/ixp4xx-qmgr.c
88
(queue - HALF_QUEUES)) & 0x01;
drivers/soc/ixp4xx/ixp4xx-qmgr.c
89
return __qmgr_get_stat1(queue) & QUEUE_STAT1_NEARLY_EMPTY;
drivers/soc/ixp4xx/ixp4xx-qmgr.c
98
int qmgr_stat_full(unsigned int queue)
drivers/soc/ti/knav_qmss.h
223
struct knav_queue *queue;
drivers/soc/ti/knav_qmss_acc.c
101
kq = knav_range_offset_to_inst(kdev, range, queue);
drivers/soc/ti/knav_qmss_acc.c
102
acc += queue;
drivers/soc/ti/knav_qmss_acc.c
148
queue = list[ACC_LIST_ENTRY_QUEUE_IDX] >> 16;
drivers/soc/ti/knav_qmss_acc.c
149
if (queue < range_base ||
drivers/soc/ti/knav_qmss_acc.c
150
queue >= range_base + range->num_queues) {
drivers/soc/ti/knav_qmss_acc.c
153
queue, range_base,
drivers/soc/ti/knav_qmss_acc.c
157
queue -= range_base;
drivers/soc/ti/knav_qmss_acc.c
159
queue);
drivers/soc/ti/knav_qmss_acc.c
166
queue + range_base);
drivers/soc/ti/knav_qmss_acc.c
174
val, idx, queue + range_base);
drivers/soc/ti/knav_qmss_acc.c
196
int queue, bool enabled)
drivers/soc/ti/knav_qmss_acc.c
209
acc = range->acc + queue;
drivers/soc/ti/knav_qmss_acc.c
210
irq = range->irqs[queue].irq;
drivers/soc/ti/knav_qmss_acc.c
211
cpu_mask = range->irqs[queue].cpu_mask;
drivers/soc/ti/knav_qmss_acc.c
216
new = old | BIT(queue);
drivers/soc/ti/knav_qmss_acc.c
218
new = old & ~BIT(queue);
drivers/soc/ti/knav_qmss_acc.c
28
int range_base, queue;
drivers/soc/ti/knav_qmss_acc.c
303
int queue)
drivers/soc/ti/knav_qmss_acc.c
315
acc = range->acc + queue;
drivers/soc/ti/knav_qmss_acc.c
316
queue_base = range->queue_base + queue;
drivers/soc/ti/knav_qmss_acc.c
33
for (queue = 0; queue < range->num_queues; queue++) {
drivers/soc/ti/knav_qmss_acc.c
336
int queue)
drivers/soc/ti/knav_qmss_acc.c
342
acc = range->acc + queue;
drivers/soc/ti/knav_qmss_acc.c
344
knav_acc_setup_cmd(kdev, range, &cmd, queue);
drivers/soc/ti/knav_qmss_acc.c
35
queue);
drivers/soc/ti/knav_qmss_acc.c
354
int queue)
drivers/soc/ti/knav_qmss_acc.c
360
acc = range->acc + queue;
drivers/soc/ti/knav_qmss_acc.c
362
knav_acc_setup_cmd(kdev, range, &cmd, queue);
drivers/soc/ti/knav_qmss_acc.c
377
int queue;
drivers/soc/ti/knav_qmss_acc.c
379
for (queue = 0; queue < range->num_queues; queue++) {
drivers/soc/ti/knav_qmss_acc.c
380
acc = range->acc + queue;
drivers/soc/ti/knav_qmss_acc.c
382
knav_acc_stop(kdev, range, queue);
drivers/soc/ti/knav_qmss_acc.c
384
result = knav_acc_start(kdev, range, queue);
drivers/soc/ti/knav_qmss_acc.c
39
range_base + queue);
drivers/soc/ti/knav_qmss_acc.c
44
queue = acc->channel - range->acc_info.start_channel;
drivers/soc/ti/knav_qmss_acc.c
45
inst = knav_range_offset_to_inst(kdev, range, queue);
drivers/soc/ti/knav_qmss_acc.c
47
range_base + queue);
drivers/soc/ti/knav_qmss_acc.c
87
int range_base, channel, queue = 0;
drivers/soc/ti/knav_qmss_acc.c
98
for (queue = 0; queue < range->num_irqs; queue++)
drivers/soc/ti/knav_qmss_acc.c
99
if (range->irqs[queue].irq == irq)
drivers/soc/ti/knav_qmss_queue.c
117
unsigned queue = inst->id - range->queue_base;
drivers/soc/ti/knav_qmss_queue.c
121
irq = range->irqs[queue].irq;
drivers/soc/ti/knav_qmss_queue.c
126
if (range->irqs[queue].cpu_mask) {
drivers/soc/ti/knav_qmss_queue.c
127
ret = irq_set_affinity_hint(irq, range->irqs[queue].cpu_mask);
drivers/soc/ti/knav_qmss_queue.c
141
unsigned queue = inst->id - inst->range->queue_base;
drivers/soc/ti/knav_qmss_queue.c
145
irq = range->irqs[queue].irq;
drivers/soc/ti/knav_qmss_queue.c
387
unsigned queue;
drivers/soc/ti/knav_qmss_queue.c
390
queue = inst->id - range->queue_base;
drivers/soc/ti/knav_qmss_queue.c
392
enable_irq(range->irqs[queue].irq);
drivers/soc/ti/knav_qmss_queue.c
394
disable_irq_nosync(range->irqs[queue].irq);
drivers/soc/ti/knav_qmss_queue.c
701
knav_queue_push(pool->queue, dma_addr, dma_size, 0);
drivers/soc/ti/knav_qmss_queue.c
713
if (!pool->queue)
drivers/soc/ti/knav_qmss_queue.c
717
dma = knav_queue_pop(pool->queue, &size);
drivers/soc/ti/knav_qmss_queue.c
727
knav_queue_close(pool->queue);
drivers/soc/ti/knav_qmss_queue.c
790
pool->queue = knav_queue_open(name, KNAV_QUEUE_GP, 0);
drivers/soc/ti/knav_qmss_queue.c
791
if (IS_ERR(pool->queue)) {
drivers/soc/ti/knav_qmss_queue.c
794
name, PTR_ERR(pool->queue));
drivers/soc/ti/knav_qmss_queue.c
795
ret = PTR_ERR(pool->queue);
drivers/soc/ti/knav_qmss_queue.c
895
dma = knav_queue_pop(pool->queue, &size);
drivers/soc/ti/knav_qmss_queue.c
913
knav_queue_push(pool->queue, dma, pool->region->desc_size, 0);
drivers/soc/ti/knav_qmss_queue.c
975
return knav_queue_get_count(pool->queue);
drivers/spi/spi-mpc52xx.c
160
if (list_empty(&ms->queue))
drivers/spi/spi-mpc52xx.c
164
ms->message = list_first_entry(&ms->queue, struct spi_message, queue);
drivers/spi/spi-mpc52xx.c
165
list_del_init(&ms->message->queue);
drivers/spi/spi-mpc52xx.c
377
list_add_tail(&m->queue, &ms->queue);
drivers/spi/spi-mpc52xx.c
468
INIT_LIST_HEAD(&ms->queue);
drivers/spi/spi-mpc52xx.c
80
struct list_head queue; /* queue of pending messages */
drivers/spi/spi-topcliff-pch.c
1117
list_for_each_entry_safe(pmsg, tmp, data->queue.next, queue) {
drivers/spi/spi-topcliff-pch.c
1127
list_del_init(&pmsg->queue);
drivers/spi/spi-topcliff-pch.c
1139
data->current_msg = list_entry(data->queue.next, struct spi_message,
drivers/spi/spi-topcliff-pch.c
1140
queue);
drivers/spi/spi-topcliff-pch.c
1142
list_del_init(&data->current_msg->queue);
drivers/spi/spi-topcliff-pch.c
1348
INIT_LIST_HEAD(&data->queue);
drivers/spi/spi-topcliff-pch.c
1417
while ((list_empty(&data->queue) == 0) && --count) {
drivers/spi/spi-topcliff-pch.c
170
struct list_head queue;
drivers/spi/spi-topcliff-pch.c
483
list_add_tail(&pmsg->queue, &data->queue);
drivers/spi/spi-topcliff-pch.c
563
list_for_each_entry_safe(pmsg, tmp, data->queue.next, queue) {
drivers/spi/spi-topcliff-pch.c
570
list_del_init(&pmsg->queue);
drivers/spi/spi-topcliff-pch.c
634
if ((list_empty(&data->queue) == 0) &&
drivers/spi/spi-topcliff-pch.c
648
list_for_each_entry_safe(pmsg, tmp, data->queue.next, queue) {
drivers/spi/spi-topcliff-pch.c
655
list_del_init(&pmsg->queue);
drivers/spi/spi.c
1853
if (list_empty(&ctlr->queue) || !ctlr->running) {
drivers/spi/spi.c
1892
msg = list_first_entry(&ctlr->queue, struct spi_message, queue);
drivers/spi/spi.c
1895
list_del_init(&msg->queue);
drivers/spi/spi.c
2082
next = list_first_entry_or_null(&ctlr->queue, struct spi_message,
drivers/spi/spi.c
2083
queue);
drivers/spi/spi.c
2214
if (list_empty(&ctlr->queue) && !ctlr->busy) {
drivers/spi/spi.c
2264
list_add_tail(&msg->queue, &ctlr->queue);
drivers/spi/spi.c
3199
INIT_LIST_HEAD(&ctlr->queue);
drivers/staging/media/atomisp/pci/atomisp_cmd.c
2866
list_move_tail(&frame->queue, &pipe->activeq);
drivers/staging/media/atomisp/pci/atomisp_cmd.c
2899
&pipe->buffers_waiting_for_param, queue) {
drivers/staging/media/atomisp/pci/atomisp_cmd.c
586
list_del(&frame->queue);
drivers/staging/media/atomisp/pci/atomisp_cmd.c
600
list_for_each_entry_safe(frame, _frame, &pipe->buffers_in_css, queue) {
drivers/staging/media/atomisp/pci/atomisp_cmd.c
606
list_for_each_entry_safe(frame, _frame, &pipe->activeq, queue)
drivers/staging/media/atomisp/pci/atomisp_cmd.c
609
list_for_each_entry_safe(frame, _frame, &pipe->buffers_waiting_for_param, queue) {
drivers/staging/media/atomisp/pci/atomisp_fops.c
242
frame = list_first_entry_or_null(&pipe->activeq, struct ia_css_frame, queue);
drivers/staging/media/atomisp/pci/atomisp_fops.c
244
list_move_tail(&frame->queue, &pipe->buffers_in_css);
drivers/staging/media/atomisp/pci/atomisp_fops.c
300
list_move_tail(&frame->queue, &pipe->activeq);
drivers/staging/media/atomisp/pci/atomisp_fops.c
393
list_add_tail(&frame->queue, &pipe->buffers_waiting_for_param);
drivers/staging/media/atomisp/pci/atomisp_fops.c
395
list_add_tail(&frame->queue, &pipe->activeq);
drivers/staging/media/atomisp/pci/atomisp_ioctl.c
795
if (buf->index >= vb2_get_num_buffers(vdev->queue))
drivers/staging/media/atomisp/pci/atomisp_subdev.c
787
pipe->vdev.queue = &pipe->vb_queue;
drivers/staging/media/atomisp/pci/atomisp_subdev.h
86
#define vq_to_pipe(queue) \
drivers/staging/media/atomisp/pci/atomisp_subdev.h
87
container_of(queue, struct atomisp_video_pipe, vb_queue)
drivers/staging/media/atomisp/pci/ia_css_frame_public.h
145
struct list_head queue;
drivers/staging/media/av7110/av7110.c
554
wake_up(&cibuf->queue);
drivers/staging/media/av7110/av7110_av.c
230
wake_up(&buf->queue);
drivers/staging/media/av7110/av7110_av.c
236
wake_up(&buf->queue);
drivers/staging/media/av7110/av7110_av.c
256
wake_up(&buf->queue);
drivers/staging/media/av7110/av7110_av.c
264
wake_up(&buf->queue);
drivers/staging/media/av7110/av7110_av.c
394
if (wait_event_interruptible(rbuf->queue,
drivers/staging/media/av7110/av7110_av.c
456
if (wait_event_interruptible(rb->queue, FREE_COND_TS))
drivers/staging/media/av7110/av7110_av.c
489
if (wait_event_interruptible(av7110->avout.queue,
drivers/staging/media/av7110/av7110_av.c
523
if (wait_event_interruptible(av7110->avout.queue,
drivers/staging/media/av7110/av7110_av.c
553
if (wait_event_interruptible(av7110->aout.queue,
drivers/staging/media/av7110/av7110_av.c
934
poll_wait(file, &av7110->avout.queue, wait);
drivers/staging/media/av7110/av7110_av.c
985
poll_wait(file, &av7110->aout.queue, wait);
drivers/staging/media/av7110/av7110_ca.c
161
if (wait_event_interruptible(cibuf->queue,
drivers/staging/media/av7110/av7110_ca.c
186
if (wait_event_interruptible(cibuf->queue,
drivers/staging/media/av7110/av7110_ca.c
225
poll_wait(file, &rbuf->queue, wait);
drivers/staging/media/av7110/av7110_ca.c
226
poll_wait(file, &wbuf->queue, wait);
drivers/staging/media/av7110/av7110_ca.c
70
wake_up_interruptible(&cibuf->queue);
drivers/staging/media/deprecated/atmel/atmel-isc-base.c
1913
vdev->queue = q;
drivers/staging/media/imx/imx-media-capture.c
999
vfd->queue = &priv->q;
drivers/staging/media/imx/imx-media-utils.c
687
if (buftype == vfd->queue->type)
drivers/staging/media/imx/imx-media-utils.c
720
if (buftype == vfd->queue->type)
drivers/staging/media/ipu3/ipu3-abi.h
277
#define IMGU_ABI_EVENT_BUFFER_ENQUEUED(thread, queue) \
drivers/staging/media/ipu3/ipu3-abi.h
278
(0 << 24 | (thread) << 16 | (queue) << 8)
drivers/staging/media/ipu3/ipu3-abi.h
279
#define IMGU_ABI_EVENT_BUFFER_DEQUEUED(queue) (1 << 24 | (queue) << 8)
drivers/staging/media/ipu3/ipu3-css-params.c
1759
css_pipe->queue[IPU3_CSS_QUEUE_IN].fmt.mpix.width;
drivers/staging/media/ipu3/ipu3-css-params.c
1761
css_pipe->queue[IPU3_CSS_QUEUE_IN].fmt.mpix.height;
drivers/staging/media/ipu3/ipu3-css-params.c
1763
css_pipe->queue[IPU3_CSS_QUEUE_IN].css_fmt->bayer_order;
drivers/staging/media/ipu3/ipu3-css-params.c
1802
css_pipe->queue[IPU3_CSS_QUEUE_OUT].fmt.mpix.height;
drivers/staging/media/ipu3/ipu3-css-params.c
1817
css_pipe->queue[IPU3_CSS_QUEUE_OUT].fmt.mpix.width;
drivers/staging/media/ipu3/ipu3-css-params.c
1858
css_pipe->queue[IPU3_CSS_QUEUE_OUT].fmt.mpix.height;
drivers/staging/media/ipu3/ipu3-css-params.c
1872
css_pipe->queue[IPU3_CSS_QUEUE_OUT].fmt.mpix.width;
drivers/staging/media/ipu3/ipu3-css-params.c
1874
css_pipe->queue[IPU3_CSS_QUEUE_OUT].fmt.mpix.height;
drivers/staging/media/ipu3/ipu3-css-params.c
1883
css_pipe->queue[IPU3_CSS_QUEUE_VF].fmt.mpix.width;
drivers/staging/media/ipu3/ipu3-css-params.c
1885
css_pipe->queue[IPU3_CSS_QUEUE_VF].fmt.mpix.height;
drivers/staging/media/ipu3/ipu3-css-params.c
1957
&css_pipe->queue[IPU3_CSS_QUEUE_IN].fmt.mpix;
drivers/staging/media/ipu3/ipu3-css-params.c
1973
ofs_x += css_pipe->queue[IPU3_CSS_QUEUE_IN].css_fmt->bayer_order ==
drivers/staging/media/ipu3/ipu3-css-params.c
1975
css_pipe->queue[IPU3_CSS_QUEUE_IN].css_fmt->bayer_order ==
drivers/staging/media/ipu3/ipu3-css-params.c
1979
ofs_y += css_pipe->queue[IPU3_CSS_QUEUE_IN].css_fmt->bayer_order ==
drivers/staging/media/ipu3/ipu3-css-params.c
1981
css_pipe->queue[IPU3_CSS_QUEUE_IN].css_fmt->bayer_order ==
drivers/staging/media/ipu3/ipu3-css-params.c
377
target_width = css_pipe->queue[IPU3_CSS_QUEUE_VF].fmt.mpix.width;
drivers/staging/media/ipu3/ipu3-css-params.c
378
target_height = css_pipe->queue[IPU3_CSS_QUEUE_VF].fmt.mpix.height;
drivers/staging/media/ipu3/ipu3-css-params.c
389
css_pipe->queue[IPU3_CSS_QUEUE_OUT].css_fmt->frame_format;
drivers/staging/media/ipu3/ipu3-css-params.c
392
css_pipe->queue[IPU3_CSS_QUEUE_OUT].fmt.mpix.width;
drivers/staging/media/ipu3/ipu3-css-params.c
394
css_pipe->queue[IPU3_CSS_QUEUE_OUT].fmt.mpix.height;
drivers/staging/media/ipu3/ipu3-css-params.c
396
css_pipe->queue[IPU3_CSS_QUEUE_OUT].width_pad;
drivers/staging/media/ipu3/ipu3-css-params.c
398
css_pipe->queue[IPU3_CSS_QUEUE_OUT].css_fmt->frame_format;
drivers/staging/media/ipu3/ipu3-css-params.c
401
css_pipe->queue[IPU3_CSS_QUEUE_VF].fmt.mpix.width;
drivers/staging/media/ipu3/ipu3-css-params.c
403
css_pipe->queue[IPU3_CSS_QUEUE_VF].fmt.mpix.height;
drivers/staging/media/ipu3/ipu3-css-params.c
405
css_pipe->queue[IPU3_CSS_QUEUE_VF].width_pad;
drivers/staging/media/ipu3/ipu3-css-params.c
407
css_pipe->queue[IPU3_CSS_QUEUE_VF].css_fmt->frame_format;
drivers/staging/media/ipu3/ipu3-css-params.c
412
css_pipe->queue[IPU3_CSS_QUEUE_OUT].fmt.mpix.width;
drivers/staging/media/ipu3/ipu3-css-params.c
414
css_pipe->queue[IPU3_CSS_QUEUE_OUT].fmt.mpix.height;
drivers/staging/media/ipu3/ipu3-css-params.c
416
css_pipe->queue[IPU3_CSS_QUEUE_VF].fmt.mpix.width;
drivers/staging/media/ipu3/ipu3-css-params.c
418
css_pipe->queue[IPU3_CSS_QUEUE_VF].fmt.mpix.height;
drivers/staging/media/ipu3/ipu3-css.c
1058
static u8 imgu_css_queue_pos(struct imgu_css *css, int queue, int thread)
drivers/staging/media/ipu3/ipu3-css.c
1066
return queue >= 0 ? readb(&q->host2sp_bufq_info[thread][queue].end) :
drivers/staging/media/ipu3/ipu3-css.c
1072
int queue, int thread, u32 data)
drivers/staging/media/ipu3/ipu3-css.c
1081
if (queue >= 0) {
drivers/staging/media/ipu3/ipu3-css.c
1082
size = readb(&q->host2sp_bufq_info[thread][queue].size);
drivers/staging/media/ipu3/ipu3-css.c
1083
start = readb(&q->host2sp_bufq_info[thread][queue].start);
drivers/staging/media/ipu3/ipu3-css.c
1084
end = readb(&q->host2sp_bufq_info[thread][queue].end);
drivers/staging/media/ipu3/ipu3-css.c
1098
if (queue >= 0) {
drivers/staging/media/ipu3/ipu3-css.c
1099
writel(data, &q->host2sp_bufq[thread][queue][end]);
drivers/staging/media/ipu3/ipu3-css.c
1100
writeb(end2, &q->host2sp_bufq_info[thread][queue].end);
drivers/staging/media/ipu3/ipu3-css.c
1110
static int imgu_css_dequeue_data(struct imgu_css *css, int queue, u32 *data)
drivers/staging/media/ipu3/ipu3-css.c
1119
if (queue >= 0) {
drivers/staging/media/ipu3/ipu3-css.c
1120
size = readb(&q->sp2host_bufq_info[queue].size);
drivers/staging/media/ipu3/ipu3-css.c
1121
start = readb(&q->sp2host_bufq_info[queue].start);
drivers/staging/media/ipu3/ipu3-css.c
1122
end = readb(&q->sp2host_bufq_info[queue].end);
drivers/staging/media/ipu3/ipu3-css.c
1137
if (queue >= 0) {
drivers/staging/media/ipu3/ipu3-css.c
1138
*data = readl(&q->sp2host_bufq[queue][start]);
drivers/staging/media/ipu3/ipu3-css.c
1139
writeb(start2, &q->sp2host_bufq_info[queue].start);
drivers/staging/media/ipu3/ipu3-css.c
1147
r = imgu_css_queue_data(css, queue, 0,
drivers/staging/media/ipu3/ipu3-css.c
120
static int imgu_css_queue_init(struct imgu_css_queue *queue,
drivers/staging/media/ipu3/ipu3-css.c
123
struct v4l2_pix_format_mplane *const f = &queue->fmt.mpix;
drivers/staging/media/ipu3/ipu3-css.c
127
INIT_LIST_HEAD(&queue->bufs);
drivers/staging/media/ipu3/ipu3-css.c
129
queue->css_fmt = NULL; /* Disable */
drivers/staging/media/ipu3/ipu3-css.c
136
queue->css_fmt = &imgu_css_formats[i];
drivers/staging/media/ipu3/ipu3-css.c
1382
&css_pipe->queue[q].bufs,
drivers/staging/media/ipu3/ipu3-css.c
140
if (!queue->css_fmt)
drivers/staging/media/ipu3/ipu3-css.c
1400
if (!list_empty(&css_pipe->queue[q].bufs))
drivers/staging/media/ipu3/ipu3-css.c
143
queue->fmt.mpix = *fmt;
drivers/staging/media/ipu3/ipu3-css.c
149
queue->width_pad = ALIGN(f->width, queue->css_fmt->width_align);
drivers/staging/media/ipu3/ipu3-css.c
151
imgu_bytesperline(f->width, queue->css_fmt->frame_format);
drivers/staging/media/ipu3/ipu3-css.c
1520
r = imgu_css_queue_init(&css_pipe->queue[q], NULL, 0);
drivers/staging/media/ipu3/ipu3-css.c
153
if (queue->css_fmt->chroma_decim)
drivers/staging/media/ipu3/ipu3-css.c
154
sizeimage += 2 * sizeimage / queue->css_fmt->chroma_decim;
drivers/staging/media/ipu3/ipu3-css.c
1551
struct imgu_css_queue queue[IPU3_CSS_QUEUES],
drivers/staging/media/ipu3/ipu3-css.c
1559
&queue[IPU3_CSS_QUEUE_IN].fmt.mpix;
drivers/staging/media/ipu3/ipu3-css.c
1561
&queue[IPU3_CSS_QUEUE_OUT].fmt.mpix;
drivers/staging/media/ipu3/ipu3-css.c
1563
&queue[IPU3_CSS_QUEUE_VF].fmt.mpix;
drivers/staging/media/ipu3/ipu3-css.c
1568
if (!imgu_css_queue_enabled(&queue[IPU3_CSS_QUEUE_IN]))
drivers/staging/media/ipu3/ipu3-css.c
159
f->colorspace = queue->css_fmt->colorspace;
drivers/staging/media/ipu3/ipu3-css.c
1616
if (imgu_css_queue_enabled(&queue[IPU3_CSS_QUEUE_OUT])) {
drivers/staging/media/ipu3/ipu3-css.c
1620
q_fmt = queue[IPU3_CSS_QUEUE_OUT].css_fmt->frame_format;
drivers/staging/media/ipu3/ipu3-css.c
1638
if (imgu_css_queue_enabled(&queue[IPU3_CSS_QUEUE_VF])) {
drivers/staging/media/ipu3/ipu3-css.c
1642
q_fmt = queue[IPU3_CSS_QUEUE_VF].css_fmt->frame_format;
drivers/staging/media/ipu3/ipu3-css.c
1851
if (imgu_css_queue_init(&css_pipe->queue[i], fmts[i],
drivers/staging/media/ipu3/ipu3-css.c
1904
if (b->queue >= IPU3_CSS_QUEUES || !imgu_css_queues[b->queue].qid)
drivers/staging/media/ipu3/ipu3-css.c
1907
b->queue_pos = imgu_css_queue_pos(css, imgu_css_queues[b->queue].qid,
drivers/staging/media/ipu3/ipu3-css.c
1910
if (b->queue_pos >= ARRAY_SIZE(css->pipes[pipe].abi_buffers[b->queue]))
drivers/staging/media/ipu3/ipu3-css.c
1912
abi_buf = css->pipes[pipe].abi_buffers[b->queue][b->queue_pos].vaddr;
drivers/staging/media/ipu3/ipu3-css.c
1917
buf_addr = (void *)abi_buf + imgu_css_queues[b->queue].ptr_ofs;
drivers/staging/media/ipu3/ipu3-css.c
1920
if (b->queue == IPU3_CSS_QUEUE_STAT_3A)
drivers/staging/media/ipu3/ipu3-css.c
1923
if (b->queue == IPU3_CSS_QUEUE_OUT)
drivers/staging/media/ipu3/ipu3-css.c
1925
css_pipe->queue[IPU3_CSS_QUEUE_OUT].width_pad;
drivers/staging/media/ipu3/ipu3-css.c
1927
if (b->queue == IPU3_CSS_QUEUE_VF)
drivers/staging/media/ipu3/ipu3-css.c
1929
css_pipe->queue[IPU3_CSS_QUEUE_VF].width_pad;
drivers/staging/media/ipu3/ipu3-css.c
1932
list_add_tail(&b->list, &css_pipe->queue[b->queue].bufs);
drivers/staging/media/ipu3/ipu3-css.c
1936
data = css->pipes[pipe].abi_buffers[b->queue][b->queue_pos].daddr;
drivers/staging/media/ipu3/ipu3-css.c
1937
r = imgu_css_queue_data(css, imgu_css_queues[b->queue].qid,
drivers/staging/media/ipu3/ipu3-css.c
1943
imgu_css_queues[b->queue].qid);
drivers/staging/media/ipu3/ipu3-css.c
1949
b, b->queue, pipe);
drivers/staging/media/ipu3/ipu3-css.c
1976
int evtype, pipe, pipeid, queue, qid, r;
drivers/staging/media/ipu3/ipu3-css.c
1998
queue = evtype_to_queue[evtype];
drivers/staging/media/ipu3/ipu3-css.c
1999
qid = imgu_css_queues[queue].qid;
drivers/staging/media/ipu3/ipu3-css.c
2013
event, queue, pipe, pipeid);
drivers/staging/media/ipu3/ipu3-css.c
2030
if (list_empty(&css_pipe->queue[queue].bufs)) {
drivers/staging/media/ipu3/ipu3-css.c
2035
b = list_first_entry(&css_pipe->queue[queue].bufs,
drivers/staging/media/ipu3/ipu3-css.c
2037
if (queue != b->queue ||
drivers/staging/media/ipu3/ipu3-css.c
2039
[b->queue][b->queue_pos].daddr) {
drivers/staging/media/ipu3/ipu3-css.c
689
&css_pipe->queue[IPU3_CSS_QUEUE_IN];
drivers/staging/media/ipu3/ipu3-css.c
691
&css_pipe->queue[IPU3_CSS_QUEUE_OUT];
drivers/staging/media/ipu3/ipu3-css.c
693
&css_pipe->queue[IPU3_CSS_QUEUE_VF];
drivers/staging/media/ipu3/ipu3-css.h
106
struct imgu_css_queue queue[IPU3_CSS_QUEUES];
drivers/staging/media/ipu3/ipu3-css.h
207
unsigned int queue, dma_addr_t daddr)
drivers/staging/media/ipu3/ipu3-css.h
210
b->queue = queue;
drivers/staging/media/ipu3/ipu3-css.h
71
unsigned int queue;
drivers/staging/media/ipu3/ipu3-v4l2.c
1239
vdev->queue = &node->vbq;
drivers/staging/media/ipu3/ipu3-v4l2.c
321
unsigned int queue = imgu_node_to_queue(node->id);
drivers/staging/media/ipu3/ipu3-v4l2.c
323
if (queue == IPU3_CSS_QUEUE_PARAMS)
drivers/staging/media/ipu3/ipu3-v4l2.c
337
unsigned int queue = imgu_node_to_queue(node->id);
drivers/staging/media/ipu3/ipu3-v4l2.c
339
if (queue == IPU3_CSS_QUEUE_PARAMS)
drivers/staging/media/ipu3/ipu3-v4l2.c
351
unsigned int queue = imgu_node_to_queue(node->id);
drivers/staging/media/ipu3/ipu3-v4l2.c
363
if (queue == IPU3_CSS_QUEUE_PARAMS && payload && payload < need_bytes) {
drivers/staging/media/ipu3/ipu3-v4l2.c
370
if (queue != IPU3_CSS_QUEUE_PARAMS)
drivers/staging/media/ipu3/ipu3-v4l2.c
371
imgu_css_buf_init(&buf->css_buf, queue, buf->map.daddr);
drivers/staging/media/ipu3/ipu3.c
154
int queue, unsigned int pipe)
drivers/staging/media/ipu3/ipu3.c
160
if (queue == IPU3_CSS_QUEUE_IN)
drivers/staging/media/ipu3/ipu3.c
163
if (WARN_ON(!imgu_pipe->queues[queue].dmap.vaddr))
drivers/staging/media/ipu3/ipu3.c
168
if (imgu_css_buf_state(&imgu_pipe->queues[queue].dummybufs[i]) !=
drivers/staging/media/ipu3/ipu3.c
175
imgu_css_buf_init(&imgu_pipe->queues[queue].dummybufs[i], queue,
drivers/staging/media/ipu3/ipu3.c
176
imgu_pipe->queues[queue].dmap.daddr);
drivers/staging/media/ipu3/ipu3.c
178
return &imgu_pipe->queues[queue].dummybufs[i];
drivers/staging/media/ipu3/ipu3.c
190
if (buf == &imgu_pipe->queues[buf->queue].dummybufs[i])
drivers/staging/media/ipu3/ipu3.c
537
node = imgu_map_node(imgu, b->queue);
drivers/staging/media/ipu7/ipu7-isys-video.c
1046
av->vdev.queue = &av->aq.vbq;
drivers/staging/media/starfive/camss/stf-capture.c
103
queue);
drivers/staging/media/starfive/camss/stf-capture.c
104
list_del(&buffer->queue);
drivers/staging/media/starfive/camss/stf-capture.c
271
INIT_LIST_HEAD(&buffer->queue);
drivers/staging/media/starfive/camss/stf-capture.c
272
list_add_tail(&buffer->queue, &output->ready_bufs);
drivers/staging/media/starfive/camss/stf-capture.c
282
queue);
drivers/staging/media/starfive/camss/stf-capture.c
283
list_del(&buffer->queue);
drivers/staging/media/starfive/camss/stf-capture.c
292
INIT_LIST_HEAD(&buffer->queue);
drivers/staging/media/starfive/camss/stf-capture.c
293
list_add_tail(&buffer->queue, &output->pending_bufs);
drivers/staging/media/starfive/camss/stf-capture.c
363
list_for_each_entry_safe(buf, t, &output->pending_bufs, queue) {
drivers/staging/media/starfive/camss/stf-capture.c
365
list_del(&buf->queue);
drivers/staging/media/starfive/camss/stf-capture.c
367
list_for_each_entry_safe(buf, t, &output->ready_bufs, queue) {
drivers/staging/media/starfive/camss/stf-capture.c
369
list_del(&buf->queue);
drivers/staging/media/starfive/camss/stf-video.c
544
vdev->queue = &video->vb2_q;
drivers/staging/media/starfive/camss/stf-video.h
46
struct list_head queue;
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_capture.c
175
list_for_each_entry(isp_buffer, &state->queue, list) {
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_capture.c
181
INIT_LIST_HEAD(&state->queue);
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_capture.c
195
if (list_empty(&state->queue))
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_capture.c
201
isp_buffer = list_first_entry(&state->queue, struct sun6i_isp_buffer,
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_capture.c
260
static int sun6i_isp_capture_queue_setup(struct vb2_queue *queue,
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_capture.c
266
struct sun6i_isp_device *isp_dev = vb2_get_drv_priv(queue);
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_capture.c
307
list_add_tail(&isp_buffer->list, &state->queue);
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_capture.c
315
static int sun6i_isp_capture_start_streaming(struct vb2_queue *queue,
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_capture.c
318
struct sun6i_isp_device *isp_dev = vb2_get_drv_priv(queue);
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_capture.c
349
static void sun6i_isp_capture_stop_streaming(struct vb2_queue *queue)
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_capture.c
351
struct sun6i_isp_device *isp_dev = vb2_get_drv_priv(queue);
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_capture.c
464
if (vb2_is_busy(&isp_dev->capture.queue))
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_capture.c
629
struct vb2_queue *queue = &capture->queue;
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_capture.c
637
INIT_LIST_HEAD(&state->queue);
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_capture.c
656
queue->type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_capture.c
657
queue->io_modes = VB2_MMAP | VB2_DMABUF;
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_capture.c
658
queue->buf_struct_size = sizeof(struct sun6i_isp_buffer);
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_capture.c
659
queue->ops = &sun6i_isp_capture_queue_ops;
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_capture.c
660
queue->mem_ops = &vb2_dma_contig_memops;
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_capture.c
661
queue->min_queued_buffers = 2;
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_capture.c
662
queue->timestamp_flags = V4L2_BUF_FLAG_TIMESTAMP_MONOTONIC;
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_capture.c
663
queue->lock = &capture->lock;
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_capture.c
664
queue->dev = isp_dev->dev;
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_capture.c
665
queue->drv_priv = isp_dev;
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_capture.c
667
ret = vb2_queue_init(queue);
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_capture.c
675
format->type = queue->type;
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_capture.c
692
video_dev->queue = queue;
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_capture.h
28
struct list_head queue;
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_capture.h
43
struct vb2_queue queue;
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_params.c
215
list_for_each_entry(isp_buffer, &state->queue, list) {
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_params.c
221
INIT_LIST_HEAD(&state->queue);
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_params.c
237
if (list_empty(&state->queue))
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_params.c
243
isp_buffer = list_first_entry(&state->queue, struct sun6i_isp_buffer,
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_params.c
292
static int sun6i_isp_params_queue_setup(struct vb2_queue *queue,
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_params.c
298
struct sun6i_isp_device *isp_dev = vb2_get_drv_priv(queue);
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_params.c
340
list_add_tail(&isp_buffer->list, &state->queue);
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_params.c
347
static int sun6i_isp_params_start_streaming(struct vb2_queue *queue,
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_params.c
350
struct sun6i_isp_device *isp_dev = vb2_get_drv_priv(queue);
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_params.c
367
static void sun6i_isp_params_stop_streaming(struct vb2_queue *queue)
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_params.c
369
struct sun6i_isp_device *isp_dev = vb2_get_drv_priv(queue);
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_params.c
462
struct vb2_queue *queue = &isp_dev->params.queue;
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_params.c
470
INIT_LIST_HEAD(&state->queue);
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_params.c
485
queue->type = V4L2_BUF_TYPE_META_OUTPUT;
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_params.c
486
queue->io_modes = VB2_MMAP | VB2_USERPTR | VB2_DMABUF;
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_params.c
487
queue->buf_struct_size = sizeof(struct sun6i_isp_buffer);
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_params.c
488
queue->ops = &sun6i_isp_params_queue_ops;
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_params.c
489
queue->mem_ops = &vb2_vmalloc_memops;
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_params.c
490
queue->min_queued_buffers = 1;
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_params.c
491
queue->timestamp_flags = V4L2_BUF_FLAG_TIMESTAMP_MONOTONIC;
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_params.c
492
queue->lock = ¶ms->lock;
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_params.c
493
queue->dev = isp_dev->dev;
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_params.c
494
queue->drv_priv = isp_dev;
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_params.c
496
ret = vb2_queue_init(queue);
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_params.c
504
format->type = queue->type;
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_params.c
518
video_dev->queue = queue;
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_params.h
17
struct list_head queue; /* Queue and buffers lock. */
drivers/staging/media/sunxi/sun6i-isp/sun6i_isp_params.h
30
struct vb2_queue queue;
drivers/staging/media/tegra-video/tegra20.c
432
buf = list_first_entry(&chan->capture, struct tegra_channel_buffer, queue);
drivers/staging/media/tegra-video/tegra20.c
433
list_del_init(&buf->queue);
drivers/staging/media/tegra-video/tegra20.c
443
vb2_queue_error(&chan->queue);
drivers/staging/media/tegra-video/tegra210.c
364
struct tegra_channel_buffer, queue);
drivers/staging/media/tegra-video/tegra210.c
366
list_del_init(&buf->queue);
drivers/staging/media/tegra-video/tegra210.c
498
list_add_tail(&buf->queue, &chan->done);
drivers/staging/media/tegra-video/tegra210.c
567
struct tegra_channel_buffer, queue);
drivers/staging/media/tegra-video/tegra210.c
568
list_del_init(&buf->queue);
drivers/staging/media/tegra-video/tegra210.c
578
vb2_queue_error(&chan->queue);
drivers/staging/media/tegra-video/vi.c
1157
chan->video.queue = &chan->queue;
drivers/staging/media/tegra-video/vi.c
1170
chan->queue.type = V4L2_BUF_TYPE_VIDEO_CAPTURE;
drivers/staging/media/tegra-video/vi.c
1171
chan->queue.io_modes = VB2_MMAP | VB2_DMABUF | VB2_READ;
drivers/staging/media/tegra-video/vi.c
1172
chan->queue.lock = &chan->video_lock;
drivers/staging/media/tegra-video/vi.c
1173
chan->queue.drv_priv = chan;
drivers/staging/media/tegra-video/vi.c
1174
chan->queue.buf_struct_size = sizeof(struct tegra_channel_buffer);
drivers/staging/media/tegra-video/vi.c
1175
chan->queue.ops = &tegra_channel_queue_qops;
drivers/staging/media/tegra-video/vi.c
1176
chan->queue.mem_ops = &vb2_dma_contig_memops;
drivers/staging/media/tegra-video/vi.c
1177
chan->queue.timestamp_flags = V4L2_BUF_FLAG_TIMESTAMP_MONOTONIC;
drivers/staging/media/tegra-video/vi.c
1178
chan->queue.min_queued_buffers = 2;
drivers/staging/media/tegra-video/vi.c
1179
chan->queue.dev = vi->dev;
drivers/staging/media/tegra-video/vi.c
1180
ret = vb2_queue_init(&chan->queue);
drivers/staging/media/tegra-video/vi.c
156
list_add_tail(&buf->queue, &chan->capture);
drivers/staging/media/tegra-video/vi.c
246
list_for_each_entry_safe(buf, nbuf, &chan->capture, queue) {
drivers/staging/media/tegra-video/vi.c
248
list_del(&buf->queue);
drivers/staging/media/tegra-video/vi.c
253
list_for_each_entry_safe(buf, nbuf, &chan->done, queue) {
drivers/staging/media/tegra-video/vi.c
255
list_del(&buf->queue);
drivers/staging/media/tegra-video/vi.c
539
if (vb2_is_busy(&chan->queue))
drivers/staging/media/tegra-video/vi.c
673
if (vb2_is_busy(&chan->queue))
drivers/staging/media/tegra-video/vi.c
750
if (vb2_is_busy(&chan->queue))
drivers/staging/media/tegra-video/vi.h
202
struct vb2_queue queue;
drivers/staging/media/tegra-video/vi.h
247
struct list_head queue;
drivers/staging/media/tegra-video/video.c
40
if (ev->type == V4L2_EVENT_SOURCE_CHANGE && vb2_is_streaming(&chan->queue))
drivers/staging/media/tegra-video/video.c
41
vb2_queue_error(&chan->queue);
drivers/staging/octeon/ethernet-mdio.c
72
priv->port, priv->queue);
drivers/staging/octeon/ethernet-tx.c
425
cvmx_pko_send_packet_prepare(priv->port, priv->queue + qos,
drivers/staging/octeon/ethernet-tx.c
430
priv->queue + qos,
drivers/staging/octeon/ethernet.c
422
priv->queue = -1;
drivers/staging/octeon/ethernet.c
424
if (priv->queue != -1)
drivers/staging/octeon/ethernet.c
778
priv->queue = -1;
drivers/staging/octeon/ethernet.c
831
priv->queue = cvmx_pko_get_base_queue(priv->port);
drivers/staging/octeon/octeon-ethernet.h
53
int queue;
drivers/staging/octeon/octeon-stubs.h
1385
static inline void cvmx_pko_send_packet_prepare(uint64_t port, uint64_t queue,
drivers/staging/octeon/octeon-stubs.h
1390
uint64_t queue, union cvmx_pko_command_word0 pko_command,
drivers/staging/rtl8723bs/core/rtw_ap.c
20
INIT_LIST_HEAD(&pacl_list->acl_node_q.queue);
drivers/staging/rtl8723bs/core/rtw_ap.c
2055
INIT_LIST_HEAD(&pacl_list->acl_node_q.queue);
drivers/staging/rtl8723bs/core/rtw_cmd.c
167
INIT_LIST_HEAD(&pcmdpriv->cmd_queue.queue);
drivers/staging/rtl8723bs/core/rtw_cmd.c
1860
list_add_tail(&pwlan->list, &pmlmepriv->scanned_queue.queue);
drivers/staging/rtl8723bs/core/rtw_cmd.c
249
int _rtw_enqueue_cmd(struct __queue *queue, struct cmd_obj *obj)
drivers/staging/rtl8723bs/core/rtw_cmd.c
257
spin_lock_irqsave(&queue->lock, irqL);
drivers/staging/rtl8723bs/core/rtw_cmd.c
259
list_add_tail(&obj->list, &queue->queue);
drivers/staging/rtl8723bs/core/rtw_cmd.c
262
spin_unlock_irqrestore(&queue->lock, irqL);
drivers/staging/rtl8723bs/core/rtw_cmd.c
268
struct cmd_obj *_rtw_dequeue_cmd(struct __queue *queue)
drivers/staging/rtl8723bs/core/rtw_cmd.c
274
spin_lock_irqsave(&queue->lock, irqL);
drivers/staging/rtl8723bs/core/rtw_cmd.c
275
if (list_empty(&queue->queue)) {
drivers/staging/rtl8723bs/core/rtw_cmd.c
278
obj = container_of(get_next(&queue->queue), struct cmd_obj, list);
drivers/staging/rtl8723bs/core/rtw_cmd.c
283
spin_unlock_irqrestore(&queue->lock, irqL);
drivers/staging/rtl8723bs/core/rtw_cmd.c
416
if (list_empty(&pcmdpriv->cmd_queue.queue))
drivers/staging/rtl8723bs/core/rtw_ioctl_set.c
42
struct __queue *queue = &(pmlmepriv->scanned_queue);
drivers/staging/rtl8723bs/core/rtw_ioctl_set.c
46
phead = get_list_head(queue);
drivers/staging/rtl8723bs/core/rtw_ioctl_set.c
57
if (list_empty(&queue->queue)) {
drivers/staging/rtl8723bs/core/rtw_mlme.c
148
if (list_empty(&free_queue->queue)) {
drivers/staging/rtl8723bs/core/rtw_mlme.c
152
plist = get_next(&free_queue->queue);
drivers/staging/rtl8723bs/core/rtw_mlme.c
1733
struct __queue *queue = &mlme->scanned_queue;
drivers/staging/rtl8723bs/core/rtw_mlme.c
1743
phead = get_list_head(queue);
drivers/staging/rtl8723bs/core/rtw_mlme.c
1825
struct __queue *queue = &pmlmepriv->scanned_queue;
drivers/staging/rtl8723bs/core/rtw_mlme.c
1839
phead = get_list_head(queue);
drivers/staging/rtl8723bs/core/rtw_mlme.c
196
list_add_tail(&pnetwork->list, &free_queue->queue);
drivers/staging/rtl8723bs/core/rtw_mlme.c
485
struct __queue *queue = &pmlmepriv->scanned_queue;
drivers/staging/rtl8723bs/core/rtw_mlme.c
491
spin_lock_bh(&queue->lock);
drivers/staging/rtl8723bs/core/rtw_mlme.c
492
phead = get_list_head(queue);
drivers/staging/rtl8723bs/core/rtw_mlme.c
516
if (list_empty(&pmlmepriv->free_bss_pool.queue)) {
drivers/staging/rtl8723bs/core/rtw_mlme.c
553
list_add_tail(&pnetwork->list, &queue->queue);
drivers/staging/rtl8723bs/core/rtw_mlme.c
58
INIT_LIST_HEAD(&pmlmepriv->free_bss_pool.queue);
drivers/staging/rtl8723bs/core/rtw_mlme.c
580
spin_unlock_bh(&queue->lock);
drivers/staging/rtl8723bs/core/rtw_mlme.c
60
INIT_LIST_HEAD(&pmlmepriv->scanned_queue.queue);
drivers/staging/rtl8723bs/core/rtw_mlme.c
78
list_add_tail(&pnetwork->list, &pmlmepriv->free_bss_pool.queue);
drivers/staging/rtl8723bs/core/rtw_mlme.c
814
list_add_tail(plist, &free_queue->queue);
drivers/staging/rtl8723bs/core/rtw_mlme_ext.c
3569
struct __queue *queue = &(pmlmepriv->scanned_queue);
drivers/staging/rtl8723bs/core/rtw_mlme_ext.c
3631
phead = get_list_head(queue);
drivers/staging/rtl8723bs/core/rtw_recv.c
1196
if (!list_empty(&pdefrag_q->queue))
drivers/staging/rtl8723bs/core/rtw_recv.c
124
if (list_empty(&pfree_recv_queue->queue))
drivers/staging/rtl8723bs/core/rtw_recv.c
186
signed int _rtw_enqueue_recvframe(union recv_frame *precvframe, struct __queue *queue)
drivers/staging/rtl8723bs/core/rtw_recv.c
196
list_add_tail(&(precvframe->u.hdr.list), get_list_head(queue));
drivers/staging/rtl8723bs/core/rtw_recv.c
199
if (queue == &precvpriv->free_recv_queue)
drivers/staging/rtl8723bs/core/rtw_recv.c
205
signed int rtw_enqueue_recvframe(union recv_frame *precvframe, struct __queue *queue)
drivers/staging/rtl8723bs/core/rtw_recv.c
210
spin_lock_bh(&queue->lock);
drivers/staging/rtl8723bs/core/rtw_recv.c
211
ret = _rtw_enqueue_recvframe(precvframe, queue);
drivers/staging/rtl8723bs/core/rtw_recv.c
213
spin_unlock_bh(&queue->lock);
drivers/staging/rtl8723bs/core/rtw_recv.c
261
signed int rtw_enqueue_recvbuf_to_head(struct recv_buf *precvbuf, struct __queue *queue)
drivers/staging/rtl8723bs/core/rtw_recv.c
263
spin_lock_bh(&queue->lock);
drivers/staging/rtl8723bs/core/rtw_recv.c
266
list_add(&precvbuf->list, get_list_head(queue));
drivers/staging/rtl8723bs/core/rtw_recv.c
268
spin_unlock_bh(&queue->lock);
drivers/staging/rtl8723bs/core/rtw_recv.c
27
INIT_LIST_HEAD(&psta_recvpriv->defrag_q.queue);
drivers/staging/rtl8723bs/core/rtw_recv.c
273
signed int rtw_enqueue_recvbuf(struct recv_buf *precvbuf, struct __queue *queue)
drivers/staging/rtl8723bs/core/rtw_recv.c
275
spin_lock_bh(&queue->lock);
drivers/staging/rtl8723bs/core/rtw_recv.c
279
list_add_tail(&precvbuf->list, get_list_head(queue));
drivers/staging/rtl8723bs/core/rtw_recv.c
280
spin_unlock_bh(&queue->lock);
drivers/staging/rtl8723bs/core/rtw_recv.c
285
struct recv_buf *rtw_dequeue_recvbuf(struct __queue *queue)
drivers/staging/rtl8723bs/core/rtw_recv.c
290
spin_lock_bh(&queue->lock);
drivers/staging/rtl8723bs/core/rtw_recv.c
292
if (list_empty(&queue->queue))
drivers/staging/rtl8723bs/core/rtw_recv.c
295
phead = get_list_head(queue);
drivers/staging/rtl8723bs/core/rtw_recv.c
305
spin_unlock_bh(&queue->lock);
drivers/staging/rtl8723bs/core/rtw_recv.c
39
INIT_LIST_HEAD(&precvpriv->free_recv_queue.queue);
drivers/staging/rtl8723bs/core/rtw_recv.c
41
INIT_LIST_HEAD(&precvpriv->recv_pending_queue.queue);
drivers/staging/rtl8723bs/core/rtw_recv.c
43
INIT_LIST_HEAD(&precvpriv->uc_swdec_pending_queue.queue);
drivers/staging/rtl8723bs/core/rtw_recv.c
67
list_add_tail(&(precvframe->u.list), &(precvpriv->free_recv_queue.queue));
drivers/staging/rtl8723bs/core/rtw_sta_mgt.c
190
if (list_empty(&pfree_sta_queue->queue)) {
drivers/staging/rtl8723bs/core/rtw_sta_mgt.c
195
psta = container_of(get_next(&pfree_sta_queue->queue), struct sta_info, list);
drivers/staging/rtl8723bs/core/rtw_sta_mgt.c
21
INIT_LIST_HEAD(&psta->sleep_q.queue);
drivers/staging/rtl8723bs/core/rtw_sta_mgt.c
247
INIT_LIST_HEAD(&preorder_ctrl->pending_recvframe_queue.queue);
drivers/staging/rtl8723bs/core/rtw_sta_mgt.c
65
INIT_LIST_HEAD(&pstapriv->free_sta_queue.queue);
drivers/staging/rtl8723bs/core/rtw_sta_mgt.c
72
INIT_LIST_HEAD(&pstapriv->sleep_q.queue);
drivers/staging/rtl8723bs/core/rtw_sta_mgt.c
74
INIT_LIST_HEAD(&pstapriv->wakeup_q.queue);
drivers/staging/rtl8723bs/core/rtw_xmit.c
1017
return ((!list_empty(&pxmitpriv->be_pending.queue)) ||
drivers/staging/rtl8723bs/core/rtw_xmit.c
1018
(!list_empty(&pxmitpriv->bk_pending.queue)) ||
drivers/staging/rtl8723bs/core/rtw_xmit.c
1019
(!list_empty(&pxmitpriv->vi_pending.queue)) ||
drivers/staging/rtl8723bs/core/rtw_xmit.c
1020
(!list_empty(&pxmitpriv->vo_pending.queue)));
drivers/staging/rtl8723bs/core/rtw_xmit.c
105
INIT_LIST_HEAD(&pxmitpriv->free_xmitbuf_queue.queue);
drivers/staging/rtl8723bs/core/rtw_xmit.c
107
INIT_LIST_HEAD(&pxmitpriv->pending_xmitbuf_queue.queue);
drivers/staging/rtl8723bs/core/rtw_xmit.c
145
&pxmitpriv->free_xmitbuf_queue.queue);
drivers/staging/rtl8723bs/core/rtw_xmit.c
1491
if (list_empty(&pfree_queue->queue)) {
drivers/staging/rtl8723bs/core/rtw_xmit.c
15
INIT_LIST_HEAD(&ptxservq->sta_pending.queue);
drivers/staging/rtl8723bs/core/rtw_xmit.c
1551
if (list_empty(&pfree_xmitbuf_queue->queue)) {
drivers/staging/rtl8723bs/core/rtw_xmit.c
156
INIT_LIST_HEAD(&pxmitpriv->free_xframe_ext_queue.queue);
drivers/staging/rtl8723bs/core/rtw_xmit.c
1651
if (list_empty(&pfree_xmit_queue->queue)) {
drivers/staging/rtl8723bs/core/rtw_xmit.c
1674
struct __queue *queue = &pxmitpriv->free_xframe_ext_queue;
drivers/staging/rtl8723bs/core/rtw_xmit.c
1676
spin_lock_bh(&queue->lock);
drivers/staging/rtl8723bs/core/rtw_xmit.c
1678
if (list_empty(&queue->queue)) {
drivers/staging/rtl8723bs/core/rtw_xmit.c
1681
phead = get_list_head(queue);
drivers/staging/rtl8723bs/core/rtw_xmit.c
1689
spin_unlock_bh(&queue->lock);
drivers/staging/rtl8723bs/core/rtw_xmit.c
1724
struct __queue *queue = NULL;
drivers/staging/rtl8723bs/core/rtw_xmit.c
1742
queue = &pxmitpriv->free_xmit_queue;
drivers/staging/rtl8723bs/core/rtw_xmit.c
1744
queue = &pxmitpriv->free_xframe_ext_queue;
drivers/staging/rtl8723bs/core/rtw_xmit.c
1746
spin_lock_bh(&queue->lock);
drivers/staging/rtl8723bs/core/rtw_xmit.c
1749
list_add_tail(&pxmitframe->list, get_list_head(queue));
drivers/staging/rtl8723bs/core/rtw_xmit.c
1755
spin_unlock_bh(&queue->lock);
drivers/staging/rtl8723bs/core/rtw_xmit.c
183
&pxmitpriv->free_xframe_ext_queue.queue);
drivers/staging/rtl8723bs/core/rtw_xmit.c
190
INIT_LIST_HEAD(&pxmitpriv->free_xmit_extbuf_queue.queue);
drivers/staging/rtl8723bs/core/rtw_xmit.c
223
&pxmitpriv->free_xmit_extbuf_queue.queue);
drivers/staging/rtl8723bs/core/rtw_xmit.c
2446
if (!list_empty(&pqueue->queue)) {
drivers/staging/rtl8723bs/core/rtw_xmit.c
2470
if (!list_empty(&pqueue->queue)) {
drivers/staging/rtl8723bs/core/rtw_xmit.c
2509
if (!list_empty(&pqueue->queue))
drivers/staging/rtl8723bs/core/rtw_xmit.c
52
INIT_LIST_HEAD(&pxmitpriv->be_pending.queue);
drivers/staging/rtl8723bs/core/rtw_xmit.c
54
INIT_LIST_HEAD(&pxmitpriv->bk_pending.queue);
drivers/staging/rtl8723bs/core/rtw_xmit.c
56
INIT_LIST_HEAD(&pxmitpriv->vi_pending.queue);
drivers/staging/rtl8723bs/core/rtw_xmit.c
58
INIT_LIST_HEAD(&pxmitpriv->vo_pending.queue);
drivers/staging/rtl8723bs/core/rtw_xmit.c
60
INIT_LIST_HEAD(&pxmitpriv->bm_pending.queue);
drivers/staging/rtl8723bs/core/rtw_xmit.c
63
INIT_LIST_HEAD(&pxmitpriv->free_xmit_queue.queue);
drivers/staging/rtl8723bs/core/rtw_xmit.c
95
&pxmitpriv->free_xmit_queue.queue);
drivers/staging/rtl8723bs/hal/rtl8723bs_recv.c
377
INIT_LIST_HEAD(&precvpriv->free_recv_buf_queue.queue);
drivers/staging/rtl8723bs/hal/rtl8723bs_recv.c
379
INIT_LIST_HEAD(&precvpriv->recv_buf_pending_queue.queue);
drivers/staging/rtl8723bs/hal/rtl8723bs_recv.c
410
list_add_tail(&precvbuf->list, &precvpriv->free_recv_buf_queue.queue);
drivers/staging/rtl8723bs/hal/rtl8723bs_xmit.c
315
if (list_empty(&pframe_queue->queue))
drivers/staging/rtl8723bs/hal/rtl8723bs_xmit.c
558
if (!list_empty(&pqueue->queue)) {
drivers/staging/rtl8723bs/include/cmd_osdep.h
15
int _rtw_enqueue_cmd(struct __queue *queue, struct cmd_obj *obj);
drivers/staging/rtl8723bs/include/cmd_osdep.h
16
extern struct cmd_obj *_rtw_dequeue_cmd(struct __queue *queue);
drivers/staging/rtl8723bs/include/osdep_service_linux.h
42
struct list_head queue;
drivers/staging/rtl8723bs/include/osdep_service_linux.h
51
static inline struct list_head *get_list_head(struct __queue *queue)
drivers/staging/rtl8723bs/include/osdep_service_linux.h
53
return (&(queue->queue));
drivers/staging/rtl8723bs/include/rtw_recv.h
333
extern int _rtw_enqueue_recvframe(union recv_frame *precvframe, struct __queue *queue);
drivers/staging/rtl8723bs/include/rtw_recv.h
334
extern int rtw_enqueue_recvframe(union recv_frame *precvframe, struct __queue *queue);
drivers/staging/rtl8723bs/include/rtw_recv.h
339
signed int rtw_enqueue_recvbuf_to_head(struct recv_buf *precvbuf, struct __queue *queue);
drivers/staging/rtl8723bs/include/rtw_recv.h
340
signed int rtw_enqueue_recvbuf(struct recv_buf *precvbuf, struct __queue *queue);
drivers/staging/rtl8723bs/include/rtw_recv.h
341
struct recv_buf *rtw_dequeue_recvbuf(struct __queue *queue);
drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c
1129
struct __queue *queue = &(pmlmepriv->scanned_queue);
drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c
1134
phead = get_list_head(queue);
drivers/staging/rtl8723bs/os_dep/xmit_linux.c
103
if (!list_empty(&padapter->xmitpriv.pending_xmitbuf_queue.queue))
drivers/staging/rtl8723bs/os_dep/xmit_linux.c
110
u16 queue;
drivers/staging/rtl8723bs/os_dep/xmit_linux.c
112
queue = skb_get_queue_mapping(pkt);
drivers/staging/rtl8723bs/os_dep/xmit_linux.c
115
if (pxmitpriv->hwxmits[queue].accnt > WMM_XMIT_THRESHOLD)
drivers/staging/rtl8723bs/os_dep/xmit_linux.c
116
netif_stop_subqueue(padapter->pnetdev, queue);
drivers/staging/rtl8723bs/os_dep/xmit_linux.c
119
if (!netif_tx_queue_stopped(netdev_get_tx_queue(padapter->pnetdev, queue)))
drivers/staging/rtl8723bs/os_dep/xmit_linux.c
120
netif_stop_subqueue(padapter->pnetdev, queue);
drivers/staging/rtl8723bs/os_dep/xmit_linux.c
72
u16 queue;
drivers/staging/rtl8723bs/os_dep/xmit_linux.c
75
queue = skb_get_queue_mapping(pkt);
drivers/staging/rtl8723bs/os_dep/xmit_linux.c
77
if (__netif_subqueue_stopped(padapter->pnetdev, queue) &&
drivers/staging/rtl8723bs/os_dep/xmit_linux.c
78
(pxmitpriv->hwxmits[queue].accnt < WMM_XMIT_THRESHOLD))
drivers/staging/rtl8723bs/os_dep/xmit_linux.c
79
netif_wake_subqueue(padapter->pnetdev, queue);
drivers/staging/rtl8723bs/os_dep/xmit_linux.c
81
if (__netif_subqueue_stopped(padapter->pnetdev, queue))
drivers/staging/rtl8723bs/os_dep/xmit_linux.c
82
netif_wake_subqueue(padapter->pnetdev, queue);
drivers/target/target_core_user.c
1040
goto queue;
drivers/target/target_core_user.c
1139
queue:
drivers/target/target_core_user.c
1226
static void tcmu_set_next_deadline(struct list_head *queue,
drivers/target/target_core_user.c
1231
if (!list_empty(queue)) {
drivers/target/target_core_user.c
1232
cmd = list_first_entry(queue, struct tcmu_cmd, queue_entry);
drivers/thunderbolt/nhi.c
244
list_for_each_entry_safe(frame, n, &ring->queue, list) {
drivers/thunderbolt/nhi.c
287
list_splice_tail_init(&ring->queue, &done);
drivers/thunderbolt/nhi.c
331
list_add_tail(&frame->list, &ring->queue);
drivers/thunderbolt/nhi.c
595
INIT_LIST_HEAD(&ring->queue);
drivers/tty/ipwireless/hardware.c
1243
list_add_tail(&packet->queue, &hw->tx_queue[priority]);
drivers/tty/ipwireless/hardware.c
1262
INIT_LIST_HEAD(&packet->queue);
drivers/tty/ipwireless/hardware.c
1285
INIT_LIST_HEAD(&packet->header.queue);
drivers/tty/ipwireless/hardware.c
1745
list_for_each_entry_safe(tp, tq, &hw->tx_queue[i], queue) {
drivers/tty/ipwireless/hardware.c
1746
list_del(&tp->queue);
drivers/tty/ipwireless/hardware.c
1750
list_for_each_entry_safe(rp, rq, &hw->rx_queue, queue) {
drivers/tty/ipwireless/hardware.c
1751
list_del(&rp->queue);
drivers/tty/ipwireless/hardware.c
1755
list_for_each_entry_safe(rp, rq, &hw->rx_pool, queue) {
drivers/tty/ipwireless/hardware.c
1756
list_del(&rp->queue);
drivers/tty/ipwireless/hardware.c
295
struct list_head queue;
drivers/tty/ipwireless/hardware.c
336
struct list_head queue;
drivers/tty/ipwireless/hardware.c
510
list_add(&packet->queue, &hw->tx_queue[0]);
drivers/tty/ipwireless/hardware.c
566
struct ipw_rx_packet, queue);
drivers/tty/ipwireless/hardware.c
569
list_del(&packet->queue);
drivers/tty/ipwireless/hardware.c
615
list_add(&packet->queue, &hw->rx_pool);
drivers/tty/ipwireless/hardware.c
690
list_add_tail(&packet->queue, &hw->rx_queue);
drivers/tty/ipwireless/hardware.c
713
struct ipw_rx_packet, queue);
drivers/tty/ipwireless/hardware.c
717
list_del(&packet->queue);
drivers/tty/ipwireless/hardware.c
971
queue);
drivers/tty/ipwireless/hardware.c
974
list_del(&packet->queue);
drivers/usb/c67x00/c67x00-sched.c
253
INIT_LIST_HEAD(&ep_data->queue);
drivers/usb/c67x00/c67x00-sched.c
292
if (!list_empty(&ep_data->queue))
drivers/usb/c67x00/c67x00-sched.c
296
list_del(&ep_data->queue);
drivers/usb/c67x00/c67x00-sched.c
30
struct list_head queue;
drivers/usb/c67x00/c67x00-sched.c
401
if (list_empty(&urbp->ep_data->queue))
drivers/usb/c67x00/c67x00-sched.c
407
last_urb = list_entry(urbp->ep_data->queue.prev,
drivers/usb/c67x00/c67x00-sched.c
420
list_add_tail(&urbp->hep_node, &urbp->ep_data->queue);
drivers/usb/c67x00/c67x00-sched.c
785
if (!list_empty(&ep_data->queue)) {
drivers/usb/c67x00/c67x00-sched.c
788
urb = list_entry(ep_data->queue.next,
drivers/usb/cdns3/cdns3-ep0.c
798
.queue = cdns3_gadget_ep0_queue,
drivers/usb/cdns3/cdns3-gadget.c
2870
.queue = cdns3_gadget_ep_queue,
drivers/usb/cdns3/cdnsp-gadget.c
1231
.queue = cdnsp_gadget_ep_queue,
drivers/usb/cdns3/cdnsp-gadget.c
1242
.queue = cdnsp_gadget_ep_queue,
drivers/usb/chipidea/ci.h
96
struct list_head queue;
drivers/usb/chipidea/debug.c
168
list_for_each_entry(req, &ci->ci_hw_ep[i].qh.queue, queue) {
drivers/usb/chipidea/udc.c
1114
if (!list_empty(&hwep->qh.queue)) {
drivers/usb/chipidea/udc.c
1134
if (!list_empty(&hwreq->queue)) {
drivers/usb/chipidea/udc.c
1148
list_add_tail(&hwreq->queue, &hwep->qh.queue);
drivers/usb/chipidea/udc.c
1285
list_for_each_entry_safe(hwreq, hwreqtemp, &hwep->qh.queue,
drivers/usb/chipidea/udc.c
1286
queue) {
drivers/usb/chipidea/udc.c
1290
list_del_init(&hwreq->queue);
drivers/usb/chipidea/udc.c
1548
if (!list_empty(&hwep->qh.queue)) {
drivers/usb/chipidea/udc.c
1650
INIT_LIST_HEAD(&hwreq->queue);
drivers/usb/chipidea/udc.c
1671
} else if (!list_empty(&hwreq->queue)) {
drivers/usb/chipidea/udc.c
1728
hwep->ep.desc == NULL || list_empty(&hwreq->queue) ||
drivers/usb/chipidea/udc.c
1729
list_empty(&hwep->qh.queue))
drivers/usb/chipidea/udc.c
1743
list_del_init(&hwreq->queue);
drivers/usb/chipidea/udc.c
1827
.queue = ep_queue,
drivers/usb/chipidea/udc.c
2068
INIT_LIST_HEAD(&hwep->qh.queue);
drivers/usb/chipidea/udc.c
731
if (!list_empty(&hwep->qh.queue)) {
drivers/usb/chipidea/udc.c
738
hwreqprev = list_entry(hwep->qh.queue.prev,
drivers/usb/chipidea/udc.c
739
struct ci_hw_req, queue);
drivers/usb/chipidea/udc.c
918
while (!list_empty(&hwep->qh.queue)) {
drivers/usb/chipidea/udc.c
921
struct ci_hw_req *hwreq = list_entry(hwep->qh.queue.next,
drivers/usb/chipidea/udc.c
922
struct ci_hw_req, queue);
drivers/usb/chipidea/udc.c
931
list_del_init(&hwreq->queue);
drivers/usb/chipidea/udc.c
969
!list_empty(&hwep->qh.queue) &&
drivers/usb/chipidea/udc.h
76
struct list_head queue;
drivers/usb/dwc2/core.h
124
struct list_head queue;
drivers/usb/dwc2/core.h
169
struct list_head queue;
drivers/usb/dwc2/debugfs.c
260
ep->queue.next, ep->queue.prev);
drivers/usb/dwc2/debugfs.c
264
list_for_each_entry(req, &ep->queue, queue) {
drivers/usb/dwc2/gadget.c
1007
list_for_each_entry_safe(hs_req, treq, &hs_ep->queue, queue) {
drivers/usb/dwc2/gadget.c
1430
INIT_LIST_HEAD(&hs_req->queue);
drivers/usb/dwc2/gadget.c
1476
first = list_empty(&hs_ep->queue);
drivers/usb/dwc2/gadget.c
1477
list_add_tail(&hs_req->queue, &hs_ep->queue);
drivers/usb/dwc2/gadget.c
1734
return list_first_entry_or_null(&hs_ep->queue, struct dwc2_hsotg_req,
drivers/usb/dwc2/gadget.c
1735
queue);
drivers/usb/dwc2/gadget.c
1752
if (!list_empty(&hs_ep->queue)) {
drivers/usb/dwc2/gadget.c
1859
list_del_init(&hs_req->queue);
drivers/usb/dwc2/gadget.c
2051
if (!list_empty(&hs_req->queue)) {
drivers/usb/dwc2/gadget.c
2145
list_del_init(&hs_req->queue);
drivers/usb/dwc2/gadget.c
2857
if (list_empty(&hs_ep->queue)) {
drivers/usb/dwc2/gadget.c
3294
while (!list_empty(&ep->queue)) {
drivers/usb/dwc2/gadget.c
404
INIT_LIST_HEAD(&req->queue);
drivers/usb/dwc2/gadget.c
4317
list_for_each_entry_safe(req, treq, &ep->queue, queue) {
drivers/usb/dwc2/gadget.c
4412
if (!now && value && !list_empty(&hs_ep->queue)) {
drivers/usb/dwc2/gadget.c
4485
.queue = dwc2_hsotg_ep_queue_lock,
drivers/usb/dwc2/gadget.c
4828
INIT_LIST_HEAD(&hs_ep->queue);
drivers/usb/dwc2/gadget.c
992
if (list_empty(&hs_ep->queue)) {
drivers/usb/dwc3/gadget.c
2328
.queue = dwc3_gadget_ep0_queue,
drivers/usb/dwc3/gadget.c
2339
.queue = dwc3_gadget_ep_queue,
drivers/usb/fotg210/fotg210-udc.c
1239
INIT_LIST_HEAD(&ep->queue);
drivers/usb/fotg210/fotg210-udc.c
231
while (!list_empty(&ep->queue)) {
drivers/usb/fotg210/fotg210-udc.c
232
req = list_entry(ep->queue.next,
drivers/usb/fotg210/fotg210-udc.c
233
struct fotg210_request, queue);
drivers/usb/fotg210/fotg210-udc.c
251
INIT_LIST_HEAD(&req->queue);
drivers/usb/fotg210/fotg210-udc.c
420
if (list_empty(&ep->queue))
drivers/usb/fotg210/fotg210-udc.c
423
list_add_tail(&req->queue, &ep->queue);
drivers/usb/fotg210/fotg210-udc.c
448
if (!list_empty(&ep->queue))
drivers/usb/fotg210/fotg210-udc.c
512
if (!list_empty(&ep->queue))
drivers/usb/fotg210/fotg210-udc.c
541
.queue = fotg210_ep_queue,
drivers/usb/fotg210/fotg210-udc.c
73
list_del_init(&req->queue);
drivers/usb/fotg210/fotg210-udc.c
819
if (!list_empty(&ep->queue) && !ep->dir_in) {
drivers/usb/fotg210/fotg210-udc.c
822
req = list_first_entry(&ep->queue,
drivers/usb/fotg210/fotg210-udc.c
823
struct fotg210_request, queue);
drivers/usb/fotg210/fotg210-udc.c
839
if ((!list_empty(&ep->queue)) && (ep->dir_in)) {
drivers/usb/fotg210/fotg210-udc.c
842
req = list_entry(ep->queue.next,
drivers/usb/fotg210/fotg210-udc.c
843
struct fotg210_request, queue);
drivers/usb/fotg210/fotg210-udc.c
857
struct fotg210_request *req = list_entry(ep->queue.next,
drivers/usb/fotg210/fotg210-udc.c
858
struct fotg210_request, queue);
drivers/usb/fotg210/fotg210-udc.c
86
if (list_empty(&ep->queue))
drivers/usb/fotg210/fotg210-udc.c
867
struct fotg210_request *req = list_entry(ep->queue.next,
drivers/usb/fotg210/fotg210-udc.c
868
struct fotg210_request, queue);
drivers/usb/fotg210/fotg210-udc.h
214
struct list_head queue;
drivers/usb/fotg210/fotg210-udc.h
221
struct list_head queue;
drivers/usb/gadget/function/u_serial.c
369
struct list_head *queue = &port->read_queue;
drivers/usb/gadget/function/u_serial.c
376
while (!list_empty(queue)) {
drivers/usb/gadget/function/u_serial.c
379
req = list_first_entry(queue, struct usb_request, list);
drivers/usb/gadget/function/u_serial.c
448
if (!list_empty(queue) && !tty_throttled(tty))
drivers/usb/gadget/function/uvc.h
142
struct uvc_video_queue queue;
drivers/usb/gadget/function/uvc_queue.c
107
struct uvc_video_queue *queue = vb2_get_drv_priv(vb->vb2_queue);
drivers/usb/gadget/function/uvc_queue.c
112
spin_lock_irqsave(&queue->irqlock, flags);
drivers/usb/gadget/function/uvc_queue.c
114
if (likely(!(queue->flags & UVC_QUEUE_DISCONNECTED))) {
drivers/usb/gadget/function/uvc_queue.c
115
list_add_tail(&buf->queue, &queue->irqqueue);
drivers/usb/gadget/function/uvc_queue.c
125
spin_unlock_irqrestore(&queue->irqlock, flags);
drivers/usb/gadget/function/uvc_queue.c
134
int uvcg_queue_init(struct uvc_video_queue *queue, struct device *dev, enum v4l2_buf_type type,
drivers/usb/gadget/function/uvc_queue.c
137
struct uvc_video *video = container_of(queue, struct uvc_video, queue);
drivers/usb/gadget/function/uvc_queue.c
141
queue->queue.type = type;
drivers/usb/gadget/function/uvc_queue.c
142
queue->queue.io_modes = VB2_MMAP | VB2_USERPTR | VB2_DMABUF;
drivers/usb/gadget/function/uvc_queue.c
143
queue->queue.drv_priv = queue;
drivers/usb/gadget/function/uvc_queue.c
144
queue->queue.buf_struct_size = sizeof(struct uvc_buffer);
drivers/usb/gadget/function/uvc_queue.c
145
queue->queue.ops = &uvc_queue_qops;
drivers/usb/gadget/function/uvc_queue.c
146
queue->queue.lock = lock;
drivers/usb/gadget/function/uvc_queue.c
148
queue->queue.mem_ops = &vb2_dma_sg_memops;
drivers/usb/gadget/function/uvc_queue.c
149
queue->use_sg = 1;
drivers/usb/gadget/function/uvc_queue.c
151
queue->queue.mem_ops = &vb2_vmalloc_memops;
drivers/usb/gadget/function/uvc_queue.c
154
queue->queue.timestamp_flags = V4L2_BUF_FLAG_TIMESTAMP_COPY
drivers/usb/gadget/function/uvc_queue.c
156
queue->queue.dev = dev;
drivers/usb/gadget/function/uvc_queue.c
158
ret = vb2_queue_init(&queue->queue);
drivers/usb/gadget/function/uvc_queue.c
162
spin_lock_init(&queue->irqlock);
drivers/usb/gadget/function/uvc_queue.c
163
INIT_LIST_HEAD(&queue->irqqueue);
drivers/usb/gadget/function/uvc_queue.c
164
queue->flags = 0;
drivers/usb/gadget/function/uvc_queue.c
172
void uvcg_free_buffers(struct uvc_video_queue *queue)
drivers/usb/gadget/function/uvc_queue.c
174
vb2_queue_release(&queue->queue);
drivers/usb/gadget/function/uvc_queue.c
180
int uvcg_alloc_buffers(struct uvc_video_queue *queue,
drivers/usb/gadget/function/uvc_queue.c
186
ret = vb2_reqbufs(&queue->queue, rb);
drivers/usb/gadget/function/uvc_queue.c
187
if (ret < 0 && queue->use_sg) {
drivers/usb/gadget/function/uvc_queue.c
190
queue->use_sg = 0;
drivers/usb/gadget/function/uvc_queue.c
191
queue->queue.mem_ops = &vb2_vmalloc_memops;
drivers/usb/gadget/function/uvc_queue.c
198
int uvcg_query_buffer(struct uvc_video_queue *queue, struct v4l2_buffer *buf)
drivers/usb/gadget/function/uvc_queue.c
200
return vb2_querybuf(&queue->queue, buf);
drivers/usb/gadget/function/uvc_queue.c
203
int uvcg_queue_buffer(struct uvc_video_queue *queue, struct v4l2_buffer *buf)
drivers/usb/gadget/function/uvc_queue.c
205
return vb2_qbuf(&queue->queue, NULL, buf);
drivers/usb/gadget/function/uvc_queue.c
212
int uvcg_dequeue_buffer(struct uvc_video_queue *queue, struct v4l2_buffer *buf,
drivers/usb/gadget/function/uvc_queue.c
215
return vb2_dqbuf(&queue->queue, buf, nonblocking);
drivers/usb/gadget/function/uvc_queue.c
224
__poll_t uvcg_queue_poll(struct uvc_video_queue *queue, struct file *file,
drivers/usb/gadget/function/uvc_queue.c
227
return vb2_poll(&queue->queue, file, wait);
drivers/usb/gadget/function/uvc_queue.c
230
int uvcg_queue_mmap(struct uvc_video_queue *queue, struct vm_area_struct *vma)
drivers/usb/gadget/function/uvc_queue.c
232
return vb2_mmap(&queue->queue, vma);
drivers/usb/gadget/function/uvc_queue.c
241
unsigned long uvcg_queue_get_unmapped_area(struct uvc_video_queue *queue,
drivers/usb/gadget/function/uvc_queue.c
244
return vb2_get_unmapped_area(&queue->queue, 0, 0, pgoff, 0);
drivers/usb/gadget/function/uvc_queue.c
260
void uvcg_queue_cancel(struct uvc_video_queue *queue, int disconnect)
drivers/usb/gadget/function/uvc_queue.c
265
spin_lock_irqsave(&queue->irqlock, flags);
drivers/usb/gadget/function/uvc_queue.c
266
while (!list_empty(&queue->irqqueue)) {
drivers/usb/gadget/function/uvc_queue.c
267
buf = list_first_entry(&queue->irqqueue, struct uvc_buffer,
drivers/usb/gadget/function/uvc_queue.c
268
queue);
drivers/usb/gadget/function/uvc_queue.c
269
list_del(&buf->queue);
drivers/usb/gadget/function/uvc_queue.c
273
queue->buf_used = 0;
drivers/usb/gadget/function/uvc_queue.c
283
queue->flags |= UVC_QUEUE_DISCONNECTED;
drivers/usb/gadget/function/uvc_queue.c
284
spin_unlock_irqrestore(&queue->irqlock, flags);
drivers/usb/gadget/function/uvc_queue.c
304
int uvcg_queue_enable(struct uvc_video_queue *queue, int enable)
drivers/usb/gadget/function/uvc_queue.c
310
ret = vb2_streamon(&queue->queue, queue->queue.type);
drivers/usb/gadget/function/uvc_queue.c
314
queue->sequence = 0;
drivers/usb/gadget/function/uvc_queue.c
315
queue->buf_used = 0;
drivers/usb/gadget/function/uvc_queue.c
316
queue->flags &= ~UVC_QUEUE_DROP_INCOMPLETE;
drivers/usb/gadget/function/uvc_queue.c
318
ret = vb2_streamoff(&queue->queue, queue->queue.type);
drivers/usb/gadget/function/uvc_queue.c
322
spin_lock_irqsave(&queue->irqlock, flags);
drivers/usb/gadget/function/uvc_queue.c
323
INIT_LIST_HEAD(&queue->irqqueue);
drivers/usb/gadget/function/uvc_queue.c
331
queue->flags &= ~UVC_QUEUE_DISCONNECTED;
drivers/usb/gadget/function/uvc_queue.c
332
spin_unlock_irqrestore(&queue->irqlock, flags);
drivers/usb/gadget/function/uvc_queue.c
339
void uvcg_complete_buffer(struct uvc_video_queue *queue,
drivers/usb/gadget/function/uvc_queue.c
342
if (queue->flags & UVC_QUEUE_DROP_INCOMPLETE) {
drivers/usb/gadget/function/uvc_queue.c
343
queue->flags &= ~UVC_QUEUE_DROP_INCOMPLETE;
drivers/usb/gadget/function/uvc_queue.c
351
buf->buf.sequence = queue->sequence++;
drivers/usb/gadget/function/uvc_queue.c
358
struct uvc_buffer *uvcg_queue_head(struct uvc_video_queue *queue)
drivers/usb/gadget/function/uvc_queue.c
362
if (!list_empty(&queue->irqqueue))
drivers/usb/gadget/function/uvc_queue.c
363
buf = list_first_entry(&queue->irqqueue, struct uvc_buffer,
drivers/usb/gadget/function/uvc_queue.c
364
queue);
drivers/usb/gadget/function/uvc_queue.c
46
struct uvc_video_queue *queue = vb2_get_drv_priv(vq);
drivers/usb/gadget/function/uvc_queue.c
47
struct uvc_video *video = container_of(queue, struct uvc_video, queue);
drivers/usb/gadget/function/uvc_queue.c
63
struct uvc_video_queue *queue = vb2_get_drv_priv(vb->vb2_queue);
drivers/usb/gadget/function/uvc_queue.c
64
struct uvc_video *video = container_of(queue, struct uvc_video, queue);
drivers/usb/gadget/function/uvc_queue.c
74
if (unlikely(queue->flags & UVC_QUEUE_DISCONNECTED))
drivers/usb/gadget/function/uvc_queue.c
78
if (queue->use_sg) {
drivers/usb/gadget/function/uvc_queue.h
100
struct uvc_buffer *uvcg_queue_head(struct uvc_video_queue *queue);
drivers/usb/gadget/function/uvc_queue.h
33
struct list_head queue;
drivers/usb/gadget/function/uvc_queue.h
50
struct vb2_queue queue;
drivers/usb/gadget/function/uvc_queue.h
63
static inline int uvc_queue_streaming(struct uvc_video_queue *queue)
drivers/usb/gadget/function/uvc_queue.h
65
return vb2_is_streaming(&queue->queue);
drivers/usb/gadget/function/uvc_queue.h
68
int uvcg_queue_init(struct uvc_video_queue *queue, struct device *dev, enum v4l2_buf_type type,
drivers/usb/gadget/function/uvc_queue.h
71
void uvcg_free_buffers(struct uvc_video_queue *queue);
drivers/usb/gadget/function/uvc_queue.h
73
int uvcg_alloc_buffers(struct uvc_video_queue *queue,
drivers/usb/gadget/function/uvc_queue.h
76
int uvcg_query_buffer(struct uvc_video_queue *queue, struct v4l2_buffer *buf);
drivers/usb/gadget/function/uvc_queue.h
78
int uvcg_queue_buffer(struct uvc_video_queue *queue, struct v4l2_buffer *buf);
drivers/usb/gadget/function/uvc_queue.h
80
int uvcg_dequeue_buffer(struct uvc_video_queue *queue,
drivers/usb/gadget/function/uvc_queue.h
83
__poll_t uvcg_queue_poll(struct uvc_video_queue *queue,
drivers/usb/gadget/function/uvc_queue.h
86
int uvcg_queue_mmap(struct uvc_video_queue *queue, struct vm_area_struct *vma);
drivers/usb/gadget/function/uvc_queue.h
89
unsigned long uvcg_queue_get_unmapped_area(struct uvc_video_queue *queue,
drivers/usb/gadget/function/uvc_queue.h
93
void uvcg_queue_cancel(struct uvc_video_queue *queue, int disconnect);
drivers/usb/gadget/function/uvc_queue.h
95
int uvcg_queue_enable(struct uvc_video_queue *queue, int enable);
drivers/usb/gadget/function/uvc_queue.h
97
void uvcg_complete_buffer(struct uvc_video_queue *queue,
drivers/usb/gadget/function/uvc_v4l2.c
258
if (fmt->type != video->queue.queue.type)
drivers/usb/gadget/function/uvc_v4l2.c
473
if (b->type != video->queue.queue.type)
drivers/usb/gadget/function/uvc_v4l2.c
476
return uvcg_alloc_buffers(&video->queue, b);
drivers/usb/gadget/function/uvc_v4l2.c
486
return uvcg_query_buffer(&video->queue, b);
drivers/usb/gadget/function/uvc_v4l2.c
497
ret = uvcg_queue_buffer(&video->queue, b);
drivers/usb/gadget/function/uvc_v4l2.c
514
return uvcg_dequeue_buffer(&video->queue, b, file->f_flags & O_NONBLOCK);
drivers/usb/gadget/function/uvc_v4l2.c
525
if (type != video->queue.queue.type)
drivers/usb/gadget/function/uvc_v4l2.c
551
if (type != video->queue.queue.type)
drivers/usb/gadget/function/uvc_v4l2.c
599
uvcg_free_buffers(&uvc->video.queue);
drivers/usb/gadget/function/uvc_v4l2.c
711
return uvcg_queue_mmap(&uvc->video.queue, vma);
drivers/usb/gadget/function/uvc_v4l2.c
720
return uvcg_queue_poll(&uvc->video.queue, file, wait);
drivers/usb/gadget/function/uvc_v4l2.c
731
return uvcg_queue_get_unmapped_area(&uvc->video.queue, pgoff);
drivers/usb/gadget/function/uvc_video.c
117
if (buf->bytesused == video->queue.buf_used) {
drivers/usb/gadget/function/uvc_video.c
118
video->queue.buf_used = 0;
drivers/usb/gadget/function/uvc_video.c
120
list_del(&buf->queue);
drivers/usb/gadget/function/uvc_video.c
128
video->queue.flags & UVC_QUEUE_DROP_INCOMPLETE ||
drivers/usb/gadget/function/uvc_video.c
129
buf->bytesused == video->queue.buf_used)
drivers/usb/gadget/function/uvc_video.c
137
unsigned int pending = buf->bytesused - video->queue.buf_used;
drivers/usb/gadget/function/uvc_video.c
187
video->queue.buf_used += req->length - header_len;
drivers/usb/gadget/function/uvc_video.c
189
if (buf->bytesused == video->queue.buf_used || !buf->sg ||
drivers/usb/gadget/function/uvc_video.c
190
video->queue.flags & UVC_QUEUE_DROP_INCOMPLETE) {
drivers/usb/gadget/function/uvc_video.c
191
video->queue.buf_used = 0;
drivers/usb/gadget/function/uvc_video.c
194
list_del(&buf->queue);
drivers/usb/gadget/function/uvc_video.c
220
if (buf->bytesused == video->queue.buf_used ||
drivers/usb/gadget/function/uvc_video.c
221
video->queue.flags & UVC_QUEUE_DROP_INCOMPLETE) {
drivers/usb/gadget/function/uvc_video.c
222
video->queue.buf_used = 0;
drivers/usb/gadget/function/uvc_video.c
224
list_del(&buf->queue);
drivers/usb/gadget/function/uvc_video.c
335
struct uvc_video_queue *queue = &video->queue;
drivers/usb/gadget/function/uvc_video.c
364
queue->flags |= UVC_QUEUE_DROP_INCOMPLETE;
drivers/usb/gadget/function/uvc_video.c
369
uvcg_queue_cancel(queue, 1);
drivers/usb/gadget/function/uvc_video.c
376
uvcg_queue_cancel(queue, 0);
drivers/usb/gadget/function/uvc_video.c
380
spin_lock_irqsave(&queue->irqlock, flags);
drivers/usb/gadget/function/uvc_video.c
381
uvcg_complete_buffer(queue, last_buf);
drivers/usb/gadget/function/uvc_video.c
382
spin_unlock_irqrestore(&queue->irqlock, flags);
drivers/usb/gadget/function/uvc_video.c
39
if (video->queue.flags & UVC_QUEUE_DROP_INCOMPLETE)
drivers/usb/gadget/function/uvc_video.c
395
uvcg_queue_cancel(queue, 0);
drivers/usb/gadget/function/uvc_video.c
42
if (video->queue.buf_used == 0 && ts.tv_sec) {
drivers/usb/gadget/function/uvc_video.c
612
struct uvc_video_queue *queue = &video->queue;
drivers/usb/gadget/function/uvc_video.c
642
spin_lock_irqsave(&queue->irqlock, flags);
drivers/usb/gadget/function/uvc_video.c
643
buf = uvcg_queue_head(queue);
drivers/usb/gadget/function/uvc_video.c
650
spin_unlock_irqrestore(&queue->irqlock, flags);
drivers/usb/gadget/function/uvc_video.c
656
spin_unlock_irqrestore(&queue->irqlock, flags);
drivers/usb/gadget/function/uvc_video.c
66
if (buf->bytesused - video->queue.buf_used <= len - pos)
drivers/usb/gadget/function/uvc_video.c
669
uvcg_queue_cancel(queue, 0);
drivers/usb/gadget/function/uvc_video.c
712
list_add_tail(&ureq->last_buf->queue, &inflight_bufs);
drivers/usb/gadget/function/uvc_video.c
719
uvcg_queue_cancel(&video->queue, 0);
drivers/usb/gadget/function/uvc_video.c
749
spin_lock_irqsave(&video->queue.irqlock, flags);
drivers/usb/gadget/function/uvc_video.c
750
list_for_each_entry_safe(buf, btemp, &inflight_bufs, queue) {
drivers/usb/gadget/function/uvc_video.c
751
list_del(&buf->queue);
drivers/usb/gadget/function/uvc_video.c
752
uvcg_complete_buffer(&video->queue, buf);
drivers/usb/gadget/function/uvc_video.c
754
spin_unlock_irqrestore(&video->queue.irqlock, flags);
drivers/usb/gadget/function/uvc_video.c
756
uvcg_queue_enable(&video->queue, 0);
drivers/usb/gadget/function/uvc_video.c
76
struct uvc_video_queue *queue = &video->queue;
drivers/usb/gadget/function/uvc_video.c
781
if ((ret = uvcg_queue_enable(&video->queue, 1)) < 0)
drivers/usb/gadget/function/uvc_video.c
791
video->encode = video->queue.use_sg ?
drivers/usb/gadget/function/uvc_video.c
81
mem = buf->mem + queue->buf_used;
drivers/usb/gadget/function/uvc_video.c
82
nbytes = min_t(unsigned int, len, buf->bytesused - queue->buf_used);
drivers/usb/gadget/function/uvc_video.c
841
return uvcg_queue_init(&video->queue, uvc->v4l2_dev.dev->parent,
drivers/usb/gadget/function/uvc_video.c
85
queue->buf_used += nbytes;
drivers/usb/gadget/legacy/raw_gadget.c
101
ret = down_interruptible(&queue->sema);
drivers/usb/gadget/legacy/raw_gadget.c
104
spin_lock_irqsave(&queue->lock, flags);
drivers/usb/gadget/legacy/raw_gadget.c
109
if (WARN_ON(!queue->size)) {
drivers/usb/gadget/legacy/raw_gadget.c
110
spin_unlock_irqrestore(&queue->lock, flags);
drivers/usb/gadget/legacy/raw_gadget.c
113
event = queue->events[0];
drivers/usb/gadget/legacy/raw_gadget.c
114
queue->size--;
drivers/usb/gadget/legacy/raw_gadget.c
115
memmove(&queue->events[0], &queue->events[1],
drivers/usb/gadget/legacy/raw_gadget.c
116
queue->size * sizeof(queue->events[0]));
drivers/usb/gadget/legacy/raw_gadget.c
117
spin_unlock_irqrestore(&queue->lock, flags);
drivers/usb/gadget/legacy/raw_gadget.c
121
static void raw_event_queue_destroy(struct raw_event_queue *queue)
drivers/usb/gadget/legacy/raw_gadget.c
125
for (i = 0; i < queue->size; i++)
drivers/usb/gadget/legacy/raw_gadget.c
126
kfree(queue->events[i]);
drivers/usb/gadget/legacy/raw_gadget.c
127
queue->size = 0;
drivers/usb/gadget/legacy/raw_gadget.c
186
struct raw_event_queue queue;
drivers/usb/gadget/legacy/raw_gadget.c
200
raw_event_queue_init(&dev->queue);
drivers/usb/gadget/legacy/raw_gadget.c
220
raw_event_queue_destroy(&dev->queue);
drivers/usb/gadget/legacy/raw_gadget.c
240
ret = raw_event_queue_add(&dev->queue, type, length, data);
drivers/usb/gadget/legacy/raw_gadget.c
55
static void raw_event_queue_init(struct raw_event_queue *queue)
drivers/usb/gadget/legacy/raw_gadget.c
57
spin_lock_init(&queue->lock);
drivers/usb/gadget/legacy/raw_gadget.c
58
sema_init(&queue->sema, 0);
drivers/usb/gadget/legacy/raw_gadget.c
59
queue->size = 0;
drivers/usb/gadget/legacy/raw_gadget.c
62
static int raw_event_queue_add(struct raw_event_queue *queue,
drivers/usb/gadget/legacy/raw_gadget.c
638
event = raw_event_queue_fetch(&dev->queue);
drivers/usb/gadget/legacy/raw_gadget.c
68
spin_lock_irqsave(&queue->lock, flags);
drivers/usb/gadget/legacy/raw_gadget.c
69
if (queue->size >= RAW_EVENT_QUEUE_SIZE) {
drivers/usb/gadget/legacy/raw_gadget.c
70
spin_unlock_irqrestore(&queue->lock, flags);
drivers/usb/gadget/legacy/raw_gadget.c
75
spin_unlock_irqrestore(&queue->lock, flags);
drivers/usb/gadget/legacy/raw_gadget.c
82
queue->events[queue->size] = event;
drivers/usb/gadget/legacy/raw_gadget.c
83
queue->size++;
drivers/usb/gadget/legacy/raw_gadget.c
84
up(&queue->sema);
drivers/usb/gadget/legacy/raw_gadget.c
85
spin_unlock_irqrestore(&queue->lock, flags);
drivers/usb/gadget/legacy/raw_gadget.c
90
struct raw_event_queue *queue)
drivers/usb/gadget/udc/amd5536udc.h
489
struct list_head queue;
drivers/usb/gadget/udc/amd5536udc.h
520
struct list_head queue;
drivers/usb/gadget/udc/aspeed-vhub/core.c
38
list_del_init(&req->queue);
drivers/usb/gadget/udc/aspeed-vhub/core.c
67
while (!list_empty(&ep->queue)) {
drivers/usb/gadget/udc/aspeed-vhub/core.c
68
req = list_first_entry(&ep->queue, struct ast_vhub_req, queue);
drivers/usb/gadget/udc/aspeed-vhub/ep0.c
285
req = list_first_entry_or_null(&ep->queue, struct ast_vhub_req, queue);
drivers/usb/gadget/udc/aspeed-vhub/ep0.c
406
if (!list_empty(&ep->queue) ||
drivers/usb/gadget/udc/aspeed-vhub/ep0.c
411
list_empty(&ep->queue), ep->ep0.state);
drivers/usb/gadget/udc/aspeed-vhub/ep0.c
417
list_add_tail(&req->queue, &ep->queue);
drivers/usb/gadget/udc/aspeed-vhub/ep0.c
449
req = list_first_entry_or_null(&ep->queue, struct ast_vhub_req, queue);
drivers/usb/gadget/udc/aspeed-vhub/ep0.c
473
.queue = ast_vhub_ep0_queue,
drivers/usb/gadget/udc/aspeed-vhub/ep0.c
494
INIT_LIST_HEAD(&ep->queue);
drivers/usb/gadget/udc/aspeed-vhub/ep0.c
503
INIT_LIST_HEAD(&ep->ep0.req.queue);
drivers/usb/gadget/udc/aspeed-vhub/ep0.c
54
if (ep->ep.ops->queue(&ep->ep, req, GFP_ATOMIC))
drivers/usb/gadget/udc/aspeed-vhub/epn.c
142
req = list_first_entry_or_null(&ep->queue, struct ast_vhub_req,
drivers/usb/gadget/udc/aspeed-vhub/epn.c
143
queue);
drivers/usb/gadget/udc/aspeed-vhub/epn.c
264
req = list_first_entry_or_null(&ep->queue, struct ast_vhub_req, queue);
drivers/usb/gadget/udc/aspeed-vhub/epn.c
315
req = list_first_entry_or_null(&ep->queue,
drivers/usb/gadget/udc/aspeed-vhub/epn.c
317
queue);
drivers/usb/gadget/udc/aspeed-vhub/epn.c
404
empty = list_empty(&ep->queue);
drivers/usb/gadget/udc/aspeed-vhub/epn.c
407
list_add_tail(&req->queue, &ep->queue);
drivers/usb/gadget/udc/aspeed-vhub/epn.c
483
list_for_each_entry(iter, &ep->queue, queue) {
drivers/usb/gadget/udc/aspeed-vhub/epn.c
540
if (halt && ep->epn.is_in && !list_empty(&ep->queue)) {
drivers/usb/gadget/udc/aspeed-vhub/epn.c
793
.queue = ast_vhub_epn_queue,
drivers/usb/gadget/udc/aspeed-vhub/epn.c
824
INIT_LIST_HEAD(&ep->queue);
drivers/usb/gadget/udc/aspeed-vhub/epn.c
93
req = list_first_entry_or_null(&ep->queue, struct ast_vhub_req, queue);
drivers/usb/gadget/udc/aspeed-vhub/vhub.h
236
struct list_head queue;
drivers/usb/gadget/udc/aspeed-vhub/vhub.h
272
struct list_head queue;
drivers/usb/gadget/udc/aspeed_udc.c
1051
list_for_each_entry(req, &udc->ep[0].queue, queue) {
drivers/usb/gadget/udc/aspeed_udc.c
1374
INIT_LIST_HEAD(&ep->queue);
drivers/usb/gadget/udc/aspeed_udc.c
163
struct list_head queue;
drivers/usb/gadget/udc/aspeed_udc.c
180
struct list_head queue;
drivers/usb/gadget/udc/aspeed_udc.c
290
list_del(&req->queue);
drivers/usb/gadget/udc/aspeed_udc.c
309
while (!list_empty(&ep->queue)) {
drivers/usb/gadget/udc/aspeed_udc.c
312
req = list_entry(ep->queue.next, struct ast_udc_request,
drivers/usb/gadget/udc/aspeed_udc.c
313
queue);
drivers/usb/gadget/udc/aspeed_udc.c
461
INIT_LIST_HEAD(&req->queue);
drivers/usb/gadget/udc/aspeed_udc.c
648
list_add_tail(&req->queue, &ep->queue);
drivers/usb/gadget/udc/aspeed_udc.c
678
if (list_is_singular(&ep->queue)) {
drivers/usb/gadget/udc/aspeed_udc.c
702
list_for_each_entry(req, &ep->queue, queue) {
drivers/usb/gadget/udc/aspeed_udc.c
704
list_del_init(&req->queue);
drivers/usb/gadget/udc/aspeed_udc.c
768
.queue = ast_udc_ep_queue,
drivers/usb/gadget/udc/aspeed_udc.c
793
if (list_empty(&ep->queue))
drivers/usb/gadget/udc/aspeed_udc.c
796
req = list_entry(ep->queue.next, struct ast_udc_request, queue);
drivers/usb/gadget/udc/aspeed_udc.c
832
if (list_empty(&ep->queue)) {
drivers/usb/gadget/udc/aspeed_udc.c
841
req = list_entry(ep->queue.next, struct ast_udc_request, queue);
drivers/usb/gadget/udc/aspeed_udc.c
865
if (list_empty(&ep->queue))
drivers/usb/gadget/udc/aspeed_udc.c
868
req = list_first_entry(&ep->queue, struct ast_udc_request, queue);
drivers/usb/gadget/udc/aspeed_udc.c
879
req = list_first_entry_or_null(&ep->queue,
drivers/usb/gadget/udc/aspeed_udc.c
881
queue);
drivers/usb/gadget/udc/aspeed_udc.c
887
req = list_first_entry_or_null(&ep->queue,
drivers/usb/gadget/udc/aspeed_udc.c
889
queue);
drivers/usb/gadget/udc/aspeed_udc.c
908
if (list_empty(&ep->queue)) {
drivers/usb/gadget/udc/aspeed_udc.c
913
req = list_first_entry(&ep->queue, struct ast_udc_request, queue);
drivers/usb/gadget/udc/aspeed_udc.c
958
req = list_first_entry_or_null(&ep->queue,
drivers/usb/gadget/udc/aspeed_udc.c
960
queue);
drivers/usb/gadget/udc/aspeed_udc.c
966
req = list_first_entry_or_null(&ep->queue,
drivers/usb/gadget/udc/aspeed_udc.c
968
queue);
drivers/usb/gadget/udc/at91_udc.c
1012
if (!list_empty(&ep->queue))
drivers/usb/gadget/udc/at91_udc.c
1013
req = list_entry(ep->queue.next,
drivers/usb/gadget/udc/at91_udc.c
1014
struct at91_request, queue);
drivers/usb/gadget/udc/at91_udc.c
1238
if (!list_empty(&ep->queue))
drivers/usb/gadget/udc/at91_udc.c
1297
if (list_empty(&ep0->queue))
drivers/usb/gadget/udc/at91_udc.c
1300
req = list_entry(ep0->queue.next, struct at91_request, queue);
drivers/usb/gadget/udc/at91_udc.c
151
if (list_empty (&ep->queue))
drivers/usb/gadget/udc/at91_udc.c
154
else list_for_each_entry (req, &ep->queue, queue) {
drivers/usb/gadget/udc/at91_udc.c
265
list_del_init(&req->queue);
drivers/usb/gadget/udc/at91_udc.c
280
if (list_empty(&ep->queue) && ep->int_mask != (1 << 0))
drivers/usb/gadget/udc/at91_udc.c
460
if (list_empty(&ep->queue))
drivers/usb/gadget/udc/at91_udc.c
464
while (!list_empty(&ep->queue)) {
drivers/usb/gadget/udc/at91_udc.c
465
req = list_entry(ep->queue.next, struct at91_request, queue);
drivers/usb/gadget/udc/at91_udc.c
592
INIT_LIST_HEAD(&req->queue);
drivers/usb/gadget/udc/at91_udc.c
601
BUG_ON(!list_empty(&req->queue));
drivers/usb/gadget/udc/at91_udc.c
618
|| !_req->buf || !list_empty(&req->queue)) {
drivers/usb/gadget/udc/at91_udc.c
641
if (list_empty(&ep->queue) && !ep->stopped) {
drivers/usb/gadget/udc/at91_udc.c
697
list_add_tail (&req->queue, &ep->queue);
drivers/usb/gadget/udc/at91_udc.c
721
list_for_each_entry(iter, &ep->queue, queue) {
drivers/usb/gadget/udc/at91_udc.c
759
if (ep->is_in && (!list_empty(&ep->queue) || (csr >> 16) != 0))
drivers/usb/gadget/udc/at91_udc.c
784
.queue = at91_ep_queue,
drivers/usb/gadget/udc/at91_udc.c
846
INIT_LIST_HEAD(&ep->queue);
drivers/usb/gadget/udc/at91_udc.h
154
struct list_head queue;
drivers/usb/gadget/udc/at91_udc.h
92
struct list_head queue;
drivers/usb/gadget/udc/atmel_usba_udc.c
110
list_del(&req->queue);
drivers/usb/gadget/udc/atmel_usba_udc.c
1141
list_for_each_entry_safe(req, tmp_req, &ep->queue, queue) {
drivers/usb/gadget/udc/atmel_usba_udc.c
1142
list_del_init(&req->queue);
drivers/usb/gadget/udc/atmel_usba_udc.c
131
list_for_each_entry_safe(req, tmp_req, queue_data, queue) {
drivers/usb/gadget/udc/atmel_usba_udc.c
132
list_del(&req->queue);
drivers/usb/gadget/udc/atmel_usba_udc.c
1442
if (!list_empty(&ep->queue))
drivers/usb/gadget/udc/atmel_usba_udc.c
1443
req = list_entry(ep->queue.next,
drivers/usb/gadget/udc/atmel_usba_udc.c
1444
struct usba_request, queue);
drivers/usb/gadget/udc/atmel_usba_udc.c
1476
list_del_init(&req->queue);
drivers/usb/gadget/udc/atmel_usba_udc.c
1506
list_del_init(&req->queue);
drivers/usb/gadget/udc/atmel_usba_udc.c
1556
list_del_init(&req->queue);
drivers/usb/gadget/udc/atmel_usba_udc.c
1630
if (list_empty(&ep->queue)) {
drivers/usb/gadget/udc/atmel_usba_udc.c
1636
req = list_entry(ep->queue.next, struct usba_request, queue);
drivers/usb/gadget/udc/atmel_usba_udc.c
1644
list_del_init(&req->queue);
drivers/usb/gadget/udc/atmel_usba_udc.c
1654
list_del_init(&req->queue);
drivers/usb/gadget/udc/atmel_usba_udc.c
1695
if (list_empty(&ep->queue))
drivers/usb/gadget/udc/atmel_usba_udc.c
1700
req = list_entry(ep->queue.next, struct usba_request, queue);
drivers/usb/gadget/udc/atmel_usba_udc.c
1703
list_del_init(&req->queue);
drivers/usb/gadget/udc/atmel_usba_udc.c
2227
INIT_LIST_HEAD(&ep->queue);
drivers/usb/gadget/udc/atmel_usba_udc.c
465
if (list_empty(&ep->queue)) {
drivers/usb/gadget/udc/atmel_usba_udc.c
470
req = list_entry(ep->queue.next, struct usba_request, queue);
drivers/usb/gadget/udc/atmel_usba_udc.c
49
list_for_each_entry(req, &ep->queue, queue) {
drivers/usb/gadget/udc/atmel_usba_udc.c
496
if (list_empty(&ep->queue)) {
drivers/usb/gadget/udc/atmel_usba_udc.c
500
req = list_entry(ep->queue.next,
drivers/usb/gadget/udc/atmel_usba_udc.c
501
struct usba_request, queue);
drivers/usb/gadget/udc/atmel_usba_udc.c
521
list_del_init(&req->queue);
drivers/usb/gadget/udc/atmel_usba_udc.c
53
list_add_tail(&req_copy->queue, queue_data);
drivers/usb/gadget/udc/atmel_usba_udc.c
543
WARN_ON(!list_empty(&req->queue));
drivers/usb/gadget/udc/atmel_usba_udc.c
565
list_for_each_entry_safe(req, tmp_req, list, queue) {
drivers/usb/gadget/udc/atmel_usba_udc.c
566
list_del_init(&req->queue);
drivers/usb/gadget/udc/atmel_usba_udc.c
62
list_for_each_entry_safe(req, req_copy, queue_data, queue) {
drivers/usb/gadget/udc/atmel_usba_udc.c
63
list_del(&req->queue);
drivers/usb/gadget/udc/atmel_usba_udc.c
682
list_splice_init(&ep->queue, &req_list);
drivers/usb/gadget/udc/atmel_usba_udc.c
709
INIT_LIST_HEAD(&req->queue);
drivers/usb/gadget/udc/atmel_usba_udc.c
762
if (list_empty(&ep->queue))
drivers/usb/gadget/udc/atmel_usba_udc.c
765
list_add_tail(&req->queue, &ep->queue);
drivers/usb/gadget/udc/atmel_usba_udc.c
803
list_add_tail(&req->queue, &ep->queue);
drivers/usb/gadget/udc/atmel_usba_udc.c
87
struct list_head *queue = file->private_data;
drivers/usb/gadget/udc/atmel_usba_udc.c
871
list_for_each_entry(iter, &ep->queue, queue) {
drivers/usb/gadget/udc/atmel_usba_udc.c
888
if (ep->queue.next == &req->queue) {
drivers/usb/gadget/udc/atmel_usba_udc.c
907
list_del_init(&req->queue);
drivers/usb/gadget/udc/atmel_usba_udc.c
945
if (!list_empty(&ep->queue)
drivers/usb/gadget/udc/atmel_usba_udc.c
96
list_for_each_entry_safe(req, tmp_req, queue, queue) {
drivers/usb/gadget/udc/atmel_usba_udc.c
983
.queue = usba_ep_queue,
drivers/usb/gadget/udc/atmel_usba_udc.h
277
struct list_head queue;
drivers/usb/gadget/udc/atmel_usba_udc.h
301
struct list_head queue;
drivers/usb/gadget/udc/bcm63xx_udc.c
1051
BUG_ON(!list_empty(&bep->queue));
drivers/usb/gadget/udc/bcm63xx_udc.c
1090
if (!list_empty(&bep->queue)) {
drivers/usb/gadget/udc/bcm63xx_udc.c
1091
list_for_each_entry_safe(breq, n, &bep->queue, queue) {
drivers/usb/gadget/udc/bcm63xx_udc.c
1094
list_del(&breq->queue);
drivers/usb/gadget/udc/bcm63xx_udc.c
1184
list_add_tail(&breq->queue, &bep->queue);
drivers/usb/gadget/udc/bcm63xx_udc.c
1185
if (list_is_singular(&bep->queue))
drivers/usb/gadget/udc/bcm63xx_udc.c
1212
if (list_empty(&bep->queue)) {
drivers/usb/gadget/udc/bcm63xx_udc.c
1217
cur = list_first_entry(&bep->queue, struct bcm63xx_req, queue);
drivers/usb/gadget/udc/bcm63xx_udc.c
1222
list_del(&breq->queue);
drivers/usb/gadget/udc/bcm63xx_udc.c
1224
if (!list_empty(&bep->queue)) {
drivers/usb/gadget/udc/bcm63xx_udc.c
1227
next = list_first_entry(&bep->queue,
drivers/usb/gadget/udc/bcm63xx_udc.c
1228
struct bcm63xx_req, queue);
drivers/usb/gadget/udc/bcm63xx_udc.c
1232
list_del(&breq->queue);
drivers/usb/gadget/udc/bcm63xx_udc.c
1292
.queue = bcm63xx_udc_queue,
drivers/usb/gadget/udc/bcm63xx_udc.c
2077
} else if (!list_empty(&bep->queue)) {
drivers/usb/gadget/udc/bcm63xx_udc.c
2078
breq = list_first_entry(&bep->queue, struct bcm63xx_req, queue);
drivers/usb/gadget/udc/bcm63xx_udc.c
2086
list_del(&breq->queue);
drivers/usb/gadget/udc/bcm63xx_udc.c
2090
if (!list_empty(&bep->queue)) {
drivers/usb/gadget/udc/bcm63xx_udc.c
2093
next = list_first_entry(&bep->queue,
drivers/usb/gadget/udc/bcm63xx_udc.c
2094
struct bcm63xx_req, queue);
drivers/usb/gadget/udc/bcm63xx_udc.c
2207
seq_printf(s, "; %zu queued\n", list_count_nodes(&iudma->bep->queue));
drivers/usb/gadget/udc/bcm63xx_udc.c
243
struct list_head queue;
drivers/usb/gadget/udc/bcm63xx_udc.c
256
struct list_head queue; /* ep's requests */
drivers/usb/gadget/udc/bcm63xx_udc.c
771
INIT_LIST_HEAD(&bep->queue);
drivers/usb/gadget/udc/bcm63xx_udc.c
964
INIT_LIST_HEAD(&bep->queue);
drivers/usb/gadget/udc/bdc/bdc.h
335
struct list_head queue;
drivers/usb/gadget/udc/bdc/bdc.h
352
struct list_head queue;
drivers/usb/gadget/udc/bdc/bdc_ep.c
1097
while (!list_empty(&ep->queue)) {
drivers/usb/gadget/udc/bdc/bdc_ep.c
1098
req = list_entry(ep->queue.next, struct bdc_req,
drivers/usb/gadget/udc/bdc/bdc_ep.c
1099
queue);
drivers/usb/gadget/udc/bdc/bdc_ep.c
1777
list_for_each_entry(iter, &ep->queue, queue) {
drivers/usb/gadget/udc/bdc/bdc_ep.c
1815
else if (!list_empty(&ep->queue))
drivers/usb/gadget/udc/bdc/bdc_ep.c
1937
.queue = bdc_gadget_ep_queue,
drivers/usb/gadget/udc/bdc/bdc_ep.c
1993
INIT_LIST_HEAD(&ep->queue);
drivers/usb/gadget/udc/bdc/bdc_ep.c
529
list_add_tail(&req->queue, &ep->queue);
drivers/usb/gadget/udc/bdc/bdc_ep.c
546
list_del(&req->queue);
drivers/usb/gadget/udc/bdc/bdc_ep.c
574
while (!list_empty(&ep->queue)) {
drivers/usb/gadget/udc/bdc/bdc_ep.c
575
req = list_entry(ep->queue.next, struct bdc_req,
drivers/usb/gadget/udc/bdc/bdc_ep.c
576
queue);
drivers/usb/gadget/udc/bdc/bdc_ep.c
827
first_req = list_first_entry(&ep->queue, struct bdc_req,
drivers/usb/gadget/udc/bdc/bdc_ep.c
828
queue);
drivers/usb/gadget/udc/bdc/bdc_ep.c
973
if (unlikely(list_empty(&ep->queue))) {
drivers/usb/gadget/udc/bdc/bdc_ep.c
977
req = list_entry(ep->queue.next, struct bdc_req,
drivers/usb/gadget/udc/bdc/bdc_ep.c
978
queue);
drivers/usb/gadget/udc/cdns2/cdns2-ep0.c
613
.queue = cdns2_gadget_ep0_queue,
drivers/usb/gadget/udc/cdns2/cdns2-gadget.c
1994
.queue = cdns2_gadget_ep_queue,
drivers/usb/gadget/udc/core.c
303
ret = ep->ops->queue(ep, req, gfp_flags);
drivers/usb/gadget/udc/dummy_hcd.c
1071
INIT_LIST_HEAD(&ep->queue);
drivers/usb/gadget/udc/dummy_hcd.c
1076
INIT_LIST_HEAD(&dum->fifo_req.queue);
drivers/usb/gadget/udc/dummy_hcd.c
1420
list_for_each_entry(req, &ep->queue, queue) {
drivers/usb/gadget/udc/dummy_hcd.c
1524
list_del_init(&req->queue);
drivers/usb/gadget/udc/dummy_hcd.c
1907
list_for_each_entry(req, &ep->queue, queue) {
drivers/usb/gadget/udc/dummy_hcd.c
1908
list_del_init(&req->queue);
drivers/usb/gadget/udc/dummy_hcd.c
335
while (!list_empty(&ep->queue)) {
drivers/usb/gadget/udc/dummy_hcd.c
338
req = list_entry(ep->queue.next, struct dummy_request, queue);
drivers/usb/gadget/udc/dummy_hcd.c
339
list_del_init(&req->queue);
drivers/usb/gadget/udc/dummy_hcd.c
677
INIT_LIST_HEAD(&req->queue);
drivers/usb/gadget/udc/dummy_hcd.c
691
WARN_ON(!list_empty(&req->queue));
drivers/usb/gadget/udc/dummy_hcd.c
709
if (!_req || !list_empty(&req->queue) || !_req->complete)
drivers/usb/gadget/udc/dummy_hcd.c
731
list_empty(&dum->fifo_req.queue) &&
drivers/usb/gadget/udc/dummy_hcd.c
732
list_empty(&ep->queue) &&
drivers/usb/gadget/udc/dummy_hcd.c
741
list_add_tail(&req->queue, &ep->queue);
drivers/usb/gadget/udc/dummy_hcd.c
748
list_add_tail(&req->queue, &ep->queue);
drivers/usb/gadget/udc/dummy_hcd.c
774
list_for_each_entry(iter, &ep->queue, queue) {
drivers/usb/gadget/udc/dummy_hcd.c
777
list_del_init(&iter->queue);
drivers/usb/gadget/udc/dummy_hcd.c
811
!list_empty(&ep->queue))
drivers/usb/gadget/udc/dummy_hcd.c
842
.queue = dummy_queue,
drivers/usb/gadget/udc/dummy_hcd.c
86
struct list_head queue;
drivers/usb/gadget/udc/dummy_hcd.c
99
struct list_head queue; /* ep's requests */
drivers/usb/gadget/udc/fsl_qe_udc.c
1053
if (list_empty(&ep->queue)) {
drivers/usb/gadget/udc/fsl_qe_udc.c
1177
if (!list_empty(&ep->queue)) {
drivers/usb/gadget/udc/fsl_qe_udc.c
1178
ep->tx_req = list_entry(ep->queue.next, struct qe_req,
drivers/usb/gadget/udc/fsl_qe_udc.c
1179
queue);
drivers/usb/gadget/udc/fsl_qe_udc.c
131
while (!list_empty(&ep->queue)) {
drivers/usb/gadget/udc/fsl_qe_udc.c
133
req = list_entry(ep->queue.next, struct qe_req, queue);
drivers/usb/gadget/udc/fsl_qe_udc.c
1481
if (list_empty(&ep->queue)) {
drivers/usb/gadget/udc/fsl_qe_udc.c
1528
if (list_empty(&ep->queue))
drivers/usb/gadget/udc/fsl_qe_udc.c
1679
INIT_LIST_HEAD(&req->queue);
drivers/usb/gadget/udc/fsl_qe_udc.c
1704
|| !list_empty(&req->queue)) {
drivers/usb/gadget/udc/fsl_qe_udc.c
1739
list_add_tail(&req->queue, &ep->queue);
drivers/usb/gadget/udc/fsl_qe_udc.c
1790
list_for_each_entry(iter, &ep->queue, queue) {
drivers/usb/gadget/udc/fsl_qe_udc.c
1830
if (value && ep_is_in(ep) && !list_empty(&ep->queue)) {
drivers/usb/gadget/udc/fsl_qe_udc.c
1863
.queue = qe_ep_queue,
drivers/usb/gadget/udc/fsl_qe_udc.c
2443
INIT_LIST_HEAD(&ep->queue);
drivers/usb/gadget/udc/fsl_qe_udc.c
780
if (ep->has_data <= 0 && (!list_empty(&ep->queue)))
drivers/usb/gadget/udc/fsl_qe_udc.c
86
list_del_init(&req->queue);
drivers/usb/gadget/udc/fsl_qe_udc.c
907
if (list_empty(&ep->queue)) {
drivers/usb/gadget/udc/fsl_qe_udc.c
910
req = list_entry(ep->queue.next, struct qe_req, queue);
drivers/usb/gadget/udc/fsl_qe_udc.c
922
if (list_empty(&ep->queue) && ep->epnum != 0)
drivers/usb/gadget/udc/fsl_qe_udc.c
960
if (list_empty(&ep->queue)) {
drivers/usb/gadget/udc/fsl_qe_udc.h
254
struct list_head queue;
drivers/usb/gadget/udc/fsl_qe_udc.h
263
struct list_head queue;
drivers/usb/gadget/udc/fsl_udc_core.c
1026
if (value && ep_is_in(ep) && !list_empty(&ep->queue)) {
drivers/usb/gadget/udc/fsl_udc_core.c
1129
.queue = fsl_ep_queue,
drivers/usb/gadget/udc/fsl_udc_core.c
1304
list_add_tail(&req->queue, &ep->queue);
drivers/usb/gadget/udc/fsl_udc_core.c
1388
list_add_tail(&req->queue, &ep->queue);
drivers/usb/gadget/udc/fsl_udc_core.c
166
list_del_init(&req->queue);
drivers/usb/gadget/udc/fsl_udc_core.c
1707
list_for_each_entry_safe(curr_req, temp_req, &curr_ep->queue,
drivers/usb/gadget/udc/fsl_udc_core.c
1708
queue) {
drivers/usb/gadget/udc/fsl_udc_core.c
213
while (!list_empty(&ep->queue)) {
drivers/usb/gadget/udc/fsl_udc_core.c
216
req = list_entry(ep->queue.next, struct fsl_req, queue);
drivers/usb/gadget/udc/fsl_udc_core.c
2176
if (list_empty(&ep->queue)) {
drivers/usb/gadget/udc/fsl_udc_core.c
2179
list_for_each_entry(req, &ep->queue, queue) {
drivers/usb/gadget/udc/fsl_udc_core.c
2195
if (list_empty(&ep->queue)) {
drivers/usb/gadget/udc/fsl_udc_core.c
2198
list_for_each_entry(req, &ep->queue, queue) {
drivers/usb/gadget/udc/fsl_udc_core.c
2349
INIT_LIST_HEAD(&ep->queue);
drivers/usb/gadget/udc/fsl_udc_core.c
686
INIT_LIST_HEAD(&req->queue);
drivers/usb/gadget/udc/fsl_udc_core.c
735
if (!(list_empty(&ep->queue)) && !(ep_index(ep) == 0)) {
drivers/usb/gadget/udc/fsl_udc_core.c
738
lastreq = list_entry(ep->queue.prev, struct fsl_req, queue);
drivers/usb/gadget/udc/fsl_udc_core.c
878
|| !list_empty(&req->queue)) {
drivers/usb/gadget/udc/fsl_udc_core.c
914
list_add_tail(&req->queue, &ep->queue);
drivers/usb/gadget/udc/fsl_udc_core.c
947
list_for_each_entry(iter, &ep->queue, queue) {
drivers/usb/gadget/udc/fsl_udc_core.c
959
if (ep->queue.next == &req->queue) {
drivers/usb/gadget/udc/fsl_udc_core.c
964
if (req->queue.next != &ep->queue) {
drivers/usb/gadget/udc/fsl_udc_core.c
967
next_req = list_entry(req->queue.next, struct fsl_req,
drivers/usb/gadget/udc/fsl_udc_core.c
968
queue);
drivers/usb/gadget/udc/fsl_udc_core.c
977
prev_req = list_entry(req->queue.prev, struct fsl_req, queue);
drivers/usb/gadget/udc/fsl_usb2_udc.h
445
struct list_head queue;
drivers/usb/gadget/udc/fsl_usb2_udc.h
460
struct list_head queue;
drivers/usb/gadget/udc/goku_udc.c
1220
if (list_empty(&ep->queue)) {
drivers/usb/gadget/udc/goku_udc.c
1226
list_for_each_entry(req, &ep->queue, queue) {
drivers/usb/gadget/udc/goku_udc.c
1227
if (ep->dma && req->queue.prev == &ep->queue) {
drivers/usb/gadget/udc/goku_udc.c
1277
INIT_LIST_HEAD (&ep->queue);
drivers/usb/gadget/udc/goku_udc.c
1678
if (list_empty (&ep->queue))
drivers/usb/gadget/udc/goku_udc.c
280
INIT_LIST_HEAD(&req->queue);
drivers/usb/gadget/udc/goku_udc.c
293
WARN_ON(!list_empty(&req->queue));
drivers/usb/gadget/udc/goku_udc.c
305
list_del_init(&req->queue);
drivers/usb/gadget/udc/goku_udc.c
491
if (dbuff && !list_empty(&ep->queue)) {
drivers/usb/gadget/udc/goku_udc.c
492
req = list_entry(ep->queue.next,
drivers/usb/gadget/udc/goku_udc.c
493
struct goku_request, queue);
drivers/usb/gadget/udc/goku_udc.c
525
if (unlikely(list_empty (&ep->queue)))
drivers/usb/gadget/udc/goku_udc.c
527
req = list_entry(ep->queue.next, struct goku_request, queue);
drivers/usb/gadget/udc/goku_udc.c
598
if (unlikely(list_empty(&ep->queue))) {
drivers/usb/gadget/udc/goku_udc.c
607
req = list_entry(ep->queue.next, struct goku_request, queue);
drivers/usb/gadget/udc/goku_udc.c
632
if (list_empty(&ep->queue))
drivers/usb/gadget/udc/goku_udc.c
634
req = list_entry(ep->queue.next, struct goku_request, queue);
drivers/usb/gadget/udc/goku_udc.c
652
req = list_entry(ep->queue.next, struct goku_request, queue);
drivers/usb/gadget/udc/goku_udc.c
722
|| !_req->buf || !list_empty(&req->queue)))
drivers/usb/gadget/udc/goku_udc.c
761
if (list_empty(&ep->queue) && likely(!ep->stopped)) {
drivers/usb/gadget/udc/goku_udc.c
779
list_add_tail(&req->queue, &ep->queue);
drivers/usb/gadget/udc/goku_udc.c
781
if (likely(!list_empty(&ep->queue))
drivers/usb/gadget/udc/goku_udc.c
799
if (list_empty(&ep->queue))
drivers/usb/gadget/udc/goku_udc.c
803
while (!list_empty(&ep->queue)) {
drivers/usb/gadget/udc/goku_udc.c
804
req = list_entry(ep->queue.next, struct goku_request, queue);
drivers/usb/gadget/udc/goku_udc.c
836
list_for_each_entry(iter, &ep->queue, queue) {
drivers/usb/gadget/udc/goku_udc.c
847
if (ep->dma && ep->queue.next == &req->queue && !ep->stopped) {
drivers/usb/gadget/udc/goku_udc.c
851
} else if (!list_empty(&req->queue))
drivers/usb/gadget/udc/goku_udc.c
873
if (list_empty(&ep->queue))
drivers/usb/gadget/udc/goku_udc.c
875
req = list_entry(ep->queue.next, struct goku_request,
drivers/usb/gadget/udc/goku_udc.c
876
queue);
drivers/usb/gadget/udc/goku_udc.c
907
if (!list_empty(&ep->queue))
drivers/usb/gadget/udc/goku_udc.c
984
.queue = goku_queue,
drivers/usb/gadget/udc/goku_udc.h
215
struct list_head queue;
drivers/usb/gadget/udc/goku_udc.h
224
struct list_head queue;
drivers/usb/gadget/udc/gr_udc.c
1240
req = list_first_entry(&ep->queue, struct gr_request, queue);
drivers/usb/gadget/udc/gr_udc.c
1269
req = list_first_entry(&ep->queue, struct gr_request, queue);
drivers/usb/gadget/udc/gr_udc.c
1415
if (!ep->stopped && !ep->callback && !list_empty(&ep->queue))
drivers/usb/gadget/udc/gr_udc.c
1422
if (!ep->stopped && !ep->callback && !list_empty(&ep->queue))
drivers/usb/gadget/udc/gr_udc.c
156
if (list_empty(&ep->queue)) {
drivers/usb/gadget/udc/gr_udc.c
162
list_for_each_entry(req, &ep->queue, queue) {
drivers/usb/gadget/udc/gr_udc.c
1644
WARN(!list_empty(&req->queue),
drivers/usb/gadget/udc/gr_udc.c
1712
list_for_each_entry(iter, &ep->queue, queue) {
drivers/usb/gadget/udc/gr_udc.c
1723
if (list_first_entry(&ep->queue, struct gr_request, queue) == req) {
drivers/usb/gadget/udc/gr_udc.c
1730
} else if (!list_empty(&req->queue)) {
drivers/usb/gadget/udc/gr_udc.c
1756
if (halt && ep->is_in && !list_empty(&ep->queue)) {
drivers/usb/gadget/udc/gr_udc.c
1833
.queue = gr_queue_ext,
drivers/usb/gadget/udc/gr_udc.c
1980
INIT_LIST_HEAD(&ep->queue);
drivers/usb/gadget/udc/gr_udc.c
289
list_del_init(&req->queue);
drivers/usb/gadget/udc/gr_udc.c
354
INIT_LIST_HEAD(&req->queue);
drivers/usb/gadget/udc/gr_udc.c
369
if (list_empty(&ep->queue)) {
drivers/usb/gadget/udc/gr_udc.c
374
req = list_first_entry(&ep->queue, struct gr_request, queue);
drivers/usb/gadget/udc/gr_udc.c
410
req = list_first_entry(&ep->queue, struct gr_request, queue);
drivers/usb/gadget/udc/gr_udc.c
586
if (unlikely(!req->req.buf || !list_empty(&req->queue))) {
drivers/usb/gadget/udc/gr_udc.c
589
ep->ep.name, req->req.buf, list_empty(&req->queue));
drivers/usb/gadget/udc/gr_udc.c
620
list_add_tail(&req->queue, &ep->queue);
drivers/usb/gadget/udc/gr_udc.c
659
while (!list_empty(&ep->queue)) {
drivers/usb/gadget/udc/gr_udc.c
660
req = list_first_entry(&ep->queue, struct gr_request, queue);
drivers/usb/gadget/udc/gr_udc.h
152
struct list_head queue;
drivers/usb/gadget/udc/gr_udc.h
163
struct list_head queue;
drivers/usb/gadget/udc/lpc32xx_udc.c
100
struct list_head queue;
drivers/usb/gadget/udc/lpc32xx_udc.c
1009
req = list_entry(ep->queue.next, struct lpc32xx_request, queue);
drivers/usb/gadget/udc/lpc32xx_udc.c
1389
INIT_LIST_HEAD(&ep->queue);
drivers/usb/gadget/udc/lpc32xx_udc.c
1401
list_del_init(&req->queue);
drivers/usb/gadget/udc/lpc32xx_udc.c
1428
while (!list_empty(&ep->queue)) {
drivers/usb/gadget/udc/lpc32xx_udc.c
1429
req = list_entry(ep->queue.next, struct lpc32xx_request, queue);
drivers/usb/gadget/udc/lpc32xx_udc.c
1446
if (list_empty(&ep0->queue))
drivers/usb/gadget/udc/lpc32xx_udc.c
1450
req = list_entry(ep0->queue.next, struct lpc32xx_request,
drivers/usb/gadget/udc/lpc32xx_udc.c
1451
queue);
drivers/usb/gadget/udc/lpc32xx_udc.c
1484
if (list_empty(&ep0->queue))
drivers/usb/gadget/udc/lpc32xx_udc.c
1487
req = list_entry(ep0->queue.next, struct lpc32xx_request,
drivers/usb/gadget/udc/lpc32xx_udc.c
1488
queue);
drivers/usb/gadget/udc/lpc32xx_udc.c
1712
INIT_LIST_HEAD(&req->queue);
drivers/usb/gadget/udc/lpc32xx_udc.c
1726
BUG_ON(!list_empty(&req->queue));
drivers/usb/gadget/udc/lpc32xx_udc.c
1744
!list_empty(&req->queue))
drivers/usb/gadget/udc/lpc32xx_udc.c
176
struct list_head queue;
drivers/usb/gadget/udc/lpc32xx_udc.c
1797
if (list_empty(&ep->queue)) {
drivers/usb/gadget/udc/lpc32xx_udc.c
1798
list_add_tail(&req->queue, &ep->queue);
drivers/usb/gadget/udc/lpc32xx_udc.c
1820
list_add_tail(&req->queue, &ep->queue);
drivers/usb/gadget/udc/lpc32xx_udc.c
1841
list_for_each_entry(iter, &ep->queue, queue) {
drivers/usb/gadget/udc/lpc32xx_udc.c
1910
.queue = lpc32xx_ep_queue,
drivers/usb/gadget/udc/lpc32xx_udc.c
1962
req = list_entry(ep->queue.next, struct lpc32xx_request, queue);
drivers/usb/gadget/udc/lpc32xx_udc.c
1967
if (!list_empty(&ep->queue)) {
drivers/usb/gadget/udc/lpc32xx_udc.c
1988
req = list_entry(ep->queue.next, struct lpc32xx_request, queue);
drivers/usb/gadget/udc/lpc32xx_udc.c
2083
if (!list_empty((&ep->queue))) {
drivers/usb/gadget/udc/lpc32xx_udc.c
484
if (list_empty(&ep->queue))
drivers/usb/gadget/udc/lpc32xx_udc.c
487
list_for_each_entry(req, &ep->queue, queue) {
drivers/usb/gadget/udc/lpc32xx_udc.c
983
req = list_entry(ep->queue.next, struct lpc32xx_request, queue);
drivers/usb/gadget/udc/m66592-udc.c
1020
req = list_entry(ep->queue.next,
drivers/usb/gadget/udc/m66592-udc.c
1021
struct m66592_request, queue);
drivers/usb/gadget/udc/m66592-udc.c
1024
if (list_empty(&ep->queue))
drivers/usb/gadget/udc/m66592-udc.c
1027
} else if (!list_empty(&ep->queue))
drivers/usb/gadget/udc/m66592-udc.c
1169
req = list_entry(ep->queue.next, struct m66592_request, queue);
drivers/usb/gadget/udc/m66592-udc.c
1317
while (!list_empty(&ep->queue)) {
drivers/usb/gadget/udc/m66592-udc.c
1318
req = list_entry(ep->queue.next, struct m66592_request, queue);
drivers/usb/gadget/udc/m66592-udc.c
1337
INIT_LIST_HEAD(&req->queue);
drivers/usb/gadget/udc/m66592-udc.c
1366
if (list_empty(&ep->queue))
drivers/usb/gadget/udc/m66592-udc.c
1369
list_add_tail(&req->queue, &ep->queue);
drivers/usb/gadget/udc/m66592-udc.c
1395
if (!list_empty(&ep->queue))
drivers/usb/gadget/udc/m66592-udc.c
1409
if (!list_empty(&ep->queue)) {
drivers/usb/gadget/udc/m66592-udc.c
1429
if (list_empty(&ep->queue) && !ep->busy) {
drivers/usb/gadget/udc/m66592-udc.c
1443
.queue = m66592_queue,
drivers/usb/gadget/udc/m66592-udc.c
1624
INIT_LIST_HEAD(&ep->queue);
drivers/usb/gadget/udc/m66592-udc.c
363
INIT_LIST_HEAD(&ep->queue);
drivers/usb/gadget/udc/m66592-udc.c
719
list_del_init(&req->queue);
drivers/usb/gadget/udc/m66592-udc.c
725
if (!list_empty(&ep->queue))
drivers/usb/gadget/udc/m66592-udc.c
733
req = list_entry(ep->queue.next, struct m66592_request, queue);
drivers/usb/gadget/udc/m66592-udc.c
907
req = list_entry(ep->queue.next, struct m66592_request, queue);
drivers/usb/gadget/udc/m66592-udc.c
915
req = list_entry(ep->queue.next,
drivers/usb/gadget/udc/m66592-udc.c
916
struct m66592_request, queue);
drivers/usb/gadget/udc/m66592-udc.c
938
req = list_entry(ep->queue.next, struct m66592_request, queue);
drivers/usb/gadget/udc/m66592-udc.c
951
req = list_entry(ep->queue.next,
drivers/usb/gadget/udc/m66592-udc.c
953
queue);
drivers/usb/gadget/udc/m66592-udc.c
954
if (!list_empty(&ep->queue))
drivers/usb/gadget/udc/m66592-udc.c
99
INIT_LIST_HEAD(&m66592->ep[0].queue);
drivers/usb/gadget/udc/m66592-udc.h
438
struct list_head queue;
drivers/usb/gadget/udc/m66592-udc.h
445
struct list_head queue;
drivers/usb/gadget/udc/max3420_udc.c
1036
list_add_tail(&req->queue, &ep->queue);
drivers/usb/gadget/udc/max3420_udc.c
1054
list_for_each_entry(iter, &ep->queue, queue) {
drivers/usb/gadget/udc/max3420_udc.c
1057
list_del_init(&req->queue);
drivers/usb/gadget/udc/max3420_udc.c
1075
.queue = max3420_ep_queue,
drivers/usb/gadget/udc/max3420_udc.c
1159
INIT_LIST_HEAD(&ep->queue);
drivers/usb/gadget/udc/max3420_udc.c
1238
INIT_LIST_HEAD(&udc->ep0req.queue);
drivers/usb/gadget/udc/max3420_udc.c
151
struct list_head queue;
drivers/usb/gadget/udc/max3420_udc.c
158
struct list_head queue;
drivers/usb/gadget/udc/max3420_udc.c
691
if (list_empty(&ep->queue))
drivers/usb/gadget/udc/max3420_udc.c
694
req = list_first_entry(&ep->queue, struct max3420_req, queue);
drivers/usb/gadget/udc/max3420_udc.c
732
list_del_init(&req->queue);
drivers/usb/gadget/udc/max3420_udc.c
962
list_for_each_entry_safe(req, r, &ep->queue, queue) {
drivers/usb/gadget/udc/max3420_udc.c
963
list_del_init(&req->queue);
drivers/usb/gadget/udc/net2280.c
1032
if (list_empty(&ep->queue) && !ep->stopped &&
drivers/usb/gadget/udc/net2280.c
1104
list_add_tail(&req->queue, &ep->queue);
drivers/usb/gadget/udc/net2280.c
1131
while (!list_empty(&ep->queue)) {
drivers/usb/gadget/udc/net2280.c
1135
req = list_entry(ep->queue.next,
drivers/usb/gadget/udc/net2280.c
1136
struct net2280_request, queue);
drivers/usb/gadget/udc/net2280.c
1197
req = list_entry(ep->queue.next, struct net2280_request, queue);
drivers/usb/gadget/udc/net2280.c
1205
if (likely(!list_empty(&ep->queue))) {
drivers/usb/gadget/udc/net2280.c
1223
while (!list_empty(&ep->queue)) {
drivers/usb/gadget/udc/net2280.c
1224
req = list_entry(ep->queue.next,
drivers/usb/gadget/udc/net2280.c
1226
queue);
drivers/usb/gadget/udc/net2280.c
1262
list_for_each_entry(iter, &ep->queue, queue) {
drivers/usb/gadget/udc/net2280.c
1276
if (ep->queue.next == &req->queue) {
drivers/usb/gadget/udc/net2280.c
1281
if (likely(ep->queue.next == &req->queue)) {
drivers/usb/gadget/udc/net2280.c
1301
if (list_empty(&ep->queue))
drivers/usb/gadget/udc/net2280.c
1308
start_dma(ep, list_entry(ep->queue.next,
drivers/usb/gadget/udc/net2280.c
1309
struct net2280_request, queue));
drivers/usb/gadget/udc/net2280.c
1344
if (!list_empty(&ep->queue)) {
drivers/usb/gadget/udc/net2280.c
1365
!list_empty(&ep->queue) && ep->td_dma)
drivers/usb/gadget/udc/net2280.c
1450
.queue = net2280_queue,
drivers/usb/gadget/udc/net2280.c
1843
if (list_empty(&ep->queue)) {
drivers/usb/gadget/udc/net2280.c
1851
list_for_each_entry(req, &ep->queue, queue) {
drivers/usb/gadget/udc/net2280.c
2523
if (!list_empty(&ep->queue))
drivers/usb/gadget/udc/net2280.c
2524
req = list_entry(ep->queue.next,
drivers/usb/gadget/udc/net2280.c
2525
struct net2280_request, queue);
drivers/usb/gadget/udc/net2280.c
2619
if (unlikely(list_empty(&ep->queue) ||
drivers/usb/gadget/udc/net2280.c
2624
req = list_entry(ep->queue.next,
drivers/usb/gadget/udc/net2280.c
2625
struct net2280_request, queue);
drivers/usb/gadget/udc/net2280.c
2689
if (!list_empty(&ep->queue))
drivers/usb/gadget/udc/net2280.c
2735
if (!list_empty(&ep->queue) && !ep->stopped)
drivers/usb/gadget/udc/net2280.c
2736
req = list_entry(ep->queue.next,
drivers/usb/gadget/udc/net2280.c
2737
struct net2280_request, queue);
drivers/usb/gadget/udc/net2280.c
2966
if (!list_empty(&e->queue) && e->td_dma)
drivers/usb/gadget/udc/net2280.c
3137
while (!list_empty(&ep->queue)) {
drivers/usb/gadget/udc/net2280.c
3138
req = list_entry(ep->queue.next,
drivers/usb/gadget/udc/net2280.c
3139
struct net2280_request, queue);
drivers/usb/gadget/udc/net2280.c
3256
!list_empty(&e->queue) && e->td_dma)
drivers/usb/gadget/udc/net2280.c
3512
if (!list_empty(&ep->queue)) {
drivers/usb/gadget/udc/net2280.c
384
INIT_LIST_HEAD(&ep->queue);
drivers/usb/gadget/udc/net2280.c
460
INIT_LIST_HEAD(&ep->queue);
drivers/usb/gadget/udc/net2280.c
561
INIT_LIST_HEAD(&req->queue);
drivers/usb/gadget/udc/net2280.c
593
WARN_ON(!list_empty(&req->queue));
drivers/usb/gadget/udc/net2280.c
950
list_del_init(&req->queue);
drivers/usb/gadget/udc/net2280.c
995
!list_empty(&req->queue)) {
drivers/usb/gadget/udc/net2280.h
101
struct list_head queue;
drivers/usb/gadget/udc/net2280.h
143
struct list_head queue;
drivers/usb/gadget/udc/omap_udc.c
1017
list_for_each_entry(iter, &ep->queue, queue) {
drivers/usb/gadget/udc/omap_udc.c
1028
if (use_dma && ep->dma_channel && ep->queue.next == &req->queue) {
drivers/usb/gadget/udc/omap_udc.c
1072
&& !list_empty(&ep->queue)) {
drivers/usb/gadget/udc/omap_udc.c
1081
&& !list_empty(&ep->queue)) {
drivers/usb/gadget/udc/omap_udc.c
1122
.queue = omap_ep_queue,
drivers/usb/gadget/udc/omap_udc.c
1345
while (!list_empty(&ep->queue)) {
drivers/usb/gadget/udc/omap_udc.c
1346
req = list_entry(ep->queue.next, struct omap_req, queue);
drivers/usb/gadget/udc/omap_udc.c
1421
if (!list_empty(&ep0->queue))
drivers/usb/gadget/udc/omap_udc.c
1422
req = container_of(ep0->queue.next, struct omap_req, queue);
drivers/usb/gadget/udc/omap_udc.c
1868
if (!list_empty(&ep->queue) && ep->ackwait) {
drivers/usb/gadget/udc/omap_udc.c
1877
req = container_of(ep->queue.next,
drivers/usb/gadget/udc/omap_udc.c
1878
struct omap_req, queue);
drivers/usb/gadget/udc/omap_udc.c
1916
if (!list_empty(&ep->queue)) {
drivers/usb/gadget/udc/omap_udc.c
1918
req = container_of(ep->queue.next,
drivers/usb/gadget/udc/omap_udc.c
1919
struct omap_req, queue);
drivers/usb/gadget/udc/omap_udc.c
1951
if (!list_empty(&ep->queue)) {
drivers/usb/gadget/udc/omap_udc.c
1952
req = container_of(ep->queue.next,
drivers/usb/gadget/udc/omap_udc.c
1953
struct omap_req, queue);
drivers/usb/gadget/udc/omap_udc.c
1983
if (ep->has_dma || list_empty(&ep->queue))
drivers/usb/gadget/udc/omap_udc.c
1985
req = list_entry(ep->queue.next, struct omap_req, queue);
drivers/usb/gadget/udc/omap_udc.c
2017
if (!list_empty(&ep->queue))
drivers/usb/gadget/udc/omap_udc.c
2190
if (list_empty(&ep->queue))
drivers/usb/gadget/udc/omap_udc.c
2193
list_for_each_entry(req, &ep->queue, queue) {
drivers/usb/gadget/udc/omap_udc.c
2554
INIT_LIST_HEAD(&ep->queue);
drivers/usb/gadget/udc/omap_udc.c
274
INIT_LIST_HEAD(&req->queue);
drivers/usb/gadget/udc/omap_udc.c
295
list_del_init(&req->queue);
drivers/usb/gadget/udc/omap_udc.c
646
if (!list_empty(&ep->queue)) {
drivers/usb/gadget/udc/omap_udc.c
647
req = container_of(ep->queue.next,
drivers/usb/gadget/udc/omap_udc.c
648
struct omap_req, queue);
drivers/usb/gadget/udc/omap_udc.c
653
if (!list_empty(&ep->queue)) {
drivers/usb/gadget/udc/omap_udc.c
654
req = container_of(ep->queue.next,
drivers/usb/gadget/udc/omap_udc.c
655
struct omap_req, queue);
drivers/usb/gadget/udc/omap_udc.c
665
if (!list_empty(&ep->queue)) {
drivers/usb/gadget/udc/omap_udc.c
666
req = container_of(ep->queue.next,
drivers/usb/gadget/udc/omap_udc.c
667
struct omap_req, queue);
drivers/usb/gadget/udc/omap_udc.c
672
if (!list_empty(&ep->queue)) {
drivers/usb/gadget/udc/omap_udc.c
673
req = container_of(ep->queue.next,
drivers/usb/gadget/udc/omap_udc.c
674
struct omap_req, queue);
drivers/usb/gadget/udc/omap_udc.c
777
restart = !ep->stopped && !list_empty(&ep->queue);
drivers/usb/gadget/udc/omap_udc.c
790
req = container_of(ep->queue.next, struct omap_req, queue);
drivers/usb/gadget/udc/omap_udc.c
814
if (!list_empty(&ep->queue))
drivers/usb/gadget/udc/omap_udc.c
815
req = container_of(ep->queue.next, struct omap_req, queue);
drivers/usb/gadget/udc/omap_udc.c
875
|| !list_empty(&req->queue)) {
drivers/usb/gadget/udc/omap_udc.c
924
} else if (list_empty(&ep->queue) && !ep->stopped && !ep->ackwait) {
drivers/usb/gadget/udc/omap_udc.c
928
if (!udc->ep0_pending || !list_empty(&ep->queue)) {
drivers/usb/gadget/udc/omap_udc.c
999
list_add_tail(&req->queue, &ep->queue);
drivers/usb/gadget/udc/omap_udc.h
134
struct list_head queue;
drivers/usb/gadget/udc/omap_udc.h
141
struct list_head queue;
drivers/usb/gadget/udc/pch_udc.c
1429
list_del_init(&req->queue);
drivers/usb/gadget/udc/pch_udc.c
1457
while (!list_empty(&ep->queue)) {
drivers/usb/gadget/udc/pch_udc.c
1458
req = list_entry(ep->queue.next, struct pch_udc_request, queue);
drivers/usb/gadget/udc/pch_udc.c
1695
INIT_LIST_HEAD(&ep->queue);
drivers/usb/gadget/udc/pch_udc.c
1724
INIT_LIST_HEAD(&req->queue);
drivers/usb/gadget/udc/pch_udc.c
1761
if (!list_empty(&req->queue))
drivers/usb/gadget/udc/pch_udc.c
1800
if (!list_empty(&req->queue))
drivers/usb/gadget/udc/pch_udc.c
1817
if (list_empty(&ep->queue) && !ep->halted) {
drivers/usb/gadget/udc/pch_udc.c
1839
list_add_tail(&req->queue, &ep->queue);
drivers/usb/gadget/udc/pch_udc.c
1870
list_for_each_entry(req, &ep->queue, queue) {
drivers/usb/gadget/udc/pch_udc.c
1873
if (!list_empty(&req->queue))
drivers/usb/gadget/udc/pch_udc.c
1907
if (list_empty(&ep->queue)) {
drivers/usb/gadget/udc/pch_udc.c
1948
if (!list_empty(&ep->queue)) {
drivers/usb/gadget/udc/pch_udc.c
1984
.queue = pch_udc_pcd_queue,
drivers/usb/gadget/udc/pch_udc.c
2020
if (list_empty(&ep->queue))
drivers/usb/gadget/udc/pch_udc.c
2024
req = list_entry(ep->queue.next, struct pch_udc_request, queue);
drivers/usb/gadget/udc/pch_udc.c
2056
if (list_empty(&ep->queue))
drivers/usb/gadget/udc/pch_udc.c
2058
req = list_entry(ep->queue.next, struct pch_udc_request, queue);
drivers/usb/gadget/udc/pch_udc.c
2076
if (!list_empty(&ep->queue)) {
drivers/usb/gadget/udc/pch_udc.c
2099
if (list_empty(&ep->queue))
drivers/usb/gadget/udc/pch_udc.c
2102
req = list_entry(ep->queue.next, struct pch_udc_request, queue);
drivers/usb/gadget/udc/pch_udc.c
2141
if (!list_empty(&ep->queue)) {
drivers/usb/gadget/udc/pch_udc.c
2142
req = list_entry(ep->queue.next, struct pch_udc_request, queue);
drivers/usb/gadget/udc/pch_udc.c
2207
if ((epsts & UDC_EPSTS_BNA) && (!list_empty(&ep->queue))) {
drivers/usb/gadget/udc/pch_udc.c
2209
req = list_entry(ep->queue.next, struct pch_udc_request,
drivers/usb/gadget/udc/pch_udc.c
2210
queue);
drivers/usb/gadget/udc/pch_udc.c
2244
if (list_empty(&ep->queue))
drivers/usb/gadget/udc/pch_udc.c
2371
if (!list_empty(&ep->queue)) {
drivers/usb/gadget/udc/pch_udc.c
2390
if (list_empty(&ep->queue))
drivers/usb/gadget/udc/pch_udc.c
2810
INIT_LIST_HEAD(&ep->queue);
drivers/usb/gadget/udc/pch_udc.c
291
struct list_head queue;
drivers/usb/gadget/udc/pch_udc.c
394
struct list_head queue;
drivers/usb/gadget/udc/pxa25x_udc.c
1021
|| !list_empty(&ep->queue))) {
drivers/usb/gadget/udc/pxa25x_udc.c
1074
if (!_ep || ep->ep.name == ep0name || !list_empty(&ep->queue)) {
drivers/usb/gadget/udc/pxa25x_udc.c
1102
.queue = pxa25x_ep_queue,
drivers/usb/gadget/udc/pxa25x_udc.c
1320
if (list_empty(&ep->queue)) {
drivers/usb/gadget/udc/pxa25x_udc.c
1324
list_for_each_entry(req, &ep->queue, queue) {
drivers/usb/gadget/udc/pxa25x_udc.c
1397
INIT_LIST_HEAD (&ep->queue);
drivers/usb/gadget/udc/pxa25x_udc.c
1601
if (list_empty(&ep->queue))
drivers/usb/gadget/udc/pxa25x_udc.c
1604
req = list_entry(ep->queue.next, struct pxa25x_request, queue);
drivers/usb/gadget/udc/pxa25x_udc.c
1823
if (likely (!list_empty(&ep->queue)))
drivers/usb/gadget/udc/pxa25x_udc.c
1824
req = list_entry(ep->queue.next,
drivers/usb/gadget/udc/pxa25x_udc.c
1825
struct pxa25x_request, queue);
drivers/usb/gadget/udc/pxa25x_udc.c
515
INIT_LIST_HEAD (&req->queue);
drivers/usb/gadget/udc/pxa25x_udc.c
529
WARN_ON(!list_empty (&req->queue));
drivers/usb/gadget/udc/pxa25x_udc.c
542
list_del_init(&req->queue);
drivers/usb/gadget/udc/pxa25x_udc.c
632
if (list_empty(&ep->queue))
drivers/usb/gadget/udc/pxa25x_udc.c
780
if (list_empty(&ep->queue))
drivers/usb/gadget/udc/pxa25x_udc.c
845
|| !list_empty(&req->queue))) {
drivers/usb/gadget/udc/pxa25x_udc.c
879
if (list_empty(&ep->queue) && !ep->stopped) {
drivers/usb/gadget/udc/pxa25x_udc.c
935
list_add_tail(&req->queue, &ep->queue);
drivers/usb/gadget/udc/pxa25x_udc.c
950
while (!list_empty(&ep->queue)) {
drivers/usb/gadget/udc/pxa25x_udc.c
951
req = list_entry(ep->queue.next,
drivers/usb/gadget/udc/pxa25x_udc.c
953
queue);
drivers/usb/gadget/udc/pxa25x_udc.c
976
list_for_each_entry(iter, &ep->queue, queue) {
drivers/usb/gadget/udc/pxa25x_udc.h
39
struct list_head queue;
drivers/usb/gadget/udc/pxa25x_udc.h
61
struct list_head queue;
drivers/usb/gadget/udc/pxa27x_udc.c
1085
is_first_req = list_empty(&ep->queue);
drivers/usb/gadget/udc/pxa27x_udc.c
1176
list_for_each_entry(iter, &ep->queue, queue) {
drivers/usb/gadget/udc/pxa27x_udc.c
1226
if (ep->dir_in && (ep_is_full(ep) || !list_empty(&ep->queue)))
drivers/usb/gadget/udc/pxa27x_udc.c
1287
if (unlikely(!list_empty(&ep->queue)))
drivers/usb/gadget/udc/pxa27x_udc.c
1387
if (!ep || is_ep0(ep) || !list_empty(&ep->queue))
drivers/usb/gadget/udc/pxa27x_udc.c
1407
.queue = pxa_ep_queue,
drivers/usb/gadget/udc/pxa27x_udc.c
153
if (list_empty(&ep->queue)) {
drivers/usb/gadget/udc/pxa27x_udc.c
158
list_for_each_entry(req, &ep->queue, queue) {
drivers/usb/gadget/udc/pxa27x_udc.c
1674
INIT_LIST_HEAD(&ep->queue);
drivers/usb/gadget/udc/pxa27x_udc.c
1938
if (!list_empty(&ep->queue))
drivers/usb/gadget/udc/pxa27x_udc.c
1939
req = list_entry(ep->queue.next, struct pxa27x_request, queue);
drivers/usb/gadget/udc/pxa27x_udc.c
2034
if (likely(!list_empty(&ep->queue)))
drivers/usb/gadget/udc/pxa27x_udc.c
2035
req = list_entry(ep->queue.next,
drivers/usb/gadget/udc/pxa27x_udc.c
2036
struct pxa27x_request, queue);
drivers/usb/gadget/udc/pxa27x_udc.c
580
INIT_LIST_HEAD(&req->queue);
drivers/usb/gadget/udc/pxa27x_udc.c
599
WARN_ON(!list_empty(&req->queue));
drivers/usb/gadget/udc/pxa27x_udc.c
621
list_add_tail(&req->queue, &ep->queue);
drivers/usb/gadget/udc/pxa27x_udc.c
643
list_del_init(&req->queue);
drivers/usb/gadget/udc/pxa27x_udc.c
645
if (!is_ep0(ep) && list_empty(&ep->queue))
drivers/usb/gadget/udc/pxa27x_udc.c
772
while (!list_empty(&ep->queue)) {
drivers/usb/gadget/udc/pxa27x_udc.c
773
req = list_entry(ep->queue.next, struct pxa27x_request, queue);
drivers/usb/gadget/udc/pxa27x_udc.h
347
struct list_head queue;
drivers/usb/gadget/udc/pxa27x_udc.h
385
struct list_head queue;
drivers/usb/gadget/udc/r8a66597-udc.c
1151
if (!list_empty(&ep->queue))
drivers/usb/gadget/udc/r8a66597-udc.c
1226
if (list_empty(&ep->queue))
drivers/usb/gadget/udc/r8a66597-udc.c
1229
} else if (!list_empty(&ep->queue))
drivers/usb/gadget/udc/r8a66597-udc.c
1567
while (!list_empty(&ep->queue)) {
drivers/usb/gadget/udc/r8a66597-udc.c
1587
INIT_LIST_HEAD(&req->queue);
drivers/usb/gadget/udc/r8a66597-udc.c
1616
if (list_empty(&ep->queue))
drivers/usb/gadget/udc/r8a66597-udc.c
1619
list_add_tail(&req->queue, &ep->queue);
drivers/usb/gadget/udc/r8a66597-udc.c
1645
if (!list_empty(&ep->queue))
drivers/usb/gadget/udc/r8a66597-udc.c
1659
if (!list_empty(&ep->queue)) {
drivers/usb/gadget/udc/r8a66597-udc.c
1697
if (list_empty(&ep->queue) && !ep->busy) {
drivers/usb/gadget/udc/r8a66597-udc.c
1713
.queue = r8a66597_queue,
drivers/usb/gadget/udc/r8a66597-udc.c
1914
INIT_LIST_HEAD(&ep->queue);
drivers/usb/gadget/udc/r8a66597-udc.c
432
INIT_LIST_HEAD(&ep->queue);
drivers/usb/gadget/udc/r8a66597-udc.c
894
return list_entry(ep->queue.next, struct r8a66597_request, queue);
drivers/usb/gadget/udc/r8a66597-udc.c
912
list_del_init(&req->queue);
drivers/usb/gadget/udc/r8a66597-udc.c
918
if (!list_empty(&ep->queue))
drivers/usb/gadget/udc/r8a66597-udc.c
98
INIT_LIST_HEAD(&r8a66597->ep[0].queue);
drivers/usb/gadget/udc/r8a66597-udc.h
52
struct list_head queue;
drivers/usb/gadget/udc/r8a66597-udc.h
60
struct list_head queue;
drivers/usb/gadget/udc/renesas_usb3.c
1601
list_add_tail(&usb3_req->queue, &usb3_ep->queue);
drivers/usb/gadget/udc/renesas_usb3.c
2272
INIT_LIST_HEAD(&usb3_req->queue);
drivers/usb/gadget/udc/renesas_usb3.c
2349
.queue = renesas_usb3_ep_queue,
drivers/usb/gadget/udc/renesas_usb3.c
2705
INIT_LIST_HEAD(&usb3_ep->queue);
drivers/usb/gadget/udc/renesas_usb3.c
310
struct list_head queue;
drivers/usb/gadget/udc/renesas_usb3.c
320
struct list_head queue;
drivers/usb/gadget/udc/renesas_usb3.c
946
return list_first_entry_or_null(&usb3_ep->queue,
drivers/usb/gadget/udc/renesas_usb3.c
947
struct renesas_usb3_request, queue);
drivers/usb/gadget/udc/renesas_usb3.c
975
list_del_init(&usb3_req->queue);
drivers/usb/gadget/udc/renesas_usbf.c
1540
list_del_init(&req->queue);
drivers/usb/gadget/udc/renesas_usbf.c
1569
while (!list_empty(&ep->queue)) {
drivers/usb/gadget/udc/renesas_usbf.c
1570
req = list_first_entry(&ep->queue, struct usbf_req, queue);
drivers/usb/gadget/udc/renesas_usbf.c
1604
req = list_first_entry_or_null(&epn->queue, struct usbf_req, queue);
drivers/usb/gadget/udc/renesas_usbf.c
1664
req = list_first_entry_or_null(&ep->queue, struct usbf_req, queue);
drivers/usb/gadget/udc/renesas_usbf.c
1711
req = list_first_entry_or_null(&ep->queue, struct usbf_req,
drivers/usb/gadget/udc/renesas_usbf.c
1712
queue);
drivers/usb/gadget/udc/renesas_usbf.c
1764
first = list_first_entry_or_null(&ep->queue, struct usbf_req, queue);
drivers/usb/gadget/udc/renesas_usbf.c
1925
list_add_tail(&req->queue, &ep0->queue);
drivers/usb/gadget/udc/renesas_usbf.c
1974
was_empty = list_empty(&ep->queue);
drivers/usb/gadget/udc/renesas_usbf.c
1975
list_add_tail(&req->queue, &ep->queue);
drivers/usb/gadget/udc/renesas_usbf.c
2027
first = list_is_first(&req->queue, &ep->queue);
drivers/usb/gadget/udc/renesas_usbf.c
2088
INIT_LIST_HEAD(&req->queue);
drivers/usb/gadget/udc/renesas_usbf.c
2106
list_del_init(&req->queue);
drivers/usb/gadget/udc/renesas_usbf.c
2122
if (!list_empty(&ep->queue)) {
drivers/usb/gadget/udc/renesas_usbf.c
2148
if (!list_empty(&ep->queue)) {
drivers/usb/gadget/udc/renesas_usbf.c
2164
.queue = usbf_ep_queue,
drivers/usb/gadget/udc/renesas_usbf.c
2561
req = list_first_entry_or_null(&ep0->queue, struct usbf_req, queue);
drivers/usb/gadget/udc/renesas_usbf.c
2588
req = list_first_entry_or_null(&ep0->queue, struct usbf_req, queue);
drivers/usb/gadget/udc/renesas_usbf.c
260
struct list_head queue;
drivers/usb/gadget/udc/renesas_usbf.c
2602
req = list_first_entry_or_null(&ep0->queue, struct usbf_req, queue);
drivers/usb/gadget/udc/renesas_usbf.c
278
struct list_head queue;
drivers/usb/gadget/udc/renesas_usbf.c
3298
INIT_LIST_HEAD(&udc->setup_reply.queue);
drivers/usb/gadget/udc/renesas_usbf.c
3308
INIT_LIST_HEAD(&ep->queue);
drivers/usb/gadget/udc/snps_udc_core.c
1062
|| !list_empty(&req->queue))
drivers/usb/gadget/udc/snps_udc_core.c
1093
if (list_empty(&ep->queue)) {
drivers/usb/gadget/udc/snps_udc_core.c
1200
list_add_tail(&req->queue, &ep->queue);
drivers/usb/gadget/udc/snps_udc_core.c
1240
while (!list_empty(&ep->queue)) {
drivers/usb/gadget/udc/snps_udc_core.c
1241
req = list_entry(ep->queue.next,
drivers/usb/gadget/udc/snps_udc_core.c
1243
queue);
drivers/usb/gadget/udc/snps_udc_core.c
1268
if (ep->queue.next == &req->queue) {
drivers/usb/gadget/udc/snps_udc_core.c
1374
.queue = udc_queue,
drivers/usb/gadget/udc/snps_udc_core.c
2095
if (!list_empty(&ep->queue)) {
drivers/usb/gadget/udc/snps_udc_core.c
2098
req = list_entry(ep->queue.next,
drivers/usb/gadget/udc/snps_udc_core.c
2099
struct udc_request, queue);
drivers/usb/gadget/udc/snps_udc_core.c
2115
if (!list_empty(&ep->queue) && !ep->halted) {
drivers/usb/gadget/udc/snps_udc_core.c
2116
req = list_entry(ep->queue.next,
drivers/usb/gadget/udc/snps_udc_core.c
2117
struct udc_request, queue);
drivers/usb/gadget/udc/snps_udc_core.c
2192
if (!list_empty(&ep->queue) && !ep->halted) {
drivers/usb/gadget/udc/snps_udc_core.c
2193
req = list_entry(ep->queue.next,
drivers/usb/gadget/udc/snps_udc_core.c
2195
queue);
drivers/usb/gadget/udc/snps_udc_core.c
2313
if (!ep->cancel_transfer && !list_empty(&ep->queue)) {
drivers/usb/gadget/udc/snps_udc_core.c
2314
req = list_entry(ep->queue.next,
drivers/usb/gadget/udc/snps_udc_core.c
2315
struct udc_request, queue);
drivers/usb/gadget/udc/snps_udc_core.c
2334
if (list_empty(&ep->queue)) {
drivers/usb/gadget/udc/snps_udc_core.c
2352
if (!list_empty(&ep->queue)) {
drivers/usb/gadget/udc/snps_udc_core.c
2354
req = list_entry(ep->queue.next,
drivers/usb/gadget/udc/snps_udc_core.c
2355
struct udc_request, queue);
drivers/usb/gadget/udc/snps_udc_core.c
2586
if (list_empty(&dev->ep[UDC_EP0OUT_IX].queue)) {
drivers/usb/gadget/udc/snps_udc_core.c
2678
if (!list_empty(&ep->queue)) {
drivers/usb/gadget/udc/snps_udc_core.c
2680
req = list_entry(ep->queue.next,
drivers/usb/gadget/udc/snps_udc_core.c
2681
struct udc_request, queue);
drivers/usb/gadget/udc/snps_udc_core.c
454
INIT_LIST_HEAD(&ep->queue);
drivers/usb/gadget/udc/snps_udc_core.c
531
INIT_LIST_HEAD(&req->queue);
drivers/usb/gadget/udc/snps_udc_core.c
592
BUG_ON(!list_empty(&req->queue));
drivers/usb/gadget/udc/snps_udc_core.c
982
list_del_init(&req->queue);
drivers/usb/gadget/udc/tegra-xudc.c
1065
while (!list_empty(&ep->queue)) {
drivers/usb/gadget/udc/tegra-xudc.c
1066
req = list_first_entry(&ep->queue, struct tegra_xudc_request,
drivers/usb/gadget/udc/tegra-xudc.c
1218
if (list_empty(&ep->queue))
drivers/usb/gadget/udc/tegra-xudc.c
1231
req = list_first_entry(&ep->queue, struct tegra_xudc_request,
drivers/usb/gadget/udc/tegra-xudc.c
1245
list_for_each_entry(req, &ep->queue, list) {
drivers/usb/gadget/udc/tegra-xudc.c
1263
if (usb_endpoint_xfer_control(ep->desc) && !list_empty(&ep->queue)) {
drivers/usb/gadget/udc/tegra-xudc.c
1302
list_add_tail(&req->list, &ep->queue);
drivers/usb/gadget/udc/tegra-xudc.c
1368
list_for_each_entry_continue(req, &ep->queue, list) {
drivers/usb/gadget/udc/tegra-xudc.c
1435
list_for_each_entry(iter, &ep->queue, list) {
drivers/usb/gadget/udc/tegra-xudc.c
1604
!list_empty(&ep->queue)) {
drivers/usb/gadget/udc/tegra-xudc.c
1931
.queue = tegra_xudc_ep_queue,
drivers/usb/gadget/udc/tegra-xudc.c
1952
.queue = tegra_xudc_ep_queue,
drivers/usb/gadget/udc/tegra-xudc.c
2641
list_for_each_entry(req, &ep->queue, list) {
drivers/usb/gadget/udc/tegra-xudc.c
3160
INIT_LIST_HEAD(&ep->queue);
drivers/usb/gadget/udc/tegra-xudc.c
413
struct list_head queue;
drivers/usb/gadget/udc/udc-xilinx.c
1011
if (!list_empty(&ep0->queue)) {
drivers/usb/gadget/udc/udc-xilinx.c
1019
list_add_tail(&req->queue, &ep0->queue);
drivers/usb/gadget/udc/udc-xilinx.c
1112
if (list_empty(&ep->queue)) {
drivers/usb/gadget/udc/udc-xilinx.c
1125
list_add_tail(&req->queue, &ep->queue);
drivers/usb/gadget/udc/udc-xilinx.c
1148
list_for_each_entry(iter, &ep->queue, queue) {
drivers/usb/gadget/udc/udc-xilinx.c
116
struct list_head queue;
drivers/usb/gadget/udc/udc-xilinx.c
1197
.queue = xudc_ep0_queue,
drivers/usb/gadget/udc/udc-xilinx.c
1207
.queue = xudc_ep_queue,
drivers/usb/gadget/udc/udc-xilinx.c
1348
INIT_LIST_HEAD(&ep->queue);
drivers/usb/gadget/udc/udc-xilinx.c
141
struct list_head queue;
drivers/usb/gadget/udc/udc-xilinx.c
1825
req = list_first_entry(&ep0->queue, struct xusb_req, queue);
drivers/usb/gadget/udc/udc-xilinx.c
1878
req = list_first_entry(&ep0->queue, struct xusb_req, queue);
drivers/usb/gadget/udc/udc-xilinx.c
1975
if (list_empty(&ep->queue))
drivers/usb/gadget/udc/udc-xilinx.c
1978
req = list_first_entry(&ep->queue, struct xusb_req, queue);
drivers/usb/gadget/udc/udc-xilinx.c
555
list_del_init(&req->queue);
drivers/usb/gadget/udc/udc-xilinx.c
738
while (!list_empty(&ep->queue)) {
drivers/usb/gadget/udc/udc-xilinx.c
739
req = list_first_entry(&ep->queue, struct xusb_req, queue);
drivers/usb/gadget/udc/udc-xilinx.c
764
if (ep->is_in && (!list_empty(&ep->queue)) && value) {
drivers/usb/gadget/udc/udc-xilinx.c
978
INIT_LIST_HEAD(&req->queue);
drivers/usb/host/r8a66597-hcd.c
1281
list_del(&td->queue);
drivers/usb/host/r8a66597-hcd.c
1779
list_move_tail(&new_td->queue,
drivers/usb/host/r8a66597-hcd.c
1878
INIT_LIST_HEAD(&td->queue);
drivers/usb/host/r8a66597-hcd.c
1925
list_add_tail(&td->queue, &r8a66597->pipe_queue[td->pipenum]);
drivers/usb/host/r8a66597-hcd.c
1937
list_del(&td->queue);
drivers/usb/host/r8a66597-hcd.c
842
list_for_each_entry_safe(td, next, list, queue) {
drivers/usb/host/r8a66597-hcd.c
847
list_del(&td->queue);
drivers/usb/host/r8a66597.h
148
struct r8a66597_td, queue);
drivers/usb/host/r8a66597.h
58
struct list_head queue;
drivers/usb/host/uhci-debug.c
211
if (list_empty(&qh->queue)) {
drivers/usb/host/uhci-debug.c
220
struct urb_priv *urbp = list_entry(qh->queue.next,
drivers/usb/host/uhci-debug.c
229
list_for_each_entry(urbp, &qh->queue, node) {
drivers/usb/host/uhci-hcd.h
162
struct list_head queue; /* Queue of urbps for this QH */
drivers/usb/host/uhci-q.c
1292
if (list_empty(&qh->queue)) {
drivers/usb/host/uhci-q.c
1297
lurb = list_entry(qh->queue.prev,
drivers/usb/host/uhci-q.c
1358
if (list_empty(&qh->queue)) {
drivers/usb/host/uhci-q.c
1456
list_add_tail(&urbp->node, &qh->queue);
drivers/usb/host/uhci-q.c
1462
if (qh->queue.next == &urbp->node && !qh->is_stopped) {
drivers/usb/host/uhci-q.c
1534
urbp->node.prev == &qh->queue &&
drivers/usb/host/uhci-q.c
1535
urbp->node.next != &qh->queue) {
drivers/usb/host/uhci-q.c
1546
if (list_empty(&qh->queue) && qh->needs_fixup) {
drivers/usb/host/uhci-q.c
1561
if (list_empty(&qh->queue)) {
drivers/usb/host/uhci-q.c
1581
while (!list_empty(&qh->queue)) {
drivers/usb/host/uhci-q.c
1582
urbp = list_entry(qh->queue.next, struct urb_priv, node);
drivers/usb/host/uhci-q.c
1615
list_for_each_entry(urbp, &qh->queue, node) {
drivers/usb/host/uhci-q.c
1634
if (!list_empty(&qh->queue)) {
drivers/usb/host/uhci-q.c
1641
urbp = list_entry(qh->queue.next, struct urb_priv, node);
drivers/usb/host/uhci-q.c
1691
urbp = list_entry(qh->queue.next, struct urb_priv, node);
drivers/usb/host/uhci-q.c
1772
list_entry(qh->queue.next, struct urb_priv, node));
drivers/usb/host/uhci-q.c
260
INIT_LIST_HEAD(&qh->queue);
drivers/usb/host/uhci-q.c
295
if (!list_empty(&qh->queue))
drivers/usb/host/uhci-q.c
334
if (qh->queue.next != &urbp->node) {
drivers/usb/host/uhci-q.c
383
urbp = list_entry(qh->queue.next, struct urb_priv, node);
drivers/usb/host/uhci-q.c
393
urbp = list_prepare_entry(urbp, &qh->queue, node);
drivers/usb/host/uhci-q.c
394
list_for_each_entry_continue(urbp, &qh->queue, node) {
drivers/usb/host/uhci-q.c
415
pipe = list_entry(qh->queue.next, struct urb_priv, node)->urb->pipe;
drivers/usb/host/uhci-q.c
482
WARN_ON(list_empty(&qh->queue));
drivers/usb/host/uhci-q.c
487
struct urb_priv *urbp = list_entry(qh->queue.next,
drivers/usb/host/xhci-dbgtty.c
410
struct list_head *queue = &port->read_queue;
drivers/usb/host/xhci-dbgtty.c
415
while (!list_empty(queue)) {
drivers/usb/host/xhci-dbgtty.c
416
req = list_first_entry(queue, struct dbc_request, list_pool);
drivers/usb/host/xhci-dbgtty.c
440
if (!list_empty(queue))
drivers/usb/isp1760/isp1760-udc.c
1029
if (!list_empty(&uep->queue)) {
drivers/usb/isp1760/isp1760-udc.c
1115
.queue = isp1760_ep_queue,
drivers/usb/isp1760/isp1760-udc.c
1457
INIT_LIST_HEAD(&ep->queue);
drivers/usb/isp1760/isp1760-udc.c
293
list_del(&req->queue);
drivers/usb/isp1760/isp1760-udc.c
30
struct list_head queue;
drivers/usb/isp1760/isp1760-udc.c
358
if (list_empty(&ep->queue)) {
drivers/usb/isp1760/isp1760-udc.c
366
req = list_first_entry(&ep->queue, struct isp1760_request,
drivers/usb/isp1760/isp1760-udc.c
367
queue);
drivers/usb/isp1760/isp1760-udc.c
392
if (list_empty(&ep->queue)) {
drivers/usb/isp1760/isp1760-udc.c
410
req = list_first_entry(&ep->queue, struct isp1760_request,
drivers/usb/isp1760/isp1760-udc.c
411
queue);
drivers/usb/isp1760/isp1760-udc.c
429
list_del(&req->queue);
drivers/usb/isp1760/isp1760-udc.c
434
if (!list_empty(&ep->queue))
drivers/usb/isp1760/isp1760-udc.c
435
req = list_first_entry(&ep->queue,
drivers/usb/isp1760/isp1760-udc.c
436
struct isp1760_request, queue);
drivers/usb/isp1760/isp1760-udc.c
495
if ((ep->addr & USB_DIR_IN) && !list_empty(&ep->queue)) {
drivers/usb/isp1760/isp1760-udc.c
498
req = list_first_entry(&ep->queue,
drivers/usb/isp1760/isp1760-udc.c
499
struct isp1760_request, queue);
drivers/usb/isp1760/isp1760-udc.c
864
list_splice_init(&uep->queue, &req_list);
drivers/usb/isp1760/isp1760-udc.c
868
list_for_each_entry_safe(req, nreq, &req_list, queue) {
drivers/usb/isp1760/isp1760-udc.c
869
list_del(&req->queue);
drivers/usb/isp1760/isp1760-udc.c
931
list_add_tail(&req->queue, &uep->queue);
drivers/usb/isp1760/isp1760-udc.c
936
list_add_tail(&req->queue, &uep->queue);
drivers/usb/isp1760/isp1760-udc.c
952
bool empty = list_empty(&uep->queue);
drivers/usb/isp1760/isp1760-udc.c
954
list_add_tail(&req->queue, &uep->queue);
drivers/usb/isp1760/isp1760-udc.c
992
list_del(&req->queue);
drivers/usb/isp1760/isp1760-udc.h
38
struct list_head queue;
drivers/usb/mtu3/mtu3_gadget.c
423
.queue = mtu3_gadget_queue,
drivers/usb/mtu3/mtu3_gadget_ep0.c
913
.queue = mtu3_ep0_queue,
drivers/usb/musb/musb_gadget.c
1496
.queue = musb_gadget_queue,
drivers/usb/musb/musb_gadget.h
103
struct list_head *queue = &ep->req_list;
drivers/usb/musb/musb_gadget.h
105
if (list_empty(queue))
drivers/usb/musb/musb_gadget.h
107
return container_of(queue->next, struct musb_request, list);
drivers/usb/musb/musb_gadget_ep0.c
1055
.queue = musb_g_ep0_queue,
drivers/usb/musb/musb_host.h
114
struct list_head *queue;
drivers/usb/musb/musb_host.h
118
queue = &qh->hep->urb_list;
drivers/usb/musb/musb_host.h
119
if (list_empty(queue))
drivers/usb/musb/musb_host.h
121
return list_entry(queue->next, struct urb, urb_list);
drivers/usb/renesas_usbhs/mod_gadget.c
785
.queue = usbhsg_ep_queue,
drivers/usb/usbip/vudc_dev.c
419
.queue = vep_queue,
drivers/vhost/net.c
107
void **queue;
drivers/vhost/net.c
1325
void **queue;
drivers/vhost/net.c
1338
queue = kmalloc_array(VHOST_NET_BATCH, sizeof(void *),
drivers/vhost/net.c
1340
if (!queue) {
drivers/vhost/net.c
1345
n->vqs[VHOST_NET_VQ_RX].rxq.queue = queue;
drivers/vhost/net.c
1351
kfree(queue);
drivers/vhost/net.c
1448
kfree(n->vqs[VHOST_NET_VQ_RX].rxq.queue);
drivers/vhost/net.c
157
return rxq->queue[rxq->head];
drivers/vhost/net.c
184
rxq->tail = ptr_ring_consume_batched(nvq->rx_ring, rxq->queue,
drivers/vhost/net.c
194
ptr_ring_unconsume(nvq->rx_ring, rxq->queue + rxq->head,
drivers/virtio/virtio_ring.c
1252
void *queue = NULL;
drivers/virtio/virtio_ring.c
1263
queue = vring_alloc_queue(vdev, vring_size(num, vring_align),
drivers/virtio/virtio_ring.c
1267
if (queue)
drivers/virtio/virtio_ring.c
1276
if (!queue) {
drivers/virtio/virtio_ring.c
1278
queue = vring_alloc_queue(vdev, vring_size(num, vring_align),
drivers/virtio/virtio_ring.c
1282
if (!queue)
drivers/virtio/virtio_ring.c
1285
vring_init(&vring_split->vring, num, queue, vring_align);
drivers/virtio/virtio_ring.c
384
void *queue = alloc_pages_exact(PAGE_ALIGN(size), flag);
drivers/virtio/virtio_ring.c
386
if (queue) {
drivers/virtio/virtio_ring.c
387
phys_addr_t phys_addr = virt_to_phys(queue);
drivers/virtio/virtio_ring.c
402
free_pages_exact(queue, PAGE_ALIGN(size));
drivers/virtio/virtio_ring.c
406
return queue;
drivers/virtio/virtio_ring.c
411
void *queue, dma_addr_t map_handle,
drivers/virtio/virtio_ring.c
416
queue, map_handle);
drivers/virtio/virtio_ring.c
418
free_pages_exact(queue, PAGE_ALIGN(size));
drivers/watchdog/mtx-1_wdt.c
103
if (mtx1_wdt_device.queue) {
drivers/watchdog/mtx-1_wdt.c
104
mtx1_wdt_device.queue = 0;
drivers/watchdog/mtx-1_wdt.c
208
mtx1_wdt_device.queue = 0;
drivers/watchdog/mtx-1_wdt.c
226
if (mtx1_wdt_device.queue) {
drivers/watchdog/mtx-1_wdt.c
227
mtx1_wdt_device.queue = 0;
drivers/watchdog/mtx-1_wdt.c
53
int queue;
drivers/watchdog/mtx-1_wdt.c
70
if (mtx1_wdt_device.queue && ticks)
drivers/watchdog/mtx-1_wdt.c
88
if (!mtx1_wdt_device.queue) {
drivers/watchdog/mtx-1_wdt.c
89
mtx1_wdt_device.queue = 1;
drivers/watchdog/rdc321x_wdt.c
233
rdc321x_wdt_device.queue = 0;
drivers/watchdog/rdc321x_wdt.c
261
if (rdc321x_wdt_device.queue) {
drivers/watchdog/rdc321x_wdt.c
262
rdc321x_wdt_device.queue = 0;
drivers/watchdog/rdc321x_wdt.c
46
int queue;
drivers/watchdog/rdc321x_wdt.c
74
if (rdc321x_wdt_device.queue && ticks)
drivers/watchdog/rdc321x_wdt.c
93
if (!rdc321x_wdt_device.queue) {
drivers/watchdog/rdc321x_wdt.c
94
rdc321x_wdt_device.queue = 1;
drivers/xen/pvcalls-back.c
769
struct request_sock_queue *queue;
drivers/xen/pvcalls-back.c
795
queue = &icsk->icsk_accept_queue;
drivers/xen/pvcalls-back.c
796
data = READ_ONCE(queue->rskq_accept_head) != NULL;
drivers/xen/xenbus/xenbus_dev_frontend.c
189
static int queue_reply(struct list_head *queue, const void *data, size_t len)
drivers/xen/xenbus/xenbus_dev_frontend.c
207
list_add_tail(&rb->list, queue);
fs/autofs/autofs_i.h
93
wait_queue_head_t queue;
fs/autofs/waitq.c
35
wake_up(&wq->queue);
fs/autofs/waitq.c
391
init_waitqueue_head(&wq->queue);
fs/autofs/waitq.c
440
wait_event_killable(wq->queue, wq->name.name == NULL);
fs/autofs/waitq.c
507
wake_up(&wq->queue);
fs/dlm/recover.c
405
static void set_lock_master(struct list_head *queue, int nodeid)
fs/dlm/recover.c
409
list_for_each_entry(lkb, queue, lkb_statequeue) {
fs/fuse/dev_uring.c
1037
struct fuse_ring_queue *queue)
fs/fuse/dev_uring.c
1039
struct fuse_ring *ring = queue->ring;
fs/fuse/dev_uring.c
107
static void fuse_uring_abort_end_queue_requests(struct fuse_ring_queue *queue)
fs/fuse/dev_uring.c
1072
ent->queue = queue;
fs/fuse/dev_uring.c
1090
struct fuse_ring_queue *queue;
fs/fuse/dev_uring.c
1107
queue = ring->queues[qid];
fs/fuse/dev_uring.c
1108
if (!queue) {
fs/fuse/dev_uring.c
1109
queue = fuse_uring_create_queue(ring, qid);
fs/fuse/dev_uring.c
1110
if (!queue)
fs/fuse/dev_uring.c
1119
ent = fuse_uring_create_ring_ent(cmd, queue);
fs/fuse/dev_uring.c
112
spin_lock(&queue->lock);
fs/fuse/dev_uring.c
113
list_for_each_entry(req, &queue->fuse_req_queue, list)
fs/fuse/dev_uring.c
115
list_splice_init(&queue->fuse_req_queue, &req_list);
fs/fuse/dev_uring.c
116
spin_unlock(&queue->lock);
fs/fuse/dev_uring.c
1202
struct fuse_ring_queue *queue = ent->queue;
fs/fuse/dev_uring.c
1204
spin_lock(&queue->lock);
fs/fuse/dev_uring.c
1206
list_move_tail(&ent->list, &queue->ent_in_userspace);
fs/fuse/dev_uring.c
1208
spin_unlock(&queue->lock);
fs/fuse/dev_uring.c
1223
struct fuse_ring_queue *queue = ent->queue;
fs/fuse/dev_uring.c
1229
fuse_uring_next_fuse_req(ent, queue, issue_flags);
fs/fuse/dev_uring.c
1242
struct fuse_ring_queue *queue;
fs/fuse/dev_uring.c
125
struct fuse_ring_queue *queue;
fs/fuse/dev_uring.c
1251
queue = ring->queues[qid];
fs/fuse/dev_uring.c
1252
WARN_ONCE(!queue, "Missing queue for qid %d\n", qid);
fs/fuse/dev_uring.c
1254
return queue;
fs/fuse/dev_uring.c
1270
struct fuse_ring_queue *queue;
fs/fuse/dev_uring.c
1275
queue = fuse_uring_task_to_queue(ring);
fs/fuse/dev_uring.c
1276
if (!queue)
fs/fuse/dev_uring.c
1281
spin_lock(&queue->lock);
fs/fuse/dev_uring.c
1283
if (unlikely(queue->stopped))
fs/fuse/dev_uring.c
1287
req->ring_queue = queue;
fs/fuse/dev_uring.c
1288
ent = list_first_entry_or_null(&queue->ent_avail_queue,
fs/fuse/dev_uring.c
129
queue = READ_ONCE(ring->queues[qid]);
fs/fuse/dev_uring.c
1293
list_add_tail(&req->list, &queue->fuse_req_queue);
fs/fuse/dev_uring.c
1294
spin_unlock(&queue->lock);
fs/fuse/dev_uring.c
130
if (!queue)
fs/fuse/dev_uring.c
1302
spin_unlock(&queue->lock);
fs/fuse/dev_uring.c
1313
struct fuse_ring_queue *queue;
fs/fuse/dev_uring.c
1316
queue = fuse_uring_task_to_queue(ring);
fs/fuse/dev_uring.c
1317
if (!queue)
fs/fuse/dev_uring.c
1320
spin_lock(&queue->lock);
fs/fuse/dev_uring.c
1321
if (unlikely(queue->stopped)) {
fs/fuse/dev_uring.c
1322
spin_unlock(&queue->lock);
fs/fuse/dev_uring.c
1327
req->ring_queue = queue;
fs/fuse/dev_uring.c
1328
list_add_tail(&req->list, &queue->fuse_req_bg_queue);
fs/fuse/dev_uring.c
133
queue->stopped = true;
fs/fuse/dev_uring.c
1330
ent = list_first_entry_or_null(&queue->ent_avail_queue,
fs/fuse/dev_uring.c
1336
fuse_uring_flush_bg(queue);
fs/fuse/dev_uring.c
1344
req = list_first_entry_or_null(&queue->fuse_req_queue, struct fuse_req,
fs/fuse/dev_uring.c
1348
spin_unlock(&queue->lock);
fs/fuse/dev_uring.c
1352
spin_unlock(&queue->lock);
fs/fuse/dev_uring.c
136
spin_lock(&queue->lock);
fs/fuse/dev_uring.c
1360
struct fuse_ring_queue *queue = req->ring_queue;
fs/fuse/dev_uring.c
1362
return fuse_remove_pending_req(req, &queue->lock);
fs/fuse/dev_uring.c
138
fuse_uring_flush_bg(queue);
fs/fuse/dev_uring.c
140
spin_unlock(&queue->lock);
fs/fuse/dev_uring.c
141
fuse_uring_abort_end_queue_requests(queue);
fs/fuse/dev_uring.c
163
struct fuse_ring_queue *queue;
fs/fuse/dev_uring.c
170
queue = READ_ONCE(ring->queues[qid]);
fs/fuse/dev_uring.c
171
if (!queue)
fs/fuse/dev_uring.c
174
spin_lock(&queue->lock);
fs/fuse/dev_uring.c
175
if (fuse_request_expired(fc, &queue->fuse_req_queue) ||
fs/fuse/dev_uring.c
176
fuse_request_expired(fc, &queue->fuse_req_bg_queue) ||
fs/fuse/dev_uring.c
177
ent_list_request_expired(fc, &queue->ent_w_req_queue) ||
fs/fuse/dev_uring.c
178
ent_list_request_expired(fc, &queue->ent_in_userspace)) {
fs/fuse/dev_uring.c
179
spin_unlock(&queue->lock);
fs/fuse/dev_uring.c
182
spin_unlock(&queue->lock);
fs/fuse/dev_uring.c
197
struct fuse_ring_queue *queue = ring->queues[qid];
fs/fuse/dev_uring.c
200
if (!queue)
fs/fuse/dev_uring.c
203
WARN_ON(!list_empty(&queue->ent_avail_queue));
fs/fuse/dev_uring.c
204
WARN_ON(!list_empty(&queue->ent_w_req_queue));
fs/fuse/dev_uring.c
205
WARN_ON(!list_empty(&queue->ent_commit_queue));
fs/fuse/dev_uring.c
206
WARN_ON(!list_empty(&queue->ent_in_userspace));
fs/fuse/dev_uring.c
208
list_for_each_entry_safe(ent, next, &queue->ent_released,
fs/fuse/dev_uring.c
214
kfree(queue->fpq.processing);
fs/fuse/dev_uring.c
215
kfree(queue);
fs/fuse/dev_uring.c
274
struct fuse_ring_queue *queue;
fs/fuse/dev_uring.c
277
queue = kzalloc_obj(*queue, GFP_KERNEL_ACCOUNT);
fs/fuse/dev_uring.c
278
if (!queue)
fs/fuse/dev_uring.c
282
kfree(queue);
fs/fuse/dev_uring.c
286
queue->qid = qid;
fs/fuse/dev_uring.c
287
queue->ring = ring;
fs/fuse/dev_uring.c
288
spin_lock_init(&queue->lock);
fs/fuse/dev_uring.c
290
INIT_LIST_HEAD(&queue->ent_avail_queue);
fs/fuse/dev_uring.c
291
INIT_LIST_HEAD(&queue->ent_commit_queue);
fs/fuse/dev_uring.c
292
INIT_LIST_HEAD(&queue->ent_w_req_queue);
fs/fuse/dev_uring.c
293
INIT_LIST_HEAD(&queue->ent_in_userspace);
fs/fuse/dev_uring.c
294
INIT_LIST_HEAD(&queue->fuse_req_queue);
fs/fuse/dev_uring.c
295
INIT_LIST_HEAD(&queue->fuse_req_bg_queue);
fs/fuse/dev_uring.c
296
INIT_LIST_HEAD(&queue->ent_released);
fs/fuse/dev_uring.c
298
queue->fpq.processing = pq;
fs/fuse/dev_uring.c
299
fuse_pqueue_init(&queue->fpq);
fs/fuse/dev_uring.c
304
kfree(queue->fpq.processing);
fs/fuse/dev_uring.c
305
kfree(queue);
fs/fuse/dev_uring.c
312
WRITE_ONCE(ring->queues[qid], queue);
fs/fuse/dev_uring.c
315
return queue;
fs/fuse/dev_uring.c
333
struct fuse_ring_queue *queue = ent->queue;
fs/fuse/dev_uring.c
335
spin_lock(&queue->lock);
fs/fuse/dev_uring.c
351
list_move(&ent->list, &queue->ent_released);
fs/fuse/dev_uring.c
353
spin_unlock(&queue->lock);
fs/fuse/dev_uring.c
363
struct fuse_ring_queue *queue,
fs/fuse/dev_uring.c
366
struct fuse_ring *ring = queue->ring;
fs/fuse/dev_uring.c
371
spin_lock(&queue->lock);
fs/fuse/dev_uring.c
375
queue->qid, ent->state, exp_state);
fs/fuse/dev_uring.c
382
spin_unlock(&queue->lock);
fs/fuse/dev_uring.c
392
static void fuse_uring_teardown_entries(struct fuse_ring_queue *queue)
fs/fuse/dev_uring.c
394
fuse_uring_stop_list_entries(&queue->ent_in_userspace, queue,
fs/fuse/dev_uring.c
396
fuse_uring_stop_list_entries(&queue->ent_avail_queue, queue,
fs/fuse/dev_uring.c
409
struct fuse_ring_queue *queue = ring->queues[qid];
fs/fuse/dev_uring.c
411
if (!queue)
fs/fuse/dev_uring.c
414
spin_lock(&queue->lock);
fs/fuse/dev_uring.c
419
list_for_each_entry(ent, &queue->ent_w_req_queue, list) {
fs/fuse/dev_uring.c
423
list_for_each_entry(ent, &queue->ent_commit_queue, list) {
fs/fuse/dev_uring.c
427
spin_unlock(&queue->lock);
fs/fuse/dev_uring.c
440
struct fuse_ring_queue *queue = READ_ONCE(ring->queues[qid]);
fs/fuse/dev_uring.c
442
if (!queue)
fs/fuse/dev_uring.c
445
fuse_uring_teardown_entries(queue);
fs/fuse/dev_uring.c
475
struct fuse_ring_queue *queue = READ_ONCE(ring->queues[qid]);
fs/fuse/dev_uring.c
477
if (!queue)
fs/fuse/dev_uring.c
480
fuse_uring_teardown_entries(queue);
fs/fuse/dev_uring.c
504
struct fuse_ring_queue *queue;
fs/fuse/dev_uring.c
51
static void fuse_uring_flush_bg(struct fuse_ring_queue *queue)
fs/fuse/dev_uring.c
511
queue = ent->queue;
fs/fuse/dev_uring.c
512
spin_lock(&queue->lock);
fs/fuse/dev_uring.c
515
list_move_tail(&ent->list, &queue->ent_in_userspace);
fs/fuse/dev_uring.c
519
spin_unlock(&queue->lock);
fs/fuse/dev_uring.c
53
struct fuse_ring *ring = queue->ring;
fs/fuse/dev_uring.c
56
lockdep_assert_held(&queue->lock);
fs/fuse/dev_uring.c
66
!queue->active_background) &&
fs/fuse/dev_uring.c
67
(!list_empty(&queue->fuse_req_bg_queue))) {
fs/fuse/dev_uring.c
670
struct fuse_ring_queue *queue = ent->queue;
fs/fuse/dev_uring.c
671
struct fuse_ring *ring = queue->ring;
fs/fuse/dev_uring.c
677
queue->qid, ent, ent->state);
fs/fuse/dev_uring.c
70
req = list_first_entry(&queue->fuse_req_bg_queue,
fs/fuse/dev_uring.c
726
struct fuse_ring_queue *queue = ent->queue;
fs/fuse/dev_uring.c
73
queue->active_background++;
fs/fuse/dev_uring.c
734
spin_lock(&queue->lock);
fs/fuse/dev_uring.c
738
list_move_tail(&ent->list, &queue->ent_in_userspace);
fs/fuse/dev_uring.c
739
spin_unlock(&queue->lock);
fs/fuse/dev_uring.c
749
struct fuse_ring_queue *queue)
fs/fuse/dev_uring.c
75
list_move_tail(&req->list, &queue->fuse_req_queue);
fs/fuse/dev_uring.c
752
list_move(&ent->list, &queue->ent_avail_queue);
fs/fuse/dev_uring.c
760
struct fuse_ring_queue *queue = ent->queue;
fs/fuse/dev_uring.c
761
struct fuse_pqueue *fpq = &queue->fpq;
fs/fuse/dev_uring.c
775
struct fuse_ring_queue *queue = ent->queue;
fs/fuse/dev_uring.c
777
lockdep_assert_held(&queue->lock);
fs/fuse/dev_uring.c
781
pr_warn("%s qid=%d state=%d\n", __func__, ent->queue->qid,
fs/fuse/dev_uring.c
788
list_move_tail(&ent->list, &queue->ent_w_req_queue);
fs/fuse/dev_uring.c
794
__must_hold(&queue->lock)
fs/fuse/dev_uring.c
797
struct fuse_ring_queue *queue = ent->queue;
fs/fuse/dev_uring.c
798
struct list_head *req_queue = &queue->fuse_req_queue;
fs/fuse/dev_uring.c
800
lockdep_assert_held(&queue->lock);
fs/fuse/dev_uring.c
818
struct fuse_ring *ring = ent->queue->ring;
fs/fuse/dev_uring.c
82
struct fuse_ring_queue *queue = ent->queue;
fs/fuse/dev_uring.c
83
struct fuse_ring *ring = queue->ring;
fs/fuse/dev_uring.c
844
struct fuse_ring_queue *queue,
fs/fuse/dev_uring.c
851
spin_lock(&queue->lock);
fs/fuse/dev_uring.c
852
fuse_uring_ent_avail(ent, queue);
fs/fuse/dev_uring.c
854
spin_unlock(&queue->lock);
fs/fuse/dev_uring.c
86
lockdep_assert_not_held(&queue->lock);
fs/fuse/dev_uring.c
865
struct fuse_ring_queue *queue = ent->queue;
fs/fuse/dev_uring.c
867
lockdep_assert_held(&queue->lock);
fs/fuse/dev_uring.c
87
spin_lock(&queue->lock);
fs/fuse/dev_uring.c
873
list_move(&ent->list, &queue->ent_commit_queue);
fs/fuse/dev_uring.c
887
struct fuse_ring_queue *queue;
fs/fuse/dev_uring.c
900
queue = ring->queues[qid];
fs/fuse/dev_uring.c
901
if (!queue)
fs/fuse/dev_uring.c
903
fpq = &queue->fpq;
fs/fuse/dev_uring.c
905
if (!READ_ONCE(fc->connected) || READ_ONCE(queue->stopped))
fs/fuse/dev_uring.c
908
spin_lock(&queue->lock);
fs/fuse/dev_uring.c
91
queue->active_background--;
fs/fuse/dev_uring.c
917
pr_info("qid=%d commit_id %llu not found\n", queue->qid,
fs/fuse/dev_uring.c
919
spin_unlock(&queue->lock);
fs/fuse/dev_uring.c
929
queue->qid, commit_id, ent->state);
fs/fuse/dev_uring.c
93
fuse_uring_flush_bg(queue);
fs/fuse/dev_uring.c
930
spin_unlock(&queue->lock);
fs/fuse/dev_uring.c
938
spin_unlock(&queue->lock);
fs/fuse/dev_uring.c
950
fuse_uring_next_fuse_req(ent, queue, issue_flags);
fs/fuse/dev_uring.c
957
struct fuse_ring_queue *queue;
fs/fuse/dev_uring.c
964
queue = ring->queues[qid];
fs/fuse/dev_uring.c
965
if (!queue) {
fs/fuse/dev_uring.c
97
spin_unlock(&queue->lock);
fs/fuse/dev_uring.c
970
spin_lock(&queue->lock);
fs/fuse/dev_uring.c
971
if (list_empty(&queue->ent_avail_queue))
fs/fuse/dev_uring.c
973
spin_unlock(&queue->lock);
fs/fuse/dev_uring.c
986
struct fuse_ring_queue *queue = ent->queue;
fs/fuse/dev_uring.c
987
struct fuse_ring *ring = queue->ring;
fs/fuse/dev_uring.c
993
spin_lock(&queue->lock);
fs/fuse/dev_uring.c
995
fuse_uring_ent_avail(ent, queue);
fs/fuse/dev_uring.c
996
spin_unlock(&queue->lock);
fs/fuse/dev_uring.c
999
bool ready = is_ring_ready(ring, queue->qid);
fs/fuse/dev_uring_i.h
46
struct fuse_ring_queue *queue;
fs/gfs2/trace_gfs2.h
233
TP_PROTO(const struct gfs2_holder *gh, int queue),
fs/gfs2/trace_gfs2.h
235
TP_ARGS(gh, queue),
fs/gfs2/trace_gfs2.h
241
__field( int, queue )
fs/gfs2/trace_gfs2.h
249
__entry->queue = queue;
fs/gfs2/trace_gfs2.h
256
__entry->queue ? "" : "de",
fs/lockd/clntproc.c
245
static int nlm_wait_on_grace(wait_queue_head_t *queue)
fs/lockd/clntproc.c
250
prepare_to_wait(queue, &wait, TASK_INTERRUPTIBLE);
fs/lockd/clntproc.c
257
finish_wait(queue, &wait);
fs/netfs/fscache_cookie.c
572
bool queue = false;
fs/netfs/fscache_cookie.c
593
queue = fscache_begin_lookup(cookie, will_modify);
fs/netfs/fscache_cookie.c
606
queue = true;
fs/netfs/fscache_cookie.c
637
if (queue)
fs/nfs/nfs4session.c
27
static void nfs4_init_slot_table(struct nfs4_slot_table *tbl, const char *queue)
fs/nfs/nfs4session.c
31
rpc_init_priority_wait_queue(&tbl->slot_tbl_waitq, queue);
fs/nfs/nfs4session.c
351
const char *queue)
fs/nfs/nfs4session.c
353
nfs4_init_slot_table(tbl, queue);
fs/nfs/nfs4session.h
84
unsigned int max_reqs, const char *queue);
fs/nfsd/nfs4state.c
7995
bool queue = false;
fs/nfsd/nfs4state.c
8002
queue = true;
fs/nfsd/nfs4state.c
8006
if (queue) {
fs/nfsd/trace.h
1720
DEFINE_NFSD_CB_LIFETIME_EVENT(queue);
fs/notify/notification.c
109
goto queue;
fs/notify/notification.c
120
queue:
fs/ocfs2/dlm/dlmmaster.c
2476
struct list_head *queue;
fs/ocfs2/dlm/dlmmaster.c
2495
queue = dlm_list_idx_to_ptr(res, idx);
fs/ocfs2/dlm/dlmmaster.c
2496
list_for_each_entry(lock, queue, list) {
fs/ocfs2/dlm/dlmmaster.c
2914
struct list_head *queue = &res->granted;
fs/ocfs2/dlm/dlmmaster.c
2923
list_for_each_entry_safe(lock, next, queue, list) {
fs/ocfs2/dlm/dlmmaster.c
2941
queue++;
fs/ocfs2/dlm/dlmmaster.c
2969
struct list_head *queue;
fs/ocfs2/dlm/dlmmaster.c
2979
queue = dlm_list_idx_to_ptr(res, idx);
fs/ocfs2/dlm/dlmmaster.c
2980
list_for_each_entry(lock, queue, list) {
fs/ocfs2/dlm/dlmrecovery.c
1081
struct list_head *iter, *queue = &res->granted;
fs/ocfs2/dlm/dlmrecovery.c
1085
list_for_each(iter, queue)
fs/ocfs2/dlm/dlmrecovery.c
1087
queue++;
fs/ocfs2/dlm/dlmrecovery.c
1171
int queue)
fs/ocfs2/dlm/dlmrecovery.c
1177
if (queue == DLM_BLOCKED_LIST)
fs/ocfs2/dlm/dlmrecovery.c
1206
struct dlm_migratable_lockres *mres, int queue)
fs/ocfs2/dlm/dlmrecovery.c
1216
ml->list = queue;
fs/ocfs2/dlm/dlmrecovery.c
1219
dlm_prepare_lvb_for_migration(lock, mres, queue);
fs/ocfs2/dlm/dlmrecovery.c
1262
struct list_head *queue;
fs/ocfs2/dlm/dlmrecovery.c
1287
queue = dlm_list_idx_to_ptr(res, i);
fs/ocfs2/dlm/dlmrecovery.c
1288
list_for_each_entry(lock, queue, list) {
fs/ocfs2/dlm/dlmrecovery.c
1790
struct list_head *queue, *iter;
fs/ocfs2/dlm/dlmrecovery.c
1819
queue = dlm_list_num_to_pointer(res, ml->list);
fs/ocfs2/dlm/dlmrecovery.c
1880
if (tmpq != queue) {
fs/ocfs2/dlm/dlmrecovery.c
1898
list_move_tail(&lock->list, queue);
fs/ocfs2/dlm/dlmrecovery.c
1916
BUG_ON(queue != &res->converting);
fs/ocfs2/dlm/dlmrecovery.c
1989
list_for_each_entry(lock, queue, list) {
fs/ocfs2/dlm/dlmrecovery.c
2018
list_add(&newlock->list, queue);
fs/ocfs2/dlm/dlmrecovery.c
2020
list_add_tail(&newlock->list, queue);
fs/ocfs2/dlm/dlmrecovery.c
2046
struct list_head *queue;
fs/ocfs2/dlm/dlmrecovery.c
2065
queue = dlm_list_idx_to_ptr(res, i);
fs/ocfs2/dlm/dlmrecovery.c
2066
list_for_each_entry_safe(lock, next, queue, list) {
fs/ocfs2/dlm/dlmrecovery.c
2210
struct list_head *queue;
fs/ocfs2/dlm/dlmrecovery.c
2231
queue = dlm_list_idx_to_ptr(res, i);
fs/ocfs2/dlm/dlmrecovery.c
2232
list_for_each_entry(lock, queue, list) {
fs/ocfs2/dlm/dlmunlock.c
401
struct list_head *queue;
fs/ocfs2/dlm/dlmunlock.c
439
queue=&res->granted;
fs/ocfs2/dlm/dlmunlock.c
463
list_for_each_entry(iter, queue, list) {
fs/ocfs2/dlm/dlmunlock.c
474
queue++;
fs/ocfs2/dlmfs/userdlm.c
177
int queue = 0;
fs/ocfs2/dlmfs/userdlm.c
185
queue = 1;
fs/ocfs2/dlmfs/userdlm.c
189
queue = 1;
fs/ocfs2/dlmfs/userdlm.c
195
if (queue)
include/crypto/algapi.h
172
void crypto_init_queue(struct crypto_queue *queue, unsigned int max_qlen);
include/crypto/algapi.h
173
int crypto_enqueue_request(struct crypto_queue *queue,
include/crypto/algapi.h
175
void crypto_enqueue_request_head(struct crypto_queue *queue,
include/crypto/algapi.h
177
struct crypto_async_request *crypto_dequeue_request(struct crypto_queue *queue);
include/crypto/algapi.h
178
static inline unsigned int crypto_queue_len(struct crypto_queue *queue)
include/crypto/algapi.h
180
return queue->qlen;
include/crypto/algapi.h
226
struct crypto_queue *queue)
include/crypto/algapi.h
228
return queue->backlog == &queue->list ? NULL :
include/crypto/algapi.h
229
container_of(queue->backlog, struct crypto_async_request, list);
include/crypto/internal/aead.h
133
static inline void aead_init_queue(struct aead_queue *queue,
include/crypto/internal/aead.h
136
crypto_init_queue(&queue->base, max_qlen);
include/crypto/internal/engine.h
48
struct crypto_queue queue __guarded_by(&queue_lock);
include/crypto/internal/hash.h
262
static inline int ahash_enqueue_request(struct crypto_queue *queue,
include/crypto/internal/hash.h
265
return crypto_enqueue_request(queue, &request->base);
include/crypto/internal/hash.h
269
struct crypto_queue *queue)
include/crypto/internal/hash.h
271
return ahash_request_cast(crypto_dequeue_request(queue));
include/drm/drm_vblank.h
147
wait_queue_head_t queue;
include/drm/spsc_queue.h
104
WRITE_ONCE(queue->head, next);
include/drm/spsc_queue.h
109
if (atomic_long_cmpxchg(&queue->tail,
include/drm/spsc_queue.h
110
(long)&node->next, (long) &queue->head) != (long)&node->next) {
include/drm/spsc_queue.h
114
} while (unlikely(!(queue->head = READ_ONCE(node->next))));
include/drm/spsc_queue.h
118
atomic_dec(&queue->job_count);
include/drm/spsc_queue.h
48
static inline void spsc_queue_init(struct spsc_queue *queue)
include/drm/spsc_queue.h
50
queue->head = NULL;
include/drm/spsc_queue.h
51
atomic_long_set(&queue->tail, (long)&queue->head);
include/drm/spsc_queue.h
52
atomic_set(&queue->job_count, 0);
include/drm/spsc_queue.h
55
static inline struct spsc_node *spsc_queue_peek(struct spsc_queue *queue)
include/drm/spsc_queue.h
57
return queue->head;
include/drm/spsc_queue.h
60
static inline int spsc_queue_count(struct spsc_queue *queue)
include/drm/spsc_queue.h
62
return atomic_read(&queue->job_count);
include/drm/spsc_queue.h
65
static inline bool spsc_queue_push(struct spsc_queue *queue, struct spsc_node *node)
include/drm/spsc_queue.h
73
atomic_inc(&queue->job_count);
include/drm/spsc_queue.h
76
tail = (struct spsc_node **)atomic_long_xchg(&queue->tail, (long)&node->next);
include/drm/spsc_queue.h
87
return tail == &queue->head;
include/drm/spsc_queue.h
91
static inline struct spsc_node *spsc_queue_pop(struct spsc_queue *queue)
include/drm/spsc_queue.h
98
node = READ_ONCE(queue->head);
include/linux/avf/virtchnl.h
1430
} queue;
include/linux/blk-integrity.h
30
return queue_limits_stack_integrity(t, &bdev->bd_disk->queue->limits);
include/linux/blk-integrity.h
55
if (!blk_integrity_queue_supports_integrity(disk->queue))
include/linux/blk-integrity.h
57
return &disk->queue->limits.integrity;
include/linux/blk-mq.h
369
struct request_queue *queue;
include/linux/blkdev.h
1477
return bdev->bd_disk->queue->limits.features & BLK_FEAT_SYNCHRONOUS;
include/linux/blkdev.h
1508
return bdev->bd_disk->queue->limits.features & BLK_FEAT_NOWAIT;
include/linux/blkdev.h
162
struct request_queue *queue;
include/linux/blkdev.h
745
if (!blk_queue_is_zoned(disk->queue))
include/linux/blkdev.h
747
return sector >> ilog2(disk->queue->limits.chunk_sectors);
include/linux/blkdev.h
752
return bdev->bd_disk->queue->limits.max_open_zones;
include/linux/blkdev.h
757
return bdev->bd_disk->queue->limits.max_active_zones;
include/linux/blkdev.h
923
sector_t zone_sectors = disk->queue->limits.chunk_sectors;
include/linux/hisi_acc_qm.h
572
int hisi_qm_mb(struct hisi_qm *qm, u8 cmd, dma_addr_t dma_addr, u16 queue,
include/linux/hisi_acc_qm.h
574
int hisi_qm_mb_read(struct hisi_qm *qm, u64 *base, u8 cmd, u16 queue);
include/linux/if_eql.h
45
slave_queue_t queue;
include/linux/iio/buffer-dma.h
143
int (*submit)(struct iio_dma_buffer_queue *queue,
include/linux/iio/buffer-dma.h
145
void (*abort)(struct iio_dma_buffer_queue *queue);
include/linux/iio/buffer-dma.h
149
void iio_dma_buffer_block_list_abort(struct iio_dma_buffer_queue *queue,
include/linux/iio/buffer-dma.h
164
void iio_dma_buffer_init(struct iio_dma_buffer_queue *queue, struct device *dev,
include/linux/iio/buffer-dma.h
166
void iio_dma_buffer_exit(struct iio_dma_buffer_queue *queue);
include/linux/iio/buffer-dma.h
167
void iio_dma_buffer_release(struct iio_dma_buffer_queue *queue);
include/linux/iio/buffer-dma.h
65
struct iio_dma_buffer_queue *queue;
include/linux/iommu.h
1665
int iopf_queue_add_device(struct iopf_queue *queue, struct device *dev);
include/linux/iommu.h
1666
void iopf_queue_remove_device(struct iopf_queue *queue, struct device *dev);
include/linux/iommu.h
1669
void iopf_queue_free(struct iopf_queue *queue);
include/linux/iommu.h
1670
int iopf_queue_discard_partial(struct iopf_queue *queue);
include/linux/iommu.h
1677
iopf_queue_add_device(struct iopf_queue *queue, struct device *dev)
include/linux/iommu.h
1683
iopf_queue_remove_device(struct iopf_queue *queue, struct device *dev)
include/linux/iommu.h
1697
static inline void iopf_queue_free(struct iopf_queue *queue)
include/linux/iommu.h
1701
static inline int iopf_queue_discard_partial(struct iopf_queue *queue)
include/linux/iommu.h
822
struct iopf_queue *queue;
include/linux/kvm_host.h
240
struct list_head queue;
include/linux/kvm_host.h
365
struct list_head queue;
include/linux/mfd/ipaq-micro.h
118
struct list_head queue;
include/linux/netdevice.h
1483
int queue, u8 *mac);
include/linux/netdevice.h
1485
int queue, u16 vlan,
include/linux/netdevice.h
922
u8 queue;
include/linux/pps_gen_kernel.h
53
wait_queue_head_t queue; /* PPS event queue */
include/linux/pps_kernel.h
56
wait_queue_head_t queue; /* PPS event queue */
include/linux/ptr_ring.h
106
if (unlikely(!r->size) || r->queue[r->producer])
include/linux/ptr_ring.h
113
WRITE_ONCE(r->queue[r->producer++], ptr);
include/linux/ptr_ring.h
172
return READ_ONCE(r->queue[r->consumer_head]);
include/linux/ptr_ring.h
197
return !r->queue[READ_ONCE(r->consumer_head)];
include/linux/ptr_ring.h
259
r->queue[--head] = NULL;
include/linux/ptr_ring.h
40
void **queue;
include/linux/ptr_ring.h
487
r->batch = SMP_CACHE_BYTES * 2 / sizeof(*(r->queue));
include/linux/ptr_ring.h
499
r->queue = __ptr_ring_init_queue_alloc_noprof(size, gfp);
include/linux/ptr_ring.h
500
if (!r->queue)
include/linux/ptr_ring.h
51
return r->queue[r->producer];
include/linux/ptr_ring.h
547
if (r->queue[head]) {
include/linux/ptr_ring.h
551
r->queue[head] = batch[--n];
include/linux/ptr_ring.h
565
static inline void **__ptr_ring_swap_queue(struct ptr_ring *r, void **queue,
include/linux/ptr_ring.h
575
queue[producer++] = ptr;
include/linux/ptr_ring.h
585
old = r->queue;
include/linux/ptr_ring.h
586
r->queue = queue;
include/linux/ptr_ring.h
601
void **queue = __ptr_ring_init_queue_alloc_noprof(size, gfp);
include/linux/ptr_ring.h
604
if (!queue)
include/linux/ptr_ring.h
610
old = __ptr_ring_swap_queue(r, queue, size, gfp, destroy);
include/linux/ptr_ring.h
680
kvfree(r->queue);
include/linux/signal.h
269
extern void flush_sigqueue(struct sigpending *queue);
include/linux/skbuff.h
4133
#define skb_queue_walk(queue, skb) \
include/linux/skbuff.h
4134
for (skb = (queue)->next; \
include/linux/skbuff.h
4135
skb != (struct sk_buff *)(queue); \
include/linux/skbuff.h
4138
#define skb_queue_walk_safe(queue, skb, tmp) \
include/linux/skbuff.h
4139
for (skb = (queue)->next, tmp = skb->next; \
include/linux/skbuff.h
4140
skb != (struct sk_buff *)(queue); \
include/linux/skbuff.h
4143
#define skb_queue_walk_from(queue, skb) \
include/linux/skbuff.h
4144
for (; skb != (struct sk_buff *)(queue); \
include/linux/skbuff.h
4159
#define skb_queue_walk_from_safe(queue, skb, tmp) \
include/linux/skbuff.h
4161
skb != (struct sk_buff *)(queue); \
include/linux/skbuff.h
4164
#define skb_queue_reverse_walk(queue, skb) \
include/linux/skbuff.h
4165
for (skb = (queue)->prev; \
include/linux/skbuff.h
4166
skb != (struct sk_buff *)(queue); \
include/linux/skbuff.h
4169
#define skb_queue_reverse_walk_safe(queue, skb, tmp) \
include/linux/skbuff.h
4170
for (skb = (queue)->prev, tmp = skb->prev; \
include/linux/skbuff.h
4171
skb != (struct sk_buff *)(queue); \
include/linux/skbuff.h
4174
#define skb_queue_reverse_walk_from_safe(queue, skb, tmp) \
include/linux/skbuff.h
4176
skb != (struct sk_buff *)(queue); \
include/linux/skbuff.h
4193
int __skb_wait_for_more_packets(struct sock *sk, struct sk_buff_head *queue,
include/linux/skbuff.h
4196
struct sk_buff *__skb_try_recv_from_queue(struct sk_buff_head *queue,
include/linux/skbuff.h
4201
struct sk_buff_head *queue,
include/linux/soc/ixp4xx/qmgr.h
57
void qmgr_put_entry(unsigned int queue, u32 val);
include/linux/soc/ixp4xx/qmgr.h
58
u32 qmgr_get_entry(unsigned int queue);
include/linux/soc/ixp4xx/qmgr.h
59
int qmgr_stat_empty(unsigned int queue);
include/linux/soc/ixp4xx/qmgr.h
60
int qmgr_stat_below_low_watermark(unsigned int queue);
include/linux/soc/ixp4xx/qmgr.h
61
int qmgr_stat_full(unsigned int queue);
include/linux/soc/ixp4xx/qmgr.h
62
int qmgr_stat_overflow(unsigned int queue);
include/linux/soc/ixp4xx/qmgr.h
63
void qmgr_release_queue(unsigned int queue);
include/linux/soc/ixp4xx/qmgr.h
64
void qmgr_set_irq(unsigned int queue, int src,
include/linux/soc/ixp4xx/qmgr.h
66
void qmgr_enable_irq(unsigned int queue);
include/linux/soc/ixp4xx/qmgr.h
67
void qmgr_disable_irq(unsigned int queue);
include/linux/soc/ixp4xx/qmgr.h
74
int qmgr_request_queue(unsigned int queue, unsigned int len /* dwords */,
include/linux/soc/ixp4xx/qmgr.h
79
int __qmgr_request_queue(unsigned int queue, unsigned int len /* dwords */,
include/linux/soc/ixp4xx/qmgr.h
82
#define qmgr_request_queue(queue, len, nearly_empty_watermark, \
include/linux/soc/ixp4xx/qmgr.h
84
__qmgr_request_queue(queue, len, nearly_empty_watermark, \
include/linux/spi/spi.h
1227
struct list_head queue;
include/linux/spi/spi.h
730
struct list_head queue;
include/linux/sunrpc/cache.h
115
struct list_head queue;
include/linux/sunrpc/sched.h
241
void rpc_sleep_on_timeout(struct rpc_wait_queue *queue,
include/linux/sunrpc/sched.h
247
void rpc_sleep_on_priority_timeout(struct rpc_wait_queue *queue,
include/linux/tcp.h
592
struct request_sock_queue *queue = &inet_csk(sk)->icsk_accept_queue;
include/linux/tcp.h
595
WRITE_ONCE(queue->fastopenq.max_qlen, min_t(unsigned int, backlog, somaxconn));
include/linux/thunderbolt.h
548
struct list_head queue;
include/linux/usb/gadget.h
150
int (*queue) (struct usb_ep *ep, struct usb_request *req,
include/linux/watch_queue.h
59
struct watch_queue __rcu *queue; /* Queue to post events to */
include/media/drv-intf/saa7146_vv.h
75
struct list_head queue;
include/media/dvb_ringbuffer.h
48
wait_queue_head_t queue;
include/media/media-request.h
286
void (*queue)(struct media_request_object *object);
include/media/v4l2-dev.h
283
struct vb2_queue *queue;
include/media/v4l2-mem2mem.h
125
struct list_head queue;
include/net/bluetooth/hci_core.h
274
struct sk_buff_head queue;
include/net/flow_offload.h
286
} queue;
include/net/fq.h
34
struct sk_buff_head queue;
include/net/fq_impl.h
225
__skb_queue_tail(&flow->queue, skb);
include/net/fq_impl.h
263
skb_queue_walk_safe(&flow->queue, skb, tmp) {
include/net/fq_impl.h
267
__skb_unlink(skb, &flow->queue);
include/net/fq_impl.h
338
__skb_queue_head_init(&flow->queue);
include/net/fq_impl.h
54
skb = __skb_dequeue(&flow->queue);
include/net/fq_impl.h
73
pending = min_t(int, 32, skb_queue_len(&flow->queue) / 2);
include/net/fq_impl.h
75
skb = __skb_dequeue(&flow->queue);
include/net/garp.h
107
struct sk_buff_head queue;
include/net/ip.h
231
struct sk_buff_head *queue,
include/net/ipv6.h
1030
struct sk_buff *__ip6_make_skb(struct sock *sk, struct sk_buff_head *queue,
include/net/libeth/xdp.h
1407
bool (*queue)(struct libeth_xdp_tx_bulk *bq,
include/net/libeth/xdp.h
1418
if (act == LIBETH_XDP_TX && unlikely(!queue(bq, xdp, flush_bulk)))
include/net/mac80211.h
6165
void ieee80211_wake_queue(struct ieee80211_hw *hw, int queue);
include/net/mac80211.h
6174
void ieee80211_stop_queue(struct ieee80211_hw *hw, int queue);
include/net/mac80211.h
6186
int ieee80211_queue_stopped(struct ieee80211_hw *hw, int queue);
include/net/mana/gdma.h
470
void mana_gd_destroy_queue(struct gdma_context *gc, struct gdma_queue *queue);
include/net/mana/gdma.h
973
int mana_gd_post_and_ring(struct gdma_queue *queue,
include/net/mana/gdma.h
981
struct gdma_queue *queue);
include/net/mrp.h
123
struct sk_buff_head queue;
include/net/netdev_rx_queue.h
40
ssize_t (*show)(struct netdev_rx_queue *queue, char *buf);
include/net/netdev_rx_queue.h
41
ssize_t (*store)(struct netdev_rx_queue *queue,
include/net/netdev_rx_queue.h
52
get_netdev_rx_queue_index(struct netdev_rx_queue *queue)
include/net/netdev_rx_queue.h
54
struct net_device *dev = queue->dev;
include/net/netdev_rx_queue.h
55
int index = queue - dev->_rx;
include/net/netfilter/nf_queue.h
108
nfqueue_hash(const struct sk_buff *skb, u16 queue, u16 queues_total, u8 family,
include/net/netfilter/nf_queue.h
113
queue += reciprocal_scale(hash_v4(ip_hdr(skb), initval),
include/net/netfilter/nf_queue.h
117
queue += reciprocal_scale(hash_v6(ipv6_hdr(skb), initval),
include/net/netfilter/nf_queue.h
121
queue += reciprocal_scale(hash_bridge(skb, initval),
include/net/netfilter/nf_queue.h
126
return queue;
include/net/pkt_cls.h
837
unsigned long queue;
include/net/pkt_sched.h
148
s32 queue;
include/net/pkt_sched.h
157
s32 queue;
include/net/pkt_sched.h
208
int queue;
include/net/request_sock.h
195
static inline bool reqsk_queue_empty(const struct request_sock_queue *queue)
include/net/request_sock.h
197
return READ_ONCE(queue->rskq_accept_head) == NULL;
include/net/request_sock.h
200
static inline struct request_sock *reqsk_queue_remove(struct request_sock_queue *queue,
include/net/request_sock.h
205
spin_lock_bh(&queue->rskq_lock);
include/net/request_sock.h
206
req = queue->rskq_accept_head;
include/net/request_sock.h
209
WRITE_ONCE(queue->rskq_accept_head, req->dl_next);
include/net/request_sock.h
210
if (queue->rskq_accept_head == NULL)
include/net/request_sock.h
211
queue->rskq_accept_tail = NULL;
include/net/request_sock.h
213
spin_unlock_bh(&queue->rskq_lock);
include/net/request_sock.h
217
static inline void reqsk_queue_removed(struct request_sock_queue *queue,
include/net/request_sock.h
221
atomic_dec(&queue->young);
include/net/request_sock.h
222
atomic_dec(&queue->qlen);
include/net/request_sock.h
225
static inline void reqsk_queue_added(struct request_sock_queue *queue)
include/net/request_sock.h
227
atomic_inc(&queue->young);
include/net/request_sock.h
228
atomic_inc(&queue->qlen);
include/net/request_sock.h
231
static inline int reqsk_queue_len(const struct request_sock_queue *queue)
include/net/request_sock.h
233
return atomic_read(&queue->qlen);
include/net/request_sock.h
236
static inline int reqsk_queue_len_young(const struct request_sock_queue *queue)
include/net/request_sock.h
238
return atomic_read(&queue->young);
include/net/rose.h
105
struct sk_buff_head queue;
include/net/sctp/ulpevent.h
142
struct net *net, struct sk_buff_head *queue,
include/net/x25.h
143
struct sk_buff_head queue;
include/rdma/rdmavt_cq.h
55
struct rvt_cq_wc *queue;
include/rdma/rdmavt_qp.h
928
RDMA_READ_UAPI_ATOMIC(cq->queue->tail) :
include/rdma/rdmavt_qp.h
944
RDMA_READ_UAPI_ATOMIC(cq->queue->head) :
include/scsi/libiscsi.h
262
struct kfifo queue; /* FIFO Queue */
include/sound/asequencer.h
44
#define snd_seq_ev_is_direct(ev) ((ev)->queue == SNDRV_SEQ_QUEUE_DIRECT)
include/trace/events/sunrpc.h
2206
DEFINE_SVC_DEFERRED_EVENT(queue);
include/uapi/sound/asequencer.h
240
unsigned char queue; /* affected queue */
include/uapi/sound/asequencer.h
271
struct snd_seq_ev_queue_control queue;
include/uapi/sound/asequencer.h
286
unsigned char queue; /* schedule queue */
include/uapi/sound/asequencer.h
300
unsigned char queue; /* schedule queue */
include/uapi/sound/asequencer.h
417
unsigned char queue; /* Queue for REMOVE_DEST */
include/uapi/sound/asequencer.h
509
int queue; /* queue id */
include/uapi/sound/asequencer.h
525
int queue; /* queue id */
include/uapi/sound/asequencer.h
537
int queue; /* sequencer queue */
include/uapi/sound/asequencer.h
554
int queue; /* sequencer queue */
include/uapi/sound/asequencer.h
567
int queue; /* sequencer queue */
include/uapi/sound/asequencer.h
585
unsigned char queue; /* input time-stamp queue (optional) */
include/uapi/sound/asequencer.h
600
unsigned char queue; /* R/O: result */
include/uapi/sound/asound.h
977
unsigned int queue; /* used queue size */
ipc/sem.c
1993
struct sem_queue queue;
ipc/sem.c
2094
queue.sops = sops;
ipc/sem.c
2095
queue.nsops = nsops;
ipc/sem.c
2096
queue.undo = un;
ipc/sem.c
2097
queue.pid = task_tgid(current);
ipc/sem.c
2098
queue.alter = alter;
ipc/sem.c
2099
queue.dupsop = dupsop;
ipc/sem.c
2101
error = perform_atomic_semop(sma, &queue);
ipc/sem.c
2134
list_add_tail(&queue.list,
ipc/sem.c
2138
list_add_tail(&queue.list,
ipc/sem.c
2142
list_add_tail(&queue.list, &curr->pending_const);
ipc/sem.c
2149
list_add_tail(&queue.list, &sma->pending_alter);
ipc/sem.c
2151
list_add_tail(&queue.list, &sma->pending_const);
ipc/sem.c
2158
WRITE_ONCE(queue.status, -EINTR);
ipc/sem.c
2159
queue.sleeper = current;
ipc/sem.c
2181
error = READ_ONCE(queue.status);
ipc/sem.c
2197
error = READ_ONCE(queue.status);
ipc/sem.c
2213
unlink_queue(sma, &queue);
kernel/audit.c
785
struct sk_buff_head *queue,
kernel/audit.c
798
skb_tail = skb_peek_tail(queue);
kernel/audit.c
799
while ((skb != skb_tail) && (skb = skb_dequeue(queue))) {
kernel/bpf/cpumap.c
304
while (!kthread_should_stop() || !__ptr_ring_empty(rcpu->queue)) {
kernel/bpf/cpumap.c
314
if (__ptr_ring_empty(rcpu->queue)) {
kernel/bpf/cpumap.c
317
if (__ptr_ring_empty(rcpu->queue)) {
kernel/bpf/cpumap.c
334
n = __ptr_ring_consume_batched(rcpu->queue, frames,
kernel/bpf/cpumap.c
397
empty = __ptr_ring_empty(rcpu->queue);
kernel/bpf/cpumap.c
460
rcpu->queue = bpf_map_kmalloc_node(map, sizeof(*rcpu->queue), gfp,
kernel/bpf/cpumap.c
462
if (!rcpu->queue)
kernel/bpf/cpumap.c
465
err = ptr_ring_init(rcpu->queue, value->qsize, gfp);
kernel/bpf/cpumap.c
507
ptr_ring_cleanup(rcpu->queue, NULL);
kernel/bpf/cpumap.c
509
kfree(rcpu->queue);
kernel/bpf/cpumap.c
538
__cpu_map_ring_cleanup(rcpu->queue);
kernel/bpf/cpumap.c
539
ptr_ring_cleanup(rcpu->queue, NULL);
kernel/bpf/cpumap.c
540
kfree(rcpu->queue);
kernel/bpf/cpumap.c
68
struct ptr_ring *queue;
kernel/bpf/cpumap.c
733
q = rcpu->queue;
kernel/bpf/cpumap.c
809
ret = ptr_ring_produce(rcpu->queue, skb);
kernel/bpf/rqspinlock.c
369
goto queue;
kernel/bpf/rqspinlock.c
391
goto queue;
kernel/bpf/rqspinlock.c
440
queue:
kernel/locking/qspinlock.c
160
goto queue;
kernel/locking/qspinlock.c
182
goto queue;
kernel/locking/qspinlock.c
212
queue:
kernel/locking/rwsem.c
1007
goto queue;
kernel/locking/rwsem.c
1031
queue:
kernel/sched/rt.c
1216
if (list_empty(array->queue + rt_se_prio(rt_se)))
kernel/sched/rt.c
1331
struct list_head *queue = array->queue + rt_se_prio(rt_se);
kernel/sched/rt.c
1348
list_add(&rt_se->run_list, queue);
kernel/sched/rt.c
1350
list_add_tail(&rt_se->run_list, queue);
kernel/sched/rt.c
1471
struct list_head *queue = array->queue + rt_se_prio(rt_se);
kernel/sched/rt.c
1474
list_move(&rt_se->run_list, queue);
kernel/sched/rt.c
1476
list_move_tail(&rt_se->run_list, queue);
kernel/sched/rt.c
1675
struct list_head *queue;
kernel/sched/rt.c
1681
queue = array->queue + idx;
kernel/sched/rt.c
1682
if (WARN_ON_ONCE(list_empty(queue)))
kernel/sched/rt.c
1684
next = list_entry(queue->next, struct sched_rt_entity, run_list);
kernel/sched/rt.c
75
INIT_LIST_HEAD(array->queue + i);
kernel/sched/sched.h
313
struct list_head queue[MAX_RT_PRIO];
kernel/signal.c
478
void flush_sigqueue(struct sigpending *queue)
kernel/signal.c
482
sigemptyset(&queue->signal);
kernel/signal.c
483
while (!list_empty(&queue->list)) {
kernel/signal.c
484
q = list_entry(queue->list.next, struct sigqueue , list);
kernel/trace/blktrace.c
1128
blk_add_trace_bio(bio->bi_bdev->bd_disk->queue, bio, BLK_TA_BACKMERGE,
kernel/trace/blktrace.c
1134
blk_add_trace_bio(bio->bi_bdev->bd_disk->queue, bio, BLK_TA_FRONTMERGE,
kernel/trace/blktrace.c
1140
blk_add_trace_bio(bio->bi_bdev->bd_disk->queue, bio, BLK_TA_QUEUE, 0);
kernel/trace/blktrace.c
1145
blk_add_trace_bio(bio->bi_bdev->bd_disk->queue, bio, BLK_TA_GETRQ, 0);
kernel/trace/blktrace.c
1213
struct request_queue *q = bio->bi_bdev->bd_disk->queue;
kernel/trace/blktrace.c
1242
struct request_queue *q = bio->bi_bdev->bd_disk->queue;
kernel/watch_queue.c
215
wqueue = rcu_dereference(watch->queue);
kernel/watch_queue.c
426
put_watch_queue(rcu_access_pointer(watch->queue));
kernel/watch_queue.c
459
rcu_assign_pointer(watch->queue, wqueue);
kernel/watch_queue.c
468
struct watch_queue *wq = rcu_access_pointer(w->queue);
kernel/watch_queue.c
508
wqueue = rcu_access_pointer(watch->queue);
kernel/watch_queue.c
545
(watch->id == id && rcu_access_pointer(watch->queue) == wq))
kernel/watch_queue.c
566
wqueue = rcu_dereference(watch->queue);
lib/plist.c
182
goto queue;
lib/plist.c
191
queue:
mm/huge_memory.c
1093
struct deferred_split *queue)
mm/huge_memory.c
1097
if (split_queue_node(folio_nid(folio)) == queue)
mm/huge_memory.c
1099
return container_of(queue, struct mem_cgroup, deferred_split_queue);
mm/huge_memory.c
1109
struct deferred_split *queue)
mm/huge_memory.c
1122
struct deferred_split *queue;
mm/huge_memory.c
1125
queue = memcg_split_queue(nid, memcg);
mm/huge_memory.c
1126
spin_lock(&queue->split_queue_lock);
mm/huge_memory.c
1133
spin_unlock(&queue->split_queue_lock);
mm/huge_memory.c
1138
return queue;
mm/huge_memory.c
1144
struct deferred_split *queue;
mm/huge_memory.c
1147
queue = memcg_split_queue(nid, memcg);
mm/huge_memory.c
1148
spin_lock_irqsave(&queue->split_queue_lock, *flags);
mm/huge_memory.c
1150
spin_unlock_irqrestore(&queue->split_queue_lock, *flags);
mm/huge_memory.c
1155
return queue;
mm/huge_memory.c
1169
static inline void split_queue_unlock(struct deferred_split *queue)
mm/huge_memory.c
1171
spin_unlock(&queue->split_queue_lock);
mm/huge_memory.c
1174
static inline void split_queue_unlock_irqrestore(struct deferred_split *queue,
mm/huge_memory.c
1177
spin_unlock_irqrestore(&queue->split_queue_lock, flags);
net/802/garp.c
264
skb_queue_tail(&app->queue, app->pdu);
net/802/garp.c
272
while ((skb = skb_dequeue(&app->queue)))
net/802/garp.c
305
goto queue;
net/802/garp.c
307
goto queue;
net/802/garp.c
312
goto queue;
net/802/garp.c
319
queue:
net/802/garp.c
596
skb_queue_head_init(&app->queue);
net/802/mrp.c
356
skb_queue_tail(&app->queue, app->pdu);
net/802/mrp.c
364
while ((skb = skb_dequeue(&app->queue)))
net/802/mrp.c
426
goto queue;
net/802/mrp.c
436
goto queue;
net/802/mrp.c
447
goto queue;
net/802/mrp.c
476
queue:
net/802/mrp.c
882
skb_queue_head_init(&app->queue);
net/atm/common.c
227
struct sk_buff_head queue, *rq;
net/atm/common.c
231
__skb_queue_head_init(&queue);
net/atm/common.c
235
skb_queue_splice_init(rq, &queue);
net/atm/common.c
238
skb_queue_walk_safe(&queue, skb, tmp) {
net/atm/common.c
239
__skb_unlink(skb, &queue);
net/bluetooth/hci_conn.c
1050
skb_queue_head_init(&conn->tx_q.queue);
net/bluetooth/hci_conn.c
1233
skb_queue_purge(&conn->tx_q.queue);
net/bluetooth/hci_conn.c
3211
skb_queue_tail(&comp->queue, skb);
net/bluetooth/hci_conn.c
3219
comp->extra += skb_queue_len(&comp->queue) + 1;
net/bluetooth/hci_conn.c
3220
skb_queue_purge(&comp->queue);
net/bluetooth/hci_conn.c
3232
if (comp->extra && (comp->tracked || skb_queue_empty(&comp->queue))) {
net/bluetooth/hci_conn.c
3237
skb = skb_dequeue(&comp->queue);
net/bluetooth/hci_core.c
3221
static void hci_queue_acl(struct hci_chan *chan, struct sk_buff_head *queue,
net/bluetooth/hci_core.c
3240
skb_queue_tail(queue, skb);
net/bluetooth/hci_core.c
3252
spin_lock_bh(&queue->lock);
net/bluetooth/hci_core.c
3254
__skb_queue_tail(queue, skb);
net/bluetooth/hci_core.c
3266
__skb_queue_tail(queue, skb);
net/bluetooth/hci_core.c
3269
spin_unlock_bh(&queue->lock);
net/bluetooth/hci_core.c
3272
bt_dev_dbg(hdev, "chan %p queued %d", chan, skb_queue_len(queue));
net/bluetooth/hci_core.c
3324
static void hci_queue_iso(struct hci_conn *conn, struct sk_buff_head *queue,
net/bluetooth/hci_core.c
3345
skb_queue_tail(queue, skb);
net/bluetooth/hci_core.c
3352
__skb_queue_tail(queue, skb);
net/bluetooth/hci_core.c
3364
__skb_queue_tail(queue, skb);
net/bluetooth/hci_core.c
3368
bt_dev_dbg(hdev, "hcon %p queued %d", conn, skb_queue_len(queue));
net/bluetooth/l2cap_core.c
2546
static void l2cap_tx_timestamp_seg(struct sk_buff_head *queue,
net/bluetooth/l2cap_core.c
2550
struct sk_buff *skb = skb_peek(queue);
net/bluetooth/l2cap_core.c
2554
l2cap_tx_timestamp(skb_peek_tail(queue), sockc, len);
net/core/datagram.c
103
if (READ_ONCE(queue->prev) != skb)
net/core/datagram.c
168
struct sk_buff *__skb_try_recv_from_queue(struct sk_buff_head *queue,
net/core/datagram.c
182
*last = queue->prev;
net/core/datagram.c
183
skb_queue_walk(queue, skb) {
net/core/datagram.c
199
__skb_unlink(skb, queue);
net/core/datagram.c
243
struct sk_buff_head *queue,
net/core/datagram.c
264
spin_lock_irqsave(&queue->lock, cpu_flags);
net/core/datagram.c
265
skb = __skb_try_recv_from_queue(queue, flags, off, &error,
net/core/datagram.c
267
spin_unlock_irqrestore(&queue->lock, cpu_flags);
net/core/datagram.c
277
} while (READ_ONCE(queue->prev) != *last);
net/core/datagram.c
89
int __skb_wait_for_more_packets(struct sock *sk, struct sk_buff_head *queue,
net/core/dev.c
11208
struct netdev_queue *queue, void *_unused)
net/core/dev.c
11211
spin_lock_init(&queue->_xmit_lock);
net/core/dev.c
11212
netdev_set_xmit_lockdep_class(&queue->_xmit_lock, dev->type);
net/core/dev.c
11213
queue->xmit_lock_owner = -1;
net/core/dev.c
11214
netdev_queue_numa_node_write(queue, NUMA_NO_NODE);
net/core/dev.c
11215
queue->dev = dev;
net/core/dev.c
11217
dql_init(&queue->dql, HZ);
net/core/dev.c
11967
struct netdev_queue *queue = dev_ingress_queue(dev);
net/core/dev.c
11970
if (queue)
net/core/dev.c
11971
return queue;
net/core/dev.c
11972
queue = kzalloc_obj(*queue);
net/core/dev.c
11973
if (!queue)
net/core/dev.c
11975
netdev_init_one_queue(dev, queue, NULL);
net/core/dev.c
11976
RCU_INIT_POINTER(queue->qdisc, &noop_qdisc);
net/core/dev.c
11977
RCU_INIT_POINTER(queue->qdisc_sleeping, &noop_qdisc);
net/core/dev.c
11978
rcu_assign_pointer(dev->ingress_queue, queue);
net/core/dev.c
11980
return queue;
net/core/net-sysfs.c
1006
old_map = rcu_dereference_protected(queue->rps_map,
net/core/net-sysfs.c
1008
rcu_assign_pointer(queue->rps_map, map);
net/core/net-sysfs.c
1033
static ssize_t store_rps_map(struct netdev_rx_queue *queue,
net/core/net-sysfs.c
1053
err = netdev_rx_queue_set_rps_mask(queue, mask);
net/core/net-sysfs.c
1060
static ssize_t show_rps_dev_flow_table_cnt(struct netdev_rx_queue *queue,
net/core/net-sysfs.c
1067
flow_table = rcu_dereference(queue->rps_flow_table);
net/core/net-sysfs.c
1082
static ssize_t store_rps_dev_flow_table_cnt(struct netdev_rx_queue *queue,
net/core/net-sysfs.c
1132
old_table = rcu_dereference_protected(queue->rps_flow_table,
net/core/net-sysfs.c
1134
rcu_assign_pointer(queue->rps_flow_table, table);
net/core/net-sysfs.c
1162
struct netdev_rx_queue *queue = to_rx_queue(kobj);
net/core/net-sysfs.c
1167
map = rcu_dereference_protected(queue->rps_map, 1);
net/core/net-sysfs.c
1169
RCU_INIT_POINTER(queue->rps_map, NULL);
net/core/net-sysfs.c
1173
flow_table = rcu_dereference_protected(queue->rps_flow_table, 1);
net/core/net-sysfs.c
1175
RCU_INIT_POINTER(queue->rps_flow_table, NULL);
net/core/net-sysfs.c
1181
netdev_put(queue->dev, &queue->dev_tracker);
net/core/net-sysfs.c
1186
struct netdev_rx_queue *queue = to_rx_queue(kobj);
net/core/net-sysfs.c
1187
struct device *dev = &queue->dev->dev;
net/core/net-sysfs.c
1212
struct netdev_rx_queue *queue)
net/core/net-sysfs.c
1222
res = netdev_rx_queue_set_rps_mask(queue, rps_default_mask);
net/core/net-sysfs.c
1234
struct netdev_rx_queue *queue = dev->_rx + index;
net/core/net-sysfs.c
1235
struct kobject *kobj = &queue->kobj;
net/core/net-sysfs.c
1257
netdev_hold(queue->dev, &queue->dev_tracker, GFP_KERNEL);
net/core/net-sysfs.c
1265
queue->groups = rx_queue_default_groups;
net/core/net-sysfs.c
1266
error = sysfs_create_groups(kobj, queue->groups);
net/core/net-sysfs.c
1276
error = rx_queue_default_mask(dev, queue);
net/core/net-sysfs.c
1285
sysfs_remove_groups(kobj, queue->groups);
net/core/net-sysfs.c
1294
struct netdev_rx_queue *queue = dev->_rx + index;
net/core/net-sysfs.c
1295
struct kobject *kobj = &queue->kobj;
net/core/net-sysfs.c
1330
struct netdev_rx_queue *queue = &dev->_rx[i];
net/core/net-sysfs.c
1331
struct kobject *kobj = &queue->kobj;
net/core/net-sysfs.c
1337
sysfs_remove_groups(kobj, queue->groups);
net/core/net-sysfs.c
1377
struct netdev_queue *queue, char *buf);
net/core/net-sysfs.c
1379
struct netdev_queue *queue, const char *buf,
net/core/net-sysfs.c
1392
struct netdev_queue *queue = to_netdev_queue(kobj);
net/core/net-sysfs.c
1397
return attribute->show(kobj, attr, queue, buf);
net/core/net-sysfs.c
1406
struct netdev_queue *queue = to_netdev_queue(kobj);
net/core/net-sysfs.c
1411
return attribute->store(kobj, attr, queue, buf, count);
net/core/net-sysfs.c
1420
struct netdev_queue *queue, char *buf)
net/core/net-sysfs.c
1422
unsigned long trans_timeout = atomic_long_read(&queue->trans_timeout);
net/core/net-sysfs.c
1427
static unsigned int get_netdev_queue_index(struct netdev_queue *queue)
net/core/net-sysfs.c
1429
struct net_device *dev = queue->dev;
net/core/net-sysfs.c
1432
i = queue - dev->_tx;
net/core/net-sysfs.c
1439
struct netdev_queue *queue, char *buf)
net/core/net-sysfs.c
1441
struct net_device *dev = queue->dev;
net/core/net-sysfs.c
1447
ret = sysfs_rtnl_lock(kobj, attr, queue->dev);
net/core/net-sysfs.c
1451
index = get_netdev_queue_index(queue);
net/core/net-sysfs.c
1477
struct netdev_queue *queue, char *buf)
net/core/net-sysfs.c
1479
return sysfs_emit(buf, "%lu\n", queue->tx_maxrate);
net/core/net-sysfs.c
1483
struct netdev_queue *queue, const char *buf,
net/core/net-sysfs.c
1486
int err, index = get_netdev_queue_index(queue);
net/core/net-sysfs.c
1487
struct net_device *dev = queue->dev;
net/core/net-sysfs.c
1514
queue->tx_maxrate = rate;
net/core/net-sysfs.c
1564
struct netdev_queue *queue, char *buf)
net/core/net-sysfs.c
1566
struct dql *dql = &queue->dql;
net/core/net-sysfs.c
1572
struct netdev_queue *queue, const char *buf,
net/core/net-sysfs.c
1575
struct dql *dql = &queue->dql;
net/core/net-sysfs.c
1593
struct netdev_queue *queue, char *buf)
net/core/net-sysfs.c
1595
struct dql *dql = &queue->dql;
net/core/net-sysfs.c
1601
struct netdev_queue *queue, const char *buf,
net/core/net-sysfs.c
1604
struct dql *dql = &queue->dql;
net/core/net-sysfs.c
1629
struct netdev_queue *queue, char *buf)
net/core/net-sysfs.c
1631
return sysfs_emit(buf, "%u\n", READ_ONCE(queue->dql.stall_max));
net/core/net-sysfs.c
1635
struct netdev_queue *queue, const char *buf,
net/core/net-sysfs.c
1638
WRITE_ONCE(queue->dql.stall_max, 0);
net/core/net-sysfs.c
1646
struct netdev_queue *queue, char *buf)
net/core/net-sysfs.c
1648
struct dql *dql = &queue->dql;
net/core/net-sysfs.c
1657
struct netdev_queue *queue, char *buf)
net/core/net-sysfs.c
1659
struct dql *dql = &queue->dql;
net/core/net-sysfs.c
1670
struct netdev_queue *queue, char *buf) \
net/core/net-sysfs.c
1672
return bql_show(buf, queue->dql.FIELD); \
net/core/net-sysfs.c
1677
struct netdev_queue *queue, \
net/core/net-sysfs.c
1680
return bql_set(buf, len, &queue->dql.FIELD); \
net/core/net-sysfs.c
1764
struct netdev_queue *queue, char *buf)
net/core/net-sysfs.c
1766
struct net_device *dev = queue->dev;
net/core/net-sysfs.c
1773
index = get_netdev_queue_index(queue);
net/core/net-sysfs.c
1775
ret = sysfs_rtnl_lock(kobj, attr, queue->dev);
net/core/net-sysfs.c
1801
struct netdev_queue *queue, const char *buf,
net/core/net-sysfs.c
1804
struct net_device *dev = queue->dev;
net/core/net-sysfs.c
1818
index = get_netdev_queue_index(queue);
net/core/net-sysfs.c
1844
struct netdev_queue *queue, char *buf)
net/core/net-sysfs.c
1846
struct net_device *dev = queue->dev;
net/core/net-sysfs.c
1850
index = get_netdev_queue_index(queue);
net/core/net-sysfs.c
1870
struct netdev_queue *queue, const char *buf,
net/core/net-sysfs.c
1873
struct net_device *dev = queue->dev;
net/core/net-sysfs.c
1886
index = get_netdev_queue_index(queue);
net/core/net-sysfs.c
1928
struct netdev_queue *queue = to_netdev_queue(kobj);
net/core/net-sysfs.c
1931
netdev_put(queue->dev, &queue->dev_tracker);
net/core/net-sysfs.c
1936
struct netdev_queue *queue = to_netdev_queue(kobj);
net/core/net-sysfs.c
1937
struct device *dev = &queue->dev->dev;
net/core/net-sysfs.c
1971
struct netdev_queue *queue = dev->_tx + index;
net/core/net-sysfs.c
1972
struct kobject *kobj = &queue->kobj;
net/core/net-sysfs.c
1994
netdev_hold(queue->dev, &queue->dev_tracker, GFP_KERNEL);
net/core/net-sysfs.c
2002
queue->groups = netdev_queue_default_groups;
net/core/net-sysfs.c
2003
error = sysfs_create_groups(kobj, queue->groups);
net/core/net-sysfs.c
2017
sysfs_remove_groups(kobj, queue->groups);
net/core/net-sysfs.c
2026
struct netdev_queue *queue = ndev->_tx + index;
net/core/net-sysfs.c
2027
struct kobject *kobj = &queue->kobj;
net/core/net-sysfs.c
2064
struct netdev_queue *queue = dev->_tx + i;
net/core/net-sysfs.c
2067
queue->kobj.uevent_suppress = 1;
net/core/net-sysfs.c
2070
sysfs_remove_group(&queue->kobj, &dql_group);
net/core/net-sysfs.c
2072
sysfs_remove_groups(&queue->kobj, queue->groups);
net/core/net-sysfs.c
2073
kobject_put(&queue->kobj);
net/core/net-sysfs.c
933
struct netdev_rx_queue *queue = to_rx_queue(kobj);
net/core/net-sysfs.c
938
return attribute->show(queue, buf);
net/core/net-sysfs.c
945
struct netdev_rx_queue *queue = to_rx_queue(kobj);
net/core/net-sysfs.c
950
return attribute->store(queue, buf, count);
net/core/net-sysfs.c
959
static ssize_t show_rps_map(struct netdev_rx_queue *queue, char *buf)
net/core/net-sysfs.c
969
map = rcu_dereference(queue->rps_map);
net/core/net-sysfs.c
981
static int netdev_rx_queue_set_rps_mask(struct netdev_rx_queue *queue,
net/core/pktgen.c
3708
init_waitqueue_head(&t->queue);
net/core/pktgen.c
3721
wait_event_freezable_timeout(t->queue,
net/core/pktgen.c
472
wait_queue_head_t queue;
net/dsa/tag_brcm.c
121
((queue & BRCM_IG_TC_MASK) << BRCM_IG_TC_SHIFT);
net/dsa/tag_brcm.c
130
skb_set_queue_mapping(skb, BRCM_TAG_SET_PORT_QUEUE(dp->index, queue));
net/dsa/tag_brcm.c
94
u16 queue = skb_get_queue_mapping(skb);
net/ethtool/ioctl.c
3855
act->queue.ctx = input->rss_ctx;
net/ethtool/ioctl.c
3857
act->queue.vf = ethtool_get_flow_spec_ring_vf(fs->ring_cookie);
net/ethtool/ioctl.c
3858
act->queue.index = ethtool_get_flow_spec_ring(fs->ring_cookie);
net/ipv4/inet_connection_sock.c
1033
struct request_sock_queue *queue;
net/ipv4/inet_connection_sock.c
1082
queue = &icsk->icsk_accept_queue;
net/ipv4/inet_connection_sock.c
1083
qlen = reqsk_queue_len(queue);
net/ipv4/inet_connection_sock.c
1085
int young = reqsk_queue_len_young(queue) << 1;
net/ipv4/inet_connection_sock.c
1095
syn_ack_recalc(req, max_syn_ack_retries, READ_ONCE(queue->rskq_defer_accept),
net/ipv4/inet_connection_sock.c
1106
atomic_dec(&queue->young);
net/ipv4/inet_connection_sock.c
1135
reqsk_queue_removed(queue, nreq);
net/ipv4/inet_connection_sock.c
1313
static void reqsk_queue_alloc(struct request_sock_queue *queue)
net/ipv4/inet_connection_sock.c
1315
queue->fastopenq.rskq_rst_head = NULL;
net/ipv4/inet_connection_sock.c
1316
queue->fastopenq.rskq_rst_tail = NULL;
net/ipv4/inet_connection_sock.c
1317
queue->fastopenq.qlen = 0;
net/ipv4/inet_connection_sock.c
1319
queue->rskq_accept_head = NULL;
net/ipv4/inet_connection_sock.c
1386
struct request_sock_queue *queue = &inet_csk(sk)->icsk_accept_queue;
net/ipv4/inet_connection_sock.c
1388
spin_lock(&queue->rskq_lock);
net/ipv4/inet_connection_sock.c
1395
if (queue->rskq_accept_head == NULL)
net/ipv4/inet_connection_sock.c
1396
WRITE_ONCE(queue->rskq_accept_head, req);
net/ipv4/inet_connection_sock.c
1398
queue->rskq_accept_tail->dl_next = req;
net/ipv4/inet_connection_sock.c
1399
queue->rskq_accept_tail = req;
net/ipv4/inet_connection_sock.c
1402
spin_unlock(&queue->rskq_lock);
net/ipv4/inet_connection_sock.c
1459
struct request_sock_queue *queue = &icsk->icsk_accept_queue;
net/ipv4/inet_connection_sock.c
1470
while ((req = reqsk_queue_remove(queue, sk)) != NULL) {
net/ipv4/inet_connection_sock.c
1512
if (queue->fastopenq.rskq_rst_head) {
net/ipv4/inet_connection_sock.c
1514
spin_lock_bh(&queue->fastopenq.lock);
net/ipv4/inet_connection_sock.c
1515
req = queue->fastopenq.rskq_rst_head;
net/ipv4/inet_connection_sock.c
1516
queue->fastopenq.rskq_rst_head = NULL;
net/ipv4/inet_connection_sock.c
1517
spin_unlock_bh(&queue->fastopenq.lock);
net/ipv4/inet_connection_sock.c
653
struct request_sock_queue *queue = &icsk->icsk_accept_queue;
net/ipv4/inet_connection_sock.c
668
if (reqsk_queue_empty(queue)) {
net/ipv4/inet_connection_sock.c
680
req = reqsk_queue_remove(queue, sk);
net/ipv4/inet_connection_sock.c
681
arg->is_empty = reqsk_queue_empty(queue);
net/ipv4/inet_connection_sock.c
686
spin_lock_bh(&queue->fastopenq.lock);
net/ipv4/inet_connection_sock.c
697
spin_unlock_bh(&queue->fastopenq.lock);
net/ipv4/inet_connection_sock.c
960
static void reqsk_queue_migrated(struct request_sock_queue *queue,
net/ipv4/inet_connection_sock.c
964
atomic_inc(&queue->young);
net/ipv4/inet_connection_sock.c
965
atomic_inc(&queue->qlen);
net/ipv4/ip_output.c
1206
__skb_queue_tail(queue, skb);
net/ipv4/ip_output.c
1398
struct sk_buff_head *queue,
net/ipv4/ip_output.c
1411
skb = __skb_dequeue(queue);
net/ipv4/ip_output.c
1419
while ((tmp_skb = __skb_dequeue(queue)) != NULL) {
net/ipv4/ip_output.c
1535
struct sk_buff_head *queue,
net/ipv4/ip_output.c
1540
while ((skb = __skb_dequeue_tail(queue)) != NULL)
net/ipv4/ip_output.c
1559
struct sk_buff_head queue;
net/ipv4/ip_output.c
1565
__skb_queue_head_init(&queue);
net/ipv4/ip_output.c
1574
err = __ip_append_data(sk, fl4, &queue, cork,
net/ipv4/ip_output.c
1578
__ip_flush_pending_frames(sk, &queue, cork);
net/ipv4/ip_output.c
1582
return __ip_make_skb(sk, fl4, &queue, cork);
net/ipv4/ip_output.c
952
struct sk_buff_head *queue,
net/ipv4/ip_output.c
978
skb = skb_peek_tail(queue);
net/ipv4/tcp_fastopen.c
331
struct request_sock_queue *queue = &inet_csk(sk)->icsk_accept_queue;
net/ipv4/tcp_fastopen.c
340
spin_lock(&queue->fastopenq.lock);
net/ipv4/tcp_fastopen.c
341
queue->fastopenq.qlen++;
net/ipv4/tcp_fastopen.c
342
spin_unlock(&queue->fastopenq.lock);
net/ipv4/tcp_input.c
7547
struct request_sock_queue *queue = &inet_csk(sk)->icsk_accept_queue;
net/ipv4/tcp_input.c
7564
if (syncookies != 2 && !READ_ONCE(queue->synflood_warned)) {
net/ipv4/tcp_input.c
7565
WRITE_ONCE(queue->synflood_warned, 1);
net/ipv4/tcp_yeah.c
114
u32 rtt, queue;
net/ipv4/tcp_yeah.c
136
queue = bw;
net/ipv4/tcp_yeah.c
138
if (queue > TCP_YEAH_ALPHA ||
net/ipv4/tcp_yeah.c
140
if (queue > TCP_YEAH_ALPHA &&
net/ipv4/tcp_yeah.c
142
u32 reduction = min(queue / TCP_YEAH_GAMMA ,
net/ipv4/tcp_yeah.c
171
yeah->lastQ = queue;
net/ipv4/udp.c
1976
struct sk_buff_head *queue;
net/ipv4/udp.c
1981
queue = &udp_sk(sk)->reader_queue;
net/ipv4/udp.c
1992
spin_lock_bh(&queue->lock);
net/ipv4/udp.c
1993
skb = __skb_try_recv_from_queue(queue, flags, off, err,
net/ipv4/udp.c
1998
spin_unlock_bh(&queue->lock);
net/ipv4/udp.c
2003
spin_unlock_bh(&queue->lock);
net/ipv4/udp.c
2013
skb_queue_splice_tail_init(sk_queue, queue);
net/ipv4/udp.c
2015
skb = __skb_try_recv_from_queue(queue, flags, off, err,
net/ipv4/udp.c
2020
spin_unlock_bh(&queue->lock);
net/ipv6/ioam6.c
802
struct netdev_queue *queue;
net/ipv6/ioam6.c
810
queue = skb_get_tx_queue(dev, skb);
net/ipv6/ioam6.c
811
qdisc = rcu_dereference(queue->qdisc);
net/ipv6/ip6_output.c
1419
struct sk_buff_head *queue,
net/ipv6/ip6_output.c
1448
skb = skb_peek_tail(queue);
net/ipv6/ip6_output.c
1738
__skb_queue_tail(queue, skb);
net/ipv6/ip6_output.c
1893
struct sk_buff_head *queue,
net/ipv6/ip6_output.c
1906
skb = __skb_dequeue(queue);
net/ipv6/ip6_output.c
1914
while ((tmp_skb = __skb_dequeue(queue)) != NULL) {
net/ipv6/ip6_output.c
2010
struct sk_buff_head *queue,
net/ipv6/ip6_output.c
2015
while ((skb = __skb_dequeue_tail(queue)) != NULL) {
net/ipv6/ip6_output.c
2040
struct sk_buff_head queue;
net/ipv6/ip6_output.c
2048
__skb_queue_head_init(&queue);
net/ipv6/ip6_output.c
2060
err = __ip6_append_data(sk, &queue, cork,
net/ipv6/ip6_output.c
2065
__ip6_flush_pending_frames(sk, &queue, cork);
net/ipv6/ip6_output.c
2069
return __ip6_make_skb(sk, &queue, cork);
net/mac80211/agg-tx.c
155
int queue = sdata->vif.hw_queue[ieee80211_ac_from_tid(tid)];
net/mac80211/agg-tx.c
159
if (atomic_inc_return(&sdata->local->agg_queue_stop[queue]) == 1)
net/mac80211/agg-tx.c
161
&sdata->local->hw, queue,
net/mac80211/agg-tx.c
170
int queue = sdata->vif.hw_queue[ieee80211_ac_from_tid(tid)];
net/mac80211/agg-tx.c
172
if (atomic_dec_return(&sdata->local->agg_queue_stop[queue]) == 0)
net/mac80211/agg-tx.c
174
&sdata->local->hw, queue,
net/mac80211/agg-tx.c
236
int queue = sdata->vif.hw_queue[ieee80211_ac_from_tid(tid)];
net/mac80211/agg-tx.c
250
&local->pending[queue]);
net/mac80211/ieee80211_i.h
2513
void ieee80211_wake_queue_by_reason(struct ieee80211_hw *hw, int queue,
net/mac80211/ieee80211_i.h
2516
void ieee80211_stop_queue_by_reason(struct ieee80211_hw *hw, int queue,
net/mac80211/rx.c
2394
int queue = rx->security_idx;
net/mac80211/rx.c
2403
rx->key->u.ccmp.rx_pn[queue],
net/mac80211/rx.c
2409
BUILD_BUG_ON(sizeof(rx->key->u.ccmp.rx_pn[queue]) !=
net/mac80211/rx.c
2410
sizeof(rx->key->u.gcmp.rx_pn[queue]));
net/mac80211/rx.c
3693
goto queue;
net/mac80211/rx.c
3716
goto queue;
net/mac80211/rx.c
3734
goto queue;
net/mac80211/rx.c
3739
goto queue;
net/mac80211/rx.c
3777
goto queue;
net/mac80211/rx.c
3820
goto queue;
net/mac80211/rx.c
3838
goto queue;
net/mac80211/rx.c
3856
goto queue;
net/mac80211/rx.c
3866
goto queue;
net/mac80211/rx.c
3885
goto queue;
net/mac80211/rx.c
3893
goto queue;
net/mac80211/rx.c
3901
goto queue;
net/mac80211/rx.c
3913
goto queue;
net/mac80211/rx.c
3922
goto queue;
net/mac80211/rx.c
3930
goto queue;
net/mac80211/rx.c
3938
goto queue;
net/mac80211/rx.c
3958
queue:
net/mac80211/tkip.c
242
u8 *ra, int only_iv, int queue,
net/mac80211/tkip.c
250
struct tkip_ctx_rx *rx_ctx = &key->u.tkip.rx[queue];
net/mac80211/tkip.h
27
u8 *ra, int only_iv, int queue,
net/mac80211/trace.h
3232
TP_PROTO(struct ieee80211_local *local, u16 queue,
net/mac80211/trace.h
3235
TP_ARGS(local, queue, reason, refcount),
net/mac80211/trace.h
3239
__field(u16, queue)
net/mac80211/trace.h
3246
__entry->queue = queue;
net/mac80211/trace.h
3253
LOCAL_PR_ARG, __entry->queue, __entry->reason,
net/mac80211/trace.h
3259
TP_PROTO(struct ieee80211_local *local, u16 queue,
net/mac80211/trace.h
3262
TP_ARGS(local, queue, reason, refcount),
net/mac80211/trace.h
3266
__field(u16, queue)
net/mac80211/trace.h
3273
__entry->queue = queue;
net/mac80211/trace.h
3280
LOCAL_PR_ARG, __entry->queue, __entry->reason,
net/mac80211/tx.c
3467
head = skb_peek_tail(&flow->queue);
net/mac80211/tx.c
4418
struct sk_buff_head *queue)
net/mac80211/tx.c
4446
__skb_queue_tail(queue, cloned_skb);
net/mac80211/tx.c
4452
__skb_queue_tail(queue, skb);
net/mac80211/tx.c
4461
__skb_queue_purge(queue);
net/mac80211/tx.c
4462
__skb_queue_tail(queue, skb);
net/mac80211/tx.c
4529
struct sk_buff_head queue;
net/mac80211/tx.c
4531
__skb_queue_head_init(&queue);
net/mac80211/tx.c
4532
ieee80211_convert_to_unicast(skb, dev, &queue);
net/mac80211/tx.c
4533
while ((skb = __skb_dequeue(&queue)))
net/mac80211/tx.c
4621
u16 queue;
net/mac80211/tx.c
4624
queue = ieee80211_select_queue(sdata, sta, skb);
net/mac80211/tx.c
4625
skb_set_queue_mapping(skb, queue);
net/mac80211/tx.c
4659
info->hw_queue = sdata->vif.hw_queue[queue];
net/mac80211/tx.c
4686
sta->deflink.tx_stats.packets[queue] += skbs;
net/mac80211/tx.c
4687
sta->deflink.tx_stats.bytes[queue] += len;
net/mac80211/tx.c
6417
u16 queue = ieee80211_select_queue(sdata, sta, skb);
net/mac80211/tx.c
6419
skb_set_queue_mapping(skb, queue);
net/mac80211/util.c
285
struct ieee80211_txq *queue)
net/mac80211/util.c
288
.sta = queue->sta,
net/mac80211/util.c
293
skb = ieee80211_tx_dequeue(&local->hw, queue);
net/mac80211/util.c
307
struct ieee80211_txq *queue;
net/mac80211/util.c
313
while ((queue = ieee80211_next_txq(hw, txq->ac))) {
net/mac80211/util.c
314
wake_tx_push_queue(local, sdata, queue);
net/mac80211/util.c
315
ieee80211_return_txq(hw, queue, false);
net/mac80211/util.c
432
static void __ieee80211_wake_queue(struct ieee80211_hw *hw, int queue,
net/mac80211/util.c
439
if (WARN_ON(queue >= hw->queues))
net/mac80211/util.c
442
if (!test_bit(reason, &local->queue_stop_reasons[queue]))
net/mac80211/util.c
446
local->q_stop_reasons[queue][reason] = 0;
net/mac80211/util.c
448
local->q_stop_reasons[queue][reason]--;
net/mac80211/util.c
449
if (WARN_ON(local->q_stop_reasons[queue][reason] < 0))
net/mac80211/util.c
450
local->q_stop_reasons[queue][reason] = 0;
net/mac80211/util.c
453
if (local->q_stop_reasons[queue][reason] == 0)
net/mac80211/util.c
454
__clear_bit(reason, &local->queue_stop_reasons[queue]);
net/mac80211/util.c
456
trace_wake_queue(local, queue, reason,
net/mac80211/util.c
457
local->q_stop_reasons[queue][reason]);
net/mac80211/util.c
459
if (local->queue_stop_reasons[queue] != 0)
net/mac80211/util.c
463
if (!skb_queue_empty(&local->pending[queue]))
net/mac80211/util.c
479
void ieee80211_wake_queue_by_reason(struct ieee80211_hw *hw, int queue,
net/mac80211/util.c
487
__ieee80211_wake_queue(hw, queue, reason, refcounted, &flags);
net/mac80211/util.c
491
void ieee80211_wake_queue(struct ieee80211_hw *hw, int queue)
net/mac80211/util.c
493
ieee80211_wake_queue_by_reason(hw, queue,
net/mac80211/util.c
499
static void __ieee80211_stop_queue(struct ieee80211_hw *hw, int queue,
net/mac80211/util.c
505
if (WARN_ON(queue >= hw->queues))
net/mac80211/util.c
509
local->q_stop_reasons[queue][reason] = 1;
net/mac80211/util.c
511
local->q_stop_reasons[queue][reason]++;
net/mac80211/util.c
513
trace_stop_queue(local, queue, reason,
net/mac80211/util.c
514
local->q_stop_reasons[queue][reason]);
net/mac80211/util.c
516
set_bit(reason, &local->queue_stop_reasons[queue]);
net/mac80211/util.c
519
void ieee80211_stop_queue_by_reason(struct ieee80211_hw *hw, int queue,
net/mac80211/util.c
527
__ieee80211_stop_queue(hw, queue, reason, refcounted);
net/mac80211/util.c
531
void ieee80211_stop_queue(struct ieee80211_hw *hw, int queue)
net/mac80211/util.c
533
ieee80211_stop_queue_by_reason(hw, queue,
net/mac80211/util.c
545
int queue = info->hw_queue;
net/mac80211/util.c
553
__ieee80211_stop_queue(hw, queue, IEEE80211_QUEUE_STOP_REASON_SKB_ADD,
net/mac80211/util.c
555
__skb_queue_tail(&local->pending[queue], skb);
net/mac80211/util.c
556
__ieee80211_wake_queue(hw, queue, IEEE80211_QUEUE_STOP_REASON_SKB_ADD,
net/mac80211/util.c
567
int queue, i;
net/mac80211/util.c
578
queue = info->hw_queue;
net/mac80211/util.c
580
__ieee80211_stop_queue(hw, queue,
net/mac80211/util.c
584
__skb_queue_tail(&local->pending[queue], skb);
net/mac80211/util.c
619
int ieee80211_queue_stopped(struct ieee80211_hw *hw, int queue)
net/mac80211/util.c
625
if (WARN_ON(queue >= hw->queues))
net/mac80211/util.c
630
&local->queue_stop_reasons[queue]);
net/mac80211/wpa.c
525
int queue;
net/mac80211/wpa.c
556
queue = rx->security_idx;
net/mac80211/wpa.c
558
res = memcmp(pn, key->u.ccmp.rx_pn[queue],
net/mac80211/wpa.c
581
memcpy(key->u.ccmp.rx_pn[queue], pn, IEEE80211_CCMP_PN_LEN);
net/mac80211/wpa.c
722
int data_len, queue, mic_len = IEEE80211_GCMP_MIC_LEN;
net/mac80211/wpa.c
753
queue = rx->security_idx;
net/mac80211/wpa.c
755
res = memcmp(pn, key->u.gcmp.rx_pn[queue],
net/mac80211/wpa.c
779
memcpy(key->u.gcmp.rx_pn[queue], pn, IEEE80211_GCMP_PN_LEN);
net/mptcp/subflow.c
1895
struct request_sock_queue *queue = &inet_csk(listener_ssk)->icsk_accept_queue;
net/mptcp/subflow.c
1904
spin_lock_bh(&queue->rskq_lock);
net/mptcp/subflow.c
1905
head = queue->rskq_accept_head;
net/mptcp/subflow.c
1906
tail = queue->rskq_accept_tail;
net/mptcp/subflow.c
1907
queue->rskq_accept_head = NULL;
net/mptcp/subflow.c
1908
queue->rskq_accept_tail = NULL;
net/mptcp/subflow.c
1909
spin_unlock_bh(&queue->rskq_lock);
net/mptcp/subflow.c
1957
spin_lock_bh(&queue->rskq_lock);
net/mptcp/subflow.c
1958
WARN_ON_ONCE(queue->rskq_accept_head);
net/mptcp/subflow.c
1959
queue->rskq_accept_head = head;
net/mptcp/subflow.c
1960
queue->rskq_accept_tail = tail;
net/mptcp/subflow.c
1961
spin_unlock_bh(&queue->rskq_lock);
net/netfilter/nfnetlink_queue.c
1007
__nfqnl_enqueue_packet_gso(struct net *net, struct nfqnl_instance *queue,
net/netfilter/nfnetlink_queue.c
1018
ret = __nfqnl_enqueue_packet(net, queue, entry);
net/netfilter/nfnetlink_queue.c
1029
ret = __nfqnl_enqueue_packet(net, queue, entry_seg);
net/netfilter/nfnetlink_queue.c
1041
struct nfqnl_instance *queue;
net/netfilter/nfnetlink_queue.c
1048
queue = instance_lookup(q, queuenum);
net/netfilter/nfnetlink_queue.c
1049
if (!queue)
net/netfilter/nfnetlink_queue.c
1052
if (queue->copy_mode == NFQNL_COPY_NONE)
net/netfilter/nfnetlink_queue.c
1075
if (!skb_is_gso(skb) || ((queue->flags & NFQA_CFG_F_GSO) && !skb_is_gso_sctp(skb)))
net/netfilter/nfnetlink_queue.c
1076
return __nfqnl_enqueue_packet(net, queue, entry);
net/netfilter/nfnetlink_queue.c
1106
err = __nfqnl_enqueue_packet_gso(net, queue,
net/netfilter/nfnetlink_queue.c
1159
nfqnl_set_mode(struct nfqnl_instance *queue,
net/netfilter/nfnetlink_queue.c
1164
spin_lock_bh(&queue->lock);
net/netfilter/nfnetlink_queue.c
1168
queue->copy_mode = mode;
net/netfilter/nfnetlink_queue.c
1169
queue->copy_range = 0;
net/netfilter/nfnetlink_queue.c
1173
queue->copy_mode = mode;
net/netfilter/nfnetlink_queue.c
1175
queue->copy_range = NFQNL_MAX_COPY_RANGE;
net/netfilter/nfnetlink_queue.c
1177
queue->copy_range = range;
net/netfilter/nfnetlink_queue.c
1184
spin_unlock_bh(&queue->lock);
net/netfilter/nfnetlink_queue.c
1327
struct nfqnl_instance *queue;
net/netfilter/nfnetlink_queue.c
1329
queue = instance_lookup(q, queue_num);
net/netfilter/nfnetlink_queue.c
1330
if (!queue)
net/netfilter/nfnetlink_queue.c
1333
if (queue->peer_portid != nlportid)
net/netfilter/nfnetlink_queue.c
1336
return queue;
net/netfilter/nfnetlink_queue.c
1368
struct nfqnl_instance *queue;
net/netfilter/nfnetlink_queue.c
1372
queue = verdict_instance_lookup(q, queue_num,
net/netfilter/nfnetlink_queue.c
1374
if (IS_ERR(queue))
net/netfilter/nfnetlink_queue.c
1375
return PTR_ERR(queue);
net/netfilter/nfnetlink_queue.c
1384
spin_lock_bh(&queue->lock);
net/netfilter/nfnetlink_queue.c
1386
list_for_each_entry_safe(entry, tmp, &queue->queue_list, list) {
net/netfilter/nfnetlink_queue.c
1389
__dequeue_entry(queue, entry);
net/netfilter/nfnetlink_queue.c
1393
spin_unlock_bh(&queue->lock);
net/netfilter/nfnetlink_queue.c
1480
struct nfqnl_instance *queue;
net/netfilter/nfnetlink_queue.c
1486
queue = verdict_instance_lookup(q, queue_num,
net/netfilter/nfnetlink_queue.c
1488
if (IS_ERR(queue))
net/netfilter/nfnetlink_queue.c
1489
return PTR_ERR(queue);
net/netfilter/nfnetlink_queue.c
1497
entry = find_dequeue_entry(queue, ntohl(vhdr->id));
net/netfilter/nfnetlink_queue.c
1565
struct nfqnl_instance *queue;
net/netfilter/nfnetlink_queue.c
1621
queue = instance_lookup(q, queue_num);
net/netfilter/nfnetlink_queue.c
1622
if (queue && queue->peer_portid != NETLINK_CB(skb).portid) {
net/netfilter/nfnetlink_queue.c
1631
if (queue)
net/netfilter/nfnetlink_queue.c
1633
queue = instance_create(q, queue_num, NETLINK_CB(skb).portid);
net/netfilter/nfnetlink_queue.c
1634
if (IS_ERR(queue))
net/netfilter/nfnetlink_queue.c
1635
return PTR_ERR(queue);
net/netfilter/nfnetlink_queue.c
1638
if (!queue)
net/netfilter/nfnetlink_queue.c
1640
instance_destroy(q, queue);
net/netfilter/nfnetlink_queue.c
1650
if (!queue)
net/netfilter/nfnetlink_queue.c
1657
nfqnl_set_mode(queue, params->copy_mode,
net/netfilter/nfnetlink_queue.c
1664
spin_lock_bh(&queue->lock);
net/netfilter/nfnetlink_queue.c
1665
queue->queue_maxlen = ntohl(*queue_maxlen);
net/netfilter/nfnetlink_queue.c
1666
spin_unlock_bh(&queue->lock);
net/netfilter/nfnetlink_queue.c
1670
spin_lock_bh(&queue->lock);
net/netfilter/nfnetlink_queue.c
1671
queue->flags &= ~mask;
net/netfilter/nfnetlink_queue.c
1672
queue->flags |= flags & mask;
net/netfilter/nfnetlink_queue.c
1673
spin_unlock_bh(&queue->lock);
net/netfilter/nfnetlink_queue.c
181
static void nfqnl_flush(struct nfqnl_instance *queue, nfqnl_cmpfn cmpfn,
net/netfilter/nfnetlink_queue.c
218
__enqueue_entry(struct nfqnl_instance *queue, struct nf_queue_entry *entry)
net/netfilter/nfnetlink_queue.c
222
err = rhashtable_insert_fast(&queue->nfqnl_packet_map, &entry->hash_node,
net/netfilter/nfnetlink_queue.c
227
list_add_tail(&entry->list, &queue->queue_list);
net/netfilter/nfnetlink_queue.c
228
queue->queue_total++;
net/netfilter/nfnetlink_queue.c
234
__dequeue_entry(struct nfqnl_instance *queue, struct nf_queue_entry *entry)
net/netfilter/nfnetlink_queue.c
236
rhashtable_remove_fast(&queue->nfqnl_packet_map, &entry->hash_node,
net/netfilter/nfnetlink_queue.c
239
queue->queue_total--;
net/netfilter/nfnetlink_queue.c
243
find_dequeue_entry(struct nfqnl_instance *queue, unsigned int id)
net/netfilter/nfnetlink_queue.c
247
spin_lock_bh(&queue->lock);
net/netfilter/nfnetlink_queue.c
248
entry = rhashtable_lookup_fast(&queue->nfqnl_packet_map, &id,
net/netfilter/nfnetlink_queue.c
252
__dequeue_entry(queue, entry);
net/netfilter/nfnetlink_queue.c
254
spin_unlock_bh(&queue->lock);
net/netfilter/nfnetlink_queue.c
478
nfqnl_flush(struct nfqnl_instance *queue, nfqnl_cmpfn cmpfn, unsigned long data)
net/netfilter/nfnetlink_queue.c
482
spin_lock_bh(&queue->lock);
net/netfilter/nfnetlink_queue.c
483
list_for_each_entry_safe(entry, next, &queue->queue_list, list) {
net/netfilter/nfnetlink_queue.c
485
__dequeue_entry(queue, entry);
net/netfilter/nfnetlink_queue.c
489
spin_unlock_bh(&queue->lock);
net/netfilter/nfnetlink_queue.c
626
nfqnl_build_packet_message(struct net *net, struct nfqnl_instance *queue,
net/netfilter/nfnetlink_queue.c
679
switch ((enum nfqnl_config_mode)READ_ONCE(queue->copy_mode)) {
net/netfilter/nfnetlink_queue.c
685
if (!(queue->flags & NFQA_CFG_F_GSO) &&
net/netfilter/nfnetlink_queue.c
690
data_len = READ_ONCE(queue->copy_range);
net/netfilter/nfnetlink_queue.c
704
if (queue->flags & NFQA_CFG_F_CONNTRACK) {
net/netfilter/nfnetlink_queue.c
713
if (queue->flags & NFQA_CFG_F_UID_GID) {
net/netfilter/nfnetlink_queue.c
718
if ((queue->flags & NFQA_CFG_F_SECCTX) && entskb->sk) {
net/netfilter/nfnetlink_queue.c
735
htons(queue->queue_num));
net/netfilter/nfnetlink_queue.c
854
if ((queue->flags & NFQA_CFG_F_UID_GID) && entskb->sk &&
net/netfilter/nfnetlink_queue.c
904
__nfqnl_enqueue_packet(struct net *net, struct nfqnl_instance *queue,
net/netfilter/nfnetlink_queue.c
912
nskb = nfqnl_build_packet_message(net, queue, entry, &packet_id_ptr);
net/netfilter/nfnetlink_queue.c
917
spin_lock_bh(&queue->lock);
net/netfilter/nfnetlink_queue.c
919
if (queue->queue_total >= queue->queue_maxlen)
net/netfilter/nfnetlink_queue.c
922
entry->id = ++queue->id_sequence;
net/netfilter/nfnetlink_queue.c
926
err = __enqueue_entry(queue, entry);
net/netfilter/nfnetlink_queue.c
931
err = nfnetlink_unicast(nskb, net, queue->peer_portid);
net/netfilter/nfnetlink_queue.c
934
__dequeue_entry(queue, entry);
net/netfilter/nfnetlink_queue.c
936
if (queue->flags & NFQA_CFG_F_FAIL_OPEN) {
net/netfilter/nfnetlink_queue.c
940
queue->queue_user_dropped++;
net/netfilter/nfnetlink_queue.c
945
spin_unlock_bh(&queue->lock);
net/netfilter/nfnetlink_queue.c
949
if (queue->flags & NFQA_CFG_F_FAIL_OPEN) {
net/netfilter/nfnetlink_queue.c
953
queue->queue_dropped++;
net/netfilter/nfnetlink_queue.c
955
if (queue->queue_total >= queue->queue_maxlen)
net/netfilter/nfnetlink_queue.c
957
queue->queue_total);
net/netfilter/nfnetlink_queue.c
963
spin_unlock_bh(&queue->lock);
net/netfilter/nft_queue.c
33
u32 queue = priv->queuenum;
net/netfilter/nft_queue.c
40
queue = priv->queuenum + cpu % priv->queues_total;
net/netfilter/nft_queue.c
42
queue = nfqueue_hash(pkt->skb, queue,
net/netfilter/nft_queue.c
48
ret = NF_QUEUE_NR(queue);
net/netfilter/nft_queue.c
60
u32 queue, ret;
net/netfilter/nft_queue.c
62
queue = regs->data[priv->sreg_qnum];
net/netfilter/nft_queue.c
64
ret = NF_QUEUE_NR(queue);
net/netfilter/xt_NFQUEUE.c
103
ret = NF_QUEUE_NR(queue);
net/netfilter/xt_NFQUEUE.c
40
u32 queue = info->queuenum;
net/netfilter/xt_NFQUEUE.c
43
queue = nfqueue_hash(skb, queue, info->queues_total,
net/netfilter/xt_NFQUEUE.c
46
return NF_QUEUE_NR(queue);
net/netfilter/xt_NFQUEUE.c
89
u32 queue = info->queuenum;
net/netfilter/xt_NFQUEUE.c
96
queue = info->queuenum + cpu % info->queues_total;
net/netfilter/xt_NFQUEUE.c
98
queue = nfqueue_hash(skb, queue, info->queues_total,
net/nfc/digital_core.c
120
queue);
net/nfc/digital_core.c
126
list_del(&cmd->queue);
net/nfc/digital_core.c
164
queue);
net/nfc/digital_core.c
217
list_del(&cmd->queue);
net/nfc/digital_core.c
244
INIT_LIST_HEAD(&cmd->queue);
net/nfc/digital_core.c
247
list_add_tail(&cmd->queue, &ddev->cmd_queue);
net/nfc/digital_core.c
28
struct list_head queue;
net/nfc/digital_core.c
849
list_for_each_entry_safe(cmd, n, &ddev->cmd_queue, queue) {
net/nfc/digital_core.c
850
list_del(&cmd->queue);
net/phonet/pep.c
337
struct sk_buff_head *queue;
net/phonet/pep.c
383
queue = &pn->ctrlreq_queue;
net/phonet/pep.c
384
goto queue;
net/phonet/pep.c
405
queue = &sk->sk_receive_queue;
net/phonet/pep.c
406
goto queue;
net/phonet/pep.c
450
queue:
net/phonet/pep.c
453
skb_queue_tail(queue, skb);
net/qrtr/tun.c
121
if (!skb_queue_empty(&tun->queue))
net/qrtr/tun.c
134
skb_queue_purge(&tun->queue);
net/qrtr/tun.c
15
struct sk_buff_head queue;
net/qrtr/tun.c
23
skb_queue_tail(&tun->queue, skb);
net/qrtr/tun.c
40
skb_queue_head_init(&tun->queue);
net/qrtr/tun.c
66
while (!(skb = skb_dequeue(&tun->queue))) {
net/qrtr/tun.c
72
!skb_queue_empty(&tun->queue)))
net/rose/rose_link.c
167
while ((skbn = skb_dequeue(&neigh->queue)) != NULL)
net/rose/rose_link.c
281
skb_queue_tail(&neigh->queue, skb);
net/rose/rose_route.c
102
skb_queue_head_init(&rose_neigh->queue);
net/rose/rose_route.c
235
skb_queue_purge(&rose_neigh->queue);
net/rose/rose_route.c
387
skb_queue_head_init(&sn->queue);
net/rose/rose_route.c
778
skb_queue_purge(&rose_neigh->queue);
net/sched/sch_cbs.c
267
cbs.queue = q->queue;
net/sched/sch_cbs.c
273
cbs.queue);
net/sched/sch_cbs.c
289
cbs.queue = q->queue;
net/sched/sch_cbs.c
423
q->queue = sch->dev_queue - netdev_get_tx_queue(dev, 0);
net/sched/sch_cbs.c
72
int queue;
net/sched/sch_etf.c
310
etf.queue = q->queue;
net/sched/sch_etf.c
316
etf.queue);
net/sched/sch_etf.c
32
int queue;
net/sched/sch_etf.c
331
etf.queue = q->queue;
net/sched/sch_etf.c
379
q->queue = sch->dev_queue - netdev_get_tx_queue(dev, 0);
net/sched/sch_generic.c
824
if (!q->ring.queue)
net/sched/sch_generic.c
891
if (!q->ring.queue)
net/sched/sch_htb.c
1402
struct netdev_queue *queue;
net/sched/sch_htb.c
1404
queue = cl->leaf.offload_queue;
net/sched/sch_htb.c
1406
WARN_ON(cl->leaf.q->dev_queue != queue);
net/sched/sch_htb.c
1408
return queue;
net/sched/sch_mq.c
213
graft_offload.graft_params.queue = cl - 1;
net/sched/sch_taprio.c
2489
.queue = cl - 1,
net/sched/sch_taprio.c
632
int queue;
net/sched/sch_taprio.c
634
queue = skb_get_queue_mapping(skb);
net/sched/sch_taprio.c
636
child = q->qdiscs[queue];
net/sctp/inqueue.c
115
struct sctp_chunk *sctp_inq_pop(struct sctp_inq *queue)
net/sctp/inqueue.c
124
chunk = queue->in_progress;
net/sctp/inqueue.c
142
chunk = queue->in_progress = NULL;
net/sctp/inqueue.c
158
entry = sctp_list_dequeue(&queue->in_chunk_list);
net/sctp/inqueue.c
186
queue->in_progress = chunk;
net/sctp/inqueue.c
32
void sctp_inq_init(struct sctp_inq *queue)
net/sctp/inqueue.c
34
INIT_LIST_HEAD(&queue->in_chunk_list);
net/sctp/inqueue.c
35
queue->in_progress = NULL;
net/sctp/inqueue.c
38
INIT_WORK(&queue->immediate, NULL);
net/sctp/inqueue.c
50
void sctp_inq_free(struct sctp_inq *queue)
net/sctp/inqueue.c
55
list_for_each_entry_safe(chunk, tmp, &queue->in_chunk_list, list) {
net/sctp/inqueue.c
63
if (queue->in_progress) {
net/sctp/inqueue.c
64
sctp_inq_chunk_free(queue->in_progress);
net/sctp/inqueue.c
65
queue->in_progress = NULL;
net/sctp/inqueue.c
92
struct sctp_chunkhdr *sctp_inq_peek(struct sctp_inq *queue)
net/sctp/inqueue.c
97
chunk = queue->in_progress;
net/sctp/outqueue.c
344
struct list_head *queue, int msg_len)
net/sctp/outqueue.c
348
list_for_each_entry_safe(chk, temp, queue, transmitted_list) {
net/sctp/outqueue.c
366
if (queue != &asoc->outqueue.retransmit &&
net/sctp/socket.c
9559
struct sk_buff_head *queue;
net/sctp/socket.c
9563
queue = &newsp->pd_lobby;
net/sctp/socket.c
9565
queue = &newsk->sk_receive_queue;
net/sctp/socket.c
9574
__skb_queue_tail(queue, skb);
net/sctp/ulpqueue.c
184
struct sk_buff_head *queue;
net/sctp/ulpqueue.c
213
queue = &sk->sk_receive_queue;
net/sctp/ulpqueue.c
224
queue = &sp->pd_lobby;
net/sctp/ulpqueue.c
227
queue = &sk->sk_receive_queue;
net/sctp/ulpqueue.c
236
queue = &sk->sk_receive_queue;
net/sctp/ulpqueue.c
238
queue = &sp->pd_lobby;
net/sctp/ulpqueue.c
242
skb_queue_splice_tail_init(skb_list, queue);
net/sctp/ulpqueue.c
251
if (queue == &sk->sk_receive_queue && !sp->data_ready_signalled) {
net/sctp/ulpqueue.c
313
struct sk_buff_head *queue,
net/sctp/ulpqueue.c
357
__skb_unlink(f_frag, queue);
net/sctp/ulpqueue.c
374
__skb_unlink(pos, queue);
net/sunrpc/cache.c
1019
for (cq= &rp->q; &cq->list != &cd->queue;
net/sunrpc/cache.c
1050
list_add(&rp->q.list, &cd->queue);
net/sunrpc/cache.c
1070
for (cq = &rp->q; &cq->list != &cd->queue;
net/sunrpc/cache.c
1117
list_for_each_entry_safe(cq, tmp, &detail->queue, list)
net/sunrpc/cache.c
1257
list_add_tail(&crq->q.list, &detail->queue);
net/sunrpc/cache.c
402
INIT_LIST_HEAD(&cd->queue);
net/sunrpc/cache.c
852
while (rp->q.list.next != &cd->queue &&
net/sunrpc/cache.c
858
if (rp->q.list.next == &cd->queue) {
net/sunrpc/cache.c
993
for (cq= &rp->q; &cq->list != &cd->queue;
net/sunrpc/sched.c
101
if (list_empty(&queue->timer_list.list))
net/sunrpc/sched.c
102
cancel_delayed_work(&queue->timer_list.dwork);
net/sunrpc/sched.c
106
rpc_set_queue_timer(struct rpc_wait_queue *queue, unsigned long expires)
net/sunrpc/sched.c
109
queue->timer_list.expires = expires;
net/sunrpc/sched.c
114
mod_delayed_work(rpciod_workqueue, &queue->timer_list.dwork, expires);
net/sunrpc/sched.c
121
__rpc_add_timer(struct rpc_wait_queue *queue, struct rpc_task *task,
net/sunrpc/sched.c
125
if (list_empty(&queue->timer_list.list) || time_before(timeout, queue->timer_list.expires))
net/sunrpc/sched.c
126
rpc_set_queue_timer(queue, timeout);
net/sunrpc/sched.c
127
list_add(&task->u.tk_wait.timer_list, &queue->timer_list.list);
net/sunrpc/sched.c
130
static void rpc_set_waitqueue_priority(struct rpc_wait_queue *queue, int priority)
net/sunrpc/sched.c
132
if (queue->priority != priority) {
net/sunrpc/sched.c
133
queue->priority = priority;
net/sunrpc/sched.c
134
queue->nr = 1U << priority;
net/sunrpc/sched.c
138
static void rpc_reset_waitqueue_priority(struct rpc_wait_queue *queue)
net/sunrpc/sched.c
140
rpc_set_waitqueue_priority(queue, queue->maxpriority);
net/sunrpc/sched.c
193
static void __rpc_add_wait_queue_priority(struct rpc_wait_queue *queue,
net/sunrpc/sched.c
197
if (unlikely(queue_priority > queue->maxpriority))
net/sunrpc/sched.c
198
queue_priority = queue->maxpriority;
net/sunrpc/sched.c
199
__rpc_list_enqueue_task(&queue->tasks[queue_priority], task);
net/sunrpc/sched.c
205
static void __rpc_add_wait_queue(struct rpc_wait_queue *queue,
net/sunrpc/sched.c
210
if (RPC_IS_PRIORITY(queue))
net/sunrpc/sched.c
211
__rpc_add_wait_queue_priority(queue, task, queue_priority);
net/sunrpc/sched.c
213
list_add_tail(&task->u.tk_wait.list, &queue->tasks[0]);
net/sunrpc/sched.c
214
task->tk_waitqueue = queue;
net/sunrpc/sched.c
215
queue->qlen++;
net/sunrpc/sched.c
233
static void __rpc_remove_wait_queue(struct rpc_wait_queue *queue, struct rpc_task *task)
net/sunrpc/sched.c
235
__rpc_disable_timer(queue, task);
net/sunrpc/sched.c
236
if (RPC_IS_PRIORITY(queue))
net/sunrpc/sched.c
240
queue->qlen--;
net/sunrpc/sched.c
243
static void __rpc_init_priority_wait_queue(struct rpc_wait_queue *queue, const char *qname, unsigned char nr_queues)
net/sunrpc/sched.c
247
spin_lock_init(&queue->lock);
net/sunrpc/sched.c
248
for (i = 0; i < ARRAY_SIZE(queue->tasks); i++)
net/sunrpc/sched.c
249
INIT_LIST_HEAD(&queue->tasks[i]);
net/sunrpc/sched.c
250
queue->maxpriority = nr_queues - 1;
net/sunrpc/sched.c
251
rpc_reset_waitqueue_priority(queue);
net/sunrpc/sched.c
252
queue->qlen = 0;
net/sunrpc/sched.c
253
queue->timer_list.expires = 0;
net/sunrpc/sched.c
254
INIT_DELAYED_WORK(&queue->timer_list.dwork, __rpc_queue_timer_fn);
net/sunrpc/sched.c
255
INIT_LIST_HEAD(&queue->timer_list.list);
net/sunrpc/sched.c
256
rpc_assign_waitqueue_name(queue, qname);
net/sunrpc/sched.c
259
void rpc_init_priority_wait_queue(struct rpc_wait_queue *queue, const char *qname)
net/sunrpc/sched.c
261
__rpc_init_priority_wait_queue(queue, qname, RPC_NR_PRIORITY);
net/sunrpc/sched.c
265
void rpc_init_wait_queue(struct rpc_wait_queue *queue, const char *qname)
net/sunrpc/sched.c
267
__rpc_init_priority_wait_queue(queue, qname, 1);
net/sunrpc/sched.c
271
void rpc_destroy_wait_queue(struct rpc_wait_queue *queue)
net/sunrpc/sched.c
273
cancel_delayed_work_sync(&queue->timer_list.dwork);
net/sunrpc/sched.c
509
struct rpc_wait_queue *queue,
net/sunrpc/sched.c
518
trace_rpc_task_wakeup(task, queue);
net/sunrpc/sched.c
520
__rpc_remove_wait_queue(queue, task);
net/sunrpc/sched.c
530
struct rpc_wait_queue *queue, struct rpc_task *task,
net/sunrpc/sched.c
535
if (task->tk_waitqueue == queue) {
net/sunrpc/sched.c
537
__rpc_do_wake_up_task_on_wq(wq, queue, task);
net/sunrpc/sched.c
548
static void rpc_wake_up_task_queue_locked(struct rpc_wait_queue *queue,
net/sunrpc/sched.c
551
rpc_wake_up_task_on_wq_queue_action_locked(rpciod_workqueue, queue,
net/sunrpc/sched.c
558
void rpc_wake_up_queued_task(struct rpc_wait_queue *queue, struct rpc_task *task)
net/sunrpc/sched.c
562
spin_lock(&queue->lock);
net/sunrpc/sched.c
563
rpc_wake_up_task_queue_locked(queue, task);
net/sunrpc/sched.c
564
spin_unlock(&queue->lock);
net/sunrpc/sched.c
575
rpc_wake_up_task_queue_set_status_locked(struct rpc_wait_queue *queue,
net/sunrpc/sched.c
578
rpc_wake_up_task_on_wq_queue_action_locked(rpciod_workqueue, queue,
net/sunrpc/sched.c
592
rpc_wake_up_queued_task_set_status(struct rpc_wait_queue *queue,
net/sunrpc/sched.c
597
spin_lock(&queue->lock);
net/sunrpc/sched.c
598
rpc_wake_up_task_queue_set_status_locked(queue, task, status);
net/sunrpc/sched.c
599
spin_unlock(&queue->lock);
net/sunrpc/sched.c
605
static struct rpc_task *__rpc_find_next_queued_priority(struct rpc_wait_queue *queue)
net/sunrpc/sched.c
613
q = &queue->tasks[RPC_NR_PRIORITY - 1];
net/sunrpc/sched.c
614
if (queue->maxpriority > RPC_PRIORITY_PRIVILEGED && !list_empty(q)) {
net/sunrpc/sched.c
622
q = &queue->tasks[queue->priority];
net/sunrpc/sched.c
623
if (!list_empty(q) && queue->nr) {
net/sunrpc/sched.c
624
queue->nr--;
net/sunrpc/sched.c
633
if (q == &queue->tasks[0])
net/sunrpc/sched.c
634
q = &queue->tasks[queue->maxpriority];
net/sunrpc/sched.c
641
} while (q != &queue->tasks[queue->priority]);
net/sunrpc/sched.c
643
rpc_reset_waitqueue_priority(queue);
net/sunrpc/sched.c
647
rpc_set_waitqueue_priority(queue, (unsigned int)(q - &queue->tasks[0]));
net/sunrpc/sched.c
652
static struct rpc_task *__rpc_find_next_queued(struct rpc_wait_queue *queue)
net/sunrpc/sched.c
654
if (RPC_IS_PRIORITY(queue))
net/sunrpc/sched.c
655
return __rpc_find_next_queued_priority(queue);
net/sunrpc/sched.c
656
if (!list_empty(&queue->tasks[0]))
net/sunrpc/sched.c
657
return list_first_entry(&queue->tasks[0], struct rpc_task, u.tk_wait.list);
net/sunrpc/sched.c
665
struct rpc_wait_queue *queue,
net/sunrpc/sched.c
670
spin_lock(&queue->lock);
net/sunrpc/sched.c
671
task = __rpc_find_next_queued(queue);
net/sunrpc/sched.c
673
task = rpc_wake_up_task_on_wq_queue_action_locked(wq, queue,
net/sunrpc/sched.c
675
spin_unlock(&queue->lock);
net/sunrpc/sched.c
683
struct rpc_task *rpc_wake_up_first(struct rpc_wait_queue *queue,
net/sunrpc/sched.c
686
return rpc_wake_up_first_on_wq(rpciod_workqueue, queue, func, data);
net/sunrpc/sched.c
698
struct rpc_task *rpc_wake_up_next(struct rpc_wait_queue *queue)
net/sunrpc/sched.c
700
return rpc_wake_up_first(queue, rpc_wake_up_next_func, NULL);
net/sunrpc/sched.c
709
static void rpc_wake_up_locked(struct rpc_wait_queue *queue)
net/sunrpc/sched.c
714
task = __rpc_find_next_queued(queue);
net/sunrpc/sched.c
717
rpc_wake_up_task_queue_locked(queue, task);
net/sunrpc/sched.c
727
void rpc_wake_up(struct rpc_wait_queue *queue)
net/sunrpc/sched.c
729
spin_lock(&queue->lock);
net/sunrpc/sched.c
730
rpc_wake_up_locked(queue);
net/sunrpc/sched.c
731
spin_unlock(&queue->lock);
net/sunrpc/sched.c
740
static void rpc_wake_up_status_locked(struct rpc_wait_queue *queue, int status)
net/sunrpc/sched.c
745
task = __rpc_find_next_queued(queue);
net/sunrpc/sched.c
748
rpc_wake_up_task_queue_set_status_locked(queue, task, status);
net/sunrpc/sched.c
759
void rpc_wake_up_status(struct rpc_wait_queue *queue, int status)
net/sunrpc/sched.c
761
spin_lock(&queue->lock);
net/sunrpc/sched.c
762
rpc_wake_up_status_locked(queue, status);
net/sunrpc/sched.c
763
spin_unlock(&queue->lock);
net/sunrpc/sched.c
769
struct rpc_wait_queue *queue = container_of(work,
net/sunrpc/sched.c
775
spin_lock(&queue->lock);
net/sunrpc/sched.c
777
list_for_each_entry_safe(task, n, &queue->timer_list.list, u.tk_wait.timer_list) {
net/sunrpc/sched.c
782
rpc_wake_up_task_queue_locked(queue, task);
net/sunrpc/sched.c
788
if (!list_empty(&queue->timer_list.list))
net/sunrpc/sched.c
789
rpc_set_queue_timer(queue, expires);
net/sunrpc/sched.c
790
spin_unlock(&queue->lock);
net/sunrpc/sched.c
859
struct rpc_wait_queue *queue;
net/sunrpc/sched.c
867
queue = READ_ONCE(task->tk_waitqueue);
net/sunrpc/sched.c
868
if (queue)
net/sunrpc/sched.c
869
rpc_wake_up_queued_task(queue, task);
net/sunrpc/sched.c
874
struct rpc_wait_queue *queue;
net/sunrpc/sched.c
878
queue = READ_ONCE(task->tk_waitqueue);
net/sunrpc/sched.c
879
if (queue)
net/sunrpc/sched.c
880
rpc_wake_up_queued_task(queue, task);
net/sunrpc/sched.c
911
struct rpc_wait_queue *queue;
net/sunrpc/sched.c
95
__rpc_disable_timer(struct rpc_wait_queue *queue, struct rpc_task *task)
net/sunrpc/sched.c
966
queue = task->tk_waitqueue;
net/sunrpc/sched.c
967
spin_lock(&queue->lock);
net/sunrpc/sched.c
969
spin_unlock(&queue->lock);
net/sunrpc/sched.c
974
rpc_wake_up_task_queue_locked(queue, task);
net/sunrpc/sched.c
975
spin_unlock(&queue->lock);
net/sunrpc/sched.c
979
spin_unlock(&queue->lock);
net/tipc/link.c
1954
struct sk_buff_head *queue = &l->transmq;
net/tipc/link.c
2016
skb_queue_walk(queue, skb) {
net/tipc/link.c
2018
if (queue == &l->backlogq)
net/tipc/link.c
2064
if (queue != &l->backlogq) {
net/tipc/link.c
2065
queue = &l->backlogq;
net/tipc/topsrv.c
251
struct list_head *queue = &con->outqueue;
net/tipc/topsrv.c
262
while (!list_empty(queue)) {
net/tipc/topsrv.c
263
e = list_first_entry(queue, struct outqueue_entry, list);
net/unix/af_unix.c
2849
struct sk_buff_head *queue = &sk->sk_receive_queue;
net/unix/af_unix.c
2862
spin_lock(&queue->lock);
net/unix/af_unix.c
2864
skb = __skb_dequeue(queue);
net/unix/af_unix.c
2866
spin_unlock(&queue->lock);
net/unix/af_unix.c
2876
spin_unlock(&queue->lock);
net/unix/af_unix.c
2884
spin_unlock(&queue->lock);
net/unix/garbage.c
408
struct sk_buff_head *queue;
net/unix/garbage.c
414
queue = &u->sk.sk_receive_queue;
net/unix/garbage.c
416
spin_lock(&queue->lock);
net/unix/garbage.c
421
skb_queue_walk(queue, skb) {
net/unix/garbage.c
429
skb_queue_splice_init(queue, hitlist);
net/unix/garbage.c
432
spin_unlock(&queue->lock);
net/vmw_vsock/virtio_transport_common.c
1724
int virtio_transport_purge_skbs(void *vsk, struct sk_buff_head *queue)
net/vmw_vsock/virtio_transport_common.c
1732
spin_lock_bh(&queue->lock);
net/vmw_vsock/virtio_transport_common.c
1733
skb_queue_walk_safe(queue, skb, tmp) {
net/vmw_vsock/virtio_transport_common.c
1737
__skb_unlink(skb, queue);
net/vmw_vsock/virtio_transport_common.c
1743
spin_unlock_bh(&queue->lock);
net/x25/x25_link.c
125
while ((skbn = skb_dequeue(&nb->queue)) != NULL)
net/x25/x25_link.c
216
skb_queue_tail(&nb->queue, skb);
net/x25/x25_link.c
222
skb_queue_tail(&nb->queue, skb);
net/x25/x25_link.c
253
skb_queue_purge(&nb->queue);
net/x25/x25_link.c
270
skb_queue_head_init(&nb->queue);
net/xdp/xsk.c
1143
static int xsk_init_queue(u32 entries, struct xsk_queue **queue,
net/xdp/xsk.c
1148
if (entries == 0 || *queue || !is_power_of_2(entries))
net/xdp/xsk.c
1157
WRITE_ONCE(*queue, q);
net/xdp/xsk_diag.c
26
static int xsk_diag_put_ring(const struct xsk_queue *queue, int nl_type,
net/xdp/xsk_diag.c
31
dr.entries = queue->nentries;
net/xfrm/espintcp.c
517
struct sk_buff_head queue;
net/xfrm/espintcp.c
520
__skb_queue_head_init(&queue);
net/xfrm/espintcp.c
521
skb_queue_splice_init(&ctx->out_queue, &queue);
net/xfrm/espintcp.c
523
while ((skb = __skb_dequeue(&queue)))
net/xfrm/xfrm_input.c
31
struct sk_buff_head queue;
net/xfrm/xfrm_input.c
791
struct sk_buff_head queue;
net/xfrm/xfrm_input.c
794
__skb_queue_head_init(&queue);
net/xfrm/xfrm_input.c
796
skb_queue_splice_init(&trans->queue, &queue);
net/xfrm/xfrm_input.c
800
while ((skb = __skb_dequeue(&queue)))
net/xfrm/xfrm_input.c
814
if (skb_queue_len(&trans->queue) >= READ_ONCE(net_hotdata.max_backlog))
net/xfrm/xfrm_input.c
822
__skb_queue_tail(&trans->queue, skb);
net/xfrm/xfrm_input.c
855
__skb_queue_head_init(&trans->queue);
net/xfrm/xfrm_iptfs.c
154
struct sk_buff_head queue; /* output queue */
net/xfrm/xfrm_iptfs.c
1787
__skb_queue_tail(&xtfs->queue, skb);
net/xfrm/xfrm_iptfs.c
2336
skb_queue_splice_init(&xtfs->queue, &list);
net/xfrm/xfrm_iptfs.c
2638
__skb_queue_head_init(&xtfs->queue);
net/xfrm/xfrm_iptfs.c
2714
skb_queue_splice_init(&xtfs->queue, &list);
samples/v4l/v4l2-pci-skeleton.c
368
if (vb2_is_busy(&skel->queue))
samples/v4l/v4l2-pci-skeleton.c
415
if (vb2_is_busy(&skel->queue))
samples/v4l/v4l2-pci-skeleton.c
499
if (vb2_is_busy(&skel->queue))
samples/v4l/v4l2-pci-skeleton.c
618
if (vb2_is_busy(&skel->queue))
samples/v4l/v4l2-pci-skeleton.c
63
struct vb2_queue queue;
samples/v4l/v4l2-pci-skeleton.c
806
q = &skel->queue;
samples/v4l/v4l2-pci-skeleton.c
861
vdev->queue = q;
security/integrity/ima/ima.h
303
struct hlist_head queue[IMA_MEASURE_HTABLE_SIZE];
security/integrity/ima/ima_queue.c
119
hlist_add_head_rcu(&qe->hnext, &ima_htable.queue[key]);
security/integrity/ima/ima_queue.c
39
.queue[0 ... IMA_MEASURE_HTABLE_SIZE - 1] = HLIST_HEAD_INIT
security/integrity/ima/ima_queue.c
64
hlist_for_each_entry_rcu(qe, &ima_htable.queue[key], hnext) {
sound/core/seq/oss/seq_oss_device.h
148
ev->queue = dp->queue;
sound/core/seq/oss/seq_oss_device.h
74
int queue; /* sequencer queue number */
sound/core/seq/oss/seq_oss_init.c
177
dp->queue = -1;
sound/core/seq/oss/seq_oss_init.c
267
delete_seq_queue(dp->queue);
sound/core/seq/oss/seq_oss_init.c
354
dp->queue = qinfo.queue;
sound/core/seq/oss/seq_oss_init.c
362
delete_seq_queue(int queue)
sound/core/seq/oss/seq_oss_init.c
367
if (queue < 0)
sound/core/seq/oss/seq_oss_init.c
370
qinfo.queue = queue;
sound/core/seq/oss/seq_oss_init.c
373
pr_err("ALSA: seq_oss: unable to delete queue %d (%d)\n", queue, rc);
sound/core/seq/oss/seq_oss_init.c
402
int queue;
sound/core/seq/oss/seq_oss_init.c
413
queue = dp->queue;
sound/core/seq/oss/seq_oss_init.c
416
delete_seq_queue(queue);
sound/core/seq/oss/seq_oss_init.c
45
static int delete_seq_queue(int queue);
sound/core/seq/oss/seq_oss_init.c
482
snd_iprintf(buf, "port %d : queue %d\n", dp->port, dp->queue);
sound/core/seq/oss/seq_oss_midi.c
349
subs.queue = dp->queue; /* queue for timestamps */
sound/core/seq/oss/seq_oss_midi.c
440
ev.queue = dp->queue;
sound/core/seq/oss/seq_oss_timer.c
137
ev.queue = dp->queue;
sound/core/seq/oss/seq_oss_timer.c
138
ev.data.queue.queue = dp->queue;
sound/core/seq/oss/seq_oss_timer.c
139
ev.data.queue.param.value = value;
sound/core/seq/oss/seq_oss_timer.c
156
tmprec.queue = dp->queue;
sound/core/seq/seq_clientmgr.c
1529
info->queue = q->queue;
sound/core/seq/seq_clientmgr.c
1535
snprintf(info->name, sizeof(info->name), "Queue-%d", q->queue);
sound/core/seq/seq_clientmgr.c
1546
return snd_seq_queue_delete(client->number, info->queue);
sound/core/seq/seq_clientmgr.c
1555
queueptr(info->queue);
sound/core/seq/seq_clientmgr.c
1561
info->queue = q->queue;
sound/core/seq/seq_clientmgr.c
1579
if (snd_seq_queue_check_access(info->queue, client->number)) {
sound/core/seq/seq_clientmgr.c
1580
if (snd_seq_queue_set_owner(info->queue, client->number, info->locked) < 0)
sound/core/seq/seq_clientmgr.c
1583
snd_seq_queue_use(info->queue, client->number, 1);
sound/core/seq/seq_clientmgr.c
1589
queueptr(info->queue);
sound/core/seq/seq_clientmgr.c
1609
info->queue = q->queue;
sound/core/seq/seq_clientmgr.c
1622
struct snd_seq_queue *queue __free(snd_seq_queue) =
sound/core/seq/seq_clientmgr.c
1623
queueptr(status->queue);
sound/core/seq/seq_clientmgr.c
1625
if (queue == NULL)
sound/core/seq/seq_clientmgr.c
1628
status->queue = queue->queue;
sound/core/seq/seq_clientmgr.c
1630
tmr = queue->timer;
sound/core/seq/seq_clientmgr.c
1631
status->events = queue->tickq->cells + queue->timeq->cells;
sound/core/seq/seq_clientmgr.c
1638
status->flags = queue->flags;
sound/core/seq/seq_clientmgr.c
1650
struct snd_seq_queue *queue __free(snd_seq_queue) =
sound/core/seq/seq_clientmgr.c
1651
queueptr(tempo->queue);
sound/core/seq/seq_clientmgr.c
1653
if (queue == NULL)
sound/core/seq/seq_clientmgr.c
1656
tempo->queue = queue->queue;
sound/core/seq/seq_clientmgr.c
1658
tmr = queue->timer;
sound/core/seq/seq_clientmgr.c
1674
if (!snd_seq_queue_check_access(tempo->queue, client))
sound/core/seq/seq_clientmgr.c
1676
return snd_seq_queue_timer_set_tempo(tempo->queue, client, tempo);
sound/core/seq/seq_clientmgr.c
1699
struct snd_seq_queue *queue __free(snd_seq_queue) =
sound/core/seq/seq_clientmgr.c
1700
queueptr(timer->queue);
sound/core/seq/seq_clientmgr.c
1702
if (queue == NULL)
sound/core/seq/seq_clientmgr.c
1705
guard(mutex)(&queue->timer_mutex);
sound/core/seq/seq_clientmgr.c
1706
tmr = queue->timer;
sound/core/seq/seq_clientmgr.c
1708
timer->queue = queue->queue;
sound/core/seq/seq_clientmgr.c
1730
if (snd_seq_queue_check_access(timer->queue, client->number)) {
sound/core/seq/seq_clientmgr.c
1733
queueptr(timer->queue);
sound/core/seq/seq_clientmgr.c
1739
snd_seq_queue_timer_close(timer->queue);
sound/core/seq/seq_clientmgr.c
1745
result = snd_seq_queue_timer_open(timer->queue);
sound/core/seq/seq_clientmgr.c
1761
used = snd_seq_queue_is_used(info->queue, client->number);
sound/core/seq/seq_clientmgr.c
1779
err = snd_seq_queue_use(info->queue, client->number, info->used);
sound/core/seq/seq_clientmgr.c
1952
subs->queue = s->info.queue;
sound/core/seq/seq_clientmgr.c
2394
ev->queue = SNDRV_SEQ_QUEUE_DIRECT;
sound/core/seq/seq_clientmgr.c
2523
snd_iprintf(buffer, "[%c:%d]", ((s->info.flags & SNDRV_SEQ_PORT_SUBS_TIME_REAL) ? 'r' : 't'), s->info.queue);
sound/core/seq/seq_clientmgr.c
542
bounce_ev.queue = SNDRV_SEQ_QUEUE_DIRECT;
sound/core/seq/seq_clientmgr.c
566
int queue, int real_time)
sound/core/seq/seq_clientmgr.c
569
queueptr(queue);
sound/core/seq/seq_clientmgr.c
573
event->queue = queue;
sound/core/seq/seq_clientmgr.c
706
update_timestamp_of_queue(event, subs->info.queue,
sound/core/seq/seq_clientmgr.c
784
if (event->queue == SNDRV_SEQ_ADDRESS_SUBSCRIBERS ||
sound/core/seq/seq_clientmgr.c
885
if (event->queue == SNDRV_SEQ_ADDRESS_SUBSCRIBERS) {
sound/core/seq/seq_clientmgr.c
887
event->queue = SNDRV_SEQ_QUEUE_DIRECT;
sound/core/seq/seq_clientmgr.c
905
if (snd_seq_queue_is_used(event->queue, client->number) <= 0)
sound/core/seq/seq_ports.c
477
return r->queue == s->queue;
sound/core/seq/seq_queue.c
100
q->queue = -1;
sound/core/seq/seq_queue.c
153
static void queue_use(struct snd_seq_queue *queue, int client, int use);
sound/core/seq/seq_queue.c
292
dest = cell->event.queue; /* destination queue */
sound/core/seq/seq_queue.c
409
struct snd_seq_queue *queue __free(snd_seq_queue) =
sound/core/seq/seq_queue.c
412
if (queue == NULL)
sound/core/seq/seq_queue.c
414
tmr = queue->timer;
sound/core/seq/seq_queue.c
415
result = snd_seq_timer_open(queue);
sound/core/seq/seq_queue.c
418
result = snd_seq_timer_open(queue);
sound/core/seq/seq_queue.c
429
struct snd_seq_queue *queue __free(snd_seq_queue) =
sound/core/seq/seq_queue.c
432
if (queue == NULL)
sound/core/seq/seq_queue.c
434
snd_seq_timer_close(queue);
sound/core/seq/seq_queue.c
460
static void queue_use(struct snd_seq_queue *queue, int client, int use)
sound/core/seq/seq_queue.c
463
if (!test_and_set_bit(client, queue->clients_bitmap))
sound/core/seq/seq_queue.c
464
queue->clients++;
sound/core/seq/seq_queue.c
466
if (test_and_clear_bit(client, queue->clients_bitmap))
sound/core/seq/seq_queue.c
467
queue->clients--;
sound/core/seq/seq_queue.c
469
if (queue->clients) {
sound/core/seq/seq_queue.c
470
if (use && queue->clients == 1)
sound/core/seq/seq_queue.c
471
snd_seq_timer_defaults(queue->timer);
sound/core/seq/seq_queue.c
472
snd_seq_timer_open(queue);
sound/core/seq/seq_queue.c
474
snd_seq_timer_close(queue);
sound/core/seq/seq_queue.c
484
struct snd_seq_queue *queue __free(snd_seq_queue) =
sound/core/seq/seq_queue.c
487
if (queue == NULL)
sound/core/seq/seq_queue.c
489
guard(mutex)(&queue->timer_mutex);
sound/core/seq/seq_queue.c
490
queue_use(queue, client, use);
sound/core/seq/seq_queue.c
537
snd_seq_queue_use(q->queue, client, 0);
sound/core/seq/seq_queue.c
557
q->queue == info->queue)) {
sound/core/seq/seq_queue.c
578
sev.queue = q->queue;
sound/core/seq/seq_queue.c
579
sev.data.queue.queue = q->queue;
sound/core/seq/seq_queue.c
58
q->queue = i;
sound/core/seq/seq_queue.c
615
snd_seq_timer_set_tempo(q->timer, ev->data.queue.param.value);
sound/core/seq/seq_queue.c
620
if (snd_seq_timer_set_position_tick(q->timer, ev->data.queue.param.time.tick) == 0) {
sound/core/seq/seq_queue.c
626
if (snd_seq_timer_set_position_time(q->timer, ev->data.queue.param.time.time) == 0) {
sound/core/seq/seq_queue.c
632
ev->data.queue.param.skew.value,
sound/core/seq/seq_queue.c
633
ev->data.queue.param.skew.base) == 0) {
sound/core/seq/seq_queue.c
651
queueptr(ev->data.queue.queue);
sound/core/seq/seq_queue.c
694
snd_iprintf(buffer, "queue %d: [%s]\n", q->queue, q->name);
sound/core/seq/seq_queue.h
20
int queue; /* queue number */
sound/core/seq/seq_timer.c
255
sprintf(str, "sequencer queue %i", q->queue);
sound/core/seq/seq_timer.c
266
err = snd_timer_open(t, &tmr->alsa_id, q->queue);
sound/core/seq/seq_timer.c
276
err = snd_timer_open(t, &tid, q->queue);
sound/core/seq/seq_timer.c
459
snd_iprintf(buffer, "Timer for queue %i : %s\n", q->queue, ti->timer->name);
sound/core/timer.c
110
unsigned int queue; /* used queue size */
sound/core/timer.c
1306
r = &tu->queue[prev];
sound/core/timer.c
1315
r = &tu->queue[tu->qtail++];
sound/core/timer.c
1431
struct snd_timer_read *queue = NULL;
sound/core/timer.c
1439
queue = kzalloc_objs(*queue, size);
sound/core/timer.c
1440
if (!queue)
sound/core/timer.c
1445
kfree(tu->queue);
sound/core/timer.c
1448
tu->queue = queue;
sound/core/timer.c
1493
kfree(tu->queue);
sound/core/timer.c
1869
struct snd_timer_read *r = &tu->queue[0];
sound/core/timer.c
1903
status.queue = tu->qused;
sound/core/timer.c
1925
status.queue = tu->qused;
sound/core/timer.c
2394
if (copy_to_user(buffer, &tu->queue[qhead],
sound/core/timer.c
81
struct snd_timer_read *queue;
sound/core/timer.c
98
unsigned int queue; /* used queue size */
sound/firewire/tascam/amdtp-tascam.c
159
&tscm->queue[tscm->push_pos];
sound/firewire/tascam/tascam-hwdep.c
44
struct snd_firewire_tascam_change *entries = tscm->queue;
sound/firewire/tascam/tascam.h
95
struct snd_firewire_tascam_change queue[SND_TSCM_QUEUE_COUNT];
sound/oss/dmasound/dmasound.h
237
#define WAKE_UP(queue) (wake_up_interruptible(&queue))
sound/usb/fcp.c
1054
init_waitqueue_head(&private->notify.queue);
sound/usb/fcp.c
74
wait_queue_head_t queue;
sound/usb/fcp.c
826
ret = wait_event_interruptible(private->notify.queue,
sound/usb/fcp.c
850
poll_wait(file, &private->notify.queue, wait);
sound/usb/fcp.c
949
wake_up_interruptible(&private->notify.queue);
sound/virtio/virtio_card.c
159
struct virtio_snd_queue *queue = virtsnd_event_queue(snd);
sound/virtio/virtio_card.c
161
if (!virtqueue_enable_cb(queue->vqueue))
sound/virtio/virtio_card.c
162
virtsnd_event_notify_cb(queue->vqueue);
sound/virtio/virtio_card.c
173
struct virtio_snd_queue *queue = virtsnd_event_queue(snd);
sound/virtio/virtio_card.c
177
if (queue->vqueue) {
sound/virtio/virtio_card.c
178
guard(spinlock_irqsave)(&queue->lock);
sound/virtio/virtio_card.c
179
virtqueue_disable_cb(queue->vqueue);
sound/virtio/virtio_card.c
180
while ((event = virtqueue_get_buf(queue->vqueue, &length)))
sound/virtio/virtio_card.c
85
struct virtio_snd_queue *queue = virtsnd_event_queue(snd);
sound/virtio/virtio_card.c
89
guard(spinlock_irqsave)(&queue->lock);
sound/virtio/virtio_ctl_msg.c
126
struct virtio_snd_queue *queue = virtsnd_control_queue(snd);
sound/virtio/virtio_ctl_msg.c
149
scoped_guard(spinlock_irqsave, &queue->lock) {
sound/virtio/virtio_ctl_msg.c
150
rc = virtqueue_add_sgs(queue->vqueue, psgs, nouts, nins, msg,
sound/virtio/virtio_ctl_msg.c
153
notify = virtqueue_kick_prepare(queue->vqueue);
sound/virtio/virtio_ctl_msg.c
173
virtqueue_notify(queue->vqueue);
sound/virtio/virtio_ctl_msg.c
234
struct virtio_snd_queue *queue = virtsnd_control_queue(snd);
sound/virtio/virtio_ctl_msg.c
236
guard(spinlock_irqsave)(&queue->lock);
sound/virtio/virtio_ctl_msg.c
293
struct virtio_snd_queue *queue = virtsnd_control_queue(snd);
sound/virtio/virtio_ctl_msg.c
297
guard(spinlock_irqsave)(&queue->lock);
sound/virtio/virtio_pcm_msg.c
343
static inline void virtsnd_pcm_notify_cb(struct virtio_snd_queue *queue)
sound/virtio/virtio_pcm_msg.c
348
guard(spinlock_irqsave)(&queue->lock);
sound/virtio/virtio_pcm_msg.c
350
virtqueue_disable_cb(queue->vqueue);
sound/virtio/virtio_pcm_msg.c
351
while ((msg = virtqueue_get_buf(queue->vqueue, &written_bytes)))
sound/virtio/virtio_pcm_msg.c
353
} while (!virtqueue_enable_cb(queue->vqueue));
sound/virtio/virtio_pcm_ops.c
328
struct virtio_snd_queue *queue;
sound/virtio/virtio_pcm_ops.c
335
queue = virtsnd_pcm_queue(vss);
sound/virtio/virtio_pcm_ops.c
337
scoped_guard(spinlock_irqsave, &queue->lock) {
sound/virtio/virtio_pcm_ops.c
478
struct virtio_snd_queue *queue = virtsnd_pcm_queue(vss);
sound/virtio/virtio_pcm_ops.c
480
guard(spinlock_irqsave)(&queue->lock);
sound/virtio/virtio_pcm_ops.c
490
struct virtio_snd_queue *queue = virtsnd_pcm_queue(vss);
sound/virtio/virtio_pcm_ops.c
492
guard(spinlock_irqsave)(&queue->lock);
tools/perf/trace/beauty/include/uapi/sound/asound.h
977
unsigned int queue; /* used queue size */
tools/perf/util/annotate.c
1228
struct annotation_line *pos, *queue = NULL;
tools/perf/util/annotate.c
1276
if (context && queue == NULL) {
tools/perf/util/annotate.c
1277
queue = pos;
tools/perf/util/annotate.c
1281
err = annotation_line__print(pos, &apd, opts, printed, queue);
tools/perf/util/annotate.c
1288
queue = NULL;
tools/perf/util/annotate.c
1305
queue = list_entry(queue->node.next, typeof(*queue), node);
tools/perf/util/annotate.c
2986
struct list_head queue;
tools/perf/util/annotate.c
3023
if (basic_block_has_offset(&bb_data->queue, offset))
tools/perf/util/annotate.c
3058
list_add_tail(&link->node, &bb_data->queue);
tools/perf/util/annotate.c
3150
list_for_each_entry_safe(link, tmp, &bb_data->queue, node) {
tools/perf/util/annotate.c
3179
.queue = LIST_HEAD_INIT(bb_data.queue),
tools/perf/util/annotate.c
3194
while (!list_empty(&bb_data.queue)) {
tools/perf/util/annotate.c
3195
link = list_first_entry(&bb_data.queue, struct basic_block_link, node);
tools/perf/util/annotate.c
785
struct annotation_line *queue)
tools/perf/util/annotate.c
818
if (queue != NULL) {
tools/perf/util/annotate.c
824
list_for_each_entry_from(queue, ¬es->src->source, node) {
tools/perf/util/annotate.c
825
if (queue == al)
tools/perf/util/annotate.c
827
annotation_line__print(queue, apd, &queue_opts,
tools/perf/util/annotate.c
897
if (queue)
tools/perf/util/arm-spe.c
1204
struct auxtrace_queue *queue,
tools/perf/util/arm-spe.c
1207
struct arm_spe_queue *speq = queue->priv;
tools/perf/util/arm-spe.c
1210
if (list_empty(&queue->head) || speq)
tools/perf/util/arm-spe.c
1218
queue->priv = speq;
tools/perf/util/arm-spe.c
1220
if (queue->cpu != -1)
tools/perf/util/arm-spe.c
1221
speq->cpu = queue->cpu;
tools/perf/util/arm-spe.c
1299
struct auxtrace_queue *queue;
tools/perf/util/arm-spe.c
1309
queue = &spe->queues.queue_array[queue_nr];
tools/perf/util/arm-spe.c
1310
speq = queue->priv;
tools/perf/util/arm-spe.c
1327
arm_spe_set_pid_tid_cpu(spe, queue);
tools/perf/util/arm-spe.c
1355
struct auxtrace_queue *queue = &spe->queues.queue_array[i];
tools/perf/util/arm-spe.c
1356
struct arm_spe_queue *speq = queue->priv;
tools/perf/util/arm-spe.c
1360
arm_spe_set_pid_tid_cpu(spe, queue);
tools/perf/util/arm-spe.c
188
struct auxtrace_queue *queue;
tools/perf/util/arm-spe.c
190
queue = &speq->spe->queues.queue_array[speq->queue_nr];
tools/perf/util/arm-spe.c
192
buffer = auxtrace_buffer__next(queue, buffer);
tools/perf/util/arm-spe.c
286
struct auxtrace_queue *queue)
tools/perf/util/arm-spe.c
288
struct arm_spe_queue *speq = queue->priv;
tools/perf/util/arm-spe.c
296
speq->tid = queue->tid;
tools/perf/util/arm-spe.c
305
if (queue->cpu == -1)
tools/perf/util/auxtrace.c
1097
struct auxtrace_buffer *auxtrace_buffer__next(struct auxtrace_queue *queue,
tools/perf/util/auxtrace.c
1101
if (list_is_last(&buffer->list, &queue->head))
tools/perf/util/auxtrace.c
1106
if (list_empty(&queue->head))
tools/perf/util/auxtrace.c
1108
return list_entry(queue->head.next, struct auxtrace_buffer,
tools/perf/util/auxtrace.c
310
struct auxtrace_queue *queue;
tools/perf/util/auxtrace.c
319
queue = &queues->queue_array[idx];
tools/perf/util/auxtrace.c
321
if (!queue->set) {
tools/perf/util/auxtrace.c
322
queue->set = true;
tools/perf/util/auxtrace.c
323
queue->tid = buffer->tid;
tools/perf/util/auxtrace.c
324
queue->cpu = buffer->cpu.cpu;
tools/perf/util/auxtrace.c
329
list_add_tail(&buffer->list, &queue->head);
tools/perf/util/auxtrace.h
537
struct auxtrace_buffer *auxtrace_buffer__next(struct auxtrace_queue *queue,
tools/perf/util/cs-etm-decoder/cs-etm-decoder.c
501
struct cs_etm_packet_queue *queue,
tools/perf/util/cs-etm-decoder/cs-etm-decoder.c
508
cs_etm_decoder__reset_timestamp(queue);
tools/perf/util/cs-etm-decoder/cs-etm-decoder.c
509
return cs_etm_decoder__buffer_packet(etmq, queue, trace_chan_id,
tools/perf/util/cs-etm-decoder/cs-etm-decoder.c
515
struct cs_etm_packet_queue *queue,
tools/perf/util/cs-etm-decoder/cs-etm-decoder.c
521
ret = cs_etm_decoder__buffer_packet(etmq, queue, trace_chan_id,
tools/perf/util/cs-etm-decoder/cs-etm-decoder.c
526
packet = &queue->packet_buffer[queue->tail];
tools/perf/util/cs-etm-decoder/cs-etm-decoder.c
534
struct cs_etm_packet_queue *queue,
tools/perf/util/cs-etm-decoder/cs-etm-decoder.c
537
return cs_etm_decoder__buffer_packet(etmq, queue, trace_chan_id,
tools/perf/util/cs-etm.c
1190
struct auxtrace_queue *queue,
tools/perf/util/cs-etm.c
1193
struct cs_etm_queue *etmq = queue->priv;
tools/perf/util/cs-etm.c
1203
queue->priv = etmq;
tools/perf/util/cs-etm.c
1206
queue->cpu = queue_nr; /* Placeholder, may be reset to -1 in per-thread mode */
tools/perf/util/cs-etm.c
1438
struct auxtrace_queue *queue;
tools/perf/util/cs-etm.c
1440
queue = &etmq->etm->queues.queue_array[etmq->queue_nr];
tools/perf/util/cs-etm.c
1442
aux_buffer = auxtrace_buffer__next(queue, aux_buffer);
tools/perf/util/cs-etm.c
2591
struct auxtrace_queue *queue = &etm->queues.queue_array[i];
tools/perf/util/cs-etm.c
2592
struct cs_etm_queue *etmq = queue->priv;
tools/perf/util/cs-etm.c
2620
struct auxtrace_queue *queue;
tools/perf/util/cs-etm.c
2646
queue = &etm->queues.queue_array[queue_nr];
tools/perf/util/cs-etm.c
2647
etmq = queue->priv;
tools/perf/util/cs-etm.c
567
static void cs_etm__clear_packet_queue(struct cs_etm_packet_queue *queue)
tools/perf/util/cs-etm.c
571
queue->head = 0;
tools/perf/util/cs-etm.c
572
queue->tail = 0;
tools/perf/util/cs-etm.c
573
queue->packet_count = 0;
tools/perf/util/cs-etm.c
575
queue->packet_buffer[i].isa = CS_ETM_ISA_UNKNOWN;
tools/perf/util/cs-etm.c
576
queue->packet_buffer[i].start_addr = CS_ETM_INVAL_ADDR;
tools/perf/util/cs-etm.c
577
queue->packet_buffer[i].end_addr = CS_ETM_INVAL_ADDR;
tools/perf/util/cs-etm.c
578
queue->packet_buffer[i].instr_count = 0;
tools/perf/util/cs-etm.c
579
queue->packet_buffer[i].last_instr_taken_branch = false;
tools/perf/util/cs-etm.c
580
queue->packet_buffer[i].last_instr_size = 0;
tools/perf/util/cs-etm.c
581
queue->packet_buffer[i].last_instr_type = 0;
tools/perf/util/cs-etm.c
582
queue->packet_buffer[i].last_instr_subtype = 0;
tools/perf/util/cs-etm.c
583
queue->packet_buffer[i].last_instr_cond = 0;
tools/perf/util/cs-etm.c
584
queue->packet_buffer[i].flags = 0;
tools/perf/util/cs-etm.c
585
queue->packet_buffer[i].exception_number = UINT32_MAX;
tools/perf/util/cs-etm.c
586
queue->packet_buffer[i].trace_chan_id = UINT8_MAX;
tools/perf/util/cs-etm.c
587
queue->packet_buffer[i].cpu = INT_MIN;
tools/perf/util/cs-etm.c
610
struct auxtrace_queue *queue;
tools/perf/util/cs-etm.c
615
queue = &etmq->etm->queues.queue_array[etmq->queue_nr];
tools/perf/util/cs-etm.c
619
queue->tid);
tools/perf/util/intel-bts.c
167
struct auxtrace_queue *queue,
tools/perf/util/intel-bts.c
170
struct intel_bts_queue *btsq = queue->priv;
tools/perf/util/intel-bts.c
172
if (list_empty(&queue->head))
tools/perf/util/intel-bts.c
179
queue->priv = btsq;
tools/perf/util/intel-bts.c
181
if (queue->cpu != -1)
tools/perf/util/intel-bts.c
182
btsq->cpu = queue->cpu;
tools/perf/util/intel-bts.c
183
btsq->tid = queue->tid;
tools/perf/util/intel-bts.c
192
btsq->buffer = auxtrace_buffer__next(queue, NULL);
tools/perf/util/intel-bts.c
248
static int intel_bts_do_fix_overlap(struct auxtrace_queue *queue,
tools/perf/util/intel-bts.c
254
if (b->list.prev == &queue->head)
tools/perf/util/intel-bts.c
450
struct auxtrace_queue *queue;
tools/perf/util/intel-bts.c
467
queue = &btsq->bts->queues.queue_array[btsq->queue_nr];
tools/perf/util/intel-bts.c
470
buffer = auxtrace_buffer__next(queue, NULL);
tools/perf/util/intel-bts.c
496
intel_bts_do_fix_overlap(queue, buffer)) {
tools/perf/util/intel-bts.c
511
btsq->buffer = auxtrace_buffer__next(queue, buffer);
tools/perf/util/intel-bts.c
545
struct auxtrace_queue *queue = &bts->queues.queue_array[i];
tools/perf/util/intel-bts.c
546
struct intel_bts_queue *btsq = queue->priv;
tools/perf/util/intel-bts.c
558
struct auxtrace_queue *queue;
tools/perf/util/intel-bts.c
570
queue = &bts->queues.queue_array[queue_nr];
tools/perf/util/intel-bts.c
571
btsq = queue->priv;
tools/perf/util/intel-pt.c
1423
struct auxtrace_queue *queue = &pt->queues.queue_array[i];
tools/perf/util/intel-pt.c
1424
struct intel_pt_queue *ptq = queue->priv;
tools/perf/util/intel-pt.c
1481
struct auxtrace_queue *queue)
tools/perf/util/intel-pt.c
1483
struct intel_pt_queue *ptq = queue->priv;
tools/perf/util/intel-pt.c
1485
if (queue->tid == -1 || pt->have_sched_switch) {
tools/perf/util/intel-pt.c
1497
if (queue->cpu == -1)
tools/perf/util/intel-pt.c
1576
struct auxtrace_queue *queue,
tools/perf/util/intel-pt.c
1579
struct intel_pt_queue *ptq = queue->priv;
tools/perf/util/intel-pt.c
1581
if (list_empty(&queue->head))
tools/perf/util/intel-pt.c
1588
queue->priv = ptq;
tools/perf/util/intel-pt.c
1590
if (queue->cpu != -1)
tools/perf/util/intel-pt.c
1591
ptq->cpu = queue->cpu;
tools/perf/util/intel-pt.c
1592
ptq->tid = queue->tid;
tools/perf/util/intel-pt.c
2752
struct auxtrace_queue *queue;
tools/perf/util/intel-pt.c
2763
queue = &pt->queues.queue_array[ptq->queue_nr];
tools/perf/util/intel-pt.c
2764
intel_pt_set_pid_tid_cpu(pt, queue);
tools/perf/util/intel-pt.c
3023
struct auxtrace_queue *queue = &pt->queues.queue_array[i];
tools/perf/util/intel-pt.c
3024
struct intel_pt_queue *ptq = queue->priv;
tools/perf/util/intel-pt.c
3038
struct auxtrace_queue *queue = &pt->queues.queue_array[i];
tools/perf/util/intel-pt.c
3039
struct intel_pt_queue *ptq = queue->priv;
tools/perf/util/intel-pt.c
3220
struct auxtrace_queue *queue;
tools/perf/util/intel-pt.c
3230
queue = &pt->queues.queue_array[queue_nr];
tools/perf/util/intel-pt.c
3231
ptq = queue->priv;
tools/perf/util/intel-pt.c
3247
intel_pt_set_pid_tid_cpu(pt, queue);
tools/perf/util/intel-pt.c
3276
struct auxtrace_queue *queue = &pt->queues.queue_array[i];
tools/perf/util/intel-pt.c
3277
struct intel_pt_queue *ptq = queue->priv;
tools/perf/util/intel-pt.c
3281
intel_pt_set_pid_tid_cpu(pt, queue);
tools/perf/util/intel-pt.c
3289
struct auxtrace_queue *queue,
tools/perf/util/intel-pt.c
3296
ptq->cpu = queue->cpu;
tools/perf/util/intel-pt.c
3319
struct auxtrace_queue *queue;
tools/perf/util/intel-pt.c
3323
queue = auxtrace_queues__sample_queue(&pt->queues, sample, pt->session);
tools/perf/util/intel-pt.c
3324
if (!queue)
tools/perf/util/intel-pt.c
3327
ptq = queue->priv;
tools/perf/util/intel-pt.c
3333
intel_pt_sample_set_pid_tid_cpu(ptq, queue, sample);
tools/perf/util/intel-pt.c
3612
struct auxtrace_queue *queue;
tools/perf/util/intel-pt.c
3616
queue = auxtrace_queues__sample_queue(&pt->queues, sample, pt->session);
tools/perf/util/intel-pt.c
3618
if (!queue || !queue->priv || !evsel || hw_id > INTEL_PT_MAX_PEBS) {
tools/perf/util/intel-pt.c
3623
ptq = queue->priv;
tools/perf/util/intel-pt.c
458
struct auxtrace_queue *queue;
tools/perf/util/intel-pt.c
461
queue = &ptq->pt->queues.queue_array[ptq->queue_nr];
tools/perf/util/intel-pt.c
466
buffer = auxtrace_buffer__next(queue, buffer);
tools/perf/util/intel-pt.c
503
struct auxtrace_queue *queue;
tools/perf/util/intel-pt.c
511
queue = &ptq->pt->queues.queue_array[ptq->queue_nr];
tools/perf/util/intel-pt.c
513
buffer = auxtrace_buffer__next(queue, buffer);
tools/perf/util/powerpc-vpadtl.c
213
struct auxtrace_queue *queue;
tools/perf/util/powerpc-vpadtl.c
215
queue = &queues->queue_array[vpaq->queue_nr];
tools/perf/util/powerpc-vpadtl.c
216
buffer = auxtrace_buffer__next(queue, buffer);
tools/perf/util/powerpc-vpadtl.c
377
struct auxtrace_queue *queue;
tools/perf/util/powerpc-vpadtl.c
387
queue = &vpa->queues.queue_array[queue_nr];
tools/perf/util/powerpc-vpadtl.c
388
vpaq = queue->priv;
tools/perf/util/powerpc-vpadtl.c
448
struct auxtrace_queue *queue,
tools/perf/util/powerpc-vpadtl.c
451
struct powerpc_vpadtl_queue *vpaq = queue->priv;
tools/perf/util/powerpc-vpadtl.c
453
if (list_empty(&queue->head) || vpaq)
tools/perf/util/powerpc-vpadtl.c
460
queue->priv = vpaq;
tools/perf/util/powerpc-vpadtl.c
462
if (queue->cpu != -1)
tools/perf/util/powerpc-vpadtl.c
463
vpaq->cpu = queue->cpu;
tools/perf/util/s390-cpumsf.c
700
struct auxtrace_queue *queue;
tools/perf/util/s390-cpumsf.c
703
queue = &sfq->sf->queues.queue_array[sfq->queue_nr];
tools/perf/util/s390-cpumsf.c
712
sfq->buffer = buffer = auxtrace_buffer__next(queue,
tools/perf/util/s390-cpumsf.c
802
struct auxtrace_queue *queue,
tools/perf/util/s390-cpumsf.c
805
struct s390_cpumsf_queue *sfq = queue->priv;
tools/perf/util/s390-cpumsf.c
807
if (list_empty(&queue->head))
tools/perf/util/s390-cpumsf.c
814
queue->priv = sfq;
tools/perf/util/s390-cpumsf.c
816
if (queue->cpu != -1)
tools/perf/util/s390-cpumsf.c
817
sfq->cpu = queue->cpu;
tools/perf/util/s390-cpumsf.c
852
struct auxtrace_queue *queue;
tools/perf/util/s390-cpumsf.c
862
queue = &sf->queues.queue_array[queue_nr];
tools/perf/util/s390-cpumsf.c
863
sfq = queue->priv;
tools/testing/selftests/bpf/progs/bpf_arena_spin_lock.h
268
goto queue;
tools/testing/selftests/bpf/progs/bpf_arena_spin_lock.h
290
goto queue;
tools/testing/selftests/bpf/progs/bpf_arena_spin_lock.h
319
queue:
tools/testing/selftests/bpf/progs/bpf_qdisc_fq.c
378
bpf_rbtree_add(&flow->queue, &skbn->node, skbn_tstamp_less);
tools/testing/selftests/bpf/progs/bpf_qdisc_fq.c
498
rb_node = bpf_rbtree_first(&flow->queue);
tools/testing/selftests/bpf/progs/bpf_qdisc_fq.c
523
rb_node = bpf_rbtree_remove(&flow->queue, rb_node);
tools/testing/selftests/bpf/progs/bpf_qdisc_fq.c
565
rb_node = bpf_rbtree_first(&flow->queue);
tools/testing/selftests/bpf/progs/bpf_qdisc_fq.c
572
rb_node = bpf_rbtree_remove(&flow->queue, &skbn->node);
tools/testing/selftests/bpf/progs/bpf_qdisc_fq.c
83
struct bpf_rb_root queue __contains(skb_node, node);
tools/testing/selftests/bpf/progs/map_ptr_kern.c
587
struct bpf_queue_stack *queue = (struct bpf_queue_stack *)&m_queue;
tools/testing/selftests/bpf/progs/map_ptr_kern.c
590
VERIFY(check(&queue->map, map, 0, sizeof(__u32), MAX_ENTRIES));
tools/testing/selftests/drivers/net/hw/toeplitz.c
141
int queue;
tools/testing/selftests/drivers/net/hw/toeplitz.c
144
queue = rss_indir_tbl[rx_hash % rss_indir_tbl_size];
tools/testing/selftests/drivers/net/hw/toeplitz.c
146
queue = rx_hash % cfg_num_queues;
tools/testing/selftests/drivers/net/hw/toeplitz.c
148
log_verbose(" rxq %d (cpu %d)", queue, rx_irq_cpus[queue]);
tools/testing/selftests/drivers/net/hw/toeplitz.c
149
if (rx_irq_cpus[queue] != cpu) {
tools/testing/selftests/mqueue/mq_open_tests.c
205
if ((queue = mq_open(queue_path, flags, perms, attr)) == -1)
tools/testing/selftests/mqueue/mq_open_tests.c
207
if (mq_getattr(queue, result))
tools/testing/selftests/mqueue/mq_open_tests.c
209
if (mq_close(queue))
tools/testing/selftests/mqueue/mq_open_tests.c
211
queue = -1;
tools/testing/selftests/mqueue/mq_open_tests.c
227
if ((queue = mq_open(queue_path, flags, perms, attr)) == -1)
tools/testing/selftests/mqueue/mq_open_tests.c
229
if (mq_getattr(queue, result))
tools/testing/selftests/mqueue/mq_open_tests.c
231
if (mq_close(queue))
tools/testing/selftests/mqueue/mq_open_tests.c
233
queue = -1;
tools/testing/selftests/mqueue/mq_open_tests.c
59
mqd_t queue = -1;
tools/testing/selftests/mqueue/mq_open_tests.c
90
if (queue != -1)
tools/testing/selftests/mqueue/mq_open_tests.c
91
if (mq_close(queue))
tools/testing/selftests/mqueue/mq_perf_tests.c
192
if (queue != -1)
tools/testing/selftests/mqueue/mq_perf_tests.c
193
if (mq_close(queue))
tools/testing/selftests/mqueue/mq_perf_tests.c
297
queue = mq_open(queue_path, flags, perms, attr);
tools/testing/selftests/mqueue/mq_perf_tests.c
298
if (queue == -1)
tools/testing/selftests/mqueue/mq_perf_tests.c
300
if (mq_getattr(queue, &result))
tools/testing/selftests/mqueue/mq_perf_tests.c
335
while (mq_send(queue, buff, sizeof(buff), 0) == 0)
tools/testing/selftests/mqueue/mq_perf_tests.c
337
mq_receive(queue, buff, sizeof(buff), &priority);
tools/testing/selftests/mqueue/mq_perf_tests.c
342
while (mq_receive(queue, buff, MSG_SIZE, &prio_in) == MSG_SIZE)
tools/testing/selftests/mqueue/mq_perf_tests.c
346
if (mq_send(queue, buff, MSG_SIZE, prio_out)) \
tools/testing/selftests/mqueue/mq_perf_tests.c
353
if (mq_send(queue, buff, MSG_SIZE, prio_out)) \
tools/testing/selftests/mqueue/mq_perf_tests.c
356
if (mq_receive(queue, buff, MSG_SIZE, &prio_in) != MSG_SIZE) \
tools/testing/selftests/mqueue/mq_perf_tests.c
99
mqd_t queue = -1;
tools/testing/selftests/net/lib/xdp_helper.c
44
int queue;
tools/testing/selftests/net/lib/xdp_helper.c
70
queue = atoi(argv[2]);
tools/testing/selftests/net/lib/xdp_helper.c
94
sxdp.sxdp_queue_id = queue;
tools/testing/selftests/net/tcp_ao/lib/repair.c
127
static void test_sock_restore_seq(int sk, int queue, uint32_t seq)
tools/testing/selftests/net/tcp_ao/lib/repair.c
129
if (setsockopt(sk, SOL_TCP, TCP_REPAIR_QUEUE, &queue, sizeof(queue)))
tools/testing/selftests/net/tcp_ao/lib/repair.c
136
static void test_sock_restore_queue(int sk, int queue, void *buf, int len)
tools/testing/selftests/net/tcp_ao/lib/repair.c
144
if (setsockopt(sk, SOL_TCP, TCP_REPAIR_QUEUE, &queue, sizeof(queue)))
tools/testing/selftests/net/tcp_ao/lib/repair.c
42
static void test_sock_checkpoint_queue(int sk, int queue, int qlen,
tools/testing/selftests/net/tcp_ao/lib/repair.c
48
if (setsockopt(sk, SOL_TCP, TCP_REPAIR_QUEUE, &queue, sizeof(queue)))
tools/testing/selftests/sched_ext/dsp_local_on.bpf.c
17
} queue SEC(".maps");
tools/testing/selftests/sched_ext/dsp_local_on.bpf.c
30
if (bpf_map_push_elem(&queue, &pid, 0))
tools/testing/selftests/sched_ext/dsp_local_on.bpf.c
39
if (bpf_map_pop_elem(&queue, &pid))
tools/testing/selftests/x86/lam.c
107
} queue;
tools/testing/selftests/x86/lam.c
529
s->sq_ring.queue.sqes = mmap(0, p.sq_entries * sizeof(struct io_uring_sqe),
tools/testing/selftests/x86/lam.c
532
if (s->sq_ring.queue.sqes == MAP_FAILED) {
tools/testing/selftests/x86/lam.c
545
cring->queue.cqes = cq_ptr + p.cq_off.cqes;
tools/testing/selftests/x86/lam.c
582
cqe = &cring->queue.cqes[head & *s->cq_ring.ring_mask];
tools/testing/selftests/x86/lam.c
640
sqe = &ring->sq_ring.queue.sqes[index];
tools/tracing/latency/latency-collector.c
231
struct queue printqueue;
tools/tracing/latency/latency-collector.c
787
static void init_queue(struct queue *q)
tools/tracing/latency/latency-collector.c
797
static __always_inline int queue_len(const struct queue *q)
tools/tracing/latency/latency-collector.c
805
static __always_inline int queue_nr_free(const struct queue *q)
tools/tracing/latency/latency-collector.c
823
static __always_inline void queue_push_to_back(struct queue *q,
tools/tracing/latency/latency-collector.c
830
static __always_inline struct entry queue_pop_from_front(struct queue *q)
tools/tracing/latency/latency-collector.c
838
static __always_inline void queue_cond_signal(struct queue *q)
tools/tracing/latency/latency-collector.c
843
static __always_inline void queue_cond_wait(struct queue *q)
tools/tracing/latency/latency-collector.c
848
static __always_inline int queue_try_to_add_entry(struct queue *q,
tools/tracing/latency/latency-collector.c
863
static struct entry queue_wait_for_entry(struct queue *q)
virt/kvm/async_pf.c
123
while (!list_empty(&vcpu->async_pf.queue)) {
virt/kvm/async_pf.c
125
list_first_entry(&vcpu->async_pf.queue,
virt/kvm/async_pf.c
126
typeof(*work), queue);
virt/kvm/async_pf.c
127
list_del(&work->queue);
virt/kvm/async_pf.c
169
list_del(&work->queue);
virt/kvm/async_pf.c
207
list_add_tail(&work->queue, &vcpu->async_pf.queue);
virt/kvm/async_pf.c
229
INIT_LIST_HEAD(&work->queue); /* for list_del to work */
virt/kvm/async_pf.c
41
INIT_LIST_HEAD(&vcpu->async_pf.queue);