arch/arm64/kvm/hyp/nvhe/ffa.c
152
.a1 = hyp_virt_to_phys(hyp_buffers.tx),
arch/arm64/kvm/hyp/nvhe/ffa.c
236
DECLARE_REG(phys_addr_t, tx, ctxt, 1);
arch/arm64/kvm/hyp/nvhe/ffa.c
247
if (!PAGE_ALIGNED(tx) || !PAGE_ALIGNED(rx)) {
arch/arm64/kvm/hyp/nvhe/ffa.c
253
if (host_buffers.tx) {
arch/arm64/kvm/hyp/nvhe/ffa.c
266
ret = __pkvm_host_share_hyp(hyp_phys_to_pfn(tx));
arch/arm64/kvm/hyp/nvhe/ffa.c
278
tx_virt = hyp_phys_to_virt(tx);
arch/arm64/kvm/hyp/nvhe/ffa.c
292
host_buffers.tx = tx_virt;
arch/arm64/kvm/hyp/nvhe/ffa.c
306
__pkvm_host_unshare_hyp(hyp_phys_to_pfn(tx));
arch/arm64/kvm/hyp/nvhe/ffa.c
324
if (!host_buffers.tx) {
arch/arm64/kvm/hyp/nvhe/ffa.c
329
hyp_unpin_shared_mem(host_buffers.tx, host_buffers.tx + 1);
arch/arm64/kvm/hyp/nvhe/ffa.c
330
WARN_ON(__pkvm_host_unshare_hyp(hyp_virt_to_pfn(host_buffers.tx)));
arch/arm64/kvm/hyp/nvhe/ffa.c
331
host_buffers.tx = NULL;
arch/arm64/kvm/hyp/nvhe/ffa.c
431
if (!host_buffers.tx)
arch/arm64/kvm/hyp/nvhe/ffa.c
434
buf = hyp_buffers.tx;
arch/arm64/kvm/hyp/nvhe/ffa.c
435
memcpy(buf, host_buffers.tx, fraglen);
arch/arm64/kvm/hyp/nvhe/ffa.c
498
if (!host_buffers.tx) {
arch/arm64/kvm/hyp/nvhe/ffa.c
508
buf = hyp_buffers.tx;
arch/arm64/kvm/hyp/nvhe/ffa.c
509
memcpy(buf, host_buffers.tx, fraglen);
arch/arm64/kvm/hyp/nvhe/ffa.c
588
buf = hyp_buffers.tx;
arch/arm64/kvm/hyp/nvhe/ffa.c
59
void *tx;
arch/arm64/kvm/hyp/nvhe/ffa.c
937
void *tx, *rx;
arch/arm64/kvm/hyp/nvhe/ffa.c
970
tx = pages;
arch/arm64/kvm/hyp/nvhe/ffa.c
983
.tx = tx,
arch/mips/include/asm/ip32/mace.h
238
volatile unsigned long tx;
arch/mips/include/asm/octeon/cvmx-agl-defs.h
1064
uint64_t tx:2;
arch/mips/include/asm/octeon/cvmx-agl-defs.h
1070
uint64_t tx:2;
arch/mips/include/asm/octeon/cvmx-agl-defs.h
1077
uint64_t tx:1;
arch/mips/include/asm/octeon/cvmx-agl-defs.h
1083
uint64_t tx:1;
arch/powerpc/platforms/powernv/vas-window.c
1347
void vas_return_credit(struct pnv_vas_window *window, bool tx)
arch/powerpc/platforms/powernv/vas-window.c
1352
if (tx) { /* send window */
arch/powerpc/platforms/powernv/vas.h
430
extern void vas_return_credit(struct pnv_vas_window *window, bool tx);
arch/s390/include/asm/machine.h
92
DEFINE_MACHINE_HAS_FEATURE(tx, MFEATURE_TX)
arch/s390/include/uapi/asm/guarded_storage.h
27
__u8 tx : 1;
arch/x86/events/intel/core.c
6623
EVENT_ATTR_STR(tx-start, tx_start, "event=0xc9,umask=0x1");
arch/x86/events/intel/core.c
6624
EVENT_ATTR_STR(tx-commit, tx_commit, "event=0xc9,umask=0x2");
arch/x86/events/intel/core.c
6625
EVENT_ATTR_STR(tx-abort, tx_abort, "event=0xc9,umask=0x4");
arch/x86/events/intel/core.c
6626
EVENT_ATTR_STR(tx-capacity, tx_capacity, "event=0x54,umask=0x2");
arch/x86/events/intel/core.c
6627
EVENT_ATTR_STR(tx-conflict, tx_conflict, "event=0x54,umask=0x1");
arch/x86/events/intel/core.c
6669
EVENT_ATTR_STR(tx-capacity-read, tx_capacity_read, "event=0x54,umask=0x80");
arch/x86/events/intel/core.c
6670
EVENT_ATTR_STR(tx-capacity-write, tx_capacity_write, "event=0x54,umask=0x2");
arch/x86/events/intel/core.c
7126
EVENT_ATTR_STR_HYBRID(tx-start, tx_start_adl, "event=0xc9,umask=0x1", hybrid_big);
arch/x86/events/intel/core.c
7127
EVENT_ATTR_STR_HYBRID(tx-commit, tx_commit_adl, "event=0xc9,umask=0x2", hybrid_big);
arch/x86/events/intel/core.c
7128
EVENT_ATTR_STR_HYBRID(tx-abort, tx_abort_adl, "event=0xc9,umask=0x4", hybrid_big);
arch/x86/events/intel/core.c
7129
EVENT_ATTR_STR_HYBRID(tx-conflict, tx_conflict_adl, "event=0x54,umask=0x1", hybrid_big);
arch/x86/events/intel/core.c
7132
EVENT_ATTR_STR_HYBRID(tx-capacity-read, tx_capacity_read_adl, "event=0x54,umask=0x80", hybrid_big);
arch/x86/events/intel/core.c
7133
EVENT_ATTR_STR_HYBRID(tx-capacity-write, tx_capacity_write_adl, "event=0x54,umask=0x2", hybrid_big);
crypto/async_tx/async_memcpy.c
39
struct dma_async_tx_descriptor *tx = NULL;
crypto/async_tx/async_memcpy.c
61
tx = device->device_prep_dma_memcpy(chan, unmap->addr[1],
crypto/async_tx/async_memcpy.c
66
if (tx) {
crypto/async_tx/async_memcpy.c
69
dma_set_unmap(tx, unmap);
crypto/async_tx/async_memcpy.c
70
async_tx_submit(chan, tx, submit);
crypto/async_tx/async_memcpy.c
91
return tx;
crypto/async_tx/async_pq.c
100
return tx;
crypto/async_tx/async_pq.c
197
struct dma_async_tx_descriptor *tx;
crypto/async_tx/async_pq.c
244
tx = do_async_gen_syndrome(chan, coefs, j, unmap, dma_flags, submit);
crypto/async_tx/async_pq.c
246
return tx;
crypto/async_tx/async_pq.c
304
struct dma_async_tx_descriptor *tx;
crypto/async_tx/async_pq.c
359
tx = device->device_prep_dma_pq_val(chan, pq,
crypto/async_tx/async_pq.c
365
if (likely(tx))
crypto/async_tx/async_pq.c
371
dma_set_unmap(tx, unmap);
crypto/async_tx/async_pq.c
372
async_tx_submit(chan, tx, submit);
crypto/async_tx/async_pq.c
398
tx = NULL;
crypto/async_tx/async_pq.c
403
tx = async_xor_offs(spare, s_off,
crypto/async_tx/async_pq.c
405
async_tx_quiesce(&tx);
crypto/async_tx/async_pq.c
41
struct dma_async_tx_descriptor *tx = NULL;
crypto/async_tx/async_pq.c
416
tx = async_gen_syndrome(blocks, offsets, disks,
crypto/async_tx/async_pq.c
418
async_tx_quiesce(&tx);
crypto/async_tx/async_pq.c
434
tx = NULL;
crypto/async_tx/async_pq.c
438
return tx;
crypto/async_tx/async_pq.c
78
tx = dma->device_prep_dma_pq(chan, dma_dest,
crypto/async_tx/async_pq.c
83
if (likely(tx))
crypto/async_tx/async_pq.c
89
dma_set_unmap(tx, unmap);
crypto/async_tx/async_pq.c
90
async_tx_submit(chan, tx, submit);
crypto/async_tx/async_pq.c
91
submit->depend_tx = tx;
crypto/async_tx/async_raid6_recov.c
104
struct dma_async_tx_descriptor *tx;
crypto/async_tx/async_raid6_recov.c
122
tx = dma->device_prep_dma_pq(chan, dma_dest, unmap->addr,
crypto/async_tx/async_raid6_recov.c
125
if (tx) {
crypto/async_tx/async_raid6_recov.c
126
dma_set_unmap(tx, unmap);
crypto/async_tx/async_raid6_recov.c
128
async_tx_submit(chan, tx, submit);
crypto/async_tx/async_raid6_recov.c
129
return tx;
crypto/async_tx/async_raid6_recov.c
157
struct dma_async_tx_descriptor *tx = NULL;
crypto/async_tx/async_raid6_recov.c
186
init_async_submit(submit, ASYNC_TX_FENCE, tx, NULL, NULL, scribble);
crypto/async_tx/async_raid6_recov.c
187
tx = async_sum_product(b, b_off, srcs, src_offs, coef, bytes, submit);
crypto/async_tx/async_raid6_recov.c
194
init_async_submit(submit, flags | ASYNC_TX_XOR_ZERO_DST, tx, cb_fn,
crypto/async_tx/async_raid6_recov.c
196
tx = async_xor_offs(a, a_off, srcs, src_offs, 2, bytes, submit);
crypto/async_tx/async_raid6_recov.c
198
return tx;
crypto/async_tx/async_raid6_recov.c
207
struct dma_async_tx_descriptor *tx = NULL;
crypto/async_tx/async_raid6_recov.c
247
init_async_submit(submit, ASYNC_TX_FENCE, tx, NULL, NULL, scribble);
crypto/async_tx/async_raid6_recov.c
248
tx = async_memcpy(dp, g, dp_off, g_off, bytes, submit);
crypto/async_tx/async_raid6_recov.c
249
init_async_submit(submit, ASYNC_TX_FENCE, tx, NULL, NULL, scribble);
crypto/async_tx/async_raid6_recov.c
250
tx = async_mult(dq, dq_off, g, g_off,
crypto/async_tx/async_raid6_recov.c
258
init_async_submit(submit, ASYNC_TX_FENCE|ASYNC_TX_XOR_DROP_DST, tx,
crypto/async_tx/async_raid6_recov.c
260
tx = async_xor_offs(dp, dp_off, srcs, src_offs, 2, bytes, submit);
crypto/async_tx/async_raid6_recov.c
267
init_async_submit(submit, ASYNC_TX_FENCE|ASYNC_TX_XOR_DROP_DST, tx,
crypto/async_tx/async_raid6_recov.c
269
tx = async_xor_offs(dq, dq_off, srcs, src_offs, 2, bytes, submit);
crypto/async_tx/async_raid6_recov.c
278
init_async_submit(submit, ASYNC_TX_FENCE, tx, NULL, NULL, scribble);
crypto/async_tx/async_raid6_recov.c
279
tx = async_sum_product(dq, dq_off, srcs, src_offs, coef, bytes, submit);
crypto/async_tx/async_raid6_recov.c
286
init_async_submit(submit, flags | ASYNC_TX_XOR_DROP_DST, tx, cb_fn,
crypto/async_tx/async_raid6_recov.c
288
tx = async_xor_offs(dp, dp_off, srcs, src_offs, 2, bytes, submit);
crypto/async_tx/async_raid6_recov.c
290
return tx;
crypto/async_tx/async_raid6_recov.c
298
struct dma_async_tx_descriptor *tx = NULL;
crypto/async_tx/async_raid6_recov.c
329
init_async_submit(submit, ASYNC_TX_FENCE, tx, NULL, NULL, scribble);
crypto/async_tx/async_raid6_recov.c
330
tx = async_gen_syndrome(blocks, offs, disks, bytes, submit);
crypto/async_tx/async_raid6_recov.c
347
init_async_submit(submit, ASYNC_TX_FENCE|ASYNC_TX_XOR_DROP_DST, tx,
crypto/async_tx/async_raid6_recov.c
349
tx = async_xor_offs(dp, dp_off, srcs, src_offs, 2, bytes, submit);
crypto/async_tx/async_raid6_recov.c
356
init_async_submit(submit, ASYNC_TX_FENCE|ASYNC_TX_XOR_DROP_DST, tx,
crypto/async_tx/async_raid6_recov.c
358
tx = async_xor_offs(dq, dq_off, srcs, src_offs, 2, bytes, submit);
crypto/async_tx/async_raid6_recov.c
36
struct dma_async_tx_descriptor *tx;
crypto/async_tx/async_raid6_recov.c
367
init_async_submit(submit, ASYNC_TX_FENCE, tx, NULL, NULL, scribble);
crypto/async_tx/async_raid6_recov.c
368
tx = async_sum_product(dq, dq_off, srcs, src_offs, coef, bytes, submit);
crypto/async_tx/async_raid6_recov.c
375
init_async_submit(submit, flags | ASYNC_TX_XOR_DROP_DST, tx, cb_fn,
crypto/async_tx/async_raid6_recov.c
377
tx = async_xor_offs(dp, dp_off, srcs, src_offs, 2, bytes, submit);
crypto/async_tx/async_raid6_recov.c
379
return tx;
crypto/async_tx/async_raid6_recov.c
475
struct dma_async_tx_descriptor *tx = NULL;
crypto/async_tx/async_raid6_recov.c
54
tx = dma->device_prep_dma_pq(chan, pq, unmap->addr, 2, coef,
crypto/async_tx/async_raid6_recov.c
546
init_async_submit(submit, ASYNC_TX_FENCE, tx, NULL, NULL,
crypto/async_tx/async_raid6_recov.c
548
tx = async_memcpy(p, g, p_off, g_off, bytes, submit);
crypto/async_tx/async_raid6_recov.c
550
init_async_submit(submit, ASYNC_TX_FENCE, tx, NULL, NULL,
crypto/async_tx/async_raid6_recov.c
552
tx = async_mult(dq, dq_off, g, g_off,
crypto/async_tx/async_raid6_recov.c
555
init_async_submit(submit, ASYNC_TX_FENCE, tx, NULL, NULL,
crypto/async_tx/async_raid6_recov.c
557
tx = async_gen_syndrome(blocks, offs, disks, bytes, submit);
crypto/async_tx/async_raid6_recov.c
56
if (tx) {
crypto/async_tx/async_raid6_recov.c
57
dma_set_unmap(tx, unmap);
crypto/async_tx/async_raid6_recov.c
573
init_async_submit(submit, ASYNC_TX_FENCE|ASYNC_TX_XOR_DROP_DST, tx,
crypto/async_tx/async_raid6_recov.c
575
tx = async_xor_offs(dq, dq_off, srcs, src_offs, 2, bytes, submit);
crypto/async_tx/async_raid6_recov.c
577
init_async_submit(submit, ASYNC_TX_FENCE, tx, NULL, NULL, scribble);
crypto/async_tx/async_raid6_recov.c
578
tx = async_mult(dq, dq_off, dq, dq_off, coef, bytes, submit);
crypto/async_tx/async_raid6_recov.c
58
async_tx_submit(chan, tx, submit);
crypto/async_tx/async_raid6_recov.c
584
init_async_submit(submit, flags | ASYNC_TX_XOR_DROP_DST, tx, cb_fn,
crypto/async_tx/async_raid6_recov.c
586
tx = async_xor_offs(p, p_off, srcs, src_offs, 2, bytes, submit);
crypto/async_tx/async_raid6_recov.c
588
return tx;
crypto/async_tx/async_raid6_recov.c
60
return tx;
crypto/async_tx/async_tx.c
100
txd_chain(intr_tx, tx);
crypto/async_tx/async_tx.c
121
tx->tx_submit(tx);
crypto/async_tx/async_tx.c
143
async_tx_submit(struct dma_chan *chan, struct dma_async_tx_descriptor *tx,
crypto/async_tx/async_tx.c
148
tx->callback = submit->cb_fn;
crypto/async_tx/async_tx.c
149
tx->callback_param = submit->cb_param;
crypto/async_tx/async_tx.c
161
txd_parent(tx));
crypto/async_tx/async_tx.c
173
txd_chain(depend_tx, tx);
crypto/async_tx/async_tx.c
192
async_tx_channel_switch(depend_tx, tx);
crypto/async_tx/async_tx.c
195
txd_clear_parent(tx);
crypto/async_tx/async_tx.c
196
tx->tx_submit(tx);
crypto/async_tx/async_tx.c
200
txd_clear_parent(tx);
crypto/async_tx/async_tx.c
201
tx->tx_submit(tx);
crypto/async_tx/async_tx.c
205
async_tx_ack(tx);
crypto/async_tx/async_tx.c
225
struct dma_async_tx_descriptor *tx;
crypto/async_tx/async_tx.c
238
tx = device ? device->device_prep_dma_interrupt(chan, 0) : NULL;
crypto/async_tx/async_tx.c
240
tx = NULL;
crypto/async_tx/async_tx.c
242
if (tx) {
crypto/async_tx/async_tx.c
245
async_tx_submit(chan, tx, submit);
crypto/async_tx/async_tx.c
255
return tx;
crypto/async_tx/async_tx.c
263
void async_tx_quiesce(struct dma_async_tx_descriptor **tx)
crypto/async_tx/async_tx.c
265
if (*tx) {
crypto/async_tx/async_tx.c
269
BUG_ON(async_tx_test_ack(*tx));
crypto/async_tx/async_tx.c
270
if (dma_wait_for_async_tx(*tx) != DMA_COMPLETE)
crypto/async_tx/async_tx.c
273
async_tx_ack(*tx);
crypto/async_tx/async_tx.c
274
*tx = NULL;
crypto/async_tx/async_tx.c
66
struct dma_async_tx_descriptor *tx)
crypto/async_tx/async_tx.c
74
if (txd_parent(depend_tx) && depend_tx->chan == tx->chan) {
crypto/async_tx/async_tx.c
75
txd_chain(depend_tx, tx);
crypto/async_tx/async_xor.c
199
struct dma_async_tx_descriptor *tx;
crypto/async_tx/async_xor.c
220
tx = do_async_xor(chan, unmap, submit);
crypto/async_tx/async_xor.c
222
return tx;
crypto/async_tx/async_xor.c
27
struct dma_async_tx_descriptor *tx = NULL;
crypto/async_tx/async_xor.c
322
struct dma_async_tx_descriptor *tx = NULL;
crypto/async_tx/async_xor.c
350
tx = device->device_prep_dma_xor_val(chan, unmap->addr, src_cnt,
crypto/async_tx/async_xor.c
353
if (unlikely(!tx)) {
crypto/async_tx/async_xor.c
356
while (!tx) {
crypto/async_tx/async_xor.c
358
tx = device->device_prep_dma_xor_val(chan,
crypto/async_tx/async_xor.c
363
dma_set_unmap(tx, unmap);
crypto/async_tx/async_xor.c
364
async_tx_submit(chan, tx, submit);
crypto/async_tx/async_xor.c
376
tx = async_xor_offs(dest, offset, src_list, src_offs,
crypto/async_tx/async_xor.c
379
async_tx_quiesce(&tx);
crypto/async_tx/async_xor.c
388
return tx;
crypto/async_tx/async_xor.c
65
tx = dma->device_prep_dma_xor(chan, dma_dest, src_list,
crypto/async_tx/async_xor.c
69
if (unlikely(!tx))
crypto/async_tx/async_xor.c
73
while (unlikely(!tx)) {
crypto/async_tx/async_xor.c
75
tx = dma->device_prep_dma_xor(chan, dma_dest,
crypto/async_tx/async_xor.c
82
dma_set_unmap(tx, unmap);
crypto/async_tx/async_xor.c
83
async_tx_submit(chan, tx, submit);
crypto/async_tx/async_xor.c
84
submit->depend_tx = tx;
crypto/async_tx/async_xor.c
96
return tx;
crypto/async_tx/raid6test.c
103
tx = async_raid6_datap_recov(disks, bytes,
crypto/async_tx/raid6test.c
108
tx = async_raid6_2data_recov(disks, bytes,
crypto/async_tx/raid6test.c
113
init_async_submit(&submit, ASYNC_TX_ACK, tx, callback, &cmp, addr_conv);
crypto/async_tx/raid6test.c
114
tx = async_syndrome_val(ptrs, offs,
crypto/async_tx/raid6test.c
116
async_tx_issue_pending(tx);
crypto/async_tx/raid6test.c
154
struct dma_async_tx_descriptor *tx;
crypto/async_tx/raid6test.c
173
tx = async_gen_syndrome(dataptrs, dataoffs, disks, PAGE_SIZE, &submit);
crypto/async_tx/raid6test.c
174
async_tx_issue_pending(tx);
crypto/async_tx/raid6test.c
62
struct dma_async_tx_descriptor *tx = NULL;
crypto/async_tx/raid6test.c
72
tx = async_gen_syndrome(ptrs, offs,
crypto/async_tx/raid6test.c
93
tx = async_xor(dest, blocks, 0, count, bytes, &submit);
crypto/async_tx/raid6test.c
95
init_async_submit(&submit, 0, tx, NULL, NULL, addr_conv);
crypto/async_tx/raid6test.c
96
tx = async_gen_syndrome(ptrs, offs,
crypto/ecc.c
1445
u64 tx[ECC_MAX_DIGITS];
crypto/ecc.c
1449
vli_set(tx, point->x, ndigits);
crypto/ecc.c
1451
apply_z(tx, ty, z, curve);
crypto/ecc.c
1452
vli_mod_sub(tz, rx, tx, curve->p, ndigits);
crypto/ecc.c
1453
xycz_add(tx, ty, rx, ry, curve);
drivers/accel/qaic/sahara.c
158
struct sahara_packet *tx[SAHARA_NUM_TX_BUF];
drivers/accel/qaic/sahara.c
283
context->tx[0]->cmd = cpu_to_le32(SAHARA_RESET_CMD);
drivers/accel/qaic/sahara.c
284
context->tx[0]->length = cpu_to_le32(SAHARA_RESET_LENGTH);
drivers/accel/qaic/sahara.c
286
ret = mhi_queue_buf(context->mhi_dev, DMA_TO_DEVICE, context->tx[0],
drivers/accel/qaic/sahara.c
323
context->tx[0]->cmd = cpu_to_le32(SAHARA_HELLO_RESP_CMD);
drivers/accel/qaic/sahara.c
324
context->tx[0]->length = cpu_to_le32(SAHARA_HELLO_LENGTH);
drivers/accel/qaic/sahara.c
325
context->tx[0]->hello_resp.version = cpu_to_le32(SAHARA_VERSION);
drivers/accel/qaic/sahara.c
326
context->tx[0]->hello_resp.version_compat = cpu_to_le32(SAHARA_VERSION);
drivers/accel/qaic/sahara.c
327
context->tx[0]->hello_resp.status = cpu_to_le32(SAHARA_SUCCESS);
drivers/accel/qaic/sahara.c
328
context->tx[0]->hello_resp.mode = context->rx->hello_resp.mode;
drivers/accel/qaic/sahara.c
330
ret = mhi_queue_buf(context->mhi_dev, DMA_TO_DEVICE, context->tx[0],
drivers/accel/qaic/sahara.c
344
memcpy(context->tx[buf_index],
drivers/accel/qaic/sahara.c
357
context->tx[buf_index], pkt_data_len, mhi_flag);
drivers/accel/qaic/sahara.c
468
context->tx[0]->cmd = cpu_to_le32(SAHARA_DONE_CMD);
drivers/accel/qaic/sahara.c
469
context->tx[0]->length = cpu_to_le32(SAHARA_DONE_LENGTH);
drivers/accel/qaic/sahara.c
471
ret = mhi_queue_buf(context->mhi_dev, DMA_TO_DEVICE, context->tx[0],
drivers/accel/qaic/sahara.c
525
context->tx[0]->cmd = cpu_to_le32(SAHARA_MEM_READ64_CMD);
drivers/accel/qaic/sahara.c
526
context->tx[0]->length = cpu_to_le32(SAHARA_MEM_READ64_LENGTH);
drivers/accel/qaic/sahara.c
527
context->tx[0]->memory_read64.memory_address = cpu_to_le64(context->dump_table_address);
drivers/accel/qaic/sahara.c
528
context->tx[0]->memory_read64.memory_length = cpu_to_le64(context->dump_table_length);
drivers/accel/qaic/sahara.c
532
ret = mhi_queue_buf(context->mhi_dev, DMA_TO_DEVICE, context->tx[0],
drivers/accel/qaic/sahara.c
670
context->tx[0]->cmd = cpu_to_le32(SAHARA_MEM_READ64_CMD);
drivers/accel/qaic/sahara.c
671
context->tx[0]->length = cpu_to_le32(SAHARA_MEM_READ64_LENGTH);
drivers/accel/qaic/sahara.c
672
context->tx[0]->memory_read64.memory_address = cpu_to_le64(context->dump_image->address);
drivers/accel/qaic/sahara.c
673
context->tx[0]->memory_read64.memory_length = cpu_to_le64(dump_length);
drivers/accel/qaic/sahara.c
677
ret = mhi_queue_buf(context->mhi_dev, DMA_TO_DEVICE, context->tx[0],
drivers/accel/qaic/sahara.c
716
context->tx[0]->cmd = cpu_to_le32(SAHARA_MEM_READ64_CMD);
drivers/accel/qaic/sahara.c
717
context->tx[0]->length = cpu_to_le32(SAHARA_MEM_READ64_LENGTH);
drivers/accel/qaic/sahara.c
718
context->tx[0]->memory_read64.memory_address =
drivers/accel/qaic/sahara.c
720
context->tx[0]->memory_read64.memory_length = cpu_to_le64(dump_length);
drivers/accel/qaic/sahara.c
725
ret = mhi_queue_buf(context->mhi_dev, DMA_TO_DEVICE, context->tx[0],
drivers/accel/qaic/sahara.c
848
context->tx[i] = devm_kzalloc(&mhi_dev->dev,
drivers/accel/qaic/sahara.c
851
if (!context->tx[i])
drivers/acpi/processor_throttling.c
231
if (p_limit->thermal.tx > target_state)
drivers/acpi/processor_throttling.c
232
target_state = p_limit->thermal.tx;
drivers/acpi/processor_throttling.c
233
if (p_limit->user.tx > target_state)
drivers/acpi/processor_throttling.c
234
target_state = p_limit->user.tx;
drivers/acpi/processor_throttling.c
325
if (limit->thermal.tx > target_state)
drivers/acpi/processor_throttling.c
326
target_state = limit->thermal.tx;
drivers/acpi/processor_throttling.c
327
if (limit->user.tx > target_state)
drivers/acpi/processor_throttling.c
328
target_state = limit->user.tx;
drivers/acpi/processor_throttling.c
523
struct acpi_processor_tx_tss *tx =
drivers/acpi/processor_throttling.c
528
state.pointer = tx;
drivers/acpi/processor_throttling.c
542
if (!tx->freqpercentage) {
drivers/acpi/processor_throttling.c
821
struct acpi_processor_tx_tss *tx =
drivers/acpi/processor_throttling.c
824
if (tx->control == value)
drivers/acpi/processor_throttling.c
836
struct acpi_processor_tx_tss *tx =
drivers/acpi/processor_throttling.c
839
*value = tx->control;
drivers/ata/pata_arasan_cf.c
397
struct dma_async_tx_descriptor *tx;
drivers/ata/pata_arasan_cf.c
403
tx = chan->device->device_prep_dma_memcpy(chan, dest, src, len, flags);
drivers/ata/pata_arasan_cf.c
404
if (!tx) {
drivers/ata/pata_arasan_cf.c
409
tx->callback = dma_callback;
drivers/ata/pata_arasan_cf.c
410
tx->callback_param = acdev;
drivers/ata/pata_arasan_cf.c
411
cookie = tx->tx_submit(tx);
drivers/ata/pata_pxa.c
50
struct dma_async_tx_descriptor *tx;
drivers/ata/pata_pxa.c
57
tx = dmaengine_prep_slave_sg(pd->dma_chan, qc->sg, qc->n_elem, dir,
drivers/ata/pata_pxa.c
59
if (!tx) {
drivers/ata/pata_pxa.c
63
tx->callback = pxa_ata_dma_irq;
drivers/ata/pata_pxa.c
64
tx->callback_param = pd;
drivers/ata/pata_pxa.c
65
pd->dma_cookie = dmaengine_submit(tx);
drivers/atm/adummy.c
118
atomic_inc(&vcc->stats->tx);
drivers/atm/atmtcp.c
229
atomic_inc(&vcc->stats->tx);
drivers/atm/atmtcp.c
328
atomic_inc(&vcc->stats->tx);
drivers/atm/eni.c
1032
struct eni_tx *tx;
drivers/atm/eni.c
1047
tx = eni_vcc->tx;
drivers/atm/eni.c
1048
NULLCHECK(tx);
drivers/atm/eni.c
1087
if (!NEPMOK(tx->tx_pos,size+TX_GAP,
drivers/atm/eni.c
1088
eni_in(MID_TX_RDPTR(tx->index)),tx->words)) {
drivers/atm/eni.c
1105
DPRINTK("dma_wr is %d, tx_pos is %ld\n",dma_wr,tx->tx_pos);
drivers/atm/eni.c
1120
eni_dev->dma[j++] = (((tx->tx_pos+TX_DESCR_SIZE) & (tx->words-1)) <<
drivers/atm/eni.c
1121
MID_DMA_COUNT_SHIFT) | (tx->index << MID_DMA_CHAN_SHIFT) |
drivers/atm/eni.c
1125
if (aal5) put_dma(tx->index,eni_dev->dma,&j,paddr,skb->len);
drivers/atm/eni.c
1126
else put_dma(tx->index,eni_dev->dma,&j,paddr+4,skb->len-4);
drivers/atm/eni.c
1131
put_dma(tx->index,eni_dev->dma,&j,(unsigned long)
drivers/atm/eni.c
1135
put_dma(tx->index,eni_dev->dma,&j,(unsigned long)
drivers/atm/eni.c
1141
put_dma(tx->index, eni_dev->dma, &j, eni_dev->zero.dma,
drivers/atm/eni.c
1145
eni_dev->dma[j++] = (((tx->tx_pos+size) & (tx->words-1)) <<
drivers/atm/eni.c
1146
MID_DMA_COUNT_SHIFT) | (tx->index << MID_DMA_CHAN_SHIFT) |
drivers/atm/eni.c
1152
(aal5 ? MID_SEG_AAL5 : 0) | (tx->prescaler << MID_SEG_PR_SHIFT) |
drivers/atm/eni.c
1153
(tx->resolution << MID_SEG_RATE_SHIFT) |
drivers/atm/eni.c
1154
(size/(ATM_CELL_PAYLOAD/4)),tx->send+tx->tx_pos*4);
drivers/atm/eni.c
1159
tx->send+((tx->tx_pos+1) & (tx->words-1))*4);
drivers/atm/eni.c
1162
writel(skb->len,tx->send+
drivers/atm/eni.c
1163
((tx->tx_pos+size-AAL5_TRAILER) & (tx->words-1))*4);
drivers/atm/eni.c
1170
ENI_PRV_POS(skb) = tx->tx_pos;
drivers/atm/eni.c
1173
tx->tx_pos = (tx->tx_pos+size) & (tx->words-1);
drivers/atm/eni.c
1174
DPRINTK("dma_wr set to %d, tx_pos is now %ld\n",dma_wr,tx->tx_pos);
drivers/atm/eni.c
1184
struct eni_tx *tx;
drivers/atm/eni.c
1191
tx = &ENI_DEV(dev)->tx[i];
drivers/atm/eni.c
1192
if (tx->send)
drivers/atm/eni.c
1193
while ((skb = skb_dequeue(&tx->backlog))) {
drivers/atm/eni.c
1197
skb_queue_head(&tx->backlog,skb);
drivers/atm/eni.c
1211
struct eni_tx *tx;
drivers/atm/eni.c
1219
tx = ENI_VCC(vcc)->tx;
drivers/atm/eni.c
1220
NULLCHECK(ENI_VCC(vcc)->tx);
drivers/atm/eni.c
1222
(unsigned) eni_in(MID_TX_DESCRSTART(tx->index)));
drivers/atm/eni.c
1223
if (ENI_VCC(vcc)->txing < tx->words && ENI_PRV_POS(skb) ==
drivers/atm/eni.c
1224
eni_in(MID_TX_DESCRSTART(tx->index))) {
drivers/atm/eni.c
1233
atomic_inc(&vcc->stats->tx);
drivers/atm/eni.c
1245
if (!eni_dev->tx[i].send) return eni_dev->tx+i;
drivers/atm/eni.c
1292
struct eni_tx *tx;
drivers/atm/eni.c
1311
eni_vcc->tx = eni_dev->ubr;
drivers/atm/eni.c
1317
new_tx = !eni_vcc->tx;
drivers/atm/eni.c
1319
if (!new_tx) tx = eni_vcc->tx;
drivers/atm/eni.c
1323
tx = alloc_tx(eni_dev,unlimited);
drivers/atm/eni.c
1324
if (!tx) {
drivers/atm/eni.c
1328
DPRINTK("got chan %d\n",tx->index);
drivers/atm/eni.c
1329
tx->reserved = tx->shaping = 0;
drivers/atm/eni.c
1330
tx->send = mem;
drivers/atm/eni.c
1331
tx->words = size >> 2;
drivers/atm/eni.c
1332
skb_queue_head_init(&tx->backlog);
drivers/atm/eni.c
1335
((tx->send-eni_dev->ram) >> (MID_LOC_SKIP+2)),
drivers/atm/eni.c
1336
MID_TX_PLACE(tx->index));
drivers/atm/eni.c
1337
tx->tx_pos = eni_in(MID_TX_DESCRSTART(tx->index)) &
drivers/atm/eni.c
1340
error = comp_tx(eni_dev,&rate,tx->reserved,&pre,&res,unlimited);
drivers/atm/eni.c
1344
if (!error && !ubr && rate > eni_dev->tx_bw+tx->reserved)
drivers/atm/eni.c
1346
if (!error && set_rsv && !set_shp && rate < tx->shaping)
drivers/atm/eni.c
1348
if (!error && !set_rsv && rate > tx->reserved && !ubr)
drivers/atm/eni.c
1352
tx->send = NULL;
drivers/atm/eni.c
1359
eni_dev->tx_bw += tx->reserved;
drivers/atm/eni.c
1360
tx->reserved = rate;
drivers/atm/eni.c
1364
if (unlimited && new_tx) eni_dev->ubr = tx;
drivers/atm/eni.c
1365
tx->prescaler = pre;
drivers/atm/eni.c
1366
tx->resolution = res;
drivers/atm/eni.c
1367
tx->shaping = rate;
drivers/atm/eni.c
1369
if (set_shp) eni_vcc->tx = tx;
drivers/atm/eni.c
1370
DPRINTK("rsv %d shp %d\n",tx->reserved,tx->shaping);
drivers/atm/eni.c
1377
ENI_VCC(vcc)->tx = NULL;
drivers/atm/eni.c
1397
if (!eni_vcc->tx) return;
drivers/atm/eni.c
1407
txing = skb_peek(&eni_vcc->tx->backlog) || eni_vcc->txing;
drivers/atm/eni.c
1416
if (eni_vcc->tx != eni_dev->ubr) {
drivers/atm/eni.c
1422
while (eni_in(MID_TX_RDPTR(eni_vcc->tx->index)) !=
drivers/atm/eni.c
1423
eni_in(MID_TX_DESCRSTART(eni_vcc->tx->index)))
drivers/atm/eni.c
1425
eni_free_mem(eni_dev,eni_vcc->tx->send,eni_vcc->tx->words << 2);
drivers/atm/eni.c
1426
eni_vcc->tx->send = NULL;
drivers/atm/eni.c
1427
eni_dev->tx_bw += eni_vcc->tx->reserved;
drivers/atm/eni.c
1429
eni_vcc->tx = NULL;
drivers/atm/eni.c
1447
eni_dev->tx[i].send = NULL;
drivers/atm/eni.c
1448
eni_dev->tx[i].index = i;
drivers/atm/eni.c
183
if (eni_dev->tx[i].send)
drivers/atm/eni.c
185
eni_dev->tx[i].send,eni_dev->tx[i].words*4);
drivers/atm/eni.c
1930
eni_vcc->tx = NULL; /* for eni_close after open_rx */
drivers/atm/eni.c
1958
struct eni_tx *tx = ENI_VCC(vcc)->tx;
drivers/atm/eni.c
1963
if (tx == eni_dev->ubr) return -EBADFD;
drivers/atm/eni.c
1967
if ((flgs & ATM_MF_DEC_RSV) && rate && rate < tx->reserved) rsv = 1;
drivers/atm/eni.c
1968
if ((flgs & ATM_MF_INC_RSV) && (!rate || rate > tx->reserved)) rsv = 1;
drivers/atm/eni.c
1969
if ((flgs & ATM_MF_DEC_SHP) && rate && rate < tx->shaping) shp = 1;
drivers/atm/eni.c
1970
if ((flgs & ATM_MF_INC_SHP) && (!rate || rate > tx->shaping)) shp = 1;
drivers/atm/eni.c
1984
dsc = tx->send+ENI_PRV_POS(skb)*4;
drivers/atm/eni.c
1986
(tx->prescaler << MID_SEG_PR_SHIFT) |
drivers/atm/eni.c
1987
(tx->resolution << MID_SEG_RATE_SHIFT), dsc);
drivers/atm/eni.c
2012
if ((mult.tx && mult.tx <= 100) || (mult.rx &&mult.rx <= 100) ||
drivers/atm/eni.c
2013
mult.tx > 65536 || mult.rx > 65536)
drivers/atm/eni.c
2015
if (mult.tx) eni_dev->tx_mult = mult.tx;
drivers/atm/eni.c
2038
if (!ENI_VCC(vcc)->tx) {
drivers/atm/eni.c
2062
skb_queue_tail(&ENI_VCC(vcc)->tx->backlog,skb);
drivers/atm/eni.c
2142
struct eni_tx *tx = eni_dev->tx+i;
drivers/atm/eni.c
2144
if (!tx->send) continue;
drivers/atm/eni.c
2148
(unsigned long) (tx->send - eni_dev->ram),
drivers/atm/eni.c
2149
tx->send-eni_dev->ram+tx->words*4-1,tx->words*4,
drivers/atm/eni.c
2150
tx->reserved,tx->shaping,
drivers/atm/eni.c
2151
tx == eni_dev->ubr ? " (UBR)" : "");
drivers/atm/eni.c
2155
skb_queue_len(&tx->backlog));
drivers/atm/eni.c
2177
if (eni_vcc->tx) length += sprintf(page+length,", ");
drivers/atm/eni.c
2179
if (eni_vcc->tx)
drivers/atm/eni.c
2181
eni_vcc->tx->index,eni_vcc->txing);
drivers/atm/eni.h
59
struct eni_tx *tx; /* TXer, NULL if none */
drivers/atm/eni.h
85
struct eni_tx tx[NR_CHAN]; /* TX channels */
drivers/atm/fore200e.c
847
atomic_inc(&vcc->stats->tx);
drivers/atm/he.c
2600
atomic_inc(&vcc->stats->tx);
drivers/atm/idt77252.c
814
atomic_inc(&vcc->stats->tx);
drivers/atm/iphase.c
2931
atomic_inc(&vcc->stats->tx);
drivers/atm/iphase.c
3036
atomic_inc(&vcc->stats->tx);
drivers/atm/iphase.c
3043
if (atomic_read(&vcc->stats->tx) % 20 == 0) {
drivers/atm/lanai.c
1137
r -= ((unsigned long) lvcc->tx.buf.ptr) -
drivers/atm/lanai.c
1138
((unsigned long) lvcc->tx.buf.start);
drivers/atm/lanai.c
1141
r += lanai_buf_size(&lvcc->tx.buf);
drivers/atm/lanai.c
1148
return !skb_queue_empty(&lvcc->tx.backlog);
drivers/atm/lanai.c
1162
APRINTK((((unsigned long) lvcc->tx.buf.ptr) & 15) == 0,
drivers/atm/lanai.c
1163
"vcc_tx_add_aal5_descriptor: bad ptr=%p\n", lvcc->tx.buf.ptr);
drivers/atm/lanai.c
1164
lvcc->tx.buf.ptr += 4; /* Hope the values REALLY don't matter */
drivers/atm/lanai.c
1165
pos = ((unsigned char *) lvcc->tx.buf.ptr) -
drivers/atm/lanai.c
1166
(unsigned char *) lvcc->tx.buf.start;
drivers/atm/lanai.c
1170
lvcc->tx.buf.start, lvcc->tx.buf.ptr, lvcc->tx.buf.end);
drivers/atm/lanai.c
1171
pos = (pos + len) & (lanai_buf_size(&lvcc->tx.buf) - 1);
drivers/atm/lanai.c
1175
lvcc->tx.buf.start, lvcc->tx.buf.ptr, lvcc->tx.buf.end);
drivers/atm/lanai.c
1176
lvcc->tx.buf.ptr[-1] =
drivers/atm/lanai.c
1178
((lvcc->tx.atmvcc->atm_options & ATM_ATMOPT_CLP) ?
drivers/atm/lanai.c
1180
if (lvcc->tx.buf.ptr >= lvcc->tx.buf.end)
drivers/atm/lanai.c
1181
lvcc->tx.buf.ptr = lvcc->tx.buf.start;
drivers/atm/lanai.c
1188
APRINTK((((unsigned long) lvcc->tx.buf.ptr) & 15) == 8,
drivers/atm/lanai.c
1189
"vcc_tx_add_aal5_trailer: bad ptr=%p\n", lvcc->tx.buf.ptr);
drivers/atm/lanai.c
1190
lvcc->tx.buf.ptr += 2;
drivers/atm/lanai.c
1191
lvcc->tx.buf.ptr[-2] = cpu_to_be32((uu << 24) | (cpi << 16) | len);
drivers/atm/lanai.c
1192
if (lvcc->tx.buf.ptr >= lvcc->tx.buf.end)
drivers/atm/lanai.c
1193
lvcc->tx.buf.ptr = lvcc->tx.buf.start;
drivers/atm/lanai.c
1201
e = ((unsigned char *) lvcc->tx.buf.ptr) + n;
drivers/atm/lanai.c
1202
m = e - (unsigned char *) lvcc->tx.buf.end;
drivers/atm/lanai.c
1205
memcpy(lvcc->tx.buf.ptr, src, n - m);
drivers/atm/lanai.c
1207
memcpy(lvcc->tx.buf.start, src + n - m, m);
drivers/atm/lanai.c
1208
e = ((unsigned char *) lvcc->tx.buf.start) + m;
drivers/atm/lanai.c
1210
lvcc->tx.buf.ptr = (u32 *) e;
drivers/atm/lanai.c
1219
e = ((unsigned char *) lvcc->tx.buf.ptr) + n;
drivers/atm/lanai.c
1220
m = e - (unsigned char *) lvcc->tx.buf.end;
drivers/atm/lanai.c
1223
memset(lvcc->tx.buf.ptr, 0, n - m);
drivers/atm/lanai.c
1225
memset(lvcc->tx.buf.start, 0, m);
drivers/atm/lanai.c
1226
e = ((unsigned char *) lvcc->tx.buf.start) + m;
drivers/atm/lanai.c
1228
lvcc->tx.buf.ptr = (u32 *) e;
drivers/atm/lanai.c
1235
int i, ptr = ((unsigned char *) lvcc->tx.buf.ptr) -
drivers/atm/lanai.c
1236
(unsigned char *) lvcc->tx.buf.start;
drivers/atm/lanai.c
1239
ptr, lvcc->vci, lvcc->tx.buf.start, lvcc->tx.buf.ptr,
drivers/atm/lanai.c
1240
lvcc->tx.buf.end);
drivers/atm/lanai.c
1292
lanai_free_skb(lvcc->tx.atmvcc, skb);
drivers/atm/lanai.c
1293
atomic_inc(&lvcc->tx.atmvcc->stats->tx);
drivers/atm/lanai.c
1307
skb = skb_dequeue(&lvcc->tx.backlog);
drivers/atm/lanai.c
1313
skb_queue_head(&lvcc->tx.backlog, skb);
drivers/atm/lanai.c
1339
skb_queue_tail(&lvcc->tx.backlog, skb);
drivers/atm/lanai.c
1357
lanai_free_skb(lvcc->tx.atmvcc, skb);
drivers/atm/lanai.c
1469
skb_queue_head_init(&lvcc->tx.backlog);
drivers/atm/lanai.c
1511
lvcc->tx.unqueue = vcc_tx_unqueue_aal0;
drivers/atm/lanai.c
1515
lvcc->tx.unqueue = vcc_tx_unqueue_aal5;
drivers/atm/lanai.c
1519
return lanai_get_sized_buffer(lanai, &lvcc->tx.buf, max_sdu,
drivers/atm/lanai.c
1633
if (unlikely(lvcc->tx.atmvcc == NULL)) {
drivers/atm/lanai.c
1641
lvcc->tx.endptr = SERVICE_GET_END(s);
drivers/atm/lanai.c
1702
lvcc->tx.unqueue(lanai, lvcc, lvcc->tx.endptr);
drivers/atm/lanai.c
1749
if (lvcc == NULL || lvcc->tx.atmvcc == NULL ||
drivers/atm/lanai.c
1755
lvcc->tx.unqueue(lanai, lvcc, endptr);
drivers/atm/lanai.c
1998
lvcc->tx.atmvcc != NULL && lvcc->tx.atmvcc != atmvcc)
drivers/atm/lanai.c
2290
if (lvcc->tx.atmvcc == atmvcc) {
drivers/atm/lanai.c
2297
lanai_buf_deallocate(&lvcc->tx.buf, lanai->pci);
drivers/atm/lanai.c
2298
lvcc->tx.atmvcc = NULL;
drivers/atm/lanai.c
2358
APRINTK(lvcc->tx.atmvcc == NULL, "tx.atmvcc!=NULL, vci=%d\n",
drivers/atm/lanai.c
2363
lvcc->tx.atmvcc = atmvcc;
drivers/atm/lanai.c
2378
if (atmvcc == lvcc->tx.atmvcc) {
drivers/atm/lanai.c
2397
lvcc->tx.atmvcc != atmvcc))
drivers/atm/lanai.c
241
} tx;
drivers/atm/lanai.c
2521
if (lvcc->tx.atmvcc != NULL)
drivers/atm/lanai.c
2524
lvcc->tx.atmvcc->qos.aal == ATM_AAL5 ? 5 : 0,
drivers/atm/lanai.c
2525
lanai_buf_size(&lvcc->tx.buf),
drivers/atm/lanai.c
2526
lvcc->tx.atmvcc == lanai->cbrvcc ? 'C' : 'U',
drivers/atm/lanai.c
719
dma_addr_t dmaaddr = lvcc->tx.buf.dmaaddr;
drivers/atm/lanai.c
727
(lvcc->tx.atmvcc->qos.txtp.traffic_class == ATM_CBR) ?
drivers/atm/lanai.c
732
TXADDR1_SET_SIZE(lanai_buf_size_cardorder(&lvcc->tx.buf)),
drivers/atm/lanai.c
772
while ((skb = skb_dequeue(&lvcc->tx.backlog)) != NULL)
drivers/atm/lanai.c
773
lanai_free_skb(lvcc->tx.atmvcc, skb);
drivers/atm/lanai.c
783
(((lanai_buf_size(&lvcc->tx.buf) / 1024) * HZ) >> 7);
drivers/atm/lanai.c
788
(lvcc->tx.atmvcc->qos.txtp.traffic_class != ATM_CBR ||
drivers/atm/lanai.c
799
lvcc->tx.atmvcc->dev->number, lvcc->vci);
drivers/atm/nicstar.c
1256
if (vcc->qos.txtp.traffic_class != ATM_NONE && vc->tx)
drivers/atm/nicstar.c
1378
vc->tx = 1;
drivers/atm/nicstar.c
1460
vc->tx = 0;
drivers/atm/nicstar.c
1641
if (!vc->tx) {
drivers/atm/nicstar.c
1713
atomic_inc(&vcc->stats->tx);
drivers/atm/nicstar.h
698
volatile unsigned int tx:1; /* TX vc? */
drivers/atm/solos-pci.c
1121
atomic_inc(&vcc->stats->tx);
drivers/block/aoe/aoenet.c
206
kts.fn = tx;
drivers/bus/moxtet.c
390
.tx_buf = moxtet->tx,
drivers/bus/moxtet.c
433
moxtet->tx[moxtet->count - mdev->idx] = val;
drivers/bus/moxtet.c
435
ret = spi_write(to_spi_device(moxtet->dev), moxtet->tx,
drivers/bus/moxtet.c
452
return moxtet->tx[moxtet->count - mdev->idx];
drivers/bus/moxtet.c
501
p = hex_byte_pack(p, moxtet->tx[moxtet->count - i]);
drivers/bus/moxtet.c
537
moxtet->tx[moxtet->count - i] = bin[i];
drivers/bus/moxtet.c
539
err = spi_write(to_spi_device(moxtet->dev), moxtet->tx,
drivers/clk/clk-rpmi.c
172
struct rpmi_get_attrs_tx tx;
drivers/clk/clk-rpmi.c
177
tx.clkid = cpu_to_le32(clkid);
drivers/clk/clk-rpmi.c
179
&tx, sizeof(tx), &rx, sizeof(rx));
drivers/clk/clk-rpmi.c
210
struct rpmi_get_supp_rates_tx tx;
drivers/clk/clk-rpmi.c
216
tx.clkid = cpu_to_le32(clkid);
drivers/clk/clk-rpmi.c
217
tx.clk_rate_idx = 0;
drivers/clk/clk-rpmi.c
229
&tx, sizeof(tx), rx, context->max_msg_data_size);
drivers/clk/clk-rpmi.c
259
tx.clk_rate_idx = cpu_to_le32(clk_rate_idx);
drivers/clk/clk-rpmi.c
263
&tx, sizeof(tx),
drivers/clk/clk-rpmi.c
306
struct rpmi_get_rate_tx tx;
drivers/clk/clk-rpmi.c
310
tx.clkid = cpu_to_le32(rpmi_clk->id);
drivers/clk/clk-rpmi.c
313
&tx, sizeof(tx), &rx, sizeof(rx));
drivers/clk/clk-rpmi.c
368
struct rpmi_set_rate_tx tx;
drivers/clk/clk-rpmi.c
372
tx.clkid = cpu_to_le32(rpmi_clk->id);
drivers/clk/clk-rpmi.c
373
tx.lo = cpu_to_le32(lower_32_bits(rate));
drivers/clk/clk-rpmi.c
374
tx.hi = cpu_to_le32(upper_32_bits(rate));
drivers/clk/clk-rpmi.c
377
&tx, sizeof(tx), &rx, sizeof(rx));
drivers/clk/clk-rpmi.c
397
struct rpmi_set_config_tx tx;
drivers/clk/clk-rpmi.c
401
tx.config = cpu_to_le32(RPMI_CLK_ENABLE);
drivers/clk/clk-rpmi.c
402
tx.clkid = cpu_to_le32(rpmi_clk->id);
drivers/clk/clk-rpmi.c
405
&tx, sizeof(tx), &rx, sizeof(rx));
drivers/clk/clk-rpmi.c
425
struct rpmi_set_config_tx tx;
drivers/clk/clk-rpmi.c
428
tx.config = cpu_to_le32(RPMI_CLK_DISABLE);
drivers/clk/clk-rpmi.c
429
tx.clkid = cpu_to_le32(rpmi_clk->id);
drivers/clk/clk-rpmi.c
432
&tx, sizeof(tx), &rx, sizeof(rx));
drivers/clk/mediatek/clk-apmixed.c
105
struct mtk_ref2usb_tx *tx = to_mtk_ref2usb_tx(hw);
drivers/clk/mediatek/clk-apmixed.c
108
kfree(tx);
drivers/clk/mediatek/clk-apmixed.c
32
struct mtk_ref2usb_tx *tx = to_mtk_ref2usb_tx(hw);
drivers/clk/mediatek/clk-apmixed.c
34
return (readl(tx->base_addr) & REF2USB_EN_MASK) == REF2USB_EN_MASK;
drivers/clk/mediatek/clk-apmixed.c
39
struct mtk_ref2usb_tx *tx = to_mtk_ref2usb_tx(hw);
drivers/clk/mediatek/clk-apmixed.c
42
val = readl(tx->base_addr);
drivers/clk/mediatek/clk-apmixed.c
45
writel(val, tx->base_addr);
drivers/clk/mediatek/clk-apmixed.c
49
writel(val, tx->base_addr);
drivers/clk/mediatek/clk-apmixed.c
52
writel(val, tx->base_addr);
drivers/clk/mediatek/clk-apmixed.c
59
struct mtk_ref2usb_tx *tx = to_mtk_ref2usb_tx(hw);
drivers/clk/mediatek/clk-apmixed.c
62
val = readl(tx->base_addr);
drivers/clk/mediatek/clk-apmixed.c
64
writel(val, tx->base_addr);
drivers/clk/mediatek/clk-apmixed.c
76
struct mtk_ref2usb_tx *tx;
drivers/clk/mediatek/clk-apmixed.c
80
tx = kzalloc_obj(*tx);
drivers/clk/mediatek/clk-apmixed.c
81
if (!tx)
drivers/clk/mediatek/clk-apmixed.c
84
tx->base_addr = reg;
drivers/clk/mediatek/clk-apmixed.c
85
tx->hw.init = &init;
drivers/clk/mediatek/clk-apmixed.c
92
ret = clk_hw_register(NULL, &tx->hw);
drivers/clk/mediatek/clk-apmixed.c
95
kfree(tx);
drivers/clk/mediatek/clk-apmixed.c
99
return &tx->hw;
drivers/clk/tegra/clk-bpmp.c
155
msg.tx.data = &request;
drivers/clk/tegra/clk-bpmp.c
156
msg.tx.size = sizeof(request);
drivers/clk/tegra/clk-bpmp.c
185
msg.tx.data = &request;
drivers/clk/tegra/clk-bpmp.c
186
msg.tx.size = sizeof(request);
drivers/clk/tegra/clk-bpmp.c
213
msg.tx.data = &request;
drivers/clk/tegra/clk-bpmp.c
214
msg.tx.size = sizeof(request);
drivers/clk/tegra/clk-bpmp.c
269
msg.tx.data = &request;
drivers/clk/tegra/clk-bpmp.c
270
msg.tx.size = sizeof(request);
drivers/clk/tegra/clk-bpmp.c
50
} tx;
drivers/clk/tegra/clk-bpmp.c
77
memcpy(req + 4, clk->tx.data, clk->tx.size);
drivers/clk/tegra/clk-bpmp.c
81
msg.tx.data = &request;
drivers/clk/tegra/clk-bpmp.c
82
msg.tx.size = sizeof(request);
drivers/comedi/drivers/ni_usb6501.c
174
u8 *tx = devpriv->usb_tx_buf;
drivers/comedi/drivers/ni_usb6501.c
186
memcpy(tx, READ_PORT_REQUEST, request_size);
drivers/comedi/drivers/ni_usb6501.c
187
tx[14] = val & 0xff;
drivers/comedi/drivers/ni_usb6501.c
192
memcpy(tx, WRITE_PORT_REQUEST, request_size);
drivers/comedi/drivers/ni_usb6501.c
193
tx[14] = val & 0xff;
drivers/comedi/drivers/ni_usb6501.c
194
tx[17] = *bitmap;
drivers/comedi/drivers/ni_usb6501.c
199
memcpy(tx, SET_PORT_DIR_REQUEST, request_size);
drivers/comedi/drivers/ni_usb6501.c
200
tx[14] = val & 0xff;
drivers/comedi/drivers/ni_usb6501.c
201
tx[15] = (val >> 8) & 0xff;
drivers/comedi/drivers/ni_usb6501.c
202
tx[16] = (val >> 16) & 0xff;
drivers/comedi/drivers/ni_usb6501.c
256
u8 *tx = devpriv->usb_tx_buf;
drivers/comedi/drivers/ni_usb6501.c
268
memcpy(tx, START_COUNTER_REQUEST, request_size);
drivers/comedi/drivers/ni_usb6501.c
273
memcpy(tx, STOP_COUNTER_REQUEST, request_size);
drivers/comedi/drivers/ni_usb6501.c
278
memcpy(tx, READ_COUNTER_REQUEST, request_size);
drivers/comedi/drivers/ni_usb6501.c
283
memcpy(tx, WRITE_COUNTER_REQUEST, request_size);
drivers/comedi/drivers/ni_usb6501.c
286
*((__be32 *)&tx[12]) = cpu_to_be32(*val);
drivers/cpufreq/tegra186-cpufreq.c
285
msg.tx.data = &req;
drivers/cpufreq/tegra186-cpufreq.c
286
msg.tx.size = sizeof(req);
drivers/cpufreq/tegra194-cpufreq.c
630
msg.tx.data = &req;
drivers/cpufreq/tegra194-cpufreq.c
631
msg.tx.size = sizeof(req);
drivers/crypto/gemini/sl3516-ce-core.c
103
dd = &ce->tx[ce->ctx];
drivers/crypto/gemini/sl3516-ce-core.c
35
ce->tx = dma_alloc_coherent(ce->dev, sz, &ce->dtx, GFP_KERNEL);
drivers/crypto/gemini/sl3516-ce-core.c
36
if (!ce->tx)
drivers/crypto/gemini/sl3516-ce-core.c
43
ce->tx[i].frame_ctrl.bits.own = CE_CPU;
drivers/crypto/gemini/sl3516-ce-core.c
44
ce->tx[i].next_desc.next_descriptor = ce->dtx + (i + 1) * sizeof(struct descriptor);
drivers/crypto/gemini/sl3516-ce-core.c
46
ce->tx[MAXDESC - 1].next_desc.next_descriptor = ce->dtx;
drivers/crypto/gemini/sl3516-ce-core.c
63
dma_free_coherent(ce->dev, sz, ce->tx, ce->dtx);
drivers/crypto/gemini/sl3516-ce-core.c
71
dma_free_coherent(ce->dev, sz, ce->tx, ce->dtx);
drivers/crypto/gemini/sl3516-ce.h
238
struct descriptor *tx;
drivers/crypto/intel/qat/qat_common/adf_bank_state.c
76
int tx = i * (tx_rx_gap + 1);
drivers/crypto/intel/qat/qat_common/adf_bank_state.c
77
int rx = tx + tx_rx_gap;
drivers/crypto/intel/qat/qat_common/adf_bank_state.c
79
ops->write_csr_ring_head(base, bank, tx, state->rings[tx].head);
drivers/crypto/intel/qat/qat_common/adf_bank_state.c
80
ops->write_csr_ring_tail(base, bank, tx, state->rings[tx].tail);
drivers/crypto/intel/qat/qat_common/adf_bank_state.c
87
if (state->ringestat & BIT(tx)) {
drivers/crypto/intel/qat/qat_common/adf_bank_state.c
91
ops->write_csr_ring_head(base, bank, tx, state->rings[tx].head);
drivers/crypto/omap-sham.c
573
struct dma_async_tx_descriptor *tx;
drivers/crypto/omap-sham.c
597
tx = dmaengine_prep_slave_sg(dd->dma_lch, ctx->sg, ctx->sg_len,
drivers/crypto/omap-sham.c
601
if (!tx) {
drivers/crypto/omap-sham.c
606
tx->callback = omap_sham_dma_callback;
drivers/crypto/omap-sham.c
607
tx->callback_param = dd;
drivers/crypto/omap-sham.c
619
dmaengine_submit(tx);
drivers/crypto/starfive/jh7110-aes.c
423
dmaengine_slave_config(cryp->tx, &cryp->cfg_in);
drivers/crypto/starfive/jh7110-aes.c
474
in_desc = dmaengine_prep_slave_sg(cryp->tx, src, 1, DMA_MEM_TO_DEV,
drivers/crypto/starfive/jh7110-aes.c
482
dma_async_issue_pending(cryp->tx);
drivers/crypto/starfive/jh7110-cryp.c
71
cryp->tx = dma_request_chan(cryp->dev, "tx");
drivers/crypto/starfive/jh7110-cryp.c
72
if (IS_ERR(cryp->tx))
drivers/crypto/starfive/jh7110-cryp.c
73
return dev_err_probe(cryp->dev, PTR_ERR(cryp->tx),
drivers/crypto/starfive/jh7110-cryp.c
78
dma_release_channel(cryp->tx);
drivers/crypto/starfive/jh7110-cryp.c
88
dma_release_channel(cryp->tx);
drivers/crypto/starfive/jh7110-cryp.h
187
struct dma_chan *tx;
drivers/crypto/starfive/jh7110-hash.c
119
dmaengine_slave_config(cryp->tx, &cryp->cfg_in);
drivers/crypto/starfive/jh7110-hash.c
139
in_desc = dmaengine_prep_slave_sg(cryp->tx, sg, 1, DMA_MEM_TO_DEV,
drivers/crypto/starfive/jh7110-hash.c
151
dma_async_issue_pending(cryp->tx);
drivers/dma/altera-msgdma.c
202
#define tx_to_desc(tx) container_of(tx, struct msgdma_sw_desc, async_tx)
drivers/dma/altera-msgdma.c
304
static dma_cookie_t msgdma_tx_submit(struct dma_async_tx_descriptor *tx)
drivers/dma/altera-msgdma.c
306
struct msgdma_device *mdev = to_mdev(tx->chan);
drivers/dma/altera-msgdma.c
311
new = tx_to_desc(tx);
drivers/dma/altera-msgdma.c
313
cookie = dma_cookie_assign(tx);
drivers/dma/amba-pl08x.c
1510
struct pl08x_txd *txd = to_pl08x_txd(&vd->tx);
drivers/dma/amba-pl08x.c
1511
struct pl08x_dma_chan *plchan = to_pl08x_chan(vd->tx.chan);
drivers/dma/amba-pl08x.c
1513
dma_descriptor_unmap(&vd->tx);
drivers/dma/amba-pl08x.c
1572
struct pl08x_txd *txd = to_pl08x_txd(&vd->tx);
drivers/dma/amba-pl08x.c
2308
struct pl08x_txd *tx;
drivers/dma/amba-pl08x.c
2318
tx = plchan->at;
drivers/dma/amba-pl08x.c
2319
if (tx && tx->cyclic) {
drivers/dma/amba-pl08x.c
2320
vchan_cyclic_callback(&tx->vd);
drivers/dma/amba-pl08x.c
2321
} else if (tx) {
drivers/dma/amba-pl08x.c
2328
tx->done = true;
drivers/dma/amba-pl08x.c
2329
vchan_cookie_complete(&tx->vd);
drivers/dma/amba-pl08x.c
320
static inline struct pl08x_txd *to_pl08x_txd(struct dma_async_tx_descriptor *tx)
drivers/dma/amba-pl08x.c
322
return container_of(tx, struct pl08x_txd, vd.tx);
drivers/dma/amba-pl08x.c
532
struct pl08x_txd *txd = to_pl08x_txd(&vd->tx);
drivers/dma/amd/ptdma/ptdma-dmaengine.c
214
tx_desc = &desc->vd.tx;
drivers/dma/amd/ptdma/ptdma-dmaengine.c
274
dma_chan = desc->vd.tx.chan;
drivers/dma/amd/ptdma/ptdma-dmaengine.c
348
dma_chan = desc->vd.tx.chan;
drivers/dma/amd/ptdma/ptdma-dmaengine.c
354
tx_desc = &desc->vd.tx;
drivers/dma/amd/ptdma/ptdma-dmaengine.c
432
return &desc->vd.tx;
drivers/dma/amd/ptdma/ptdma-dmaengine.c
445
return &desc->vd.tx;
drivers/dma/amd/qdma/qdma.c
744
static dma_cookie_t qdma_tx_submit(struct dma_async_tx_descriptor *tx)
drivers/dma/amd/qdma/qdma.c
746
struct virt_dma_chan *vc = to_virt_chan(tx->chan);
drivers/dma/amd/qdma/qdma.c
752
vd = container_of(tx, struct virt_dma_desc, tx);
drivers/dma/amd/qdma/qdma.c
754
cookie = dma_cookie_assign(tx);
drivers/dma/amd/qdma/qdma.c
769
struct dma_async_tx_descriptor *tx;
drivers/dma/amd/qdma/qdma.c
782
tx = vchan_tx_prep(&q->vchan, &vdesc->vdesc, flags);
drivers/dma/amd/qdma/qdma.c
783
tx->tx_submit = qdma_tx_submit;
drivers/dma/amd/qdma/qdma.c
785
return tx;
drivers/dma/apple-admac.c
137
struct dma_async_tx_descriptor tx;
drivers/dma/apple-admac.c
219
static struct admac_tx *to_admac_tx(struct dma_async_tx_descriptor *tx)
drivers/dma/apple-admac.c
221
return container_of(tx, struct admac_tx, tx);
drivers/dma/apple-admac.c
230
static dma_cookie_t admac_tx_submit(struct dma_async_tx_descriptor *tx)
drivers/dma/apple-admac.c
232
struct admac_tx *adtx = to_admac_tx(tx);
drivers/dma/apple-admac.c
233
struct admac_chan *adchan = to_admac_chan(tx->chan);
drivers/dma/apple-admac.c
238
cookie = dma_cookie_assign(tx);
drivers/dma/apple-admac.c
245
static int admac_desc_free(struct dma_async_tx_descriptor *tx)
drivers/dma/apple-admac.c
247
kfree(to_admac_tx(tx));
drivers/dma/apple-admac.c
277
dma_async_tx_descriptor_init(&adtx->tx, chan);
drivers/dma/apple-admac.c
278
adtx->tx.tx_submit = admac_tx_submit;
drivers/dma/apple-admac.c
279
adtx->tx.desc_free = admac_desc_free;
drivers/dma/apple-admac.c
281
return &adtx->tx;
drivers/dma/apple-admac.c
288
struct admac_tx *tx)
drivers/dma/apple-admac.c
292
addr = tx->buf_addr + (tx->submitted_pos % tx->buf_len);
drivers/dma/apple-admac.c
295
WARN_ON_ONCE(addr + tx->period_len > tx->buf_end);
drivers/dma/apple-admac.c
298
channo, &addr, tx->period_len, FLAG_DESC_NOTIFY);
drivers/dma/apple-admac.c
302
writel_relaxed(tx->period_len, ad->base + REG_DESC_WRITE(channo));
drivers/dma/apple-admac.c
305
tx->submitted_pos += tx->period_len;
drivers/dma/apple-admac.c
306
tx->submitted_pos %= 2 * tx->buf_len;
drivers/dma/apple-admac.c
314
struct admac_tx *tx)
drivers/dma/apple-admac.c
321
admac_cyclic_write_one_desc(ad, channo, tx);
drivers/dma/apple-admac.c
392
if (adtx && adtx->tx.cookie == cookie) {
drivers/dma/apple-admac.c
399
if (adtx->tx.cookie == cookie) {
drivers/dma/apple-admac.c
477
struct admac_tx *tx;
drivers/dma/apple-admac.c
483
tx = list_first_entry(&adchan->issued, struct admac_tx, node);
drivers/dma/apple-admac.c
484
list_del(&tx->node);
drivers/dma/apple-admac.c
486
adchan->current_tx = tx;
drivers/dma/apple-admac.c
550
admac_desc_free(&adtx->tx);
drivers/dma/apple-admac.c
658
struct admac_tx *tx = adchan->current_tx;
drivers/dma/apple-admac.c
661
tx->reclaimed_pos += nreports * tx->period_len;
drivers/dma/apple-admac.c
662
tx->reclaimed_pos %= 2 * tx->buf_len;
drivers/dma/apple-admac.c
664
admac_cyclic_write_desc(ad, channo, tx);
drivers/dma/apple-admac.c
735
dmaengine_desc_get_callback(&adtx->tx, &cb);
drivers/dma/arm-dma350.c
409
dch->cookie = dch->desc->vd.tx.cookie;
drivers/dma/at_hdmac.c
368
return container_of(t, struct at_desc, vd.tx);
drivers/dma/at_hdmac.c
571
atchan->desc = desc = to_atdma_desc(&vd->tx);
drivers/dma/at_hdmac.c
594
struct at_dma *atdma = to_at_dma(vd->tx.chan->device);
drivers/dma/at_hdmac.c
595
struct at_desc *desc = to_atdma_desc(&vd->tx);
drivers/dma/at_hdmac.c
756
desc = to_atdma_desc(&vd->tx);
drivers/dma/at_hdmac.c
757
else if (atchan->desc && atchan->desc->vd.tx.cookie == cookie)
drivers/dma/at_hdmac.c
795
desc->vd.tx.cookie);
drivers/dma/at_xdmac.c
542
static dma_cookie_t at_xdmac_tx_submit(struct dma_async_tx_descriptor *tx)
drivers/dma/at_xdmac.c
544
struct at_xdmac_desc *desc = txd_to_at_desc(tx);
drivers/dma/at_xdmac.c
545
struct at_xdmac_chan *atchan = to_at_xdmac_chan(tx->chan);
drivers/dma/at_xdmac.c
550
cookie = dma_cookie_assign(tx);
drivers/dma/at_xdmac.c
555
dev_vdbg(chan2dev(tx->chan), "%s: atchan 0x%p, add desc 0x%p to xfers_list\n",
drivers/dma/bcm-sba-raid.c
1072
req->tx.flags = flags;
drivers/dma/bcm-sba-raid.c
1073
req->tx.cookie = -EBUSY;
drivers/dma/bcm-sba-raid.c
1089
dma_addr_t resp_dma = req->tx.phys;
drivers/dma/bcm-sba-raid.c
113
struct dma_async_tx_descriptor tx;
drivers/dma/bcm-sba-raid.c
1327
req->tx.flags = flags;
drivers/dma/bcm-sba-raid.c
1328
req->tx.cookie = -EBUSY;
drivers/dma/bcm-sba-raid.c
1424
return (first) ? &first->tx : NULL;
drivers/dma/bcm-sba-raid.c
1512
dma_async_tx_descriptor_init(&req->tx, &sba->dma_chan);
drivers/dma/bcm-sba-raid.c
1513
async_tx_ack(&req->tx);
drivers/dma/bcm-sba-raid.c
1514
req->tx.tx_submit = sba_tx_submit;
drivers/dma/bcm-sba-raid.c
1515
req->tx.phys = sba->resp_dma_base + i * sba->hw_resp_size;
drivers/dma/bcm-sba-raid.c
203
if (async_tx_test_ack(&req->tx)) {
drivers/dma/bcm-sba-raid.c
227
dma_async_tx_descriptor_init(&req->tx, &sba->dma_chan);
drivers/dma/bcm-sba-raid.c
228
async_tx_ack(&req->tx);
drivers/dma/bcm-sba-raid.c
411
struct dma_async_tx_descriptor *tx;
drivers/dma/bcm-sba-raid.c
416
tx = &first->tx;
drivers/dma/bcm-sba-raid.c
418
WARN_ON(tx->cookie < 0);
drivers/dma/bcm-sba-raid.c
419
if (tx->cookie > 0) {
drivers/dma/bcm-sba-raid.c
421
dma_cookie_complete(tx);
drivers/dma/bcm-sba-raid.c
423
dmaengine_desc_get_callback_invoke(tx, NULL);
drivers/dma/bcm-sba-raid.c
424
dma_descriptor_unmap(tx);
drivers/dma/bcm-sba-raid.c
425
tx->callback = NULL;
drivers/dma/bcm-sba-raid.c
426
tx->callback_result = NULL;
drivers/dma/bcm-sba-raid.c
429
dma_run_dependencies(tx);
drivers/dma/bcm-sba-raid.c
459
if (async_tx_test_ack(&req->tx))
drivers/dma/bcm-sba-raid.c
515
static dma_cookie_t sba_tx_submit(struct dma_async_tx_descriptor *tx)
drivers/dma/bcm-sba-raid.c
522
if (unlikely(!tx))
drivers/dma/bcm-sba-raid.c
525
sba = to_sba_device(tx->chan);
drivers/dma/bcm-sba-raid.c
526
req = to_sba_request(tx);
drivers/dma/bcm-sba-raid.c
530
cookie = dma_cookie_assign(tx);
drivers/dma/bcm-sba-raid.c
561
dma_addr_t resp_dma = req->tx.phys;
drivers/dma/bcm-sba-raid.c
635
req->tx.flags = flags;
drivers/dma/bcm-sba-raid.c
636
req->tx.cookie = -EBUSY;
drivers/dma/bcm-sba-raid.c
638
return &req->tx;
drivers/dma/bcm-sba-raid.c
649
dma_addr_t resp_dma = req->tx.phys;
drivers/dma/bcm-sba-raid.c
721
req->tx.flags = flags;
drivers/dma/bcm-sba-raid.c
722
req->tx.cookie = -EBUSY;
drivers/dma/bcm-sba-raid.c
757
return (first) ? &first->tx : NULL;
drivers/dma/bcm-sba-raid.c
769
dma_addr_t resp_dma = req->tx.phys;
drivers/dma/bcm-sba-raid.c
85
#define to_sba_request(tx) \
drivers/dma/bcm-sba-raid.c
86
container_of(tx, struct sba_request, tx)
drivers/dma/bcm-sba-raid.c
860
req->tx.flags = flags;
drivers/dma/bcm-sba-raid.c
861
req->tx.cookie = -EBUSY;
drivers/dma/bcm-sba-raid.c
900
return (first) ? &first->tx : NULL;
drivers/dma/bcm-sba-raid.c
914
dma_addr_t resp_dma = req->tx.phys;
drivers/dma/bcm2835-dma.c
201
return container_of(t, struct bcm2835_desc, vd.tx);
drivers/dma/bcm2835-dma.c
433
c->desc = d = to_bcm2835_dma_desc(&vd->tx);
drivers/dma/bcm2835-dma.c
561
bcm2835_dma_desc_size(to_bcm2835_dma_desc(&vd->tx));
drivers/dma/bcm2835-dma.c
562
} else if (c->desc && c->desc->vd.tx.cookie == cookie) {
drivers/dma/dma-axi-dmac.c
293
if (chan->hw_cyclic && desc->cyclic && !desc->vdesc.tx.callback) {
drivers/dma/dma-jz4780.c
265
struct jz4780_dma_chan *jzchan = to_jz4780_dma_chan(vdesc->tx.chan);
drivers/dma/dma-jz4780.c
503
if (jzchan->desc->type == DMA_CYCLIC && vdesc->tx.callback) {
drivers/dma/dma-jz4780.c
661
} else if (cookie == jzchan->desc->vdesc.tx.cookie) {
drivers/dma/dmaengine.c
1473
void dma_async_tx_descriptor_init(struct dma_async_tx_descriptor *tx,
drivers/dma/dmaengine.c
1476
tx->chan = chan;
drivers/dma/dmaengine.c
1478
spin_lock_init(&tx->lock);
drivers/dma/dmaengine.c
1561
dma_wait_for_async_tx(struct dma_async_tx_descriptor *tx)
drivers/dma/dmaengine.c
1565
if (!tx)
drivers/dma/dmaengine.c
1568
while (tx->cookie == -EBUSY) {
drivers/dma/dmaengine.c
1570
dev_err(tx->chan->device->dev,
drivers/dma/dmaengine.c
1577
return dma_sync_wait(tx->chan, tx->cookie);
drivers/dma/dmaengine.c
1588
void dma_run_dependencies(struct dma_async_tx_descriptor *tx)
drivers/dma/dmaengine.c
1590
struct dma_async_tx_descriptor *dep = txd_next(tx);
drivers/dma/dmaengine.c
1598
txd_clear_next(tx);
drivers/dma/dmaengine.h
114
dmaengine_desc_get_callback(struct dma_async_tx_descriptor *tx,
drivers/dma/dmaengine.h
117
cb->callback = tx->callback;
drivers/dma/dmaengine.h
118
cb->callback_result = tx->callback_result;
drivers/dma/dmaengine.h
119
cb->callback_param = tx->callback_param;
drivers/dma/dmaengine.h
160
dmaengine_desc_get_callback_invoke(struct dma_async_tx_descriptor *tx,
drivers/dma/dmaengine.h
165
dmaengine_desc_get_callback(tx, &cb);
drivers/dma/dmaengine.h
29
static inline dma_cookie_t dma_cookie_assign(struct dma_async_tx_descriptor *tx)
drivers/dma/dmaengine.h
31
struct dma_chan *chan = tx->chan;
drivers/dma/dmaengine.h
37
tx->cookie = chan->cookie = cookie;
drivers/dma/dmaengine.h
52
static inline void dma_cookie_complete(struct dma_async_tx_descriptor *tx)
drivers/dma/dmaengine.h
54
BUG_ON(tx->cookie < DMA_MIN_COOKIE);
drivers/dma/dmaengine.h
55
tx->chan->completed_cookie = tx->cookie;
drivers/dma/dmaengine.h
56
tx->cookie = 0;
drivers/dma/dmatest.c
699
struct dma_async_tx_descriptor *tx = NULL;
drivers/dma/dmatest.c
794
tx = dev->device_prep_dma_memcpy(chan,
drivers/dma/dmatest.c
798
tx = dev->device_prep_dma_memset(chan,
drivers/dma/dmatest.c
803
tx = dev->device_prep_dma_xor(chan,
drivers/dma/dmatest.c
810
tx = dev->device_prep_dma_pq(chan, dma_pq, srcs,
drivers/dma/dmatest.c
815
if (!tx) {
drivers/dma/dmatest.c
824
tx->callback = dmatest_callback;
drivers/dma/dmatest.c
825
tx->callback_param = done;
drivers/dma/dmatest.c
827
cookie = tx->tx_submit(tx);
drivers/dma/dw-axi-dmac/dw-axi-dmac-platform.c
1081
axi_chan_name(chan), vd->tx.cookie, status);
drivers/dma/dw-axi-dmac/dw-axi-dmac-platform.c
494
vd->tx.cookie);
drivers/dma/dw-edma/dw-edma-core.c
594
if (!vd->tx.callback_result)
drivers/dma/dw/core.c
54
static dma_cookie_t dwc_tx_submit(struct dma_async_tx_descriptor *tx)
drivers/dma/dw/core.c
56
struct dw_desc *desc = txd_to_dw_desc(tx);
drivers/dma/dw/core.c
57
struct dw_dma_chan *dwc = to_dw_dma_chan(tx->chan);
drivers/dma/dw/core.c
62
cookie = dma_cookie_assign(tx);
drivers/dma/dw/core.c
72
dev_vdbg(chan2dev(tx->chan), "%s: queued %u\n",
drivers/dma/ep93xx_dma.c
888
static dma_cookie_t ep93xx_dma_tx_submit(struct dma_async_tx_descriptor *tx)
drivers/dma/ep93xx_dma.c
890
struct ep93xx_dma_chan *edmac = to_ep93xx_dma_chan(tx->chan);
drivers/dma/ep93xx_dma.c
896
cookie = dma_cookie_assign(tx);
drivers/dma/ep93xx_dma.c
898
desc = container_of(tx, struct ep93xx_dma_desc, txd);
drivers/dma/fsl-edma-common.c
416
if (fsl_chan->edesc && cookie == fsl_chan->edesc->vdesc.tx.cookie)
drivers/dma/fsl_raid.c
429
struct dma_async_tx_descriptor *tx;
drivers/dma/fsl_raid.c
437
tx = fsl_re_prep_dma_genq(chan, dest[1], dma_src, 2, coef, len,
drivers/dma/fsl_raid.c
439
if (tx)
drivers/dma/fsl_raid.c
440
desc = to_fsl_re_dma_desc(tx);
drivers/dma/fsl_raid.c
442
return tx;
drivers/dma/fsl_raid.c
84
#define to_fsl_re_dma_desc(tx) container_of(tx, struct fsl_re_desc, async_tx)
drivers/dma/fsl_raid.c
87
static dma_cookie_t fsl_re_tx_submit(struct dma_async_tx_descriptor *tx)
drivers/dma/fsl_raid.c
94
desc = to_fsl_re_dma_desc(tx);
drivers/dma/fsl_raid.c
95
re_chan = container_of(tx->chan, struct fsl_re_chan, chan);
drivers/dma/fsl_raid.c
98
cookie = dma_cookie_assign(tx);
drivers/dma/fsldma.c
407
static dma_cookie_t fsl_dma_tx_submit(struct dma_async_tx_descriptor *tx)
drivers/dma/fsldma.c
409
struct fsldma_chan *chan = to_fsl_chan(tx->chan);
drivers/dma/fsldma.c
410
struct fsl_desc_sw *desc = tx_to_fsl_desc(tx);
drivers/dma/fsldma.h
193
#define tx_to_fsl_desc(tx) container_of(tx, struct fsl_desc_sw, async_tx)
drivers/dma/hsu/hsu.c
341
if (hsuc->desc && cookie == hsuc->desc->vdesc.tx.cookie) {
drivers/dma/idma64.c
232
struct idma64_chan *idma64c = to_idma64_chan(vdesc->tx.chan);
drivers/dma/idma64.c
388
if (idma64c->desc && cookie == idma64c->desc->vdesc.tx.cookie) {
drivers/dma/idxd/device.c
1291
struct dma_async_tx_descriptor *tx;
drivers/dma/idxd/device.c
1301
tx = &desc->txd;
drivers/dma/idxd/device.c
1302
tx->callback = NULL;
drivers/dma/idxd/device.c
1303
tx->callback_result = NULL;
drivers/dma/idxd/dma.c
171
static dma_cookie_t idxd_dma_tx_submit(struct dma_async_tx_descriptor *tx)
drivers/dma/idxd/dma.c
173
struct dma_chan *c = tx->chan;
drivers/dma/idxd/dma.c
177
struct idxd_desc *desc = container_of(tx, struct idxd_desc, txd);
drivers/dma/idxd/dma.c
179
cookie = dma_cookie_assign(tx);
drivers/dma/idxd/dma.c
28
struct dma_async_tx_descriptor *tx;
drivers/dma/idxd/dma.c
46
tx = &desc->txd;
drivers/dma/idxd/dma.c
47
if (complete && tx->cookie) {
drivers/dma/idxd/dma.c
48
dma_cookie_complete(tx);
drivers/dma/idxd/dma.c
49
dma_descriptor_unmap(tx);
drivers/dma/idxd/dma.c
50
dmaengine_desc_get_callback_invoke(tx, &res);
drivers/dma/idxd/dma.c
51
tx->callback = NULL;
drivers/dma/idxd/dma.c
52
tx->callback_result = NULL;
drivers/dma/img-mdc-dma.c
175
struct virt_dma_desc *vdesc = container_of(t, struct virt_dma_desc, tx);
drivers/dma/img-mdc-dma.c
278
struct mdc_tx_desc *mdesc = to_mdc_desc(&vd->tx);
drivers/dma/img-mdc-dma.c
536
mdesc = to_mdc_desc(&vd->tx);
drivers/dma/img-mdc-dma.c
590
mdesc = to_mdc_desc(&vd->tx);
drivers/dma/img-mdc-dma.c
592
} else if (mchan->desc && mchan->desc->vd.tx.cookie == cookie) {
drivers/dma/imx-dma.c
723
static dma_cookie_t imxdma_tx_submit(struct dma_async_tx_descriptor *tx)
drivers/dma/imx-dma.c
725
struct imxdma_channel *imxdmac = to_imxdma_chan(tx->chan);
drivers/dma/imx-dma.c
732
cookie = dma_cookie_assign(tx);
drivers/dma/imx-sdma.c
1892
desc = to_sdma_desc(&vd->tx);
drivers/dma/imx-sdma.c
1893
else if (sdmac->desc && sdmac->desc->vd.tx.cookie == cookie)
drivers/dma/imx-sdma.c
836
return container_of(t, struct sdma_desc, vd.tx);
drivers/dma/imx-sdma.c
850
sdmac->desc = desc = to_sdma_desc(&vd->tx);
drivers/dma/imx-sdma.c
900
dmaengine_desc_get_callback_invoke(&desc->vd.tx, NULL);
drivers/dma/ioat/dma.c
309
static dma_cookie_t ioat_tx_submit_unlock(struct dma_async_tx_descriptor *tx)
drivers/dma/ioat/dma.c
312
struct dma_chan *c = tx->chan;
drivers/dma/ioat/dma.c
316
cookie = dma_cookie_assign(tx);
drivers/dma/ioat/dma.c
614
struct dma_async_tx_descriptor *tx;
drivers/dma/ioat/dma.c
624
tx = &desc->txd;
drivers/dma/ioat/dma.c
625
if (tx->cookie) {
drivers/dma/ioat/dma.c
626
dma_cookie_complete(tx);
drivers/dma/ioat/dma.c
627
dma_descriptor_unmap(tx);
drivers/dma/ioat/dma.c
628
dmaengine_desc_get_callback_invoke(tx, NULL);
drivers/dma/ioat/dma.c
629
tx->callback = NULL;
drivers/dma/ioat/dma.c
630
tx->callback_result = NULL;
drivers/dma/ioat/dma.c
633
if (tx->phys == phys_complete)
drivers/dma/ioat/dma.c
737
struct dma_async_tx_descriptor *tx;
drivers/dma/ioat/dma.c
742
tx = &desc->txd;
drivers/dma/ioat/dma.c
743
if (tx->cookie) {
drivers/dma/ioat/dma.c
746
dma_cookie_complete(tx);
drivers/dma/ioat/dma.c
747
dma_descriptor_unmap(tx);
drivers/dma/ioat/dma.c
749
dmaengine_desc_get_callback_invoke(tx, &res);
drivers/dma/ioat/dma.c
750
tx->callback = NULL;
drivers/dma/ioat/dma.c
751
tx->callback_result = NULL;
drivers/dma/ioat/dma.c
778
struct dma_async_tx_descriptor *tx;
drivers/dma/ioat/dma.c
847
tx = &desc->txd;
drivers/dma/ioat/dma.c
848
if (tx->cookie) {
drivers/dma/ioat/dma.c
849
dma_cookie_complete(tx);
drivers/dma/ioat/dma.c
850
dma_descriptor_unmap(tx);
drivers/dma/ioat/dma.c
851
dmaengine_desc_get_callback_invoke(tx, &res);
drivers/dma/ioat/dma.c
852
tx->callback = NULL;
drivers/dma/ioat/dma.c
853
tx->callback_result = NULL;
drivers/dma/ioat/dma.h
222
struct dma_async_tx_descriptor *tx, int id)
drivers/dma/ioat/dma.h
228
(unsigned long long) tx->phys,
drivers/dma/ioat/dma.h
229
(unsigned long long) hw->next, tx->cookie, tx->flags,
drivers/dma/ioat/init.c
308
struct dma_async_tx_descriptor *tx;
drivers/dma/ioat/init.c
351
tx = ioat_dma->dma_dev.device_prep_dma_memcpy(dma_chan, dma_dest,
drivers/dma/ioat/init.c
354
if (!tx) {
drivers/dma/ioat/init.c
360
async_tx_ack(tx);
drivers/dma/ioat/init.c
362
tx->callback = ioat_dma_test_callback;
drivers/dma/ioat/init.c
363
tx->callback_param = &cmp;
drivers/dma/ioat/init.c
364
cookie = tx->tx_submit(tx);
drivers/dma/ioat/init.c
780
struct dma_async_tx_descriptor *tx;
drivers/dma/ioat/init.c
854
tx = dma->device_prep_dma_xor(dma_chan, dest_dma, dma_srcs,
drivers/dma/ioat/init.c
858
if (!tx) {
drivers/dma/ioat/init.c
864
async_tx_ack(tx);
drivers/dma/ioat/init.c
866
tx->callback = ioat_dma_test_callback;
drivers/dma/ioat/init.c
867
tx->callback_param = &cmp;
drivers/dma/ioat/init.c
868
cookie = tx->tx_submit(tx);
drivers/dma/ioat/init.c
923
tx = dma->device_prep_dma_xor_val(dma_chan, dma_srcs,
drivers/dma/ioat/init.c
926
if (!tx) {
drivers/dma/ioat/init.c
932
async_tx_ack(tx);
drivers/dma/ioat/init.c
934
tx->callback = ioat_dma_test_callback;
drivers/dma/ioat/init.c
935
tx->callback_param = &cmp;
drivers/dma/ioat/init.c
936
cookie = tx->tx_submit(tx);
drivers/dma/ioat/init.c
976
tx = dma->device_prep_dma_xor_val(dma_chan, dma_srcs,
drivers/dma/ioat/init.c
979
if (!tx) {
drivers/dma/ioat/init.c
985
async_tx_ack(tx);
drivers/dma/ioat/init.c
987
tx->callback = ioat_dma_test_callback;
drivers/dma/ioat/init.c
988
tx->callback_param = &cmp;
drivers/dma/ioat/init.c
989
cookie = tx->tx_submit(tx);
drivers/dma/k3dma.c
708
struct k3_dma_dev *d = to_k3_dma(vd->tx.chan->device);
drivers/dma/lgm/lgm-dma.c
1309
struct dma_async_tx_descriptor *tx = &c->ds->vdesc.tx;
drivers/dma/lgm/lgm-dma.c
1319
dmaengine_desc_get_callback(tx, &cb);
drivers/dma/lgm/lgm-dma.c
1320
dma_cookie_complete(tx);
drivers/dma/lgm/lgm-dma.c
1324
dmaengine_desc_get_callback(tx, &cb);
drivers/dma/lgm/lgm-dma.c
1325
dma_cookie_complete(tx);
drivers/dma/lgm/lgm-dma.c
663
struct dma_async_tx_descriptor *tx;
drivers/dma/lgm/lgm-dma.c
688
tx = &ds->vdesc.tx;
drivers/dma/lgm/lgm-dma.c
689
dma_async_tx_descriptor_init(tx, chan);
drivers/dma/lgm/lgm-dma.c
691
return tx;
drivers/dma/loongson1-apb-dma.c
192
struct ls1x_dma_chan *chan = to_ls1x_dma_chan(vd->tx.chan);
drivers/dma/loongson2-apb-dma.c
171
struct ls2x_dma_chan *lchan = to_ldma_chan(vdesc->tx.chan);
drivers/dma/mediatek/mtk-cqdma.c
371
dma_run_dependencies(&cvd->vd.tx);
drivers/dma/mediatek/mtk-cqdma.c
425
if (vd->tx.cookie == cookie) {
drivers/dma/mediatek/mtk-cqdma.c
430
if (vd->tx.cookie == cookie)
drivers/dma/mediatek/mtk-cqdma.c
489
struct dma_async_tx_descriptor *tx = NULL, *prev_tx = NULL;
drivers/dma/mediatek/mtk-cqdma.c
526
tx = vchan_tx_prep(to_virt_chan(c), &cvd[i]->vd, flags);
drivers/dma/mediatek/mtk-cqdma.c
527
tx->next = NULL;
drivers/dma/mediatek/mtk-cqdma.c
532
prev_tx->next = tx;
drivers/dma/mediatek/mtk-cqdma.c
542
prev_tx = tx;
drivers/dma/mediatek/mtk-cqdma.c
545
return &cvd[0]->vd.tx;
drivers/dma/mediatek/mtk-hsdma.c
583
hvc = to_hsdma_vchan(cb->vd->tx.chan);
drivers/dma/mediatek/mtk-hsdma.c
666
if (vd->tx.cookie == cookie)
drivers/dma/mediatek/mtk-hsdma.c
670
if (vd->tx.cookie == cookie)
drivers/dma/mediatek/mtk-uart-apdma.c
117
return container_of(t, struct mtk_uart_apdma_desc, vd.tx);
drivers/dma/mediatek/mtk-uart-apdma.c
374
c->desc = to_mtk_uart_apdma_desc(&vd->tx);
drivers/dma/milbeaut-hdmac.c
345
if (mc->md && mc->md->vd.tx.cookie == cookie) {
drivers/dma/mmp_pdma.c
189
#define tx_to_mmp_pdma_desc(tx) \
drivers/dma/mmp_pdma.c
190
container_of(tx, struct mmp_pdma_desc_sw, async_tx)
drivers/dma/mmp_pdma.c
498
static dma_cookie_t mmp_pdma_tx_submit(struct dma_async_tx_descriptor *tx)
drivers/dma/mmp_pdma.c
500
struct mmp_pdma_chan *chan = to_mmp_pdma_chan(tx->chan);
drivers/dma/mmp_pdma.c
501
struct mmp_pdma_desc_sw *desc = tx_to_mmp_pdma_desc(tx);
drivers/dma/mmp_tdma.c
372
static dma_cookie_t mmp_tdma_tx_submit(struct dma_async_tx_descriptor *tx)
drivers/dma/mmp_tdma.c
374
struct mmp_tdma_chan *tdmac = to_mmp_tdma_chan(tx->chan);
drivers/dma/moxart-dma.c
170
return container_of(t, struct moxart_desc, vd.tx);
drivers/dma/moxart-dma.c
424
ch->desc = to_moxart_dma_desc(&vd->tx);
drivers/dma/moxart-dma.c
486
d = to_moxart_dma_desc(&vd->tx);
drivers/dma/moxart-dma.c
488
} else if (ch->desc && ch->desc->vd.tx.cookie == cookie) {
drivers/dma/mv_xor.c
382
mv_xor_tx_submit(struct dma_async_tx_descriptor *tx)
drivers/dma/mv_xor.c
384
struct mv_xor_desc_slot *sw_desc = to_mv_xor_slot(tx);
drivers/dma/mv_xor.c
385
struct mv_xor_chan *mv_chan = to_mv_xor_chan(tx->chan);
drivers/dma/mv_xor.c
395
cookie = dma_cookie_assign(tx);
drivers/dma/mv_xor.c
42
#define to_mv_xor_slot(tx) \
drivers/dma/mv_xor.c
43
container_of(tx, struct mv_xor_desc_slot, async_tx)
drivers/dma/mv_xor.c
768
struct dma_async_tx_descriptor *tx;
drivers/dma/mv_xor.c
823
tx = mv_xor_prep_dma_memcpy(dma_chan, dest_dma, src_dma,
drivers/dma/mv_xor.c
825
if (!tx) {
drivers/dma/mv_xor.c
832
cookie = mv_xor_tx_submit(tx);
drivers/dma/mv_xor.c
841
async_tx_ack(tx);
drivers/dma/mv_xor.c
879
struct dma_async_tx_descriptor *tx;
drivers/dma/mv_xor.c
956
tx = mv_xor_prep_dma_xor(dma_chan, dest_dma, dma_srcs,
drivers/dma/mv_xor.c
958
if (!tx) {
drivers/dma/mv_xor.c
965
cookie = mv_xor_tx_submit(tx);
drivers/dma/mv_xor.c
974
async_tx_ack(tx);
drivers/dma/mv_xor_v2.c
303
mv_xor_v2_tx_submit(struct dma_async_tx_descriptor *tx)
drivers/dma/mv_xor_v2.c
308
container_of(tx, struct mv_xor_v2_sw_desc, async_tx);
drivers/dma/mv_xor_v2.c
310
container_of(tx->chan, struct mv_xor_v2_device, dmachan);
drivers/dma/mv_xor_v2.c
318
cookie = dma_cookie_assign(tx);
drivers/dma/mxs-dma.c
298
static dma_cookie_t mxs_dma_tx_submit(struct dma_async_tx_descriptor *tx)
drivers/dma/mxs-dma.c
300
return dma_cookie_assign(tx);
drivers/dma/nbpfaxi.c
672
static dma_cookie_t nbpf_tx_submit(struct dma_async_tx_descriptor *tx)
drivers/dma/nbpfaxi.c
674
struct nbpf_desc *desc = container_of(tx, struct nbpf_desc, async_tx);
drivers/dma/nbpfaxi.c
680
cookie = dma_cookie_assign(tx);
drivers/dma/owl-dma.c
305
static inline struct owl_dma_txd *to_owl_txd(struct dma_async_tx_descriptor *tx)
drivers/dma/owl-dma.c
307
return container_of(tx, struct owl_dma_txd, vd.tx);
drivers/dma/owl-dma.c
540
struct owl_dma_txd *txd = to_owl_txd(&vd->tx);
drivers/dma/owl-dma.c
691
struct owl_dma *od = to_owl_dma(vd->tx.chan->device);
drivers/dma/owl-dma.c
692
struct owl_dma_txd *txd = to_owl_txd(&vd->tx);
drivers/dma/owl-dma.c
824
txd = to_owl_txd(&vd->tx);
drivers/dma/pl330.c
2036
to_desc(struct dma_async_tx_descriptor *tx)
drivers/dma/pl330.c
2038
return container_of(tx, struct dma_pl330_desc, txd);
drivers/dma/pl330.c
2501
static dma_cookie_t pl330_tx_submit(struct dma_async_tx_descriptor *tx)
drivers/dma/pl330.c
2503
struct dma_pl330_desc *desc, *last = to_desc(tx);
drivers/dma/pl330.c
2504
struct dma_pl330_chan *pch = to_pchan(tx->chan);
drivers/dma/ppc4xx/adma.c
1225
ppc440spe_adma_tx_submit(struct dma_async_tx_descriptor *tx);
drivers/dma/ppc4xx/adma.c
1227
static void ppc440spe_adma_set_dest(struct ppc440spe_adma_desc_slot *tx,
drivers/dma/ppc4xx/adma.c
1230
ppc440spe_adma_memcpy_xor_set_src(struct ppc440spe_adma_desc_slot *tx,
drivers/dma/ppc4xx/adma.c
1234
ppc440spe_adma_pq_set_dest(struct ppc440spe_adma_desc_slot *tx,
drivers/dma/ppc4xx/adma.c
1237
ppc440spe_adma_pq_set_src(struct ppc440spe_adma_desc_slot *tx,
drivers/dma/ppc4xx/adma.c
1240
ppc440spe_adma_pq_set_src_mult(struct ppc440spe_adma_desc_slot *tx,
drivers/dma/ppc4xx/adma.c
1243
ppc440spe_adma_pqzero_sum_set_dest(struct ppc440spe_adma_desc_slot *tx,
drivers/dma/ppc4xx/adma.c
1894
static dma_cookie_t ppc440spe_adma_tx_submit(struct dma_async_tx_descriptor *tx)
drivers/dma/ppc4xx/adma.c
1897
struct ppc440spe_adma_chan *chan = to_ppc440spe_adma_chan(tx->chan);
drivers/dma/ppc4xx/adma.c
1903
sw_desc = tx_to_ppc440spe_adma_slot(tx);
drivers/dma/ppc4xx/adma.c
1910
cookie = dma_cookie_assign(tx);
drivers/dma/ppc4xx/adma.c
2765
struct dma_async_tx_descriptor *tx;
drivers/dma/ppc4xx/adma.c
2773
tx = ppc440spe_adma_prep_dma_pqzero_sum(chan, pq, &src[1],
drivers/dma/ppc4xx/adma.c
2776
return tx;
drivers/dma/ppc4xx/adma.h
19
#define tx_to_ppc440spe_adma_slot(tx) \
drivers/dma/ppc4xx/adma.h
20
container_of(tx, struct ppc440spe_adma_desc_slot, async_tx)
drivers/dma/pxa_dma.c
1092
vd, vd->tx.cookie, is_desc_completed(vd));
drivers/dma/pxa_dma.c
137
#define tx_to_pxad_desc(tx) \
drivers/dma/pxa_dma.c
138
container_of(tx, struct pxad_desc_sw, async_tx)
drivers/dma/pxa_dma.c
622
__func__, vd, vd->tx.cookie, vd_completed,
drivers/dma/pxa_dma.c
624
last_started = vd->tx.cookie;
drivers/dma/pxa_dma.c
773
static dma_cookie_t pxad_tx_submit(struct dma_async_tx_descriptor *tx)
drivers/dma/pxa_dma.c
775
struct virt_dma_chan *vc = to_virt_chan(tx->chan);
drivers/dma/pxa_dma.c
778
*vd = container_of(tx, struct virt_dma_desc, tx);
drivers/dma/pxa_dma.c
782
set_updater_desc(to_pxad_sw_desc(vd), tx->flags);
drivers/dma/pxa_dma.c
785
cookie = dma_cookie_assign(tx);
drivers/dma/pxa_dma.c
836
"%s(): txd %p[%x]", __func__, vd_first, vd_first->tx.cookie);
drivers/dma/pxa_dma.c
849
struct dma_async_tx_descriptor *tx;
drivers/dma/pxa_dma.c
853
tx = vchan_tx_prep(vc, vd, tx_flags);
drivers/dma/pxa_dma.c
854
tx->tx_submit = pxad_tx_submit;
drivers/dma/pxa_dma.c
857
vc, vd, vd->tx.cookie,
drivers/dma/pxa_dma.c
860
return tx;
drivers/dma/qcom/bam_dma.c
1049
dmaengine_desc_get_callback(&async_desc->vd.tx, &cb);
drivers/dma/qcom/bam_dma.c
953
if (async_desc->vd.tx.cookie != cookie)
drivers/dma/qcom/gpi.c
1063
dma_cookie_complete(&vd->tx);
drivers/dma/qcom/gpi.c
1064
dmaengine_desc_get_callback_invoke(&vd->tx, &result);
drivers/dma/qcom/gpi.c
984
dma_cookie_complete(&vd->tx);
drivers/dma/qcom/gpi.c
985
dmaengine_desc_get_callback_invoke(&vd->tx, &result);
drivers/dma/sa11x0-dma.c
155
p->num, &txd->vd, txd->vd.tx.cookie, txd->ddar);
drivers/dma/sa11x0-dma.c
450
if (p->txd_done && p->txd_done->vd.tx.cookie == cookie)
drivers/dma/sa11x0-dma.c
452
else if (p->txd_load && p->txd_load->vd.tx.cookie == cookie)
drivers/dma/sf-pdma/sf-pdma.c
164
struct dma_async_tx_descriptor *tx = NULL;
drivers/dma/sf-pdma/sf-pdma.c
169
if (vd->tx.cookie == cookie)
drivers/dma/sf-pdma/sf-pdma.c
170
tx = &vd->tx;
drivers/dma/sf-pdma/sf-pdma.c
172
if (!tx)
drivers/dma/sf-pdma/sf-pdma.c
175
if (cookie == tx->chan->completed_cookie)
drivers/dma/sf-pdma/sf-pdma.c
178
if (cookie == tx->cookie) {
drivers/dma/sh/rcar-dmac.c
498
static dma_cookie_t rcar_dmac_tx_submit(struct dma_async_tx_descriptor *tx)
drivers/dma/sh/rcar-dmac.c
500
struct rcar_dmac_chan *chan = to_rcar_dmac_chan(tx->chan);
drivers/dma/sh/rcar-dmac.c
501
struct rcar_dmac_desc *desc = to_rcar_dmac_desc(tx);
drivers/dma/sh/rcar-dmac.c
507
cookie = dma_cookie_assign(tx);
drivers/dma/sh/rcar-dmac.c
510
chan->index, tx->cookie, desc);
drivers/dma/sh/shdma-base.c
108
tx->cookie, &chunk->async_tx, schan->id);
drivers/dma/sh/shdma-base.c
350
struct dma_async_tx_descriptor *tx = &desc->async_tx;
drivers/dma/sh/shdma-base.c
352
BUG_ON(tx->cookie > 0 && tx->cookie != desc->cookie);
drivers/dma/sh/shdma-base.c
366
if (tx->cookie > 0)
drivers/dma/sh/shdma-base.c
367
cookie = tx->cookie;
drivers/dma/sh/shdma-base.c
379
if (desc->mark == DESC_COMPLETED && tx->callback) {
drivers/dma/sh/shdma-base.c
381
dmaengine_desc_get_callback(tx, &cb);
drivers/dma/sh/shdma-base.c
382
callback = tx->callback;
drivers/dma/sh/shdma-base.c
384
tx->cookie, tx, schan->id);
drivers/dma/sh/shdma-base.c
389
if (tx->cookie > 0 || tx->cookie == -EBUSY) {
drivers/dma/sh/shdma-base.c
391
BUG_ON(tx->cookie < 0);
drivers/dma/sh/shdma-base.c
394
head_acked = async_tx_test_ack(tx);
drivers/dma/sh/shdma-base.c
407
tx, tx->cookie);
drivers/dma/sh/shdma-base.c
70
static dma_cookie_t shdma_tx_submit(struct dma_async_tx_descriptor *tx)
drivers/dma/sh/shdma-base.c
73
container_of(tx, struct shdma_desc, async_tx);
drivers/dma/sh/shdma-base.c
74
struct shdma_chan *schan = to_shdma_chan(tx->chan);
drivers/dma/sh/shdma-base.c
75
dma_async_tx_callback callback = tx->callback;
drivers/dma/sh/shdma-base.c
83
cookie = dma_cookie_assign(tx);
drivers/dma/sh/shdma-base.c
888
struct dma_async_tx_descriptor *tx = &sdesc->async_tx;
drivers/dma/sh/shdma-base.c
891
dmaengine_desc_get_callback_invoke(tx, NULL);
drivers/dma/sh/shdma-base.c
99
chunk->async_tx.callback_param = tx->callback_param;
drivers/dma/sh/shdma.h
57
#define tx_to_sh_desc(tx) container_of(tx, struct sh_desc, async_tx)
drivers/dma/sh/usb-dmac.c
568
struct usb_dmac_chan *chan = to_usb_dmac_chan(vd->tx.chan);
drivers/dma/sh/usb-dmac.c
588
desc->done_cookie = desc->vd.tx.cookie;
drivers/dma/sprd-dma.c
687
} else if (schan->cur_desc && schan->cur_desc->vd.tx.cookie == cookie) {
drivers/dma/st_fdma.c
596
if (fchan->fdesc && cookie == fchan->fdesc->vdesc.tx.cookie)
drivers/dma/ste_dma40.c
1486
static dma_cookie_t d40_tx_submit(struct dma_async_tx_descriptor *tx)
drivers/dma/ste_dma40.c
1488
struct d40_chan *d40c = container_of(tx->chan,
drivers/dma/ste_dma40.c
1491
struct d40_desc *d40d = container_of(tx, struct d40_desc, txd);
drivers/dma/ste_dma40.c
1496
cookie = dma_cookie_assign(tx);
drivers/dma/stm32/stm32-dma.c
1443
if (chan->desc && cookie == chan->desc->vdesc.tx.cookie)
drivers/dma/stm32/stm32-dma.c
508
dma_cookie_complete(&chan->desc->vdesc.tx);
drivers/dma/stm32/stm32-dma3.c
1576
else if (chan->swdesc && chan->swdesc->vdesc.tx.cookie == cookie)
drivers/dma/stm32/stm32-dma3.c
463
struct stm32_dma3_chan *chan = to_stm32_dma3_chan(vdesc->tx.chan);
drivers/dma/stm32/stm32-mdma.c
1377
if (chan->desc && cookie == chan->desc->vdesc.tx.cookie)
drivers/dma/stm32/stm32-mdma.c
348
struct stm32_mdma_chan *chan = to_stm32_mdma_chan(vdesc->tx.chan);
drivers/dma/sun6i-dma.c
232
to_sun6i_desc(struct dma_async_tx_descriptor *tx)
drivers/dma/sun6i-dma.c
234
return container_of(tx, struct sun6i_desc, vd.tx);
drivers/dma/sun6i-dma.c
411
struct sun6i_desc *txd = to_sun6i_desc(&vd->tx);
drivers/dma/sun6i-dma.c
412
struct sun6i_dma_dev *sdev = to_sun6i_dma_dev(vd->tx.chan->device);
drivers/dma/sun6i-dma.c
453
pchan->desc = to_sun6i_desc(&desc->tx);
drivers/dma/sun6i-dma.c
997
txd = to_sun6i_desc(&vd->tx);
drivers/dma/tegra186-gpc-dma.c
788
} else if (tdc->dma_desc && tdc->dma_desc->vd.tx.cookie == cookie) {
drivers/dma/tegra20-apb-dma.c
262
static dma_cookie_t tegra_dma_tx_submit(struct dma_async_tx_descriptor *tx);
drivers/dma/tegra210-adma.c
240
return container_of(td, struct tegra_adma_desc, vd.tx);
drivers/dma/tegra210-adma.c
454
desc = to_tegra_adma_desc(&vd->tx);
drivers/dma/tegra210-adma.c
628
desc = to_tegra_adma_desc(&vd->tx);
drivers/dma/tegra210-adma.c
630
} else if (tdc->desc && tdc->desc->vd.tx.cookie == cookie) {
drivers/dma/ti/cppi41.c
362
static dma_cookie_t cppi41_tx_submit(struct dma_async_tx_descriptor *tx)
drivers/dma/ti/cppi41.c
366
cookie = dma_cookie_assign(tx);
drivers/dma/ti/edma.c
1874
if (echan->edesc && echan->edesc->vdesc.tx.cookie == cookie) {
drivers/dma/ti/edma.c
1881
txstate->residue = to_edma_desc(&vdesc->tx)->residue;
drivers/dma/ti/edma.c
1892
echan->edesc->vdesc.tx.cookie == cookie) {
drivers/dma/ti/edma.c
707
static inline struct edma_desc *to_edma_desc(struct dma_async_tx_descriptor *tx)
drivers/dma/ti/edma.c
709
return container_of(tx, struct edma_desc, vdesc.tx);
drivers/dma/ti/edma.c
732
echan->edesc = to_edma_desc(&vdesc->tx);
drivers/dma/ti/k3-psil-am62.c
36
#define PSIL_SAUL(x, ch, flow_base, flow_cnt, default_flow, tx) \
drivers/dma/ti/k3-psil-am62.c
48
.notdpkt = tx, \
drivers/dma/ti/k3-psil-am62a.c
46
#define PSIL_SAUL(x, ch, flow_base, flow_cnt, default_flow, tx) \
drivers/dma/ti/k3-psil-am62a.c
58
.notdpkt = tx, \
drivers/dma/ti/k3-psil-am62p.c
46
#define PSIL_SAUL(x, ch, flow_base, flow_cnt, default_flow, tx) \
drivers/dma/ti/k3-psil-am62p.c
58
.notdpkt = tx, \
drivers/dma/ti/k3-psil-am64.c
47
#define PSIL_SAUL(x, ch, flow_base, flow_cnt, default_flow, tx) \
drivers/dma/ti/k3-psil-am64.c
59
.notdpkt = tx, \
drivers/dma/ti/k3-psil-am654.c
39
#define PSIL_SA2UL(x, tx) \
drivers/dma/ti/k3-psil-am654.c
47
.notdpkt = tx, \
drivers/dma/ti/k3-psil-j7200.c
49
#define PSIL_SA2UL(x, tx) \
drivers/dma/ti/k3-psil-j7200.c
57
.notdpkt = tx, \
drivers/dma/ti/k3-psil-j721e.c
49
#define PSIL_SA2UL(x, tx) \
drivers/dma/ti/k3-psil-j721e.c
57
.notdpkt = tx, \
drivers/dma/ti/k3-psil-j721s2.c
48
#define PSIL_SA2UL(x, tx) \
drivers/dma/ti/k3-psil-j721s2.c
56
.notdpkt = tx, \
drivers/dma/ti/k3-psil-j784s4.c
48
#define PSIL_SA2UL(x, tx) \
drivers/dma/ti/k3-psil-j784s4.c
56
.notdpkt = tx, \
drivers/dma/ti/k3-udma.c
1211
dma_cookie_complete(&d->vd.tx);
drivers/dma/ti/k3-udma.c
329
return container_of(t, struct udma_desc, vd.tx);
drivers/dma/ti/k3-udma.c
3483
d->vd.tx.metadata_ops = &metadata_ops;
drivers/dma/ti/k3-udma.c
3706
d->vd.tx.metadata_ops = &metadata_ops;
drivers/dma/ti/k3-udma.c
3797
d->vd.tx.metadata_ops = &metadata_ops;
drivers/dma/ti/k3-udma.c
3845
if (uc->desc && uc->desc->vd.tx.cookie == cookie) {
drivers/dma/ti/k3-udma.c
4015
d = to_udma_desc(&vd->tx);
drivers/dma/ti/k3-udma.c
4060
dmaengine_desc_get_callback(&vd->tx, &cb);
drivers/dma/ti/k3-udma.c
4072
dmaengine_desc_get_callback(&vd->tx, &cb);
drivers/dma/ti/k3-udma.c
567
struct udma_chan *uc = to_udma_chan(vd->tx.chan);
drivers/dma/ti/k3-udma.c
568
struct udma_desc *d = to_udma_desc(&vd->tx);
drivers/dma/ti/k3-udma.c
582
struct udma_dev *ud = to_udma_dev(vd->tx.chan->device);
drivers/dma/ti/k3-udma.c
583
struct udma_chan *uc = to_udma_chan(vd->tx.chan);
drivers/dma/ti/k3-udma.c
584
struct udma_desc *d = to_udma_desc(&vd->tx);
drivers/dma/ti/k3-udma.c
908
uc->desc = to_udma_desc(&vd->tx);
drivers/dma/ti/omap-dma.c
245
return container_of(t, struct omap_desc, vd.tx);
drivers/dma/ti/omap-dma.c
250
struct omap_desc *d = to_omap_dma_desc(&vd->tx);
drivers/dma/ti/omap-dma.c
253
struct omap_dmadev *od = to_omap_dma_dev(vd->tx.chan->device);
drivers/dma/ti/omap-dma.c
577
c->desc = d = to_omap_dma_desc(&vd->tx);
drivers/dma/ti/omap-dma.c
900
if (c->desc && c->desc->vd.tx.cookie == cookie)
drivers/dma/ti/omap-dma.c
922
to_omap_dma_desc(&vd->tx));
drivers/dma/txx9dmac.c
187
static dma_cookie_t txx9dmac_tx_submit(struct dma_async_tx_descriptor *tx);
drivers/dma/txx9dmac.c
690
static dma_cookie_t txx9dmac_tx_submit(struct dma_async_tx_descriptor *tx)
drivers/dma/txx9dmac.c
692
struct txx9dmac_desc *desc = txd_to_txx9dmac_desc(tx);
drivers/dma/txx9dmac.c
693
struct txx9dmac_chan *dc = to_txx9dmac_chan(tx->chan);
drivers/dma/txx9dmac.c
697
cookie = dma_cookie_assign(tx);
drivers/dma/txx9dmac.c
699
dev_vdbg(chan2dev(tx->chan), "tx_submit: queued %u %p\n",
drivers/dma/uniphier-mdmac.c
304
if (mc->md && mc->md->vd.tx.cookie == cookie) {
drivers/dma/virt-dma.c
104
dmaengine_desc_get_callback(&vd->tx, &cb);
drivers/dma/virt-dma.c
14
static struct virt_dma_desc *to_virt_desc(struct dma_async_tx_descriptor *tx)
drivers/dma/virt-dma.c
16
return container_of(tx, struct virt_dma_desc, tx);
drivers/dma/virt-dma.c
19
dma_cookie_t vchan_tx_submit(struct dma_async_tx_descriptor *tx)
drivers/dma/virt-dma.c
21
struct virt_dma_chan *vc = to_virt_chan(tx->chan);
drivers/dma/virt-dma.c
22
struct virt_dma_desc *vd = to_virt_desc(tx);
drivers/dma/virt-dma.c
27
cookie = dma_cookie_assign(tx);
drivers/dma/virt-dma.c
49
int vchan_tx_desc_free(struct dma_async_tx_descriptor *tx)
drivers/dma/virt-dma.c
51
struct virt_dma_chan *vc = to_virt_chan(tx->chan);
drivers/dma/virt-dma.c
52
struct virt_dma_desc *vd = to_virt_desc(tx);
drivers/dma/virt-dma.c
60
vc, vd, vd->tx.cookie);
drivers/dma/virt-dma.c
72
if (vd->tx.cookie == cookie)
drivers/dma/virt-dma.c
95
dmaengine_desc_get_callback(&vd->tx, &cb);
drivers/dma/virt-dma.h
103
cookie = vd->tx.cookie;
drivers/dma/virt-dma.h
104
dma_cookie_complete(&vd->tx);
drivers/dma/virt-dma.h
118
struct virt_dma_chan *vc = to_virt_chan(vd->tx.chan);
drivers/dma/virt-dma.h
120
if (dmaengine_desc_test_reuse(&vd->tx)) {
drivers/dma/virt-dma.h
137
struct virt_dma_chan *vc = to_virt_chan(vd->tx.chan);
drivers/dma/virt-dma.h
151
struct virt_dma_chan *vc = to_virt_chan(vd->tx.chan);
drivers/dma/virt-dma.h
16
struct dma_async_tx_descriptor tx;
drivers/dma/virt-dma.h
206
dmaengine_desc_clear_reuse(&vd->tx);
drivers/dma/virt-dma.h
61
dma_async_tx_descriptor_init(&vd->tx, &vc->chan);
drivers/dma/virt-dma.h
62
vd->tx.flags = tx_flags;
drivers/dma/virt-dma.h
63
vd->tx.tx_submit = vchan_tx_submit;
drivers/dma/virt-dma.h
64
vd->tx.desc_free = vchan_tx_desc_free;
drivers/dma/virt-dma.h
73
return &vd->tx;
drivers/dma/virt-dma.h
98
struct virt_dma_chan *vc = to_virt_chan(vd->tx.chan);
drivers/dma/xgene-dma.c
189
#define to_dma_desc_sw(tx) \
drivers/dma/xgene-dma.c
190
container_of(tx, struct xgene_dma_desc_sw, tx)
drivers/dma/xgene-dma.c
242
struct dma_async_tx_descriptor tx;
drivers/dma/xgene-dma.c
457
static dma_cookie_t xgene_dma_tx_submit(struct dma_async_tx_descriptor *tx)
drivers/dma/xgene-dma.c
463
if (unlikely(!tx))
drivers/dma/xgene-dma.c
466
chan = to_dma_chan(tx->chan);
drivers/dma/xgene-dma.c
467
desc = to_dma_desc_sw(tx);
drivers/dma/xgene-dma.c
471
cookie = dma_cookie_assign(tx);
drivers/dma/xgene-dma.c
486
dma_pool_free(chan->desc_pool, desc, desc->tx.phys);
drivers/dma/xgene-dma.c
502
desc->tx.phys = phys;
drivers/dma/xgene-dma.c
503
desc->tx.tx_submit = xgene_dma_tx_submit;
drivers/dma/xgene-dma.c
504
dma_async_tx_descriptor_init(&desc->tx, &chan->dma_chan);
drivers/dma/xgene-dma.c
524
if (async_tx_test_ack(&desc->tx))
drivers/dma/xgene-dma.c
540
struct dma_async_tx_descriptor *tx = &desc->tx;
drivers/dma/xgene-dma.c
549
if (tx->cookie == 0)
drivers/dma/xgene-dma.c
552
dma_cookie_complete(tx);
drivers/dma/xgene-dma.c
553
dma_descriptor_unmap(tx);
drivers/dma/xgene-dma.c
556
dmaengine_desc_get_callback_invoke(tx, NULL);
drivers/dma/xgene-dma.c
559
dma_run_dependencies(tx);
drivers/dma/xgene-dma.c
581
if (!async_tx_test_ack(&desc->tx)) {
drivers/dma/xgene-dma.c
591
dma_pool_free(chan->desc_pool, desc, desc->tx.phys);
drivers/dma/xgene-dma.c
865
new->tx.cookie = 0;
drivers/dma/xgene-dma.c
866
async_tx_ack(&new->tx);
drivers/dma/xgene-dma.c
872
new->tx.flags = flags; /* client is in control of this ack */
drivers/dma/xgene-dma.c
873
new->tx.cookie = -EBUSY;
drivers/dma/xgene-dma.c
876
return &new->tx;
drivers/dma/xgene-dma.c
923
new->tx.cookie = 0;
drivers/dma/xgene-dma.c
924
async_tx_ack(&new->tx);
drivers/dma/xgene-dma.c
949
new->tx.flags = flags; /* client is in control of this ack */
drivers/dma/xgene-dma.c
950
new->tx.cookie = -EBUSY;
drivers/dma/xgene-dma.c
953
return &new->tx;
drivers/dma/xilinx/xdma.c
523
dma_cookie_complete(&vd->tx);
drivers/dma/xilinx/xdma.c
926
repeat_tx = vd->tx.flags & DMA_PREP_REPEAT;
drivers/dma/xilinx/xdma.c
929
repeat_tx = repeat_tx && !(next_vd->tx.flags & DMA_PREP_LOAD_EOT);
drivers/dma/xilinx/xilinx_dma.c
2029
static dma_cookie_t xilinx_dma_tx_submit(struct dma_async_tx_descriptor *tx)
drivers/dma/xilinx/xilinx_dma.c
2031
struct xilinx_dma_tx_descriptor *desc = to_dma_tx_descriptor(tx);
drivers/dma/xilinx/xilinx_dma.c
2032
struct xilinx_dma_chan *chan = to_xilinx_chan(tx->chan);
drivers/dma/xilinx/xilinx_dma.c
2054
cookie = dma_cookie_assign(tx);
drivers/dma/xilinx/xilinx_dma.c
530
#define to_dma_tx_descriptor(tx) \
drivers/dma/xilinx/xilinx_dma.c
531
container_of(tx, struct xilinx_dma_tx_descriptor, async_tx)
drivers/dma/xilinx/xilinx_dma.c
644
static void *xilinx_dma_get_metadata_ptr(struct dma_async_tx_descriptor *tx,
drivers/dma/xilinx/xilinx_dma.c
647
struct xilinx_dma_tx_descriptor *desc = to_dma_tx_descriptor(tx);
drivers/dma/xilinx/xilinx_dpdma.c
1311
return &desc->vdesc.tx;
drivers/dma/xilinx/xilinx_dpdma.c
938
sw_desc->hw.desc_id = desc->vdesc.tx.cookie
drivers/dma/xilinx/zynqmp_dma.c
145
#define tx_to_desc(tx) container_of(tx, struct zynqmp_dma_desc_sw, \
drivers/dma/xilinx/zynqmp_dma.c
382
static dma_cookie_t zynqmp_dma_tx_submit(struct dma_async_tx_descriptor *tx)
drivers/dma/xilinx/zynqmp_dma.c
384
struct zynqmp_dma_chan *chan = to_chan(tx->chan);
drivers/dma/xilinx/zynqmp_dma.c
389
new = tx_to_desc(tx);
drivers/dma/xilinx/zynqmp_dma.c
391
cookie = dma_cookie_assign(tx);
drivers/firmware/arm_scmi/base.c
189
num_skip = t->tx.buf;
drivers/firmware/arm_scmi/base.c
275
put_unaligned_le32(id, t->tx.buf);
drivers/firmware/arm_scmi/base.c
301
cfg = t->tx.buf;
drivers/firmware/arm_scmi/clock.c
331
put_unaligned_le32(clk_id, t->tx.buf);
drivers/firmware/arm_scmi/clock.c
361
put_unaligned_le32(clk_id, t->tx.buf);
drivers/firmware/arm_scmi/clock.c
562
put_unaligned_le32(clk_id, t->tx.buf);
drivers/firmware/arm_scmi/clock.c
597
cfg = t->tx.buf;
drivers/firmware/arm_scmi/clock.c
647
cfg = t->tx.buf;
drivers/firmware/arm_scmi/clock.c
684
cfg = t->tx.buf;
drivers/firmware/arm_scmi/clock.c
707
put_unaligned_le32(clk_id, t->tx.buf);
drivers/firmware/arm_scmi/clock.c
743
cfg = t->tx.buf;
drivers/firmware/arm_scmi/clock.c
811
cfg = t->tx.buf;
drivers/firmware/arm_scmi/clock.c
852
put_unaligned_le32(clk_id, t->tx.buf);
drivers/firmware/arm_scmi/clock.c
981
notify = t->tx.buf;
drivers/firmware/arm_scmi/common.h
208
bool tx);
drivers/firmware/arm_scmi/common.h
409
bool tx, struct resource *res,
drivers/firmware/arm_scmi/driver.c
1453
xfer->hdr.status, xfer->tx.buf, xfer->tx.len);
drivers/firmware/arm_scmi/driver.c
1584
xfer->tx.len = tx_size;
drivers/firmware/arm_scmi/driver.c
1696
put_unaligned_le32(res_id, t->tx.buf);
drivers/firmware/arm_scmi/driver.c
1698
put_unaligned_le32(*flags, t->tx.buf + sizeof(res_id));
drivers/firmware/arm_scmi/driver.c
1752
put_unaligned_le32(message_id, t->tx.buf);
drivers/firmware/arm_scmi/driver.c
1809
i->msg = i->t->tx.buf;
drivers/firmware/arm_scmi/driver.c
1938
info = t->tx.buf;
drivers/firmware/arm_scmi/driver.c
2102
put_unaligned_le32(pi->proto->supported_version, t->tx.buf);
drivers/firmware/arm_scmi/driver.c
2658
xfer->tx.buf = xfer->rx.buf;
drivers/firmware/arm_scmi/driver.c
2712
int prot_id, bool tx)
drivers/firmware/arm_scmi/driver.c
2721
idx = tx ? 0 : 1;
drivers/firmware/arm_scmi/driver.c
2722
idr = tx ? &info->tx_idr : &info->rx_idr;
drivers/firmware/arm_scmi/driver.c
2735
cinfo->is_p2a = !tx;
drivers/firmware/arm_scmi/driver.c
2754
ret = info->desc->ops->chan_setup(cinfo, info->dev, tx);
drivers/firmware/arm_scmi/driver.c
2762
if (tx && is_polling_required(cinfo, info->desc)) {
drivers/firmware/arm_scmi/msg.c
35
return sizeof(struct scmi_msg_payld) + xfer->tx.len;
drivers/firmware/arm_scmi/msg.c
59
if (xfer->tx.buf)
drivers/firmware/arm_scmi/msg.c
60
memcpy(msg->msg_payload, xfer->tx.buf, xfer->tx.len);
drivers/firmware/arm_scmi/perf.c
265
put_unaligned_le32(dom_info->id, t->tx.buf);
drivers/firmware/arm_scmi/perf.c
536
limits = t->tx.buf;
drivers/firmware/arm_scmi/perf.c
615
put_unaligned_le32(domain, t->tx.buf);
drivers/firmware/arm_scmi/perf.c
691
lvl = t->tx.buf;
drivers/firmware/arm_scmi/perf.c
755
put_unaligned_le32(domain, t->tx.buf);
drivers/firmware/arm_scmi/perf.c
818
notify = t->tx.buf;
drivers/firmware/arm_scmi/pinctrl.c
196
struct scmi_msg_pinctrl_attributes *tx;
drivers/firmware/arm_scmi/pinctrl.c
207
ret = ph->xops->xfer_get_init(ph, PINCTRL_ATTRIBUTES, sizeof(*tx),
drivers/firmware/arm_scmi/pinctrl.c
212
tx = t->tx.buf;
drivers/firmware/arm_scmi/pinctrl.c
214
tx->identifier = cpu_to_le32(selector);
drivers/firmware/arm_scmi/pinctrl.c
215
tx->flags = cpu_to_le32(type);
drivers/firmware/arm_scmi/pinctrl.c
470
struct scmi_msg_settings_conf *tx;
drivers/firmware/arm_scmi/pinctrl.c
484
configs_in_chunk = (max_msg_size - sizeof(*tx)) / (sizeof(__le32) * 2);
drivers/firmware/arm_scmi/pinctrl.c
490
sizeof(*tx) +
drivers/firmware/arm_scmi/pinctrl.c
495
tx = t->tx.buf;
drivers/firmware/arm_scmi/pinctrl.c
496
tx->identifier = cpu_to_le32(selector);
drivers/firmware/arm_scmi/pinctrl.c
497
tx->function_id = cpu_to_le32(0xFFFFFFFF);
drivers/firmware/arm_scmi/pinctrl.c
500
tx->attributes = cpu_to_le32(attributes);
drivers/firmware/arm_scmi/pinctrl.c
503
tx->configs[i * 2] =
drivers/firmware/arm_scmi/pinctrl.c
505
tx->configs[i * 2 + 1] =
drivers/firmware/arm_scmi/pinctrl.c
529
struct scmi_msg_settings_conf *tx;
drivers/firmware/arm_scmi/pinctrl.c
537
sizeof(*tx), 0, &t);
drivers/firmware/arm_scmi/pinctrl.c
541
tx = t->tx.buf;
drivers/firmware/arm_scmi/pinctrl.c
542
tx->identifier = cpu_to_le32(group);
drivers/firmware/arm_scmi/pinctrl.c
543
tx->function_id = cpu_to_le32(function_id);
drivers/firmware/arm_scmi/pinctrl.c
545
tx->attributes = cpu_to_le32(attributes);
drivers/firmware/arm_scmi/pinctrl.c
560
struct scmi_msg_request *tx;
drivers/firmware/arm_scmi/pinctrl.c
572
ret = ph->xops->xfer_get_init(ph, cmd, sizeof(*tx), 0, &t);
drivers/firmware/arm_scmi/pinctrl.c
576
tx = t->tx.buf;
drivers/firmware/arm_scmi/pinctrl.c
577
tx->identifier = cpu_to_le32(identifier);
drivers/firmware/arm_scmi/pinctrl.c
578
tx->flags = cpu_to_le32(type);
drivers/firmware/arm_scmi/power.c
123
put_unaligned_le32(domain, t->tx.buf);
drivers/firmware/arm_scmi/power.c
164
st = t->tx.buf;
drivers/firmware/arm_scmi/power.c
185
put_unaligned_le32(domain, t->tx.buf);
drivers/firmware/arm_scmi/power.c
231
notify = t->tx.buf;
drivers/firmware/arm_scmi/powercap.c
208
put_unaligned_le32(domain, t->tx.buf);
drivers/firmware/arm_scmi/powercap.c
326
put_unaligned_le32(domain_id, t->tx.buf);
drivers/firmware/arm_scmi/powercap.c
378
msg = t->tx.buf;
drivers/firmware/arm_scmi/powercap.c
477
put_unaligned_le32(domain_id, t->tx.buf);
drivers/firmware/arm_scmi/powercap.c
519
msg = t->tx.buf;
drivers/firmware/arm_scmi/powercap.c
572
put_unaligned_le32(domain_id, t->tx.buf);
drivers/firmware/arm_scmi/powercap.c
759
notify = t->tx.buf;
drivers/firmware/arm_scmi/powercap.c
793
notify = t->tx.buf;
drivers/firmware/arm_scmi/protocols.h
130
struct scmi_msg tx;
drivers/firmware/arm_scmi/raw_mode.c
241
struct scmi_msg tx;
drivers/firmware/arm_scmi/raw_mode.c
558
xfer->tx.len = tx_size;
drivers/firmware/arm_scmi/raw_mode.c
561
memset(xfer->tx.buf, 0x00, raw->desc->max_msg_size);
drivers/firmware/arm_scmi/raw_mode.c
562
if (xfer->tx.len)
drivers/firmware/arm_scmi/raw_mode.c
563
memcpy(xfer->tx.buf, (u8 *)buf + sizeof(msg_hdr), xfer->tx.len);
drivers/firmware/arm_scmi/raw_mode.c
659
xfer->tx.buf, xfer->tx.len);
drivers/firmware/arm_scmi/raw_mode.c
823
if (count > rd->tx.len - rd->tx_size)
drivers/firmware/arm_scmi/raw_mode.c
837
cnt = simple_write_to_buffer(rd->tx.buf, rd->tx.len, ppos,
drivers/firmware/arm_scmi/raw_mode.c
847
ret = scmi_raw_message_send(rd->raw, rd->tx.buf, rd->tx_size,
drivers/firmware/arm_scmi/raw_mode.c
922
rd->tx.len = raw->desc->max_msg_size + sizeof(u32);
drivers/firmware/arm_scmi/raw_mode.c
923
rd->tx.buf = kzalloc(rd->tx.len, GFP_KERNEL);
drivers/firmware/arm_scmi/raw_mode.c
924
if (!rd->tx.buf) {
drivers/firmware/arm_scmi/raw_mode.c
945
kfree(rd->tx.buf);
drivers/firmware/arm_scmi/reset.c
126
put_unaligned_le32(domain, t->tx.buf);
drivers/firmware/arm_scmi/reset.c
208
dom = t->tx.buf;
drivers/firmware/arm_scmi/reset.c
278
cfg = t->tx.buf;
drivers/firmware/arm_scmi/sensors.c
723
cfg = t->tx.buf;
drivers/firmware/arm_scmi/sensors.c
764
trip = t->tx.buf;
drivers/firmware/arm_scmi/sensors.c
791
put_unaligned_le32(sensor_id, t->tx.buf);
drivers/firmware/arm_scmi/sensors.c
820
msg = t->tx.buf;
drivers/firmware/arm_scmi/sensors.c
867
sensor = t->tx.buf;
drivers/firmware/arm_scmi/sensors.c
940
sensor = t->tx.buf;
drivers/firmware/arm_scmi/shmem.c
125
iowrite32(sizeof(shmem->msg_header) + xfer->tx.len, &shmem->length);
drivers/firmware/arm_scmi/shmem.c
127
if (xfer->tx.buf)
drivers/firmware/arm_scmi/shmem.c
128
copy_toio(shmem->msg_payload, xfer->tx.buf, xfer->tx.len);
drivers/firmware/arm_scmi/shmem.c
195
struct device *dev, bool tx,
drivers/firmware/arm_scmi/shmem.c
199
const char *desc = tx ? "Tx" : "Rx";
drivers/firmware/arm_scmi/shmem.c
200
int ret, idx = tx ? 0 : 1;
drivers/firmware/arm_scmi/system.c
64
notify = t->tx.buf;
drivers/firmware/arm_scmi/transports/mailbox.c
183
bool tx)
drivers/firmware/arm_scmi/transports/mailbox.c
185
const char *desc = tx ? "Tx" : "Rx";
drivers/firmware/arm_scmi/transports/mailbox.c
195
if (!tx && !p2a_chan)
drivers/firmware/arm_scmi/transports/mailbox.c
202
smbox->shmem = core->shmem->setup_iomap(cinfo, dev, tx, NULL,
drivers/firmware/arm_scmi/transports/mailbox.c
209
cl->tx_prepare = tx ? tx_prepare : NULL;
drivers/firmware/arm_scmi/transports/mailbox.c
212
cl->knows_txdone = tx;
drivers/firmware/arm_scmi/transports/mailbox.c
214
smbox->chan = mbox_request_channel(cl, tx ? 0 : p2a_chan);
drivers/firmware/arm_scmi/transports/mailbox.c
224
if (tx && a2p_rx_chan) {
drivers/firmware/arm_scmi/transports/mailbox.c
234
if (!tx && p2a_rx_chan) {
drivers/firmware/arm_scmi/transports/optee.c
370
static int scmi_optee_chan_setup(struct scmi_chan_info *cinfo, struct device *dev, bool tx)
drivers/firmware/arm_scmi/transports/optee.c
376
if (!tx)
drivers/firmware/arm_scmi/transports/smc.c
133
bool tx)
drivers/firmware/arm_scmi/transports/smc.c
142
if (!tx)
drivers/firmware/arm_scmi/transports/smc.c
149
scmi_info->shmem = core->shmem->setup_iomap(cinfo, dev, tx, &res,
drivers/firmware/arm_scmi/transports/virtio.c
406
bool tx)
drivers/firmware/arm_scmi/transports/virtio.c
409
int index = tx ? VIRTIO_SCMI_VQ_TX : VIRTIO_SCMI_VQ_RX;
drivers/firmware/arm_scmi/transports/virtio.c
418
if (tx && !vioch->deferred_tx_wq) {
drivers/firmware/arm_scmi/transports/virtio.c
445
if (tx) {
drivers/firmware/arm_scmi/vendors/imx/imx-sm-bbm.c
129
rtc_notify = t->tx.buf;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-bbm.c
141
button_notify = t->tx.buf;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-bbm.c
249
cfg = t->tx.buf;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-bbm.c
278
cfg = t->tx.buf;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-bbm.c
306
cfg = t->tx.buf;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-cpu.c
121
in = t->tx.buf;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-cpu.c
160
put_unaligned_le32(cpuid, t->tx.buf);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-cpu.c
218
put_unaligned_le32(cpuid, t->tx.buf);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-cpu.c
96
put_unaligned_le32(cpuid, t->tx.buf);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-lmm.c
126
put_unaligned_le32(lmid, t->tx.buf);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-lmm.c
146
in = t->tx.buf;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-lmm.c
175
in = t->tx.buf;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-lmm.c
87
put_unaligned_le32(lmid, t->tx.buf);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-misc.c
166
in = t->tx.buf;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-misc.c
257
put_unaligned_le32(ctrl_id, t->tx.buf);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-misc.c
299
in = t->tx.buf;
drivers/firmware/arm_scmi/voltage.c
229
put_unaligned_le32(dom, td->tx.buf);
drivers/firmware/arm_scmi/voltage.c
278
put_unaligned_le32(domain_id, t->tx.buf);
drivers/firmware/arm_scmi/voltage.c
303
cmd = t->tx.buf;
drivers/firmware/arm_scmi/voltage.c
341
cmd = t->tx.buf;
drivers/firmware/samsung/exynos-acpm.c
144
struct acpm_queue tx;
drivers/firmware/samsung/exynos-acpm.c
354
ret = readl_poll_timeout(achan->tx.rear, val, next_tx_front != val, 0,
drivers/firmware/samsung/exynos-acpm.c
440
tx_front = readl(achan->tx.front);
drivers/firmware/samsung/exynos-acpm.c
450
__iowrite32_copy(achan->tx.base + achan->mlen * tx_front,
drivers/firmware/samsung/exynos-acpm.c
454
writel(idx, achan->tx.front);
drivers/firmware/samsung/exynos-acpm.c
477
struct acpm_queue *tx = &achan->tx;
drivers/firmware/samsung/exynos-acpm.c
484
tx->base = base + readl(&chan_shmem->rx_base);
drivers/firmware/samsung/exynos-acpm.c
485
tx->rear = base + readl(&chan_shmem->rx_rear);
drivers/firmware/samsung/exynos-acpm.c
486
tx->front = base + readl(&chan_shmem->rx_front);
drivers/firmware/tegra/bpmp-debugfs.c
116
.tx = {
drivers/firmware/tegra/bpmp-debugfs.c
157
.tx = {
drivers/firmware/tegra/bpmp-debugfs.c
186
.tx = {
drivers/firmware/tegra/bpmp-debugfs.c
251
.tx = {
drivers/firmware/tegra/bpmp-debugfs.c
314
.tx = {
drivers/firmware/tegra/bpmp-debugfs.c
501
.tx = {
drivers/firmware/tegra/bpmp-debugfs.c
538
.tx = {
drivers/firmware/tegra/bpmp-debugfs.c
560
.tx = {
drivers/firmware/tegra/bpmp-tegra186.c
113
struct iosys_map rx, tx;
drivers/firmware/tegra/bpmp-tegra186.c
128
iosys_map_set_vaddr_iomem(&tx, priv->tx.sram + offset);
drivers/firmware/tegra/bpmp-tegra186.c
131
iosys_map_set_vaddr(&tx, priv->tx.dram + offset);
drivers/firmware/tegra/bpmp-tegra186.c
134
err = tegra_ivc_init(channel->ivc, NULL, &rx, priv->rx.phys + offset, &tx,
drivers/firmware/tegra/bpmp-tegra186.c
135
priv->tx.phys + offset, 1, message_size, tegra186_bpmp_ivc_notify,
drivers/firmware/tegra/bpmp-tegra186.c
186
if (priv->tx.pool) {
drivers/firmware/tegra/bpmp-tegra186.c
187
gen_pool_free(priv->tx.pool, (unsigned long)priv->tx.sram, 4096);
drivers/firmware/tegra/bpmp-tegra186.c
215
priv->tx.phys = res.start;
drivers/firmware/tegra/bpmp-tegra186.c
218
priv->tx.dram = devm_memremap(bpmp->dev, priv->tx.phys, size,
drivers/firmware/tegra/bpmp-tegra186.c
220
if (IS_ERR(priv->tx.dram)) {
drivers/firmware/tegra/bpmp-tegra186.c
221
err = PTR_ERR(priv->tx.dram);
drivers/firmware/tegra/bpmp-tegra186.c
226
priv->rx.dram = priv->tx.dram + SZ_4K;
drivers/firmware/tegra/bpmp-tegra186.c
236
priv->tx.pool = of_gen_pool_get(bpmp->dev->of_node, "shmem", 0);
drivers/firmware/tegra/bpmp-tegra186.c
237
if (!priv->tx.pool) {
drivers/firmware/tegra/bpmp-tegra186.c
242
priv->tx.sram = (void __iomem *)gen_pool_dma_alloc(priv->tx.pool, 4096,
drivers/firmware/tegra/bpmp-tegra186.c
243
&priv->tx.phys);
drivers/firmware/tegra/bpmp-tegra186.c
244
if (!priv->tx.sram) {
drivers/firmware/tegra/bpmp-tegra186.c
267
gen_pool_free(priv->tx.pool, (unsigned long)priv->tx.sram, 4096);
drivers/firmware/tegra/bpmp-tegra186.c
28
} tx, rx;
drivers/firmware/tegra/bpmp.c
337
msg->tx.data, msg->tx.size);
drivers/firmware/tegra/bpmp.c
379
channel = tegra_bpmp_write_threaded(bpmp, msg->mrq, msg->tx.data,
drivers/firmware/tegra/bpmp.c
380
msg->tx.size);
drivers/firmware/tegra/bpmp.c
520
.tx = {
drivers/firmware/tegra/bpmp.c
570
msg.tx.data = &request;
drivers/firmware/tegra/bpmp.c
571
msg.tx.size = sizeof(request);
drivers/firmware/tegra/bpmp.c
614
msg.tx.data = &request;
drivers/firmware/tegra/bpmp.c
615
msg.tx.size = sizeof(request);
drivers/firmware/tegra/bpmp.c
89
return (msg->tx.size <= MSG_DATA_MIN_SZ) &&
drivers/firmware/tegra/bpmp.c
91
(msg->tx.size == 0 || msg->tx.data) &&
drivers/firmware/tegra/ivc.c
102
u32 tx = tegra_ivc_header_read_field(map, tx.count);
drivers/firmware/tegra/ivc.c
115
if (tx - rx > ivc->num_frames)
drivers/firmware/tegra/ivc.c
118
return tx == rx;
drivers/firmware/tegra/ivc.c
123
u32 tx = tegra_ivc_header_read_field(map, tx.count);
drivers/firmware/tegra/ivc.c
130
return tx - rx >= ivc->num_frames;
drivers/firmware/tegra/ivc.c
135
u32 tx = tegra_ivc_header_read_field(map, tx.count);
drivers/firmware/tegra/ivc.c
144
return tx - rx;
drivers/firmware/tegra/ivc.c
149
unsigned int count = tegra_ivc_header_read_field(&ivc->tx.map, tx.count);
drivers/firmware/tegra/ivc.c
151
tegra_ivc_header_write_field(&ivc->tx.map, tx.count, count + 1);
drivers/firmware/tegra/ivc.c
153
if (ivc->tx.position == ivc->num_frames - 1)
drivers/firmware/tegra/ivc.c
154
ivc->tx.position = 0;
drivers/firmware/tegra/ivc.c
156
ivc->tx.position++;
drivers/firmware/tegra/ivc.c
173
unsigned int offset = offsetof(struct tegra_ivc_header, tx.count);
drivers/firmware/tegra/ivc.c
184
state = tegra_ivc_header_read_field(&ivc->tx.map, tx.state);
drivers/firmware/tegra/ivc.c
211
state = tegra_ivc_header_read_field(&ivc->tx.map, tx.state);
drivers/firmware/tegra/ivc.c
215
if (!tegra_ivc_full(ivc, &ivc->tx.map))
drivers/firmware/tegra/ivc.c
218
tegra_ivc_invalidate(ivc, ivc->tx.phys + offset);
drivers/firmware/tegra/ivc.c
220
if (tegra_ivc_full(ivc, &ivc->tx.map))
drivers/firmware/tegra/ivc.c
306
unsigned int tx = offsetof(struct tegra_ivc_header, tx.count);
drivers/firmware/tegra/ivc.c
333
tegra_ivc_invalidate(ivc, ivc->rx.phys + tx);
drivers/firmware/tegra/ivc.c
351
return tegra_ivc_frame_virt(ivc, &ivc->tx.map, ivc->tx.position, map);
drivers/firmware/tegra/ivc.c
358
unsigned int tx = offsetof(struct tegra_ivc_header, tx.count);
drivers/firmware/tegra/ivc.c
366
tegra_ivc_flush_frame(ivc, ivc->tx.phys, ivc->tx.position, 0,
drivers/firmware/tegra/ivc.c
376
tegra_ivc_flush(ivc, ivc->tx.phys + tx);
drivers/firmware/tegra/ivc.c
389
tegra_ivc_invalidate(ivc, ivc->tx.phys + rx);
drivers/firmware/tegra/ivc.c
391
if (tegra_ivc_available(ivc, &ivc->tx.map) == 1)
drivers/firmware/tegra/ivc.c
400
unsigned int offset = offsetof(struct tegra_ivc_header, tx.count);
drivers/firmware/tegra/ivc.c
402
tegra_ivc_header_write_field(&ivc->tx.map, tx.state, TEGRA_IVC_STATE_SYNC);
drivers/firmware/tegra/ivc.c
403
tegra_ivc_flush(ivc, ivc->tx.phys + offset);
drivers/firmware/tegra/ivc.c
430
unsigned int offset = offsetof(struct tegra_ivc_header, tx.count);
drivers/firmware/tegra/ivc.c
435
rx_state = tegra_ivc_header_read_field(&ivc->rx.map, tx.state);
drivers/firmware/tegra/ivc.c
436
tx_state = tegra_ivc_header_read_field(&ivc->tx.map, tx.state);
drivers/firmware/tegra/ivc.c
439
offset = offsetof(struct tegra_ivc_header, tx.count);
drivers/firmware/tegra/ivc.c
452
tegra_ivc_header_write_field(&ivc->tx.map, tx.count, 0);
drivers/firmware/tegra/ivc.c
455
ivc->tx.position = 0;
drivers/firmware/tegra/ivc.c
468
tegra_ivc_header_write_field(&ivc->tx.map, tx.state, TEGRA_IVC_STATE_ACK);
drivers/firmware/tegra/ivc.c
469
tegra_ivc_flush(ivc, ivc->tx.phys + offset);
drivers/firmware/tegra/ivc.c
478
offset = offsetof(struct tegra_ivc_header, tx.count);
drivers/firmware/tegra/ivc.c
491
tegra_ivc_header_write_field(&ivc->tx.map, tx.count, 0);
drivers/firmware/tegra/ivc.c
494
ivc->tx.position = 0;
drivers/firmware/tegra/ivc.c
508
tegra_ivc_header_write_field(&ivc->tx.map, tx.state, TEGRA_IVC_STATE_ESTABLISHED);
drivers/firmware/tegra/ivc.c
509
tegra_ivc_flush(ivc, ivc->tx.phys + offset);
drivers/firmware/tegra/ivc.c
517
offset = offsetof(struct tegra_ivc_header, tx.count);
drivers/firmware/tegra/ivc.c
532
tegra_ivc_header_write_field(&ivc->tx.map, tx.state, TEGRA_IVC_STATE_ESTABLISHED);
drivers/firmware/tegra/ivc.c
533
tegra_ivc_flush(ivc, ivc->tx.phys + offset);
drivers/firmware/tegra/ivc.c
574
static int tegra_ivc_check_params(unsigned long rx, unsigned long tx,
drivers/firmware/tegra/ivc.c
577
BUILD_BUG_ON(!IS_ALIGNED(offsetof(struct tegra_ivc_header, tx.count),
drivers/firmware/tegra/ivc.c
603
if (!IS_ALIGNED(tx, TEGRA_IVC_ALIGN)) {
drivers/firmware/tegra/ivc.c
604
pr_err("IVC channel start not aligned: %#lx\n", tx);
drivers/firmware/tegra/ivc.c
608
if (rx < tx) {
drivers/firmware/tegra/ivc.c
609
if (rx + frame_size * num_frames > tx) {
drivers/firmware/tegra/ivc.c
611
rx, frame_size * num_frames, tx);
drivers/firmware/tegra/ivc.c
615
if (tx + frame_size * num_frames > rx) {
drivers/firmware/tegra/ivc.c
617
tx, frame_size * num_frames, rx);
drivers/firmware/tegra/ivc.c
62
} tx;
drivers/firmware/tegra/ivc.c
647
dma_addr_t rx_phys, const struct iosys_map *tx, dma_addr_t tx_phys,
drivers/firmware/tegra/ivc.c
665
err = tegra_ivc_check_params(iosys_map_get_address(rx), iosys_map_get_address(tx),
drivers/firmware/tegra/ivc.c
678
ivc->tx.phys = dma_map_single(peer, iosys_map_get_vaddr(tx), queue_size,
drivers/firmware/tegra/ivc.c
680
if (dma_mapping_error(peer, ivc->tx.phys)) {
drivers/firmware/tegra/ivc.c
687
ivc->tx.phys = tx_phys;
drivers/firmware/tegra/ivc.c
691
iosys_map_copy(&ivc->tx.map, tx);
drivers/firmware/tegra/ivc.c
702
ivc->tx.position = 0;
drivers/firmware/tegra/ivc.c
717
dma_unmap_single(ivc->peer, ivc->tx.phys, size,
drivers/fpga/machxo2-spi.c
139
struct spi_transfer tx[2];
drivers/fpga/machxo2-spi.c
144
memset(tx, 0, sizeof(tx));
drivers/fpga/machxo2-spi.c
146
tx[0].tx_buf = &erase;
drivers/fpga/machxo2-spi.c
147
tx[0].len = sizeof(erase);
drivers/fpga/machxo2-spi.c
148
spi_message_add_tail(&tx[0], &msg);
drivers/fpga/machxo2-spi.c
158
tx[1].tx_buf = &refresh;
drivers/fpga/machxo2-spi.c
159
tx[1].len = sizeof(refresh);
drivers/fpga/machxo2-spi.c
160
tx[1].delay.value = MACHXO2_REFRESH_USEC;
drivers/fpga/machxo2-spi.c
161
tx[1].delay.unit = SPI_DELAY_UNIT_USECS;
drivers/fpga/machxo2-spi.c
162
spi_message_add_tail(&tx[1], &msg);
drivers/fpga/machxo2-spi.c
193
struct spi_transfer tx[3];
drivers/fpga/machxo2-spi.c
208
memset(tx, 0, sizeof(tx));
drivers/fpga/machxo2-spi.c
210
tx[0].tx_buf = &enable;
drivers/fpga/machxo2-spi.c
211
tx[0].len = sizeof(enable);
drivers/fpga/machxo2-spi.c
212
tx[0].delay.value = MACHXO2_LOW_DELAY_USEC;
drivers/fpga/machxo2-spi.c
213
tx[0].delay.unit = SPI_DELAY_UNIT_USECS;
drivers/fpga/machxo2-spi.c
214
spi_message_add_tail(&tx[0], &msg);
drivers/fpga/machxo2-spi.c
216
tx[1].tx_buf = &erase;
drivers/fpga/machxo2-spi.c
217
tx[1].len = sizeof(erase);
drivers/fpga/machxo2-spi.c
218
spi_message_add_tail(&tx[1], &msg);
drivers/fpga/machxo2-spi.c
235
tx[2].tx_buf = &initaddr;
drivers/fpga/machxo2-spi.c
236
tx[2].len = sizeof(initaddr);
drivers/fpga/machxo2-spi.c
237
spi_message_add_tail(&tx[2], &msg);
drivers/fpga/machxo2-spi.c
257
struct spi_transfer tx;
drivers/fpga/machxo2-spi.c
272
memset(&tx, 0, sizeof(tx));
drivers/fpga/machxo2-spi.c
274
tx.tx_buf = payload;
drivers/fpga/machxo2-spi.c
275
tx.len = MACHXO2_BUF_SIZE;
drivers/fpga/machxo2-spi.c
276
tx.delay.value = MACHXO2_HIGH_DELAY_USEC;
drivers/fpga/machxo2-spi.c
277
tx.delay.unit = SPI_DELAY_UNIT_USECS;
drivers/fpga/machxo2-spi.c
278
spi_message_add_tail(&tx, &msg);
drivers/fpga/machxo2-spi.c
296
struct spi_transfer tx[2];
drivers/fpga/machxo2-spi.c
302
memset(tx, 0, sizeof(tx));
drivers/fpga/machxo2-spi.c
304
tx[0].tx_buf = &progdone;
drivers/fpga/machxo2-spi.c
305
tx[0].len = sizeof(progdone);
drivers/fpga/machxo2-spi.c
306
spi_message_add_tail(&tx[0], &msg);
drivers/fpga/machxo2-spi.c
324
tx[1].tx_buf = &refresh;
drivers/fpga/machxo2-spi.c
325
tx[1].len = sizeof(refresh);
drivers/fpga/machxo2-spi.c
326
tx[1].delay.value = MACHXO2_REFRESH_USEC;
drivers/fpga/machxo2-spi.c
327
tx[1].delay.unit = SPI_DELAY_UNIT_USECS;
drivers/fpga/machxo2-spi.c
328
spi_message_add_tail(&tx[1], &msg);
drivers/fpga/machxo2-spi.c
69
struct spi_transfer rx, tx;
drivers/fpga/machxo2-spi.c
74
memset(&tx, 0, sizeof(tx));
drivers/fpga/machxo2-spi.c
75
tx.tx_buf = cmd;
drivers/fpga/machxo2-spi.c
76
tx.len = sizeof(cmd);
drivers/fpga/machxo2-spi.c
80
spi_message_add_tail(&tx, &msg);
drivers/fpga/microchip-spi.c
287
.tx_buf = &priv->tx,
drivers/fpga/microchip-spi.c
300
priv->tx = MPF_SPI_FRAME;
drivers/fpga/microchip-spi.c
46
u8 tx __aligned(ARCH_KMALLOC_MINALIGN);
drivers/fpga/microchip-spi.c
59
.tx_buf = &priv->tx,
drivers/fpga/microchip-spi.c
64
.tx_buf = &priv->tx,
drivers/fpga/microchip-spi.c
72
priv->tx = MPF_SPI_READ_STATUS;
drivers/gpio/gpio-mpsse.c
110
if (desc->tx)
drivers/gpio/gpio-mpsse.c
130
desc.tx = true;
drivers/gpio/gpio-mpsse.c
147
desc.tx = false;
drivers/gpio/gpio-mpsse.c
52
bool tx; /* direction of bulk transfer */
drivers/gpu/drm/amd/pm/powerplay/hwmgr/smu8_hwmgr.h
78
#define DDI_POWERGATING_ARG(phyID, lanemask, rx, tx, core) \
drivers/gpu/drm/amd/pm/powerplay/hwmgr/smu8_hwmgr.h
82
((tx) ? DISPLAYPHY_TX_SELECT : 0) | \
drivers/gpu/drm/display/drm_dp_mst_topology.c
542
struct drm_dp_remote_i2c_read_tx *tx;
drivers/gpu/drm/display/drm_dp_mst_topology.c
548
tx = &r->transactions[i];
drivers/gpu/drm/display/drm_dp_mst_topology.c
550
tx->i2c_dev_id = buf[++idx] & 0x7f;
drivers/gpu/drm/display/drm_dp_mst_topology.c
551
tx->num_bytes = buf[++idx];
drivers/gpu/drm/display/drm_dp_mst_topology.c
552
tx->bytes = kmemdup(&buf[++idx],
drivers/gpu/drm/display/drm_dp_mst_topology.c
553
tx->num_bytes,
drivers/gpu/drm/display/drm_dp_mst_topology.c
555
if (!tx->bytes) {
drivers/gpu/drm/display/drm_dp_mst_topology.c
559
idx += tx->num_bytes;
drivers/gpu/drm/display/drm_dp_mst_topology.c
560
tx->no_stop_bit = (buf[idx] >> 5) & 0x1;
drivers/gpu/drm/display/drm_dp_mst_topology.c
561
tx->i2c_transaction_delay = buf[idx] & 0xf;
drivers/gpu/drm/display/drm_dp_mst_topology.c
566
tx = &r->transactions[i];
drivers/gpu/drm/display/drm_dp_mst_topology.c
567
kfree(tx->bytes);
drivers/gpu/drm/drm_mipi_dsi.c
1049
u8 *tx;
drivers/gpu/drm/drm_mipi_dsi.c
1053
tx = kmalloc(size, GFP_KERNEL);
drivers/gpu/drm/drm_mipi_dsi.c
1054
if (!tx)
drivers/gpu/drm/drm_mipi_dsi.c
1057
tx = stack_tx;
drivers/gpu/drm/drm_mipi_dsi.c
1061
tx[0] = cmd;
drivers/gpu/drm/drm_mipi_dsi.c
1063
memcpy(&tx[1], data, len);
drivers/gpu/drm/drm_mipi_dsi.c
1065
err = mipi_dsi_dcs_write_buffer(dsi, tx, size);
drivers/gpu/drm/drm_mipi_dsi.c
1067
if (tx != stack_tx)
drivers/gpu/drm/drm_mipi_dsi.c
1068
kfree(tx);
drivers/gpu/drm/drm_mipi_dsi.c
574
const u8 *tx = msg->tx_buf;
drivers/gpu/drm/drm_mipi_dsi.c
576
packet->header[1] = (msg->tx_len > 0) ? tx[0] : 0;
drivers/gpu/drm/drm_mipi_dsi.c
577
packet->header[2] = (msg->tx_len > 1) ? tx[1] : 0;
drivers/gpu/drm/drm_mipi_dsi.c
640
u8 tx[2] = { value & 0xff, value >> 8 };
drivers/gpu/drm/drm_mipi_dsi.c
644
.tx_len = sizeof(tx),
drivers/gpu/drm/drm_mipi_dsi.c
645
.tx_buf = tx,
drivers/gpu/drm/drm_mipi_dsi.c
669
u8 tx[2] = { };
drivers/gpu/drm/drm_mipi_dsi.c
673
.tx_len = sizeof(tx),
drivers/gpu/drm/drm_mipi_dsi.c
674
.tx_buf = tx,
drivers/gpu/drm/drm_mipi_dsi.c
681
tx[0] = (enable << 0) |
drivers/gpu/drm/i915/display/intel_cx0_phy.c
1021
.tx = { 0xbe88,
drivers/gpu/drm/i915/display/intel_cx0_phy.c
1046
.tx = { 0xbe88,
drivers/gpu/drm/i915/display/intel_cx0_phy.c
1071
.tx = { 0xbe88,
drivers/gpu/drm/i915/display/intel_cx0_phy.c
1096
.tx = { 0xbe88,
drivers/gpu/drm/i915/display/intel_cx0_phy.c
1134
.tx = { 0xbea0, /* tx cfg0 */
drivers/gpu/drm/i915/display/intel_cx0_phy.c
1189
.tx = 0x10,
drivers/gpu/drm/i915/display/intel_cx0_phy.c
1215
.tx = 0x10,
drivers/gpu/drm/i915/display/intel_cx0_phy.c
1241
.tx = 0x10,
drivers/gpu/drm/i915/display/intel_cx0_phy.c
1267
.tx = 0x10,
drivers/gpu/drm/i915/display/intel_cx0_phy.c
1293
.tx = 0x10,
drivers/gpu/drm/i915/display/intel_cx0_phy.c
1320
.tx = 0x10,
drivers/gpu/drm/i915/display/intel_cx0_phy.c
1330
.tx = 0x10,
drivers/gpu/drm/i915/display/intel_cx0_phy.c
1340
.tx = 0x10,
drivers/gpu/drm/i915/display/intel_cx0_phy.c
1350
.tx = 0x10,
drivers/gpu/drm/i915/display/intel_cx0_phy.c
1360
.tx = 0x10,
drivers/gpu/drm/i915/display/intel_cx0_phy.c
1370
.tx = 0x10,
drivers/gpu/drm/i915/display/intel_cx0_phy.c
1380
.tx = 0x10,
drivers/gpu/drm/i915/display/intel_cx0_phy.c
1390
.tx = 0x10,
drivers/gpu/drm/i915/display/intel_cx0_phy.c
1400
.tx = 0x10,
drivers/gpu/drm/i915/display/intel_cx0_phy.c
1410
.tx = 0x10,
drivers/gpu/drm/i915/display/intel_cx0_phy.c
1420
.tx = 0x10,
drivers/gpu/drm/i915/display/intel_cx0_phy.c
1430
.tx = 0x10,
drivers/gpu/drm/i915/display/intel_cx0_phy.c
1440
.tx = 0x10,
drivers/gpu/drm/i915/display/intel_cx0_phy.c
1450
.tx = 0x10,
drivers/gpu/drm/i915/display/intel_cx0_phy.c
1460
.tx = 0x10,
drivers/gpu/drm/i915/display/intel_cx0_phy.c
1470
.tx = 0x10,
drivers/gpu/drm/i915/display/intel_cx0_phy.c
1480
.tx = 0x10,
drivers/gpu/drm/i915/display/intel_cx0_phy.c
1490
.tx = 0x10,
drivers/gpu/drm/i915/display/intel_cx0_phy.c
1500
.tx = 0x10,
drivers/gpu/drm/i915/display/intel_cx0_phy.c
1510
.tx = 0x10,
drivers/gpu/drm/i915/display/intel_cx0_phy.c
1520
.tx = 0x10,
drivers/gpu/drm/i915/display/intel_cx0_phy.c
1530
.tx = 0x10,
drivers/gpu/drm/i915/display/intel_cx0_phy.c
1540
.tx = 0x10,
drivers/gpu/drm/i915/display/intel_cx0_phy.c
1550
.tx = 0x10,
drivers/gpu/drm/i915/display/intel_cx0_phy.c
1560
.tx = 0x10,
drivers/gpu/drm/i915/display/intel_cx0_phy.c
1570
.tx = 0x10,
drivers/gpu/drm/i915/display/intel_cx0_phy.c
1580
.tx = 0x10,
drivers/gpu/drm/i915/display/intel_cx0_phy.c
1590
.tx = 0x10,
drivers/gpu/drm/i915/display/intel_cx0_phy.c
1600
.tx = 0x10,
drivers/gpu/drm/i915/display/intel_cx0_phy.c
1610
.tx = 0x10,
drivers/gpu/drm/i915/display/intel_cx0_phy.c
1620
.tx = 0x10,
drivers/gpu/drm/i915/display/intel_cx0_phy.c
1630
.tx = 0x10,
drivers/gpu/drm/i915/display/intel_cx0_phy.c
1640
.tx = 0x10,
drivers/gpu/drm/i915/display/intel_cx0_phy.c
1650
.tx = 0x10,
drivers/gpu/drm/i915/display/intel_cx0_phy.c
1660
.tx = 0x10,
drivers/gpu/drm/i915/display/intel_cx0_phy.c
1670
.tx = 0x10,
drivers/gpu/drm/i915/display/intel_cx0_phy.c
1680
.tx = 0x10,
drivers/gpu/drm/i915/display/intel_cx0_phy.c
1690
.tx = 0x10,
drivers/gpu/drm/i915/display/intel_cx0_phy.c
1700
.tx = 0x10,
drivers/gpu/drm/i915/display/intel_cx0_phy.c
1710
.tx = 0x10,
drivers/gpu/drm/i915/display/intel_cx0_phy.c
1769
.tx = { 0xbe88, /* tx cfg0 */
drivers/gpu/drm/i915/display/intel_cx0_phy.c
1794
.tx = { 0xbe88, /* tx cfg0 */
drivers/gpu/drm/i915/display/intel_cx0_phy.c
1819
.tx = { 0xbe88, /* tx cfg0 */
drivers/gpu/drm/i915/display/intel_cx0_phy.c
1844
.tx = { 0xbe88, /* tx cfg0 */
drivers/gpu/drm/i915/display/intel_cx0_phy.c
1869
.tx = { 0xbe88, /* tx cfg0 */
drivers/gpu/drm/i915/display/intel_cx0_phy.c
1894
.tx = { 0xbe98, /* tx cfg0 */
drivers/gpu/drm/i915/display/intel_cx0_phy.c
1919
.tx = { 0xbe98, /* tx cfg0 */
drivers/gpu/drm/i915/display/intel_cx0_phy.c
1944
.tx = { 0xbe98, /* tx cfg0 */
drivers/gpu/drm/i915/display/intel_cx0_phy.c
1969
.tx = { 0xbe98, /* tx cfg0 */
drivers/gpu/drm/i915/display/intel_cx0_phy.c
1994
.tx = { 0xbe98, /* tx cfg0 */
drivers/gpu/drm/i915/display/intel_cx0_phy.c
2229
int tx = tx_lane % 2 + 1;
drivers/gpu/drm/i915/display/intel_cx0_phy.c
2232
val = intel_cx0_read(encoder, phy_lane_mask, PHY_CX0_TX_CONTROL(tx, 2));
drivers/gpu/drm/i915/display/intel_cx0_phy.c
2274
pll_state->tx = intel_cx0_read(encoder, lane, PHY_C10_VDR_TX(0));
drivers/gpu/drm/i915/display/intel_cx0_phy.c
2305
intel_cx0_write(encoder, INTEL_CX0_LANE0, PHY_C10_VDR_TX(0), pll_state->tx, MB_WRITE_COMMITTED);
drivers/gpu/drm/i915/display/intel_cx0_phy.c
2342
drm_printf(p, "tx: 0x%x, cmn: 0x%x\n", hw_state->tx, hw_state->cmn);
drivers/gpu/drm/i915/display/intel_cx0_phy.c
2431
pll_state->tx[0] = 0xbe88;
drivers/gpu/drm/i915/display/intel_cx0_phy.c
2432
pll_state->tx[1] = intel_c20_hdmi_tmds_tx_cgf_1(crtc_state);
drivers/gpu/drm/i915/display/intel_cx0_phy.c
2433
pll_state->tx[2] = 0x0000;
drivers/gpu/drm/i915/display/intel_cx0_phy.c
2710
return state->tx[0] & C20_PHY_USE_MPLLB;
drivers/gpu/drm/i915/display/intel_cx0_phy.c
2723
unsigned int tx_rate = REG_FIELD_GET(C20_PHY_TX_RATE, pll_state->tx[0]);
drivers/gpu/drm/i915/display/intel_cx0_phy.c
2777
for (i = 0; i < ARRAY_SIZE(pll_state->tx); i++) {
drivers/gpu/drm/i915/display/intel_cx0_phy.c
2779
pll_state->tx[i] = intel_c20_sram_read(encoder,
drivers/gpu/drm/i915/display/intel_cx0_phy.c
2783
pll_state->tx[i] = intel_c20_sram_read(encoder,
drivers/gpu/drm/i915/display/intel_cx0_phy.c
2841
hw_state->tx[0], hw_state->tx[1], hw_state->tx[2]);
drivers/gpu/drm/i915/display/intel_cx0_phy.c
2911
for (i = 0; i < ARRAY_SIZE(pll_state->tx); i++) {
drivers/gpu/drm/i915/display/intel_cx0_phy.c
2915
pll_state->tx[i]);
drivers/gpu/drm/i915/display/intel_cx0_phy.c
2919
pll_state->tx[i]);
drivers/gpu/drm/i915/display/intel_cx0_phy.c
3186
int tx = i % 2 + 1;
drivers/gpu/drm/i915/display/intel_cx0_phy.c
3192
intel_cx0_rmw(encoder, lane_mask, PHY_CX0_TX_CONTROL(tx, 2),
drivers/gpu/drm/i915/display/intel_cx0_phy.c
3501
int tx = i % 2 + 1;
drivers/gpu/drm/i915/display/intel_cx0_phy.c
3507
intel_cx0_rmw(encoder, lane_mask, PHY_CMN1_CONTROL(tx, 0),
drivers/gpu/drm/i915/display/intel_cx0_phy.c
3680
if (a->tx != b->tx)
drivers/gpu/drm/i915/display/intel_cx0_phy.c
3695
if (memcmp(&a->tx, &b->tx, sizeof(a->tx)) != 0)
drivers/gpu/drm/i915/display/intel_cx0_phy.c
3701
if (a->tx[0] & C20_PHY_USE_MPLLB) {
drivers/gpu/drm/i915/display/intel_cx0_phy.c
512
int tx = ln % 2;
drivers/gpu/drm/i915/display/intel_cx0_phy.c
518
intel_cx0_rmw(encoder, lane_mask, PHY_CX0_VDROVRD_CTL(lane, tx, 0),
drivers/gpu/drm/i915/display/intel_cx0_phy.c
522
intel_cx0_rmw(encoder, lane_mask, PHY_CX0_VDROVRD_CTL(lane, tx, 1),
drivers/gpu/drm/i915/display/intel_cx0_phy.c
526
intel_cx0_rmw(encoder, lane_mask, PHY_CX0_VDROVRD_CTL(lane, tx, 2),
drivers/gpu/drm/i915/display/intel_cx0_phy.c
551
.tx = 0x10,
drivers/gpu/drm/i915/display/intel_cx0_phy.c
577
.tx = 0x10,
drivers/gpu/drm/i915/display/intel_cx0_phy.c
603
.tx = 0x10,
drivers/gpu/drm/i915/display/intel_cx0_phy.c
629
.tx = 0x10,
drivers/gpu/drm/i915/display/intel_cx0_phy.c
655
.tx = 0x10,
drivers/gpu/drm/i915/display/intel_cx0_phy.c
681
.tx = 0x10,
drivers/gpu/drm/i915/display/intel_cx0_phy.c
707
.tx = 0x10,
drivers/gpu/drm/i915/display/intel_cx0_phy.c
733
.tx = 0x10,
drivers/gpu/drm/i915/display/intel_cx0_phy.c
759
.tx = 0x10,
drivers/gpu/drm/i915/display/intel_cx0_phy.c
807
.tx = { 0xbe88, /* tx cfg0 */
drivers/gpu/drm/i915/display/intel_cx0_phy.c
832
.tx = { 0xbe88, /* tx cfg0 */
drivers/gpu/drm/i915/display/intel_cx0_phy.c
857
.tx = { 0xbe88, /* tx cfg0 */
drivers/gpu/drm/i915/display/intel_cx0_phy.c
882
.tx = { 0xbe88, /* tx cfg0 */
drivers/gpu/drm/i915/display/intel_cx0_phy.c
908
.tx = { 0xbe21, /* tx cfg0 */
drivers/gpu/drm/i915/display/intel_cx0_phy.c
932
.tx = { 0xbea0, /* tx cfg0 */
drivers/gpu/drm/i915/display/intel_cx0_phy.c
957
.tx = { 0xbe20, /* tx cfg0 */
drivers/gpu/drm/i915/display/intel_cx0_phy.c
996
.tx = { 0xbe88,
drivers/gpu/drm/i915/display/intel_cx0_phy_regs.h
284
#define PHY_CX0_VDROVRD_CTL(lane, tx, control) \
drivers/gpu/drm/i915/display/intel_cx0_phy_regs.h
286
((lane) ^ (tx)) * 0x10 + (control))
drivers/gpu/drm/i915/display/intel_cx0_phy_regs.h
289
#define PHY_CX0_TX_CONTROL(tx, control) (0x400 + ((tx) - 1) * 0x200 + (control))
drivers/gpu/drm/i915/display/intel_cx0_phy_regs.h
292
#define PHY_CMN1_CONTROL(tx, control) (0x800 + ((tx) - 1) * 0x200 + (control))
drivers/gpu/drm/i915/display/intel_dpll_mgr.h
245
u8 tx;
drivers/gpu/drm/i915/display/intel_dpll_mgr.h
252
u16 tx[3];
drivers/gpu/drm/i915/display/intel_lt_phy.c
2165
int tx = ln % 2;
drivers/gpu/drm/i915/display/intel_lt_phy.c
2171
intel_lt_phy_rmw(encoder, lane_mask, LT_PHY_TXY_CTL8(tx),
drivers/gpu/drm/i915/display/intel_lt_phy.c
2177
intel_lt_phy_rmw(encoder, lane_mask, LT_PHY_TXY_CTL2(tx),
drivers/gpu/drm/i915/display/intel_lt_phy.c
2181
intel_lt_phy_rmw(encoder, lane_mask, LT_PHY_TXY_CTL3(tx),
drivers/gpu/drm/i915/display/intel_lt_phy.c
2185
intel_lt_phy_rmw(encoder, lane_mask, LT_PHY_TXY_CTL4(tx),
drivers/gpu/drm/i915/display/intel_snps_hdmi_pll.c
337
pll_state->tx = 0x10;
drivers/gpu/drm/mcde/mcde_dsi.c
298
const u8 *tx = msg->tx_buf;
drivers/gpu/drm/mcde/mcde_dsi.c
345
val |= tx[i] << (i * 8);
drivers/gpu/drm/mcde/mcde_dsi.c
351
val |= tx[i + 4] << (i * 8);
drivers/gpu/drm/mcde/mcde_dsi.c
357
val |= tx[i + 8] << (i * 8);
drivers/gpu/drm/mcde/mcde_dsi.c
363
val |= tx[i + 12] << (i * 8);
drivers/gpu/drm/msm/dsi/dsi_host.c
2208
u8 tx[2] = {pkt_size & 0xff, pkt_size >> 8};
drivers/gpu/drm/msm/dsi/dsi_host.c
2213
.tx_buf = tx,
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
1137
msgqTxHeader tx;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
1164
cmdq->tx.version = 0;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
1165
cmdq->tx.size = gsp->shm.cmdq.size;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
1166
cmdq->tx.entryOff = GSP_PAGE_SIZE;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
1167
cmdq->tx.msgSize = GSP_PAGE_SIZE;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
1168
cmdq->tx.msgCount = (cmdq->tx.size - cmdq->tx.entryOff) / cmdq->tx.msgSize;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
1169
cmdq->tx.writePtr = 0;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
1170
cmdq->tx.flags = 1;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
1171
cmdq->tx.rxHdrOff = offsetof(typeof(*cmdq), rx.readPtr);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
1175
gsp->cmdq.cnt = cmdq->tx.msgCount;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
1176
gsp->cmdq.wptr = &cmdq->tx.writePtr;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
1178
gsp->msgq.cnt = cmdq->tx.msgCount;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
1179
gsp->msgq.wptr = &msgq->tx.writePtr;
drivers/gpu/drm/nouveau/nvkm/subdev/i2c/base.c
117
u32 hi, lo, rq, tx;
drivers/gpu/drm/nouveau/nvkm/subdev/i2c/base.c
122
i2c->func->aux_stat(i2c, &hi, &lo, &rq, &tx);
drivers/gpu/drm/nouveau/nvkm/subdev/i2c/base.c
123
if (!hi && !lo && !rq && !tx)
drivers/gpu/drm/nouveau/nvkm/subdev/i2c/base.c
131
if (tx & aux->intr) mask |= NVKM_I2C_DONE;
drivers/gpu/drm/nouveau/nvkm/subdev/i2c/g94.c
28
g94_aux_stat(struct nvkm_i2c *i2c, u32 *hi, u32 *lo, u32 *rq, u32 *tx)
drivers/gpu/drm/nouveau/nvkm/subdev/i2c/g94.c
33
for (i = 0, *hi = *lo = *rq = *tx = 0; i < 8; i++) {
drivers/gpu/drm/nouveau/nvkm/subdev/i2c/g94.c
37
if ((stat & (8 << (i * 4)))) *tx |= 1 << i;
drivers/gpu/drm/nouveau/nvkm/subdev/i2c/gk104.c
28
gk104_aux_stat(struct nvkm_i2c *i2c, u32 *hi, u32 *lo, u32 *rq, u32 *tx)
drivers/gpu/drm/nouveau/nvkm/subdev/i2c/gk104.c
33
for (i = 0, *hi = *lo = *rq = *tx = 0; i < 8; i++) {
drivers/gpu/drm/nouveau/nvkm/subdev/i2c/gk104.c
37
if ((stat & (8 << (i * 4)))) *tx |= 1 << i;
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
80
struct dma_async_tx_descriptor *tx;
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
84
tx = dmaengine_prep_dma_memcpy(dmm->wa_dma_chan, dst, src, 4, 0);
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
85
if (!tx) {
drivers/gpu/drm/omapdrm/omap_dmm_tiler.c
90
cookie = tx->tx_submit(tx);
drivers/gpu/drm/rockchip/rockchip_drm_vop2.c
439
u8 tx, ty;
drivers/gpu/drm/rockchip/rockchip_drm_vop2.c
458
tx = 16 - ((act_xoffset + width) & 0xf);
drivers/gpu/drm/rockchip/rockchip_drm_vop2.c
462
tx = bottom_crop_line_num - act_yoffset;
drivers/gpu/drm/rockchip/rockchip_drm_vop2.c
466
tx = act_yoffset;
drivers/gpu/drm/rockchip/rockchip_drm_vop2.c
470
tx = 16 - ((act_xoffset + width) & 0xf);
drivers/gpu/drm/rockchip/rockchip_drm_vop2.c
474
tx = act_xoffset;
drivers/gpu/drm/rockchip/rockchip_drm_vop2.c
478
tx = bottom_crop_line_num - act_yoffset;
drivers/gpu/drm/rockchip/rockchip_drm_vop2.c
482
tx = act_yoffset;
drivers/gpu/drm/rockchip/rockchip_drm_vop2.c
486
tx = act_xoffset;
drivers/gpu/drm/rockchip/rockchip_drm_vop2.c
496
return FIELD_PREP(TRANSFORM_XOFFSET, tx) |
drivers/gpu/drm/tiny/repaper.c
104
const void *tx, void *rx, size_t len)
drivers/gpu/drm/tiny/repaper.c
120
if (tx && len <= 32) {
drivers/gpu/drm/tiny/repaper.c
121
txbuf = kmemdup(tx, len, GFP_KERNEL);
drivers/gpu/drm/tiny/repaper.c
136
tr[1].tx_buf = txbuf ? txbuf : tx;
drivers/gpu/drm/vc4/vc4_dsi.c
617
struct dma_async_tx_descriptor *tx;
drivers/gpu/drm/vc4/vc4_dsi.c
631
tx = chan->device->device_prep_dma_memcpy(chan,
drivers/gpu/drm/vc4/vc4_dsi.c
635
if (!tx) {
drivers/gpu/drm/vc4/vc4_dsi.c
640
cookie = tx->tx_submit(tx);
drivers/hid/hid-led.c
291
union delcom_packet dp = { .tx.major_cmd = 101, .tx.minor_cmd = 12 };
drivers/hid/hid-led.c
293
dp.tx.data_lsb = 1 << delcom_get_lednum(led);
drivers/hid/hid-led.c
294
dp.tx.data_msb = 0;
drivers/hid/hid-led.c
301
union delcom_packet dp = { .tx.major_cmd = 101, .tx.minor_cmd = 34 };
drivers/hid/hid-led.c
303
dp.tx.data_lsb = delcom_get_lednum(led);
drivers/hid/hid-led.c
304
dp.tx.data_msb = led->cdev.brightness;
drivers/hid/hid-led.c
52
} tx;
drivers/hwmon/sht21.c
162
u8 tx[2];
drivers/hwmon/sht21.c
170
.buf = tx,
drivers/hwmon/sht21.c
181
tx[0] = SHT21_READ_SNB_CMD1;
drivers/hwmon/sht21.c
182
tx[1] = SHT21_READ_SNB_CMD2;
drivers/hwmon/sht21.c
191
tx[0] = SHT21_READ_SNAC_CMD1;
drivers/hwmon/sht21.c
192
tx[1] = SHT21_READ_SNAC_CMD2;
drivers/i2c/busses/i2c-cpm.c
182
struct i2c_msg *pmsg, int num, int tx, int rx)
drivers/i2c/busses/i2c-cpm.c
191
tbdf = cpm->tbase + tx;
drivers/i2c/busses/i2c-cpm.c
196
tb = cpm->txbuf[tx];
drivers/i2c/busses/i2c-cpm.c
210
if (tx + 1 == num)
drivers/i2c/busses/i2c-cpm.c
241
struct i2c_msg *pmsg, int tx, int rx)
drivers/i2c/busses/i2c-cpm.c
249
tbdf = cpm->tbase + tx;
drivers/i2c/busses/i2c-cpm.c
252
tb = cpm->txbuf[tx];
drivers/i2c/busses/i2c-cpm.c
278
dev_dbg(&adap->dev, "tx sc %d 0x%04x\n", tx,
drivers/i2c/busses/i2c-dln2.c
111
} __packed tx;
drivers/i2c/busses/i2c-dln2.c
120
tx.port = dln2->port;
drivers/i2c/busses/i2c-dln2.c
121
tx.addr = addr;
drivers/i2c/busses/i2c-dln2.c
122
tx.mem_addr_len = 0;
drivers/i2c/busses/i2c-dln2.c
123
tx.mem_addr = 0;
drivers/i2c/busses/i2c-dln2.c
124
tx.buf_len = cpu_to_le16(data_len);
drivers/i2c/busses/i2c-dln2.c
126
ret = dln2_transfer(dln2->pdev, DLN2_I2C_READ, &tx, sizeof(tx),
drivers/i2c/busses/i2c-dln2.c
58
} tx;
drivers/i2c/busses/i2c-dln2.c
60
tx.port = dln2->port;
drivers/i2c/busses/i2c-dln2.c
67
return dln2_transfer_tx(dln2->pdev, cmd, &tx, sizeof(tx));
drivers/i2c/busses/i2c-dln2.c
81
} __packed *tx = dln2->buf;
drivers/i2c/busses/i2c-dln2.c
84
BUILD_BUG_ON(sizeof(*tx) > DLN2_I2C_BUF_SIZE);
drivers/i2c/busses/i2c-dln2.c
86
tx->port = dln2->port;
drivers/i2c/busses/i2c-dln2.c
87
tx->addr = addr;
drivers/i2c/busses/i2c-dln2.c
88
tx->mem_addr_len = 0;
drivers/i2c/busses/i2c-dln2.c
89
tx->mem_addr = 0;
drivers/i2c/busses/i2c-dln2.c
90
tx->buf_len = cpu_to_le16(data_len);
drivers/i2c/busses/i2c-dln2.c
91
memcpy(tx->buf, data, data_len);
drivers/i2c/busses/i2c-dln2.c
93
len = sizeof(*tx) + data_len - DLN2_I2C_MAX_XFER_SIZE;
drivers/i2c/busses/i2c-dln2.c
94
ret = dln2_transfer_tx(dln2->pdev, DLN2_I2C_WRITE, tx, len);
drivers/i2c/busses/i2c-imx-lpi2c.c
952
struct dma_slave_config rx = {}, tx = {};
drivers/i2c/busses/i2c-imx-lpi2c.c
958
tx.dst_addr = dma->phy_addr + LPI2C_MTDR;
drivers/i2c/busses/i2c-imx-lpi2c.c
959
tx.dst_addr_width = DMA_SLAVE_BUSWIDTH_2_BYTES;
drivers/i2c/busses/i2c-imx-lpi2c.c
960
tx.dst_maxburst = dma->tx_burst_num;
drivers/i2c/busses/i2c-imx-lpi2c.c
961
tx.direction = DMA_MEM_TO_DEV;
drivers/i2c/busses/i2c-imx-lpi2c.c
962
ret = dmaengine_slave_config(dma->chan_tx, &tx);
drivers/i2c/busses/i2c-imx-lpi2c.c
974
tx.dst_addr = dma->phy_addr + LPI2C_MTDR;
drivers/i2c/busses/i2c-imx-lpi2c.c
975
tx.dst_addr_width = DMA_SLAVE_BUSWIDTH_1_BYTE;
drivers/i2c/busses/i2c-imx-lpi2c.c
976
tx.dst_maxburst = dma->tx_burst_num;
drivers/i2c/busses/i2c-imx-lpi2c.c
977
tx.direction = DMA_MEM_TO_DEV;
drivers/i2c/busses/i2c-imx-lpi2c.c
978
ret = dmaengine_slave_config(dma->chan_tx, &tx);
drivers/i2c/busses/i2c-tegra-bpmp.c
187
msg.tx.data = request;
drivers/i2c/busses/i2c-tegra-bpmp.c
188
msg.tx.size = sizeof(*request);
drivers/iio/accel/sca3000.c
169
u8 tx[6] __aligned(IIO_DMA_MINALIGN);
drivers/iio/accel/sca3000.c
277
st->tx[0] = SCA3000_WRITE_REG(address);
drivers/iio/accel/sca3000.c
278
st->tx[1] = val;
drivers/iio/accel/sca3000.c
279
return spi_write(st->us, st->tx, 2);
drivers/iio/accel/sca3000.c
289
.tx_buf = st->tx,
drivers/iio/accel/sca3000.c
295
st->tx[0] = SCA3000_READ_REG(reg_address_high);
drivers/iio/accel/sca3000.c
331
.tx_buf = st->tx,
drivers/iio/accel/sca3000.c
335
.tx_buf = st->tx + 2,
drivers/iio/accel/sca3000.c
338
.tx_buf = st->tx + 4,
drivers/iio/accel/sca3000.c
341
st->tx[0] = SCA3000_WRITE_REG(SCA3000_REG_UNLOCK_ADDR);
drivers/iio/accel/sca3000.c
342
st->tx[1] = 0x00;
drivers/iio/accel/sca3000.c
343
st->tx[2] = SCA3000_WRITE_REG(SCA3000_REG_UNLOCK_ADDR);
drivers/iio/accel/sca3000.c
344
st->tx[3] = 0x50;
drivers/iio/accel/sca3000.c
345
st->tx[4] = SCA3000_WRITE_REG(SCA3000_REG_UNLOCK_ADDR);
drivers/iio/accel/sca3000.c
346
st->tx[5] = 0xA0;
drivers/iio/accel/sca3000.c
972
.tx_buf = st->tx,
drivers/iio/accel/sca3000.c
979
st->tx[0] = SCA3000_READ_REG(reg_address_high);
drivers/iio/adc/ad7280a.c
187
__be32 tx __aligned(IIO_DMA_MINALIGN);
drivers/iio/adc/ad7280a.c
231
.tx_buf = &st->tx,
drivers/iio/adc/ad7280a.c
233
.len = sizeof(st->tx),
drivers/iio/adc/ad7280a.c
236
st->tx = cpu_to_be32(AD7280A_READ_TXVAL);
drivers/iio/adc/ad7280a.c
260
st->tx = cpu_to_be32(reg);
drivers/iio/adc/ad7280a.c
262
return spi_write(st->spi, &st->tx, sizeof(st->tx));
drivers/iio/adc/ad7380.c
915
u16 tx;
drivers/iio/adc/ad7380.c
927
.tx_buf = &st->tx,
drivers/iio/adc/ad7380.c
930
st->tx = FIELD_PREP(AD7380_REG_WR, 1) |
drivers/iio/adc/ad7380.c
946
.tx_buf = &st->tx,
drivers/iio/adc/ad7380.c
961
st->tx = FIELD_PREP(AD7380_REG_WR, 0) |
drivers/iio/adc/max1363.c
797
u8 tx[2] = { st->setupbyte,
drivers/iio/adc/max1363.c
804
st->send(st->client, tx, 2);
drivers/iio/adc/ti-tsc2046.c
172
struct tsc2046_adc_atom *tx;
drivers/iio/adc/ti-tsc2046.c
362
return sizeof(*priv->tx) * max_count;
drivers/iio/adc/ti-tsc2046.c
379
priv->tx[l->offset + i].cmd = cmd;
drivers/iio/adc/ti-tsc2046.c
382
priv->tx[l->offset + i].cmd = tsc2046_adc_get_cmd(priv, ch_idx, false);
drivers/iio/adc/ti-tsc2046.c
685
priv->tx = devm_kzalloc(&priv->spi->dev, size, GFP_KERNEL);
drivers/iio/adc/ti-tsc2046.c
686
if (!priv->tx)
drivers/iio/adc/ti-tsc2046.c
693
priv->xfer.tx_buf = priv->tx;
drivers/iio/gyro/adxrs290.c
506
u8 tx = ADXRS290_READ_REG(ADXRS290_REG_DATAX0);
drivers/iio/gyro/adxrs290.c
512
ret = spi_write_then_read(st->spi, &tx, sizeof(tx), st->buffer.channels,
drivers/iio/gyro/adxrs450.c
107
tx = ADXRS450_READ_DATA | (reg_address << 17);
drivers/iio/gyro/adxrs450.c
109
if (!(hweight32(tx) & 1))
drivers/iio/gyro/adxrs450.c
110
tx |= ADXRS450_P;
drivers/iio/gyro/adxrs450.c
112
st->tx = cpu_to_be32(tx);
drivers/iio/gyro/adxrs450.c
139
u32 tx;
drivers/iio/gyro/adxrs450.c
143
tx = ADXRS450_WRITE_DATA | (reg_address << 17) | (val << 1);
drivers/iio/gyro/adxrs450.c
145
if (!(hweight32(tx) & 1))
drivers/iio/gyro/adxrs450.c
146
tx |= ADXRS450_P;
drivers/iio/gyro/adxrs450.c
148
st->tx = cpu_to_be32(tx);
drivers/iio/gyro/adxrs450.c
149
ret = spi_write(st->us, &st->tx, sizeof(st->tx));
drivers/iio/gyro/adxrs450.c
169
.tx_buf = &st->tx,
drivers/iio/gyro/adxrs450.c
170
.len = sizeof(st->tx),
drivers/iio/gyro/adxrs450.c
179
st->tx = cpu_to_be32(ADXRS450_SENSOR_DATA);
drivers/iio/gyro/adxrs450.c
204
u32 tx;
drivers/iio/gyro/adxrs450.c
206
.tx_buf = &st->tx,
drivers/iio/gyro/adxrs450.c
208
.len = sizeof(st->tx),
drivers/iio/gyro/adxrs450.c
212
tx = ADXRS450_SENSOR_DATA;
drivers/iio/gyro/adxrs450.c
214
tx |= (ADXRS450_CHK | ADXRS450_P);
drivers/iio/gyro/adxrs450.c
215
st->tx = cpu_to_be32(tx);
drivers/iio/gyro/adxrs450.c
76
__be32 tx __aligned(IIO_DMA_MINALIGN);
drivers/iio/gyro/adxrs450.c
93
u32 tx;
drivers/iio/gyro/adxrs450.c
97
.tx_buf = &st->tx,
drivers/iio/gyro/adxrs450.c
98
.len = sizeof(st->tx),
drivers/iio/gyro/itg3200_buffer.c
25
u8 tx = 0x80 | ITG3200_REG_TEMP_OUT_H;
drivers/iio/gyro/itg3200_buffer.c
31
.buf = &tx,
drivers/iio/health/afe4403.c
217
u8 tx[4] = {AFE440X_CONTROL0, 0x0, 0x0, AFE440X_CONTROL0_READ};
drivers/iio/health/afe4403.c
222
ret = spi_write_then_read(afe->spi, tx, 4, NULL, 0);
drivers/iio/health/afe4403.c
233
tx[3] = AFE440X_CONTROL0_WRITE;
drivers/iio/health/afe4403.c
234
ret = spi_write_then_read(afe->spi, tx, 4, NULL, 0);
drivers/iio/health/afe4403.c
314
u8 tx[4] = {AFE440X_CONTROL0, 0x0, 0x0, AFE440X_CONTROL0_READ};
drivers/iio/health/afe4403.c
318
ret = spi_write_then_read(afe->spi, tx, 4, NULL, 0);
drivers/iio/health/afe4403.c
333
tx[3] = AFE440X_CONTROL0_WRITE;
drivers/iio/health/afe4403.c
334
ret = spi_write_then_read(afe->spi, tx, 4, NULL, 0);
drivers/iio/imu/adis.c
130
.tx_buf = adis->tx,
drivers/iio/imu/adis.c
136
.tx_buf = adis->tx + 2,
drivers/iio/imu/adis.c
142
.tx_buf = adis->tx + 4,
drivers/iio/imu/adis.c
159
adis->tx[0] = ADIS_WRITE_REG(ADIS_REG_PAGE_ID);
drivers/iio/imu/adis.c
160
adis->tx[1] = page;
drivers/iio/imu/adis.c
166
adis->tx[2] = ADIS_READ_REG(reg + 2);
drivers/iio/imu/adis.c
167
adis->tx[3] = 0;
drivers/iio/imu/adis.c
171
adis->tx[4] = ADIS_READ_REG(reg);
drivers/iio/imu/adis.c
172
adis->tx[5] = 0;
drivers/iio/imu/adis.c
41
.tx_buf = adis->tx,
drivers/iio/imu/adis.c
47
.tx_buf = adis->tx + 2,
drivers/iio/imu/adis.c
53
.tx_buf = adis->tx + 4,
drivers/iio/imu/adis.c
59
.tx_buf = adis->tx + 6,
drivers/iio/imu/adis.c
64
.tx_buf = adis->tx + 8,
drivers/iio/imu/adis.c
74
adis->tx[0] = ADIS_WRITE_REG(ADIS_REG_PAGE_ID);
drivers/iio/imu/adis.c
75
adis->tx[1] = page;
drivers/iio/imu/adis.c
81
adis->tx[8] = ADIS_WRITE_REG(reg + 3);
drivers/iio/imu/adis.c
82
adis->tx[9] = (value >> 24) & 0xff;
drivers/iio/imu/adis.c
83
adis->tx[6] = ADIS_WRITE_REG(reg + 2);
drivers/iio/imu/adis.c
84
adis->tx[7] = (value >> 16) & 0xff;
drivers/iio/imu/adis.c
87
adis->tx[4] = ADIS_WRITE_REG(reg + 1);
drivers/iio/imu/adis.c
88
adis->tx[5] = (value >> 8) & 0xff;
drivers/iio/imu/adis.c
91
adis->tx[2] = ADIS_WRITE_REG(reg);
drivers/iio/imu/adis.c
92
adis->tx[3] = value & 0xff;
drivers/iio/imu/adis16475.c
1681
u8 *tx;
drivers/iio/imu/adis16475.c
1688
tx = adis->buffer + burst_max_length;
drivers/iio/imu/adis16475.c
1689
tx[0] = ADIS_READ_REG(burst_req);
drivers/iio/imu/adis16480.c
1407
adis->tx[0] = ADIS_WRITE_REG(ADIS_REG_PAGE_ID);
drivers/iio/imu/adis16480.c
1408
adis->tx[1] = 0;
drivers/iio/imu/adis16480.c
1409
ret = spi_write(adis->spi, adis->tx, 2);
drivers/iio/imu/adis16550.c
885
u8 *tx;
drivers/iio/imu/adis16550.c
894
tx = st->buffer + burst_length;
drivers/iio/imu/adis16550.c
895
tx[0] = 0x00;
drivers/iio/imu/adis16550.c
896
tx[1] = 0x00;
drivers/iio/imu/adis16550.c
897
tx[2] = burst_cmd;
drivers/iio/imu/adis16550.c
899
tx[3] = spi_crc4(get_unaligned_le32(tx));
drivers/iio/imu/adis_buffer.c
107
adis->xfer[j].tx_buf = &tx[j];
drivers/iio/imu/adis_buffer.c
118
*tx++ = cpu_to_be16((chan->address + 2) << 8);
drivers/iio/imu/adis_buffer.c
119
*tx++ = cpu_to_be16(chan->address << 8);
drivers/iio/imu/adis_buffer.c
132
adis->tx[0] = ADIS_WRITE_REG(ADIS_REG_PAGE_ID);
drivers/iio/imu/adis_buffer.c
133
adis->tx[1] = 0;
drivers/iio/imu/adis_buffer.c
134
ret = spi_write(adis->spi, adis->tx, 2);
drivers/iio/imu/adis_buffer.c
27
u8 *tx;
drivers/iio/imu/adis_buffer.c
47
tx = adis->buffer + burst_max_length;
drivers/iio/imu/adis_buffer.c
48
tx[0] = ADIS_READ_REG(adis->data->burst_reg_cmd);
drivers/iio/imu/adis_buffer.c
49
tx[1] = 0;
drivers/iio/imu/adis_buffer.c
51
adis->xfer[0].tx_buf = tx;
drivers/iio/imu/adis_buffer.c
74
__be16 *tx, *rx;
drivers/iio/imu/adis_buffer.c
96
tx = rx + scan_count;
drivers/iio/pressure/mpl115_spi.c
19
u8 tx[4];
drivers/iio/pressure/mpl115_spi.c
42
.tx_buf = buf->tx,
drivers/iio/pressure/mpl115_spi.c
48
buf->tx[0] = MPL115_SPI_READ(address);
drivers/iio/pressure/mpl115_spi.c
49
buf->tx[2] = MPL115_SPI_READ(address + 1);
drivers/iio/pressure/mpl115_spi.c
63
.tx_buf = buf->tx,
drivers/iio/pressure/mpl115_spi.c
67
buf->tx[0] = MPL115_SPI_WRITE(address);
drivers/iio/pressure/mpl115_spi.c
68
buf->tx[1] = value;
drivers/iio/resolver/ad2s1210.c
173
u8 tx[2];
drivers/iio/resolver/ad2s1210.c
203
.tx_buf = &st->tx[0],
drivers/iio/resolver/ad2s1210.c
208
.tx_buf = &st->tx[1],
drivers/iio/resolver/ad2s1210.c
217
st->tx[0] = reg;
drivers/iio/resolver/ad2s1210.c
218
st->tx[1] = val;
drivers/iio/resolver/ad2s1210.c
249
.tx_buf = &st->tx[0],
drivers/iio/resolver/ad2s1210.c
254
.tx_buf = &st->tx[1],
drivers/iio/resolver/ad2s1210.c
263
st->tx[0] = reg;
drivers/iio/resolver/ad2s1210.c
269
st->tx[1] = AD2S1210_REG_CONTROL;
drivers/infiniband/hw/hfi1/chip.c
7312
u16 tx, rx;
drivers/infiniband/hw/hfi1/chip.c
7325
tx = nibble_to_count(enable_lane_tx);
drivers/infiniband/hw/hfi1/chip.c
7354
enable_lane_tx, tx, enable_lane_rx, rx);
drivers/infiniband/hw/hfi1/chip.c
7355
*tx_width = link_width_to_bits(dd, tx);
drivers/infiniband/hw/hfi1/chip.c
7375
u16 widths, tx, rx;
drivers/infiniband/hw/hfi1/chip.c
7380
tx = widths >> 12;
drivers/infiniband/hw/hfi1/chip.c
7383
*tx_width = link_width_to_bits(dd, tx);
drivers/infiniband/hw/hfi1/chip.c
7605
u16 tx, rx;
drivers/infiniband/hw/hfi1/chip.c
7631
get_link_widths(ppd->dd, &tx, &rx);
drivers/infiniband/hw/hfi1/chip.c
7632
ppd->link_width_downgrade_tx_active = tx;
drivers/infiniband/hw/hfi1/iowait.c
46
struct sdma_txreq *tx,
drivers/infiniband/hw/hfi1/iowait.h
103
struct sdma_txreq *tx,
drivers/infiniband/hw/hfi1/iowait.h
135
struct sdma_txreq *tx,
drivers/infiniband/hw/hfi1/iowait.h
287
struct sdma_txreq *tx = NULL;
drivers/infiniband/hw/hfi1/iowait.h
290
tx = list_first_entry(
drivers/infiniband/hw/hfi1/iowait.h
294
list_del_init(&tx->list);
drivers/infiniband/hw/hfi1/iowait.h
296
return tx;
drivers/infiniband/hw/hfi1/iowait.h
302
struct sdma_txreq *tx = NULL;
drivers/infiniband/hw/hfi1/iowait.h
305
tx = list_first_entry(&w->tx_head, struct sdma_txreq,
drivers/infiniband/hw/hfi1/iowait.h
307
num_desc = tx->num_desc;
drivers/infiniband/hw/hfi1/iowait.h
308
if (tx->flags & SDMA_TXREQ_F_VIP)
drivers/infiniband/hw/hfi1/iowait.h
325
struct sdma_txreq *tx = NULL;
drivers/infiniband/hw/hfi1/iowait.h
328
tx = list_first_entry(&w->tx_head, struct sdma_txreq,
drivers/infiniband/hw/hfi1/iowait.h
330
if (tx->flags & SDMA_TXREQ_F_VIP)
drivers/infiniband/hw/hfi1/ipoib_tx.c
114
static void hfi1_ipoib_free_tx(struct ipoib_txreq *tx, int budget)
drivers/infiniband/hw/hfi1/ipoib_tx.c
116
struct hfi1_ipoib_dev_priv *priv = tx->txq->priv;
drivers/infiniband/hw/hfi1/ipoib_tx.c
118
if (likely(!tx->sdma_status)) {
drivers/infiniband/hw/hfi1/ipoib_tx.c
119
dev_sw_netstats_tx_add(priv->netdev, 1, tx->skb->len);
drivers/infiniband/hw/hfi1/ipoib_tx.c
124
__func__, tx->sdma_status,
drivers/infiniband/hw/hfi1/ipoib_tx.c
125
le64_to_cpu(tx->sdma_hdr->pbc), tx->txq->q_idx,
drivers/infiniband/hw/hfi1/ipoib_tx.c
126
tx->txq->sde->this_idx);
drivers/infiniband/hw/hfi1/ipoib_tx.c
129
napi_consume_skb(tx->skb, budget);
drivers/infiniband/hw/hfi1/ipoib_tx.c
130
tx->skb = NULL;
drivers/infiniband/hw/hfi1/ipoib_tx.c
131
sdma_txclean(priv->dd, &tx->txreq);
drivers/infiniband/hw/hfi1/ipoib_tx.c
138
struct ipoib_txreq *tx;
drivers/infiniband/hw/hfi1/ipoib_tx.c
141
tx = hfi1_txreq_from_idx(tx_ring, i);
drivers/infiniband/hw/hfi1/ipoib_tx.c
142
tx->complete = 0;
drivers/infiniband/hw/hfi1/ipoib_tx.c
143
dev_kfree_skb_any(tx->skb);
drivers/infiniband/hw/hfi1/ipoib_tx.c
144
tx->skb = NULL;
drivers/infiniband/hw/hfi1/ipoib_tx.c
145
sdma_txclean(txq->priv->dd, &tx->txreq);
drivers/infiniband/hw/hfi1/ipoib_tx.c
162
struct ipoib_txreq *tx = hfi1_txreq_from_idx(tx_ring, head);
drivers/infiniband/hw/hfi1/ipoib_tx.c
167
if (!smp_load_acquire(&tx->complete))
drivers/infiniband/hw/hfi1/ipoib_tx.c
169
tx->complete = 0;
drivers/infiniband/hw/hfi1/ipoib_tx.c
170
trace_hfi1_tx_produce(tx, head);
drivers/infiniband/hw/hfi1/ipoib_tx.c
171
hfi1_ipoib_free_tx(tx, budget);
drivers/infiniband/hw/hfi1/ipoib_tx.c
173
tx = hfi1_txreq_from_idx(tx_ring, head);
drivers/infiniband/hw/hfi1/ipoib_tx.c
190
struct ipoib_txreq *tx = container_of(txreq, struct ipoib_txreq, txreq);
drivers/infiniband/hw/hfi1/ipoib_tx.c
192
trace_hfi1_txq_complete(tx->txq);
drivers/infiniband/hw/hfi1/ipoib_tx.c
193
tx->sdma_status = status;
drivers/infiniband/hw/hfi1/ipoib_tx.c
195
smp_store_release(&tx->complete, 1);
drivers/infiniband/hw/hfi1/ipoib_tx.c
196
napi_schedule_irqoff(&tx->txq->napi);
drivers/infiniband/hw/hfi1/ipoib_tx.c
199
static int hfi1_ipoib_build_ulp_payload(struct ipoib_txreq *tx,
drivers/infiniband/hw/hfi1/ipoib_tx.c
203
struct sdma_txreq *txreq = &tx->txreq;
drivers/infiniband/hw/hfi1/ipoib_tx.c
204
struct sk_buff *skb = tx->skb;
drivers/infiniband/hw/hfi1/ipoib_tx.c
230
static int hfi1_ipoib_build_tx_desc(struct ipoib_txreq *tx,
drivers/infiniband/hw/hfi1/ipoib_tx.c
234
struct sdma_txreq *txreq = &tx->txreq;
drivers/infiniband/hw/hfi1/ipoib_tx.c
235
struct hfi1_sdma_header *sdma_hdr = tx->sdma_hdr;
drivers/infiniband/hw/hfi1/ipoib_tx.c
237
sizeof(sdma_hdr->pbc) + (txp->hdr_dwords << 2) + tx->skb->len;
drivers/infiniband/hw/hfi1/ipoib_tx.c
253
return hfi1_ipoib_build_ulp_payload(tx, txp);
drivers/infiniband/hw/hfi1/ipoib_tx.c
256
static void hfi1_ipoib_build_ib_tx_headers(struct ipoib_txreq *tx,
drivers/infiniband/hw/hfi1/ipoib_tx.c
259
struct hfi1_ipoib_dev_priv *priv = tx->txq->priv;
drivers/infiniband/hw/hfi1/ipoib_tx.c
260
struct hfi1_sdma_header *sdma_hdr = tx->sdma_hdr;
drivers/infiniband/hw/hfi1/ipoib_tx.c
261
struct sk_buff *skb = tx->skb;
drivers/infiniband/hw/hfi1/ipoib_tx.c
354
struct ipoib_txreq *tx;
drivers/infiniband/hw/hfi1/ipoib_tx.c
373
tx = hfi1_txreq_from_idx(tx_ring, tail);
drivers/infiniband/hw/hfi1/ipoib_tx.c
377
tx->txreq.num_desc = 0;
drivers/infiniband/hw/hfi1/ipoib_tx.c
378
tx->txq = txq;
drivers/infiniband/hw/hfi1/ipoib_tx.c
379
tx->skb = skb;
drivers/infiniband/hw/hfi1/ipoib_tx.c
380
INIT_LIST_HEAD(&tx->txreq.list);
drivers/infiniband/hw/hfi1/ipoib_tx.c
382
hfi1_ipoib_build_ib_tx_headers(tx, txp);
drivers/infiniband/hw/hfi1/ipoib_tx.c
384
ret = hfi1_ipoib_build_tx_desc(tx, txp);
drivers/infiniband/hw/hfi1/ipoib_tx.c
396
return tx;
drivers/infiniband/hw/hfi1/ipoib_tx.c
399
sdma_txclean(priv->dd, &tx->txreq);
drivers/infiniband/hw/hfi1/ipoib_tx.c
440
struct ipoib_txreq *tx)
drivers/infiniband/hw/hfi1/ipoib_tx.c
446
&tx->txreq,
drivers/infiniband/hw/hfi1/ipoib_tx.c
462
struct ipoib_txreq *tx;
drivers/infiniband/hw/hfi1/ipoib_tx.c
465
tx = hfi1_ipoib_send_dma_common(dev, skb, txp);
drivers/infiniband/hw/hfi1/ipoib_tx.c
466
if (IS_ERR(tx)) {
drivers/infiniband/hw/hfi1/ipoib_tx.c
467
int ret = PTR_ERR(tx);
drivers/infiniband/hw/hfi1/ipoib_tx.c
480
trace_hfi1_tx_consume(tx, tx_ring->tail);
drivers/infiniband/hw/hfi1/ipoib_tx.c
483
ret = hfi1_ipoib_submit_tx(txq, tx);
drivers/infiniband/hw/hfi1/ipoib_tx.c
487
&tx->sdma_hdr->hdr,
drivers/infiniband/hw/hfi1/ipoib_tx.c
499
smp_store_release(&tx->complete, 1);
drivers/infiniband/hw/hfi1/ipoib_tx.c
500
napi_schedule(&tx->txq->napi);
drivers/infiniband/hw/hfi1/ipoib_tx.c
513
struct ipoib_txreq *tx;
drivers/infiniband/hw/hfi1/ipoib_tx.c
528
tx = hfi1_ipoib_send_dma_common(dev, skb, txp);
drivers/infiniband/hw/hfi1/ipoib_tx.c
529
if (IS_ERR(tx)) {
drivers/infiniband/hw/hfi1/ipoib_tx.c
530
int ret = PTR_ERR(tx);
drivers/infiniband/hw/hfi1/ipoib_tx.c
543
trace_hfi1_tx_consume(tx, tx_ring->tail);
drivers/infiniband/hw/hfi1/ipoib_tx.c
546
list_add_tail(&tx->txreq.list, &txq->tx_list);
drivers/infiniband/hw/hfi1/ipoib_tx.c
551
&tx->sdma_hdr->hdr,
drivers/infiniband/hw/hfi1/ipoib_tx.c
706
struct ipoib_txreq *tx;
drivers/infiniband/hw/hfi1/ipoib_tx.c
743
kzalloc_node(sizeof(*tx->sdma_hdr),
drivers/infiniband/hw/hfi1/ipoib_tx.c
776
struct ipoib_txreq *tx =
drivers/infiniband/hw/hfi1/ipoib_tx.c
780
sdma_txclean(txq->priv->dd, &tx->txreq);
drivers/infiniband/hw/hfi1/ipoib_tx.c
781
dev_kfree_skb_any(tx->skb);
drivers/infiniband/hw/hfi1/ipoib_tx.c
782
tx->skb = NULL;
drivers/infiniband/hw/hfi1/pin_system.c
283
struct user_sdma_txreq *tx,
drivers/infiniband/hw/hfi1/pin_system.c
325
ret = sdma_txadd_page(pq->dd, &tx->txreq,
drivers/infiniband/hw/hfi1/pin_system.c
348
struct user_sdma_txreq *tx,
drivers/infiniband/hw/hfi1/pin_system.c
370
ret = add_mapping_to_sdma_packet(req, tx, cache_entry, start,
drivers/infiniband/hw/hfi1/pin_system.c
402
struct user_sdma_txreq *tx,
drivers/infiniband/hw/hfi1/pin_system.c
428
ret = add_system_iovec_to_sdma_packet(req, tx, cur_iovec,
drivers/infiniband/hw/hfi1/pinning.h
16
struct user_sdma_txreq *tx,
drivers/infiniband/hw/hfi1/qp.c
110
struct sdma_txreq *tx;
drivers/infiniband/hw/hfi1/qp.c
112
tx = list_first_entry(
drivers/infiniband/hw/hfi1/qp.c
116
list_del_init(&tx->list);
drivers/infiniband/hw/hfi1/qp.c
118
container_of(tx, struct verbs_txreq, txreq));
drivers/infiniband/hw/hfi1/qp.c
443
struct verbs_txreq *tx = container_of(stx, struct verbs_txreq, txreq);
drivers/infiniband/hw/hfi1/qp.c
449
qp = tx->qp;
drivers/infiniband/hw/hfi1/qp.c
483
hfi1_put_txreq(tx);
drivers/infiniband/hw/hfi1/sdma.c
1599
static inline u8 ahg_mode(struct sdma_txreq *tx)
drivers/infiniband/hw/hfi1/sdma.c
1601
return (tx->descp[0].qw[1] & SDMA_DESC1_HEADER_MODE_SMASK)
drivers/infiniband/hw/hfi1/sdma.c
1618
struct sdma_txreq *tx)
drivers/infiniband/hw/hfi1/sdma.c
1622
if (tx->num_desc) {
drivers/infiniband/hw/hfi1/sdma.c
1623
u8 skip = 0, mode = ahg_mode(tx);
drivers/infiniband/hw/hfi1/sdma.c
1626
sdma_unmap_desc(dd, &tx->descp[0]);
drivers/infiniband/hw/hfi1/sdma.c
1630
for (i = 1 + skip; i < tx->num_desc; i++)
drivers/infiniband/hw/hfi1/sdma.c
1631
sdma_unmap_desc(dd, &tx->descp[i]);
drivers/infiniband/hw/hfi1/sdma.c
1632
tx->num_desc = 0;
drivers/infiniband/hw/hfi1/sdma.c
1634
kfree(tx->coalesce_buf);
drivers/infiniband/hw/hfi1/sdma.c
1635
tx->coalesce_buf = NULL;
drivers/infiniband/hw/hfi1/sdma.c
1637
if (unlikely(tx->desc_limit > ARRAY_SIZE(tx->descs))) {
drivers/infiniband/hw/hfi1/sdma.c
1638
tx->desc_limit = ARRAY_SIZE(tx->descs);
drivers/infiniband/hw/hfi1/sdma.c
1639
kfree(tx->descp);
drivers/infiniband/hw/hfi1/sdma.c
2248
static inline u16 submit_tx(struct sdma_engine *sde, struct sdma_txreq *tx)
drivers/infiniband/hw/hfi1/sdma.c
2252
struct sdma_desc *descp = tx->descp;
drivers/infiniband/hw/hfi1/sdma.c
2253
u8 skip = 0, mode = ahg_mode(tx);
drivers/infiniband/hw/hfi1/sdma.c
2264
for (i = 1; i < tx->num_desc; i++, descp++) {
drivers/infiniband/hw/hfi1/sdma.c
2281
tx->next_descq_idx = tail;
drivers/infiniband/hw/hfi1/sdma.c
2283
tx->sn = sde->tail_sn++;
drivers/infiniband/hw/hfi1/sdma.c
2284
trace_hfi1_sdma_in_sn(sde, tx->sn);
drivers/infiniband/hw/hfi1/sdma.c
2287
sde->tx_ring[sde->tx_tail++ & sde->sdma_mask] = tx;
drivers/infiniband/hw/hfi1/sdma.c
2288
sde->desc_avail -= tx->num_desc;
drivers/infiniband/hw/hfi1/sdma.c
2298
struct sdma_txreq *tx,
drivers/infiniband/hw/hfi1/sdma.c
2304
if (tx->num_desc <= sde->desc_avail)
drivers/infiniband/hw/hfi1/sdma.c
2312
ret = wait->iow->sleep(sde, wait, tx, seq, pkts_sent);
drivers/infiniband/hw/hfi1/sdma.c
2338
struct sdma_txreq *tx,
drivers/infiniband/hw/hfi1/sdma.c
2346
if (unlikely(tx->tlen))
drivers/infiniband/hw/hfi1/sdma.c
2348
tx->wait = iowait_ioww_to_iow(wait);
drivers/infiniband/hw/hfi1/sdma.c
2353
if (unlikely(tx->num_desc > sde->desc_avail))
drivers/infiniband/hw/hfi1/sdma.c
2355
tail = submit_tx(sde, tx);
drivers/infiniband/hw/hfi1/sdma.c
2365
tx->next_descq_idx = 0;
drivers/infiniband/hw/hfi1/sdma.c
2367
tx->sn = sde->tail_sn++;
drivers/infiniband/hw/hfi1/sdma.c
2368
trace_hfi1_sdma_in_sn(sde, tx->sn);
drivers/infiniband/hw/hfi1/sdma.c
2371
list_add_tail(&tx->list, &sde->flushlist);
drivers/infiniband/hw/hfi1/sdma.c
2373
iowait_inc_wait_count(wait, tx->num_desc);
drivers/infiniband/hw/hfi1/sdma.c
2378
ret = sdma_check_progress(sde, wait, tx, pkts_sent);
drivers/infiniband/hw/hfi1/sdma.c
2418
struct sdma_txreq *tx, *tx_next;
drivers/infiniband/hw/hfi1/sdma.c
2426
list_for_each_entry_safe(tx, tx_next, tx_list, list) {
drivers/infiniband/hw/hfi1/sdma.c
2427
tx->wait = iowait_ioww_to_iow(wait);
drivers/infiniband/hw/hfi1/sdma.c
2430
if (unlikely(tx->num_desc > sde->desc_avail))
drivers/infiniband/hw/hfi1/sdma.c
2432
if (unlikely(tx->tlen)) {
drivers/infiniband/hw/hfi1/sdma.c
2436
list_del_init(&tx->list);
drivers/infiniband/hw/hfi1/sdma.c
2437
tail = submit_tx(sde, tx);
drivers/infiniband/hw/hfi1/sdma.c
2459
list_for_each_entry_safe(tx, tx_next, tx_list, list) {
drivers/infiniband/hw/hfi1/sdma.c
2460
tx->wait = iowait_ioww_to_iow(wait);
drivers/infiniband/hw/hfi1/sdma.c
2461
list_del_init(&tx->list);
drivers/infiniband/hw/hfi1/sdma.c
2462
tx->next_descq_idx = 0;
drivers/infiniband/hw/hfi1/sdma.c
2464
tx->sn = sde->tail_sn++;
drivers/infiniband/hw/hfi1/sdma.c
2465
trace_hfi1_sdma_in_sn(sde, tx->sn);
drivers/infiniband/hw/hfi1/sdma.c
2467
list_add_tail(&tx->list, &sde->flushlist);
drivers/infiniband/hw/hfi1/sdma.c
2469
iowait_inc_wait_count(wait, tx->num_desc);
drivers/infiniband/hw/hfi1/sdma.c
2476
ret = sdma_check_progress(sde, wait, tx, submit_count > 0);
drivers/infiniband/hw/hfi1/sdma.c
2996
static int _extend_sdma_tx_descs(struct hfi1_devdata *dd, struct sdma_txreq *tx)
drivers/infiniband/hw/hfi1/sdma.c
3002
if (unlikely((tx->num_desc == (MAX_DESC - 1)))) {
drivers/infiniband/hw/hfi1/sdma.c
3004
if (!tx->tlen) {
drivers/infiniband/hw/hfi1/sdma.c
3005
tx->desc_limit = MAX_DESC;
drivers/infiniband/hw/hfi1/sdma.c
3006
} else if (!tx->coalesce_buf) {
drivers/infiniband/hw/hfi1/sdma.c
3008
tx->coalesce_buf = kmalloc(tx->tlen + sizeof(u32),
drivers/infiniband/hw/hfi1/sdma.c
3010
if (!tx->coalesce_buf)
drivers/infiniband/hw/hfi1/sdma.c
3012
tx->coalesce_idx = 0;
drivers/infiniband/hw/hfi1/sdma.c
3017
if (unlikely(tx->num_desc == MAX_DESC))
drivers/infiniband/hw/hfi1/sdma.c
3023
tx->descp = descp;
drivers/infiniband/hw/hfi1/sdma.c
3026
tx->desc_limit = MAX_DESC - 1;
drivers/infiniband/hw/hfi1/sdma.c
3028
for (i = 0; i < tx->num_desc; i++)
drivers/infiniband/hw/hfi1/sdma.c
3029
tx->descp[i] = tx->descs[i];
drivers/infiniband/hw/hfi1/sdma.c
3032
__sdma_txclean(dd, tx);
drivers/infiniband/hw/hfi1/sdma.c
3052
int ext_coal_sdma_tx_descs(struct hfi1_devdata *dd, struct sdma_txreq *tx,
drivers/infiniband/hw/hfi1/sdma.c
3059
rval = _extend_sdma_tx_descs(dd, tx);
drivers/infiniband/hw/hfi1/sdma.c
3061
__sdma_txclean(dd, tx);
drivers/infiniband/hw/hfi1/sdma.c
3066
if (tx->coalesce_buf) {
drivers/infiniband/hw/hfi1/sdma.c
3068
__sdma_txclean(dd, tx);
drivers/infiniband/hw/hfi1/sdma.c
3076
__sdma_txclean(dd, tx);
drivers/infiniband/hw/hfi1/sdma.c
3080
memcpy(tx->coalesce_buf + tx->coalesce_idx, kvaddr, len);
drivers/infiniband/hw/hfi1/sdma.c
3081
tx->coalesce_idx += len;
drivers/infiniband/hw/hfi1/sdma.c
3086
if (tx->tlen - tx->coalesce_idx)
drivers/infiniband/hw/hfi1/sdma.c
3090
pad_len = tx->packet_len & (sizeof(u32) - 1);
drivers/infiniband/hw/hfi1/sdma.c
3093
memset(tx->coalesce_buf + tx->coalesce_idx, 0, pad_len);
drivers/infiniband/hw/hfi1/sdma.c
3095
tx->packet_len += pad_len;
drivers/infiniband/hw/hfi1/sdma.c
3096
tx->tlen += pad_len;
drivers/infiniband/hw/hfi1/sdma.c
3101
tx->coalesce_buf,
drivers/infiniband/hw/hfi1/sdma.c
3102
tx->tlen,
drivers/infiniband/hw/hfi1/sdma.c
3106
__sdma_txclean(dd, tx);
drivers/infiniband/hw/hfi1/sdma.c
3111
tx->desc_limit = MAX_DESC;
drivers/infiniband/hw/hfi1/sdma.c
3112
return _sdma_txadd_daddr(dd, SDMA_MAP_SINGLE, tx,
drivers/infiniband/hw/hfi1/sdma.c
3113
addr, tx->tlen, NULL, NULL, NULL);
drivers/infiniband/hw/hfi1/sdma.c
3140
int _pad_sdma_tx_descs(struct hfi1_devdata *dd, struct sdma_txreq *tx)
drivers/infiniband/hw/hfi1/sdma.c
3144
if ((unlikely(tx->num_desc == tx->desc_limit))) {
drivers/infiniband/hw/hfi1/sdma.c
3145
rval = _extend_sdma_tx_descs(dd, tx);
drivers/infiniband/hw/hfi1/sdma.c
3147
__sdma_txclean(dd, tx);
drivers/infiniband/hw/hfi1/sdma.c
3154
tx,
drivers/infiniband/hw/hfi1/sdma.c
3157
sizeof(u32) - (tx->packet_len & (sizeof(u32) - 1)),
drivers/infiniband/hw/hfi1/sdma.c
3159
tx->num_desc++;
drivers/infiniband/hw/hfi1/sdma.c
3160
_sdma_close_tx(dd, tx);
drivers/infiniband/hw/hfi1/sdma.c
3172
struct sdma_txreq *tx,
drivers/infiniband/hw/hfi1/sdma.c
3189
tx->num_desc++;
drivers/infiniband/hw/hfi1/sdma.c
3193
tx->num_desc++;
drivers/infiniband/hw/hfi1/sdma.c
3194
tx->descs[2].qw[0] = 0;
drivers/infiniband/hw/hfi1/sdma.c
3195
tx->descs[2].qw[1] = 0;
drivers/infiniband/hw/hfi1/sdma.c
3198
tx->num_desc++;
drivers/infiniband/hw/hfi1/sdma.c
3199
tx->descs[1].qw[0] = 0;
drivers/infiniband/hw/hfi1/sdma.c
3200
tx->descs[1].qw[1] = 0;
drivers/infiniband/hw/hfi1/sdma.c
3204
tx->descs[0].qw[1] |=
drivers/infiniband/hw/hfi1/sdma.c
3216
tx->descs[desc].qw[!!(i & 2)] |=
drivers/infiniband/hw/hfi1/sdma.c
323
struct sdma_txreq *tx,
drivers/infiniband/hw/hfi1/sdma.c
327
struct iowait *wait = tx->wait;
drivers/infiniband/hw/hfi1/sdma.c
328
callback_t complete = tx->complete;
drivers/infiniband/hw/hfi1/sdma.c
331
trace_hfi1_sdma_out_sn(sde, tx->sn);
drivers/infiniband/hw/hfi1/sdma.c
332
if (WARN_ON_ONCE(sde->head_sn != tx->sn))
drivers/infiniband/hw/hfi1/sdma.c
334
sde->head_sn, tx->sn);
drivers/infiniband/hw/hfi1/sdma.c
337
__sdma_txclean(sde->dd, tx);
drivers/infiniband/hw/hfi1/sdma.c
339
(*complete)(tx, res);
drivers/infiniband/hw/hfi1/sdma.h
439
struct sdma_txreq *tx,
drivers/infiniband/hw/hfi1/sdma.h
498
struct sdma_txreq *tx,
drivers/infiniband/hw/hfi1/sdma.h
511
tx->desc_limit = ARRAY_SIZE(tx->descs);
drivers/infiniband/hw/hfi1/sdma.h
512
tx->descp = &tx->descs[0];
drivers/infiniband/hw/hfi1/sdma.h
513
INIT_LIST_HEAD(&tx->list);
drivers/infiniband/hw/hfi1/sdma.h
514
tx->num_desc = 0;
drivers/infiniband/hw/hfi1/sdma.h
515
tx->flags = flags;
drivers/infiniband/hw/hfi1/sdma.h
516
tx->complete = cb;
drivers/infiniband/hw/hfi1/sdma.h
517
tx->coalesce_buf = NULL;
drivers/infiniband/hw/hfi1/sdma.h
518
tx->wait = NULL;
drivers/infiniband/hw/hfi1/sdma.h
519
tx->packet_len = tlen;
drivers/infiniband/hw/hfi1/sdma.h
520
tx->tlen = tx->packet_len;
drivers/infiniband/hw/hfi1/sdma.h
521
tx->descs[0].qw[0] = SDMA_DESC0_FIRST_DESC_FLAG;
drivers/infiniband/hw/hfi1/sdma.h
522
tx->descs[0].qw[1] = 0;
drivers/infiniband/hw/hfi1/sdma.h
524
tx->descs[0].qw[1] |=
drivers/infiniband/hw/hfi1/sdma.h
530
_sdma_txreq_ahgadd(tx, num_ahg, ahg_entry, ahg, ahg_hlen);
drivers/infiniband/hw/hfi1/sdma.h
566
struct sdma_txreq *tx,
drivers/infiniband/hw/hfi1/sdma.h
571
return sdma_txinit_ahg(tx, flags, tlen, 0, 0, NULL, 0, cb);
drivers/infiniband/hw/hfi1/sdma.h
594
struct sdma_txreq *tx,
drivers/infiniband/hw/hfi1/sdma.h
602
struct sdma_desc *desc = &tx->descp[tx->num_desc];
drivers/infiniband/hw/hfi1/sdma.h
604
if (!tx->num_desc) {
drivers/infiniband/hw/hfi1/sdma.h
625
int ext_coal_sdma_tx_descs(struct hfi1_devdata *dd, struct sdma_txreq *tx,
drivers/infiniband/hw/hfi1/sdma.h
631
static inline void sdma_txclean(struct hfi1_devdata *dd, struct sdma_txreq *tx)
drivers/infiniband/hw/hfi1/sdma.h
633
if (tx->num_desc)
drivers/infiniband/hw/hfi1/sdma.h
634
__sdma_txclean(dd, tx);
drivers/infiniband/hw/hfi1/sdma.h
639
struct sdma_txreq *tx)
drivers/infiniband/hw/hfi1/sdma.h
641
u16 last_desc = tx->num_desc - 1;
drivers/infiniband/hw/hfi1/sdma.h
643
tx->descp[last_desc].qw[0] |= SDMA_DESC0_LAST_DESC_FLAG;
drivers/infiniband/hw/hfi1/sdma.h
644
tx->descp[last_desc].qw[1] |= dd->default_desc1;
drivers/infiniband/hw/hfi1/sdma.h
645
if (tx->flags & SDMA_TXREQ_F_URGENT)
drivers/infiniband/hw/hfi1/sdma.h
646
tx->descp[last_desc].qw[1] |= (SDMA_DESC1_HEAD_TO_HOST_FLAG |
drivers/infiniband/hw/hfi1/sdma.h
653
struct sdma_txreq *tx,
drivers/infiniband/hw/hfi1/sdma.h
663
tx,
drivers/infiniband/hw/hfi1/sdma.h
667
WARN_ON(len > tx->tlen);
drivers/infiniband/hw/hfi1/sdma.h
668
tx->num_desc++;
drivers/infiniband/hw/hfi1/sdma.h
669
tx->tlen -= len;
drivers/infiniband/hw/hfi1/sdma.h
671
if (!tx->tlen) {
drivers/infiniband/hw/hfi1/sdma.h
672
if (tx->packet_len & (sizeof(u32) - 1)) {
drivers/infiniband/hw/hfi1/sdma.h
673
rval = _pad_sdma_tx_descs(dd, tx);
drivers/infiniband/hw/hfi1/sdma.h
677
_sdma_close_tx(dd, tx);
drivers/infiniband/hw/hfi1/sdma.h
709
struct sdma_txreq *tx,
drivers/infiniband/hw/hfi1/sdma.h
720
if ((unlikely(tx->num_desc == tx->desc_limit))) {
drivers/infiniband/hw/hfi1/sdma.h
721
rval = ext_coal_sdma_tx_descs(dd, tx, SDMA_MAP_PAGE,
drivers/infiniband/hw/hfi1/sdma.h
735
__sdma_txclean(dd, tx);
drivers/infiniband/hw/hfi1/sdma.h
739
return _sdma_txadd_daddr(dd, SDMA_MAP_PAGE, tx, addr, len,
drivers/infiniband/hw/hfi1/sdma.h
761
struct sdma_txreq *tx,
drivers/infiniband/hw/hfi1/sdma.h
767
if ((unlikely(tx->num_desc == tx->desc_limit))) {
drivers/infiniband/hw/hfi1/sdma.h
768
rval = ext_coal_sdma_tx_descs(dd, tx, SDMA_MAP_NONE,
drivers/infiniband/hw/hfi1/sdma.h
774
return _sdma_txadd_daddr(dd, SDMA_MAP_NONE, tx, addr, len,
drivers/infiniband/hw/hfi1/sdma.h
796
struct sdma_txreq *tx,
drivers/infiniband/hw/hfi1/sdma.h
803
if ((unlikely(tx->num_desc == tx->desc_limit))) {
drivers/infiniband/hw/hfi1/sdma.h
804
rval = ext_coal_sdma_tx_descs(dd, tx, SDMA_MAP_SINGLE,
drivers/infiniband/hw/hfi1/sdma.h
817
__sdma_txclean(dd, tx);
drivers/infiniband/hw/hfi1/sdma.h
821
return _sdma_txadd_daddr(dd, SDMA_MAP_SINGLE, tx, addr, len,
drivers/infiniband/hw/hfi1/sdma.h
829
struct sdma_txreq *tx,
drivers/infiniband/hw/hfi1/sdma.h
880
struct sdma_txreq *tx)
drivers/infiniband/hw/hfi1/sdma.h
884
if (tx->num_desc > sde->desc_avail)
drivers/infiniband/hw/hfi1/sdma_txreq.h
92
static inline int sdma_txreq_built(struct sdma_txreq *tx)
drivers/infiniband/hw/hfi1/sdma_txreq.h
94
return tx->num_desc;
drivers/infiniband/hw/hfi1/trace_tx.h
1002
TP_PROTO(struct ipoib_txreq *tx, u32 idx),
drivers/infiniband/hw/hfi1/trace_tx.h
1003
TP_ARGS(tx, idx),
drivers/infiniband/hw/hfi1/trace_tx.h
1005
DD_DEV_ENTRY(tx->txq->priv->dd)
drivers/infiniband/hw/hfi1/trace_tx.h
1006
__field(struct ipoib_txreq *, tx)
drivers/infiniband/hw/hfi1/trace_tx.h
1012
DD_DEV_ASSIGN(tx->txq->priv->dd);
drivers/infiniband/hw/hfi1/trace_tx.h
1013
__entry->tx = tx;
drivers/infiniband/hw/hfi1/trace_tx.h
1014
__entry->skb = tx->skb;
drivers/infiniband/hw/hfi1/trace_tx.h
1015
__entry->txq = tx->txq;
drivers/infiniband/hw/hfi1/trace_tx.h
1021
(unsigned long long)__entry->tx,
drivers/infiniband/hw/hfi1/trace_tx.h
1031
TP_PROTO(struct ipoib_txreq *tx, u32 idx),
drivers/infiniband/hw/hfi1/trace_tx.h
1032
TP_ARGS(tx, idx)
drivers/infiniband/hw/hfi1/trace_tx.h
1037
TP_PROTO(struct ipoib_txreq *tx, u32 idx),
drivers/infiniband/hw/hfi1/trace_tx.h
1038
TP_ARGS(tx, idx)
drivers/infiniband/hw/hfi1/user_sdma.c
1004
if (unlikely(tx->flags & TXREQ_FLAGS_REQ_DISABLE_SH))
drivers/infiniband/hw/hfi1/user_sdma.c
1022
return sdma_txadd_kvaddr(pq->dd, &tx->txreq, hdr, sizeof(*hdr));
drivers/infiniband/hw/hfi1/user_sdma.c
1026
struct user_sdma_txreq *tx, u32 datalen)
drivers/infiniband/hw/hfi1/user_sdma.c
1056
if (unlikely(tx->flags & TXREQ_FLAGS_REQ_ACK))
drivers/infiniband/hw/hfi1/user_sdma.c
1112
if (unlikely(tx->flags & TXREQ_FLAGS_REQ_DISABLE_SH)) {
drivers/infiniband/hw/hfi1/user_sdma.c
1133
sdma_txinit_ahg(&tx->txreq,
drivers/infiniband/hw/hfi1/user_sdma.c
1154
struct user_sdma_txreq *tx =
drivers/infiniband/hw/hfi1/user_sdma.c
1161
if (!tx->req)
drivers/infiniband/hw/hfi1/user_sdma.c
1164
req = tx->req;
drivers/infiniband/hw/hfi1/user_sdma.c
1175
req->seqcomp = tx->seqnum;
drivers/infiniband/hw/hfi1/user_sdma.c
1176
kmem_cache_free(pq->txreq_cache, tx);
drivers/infiniband/hw/hfi1/user_sdma.c
1199
struct user_sdma_txreq *tx =
drivers/infiniband/hw/hfi1/user_sdma.c
1203
kmem_cache_free(req->pq->txreq_cache, tx);
drivers/infiniband/hw/hfi1/user_sdma.c
46
struct user_sdma_txreq *tx, u32 datalen);
drivers/infiniband/hw/hfi1/user_sdma.c
48
struct user_sdma_txreq *tx, u32 len);
drivers/infiniband/hw/hfi1/user_sdma.c
569
struct user_sdma_txreq *tx)
drivers/infiniband/hw/hfi1/user_sdma.c
590
(sizeof(tx->hdr) - 4));
drivers/infiniband/hw/hfi1/user_sdma.c
638
struct user_sdma_txreq *tx,
drivers/infiniband/hw/hfi1/user_sdma.c
654
memcpy(&tx->hdr, &req->hdr, sizeof(tx->hdr));
drivers/infiniband/hw/hfi1/user_sdma.c
657
tx->hdr.pbc[0] = cpu_to_le16(pbclen);
drivers/infiniband/hw/hfi1/user_sdma.c
659
ret = check_header_template(req, &tx->hdr, lrhlen, datalen);
drivers/infiniband/hw/hfi1/user_sdma.c
662
ret = sdma_txinit_ahg(&tx->txreq, SDMA_TXREQ_F_AHG_COPY,
drivers/infiniband/hw/hfi1/user_sdma.c
663
sizeof(tx->hdr) + datalen, req->ahg_idx,
drivers/infiniband/hw/hfi1/user_sdma.c
667
ret = sdma_txadd_kvaddr(pq->dd, &tx->txreq, &tx->hdr, sizeof(tx->hdr));
drivers/infiniband/hw/hfi1/user_sdma.c
669
sdma_txclean(pq->dd, &tx->txreq);
drivers/infiniband/hw/hfi1/user_sdma.c
678
struct user_sdma_txreq *tx = NULL;
drivers/infiniband/hw/hfi1/user_sdma.c
714
tx = kmem_cache_alloc(pq->txreq_cache, GFP_KERNEL);
drivers/infiniband/hw/hfi1/user_sdma.c
715
if (!tx)
drivers/infiniband/hw/hfi1/user_sdma.c
718
tx->flags = 0;
drivers/infiniband/hw/hfi1/user_sdma.c
719
tx->req = req;
drivers/infiniband/hw/hfi1/user_sdma.c
720
INIT_LIST_HEAD(&tx->list);
drivers/infiniband/hw/hfi1/user_sdma.c
727
tx->flags |= (TXREQ_FLAGS_REQ_ACK |
drivers/infiniband/hw/hfi1/user_sdma.c
746
datalen = compute_data_length(req, tx);
drivers/infiniband/hw/hfi1/user_sdma.c
762
tx->flags |= TXREQ_FLAGS_REQ_DISABLE_SH;
drivers/infiniband/hw/hfi1/user_sdma.c
768
ret = user_sdma_txadd_ahg(req, tx, datalen);
drivers/infiniband/hw/hfi1/user_sdma.c
774
changes = set_txreq_header_ahg(req, tx,
drivers/infiniband/hw/hfi1/user_sdma.c
782
ret = sdma_txinit(&tx->txreq, 0, sizeof(req->hdr) +
drivers/infiniband/hw/hfi1/user_sdma.c
792
ret = set_txreq_header(req, tx, datalen);
drivers/infiniband/hw/hfi1/user_sdma.c
802
ret = hfi1_add_pages_to_sdma_packet(req, tx, iovec,
drivers/infiniband/hw/hfi1/user_sdma.c
808
list_add_tail(&tx->txreq.list, &req->txps);
drivers/infiniband/hw/hfi1/user_sdma.c
814
tx->seqnum = req->seqnum++;
drivers/infiniband/hw/hfi1/user_sdma.c
835
sdma_txclean(pq->dd, &tx->txreq);
drivers/infiniband/hw/hfi1/user_sdma.c
837
kmem_cache_free(pq->txreq_cache, tx);
drivers/infiniband/hw/hfi1/user_sdma.c
912
struct user_sdma_txreq *tx, u32 datalen)
drivers/infiniband/hw/hfi1/user_sdma.c
915
struct hfi1_pkt_header *hdr = &tx->hdr;
drivers/infiniband/hw/hfi1/user_sdma.c
969
if (unlikely(tx->flags & TXREQ_FLAGS_REQ_ACK))
drivers/infiniband/hw/hfi1/verbs.c
1178
struct verbs_txreq *tx = ps->s_txreq;
drivers/infiniband/hw/hfi1/verbs.c
1191
(tx->s_cur_size + priv->s_running_pkt_size) / 2;
drivers/infiniband/hw/hfi1/verbs.c
1196
!sdma_txreq_built(&tx->txreq))
drivers/infiniband/hw/hfi1/verbs.c
589
struct verbs_txreq *tx =
drivers/infiniband/hw/hfi1/verbs.c
591
struct rvt_qp *qp = tx->qp;
drivers/infiniband/hw/hfi1/verbs.c
594
if (tx->wqe) {
drivers/infiniband/hw/hfi1/verbs.c
595
rvt_send_complete(qp, tx->wqe, IB_WC_SUCCESS);
drivers/infiniband/hw/hfi1/verbs.c
599
hdr = &tx->phdr.hdr;
drivers/infiniband/hw/hfi1/verbs.c
606
hfi1_put_txreq(tx);
drivers/infiniband/hw/hfi1/verbs.c
657
struct verbs_txreq *tx)
drivers/infiniband/hw/hfi1/verbs.c
659
struct rvt_sge_state *ss = tx->ss;
drivers/infiniband/hw/hfi1/verbs.c
671
&tx->txreq,
drivers/infiniband/hw/hfi1/verbs.c
721
struct verbs_txreq *tx,
drivers/infiniband/hw/hfi1/verbs.c
726
struct hfi1_sdma_header *phdr = &tx->phdr;
drivers/infiniband/hw/hfi1/verbs.c
727
u16 hdrbytes = (tx->hdr_dwords + sizeof(pbc) / 4) << 2;
drivers/infiniband/hw/hfi1/verbs.c
730
if (tx->phdr.hdr.hdr_type) {
drivers/infiniband/hw/hfi1/verbs.c
740
&tx->txreq,
drivers/infiniband/hw/hfi1/verbs.c
754
&tx->txreq,
drivers/infiniband/hw/hfi1/verbs.c
761
&tx->txreq,
drivers/infiniband/hw/hfi1/verbs.c
773
if (tx->ss) {
drivers/infiniband/hw/hfi1/verbs.c
774
ret = build_verbs_ulp_payload(sde, length, tx);
drivers/infiniband/hw/hfi1/verbs.c
781
ret = sdma_txadd_daddr(sde->dd, &tx->txreq, sde->dd->sdma_pad_phys,
drivers/infiniband/hw/hfi1/verbs.c
807
struct verbs_txreq *tx;
drivers/infiniband/hw/hfi1/verbs.c
822
tx = ps->s_txreq;
drivers/infiniband/hw/hfi1/verbs.c
823
if (!sdma_txreq_built(&tx->txreq)) {
drivers/infiniband/hw/hfi1/verbs.c
847
tx->wqe = qp->s_wqe;
drivers/infiniband/hw/hfi1/verbs.c
848
ret = build_verbs_tx_desc(tx->sde, len, tx, ahg_info, pbc);
drivers/infiniband/hw/hfi1/verbs.c
852
ret = sdma_send_txreq(tx->sde, ps->wait, &tx->txreq, ps->pkts_sent);
drivers/infiniband/hw/hfi1/verbs.h
359
void hfi1_put_txreq(struct verbs_txreq *tx);
drivers/infiniband/hw/hfi1/verbs_txreq.c
13
void hfi1_put_txreq(struct verbs_txreq *tx)
drivers/infiniband/hw/hfi1/verbs_txreq.c
21
qp = tx->qp;
drivers/infiniband/hw/hfi1/verbs_txreq.c
24
if (tx->mr)
drivers/infiniband/hw/hfi1/verbs_txreq.c
25
rvt_put_mr(tx->mr);
drivers/infiniband/hw/hfi1/verbs_txreq.c
27
sdma_txclean(dd_from_dev(dev), &tx->txreq);
drivers/infiniband/hw/hfi1/verbs_txreq.c
30
kmem_cache_free(dev->verbs_txreq_cache, tx);
drivers/infiniband/hw/hfi1/verbs_txreq.c
55
struct verbs_txreq *tx = NULL;
drivers/infiniband/hw/hfi1/verbs_txreq.c
61
tx = kmem_cache_alloc(dev->verbs_txreq_cache, VERBS_TXREQ_GFP);
drivers/infiniband/hw/hfi1/verbs_txreq.c
62
if (tx)
drivers/infiniband/hw/hfi1/verbs_txreq.c
77
return tx;
drivers/infiniband/hw/hfi1/verbs_txreq.h
38
struct verbs_txreq *tx;
drivers/infiniband/hw/hfi1/verbs_txreq.h
41
tx = kmem_cache_alloc(dev->verbs_txreq_cache, VERBS_TXREQ_GFP);
drivers/infiniband/hw/hfi1/verbs_txreq.h
42
if (unlikely(!tx)) {
drivers/infiniband/hw/hfi1/verbs_txreq.h
44
tx = __get_txreq(dev, qp);
drivers/infiniband/hw/hfi1/verbs_txreq.h
45
if (!tx)
drivers/infiniband/hw/hfi1/verbs_txreq.h
46
return tx;
drivers/infiniband/hw/hfi1/verbs_txreq.h
48
tx->qp = qp;
drivers/infiniband/hw/hfi1/verbs_txreq.h
49
tx->mr = NULL;
drivers/infiniband/hw/hfi1/verbs_txreq.h
50
tx->sde = priv->s_sde;
drivers/infiniband/hw/hfi1/verbs_txreq.h
51
tx->psc = priv->s_sendcontext;
drivers/infiniband/hw/hfi1/verbs_txreq.h
53
tx->txreq.num_desc = 0;
drivers/infiniband/hw/hfi1/verbs_txreq.h
55
tx->phdr.hdr.hdr_type = priv->hdr_type;
drivers/infiniband/hw/hfi1/verbs_txreq.h
56
tx->txreq.flags = 0;
drivers/infiniband/hw/hfi1/verbs_txreq.h
57
return tx;
drivers/infiniband/hw/hfi1/verbs_txreq.h
75
void hfi1_put_txreq(struct verbs_txreq *tx);
drivers/infiniband/hw/hfi1/vnic_sdma.c
105
tx->pbc_val = cpu_to_le64(pbc);
drivers/infiniband/hw/hfi1/vnic_sdma.c
108
&tx->txreq,
drivers/infiniband/hw/hfi1/vnic_sdma.c
109
&tx->pbc_val,
drivers/infiniband/hw/hfi1/vnic_sdma.c
115
ret = build_vnic_ulp_payload(sde, tx);
drivers/infiniband/hw/hfi1/vnic_sdma.c
132
struct vnic_txreq *tx;
drivers/infiniband/hw/hfi1/vnic_sdma.c
141
tx = kmem_cache_alloc(dd->vnic.txreq_cache, GFP_ATOMIC);
drivers/infiniband/hw/hfi1/vnic_sdma.c
142
if (unlikely(!tx)) {
drivers/infiniband/hw/hfi1/vnic_sdma.c
147
tx->sdma = vnic_sdma;
drivers/infiniband/hw/hfi1/vnic_sdma.c
148
tx->skb = skb;
drivers/infiniband/hw/hfi1/vnic_sdma.c
149
hfi1_vnic_update_pad(tx->pad, plen);
drivers/infiniband/hw/hfi1/vnic_sdma.c
150
tx->plen = plen;
drivers/infiniband/hw/hfi1/vnic_sdma.c
151
ret = build_vnic_tx_desc(sde, tx, pbc);
drivers/infiniband/hw/hfi1/vnic_sdma.c
156
&tx->txreq, vnic_sdma->pkts_sent);
drivers/infiniband/hw/hfi1/vnic_sdma.c
168
sdma_txclean(dd, &tx->txreq);
drivers/infiniband/hw/hfi1/vnic_sdma.c
169
kmem_cache_free(dd->vnic.txreq_cache, tx);
drivers/infiniband/hw/hfi1/vnic_sdma.c
41
struct vnic_txreq *tx = container_of(txreq, struct vnic_txreq, txreq);
drivers/infiniband/hw/hfi1/vnic_sdma.c
42
struct hfi1_vnic_sdma *vnic_sdma = tx->sdma;
drivers/infiniband/hw/hfi1/vnic_sdma.c
45
dev_kfree_skb_any(tx->skb);
drivers/infiniband/hw/hfi1/vnic_sdma.c
46
kmem_cache_free(vnic_sdma->dd->vnic.txreq_cache, tx);
drivers/infiniband/hw/hfi1/vnic_sdma.c
50
struct vnic_txreq *tx)
drivers/infiniband/hw/hfi1/vnic_sdma.c
56
&tx->txreq,
drivers/infiniband/hw/hfi1/vnic_sdma.c
57
tx->skb->data,
drivers/infiniband/hw/hfi1/vnic_sdma.c
58
skb_headlen(tx->skb));
drivers/infiniband/hw/hfi1/vnic_sdma.c
62
for (i = 0; i < skb_shinfo(tx->skb)->nr_frags; i++) {
drivers/infiniband/hw/hfi1/vnic_sdma.c
63
skb_frag_t *frag = &skb_shinfo(tx->skb)->frags[i];
drivers/infiniband/hw/hfi1/vnic_sdma.c
67
&tx->txreq,
drivers/infiniband/hw/hfi1/vnic_sdma.c
76
if (tx->plen)
drivers/infiniband/hw/hfi1/vnic_sdma.c
77
ret = sdma_txadd_kvaddr(sde->dd, &tx->txreq,
drivers/infiniband/hw/hfi1/vnic_sdma.c
78
tx->pad + HFI1_VNIC_MAX_PAD - tx->plen,
drivers/infiniband/hw/hfi1/vnic_sdma.c
79
tx->plen);
drivers/infiniband/hw/hfi1/vnic_sdma.c
86
struct vnic_txreq *tx,
drivers/infiniband/hw/hfi1/vnic_sdma.c
93
&tx->txreq,
drivers/infiniband/hw/hfi1/vnic_sdma.c
95
hdrbytes + tx->skb->len + tx->plen,
drivers/infiniband/hw/vmw_pvrdma/pvrdma_ring.h
59
struct pvrdma_ring tx; /* Tx ring. */
drivers/infiniband/sw/siw/siw.h
471
#define tx_qp(tx) container_of(tx, struct siw_qp, tx_ctx)
drivers/infiniband/ulp/ipoib/ipoib.h
646
static inline void ipoib_cm_set(struct ipoib_neigh *neigh, struct ipoib_cm_tx *tx)
drivers/infiniband/ulp/ipoib/ipoib.h
648
neigh->cm = tx;
drivers/infiniband/ulp/ipoib/ipoib.h
663
void ipoib_cm_send(struct net_device *dev, struct sk_buff *skb, struct ipoib_cm_tx *tx);
drivers/infiniband/ulp/ipoib/ipoib.h
671
void ipoib_cm_destroy_tx(struct ipoib_cm_tx *tx);
drivers/infiniband/ulp/ipoib/ipoib.h
701
static inline void ipoib_cm_set(struct ipoib_neigh *neigh, struct ipoib_cm_tx *tx)
drivers/infiniband/ulp/ipoib/ipoib.h
716
void ipoib_cm_send(struct net_device *dev, struct sk_buff *skb, struct ipoib_cm_tx *tx)
drivers/infiniband/ulp/ipoib/ipoib.h
753
void ipoib_cm_destroy_tx(struct ipoib_cm_tx *tx)
drivers/infiniband/ulp/ipoib/ipoib_cm.c
1057
static struct ib_qp *ipoib_cm_create_tx_qp(struct net_device *dev, struct ipoib_cm_tx *tx)
drivers/infiniband/ulp/ipoib/ipoib_cm.c
1068
.qp_context = tx,
drivers/infiniband/ulp/ipoib/ipoib_cm.c
1078
tx->max_send_sge = attr.cap.max_send_sge;
drivers/infiniband/ulp/ipoib/ipoib_cm.c
1252
struct ipoib_cm_tx *tx = cm_id->context;
drivers/infiniband/ulp/ipoib/ipoib_cm.c
1253
struct ipoib_dev_priv *priv = ipoib_priv(tx->dev);
drivers/infiniband/ulp/ipoib/ipoib_cm.c
1277
neigh = tx->neigh;
drivers/infiniband/ulp/ipoib/ipoib_cm.c
1283
tx->neigh = NULL;
drivers/infiniband/ulp/ipoib/ipoib_cm.c
1286
if (test_and_clear_bit(IPOIB_FLAG_INITIALIZED, &tx->flags)) {
drivers/infiniband/ulp/ipoib/ipoib_cm.c
1287
list_move(&tx->list, &priv->cm.reap_list);
drivers/infiniband/ulp/ipoib/ipoib_cm.c
1305
struct ipoib_cm_tx *tx;
drivers/infiniband/ulp/ipoib/ipoib_cm.c
1307
tx = kzalloc_obj(*tx, GFP_ATOMIC);
drivers/infiniband/ulp/ipoib/ipoib_cm.c
1308
if (!tx)
drivers/infiniband/ulp/ipoib/ipoib_cm.c
1311
neigh->cm = tx;
drivers/infiniband/ulp/ipoib/ipoib_cm.c
1312
tx->neigh = neigh;
drivers/infiniband/ulp/ipoib/ipoib_cm.c
1313
tx->dev = dev;
drivers/infiniband/ulp/ipoib/ipoib_cm.c
1314
list_add(&tx->list, &priv->cm.start_list);
drivers/infiniband/ulp/ipoib/ipoib_cm.c
1315
set_bit(IPOIB_FLAG_INITIALIZED, &tx->flags);
drivers/infiniband/ulp/ipoib/ipoib_cm.c
1317
return tx;
drivers/infiniband/ulp/ipoib/ipoib_cm.c
1320
void ipoib_cm_destroy_tx(struct ipoib_cm_tx *tx)
drivers/infiniband/ulp/ipoib/ipoib_cm.c
1322
struct ipoib_dev_priv *priv = ipoib_priv(tx->dev);
drivers/infiniband/ulp/ipoib/ipoib_cm.c
1324
if (test_and_clear_bit(IPOIB_FLAG_INITIALIZED, &tx->flags)) {
drivers/infiniband/ulp/ipoib/ipoib_cm.c
1326
list_move(&tx->list, &priv->cm.reap_list);
drivers/infiniband/ulp/ipoib/ipoib_cm.c
1329
tx->neigh->daddr + 4);
drivers/infiniband/ulp/ipoib/ipoib_cm.c
1330
tx->neigh = NULL;
drivers/infiniband/ulp/ipoib/ipoib_cm.c
697
struct ipoib_cm_tx *tx,
drivers/infiniband/ulp/ipoib/ipoib_cm.c
705
return ib_post_send(tx->qp, &priv->tx_wr.wr, NULL);
drivers/infiniband/ulp/ipoib/ipoib_cm.c
708
void ipoib_cm_send(struct net_device *dev, struct sk_buff *skb, struct ipoib_cm_tx *tx)
drivers/infiniband/ulp/ipoib/ipoib_cm.c
713
unsigned int usable_sge = tx->max_send_sge - !!skb_headlen(skb);
drivers/infiniband/ulp/ipoib/ipoib_cm.c
715
if (unlikely(skb->len > tx->mtu)) {
drivers/infiniband/ulp/ipoib/ipoib_cm.c
717
skb->len, tx->mtu);
drivers/infiniband/ulp/ipoib/ipoib_cm.c
720
ipoib_cm_skb_too_long(dev, skb, tx->mtu - IPOIB_ENCAP_LEN);
drivers/infiniband/ulp/ipoib/ipoib_cm.c
741
tx->tx_head, skb->len, tx->qp->qp_num);
drivers/infiniband/ulp/ipoib/ipoib_cm.c
750
tx_req = &tx->tx_ring[tx->tx_head & (ipoib_sendq_size - 1)];
drivers/infiniband/ulp/ipoib/ipoib_cm.c
762
tx->qp->qp_num);
drivers/infiniband/ulp/ipoib/ipoib_cm.c
778
rc = post_send(priv, tx, tx->tx_head & (ipoib_sendq_size - 1), tx_req);
drivers/infiniband/ulp/ipoib/ipoib_cm.c
789
++tx->tx_head;
drivers/infiniband/ulp/ipoib/ipoib_cm.c
797
struct ipoib_cm_tx *tx = wc->qp->qp_context;
drivers/infiniband/ulp/ipoib/ipoib_cm.c
811
tx_req = &tx->tx_ring[wr_id];
drivers/infiniband/ulp/ipoib/ipoib_cm.c
823
++tx->tx_tail;
drivers/infiniband/ulp/ipoib/ipoib_cm.c
850
neigh = tx->neigh;
drivers/infiniband/ulp/ipoib/ipoib_cm.c
856
tx->neigh = NULL;
drivers/infiniband/ulp/ipoib/ipoib_cm.c
859
if (test_and_clear_bit(IPOIB_FLAG_INITIALIZED, &tx->flags)) {
drivers/infiniband/ulp/ipoib/ipoib_cm.c
860
list_move(&tx->list, &priv->cm.reap_list);
drivers/infiniband/ulp/ipoib/ipoib_cm.c
864
clear_bit(IPOIB_FLAG_OPER_UP, &tx->flags);
drivers/input/gameport/gameport.c
100
tx = t;
drivers/input/gameport/gameport.c
105
if (tx)
drivers/input/gameport/gameport.c
106
t /= tx;
drivers/input/gameport/gameport.c
114
unsigned int i, t, t1, t2, t3, tx;
drivers/input/gameport/gameport.c
120
tx = 1 << 30;
drivers/input/gameport/gameport.c
130
if ((t = DELTA(t2,t1) - DELTA(t3,t2)) < tx) tx = t;
drivers/input/gameport/gameport.c
134
return 59659 / (tx < 1 ? 1 : tx);
drivers/input/gameport/gameport.c
139
unsigned long tx, t1, t2, flags;
drivers/input/gameport/gameport.c
144
tx = 1 << 30;
drivers/input/gameport/gameport.c
153
if (t2 - t1 < tx) tx = t2 - t1;
drivers/input/gameport/gameport.c
158
(unsigned long)HZ / (1000 / 50)) / (tx < 1 ? 1 : tx);
drivers/input/gameport/gameport.c
80
unsigned int i, t, tx;
drivers/input/gameport/gameport.c
87
tx = ~0;
drivers/input/gameport/gameport.c
99
if (t < tx)
drivers/input/joystick/analog.c
310
unsigned int i, t, tx;
drivers/input/joystick/analog.c
314
tx = ~0;
drivers/input/joystick/analog.c
327
if (t < tx) tx = t;
drivers/input/joystick/analog.c
330
port->loop = tx / 50;
drivers/input/rmi4/rmi_f54.c
223
u8 tx = drv_data->num_tx_electrodes ? : f54->num_tx_electrodes;
drivers/input/rmi4/rmi_f54.c
228
size = rx * tx;
drivers/input/rmi4/rmi_f54.c
235
size = sizeof(u16) * rx * tx;
drivers/input/rmi4/rmi_f54.c
420
u8 tx = drv_data->num_tx_electrodes ? : f54->num_tx_electrodes;
drivers/input/rmi4/rmi_f54.c
436
f->height = tx;
drivers/input/rmi4/rmi_f54.c
663
u8 rx, tx;
drivers/input/rmi4/rmi_f54.c
680
tx = f54->num_tx_electrodes;
drivers/input/rmi4/rmi_f54.c
682
array3_size(tx, rx, sizeof(u16)),
drivers/input/serio/maceps2.c
62
port->tx = val;
drivers/input/serio/ps2-gpio.c
101
drvdata->tx.t_xfer_end = 0;
drivers/input/serio/ps2-gpio.c
111
flush_delayed_work(&drvdata->tx.work);
drivers/input/serio/ps2-gpio.c
123
drvdata->tx.byte = val;
drivers/input/serio/ps2-gpio.c
125
schedule_delayed_work(&drvdata->tx.work, usecs_to_jiffies(200));
drivers/input/serio/ps2-gpio.c
136
guard(mutex)(&drvdata->tx.mutex);
drivers/input/serio/ps2-gpio.c
139
if (!wait_for_completion_timeout(&drvdata->tx.complete,
drivers/input/serio/ps2-gpio.c
154
tx.work);
drivers/input/serio/ps2-gpio.c
156
drvdata->tx.t_xfer_start = ktime_get();
drivers/input/serio/ps2-gpio.c
178
us_delta = ktime_us_delta(drvdata->t_irq_now, drvdata->tx.t_xfer_end);
drivers/input/serio/ps2-gpio.c
276
cnt = drvdata->tx.cnt;
drivers/input/serio/ps2-gpio.c
277
byte = drvdata->tx.byte;
drivers/input/serio/ps2-gpio.c
288
drvdata->tx.t_xfer_start);
drivers/input/serio/ps2-gpio.c
336
drvdata->tx.t_xfer_end = ktime_get();
drivers/input/serio/ps2-gpio.c
338
complete(&drvdata->tx.complete);
drivers/input/serio/ps2-gpio.c
358
__ps2_gpio_write(drvdata->serio, drvdata->tx.byte);
drivers/input/serio/ps2-gpio.c
360
drvdata->tx.cnt = cnt;
drivers/input/serio/ps2-gpio.c
461
drvdata->tx.cnt = 1;
drivers/input/serio/ps2-gpio.c
463
INIT_DELAYED_WORK(&drvdata->tx.work, ps2_gpio_tx_work_fn);
drivers/input/serio/ps2-gpio.c
464
init_completion(&drvdata->tx.complete);
drivers/input/serio/ps2-gpio.c
465
mutex_init(&drvdata->tx.mutex);
drivers/input/serio/ps2-gpio.c
93
} tx;
drivers/input/touchscreen/ads7846.c
1087
size += sizeof(*packet->tx) * max_count;
drivers/input/touchscreen/ads7846.c
1090
packet->tx = devm_kzalloc(&ts->spi->dev, size, GFP_KERNEL);
drivers/input/touchscreen/ads7846.c
1091
if (!packet->tx)
drivers/input/touchscreen/ads7846.c
1122
packet->tx[l->offset + b].cmd = cmd;
drivers/input/touchscreen/ads7846.c
1125
x->tx_buf = packet->tx;
drivers/input/touchscreen/ads7846.c
86
struct ads7846_buf *tx;
drivers/input/touchscreen/tsc2004.c
23
u8 tx = TSC200X_CMD | TSC200X_CMD_12BIT | cmd;
drivers/input/touchscreen/tsc2004.c
27
data = i2c_smbus_write_byte(i2c, tx);
drivers/input/touchscreen/tsc2005.c
26
u8 tx = TSC200X_CMD | TSC200X_CMD_12BIT | cmd;
drivers/input/touchscreen/tsc2005.c
28
.tx_buf = &tx,
drivers/ipack/devices/ipoctal.c
117
stats->tx = 0;
drivers/ipack/devices/ipoctal.c
151
icount->tx = channel->stats.tx;
drivers/ipack/devices/ipoctal.c
212
channel->stats.tx++;
drivers/ipack/devices/ipoctal.h
28
unsigned long tx;
drivers/irqchip/irq-riscv-rpmi-sysmsi.c
103
struct rpmi_sysmsi_set_msi_target_tx tx;
drivers/irqchip/irq-riscv-rpmi-sysmsi.c
108
tx.sys_msi_index = cpu_to_le32(sys_msi_index);
drivers/irqchip/irq-riscv-rpmi-sysmsi.c
109
tx.sys_msi_address_low = cpu_to_le32(m->address_lo);
drivers/irqchip/irq-riscv-rpmi-sysmsi.c
110
tx.sys_msi_address_high = cpu_to_le32(m->address_hi);
drivers/irqchip/irq-riscv-rpmi-sysmsi.c
111
tx.sys_msi_data = cpu_to_le32(m->data);
drivers/irqchip/irq-riscv-rpmi-sysmsi.c
113
&tx, sizeof(tx), &rx, sizeof(rx));
drivers/irqchip/irq-riscv-rpmi-sysmsi.c
82
struct rpmi_sysmsi_set_msi_state_tx tx;
drivers/irqchip/irq-riscv-rpmi-sysmsi.c
87
tx.sys_msi_index = cpu_to_le32(sys_msi_index);
drivers/irqchip/irq-riscv-rpmi-sysmsi.c
88
tx.sys_msi_state = cpu_to_le32(sys_msi_state);
drivers/irqchip/irq-riscv-rpmi-sysmsi.c
90
&tx, sizeof(tx), &rx, sizeof(rx));
drivers/leds/trigger/ledtrig-netdev.c
438
DEFINE_NETDEV_TRIGGER(tx, TRIGGER_NETDEV_TX);
drivers/leds/trigger/ledtrig-tty.c
189
DEFINE_TTY_TRIGGER(tx, TRIGGER_TTY_TX);
drivers/leds/trigger/ledtrig-tty.c
19
int rx, tx;
drivers/leds/trigger/ledtrig-tty.c
264
if (trigger_data->mode_tx && (icount.tx != trigger_data->tx)) {
drivers/leds/trigger/ledtrig-tty.c
265
trigger_data->tx = icount.tx;
drivers/mailbox/bcm-pdc-mailbox.c
1003
dev_dbg(dev, " - base DMA addr of tx ring %pad", &tx.dmabase);
drivers/mailbox/bcm-pdc-mailbox.c
1004
dev_dbg(dev, " - base virtual addr of tx ring %p", tx.vbase);
drivers/mailbox/bcm-pdc-mailbox.c
1008
memcpy(&pdcs->tx_ring_alloc, &tx, sizeof(tx));
drivers/mailbox/bcm-pdc-mailbox.c
1074
dma_pool_free(pdcs->ring_pool, tx.vbase, tx.dmabase);
drivers/mailbox/bcm-pdc-mailbox.c
986
struct pdc_ring_alloc tx;
drivers/mailbox/bcm-pdc-mailbox.c
990
tx.vbase = dma_pool_zalloc(pdcs->ring_pool, GFP_KERNEL, &tx.dmabase);
drivers/mailbox/bcm-pdc-mailbox.c
991
if (unlikely(!tx.vbase)) {
drivers/mailbox/bcm74110-mailbox.c
120
BCM74110_OFFSET_IO_WRITEL_MACRO(tx, BCM_MBOX_BASE(mbox->tx_chan));
drivers/mailbox/bcm74110-mailbox.c
129
BCM74110_OFFSET_IO_READL_MACRO(tx, BCM_MBOX_BASE(mbox->tx_chan));
drivers/mailbox/imx-mailbox.c
1008
.tx = imx_mu_specific_tx,
drivers/mailbox/imx-mailbox.c
1019
.tx = imx_mu_specific_tx,
drivers/mailbox/imx-mailbox.c
1030
.tx = imx_mu_specific_tx,
drivers/mailbox/imx-mailbox.c
1041
.tx = imx_mu_seco_tx,
drivers/mailbox/imx-mailbox.c
116
int (*tx)(struct imx_mu_priv *priv, struct imx_mu_con_priv *cp, void *data);
drivers/mailbox/imx-mailbox.c
576
return priv->dcfg->tx(priv, cp, data);
drivers/mailbox/imx-mailbox.c
973
.tx = imx_mu_generic_tx,
drivers/mailbox/imx-mailbox.c
984
.tx = imx_mu_generic_tx,
drivers/mailbox/imx-mailbox.c
996
.tx = imx_mu_generic_tx,
drivers/mailbox/riscv-sbi-mpxy-mbox.c
238
void *tx, unsigned long tx_len,
drivers/mailbox/riscv-sbi-mpxy-mbox.c
248
if (!tx && tx_len)
drivers/mailbox/riscv-sbi-mpxy-mbox.c
255
memcpy(mpxy->shmem, tx, tx_len);
drivers/mailbox/riscv-sbi-mpxy-mbox.c
276
void *tx, unsigned long tx_len)
drivers/mailbox/riscv-sbi-mpxy-mbox.c
283
if (!tx && tx_len)
drivers/mailbox/riscv-sbi-mpxy-mbox.c
290
memcpy(mpxy->shmem, tx, tx_len);
drivers/md/raid5-log.h
36
struct dma_async_tx_descriptor *tx);
drivers/md/raid5-ppl.c
159
struct dma_async_tx_descriptor *tx)
drivers/md/raid5-ppl.c
189
return tx;
drivers/md/raid5-ppl.c
192
init_async_submit(&submit, ASYNC_TX_FENCE|ASYNC_TX_XOR_ZERO_DST, tx,
drivers/md/raid5-ppl.c
196
tx = async_memcpy(sh->ppl_page, srcs[0], 0, 0, PAGE_SIZE,
drivers/md/raid5-ppl.c
199
tx = async_xor(sh->ppl_page, srcs, 0, count, PAGE_SIZE,
drivers/md/raid5-ppl.c
202
return tx;
drivers/md/raid5-ppl.c
708
struct dma_async_tx_descriptor *tx;
drivers/md/raid5-ppl.c
713
tx = async_xor(page1, xor_srcs, 0, 2, size, &submit);
drivers/md/raid5-ppl.c
715
async_tx_quiesce(&tx);
drivers/md/raid5.c
1363
unsigned int poff, sector_t sector, struct dma_async_tx_descriptor *tx,
drivers/md/raid5.c
1381
init_async_submit(&submit, flags, tx, NULL, NULL, NULL);
drivers/md/raid5.c
1409
tx = async_memcpy(*page, bio_page, page_offset + poff,
drivers/md/raid5.c
1412
tx = async_memcpy(bio_page, *page, b_offset,
drivers/md/raid5.c
1416
submit.depend_tx = tx;
drivers/md/raid5.c
1423
return tx;
drivers/md/raid5.c
1466
struct dma_async_tx_descriptor *tx = NULL;
drivers/md/raid5.c
1485
tx = async_copy_data(0, rbi, &dev->page,
drivers/md/raid5.c
1487
dev->sector, tx, sh, 0);
drivers/md/raid5.c
1494
init_async_submit(&submit, ASYNC_TX_ACK, tx, ops_complete_biofill, sh, NULL);
drivers/md/raid5.c
1562
struct dma_async_tx_descriptor *tx;
drivers/md/raid5.c
1584
tx = async_memcpy(xor_dest, xor_srcs[0], off_dest, off_srcs[0],
drivers/md/raid5.c
1587
tx = async_xor_offs(xor_dest, off_dest, xor_srcs, off_srcs, count,
drivers/md/raid5.c
1590
return tx;
drivers/md/raid5.c
1656
struct dma_async_tx_descriptor *tx;
drivers/md/raid5.c
1690
tx = async_gen_syndrome(blocks, offs, count+2,
drivers/md/raid5.c
1705
tx = async_xor_offs(dest, dest_off, blocks, offs, count,
drivers/md/raid5.c
1709
return tx;
drivers/md/raid5.c
1723
struct dma_async_tx_descriptor *tx;
drivers/md/raid5.c
1800
tx = async_xor_offs(dest, dest_off, blocks, offs, count,
drivers/md/raid5.c
1805
init_async_submit(&submit, ASYNC_TX_FENCE, tx,
drivers/md/raid5.c
1849
struct dma_async_tx_descriptor *tx)
drivers/md/raid5.c
1881
init_async_submit(&submit, ASYNC_TX_FENCE|ASYNC_TX_XOR_DROP_DST, tx,
drivers/md/raid5.c
1883
tx = async_xor_offs(xor_dest, off_dest, xor_srcs, off_srcs, count,
drivers/md/raid5.c
1886
return tx;
drivers/md/raid5.c
1891
struct dma_async_tx_descriptor *tx)
drivers/md/raid5.c
1903
init_async_submit(&submit, ASYNC_TX_FENCE|ASYNC_TX_PQ_XOR_DST, tx,
drivers/md/raid5.c
1905
tx = async_gen_syndrome(blocks, offs, count+2,
drivers/md/raid5.c
1908
return tx;
drivers/md/raid5.c
1912
ops_run_biodrain(struct stripe_head *sh, struct dma_async_tx_descriptor *tx)
drivers/md/raid5.c
1955
tx = async_copy_data(1, wbi, &dev->page,
drivers/md/raid5.c
1957
dev->sector, tx, sh,
drivers/md/raid5.c
1980
return tx;
drivers/md/raid5.c
2032
struct dma_async_tx_descriptor *tx)
drivers/md/raid5.c
2106
init_async_submit(&submit, flags, tx, ops_complete_reconstruct, head_sh,
drivers/md/raid5.c
2110
init_async_submit(&submit, flags, tx, NULL, NULL,
drivers/md/raid5.c
2115
tx = async_memcpy(xor_dest, xor_srcs[0], off_dest, off_srcs[0],
drivers/md/raid5.c
2118
tx = async_xor_offs(xor_dest, off_dest, xor_srcs, off_srcs, count,
drivers/md/raid5.c
2130
struct dma_async_tx_descriptor *tx)
drivers/md/raid5.c
2176
init_async_submit(&submit, txflags, tx, ops_complete_reconstruct,
drivers/md/raid5.c
2179
init_async_submit(&submit, 0, tx, NULL, NULL,
drivers/md/raid5.c
2181
tx = async_gen_syndrome(blocks, offs, count+2,
drivers/md/raid5.c
2212
struct dma_async_tx_descriptor *tx;
drivers/md/raid5.c
2235
tx = async_xor_val_offs(xor_dest, off_dest, xor_srcs, off_srcs, count,
drivers/md/raid5.c
2240
init_async_submit(&submit, ASYNC_TX_ACK, tx, ops_complete_check, sh, NULL);
drivers/md/raid5.c
2241
tx = async_trigger_callback(&submit);
drivers/md/raid5.c
2270
struct dma_async_tx_descriptor *tx = NULL;
drivers/md/raid5.c
2284
tx = ops_run_compute5(sh, percpu);
drivers/md/raid5.c
2287
tx = ops_run_compute6_1(sh, percpu);
drivers/md/raid5.c
2289
tx = ops_run_compute6_2(sh, percpu);
drivers/md/raid5.c
2292
if (tx && !test_bit(STRIPE_OP_RECONSTRUCT, &ops_request))
drivers/md/raid5.c
2293
async_tx_ack(tx);
drivers/md/raid5.c
2298
tx = ops_run_prexor5(sh, percpu, tx);
drivers/md/raid5.c
2300
tx = ops_run_prexor6(sh, percpu, tx);
drivers/md/raid5.c
2304
tx = ops_run_partial_parity(sh, percpu, tx);
drivers/md/raid5.c
2307
tx = ops_run_biodrain(sh, tx);
drivers/md/raid5.c
2313
ops_run_reconstruct5(sh, percpu, tx);
drivers/md/raid5.c
2315
ops_run_reconstruct6(sh, percpu, tx);
drivers/md/raid5.c
4545
struct dma_async_tx_descriptor *tx = NULL;
drivers/md/raid5.c
4573
init_async_submit(&submit, 0, tx, NULL, NULL, NULL);
drivers/md/raid5.c
4574
tx = async_memcpy(sh2->dev[dd_idx].page,
drivers/md/raid5.c
4594
async_tx_quiesce(&tx);
drivers/media/cec/platform/s5p/s5p_cec.c
102
cec->tx = STATE_DONE;
drivers/media/cec/platform/s5p/s5p_cec.c
137
switch (cec->tx) {
drivers/media/cec/platform/s5p/s5p_cec.c
140
cec->tx = STATE_IDLE;
drivers/media/cec/platform/s5p/s5p_cec.c
146
cec->tx = STATE_IDLE;
drivers/media/cec/platform/s5p/s5p_cec.c
152
cec->tx = STATE_IDLE;
drivers/media/cec/platform/s5p/s5p_cec.c
96
cec->tx = STATE_NACK;
drivers/media/cec/platform/s5p/s5p_cec.c
99
cec->tx = STATE_ERROR;
drivers/media/cec/platform/s5p/s5p_cec.h
72
enum cec_state tx;
drivers/media/dvb-frontends/cxd2880/cxd2880_spi_device.c
19
struct spi_transfer tx;
drivers/media/dvb-frontends/cxd2880/cxd2880_spi_device.c
27
memset(&tx, 0, sizeof(tx));
drivers/media/dvb-frontends/cxd2880/cxd2880_spi_device.c
28
tx.tx_buf = data;
drivers/media/dvb-frontends/cxd2880/cxd2880_spi_device.c
29
tx.len = size;
drivers/media/dvb-frontends/cxd2880/cxd2880_spi_device.c
32
spi_message_add_tail(&tx, &msg);
drivers/media/dvb-frontends/dib7000p.c
2009
u8 *tx, *rx;
drivers/media/dvb-frontends/dib7000p.c
2016
tx = kzalloc(2, GFP_KERNEL);
drivers/media/dvb-frontends/dib7000p.c
2017
if (!tx)
drivers/media/dvb-frontends/dib7000p.c
2025
msg[0].buf = tx;
drivers/media/dvb-frontends/dib7000p.c
2028
tx[0] = 0x03;
drivers/media/dvb-frontends/dib7000p.c
2029
tx[1] = 0x00;
drivers/media/dvb-frontends/dib7000p.c
2052
kfree(tx);
drivers/media/dvb-frontends/dib9000.c
1012
static int dib9000_risc_apb_access_read(struct dib9000_state *state, u32 address, u16 attribute, const u8 * tx, u32 txlen, u8 * b, u32 len)
drivers/media/dvb-frontends/dib9000.c
208
static int dib9000_risc_apb_access_read(struct dib9000_state *state, u32 address, u16 attribute, const u8 * tx, u32 txlen, u8 * b, u32 len);
drivers/media/dvb-frontends/dibx000_common.c
263
static int dibx000_i2c_gate_ctrl(struct dibx000_i2c_master *mst, u8 tx[4],
drivers/media/dvb-frontends/dibx000_common.c
277
tx[0] = (((mst->base_reg + 1) >> 8) & 0xff);
drivers/media/dvb-frontends/dibx000_common.c
278
tx[1] = ((mst->base_reg + 1) & 0xff);
drivers/media/dvb-frontends/dibx000_common.c
279
tx[2] = val >> 8;
drivers/media/dvb-frontends/dibx000_common.c
280
tx[3] = val & 0xff;
drivers/media/dvb-frontends/dibx000_common.c
413
u8 tx[4];
drivers/media/dvb-frontends/dibx000_common.c
414
struct i2c_msg m = {.addr = mst->i2c_addr,.buf = tx,.len = 4 };
drivers/media/dvb-frontends/dibx000_common.c
416
dibx000_i2c_gate_ctrl(mst, tx, 0, 0);
drivers/media/i2c/adv748x/adv748x-afe.c
269
ret = adv748x_tx_power(afe->tx, enable);
drivers/media/i2c/adv748x/adv748x-afe.c
302
struct v4l2_subdev *tx;
drivers/media/i2c/adv748x/adv748x-afe.c
304
tx = adv748x_get_remote_sd(&afe->pads[ADV748X_AFE_SOURCE]);
drivers/media/i2c/adv748x/adv748x-afe.c
305
if (!tx)
drivers/media/i2c/adv748x/adv748x-afe.c
313
return adv748x_csi2_set_pixelrate(tx, 14318180);
drivers/media/i2c/adv748x/adv748x-core.c
237
static int adv748x_power_up_tx(struct adv748x_csi2 *tx)
drivers/media/i2c/adv748x/adv748x-core.c
239
struct adv748x_state *state = tx->state;
drivers/media/i2c/adv748x/adv748x-core.c
240
u8 page = is_txa(tx) ? ADV748X_PAGE_TXA : ADV748X_PAGE_TXB;
drivers/media/i2c/adv748x/adv748x-core.c
244
adv748x_write_check(state, page, 0x00, 0x80 | tx->active_lanes, &ret);
drivers/media/i2c/adv748x/adv748x-core.c
247
adv748x_write_check(state, page, 0x00, 0xa0 | tx->active_lanes, &ret);
drivers/media/i2c/adv748x/adv748x-core.c
250
if (tx->src == &state->hdmi.sd) {
drivers/media/i2c/adv748x/adv748x-core.c
273
adv748x_write_check(state, page, 0x00, 0x20 | tx->active_lanes, &ret);
drivers/media/i2c/adv748x/adv748x-core.c
284
static int adv748x_power_down_tx(struct adv748x_csi2 *tx)
drivers/media/i2c/adv748x/adv748x-core.c
286
struct adv748x_state *state = tx->state;
drivers/media/i2c/adv748x/adv748x-core.c
287
u8 page = is_txa(tx) ? ADV748X_PAGE_TXA : ADV748X_PAGE_TXB;
drivers/media/i2c/adv748x/adv748x-core.c
295
adv748x_write_check(state, page, 0x00, 0x80 | tx->active_lanes, &ret);
drivers/media/i2c/adv748x/adv748x-core.c
306
int adv748x_tx_power(struct adv748x_csi2 *tx, bool on)
drivers/media/i2c/adv748x/adv748x-core.c
310
if (!is_tx_enabled(tx))
drivers/media/i2c/adv748x/adv748x-core.c
313
val = tx_read(tx, ADV748X_CSI_FS_AS_LS);
drivers/media/i2c/adv748x/adv748x-core.c
325
return on ? adv748x_power_up_tx(tx) : adv748x_power_down_tx(tx);
drivers/media/i2c/adv748x/adv748x-core.c
338
struct adv748x_csi2 *tx = adv748x_sd_to_csi2(sd);
drivers/media/i2c/adv748x/adv748x-core.c
346
if (enable && tx->src)
drivers/media/i2c/adv748x/adv748x-core.c
351
state->afe.tx = enable ? tx : NULL;
drivers/media/i2c/adv748x/adv748x-core.c
353
state->hdmi.tx = enable ? tx : NULL;
drivers/media/i2c/adv748x/adv748x-core.c
355
tx->src = enable ? rsd : NULL;
drivers/media/i2c/adv748x/adv748x-core.c
357
if (state->afe.tx) {
drivers/media/i2c/adv748x/adv748x-core.c
360
if (is_txa(tx)) {
drivers/media/i2c/adv748x/adv748x-core.c
367
tx->active_lanes = min(tx->num_lanes, 2U);
drivers/media/i2c/adv748x/adv748x-core.c
375
if (state->hdmi.tx) {
drivers/media/i2c/adv748x/adv748x-core.c
380
tx->active_lanes = tx->num_lanes;
drivers/media/i2c/adv748x/adv748x-csi2.c
104
struct adv748x_csi2 *tx = adv748x_sd_to_csi2(sd);
drivers/media/i2c/adv748x/adv748x-csi2.c
105
struct adv748x_state *state = tx->state;
drivers/media/i2c/adv748x/adv748x-csi2.c
108
adv_dbg(state, "Registered %s (%s)", is_txa(tx) ? "TXA":"TXB",
drivers/media/i2c/adv748x/adv748x-csi2.c
118
ret = adv748x_csi2_register_link(tx, sd->v4l2_dev,
drivers/media/i2c/adv748x/adv748x-csi2.c
121
is_txb(tx));
drivers/media/i2c/adv748x/adv748x-csi2.c
126
if (is_txb(tx))
drivers/media/i2c/adv748x/adv748x-csi2.c
127
state->afe.tx = tx;
drivers/media/i2c/adv748x/adv748x-csi2.c
131
if (is_txb(tx) || !is_hdmi_enabled(state))
drivers/media/i2c/adv748x/adv748x-csi2.c
134
ret = adv748x_csi2_register_link(tx, sd->v4l2_dev, &state->hdmi.sd,
drivers/media/i2c/adv748x/adv748x-csi2.c
140
state->hdmi.tx = tx;
drivers/media/i2c/adv748x/adv748x-csi2.c
156
struct adv748x_csi2 *tx = adv748x_sd_to_csi2(sd);
drivers/media/i2c/adv748x/adv748x-csi2.c
159
src = adv748x_get_remote_sd(&tx->pads[ADV748X_CSI2_SINK]);
drivers/media/i2c/adv748x/adv748x-csi2.c
181
struct adv748x_csi2 *tx = adv748x_sd_to_csi2(sd);
drivers/media/i2c/adv748x/adv748x-csi2.c
182
const unsigned int *codes = is_txa(tx) ?
drivers/media/i2c/adv748x/adv748x-csi2.c
185
size_t num_fmts = is_txa(tx) ? ARRAY_SIZE(adv748x_csi2_txa_fmts)
drivers/media/i2c/adv748x/adv748x-csi2.c
212
static bool adv748x_csi2_is_fmt_supported(struct adv748x_csi2 *tx, u32 code)
drivers/media/i2c/adv748x/adv748x-csi2.c
214
const unsigned int *codes = is_txa(tx) ?
drivers/media/i2c/adv748x/adv748x-csi2.c
217
size_t num_fmts = is_txa(tx) ? ARRAY_SIZE(adv748x_csi2_txa_fmts)
drivers/media/i2c/adv748x/adv748x-csi2.c
232
struct adv748x_csi2 *tx = adv748x_sd_to_csi2(sd);
drivers/media/i2c/adv748x/adv748x-csi2.c
242
if (!adv748x_csi2_is_fmt_supported(tx, sdformat->format.code))
drivers/media/i2c/adv748x/adv748x-csi2.c
25
int adv748x_csi2_set_virtual_channel(struct adv748x_csi2 *tx, unsigned int vc)
drivers/media/i2c/adv748x/adv748x-csi2.c
258
struct adv748x_csi2 *tx = adv748x_sd_to_csi2(sd);
drivers/media/i2c/adv748x/adv748x-csi2.c
264
config->bus.mipi_csi2.num_data_lanes = tx->active_lanes;
drivers/media/i2c/adv748x/adv748x-csi2.c
27
return tx_write(tx, ADV748X_CSI_VC_REF, vc << ADV748X_CSI_VC_REF_SHIFT);
drivers/media/i2c/adv748x/adv748x-csi2.c
291
struct adv748x_csi2 *tx = adv748x_sd_to_csi2(sd);
drivers/media/i2c/adv748x/adv748x-csi2.c
293
if (!tx->pixel_rate)
drivers/media/i2c/adv748x/adv748x-csi2.c
296
return v4l2_ctrl_s_ctrl_int64(tx->pixel_rate, rate);
drivers/media/i2c/adv748x/adv748x-csi2.c
313
static int adv748x_csi2_init_controls(struct adv748x_csi2 *tx)
drivers/media/i2c/adv748x/adv748x-csi2.c
316
v4l2_ctrl_handler_init(&tx->ctrl_hdl, 1);
drivers/media/i2c/adv748x/adv748x-csi2.c
318
tx->pixel_rate = v4l2_ctrl_new_std(&tx->ctrl_hdl,
drivers/media/i2c/adv748x/adv748x-csi2.c
323
tx->sd.ctrl_handler = &tx->ctrl_hdl;
drivers/media/i2c/adv748x/adv748x-csi2.c
324
if (tx->ctrl_hdl.error) {
drivers/media/i2c/adv748x/adv748x-csi2.c
325
v4l2_ctrl_handler_free(&tx->ctrl_hdl);
drivers/media/i2c/adv748x/adv748x-csi2.c
326
return tx->ctrl_hdl.error;
drivers/media/i2c/adv748x/adv748x-csi2.c
329
return v4l2_ctrl_handler_setup(&tx->ctrl_hdl);
drivers/media/i2c/adv748x/adv748x-csi2.c
332
int adv748x_csi2_init(struct adv748x_state *state, struct adv748x_csi2 *tx)
drivers/media/i2c/adv748x/adv748x-csi2.c
336
if (!is_tx_enabled(tx))
drivers/media/i2c/adv748x/adv748x-csi2.c
339
adv748x_subdev_init(&tx->sd, state, &adv748x_csi2_ops,
drivers/media/i2c/adv748x/adv748x-csi2.c
341
is_txa(tx) ? "txa" : "txb");
drivers/media/i2c/adv748x/adv748x-csi2.c
344
tx->sd.internal_ops = &adv748x_csi2_internal_ops;
drivers/media/i2c/adv748x/adv748x-csi2.c
346
tx->pads[ADV748X_CSI2_SINK].flags = MEDIA_PAD_FL_SINK;
drivers/media/i2c/adv748x/adv748x-csi2.c
347
tx->pads[ADV748X_CSI2_SOURCE].flags = MEDIA_PAD_FL_SOURCE;
drivers/media/i2c/adv748x/adv748x-csi2.c
349
ret = media_entity_pads_init(&tx->sd.entity, ADV748X_CSI2_NR_PADS,
drivers/media/i2c/adv748x/adv748x-csi2.c
350
tx->pads);
drivers/media/i2c/adv748x/adv748x-csi2.c
354
ret = v4l2_async_subdev_endpoint_add(&tx->sd,
drivers/media/i2c/adv748x/adv748x-csi2.c
355
of_fwnode_handle(state->endpoints[tx->port]));
drivers/media/i2c/adv748x/adv748x-csi2.c
359
ret = adv748x_csi2_init_controls(tx);
drivers/media/i2c/adv748x/adv748x-csi2.c
363
tx->sd.state_lock = &state->mutex;
drivers/media/i2c/adv748x/adv748x-csi2.c
364
ret = v4l2_subdev_init_finalize(&tx->sd);
drivers/media/i2c/adv748x/adv748x-csi2.c
368
ret = v4l2_async_register_subdev(&tx->sd);
drivers/media/i2c/adv748x/adv748x-csi2.c
375
v4l2_ctrl_handler_free(&tx->ctrl_hdl);
drivers/media/i2c/adv748x/adv748x-csi2.c
377
v4l2_subdev_cleanup(&tx->sd);
drivers/media/i2c/adv748x/adv748x-csi2.c
379
media_entity_cleanup(&tx->sd.entity);
drivers/media/i2c/adv748x/adv748x-csi2.c
384
void adv748x_csi2_cleanup(struct adv748x_csi2 *tx)
drivers/media/i2c/adv748x/adv748x-csi2.c
386
if (!is_tx_enabled(tx))
drivers/media/i2c/adv748x/adv748x-csi2.c
389
v4l2_async_unregister_subdev(&tx->sd);
drivers/media/i2c/adv748x/adv748x-csi2.c
390
media_entity_cleanup(&tx->sd.entity);
drivers/media/i2c/adv748x/adv748x-csi2.c
391
v4l2_ctrl_handler_free(&tx->ctrl_hdl);
drivers/media/i2c/adv748x/adv748x-csi2.c
392
v4l2_subdev_cleanup(&tx->sd);
drivers/media/i2c/adv748x/adv748x-csi2.c
42
static int adv748x_csi2_register_link(struct adv748x_csi2 *tx,
drivers/media/i2c/adv748x/adv748x-csi2.c
57
&tx->sd.entity, ADV748X_CSI2_SINK,
drivers/media/i2c/adv748x/adv748x-csi2.c
63
tx->src = src;
drivers/media/i2c/adv748x/adv748x-hdmi.c
371
ret = adv748x_tx_power(hdmi->tx, enable);
drivers/media/i2c/adv748x/adv748x-hdmi.c
396
struct v4l2_subdev *tx;
drivers/media/i2c/adv748x/adv748x-hdmi.c
399
tx = adv748x_get_remote_sd(&hdmi->pads[ADV748X_HDMI_SOURCE]);
drivers/media/i2c/adv748x/adv748x-hdmi.c
400
if (!tx)
drivers/media/i2c/adv748x/adv748x-hdmi.c
405
return adv748x_csi2_set_pixelrate(tx, timings.bt.pixelclock);
drivers/media/i2c/adv748x/adv748x.h
124
struct adv748x_csi2 *tx;
drivers/media/i2c/adv748x/adv748x.h
156
struct adv748x_csi2 *tx;
drivers/media/i2c/adv748x/adv748x.h
429
int adv748x_tx_power(struct adv748x_csi2 *tx, bool on);
drivers/media/i2c/adv748x/adv748x.h
435
int adv748x_csi2_init(struct adv748x_state *state, struct adv748x_csi2 *tx);
drivers/media/i2c/adv748x/adv748x.h
436
void adv748x_csi2_cleanup(struct adv748x_csi2 *tx);
drivers/media/i2c/adv748x/adv748x.h
437
int adv748x_csi2_set_virtual_channel(struct adv748x_csi2 *tx, unsigned int vc);
drivers/media/i2c/ds90ub960.c
3421
unsigned int rx, tx;
drivers/media/i2c/ds90ub960.c
3427
tx = ub960_pad_to_port(priv, route->source_pad);
drivers/media/i2c/ds90ub960.c
3429
vc[rx] = cur_vc[tx]++;
drivers/media/i2c/st-mipid02.c
411
struct v4l2_fwnode_endpoint *ep = &bridge->tx;
drivers/media/i2c/st-mipid02.c
792
bridge->tx = ep;
drivers/media/i2c/st-mipid02.c
97
struct v4l2_fwnode_endpoint tx;
drivers/media/pci/mgb4/mgb4_dma.c
27
struct dma_async_tx_descriptor *tx;
drivers/media/pci/mgb4/mgb4_dma.c
51
tx = dmaengine_prep_slave_sg(chan->chan, sgt->sgl, sgt->nents,
drivers/media/pci/mgb4/mgb4_dma.c
53
if (!tx) {
drivers/media/pci/mgb4/mgb4_dma.c
58
tx->callback = chan_irq;
drivers/media/pci/mgb4/mgb4_dma.c
59
tx->callback_param = chan;
drivers/media/pci/mgb4/mgb4_dma.c
61
ret = dma_submit_error(dmaengine_submit(tx));
drivers/media/platform/intel/pxa_camera.c
825
struct dma_async_tx_descriptor *tx;
drivers/media/platform/intel/pxa_camera.c
827
tx = dmaengine_prep_slave_sg(dma_chan, sg, sglen, DMA_DEV_TO_MEM,
drivers/media/platform/intel/pxa_camera.c
829
if (!tx) {
drivers/media/platform/intel/pxa_camera.c
835
tx->callback_param = pcdev;
drivers/media/platform/intel/pxa_camera.c
838
tx->callback = pxa_camera_dma_irq_y;
drivers/media/platform/intel/pxa_camera.c
841
tx->callback = pxa_camera_dma_irq_u;
drivers/media/platform/intel/pxa_camera.c
844
tx->callback = pxa_camera_dma_irq_v;
drivers/media/platform/intel/pxa_camera.c
848
buf->descs[channel] = tx;
drivers/media/platform/intel/pxa_camera.c
853
__func__, buf, tx);
drivers/media/platform/m2m-deinterlace.c
215
struct dma_async_tx_descriptor *tx;
drivers/media/platform/m2m-deinterlace.c
333
tx = dmadev->device_prep_interleaved_dma(chan, ctx->xt, flags);
drivers/media/platform/m2m-deinterlace.c
334
if (tx == NULL) {
drivers/media/platform/m2m-deinterlace.c
340
tx->callback = dma_callback;
drivers/media/platform/m2m-deinterlace.c
341
tx->callback_param = ctx;
drivers/media/platform/m2m-deinterlace.c
344
ctx->cookie = dmaengine_submit(tx);
drivers/media/platform/samsung/exynos-gsc/gsc-core.c
276
u32 tx, u32 ty, u32 *hratio)
drivers/media/platform/samsung/exynos-gsc/gsc-core.c
294
gsc_cal_prescaler_ratio(var, s_frame->crop.width, tx, wratio);
drivers/media/platform/samsung/exynos-gsc/gsc-core.c
639
int tx, ty;
drivers/media/platform/samsung/exynos-gsc/gsc-core.c
653
tx = d_frame->crop.height;
drivers/media/platform/samsung/exynos-gsc/gsc-core.c
655
tx = d_frame->crop.width;
drivers/media/platform/samsung/exynos-gsc/gsc-core.c
659
if (tx <= 0 || ty <= 0) {
drivers/media/platform/samsung/exynos-gsc/gsc-core.c
660
dev_err(dev, "Invalid target size: %dx%d", tx, ty);
drivers/media/platform/samsung/exynos-gsc/gsc-core.c
665
tx, &sc->pre_hratio);
drivers/media/platform/samsung/exynos-gsc/gsc-core.c
679
tx, ty, &sc->pre_vratio);
drivers/media/platform/samsung/exynos-gsc/gsc-core.c
684
sc->main_hratio = (s_frame->crop.width << 16) / tx;
drivers/media/platform/samsung/exynos-gsc/gsc-core.c
688
s_frame->crop.width, s_frame->crop.height, tx, ty);
drivers/media/platform/samsung/exynos-gsc/gsc-core.h
407
u32 *wratio, u32 tx, u32 ty, u32 *hratio);
drivers/media/platform/samsung/exynos4-is/fimc-core.c
233
int tx, ty, sx, sy;
drivers/media/platform/samsung/exynos4-is/fimc-core.c
238
tx = d_frame->height;
drivers/media/platform/samsung/exynos4-is/fimc-core.c
240
tx = d_frame->width;
drivers/media/platform/samsung/exynos4-is/fimc-core.c
243
if (tx <= 0 || ty <= 0) {
drivers/media/platform/samsung/exynos4-is/fimc-core.c
244
dev_err(dev, "Invalid target size: %dx%d\n", tx, ty);
drivers/media/platform/samsung/exynos4-is/fimc-core.c
257
ret = fimc_get_scaler_factor(sx, tx, &sc->pre_hratio, &sc->hfactor);
drivers/media/platform/samsung/exynos4-is/fimc-core.c
269
sc->main_hratio = (sx << 14) / (tx << sc->hfactor);
drivers/media/platform/samsung/exynos4-is/fimc-core.c
272
sc->main_hratio = (sx << 8) / (tx << sc->hfactor);
drivers/media/platform/samsung/exynos4-is/fimc-core.c
277
sc->scaleup_h = (tx >= sx) ? 1 : 0;
drivers/media/platform/ti/omap/omap_vout_vrfb.c
233
struct dma_async_tx_descriptor *tx;
drivers/media/platform/ti/omap/omap_vout_vrfb.c
270
tx = dmaengine_prep_interleaved_dma(chan, xt, flags);
drivers/media/platform/ti/omap/omap_vout_vrfb.c
271
if (tx == NULL) {
drivers/media/platform/ti/omap/omap_vout_vrfb.c
276
tx->callback = omap_vout_vrfb_dma_tx_callback;
drivers/media/platform/ti/omap/omap_vout_vrfb.c
277
tx->callback_param = &vout->vrfb_dma_tx;
drivers/media/platform/ti/omap/omap_vout_vrfb.c
279
cookie = dmaengine_submit(tx);
drivers/media/platform/ti/omap3isp/isphist.c
176
struct dma_async_tx_descriptor *tx;
drivers/media/platform/ti/omap3isp/isphist.c
203
tx = dmaengine_prep_slave_single(hist->dma_ch, dma_addr,
drivers/media/platform/ti/omap3isp/isphist.c
206
if (tx == NULL) {
drivers/media/platform/ti/omap3isp/isphist.c
212
tx->callback = hist_dma_cb;
drivers/media/platform/ti/omap3isp/isphist.c
213
tx->callback_param = hist;
drivers/media/platform/ti/omap3isp/isphist.c
214
cookie = tx->tx_submit(tx);
drivers/media/radio/radio-keene.c
109
radio->buffer[2] = radio->tx;
drivers/media/radio/radio-keene.c
253
radio->tx = db2tx[(ctrl->val - (s32)ctrl->minimum) / (s32)ctrl->step];
drivers/media/radio/radio-keene.c
336
radio->tx = 0x32;
drivers/media/radio/radio-keene.c
57
u8 tx;
drivers/media/rc/fintek-cir.h
62
} tx;
drivers/media/rc/imon.c
1090
ictx->tx.status = urb->status;
drivers/media/rc/imon.c
1093
ictx->tx.busy = false;
drivers/media/rc/imon.c
1095
complete(&ictx->tx.finished);
drivers/media/rc/imon.c
117
} tx;
drivers/media/rc/imon.c
2258
init_completion(&ictx->tx.finished);
drivers/media/rc/imon.c
2537
if (ictx->tx.busy) {
drivers/media/rc/imon.c
2539
complete(&ictx->tx.finished);
drivers/media/rc/imon.c
640
reinit_completion(&ictx->tx.finished);
drivers/media/rc/imon.c
641
ictx->tx.busy = true;
drivers/media/rc/imon.c
646
ictx->tx.busy = false;
drivers/media/rc/imon.c
651
retval = wait_for_completion_interruptible_timeout(&ictx->tx.finished, 10 * HZ);
drivers/media/rc/imon.c
656
ictx->tx.status = retval;
drivers/media/rc/imon.c
658
ictx->tx.status = -ETIMEDOUT;
drivers/media/rc/imon.c
661
ictx->tx.busy = false;
drivers/media/rc/imon.c
662
retval = ictx->tx.status;
drivers/media/rc/imon.c
977
if (copy_from_user(ictx->tx.data_buf, buf, n_bytes)) {
drivers/media/rc/imon.c
984
ictx->tx.data_buf[i] = ' ';
drivers/media/rc/imon.c
987
ictx->tx.data_buf[i] = 0xFF;
drivers/media/rc/imon.c
993
memcpy(ictx->usb_tx_buf, ictx->tx.data_buf + offset, 7);
drivers/media/spi/cxd2880-spi.c
65
struct spi_transfer tx = {};
drivers/media/spi/cxd2880-spi.c
72
tx.tx_buf = data;
drivers/media/spi/cxd2880-spi.c
73
tx.len = size;
drivers/media/spi/cxd2880-spi.c
76
spi_message_add_tail(&tx, &msg);
drivers/media/spi/gs1662.c
138
struct spi_transfer tx[] = {
drivers/media/spi/gs1662.c
157
spi_message_add_tail(&tx[0], &msg);
drivers/media/spi/gs1662.c
158
spi_message_add_tail(&tx[1], &msg);
drivers/media/spi/gs1662.c
172
struct spi_transfer tx[] = {
drivers/media/spi/gs1662.c
191
spi_message_add_tail(&tx[0], &msg);
drivers/media/spi/gs1662.c
192
spi_message_add_tail(&tx[1], &msg);
drivers/media/usb/dvb-usb/dib0700.h
60
int dib0700_ctrl_rd(struct dvb_usb_device *d, u8 *tx, u8 txlen, u8 *rx,
drivers/media/usb/dvb-usb/dib0700_core.c
53
static int dib0700_ctrl_wr(struct dvb_usb_device *d, u8 *tx, u8 txlen)
drivers/media/usb/dvb-usb/dib0700_core.c
58
debug_dump(tx, txlen, deb_data);
drivers/media/usb/dvb-usb/dib0700_core.c
61
tx[0], USB_TYPE_VENDOR | USB_DIR_OUT, 0, 0, tx, txlen,
drivers/media/usb/dvb-usb/dib0700_core.c
71
int dib0700_ctrl_rd(struct dvb_usb_device *d, u8 *tx, u8 txlen, u8 *rx, u8 rxlen)
drivers/media/usb/dvb-usb/dib0700_core.c
86
debug_dump(tx,txlen,deb_data);
drivers/media/usb/dvb-usb/dib0700_core.c
88
value = ((txlen - 2) << 8) | tx[1];
drivers/media/usb/dvb-usb/dib0700_core.c
91
index |= (tx[2] << 8);
drivers/media/usb/dvb-usb/dib0700_core.c
93
index |= tx[3];
drivers/media/usb/dvb-usb/dib0700_core.c
95
status = usb_control_msg(d->udev, usb_rcvctrlpipe(d->udev,0), tx[0],
drivers/media/usb/dvb-usb/technisat-usb2.c
117
memcpy(&b[2], tx, txlen);
drivers/media/usb/dvb-usb/technisat-usb2.c
85
u8 device_addr, u8 *tx, u8 txlen, u8 *rx, u8 rxlen)
drivers/media/usb/dvb-usb/technisat-usb2.c
95
debug_dump(tx, txlen, deb_i2c);
drivers/media/usb/hackrf/hackrf.c
242
const bool tx = test_bit(TX_ON, &dev->flags);
drivers/media/usb/hackrf/hackrf.c
264
if (!rx && !tx) {
drivers/media/usb/hackrf/hackrf.c
275
} else if (tx && test_and_clear_bit(TX_DAC_FREQUENCY, &dev->flags)) {
drivers/media/usb/hackrf/hackrf.c
314
} else if (tx && test_and_clear_bit(TX_BANDWIDTH, &dev->flags)) {
drivers/media/usb/hackrf/hackrf.c
351
} else if (tx && test_and_clear_bit(TX_RF_FREQUENCY, &dev->flags)) {
drivers/media/usb/hackrf/hackrf.c
386
if (tx && test_and_clear_bit(TX_RF_GAIN, &dev->flags)) {
drivers/media/usb/hackrf/hackrf.c
420
if (tx && test_and_clear_bit(TX_LNA_GAIN, &dev->flags)) {
drivers/memory/tegra/tegra186-emc.c
162
msg.tx.data = NULL;
drivers/memory/tegra/tegra186-emc.c
163
msg.tx.size = 0;
drivers/memory/tegra/tegra234.c
1074
msg.tx.data = &bwmgr_req;
drivers/memory/tegra/tegra234.c
1075
msg.tx.size = sizeof(bwmgr_req);
drivers/memory/tegra/tegra264.c
243
msg.tx.data = &bwmgr_req;
drivers/memory/tegra/tegra264.c
244
msg.tx.size = sizeof(bwmgr_req);
drivers/mfd/ipaq-micro.c
280
struct ipaq_micro_txdev *tx = µ->tx;
drivers/mfd/ipaq-micro.c
283
while ((tx->index < tx->len) &&
drivers/mfd/ipaq-micro.c
285
writel(tx->buf[tx->index], micro->base + UTDR);
drivers/mfd/ipaq-micro.c
286
tx->index++;
drivers/mfd/ipaq-micro.c
332
struct ipaq_micro_txdev *tx = µ->tx;
drivers/mfd/ipaq-micro.c
34
struct ipaq_micro_txdev *tx = µ->tx;
drivers/mfd/ipaq-micro.c
354
} while (((tx->index < tx->len) && (status & UTSR0_TFS)) ||
drivers/mfd/ipaq-micro.c
41
tx->buf[bp++] = CHAR_SOF;
drivers/mfd/ipaq-micro.c
44
tx->buf[bp++] = checksum;
drivers/mfd/ipaq-micro.c
47
tx->buf[bp++] = msg->tx_data[i];
drivers/mfd/ipaq-micro.c
51
tx->buf[bp++] = checksum;
drivers/mfd/ipaq-micro.c
52
tx->len = bp;
drivers/mfd/ipaq-micro.c
53
tx->index = 0;
drivers/mfd/qnap-mcu.c
100
tx[data_size] = qnap_mcu_csum(data, data_size);
drivers/mfd/qnap-mcu.c
104
return serdev_device_write(mcu->serdev, tx, length, HZ);
drivers/mfd/qnap-mcu.c
91
unsigned char tx[QNAP_MCU_TX_BUFFER_SIZE];
drivers/mfd/qnap-mcu.c
94
if (length > sizeof(tx)) {
drivers/mfd/qnap-mcu.c
99
memcpy(tx, data, data_size);
drivers/mmc/core/sdio_uart.c
417
port->icount.tx++;
drivers/mmc/core/sdio_uart.c
434
port->icount.tx++;
drivers/mmc/core/sdio_uart.c
57
__u32 tx;
drivers/mmc/core/sdio_uart.c
956
port->icount.tx, port->icount.rx);
drivers/mmc/host/omap.c
1045
tx = dmaengine_prep_slave_sg(c, data->sg, host->sg_len,
drivers/mmc/host/omap.c
1048
if (!tx)
drivers/mmc/host/omap.c
1053
tx->callback = mmc_omap_dma_callback;
drivers/mmc/host/omap.c
1054
tx->callback_param = host;
drivers/mmc/host/omap.c
1055
dmaengine_submit(tx);
drivers/mmc/host/omap.c
989
struct dma_async_tx_descriptor *tx;
drivers/mmc/host/omap_hsmmc.c
1234
struct dma_async_tx_descriptor *tx;
drivers/mmc/host/omap_hsmmc.c
1273
tx = dmaengine_prep_slave_sg(chan, data->sg, data->sg_len,
drivers/mmc/host/omap_hsmmc.c
1276
if (!tx) {
drivers/mmc/host/omap_hsmmc.c
1282
tx->callback = omap_hsmmc_dma_callback;
drivers/mmc/host/omap_hsmmc.c
1283
tx->callback_param = host;
drivers/mmc/host/omap_hsmmc.c
1286
dmaengine_submit(tx);
drivers/mmc/host/pxamci.c
160
struct dma_async_tx_descriptor *tx;
drivers/mmc/host/pxamci.c
208
tx = dmaengine_prep_slave_sg(chan, data->sg, host->dma_len, direction,
drivers/mmc/host/pxamci.c
210
if (!tx) {
drivers/mmc/host/pxamci.c
216
tx->callback = pxamci_dma_irq;
drivers/mmc/host/pxamci.c
217
tx->callback_param = host;
drivers/mmc/host/pxamci.c
220
host->dma_cookie = dmaengine_submit(tx);
drivers/mmc/host/rtsx_usb_sdmmc.c
576
static int sd_change_phase(struct rtsx_usb_sdmmc *host, u8 sample_point, int tx)
drivers/mmc/host/rtsx_usb_sdmmc.c
581
__func__, tx ? "TX" : "RX", sample_point);
drivers/mmc/host/rtsx_usb_sdmmc.c
587
if (tx)
drivers/mtd/hyperbus/hbmc-am654.c
102
tx->callback = am654_hbmc_dma_callback;
drivers/mtd/hyperbus/hbmc-am654.c
103
tx->callback_param = priv;
drivers/mtd/hyperbus/hbmc-am654.c
104
cookie = dmaengine_submit(tx);
drivers/mtd/hyperbus/hbmc-am654.c
79
struct dma_async_tx_descriptor *tx;
drivers/mtd/hyperbus/hbmc-am654.c
94
tx = dmaengine_prep_dma_memcpy(rx_chan, dma_dst, dma_src, len, flags);
drivers/mtd/hyperbus/hbmc-am654.c
95
if (!tx) {
drivers/mtd/nand/onenand/onenand_omap2.c
328
struct dma_async_tx_descriptor *tx;
drivers/mtd/nand/onenand/onenand_omap2.c
331
tx = dmaengine_prep_dma_memcpy(c->dma_chan, dst, src, count,
drivers/mtd/nand/onenand/onenand_omap2.c
333
if (!tx) {
drivers/mtd/nand/onenand/onenand_omap2.c
340
tx->callback = omap2_onenand_dma_complete_func;
drivers/mtd/nand/onenand/onenand_omap2.c
341
tx->callback_param = &c->dma_done;
drivers/mtd/nand/onenand/onenand_omap2.c
343
cookie = tx->tx_submit(tx);
drivers/mtd/nand/raw/atmel/nand-controller.c
372
struct dma_async_tx_descriptor *tx;
drivers/mtd/nand/raw/atmel/nand-controller.c
390
tx = dmaengine_prep_dma_memcpy(nc->dmac, dst_dma, src_dma, len,
drivers/mtd/nand/raw/atmel/nand-controller.c
392
if (!tx) {
drivers/mtd/nand/raw/atmel/nand-controller.c
397
tx->callback = atmel_nand_dma_transfer_finished;
drivers/mtd/nand/raw/atmel/nand-controller.c
398
tx->callback_param = &finished;
drivers/mtd/nand/raw/atmel/nand-controller.c
400
cookie = dmaengine_submit(tx);
drivers/mtd/nand/raw/cadence-nand-controller.c
1878
struct dma_async_tx_descriptor *tx;
drivers/mtd/nand/raw/cadence-nand-controller.c
1898
tx = dmaengine_prep_dma_memcpy(cdns_ctrl->dmac, dst_dma, src_dma, len,
drivers/mtd/nand/raw/cadence-nand-controller.c
1900
if (!tx) {
drivers/mtd/nand/raw/cadence-nand-controller.c
1905
tx->callback = cadence_nand_slave_dma_transfer_finished;
drivers/mtd/nand/raw/cadence-nand-controller.c
1906
tx->callback_param = &finished;
drivers/mtd/nand/raw/cadence-nand-controller.c
1908
cookie = dmaengine_submit(tx);
drivers/mtd/nand/raw/fsmc_nand.c
490
struct dma_async_tx_descriptor *tx;
drivers/mtd/nand/raw/fsmc_nand.c
517
tx = dma_dev->device_prep_dma_memcpy(chan, dma_dst, dma_src,
drivers/mtd/nand/raw/fsmc_nand.c
519
if (!tx) {
drivers/mtd/nand/raw/fsmc_nand.c
525
tx->callback = dma_complete;
drivers/mtd/nand/raw/fsmc_nand.c
526
tx->callback_param = host;
drivers/mtd/nand/raw/fsmc_nand.c
527
cookie = tx->tx_submit(tx);
drivers/mtd/nand/raw/intel-nand-controller.c
290
struct dma_async_tx_descriptor *tx;
drivers/mtd/nand/raw/intel-nand-controller.c
317
tx = dmaengine_prep_slave_single(chan, buf_dma, len, dir, flags);
drivers/mtd/nand/raw/intel-nand-controller.c
318
if (!tx) {
drivers/mtd/nand/raw/intel-nand-controller.c
323
tx->callback = callback;
drivers/mtd/nand/raw/intel-nand-controller.c
324
tx->callback_param = ebu_host;
drivers/mtd/nand/raw/intel-nand-controller.c
325
cookie = tx->tx_submit(tx);
drivers/mtd/nand/raw/marvell_nand.c
871
struct dma_async_tx_descriptor *tx;
drivers/mtd/nand/raw/marvell_nand.c
885
tx = dmaengine_prep_slave_sg(nfc->dma_chan, &sg, 1,
drivers/mtd/nand/raw/marvell_nand.c
889
if (!tx) {
drivers/mtd/nand/raw/marvell_nand.c
896
cookie = dmaengine_submit(tx);
drivers/mtd/nand/raw/nandsim.c
1343
ssize_t tx;
drivers/mtd/nand/raw/nandsim.c
1351
tx = kernel_read(file, buf, count, &pos);
drivers/mtd/nand/raw/nandsim.c
1354
return tx;
drivers/mtd/nand/raw/nandsim.c
1360
ssize_t tx;
drivers/mtd/nand/raw/nandsim.c
1368
tx = kernel_write(file, buf, count, &pos);
drivers/mtd/nand/raw/nandsim.c
1371
return tx;
drivers/mtd/nand/raw/nandsim.c
1432
ssize_t tx;
drivers/mtd/nand/raw/nandsim.c
1439
tx = ns_read_file(ns, ns->cfile, ns->buf.byte, num,
drivers/mtd/nand/raw/nandsim.c
1441
if (tx != num) {
drivers/mtd/nand/raw/nandsim.c
1442
NS_ERR("read_page: read error for page %d ret %ld\n", ns->regs.row, (long)tx);
drivers/mtd/nand/raw/nandsim.c
1503
ssize_t tx;
drivers/mtd/nand/raw/nandsim.c
1514
tx = ns_read_file(ns, ns->cfile, pg_off, num, off);
drivers/mtd/nand/raw/nandsim.c
1515
if (tx != num) {
drivers/mtd/nand/raw/nandsim.c
1516
NS_ERR("prog_page: read error for page %d ret %ld\n", ns->regs.row, (long)tx);
drivers/mtd/nand/raw/nandsim.c
1524
tx = ns_write_file(ns, ns->cfile, ns->file_buf,
drivers/mtd/nand/raw/nandsim.c
1526
if (tx != ns->geom.pgszoob) {
drivers/mtd/nand/raw/nandsim.c
1527
NS_ERR("prog_page: write error for page %d ret %ld\n", ns->regs.row, (long)tx);
drivers/mtd/nand/raw/nandsim.c
1532
tx = ns_write_file(ns, ns->cfile, pg_off, num, off);
drivers/mtd/nand/raw/nandsim.c
1533
if (tx != num) {
drivers/mtd/nand/raw/nandsim.c
1534
NS_ERR("prog_page: write error for page %d ret %ld\n", ns->regs.row, (long)tx);
drivers/mtd/nand/raw/omap2.c
382
struct dma_async_tx_descriptor *tx;
drivers/mtd/nand/raw/omap2.c
402
tx = dmaengine_prep_slave_sg(info->dma, &sg, n,
drivers/mtd/nand/raw/omap2.c
405
if (!tx)
drivers/mtd/nand/raw/omap2.c
408
tx->callback = omap_nand_dma_callback;
drivers/mtd/nand/raw/omap2.c
409
tx->callback_param = &info->comp;
drivers/mtd/nand/raw/omap2.c
410
dmaengine_submit(tx);
drivers/net/can/c_can/c_can.h
203
struct c_can_tx_ring tx;
drivers/net/can/c_can/c_can_main.c
1247
priv->tx.head = 0;
drivers/net/can/c_can/c_can_main.c
1248
priv->tx.tail = 0;
drivers/net/can/c_can/c_can_main.c
1249
priv->tx.obj_num = msg_obj_tx_num;
drivers/net/can/c_can/c_can_main.c
457
struct c_can_tx_ring *tx_ring = &priv->tx;
drivers/net/can/c_can/c_can_main.c
592
struct c_can_tx_ring *tx_ring = &priv->tx;
drivers/net/can/c_can/c_can_main.c
723
struct c_can_tx_ring *tx_ring = &priv->tx;
drivers/net/can/grcan.c
1057
priv->echo_skb = kzalloc_objs(*priv->echo_skb, dma->tx.size);
drivers/net/can/grcan.c
1062
priv->can.echo_skb_max = dma->tx.size;
drivers/net/can/grcan.c
1293
if (grcan_txspace(dma->tx.size, txwr, priv->eskbp))
drivers/net/can/grcan.c
1365
space = grcan_txspace(dma->tx.size, txwr, priv->eskbp);
drivers/net/can/grcan.c
1368
slot = dma->tx.buf + txwr;
drivers/net/can/grcan.c
1420
if (unlikely(grcan_ring_sub(txwr, txrd, dma->tx.size) == 1)) {
drivers/net/can/grcan.c
1445
grcan_ring_add(txwr, GRCAN_MSG_SIZE, dma->tx.size));
drivers/net/can/grcan.c
225
struct grcan_dma_buffer tx;
drivers/net/can/grcan.c
527
dma->tx.size);
drivers/net/can/grcan.c
558
txrd = grcan_ring_add(txrd, GRCAN_MSG_SIZE, dma->tx.size);
drivers/net/can/grcan.c
696
if (grcan_txspace(dma->tx.size, txwr,
drivers/net/can/grcan.c
864
if (grcan_txspace(priv->dma.tx.size, txwr, priv->eskbp) &&
drivers/net/can/grcan.c
938
struct grcan_dma_buffer *large = rsize > tsize ? &dma->rx : &dma->tx;
drivers/net/can/grcan.c
939
struct grcan_dma_buffer *small = rsize > tsize ? &dma->tx : &dma->rx;
drivers/net/can/grcan.c
961
dma->tx.size = tsize;
drivers/net/can/grcan.c
983
grcan_write_reg(®s->txaddr, priv->dma.tx.handle);
drivers/net/can/grcan.c
984
grcan_write_reg(®s->txsize, priv->dma.tx.size);
drivers/net/can/mscan/mscan.c
229
out_be16(®s->tx.idr3_2, can_id);
drivers/net/can/mscan/mscan.c
241
out_be16(®s->tx.idr1_0, can_id);
drivers/net/can/mscan/mscan.c
244
void __iomem *data = ®s->tx.dsr1_0;
drivers/net/can/mscan/mscan.c
256
out_8(®s->tx.dlr, frame->len);
drivers/net/can/mscan/mscan.c
257
out_8(®s->tx.tbpr, priv->cur_pri);
drivers/net/can/mscan/mscan.h
217
} tx;
drivers/net/can/softing/softing.h
29
} tx;
drivers/net/can/softing/softing.h
64
} tx;
drivers/net/can/softing/softing_fw.c
334
card->tx.pending = 0;
drivers/net/can/softing/softing_fw.c
455
priv->tx.pending = 0;
drivers/net/can/softing/softing_fw.c
456
priv->tx.echo_put = 0;
drivers/net/can/softing/softing_fw.c
457
priv->tx.echo_get = 0;
drivers/net/can/softing/softing_fw.c
470
card->tx.pending = 0;
drivers/net/can/softing/softing_main.c
105
card->tx.last_bus = priv->index;
drivers/net/can/softing/softing_main.c
106
++card->tx.pending;
drivers/net/can/softing/softing_main.c
107
++priv->tx.pending;
drivers/net/can/softing/softing_main.c
108
can_put_echo_skb(skb, dev, priv->tx.echo_put, 0);
drivers/net/can/softing/softing_main.c
109
++priv->tx.echo_put;
drivers/net/can/softing/softing_main.c
110
if (priv->tx.echo_put >= TX_ECHO_SKB_MAX)
drivers/net/can/softing/softing_main.c
111
priv->tx.echo_put = 0;
drivers/net/can/softing/softing_main.c
116
if (card->tx.pending >= TXMAX) {
drivers/net/can/softing/softing_main.c
283
skb = priv->can.echo_skb[priv->tx.echo_get];
drivers/net/can/softing/softing_main.c
288
can_get_echo_skb(netdev, priv->tx.echo_get,
drivers/net/can/softing/softing_main.c
290
++priv->tx.echo_get;
drivers/net/can/softing/softing_main.c
291
if (priv->tx.echo_get >= TX_ECHO_SKB_MAX)
drivers/net/can/softing/softing_main.c
292
priv->tx.echo_get = 0;
drivers/net/can/softing/softing_main.c
293
if (priv->tx.pending)
drivers/net/can/softing/softing_main.c
294
--priv->tx.pending;
drivers/net/can/softing/softing_main.c
295
if (card->tx.pending)
drivers/net/can/softing/softing_main.c
296
--card->tx.pending;
drivers/net/can/softing/softing_main.c
332
offset = card->tx.last_bus;
drivers/net/can/softing/softing_main.c
334
if (card->tx.pending >= TXMAX)
drivers/net/can/softing/softing_main.c
343
if (priv->tx.pending >= TX_ECHO_SKB_MAX)
drivers/net/can/softing/softing_main.c
70
(card->tx.pending >= TXMAX) ||
drivers/net/can/softing/softing_main.c
71
(priv->tx.pending >= TX_ECHO_SKB_MAX))
drivers/net/can/spi/mcp251xfd/mcp251xfd-chip-fifo.c
102
MCP251XFD_REG_FIFOCON(priv->tx->fifo_nr),
drivers/net/can/spi/mcp251xfd/mcp251xfd-chip-fifo.c
65
const struct mcp251xfd_tx_ring *tx_ring = priv->tx;
drivers/net/can/spi/mcp251xfd/mcp251xfd-core.c
1267
struct mcp251xfd_tx_ring *tx_ring = priv->tx;
drivers/net/can/spi/mcp251xfd/mcp251xfd-core.c
1336
err = mcp251xfd_get_tx_nr_by_addr(priv->tx, &nr, addr);
drivers/net/can/spi/mcp251xfd/mcp251xfd-dump.c
121
const struct mcp251xfd_tx_ring *tx = priv->tx;
drivers/net/can/spi/mcp251xfd/mcp251xfd-dump.c
140
.val = tx->obj_num,
drivers/net/can/spi/mcp251xfd/mcp251xfd-dump.c
197
const struct mcp251xfd_tx_ring *tx = priv->tx;
drivers/net/can/spi/mcp251xfd/mcp251xfd-dump.c
201
.val = tx->head,
drivers/net/can/spi/mcp251xfd/mcp251xfd-dump.c
204
.val = tx->tail,
drivers/net/can/spi/mcp251xfd/mcp251xfd-dump.c
207
.val = tx->base,
drivers/net/can/spi/mcp251xfd/mcp251xfd-dump.c
210
.val = tx->nr,
drivers/net/can/spi/mcp251xfd/mcp251xfd-dump.c
213
.val = tx->fifo_nr,
drivers/net/can/spi/mcp251xfd/mcp251xfd-dump.c
216
.val = tx->obj_num,
drivers/net/can/spi/mcp251xfd/mcp251xfd-dump.c
219
.val = tx->obj_size,
drivers/net/can/spi/mcp251xfd/mcp251xfd-ethtool.c
112
priv->tx->obj_num = layout.cur_tx;
drivers/net/can/spi/mcp251xfd/mcp251xfd-ethtool.c
139
priv->tx->obj_num = layout.default_tx;
drivers/net/can/spi/mcp251xfd/mcp251xfd-ethtool.c
29
ring->tx_pending = priv->tx->obj_num;
drivers/net/can/spi/mcp251xfd/mcp251xfd-ethtool.c
44
layout.cur_tx != priv->tx->obj_num) &&
drivers/net/can/spi/mcp251xfd/mcp251xfd-ethtool.c
50
priv->tx->obj_num = layout.cur_tx;
drivers/net/can/spi/mcp251xfd/mcp251xfd-ethtool.c
95
.tx_pending = priv->tx->obj_num,
drivers/net/can/spi/mcp251xfd/mcp251xfd-ram.c
130
num_tx = ram_free / config->tx.size[fd_mode];
drivers/net/can/spi/mcp251xfd/mcp251xfd-ram.c
132
num_tx = can_ram_rounddown_pow_of_two(config, &config->tx, 0, num_tx);
drivers/net/can/spi/mcp251xfd/mcp251xfd-ram.c
144
(u32)config->tx.fifo_depth_coalesce_min,
drivers/net/can/spi/mcp251xfd/mcp251xfd-ram.c
148
num_tx = can_ram_rounddown_pow_of_two(config, &config->tx,
drivers/net/can/spi/mcp251xfd/mcp251xfd-ram.c
71
num_tx = config->tx.def[fd_mode];
drivers/net/can/spi/mcp251xfd/mcp251xfd-ram.c
72
num_tx = can_ram_rounddown_pow_of_two(config, &config->tx, 0, num_tx);
drivers/net/can/spi/mcp251xfd/mcp251xfd-ram.c
75
ram_free -= config->tx.size[fd_mode] * num_tx;
drivers/net/can/spi/mcp251xfd/mcp251xfd-ram.c
85
ram_free -= config->tx.size[fd_mode] * config->tx.min;
drivers/net/can/spi/mcp251xfd/mcp251xfd-ram.c
90
num_tx = ram_free / config->tx.size[fd_mode];
drivers/net/can/spi/mcp251xfd/mcp251xfd-ram.c
93
layout->max_tx = can_ram_rounddown_pow_of_two(config, &config->tx, 0, num_tx);
drivers/net/can/spi/mcp251xfd/mcp251xfd-ram.h
36
const struct can_ram_obj_config tx;
drivers/net/can/spi/mcp251xfd/mcp251xfd-ring.c
178
tx_ring = priv->tx;
drivers/net/can/spi/mcp251xfd/mcp251xfd-ring.c
331
priv->tx->obj_num - priv->tx_obj_num_coalesce_irq,
drivers/net/can/spi/mcp251xfd/mcp251xfd-ring.c
333
(priv->tx->obj_num - priv->tx_obj_num_coalesce_irq) *
drivers/net/can/spi/mcp251xfd/mcp251xfd-ring.c
339
priv->tx->obj_num, sizeof(struct mcp251xfd_hw_tef_obj),
drivers/net/can/spi/mcp251xfd/mcp251xfd-ring.c
340
priv->tx->obj_num * sizeof(struct mcp251xfd_hw_tef_obj));
drivers/net/can/spi/mcp251xfd/mcp251xfd-ring.c
375
priv->tx->fifo_nr,
drivers/net/can/spi/mcp251xfd/mcp251xfd-ring.c
376
mcp251xfd_get_tx_obj_addr(priv->tx, 0),
drivers/net/can/spi/mcp251xfd/mcp251xfd-ring.c
377
priv->tx->obj_num, priv->tx->obj_size,
drivers/net/can/spi/mcp251xfd/mcp251xfd-ring.c
378
priv->tx->obj_num * priv->tx->obj_size);
drivers/net/can/spi/mcp251xfd/mcp251xfd-ring.c
393
priv->tx_obj_num_coalesce_irq * 2 != priv->tx->obj_num) {
drivers/net/can/spi/mcp251xfd/mcp251xfd-ring.c
396
priv->tx_obj_num_coalesce_irq, priv->tx->obj_num);
drivers/net/can/spi/mcp251xfd/mcp251xfd-ring.c
453
.tx = {
drivers/net/can/spi/mcp251xfd/mcp251xfd-ring.c
473
struct mcp251xfd_tx_ring *tx_ring = priv->tx;
drivers/net/can/spi/mcp251xfd/mcp251xfd-ring.c
482
.tx_pending = priv->tx->obj_num,
drivers/net/can/spi/mcp251xfd/mcp251xfd-ring.c
82
*base = mcp251xfd_get_tef_obj_addr(priv->tx->obj_num);
drivers/net/can/spi/mcp251xfd/mcp251xfd-tef.c
119
const struct mcp251xfd_tx_ring *tx_ring = priv->tx;
drivers/net/can/spi/mcp251xfd/mcp251xfd-tef.c
125
err = regmap_read(priv->map_reg, MCP251XFD_REG_FIFOSTA(priv->tx->fifo_nr),
drivers/net/can/spi/mcp251xfd/mcp251xfd-tef.c
187
const struct mcp251xfd_tx_ring *tx_ring = priv->tx;
drivers/net/can/spi/mcp251xfd/mcp251xfd-tef.c
257
struct mcp251xfd_tx_ring *tx_ring = priv->tx;
drivers/net/can/spi/mcp251xfd/mcp251xfd-tef.c
287
if (mcp251xfd_get_tx_free(priv->tx)) {
drivers/net/can/spi/mcp251xfd/mcp251xfd-tx.c
159
struct mcp251xfd_tx_ring *tx_ring = priv->tx;
drivers/net/can/spi/mcp251xfd/mcp251xfd-tx.c
207
struct mcp251xfd_tx_ring *tx_ring = priv->tx;
drivers/net/can/spi/mcp251xfd/mcp251xfd.h
647
struct mcp251xfd_tx_ring tx[MCP251XFD_FIFO_TX_NUM];
drivers/net/can/spi/mcp251xfd/mcp251xfd.h
875
MCP251XFD_REG_FIFOSTA(priv->tx->fifo_nr),
drivers/net/can/spi/mcp251xfd/mcp251xfd.h
887
return priv->tef->head & (priv->tx->obj_num - 1);
drivers/net/can/spi/mcp251xfd/mcp251xfd.h
892
return priv->tef->tail & (priv->tx->obj_num - 1);
drivers/net/can/spi/mcp251xfd/mcp251xfd.h
897
return min_t(u8, len, priv->tx->obj_num - mcp251xfd_get_tef_tail(priv));
drivers/net/can/usb/ems_usb.c
187
u8 tx;
drivers/net/can/usb/esd_usb.c
247
struct esd_usb_tx_msg tx;
drivers/net/can/usb/esd_usb.c
857
msg->tx.net = priv->index;
drivers/net/can/usb/esd_usb.c
860
msg->tx.dlc = can_fd_len2dlc(cfd->len);
drivers/net/can/usb/esd_usb.c
861
msg->tx.dlc |= ESD_USB_FD;
drivers/net/can/usb/esd_usb.c
864
msg->tx.dlc |= ESD_USB_NO_BRS;
drivers/net/can/usb/esd_usb.c
866
msg->tx.dlc = can_get_cc_dlc((struct can_frame *)cfd, priv->can.ctrlmode);
drivers/net/can/usb/esd_usb.c
869
msg->tx.dlc |= ESD_USB_RTR;
drivers/net/can/usb/esd_usb.c
872
msg->tx.id = cpu_to_le32(cfd->can_id & CAN_ERR_MASK);
drivers/net/can/usb/esd_usb.c
875
msg->tx.id |= cpu_to_le32(ESD_USB_EXTID);
drivers/net/can/usb/esd_usb.c
877
memcpy(msg->tx.data_fd, cfd->data, cfd->len);
drivers/net/can/usb/esd_usb.c
900
msg->tx.hnd = BIT(31) | i; /* returned in TX done message */
drivers/net/can/usb/nct6694_canfd.c
140
struct nct6694_canfd_frame tx;
drivers/net/can/usb/nct6694_canfd.c
438
struct nct6694_canfd_frame *frame = &priv->tx;
drivers/net/dsa/microchip/ksz_common.c
101
u64 tx;
drivers/net/dsa/microchip/ksz_common.c
2316
stats->tx_bytes = raw->tx + raw->tx_hi - stats->tx_packets * ETH_FCS_LEN;
drivers/net/dsa/mxl862xx/mxl862xx-api.h
619
u8 tx; /* enum mxl862xx_ss_sp_tag_tx */
drivers/net/dsa/mxl862xx/mxl862xx.c
254
.tx = enable ? MXL862XX_SS_SP_TAG_TX_TAG_NO_REMOVE :
drivers/net/dsa/vitesse-vsc73xx-core.c
1453
bool tx)
drivers/net/dsa/vitesse-vsc73xx-core.c
1459
if (tx) {
drivers/net/ethernet/3com/typhoon.c
1506
struct tx_desc *tx;
drivers/net/ethernet/3com/typhoon.c
1512
tx = (struct tx_desc *) (txRing->ringBase + lastRead);
drivers/net/ethernet/3com/typhoon.c
1513
type = tx->flags & TYPHOON_TYPE_MASK;
drivers/net/ethernet/3com/typhoon.c
1518
unsigned long ptr = tx->tx_addr;
drivers/net/ethernet/3com/typhoon.c
1524
skb_dma = (dma_addr_t) le32_to_cpu(tx->frag.addr);
drivers/net/ethernet/3com/typhoon.c
1525
dma_len = le16_to_cpu(tx->len);
drivers/net/ethernet/3com/typhoon.c
1530
tx->flags = 0;
drivers/net/ethernet/alteon/acenic.c
2324
struct tx_desc __iomem *tx;
drivers/net/ethernet/alteon/acenic.c
2325
tx = (__force struct tx_desc __iomem *) &ap->tx_ring[i];
drivers/net/ethernet/alteon/acenic.c
2326
writel(0, &tx->addr.addrhi);
drivers/net/ethernet/alteon/acenic.c
2327
writel(0, &tx->addr.addrlo);
drivers/net/ethernet/alteon/acenic.c
2328
writel(0, &tx->flagsize);
drivers/net/ethernet/amazon/ena/ena_admin_defs.h
790
u32 tx;
drivers/net/ethernet/amazon/ena/ena_ethtool.c
42
ENA_STAT_ENTRY(stat, tx)
drivers/net/ethernet/amazon/ena/ena_netdev.c
3758
if (feat->offload.tx &
drivers/net/ethernet/amazon/ena/ena_netdev.c
3762
if (feat->offload.tx &
drivers/net/ethernet/amazon/ena/ena_netdev.c
3766
if (feat->offload.tx & ENA_ADMIN_FEATURE_OFFLOAD_DESC_TSO_IPV4_MASK)
drivers/net/ethernet/amazon/ena/ena_netdev.c
3769
if (feat->offload.tx & ENA_ADMIN_FEATURE_OFFLOAD_DESC_TSO_IPV6_MASK)
drivers/net/ethernet/amazon/ena/ena_netdev.c
3772
if (feat->offload.tx & ENA_ADMIN_FEATURE_OFFLOAD_DESC_TSO_ECN_MASK)
drivers/net/ethernet/amd/sunlance.c
1358
lp->tx = lance_tx_pio;
drivers/net/ethernet/amd/sunlance.c
1370
lp->tx = lance_tx_dvma;
drivers/net/ethernet/amd/sunlance.c
257
void (*tx)(struct net_device *);
drivers/net/ethernet/amd/sunlance.c
834
lp->tx(dev);
drivers/net/ethernet/amd/xgbe/xgbe-desc.c
315
memset(&ring->tx, 0, sizeof(ring->tx));
drivers/net/ethernet/amd/xgbe/xgbe-desc.c
406
memset(&rdata->tx, 0, sizeof(rdata->tx));
drivers/net/ethernet/amd/xgbe/xgbe-desc.c
447
if ((tso && (packet->mss != ring->tx.cur_mss)) ||
drivers/net/ethernet/amd/xgbe/xgbe-desc.c
448
(vlan && (packet->vlan_ctag != ring->tx.cur_vlan_ctag)))
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
1585
ring->tx.xmit_more = 0;
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
1617
if (tso && (packet->mss != ring->tx.cur_mss))
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
1622
if (vlan && (packet->vlan_ctag != ring->tx.cur_vlan_ctag))
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
1669
ring->tx.cur_mss = packet->mss;
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
1689
ring->tx.cur_vlan_ctag = packet->vlan_ctag;
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
1787
rdata->tx.packets = tx_packets;
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
1788
rdata->tx.bytes = tx_bytes;
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
1817
ring->tx.xmit_more = 1;
drivers/net/ethernet/amd/xgbe/xgbe-drv.c
1525
if (skb_shinfo(skb)->gso_size != ring->tx.cur_mss) {
drivers/net/ethernet/amd/xgbe/xgbe-drv.c
1547
if (skb_vlan_tag_get(skb) != ring->tx.cur_vlan_ctag)
drivers/net/ethernet/amd/xgbe/xgbe-drv.c
173
ring->tx.queue_stopped = 1;
drivers/net/ethernet/amd/xgbe/xgbe-drv.c
178
if (ring->tx.xmit_more)
drivers/net/ethernet/amd/xgbe/xgbe-drv.c
2189
tx_packets += rdata->tx.packets;
drivers/net/ethernet/amd/xgbe/xgbe-drv.c
2190
tx_bytes += rdata->tx.bytes;
drivers/net/ethernet/amd/xgbe/xgbe-drv.c
2206
if ((ring->tx.queue_stopped == 1) &&
drivers/net/ethernet/amd/xgbe/xgbe-drv.c
2208
ring->tx.queue_stopped = 0;
drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c
555
unsigned int rx, tx;
drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c
586
tx = __rounddown_pow_of_two(ringparam->tx_pending);
drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c
587
if (tx != ringparam->tx_pending)
drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c
590
tx);
drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c
593
(tx == pdata->tx_desc_count))
drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c
597
pdata->tx_desc_count = tx;
drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c
609
unsigned int rx, tx, combined;
drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c
619
tx = min(pdata->hw_feat.tx_ch_cnt, pdata->tx_max_channel_count);
drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c
620
tx = min(tx, pdata->channel_irq_count);
drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c
621
tx = min(tx, pdata->tx_max_q_count);
drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c
623
combined = min(rx, tx);
drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c
627
channels->max_tx = tx ? tx - 1 : 0;
drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c
631
tx = pdata->new_tx_ring_count ? : pdata->tx_ring_count;
drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c
633
combined = min(rx, tx);
drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c
635
tx -= combined;
drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c
639
channels->tx_count = tx;
drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c
654
unsigned int rx, rx_curr, tx, tx_curr, combined;
drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c
664
tx = min(pdata->hw_feat.tx_ch_cnt, pdata->tx_max_channel_count);
drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c
665
tx = min(tx, pdata->tx_max_q_count);
drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c
666
tx = min(tx, pdata->channel_irq_count);
drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c
668
combined = min(rx, tx);
drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c
711
if ((channels->combined_count + channels->tx_count) > tx) {
drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c
714
channels->combined_count + channels->tx_count, tx);
drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c
720
tx = channels->combined_count + channels->tx_count;
drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c
725
if ((rx == rx_curr) && (tx == tx_curr))
drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c
729
pdata->new_tx_ring_count = tx;
drivers/net/ethernet/amd/xgbe/xgbe.h
352
struct xgbe_tx_ring_data tx; /* Tx-related data */
drivers/net/ethernet/amd/xgbe/xgbe.h
410
} tx;
drivers/net/ethernet/apm/xgene/xgene_enet_hw.c
601
u32 *rx, u32 *tx)
drivers/net/ethernet/apm/xgene/xgene_enet_hw.c
607
*tx = ECM_DROP_COUNT(count);
drivers/net/ethernet/apm/xgene/xgene_enet_main.h
149
void (*get_drop_cnt)(struct xgene_enet_pdata *pdata, u32 *rx, u32 *tx);
drivers/net/ethernet/apm/xgene/xgene_enet_sgmac.c
87
u32 *rx, u32 *tx)
drivers/net/ethernet/apm/xgene/xgene_enet_sgmac.c
96
*tx = ECM_DROP_COUNT(count);
drivers/net/ethernet/apm/xgene/xgene_enet_xgmac.c
178
u32 *rx, u32 *tx)
drivers/net/ethernet/apm/xgene/xgene_enet_xgmac.c
184
*tx = ECM_DROP_COUNT(count);
drivers/net/ethernet/aquantia/atlantic/aq_ptp.c
787
u64_stats_update_begin(&ring->stats.tx.syncp);
drivers/net/ethernet/aquantia/atlantic/aq_ptp.c
788
++ring->stats.tx.packets;
drivers/net/ethernet/aquantia/atlantic/aq_ptp.c
789
ring->stats.tx.bytes += skb->len;
drivers/net/ethernet/aquantia/atlantic/aq_ptp.c
790
u64_stats_update_end(&ring->stats.tx.syncp);
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
240
u64_stats_init(&self->stats.tx.syncp);
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
268
u64_stats_update_begin(&ring->stats.tx.syncp);
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
269
ring->stats.tx.queue_restarts++;
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
270
u64_stats_update_end(&ring->stats.tx.syncp);
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
315
u64_stats_update_begin(&self->stats.tx.syncp);
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
316
++self->stats.tx.packets;
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
317
self->stats.tx.bytes += buff->skb->len;
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
318
u64_stats_update_end(&self->stats.tx.syncp);
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
321
u64_stats_update_begin(&self->stats.tx.syncp);
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
322
++self->stats.tx.packets;
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
323
self->stats.tx.bytes += xdp_get_frame_len(buff->xdpf);
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
324
u64_stats_update_end(&self->stats.tx.syncp);
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
971
start = u64_stats_fetch_begin(&self->stats.tx.syncp);
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
972
data[count] = self->stats.tx.packets;
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
973
data[++count] = self->stats.tx.queue_restarts;
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
974
} while (u64_stats_fetch_retry(&self->stats.tx.syncp, start));
drivers/net/ethernet/aquantia/atlantic/aq_ring.h
127
struct aq_ring_stats_tx_s tx;
drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_llh.c
217
u32 tx)
drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_llh.c
251
aq_hw_write_reg_bit(aq_hw, itr_imr_txten_adr[tx],
drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_llh.c
252
itr_imr_txten_msk[tx],
drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_llh.c
253
itr_imr_txten_shift[tx],
drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_llh.c
297
void hw_atl_itr_irq_map_tx_set(struct aq_hw_s *aq_hw, u32 irq_map_tx, u32 tx)
drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_llh.c
331
aq_hw_write_reg_bit(aq_hw, itr_imr_txt_adr[tx],
drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_llh.c
332
itr_imr_txt_msk[tx],
drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_llh.c
333
itr_imr_txt_shift[tx],
drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_llh.h
146
u32 tx);
drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_llh.h
152
void hw_atl_itr_irq_map_tx_set(struct aq_hw_s *aq_hw, u32 irq_map_tx, u32 tx);
drivers/net/ethernet/atheros/ag71xx.c
1135
if (ring->buf[i].tx.skb) {
drivers/net/ethernet/atheros/ag71xx.c
1136
bytes_compl += ring->buf[i].tx.len;
drivers/net/ethernet/atheros/ag71xx.c
1138
dev_kfree_skb_any(ring->buf[i].tx.skb);
drivers/net/ethernet/atheros/ag71xx.c
1140
ring->buf[i].tx.skb = NULL;
drivers/net/ethernet/atheros/ag71xx.c
1164
ring->buf[i].tx.skb = NULL;
drivers/net/ethernet/atheros/ag71xx.c
1301
struct ag71xx_ring *tx = &ag->tx_ring;
drivers/net/ethernet/atheros/ag71xx.c
1305
ring_size = BIT(tx->order) + BIT(rx->order);
drivers/net/ethernet/atheros/ag71xx.c
1306
tx_size = BIT(tx->order);
drivers/net/ethernet/atheros/ag71xx.c
1308
tx->buf = kzalloc_objs(*tx->buf, ring_size);
drivers/net/ethernet/atheros/ag71xx.c
1309
if (!tx->buf)
drivers/net/ethernet/atheros/ag71xx.c
1312
tx->descs_cpu = dma_alloc_coherent(&ag->pdev->dev,
drivers/net/ethernet/atheros/ag71xx.c
1314
&tx->descs_dma, GFP_KERNEL);
drivers/net/ethernet/atheros/ag71xx.c
1315
if (!tx->descs_cpu) {
drivers/net/ethernet/atheros/ag71xx.c
1316
kfree(tx->buf);
drivers/net/ethernet/atheros/ag71xx.c
1317
tx->buf = NULL;
drivers/net/ethernet/atheros/ag71xx.c
1321
rx->buf = &tx->buf[tx_size];
drivers/net/ethernet/atheros/ag71xx.c
1322
rx->descs_cpu = ((void *)tx->descs_cpu) + tx_size * AG71XX_DESC_SIZE;
drivers/net/ethernet/atheros/ag71xx.c
1323
rx->descs_dma = tx->descs_dma + tx_size * AG71XX_DESC_SIZE;
drivers/net/ethernet/atheros/ag71xx.c
1331
struct ag71xx_ring *tx = &ag->tx_ring;
drivers/net/ethernet/atheros/ag71xx.c
1335
ring_size = BIT(tx->order) + BIT(rx->order);
drivers/net/ethernet/atheros/ag71xx.c
1337
if (tx->descs_cpu)
drivers/net/ethernet/atheros/ag71xx.c
1339
tx->descs_cpu, tx->descs_dma);
drivers/net/ethernet/atheros/ag71xx.c
1341
kfree(tx->buf);
drivers/net/ethernet/atheros/ag71xx.c
1343
tx->descs_cpu = NULL;
drivers/net/ethernet/atheros/ag71xx.c
1345
tx->buf = NULL;
drivers/net/ethernet/atheros/ag71xx.c
1534
ring->buf[i].tx.len = skb->len;
drivers/net/ethernet/atheros/ag71xx.c
1535
ring->buf[i].tx.skb = skb;
drivers/net/ethernet/atheros/ag71xx.c
309
} tx;
drivers/net/ethernet/atheros/ag71xx.c
789
skb = ring->buf[i].tx.skb;
drivers/net/ethernet/atheros/ag71xx.c
809
ring->buf[i].tx.skb = NULL;
drivers/net/ethernet/atheros/ag71xx.c
811
bytes_compl += ring->buf[i].tx.len;
drivers/net/ethernet/atheros/ag71xx.c
846
u32 rx, tx;
drivers/net/ethernet/atheros/ag71xx.c
851
tx = ag71xx_rr(ag, AG71XX_REG_TX_CTRL) & TX_CTRL_TXE;
drivers/net/ethernet/atheros/ag71xx.c
852
if (!rx && !tx)
drivers/net/ethernet/broadcom/bcmsysport.c
233
STAT_MIB_TX("tx_64_octets", mib.tx.pkt_cnt.cnt_64),
drivers/net/ethernet/broadcom/bcmsysport.c
234
STAT_MIB_TX("tx_65_127_oct", mib.tx.pkt_cnt.cnt_127),
drivers/net/ethernet/broadcom/bcmsysport.c
235
STAT_MIB_TX("tx_128_255_oct", mib.tx.pkt_cnt.cnt_255),
drivers/net/ethernet/broadcom/bcmsysport.c
236
STAT_MIB_TX("tx_256_511_oct", mib.tx.pkt_cnt.cnt_511),
drivers/net/ethernet/broadcom/bcmsysport.c
237
STAT_MIB_TX("tx_512_1023_oct", mib.tx.pkt_cnt.cnt_1023),
drivers/net/ethernet/broadcom/bcmsysport.c
238
STAT_MIB_TX("tx_1024_1518_oct", mib.tx.pkt_cnt.cnt_1518),
drivers/net/ethernet/broadcom/bcmsysport.c
239
STAT_MIB_TX("tx_vlan_1519_1522_oct", mib.tx.pkt_cnt.cnt_mgv),
drivers/net/ethernet/broadcom/bcmsysport.c
240
STAT_MIB_TX("tx_1522_2047_oct", mib.tx.pkt_cnt.cnt_2047),
drivers/net/ethernet/broadcom/bcmsysport.c
241
STAT_MIB_TX("tx_2048_4095_oct", mib.tx.pkt_cnt.cnt_4095),
drivers/net/ethernet/broadcom/bcmsysport.c
242
STAT_MIB_TX("tx_4096_9216_oct", mib.tx.pkt_cnt.cnt_9216),
drivers/net/ethernet/broadcom/bcmsysport.c
243
STAT_MIB_TX("tx_pkts", mib.tx.pkts),
drivers/net/ethernet/broadcom/bcmsysport.c
244
STAT_MIB_TX("tx_multicast", mib.tx.mca),
drivers/net/ethernet/broadcom/bcmsysport.c
245
STAT_MIB_TX("tx_broadcast", mib.tx.bca),
drivers/net/ethernet/broadcom/bcmsysport.c
246
STAT_MIB_TX("tx_pause", mib.tx.pf),
drivers/net/ethernet/broadcom/bcmsysport.c
247
STAT_MIB_TX("tx_control", mib.tx.cf),
drivers/net/ethernet/broadcom/bcmsysport.c
248
STAT_MIB_TX("tx_fcs_err", mib.tx.fcs),
drivers/net/ethernet/broadcom/bcmsysport.c
249
STAT_MIB_TX("tx_oversize", mib.tx.ovr),
drivers/net/ethernet/broadcom/bcmsysport.c
250
STAT_MIB_TX("tx_defer", mib.tx.drf),
drivers/net/ethernet/broadcom/bcmsysport.c
251
STAT_MIB_TX("tx_excess_defer", mib.tx.edf),
drivers/net/ethernet/broadcom/bcmsysport.c
252
STAT_MIB_TX("tx_single_col", mib.tx.scl),
drivers/net/ethernet/broadcom/bcmsysport.c
253
STAT_MIB_TX("tx_multi_col", mib.tx.mcl),
drivers/net/ethernet/broadcom/bcmsysport.c
254
STAT_MIB_TX("tx_late_col", mib.tx.lcl),
drivers/net/ethernet/broadcom/bcmsysport.c
255
STAT_MIB_TX("tx_excess_col", mib.tx.ecl),
drivers/net/ethernet/broadcom/bcmsysport.c
256
STAT_MIB_TX("tx_frags", mib.tx.frg),
drivers/net/ethernet/broadcom/bcmsysport.c
257
STAT_MIB_TX("tx_total_col", mib.tx.ncl),
drivers/net/ethernet/broadcom/bcmsysport.c
258
STAT_MIB_TX("tx_jabber", mib.tx.jbr),
drivers/net/ethernet/broadcom/bcmsysport.c
259
STAT_MIB_TX("tx_bytes", mib.tx.bytes),
drivers/net/ethernet/broadcom/bcmsysport.c
260
STAT_MIB_TX("tx_good_pkts", mib.tx.pok),
drivers/net/ethernet/broadcom/bcmsysport.c
261
STAT_MIB_TX("tx_unicast", mib.tx.uc),
drivers/net/ethernet/broadcom/bcmsysport.h
560
struct bcm_sysport_tx_counters tx;
drivers/net/ethernet/broadcom/bnge/bnge_hwrm_lib.c
392
u16 rx, tx;
drivers/net/ethernet/broadcom/bnge/bnge_hwrm_lib.c
417
tx = hw_resc->resv_tx_rings;
drivers/net/ethernet/broadcom/bnge/bnge_hwrm_lib.c
420
if (cp < (rx + tx)) {
drivers/net/ethernet/broadcom/bnge/bnge_hwrm_lib.c
421
rc = bnge_fix_rings_count(&rx, &tx, cp, false);
drivers/net/ethernet/broadcom/bnge/bnge_hwrm_lib.c
427
hw_resc->resv_tx_rings = tx;
drivers/net/ethernet/broadcom/bnge/bnge_hwrm_lib.c
449
enables |= hwr->tx ? FUNC_CFG_REQ_ENABLES_NUM_TX_RINGS : 0;
drivers/net/ethernet/broadcom/bnge/bnge_hwrm_lib.c
450
req->num_tx_rings = cpu_to_le16(hwr->tx);
drivers/net/ethernet/broadcom/bnge/bnge_netdev.c
267
int rx = 0, tx = 0;
drivers/net/ethernet/broadcom/bnge/bnge_netdev.c
28
#define BNGE_RING_TO_TC_OFF(bd, tx) \
drivers/net/ethernet/broadcom/bnge/bnge_netdev.c
287
tx = 1;
drivers/net/ethernet/broadcom/bnge/bnge_netdev.c
29
((tx) % (bd)->tx_nr_rings_per_tc)
drivers/net/ethernet/broadcom/bnge/bnge_netdev.c
31
#define BNGE_RING_TO_TC(bd, tx) \
drivers/net/ethernet/broadcom/bnge/bnge_netdev.c
317
if (tx)
drivers/net/ethernet/broadcom/bnge/bnge_netdev.c
32
((tx) / (bd)->tx_nr_rings_per_tc)
drivers/net/ethernet/broadcom/bnge/bnge_resc.c
137
int bnge_fix_rings_count(u16 *rx, u16 *tx, u16 max, bool shared)
drivers/net/ethernet/broadcom/bnge/bnge_resc.c
139
u16 _rx = *rx, _tx = *tx;
drivers/net/ethernet/broadcom/bnge/bnge_resc.c
143
*tx = min_t(u16, _tx, max);
drivers/net/ethernet/broadcom/bnge/bnge_resc.c
15
static u16 bnge_num_tx_to_cp(struct bnge_dev *bd, u16 tx)
drivers/net/ethernet/broadcom/bnge/bnge_resc.c
154
*tx = _tx;
drivers/net/ethernet/broadcom/bnge/bnge_resc.c
161
u16 *tx, u16 max_nq, bool sh)
drivers/net/ethernet/broadcom/bnge/bnge_resc.c
163
u16 tx_chunks = bnge_num_tx_to_cp(bd, *tx);
drivers/net/ethernet/broadcom/bnge/bnge_resc.c
165
if (tx_chunks != *tx) {
drivers/net/ethernet/broadcom/bnge/bnge_resc.c
172
*tx = bnge_num_cp_to_tx(bd, tx_chunks);
drivers/net/ethernet/broadcom/bnge/bnge_resc.c
176
return bnge_fix_rings_count(rx, tx, max_nq, sh);
drivers/net/ethernet/broadcom/bnge/bnge_resc.c
22
return tx / tcs;
drivers/net/ethernet/broadcom/bnge/bnge_resc.c
228
hwr->tx = hw_resc->resv_tx_rings;
drivers/net/ethernet/broadcom/bnge/bnge_resc.c
240
return hwr->tx && hwr->rx && hwr->nq && hwr->grp && hwr->vnic &&
drivers/net/ethernet/broadcom/bnge/bnge_resc.c
295
hwr.tx = bd->tx_nr_rings;
drivers/net/ethernet/broadcom/bnge/bnge_resc.c
299
hwr.cmpl = hwr.rx + hwr.tx;
drivers/net/ethernet/broadcom/bnge/bnge_resc.c
331
rc = bnge_adjust_rings(bd, &rx_rings, &hwr.tx, hwr.nq, sh);
drivers/net/ethernet/broadcom/bnge/bnge_resc.c
334
tx_cp = hwr.tx;
drivers/net/ethernet/broadcom/bnge/bnge_resc.c
336
bd->tx_nr_rings = hwr.tx;
drivers/net/ethernet/broadcom/bnge/bnge_resc.c
466
u16 rx, tx, nq;
drivers/net/ethernet/broadcom/bnge/bnge_resc.c
468
_bnge_get_max_rings(bd, &rx, &tx, &nq);
drivers/net/ethernet/broadcom/bnge/bnge_resc.c
470
*max_tx = tx;
drivers/net/ethernet/broadcom/bnge/bnge_resc.c
471
if (!rx || !tx || !nq)
drivers/net/ethernet/broadcom/bnge/bnge_resc.h
47
u16 tx;
drivers/net/ethernet/broadcom/bnge/bnge_resc.h
69
int bnge_fix_rings_count(u16 *rx, u16 *tx, u16 max, bool shared);
drivers/net/ethernet/broadcom/bnx2.c
7322
bnx2_change_ring_size(struct bnx2 *bp, u32 rx, u32 tx, bool reset_irq)
drivers/net/ethernet/broadcom/bnx2.c
7341
bp->tx_ring_size = tx;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
1995
int rc, tx, rx;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
1997
tx = BNX2X_NUM_ETH_QUEUES(bp) * bp->max_cos;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
2003
tx++;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
2006
rc = netif_set_real_num_tx_queues(bp->dev, tx);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
2018
tx, rx);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_hsi.h
3742
struct client_init_tx_data tx;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_hsi.h
4815
struct client_init_tx_data tx;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c
12151
u32 rx = 0, tx = 0, i;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c
12162
tx = REG_RD(bp, shmem_base +
drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c
12170
tx = REG_RD(bp, shmem_base +
drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c
12178
phy->tx_preemphasis[i << 1] = ((tx>>16) & 0xffff);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c
12179
phy->tx_preemphasis[(i << 1) + 1] = (tx & 0xffff);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c
8749
__set_bit(BNX2X_Q_FLG_HC, &init_params->tx.flags);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c
8755
__set_bit(BNX2X_Q_FLG_HC_EN, &init_params->tx.flags);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c
8760
init_params->tx.hc_rate = bp->tx_ticks ?
drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c
8764
init_params->rx.fw_sb_id = init_params->tx.fw_sb_id =
drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c
8772
init_params->tx.sb_cq_index = HC_INDEX_ETH_FIRST_TX_CQ_CONS;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c
4958
&data->tx,
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c
4983
&data->tx,
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c
4988
data->tx.tx_bd_page_base.lo,
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c
4989
data->tx.tx_bd_page_base.hi);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c
5013
test_bit(BNX2X_Q_FLG_HC, &init->tx.flags)) {
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c
5014
hc_usec = init->tx.hc_rate ? 1000000 / init->tx.hc_rate : 0;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c
5016
bnx2x_update_coalesce_sb_index(bp, init->tx.fw_sb_id,
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c
5017
init->tx.sb_cq_index,
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c
5018
!test_bit(BNX2X_Q_FLG_HC_EN, &init->tx.flags),
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.h
886
} tx;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c
136
init_params->tx.sb_cq_index,
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c
137
init_params->tx.hc_rate,
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c
181
if (test_bit(BNX2X_Q_FLG_HC, &init_p->tx.flags))
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c
182
__set_bit(BNX2X_Q_FLG_HC_EN, &init_p->tx.flags);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c
186
init_p->tx.fw_sb_id = vf_igu_sb(vf, q->sb_idx);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.c
1567
init_p->tx.hc_rate = setup_q->txq.hc_rate;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.c
1568
init_p->tx.sb_cq_index = setup_q->txq.sb_index;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.c
1571
&init_p->tx.flags);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
11210
static int __bnxt_trim_rings(struct bnxt *bp, int *rx, int *tx, int max,
drivers/net/ethernet/broadcom/bnxt/bnxt.c
11213
int _rx = *rx, _tx = *tx;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
11217
*tx = min_t(int, _tx, max);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
11229
*tx = _tx;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
11234
static int __bnxt_num_tx_to_cp(struct bnxt *bp, int tx, int tx_sets, int tx_xdp)
drivers/net/ethernet/broadcom/bnxt/bnxt.c
11236
return (tx - tx_xdp) / tx_sets + tx_xdp;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
11239
int bnxt_num_tx_to_cp(struct bnxt *bp, int tx)
drivers/net/ethernet/broadcom/bnxt/bnxt.c
11245
return __bnxt_num_tx_to_cp(bp, tx, tcs, bp->tx_nr_rings_xdp);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
11256
static int bnxt_trim_rings(struct bnxt *bp, int *rx, int *tx, int max,
drivers/net/ethernet/broadcom/bnxt/bnxt.c
11259
int tx_cp = bnxt_num_tx_to_cp(bp, *tx);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
11261
if (tx_cp != *tx) {
drivers/net/ethernet/broadcom/bnxt/bnxt.c
11268
*tx = bnxt_num_cp_to_tx(bp, tx_cp);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
11271
return __bnxt_trim_rings(bp, rx, tx, max, sh);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
11873
cpr->sw_stats->tx.tx_resets++;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
13481
u64 *tx = bp->port_stats.sw_stats +
drivers/net/ethernet/broadcom/bnxt/bnxt.c
13496
BNXT_GET_TX_PORT_STATS64(tx, tx_total_collisions);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
13498
BNXT_GET_TX_PORT_STATS64(tx, tx_fifo_underruns);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
13499
stats->tx_errors = BNXT_GET_TX_PORT_STATS64(tx, tx_err);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
13520
stats->tx_total_resets += sw_stats->tx.tx_resets;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
14723
int bnxt_check_rings(struct bnxt *bp, int tx, int rx, bool sh, int tcs,
drivers/net/ethernet/broadcom/bnxt/bnxt.c
14743
hwr.tx = tx * tx_sets + tx_xdp;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
14744
if (max_tx < hwr.tx)
drivers/net/ethernet/broadcom/bnxt/bnxt.c
14749
tx_cp = __bnxt_num_tx_to_cp(bp, hwr.tx, tx_sets, tx_xdp);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
14761
hwr.cp_p5 = hwr.tx + rx;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
15977
struct netdev_queue_stats_tx *tx)
drivers/net/ethernet/broadcom/bnxt/bnxt.c
15987
tx->packets = bp->net_stats_prev.tx_packets;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
15988
tx->bytes = bp->net_stats_prev.tx_bytes;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
16478
int rx, tx, cp;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
16480
_bnxt_get_max_rings(bp, &rx, &tx, &cp);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
16482
*max_tx = tx;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
16483
if (!rx || !tx || !cp)
drivers/net/ethernet/broadcom/bnxt/bnxt.c
4003
#define BNXT_RING_TO_TC_OFF(bp, tx) \
drivers/net/ethernet/broadcom/bnxt/bnxt.c
4004
((tx) % (bp)->tx_nr_rings_per_tc)
drivers/net/ethernet/broadcom/bnxt/bnxt.c
4006
#define BNXT_RING_TO_TC(bp, tx) \
drivers/net/ethernet/broadcom/bnxt/bnxt.c
4007
((tx) / (bp)->tx_nr_rings_per_tc)
drivers/net/ethernet/broadcom/bnxt/bnxt.c
4203
int rx = 0, tx = 0;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
4227
tx = 1;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
4231
tx = 1;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
4258
if (tx)
drivers/net/ethernet/broadcom/bnxt/bnxt.c
4900
int rx, tx;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
4902
bnxt_get_max_rings(bp, &rx, &tx, true);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
7655
static int __bnxt_trim_rings(struct bnxt *bp, int *rx, int *tx, int max,
drivers/net/ethernet/broadcom/bnxt/bnxt.c
7657
static int bnxt_trim_rings(struct bnxt *bp, int *rx, int *tx, int max,
drivers/net/ethernet/broadcom/bnxt/bnxt.c
7696
int tx = hw_resc->resv_tx_rings;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
7700
if (cp < (rx + tx)) {
drivers/net/ethernet/broadcom/bnxt/bnxt.c
7701
rc = __bnxt_trim_rings(bp, &rx, &tx, cp, false);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
7707
hw_resc->resv_tx_rings = tx;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
7755
enables |= hwr->tx ? FUNC_CFG_REQ_ENABLES_NUM_TX_RINGS : 0;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
7756
req->num_tx_rings = cpu_to_le16(hwr->tx);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
7798
enables |= hwr->tx ? FUNC_VF_CFG_REQ_ENABLES_NUM_TX_RINGS : 0;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
7815
req->num_tx_rings = cpu_to_le16(hwr->tx);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
7851
bp->hw_resc.resv_tx_rings = hwr->tx;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
7863
bp->hw_resc.resv_tx_rings = hwr->tx;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
7957
hwr->tx = bp->tx_nr_rings;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
7982
if (hw_resc->resv_tx_rings != hwr.tx && bp->hwrm_spec_code >= 0x10601)
drivers/net/ethernet/broadcom/bnxt/bnxt.c
8011
hwr->tx = hw_resc->resv_tx_rings;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
8026
return hwr->tx && hwr->rx && hwr->cp && hwr->grp && hwr->vnic &&
drivers/net/ethernet/broadcom/bnxt/bnxt.c
8058
hwr.tx = bp->tx_nr_rings;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
8063
hwr.cp_p5 = hwr.rx + hwr.tx;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
8101
rc = bnxt_trim_rings(bp, &rx_rings, &hwr.tx, hwr.cp, sh);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
8104
tx_cp = bnxt_num_tx_to_cp(bp, hwr.tx);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
8106
if (hwr.tx != bp->tx_nr_rings) {
drivers/net/ethernet/broadcom/bnxt/bnxt.c
8109
hwr.tx, bp->tx_nr_rings);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
8111
bp->tx_nr_rings = hwr.tx;
drivers/net/ethernet/broadcom/bnxt/bnxt.h
1146
struct bnxt_tx_sw_stats tx;
drivers/net/ethernet/broadcom/bnxt/bnxt.h
1332
int tx;
drivers/net/ethernet/broadcom/bnxt/bnxt.h
2945
int bnxt_num_tx_to_cp(struct bnxt *bp, int tx);
drivers/net/ethernet/broadcom/bnxt/bnxt.h
2984
int bnxt_check_rings(struct bnxt *bp, int tx, int rx, bool sh, int tcs,
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
3371
u64 *rx, *tx;
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
3377
tx = bp->port_stats.sw_stats + BNXT_TX_PORT_STATS_BYTE_OFFSET / 8;
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
3380
epstat->tx_pause_frames = BNXT_GET_TX_PORT_STATS64(tx, tx_pause_frames);
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
526
int rx, tx, cmn;
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
530
tx = NUM_RING_TX_HW_STATS;
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
533
tx * (bp->tx_nr_rings_xdp + bp->tx_nr_rings_per_tc) +
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
5548
u64 *rx, *tx;
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
5554
tx = bp->port_stats.sw_stats + BNXT_TX_PORT_STATS_BYTE_OFFSET / 8;
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
5559
BNXT_GET_TX_PORT_STATS64(tx, tx_good_frames);
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
5601
u64 *rx, *tx;
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
5607
tx = bp->port_stats.sw_stats + BNXT_TX_PORT_STATS_BYTE_OFFSET / 8;
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
5634
BNXT_GET_TX_PORT_STATS64(tx, tx_64b_frames);
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
5636
BNXT_GET_TX_PORT_STATS64(tx, tx_65b_127b_frames);
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
5638
BNXT_GET_TX_PORT_STATS64(tx, tx_128b_255b_frames);
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
5640
BNXT_GET_TX_PORT_STATS64(tx, tx_256b_511b_frames);
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
5642
BNXT_GET_TX_PORT_STATS64(tx, tx_512b_1023b_frames);
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
5644
BNXT_GET_TX_PORT_STATS64(tx, tx_1024b_1518b_frames);
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
5646
BNXT_GET_TX_PORT_STATS64(tx, tx_1519b_2047b_frames);
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
5648
BNXT_GET_TX_PORT_STATS64(tx, tx_2048b_4095b_frames);
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
5650
BNXT_GET_TX_PORT_STATS64(tx, tx_4096b_9216b_frames);
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
5652
BNXT_GET_TX_PORT_STATS64(tx, tx_9217b_16383b_frames);
drivers/net/ethernet/broadcom/cnic.c
4961
data->tx.tx_bd_page_base.hi = cpu_to_le32(val);
drivers/net/ethernet/broadcom/cnic.c
4966
data->tx.tx_bd_page_base.lo = cpu_to_le32(val);
drivers/net/ethernet/broadcom/cnic.c
4969
data->tx.tx_sb_index_number = HC_SP_INDEX_ETH_ISCSI_CQ_CONS;
drivers/net/ethernet/broadcom/cnic.c
4970
data->tx.tx_status_block_id = BNX2X_DEF_SB_ID;
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1115
STAT_GENET_MIB_TX("tx_64_octets", mib.tx.pkt_cnt.cnt_64),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1116
STAT_GENET_MIB_TX("tx_65_127_oct", mib.tx.pkt_cnt.cnt_127),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1117
STAT_GENET_MIB_TX("tx_128_255_oct", mib.tx.pkt_cnt.cnt_255),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1118
STAT_GENET_MIB_TX("tx_256_511_oct", mib.tx.pkt_cnt.cnt_511),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1119
STAT_GENET_MIB_TX("tx_512_1023_oct", mib.tx.pkt_cnt.cnt_1023),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1120
STAT_GENET_MIB_TX("tx_1024_1518_oct", mib.tx.pkt_cnt.cnt_1518),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1121
STAT_GENET_MIB_TX("tx_vlan_1519_1522_oct", mib.tx.pkt_cnt.cnt_mgv),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1122
STAT_GENET_MIB_TX("tx_1522_2047_oct", mib.tx.pkt_cnt.cnt_2047),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1123
STAT_GENET_MIB_TX("tx_2048_4095_oct", mib.tx.pkt_cnt.cnt_4095),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1124
STAT_GENET_MIB_TX("tx_4096_9216_oct", mib.tx.pkt_cnt.cnt_9216),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1125
STAT_GENET_MIB_TX("tx_pkts", mib.tx.pkts),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1126
STAT_GENET_MIB_TX("tx_multicast", mib.tx.mca),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1127
STAT_GENET_MIB_TX("tx_broadcast", mib.tx.bca),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1128
STAT_GENET_MIB_TX("tx_pause", mib.tx.pf),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1129
STAT_GENET_MIB_TX("tx_control", mib.tx.cf),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1130
STAT_GENET_MIB_TX("tx_fcs_err", mib.tx.fcs),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1131
STAT_GENET_MIB_TX("tx_oversize", mib.tx.ovr),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1132
STAT_GENET_MIB_TX("tx_defer", mib.tx.drf),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1133
STAT_GENET_MIB_TX("tx_excess_defer", mib.tx.edf),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1134
STAT_GENET_MIB_TX("tx_single_col", mib.tx.scl),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1135
STAT_GENET_MIB_TX("tx_multi_col", mib.tx.mcl),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1136
STAT_GENET_MIB_TX("tx_late_col", mib.tx.lcl),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1137
STAT_GENET_MIB_TX("tx_excess_col", mib.tx.ecl),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1138
STAT_GENET_MIB_TX("tx_frags", mib.tx.frg),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1139
STAT_GENET_MIB_TX("tx_total_col", mib.tx.ncl),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1140
STAT_GENET_MIB_TX("tx_jabber", mib.tx.jbr),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1141
STAT_GENET_MIB_TX("tx_bytes", mib.tx.bytes),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1142
STAT_GENET_MIB_TX("tx_good_pkts", mib.tx.pok),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1143
STAT_GENET_MIB_TX("tx_unicast", mib.tx.uc),
drivers/net/ethernet/broadcom/genet/bcmgenet.h
143
struct bcmgenet_tx_counters tx;
drivers/net/ethernet/broadcom/genet/bcmgenet.h
738
void bcmgenet_phy_pause_set(struct net_device *dev, bool rx, bool tx);
drivers/net/ethernet/broadcom/genet/bcmmii.c
132
void bcmgenet_phy_pause_set(struct net_device *dev, bool rx, bool tx)
drivers/net/ethernet/broadcom/genet/bcmmii.c
138
rx | tx);
drivers/net/ethernet/brocade/bna/bfa_cs.h
51
BFA_SM_TABLE(tx, bna_tx, bna_tx_event, bna_fsm_tx_t)
drivers/net/ethernet/brocade/bna/bna.h
289
void bna_bfi_tx_enet_start_rsp(struct bna_tx *tx,
drivers/net/ethernet/brocade/bna/bna.h
291
void bna_bfi_tx_enet_stop_rsp(struct bna_tx *tx,
drivers/net/ethernet/brocade/bna/bna.h
312
void bna_tx_destroy(struct bna_tx *tx);
drivers/net/ethernet/brocade/bna/bna.h
313
void bna_tx_enable(struct bna_tx *tx);
drivers/net/ethernet/brocade/bna/bna.h
314
void bna_tx_disable(struct bna_tx *tx, enum bna_cleanup_type type,
drivers/net/ethernet/brocade/bna/bna.h
316
void bna_tx_cleanup_complete(struct bna_tx *tx);
drivers/net/ethernet/brocade/bna/bna.h
317
void bna_tx_coalescing_timeo_set(struct bna_tx *tx, int coalescing_timeo);
drivers/net/ethernet/brocade/bna/bna_enet.c
276
struct bna_tx *tx;
drivers/net/ethernet/brocade/bna/bna_enet.c
322
bna_tx_from_rid(bna, msghdr->enet_id, tx);
drivers/net/ethernet/brocade/bna/bna_enet.c
323
if (tx)
drivers/net/ethernet/brocade/bna/bna_enet.c
324
bna_bfi_tx_enet_start_rsp(tx, msghdr);
drivers/net/ethernet/brocade/bna/bna_enet.c
328
bna_tx_from_rid(bna, msghdr->enet_id, tx);
drivers/net/ethernet/brocade/bna/bna_enet.c
329
if (tx)
drivers/net/ethernet/brocade/bna/bna_enet.c
330
bna_bfi_tx_enet_stop_rsp(tx, msghdr);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2758
#define call_tx_stop_cbfn(tx) \
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2760
if ((tx)->stop_cbfn) { \
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2763
cbfn = (tx)->stop_cbfn; \
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2764
cbarg = (tx)->stop_cbarg; \
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2765
(tx)->stop_cbfn = NULL; \
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2766
(tx)->stop_cbarg = NULL; \
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2767
cbfn(cbarg, (tx)); \
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2771
static void bna_tx_mod_cb_tx_stopped(void *tx_mod, struct bna_tx *tx);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2772
static void bna_bfi_tx_enet_start(struct bna_tx *tx);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2773
static void bna_tx_enet_stop(struct bna_tx *tx);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2800
bna_tx_sm_stopped_entry(struct bna_tx *tx)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2802
call_tx_stop_cbfn(tx);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2806
bna_tx_sm_stopped(struct bna_tx *tx, enum bna_tx_event event)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2810
bfa_fsm_set_state(tx, bna_tx_sm_start_wait);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2814
call_tx_stop_cbfn(tx);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2831
bna_tx_sm_start_wait_entry(struct bna_tx *tx)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2833
bna_bfi_tx_enet_start(tx);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2837
bna_tx_sm_start_wait(struct bna_tx *tx, enum bna_tx_event event)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2841
tx->flags &= ~BNA_TX_F_BW_UPDATED;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2842
bfa_fsm_set_state(tx, bna_tx_sm_stop_wait);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2846
tx->flags &= ~BNA_TX_F_BW_UPDATED;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2847
bfa_fsm_set_state(tx, bna_tx_sm_stopped);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2851
if (tx->flags & BNA_TX_F_BW_UPDATED) {
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2852
tx->flags &= ~BNA_TX_F_BW_UPDATED;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2853
bfa_fsm_set_state(tx, bna_tx_sm_prio_stop_wait);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2855
bfa_fsm_set_state(tx, bna_tx_sm_started);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2859
tx->flags |= BNA_TX_F_BW_UPDATED;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2868
bna_tx_sm_started_entry(struct bna_tx *tx)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2871
int is_regular = (tx->type == BNA_TX_T_REGULAR);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2873
list_for_each_entry(txq, &tx->txq_q, qe) {
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2876
bna_ib_start(tx->bna, &txq->ib, is_regular);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2878
tx->tx_resume_cbfn(tx->bna->bnad, tx);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2882
bna_tx_sm_started(struct bna_tx *tx, enum bna_tx_event event)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2886
bfa_fsm_set_state(tx, bna_tx_sm_stop_wait);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2887
tx->tx_stall_cbfn(tx->bna->bnad, tx);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2888
bna_tx_enet_stop(tx);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2892
bfa_fsm_set_state(tx, bna_tx_sm_failed);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2893
tx->tx_stall_cbfn(tx->bna->bnad, tx);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2894
tx->tx_cleanup_cbfn(tx->bna->bnad, tx);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2898
bfa_fsm_set_state(tx, bna_tx_sm_prio_stop_wait);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2907
bna_tx_sm_stop_wait_entry(struct bna_tx *tx)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2912
bna_tx_sm_stop_wait(struct bna_tx *tx, enum bna_tx_event event)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2917
bfa_fsm_set_state(tx, bna_tx_sm_cleanup_wait);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2918
tx->tx_cleanup_cbfn(tx->bna->bnad, tx);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2926
bna_tx_enet_stop(tx);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2939
bna_tx_sm_cleanup_wait_entry(struct bna_tx *tx)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2944
bna_tx_sm_cleanup_wait(struct bna_tx *tx, enum bna_tx_event event)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2953
bfa_fsm_set_state(tx, bna_tx_sm_stopped);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2962
bna_tx_sm_prio_stop_wait_entry(struct bna_tx *tx)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2964
tx->tx_stall_cbfn(tx->bna->bnad, tx);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2965
bna_tx_enet_stop(tx);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2969
bna_tx_sm_prio_stop_wait(struct bna_tx *tx, enum bna_tx_event event)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2973
bfa_fsm_set_state(tx, bna_tx_sm_stop_wait);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2977
bfa_fsm_set_state(tx, bna_tx_sm_failed);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2978
tx->tx_cleanup_cbfn(tx->bna->bnad, tx);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2982
bfa_fsm_set_state(tx, bna_tx_sm_prio_cleanup_wait);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2995
bna_tx_sm_prio_cleanup_wait_entry(struct bna_tx *tx)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2997
tx->tx_cleanup_cbfn(tx->bna->bnad, tx);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3001
bna_tx_sm_prio_cleanup_wait(struct bna_tx *tx, enum bna_tx_event event)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3005
bfa_fsm_set_state(tx, bna_tx_sm_cleanup_wait);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3009
bfa_fsm_set_state(tx, bna_tx_sm_failed);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3017
bfa_fsm_set_state(tx, bna_tx_sm_start_wait);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3026
bna_tx_sm_failed_entry(struct bna_tx *tx)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3031
bna_tx_sm_failed(struct bna_tx *tx, enum bna_tx_event event)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3035
bfa_fsm_set_state(tx, bna_tx_sm_quiesce_wait);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3039
bfa_fsm_set_state(tx, bna_tx_sm_cleanup_wait);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3047
bfa_fsm_set_state(tx, bna_tx_sm_stopped);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3056
bna_tx_sm_quiesce_wait_entry(struct bna_tx *tx)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3061
bna_tx_sm_quiesce_wait(struct bna_tx *tx, enum bna_tx_event event)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3065
bfa_fsm_set_state(tx, bna_tx_sm_cleanup_wait);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3069
bfa_fsm_set_state(tx, bna_tx_sm_failed);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3073
bfa_fsm_set_state(tx, bna_tx_sm_start_wait);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3086
bna_bfi_tx_enet_start(struct bna_tx *tx)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3088
struct bfi_enet_tx_cfg_req *cfg_req = &tx->bfi_enet_cmd.cfg_req;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3093
BFI_ENET_H2I_TX_CFG_SET_REQ, 0, tx->rid);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3097
cfg_req->num_queues = tx->num_txq;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3098
for (i = 0; i < tx->num_txq; i++) {
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3100
: list_first_entry(&tx->txq_q, struct bna_txq, qe);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3125
cfg_req->tx_cfg.vlan_id = htons((u16)tx->txf_vlan_id);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3129
bfa_msgq_cmd_set(&tx->msgq_cmd, NULL, NULL,
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3131
bfa_msgq_cmd_post(&tx->bna->msgq, &tx->msgq_cmd);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3135
bna_bfi_tx_enet_stop(struct bna_tx *tx)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3137
struct bfi_enet_req *req = &tx->bfi_enet_cmd.req;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3140
BFI_ENET_H2I_TX_CFG_CLR_REQ, 0, tx->rid);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3143
bfa_msgq_cmd_set(&tx->msgq_cmd, NULL, NULL, sizeof(struct bfi_enet_req),
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3145
bfa_msgq_cmd_post(&tx->bna->msgq, &tx->msgq_cmd);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3149
bna_tx_enet_stop(struct bna_tx *tx)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3154
list_for_each_entry(txq, &tx->txq_q, qe)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3155
bna_ib_stop(tx->bna, &txq->ib);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3157
bna_bfi_tx_enet_stop(tx);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3199
struct bna_tx *tx = NULL;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3204
tx = list_first_entry(&tx_mod->tx_free_q, struct bna_tx, qe);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3206
tx = list_last_entry(&tx_mod->tx_free_q, struct bna_tx, qe);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3207
list_del(&tx->qe);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3208
tx->type = type;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3210
return tx;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3214
bna_tx_free(struct bna_tx *tx)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3216
struct bna_tx_mod *tx_mod = &tx->bna->tx_mod;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3220
while (!list_empty(&tx->txq_q)) {
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3221
txq = list_first_entry(&tx->txq_q, struct bna_txq, qe);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3223
txq->tx = NULL;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3228
if (qe == &tx->qe) {
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3229
list_del(&tx->qe);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3234
tx->bna = NULL;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3235
tx->priv = NULL;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3238
if (((struct bna_tx *)qe)->rid < tx->rid)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3241
list_add(&tx->qe, qe);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3245
bna_tx_start(struct bna_tx *tx)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3247
tx->flags |= BNA_TX_F_ENET_STARTED;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3248
if (tx->flags & BNA_TX_F_ENABLED)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3249
bfa_fsm_send_event(tx, TX_E_START);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3253
bna_tx_stop(struct bna_tx *tx)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3255
tx->stop_cbfn = bna_tx_mod_cb_tx_stopped;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3256
tx->stop_cbarg = &tx->bna->tx_mod;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3258
tx->flags &= ~BNA_TX_F_ENET_STARTED;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3259
bfa_fsm_send_event(tx, TX_E_STOP);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3263
bna_tx_fail(struct bna_tx *tx)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3265
tx->flags &= ~BNA_TX_F_ENET_STARTED;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3266
bfa_fsm_send_event(tx, TX_E_FAIL);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3270
bna_bfi_tx_enet_start_rsp(struct bna_tx *tx, struct bfi_msgq_mhdr *msghdr)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3272
struct bfi_enet_tx_cfg_rsp *cfg_rsp = &tx->bfi_enet_cmd.cfg_rsp;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3276
bfa_msgq_rsp_copy(&tx->bna->msgq, (u8 *)cfg_rsp,
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3279
tx->hw_id = cfg_rsp->hw_id;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3281
for (i = 0, txq = list_first_entry(&tx->txq_q, struct bna_txq, qe);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3282
i < tx->num_txq; i++, txq = list_next_entry(txq, qe)) {
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3285
tx->bna->pcidev.pci_bar_kva
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3288
tx->bna->pcidev.pci_bar_kva
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3297
bfa_fsm_send_event(tx, TX_E_STARTED);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3301
bna_bfi_tx_enet_stop_rsp(struct bna_tx *tx, struct bfi_msgq_mhdr *msghdr)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3303
bfa_fsm_send_event(tx, TX_E_STOPPED);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3309
struct bna_tx *tx;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3311
list_for_each_entry(tx, &tx_mod->tx_active_q, qe)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3312
bfa_fsm_send_event(tx, TX_E_BW_UPDATE);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3370
struct bna_tx *tx;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3388
tx = bna_tx_get(tx_mod, tx_cfg->tx_type);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3389
if (!tx)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3391
tx->bna = bna;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3392
tx->priv = priv;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3396
INIT_LIST_HEAD(&tx->txq_q);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3402
list_move_tail(&txq->qe, &tx->txq_q);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3403
txq->tx = tx;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3412
tx->tcb_setup_cbfn = tx_cbfn->tcb_setup_cbfn;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3413
tx->tcb_destroy_cbfn = tx_cbfn->tcb_destroy_cbfn;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3415
tx->tx_stall_cbfn = tx_cbfn->tx_stall_cbfn;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3416
tx->tx_resume_cbfn = tx_cbfn->tx_resume_cbfn;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3417
tx->tx_cleanup_cbfn = tx_cbfn->tx_cleanup_cbfn;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3419
list_add_tail(&tx->qe, &tx_mod->tx_active_q);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3421
tx->num_txq = tx_cfg->num_txq;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3423
tx->flags = 0;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3424
if (tx->bna->tx_mod.flags & BNA_TX_MOD_F_ENET_STARTED) {
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3425
switch (tx->type) {
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3427
if (!(tx->bna->tx_mod.flags &
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3429
tx->flags |= BNA_TX_F_ENET_STARTED;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3432
if (tx->bna->tx_mod.flags & BNA_TX_MOD_F_ENET_LOOPBACK)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3433
tx->flags |= BNA_TX_F_ENET_STARTED;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3441
list_for_each_entry(txq, &tx->txq_q, qe) {
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3486
if (tx->tcb_setup_cbfn)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3487
(tx->tcb_setup_cbfn)(bna->bnad, txq->tcb);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3497
tx->txf_vlan_id = 0;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3499
bfa_fsm_set_state(tx, bna_tx_sm_stopped);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3501
tx_mod->rid_mask |= BIT(tx->rid);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3503
return tx;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3506
bna_tx_free(tx);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3511
bna_tx_destroy(struct bna_tx *tx)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3515
list_for_each_entry(txq, &tx->txq_q, qe)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3516
if (tx->tcb_destroy_cbfn)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3517
(tx->tcb_destroy_cbfn)(tx->bna->bnad, txq->tcb);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3519
tx->bna->tx_mod.rid_mask &= ~BIT(tx->rid);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3520
bna_tx_free(tx);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3524
bna_tx_enable(struct bna_tx *tx)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3526
if (tx->fsm != bna_tx_sm_stopped)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3529
tx->flags |= BNA_TX_F_ENABLED;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3531
if (tx->flags & BNA_TX_F_ENET_STARTED)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3532
bfa_fsm_send_event(tx, TX_E_START);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3536
bna_tx_disable(struct bna_tx *tx, enum bna_cleanup_type type,
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3540
(*cbfn)(tx->bna->bnad, tx);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3544
tx->stop_cbfn = cbfn;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3545
tx->stop_cbarg = tx->bna->bnad;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3547
tx->flags &= ~BNA_TX_F_ENABLED;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3549
bfa_fsm_send_event(tx, TX_E_STOP);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3553
bna_tx_cleanup_complete(struct bna_tx *tx)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3555
bfa_fsm_send_event(tx, TX_E_CLEANUP_DONE);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3559
bna_tx_mod_cb_tx_stopped(void *arg, struct bna_tx *tx)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3585
tx_mod->tx = (struct bna_tx *)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3596
tx_mod->tx[i].rid = i;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3597
list_add_tail(&tx_mod->tx[i].qe, &tx_mod->tx_free_q);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3616
struct bna_tx *tx;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3622
list_for_each_entry(tx, &tx_mod->tx_active_q, qe)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3623
if (tx->type == type)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3624
bna_tx_start(tx);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3630
struct bna_tx *tx;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3639
list_for_each_entry(tx, &tx_mod->tx_active_q, qe)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3640
if (tx->type == type) {
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3642
bna_tx_stop(tx);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3651
struct bna_tx *tx;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3656
list_for_each_entry(tx, &tx_mod->tx_active_q, qe)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3657
bna_tx_fail(tx);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3661
bna_tx_coalescing_timeo_set(struct bna_tx *tx, int coalescing_timeo)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
3665
list_for_each_entry(txq, &tx->txq_q, qe)
drivers/net/ethernet/brocade/bna/bna_types.h
454
struct bna_tx *tx;
drivers/net/ethernet/brocade/bna/bna_types.h
489
void (*stop_cbfn)(void *arg, struct bna_tx *tx);
drivers/net/ethernet/brocade/bna/bna_types.h
523
struct bna_tx *tx; /* BFI_MAX_TXQ entries */
drivers/net/ethernet/brocade/bna/bnad.c
1003
(struct bnad_tx_info *)tcb->txq->tx->priv;
drivers/net/ethernet/brocade/bna/bnad.c
1013
(struct bnad_tx_info *)tcb->txq->tx->priv;
drivers/net/ethernet/brocade/bna/bnad.c
1039
bnad_cb_tx_stall(struct bnad *bnad, struct bna_tx *tx)
drivers/net/ethernet/brocade/bna/bnad.c
1041
struct bnad_tx_info *tx_info = tx->priv;
drivers/net/ethernet/brocade/bna/bnad.c
1057
bnad_cb_tx_resume(struct bnad *bnad, struct bna_tx *tx)
drivers/net/ethernet/brocade/bna/bnad.c
1059
struct bnad_tx_info *tx_info = tx->priv;
drivers/net/ethernet/brocade/bna/bnad.c
1129
bna_tx_cleanup_complete(tx_info->tx);
drivers/net/ethernet/brocade/bna/bnad.c
1134
bnad_cb_tx_cleanup(struct bnad *bnad, struct bna_tx *tx)
drivers/net/ethernet/brocade/bna/bnad.c
1136
struct bnad_tx_info *tx_info = tx->priv;
drivers/net/ethernet/brocade/bna/bnad.c
1918
if (!tx_info->tx)
drivers/net/ethernet/brocade/bna/bnad.c
1923
bna_tx_disable(tx_info->tx, BNA_HARD_CLEANUP, bnad_cb_tx_disabled);
drivers/net/ethernet/brocade/bna/bnad.c
1932
bna_tx_destroy(tx_info->tx);
drivers/net/ethernet/brocade/bna/bnad.c
1935
tx_info->tx = NULL;
drivers/net/ethernet/brocade/bna/bnad.c
1959
struct bna_tx *tx;
drivers/net/ethernet/brocade/bna/bnad.c
1988
tx = bna_tx_create(&bnad->bna, bnad, tx_config, &tx_cbfn, res_info,
drivers/net/ethernet/brocade/bna/bnad.c
1991
if (!tx) {
drivers/net/ethernet/brocade/bna/bnad.c
1995
tx_info->tx = tx;
drivers/net/ethernet/brocade/bna/bnad.c
2008
bna_tx_enable(tx);
drivers/net/ethernet/brocade/bna/bnad.c
2015
bna_tx_destroy(tx_info->tx);
drivers/net/ethernet/brocade/bna/bnad.c
2017
tx_info->tx = NULL;
drivers/net/ethernet/brocade/bna/bnad.c
2299
if (!tx_info->tx)
drivers/net/ethernet/brocade/bna/bnad.c
2302
bna_tx_coalescing_timeo_set(tx_info->tx, bnad->tx_coalescing_timeo);
drivers/net/ethernet/brocade/bna/bnad.c
992
bnad_cb_tx_disabled(void *arg, struct bna_tx *tx)
drivers/net/ethernet/brocade/bna/bnad.h
201
struct bna_tx *tx; /* 1:1 between tx_info & tx */
drivers/net/ethernet/brocade/bna/bnad_ethtool.c
487
if (!bnad->tx_info[i].tx)
drivers/net/ethernet/brocade/bna/bnad_ethtool.c
651
if (!bnad->tx_info[i].tx)
drivers/net/ethernet/brocade/bna/bnad_ethtool.c
696
if (!bnad->tx_info[i].tx)
drivers/net/ethernet/brocade/bna/bnad_ethtool.c
774
if (!bnad->tx_info[i].tx)
drivers/net/ethernet/cadence/macb_main.c
2603
void *tx, *rx;
drivers/net/ethernet/cadence/macb_main.c
2614
tx = dma_alloc_coherent(dev, size, &tx_dma, GFP_KERNEL);
drivers/net/ethernet/cadence/macb_main.c
2615
if (!tx || upper_32_bits(tx_dma) != upper_32_bits(tx_dma + size - 1))
drivers/net/ethernet/cadence/macb_main.c
2618
size, bp->num_queues, (unsigned long)tx_dma, tx);
drivers/net/ethernet/cadence/macb_main.c
2628
queue->tx_ring = tx + macb_tx_ring_size_per_queue(bp) * q;
drivers/net/ethernet/calxeda/xgmac.c
644
static int xgmac_set_flow_ctrl(struct xgmac_priv *priv, int rx, int tx)
drivers/net/ethernet/calxeda/xgmac.c
650
priv->tx_pause = tx;
drivers/net/ethernet/calxeda/xgmac.c
652
if (rx || tx) {
drivers/net/ethernet/calxeda/xgmac.c
655
if (tx)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
2677
#define T3(fmt_spec, s, v) S3(fmt_spec, s, tx[i].v)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
2678
#define T(s, v) S3("u", s, tx[i].v)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
2687
const struct sge_eth_txq *tx = &s->ethtxq[base_qset];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
2703
4*(7-tx[i].dcb_prio)) & 0xf);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
2706
1*(7-tx[i].dcb_prio)) & 0x1);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
2810
const struct sge_eohw_txq *tx = &s->eohw_txq[base_qset];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
2871
const struct sge_eosw_txq *tx;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
2874
tx = &port_mqprio->eosw_txq[4 * r];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
2918
const struct sge_uld_txq *tx;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
2921
tx = &utxq_info->uldtxq[r * 4];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3067
const struct sge_uld_txq *tx;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3071
tx = &utxq_info->uldtxq[r * 4];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3103
const struct sge_ctrl_txq *tx = &s->ctrlq[r * 4];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
283
const struct sge_eth_txq *tx = &adap->sge.ethtxq[p->first_qset];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
292
for (i = 0; i < p->nqsets; i++, rx++, tx++) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
293
s->tso += tx->tso;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
294
s->uso += tx->uso;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
295
s->tx_csum += tx->tx_cso;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
298
s->vlan_ins += tx->vlan_ins;
drivers/net/ethernet/cisco/enic/enic_ethtool.c
334
*(data++) = ((u64 *)&vstats->tx)[enic_tx_stats[i].index];
drivers/net/ethernet/cisco/enic/enic_main.c
913
net_stats->tx_packets = stats->tx.tx_frames_ok;
drivers/net/ethernet/cisco/enic/enic_main.c
914
net_stats->tx_bytes = stats->tx.tx_bytes_ok;
drivers/net/ethernet/cisco/enic/enic_main.c
915
net_stats->tx_errors = stats->tx.tx_errors;
drivers/net/ethernet/cisco/enic/enic_main.c
916
net_stats->tx_dropped = stats->tx.tx_drops;
drivers/net/ethernet/cisco/enic/vnic_stats.h
58
struct vnic_tx_stats tx;
drivers/net/ethernet/cortina/gemini.c
273
static void gmac_set_flow_control(struct net_device *netdev, bool tx, bool rx)
drivers/net/ethernet/cortina/gemini.c
283
if (tx)
drivers/net/ethernet/dec/tulip/interrupt.c
530
int tx = 0;
drivers/net/ethernet/dec/tulip/interrupt.c
648
tx++;
drivers/net/ethernet/dec/tulip/interrupt.c
739
if (tx > maxtx || rx > maxrx || oi > maxoi) {
drivers/net/ethernet/dec/tulip/interrupt.c
742
csr5, tp->nir, tx, rx, oi);
drivers/net/ethernet/engleder/tsnep.h
148
struct tsnep_tx *tx;
drivers/net/ethernet/engleder/tsnep.h
194
struct tsnep_tx tx[TSNEP_MAX_QUEUES];
drivers/net/ethernet/engleder/tsnep_ethtool.c
195
tsnep_stats.tx_packets += adapter->tx[i].packets;
drivers/net/ethernet/engleder/tsnep_ethtool.c
196
tsnep_stats.tx_bytes += adapter->tx[i].bytes;
drivers/net/ethernet/engleder/tsnep_ethtool.c
197
tsnep_stats.tx_dropped += adapter->tx[i].dropped;
drivers/net/ethernet/engleder/tsnep_ethtool.c
231
tsnep_tx_queue_stats.tx_packets += adapter->tx[i].packets;
drivers/net/ethernet/engleder/tsnep_ethtool.c
232
tsnep_tx_queue_stats.tx_bytes += adapter->tx[i].bytes;
drivers/net/ethernet/engleder/tsnep_ethtool.c
233
tsnep_tx_queue_stats.tx_dropped += adapter->tx[i].dropped;
drivers/net/ethernet/engleder/tsnep_ethtool.c
338
if (adapter->queue[i].tx) {
drivers/net/ethernet/engleder/tsnep_hw.h
185
sizeof_field(struct tsnep_tx_desc, tx))
drivers/net/ethernet/engleder/tsnep_hw.h
199
__le64 tx;
drivers/net/ethernet/engleder/tsnep_main.c
1288
struct netdev_queue *tx_nq, struct tsnep_tx *tx)
drivers/net/ethernet/engleder/tsnep_main.c
1301
if (!tsnep_xdp_xmit_back(rx->adapter, xdp, tx_nq, tx, false))
drivers/net/ethernet/engleder/tsnep_main.c
1333
struct tsnep_tx *tx)
drivers/net/ethernet/engleder/tsnep_main.c
1351
if (!tsnep_xdp_xmit_back(rx->adapter, xdp, tx_nq, tx, true))
drivers/net/ethernet/engleder/tsnep_main.c
1369
struct netdev_queue *tx_nq, struct tsnep_tx *tx)
drivers/net/ethernet/engleder/tsnep_main.c
1373
tsnep_xdp_xmit_flush(tx);
drivers/net/ethernet/engleder/tsnep_main.c
1443
struct tsnep_tx *tx;
drivers/net/ethernet/engleder/tsnep_main.c
1455
tx = &rx->adapter->tx[rx->tx_queue_index];
drivers/net/ethernet/engleder/tsnep_main.c
1516
&xdp_status, tx_nq, tx);
drivers/net/ethernet/engleder/tsnep_main.c
1532
tsnep_finalize_xdp(rx->adapter, xdp_status, tx_nq, tx);
drivers/net/ethernet/engleder/tsnep_main.c
1546
struct tsnep_tx *tx;
drivers/net/ethernet/engleder/tsnep_main.c
1558
tx = &rx->adapter->tx[rx->tx_queue_index];
drivers/net/ethernet/engleder/tsnep_main.c
1616
&xdp_status, tx_nq, tx);
drivers/net/ethernet/engleder/tsnep_main.c
1641
tsnep_finalize_xdp(rx->adapter, xdp_status, tx_nq, tx);
drivers/net/ethernet/engleder/tsnep_main.c
1798
if (queue->tx && tsnep_tx_pending(queue->tx))
drivers/net/ethernet/engleder/tsnep_main.c
1814
if (queue->tx)
drivers/net/ethernet/engleder/tsnep_main.c
1815
complete = tsnep_tx_poll(queue->tx, budget);
drivers/net/ethernet/engleder/tsnep_main.c
1861
if (queue->tx && queue->rx)
drivers/net/ethernet/engleder/tsnep_main.c
1864
else if (queue->tx)
drivers/net/ethernet/engleder/tsnep_main.c
1866
name, queue->tx->queue_index);
drivers/net/ethernet/engleder/tsnep_main.c
1919
struct tsnep_tx *tx = queue->tx;
drivers/net/ethernet/engleder/tsnep_main.c
1926
if (tx)
drivers/net/ethernet/engleder/tsnep_main.c
1927
rx->tx_queue_index = tx->queue_index;
drivers/net/ethernet/engleder/tsnep_main.c
1982
if (queue->tx) {
drivers/net/ethernet/engleder/tsnep_main.c
1983
netif_queue_set_napi(adapter->netdev, queue->tx->queue_index,
drivers/net/ethernet/engleder/tsnep_main.c
1985
tsnep_tx_enable(queue->tx);
drivers/net/ethernet/engleder/tsnep_main.c
2003
if (queue->tx) {
drivers/net/ethernet/engleder/tsnep_main.c
2004
tsnep_tx_disable(queue->tx, &queue->napi);
drivers/net/ethernet/engleder/tsnep_main.c
2005
netif_queue_set_napi(adapter->netdev, queue->tx->queue_index,
drivers/net/ethernet/engleder/tsnep_main.c
2025
if (adapter->queue[i].tx) {
drivers/net/ethernet/engleder/tsnep_main.c
2026
retval = tsnep_tx_open(adapter->queue[i].tx);
drivers/net/ethernet/engleder/tsnep_main.c
2068
if (adapter->queue[i].tx)
drivers/net/ethernet/engleder/tsnep_main.c
2069
tsnep_tx_close(adapter->queue[i].tx);
drivers/net/ethernet/engleder/tsnep_main.c
2089
if (adapter->queue[i].tx)
drivers/net/ethernet/engleder/tsnep_main.c
2090
tsnep_tx_close(adapter->queue[i].tx);
drivers/net/ethernet/engleder/tsnep_main.c
2123
queue->tx->xsk_pool = pool;
drivers/net/ethernet/engleder/tsnep_main.c
2144
queue->tx->xsk_pool = NULL;
drivers/net/ethernet/engleder/tsnep_main.c
2166
return tsnep_xmit_frame_ring(skb, &adapter->tx[queue_mapping]);
drivers/net/ethernet/engleder/tsnep_main.c
2194
stats->tx_packets += adapter->tx[i].packets;
drivers/net/ethernet/engleder/tsnep_main.c
2195
stats->tx_bytes += adapter->tx[i].bytes;
drivers/net/ethernet/engleder/tsnep_main.c
2196
stats->tx_dropped += adapter->tx[i].dropped;
drivers/net/ethernet/engleder/tsnep_main.c
2312
return &adapter->tx[cpu];
drivers/net/ethernet/engleder/tsnep_main.c
2321
struct tsnep_tx *tx;
drivers/net/ethernet/engleder/tsnep_main.c
2328
tx = tsnep_xdp_get_tx(adapter, cpu);
drivers/net/ethernet/engleder/tsnep_main.c
2329
nq = netdev_get_tx_queue(adapter->netdev, tx->queue_index);
drivers/net/ethernet/engleder/tsnep_main.c
2334
xmit = tsnep_xdp_xmit_frame_ring(xdp[nxmit], tx,
drivers/net/ethernet/engleder/tsnep_main.c
2346
tsnep_xdp_xmit_flush(tx);
drivers/net/ethernet/engleder/tsnep_main.c
2504
adapter->queue[0].tx = &adapter->tx[0];
drivers/net/ethernet/engleder/tsnep_main.c
2505
adapter->queue[0].tx->adapter = adapter;
drivers/net/ethernet/engleder/tsnep_main.c
2506
adapter->queue[0].tx->addr = adapter->addr + TSNEP_QUEUE(0);
drivers/net/ethernet/engleder/tsnep_main.c
2507
adapter->queue[0].tx->queue_index = 0;
drivers/net/ethernet/engleder/tsnep_main.c
2535
adapter->queue[i].tx = &adapter->tx[i];
drivers/net/ethernet/engleder/tsnep_main.c
2536
adapter->queue[i].tx->adapter = adapter;
drivers/net/ethernet/engleder/tsnep_main.c
2537
adapter->queue[i].tx->addr = adapter->addr + TSNEP_QUEUE(i);
drivers/net/ethernet/engleder/tsnep_main.c
2538
adapter->queue[i].tx->queue_index = i;
drivers/net/ethernet/engleder/tsnep_main.c
278
static void tsnep_tx_ring_cleanup(struct tsnep_tx *tx)
drivers/net/ethernet/engleder/tsnep_main.c
280
struct device *dmadev = tx->adapter->dmadev;
drivers/net/ethernet/engleder/tsnep_main.c
283
memset(tx->entry, 0, sizeof(tx->entry));
drivers/net/ethernet/engleder/tsnep_main.c
286
if (tx->page[i]) {
drivers/net/ethernet/engleder/tsnep_main.c
287
dma_free_coherent(dmadev, PAGE_SIZE, tx->page[i],
drivers/net/ethernet/engleder/tsnep_main.c
288
tx->page_dma[i]);
drivers/net/ethernet/engleder/tsnep_main.c
289
tx->page[i] = NULL;
drivers/net/ethernet/engleder/tsnep_main.c
290
tx->page_dma[i] = 0;
drivers/net/ethernet/engleder/tsnep_main.c
295
static int tsnep_tx_ring_create(struct tsnep_tx *tx)
drivers/net/ethernet/engleder/tsnep_main.c
297
struct device *dmadev = tx->adapter->dmadev;
drivers/net/ethernet/engleder/tsnep_main.c
304
tx->page[i] =
drivers/net/ethernet/engleder/tsnep_main.c
305
dma_alloc_coherent(dmadev, PAGE_SIZE, &tx->page_dma[i],
drivers/net/ethernet/engleder/tsnep_main.c
307
if (!tx->page[i]) {
drivers/net/ethernet/engleder/tsnep_main.c
312
entry = &tx->entry[TSNEP_RING_ENTRIES_PER_PAGE * i + j];
drivers/net/ethernet/engleder/tsnep_main.c
314
(((u8 *)tx->page[i]) + TSNEP_DESC_SIZE * j);
drivers/net/ethernet/engleder/tsnep_main.c
317
entry->desc_dma = tx->page_dma[i] + TSNEP_DESC_SIZE * j;
drivers/net/ethernet/engleder/tsnep_main.c
322
entry = &tx->entry[i];
drivers/net/ethernet/engleder/tsnep_main.c
323
next_entry = &tx->entry[(i + 1) & TSNEP_RING_MASK];
drivers/net/ethernet/engleder/tsnep_main.c
330
tsnep_tx_ring_cleanup(tx);
drivers/net/ethernet/engleder/tsnep_main.c
334
static void tsnep_tx_init(struct tsnep_tx *tx)
drivers/net/ethernet/engleder/tsnep_main.c
338
dma = tx->entry[0].desc_dma | TSNEP_RESET_OWNER_COUNTER;
drivers/net/ethernet/engleder/tsnep_main.c
339
iowrite32(DMA_ADDR_LOW(dma), tx->addr + TSNEP_TX_DESC_ADDR_LOW);
drivers/net/ethernet/engleder/tsnep_main.c
340
iowrite32(DMA_ADDR_HIGH(dma), tx->addr + TSNEP_TX_DESC_ADDR_HIGH);
drivers/net/ethernet/engleder/tsnep_main.c
341
tx->write = 0;
drivers/net/ethernet/engleder/tsnep_main.c
342
tx->read = 0;
drivers/net/ethernet/engleder/tsnep_main.c
343
tx->owner_counter = 1;
drivers/net/ethernet/engleder/tsnep_main.c
344
tx->increment_owner_counter = TSNEP_RING_SIZE - 1;
drivers/net/ethernet/engleder/tsnep_main.c
347
static void tsnep_tx_enable(struct tsnep_tx *tx)
drivers/net/ethernet/engleder/tsnep_main.c
351
nq = netdev_get_tx_queue(tx->adapter->netdev, tx->queue_index);
drivers/net/ethernet/engleder/tsnep_main.c
358
static void tsnep_tx_disable(struct tsnep_tx *tx, struct napi_struct *napi)
drivers/net/ethernet/engleder/tsnep_main.c
363
nq = netdev_get_tx_queue(tx->adapter->netdev, tx->queue_index);
drivers/net/ethernet/engleder/tsnep_main.c
370
readx_poll_timeout(ioread32, tx->addr + TSNEP_CONTROL, val,
drivers/net/ethernet/engleder/tsnep_main.c
375
while (READ_ONCE(tx->read) != tx->write) {
drivers/net/ethernet/engleder/tsnep_main.c
381
static void tsnep_tx_activate(struct tsnep_tx *tx, int index, int length,
drivers/net/ethernet/engleder/tsnep_main.c
384
struct tsnep_tx_entry *entry = &tx->entry[index];
drivers/net/ethernet/engleder/tsnep_main.c
416
if (index == tx->increment_owner_counter) {
drivers/net/ethernet/engleder/tsnep_main.c
417
tx->owner_counter++;
drivers/net/ethernet/engleder/tsnep_main.c
418
if (tx->owner_counter == 4)
drivers/net/ethernet/engleder/tsnep_main.c
419
tx->owner_counter = 1;
drivers/net/ethernet/engleder/tsnep_main.c
420
tx->increment_owner_counter--;
drivers/net/ethernet/engleder/tsnep_main.c
421
if (tx->increment_owner_counter < 0)
drivers/net/ethernet/engleder/tsnep_main.c
422
tx->increment_owner_counter = TSNEP_RING_SIZE - 1;
drivers/net/ethernet/engleder/tsnep_main.c
425
(tx->owner_counter << TSNEP_DESC_OWNER_COUNTER_SHIFT) &
drivers/net/ethernet/engleder/tsnep_main.c
442
static int tsnep_tx_desc_available(struct tsnep_tx *tx)
drivers/net/ethernet/engleder/tsnep_main.c
444
if (tx->read <= tx->write)
drivers/net/ethernet/engleder/tsnep_main.c
445
return TSNEP_RING_SIZE - tx->write + tx->read - 1;
drivers/net/ethernet/engleder/tsnep_main.c
447
return tx->read - tx->write - 1;
drivers/net/ethernet/engleder/tsnep_main.c
467
memcpy(&entry->desc->tx, fragdata, len);
drivers/net/ethernet/engleder/tsnep_main.c
472
memcpy(&entry->desc->tx, fragdata + skb_frag_off(frag),
drivers/net/ethernet/engleder/tsnep_main.c
483
static int tsnep_tx_map(struct sk_buff *skb, struct tsnep_tx *tx, int count,
drivers/net/ethernet/engleder/tsnep_main.c
486
struct device *dmadev = tx->adapter->dmadev;
drivers/net/ethernet/engleder/tsnep_main.c
494
entry = &tx->entry[(tx->write + i) & TSNEP_RING_MASK];
drivers/net/ethernet/engleder/tsnep_main.c
506
memcpy(&entry->desc->tx, skb->data, len);
drivers/net/ethernet/engleder/tsnep_main.c
525
entry->desc->tx = __cpu_to_le64(dma);
drivers/net/ethernet/engleder/tsnep_main.c
534
static int tsnep_tx_unmap(struct tsnep_tx *tx, int index, int count)
drivers/net/ethernet/engleder/tsnep_main.c
536
struct device *dmadev = tx->adapter->dmadev;
drivers/net/ethernet/engleder/tsnep_main.c
542
entry = &tx->entry[(index + i) & TSNEP_RING_MASK];
drivers/net/ethernet/engleder/tsnep_main.c
564
struct tsnep_tx *tx)
drivers/net/ethernet/engleder/tsnep_main.c
576
if (tsnep_tx_desc_available(tx) < count) {
drivers/net/ethernet/engleder/tsnep_main.c
580
netif_stop_subqueue(tx->adapter->netdev, tx->queue_index);
drivers/net/ethernet/engleder/tsnep_main.c
585
entry = &tx->entry[tx->write];
drivers/net/ethernet/engleder/tsnep_main.c
589
tx->adapter->hwtstamp_config.tx_type == HWTSTAMP_TX_ON) {
drivers/net/ethernet/engleder/tsnep_main.c
594
retval = tsnep_tx_map(skb, tx, count, do_tstamp);
drivers/net/ethernet/engleder/tsnep_main.c
596
tsnep_tx_unmap(tx, tx->write, count);
drivers/net/ethernet/engleder/tsnep_main.c
600
tx->dropped++;
drivers/net/ethernet/engleder/tsnep_main.c
607
tsnep_tx_activate(tx, (tx->write + i) & TSNEP_RING_MASK, length,
drivers/net/ethernet/engleder/tsnep_main.c
609
tx->write = (tx->write + count) & TSNEP_RING_MASK;
drivers/net/ethernet/engleder/tsnep_main.c
616
iowrite32(TSNEP_CONTROL_TX_ENABLE, tx->addr + TSNEP_CONTROL);
drivers/net/ethernet/engleder/tsnep_main.c
618
if (tsnep_tx_desc_available(tx) < (MAX_SKB_FRAGS + 1)) {
drivers/net/ethernet/engleder/tsnep_main.c
620
netif_stop_subqueue(tx->adapter->netdev, tx->queue_index);
drivers/net/ethernet/engleder/tsnep_main.c
626
static int tsnep_xdp_tx_map(struct xdp_frame *xdpf, struct tsnep_tx *tx,
drivers/net/ethernet/engleder/tsnep_main.c
629
struct device *dmadev = tx->adapter->dmadev;
drivers/net/ethernet/engleder/tsnep_main.c
642
entry = &tx->entry[(tx->write + i) & TSNEP_RING_MASK];
drivers/net/ethernet/engleder/tsnep_main.c
668
entry->desc->tx = __cpu_to_le64(dma);
drivers/net/ethernet/engleder/tsnep_main.c
683
struct tsnep_tx *tx, u32 type)
drivers/net/ethernet/engleder/tsnep_main.c
697
if (tsnep_tx_desc_available(tx) < (MAX_SKB_FRAGS + 1 + count))
drivers/net/ethernet/engleder/tsnep_main.c
700
entry = &tx->entry[tx->write];
drivers/net/ethernet/engleder/tsnep_main.c
703
retval = tsnep_xdp_tx_map(xdpf, tx, shinfo, count, type);
drivers/net/ethernet/engleder/tsnep_main.c
705
tsnep_tx_unmap(tx, tx->write, count);
drivers/net/ethernet/engleder/tsnep_main.c
708
tx->dropped++;
drivers/net/ethernet/engleder/tsnep_main.c
715
tsnep_tx_activate(tx, (tx->write + i) & TSNEP_RING_MASK, length,
drivers/net/ethernet/engleder/tsnep_main.c
717
tx->write = (tx->write + count) & TSNEP_RING_MASK;
drivers/net/ethernet/engleder/tsnep_main.c
725
static void tsnep_xdp_xmit_flush(struct tsnep_tx *tx)
drivers/net/ethernet/engleder/tsnep_main.c
727
iowrite32(TSNEP_CONTROL_TX_ENABLE, tx->addr + TSNEP_CONTROL);
drivers/net/ethernet/engleder/tsnep_main.c
732
struct netdev_queue *tx_nq, struct tsnep_tx *tx,
drivers/net/ethernet/engleder/tsnep_main.c
750
xmit = tsnep_xdp_xmit_frame_ring(xdpf, tx, type);
drivers/net/ethernet/engleder/tsnep_main.c
761
static int tsnep_xdp_tx_map_zc(struct xdp_desc *xdpd, struct tsnep_tx *tx)
drivers/net/ethernet/engleder/tsnep_main.c
766
entry = &tx->entry[tx->write];
drivers/net/ethernet/engleder/tsnep_main.c
769
dma = xsk_buff_raw_get_dma(tx->xsk_pool, xdpd->addr);
drivers/net/ethernet/engleder/tsnep_main.c
770
xsk_buff_raw_dma_sync_for_device(tx->xsk_pool, dma, xdpd->len);
drivers/net/ethernet/engleder/tsnep_main.c
775
entry->desc->tx = __cpu_to_le64(dma);
drivers/net/ethernet/engleder/tsnep_main.c
781
struct tsnep_tx *tx)
drivers/net/ethernet/engleder/tsnep_main.c
785
length = tsnep_xdp_tx_map_zc(xdpd, tx);
drivers/net/ethernet/engleder/tsnep_main.c
787
tsnep_tx_activate(tx, tx->write, length, true);
drivers/net/ethernet/engleder/tsnep_main.c
788
tx->write = (tx->write + 1) & TSNEP_RING_MASK;
drivers/net/ethernet/engleder/tsnep_main.c
791
static void tsnep_xdp_xmit_zc(struct tsnep_tx *tx)
drivers/net/ethernet/engleder/tsnep_main.c
793
int desc_available = tsnep_tx_desc_available(tx);
drivers/net/ethernet/engleder/tsnep_main.c
794
struct xdp_desc *descs = tx->xsk_pool->tx_descs;
drivers/net/ethernet/engleder/tsnep_main.c
805
batch = xsk_tx_peek_release_desc_batch(tx->xsk_pool, desc_available);
drivers/net/ethernet/engleder/tsnep_main.c
807
tsnep_xdp_xmit_frame_ring_zc(&descs[i], tx);
drivers/net/ethernet/engleder/tsnep_main.c
815
tsnep_xdp_xmit_flush(tx);
drivers/net/ethernet/engleder/tsnep_main.c
819
static bool tsnep_tx_poll(struct tsnep_tx *tx, int napi_budget)
drivers/net/ethernet/engleder/tsnep_main.c
828
nq = netdev_get_tx_queue(tx->adapter->netdev, tx->queue_index);
drivers/net/ethernet/engleder/tsnep_main.c
832
if (tx->read == tx->write)
drivers/net/ethernet/engleder/tsnep_main.c
835
entry = &tx->entry[tx->read];
drivers/net/ethernet/engleder/tsnep_main.c
854
length = tsnep_tx_unmap(tx, tx->read, count);
drivers/net/ethernet/engleder/tsnep_main.c
885
tx->read = (tx->read + count) & TSNEP_RING_MASK;
drivers/net/ethernet/engleder/tsnep_main.c
887
tx->packets++;
drivers/net/ethernet/engleder/tsnep_main.c
888
tx->bytes += length + ETH_FCS_LEN;
drivers/net/ethernet/engleder/tsnep_main.c
893
if (tx->xsk_pool) {
drivers/net/ethernet/engleder/tsnep_main.c
895
xsk_tx_completed(tx->xsk_pool, xsk_frames);
drivers/net/ethernet/engleder/tsnep_main.c
896
if (xsk_uses_need_wakeup(tx->xsk_pool))
drivers/net/ethernet/engleder/tsnep_main.c
897
xsk_set_tx_need_wakeup(tx->xsk_pool);
drivers/net/ethernet/engleder/tsnep_main.c
898
tsnep_xdp_xmit_zc(tx);
drivers/net/ethernet/engleder/tsnep_main.c
901
if ((tsnep_tx_desc_available(tx) >= ((MAX_SKB_FRAGS + 1) * 2)) &&
drivers/net/ethernet/engleder/tsnep_main.c
911
static bool tsnep_tx_pending(struct tsnep_tx *tx)
drivers/net/ethernet/engleder/tsnep_main.c
917
nq = netdev_get_tx_queue(tx->adapter->netdev, tx->queue_index);
drivers/net/ethernet/engleder/tsnep_main.c
920
if (tx->read != tx->write) {
drivers/net/ethernet/engleder/tsnep_main.c
921
entry = &tx->entry[tx->read];
drivers/net/ethernet/engleder/tsnep_main.c
933
static int tsnep_tx_open(struct tsnep_tx *tx)
drivers/net/ethernet/engleder/tsnep_main.c
937
retval = tsnep_tx_ring_create(tx);
drivers/net/ethernet/engleder/tsnep_main.c
941
tsnep_tx_init(tx);
drivers/net/ethernet/engleder/tsnep_main.c
946
static void tsnep_tx_close(struct tsnep_tx *tx)
drivers/net/ethernet/engleder/tsnep_main.c
948
tsnep_tx_ring_cleanup(tx);
drivers/net/ethernet/engleder/tsnep_xdp.c
33
queue->tx->queue_index != queue_id) {
drivers/net/ethernet/freescale/fman/fman_port.c
1000
iowrite32be(0xFFFFFF, &port->bmi_regs->tx.fmbm_tcfqid);
drivers/net/ethernet/freescale/fman/fman_port.c
1002
&port->bmi_regs->tx.fmbm_tfene);
drivers/net/ethernet/freescale/fman/fman_port.c
1376
iowrite32be(reg, &port->bmi_regs->tx.fmbm_tfp);
drivers/net/ethernet/freescale/fman/fman_port.c
1579
bmi_cfg_reg = &port->bmi_regs->tx.fmbm_tcfg;
drivers/net/ethernet/freescale/fman/fman_port.c
1580
bmi_status_reg = &port->bmi_regs->tx.fmbm_tst;
drivers/net/ethernet/freescale/fman/fman_port.c
1654
bmi_cfg_reg = &port->bmi_regs->tx.fmbm_tcfg;
drivers/net/ethernet/freescale/fman/fman_port.c
278
struct fman_port_tx_bmi_regs tx;
drivers/net/ethernet/freescale/fman/fman_port.c
542
struct fman_port_tx_bmi_regs __iomem *regs = &port->bmi_regs->tx;
drivers/net/ethernet/google/gve/gve.h
1235
int gve_xdp_xmit_one(struct gve_priv *priv, struct gve_tx_ring *tx,
drivers/net/ethernet/google/gve/gve.h
1238
int gve_xdp_xmit_one_dqo(struct gve_priv *priv, struct gve_tx_ring *tx,
drivers/net/ethernet/google/gve/gve.h
1250
struct gve_tx_ring *tx);
drivers/net/ethernet/google/gve/gve.h
1251
bool gve_tx_clean_pending(struct gve_priv *priv, struct gve_tx_ring *tx);
drivers/net/ethernet/google/gve/gve.h
673
struct gve_tx_ring *tx; /* tx rings on this block */
drivers/net/ethernet/google/gve/gve.h
721
struct gve_tx_ring *tx;
drivers/net/ethernet/google/gve/gve.h
799
struct gve_tx_ring *tx; /* array of tx_cfg.num_queues */
drivers/net/ethernet/google/gve/gve_adminq.c
697
struct gve_tx_ring *tx = &priv->tx[queue_index];
drivers/net/ethernet/google/gve/gve_adminq.c
705
cpu_to_be64(tx->q_resources_bus),
drivers/net/ethernet/google/gve/gve_adminq.c
706
.tx_ring_addr = cpu_to_be64(tx->bus),
drivers/net/ethernet/google/gve/gve_adminq.c
707
.ntfy_id = cpu_to_be32(tx->ntfy_id),
drivers/net/ethernet/google/gve/gve_adminq.c
713
GVE_RAW_ADDRESSING_QPL_ID : tx->tx_fifo.qpl->id;
drivers/net/ethernet/google/gve/gve_adminq.c
722
qpl_id = tx->dqo.qpl->id;
drivers/net/ethernet/google/gve/gve_adminq.c
725
cpu_to_be64(tx->complq_bus_dqo);
drivers/net/ethernet/google/gve/gve_dqo.h
62
int gve_clean_tx_done_dqo(struct gve_priv *priv, struct gve_tx_ring *tx,
drivers/net/ethernet/google/gve/gve_ethtool.c
237
if (priv->tx) {
drivers/net/ethernet/google/gve/gve_ethtool.c
240
u64_stats_fetch_begin(&priv->tx[ring].statss);
drivers/net/ethernet/google/gve/gve_ethtool.c
241
tmp_tx_pkts = priv->tx[ring].pkt_done;
drivers/net/ethernet/google/gve/gve_ethtool.c
242
tmp_tx_bytes = priv->tx[ring].bytes_done;
drivers/net/ethernet/google/gve/gve_ethtool.c
243
} while (u64_stats_fetch_retry(&priv->tx[ring].statss,
drivers/net/ethernet/google/gve/gve_ethtool.c
247
tx_dropped += priv->tx[ring].dropped_pkt;
drivers/net/ethernet/google/gve/gve_ethtool.c
404
if (priv->tx) {
drivers/net/ethernet/google/gve/gve_ethtool.c
406
struct gve_tx_ring *tx = &priv->tx[ring];
drivers/net/ethernet/google/gve/gve_ethtool.c
409
data[i++] = tx->req;
drivers/net/ethernet/google/gve/gve_ethtool.c
410
data[i++] = tx->done;
drivers/net/ethernet/google/gve/gve_ethtool.c
411
data[i++] = tx->req - tx->done;
drivers/net/ethernet/google/gve/gve_ethtool.c
419
(tx->dqo_tx.tail - tx->dqo_tx.head) &
drivers/net/ethernet/google/gve/gve_ethtool.c
420
tx->mask;
drivers/net/ethernet/google/gve/gve_ethtool.c
424
u64_stats_fetch_begin(&priv->tx[ring].statss);
drivers/net/ethernet/google/gve/gve_ethtool.c
425
tmp_tx_bytes = tx->bytes_done;
drivers/net/ethernet/google/gve/gve_ethtool.c
426
} while (u64_stats_fetch_retry(&priv->tx[ring].statss,
drivers/net/ethernet/google/gve/gve_ethtool.c
429
data[i++] = tx->wake_queue;
drivers/net/ethernet/google/gve/gve_ethtool.c
430
data[i++] = tx->stop_queue;
drivers/net/ethernet/google/gve/gve_ethtool.c
431
data[i++] = gve_tx_load_event_counter(priv, tx);
drivers/net/ethernet/google/gve/gve_ethtool.c
432
data[i++] = tx->dma_mapping_error;
drivers/net/ethernet/google/gve/gve_ethtool.c
447
start = u64_stats_fetch_begin(&priv->tx[ring].statss);
drivers/net/ethernet/google/gve/gve_ethtool.c
448
data[i] = tx->xdp_xsk_sent;
drivers/net/ethernet/google/gve/gve_ethtool.c
449
data[i + 1] = tx->xdp_xmit;
drivers/net/ethernet/google/gve/gve_ethtool.c
450
data[i + 2] = tx->xdp_xmit_errors;
drivers/net/ethernet/google/gve/gve_ethtool.c
451
} while (u64_stats_fetch_retry(&priv->tx[ring].statss,
drivers/net/ethernet/google/gve/gve_main.c
1184
if (!priv->tx)
drivers/net/ethernet/google/gve/gve_main.c
1186
priv->tx[gve_xdp_tx_queue_id(priv, qid)].xsk_pool = NULL;
drivers/net/ethernet/google/gve/gve_main.c
1207
priv->tx[tx_qid].xsk_pool = pool;
drivers/net/ethernet/google/gve/gve_main.c
130
if (priv->tx) {
drivers/net/ethernet/google/gve/gve_main.c
134
u64_stats_fetch_begin(&priv->tx[ring].statss);
drivers/net/ethernet/google/gve/gve_main.c
135
packets = priv->tx[ring].pkt_done;
drivers/net/ethernet/google/gve/gve_main.c
1351
priv->tx = NULL;
drivers/net/ethernet/google/gve/gve_main.c
136
bytes = priv->tx[ring].bytes_done;
drivers/net/ethernet/google/gve/gve_main.c
1366
priv->tx = tx_alloc_cfg->tx;
drivers/net/ethernet/google/gve/gve_main.c
137
} while (u64_stats_fetch_retry(&priv->tx[ring].statss,
drivers/net/ethernet/google/gve/gve_main.c
1673
napi_tx = &priv->ntfy_blocks[priv->tx[tx_qid].ntfy_id].napi;
drivers/net/ethernet/google/gve/gve_main.c
1685
if (gve_tx_clean_pending(priv, &priv->tx[tx_qid]))
drivers/net/ethernet/google/gve/gve_main.c
2035
if (block->tx->last_kick_msec + MIN_TX_TIMEOUT_GAP > current_time)
drivers/net/ethernet/google/gve/gve_main.c
2040
block->tx->last_kick_msec = current_time;
drivers/net/ethernet/google/gve/gve_main.c
2057
block->tx->queue_timeout++;
drivers/net/ethernet/google/gve/gve_main.c
2259
if (priv->tx) {
drivers/net/ethernet/google/gve/gve_main.c
2266
last_completion = priv->tx[idx].done;
drivers/net/ethernet/google/gve/gve_main.c
2267
tx_frames = priv->tx[idx].req;
drivers/net/ethernet/google/gve/gve_main.c
2271
start = u64_stats_fetch_begin(&priv->tx[idx].statss);
drivers/net/ethernet/google/gve/gve_main.c
2272
tx_bytes = priv->tx[idx].bytes_done;
drivers/net/ethernet/google/gve/gve_main.c
2273
} while (u64_stats_fetch_retry(&priv->tx[idx].statss, start));
drivers/net/ethernet/google/gve/gve_main.c
2276
.value = cpu_to_be64(priv->tx[idx].wake_queue),
drivers/net/ethernet/google/gve/gve_main.c
2281
.value = cpu_to_be64(priv->tx[idx].stop_queue),
drivers/net/ethernet/google/gve/gve_main.c
2301
.value = cpu_to_be64(priv->tx[idx].queue_timeout),
drivers/net/ethernet/google/gve/gve_main.c
2723
struct gve_tx_ring *tx = &priv->tx[idx];
drivers/net/ethernet/google/gve/gve_main.c
2727
start = u64_stats_fetch_begin(&tx->statss);
drivers/net/ethernet/google/gve/gve_main.c
2728
tx_stats->packets = tx->pkt_done;
drivers/net/ethernet/google/gve/gve_main.c
2729
tx_stats->bytes = tx->bytes_done;
drivers/net/ethernet/google/gve/gve_main.c
2730
} while (u64_stats_fetch_retry(&tx->statss, start));
drivers/net/ethernet/google/gve/gve_main.c
2735
struct netdev_queue_stats_tx *tx)
drivers/net/ethernet/google/gve/gve_main.c
2741
tx->packets = 0;
drivers/net/ethernet/google/gve/gve_main.c
2742
tx->bytes = 0;
drivers/net/ethernet/google/gve/gve_main.c
363
if (block->tx) {
drivers/net/ethernet/google/gve/gve_main.c
364
if (block->tx->q_num < priv->tx_cfg.num_queues)
drivers/net/ethernet/google/gve/gve_main.c
399
if (block->tx)
drivers/net/ethernet/google/gve/gve_main.c
400
reschedule |= gve_tx_clean_pending(priv, block->tx);
drivers/net/ethernet/google/gve/gve_main.c
418
if (block->tx) {
drivers/net/ethernet/google/gve/gve_main.c
419
if (block->tx->q_num < priv->tx_cfg.num_queues)
drivers/net/ethernet/google/gve/gve_main.c
819
struct gve_tx_ring *tx = &priv->tx[idx];
drivers/net/ethernet/google/gve/gve_main.c
822
return tx->tx_fifo.qpl;
drivers/net/ethernet/google/gve/gve_main.c
824
return tx->dqo.qpl;
drivers/net/ethernet/google/gve/gve_main.c
944
u64_stats_init(&priv->tx[i].statss);
drivers/net/ethernet/google/gve/gve_main.c
945
priv->tx[i].ntfy_id = ntfy_idx;
drivers/net/ethernet/google/gve/gve_main.c
954
u64_stats_init(&priv->tx[i].statss);
drivers/net/ethernet/google/gve/gve_main.c
970
cfg->tx = priv->tx;
drivers/net/ethernet/google/gve/gve_main.c
977
if (!priv->tx)
drivers/net/ethernet/google/gve/gve_rx.c
759
struct gve_tx_ring *tx;
drivers/net/ethernet/google/gve/gve_rx.c
770
tx = &priv->tx[tx_qid];
drivers/net/ethernet/google/gve/gve_rx.c
771
spin_lock(&tx->xdp_lock);
drivers/net/ethernet/google/gve/gve_rx.c
772
err = gve_xdp_xmit_one(priv, tx, xdp->data,
drivers/net/ethernet/google/gve/gve_rx.c
774
spin_unlock(&tx->xdp_lock);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
610
struct gve_tx_ring *tx;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
623
tx = &priv->tx[tx_qid];
drivers/net/ethernet/google/gve/gve_rx_dqo.c
624
spin_lock(&tx->dqo_tx.xdp_lock);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
625
err = gve_xdp_xmit_one_dqo(priv, tx, xdpf);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
626
spin_unlock(&tx->dqo_tx.xdp_lock);
drivers/net/ethernet/google/gve/gve_tx.c
1003
nic_done = gve_tx_load_event_counter(priv, tx);
drivers/net/ethernet/google/gve/gve_tx.c
1004
to_do = min_t(u32, (nic_done - tx->done), budget);
drivers/net/ethernet/google/gve/gve_tx.c
1005
gve_clean_xdp_done(priv, tx, to_do);
drivers/net/ethernet/google/gve/gve_tx.c
1008
return nic_done != tx->done;
drivers/net/ethernet/google/gve/gve_tx.c
1014
struct gve_tx_ring *tx = block->tx;
drivers/net/ethernet/google/gve/gve_tx.c
1026
spin_lock(&tx->clean_lock);
drivers/net/ethernet/google/gve/gve_tx.c
1028
nic_done = gve_tx_load_event_counter(priv, tx);
drivers/net/ethernet/google/gve/gve_tx.c
1029
to_do = min_t(u32, (nic_done - tx->done), budget);
drivers/net/ethernet/google/gve/gve_tx.c
1030
gve_clean_tx_done(priv, tx, to_do, true);
drivers/net/ethernet/google/gve/gve_tx.c
1031
spin_unlock(&tx->clean_lock);
drivers/net/ethernet/google/gve/gve_tx.c
1033
return nic_done != tx->done;
drivers/net/ethernet/google/gve/gve_tx.c
1036
bool gve_tx_clean_pending(struct gve_priv *priv, struct gve_tx_ring *tx)
drivers/net/ethernet/google/gve/gve_tx.c
1038
u32 nic_done = gve_tx_load_event_counter(priv, tx);
drivers/net/ethernet/google/gve/gve_tx.c
1040
return nic_done != tx->done;
drivers/net/ethernet/google/gve/gve_tx.c
157
static int gve_clean_xdp_done(struct gve_priv *priv, struct gve_tx_ring *tx,
drivers/net/ethernet/google/gve/gve_tx.c
168
idx = tx->done & tx->mask;
drivers/net/ethernet/google/gve/gve_tx.c
169
info = &tx->info[idx];
drivers/net/ethernet/google/gve/gve_tx.c
170
tx->done++;
drivers/net/ethernet/google/gve/gve_tx.c
187
gve_tx_free_fifo(&tx->tx_fifo, space_freed);
drivers/net/ethernet/google/gve/gve_tx.c
188
if (xsk_complete > 0 && tx->xsk_pool)
drivers/net/ethernet/google/gve/gve_tx.c
189
xsk_tx_completed(tx->xsk_pool, xsk_complete);
drivers/net/ethernet/google/gve/gve_tx.c
190
u64_stats_update_begin(&tx->statss);
drivers/net/ethernet/google/gve/gve_tx.c
191
tx->bytes_done += bytes;
drivers/net/ethernet/google/gve/gve_tx.c
192
tx->pkt_done += pkts;
drivers/net/ethernet/google/gve/gve_tx.c
193
u64_stats_update_end(&tx->statss);
drivers/net/ethernet/google/gve/gve_tx.c
197
static int gve_clean_tx_done(struct gve_priv *priv, struct gve_tx_ring *tx,
drivers/net/ethernet/google/gve/gve_tx.c
203
struct gve_tx_ring *tx = &priv->tx[idx];
drivers/net/ethernet/google/gve/gve_tx.c
209
if (tx->q_num < priv->tx_cfg.num_queues)
drivers/net/ethernet/google/gve/gve_tx.c
210
gve_clean_tx_done(priv, tx, priv->tx_desc_cnt, false);
drivers/net/ethernet/google/gve/gve_tx.c
212
gve_clean_xdp_done(priv, tx, priv->tx_desc_cnt);
drivers/net/ethernet/google/gve/gve_tx.c
213
netdev_tx_reset_queue(tx->netdev_txq);
drivers/net/ethernet/google/gve/gve_tx.c
217
static void gve_tx_free_ring_gqi(struct gve_priv *priv, struct gve_tx_ring *tx,
drivers/net/ethernet/google/gve/gve_tx.c
221
int idx = tx->q_num;
drivers/net/ethernet/google/gve/gve_tx.c
226
slots = tx->mask + 1;
drivers/net/ethernet/google/gve/gve_tx.c
227
dma_free_coherent(hdev, sizeof(*tx->q_resources),
drivers/net/ethernet/google/gve/gve_tx.c
228
tx->q_resources, tx->q_resources_bus);
drivers/net/ethernet/google/gve/gve_tx.c
229
tx->q_resources = NULL;
drivers/net/ethernet/google/gve/gve_tx.c
231
if (tx->tx_fifo.qpl) {
drivers/net/ethernet/google/gve/gve_tx.c
232
if (tx->tx_fifo.base)
drivers/net/ethernet/google/gve/gve_tx.c
233
gve_tx_fifo_release(priv, &tx->tx_fifo);
drivers/net/ethernet/google/gve/gve_tx.c
235
qpl_id = gve_tx_qpl_id(priv, tx->q_num);
drivers/net/ethernet/google/gve/gve_tx.c
236
gve_free_queue_page_list(priv, tx->tx_fifo.qpl, qpl_id);
drivers/net/ethernet/google/gve/gve_tx.c
237
tx->tx_fifo.qpl = NULL;
drivers/net/ethernet/google/gve/gve_tx.c
240
bytes = sizeof(*tx->desc) * slots;
drivers/net/ethernet/google/gve/gve_tx.c
241
dma_free_coherent(hdev, bytes, tx->desc, tx->bus);
drivers/net/ethernet/google/gve/gve_tx.c
242
tx->desc = NULL;
drivers/net/ethernet/google/gve/gve_tx.c
244
vfree(tx->info);
drivers/net/ethernet/google/gve/gve_tx.c
245
tx->info = NULL;
drivers/net/ethernet/google/gve/gve_tx.c
253
struct gve_tx_ring *tx = &priv->tx[idx];
drivers/net/ethernet/google/gve/gve_tx.c
257
tx->netdev_txq = netdev_get_tx_queue(priv->dev, idx);
drivers/net/ethernet/google/gve/gve_tx.c
26
struct gve_tx_ring *tx = &priv->tx[tx_qid];
drivers/net/ethernet/google/gve/gve_tx.c
263
struct gve_tx_ring *tx,
drivers/net/ethernet/google/gve/gve_tx.c
272
memset(tx, 0, sizeof(*tx));
drivers/net/ethernet/google/gve/gve_tx.c
273
spin_lock_init(&tx->clean_lock);
drivers/net/ethernet/google/gve/gve_tx.c
274
spin_lock_init(&tx->xdp_lock);
drivers/net/ethernet/google/gve/gve_tx.c
275
tx->q_num = idx;
drivers/net/ethernet/google/gve/gve_tx.c
277
tx->mask = cfg->ring_size - 1;
drivers/net/ethernet/google/gve/gve_tx.c
28
gve_tx_put_doorbell(priv, tx->q_resources, tx->req);
drivers/net/ethernet/google/gve/gve_tx.c
280
tx->info = vcalloc(cfg->ring_size, sizeof(*tx->info));
drivers/net/ethernet/google/gve/gve_tx.c
281
if (!tx->info)
drivers/net/ethernet/google/gve/gve_tx.c
285
bytes = sizeof(*tx->desc) * cfg->ring_size;
drivers/net/ethernet/google/gve/gve_tx.c
286
tx->desc = dma_alloc_coherent(hdev, bytes, &tx->bus, GFP_KERNEL);
drivers/net/ethernet/google/gve/gve_tx.c
287
if (!tx->desc)
drivers/net/ethernet/google/gve/gve_tx.c
290
tx->raw_addressing = cfg->raw_addressing;
drivers/net/ethernet/google/gve/gve_tx.c
291
tx->dev = hdev;
drivers/net/ethernet/google/gve/gve_tx.c
292
if (!tx->raw_addressing) {
drivers/net/ethernet/google/gve/gve_tx.c
293
qpl_id = gve_tx_qpl_id(priv, tx->q_num);
drivers/net/ethernet/google/gve/gve_tx.c
296
tx->tx_fifo.qpl = gve_alloc_queue_page_list(priv, qpl_id,
drivers/net/ethernet/google/gve/gve_tx.c
298
if (!tx->tx_fifo.qpl)
drivers/net/ethernet/google/gve/gve_tx.c
302
if (gve_tx_fifo_init(priv, &tx->tx_fifo))
drivers/net/ethernet/google/gve/gve_tx.c
306
tx->q_resources =
drivers/net/ethernet/google/gve/gve_tx.c
308
sizeof(*tx->q_resources),
drivers/net/ethernet/google/gve/gve_tx.c
309
&tx->q_resources_bus,
drivers/net/ethernet/google/gve/gve_tx.c
311
if (!tx->q_resources)
drivers/net/ethernet/google/gve/gve_tx.c
317
if (!tx->raw_addressing)
drivers/net/ethernet/google/gve/gve_tx.c
318
gve_tx_fifo_release(priv, &tx->tx_fifo);
drivers/net/ethernet/google/gve/gve_tx.c
320
if (!tx->raw_addressing) {
drivers/net/ethernet/google/gve/gve_tx.c
321
gve_free_queue_page_list(priv, tx->tx_fifo.qpl, qpl_id);
drivers/net/ethernet/google/gve/gve_tx.c
322
tx->tx_fifo.qpl = NULL;
drivers/net/ethernet/google/gve/gve_tx.c
325
dma_free_coherent(hdev, bytes, tx->desc, tx->bus);
drivers/net/ethernet/google/gve/gve_tx.c
326
tx->desc = NULL;
drivers/net/ethernet/google/gve/gve_tx.c
328
vfree(tx->info);
drivers/net/ethernet/google/gve/gve_tx.c
329
tx->info = NULL;
drivers/net/ethernet/google/gve/gve_tx.c
336
struct gve_tx_ring *tx = cfg->tx;
drivers/net/ethernet/google/gve/gve_tx.c
348
tx = kvzalloc_objs(struct gve_tx_ring, cfg->qcfg->max_queues);
drivers/net/ethernet/google/gve/gve_tx.c
349
if (!tx)
drivers/net/ethernet/google/gve/gve_tx.c
353
err = gve_tx_alloc_ring_gqi(priv, cfg, &tx[i], i);
drivers/net/ethernet/google/gve/gve_tx.c
362
cfg->tx = tx;
drivers/net/ethernet/google/gve/gve_tx.c
367
gve_tx_free_ring_gqi(priv, &tx[j], cfg);
drivers/net/ethernet/google/gve/gve_tx.c
368
kvfree(tx);
drivers/net/ethernet/google/gve/gve_tx.c
375
struct gve_tx_ring *tx = cfg->tx;
drivers/net/ethernet/google/gve/gve_tx.c
378
if (!tx)
drivers/net/ethernet/google/gve/gve_tx.c
382
gve_tx_free_ring_gqi(priv, &tx[i], cfg);
drivers/net/ethernet/google/gve/gve_tx.c
384
kvfree(tx);
drivers/net/ethernet/google/gve/gve_tx.c
385
cfg->tx = NULL;
drivers/net/ethernet/google/gve/gve_tx.c
395
static inline u32 gve_tx_avail(struct gve_tx_ring *tx)
drivers/net/ethernet/google/gve/gve_tx.c
397
return tx->mask + 1 - (tx->req - tx->done);
drivers/net/ethernet/google/gve/gve_tx.c
400
static inline int gve_skb_fifo_bytes_required(struct gve_tx_ring *tx,
drivers/net/ethernet/google/gve/gve_tx.c
410
pad_bytes = gve_tx_fifo_pad_alloc_one_frag(&tx->tx_fifo,
drivers/net/ethernet/google/gve/gve_tx.c
445
static inline bool gve_can_tx(struct gve_tx_ring *tx, int bytes_required)
drivers/net/ethernet/google/gve/gve_tx.c
449
if (!tx->raw_addressing)
drivers/net/ethernet/google/gve/gve_tx.c
450
can_alloc = gve_tx_fifo_can_alloc(&tx->tx_fifo, bytes_required);
drivers/net/ethernet/google/gve/gve_tx.c
452
return (gve_tx_avail(tx) >= MAX_TX_DESC_NEEDED && can_alloc);
drivers/net/ethernet/google/gve/gve_tx.c
458
static int gve_maybe_stop_tx(struct gve_priv *priv, struct gve_tx_ring *tx,
drivers/net/ethernet/google/gve/gve_tx.c
466
if (!tx->raw_addressing)
drivers/net/ethernet/google/gve/gve_tx.c
467
bytes_required = gve_skb_fifo_bytes_required(tx, skb);
drivers/net/ethernet/google/gve/gve_tx.c
469
if (likely(gve_can_tx(tx, bytes_required)))
drivers/net/ethernet/google/gve/gve_tx.c
473
spin_lock(&tx->clean_lock);
drivers/net/ethernet/google/gve/gve_tx.c
474
nic_done = gve_tx_load_event_counter(priv, tx);
drivers/net/ethernet/google/gve/gve_tx.c
475
to_do = nic_done - tx->done;
drivers/net/ethernet/google/gve/gve_tx.c
478
if (to_do + gve_tx_avail(tx) >= MAX_TX_DESC_NEEDED) {
drivers/net/ethernet/google/gve/gve_tx.c
481
gve_clean_tx_done(priv, tx, to_do, false);
drivers/net/ethernet/google/gve/gve_tx.c
483
if (likely(gve_can_tx(tx, bytes_required)))
drivers/net/ethernet/google/gve/gve_tx.c
488
tx->stop_queue++;
drivers/net/ethernet/google/gve/gve_tx.c
489
netif_tx_stop_queue(tx->netdev_txq);
drivers/net/ethernet/google/gve/gve_tx.c
491
spin_unlock(&tx->clean_lock);
drivers/net/ethernet/google/gve/gve_tx.c
561
static int gve_tx_add_skb_copy(struct gve_priv *priv, struct gve_tx_ring *tx, struct sk_buff *skb)
drivers/net/ethernet/google/gve/gve_tx.c
568
u32 idx = tx->req & tx->mask;
drivers/net/ethernet/google/gve/gve_tx.c
574
info = &tx->info[idx];
drivers/net/ethernet/google/gve/gve_tx.c
575
pkt_desc = &tx->desc[idx];
drivers/net/ethernet/google/gve/gve_tx.c
588
pad_bytes = gve_tx_fifo_pad_alloc_one_frag(&tx->tx_fifo, hlen);
drivers/net/ethernet/google/gve/gve_tx.c
589
hdr_nfrags = gve_tx_alloc_fifo(&tx->tx_fifo, hlen + pad_bytes,
drivers/net/ethernet/google/gve/gve_tx.c
592
payload_nfrags = gve_tx_alloc_fifo(&tx->tx_fifo, skb->len - hlen,
drivers/net/ethernet/google/gve/gve_tx.c
601
tx->tx_fifo.base + info->iov[hdr_nfrags - 1].iov_offset,
drivers/net/ethernet/google/gve/gve_tx.c
603
gve_dma_sync_for_device(&priv->pdev->dev, tx->tx_fifo.qpl->page_buses,
drivers/net/ethernet/google/gve/gve_tx.c
609
next_idx = (tx->req + 1) & tx->mask;
drivers/net/ethernet/google/gve/gve_tx.c
610
gve_tx_fill_mtd_desc(&tx->desc[next_idx], skb);
drivers/net/ethernet/google/gve/gve_tx.c
614
next_idx = (tx->req + 1 + mtd_desc_nr + i - payload_iov) & tx->mask;
drivers/net/ethernet/google/gve/gve_tx.c
615
seg_desc = &tx->desc[next_idx];
drivers/net/ethernet/google/gve/gve_tx.c
624
tx->tx_fifo.base + info->iov[i].iov_offset,
drivers/net/ethernet/google/gve/gve_tx.c
626
gve_dma_sync_for_device(&priv->pdev->dev, tx->tx_fifo.qpl->page_buses,
drivers/net/ethernet/google/gve/gve_tx.c
635
static int gve_tx_add_skb_no_copy(struct gve_priv *priv, struct gve_tx_ring *tx,
drivers/net/ethernet/google/gve/gve_tx.c
644
u32 idx = tx->req & tx->mask;
drivers/net/ethernet/google/gve/gve_tx.c
649
info = &tx->info[idx];
drivers/net/ethernet/google/gve/gve_tx.c
650
pkt_desc = &tx->desc[idx];
drivers/net/ethernet/google/gve/gve_tx.c
663
addr = dma_map_single(tx->dev, skb->data, len, DMA_TO_DEVICE);
drivers/net/ethernet/google/gve/gve_tx.c
664
if (unlikely(dma_mapping_error(tx->dev, addr))) {
drivers/net/ethernet/google/gve/gve_tx.c
665
tx->dma_mapping_error++;
drivers/net/ethernet/google/gve/gve_tx.c
682
idx = (idx + 1) & tx->mask;
drivers/net/ethernet/google/gve/gve_tx.c
683
mtd_desc = &tx->desc[idx];
drivers/net/ethernet/google/gve/gve_tx.c
693
idx = (idx + 1) & tx->mask;
drivers/net/ethernet/google/gve/gve_tx.c
694
seg_desc = &tx->desc[idx];
drivers/net/ethernet/google/gve/gve_tx.c
703
idx = (idx + 1) & tx->mask;
drivers/net/ethernet/google/gve/gve_tx.c
704
seg_desc = &tx->desc[idx];
drivers/net/ethernet/google/gve/gve_tx.c
706
addr = skb_frag_dma_map(tx->dev, frag, 0, len, DMA_TO_DEVICE);
drivers/net/ethernet/google/gve/gve_tx.c
707
if (unlikely(dma_mapping_error(tx->dev, addr))) {
drivers/net/ethernet/google/gve/gve_tx.c
708
tx->dma_mapping_error++;
drivers/net/ethernet/google/gve/gve_tx.c
711
tx->info[idx].skb = NULL;
drivers/net/ethernet/google/gve/gve_tx.c
712
dma_unmap_len_set(&tx->info[idx], len, len);
drivers/net/ethernet/google/gve/gve_tx.c
713
dma_unmap_addr_set(&tx->info[idx], dma, addr);
drivers/net/ethernet/google/gve/gve_tx.c
729
gve_tx_unmap_buf(tx->dev, &tx->info[idx & tx->mask]);
drivers/net/ethernet/google/gve/gve_tx.c
732
u64_stats_update_begin(&tx->statss);
drivers/net/ethernet/google/gve/gve_tx.c
733
tx->dropped_pkt++;
drivers/net/ethernet/google/gve/gve_tx.c
734
u64_stats_update_end(&tx->statss);
drivers/net/ethernet/google/gve/gve_tx.c
741
struct gve_tx_ring *tx;
drivers/net/ethernet/google/gve/gve_tx.c
746
tx = &priv->tx[skb_get_queue_mapping(skb)];
drivers/net/ethernet/google/gve/gve_tx.c
747
if (unlikely(gve_maybe_stop_tx(priv, tx, skb))) {
drivers/net/ethernet/google/gve/gve_tx.c
753
gve_tx_put_doorbell(priv, tx->q_resources, tx->req);
drivers/net/ethernet/google/gve/gve_tx.c
756
if (tx->raw_addressing)
drivers/net/ethernet/google/gve/gve_tx.c
757
nsegs = gve_tx_add_skb_no_copy(priv, tx, skb);
drivers/net/ethernet/google/gve/gve_tx.c
759
nsegs = gve_tx_add_skb_copy(priv, tx, skb);
drivers/net/ethernet/google/gve/gve_tx.c
763
netdev_tx_sent_queue(tx->netdev_txq, skb->len);
drivers/net/ethernet/google/gve/gve_tx.c
765
tx->req += nsegs;
drivers/net/ethernet/google/gve/gve_tx.c
770
if (!netif_xmit_stopped(tx->netdev_txq) && netdev_xmit_more())
drivers/net/ethernet/google/gve/gve_tx.c
776
gve_tx_put_doorbell(priv, tx->q_resources, tx->req);
drivers/net/ethernet/google/gve/gve_tx.c
780
static int gve_tx_fill_xdp(struct gve_priv *priv, struct gve_tx_ring *tx,
drivers/net/ethernet/google/gve/gve_tx.c
785
u32 reqi = tx->req;
drivers/net/ethernet/google/gve/gve_tx.c
787
pad = gve_tx_fifo_pad_alloc_one_frag(&tx->tx_fifo, len);
drivers/net/ethernet/google/gve/gve_tx.c
790
info = &tx->info[reqi & tx->mask];
drivers/net/ethernet/google/gve/gve_tx.c
795
nfrags = gve_tx_alloc_fifo(&tx->tx_fifo, pad + len,
drivers/net/ethernet/google/gve/gve_tx.c
803
gve_tx_fill_pkt_desc(&tx->desc[reqi & tx->mask], 0,
drivers/net/ethernet/google/gve/gve_tx.c
808
gve_tx_fill_seg_desc(&tx->desc[reqi & tx->mask],
drivers/net/ethernet/google/gve/gve_tx.c
813
memcpy(tx->tx_fifo.base + info->iov[iovi].iov_offset,
drivers/net/ethernet/google/gve/gve_tx.c
816
tx->tx_fifo.qpl->page_buses,
drivers/net/ethernet/google/gve/gve_tx.c
831
struct gve_tx_ring *tx;
drivers/net/ethernet/google/gve/gve_tx.c
843
tx = &priv->tx[qid];
drivers/net/ethernet/google/gve/gve_tx.c
845
spin_lock(&tx->xdp_lock);
drivers/net/ethernet/google/gve/gve_tx.c
847
err = gve_xdp_xmit_one(priv, tx, frames[i]->data,
drivers/net/ethernet/google/gve/gve_tx.c
854
gve_tx_put_doorbell(priv, tx->q_resources, tx->req);
drivers/net/ethernet/google/gve/gve_tx.c
856
spin_unlock(&tx->xdp_lock);
drivers/net/ethernet/google/gve/gve_tx.c
858
u64_stats_update_begin(&tx->statss);
drivers/net/ethernet/google/gve/gve_tx.c
859
tx->xdp_xmit += n;
drivers/net/ethernet/google/gve/gve_tx.c
860
tx->xdp_xmit_errors += n - i;
drivers/net/ethernet/google/gve/gve_tx.c
861
u64_stats_update_end(&tx->statss);
drivers/net/ethernet/google/gve/gve_tx.c
866
int gve_xdp_xmit_one(struct gve_priv *priv, struct gve_tx_ring *tx,
drivers/net/ethernet/google/gve/gve_tx.c
871
if (!gve_can_tx(tx, len + GVE_GQ_TX_MIN_PKT_DESC_BYTES - 1))
drivers/net/ethernet/google/gve/gve_tx.c
874
nsegs = gve_tx_fill_xdp(priv, tx, data, len, frame_p, false);
drivers/net/ethernet/google/gve/gve_tx.c
875
tx->req += nsegs;
drivers/net/ethernet/google/gve/gve_tx.c
882
static int gve_clean_tx_done(struct gve_priv *priv, struct gve_tx_ring *tx,
drivers/net/ethernet/google/gve/gve_tx.c
893
idx = tx->done & tx->mask;
drivers/net/ethernet/google/gve/gve_tx.c
896
tx->q_num, __func__, idx, tx->req, tx->done);
drivers/net/ethernet/google/gve/gve_tx.c
897
info = &tx->info[idx];
drivers/net/ethernet/google/gve/gve_tx.c
901
if (tx->raw_addressing)
drivers/net/ethernet/google/gve/gve_tx.c
902
gve_tx_unmap_buf(tx->dev, info);
drivers/net/ethernet/google/gve/gve_tx.c
903
tx->done++;
drivers/net/ethernet/google/gve/gve_tx.c
910
if (tx->raw_addressing)
drivers/net/ethernet/google/gve/gve_tx.c
916
if (!tx->raw_addressing)
drivers/net/ethernet/google/gve/gve_tx.c
917
gve_tx_free_fifo(&tx->tx_fifo, space_freed);
drivers/net/ethernet/google/gve/gve_tx.c
918
u64_stats_update_begin(&tx->statss);
drivers/net/ethernet/google/gve/gve_tx.c
919
tx->bytes_done += bytes;
drivers/net/ethernet/google/gve/gve_tx.c
920
tx->pkt_done += pkts;
drivers/net/ethernet/google/gve/gve_tx.c
921
u64_stats_update_end(&tx->statss);
drivers/net/ethernet/google/gve/gve_tx.c
922
netdev_tx_completed_queue(tx->netdev_txq, pkts, bytes);
drivers/net/ethernet/google/gve/gve_tx.c
929
if (try_to_wake && netif_tx_queue_stopped(tx->netdev_txq) &&
drivers/net/ethernet/google/gve/gve_tx.c
930
likely(gve_can_tx(tx, GVE_TX_START_THRESH))) {
drivers/net/ethernet/google/gve/gve_tx.c
931
tx->wake_queue++;
drivers/net/ethernet/google/gve/gve_tx.c
932
netif_tx_wake_queue(tx->netdev_txq);
drivers/net/ethernet/google/gve/gve_tx.c
939
struct gve_tx_ring *tx)
drivers/net/ethernet/google/gve/gve_tx.c
941
u32 counter_index = be32_to_cpu(tx->q_resources->counter_index);
drivers/net/ethernet/google/gve/gve_tx.c
947
static int gve_xsk_tx(struct gve_priv *priv, struct gve_tx_ring *tx,
drivers/net/ethernet/google/gve/gve_tx.c
954
spin_lock(&tx->xdp_lock);
drivers/net/ethernet/google/gve/gve_tx.c
956
if (!gve_can_tx(tx, GVE_TX_START_THRESH) ||
drivers/net/ethernet/google/gve/gve_tx.c
957
!xsk_tx_peek_desc(tx->xsk_pool, &desc))
drivers/net/ethernet/google/gve/gve_tx.c
960
data = xsk_buff_raw_get_data(tx->xsk_pool, desc.addr);
drivers/net/ethernet/google/gve/gve_tx.c
961
nsegs = gve_tx_fill_xdp(priv, tx, data, desc.len, NULL, true);
drivers/net/ethernet/google/gve/gve_tx.c
962
tx->req += nsegs;
drivers/net/ethernet/google/gve/gve_tx.c
967
gve_tx_put_doorbell(priv, tx->q_resources, tx->req);
drivers/net/ethernet/google/gve/gve_tx.c
968
xsk_tx_release(tx->xsk_pool);
drivers/net/ethernet/google/gve/gve_tx.c
970
spin_unlock(&tx->xdp_lock);
drivers/net/ethernet/google/gve/gve_tx.c
978
struct gve_tx_ring *tx;
drivers/net/ethernet/google/gve/gve_tx.c
981
tx = &priv->tx[gve_xdp_tx_queue_id(priv, rx->q_num)];
drivers/net/ethernet/google/gve/gve_tx.c
982
if (tx->xsk_pool) {
drivers/net/ethernet/google/gve/gve_tx.c
983
sent = gve_xsk_tx(priv, tx, budget);
drivers/net/ethernet/google/gve/gve_tx.c
985
u64_stats_update_begin(&tx->statss);
drivers/net/ethernet/google/gve/gve_tx.c
986
tx->xdp_xsk_sent += sent;
drivers/net/ethernet/google/gve/gve_tx.c
987
u64_stats_update_end(&tx->statss);
drivers/net/ethernet/google/gve/gve_tx.c
988
if (xsk_uses_need_wakeup(tx->xsk_pool))
drivers/net/ethernet/google/gve/gve_tx.c
989
xsk_set_tx_need_wakeup(tx->xsk_pool);
drivers/net/ethernet/google/gve/gve_tx.c
998
struct gve_tx_ring *tx = block->tx;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1001
if (unlikely(gve_tx_add_skb_dqo(tx, skb) < 0))
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1004
netdev_tx_sent_queue(tx->netdev_txq, skb->len);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1009
u64_stats_update_begin(&tx->statss);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1010
tx->dropped_pkt++;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1011
u64_stats_update_end(&tx->statss);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1016
static void gve_xsk_reorder_queue_push_dqo(struct gve_tx_ring *tx,
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1019
u32 tail = atomic_read(&tx->dqo_tx.xsk_reorder_queue_tail);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
102
static bool gve_has_pending_packet(struct gve_tx_ring *tx)
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1021
tx->dqo.xsk_reorder_queue[tail] = completion_tag;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1022
tail = (tail + 1) & tx->dqo.complq_mask;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1023
atomic_set_release(&tx->dqo_tx.xsk_reorder_queue_tail, tail);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1027
gve_xsk_reorder_queue_head(struct gve_tx_ring *tx)
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1029
u32 head = tx->dqo_compl.xsk_reorder_queue_head;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1031
if (head == tx->dqo_compl.xsk_reorder_queue_tail) {
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1032
tx->dqo_compl.xsk_reorder_queue_tail =
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1033
atomic_read_acquire(&tx->dqo_tx.xsk_reorder_queue_tail);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1035
if (head == tx->dqo_compl.xsk_reorder_queue_tail)
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1039
return &tx->dqo.pending_packets[tx->dqo.xsk_reorder_queue[head]];
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1042
static void gve_xsk_reorder_queue_pop_dqo(struct gve_tx_ring *tx)
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1044
tx->dqo_compl.xsk_reorder_queue_head++;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1045
tx->dqo_compl.xsk_reorder_queue_head &= tx->dqo.complq_mask;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
105
if (tx->dqo_tx.free_pending_packets != -1)
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1052
struct gve_tx_ring *tx;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1054
tx = &priv->tx[skb_get_queue_mapping(skb)];
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1055
if (unlikely(gve_try_tx_skb(priv, tx, skb) < 0)) {
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1060
gve_tx_put_doorbell_dqo(priv, tx->q_resources, tx->dqo_tx.tail);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1064
if (!netif_xmit_stopped(tx->netdev_txq) && netdev_xmit_more())
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1067
gve_tx_put_doorbell_dqo(priv, tx->q_resources, tx->dqo_tx.tail);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1071
static bool gve_xsk_tx_dqo(struct gve_priv *priv, struct gve_tx_ring *tx,
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1074
struct xsk_buff_pool *pool = tx->xsk_pool;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1079
spin_lock(&tx->dqo_tx.xdp_lock);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1086
if (unlikely(!gve_has_avail_slots_tx_dqo(tx, 1, 1))) {
drivers/net/ethernet/google/gve/gve_tx_dqo.c
109
if (atomic_read_acquire(&tx->dqo_compl.free_pending_packets) != -1)
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1094
pkt = gve_alloc_pending_packet(tx);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1097
completion_tag = pkt - tx->dqo.pending_packets;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1102
desc_idx = tx->dqo_tx.tail;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1103
gve_tx_fill_pkt_desc_dqo(tx, &desc_idx,
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1108
gve_tx_update_tail(tx, desc_idx);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1109
tx->dqo_tx.posted_packet_desc_cnt += pkt->num_bufs;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1110
gve_xsk_reorder_queue_push_dqo(tx, completion_tag);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1114
gve_tx_put_doorbell_dqo(priv, tx->q_resources, tx->dqo_tx.tail);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1118
spin_unlock(&tx->dqo_tx.xdp_lock);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1120
u64_stats_update_begin(&tx->statss);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1121
tx->xdp_xsk_sent += sent;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1122
u64_stats_update_end(&tx->statss);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1127
static void add_to_list(struct gve_tx_ring *tx, struct gve_index_list *list,
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1132
index = pending_packet - tx->dqo.pending_packets;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1138
tx->dqo.pending_packets[old_tail].next = index;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1144
static void remove_from_list(struct gve_tx_ring *tx,
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1157
tx->dqo.pending_packets[prev_index].next = next_index;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1163
tx->dqo.pending_packets[next_index].prev = prev_index;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1174
struct gve_tx_ring *tx, bool is_napi,
drivers/net/ethernet/google/gve/gve_tx_dqo.c
118
struct gve_tx_ring *tx = &priv->tx[tx_qid];
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1180
if (unlikely(compl_tag >= tx->dqo.num_pending_packets)) {
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1186
pending_packet = &tx->dqo.pending_packets[compl_tag];
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1196
remove_from_list(tx,
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1197
&tx->dqo_compl.timed_out_completions,
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1199
gve_free_pending_packet(tx, pending_packet);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
120
gve_tx_put_doorbell_dqo(priv, tx->q_resources, tx->dqo_tx.tail);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1213
remove_from_list(tx, &tx->dqo_compl.miss_completions,
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1224
tx->dqo_tx.completed_packet_desc_cnt += pending_packet->num_bufs;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1228
if (tx->dqo.qpl)
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1229
gve_free_tx_qpl_bufs(tx, pending_packet);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1231
gve_unmap_packet(tx->dev, pending_packet);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1237
gve_free_pending_packet(tx, pending_packet);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
124
gve_alloc_pending_packet(struct gve_tx_ring *tx)
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1240
gve_unmap_packet(tx->dev, pending_packet);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1246
gve_free_pending_packet(tx, pending_packet);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1257
struct gve_tx_ring *tx, u16 compl_tag,
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1262
if (unlikely(compl_tag >= tx->dqo.num_pending_packets)) {
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1268
pending_packet = &tx->dqo.pending_packets[compl_tag];
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1282
add_to_list(tx, &tx->dqo_compl.miss_completions, pending_packet);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1289
struct gve_tx_ring *tx)
drivers/net/ethernet/google/gve/gve_tx_dqo.c
129
index = tx->dqo_tx.free_pending_packets;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1294
next_index = tx->dqo_compl.miss_completions.head;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1296
pending_packet = &tx->dqo.pending_packets[next_index];
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1302
remove_from_list(tx, &tx->dqo_compl.miss_completions,
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1309
if (tx->dqo.qpl)
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1310
gve_free_tx_qpl_bufs(tx, pending_packet);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1312
gve_unmap_packet(tx->dev, pending_packet);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1318
u64_stats_update_begin(&tx->statss);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1319
tx->dropped_pkt++;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1320
u64_stats_update_end(&tx->statss);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1324
(int)(pending_packet - tx->dqo.pending_packets));
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1333
add_to_list(tx, &tx->dqo_compl.timed_out_completions,
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1339
struct gve_tx_ring *tx)
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1344
next_index = tx->dqo_compl.timed_out_completions.head;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1346
pending_packet = &tx->dqo.pending_packets[next_index];
drivers/net/ethernet/google/gve/gve_tx_dqo.c
135
tx->dqo_tx.free_pending_packets =
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1352
remove_from_list(tx, &tx->dqo_compl.timed_out_completions,
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1359
gve_free_pending_packet(tx, pending_packet);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
136
atomic_xchg(&tx->dqo_compl.free_pending_packets, -1);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1363
static void gve_tx_process_xsk_completions(struct gve_tx_ring *tx)
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1369
gve_xsk_reorder_queue_head(tx);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
137
index = tx->dqo_tx.free_pending_packets;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1376
gve_xsk_reorder_queue_pop_dqo(tx);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1377
gve_free_pending_packet(tx, pending_packet);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1381
xsk_tx_completed(tx->xsk_pool, num_xsks);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1384
int gve_clean_tx_done_dqo(struct gve_priv *priv, struct gve_tx_ring *tx,
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1398
&tx->dqo.compl_ring[tx->dqo_compl.head];
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1401
if (compl_desc->generation == tx->dqo_compl.cur_gen_bit)
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1405
prefetch(&tx->dqo.compl_ring[(tx->dqo_compl.head + 1) &
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1406
tx->dqo.complq_mask]);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1416
atomic_set_release(&tx->dqo_compl.hw_tx_head, tx_head);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1421
gve_handle_miss_completion(priv, tx, compl_tag,
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1425
gve_handle_packet_completion(priv, tx, !!napi,
drivers/net/ethernet/google/gve/gve_tx_dqo.c
143
pending_packet = &tx->dqo.pending_packets[index];
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1434
gve_handle_miss_completion(priv, tx, compl_tag,
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1440
gve_handle_packet_completion(priv, tx, !!napi,
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1447
tx->dqo_compl.head =
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1448
(tx->dqo_compl.head + 1) & tx->dqo.complq_mask;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1450
tx->dqo_compl.cur_gen_bit ^= tx->dqo_compl.head == 0;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1454
if (tx->netdev_txq)
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1455
netdev_tx_completed_queue(tx->netdev_txq,
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1459
remove_miss_completions(priv, tx);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
146
tx->dqo_tx.free_pending_packets = pending_packet->next;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1460
remove_timed_out_completions(priv, tx);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1462
if (tx->xsk_pool)
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1463
gve_tx_process_xsk_completions(tx);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1465
u64_stats_update_begin(&tx->statss);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1466
tx->bytes_done += pkt_compl_bytes + reinject_compl_bytes;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1467
tx->pkt_done += pkt_compl_pkts + reinject_compl_pkts;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1468
u64_stats_update_end(&tx->statss);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1475
struct gve_tx_ring *tx = block->tx;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1479
int num_descs_cleaned = gve_clean_tx_done_dqo(priv, tx,
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1485
if (netif_tx_queue_stopped(tx->netdev_txq) &&
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1487
tx->wake_queue++;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1488
netif_tx_wake_queue(tx->netdev_txq);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1493
compl_desc = &tx->dqo.compl_ring[tx->dqo_compl.head];
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1494
return compl_desc->generation != tx->dqo_compl.cur_gen_bit;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1501
struct gve_tx_ring *tx;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1503
tx = &priv->tx[gve_xdp_tx_queue_id(priv, rx->q_num)];
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1504
if (tx->xsk_pool)
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1505
return gve_xsk_tx_dqo(priv, tx, budget);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1513
struct gve_tx_ring *tx = block->tx;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1516
gve_clean_tx_done_dqo(priv, tx, &block->napi);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1519
compl_desc = &tx->dqo.compl_ring[tx->dqo_compl.head];
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1520
return compl_desc->generation != tx->dqo_compl.cur_gen_bit;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1523
int gve_xdp_xmit_one_dqo(struct gve_priv *priv, struct gve_tx_ring *tx,
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1527
u32 desc_idx = tx->dqo_tx.tail;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
153
gve_free_pending_packet(struct gve_tx_ring *tx,
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1533
if (unlikely(!gve_has_tx_slots_available(tx, num_descs)))
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1536
pkt = gve_alloc_pending_packet(tx);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1543
completion_tag = pkt - tx->dqo.pending_packets;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1546
addr = dma_map_single(tx->dev, xdpf->data, xdpf->len, DMA_TO_DEVICE);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1547
err = dma_mapping_error(tx->dev, addr);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1555
gve_tx_fill_pkt_desc_dqo(tx, &desc_idx,
drivers/net/ethernet/google/gve/gve_tx_dqo.c
156
s16 index = pending_packet - tx->dqo.pending_packets;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1560
gve_tx_update_tail(tx, desc_idx);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1566
gve_free_pending_packet(tx, pkt);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1574
struct gve_tx_ring *tx;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1583
tx = &priv->tx[qid];
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1585
spin_lock(&tx->dqo_tx.xdp_lock);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1587
err = gve_xdp_xmit_one_dqo(priv, tx, frames[i]);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1593
gve_tx_put_doorbell_dqo(priv, tx->q_resources, tx->dqo_tx.tail);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1595
spin_unlock(&tx->dqo_tx.xdp_lock);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1597
u64_stats_update_begin(&tx->statss);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1598
tx->xdp_xmit += n;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1599
tx->xdp_xmit_errors += n - i;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
160
s16 old_head = atomic_read_acquire(&tx->dqo_compl.free_pending_packets);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1600
u64_stats_update_end(&tx->statss);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
163
if (atomic_cmpxchg(&tx->dqo_compl.free_pending_packets,
drivers/net/ethernet/google/gve/gve_tx_dqo.c
19
static bool gve_has_free_tx_qpl_bufs(struct gve_tx_ring *tx, int count)
drivers/net/ethernet/google/gve/gve_tx_dqo.c
191
static void gve_tx_clean_pending_packets(struct gve_tx_ring *tx)
drivers/net/ethernet/google/gve/gve_tx_dqo.c
195
for (i = 0; i < tx->dqo.num_pending_packets; i++) {
drivers/net/ethernet/google/gve/gve_tx_dqo.c
197
&tx->dqo.pending_packets[i];
drivers/net/ethernet/google/gve/gve_tx_dqo.c
199
if (tx->dqo.qpl)
drivers/net/ethernet/google/gve/gve_tx_dqo.c
200
gve_free_tx_qpl_bufs(tx, cur_state);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
202
gve_unmap_packet(tx->dev, cur_state);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
214
struct gve_tx_ring *tx = &priv->tx[idx];
drivers/net/ethernet/google/gve/gve_tx_dqo.c
220
gve_clean_tx_done_dqo(priv, tx, /*napi=*/NULL);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
221
if (tx->netdev_txq)
drivers/net/ethernet/google/gve/gve_tx_dqo.c
222
netdev_tx_reset_queue(tx->netdev_txq);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
223
gve_tx_clean_pending_packets(tx);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
227
static void gve_tx_free_ring_dqo(struct gve_priv *priv, struct gve_tx_ring *tx,
drivers/net/ethernet/google/gve/gve_tx_dqo.c
23
if (!tx->dqo.qpl)
drivers/net/ethernet/google/gve/gve_tx_dqo.c
231
int idx = tx->q_num;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
235
if (tx->q_resources) {
drivers/net/ethernet/google/gve/gve_tx_dqo.c
236
dma_free_coherent(hdev, sizeof(*tx->q_resources),
drivers/net/ethernet/google/gve/gve_tx_dqo.c
237
tx->q_resources, tx->q_resources_bus);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
238
tx->q_resources = NULL;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
241
if (tx->dqo.compl_ring) {
drivers/net/ethernet/google/gve/gve_tx_dqo.c
242
bytes = sizeof(tx->dqo.compl_ring[0]) *
drivers/net/ethernet/google/gve/gve_tx_dqo.c
243
(tx->dqo.complq_mask + 1);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
244
dma_free_coherent(hdev, bytes, tx->dqo.compl_ring,
drivers/net/ethernet/google/gve/gve_tx_dqo.c
245
tx->complq_bus_dqo);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
246
tx->dqo.compl_ring = NULL;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
249
if (tx->dqo.tx_ring) {
drivers/net/ethernet/google/gve/gve_tx_dqo.c
250
bytes = sizeof(tx->dqo.tx_ring[0]) * (tx->mask + 1);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
251
dma_free_coherent(hdev, bytes, tx->dqo.tx_ring, tx->bus);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
252
tx->dqo.tx_ring = NULL;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
255
kvfree(tx->dqo.xsk_reorder_queue);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
256
tx->dqo.xsk_reorder_queue = NULL;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
258
kvfree(tx->dqo.pending_packets);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
259
tx->dqo.pending_packets = NULL;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
26
num_avail = tx->dqo.num_tx_qpl_bufs -
drivers/net/ethernet/google/gve/gve_tx_dqo.c
261
kvfree(tx->dqo.tx_qpl_buf_next);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
262
tx->dqo.tx_qpl_buf_next = NULL;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
264
if (tx->dqo.qpl) {
drivers/net/ethernet/google/gve/gve_tx_dqo.c
265
qpl_id = gve_tx_qpl_id(priv, tx->q_num);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
266
gve_free_queue_page_list(priv, tx->dqo.qpl, qpl_id);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
267
tx->dqo.qpl = NULL;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
27
(tx->dqo_tx.alloc_tx_qpl_buf_cnt -
drivers/net/ethernet/google/gve/gve_tx_dqo.c
273
static int gve_tx_qpl_buf_init(struct gve_tx_ring *tx)
drivers/net/ethernet/google/gve/gve_tx_dqo.c
276
tx->dqo.qpl->num_entries;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
279
tx->dqo.tx_qpl_buf_next = kvzalloc_objs(tx->dqo.tx_qpl_buf_next[0],
drivers/net/ethernet/google/gve/gve_tx_dqo.c
28
tx->dqo_tx.free_tx_qpl_buf_cnt);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
281
if (!tx->dqo.tx_qpl_buf_next)
drivers/net/ethernet/google/gve/gve_tx_dqo.c
284
tx->dqo.num_tx_qpl_bufs = num_tx_qpl_bufs;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
288
tx->dqo.tx_qpl_buf_next[i] = i + 1;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
289
tx->dqo.tx_qpl_buf_next[num_tx_qpl_bufs - 1] = -1;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
291
atomic_set_release(&tx->dqo_compl.free_tx_qpl_buf_head, -1);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
298
struct gve_tx_ring *tx = &priv->tx[idx];
drivers/net/ethernet/google/gve/gve_tx_dqo.c
303
tx->netdev_txq = netdev_get_tx_queue(priv->dev, idx);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
309
struct gve_tx_ring *tx,
drivers/net/ethernet/google/gve/gve_tx_dqo.c
319
memset(tx, 0, sizeof(*tx));
drivers/net/ethernet/google/gve/gve_tx_dqo.c
320
tx->q_num = idx;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
321
tx->dev = hdev;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
322
spin_lock_init(&tx->dqo_tx.xdp_lock);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
323
atomic_set_release(&tx->dqo_compl.hw_tx_head, 0);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
326
tx->mask = cfg->ring_size - 1;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
327
tx->dqo.complq_mask = tx->mask;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
335
num_pending_packets = tx->dqo.complq_mask + 1;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
34
tx->dqo_tx.free_tx_qpl_buf_cnt =
drivers/net/ethernet/google/gve/gve_tx_dqo.c
341
(tx->dqo.complq_mask + 1) / GVE_TX_MIN_RE_INTERVAL;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
348
tx->dqo.num_pending_packets = min_t(int, num_pending_packets, S16_MAX);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
349
tx->dqo.pending_packets = kvzalloc_objs(tx->dqo.pending_packets[0],
drivers/net/ethernet/google/gve/gve_tx_dqo.c
35
atomic_read_acquire(&tx->dqo_compl.free_tx_qpl_buf_cnt);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
350
tx->dqo.num_pending_packets);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
351
if (!tx->dqo.pending_packets)
drivers/net/ethernet/google/gve/gve_tx_dqo.c
355
for (i = 0; i < tx->dqo.num_pending_packets - 1; i++)
drivers/net/ethernet/google/gve/gve_tx_dqo.c
356
tx->dqo.pending_packets[i].next = i + 1;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
358
tx->dqo.pending_packets[tx->dqo.num_pending_packets - 1].next = -1;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
359
atomic_set_release(&tx->dqo_compl.free_pending_packets, -1);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
363
tx->dqo.xsk_reorder_queue =
drivers/net/ethernet/google/gve/gve_tx_dqo.c
364
kvcalloc(tx->dqo.complq_mask + 1,
drivers/net/ethernet/google/gve/gve_tx_dqo.c
365
sizeof(tx->dqo.xsk_reorder_queue[0]),
drivers/net/ethernet/google/gve/gve_tx_dqo.c
367
if (!tx->dqo.xsk_reorder_queue)
drivers/net/ethernet/google/gve/gve_tx_dqo.c
37
num_avail = tx->dqo.num_tx_qpl_bufs -
drivers/net/ethernet/google/gve/gve_tx_dqo.c
371
tx->dqo_compl.miss_completions.head = -1;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
372
tx->dqo_compl.miss_completions.tail = -1;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
373
tx->dqo_compl.timed_out_completions.head = -1;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
374
tx->dqo_compl.timed_out_completions.tail = -1;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
376
bytes = sizeof(tx->dqo.tx_ring[0]) * (tx->mask + 1);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
377
tx->dqo.tx_ring = dma_alloc_coherent(hdev, bytes, &tx->bus, GFP_KERNEL);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
378
if (!tx->dqo.tx_ring)
drivers/net/ethernet/google/gve/gve_tx_dqo.c
38
(tx->dqo_tx.alloc_tx_qpl_buf_cnt -
drivers/net/ethernet/google/gve/gve_tx_dqo.c
381
bytes = sizeof(tx->dqo.compl_ring[0]) * (tx->dqo.complq_mask + 1);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
382
tx->dqo.compl_ring = dma_alloc_coherent(hdev, bytes,
drivers/net/ethernet/google/gve/gve_tx_dqo.c
383
&tx->complq_bus_dqo,
drivers/net/ethernet/google/gve/gve_tx_dqo.c
385
if (!tx->dqo.compl_ring)
drivers/net/ethernet/google/gve/gve_tx_dqo.c
388
tx->q_resources = dma_alloc_coherent(hdev, sizeof(*tx->q_resources),
drivers/net/ethernet/google/gve/gve_tx_dqo.c
389
&tx->q_resources_bus, GFP_KERNEL);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
39
tx->dqo_tx.free_tx_qpl_buf_cnt);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
390
if (!tx->q_resources)
drivers/net/ethernet/google/gve/gve_tx_dqo.c
394
qpl_id = gve_tx_qpl_id(priv, tx->q_num);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
397
tx->dqo.qpl = gve_alloc_queue_page_list(priv, qpl_id,
drivers/net/ethernet/google/gve/gve_tx_dqo.c
399
if (!tx->dqo.qpl)
drivers/net/ethernet/google/gve/gve_tx_dqo.c
402
if (gve_tx_qpl_buf_init(tx))
drivers/net/ethernet/google/gve/gve_tx_dqo.c
409
gve_tx_free_ring_dqo(priv, tx, cfg);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
416
struct gve_tx_ring *tx = cfg->tx;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
428
tx = kvzalloc_objs(struct gve_tx_ring, cfg->qcfg->max_queues);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
429
if (!tx)
drivers/net/ethernet/google/gve/gve_tx_dqo.c
433
err = gve_tx_alloc_ring_dqo(priv, cfg, &tx[i], i);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
442
cfg->tx = tx;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
447
gve_tx_free_ring_dqo(priv, &tx[j], cfg);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
448
kvfree(tx);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
45
gve_alloc_tx_qpl_buf(struct gve_tx_ring *tx)
drivers/net/ethernet/google/gve/gve_tx_dqo.c
455
struct gve_tx_ring *tx = cfg->tx;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
458
if (!tx)
drivers/net/ethernet/google/gve/gve_tx_dqo.c
462
gve_tx_free_ring_dqo(priv, &tx[i], cfg);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
464
kvfree(tx);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
465
cfg->tx = NULL;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
469
static u32 num_avail_tx_slots(const struct gve_tx_ring *tx)
drivers/net/ethernet/google/gve/gve_tx_dqo.c
471
u32 num_used = (tx->dqo_tx.tail - tx->dqo_tx.head) & tx->mask;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
473
return tx->mask - num_used;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
477
static bool gve_has_tx_slots_available(struct gve_tx_ring *tx, u32 slots_req)
drivers/net/ethernet/google/gve/gve_tx_dqo.c
479
u32 num_avail = num_avail_tx_slots(tx);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
487
tx->dqo_tx.head = atomic_read_acquire(&tx->dqo_compl.hw_tx_head);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
489
return num_avail_tx_slots(tx) >= slots_req;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
49
index = tx->dqo_tx.free_tx_qpl_buf_head;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
492
static bool gve_has_avail_slots_tx_dqo(struct gve_tx_ring *tx,
drivers/net/ethernet/google/gve/gve_tx_dqo.c
495
return gve_has_pending_packet(tx) &&
drivers/net/ethernet/google/gve/gve_tx_dqo.c
496
gve_has_tx_slots_available(tx, desc_count) &&
drivers/net/ethernet/google/gve/gve_tx_dqo.c
497
gve_has_free_tx_qpl_bufs(tx, buf_count);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
503
static int gve_maybe_stop_tx_dqo(struct gve_tx_ring *tx,
drivers/net/ethernet/google/gve/gve_tx_dqo.c
506
if (likely(gve_has_avail_slots_tx_dqo(tx, desc_count, buf_count)))
drivers/net/ethernet/google/gve/gve_tx_dqo.c
510
tx->stop_queue++;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
511
netif_tx_stop_queue(tx->netdev_txq);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
519
if (likely(!gve_has_avail_slots_tx_dqo(tx, desc_count, buf_count)))
drivers/net/ethernet/google/gve/gve_tx_dqo.c
522
netif_tx_start_queue(tx->netdev_txq);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
523
tx->wake_queue++;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
544
static void gve_tx_fill_pkt_desc_dqo(struct gve_tx_ring *tx, u32 *desc_idx,
drivers/net/ethernet/google/gve/gve_tx_dqo.c
55
tx->dqo_tx.free_tx_qpl_buf_head =
drivers/net/ethernet/google/gve/gve_tx_dqo.c
550
&tx->dqo.tx_ring[*desc_idx].pkt;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
56
atomic_xchg(&tx->dqo_compl.free_tx_qpl_buf_head, -1);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
565
*desc_idx = (*desc_idx + 1) & tx->mask;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
57
index = tx->dqo_tx.free_tx_qpl_buf_head;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
64
tx->dqo_tx.free_tx_qpl_buf_head = tx->dqo.tx_qpl_buf_next[index];
drivers/net/ethernet/google/gve/gve_tx_dqo.c
655
static void gve_tx_update_tail(struct gve_tx_ring *tx, u32 desc_idx)
drivers/net/ethernet/google/gve/gve_tx_dqo.c
657
u32 last_desc_idx = (desc_idx - 1) & tx->mask;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
659
(last_desc_idx - tx->dqo_tx.last_re_idx) & tx->mask;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
662
tx->dqo_tx.tail = desc_idx;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
669
tx->dqo.tx_ring[last_desc_idx].pkt.report_event = true;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
670
tx->dqo_tx.last_re_idx = last_desc_idx;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
674
static int gve_tx_add_skb_no_copy_dqo(struct gve_tx_ring *tx,
drivers/net/ethernet/google/gve/gve_tx_dqo.c
699
addr = dma_map_single(tx->dev, skb->data, len, DMA_TO_DEVICE);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
70
gve_free_tx_qpl_bufs(struct gve_tx_ring *tx,
drivers/net/ethernet/google/gve/gve_tx_dqo.c
700
if (unlikely(dma_mapping_error(tx->dev, addr)))
drivers/net/ethernet/google/gve/gve_tx_dqo.c
707
gve_tx_fill_pkt_desc_dqo(tx, desc_idx, enable_csum, len, addr,
drivers/net/ethernet/google/gve/gve_tx_dqo.c
718
addr = skb_frag_dma_map(tx->dev, frag, 0, len, DMA_TO_DEVICE);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
719
if (unlikely(dma_mapping_error(tx->dev, addr)))
drivers/net/ethernet/google/gve/gve_tx_dqo.c
727
gve_tx_fill_pkt_desc_dqo(tx, desc_idx, enable_csum, len, addr,
drivers/net/ethernet/google/gve/gve_tx_dqo.c
735
dma_unmap_single(tx->dev,
drivers/net/ethernet/google/gve/gve_tx_dqo.c
740
dma_unmap_page(tx->dev,
drivers/net/ethernet/google/gve/gve_tx_dqo.c
754
static void gve_tx_buf_get_addr(struct gve_tx_ring *tx,
drivers/net/ethernet/google/gve/gve_tx_dqo.c
761
*va = page_address(tx->dqo.qpl->pages[page_id]) + offset;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
762
*dma_addr = tx->dqo.qpl->page_buses[page_id] + offset;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
765
static int gve_tx_add_skb_copy_dqo(struct gve_tx_ring *tx,
drivers/net/ethernet/google/gve/gve_tx_dqo.c
782
index = gve_alloc_tx_qpl_buf(tx);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
786
gve_tx_buf_get_addr(tx, index, &va, &dma_addr);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
792
dma_sync_single_for_device(tx->dev, dma_addr,
drivers/net/ethernet/google/gve/gve_tx_dqo.c
794
gve_tx_fill_pkt_desc_dqo(tx, desc_idx, enable_csum,
drivers/net/ethernet/google/gve/gve_tx_dqo.c
802
++tx->dqo_tx.alloc_tx_qpl_buf_cnt;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
809
gve_free_tx_qpl_bufs(tx, pkt);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
818
static int gve_tx_add_skb_dqo(struct gve_tx_ring *tx,
drivers/net/ethernet/google/gve/gve_tx_dqo.c
82
tx->dqo.tx_qpl_buf_next[index] = pkt->tx_qpl_buf_ids[i];
drivers/net/ethernet/google/gve/gve_tx_dqo.c
822
u32 desc_idx = tx->dqo_tx.tail;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
827
pkt = gve_alloc_pending_packet(tx);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
833
completion_tag = pkt - tx->dqo.pending_packets;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
842
gve_tx_fill_tso_ctx_desc(&tx->dqo.tx_ring[desc_idx].tso_ctx,
drivers/net/ethernet/google/gve/gve_tx_dqo.c
844
desc_idx = (desc_idx + 1) & tx->mask;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
847
gve_tx_fill_general_ctx_desc(&tx->dqo.tx_ring[desc_idx].general_ctx,
drivers/net/ethernet/google/gve/gve_tx_dqo.c
849
desc_idx = (desc_idx + 1) & tx->mask;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
851
if (tx->dqo.qpl) {
drivers/net/ethernet/google/gve/gve_tx_dqo.c
852
if (gve_tx_add_skb_copy_dqo(tx, skb, pkt,
drivers/net/ethernet/google/gve/gve_tx_dqo.c
857
if (gve_tx_add_skb_no_copy_dqo(tx, skb, pkt,
drivers/net/ethernet/google/gve/gve_tx_dqo.c
863
tx->dqo_tx.posted_packet_desc_cnt += pkt->num_bufs;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
865
gve_tx_update_tail(tx, desc_idx);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
87
s16 old_head = atomic_read_acquire(&tx->dqo_compl.free_tx_qpl_buf_head);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
870
gve_free_pending_packet(tx, pkt);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
89
tx->dqo.tx_qpl_buf_next[index] = old_head;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
90
if (atomic_cmpxchg(&tx->dqo_compl.free_tx_qpl_buf_head,
drivers/net/ethernet/google/gve/gve_tx_dqo.c
967
static int gve_try_tx_skb(struct gve_priv *priv, struct gve_tx_ring *tx,
drivers/net/ethernet/google/gve/gve_tx_dqo.c
97
atomic_add(pkt->num_bufs, &tx->dqo_compl.free_tx_qpl_buf_cnt);
drivers/net/ethernet/google/gve/gve_tx_dqo.c
973
if (tx->dqo.qpl) {
drivers/net/ethernet/google/gve/gve_tx_dqo.c
996
if (unlikely(gve_maybe_stop_tx_dqo(tx, total_num_descs,
drivers/net/ethernet/google/gve/gve_utils.c
16
return block->tx != NULL;
drivers/net/ethernet/google/gve/gve_utils.c
24
block->tx = NULL;
drivers/net/ethernet/google/gve/gve_utils.c
33
struct gve_tx_ring *tx = &priv->tx[queue_idx];
drivers/net/ethernet/google/gve/gve_utils.c
35
block->tx = tx;
drivers/net/ethernet/google/gve/gve_utils.c
36
tx->ntfy_id = ntfy_idx;
drivers/net/ethernet/hisilicon/hix5hd2_gmac.c
331
static void hix5hd2_set_desc_depth(struct hix5hd2_priv *priv, int rx, int tx)
drivers/net/ethernet/hisilicon/hix5hd2_gmac.c
342
writel_relaxed(tx << 3, priv->base + TX_BQ_DEPTH);
drivers/net/ethernet/hisilicon/hix5hd2_gmac.c
346
writel_relaxed(tx << 3, priv->base + TX_RQ_DEPTH);
drivers/net/ethernet/hisilicon/hns/hnae.h
194
} tx;
drivers/net/ethernet/hisilicon/hns/hns_dsaf_gmac.c
102
*tx = dsaf_get_bit(porten, GMAC_PORT_TX_EN_B);
drivers/net/ethernet/hisilicon/hns/hns_dsaf_gmac.c
495
u32 tx;
drivers/net/ethernet/hisilicon/hns/hns_dsaf_gmac.c
506
hns_gmac_get_en(mac_drv, &rx, &tx);
drivers/net/ethernet/hisilicon/hns/hns_dsaf_gmac.c
507
mac_info->port_en = rx && tx;
drivers/net/ethernet/hisilicon/hns/hns_dsaf_gmac.c
96
static void hns_gmac_get_en(void *mac_drv, u32 *rx, u32 *tx)
drivers/net/ethernet/hisilicon/hns/hns_enet.c
112
desc->tx.ip_offset = ip_offset;
drivers/net/ethernet/hisilicon/hns/hns_enet.c
113
desc->tx.tse_vlan_snap_v6_sctp_nth = tvsvsn;
drivers/net/ethernet/hisilicon/hns/hns_enet.c
114
desc->tx.mss = cpu_to_le16(mss);
drivers/net/ethernet/hisilicon/hns/hns_enet.c
115
desc->tx.l4_len = l4_len;
drivers/net/ethernet/hisilicon/hns/hns_enet.c
116
desc->tx.paylen = cpu_to_le16(paylen);
drivers/net/ethernet/hisilicon/hns/hns_enet.c
122
desc->tx.bn_pid = bn_pid;
drivers/net/ethernet/hisilicon/hns/hns_enet.c
123
desc->tx.ra_ri_cs_fe_vld = rrcfv;
drivers/net/ethernet/hisilicon/hns/hns_enet.c
162
desc->tx.send_size = cpu_to_le16((u16)size);
drivers/net/ethernet/hisilicon/hns/hns_enet.c
199
desc->tx.asid_bufnum_pid = cpu_to_le16(asid_bufnum_pid);
drivers/net/ethernet/hisilicon/hns/hns_enet.c
200
desc->tx.flag_ipoffset = cpu_to_le32(flag_ipoffset);
drivers/net/ethernet/hisilicon/hns/hns_enet.c
59
desc->tx.send_size = cpu_to_le16((u16)send_sz);
drivers/net/ethernet/hisilicon/hns3/hns3_debugfs.c
671
le16_to_cpu(desc->tx.vlan_tag),
drivers/net/ethernet/hisilicon/hns3/hns3_debugfs.c
672
le16_to_cpu(desc->tx.send_size),
drivers/net/ethernet/hisilicon/hns3/hns3_debugfs.c
673
le32_to_cpu(desc->tx.type_cs_vlan_tso_len),
drivers/net/ethernet/hisilicon/hns3/hns3_debugfs.c
674
le16_to_cpu(desc->tx.outer_vlan_tag),
drivers/net/ethernet/hisilicon/hns3/hns3_debugfs.c
675
le16_to_cpu(desc->tx.tv),
drivers/net/ethernet/hisilicon/hns3/hns3_debugfs.c
676
le32_to_cpu(desc->tx.ol_type_vlan_len_msec),
drivers/net/ethernet/hisilicon/hns3/hns3_debugfs.c
677
le32_to_cpu(desc->tx.paylen_ol4cs),
drivers/net/ethernet/hisilicon/hns3/hns3_debugfs.c
678
le16_to_cpu(desc->tx.bdtp_fe_sc_vld_ra_ri),
drivers/net/ethernet/hisilicon/hns3/hns3_debugfs.c
679
le16_to_cpu(desc->tx.mss_hw_csum));
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
1688
desc->tx.ol_type_vlan_len_msec =
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
1690
desc->tx.type_cs_vlan_tso_len = cpu_to_le32(param.type_cs_vlan_tso);
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
1691
desc->tx.paylen_ol4cs = cpu_to_le32(param.paylen_ol4cs);
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
1692
desc->tx.mss_hw_csum = cpu_to_le16(param.mss_hw_csum);
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
1693
desc->tx.vlan_tag = cpu_to_le16(param.inner_vtag);
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
1694
desc->tx.outer_vlan_tag = cpu_to_le16(param.out_vtag);
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
1710
desc->tx.send_size = cpu_to_le16(size);
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
1711
desc->tx.bdtp_fe_sc_vld_ra_ri =
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
1727
desc->tx.send_size = cpu_to_le16((k == frag_buf_num - 1) ?
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
1729
desc->tx.bdtp_fe_sc_vld_ra_ri =
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
2160
desc->tx.bdtp_fe_sc_vld_ra_ri |= cpu_to_le16(BIT(HNS3_TXD_TSYN_B));
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
2355
ring->desc[pre_ntu].tx.bdtp_fe_sc_vld_ra_ri |=
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
3641
if (le16_to_cpu(desc->tx.bdtp_fe_sc_vld_ra_ri) &
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
5563
ring->desc[ring->next_to_clean].tx.bdtp_fe_sc_vld_ra_ri = 0;
drivers/net/ethernet/hisilicon/hns3/hns3_enet.h
293
} tx;
drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_tm.c
138
static void hclge_pfc_stats_get(struct hclge_dev *hdev, bool tx, u64 *stats)
drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_tm.c
143
if (tx)
drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_tm.c
162
int hclge_mac_pause_en_cfg(struct hclge_dev *hdev, bool tx, bool rx)
drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_tm.c
168
desc.data[0] = cpu_to_le32((tx ? HCLGE_TX_MAC_PAUSE_EN_MSK : 0) |
drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_tm.h
246
int hclge_mac_pause_en_cfg(struct hclge_dev *hdev, bool tx, bool rx);
drivers/net/ethernet/i825xx/ether1.c
673
tx_t tx;
drivers/net/ethernet/i825xx/ether1.c
701
tx.tx_status = 0;
drivers/net/ethernet/i825xx/ether1.c
702
tx.tx_command = CMD_TX | CMD_INTR;
drivers/net/ethernet/i825xx/ether1.c
703
tx.tx_link = nopaddr;
drivers/net/ethernet/i825xx/ether1.c
704
tx.tx_tbdoffset = tbdaddr;
drivers/net/ethernet/i825xx/ether1.c
714
ether1_writebuffer (dev, &tx, txaddr, TX_SIZE);
drivers/net/ethernet/intel/fm10k/fm10k.h
189
struct fm10k_ring_container rx, tx;
drivers/net/ethernet/intel/fm10k/fm10k_debugfs.c
165
for (i = 0; i < q_vector->tx.count; i++) {
drivers/net/ethernet/intel/fm10k/fm10k_debugfs.c
166
struct fm10k_ring *ring = &q_vector->tx.ring[i];
drivers/net/ethernet/intel/fm10k/fm10k_ethtool.c
687
qv->tx.itr = tx_itr;
drivers/net/ethernet/intel/fm10k/fm10k_main.c
1185
unsigned int budget = q_vector->tx.work_limit;
drivers/net/ethernet/intel/fm10k/fm10k_main.c
1273
q_vector->tx.total_bytes += total_bytes;
drivers/net/ethernet/intel/fm10k/fm10k_main.c
1274
q_vector->tx.total_packets += total_packets;
drivers/net/ethernet/intel/fm10k/fm10k_main.c
1405
fm10k_update_itr(&q_vector->tx);
drivers/net/ethernet/intel/fm10k/fm10k_main.c
1411
itr |= (q_vector->tx.itr & FM10K_ITR_MAX);
drivers/net/ethernet/intel/fm10k/fm10k_main.c
1428
fm10k_for_each_ring(ring, q_vector->tx) {
drivers/net/ethernet/intel/fm10k/fm10k_main.c
1616
q_vector->tx.ring = ring;
drivers/net/ethernet/intel/fm10k/fm10k_main.c
1617
q_vector->tx.work_limit = FM10K_DEFAULT_TX_WORK;
drivers/net/ethernet/intel/fm10k/fm10k_main.c
1618
q_vector->tx.itr = interface->tx_itr;
drivers/net/ethernet/intel/fm10k/fm10k_main.c
1619
q_vector->tx.itr_scale = interface->hw.mac.itr_scale;
drivers/net/ethernet/intel/fm10k/fm10k_main.c
1620
q_vector->tx.count = txr_count;
drivers/net/ethernet/intel/fm10k/fm10k_main.c
1696
fm10k_for_each_ring(ring, q_vector->tx)
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1071
while (fm10k_fifo_head_len(&mbx->tx) && mbx->pulled) {
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1072
len = fm10k_fifo_head_drop(&mbx->tx);
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1107
for (len = fm10k_fifo_head_len(&mbx->tx);
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1109
len = fm10k_fifo_head_len(&mbx->tx)) {
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1110
fm10k_fifo_head_drop(&mbx->tx);
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1418
fm10k_fifo_drop_all(&mbx->tx);
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1608
fm10k_fifo_init(&mbx->tx, mbx->buffer, FM10K_MBX_TX_BUFFER_SIZE);
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1759
fm10k_fifo_drop_all(&mbx->tx);
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1921
struct fm10k_mbx_fifo *fifo = &mbx->tx;
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
2167
fm10k_fifo_init(&mbx->tx, mbx->buffer, FM10K_MBX_TX_BUFFER_SIZE);
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
333
struct fm10k_mbx_fifo *fifo = &mbx->tx;
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
386
struct fm10k_mbx_fifo *fifo = &mbx->tx;
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
622
head = fm10k_fifo_head_offset(&mbx->tx, mbx->pulled);
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
625
mbx->local = fm10k_fifo_crc(&mbx->tx, head, len, mbx->local);
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
678
u16 fifo_unused = fm10k_fifo_unused(&mbx->tx);
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
691
return fm10k_fifo_empty(&mbx->tx);
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
754
err = fm10k_fifo_enqueue(&mbx->tx, msg);
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
761
err = fm10k_fifo_enqueue(&mbx->tx, msg);
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
856
struct fm10k_mbx_fifo *fifo = &mbx->tx;
drivers/net/ethernet/intel/fm10k/fm10k_mbx.h
257
struct fm10k_mbx_fifo tx;
drivers/net/ethernet/intel/fm10k/fm10k_pci.c
1187
if (q_vector->rx.count || q_vector->tx.count)
drivers/net/ethernet/intel/fm10k/fm10k_pci.c
1734
if (!q_vector->tx.count && !q_vector->rx.count)
drivers/net/ethernet/intel/fm10k/fm10k_pci.c
1768
if (q_vector->tx.count && q_vector->rx.count) {
drivers/net/ethernet/intel/fm10k/fm10k_pci.c
1775
} else if (q_vector->tx.count) {
drivers/net/ethernet/intel/fm10k/fm10k_pci.c
1818
if (!q_vector->tx.count && !q_vector->rx.count)
drivers/net/ethernet/intel/fm10k/fm10k_pci.c
722
if (!qv->tx.count && !qv->rx.count)
drivers/net/ethernet/intel/i40e/i40e.h
956
struct i40e_ring_container tx;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2991
q_vector->tx.target_itr = ITR_TO_REG(tx_ring->itr_setting);
drivers/net/ethernet/intel/i40e/i40e_main.c
13390
if (q_vector->rx.ring || q_vector->tx.ring) {
drivers/net/ethernet/intel/i40e/i40e_main.c
3901
q_vector->tx.next_update = jiffies + 1;
drivers/net/ethernet/intel/i40e/i40e_main.c
3902
q_vector->tx.target_itr =
drivers/net/ethernet/intel/i40e/i40e_main.c
3905
q_vector->tx.target_itr >> 1);
drivers/net/ethernet/intel/i40e/i40e_main.c
3906
q_vector->tx.current_itr = q_vector->tx.target_itr;
drivers/net/ethernet/intel/i40e/i40e_main.c
4018
q_vector->tx.next_update = jiffies + 1;
drivers/net/ethernet/intel/i40e/i40e_main.c
4019
q_vector->tx.target_itr = ITR_TO_REG(vsi->tx_rings[0]->itr_setting);
drivers/net/ethernet/intel/i40e/i40e_main.c
4020
wr32(hw, I40E_PFINT_ITR0(I40E_TX_ITR), q_vector->tx.target_itr >> 1);
drivers/net/ethernet/intel/i40e/i40e_main.c
4021
q_vector->tx.current_itr = q_vector->tx.target_itr;
drivers/net/ethernet/intel/i40e/i40e_main.c
4084
if (!q_vector->tx.ring && !q_vector->rx.ring)
drivers/net/ethernet/intel/i40e/i40e_main.c
4142
if (q_vector->tx.ring && q_vector->rx.ring) {
drivers/net/ethernet/intel/i40e/i40e_main.c
4149
} else if (q_vector->tx.ring) {
drivers/net/ethernet/intel/i40e/i40e_main.c
4515
if (!q_vector->tx.ring)
drivers/net/ethernet/intel/i40e/i40e_main.c
4518
vsi = q_vector->tx.ring->vsi;
drivers/net/ethernet/intel/i40e/i40e_main.c
4519
i40e_clean_fdir_tx_irq(q_vector->tx.ring, vsi->work_limit);
drivers/net/ethernet/intel/i40e/i40e_main.c
4537
tx_ring->next = q_vector->tx.ring;
drivers/net/ethernet/intel/i40e/i40e_main.c
4538
q_vector->tx.ring = tx_ring;
drivers/net/ethernet/intel/i40e/i40e_main.c
4539
q_vector->tx.count++;
drivers/net/ethernet/intel/i40e/i40e_main.c
4546
xdp_ring->next = q_vector->tx.ring;
drivers/net/ethernet/intel/i40e/i40e_main.c
4547
q_vector->tx.ring = xdp_ring;
drivers/net/ethernet/intel/i40e/i40e_main.c
4548
q_vector->tx.count++;
drivers/net/ethernet/intel/i40e/i40e_main.c
4590
q_vector->tx.count = 0;
drivers/net/ethernet/intel/i40e/i40e_main.c
4592
q_vector->tx.ring = NULL;
drivers/net/ethernet/intel/i40e/i40e_main.c
5106
i40e_for_each_ring(ring, q_vector->tx)
drivers/net/ethernet/intel/i40e/i40e_main.c
5196
if (q_vector->rx.ring || q_vector->tx.ring)
drivers/net/ethernet/intel/i40e/i40e_main.c
5215
if (q_vector->rx.ring || q_vector->tx.ring)
drivers/net/ethernet/intel/i40e/i40e_txrx.c
1060
u16 flags = q_vector->tx.ring[0].flags;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
1205
(q_vector->tx.target_itr & I40E_ITR_ADAPTIVE_LATENCY)) {
drivers/net/ethernet/intel/i40e/i40e_txrx.c
1244
itr = min(q_vector->tx.current_itr, q_vector->rx.current_itr);
drivers/net/ethernet/intel/i40e/i40e_txrx.c
2679
i40e_update_itr(q_vector, &q_vector->tx);
drivers/net/ethernet/intel/i40e/i40e_txrx.c
2696
} else if ((q_vector->tx.target_itr < q_vector->tx.current_itr) ||
drivers/net/ethernet/intel/i40e/i40e_txrx.c
2698
(q_vector->tx.target_itr - q_vector->tx.current_itr))) {
drivers/net/ethernet/intel/i40e/i40e_txrx.c
2703
interval = q_vector->tx.target_itr;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
2704
q_vector->tx.current_itr = q_vector->tx.target_itr;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
2766
i40e_for_each_ring(ring, q_vector->tx) {
drivers/net/ethernet/intel/i40e/i40e_txrx.c
2832
q_vector->tx.ring[0].tx_stats.tx_force_wb++;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
2838
if (q_vector->tx.ring[0].flags & I40E_TXR_FLAGS_WB_ON_ITR)
drivers/net/ethernet/intel/i40e/i40e_txrx_common.h
54
tx_ring->q_vector->tx.total_bytes += total_bytes;
drivers/net/ethernet/intel/i40e/i40e_txrx_common.h
55
tx_ring->q_vector->tx.total_packets += total_packets;
drivers/net/ethernet/intel/iavf/iavf.h
109
struct iavf_ring_container tx;
drivers/net/ethernet/intel/iavf/iavf_ethtool.c
667
q_vector->tx.target_itr = ITR_TO_REG(tx_ring->itr_setting);
drivers/net/ethernet/intel/iavf/iavf_main.c
414
if (!q_vector->tx.ring && !q_vector->rx.ring)
drivers/net/ethernet/intel/iavf/iavf_main.c
462
tx_ring->next = q_vector->tx.ring;
drivers/net/ethernet/intel/iavf/iavf_main.c
464
q_vector->tx.ring = tx_ring;
drivers/net/ethernet/intel/iavf/iavf_main.c
465
q_vector->tx.count++;
drivers/net/ethernet/intel/iavf/iavf_main.c
466
q_vector->tx.next_update = jiffies + 1;
drivers/net/ethernet/intel/iavf/iavf_main.c
467
q_vector->tx.target_itr = ITR_TO_REG(tx_ring->itr_setting);
drivers/net/ethernet/intel/iavf/iavf_main.c
470
q_vector->tx.target_itr >> 1);
drivers/net/ethernet/intel/iavf/iavf_main.c
471
q_vector->tx.current_itr = q_vector->tx.target_itr;
drivers/net/ethernet/intel/iavf/iavf_main.c
530
if (q_vector->tx.ring && q_vector->rx.ring) {
drivers/net/ethernet/intel/iavf/iavf_main.c
537
} else if (q_vector->tx.ring) {
drivers/net/ethernet/intel/iavf/iavf_txrx.c
1543
iavf_update_itr(q_vector, &q_vector->tx);
drivers/net/ethernet/intel/iavf/iavf_txrx.c
1560
} else if ((q_vector->tx.target_itr < q_vector->tx.current_itr) ||
drivers/net/ethernet/intel/iavf/iavf_txrx.c
1562
(q_vector->tx.target_itr - q_vector->tx.current_itr))) {
drivers/net/ethernet/intel/iavf/iavf_txrx.c
1567
q_vector->tx.target_itr);
drivers/net/ethernet/intel/iavf/iavf_txrx.c
1568
q_vector->tx.current_itr = q_vector->tx.target_itr;
drivers/net/ethernet/intel/iavf/iavf_txrx.c
1615
iavf_for_each_ring(ring, q_vector->tx) {
drivers/net/ethernet/intel/iavf/iavf_txrx.c
1666
q_vector->tx.ring[0].tx_stats.tx_force_wb++;
drivers/net/ethernet/intel/iavf/iavf_txrx.c
328
tx_ring->q_vector->tx.total_bytes += total_bytes;
drivers/net/ethernet/intel/iavf/iavf_txrx.c
329
tx_ring->q_vector->tx.total_packets += total_packets;
drivers/net/ethernet/intel/iavf/iavf_txrx.c
378
u16 flags = q_vector->tx.ring[0].flags;
drivers/net/ethernet/intel/iavf/iavf_txrx.c
518
(q_vector->tx.target_itr & IAVF_ITR_ADAPTIVE_LATENCY)) {
drivers/net/ethernet/intel/iavf/iavf_txrx.c
557
itr = min(q_vector->tx.current_itr, q_vector->rx.current_itr);
drivers/net/ethernet/intel/ice/ice.h
484
struct ice_ring_container tx;
drivers/net/ethernet/intel/ice/ice_base.c
116
q_vector->tx.itr_setting = ICE_DFLT_TX_ITR;
drivers/net/ethernet/intel/ice/ice_base.c
118
q_vector->tx.itr_mode = ITR_DYNAMIC;
drivers/net/ethernet/intel/ice/ice_base.c
120
q_vector->tx.type = ICE_TX_CONTAINER;
drivers/net/ethernet/intel/ice/ice_base.c
1222
ice_write_itr(&q_vector->tx, q_vector->tx.itr_setting);
drivers/net/ethernet/intel/ice/ice_base.c
193
ice_for_each_tx_ring(tx_ring, vsi->q_vectors[v_idx]->tx)
drivers/net/ethernet/intel/ice/ice_base.c
947
q_vector->tx.tx_ring = NULL;
drivers/net/ethernet/intel/ice/ice_base.c
948
q_vector->tx.itr_idx = ICE_TX_ITR;
drivers/net/ethernet/intel/ice/ice_base.c
955
tx_ring->next = q_vector->tx.tx_ring;
drivers/net/ethernet/intel/ice/ice_base.c
956
q_vector->tx.tx_ring = tx_ring;
drivers/net/ethernet/intel/ice/ice_ethtool.c
3293
tx_rings[i].tx_tstamps = &pf->ptp.port.tx;
drivers/net/ethernet/intel/ice/ice_ethtool.c
4082
&vsi->tx_rings[q_num]->q_vector->tx))
drivers/net/ethernet/intel/ice/ice_ethtool.c
4090
&vsi->tx_rings[q_num]->q_vector->tx))
drivers/net/ethernet/intel/ice/ice_ethtool.c
4252
&vsi->tx_rings[q_num]->q_vector->tx,
drivers/net/ethernet/intel/ice/ice_ethtool.c
4262
&vsi->tx_rings[q_num]->q_vector->tx,
drivers/net/ethernet/intel/ice/ice_ethtool.c
4319
ice_print_if_odd_usecs(netdev, q_vector->tx.itr_setting,
drivers/net/ethernet/intel/ice/ice_ethtool.c
697
static const int tx = ICE_AQC_OP_CODE_TX_EQU;
drivers/net/ethernet/intel/ice/ice_ethtool.c
704
{ ICE_AQC_TX_EQU_PRE1, tx, &ptr->tx_equ_pre1 },
drivers/net/ethernet/intel/ice/ice_ethtool.c
705
{ ICE_AQC_TX_EQU_PRE3, tx, &ptr->tx_equ_pre3 },
drivers/net/ethernet/intel/ice/ice_ethtool.c
706
{ ICE_AQC_TX_EQU_ATTEN, tx, &ptr->tx_equ_atten },
drivers/net/ethernet/intel/ice/ice_ethtool.c
707
{ ICE_AQC_TX_EQU_POST1, tx, &ptr->tx_equ_post1 },
drivers/net/ethernet/intel/ice/ice_ethtool.c
708
{ ICE_AQC_TX_EQU_PRE2, tx, &ptr->tx_equ_pre2 },
drivers/net/ethernet/intel/ice/ice_lib.c
1361
q_vector->tx.tx_ring = NULL;
drivers/net/ethernet/intel/ice/ice_lib.c
1410
ring->tx_tstamps = &pf->ptp.port.tx;
drivers/net/ethernet/intel/ice/ice_lib.c
1895
if (ITR_IS_DYNAMIC(&q_vector->tx) || ITR_IS_DYNAMIC(&q_vector->rx)) {
drivers/net/ethernet/intel/ice/ice_lib.c
1941
q_vector->tx.itr_idx);
drivers/net/ethernet/intel/ice/ice_lib.c
2078
void ice_vsi_cfg_sw_lldp(struct ice_vsi *vsi, bool tx, bool create)
drivers/net/ethernet/intel/ice/ice_lib.c
2089
if (tx) {
drivers/net/ethernet/intel/ice/ice_lib.c
2113
create ? "add" : "remove", tx ? "Tx" : "Rx",
drivers/net/ethernet/intel/ice/ice_lib.c
2614
ice_write_itr(&q_vector->tx, 0);
drivers/net/ethernet/intel/ice/ice_lib.c
2916
coalesce[i].itr_tx = q_vector->tx.itr_settings;
drivers/net/ethernet/intel/ice/ice_lib.c
2982
rc = &vsi->q_vectors[i]->tx;
drivers/net/ethernet/intel/ice/ice_lib.c
2986
rc = &vsi->q_vectors[i]->tx;
drivers/net/ethernet/intel/ice/ice_lib.c
3000
rc = &vsi->q_vectors[i]->tx;
drivers/net/ethernet/intel/ice/ice_lib.c
488
if (!q_vector->tx.tx_ring)
drivers/net/ethernet/intel/ice/ice_lib.c
492
ice_clean_ctrl_tx_irq(q_vector->tx.tx_ring);
drivers/net/ethernet/intel/ice/ice_lib.c
506
if (!q_vector->tx.tx_ring && !q_vector->rx.rx_ring)
drivers/net/ethernet/intel/ice/ice_lib.h
39
void ice_vsi_cfg_sw_lldp(struct ice_vsi *vsi, bool tx, bool create);
drivers/net/ethernet/intel/ice/ice_main.c
163
if (ring_stats->tx.prev_pkt == packets) {
drivers/net/ethernet/intel/ice/ice_main.c
173
ring_stats->tx.prev_pkt =
drivers/net/ethernet/intel/ice/ice_main.c
2549
if (q_vector->tx.tx_ring && q_vector->rx.rx_ring) {
drivers/net/ethernet/intel/ice/ice_main.c
2556
} else if (q_vector->tx.tx_ring) {
drivers/net/ethernet/intel/ice/ice_main.c
2680
ice_for_each_tx_ring(ring, q_vector->tx)
drivers/net/ethernet/intel/ice/ice_main.c
2712
xdp_ring->next = q_vector->tx.tx_ring;
drivers/net/ethernet/intel/ice/ice_main.c
2713
q_vector->tx.tx_ring = xdp_ring;
drivers/net/ethernet/intel/ice/ice_main.c
2737
ice_for_each_tx_ring(ring, q_vector->tx)
drivers/net/ethernet/intel/ice/ice_main.c
2742
q_vector->tx.tx_ring = ring;
drivers/net/ethernet/intel/ice/ice_main.c
3137
struct ice_ptp_tx *tx;
drivers/net/ethernet/intel/ice/ice_main.c
3144
tx = &pf->ptp.port.tx;
drivers/net/ethernet/intel/ice/ice_main.c
3145
spin_lock_irqsave(&tx->lock, flags);
drivers/net/ethernet/intel/ice/ice_main.c
3146
if (tx->init) {
drivers/net/ethernet/intel/ice/ice_main.c
3147
ice_ptp_complete_tx_single_tstamp(tx);
drivers/net/ethernet/intel/ice/ice_main.c
3149
idx = find_next_bit_wrap(tx->in_use, tx->len,
drivers/net/ethernet/intel/ice/ice_main.c
3150
tx->last_ll_ts_idx_read + 1);
drivers/net/ethernet/intel/ice/ice_main.c
3151
if (idx != tx->len)
drivers/net/ethernet/intel/ice/ice_main.c
3152
ice_ptp_req_tx_single_tstamp(tx, idx);
drivers/net/ethernet/intel/ice/ice_main.c
3154
spin_unlock_irqrestore(&tx->lock, flags);
drivers/net/ethernet/intel/ice/ice_main.c
6682
ice_trace(tx_dim_work, container_of(rc, struct ice_q_vector, tx), dim);
drivers/net/ethernet/intel/ice/ice_main.c
6725
rc = &q_vector->tx;
drivers/net/ethernet/intel/ice/ice_main.c
6766
if (q_vector->rx.rx_ring || q_vector->tx.tx_ring)
drivers/net/ethernet/intel/ice/ice_main.c
7265
if (q_vector->rx.rx_ring || q_vector->tx.tx_ring)
drivers/net/ethernet/intel/ice/ice_main.c
7268
cancel_work_sync(&q_vector->tx.dim.work);
drivers/net/ethernet/intel/ice/ice_main.c
8709
rc = &tx_q_vector->tx;
drivers/net/ethernet/intel/ice/ice_ptp.c
1007
ice_ptp_mark_tx_tracker_stale(&pf->ptp.port.tx);
drivers/net/ethernet/intel/ice/ice_ptp.c
1243
spin_lock_irqsave(&ptp_port->tx.lock, flags);
drivers/net/ethernet/intel/ice/ice_ptp.c
1244
ptp_port->tx.calibrating = true;
drivers/net/ethernet/intel/ice/ice_ptp.c
1245
spin_unlock_irqrestore(&ptp_port->tx.lock, flags);
drivers/net/ethernet/intel/ice/ice_ptp.c
1254
spin_lock_irqsave(&ptp_port->tx.lock, flags);
drivers/net/ethernet/intel/ice/ice_ptp.c
1255
ptp_port->tx.calibrating = false;
drivers/net/ethernet/intel/ice/ice_ptp.c
1256
spin_unlock_irqrestore(&ptp_port->tx.lock, flags);
drivers/net/ethernet/intel/ice/ice_ptp.c
2613
s8 ice_ptp_request_ts(struct ice_ptp_tx *tx, struct sk_buff *skb)
drivers/net/ethernet/intel/ice/ice_ptp.c
2618
spin_lock_irqsave(&tx->lock, flags);
drivers/net/ethernet/intel/ice/ice_ptp.c
2621
if (!ice_ptp_is_tx_tracker_up(tx)) {
drivers/net/ethernet/intel/ice/ice_ptp.c
2622
spin_unlock_irqrestore(&tx->lock, flags);
drivers/net/ethernet/intel/ice/ice_ptp.c
2627
idx = find_next_zero_bit(tx->in_use, tx->len,
drivers/net/ethernet/intel/ice/ice_ptp.c
2628
tx->last_ll_ts_idx_read + 1);
drivers/net/ethernet/intel/ice/ice_ptp.c
2629
if (idx == tx->len)
drivers/net/ethernet/intel/ice/ice_ptp.c
2630
idx = find_first_zero_bit(tx->in_use, tx->len);
drivers/net/ethernet/intel/ice/ice_ptp.c
2632
if (idx < tx->len) {
drivers/net/ethernet/intel/ice/ice_ptp.c
2637
set_bit(idx, tx->in_use);
drivers/net/ethernet/intel/ice/ice_ptp.c
2638
clear_bit(idx, tx->stale);
drivers/net/ethernet/intel/ice/ice_ptp.c
2639
tx->tstamps[idx].start = jiffies;
drivers/net/ethernet/intel/ice/ice_ptp.c
2640
tx->tstamps[idx].skb = skb_get(skb);
drivers/net/ethernet/intel/ice/ice_ptp.c
2645
spin_unlock_irqrestore(&tx->lock, flags);
drivers/net/ethernet/intel/ice/ice_ptp.c
2650
if (idx >= tx->len)
drivers/net/ethernet/intel/ice/ice_ptp.c
2653
return idx + tx->offset;
drivers/net/ethernet/intel/ice/ice_ptp.c
2664
ice_ptp_process_tx_tstamp(&pf->ptp.port.tx);
drivers/net/ethernet/intel/ice/ice_ptp.c
2677
static bool ice_port_has_timestamps(struct ice_ptp_tx *tx)
drivers/net/ethernet/intel/ice/ice_ptp.c
2681
scoped_guard(spinlock_irqsave, &tx->lock) {
drivers/net/ethernet/intel/ice/ice_ptp.c
2682
if (!tx->init)
drivers/net/ethernet/intel/ice/ice_ptp.c
2685
more_timestamps = !bitmap_empty(tx->in_use, tx->len);
drivers/net/ethernet/intel/ice/ice_ptp.c
2698
struct ice_ptp_tx *tx = &port->tx;
drivers/net/ethernet/intel/ice/ice_ptp.c
2700
if (ice_port_has_timestamps(tx))
drivers/net/ethernet/intel/ice/ice_ptp.c
2718
if (ice_port_has_timestamps(&pf->ptp.port.tx))
drivers/net/ethernet/intel/ice/ice_ptp.c
2762
struct ice_ptp_tx *tx = &pf->ptp.port.tx;
drivers/net/ethernet/intel/ice/ice_ptp.c
2768
spin_lock(&tx->lock);
drivers/net/ethernet/intel/ice/ice_ptp.c
2769
if (tx->init) {
drivers/net/ethernet/intel/ice/ice_ptp.c
2770
last = tx->last_ll_ts_idx_read + 1;
drivers/net/ethernet/intel/ice/ice_ptp.c
2771
idx = find_next_bit_wrap(tx->in_use, tx->len,
drivers/net/ethernet/intel/ice/ice_ptp.c
2773
if (idx != tx->len)
drivers/net/ethernet/intel/ice/ice_ptp.c
2774
ice_ptp_req_tx_single_tstamp(tx, idx);
drivers/net/ethernet/intel/ice/ice_ptp.c
2776
spin_unlock(&tx->lock);
drivers/net/ethernet/intel/ice/ice_ptp.c
2828
if (!pf->ptp.port.tx.has_ready_bitmap)
drivers/net/ethernet/intel/ice/ice_ptp.c
2946
ice_ptp_release_tx_tracker(pf, &pf->ptp.port.tx);
drivers/net/ethernet/intel/ice/ice_ptp.c
3240
return ice_ptp_init_tx(pf, &ptp_port->tx, ptp_port->port_num);
drivers/net/ethernet/intel/ice/ice_ptp.c
3244
return ice_ptp_init_tx_e82x(pf, &ptp_port->tx,
drivers/net/ethernet/intel/ice/ice_ptp.c
3389
ice_ptp_release_tx_tracker(pf, &pf->ptp.port.tx);
drivers/net/ethernet/intel/ice/ice_ptp.c
367
ice_ptp_is_tx_tracker_up(struct ice_ptp_tx *tx)
drivers/net/ethernet/intel/ice/ice_ptp.c
369
lockdep_assert_held(&tx->lock);
drivers/net/ethernet/intel/ice/ice_ptp.c
371
return tx->init && !tx->calibrating;
drivers/net/ethernet/intel/ice/ice_ptp.c
379
void ice_ptp_req_tx_single_tstamp(struct ice_ptp_tx *tx, u8 idx)
drivers/net/ethernet/intel/ice/ice_ptp.c
387
if (!tx->init)
drivers/net/ethernet/intel/ice/ice_ptp.c
390
ptp_port = container_of(tx, struct ice_ptp_port, tx);
drivers/net/ethernet/intel/ice/ice_ptp.c
395
if (time_is_before_jiffies(tx->tstamps[idx].start + 2 * HZ)) {
drivers/net/ethernet/intel/ice/ice_ptp.c
399
skb = tx->tstamps[idx].skb;
drivers/net/ethernet/intel/ice/ice_ptp.c
400
tx->tstamps[idx].skb = NULL;
drivers/net/ethernet/intel/ice/ice_ptp.c
401
clear_bit(idx, tx->in_use);
drivers/net/ethernet/intel/ice/ice_ptp.c
407
ice_trace(tx_tstamp_fw_req, tx->tstamps[idx].skb, idx);
drivers/net/ethernet/intel/ice/ice_ptp.c
417
tx->last_ll_ts_idx_read = idx;
drivers/net/ethernet/intel/ice/ice_ptp.c
426
void ice_ptp_complete_tx_single_tstamp(struct ice_ptp_tx *tx)
drivers/net/ethernet/intel/ice/ice_ptp.c
429
u8 idx = tx->last_ll_ts_idx_read;
drivers/net/ethernet/intel/ice/ice_ptp.c
440
if (!tx->init || tx->last_ll_ts_idx_read < 0)
drivers/net/ethernet/intel/ice/ice_ptp.c
443
ptp_port = container_of(tx, struct ice_ptp_port, tx);
drivers/net/ethernet/intel/ice/ice_ptp.c
448
ice_trace(tx_tstamp_fw_done, tx->tstamps[idx].skb, idx);
drivers/net/ethernet/intel/ice/ice_ptp.c
480
if (raw_tstamp == tx->tstamps[idx].cached_tstamp)
drivers/net/ethernet/intel/ice/ice_ptp.c
483
tx->tstamps[idx].cached_tstamp = raw_tstamp;
drivers/net/ethernet/intel/ice/ice_ptp.c
484
clear_bit(idx, tx->in_use);
drivers/net/ethernet/intel/ice/ice_ptp.c
485
skb = tx->tstamps[idx].skb;
drivers/net/ethernet/intel/ice/ice_ptp.c
486
tx->tstamps[idx].skb = NULL;
drivers/net/ethernet/intel/ice/ice_ptp.c
487
if (test_and_clear_bit(idx, tx->stale))
drivers/net/ethernet/intel/ice/ice_ptp.c
560
static void ice_ptp_process_tx_tstamp(struct ice_ptp_tx *tx)
drivers/net/ethernet/intel/ice/ice_ptp.c
572
ptp_port = container_of(tx, struct ice_ptp_port, tx);
drivers/net/ethernet/intel/ice/ice_ptp.c
576
if (!tx->init)
drivers/net/ethernet/intel/ice/ice_ptp.c
580
if (tx->has_ready_bitmap) {
drivers/net/ethernet/intel/ice/ice_ptp.c
581
err = ice_get_phy_tx_tstamp_ready(hw, tx->block, &tstamp_ready);
drivers/net/ethernet/intel/ice/ice_ptp.c
589
for_each_set_bit(idx, tx->in_use, tx->len) {
drivers/net/ethernet/intel/ice/ice_ptp.c
591
u8 phy_idx = idx + tx->offset;
drivers/net/ethernet/intel/ice/ice_ptp.c
597
if (time_is_before_jiffies(tx->tstamps[idx].start + 2 * HZ)) {
drivers/net/ethernet/intel/ice/ice_ptp.c
612
if (tx->has_ready_bitmap &&
drivers/net/ethernet/intel/ice/ice_ptp.c
620
ice_trace(tx_tstamp_fw_req, tx->tstamps[idx].skb, idx);
drivers/net/ethernet/intel/ice/ice_ptp.c
622
err = ice_read_phy_tstamp(hw, tx->block, phy_idx, &raw_tstamp);
drivers/net/ethernet/intel/ice/ice_ptp.c
626
ice_trace(tx_tstamp_fw_done, tx->tstamps[idx].skb, idx);
drivers/net/ethernet/intel/ice/ice_ptp.c
633
if (!drop_ts && !tx->has_ready_bitmap &&
drivers/net/ethernet/intel/ice/ice_ptp.c
634
raw_tstamp == tx->tstamps[idx].cached_tstamp)
drivers/net/ethernet/intel/ice/ice_ptp.c
642
spin_lock_irqsave(&tx->lock, flags);
drivers/net/ethernet/intel/ice/ice_ptp.c
643
if (!tx->has_ready_bitmap && raw_tstamp)
drivers/net/ethernet/intel/ice/ice_ptp.c
644
tx->tstamps[idx].cached_tstamp = raw_tstamp;
drivers/net/ethernet/intel/ice/ice_ptp.c
645
clear_bit(idx, tx->in_use);
drivers/net/ethernet/intel/ice/ice_ptp.c
646
skb = tx->tstamps[idx].skb;
drivers/net/ethernet/intel/ice/ice_ptp.c
647
tx->tstamps[idx].skb = NULL;
drivers/net/ethernet/intel/ice/ice_ptp.c
648
if (test_and_clear_bit(idx, tx->stale))
drivers/net/ethernet/intel/ice/ice_ptp.c
650
spin_unlock_irqrestore(&tx->lock, flags);
drivers/net/ethernet/intel/ice/ice_ptp.c
686
struct ice_ptp_tx *tx = &port->tx;
drivers/net/ethernet/intel/ice/ice_ptp.c
688
if (!tx || !tx->init)
drivers/net/ethernet/intel/ice/ice_ptp.c
691
ice_ptp_process_tx_tstamp(tx);
drivers/net/ethernet/intel/ice/ice_ptp.c
704
ice_ptp_alloc_tx_tracker(struct ice_ptp_tx *tx)
drivers/net/ethernet/intel/ice/ice_ptp.c
709
tstamps = kzalloc_objs(*tstamps, tx->len);
drivers/net/ethernet/intel/ice/ice_ptp.c
710
in_use = bitmap_zalloc(tx->len, GFP_KERNEL);
drivers/net/ethernet/intel/ice/ice_ptp.c
711
stale = bitmap_zalloc(tx->len, GFP_KERNEL);
drivers/net/ethernet/intel/ice/ice_ptp.c
721
tx->tstamps = tstamps;
drivers/net/ethernet/intel/ice/ice_ptp.c
722
tx->in_use = in_use;
drivers/net/ethernet/intel/ice/ice_ptp.c
723
tx->stale = stale;
drivers/net/ethernet/intel/ice/ice_ptp.c
724
tx->init = 1;
drivers/net/ethernet/intel/ice/ice_ptp.c
725
tx->last_ll_ts_idx_read = -1;
drivers/net/ethernet/intel/ice/ice_ptp.c
727
spin_lock_init(&tx->lock);
drivers/net/ethernet/intel/ice/ice_ptp.c
740
ice_ptp_flush_tx_tracker(struct ice_pf *pf, struct ice_ptp_tx *tx)
drivers/net/ethernet/intel/ice/ice_ptp.c
748
err = ice_get_phy_tx_tstamp_ready(hw, tx->block, &tstamp_ready);
drivers/net/ethernet/intel/ice/ice_ptp.c
751
tx->block, err);
drivers/net/ethernet/intel/ice/ice_ptp.c
759
for_each_set_bit(idx, tx->in_use, tx->len) {
drivers/net/ethernet/intel/ice/ice_ptp.c
760
u8 phy_idx = idx + tx->offset;
drivers/net/ethernet/intel/ice/ice_ptp.c
765
ice_clear_phy_tstamp(hw, tx->block, phy_idx);
drivers/net/ethernet/intel/ice/ice_ptp.c
767
spin_lock_irqsave(&tx->lock, flags);
drivers/net/ethernet/intel/ice/ice_ptp.c
768
skb = tx->tstamps[idx].skb;
drivers/net/ethernet/intel/ice/ice_ptp.c
769
tx->tstamps[idx].skb = NULL;
drivers/net/ethernet/intel/ice/ice_ptp.c
770
clear_bit(idx, tx->in_use);
drivers/net/ethernet/intel/ice/ice_ptp.c
771
clear_bit(idx, tx->stale);
drivers/net/ethernet/intel/ice/ice_ptp.c
772
spin_unlock_irqrestore(&tx->lock, flags);
drivers/net/ethernet/intel/ice/ice_ptp.c
794
ice_ptp_mark_tx_tracker_stale(struct ice_ptp_tx *tx)
drivers/net/ethernet/intel/ice/ice_ptp.c
798
spin_lock_irqsave(&tx->lock, flags);
drivers/net/ethernet/intel/ice/ice_ptp.c
799
bitmap_or(tx->stale, tx->stale, tx->in_use, tx->len);
drivers/net/ethernet/intel/ice/ice_ptp.c
800
spin_unlock_irqrestore(&tx->lock, flags);
drivers/net/ethernet/intel/ice/ice_ptp.c
816
ice_ptp_flush_tx_tracker(ptp_port_to_pf(port), &port->tx);
drivers/net/ethernet/intel/ice/ice_ptp.c
827
ice_ptp_release_tx_tracker(struct ice_pf *pf, struct ice_ptp_tx *tx)
drivers/net/ethernet/intel/ice/ice_ptp.c
831
spin_lock_irqsave(&tx->lock, flags);
drivers/net/ethernet/intel/ice/ice_ptp.c
832
tx->init = 0;
drivers/net/ethernet/intel/ice/ice_ptp.c
833
spin_unlock_irqrestore(&tx->lock, flags);
drivers/net/ethernet/intel/ice/ice_ptp.c
838
ice_ptp_flush_tx_tracker(pf, tx);
drivers/net/ethernet/intel/ice/ice_ptp.c
840
kfree(tx->tstamps);
drivers/net/ethernet/intel/ice/ice_ptp.c
841
tx->tstamps = NULL;
drivers/net/ethernet/intel/ice/ice_ptp.c
843
bitmap_free(tx->in_use);
drivers/net/ethernet/intel/ice/ice_ptp.c
844
tx->in_use = NULL;
drivers/net/ethernet/intel/ice/ice_ptp.c
846
bitmap_free(tx->stale);
drivers/net/ethernet/intel/ice/ice_ptp.c
847
tx->stale = NULL;
drivers/net/ethernet/intel/ice/ice_ptp.c
849
tx->len = 0;
drivers/net/ethernet/intel/ice/ice_ptp.c
865
static int ice_ptp_init_tx_e82x(struct ice_pf *pf, struct ice_ptp_tx *tx,
drivers/net/ethernet/intel/ice/ice_ptp.c
868
tx->block = ICE_GET_QUAD_NUM(port);
drivers/net/ethernet/intel/ice/ice_ptp.c
869
tx->offset = (port % ICE_PORTS_PER_QUAD) * INDEX_PER_PORT_E82X;
drivers/net/ethernet/intel/ice/ice_ptp.c
870
tx->len = INDEX_PER_PORT_E82X;
drivers/net/ethernet/intel/ice/ice_ptp.c
871
tx->has_ready_bitmap = 1;
drivers/net/ethernet/intel/ice/ice_ptp.c
873
return ice_ptp_alloc_tx_tracker(tx);
drivers/net/ethernet/intel/ice/ice_ptp.c
887
static int ice_ptp_init_tx(struct ice_pf *pf, struct ice_ptp_tx *tx, u8 port)
drivers/net/ethernet/intel/ice/ice_ptp.c
889
tx->block = port;
drivers/net/ethernet/intel/ice/ice_ptp.c
890
tx->offset = 0;
drivers/net/ethernet/intel/ice/ice_ptp.c
891
tx->len = INDEX_PER_PORT;
drivers/net/ethernet/intel/ice/ice_ptp.c
897
tx->has_ready_bitmap = pf->hw.mac_type != ICE_MAC_E810;
drivers/net/ethernet/intel/ice/ice_ptp.c
899
return ice_ptp_alloc_tx_tracker(tx);
drivers/net/ethernet/intel/ice/ice_ptp.h
150
struct ice_ptp_tx tx;
drivers/net/ethernet/intel/ice/ice_ptp.h
304
s8 ice_ptp_request_ts(struct ice_ptp_tx *tx, struct sk_buff *skb);
drivers/net/ethernet/intel/ice/ice_ptp.h
305
void ice_ptp_req_tx_single_tstamp(struct ice_ptp_tx *tx, u8 idx);
drivers/net/ethernet/intel/ice/ice_ptp.h
306
void ice_ptp_complete_tx_single_tstamp(struct ice_ptp_tx *tx);
drivers/net/ethernet/intel/ice/ice_ptp.h
340
ice_ptp_request_ts(struct ice_ptp_tx *tx, struct sk_buff *skb)
drivers/net/ethernet/intel/ice/ice_ptp.h
345
static inline void ice_ptp_req_tx_single_tstamp(struct ice_ptp_tx *tx, u8 idx)
drivers/net/ethernet/intel/ice/ice_ptp.h
348
static inline void ice_ptp_complete_tx_single_tstamp(struct ice_ptp_tx *tx) { }
drivers/net/ethernet/intel/ice/ice_trace.h
103
__entry->q_vector->tx.tx_ring->q_index,
drivers/net/ethernet/intel/ice/ice_trace.h
95
__string(devname, q_vector->tx.tx_ring->netdev->name)),
drivers/net/ethernet/intel/ice/ice_txrx.c
1143
struct ice_ring_container *tx = &q_vector->tx;
drivers/net/ethernet/intel/ice/ice_txrx.c
1146
if (ITR_IS_DYNAMIC(tx)) {
drivers/net/ethernet/intel/ice/ice_txrx.c
1149
__ice_update_sample(q_vector, tx, &dim_sample, true);
drivers/net/ethernet/intel/ice/ice_txrx.c
1150
net_dim(&tx->dim, &dim_sample);
drivers/net/ethernet/intel/ice/ice_txrx.c
1279
ice_for_each_tx_ring(tx_ring, q_vector->tx) {
drivers/net/ethernet/intel/ice/ice_txrx.c
502
tx_ring->ring_stats->tx.prev_pkt = -1;
drivers/net/ethernet/intel/ice/ice_txrx.h
139
struct_group(tx,
drivers/net/ethernet/intel/ice/ice_xsk.c
88
ice_cfg_txq_interrupt(vsi, _qid, reg_idx, q_vector->tx.itr_idx);
drivers/net/ethernet/intel/ice/virt/queues.c
496
q_vector->tx.itr_idx = map->txitr_idx;
drivers/net/ethernet/intel/ice/virt/queues.c
500
q_vector->tx.itr_idx);
drivers/net/ethernet/intel/idpf/idpf.h
1061
u16 itr, bool tx);
drivers/net/ethernet/intel/idpf/idpf_singleq_txrx.c
601
q = q_vec->tx[i];
drivers/net/ethernet/intel/idpf/idpf_txrx.c
1147
qs->qs[num++].txq = qv->tx[i];
drivers/net/ethernet/intel/idpf/idpf_txrx.c
3831
kfree(q_vector->tx);
drivers/net/ethernet/intel/idpf/idpf_txrx.c
3832
q_vector->tx = NULL;
drivers/net/ethernet/intel/idpf/idpf_txrx.c
3851
netif_queue_set_napi(dev, q_vector->tx[i]->idx,
drivers/net/ethernet/intel/idpf/idpf_txrx.c
3974
struct idpf_tx_queue *txq = q_vector->tx[i];
drivers/net/ethernet/intel/idpf/idpf_txrx.c
4099
void idpf_vport_intr_write_itr(struct idpf_q_vector *q_vector, u16 itr, bool tx)
drivers/net/ethernet/intel/idpf/idpf_txrx.c
4103
if (tx && !q_vector->tx)
drivers/net/ethernet/intel/idpf/idpf_txrx.c
4105
else if (!tx && !q_vector->rx)
drivers/net/ethernet/intel/idpf/idpf_txrx.c
4110
tx ? intr_reg->tx_itr : intr_reg->rx_itr);
drivers/net/ethernet/intel/idpf/idpf_txrx.c
4441
q->q_vector->tx[q->q_vector->num_txq++] = q;
drivers/net/ethernet/intel/idpf/idpf_txrx.c
4592
q_vector->tx = kzalloc_objs(*q_vector->tx, txqs_per_vector);
drivers/net/ethernet/intel/idpf/idpf_txrx.c
4593
if (!q_vector->tx)
drivers/net/ethernet/intel/idpf/idpf_txrx.h
404
struct idpf_tx_queue **tx;
drivers/net/ethernet/intel/igb/igb.h
377
struct igb_ring_container rx, tx;
drivers/net/ethernet/intel/igb/igb_ethtool.c
2231
q_vector->tx.work_limit = adapter->tx_work_limit;
drivers/net/ethernet/intel/igb/igb_main.c
1009
if (q_vector->tx.ring)
drivers/net/ethernet/intel/igb/igb_main.c
1010
adapter->tx_ring[q_vector->tx.ring->queue_index] = NULL;
drivers/net/ethernet/intel/igb/igb_main.c
1208
q_vector->tx.work_limit = adapter->tx_work_limit;
drivers/net/ethernet/intel/igb/igb_main.c
1237
igb_add_ring(ring, &q_vector->tx);
drivers/net/ethernet/intel/igb/igb_main.c
2112
if (q_vector->tx.ring)
drivers/net/ethernet/intel/igb/igb_main.c
2114
q_vector->tx.ring->queue_index,
drivers/net/ethernet/intel/igb/igb_main.c
5844
packets = q_vector->tx.total_packets;
drivers/net/ethernet/intel/igb/igb_main.c
5847
q_vector->tx.total_bytes / packets);
drivers/net/ethernet/intel/igb/igb_main.c
5879
q_vector->tx.total_bytes = 0;
drivers/net/ethernet/intel/igb/igb_main.c
5880
q_vector->tx.total_packets = 0;
drivers/net/ethernet/intel/igb/igb_main.c
5964
igb_update_itr(q_vector, &q_vector->tx);
drivers/net/ethernet/intel/igb/igb_main.c
5967
current_itr = max(q_vector->rx.itr, q_vector->tx.itr);
drivers/net/ethernet/intel/igb/igb_main.c
7216
if (q_vector->tx.ring)
drivers/net/ethernet/intel/igb/igb_main.c
7217
igb_update_tx_dca(adapter, q_vector->tx.ring, cpu);
drivers/net/ethernet/intel/igb/igb_main.c
768
if (q_vector->tx.ring)
drivers/net/ethernet/intel/igb/igb_main.c
769
tx_queue = q_vector->tx.ring->reg_idx;
drivers/net/ethernet/intel/igb/igb_main.c
8292
if (q_vector->tx.ring)
drivers/net/ethernet/intel/igb/igb_main.c
8332
unsigned int budget = q_vector->tx.work_limit;
drivers/net/ethernet/intel/igb/igb_main.c
8333
struct igb_ring *tx_ring = q_vector->tx.ring;
drivers/net/ethernet/intel/igb/igb_main.c
8437
q_vector->tx.total_bytes += total_bytes;
drivers/net/ethernet/intel/igb/igb_main.c
8438
q_vector->tx.total_packets += total_packets;
drivers/net/ethernet/intel/igb/igb_main.c
933
if (q_vector->rx.ring && q_vector->tx.ring)
drivers/net/ethernet/intel/igb/igb_main.c
936
else if (q_vector->tx.ring)
drivers/net/ethernet/intel/igb/igb_main.c
938
q_vector->tx.ring->queue_index);
drivers/net/ethernet/intel/igc/igc.h
617
struct igc_ring_container rx, tx;
drivers/net/ethernet/intel/igc/igc_ethtool.c
948
q_vector->tx.work_limit = adapter->tx_work_limit;
drivers/net/ethernet/intel/igc/igc_main.c
2904
struct igc_ring *ring = q_vector->tx.ring;
drivers/net/ethernet/intel/igc/igc_main.c
2911
q_vector->tx.total_bytes += bytes;
drivers/net/ethernet/intel/igc/igc_main.c
2912
q_vector->tx.total_packets += packets;
drivers/net/ethernet/intel/igc/igc_main.c
3138
unsigned int budget = q_vector->tx.work_limit;
drivers/net/ethernet/intel/igc/igc_main.c
3139
struct igc_ring *tx_ring = q_vector->tx.ring;
drivers/net/ethernet/intel/igc/igc_main.c
4233
if (q_vector->tx.ring)
drivers/net/ethernet/intel/igc/igc_main.c
4234
tx_queue = q_vector->tx.ring->reg_idx;
drivers/net/ethernet/intel/igc/igc_main.c
4406
if (q_vector->tx.ring)
drivers/net/ethernet/intel/igc/igc_main.c
4407
adapter->tx_ring[q_vector->tx.ring->queue_index] = NULL;
drivers/net/ethernet/intel/igc/igc_main.c
4541
igc_update_itr(q_vector, &q_vector->tx);
drivers/net/ethernet/intel/igc/igc_main.c
4544
current_itr = max(q_vector->rx.itr, q_vector->tx.itr);
drivers/net/ethernet/intel/igc/igc_main.c
4710
packets = q_vector->tx.total_packets;
drivers/net/ethernet/intel/igc/igc_main.c
4713
q_vector->tx.total_bytes / packets);
drivers/net/ethernet/intel/igc/igc_main.c
4745
q_vector->tx.total_bytes = 0;
drivers/net/ethernet/intel/igc/igc_main.c
4746
q_vector->tx.total_packets = 0;
drivers/net/ethernet/intel/igc/igc_main.c
4813
if (q_vector->tx.ring)
drivers/net/ethernet/intel/igc/igc_main.c
4883
q_vector->tx.work_limit = adapter->tx_work_limit;
drivers/net/ethernet/intel/igc/igc_main.c
4912
igc_add_ring(ring, &q_vector->tx);
drivers/net/ethernet/intel/igc/igc_main.c
5102
if (q_vector->tx.ring)
drivers/net/ethernet/intel/igc/igc_main.c
5104
q_vector->tx.ring->queue_index,
drivers/net/ethernet/intel/igc/igc_main.c
5727
if (q_vector->rx.ring && q_vector->tx.ring)
drivers/net/ethernet/intel/igc/igc_main.c
5730
else if (q_vector->tx.ring)
drivers/net/ethernet/intel/igc/igc_main.c
5732
q_vector->tx.ring->queue_index);
drivers/net/ethernet/intel/ixgbe/ixgbe.h
511
struct ixgbe_ring_container rx, tx;
drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c
28
int tx = DCB_TX_CONFIG;
drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c
46
if (dst->path[tx].prio_type != src->path[tx].prio_type) {
drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c
47
dst->path[tx].prio_type = src->path[tx].prio_type;
drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c
51
if (dst->path[tx].bwg_id != src->path[tx].bwg_id) {
drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c
52
dst->path[tx].bwg_id = src->path[tx].bwg_id;
drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c
56
if (dst->path[tx].bwg_percent != src->path[tx].bwg_percent) {
drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c
57
dst->path[tx].bwg_percent = src->path[tx].bwg_percent;
drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c
61
if (dst->path[tx].up_to_tc_bitmap !=
drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c
62
src->path[tx].up_to_tc_bitmap) {
drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c
63
dst->path[tx].up_to_tc_bitmap =
drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c
64
src->path[tx].up_to_tc_bitmap;
drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c
93
if (dcfg->bw_percentage[tx][j] != scfg->bw_percentage[tx][j]) {
drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c
94
dcfg->bw_percentage[tx][j] = scfg->bw_percentage[tx][j];
drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c
2538
if (adapter->q_vector[0]->tx.count && adapter->q_vector[0]->rx.count)
drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c
2591
if (adapter->q_vector[0]->tx.count && adapter->q_vector[0]->rx.count) {
drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c
2625
if (adapter->q_vector[0]->tx.count && adapter->q_vector[0]->rx.count)
drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c
2645
if (q_vector->tx.count && !q_vector->rx.count)
drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c
102
*tx = tc << 2; /* 0, 4, 8, 12, 16, 20, 24, 28 */
drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c
1023
ixgbe_for_each_ring(ring, q_vector->tx) {
drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c
119
*tx = tc << 5; /* 0, 32, 64 */
drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c
121
*tx = (tc + 2) << 4; /* 80, 96 */
drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c
123
*tx = (tc + 8) << 3; /* 104, 112, 120 */
drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c
132
*tx = tc << 6; /* 0, 64 */
drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c
134
*tx = (tc + 4) << 4; /* 96, 112 */
drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c
883
q_vector->tx.work_limit = adapter->tx_work_limit;
drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c
886
q_vector->tx.itr = IXGBE_ITR_ADAPTIVE_MAX_USECS |
drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c
91
unsigned int *tx, unsigned int *rx)
drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c
918
ixgbe_add_ring(ring, &q_vector->tx);
drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c
944
ixgbe_add_ring(ring, &q_vector->tx);
drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c
96
*tx = 0;
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
1178
q_vector->tx.total_bytes += bytes;
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
1179
q_vector->tx.total_packets += pkts;
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
1356
unsigned int budget = q_vector->tx.work_limit;
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
1560
ixgbe_for_each_ring(ring, q_vector->tx)
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
2678
ixgbe_for_each_ring(ring, q_vector->tx)
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
2944
ixgbe_update_itr(q_vector, &q_vector->tx);
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
2948
new_itr = min(q_vector->rx.itr, q_vector->tx.itr);
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
3564
if (q_vector->rx.ring || q_vector->tx.ring)
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
3591
ixgbe_for_each_ring(ring, q_vector->tx) {
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
3656
if (q_vector->tx.ring && q_vector->rx.ring) {
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
3663
} else if (q_vector->tx.ring) {
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
3831
if (!q_vector->rx.ring && !q_vector->tx.ring)
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
7997
if (qv->rx.ring || qv->tx.ring)
drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c
508
return ixgbe_xmit_zc(tx_ring, q_vector->tx.work_limit);
drivers/net/ethernet/intel/ixgbevf/ethtool.c
803
if (adapter->q_vector[0]->tx.count && adapter->q_vector[0]->rx.count)
drivers/net/ethernet/intel/ixgbevf/ethtool.c
826
if (adapter->q_vector[0]->tx.count &&
drivers/net/ethernet/intel/ixgbevf/ethtool.c
858
if (q_vector->tx.count && !q_vector->rx.count)
drivers/net/ethernet/intel/ixgbevf/ixgbevf.h
239
struct ixgbevf_ring_container rx, tx;
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
1280
ixgbevf_for_each_ring(ring, q_vector->tx) {
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
1368
ixgbevf_for_each_ring(ring, q_vector->tx)
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
1371
if (q_vector->tx.ring && !q_vector->rx.ring) {
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
1471
ixgbevf_update_itr(q_vector, &q_vector->tx);
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
1474
current_itr = max(q_vector->rx.itr, q_vector->tx.itr);
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
1527
if (q_vector->rx.ring || q_vector->tx.ring)
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
1551
if (q_vector->tx.ring && q_vector->rx.ring) {
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
1558
} else if (q_vector->tx.ring) {
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
1638
!adapter->q_vector[i]->tx.ring)
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
2792
ixgbevf_add_ring(ring, &q_vector->tx);
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
2820
ixgbevf_add_ring(ring, &q_vector->tx);
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
2884
ixgbevf_for_each_ring(ring, q_vector->tx) {
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
3269
if (qv->rx.ring || qv->tx.ring)
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
367
q_vector->tx.total_bytes += total_bytes;
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
368
q_vector->tx.total_packets += total_packets;
drivers/net/ethernet/marvell/octeontx2/af/mbox.h
1182
} tx;
drivers/net/ethernet/marvell/octeontx2/af/mbox.h
1498
} tx;
drivers/net/ethernet/marvell/octeontx2/af/mcs.c
1564
err = mcs_alloc_struct_mem(mcs, &mcs->tx);
drivers/net/ethernet/marvell/octeontx2/af/mcs.c
300
map = &mcs->tx;
drivers/net/ethernet/marvell/octeontx2/af/mcs.c
502
__set_bit(flow_id, mcs->tx.flow_ids.bmap);
drivers/net/ethernet/marvell/octeontx2/af/mcs.c
515
__set_bit(secy_id, mcs->tx.secy.bmap);
drivers/net/ethernet/marvell/octeontx2/af/mcs.c
552
map = &mcs->tx;
drivers/net/ethernet/marvell/octeontx2/af/mcs.c
590
map = (req->dir == MCS_RX) ? &mcs->rx : &mcs->tx;
drivers/net/ethernet/marvell/octeontx2/af/mcs.c
720
map = &mcs->tx;
drivers/net/ethernet/marvell/octeontx2/af/mcs.c
780
map = &mcs->tx;
drivers/net/ethernet/marvell/octeontx2/af/mcs.c
817
sc_bmap = &mcs->tx.sc;
drivers/net/ethernet/marvell/octeontx2/af/mcs.c
832
event.pcifunc = mcs->tx.sa2pf_map[event.sa_id];
drivers/net/ethernet/marvell/octeontx2/af/mcs.c
844
sc_bmap = &mcs->tx.sc;
drivers/net/ethernet/marvell/octeontx2/af/mcs.c
871
event.pcifunc = mcs->tx.sa2pf_map[event.sa_id];
drivers/net/ethernet/marvell/octeontx2/af/mcs.h
139
struct mcs_rsrc_map tx;
drivers/net/ethernet/marvell/octeontx2/af/mcs_cnf10kb.c
164
sc_bmap = &mcs->tx.sc;
drivers/net/ethernet/marvell/octeontx2/af/mcs_cnf10kb.c
187
event.pcifunc = mcs->tx.sa2pf_map[event.sa_id];
drivers/net/ethernet/marvell/octeontx2/af/mcs_cnf10kb.c
199
sc_bmap = &mcs->tx.sc;
drivers/net/ethernet/marvell/octeontx2/af/mcs_cnf10kb.c
214
event.pcifunc = mcs->tx.sa2pf_map[event.sa_id];
drivers/net/ethernet/marvell/octeontx2/af/mcs_rvu_if.c
646
map = &mcs->tx;
drivers/net/ethernet/marvell/octeontx2/af/mcs_rvu_if.c
696
map = &mcs->tx;
drivers/net/ethernet/marvell/octeontx2/af/mcs_rvu_if.c
776
map = (req->dir == MCS_RX) ? &mcs->rx : &mcs->tx;
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
249
map = &mcs->tx.sa;
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
301
map = &mcs->tx.sc;
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
373
map = &mcs->tx.flow_ids;
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
405
map = &mcs->tx.secy;
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
3071
int idx0 = req->tx.vtag0_idx;
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
3072
int idx1 = req->tx.vtag1_idx;
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
3080
if (req->tx.free_vtag0 && req->tx.free_vtag1)
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
3087
if (req->tx.free_vtag0) {
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
3093
if (req->tx.free_vtag1)
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
3113
if (req->tx.cfg_vtag0) {
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
3116
req->tx.vtag0, req->vtag_size);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
3124
if (req->tx.cfg_vtag1) {
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
3127
req->tx.vtag1, req->vtag_size);
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
3138
if (req->tx.cfg_vtag0)
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
3162
if ((req->tx.cfg_vtag0 || req->tx.cfg_vtag1) &&
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
3163
(req->tx.free_vtag0 || req->tx.free_vtag1))
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
3166
if (req->tx.cfg_vtag0 || req->tx.cfg_vtag1)
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
3169
if (req->tx.free_vtag0 || req->tx.free_vtag1)
drivers/net/ethernet/marvell/octeontx2/af/rvu_rep.c
174
rsp->tx.octs = RVU_LF_TX_STATS(TX_OCTS);
drivers/net/ethernet/marvell/octeontx2/af/rvu_rep.c
175
rsp->tx.ucast = RVU_LF_TX_STATS(TX_UCAST);
drivers/net/ethernet/marvell/octeontx2/af/rvu_rep.c
176
rsp->tx.bcast = RVU_LF_TX_STATS(TX_BCAST);
drivers/net/ethernet/marvell/octeontx2/af/rvu_rep.c
177
rsp->tx.mcast = RVU_LF_TX_STATS(TX_MCAST);
drivers/net/ethernet/marvell/octeontx2/af/rvu_rep.c
178
rsp->tx.drop = RVU_LF_TX_STATS(TX_DROP);
drivers/net/ethernet/marvell/octeontx2/af/rvu_rep.c
206
req.tx.cfg_vtag0 = true;
drivers/net/ethernet/marvell/octeontx2/af/rvu_rep.c
207
req.tx.vtag0 = FIELD_PREP(NIX_VLAN_ETYPE_MASK, etype) | vlan_tci;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2603
vtag_req->tx.free_vtag0 = 1;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2604
vtag_req->tx.vtag0_idx = config->tx_vtag_idx;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2675
vtag_req->tx.cfg_vtag0 = 1;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_pf.c
2676
vtag_req->tx.vtag0 = ((u64)ntohs(proto) << 16) | vlan;
drivers/net/ethernet/marvell/octeontx2/nic/rep.c
372
stats->tx_bytes = rsp->tx.octs;
drivers/net/ethernet/marvell/octeontx2/nic/rep.c
373
stats->tx_frames = rsp->tx.ucast + rsp->tx.bcast + rsp->tx.mcast;
drivers/net/ethernet/marvell/octeontx2/nic/rep.c
374
stats->tx_drops = rsp->tx.drop +
drivers/net/ethernet/marvell/sky2.c
2225
static enum flow_control sky2_flow(int rx, int tx)
drivers/net/ethernet/marvell/sky2.c
2228
return tx ? FC_BOTH : FC_RX;
drivers/net/ethernet/marvell/sky2.c
2230
return tx ? FC_TX : FC_NONE;
drivers/net/ethernet/mediatek/mtk_eth_soc.c
1304
int cnt = soc->tx.fq_dma_size;
drivers/net/ethernet/mediatek/mtk_eth_soc.c
1308
eth->scratch_ring = mtk_dma_ring_alloc(eth, cnt * soc->tx.desc_size,
drivers/net/ethernet/mediatek/mtk_eth_soc.c
1314
phy_ring_tail = eth->phy_scratch_ring + soc->tx.desc_size * (cnt - 1);
drivers/net/ethernet/mediatek/mtk_eth_soc.c
1316
for (j = 0; j < DIV_ROUND_UP(soc->tx.fq_dma_size, MTK_FQ_DMA_LENGTH); j++) {
drivers/net/ethernet/mediatek/mtk_eth_soc.c
1333
txd = eth->scratch_ring + (j * MTK_FQ_DMA_LENGTH + i) * soc->tx.desc_size;
drivers/net/ethernet/mediatek/mtk_eth_soc.c
1337
(j * MTK_FQ_DMA_LENGTH + i + 1) * soc->tx.desc_size;
drivers/net/ethernet/mediatek/mtk_eth_soc.c
1590
itx_buf = mtk_desc_to_tx_buf(ring, itxd, soc->tx.desc_size);
drivers/net/ethernet/mediatek/mtk_eth_soc.c
1631
soc->tx.dma_max_len);
drivers/net/ethernet/mediatek/mtk_eth_soc.c
1644
soc->tx.desc_size);
drivers/net/ethernet/mediatek/mtk_eth_soc.c
1687
next_idx = NEXT_DESP_IDX(txd_to_idx(ring, txd, soc->tx.desc_size),
drivers/net/ethernet/mediatek/mtk_eth_soc.c
1696
tx_buf = mtk_desc_to_tx_buf(ring, itxd, soc->tx.desc_size);
drivers/net/ethernet/mediatek/mtk_eth_soc.c
1721
eth->soc->tx.dma_max_len);
drivers/net/ethernet/mediatek/mtk_eth_soc.c
2004
tx_buf = mtk_desc_to_tx_buf(ring, txd, soc->tx.desc_size);
drivers/net/ethernet/mediatek/mtk_eth_soc.c
2023
soc->tx.desc_size);
drivers/net/ethernet/mediatek/mtk_eth_soc.c
2061
idx = txd_to_idx(ring, txd, soc->tx.desc_size);
drivers/net/ethernet/mediatek/mtk_eth_soc.c
2072
tx_buf = mtk_desc_to_tx_buf(ring, htxd, soc->tx.desc_size);
drivers/net/ethernet/mediatek/mtk_eth_soc.c
2471
eth->soc->tx.desc_size);
drivers/net/ethernet/mediatek/mtk_eth_soc.c
2522
desc = ring->dma + cpu * eth->soc->tx.desc_size;
drivers/net/ethernet/mediatek/mtk_eth_soc.c
2640
int i, sz = soc->tx.desc_size;
drivers/net/ethernet/mediatek/mtk_eth_soc.c
2648
ring_size = soc->tx.dma_size;
drivers/net/ethernet/mediatek/mtk_eth_soc.c
2755
mtk_dma_ring_free(eth, ring->dma_size * soc->tx.desc_size,
drivers/net/ethernet/mediatek/mtk_eth_soc.c
2762
ring->dma_size * soc->tx.desc_size,
drivers/net/ethernet/mediatek/mtk_eth_soc.c
3299
mtk_dma_ring_free(eth, soc->tx.fq_dma_size * soc->tx.desc_size,
drivers/net/ethernet/mediatek/mtk_eth_soc.c
3316
for (i = 0; i < DIV_ROUND_UP(soc->tx.fq_dma_size, MTK_FQ_DMA_LENGTH); i++) {
drivers/net/ethernet/mediatek/mtk_eth_soc.c
5358
.tx = {
drivers/net/ethernet/mediatek/mtk_eth_soc.c
5386
.tx = {
drivers/net/ethernet/mediatek/mtk_eth_soc.c
5416
.tx = {
drivers/net/ethernet/mediatek/mtk_eth_soc.c
5445
.tx = {
drivers/net/ethernet/mediatek/mtk_eth_soc.c
5471
.tx = {
drivers/net/ethernet/mediatek/mtk_eth_soc.c
5501
.tx = {
drivers/net/ethernet/mediatek/mtk_eth_soc.c
5531
.tx = {
drivers/net/ethernet/mediatek/mtk_eth_soc.c
5561
.tx = {
drivers/net/ethernet/mediatek/mtk_eth_soc.c
5585
.tx = {
drivers/net/ethernet/mediatek/mtk_eth_soc.h
1229
} tx;
drivers/net/ethernet/mediatek/mtk_eth_soc.h
342
#define TX_DMA_PLEN0(x) (((x) & eth->soc->tx.dma_max_len) << eth->soc->tx.dma_len_offset)
drivers/net/ethernet/mediatek/mtk_eth_soc.h
343
#define TX_DMA_PLEN1(x) ((x) & eth->soc->tx.dma_max_len)
drivers/net/ethernet/mediatek/mtk_star_emac.c
439
bool rx, bool tx)
drivers/net/ethernet/mediatek/mtk_star_emac.c
445
if (tx)
drivers/net/ethernet/mediatek/mtk_star_emac.c
454
bool rx, bool tx)
drivers/net/ethernet/mediatek/mtk_star_emac.c
460
if (tx)
drivers/net/ethernet/mediatek/mtk_star_emac.c
774
bool rx, tx;
drivers/net/ethernet/mediatek/mtk_star_emac.c
778
tx = (intr_status & MTK_STAR_BIT_INT_STS_TNTC) &&
drivers/net/ethernet/mediatek/mtk_star_emac.c
781
if (rx || tx) {
drivers/net/ethernet/mediatek/mtk_star_emac.c
784
mtk_star_disable_dma_irq(priv, rx, tx);
drivers/net/ethernet/mediatek/mtk_star_emac.c
789
if (tx)
drivers/net/ethernet/mediatek/mtk_wed.c
1541
mtk_wed_ring_reset(struct mtk_wed_ring *ring, int size, bool tx)
drivers/net/ethernet/mediatek/mtk_wed.c
1551
if (tx)
drivers/net/ethernet/mediatek/mtk_wed.c
1860
int size, u32 desc_size, bool tx)
drivers/net/ethernet/mediatek/mtk_wed.c
1869
mtk_wed_ring_reset(ring, size, tx);
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
3131
struct netdev_queue_stats_tx *tx)
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
3147
tx->packets = 0;
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
3148
tx->bytes = 0;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec.h
252
struct mlx5e_ipsec_tx *tx;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1221
static int ipsec_counter_rule_tx(struct mlx5_core_dev *mdev, struct mlx5e_ipsec_tx *tx)
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1237
dest.counter = tx->fc->cnt;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1238
fte = mlx5_add_flow_rules(tx->ft.status, spec, &flow_act, &dest, 1);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1246
tx->status.rule = fte;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1255
static void tx_destroy(struct mlx5e_ipsec *ipsec, struct mlx5e_ipsec_tx *tx,
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1259
if (tx->chains) {
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1260
ipsec_chains_destroy(tx->chains);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1262
mlx5_del_flow_rules(tx->pol.rule);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1263
mlx5_destroy_flow_group(tx->pol.group);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1264
mlx5_destroy_flow_table(tx->ft.pol);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1267
if (tx == ipsec->tx_esw) {
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1268
mlx5_del_flow_rules(tx->sa.rule);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1269
mlx5_destroy_flow_group(tx->sa.group);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1271
mlx5_destroy_flow_table(tx->ft.sa);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1272
if (tx->allow_tunnel_mode)
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1274
mlx5_del_flow_rules(tx->status.rule);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1275
mlx5_destroy_flow_table(tx->ft.status);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1279
struct mlx5e_ipsec_tx *tx,
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1282
if (tx == ipsec->tx_esw) {
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1294
static int tx_create(struct mlx5e_ipsec *ipsec, struct mlx5e_ipsec_tx *tx,
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1304
ipsec_tx_create_attr_set(ipsec, tx, &attr);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1305
ft = ipsec_ft_create(tx->ns, attr.cnt_level, attr.prio, 1, 1, 0);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1308
tx->ft.status = ft;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1310
err = ipsec_counter_rule_tx(mdev, tx);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1315
tx->allow_tunnel_mode =
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1316
mlx5_eswitch_block_encap(mdev, tx == ipsec->tx_esw);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1318
if (tx->allow_tunnel_mode)
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1320
ft = ipsec_ft_create(tx->ns, attr.sa_level, attr.prio, 1, 4, flags);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1325
tx->ft.sa = ft;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1327
if (tx == ipsec->tx_esw) {
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1330
err = ipsec_miss_create(mdev, tx->ft.sa, &tx->sa, &dest);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1337
tx->chains = ipsec_chains_create(
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1338
mdev, tx->ft.sa, attr.chains_ns, attr.prio, attr.pol_level,
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1339
&tx->ft.pol);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1340
if (IS_ERR(tx->chains)) {
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1341
err = PTR_ERR(tx->chains);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1348
ft = ipsec_ft_create(tx->ns, attr.pol_level, attr.prio, 1, 2, 0);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1353
tx->ft.pol = ft;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1355
dest.ft = tx->ft.sa;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1356
err = ipsec_miss_create(mdev, tx->ft.pol, &tx->pol, &dest);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1358
mlx5_destroy_flow_table(tx->ft.pol);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1363
err = mlx5_ipsec_fs_roce_tx_create(mdev, roce, tx->ft.pol, false);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1369
if (tx->chains) {
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1370
ipsec_chains_destroy(tx->chains);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1372
mlx5_del_flow_rules(tx->pol.rule);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1373
mlx5_destroy_flow_group(tx->pol.group);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1374
mlx5_destroy_flow_table(tx->ft.pol);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1377
if (tx == ipsec->tx_esw) {
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1378
mlx5_del_flow_rules(tx->sa.rule);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1379
mlx5_destroy_flow_group(tx->sa.group);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1382
mlx5_destroy_flow_table(tx->ft.sa);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1384
if (tx->allow_tunnel_mode)
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1386
mlx5_del_flow_rules(tx->status.rule);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1388
mlx5_destroy_flow_table(tx->ft.status);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1412
struct mlx5e_ipsec_tx *tx)
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1416
if (tx->ft.refcnt)
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1423
err = tx_create(ipsec, tx, ipsec->roce);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1429
if (tx == ipsec->tx_esw)
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1430
ipsec_esw_tx_ft_policy_set(mdev, tx->ft.pol);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1433
tx->ft.refcnt++;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1437
static void tx_put(struct mlx5e_ipsec *ipsec, struct mlx5e_ipsec_tx *tx)
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1439
if (--tx->ft.refcnt)
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1442
if (tx == ipsec->tx_esw) {
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1447
tx_destroy(ipsec, tx, ipsec->roce);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1455
struct mlx5e_ipsec_tx *tx = ipsec_tx(ipsec, type);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1459
mutex_lock(&tx->ft.mutex);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1460
err = tx_get(mdev, ipsec, tx);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1464
ft = tx->chains ? ipsec_chains_get_table(tx->chains, prio) : tx->ft.pol;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1470
mutex_unlock(&tx->ft.mutex);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1474
tx_put(ipsec, tx);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1476
mutex_unlock(&tx->ft.mutex);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1483
struct mlx5e_ipsec_tx *tx = ipsec_tx(ipsec, type);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1486
mutex_lock(&tx->ft.mutex);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1487
err = tx_get(mdev, ipsec, tx);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1488
mutex_unlock(&tx->ft.mutex);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1492
return tx;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1497
struct mlx5e_ipsec_tx *tx = ipsec_tx(ipsec, type);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1499
mutex_lock(&tx->ft.mutex);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1500
tx_put(ipsec, tx);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1501
mutex_unlock(&tx->ft.mutex);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1506
struct mlx5e_ipsec_tx *tx = ipsec_tx(ipsec, type);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1508
mutex_lock(&tx->ft.mutex);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1509
if (tx->chains)
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1510
ipsec_chains_put_table(tx->chains, prio);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1512
tx_put(ipsec, tx);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1513
mutex_unlock(&tx->ft.mutex);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
2171
struct mlx5e_ipsec_tx *tx;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
2175
tx = tx_ft_get(mdev, ipsec, attrs->type);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
2176
if (IS_ERR(tx))
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
2177
return PTR_ERR(tx);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
2224
dest[0].ft = tx->ft.status;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
2228
rule = mlx5_add_flow_rules(tx->ft.sa, spec, &flow_act, dest, 2);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
2263
struct mlx5e_ipsec_tx *tx;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
2276
tx = ipsec_tx(ipsec, attrs->type);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
2300
dest[dstn].counter = tx->fc->drop;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
2310
if (tx == ipsec->tx_esw && tx->chains)
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
2312
dest[dstn].ft = tx->ft.sa;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
2423
ipsec_fs_destroy_single_counter(mdev, ipsec->tx->fc);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
2482
ipsec->tx->fc = fc;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
2507
ipsec_fs_destroy_single_counter(mdev, ipsec->tx->fc);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
2542
fc = ipsec->tx->fc;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
2732
if (!ipsec->tx)
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
2739
mutex_destroy(&ipsec->tx->ft.mutex);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
2740
WARN_ON(ipsec->tx->ft.refcnt);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
2741
kfree(ipsec->tx);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
2790
ipsec->tx = kzalloc_obj(*ipsec->tx);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
2791
if (!ipsec->tx)
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
2806
mutex_init(&ipsec->tx->ft.mutex);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
2809
ipsec->tx->ns = ns;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
2829
kfree(ipsec->tx);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
2860
struct mlx5e_ipsec_tx *tx = ipsec_tx(ipsec, x->xso.type);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
2862
from_fdb = (tx == ipsec->tx_esw);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
650
mutex_lock(&ipsec->tx->ft.mutex);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
651
if (ipsec->tx->ft.refcnt)
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
652
mlx5_ipsec_fs_roce_tx_create(ipsec->mdev, ipsec->roce, ipsec->tx->ft.pol,
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
654
mutex_unlock(&ipsec->tx->ft.mutex);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
667
mutex_lock(&ipsec->tx->ft.mutex);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
668
if (ipsec->tx->ft.refcnt)
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
670
mutex_unlock(&ipsec->tx->ft.mutex);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
91
return ipsec->tx;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/psp.c
991
struct psp_key_parsed *tx = &pas->tx;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/psp.c
999
err = mlx5_create_encryption_key(mdev, tx->key,
drivers/net/ethernet/mellanox/mlx5/core/en_accel/psp_rxtx.c
113
psp_st->spi = pas->tx.spi;
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
5461
struct netdev_queue_stats_tx *tx)
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
5494
tx->packets = 0;
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
5495
tx->bytes = 0;
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
5520
tx->packets += sq_stats->packets;
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
5521
tx->bytes += sq_stats->bytes;
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
5538
tx->packets += sq_stats->packets;
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
5539
tx->bytes += sq_stats->bytes;
drivers/net/ethernet/mellanox/mlx5/core/lib/ipsec_fs_roce.c
399
roce = &ipsec_roce->tx;
drivers/net/ethernet/mellanox/mlx5/core/lib/ipsec_fs_roce.c
45
struct mlx5_ipsec_tx_roce tx;
drivers/net/ethernet/mellanox/mlx5/core/lib/ipsec_fs_roce.c
572
tx_roce = &ipsec_roce->tx;
drivers/net/ethernet/mellanox/mlx5/core/lib/ipsec_fs_roce.c
616
roce = &ipsec_roce->tx;
drivers/net/ethernet/mellanox/mlx5/core/lib/ipsec_fs_roce.c
866
roce_ipsec->tx.ns = ns;
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/action_ste_pool.c
46
enum mlx5hws_pool_optimize opt, size_t log_sz, bool tx)
drivers/net/ethernet/mellanox/mlx5/core/steering/hws/action_ste_pool.c
60
if (tx) {
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_action.c
731
dest_tbl->tbl->tx.s_anchor->chunk;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_dbg.c
552
struct mlx5dr_rule_rx_tx *tx = &rule->tx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_dbg.c
574
if (tx->nic_matcher) {
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_dbg.c
575
ret = dr_dump_rule_rx_tx(file, buff, tx, false, rule_id, format_ver);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_dbg.c
748
struct mlx5dr_matcher_rx_tx *tx = &matcher->tx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_dbg.c
777
if (tx->nic_tbl) {
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_dbg.c
778
ret = dr_dump_matcher_rx_tx(file, buff, false, tx, matcher_id);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_dbg.c
835
struct mlx5dr_table_rx_tx *tx = &table->tx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_dbg.c
857
if (tx->nic_dmn) {
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_dbg.c
858
ret = dr_dump_table_rx_tx(file, buff, false, tx,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_domain.c
421
if (!DR_DOMAIN_SW_STEERING_SUPPORTED(dmn, tx))
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_domain.c
425
dmn->info.tx.type = DR_DOMAIN_NIC_TYPE_TX;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_domain.c
426
dmn->info.tx.default_icm_addr = dmn->info.caps.nic_tx_allow_address;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_domain.c
427
dmn->info.tx.drop_icm_addr = dmn->info.caps.nic_tx_drop_address;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_domain.c
437
dmn->info.tx.type = DR_DOMAIN_NIC_TYPE_TX;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_domain.c
441
dmn->info.tx.default_icm_addr = vport_cap->icm_address_tx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_domain.c
444
dmn->info.tx.drop_icm_addr = dmn->info.caps.esw_tx_drop_address;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_domain.c
478
mutex_init(&dmn->info.tx.mutex);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_domain.c
532
mutex_destroy(&dmn->info.tx.mutex);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
797
dr_matcher_uninit_nic(&matcher->tx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
809
dr_matcher_uninit_nic(&matcher->tx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
886
ret = dr_matcher_init_nic(matcher, &matcher->tx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
958
matcher->tx.nic_tbl = &tbl->tx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
959
ret = dr_matcher_init_nic(matcher, &matcher->tx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
963
matcher->tx.nic_tbl = &tbl->tx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_rule.c
1047
dr_rule_destroy_rule_nic(rule, &rule->tx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_rule.c
1062
dr_rule_destroy_rule_nic(rule, &rule->tx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_rule.c
1274
ret = dr_rule_create_rule_nic(rule, &rule->tx, ©_param,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_rule.c
1320
rule->tx.nic_matcher = &matcher->tx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_rule.c
1321
ret = dr_rule_create_rule_nic(rule, &rule->tx, ¶m,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_rule.c
1326
rule->tx.nic_matcher = &matcher->tx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_table.c
107
dr_table_uninit_nic(&tbl->tx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_table.c
167
ret = dr_table_init_nic(tbl->dmn, &tbl->tx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_table.c
194
tbl->tx.nic_dmn = &tbl->dmn->info.tx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_table.c
195
ret = dr_table_init_nic(tbl->dmn, &tbl->tx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_table.c
200
tbl->tx.nic_dmn = &tbl->dmn->info.tx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_table.c
232
if (tbl->tx.s_anchor)
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_table.c
233
icm_addr_tx = mlx5dr_icm_pool_get_chunk_icm_addr(tbl->tx.s_anchor->chunk);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_table.c
29
action->dest_tbl->tbl->tx.s_anchor->chunk;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_table.c
65
ret = dr_table_set_miss_action_nic(tbl->dmn, &tbl->tx, action);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_table.c
95
dr_table_uninit_nic(&tbl->tx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
1159
struct mlx5dr_rule_rx_tx tx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
1200
mlx5dr_domain_nic_lock(&dmn->info.tx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
1205
mlx5dr_domain_nic_unlock(&dmn->info.tx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
929
struct mlx5dr_domain_rx_tx tx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
969
struct mlx5dr_table_rx_tx tx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
999
struct mlx5dr_matcher_rx_tx tx;
drivers/net/ethernet/mellanox/mlxsw/txheader.h
11
MLXSW_ITEM32(tx, hdr, version, 0x00, 28, 4);
drivers/net/ethernet/mellanox/mlxsw/txheader.h
18
MLXSW_ITEM32(tx, hdr, ctl, 0x00, 26, 2);
drivers/net/ethernet/mellanox/mlxsw/txheader.h
23
MLXSW_ITEM32(tx, hdr, proto, 0x00, 21, 3);
drivers/net/ethernet/mellanox/mlxsw/txheader.h
28
MLXSW_ITEM32(tx, hdr, rx_is_router, 0x00, 19, 1);
drivers/net/ethernet/mellanox/mlxsw/txheader.h
34
MLXSW_ITEM32(tx, hdr, fid_valid, 0x00, 16, 1);
drivers/net/ethernet/mellanox/mlxsw/txheader.h
39
MLXSW_ITEM32(tx, hdr, swid, 0x00, 12, 3);
drivers/net/ethernet/mellanox/mlxsw/txheader.h
45
MLXSW_ITEM32(tx, hdr, control_tclass, 0x00, 6, 1);
drivers/net/ethernet/mellanox/mlxsw/txheader.h
55
MLXSW_ITEM32(tx, hdr, port_mid, 0x04, 16, 16);
drivers/net/ethernet/mellanox/mlxsw/txheader.h
62
MLXSW_ITEM32(tx, hdr, fid, 0x08, 16, 16);
drivers/net/ethernet/mellanox/mlxsw/txheader.h
68
MLXSW_ITEM32(tx, hdr, type, 0x0C, 0, 4);
drivers/net/ethernet/meta/fbnic/fbnic_ethtool.c
162
memset(clone->tx, 0, sizeof(clone->tx));
drivers/net/ethernet/meta/fbnic/fbnic_ethtool.c
1633
ring = fbn->tx[i];
drivers/net/ethernet/meta/fbnic/fbnic_ethtool.c
214
for (i = 0; i < ARRAY_SIZE(orig->tx); i++)
drivers/net/ethernet/meta/fbnic/fbnic_ethtool.c
215
swap(clone->tx[i], orig->tx[i]);
drivers/net/ethernet/meta/fbnic/fbnic_ethtool.c
561
fbnic_get_xdp_queue_stats(fbn->tx[i + FBNIC_MAX_TXQS], &data);
drivers/net/ethernet/meta/fbnic/fbnic_netdev.c
420
struct fbnic_ring *txr = fbn->tx[i];
drivers/net/ethernet/meta/fbnic/fbnic_netdev.c
474
struct fbnic_ring *xdpr = fbn->tx[FBNIC_MAX_TXQS + i];
drivers/net/ethernet/meta/fbnic/fbnic_netdev.c
620
struct netdev_queue_stats_tx *tx)
drivers/net/ethernet/meta/fbnic/fbnic_netdev.c
623
struct fbnic_ring *txr = fbn->tx[idx];
drivers/net/ethernet/meta/fbnic/fbnic_netdev.c
644
tx->bytes = bytes;
drivers/net/ethernet/meta/fbnic/fbnic_netdev.c
645
tx->packets = packets;
drivers/net/ethernet/meta/fbnic/fbnic_netdev.c
646
tx->needs_csum = csum + lso;
drivers/net/ethernet/meta/fbnic/fbnic_netdev.c
647
tx->hw_gso_wire_packets = lso;
drivers/net/ethernet/meta/fbnic/fbnic_netdev.c
648
tx->stop = stop;
drivers/net/ethernet/meta/fbnic/fbnic_netdev.c
649
tx->wake = wake;
drivers/net/ethernet/meta/fbnic/fbnic_netdev.c
651
xdpr = fbn->tx[FBNIC_MAX_TXQS + idx];
drivers/net/ethernet/meta/fbnic/fbnic_netdev.c
660
tx->bytes += bytes;
drivers/net/ethernet/meta/fbnic/fbnic_netdev.c
661
tx->packets += packets;
drivers/net/ethernet/meta/fbnic/fbnic_netdev.c
667
struct netdev_queue_stats_tx *tx)
drivers/net/ethernet/meta/fbnic/fbnic_netdev.c
671
tx->bytes = fbn->tx_stats.bytes;
drivers/net/ethernet/meta/fbnic/fbnic_netdev.c
672
tx->packets = fbn->tx_stats.packets;
drivers/net/ethernet/meta/fbnic/fbnic_netdev.c
673
tx->needs_csum = fbn->tx_stats.twq.csum_partial + fbn->tx_stats.twq.lso;
drivers/net/ethernet/meta/fbnic/fbnic_netdev.c
674
tx->hw_gso_wire_packets = fbn->tx_stats.twq.lso;
drivers/net/ethernet/meta/fbnic/fbnic_netdev.c
675
tx->stop = fbn->tx_stats.twq.stop;
drivers/net/ethernet/meta/fbnic/fbnic_netdev.c
676
tx->wake = fbn->tx_stats.twq.wake;
drivers/net/ethernet/meta/fbnic/fbnic_netdev.c
693
unsigned int tx, unsigned int rx)
drivers/net/ethernet/meta/fbnic/fbnic_netdev.c
700
tx = min(tx, max_napis);
drivers/net/ethernet/meta/fbnic/fbnic_netdev.c
701
fbn->num_tx_queues = tx;
drivers/net/ethernet/meta/fbnic/fbnic_netdev.c
706
fbn->num_napi = max(tx, rx);
drivers/net/ethernet/meta/fbnic/fbnic_netdev.h
23
struct fbnic_ring *tx[FBNIC_MAX_TXQS + FBNIC_MAX_XDPQS];
drivers/net/ethernet/meta/fbnic/fbnic_netdev.h
90
unsigned int tx, unsigned int rx);
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
1472
WARN_ON(fbn->tx[txr->q_idx] && fbn->tx[txr->q_idx] != txr);
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
1473
fbn->tx[txr->q_idx] = NULL;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
1485
WARN_ON(fbn->tx[xdpr->q_idx] && fbn->tx[xdpr->q_idx] != xdpr);
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
1486
fbn->tx[xdpr->q_idx] = NULL;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
1686
fbn->tx[txq_idx] = &qt->sub0;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
1708
fbn->tx[xdp_idx] = &qt->sub1;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
2334
static const struct fbnic_idle_regs tx[] = {
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
2348
false, fbd, tx, ARRAY_SIZE(tx));
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
2353
fbd, tx, ARRAY_SIZE(tx));
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
2357
fbnic_idle_dump(fbd, tx, ARRAY_SIZE(tx), "Tx", err);
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
460
return fbnic_xmit_frame_ring(skb, fbn->tx[q_map]);
drivers/net/ethernet/micrel/ksz884x.c
1524
desc->sw.ctrl.tx.hw_owned = 1;
drivers/net/ethernet/micrel/ksz884x.c
1568
desc->sw.buf.tx.buf_size = len;
drivers/net/ethernet/micrel/ksz884x.c
2836
static void set_flow_ctrl(struct ksz_hw *hw, int rx, int tx)
drivers/net/ethernet/micrel/ksz884x.c
2847
if (tx)
drivers/net/ethernet/micrel/ksz884x.c
2863
int tx;
drivers/net/ethernet/micrel/ksz884x.c
2868
rx = tx = 0;
drivers/net/ethernet/micrel/ksz884x.c
2870
rx = tx = 1;
drivers/net/ethernet/micrel/ksz884x.c
2873
rx = tx = 1;
drivers/net/ethernet/micrel/ksz884x.c
2878
tx = 1;
drivers/net/ethernet/micrel/ksz884x.c
2886
set_flow_ctrl(hw, rx, tx);
drivers/net/ethernet/micrel/ksz884x.c
3608
hw->tx_desc_info.cur->sw.buf.tx.first_seg = 1;
drivers/net/ethernet/micrel/ksz884x.c
3634
cur->sw.buf.tx.last_seg = 1;
drivers/net/ethernet/micrel/ksz884x.c
3638
cur->sw.buf.tx.intr = 1;
drivers/net/ethernet/micrel/ksz884x.c
3644
cur->sw.buf.tx.dest_port = hw->dst_ports;
drivers/net/ethernet/micrel/ksz884x.c
4337
(desc)->sw.buf.tx.csum_gen_tcp = 1;
drivers/net/ethernet/micrel/ksz884x.c
4338
(desc)->sw.buf.tx.csum_gen_udp = 1;
drivers/net/ethernet/micrel/ksz884x.c
4378
if (status.tx.hw_owned) {
drivers/net/ethernet/micrel/ksz884x.c
855
struct ksz_desc_tx_stat tx;
drivers/net/ethernet/micrel/ksz884x.c
861
struct ksz_desc_tx_buf tx;
drivers/net/ethernet/microchip/enc28j60.c
103
spi_message_add_tail(&tx, &msg);
drivers/net/ethernet/microchip/enc28j60.c
89
struct spi_transfer tx = {
drivers/net/ethernet/microchip/lan743x_ethtool.c
868
for (i = 0; i < ARRAY_SIZE(adapter->tx); i++) {
drivers/net/ethernet/microchip/lan743x_ethtool.c
869
pkt_cnt = (u64)(adapter->tx[i].frame_count);
drivers/net/ethernet/microchip/lan743x_main.c
1568
static void lan743x_tx_release_desc(struct lan743x_tx *tx,
drivers/net/ethernet/microchip/lan743x_main.c
1576
descriptor = &tx->ring_cpu_ptr[descriptor_index];
drivers/net/ethernet/microchip/lan743x_main.c
1577
buffer_info = &tx->buffer_info[descriptor_index];
drivers/net/ethernet/microchip/lan743x_main.c
1592
dma_unmap_page(&tx->adapter->pdev->dev,
drivers/net/ethernet/microchip/lan743x_main.c
1597
dma_unmap_single(&tx->adapter->pdev->dev,
drivers/net/ethernet/microchip/lan743x_main.c
1614
lan743x_ptp_unrequest_tx_timestamp(tx->adapter);
drivers/net/ethernet/microchip/lan743x_main.c
1619
lan743x_ptp_tx_timestamp_skb(tx->adapter,
drivers/net/ethernet/microchip/lan743x_main.c
1634
static int lan743x_tx_next_index(struct lan743x_tx *tx, int index)
drivers/net/ethernet/microchip/lan743x_main.c
1636
return ((++index) % tx->ring_size);
drivers/net/ethernet/microchip/lan743x_main.c
1639
static void lan743x_tx_release_completed_descriptors(struct lan743x_tx *tx)
drivers/net/ethernet/microchip/lan743x_main.c
1641
while (le32_to_cpu(*tx->head_cpu_ptr) != (tx->last_head)) {
drivers/net/ethernet/microchip/lan743x_main.c
1642
lan743x_tx_release_desc(tx, tx->last_head, false);
drivers/net/ethernet/microchip/lan743x_main.c
1643
tx->last_head = lan743x_tx_next_index(tx, tx->last_head);
drivers/net/ethernet/microchip/lan743x_main.c
1647
static void lan743x_tx_release_all_descriptors(struct lan743x_tx *tx)
drivers/net/ethernet/microchip/lan743x_main.c
1651
original_head = tx->last_head;
drivers/net/ethernet/microchip/lan743x_main.c
1653
lan743x_tx_release_desc(tx, tx->last_head, true);
drivers/net/ethernet/microchip/lan743x_main.c
1654
tx->last_head = lan743x_tx_next_index(tx, tx->last_head);
drivers/net/ethernet/microchip/lan743x_main.c
1655
} while (tx->last_head != original_head);
drivers/net/ethernet/microchip/lan743x_main.c
1656
memset(tx->ring_cpu_ptr, 0,
drivers/net/ethernet/microchip/lan743x_main.c
1657
sizeof(*tx->ring_cpu_ptr) * (tx->ring_size));
drivers/net/ethernet/microchip/lan743x_main.c
1658
memset(tx->buffer_info, 0,
drivers/net/ethernet/microchip/lan743x_main.c
1659
sizeof(*tx->buffer_info) * (tx->ring_size));
drivers/net/ethernet/microchip/lan743x_main.c
1662
static int lan743x_tx_get_desc_cnt(struct lan743x_tx *tx,
drivers/net/ethernet/microchip/lan743x_main.c
1675
static int lan743x_tx_get_avail_desc(struct lan743x_tx *tx)
drivers/net/ethernet/microchip/lan743x_main.c
1677
int last_head = tx->last_head;
drivers/net/ethernet/microchip/lan743x_main.c
1678
int last_tail = tx->last_tail;
drivers/net/ethernet/microchip/lan743x_main.c
1681
return tx->ring_size - last_tail + last_head - 1;
drivers/net/ethernet/microchip/lan743x_main.c
1731
void lan743x_tx_set_timestamping_mode(struct lan743x_tx *tx,
drivers/net/ethernet/microchip/lan743x_main.c
1736
tx->ts_flags |= TX_TS_FLAG_TIMESTAMPING_ENABLED;
drivers/net/ethernet/microchip/lan743x_main.c
1738
tx->ts_flags &= ~TX_TS_FLAG_TIMESTAMPING_ENABLED;
drivers/net/ethernet/microchip/lan743x_main.c
1740
tx->ts_flags |= TX_TS_FLAG_ONE_STEP_SYNC;
drivers/net/ethernet/microchip/lan743x_main.c
1742
tx->ts_flags &= ~TX_TS_FLAG_ONE_STEP_SYNC;
drivers/net/ethernet/microchip/lan743x_main.c
1745
static int lan743x_tx_frame_start(struct lan743x_tx *tx,
drivers/net/ethernet/microchip/lan743x_main.c
1757
struct lan743x_adapter *adapter = tx->adapter;
drivers/net/ethernet/microchip/lan743x_main.c
1761
tx->frame_flags |= TX_FRAME_FLAG_IN_PROGRESS;
drivers/net/ethernet/microchip/lan743x_main.c
1762
tx->frame_first = tx->last_tail;
drivers/net/ethernet/microchip/lan743x_main.c
1763
tx->frame_tail = tx->frame_first;
drivers/net/ethernet/microchip/lan743x_main.c
1765
tx_descriptor = &tx->ring_cpu_ptr[tx->frame_tail];
drivers/net/ethernet/microchip/lan743x_main.c
1766
buffer_info = &tx->buffer_info[tx->frame_tail];
drivers/net/ethernet/microchip/lan743x_main.c
1782
tx->frame_data0 = (first_buffer_length &
drivers/net/ethernet/microchip/lan743x_main.c
1788
tx->frame_data0 |= TX_DESC_DATA0_TSE_;
drivers/net/ethernet/microchip/lan743x_main.c
1791
tx->frame_data0 |= TX_DESC_DATA0_ICE_ |
drivers/net/ethernet/microchip/lan743x_main.c
1799
static void lan743x_tx_frame_add_lso(struct lan743x_tx *tx,
drivers/net/ethernet/microchip/lan743x_main.c
1810
tx->frame_data0 |= TX_DESC_DATA0_EXT_;
drivers/net/ethernet/microchip/lan743x_main.c
1812
tx->frame_data0 |= TX_DESC_DATA0_LS_;
drivers/net/ethernet/microchip/lan743x_main.c
1813
tx->frame_data0 |= TX_DESC_DATA0_IOC_;
drivers/net/ethernet/microchip/lan743x_main.c
1814
tx->frame_last = tx->frame_first;
drivers/net/ethernet/microchip/lan743x_main.c
1816
tx_descriptor = &tx->ring_cpu_ptr[tx->frame_tail];
drivers/net/ethernet/microchip/lan743x_main.c
1817
tx_descriptor->data0 = cpu_to_le32(tx->frame_data0);
drivers/net/ethernet/microchip/lan743x_main.c
1820
tx->frame_tail = lan743x_tx_next_index(tx, tx->frame_tail);
drivers/net/ethernet/microchip/lan743x_main.c
1821
tx_descriptor = &tx->ring_cpu_ptr[tx->frame_tail];
drivers/net/ethernet/microchip/lan743x_main.c
1822
buffer_info = &tx->buffer_info[tx->frame_tail];
drivers/net/ethernet/microchip/lan743x_main.c
1834
tx->frame_data0 = (frame_length & TX_DESC_DATA0_EXT_PAY_LENGTH_MASK_) |
drivers/net/ethernet/microchip/lan743x_main.c
1841
static int lan743x_tx_frame_add_fragment(struct lan743x_tx *tx,
drivers/net/ethernet/microchip/lan743x_main.c
1850
struct lan743x_adapter *adapter = tx->adapter;
drivers/net/ethernet/microchip/lan743x_main.c
1860
tx_descriptor = &tx->ring_cpu_ptr[tx->frame_tail];
drivers/net/ethernet/microchip/lan743x_main.c
1861
tx_descriptor->data0 = cpu_to_le32(tx->frame_data0);
drivers/net/ethernet/microchip/lan743x_main.c
1864
tx->frame_tail = lan743x_tx_next_index(tx, tx->frame_tail);
drivers/net/ethernet/microchip/lan743x_main.c
1865
tx_descriptor = &tx->ring_cpu_ptr[tx->frame_tail];
drivers/net/ethernet/microchip/lan743x_main.c
1866
buffer_info = &tx->buffer_info[tx->frame_tail];
drivers/net/ethernet/microchip/lan743x_main.c
1874
desc_index = tx->frame_first;
drivers/net/ethernet/microchip/lan743x_main.c
1875
while (desc_index != tx->frame_tail) {
drivers/net/ethernet/microchip/lan743x_main.c
1876
lan743x_tx_release_desc(tx, desc_index, true);
drivers/net/ethernet/microchip/lan743x_main.c
1877
desc_index = lan743x_tx_next_index(tx, desc_index);
drivers/net/ethernet/microchip/lan743x_main.c
1880
tx->frame_flags &= ~TX_FRAME_FLAG_IN_PROGRESS;
drivers/net/ethernet/microchip/lan743x_main.c
1881
tx->frame_first = 0;
drivers/net/ethernet/microchip/lan743x_main.c
1882
tx->frame_data0 = 0;
drivers/net/ethernet/microchip/lan743x_main.c
1883
tx->frame_tail = 0;
drivers/net/ethernet/microchip/lan743x_main.c
1884
tx->frame_last = 0;
drivers/net/ethernet/microchip/lan743x_main.c
1899
tx->frame_data0 = (fragment_length & TX_DESC_DATA0_BUF_LENGTH_MASK_) |
drivers/net/ethernet/microchip/lan743x_main.c
1907
static void lan743x_tx_frame_end(struct lan743x_tx *tx,
drivers/net/ethernet/microchip/lan743x_main.c
1917
struct lan743x_adapter *adapter = tx->adapter;
drivers/net/ethernet/microchip/lan743x_main.c
1921
if ((tx->frame_data0 & TX_DESC_DATA0_DTYPE_MASK_) ==
drivers/net/ethernet/microchip/lan743x_main.c
1923
tx->frame_data0 |= TX_DESC_DATA0_LS_;
drivers/net/ethernet/microchip/lan743x_main.c
1924
tx->frame_data0 |= TX_DESC_DATA0_IOC_;
drivers/net/ethernet/microchip/lan743x_main.c
1925
tx->frame_last = tx->frame_tail;
drivers/net/ethernet/microchip/lan743x_main.c
1928
tx_descriptor = &tx->ring_cpu_ptr[tx->frame_last];
drivers/net/ethernet/microchip/lan743x_main.c
1929
buffer_info = &tx->buffer_info[tx->frame_last];
drivers/net/ethernet/microchip/lan743x_main.c
1936
tx_descriptor = &tx->ring_cpu_ptr[tx->frame_tail];
drivers/net/ethernet/microchip/lan743x_main.c
1937
tx_descriptor->data0 = cpu_to_le32(tx->frame_data0);
drivers/net/ethernet/microchip/lan743x_main.c
1938
tx->frame_tail = lan743x_tx_next_index(tx, tx->frame_tail);
drivers/net/ethernet/microchip/lan743x_main.c
1939
tx->last_tail = tx->frame_tail;
drivers/net/ethernet/microchip/lan743x_main.c
1943
if (tx->vector_flags & LAN743X_VECTOR_FLAG_VECTOR_ENABLE_AUTO_SET)
drivers/net/ethernet/microchip/lan743x_main.c
1945
if (tx->vector_flags & LAN743X_VECTOR_FLAG_SOURCE_ENABLE_AUTO_SET)
drivers/net/ethernet/microchip/lan743x_main.c
1949
lan743x_csr_write(adapter, TX_TAIL(tx->channel_number),
drivers/net/ethernet/microchip/lan743x_main.c
1950
tx_tail_flags | tx->frame_tail);
drivers/net/ethernet/microchip/lan743x_main.c
1951
tx->frame_flags &= ~TX_FRAME_FLAG_IN_PROGRESS;
drivers/net/ethernet/microchip/lan743x_main.c
1954
static netdev_tx_t lan743x_tx_xmit_frame(struct lan743x_tx *tx,
drivers/net/ethernet/microchip/lan743x_main.c
1970
required_number_of_descriptors = lan743x_tx_get_desc_cnt(tx, skb);
drivers/net/ethernet/microchip/lan743x_main.c
1972
spin_lock_irqsave(&tx->ring_lock, irq_flags);
drivers/net/ethernet/microchip/lan743x_main.c
1974
lan743x_tx_get_avail_desc(tx)) {
drivers/net/ethernet/microchip/lan743x_main.c
1975
if (required_number_of_descriptors > (tx->ring_size - 1)) {
drivers/net/ethernet/microchip/lan743x_main.c
1979
tx->rqd_descriptors = required_number_of_descriptors;
drivers/net/ethernet/microchip/lan743x_main.c
1981
txq = netdev_get_tx_queue(tx->adapter->netdev,
drivers/net/ethernet/microchip/lan743x_main.c
1982
tx->channel_number);
drivers/net/ethernet/microchip/lan743x_main.c
1990
(tx->ts_flags & TX_TS_FLAG_TIMESTAMPING_ENABLED) &&
drivers/net/ethernet/microchip/lan743x_main.c
1991
(lan743x_ptp_request_tx_timestamp(tx->adapter))) {
drivers/net/ethernet/microchip/lan743x_main.c
1994
if (tx->ts_flags & TX_TS_FLAG_ONE_STEP_SYNC)
drivers/net/ethernet/microchip/lan743x_main.c
2007
if (lan743x_tx_frame_start(tx,
drivers/net/ethernet/microchip/lan743x_main.c
2015
tx->frame_count++;
drivers/net/ethernet/microchip/lan743x_main.c
2018
lan743x_tx_frame_add_lso(tx, frame_length, nr_frags);
drivers/net/ethernet/microchip/lan743x_main.c
2026
if (lan743x_tx_frame_add_fragment(tx, frag, frame_length)) {
drivers/net/ethernet/microchip/lan743x_main.c
2038
lan743x_tx_frame_end(tx, skb, do_timestamp, ignore_sync);
drivers/net/ethernet/microchip/lan743x_main.c
2041
spin_unlock_irqrestore(&tx->ring_lock, irq_flags);
drivers/net/ethernet/microchip/lan743x_main.c
2047
struct lan743x_tx *tx = container_of(napi, struct lan743x_tx, napi);
drivers/net/ethernet/microchip/lan743x_main.c
2048
struct lan743x_adapter *adapter = tx->adapter;
drivers/net/ethernet/microchip/lan743x_main.c
2053
ioc_bit = DMAC_INT_BIT_TX_IOC_(tx->channel_number);
drivers/net/ethernet/microchip/lan743x_main.c
2055
if (tx->vector_flags & LAN743X_VECTOR_FLAG_SOURCE_STATUS_W2C)
drivers/net/ethernet/microchip/lan743x_main.c
2057
spin_lock_irqsave(&tx->ring_lock, irq_flags);
drivers/net/ethernet/microchip/lan743x_main.c
2060
lan743x_tx_release_completed_descriptors(tx);
drivers/net/ethernet/microchip/lan743x_main.c
2061
txq = netdev_get_tx_queue(adapter->netdev, tx->channel_number);
drivers/net/ethernet/microchip/lan743x_main.c
2063
if (tx->rqd_descriptors) {
drivers/net/ethernet/microchip/lan743x_main.c
2064
if (tx->rqd_descriptors <=
drivers/net/ethernet/microchip/lan743x_main.c
2065
lan743x_tx_get_avail_desc(tx)) {
drivers/net/ethernet/microchip/lan743x_main.c
2066
tx->rqd_descriptors = 0;
drivers/net/ethernet/microchip/lan743x_main.c
2073
spin_unlock_irqrestore(&tx->ring_lock, irq_flags);
drivers/net/ethernet/microchip/lan743x_main.c
2080
INT_BIT_DMA_TX_(tx->channel_number));
drivers/net/ethernet/microchip/lan743x_main.c
2087
static void lan743x_tx_ring_cleanup(struct lan743x_tx *tx)
drivers/net/ethernet/microchip/lan743x_main.c
2089
if (tx->head_cpu_ptr) {
drivers/net/ethernet/microchip/lan743x_main.c
2090
dma_free_coherent(&tx->adapter->pdev->dev,
drivers/net/ethernet/microchip/lan743x_main.c
2091
sizeof(*tx->head_cpu_ptr), tx->head_cpu_ptr,
drivers/net/ethernet/microchip/lan743x_main.c
2092
tx->head_dma_ptr);
drivers/net/ethernet/microchip/lan743x_main.c
2093
tx->head_cpu_ptr = NULL;
drivers/net/ethernet/microchip/lan743x_main.c
2094
tx->head_dma_ptr = 0;
drivers/net/ethernet/microchip/lan743x_main.c
2096
kfree(tx->buffer_info);
drivers/net/ethernet/microchip/lan743x_main.c
2097
tx->buffer_info = NULL;
drivers/net/ethernet/microchip/lan743x_main.c
2099
if (tx->ring_cpu_ptr) {
drivers/net/ethernet/microchip/lan743x_main.c
2100
dma_free_coherent(&tx->adapter->pdev->dev,
drivers/net/ethernet/microchip/lan743x_main.c
2101
tx->ring_allocation_size, tx->ring_cpu_ptr,
drivers/net/ethernet/microchip/lan743x_main.c
2102
tx->ring_dma_ptr);
drivers/net/ethernet/microchip/lan743x_main.c
2103
tx->ring_allocation_size = 0;
drivers/net/ethernet/microchip/lan743x_main.c
2104
tx->ring_cpu_ptr = NULL;
drivers/net/ethernet/microchip/lan743x_main.c
2105
tx->ring_dma_ptr = 0;
drivers/net/ethernet/microchip/lan743x_main.c
2107
tx->ring_size = 0;
drivers/net/ethernet/microchip/lan743x_main.c
2110
static int lan743x_tx_ring_init(struct lan743x_tx *tx)
drivers/net/ethernet/microchip/lan743x_main.c
2117
tx->ring_size = LAN743X_TX_RING_SIZE;
drivers/net/ethernet/microchip/lan743x_main.c
2118
if (tx->ring_size & ~TX_CFG_B_TX_RING_LEN_MASK_) {
drivers/net/ethernet/microchip/lan743x_main.c
2122
if (dma_set_mask_and_coherent(&tx->adapter->pdev->dev,
drivers/net/ethernet/microchip/lan743x_main.c
2124
dev_warn(&tx->adapter->pdev->dev,
drivers/net/ethernet/microchip/lan743x_main.c
2129
ring_allocation_size = ALIGN(tx->ring_size *
drivers/net/ethernet/microchip/lan743x_main.c
2133
cpu_ptr = dma_alloc_coherent(&tx->adapter->pdev->dev,
drivers/net/ethernet/microchip/lan743x_main.c
2140
tx->ring_allocation_size = ring_allocation_size;
drivers/net/ethernet/microchip/lan743x_main.c
2141
tx->ring_cpu_ptr = (struct lan743x_tx_descriptor *)cpu_ptr;
drivers/net/ethernet/microchip/lan743x_main.c
2142
tx->ring_dma_ptr = dma_ptr;
drivers/net/ethernet/microchip/lan743x_main.c
2144
cpu_ptr = kzalloc_objs(*tx->buffer_info, tx->ring_size);
drivers/net/ethernet/microchip/lan743x_main.c
2149
tx->buffer_info = (struct lan743x_tx_buffer_info *)cpu_ptr;
drivers/net/ethernet/microchip/lan743x_main.c
2151
cpu_ptr = dma_alloc_coherent(&tx->adapter->pdev->dev,
drivers/net/ethernet/microchip/lan743x_main.c
2152
sizeof(*tx->head_cpu_ptr), &dma_ptr,
drivers/net/ethernet/microchip/lan743x_main.c
2159
tx->head_cpu_ptr = cpu_ptr;
drivers/net/ethernet/microchip/lan743x_main.c
2160
tx->head_dma_ptr = dma_ptr;
drivers/net/ethernet/microchip/lan743x_main.c
2161
if (tx->head_dma_ptr & 0x3) {
drivers/net/ethernet/microchip/lan743x_main.c
2169
lan743x_tx_ring_cleanup(tx);
drivers/net/ethernet/microchip/lan743x_main.c
2173
static void lan743x_tx_close(struct lan743x_tx *tx)
drivers/net/ethernet/microchip/lan743x_main.c
2175
struct lan743x_adapter *adapter = tx->adapter;
drivers/net/ethernet/microchip/lan743x_main.c
2179
DMAC_CMD_STOP_T_(tx->channel_number));
drivers/net/ethernet/microchip/lan743x_main.c
2180
lan743x_dmac_tx_wait_till_stopped(adapter, tx->channel_number);
drivers/net/ethernet/microchip/lan743x_main.c
2184
DMAC_INT_BIT_TX_IOC_(tx->channel_number));
drivers/net/ethernet/microchip/lan743x_main.c
2186
INT_BIT_DMA_TX_(tx->channel_number));
drivers/net/ethernet/microchip/lan743x_main.c
2187
napi_disable(&tx->napi);
drivers/net/ethernet/microchip/lan743x_main.c
2188
netif_napi_del(&tx->napi);
drivers/net/ethernet/microchip/lan743x_main.c
2191
FCT_TX_CTL_DIS_(tx->channel_number));
drivers/net/ethernet/microchip/lan743x_main.c
2193
FCT_TX_CTL_EN_(tx->channel_number),
drivers/net/ethernet/microchip/lan743x_main.c
2196
lan743x_tx_release_all_descriptors(tx);
drivers/net/ethernet/microchip/lan743x_main.c
2198
tx->rqd_descriptors = 0;
drivers/net/ethernet/microchip/lan743x_main.c
2200
lan743x_tx_ring_cleanup(tx);
drivers/net/ethernet/microchip/lan743x_main.c
2203
static int lan743x_tx_open(struct lan743x_tx *tx)
drivers/net/ethernet/microchip/lan743x_main.c
2209
adapter = tx->adapter;
drivers/net/ethernet/microchip/lan743x_main.c
221
struct lan743x_tx *tx = context;
drivers/net/ethernet/microchip/lan743x_main.c
2210
ret = lan743x_tx_ring_init(tx);
drivers/net/ethernet/microchip/lan743x_main.c
2216
FCT_TX_CTL_RESET_(tx->channel_number));
drivers/net/ethernet/microchip/lan743x_main.c
2218
FCT_TX_CTL_RESET_(tx->channel_number),
drivers/net/ethernet/microchip/lan743x_main.c
222
struct lan743x_adapter *adapter = tx->adapter;
drivers/net/ethernet/microchip/lan743x_main.c
2223
FCT_TX_CTL_EN_(tx->channel_number));
drivers/net/ethernet/microchip/lan743x_main.c
2227
DMAC_CMD_TX_SWR_(tx->channel_number));
drivers/net/ethernet/microchip/lan743x_main.c
2229
DMAC_CMD_TX_SWR_(tx->channel_number),
drivers/net/ethernet/microchip/lan743x_main.c
2234
TX_BASE_ADDRH(tx->channel_number),
drivers/net/ethernet/microchip/lan743x_main.c
2235
DMA_ADDR_HIGH32(tx->ring_dma_ptr));
drivers/net/ethernet/microchip/lan743x_main.c
2237
TX_BASE_ADDRL(tx->channel_number),
drivers/net/ethernet/microchip/lan743x_main.c
2238
DMA_ADDR_LOW32(tx->ring_dma_ptr));
drivers/net/ethernet/microchip/lan743x_main.c
2241
data = lan743x_csr_read(adapter, TX_CFG_B(tx->channel_number));
drivers/net/ethernet/microchip/lan743x_main.c
2243
data |= ((tx->ring_size) & TX_CFG_B_TX_RING_LEN_MASK_);
drivers/net/ethernet/microchip/lan743x_main.c
2246
lan743x_csr_write(adapter, TX_CFG_B(tx->channel_number), data);
drivers/net/ethernet/microchip/lan743x_main.c
2256
lan743x_csr_write(adapter, TX_CFG_A(tx->channel_number), data);
drivers/net/ethernet/microchip/lan743x_main.c
2260
TX_HEAD_WRITEBACK_ADDRH(tx->channel_number),
drivers/net/ethernet/microchip/lan743x_main.c
2261
DMA_ADDR_HIGH32(tx->head_dma_ptr));
drivers/net/ethernet/microchip/lan743x_main.c
2263
TX_HEAD_WRITEBACK_ADDRL(tx->channel_number),
drivers/net/ethernet/microchip/lan743x_main.c
2264
DMA_ADDR_LOW32(tx->head_dma_ptr));
drivers/net/ethernet/microchip/lan743x_main.c
2267
tx->last_head = lan743x_csr_read(adapter, TX_HEAD(tx->channel_number));
drivers/net/ethernet/microchip/lan743x_main.c
2270
tx->last_tail = 0;
drivers/net/ethernet/microchip/lan743x_main.c
2271
lan743x_csr_write(adapter, TX_TAIL(tx->channel_number),
drivers/net/ethernet/microchip/lan743x_main.c
2272
(u32)(tx->last_tail));
drivers/net/ethernet/microchip/lan743x_main.c
2273
tx->vector_flags = lan743x_intr_get_vector_flags(adapter,
drivers/net/ethernet/microchip/lan743x_main.c
2275
(tx->channel_number));
drivers/net/ethernet/microchip/lan743x_main.c
2277
&tx->napi, lan743x_tx_napi_poll,
drivers/net/ethernet/microchip/lan743x_main.c
2279
napi_enable(&tx->napi);
drivers/net/ethernet/microchip/lan743x_main.c
228
INT_BIT_DMA_TX_(tx->channel_number));
drivers/net/ethernet/microchip/lan743x_main.c
2282
if (tx->vector_flags & LAN743X_VECTOR_FLAG_SOURCE_ENABLE_AUTO_CLEAR)
drivers/net/ethernet/microchip/lan743x_main.c
2284
if (tx->vector_flags & LAN743X_VECTOR_FLAG_SOURCE_STATUS_AUTO_CLEAR)
drivers/net/ethernet/microchip/lan743x_main.c
2286
if (tx->vector_flags & LAN743X_VECTOR_FLAG_SOURCE_STATUS_R2C)
drivers/net/ethernet/microchip/lan743x_main.c
2288
if (tx->vector_flags & LAN743X_VECTOR_FLAG_SOURCE_ENABLE_R2C)
drivers/net/ethernet/microchip/lan743x_main.c
2290
lan743x_csr_write(adapter, TX_CFG_C(tx->channel_number), data);
drivers/net/ethernet/microchip/lan743x_main.c
2292
if (!(tx->vector_flags & LAN743X_VECTOR_FLAG_SOURCE_ENABLE_AUTO_SET))
drivers/net/ethernet/microchip/lan743x_main.c
2294
INT_BIT_DMA_TX_(tx->channel_number));
drivers/net/ethernet/microchip/lan743x_main.c
2296
DMAC_INT_BIT_TX_IOC_(tx->channel_number));
drivers/net/ethernet/microchip/lan743x_main.c
2300
DMAC_CMD_START_T_(tx->channel_number));
drivers/net/ethernet/microchip/lan743x_main.c
231
if (int_sts & INT_BIT_DMA_TX_(tx->channel_number)) {
drivers/net/ethernet/microchip/lan743x_main.c
232
u32 ioc_bit = DMAC_INT_BIT_TX_IOC_(tx->channel_number);
drivers/net/ethernet/microchip/lan743x_main.c
249
napi_schedule(&tx->napi);
drivers/net/ethernet/microchip/lan743x_main.c
257
INT_BIT_DMA_TX_(tx->channel_number));
drivers/net/ethernet/microchip/lan743x_main.c
324
lan743x_tx_isr(&adapter->tx[channel],
drivers/net/ethernet/microchip/lan743x_main.c
3241
lan743x_tx_close(&adapter->tx[index]);
drivers/net/ethernet/microchip/lan743x_main.c
3288
ret = lan743x_tx_open(&adapter->tx[index]);
drivers/net/ethernet/microchip/lan743x_main.c
3310
if (adapter->tx[index].ring_cpu_ptr)
drivers/net/ethernet/microchip/lan743x_main.c
3311
lan743x_tx_close(&adapter->tx[index]);
drivers/net/ethernet/microchip/lan743x_main.c
3344
return lan743x_tx_xmit_frame(&adapter->tx[ch], skb);
drivers/net/ethernet/microchip/lan743x_main.c
3496
struct lan743x_tx *tx;
drivers/net/ethernet/microchip/lan743x_main.c
3552
tx = &adapter->tx[index];
drivers/net/ethernet/microchip/lan743x_main.c
3553
tx->adapter = adapter;
drivers/net/ethernet/microchip/lan743x_main.c
3554
tx->channel_number = index;
drivers/net/ethernet/microchip/lan743x_main.c
3555
spin_lock_init(&tx->ring_lock);
drivers/net/ethernet/microchip/lan743x_main.c
682
&adapter->tx[index]);
drivers/net/ethernet/microchip/lan743x_main.h
1069
struct lan743x_tx tx[PCI11X1X_USED_TX_CHANNELS];
drivers/net/ethernet/microchip/lan743x_main.h
997
void lan743x_tx_set_timestamping_mode(struct lan743x_tx *tx,
drivers/net/ethernet/microchip/lan743x_ptp.c
1743
struct lan743x_tx *tx = &adapter->tx[0];
drivers/net/ethernet/microchip/lan743x_ptp.c
1745
if (tx->ts_flags & TX_TS_FLAG_ONE_STEP_SYNC)
drivers/net/ethernet/microchip/lan743x_ptp.c
1747
else if (tx->ts_flags & TX_TS_FLAG_TIMESTAMPING_ENABLED)
drivers/net/ethernet/microchip/lan743x_ptp.c
1768
lan743x_tx_set_timestamping_mode(&adapter->tx[index],
drivers/net/ethernet/microchip/lan743x_ptp.c
1775
lan743x_tx_set_timestamping_mode(&adapter->tx[index],
drivers/net/ethernet/microchip/lan743x_ptp.c
1782
lan743x_tx_set_timestamping_mode(&adapter->tx[index],
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
197
static int lan966x_fdma_tx_alloc(struct lan966x_tx *tx)
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
199
struct lan966x *lan966x = tx->lan966x;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
200
struct fdma *fdma = &tx->fdma;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
203
tx->dcbs_buf = kzalloc_objs(struct lan966x_tx_dcb_buf, fdma->n_dcbs);
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
204
if (!tx->dcbs_buf)
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
216
kfree(tx->dcbs_buf);
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
220
static void lan966x_fdma_tx_free(struct lan966x_tx *tx)
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
222
struct lan966x *lan966x = tx->lan966x;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
224
kfree(tx->dcbs_buf);
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
225
fdma_free_coherent(lan966x->dev, &tx->fdma);
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
228
static void lan966x_fdma_tx_activate(struct lan966x_tx *tx)
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
230
struct lan966x *lan966x = tx->lan966x;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
231
struct fdma *fdma = &tx->fdma;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
267
static void lan966x_fdma_tx_disable(struct lan966x_tx *tx)
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
269
struct lan966x *lan966x = tx->lan966x;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
270
struct fdma *fdma = &tx->fdma;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
286
tx->activated = false;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
289
static void lan966x_fdma_tx_reload(struct lan966x_tx *tx)
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
291
struct lan966x *lan966x = tx->lan966x;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
294
lan_rmw(FDMA_CH_RELOAD_CH_RELOAD_SET(BIT(tx->fdma.channel_id)),
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
31
*dataptr = lan966x->tx.dcbs_buf[dcb].dma_addr;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
330
struct lan966x_tx *tx = &lan966x->tx;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
333
struct fdma *fdma = &tx->fdma;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
344
dcb_buf = &tx->dcbs_buf[i];
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
41
*dataptr = lan966x->tx.dcbs_buf[dcb].dma_addr + XDP_PACKET_HEADROOM;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
570
static int lan966x_fdma_get_next_dcb(struct lan966x_tx *tx)
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
573
struct fdma *fdma = &tx->fdma;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
577
dcb_buf = &tx->dcbs_buf[i];
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
579
!fdma_is_last(&tx->fdma, &tx->fdma.dcbs[i]))
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
586
static void lan966x_fdma_tx_start(struct lan966x_tx *tx)
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
588
struct lan966x *lan966x = tx->lan966x;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
590
if (likely(lan966x->tx.activated)) {
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
591
lan966x_fdma_tx_reload(tx);
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
594
lan966x->tx.activated = true;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
595
lan966x_fdma_tx_activate(tx);
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
603
struct lan966x_tx *tx = &lan966x->tx;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
614
next_to_use = lan966x_fdma_get_next_dcb(tx);
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
622
next_dcb_buf = &tx->dcbs_buf[next_to_use];
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
675
__fdma_dcb_add(&tx->fdma,
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
687
lan966x_fdma_tx_start(tx);
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
700
struct lan966x_tx *tx = &lan966x->tx;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
708
next_to_use = lan966x_fdma_get_next_dcb(tx);
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
746
next_dcb_buf = &tx->dcbs_buf[next_to_use];
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
756
fdma_dcb_add(&tx->fdma,
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
770
lan966x_fdma_tx_start(tx);
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
941
lan966x->tx.lan966x = lan966x;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
942
lan966x->tx.fdma.channel_id = FDMA_INJ_CHANNEL;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
943
lan966x->tx.fdma.n_dcbs = FDMA_DCB_MAX;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
944
lan966x->tx.fdma.n_dbs = FDMA_TX_DCB_MAX_DBS;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
945
lan966x->tx.fdma.priv = lan966x;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
946
lan966x->tx.fdma.size = fdma_get_size(&lan966x->tx.fdma);
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
947
lan966x->tx.fdma.db_size = PAGE_SIZE << lan966x->rx.page_order;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
948
lan966x->tx.fdma.ops.nextptr_cb = &fdma_nextptr_cb;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
949
lan966x->tx.fdma.ops.dataptr_cb = &lan966x_fdma_tx_dataptr_cb;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
955
err = lan966x_fdma_tx_alloc(&lan966x->tx);
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
972
lan966x_fdma_tx_disable(&lan966x->tx);
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
980
lan966x_fdma_tx_free(&lan966x->tx);
drivers/net/ethernet/microchip/lan966x/lan966x_main.h
331
struct lan966x_tx tx;
drivers/net/ethernet/microchip/sparx5/lan969x/lan969x_fdma.c
185
struct sparx5_tx *tx = &sparx5->tx;
drivers/net/ethernet/microchip/sparx5/lan969x/lan969x_fdma.c
186
struct fdma *fdma = &tx->fdma;
drivers/net/ethernet/microchip/sparx5/lan969x/lan969x_fdma.c
189
tx->dbs = kzalloc_objs(struct sparx5_tx_buf, fdma->n_dcbs);
drivers/net/ethernet/microchip/sparx5/lan969x/lan969x_fdma.c
190
if (!tx->dbs)
drivers/net/ethernet/microchip/sparx5/lan969x/lan969x_fdma.c
195
kfree(tx->dbs);
drivers/net/ethernet/microchip/sparx5/lan969x/lan969x_fdma.c
20
*dataptr = FDMA_PRIV(fdma)->tx.dbs[dcb].dma_addr;
drivers/net/ethernet/microchip/sparx5/lan969x/lan969x_fdma.c
232
struct fdma *fdma = &sparx5->tx.fdma;
drivers/net/ethernet/microchip/sparx5/lan969x/lan969x_fdma.c
302
struct sparx5_tx *tx = &sparx5->tx;
drivers/net/ethernet/microchip/sparx5/lan969x/lan969x_fdma.c
303
struct fdma *fdma = &tx->fdma;
drivers/net/ethernet/microchip/sparx5/lan969x/lan969x_fdma.c
307
next_dcb = lan969x_fdma_get_next_dcb(tx);
drivers/net/ethernet/microchip/sparx5/lan969x/lan969x_fdma.c
324
db_buf = &tx->dbs[next_dcb];
drivers/net/ethernet/microchip/sparx5/lan969x/lan969x_fdma.c
395
struct sparx5_tx *tx = &sparx5->tx;
drivers/net/ethernet/microchip/sparx5/lan969x/lan969x_fdma.c
398
fdma_free_coherent(sparx5->dev, &tx->fdma);
drivers/net/ethernet/microchip/sparx5/lan969x/lan969x_fdma.c
42
static int lan969x_fdma_get_next_dcb(struct sparx5_tx *tx)
drivers/net/ethernet/microchip/sparx5/lan969x/lan969x_fdma.c
44
struct fdma *fdma = &tx->fdma;
drivers/net/ethernet/microchip/sparx5/lan969x/lan969x_fdma.c
47
if (!tx->dbs[i].used && !fdma_is_last(fdma, &fdma->dcbs[i]))
drivers/net/ethernet/microchip/sparx5/lan969x/lan969x_fdma.c
55
struct fdma *fdma = &sparx5->tx.fdma;
drivers/net/ethernet/microchip/sparx5/lan969x/lan969x_fdma.c
63
db = &sparx5->tx.dbs[i];
drivers/net/ethernet/microchip/sparx5/lan969x/lan969x_fdma.c
73
sparx5->tx.packets++;
drivers/net/ethernet/microchip/sparx5/sparx5_fdma.c
102
static void sparx5_fdma_tx_activate(struct sparx5 *sparx5, struct sparx5_tx *tx)
drivers/net/ethernet/microchip/sparx5/sparx5_fdma.c
104
struct fdma *fdma = &tx->fdma;
drivers/net/ethernet/microchip/sparx5/sparx5_fdma.c
126
static void sparx5_fdma_tx_deactivate(struct sparx5 *sparx5, struct sparx5_tx *tx)
drivers/net/ethernet/microchip/sparx5/sparx5_fdma.c
129
spx5_rmw(0, BIT(tx->fdma.channel_id) & FDMA_CH_ACTIVATE_CH_ACTIVATE,
drivers/net/ethernet/microchip/sparx5/sparx5_fdma.c
216
struct sparx5_tx *tx = &sparx5->tx;
drivers/net/ethernet/microchip/sparx5/sparx5_fdma.c
217
struct fdma *fdma = &tx->fdma;
drivers/net/ethernet/microchip/sparx5/sparx5_fdma.c
261
struct sparx5_tx *tx = &sparx5->tx;
drivers/net/ethernet/microchip/sparx5/sparx5_fdma.c
262
struct fdma *fdma = &tx->fdma;
drivers/net/ethernet/microchip/sparx5/sparx5_fdma.c
301
struct sparx5_tx *tx, int channel)
drivers/net/ethernet/microchip/sparx5/sparx5_fdma.c
303
struct fdma *fdma = &tx->fdma;
drivers/net/ethernet/microchip/sparx5/sparx5_fdma.c
310
fdma->size = fdma_get_size_contiguous(&sparx5->tx.fdma);
drivers/net/ethernet/microchip/sparx5/sparx5_fdma.c
421
sparx5_fdma_tx_init(sparx5, &sparx5->tx, FDMA_INJ_CHANNEL);
drivers/net/ethernet/microchip/sparx5/sparx5_fdma.c
439
fdma_free_phys(&sparx5->tx.fdma);
drivers/net/ethernet/microchip/sparx5/sparx5_fdma.c
453
struct sparx5_tx *tx = &sparx5->tx;
drivers/net/ethernet/microchip/sparx5/sparx5_fdma.c
463
sparx5_fdma_tx_activate(sparx5, tx);
drivers/net/ethernet/microchip/sparx5/sparx5_fdma.c
471
struct sparx5_tx *tx = &sparx5->tx;
drivers/net/ethernet/microchip/sparx5/sparx5_fdma.c
478
sparx5_fdma_tx_deactivate(sparx5, tx);
drivers/net/ethernet/microchip/sparx5/sparx5_main.h
420
struct sparx5_tx tx;
drivers/net/ethernet/microchip/sparx5/sparx5_packet.c
278
sparx5->tx.packets++;
drivers/net/ethernet/microchip/sparx5/sparx5_packet.c
288
sparx5->tx.dropped++;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1061
ss->tx.req = 0;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1062
ss->tx.done = 0;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1063
ss->tx.pkt_start = 0;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1064
ss->tx.pkt_done = 0;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1069
ss->tx.wake_queue = 0;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1070
ss->tx.stop_queue = 0;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1390
struct myri10ge_tx_buf *tx = &ss->tx;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1395
while (tx->pkt_done != mcp_index) {
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1396
idx = tx->done & tx->mask;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1397
skb = tx->info[idx].skb;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1400
tx->info[idx].skb = NULL;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1401
if (tx->info[idx].last) {
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1402
tx->pkt_done++;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1403
tx->info[idx].last = 0;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1405
tx->done++;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1406
len = dma_unmap_len(&tx->info[idx], len);
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1407
dma_unmap_len_set(&tx->info[idx], len, 0);
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1414
dma_unmap_addr(&tx->info[idx],
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1420
dma_unmap_addr(&tx->info[idx],
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1438
if (tx->req == tx->done) {
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1439
tx->queue_active = 0;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1440
put_be32(htonl(1), tx->send_stop);
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1448
tx->req - tx->done < (tx->mask >> 1) &&
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1450
tx->wake_queue++;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1555
struct myri10ge_tx_buf *tx = &ss->tx;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1589
if (send_done_count != tx->pkt_done)
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1726
ring->tx_max_pending = mgp->ss[0].tx.mask + 1;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
174
struct myri10ge_tx_buf tx; /* transmit ring */
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1850
data[i++] = (unsigned int)ss->tx.pkt_start;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1851
data[i++] = (unsigned int)ss->tx.pkt_done;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1852
data[i++] = (unsigned int)ss->tx.req;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1853
data[i++] = (unsigned int)ss->tx.done;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1856
data[i++] = (unsigned int)ss->tx.wake_queue;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1857
data[i++] = (unsigned int)ss->tx.stop_queue;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1858
data[i++] = (unsigned int)ss->tx.linearized;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1971
ss->tx.mask = tx_ring_entries - 1;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1979
* sizeof(*ss->tx.req_list);
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1980
ss->tx.req_bytes = kzalloc(bytes, GFP_KERNEL);
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1981
if (ss->tx.req_bytes == NULL)
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1985
ss->tx.req_list = (struct mcp_kreq_ether_send *)
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1986
ALIGN((unsigned long)ss->tx.req_bytes, 8);
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1987
ss->tx.queue_active = 0;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2001
bytes = tx_ring_entries * sizeof(*ss->tx.info);
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2002
ss->tx.info = kzalloc(bytes, GFP_KERNEL);
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2003
if (ss->tx.info == NULL)
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2071
kfree(ss->tx.info);
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2080
kfree(ss->tx.req_bytes);
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2081
ss->tx.req_bytes = NULL;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2082
ss->tx.req_list = NULL;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2092
struct myri10ge_tx_buf *tx;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2096
if (ss->tx.req_list == NULL)
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2119
tx = &ss->tx;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2120
while (tx->done != tx->req) {
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2121
idx = tx->done & tx->mask;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2122
skb = tx->info[idx].skb;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2125
tx->info[idx].skb = NULL;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2126
tx->done++;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2127
len = dma_unmap_len(&tx->info[idx], len);
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2128
dma_unmap_len_set(&tx->info[idx], len, 0);
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2134
dma_unmap_addr(&tx->info[idx],
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2140
dma_unmap_addr(&tx->info[idx],
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2149
kfree(ss->tx.info);
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2155
kfree(ss->tx.req_bytes);
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2156
ss->tx.req_bytes = NULL;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2157
ss->tx.req_list = NULL;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2257
ss->tx.lanai = (struct mcp_kreq_ether_send __iomem *)
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2273
ss->tx.send_go = (__iomem __be32 *)
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2275
ss->tx.send_stop = (__iomem __be32 *)
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2500
if (mgp->ss[0].tx.req_bytes == NULL)
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2542
myri10ge_submit_req_backwards(struct myri10ge_tx_buf *tx,
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2546
starting_slot = tx->req;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2549
idx = (starting_slot + cnt) & tx->mask;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2550
myri10ge_pio_copy(&tx->lanai[idx], &src[cnt], sizeof(*src));
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2563
myri10ge_submit_req(struct myri10ge_tx_buf *tx, struct mcp_kreq_ether_send *src,
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2571
idx = tx->req & tx->mask;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2576
dst = dstp = &tx->lanai[idx];
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2579
if ((idx + cnt) < tx->mask) {
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2589
myri10ge_submit_req_backwards(tx, src, cnt);
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2601
tx->req += cnt;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2606
struct myri10ge_tx_buf *tx, int idx)
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2612
last_idx = (idx + 1) & tx->mask;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2613
idx = tx->req & tx->mask;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2615
len = dma_unmap_len(&tx->info[idx], len);
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2617
if (tx->info[idx].skb != NULL)
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2619
dma_unmap_addr(&tx->info[idx],
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2624
dma_unmap_addr(&tx->info[idx],
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2627
dma_unmap_len_set(&tx->info[idx], len, 0);
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2628
tx->info[idx].skb = NULL;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2630
idx = (idx + 1) & tx->mask;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2650
struct myri10ge_tx_buf *tx;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2665
tx = &ss->tx;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2668
req = tx->req_list;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2669
avail = tx->mask - 1 - (tx->req - tx->done);
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2681
tx->stop_queue++;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2753
idx = tx->req & tx->mask;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2754
tx->info[idx].skb = skb;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2755
dma_unmap_addr_set(&tx->info[idx], bus, bus);
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2756
dma_unmap_len_set(&tx->info[idx], len, len);
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2859
myri10ge_unmap_tx_dma(mgp, tx, idx);
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2862
idx = (count + tx->req) & tx->mask;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2863
dma_unmap_addr_set(&tx->info[idx], bus, bus);
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2864
dma_unmap_len_set(&tx->info[idx], len, len);
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2874
idx = ((count - 1) + tx->req) & tx->mask;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2875
tx->info[idx].last = 1;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2876
myri10ge_submit_req(tx, tx->req_list, count);
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2879
if ((mgp->dev->real_num_tx_queues > 1) && tx->queue_active == 0) {
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2880
tx->queue_active = 1;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2881
put_be32(htonl(1), tx->send_go);
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2884
tx->pkt_start++;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2886
tx->stop_queue++;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2892
myri10ge_unmap_tx_dma(mgp, tx, idx);
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
2902
tx->linearized++;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
3374
if (ss->tx.req != ss->tx.done &&
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
3375
ss->tx.done == ss->watchdog_tx_done &&
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
3385
slice, ss->tx.queue_active, ss->tx.req,
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
3386
ss->tx.done, ss->tx.pkt_start,
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
3387
ss->tx.pkt_done,
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
3394
if (ss->watchdog_tx_done != ss->tx.done ||
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
3398
ss->watchdog_tx_done = ss->tx.done;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
3399
ss->watchdog_tx_req = ss->tx.req;
drivers/net/ethernet/netronome/nfp/nfp_net_debugfs.c
131
struct dentry *queues, *tx, *rx, *xdp;
drivers/net/ethernet/netronome/nfp/nfp_net_debugfs.c
148
tx = debugfs_create_dir("tx", queues);
drivers/net/ethernet/netronome/nfp/nfp_net_debugfs.c
161
debugfs_create_file(name, 0400, tx,
drivers/net/ethernet/pasemi/pasemi_mac.c
1077
mac->tx = pasemi_mac_setup_tx_resources(dev);
drivers/net/ethernet/pasemi/pasemi_mac.c
1079
if (!mac->tx) {
drivers/net/ethernet/pasemi/pasemi_mac.c
1106
write_iob_reg(PAS_IOB_DMA_TXCH_CFG(mac->tx->chan.chno),
drivers/net/ethernet/pasemi/pasemi_mac.c
112
return mac->tx;
drivers/net/ethernet/pasemi/pasemi_mac.c
1171
ret = request_irq(mac->tx->chan.irq, pasemi_mac_tx_intr, 0,
drivers/net/ethernet/pasemi/pasemi_mac.c
1172
mac->tx_irq_name, mac->tx);
drivers/net/ethernet/pasemi/pasemi_mac.c
1175
mac->tx->chan.irq, ret);
drivers/net/ethernet/pasemi/pasemi_mac.c
1193
timer_setup(&mac->tx->clean_timer, pasemi_mac_tx_timer, 0);
drivers/net/ethernet/pasemi/pasemi_mac.c
1194
mod_timer(&mac->tx->clean_timer, jiffies + HZ);
drivers/net/ethernet/pasemi/pasemi_mac.c
1199
free_irq(mac->tx->chan.irq, mac->tx);
drivers/net/ethernet/pasemi/pasemi_mac.c
1204
if (mac->tx)
drivers/net/ethernet/pasemi/pasemi_mac.c
1289
timer_delete_sync(&mac->tx->clean_timer);
drivers/net/ethernet/pasemi/pasemi_mac.c
1321
free_irq(mac->tx->chan.irq, mac->tx);
drivers/net/ethernet/pasemi/pasemi_mac.c
1585
disable_irq(mac->tx->chan.irq);
drivers/net/ethernet/pasemi/pasemi_mac.c
1586
pasemi_mac_tx_intr(mac->tx->chan.irq, mac->tx);
drivers/net/ethernet/pasemi/pasemi_mac.c
1587
enable_irq(mac->tx->chan.irq);
drivers/net/ethernet/pasemi/pasemi_mac.c
1807
pasemi_dma_free_chan(&mac->tx->chan);
drivers/net/ethernet/pasemi/pasemi_mac.h
76
struct pasemi_mac_txring *tx;
drivers/net/ethernet/pasemi/pasemi_mac.h
94
#define TX_DESC(tx, num) ((tx)->chan.ring_virt[(num) & (TX_RING_SIZE-1)])
drivers/net/ethernet/pasemi/pasemi_mac.h
95
#define TX_DESC_INFO(tx, num) ((tx)->ring_info[(num) & (TX_RING_SIZE-1)])
drivers/net/ethernet/pasemi/pasemi_mac_ethtool.c
79
ering->tx_pending = RING_USED(mac->tx)/2;
drivers/net/ethernet/qlogic/qed/qed_hsi.h
3815
u8 tx;
drivers/net/ethernet/qlogic/qed/qed_int.c
2363
u8 timer_res, u16 sb_id, bool tx)
drivers/net/ethernet/qlogic/qed/qed_int.c
2384
if (tx)
drivers/net/ethernet/qlogic/qed/qed_int.h
452
u8 timer_res, u16 sb_id, bool tx);
drivers/net/ethernet/qlogic/qed/qed_l2.c
1261
p_ramrod->filter_cmd_hdr.tx = p_filter_cmd->is_tx_filter ? 1 : 0;
drivers/net/ethernet/qlogic/qed/qed_l2.c
69
u8 rx = 0, tx = 0;
drivers/net/ethernet/qlogic/qed/qed_l2.c
72
qed_vf_get_num_txqs(p_hwfn, &tx);
drivers/net/ethernet/qlogic/qed/qed_l2.c
74
p_l2_info->queues = max_t(u8, rx, tx);
drivers/net/ethernet/qlogic/qlcnic/qlcnic.h
1504
struct __qlcnic_esw_statistics tx;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c
1318
struct qlcnic_host_tx_ring *tx, int ring)
drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c
1330
tx->producer = 0;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c
1331
tx->sw_consumer = 0;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c
1332
*(tx->hw_consumer) = 0;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c
1337
mbx.phys_addr_low = LSD(tx->phys_addr);
drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c
1338
mbx.phys_addr_high = MSD(tx->phys_addr);
drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c
1339
mbx.cnsmr_index_low = LSD(tx->hw_cons_phys_addr);
drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c
1340
mbx.cnsmr_index_high = MSD(tx->hw_cons_phys_addr);
drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c
1341
mbx.size = tx->num_desc;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c
1381
tx->crb_cmd_producer = ahw->pci_base0 + mbx_out->host_prod;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c
1382
tx->ctx_id = mbx_out->ctx_id;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c
1386
tx->crb_intr_mask = ahw->pci_base0 + intr_mask;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_hw.c
1390
tx->ctx_id, mbx_out->state);
drivers/net/ethernet/qlogic/qlcnic/qlcnic_ethtool.c
1393
QLCNIC_QUERY_TX_COUNTER, &port_stats.tx);
drivers/net/ethernet/qlogic/qlcnic/qlcnic_ethtool.c
1397
qlcnic_fill_stats(data, &port_stats.tx, QLCNIC_ESW_STATS);
drivers/net/ethernet/qlogic/qlcnic/qlcnic_sysfs.c
797
&port_stats.tx);
drivers/net/ethernet/qlogic/qlcnic/qlcnic_sysfs.c
832
&esw_stats.tx);
drivers/net/ethernet/realtek/r8169_leds.c
40
bool rx, tx;
drivers/net/ethernet/realtek/r8169_leds.c
48
tx = flags & BIT(TRIGGER_NETDEV_TX);
drivers/net/ethernet/realtek/r8169_leds.c
50
return rx == tx;
drivers/net/ethernet/renesas/rswitch_main.c
203
static bool rswitch_is_any_data_irq(struct rswitch_private *priv, u32 *dis, bool tx)
drivers/net/ethernet/renesas/rswitch_main.c
205
u32 *mask = tx ? priv->gwca.tx_irq_bits : priv->gwca.rx_irq_bits;
drivers/net/ethernet/sfc/efx.c
647
struct netdev_queue_stats_tx *tx)
drivers/net/ethernet/sfc/efx.c
658
tx->packets = 0;
drivers/net/ethernet/sfc/efx.c
659
tx->bytes = 0;
drivers/net/ethernet/sfc/efx.c
660
tx->hw_gso_packets = 0;
drivers/net/ethernet/sfc/efx.c
661
tx->hw_gso_wire_packets = 0;
drivers/net/ethernet/sfc/efx.c
683
tx->packets += tx_queue->complete_packets;
drivers/net/ethernet/sfc/efx.c
684
tx->bytes += tx_queue->complete_bytes;
drivers/net/ethernet/sfc/efx.c
685
tx->hw_gso_packets += tx_queue->tso_bursts;
drivers/net/ethernet/sfc/efx.c
686
tx->hw_gso_wire_packets += tx_queue->tso_packets;
drivers/net/ethernet/sfc/efx.c
688
tx->packets += tx_queue->old_complete_packets;
drivers/net/ethernet/sfc/efx.c
689
tx->bytes += tx_queue->old_complete_bytes;
drivers/net/ethernet/sfc/efx.c
690
tx->hw_gso_packets += tx_queue->old_tso_bursts;
drivers/net/ethernet/sfc/efx.c
691
tx->hw_gso_wire_packets += tx_queue->old_tso_packets;
drivers/net/ethernet/sfc/efx.c
694
tx->packets += tx_queue->complete_xdp_packets;
drivers/net/ethernet/sfc/efx.c
695
tx->bytes += tx_queue->complete_xdp_bytes;
drivers/net/ethernet/smsc/smc91x.h
272
struct dma_async_tx_descriptor *tx;
drivers/net/ethernet/smsc/smc91x.h
278
tx = dmaengine_prep_slave_single(lp->dma_chan, dmabuf, len,
drivers/net/ethernet/smsc/smc91x.h
280
if (tx) {
drivers/net/ethernet/smsc/smc91x.h
281
cookie = dmaengine_submit(tx);
drivers/net/ethernet/socionext/sni_ave.c
1276
priv->tx.desc = kzalloc_objs(*priv->tx.desc, priv->tx.ndesc);
drivers/net/ethernet/socionext/sni_ave.c
1277
if (!priv->tx.desc) {
drivers/net/ethernet/socionext/sni_ave.c
1284
kfree(priv->tx.desc);
drivers/net/ethernet/socionext/sni_ave.c
1290
priv->tx.proc_idx = 0;
drivers/net/ethernet/socionext/sni_ave.c
1291
priv->tx.done_idx = 0;
drivers/net/ethernet/socionext/sni_ave.c
1292
for (entry = 0; entry < priv->tx.ndesc; entry++) {
drivers/net/ethernet/socionext/sni_ave.c
1297
(((priv->tx.ndesc * priv->desc_size) << 16) & AVE_TXDC_SIZE),
drivers/net/ethernet/socionext/sni_ave.c
1367
for (entry = 0; entry < priv->tx.ndesc; entry++) {
drivers/net/ethernet/socionext/sni_ave.c
1368
if (!priv->tx.desc[entry].skbs)
drivers/net/ethernet/socionext/sni_ave.c
1371
ave_dma_unmap(ndev, &priv->tx.desc[entry], DMA_TO_DEVICE);
drivers/net/ethernet/socionext/sni_ave.c
1372
dev_kfree_skb_any(priv->tx.desc[entry].skbs);
drivers/net/ethernet/socionext/sni_ave.c
1373
priv->tx.desc[entry].skbs = NULL;
drivers/net/ethernet/socionext/sni_ave.c
1375
priv->tx.proc_idx = 0;
drivers/net/ethernet/socionext/sni_ave.c
1376
priv->tx.done_idx = 0;
drivers/net/ethernet/socionext/sni_ave.c
1390
kfree(priv->tx.desc);
drivers/net/ethernet/socionext/sni_ave.c
1403
proc_idx = priv->tx.proc_idx;
drivers/net/ethernet/socionext/sni_ave.c
1404
done_idx = priv->tx.done_idx;
drivers/net/ethernet/socionext/sni_ave.c
1405
ndesc = priv->tx.ndesc;
drivers/net/ethernet/socionext/sni_ave.c
1423
ret = ave_dma_map(ndev, &priv->tx.desc[proc_idx],
drivers/net/ethernet/socionext/sni_ave.c
1431
priv->tx.desc[proc_idx].skbs = skb;
drivers/net/ethernet/socionext/sni_ave.c
1449
priv->tx.proc_idx = (proc_idx + 1) % ndesc;
drivers/net/ethernet/socionext/sni_ave.c
1619
priv->tx.daddr = AVE_TXDM_64;
drivers/net/ethernet/socionext/sni_ave.c
1624
priv->tx.daddr = AVE_TXDM_32;
drivers/net/ethernet/socionext/sni_ave.c
1632
priv->tx.ndesc = AVE_NR_TXDESC;
drivers/net/ethernet/socionext/sni_ave.c
279
struct ave_desc_info tx;
drivers/net/ethernet/socionext/sni_ave.c
303
addr = ((id == AVE_DESCID_TX) ? priv->tx.daddr : priv->rx.daddr)
drivers/net/ethernet/socionext/sni_ave.c
321
addr = ((id == AVE_DESCID_TX) ? priv->tx.daddr : priv->rx.daddr)
drivers/net/ethernet/socionext/sni_ave.c
686
proc_idx = priv->tx.proc_idx;
drivers/net/ethernet/socionext/sni_ave.c
687
done_idx = priv->tx.done_idx;
drivers/net/ethernet/socionext/sni_ave.c
688
ndesc = priv->tx.ndesc;
drivers/net/ethernet/socionext/sni_ave.c
714
if (priv->tx.desc[done_idx].skbs) {
drivers/net/ethernet/socionext/sni_ave.c
715
ave_dma_unmap(ndev, &priv->tx.desc[done_idx],
drivers/net/ethernet/socionext/sni_ave.c
717
dev_consume_skb_any(priv->tx.desc[done_idx].skbs);
drivers/net/ethernet/socionext/sni_ave.c
718
priv->tx.desc[done_idx].skbs = NULL;
drivers/net/ethernet/socionext/sni_ave.c
724
priv->tx.done_idx = done_idx;
drivers/net/ethernet/stmicro/stmmac/dwmac-rk.c
249
#define DELAY_ENABLE(soc, tx, rx) \
drivers/net/ethernet/stmicro/stmmac/dwmac-rk.c
250
(((tx) ? soc##_GMAC_TXCLK_DLY_ENABLE : soc##_GMAC_TXCLK_DLY_DISABLE) | \
drivers/net/ethernet/stmicro/stmmac/dwmac-sun8i.c
350
bool rx, bool tx)
drivers/net/ethernet/stmicro/stmmac/dwmac-sun8i.c
356
if (tx)
drivers/net/ethernet/stmicro/stmmac/dwmac-sun8i.c
364
bool rx, bool tx)
drivers/net/ethernet/stmicro/stmmac/dwmac-sun8i.c
370
if (tx)
drivers/net/ethernet/stmicro/stmmac/dwmac4_dma.h
176
u32 chan, bool rx, bool tx);
drivers/net/ethernet/stmicro/stmmac/dwmac4_dma.h
178
u32 chan, bool rx, bool tx);
drivers/net/ethernet/stmicro/stmmac/dwmac4_dma.h
180
u32 chan, bool rx, bool tx);
drivers/net/ethernet/stmicro/stmmac/dwmac4_dma.h
182
u32 chan, bool rx, bool tx);
drivers/net/ethernet/stmicro/stmmac/dwmac4_lib.c
113
u32 chan, bool rx, bool tx)
drivers/net/ethernet/stmicro/stmmac/dwmac4_lib.c
120
if (tx)
drivers/net/ethernet/stmicro/stmmac/dwmac4_lib.c
127
u32 chan, bool rx, bool tx)
drivers/net/ethernet/stmicro/stmmac/dwmac4_lib.c
134
if (tx)
drivers/net/ethernet/stmicro/stmmac/dwmac4_lib.c
141
u32 chan, bool rx, bool tx)
drivers/net/ethernet/stmicro/stmmac/dwmac4_lib.c
148
if (tx)
drivers/net/ethernet/stmicro/stmmac/dwmac4_lib.c
155
u32 chan, bool rx, bool tx)
drivers/net/ethernet/stmicro/stmmac/dwmac4_lib.c
162
if (tx)
drivers/net/ethernet/stmicro/stmmac/dwmac_dma.h
148
u32 chan, bool rx, bool tx);
drivers/net/ethernet/stmicro/stmmac/dwmac_dma.h
150
u32 chan, bool rx, bool tx);
drivers/net/ethernet/stmicro/stmmac/dwmac_lib.c
42
u32 chan, bool rx, bool tx)
drivers/net/ethernet/stmicro/stmmac/dwmac_lib.c
48
if (tx)
drivers/net/ethernet/stmicro/stmmac/dwmac_lib.c
55
u32 chan, bool rx, bool tx)
drivers/net/ethernet/stmicro/stmmac/dwmac_lib.c
61
if (tx)
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c
21
u32 tx, rx;
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c
23
tx = readl(ioaddr + XGMAC_TX_CONFIG);
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c
26
writel(tx | XGMAC_CORE_INIT_TX, ioaddr + XGMAC_TX_CONFIG);
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c
53
u32 tx = readl(ioaddr + XGMAC_TX_CONFIG);
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c
57
tx |= XGMAC_CONFIG_TE;
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c
60
tx &= ~XGMAC_CONFIG_TE;
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c
64
writel(tx, ioaddr + XGMAC_TX_CONFIG);
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_dma.c
223
bool rx, bool tx)
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_dma.c
229
if (tx)
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_dma.c
237
bool rx, bool tx)
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_dma.c
243
if (tx)
drivers/net/ethernet/stmicro/stmmac/hwif.h
206
u32 chan, bool rx, bool tx);
drivers/net/ethernet/stmicro/stmmac/hwif.h
208
u32 chan, bool rx, bool tx);
drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c
448
memset(&ring->tx, 0, sizeof(ring->tx));
drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c
523
if ((tso && (pkt_info->mss != ring->tx.cur_mss)) ||
drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c
524
(vlan && (pkt_info->vlan_ctag != ring->tx.cur_vlan_ctag)))
drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c
61
memset(&desc_data->tx, 0, sizeof(desc_data->tx));
drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c
1003
ring->tx.xmit_more = 1;
drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c
699
ring->tx.xmit_more = 0;
drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c
727
if (tso && (pkt_info->mss != ring->tx.cur_mss))
drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c
732
if (vlan && (pkt_info->vlan_ctag != ring->tx.cur_vlan_ctag))
drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c
789
ring->tx.cur_mss = pkt_info->mss;
drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c
818
ring->tx.cur_vlan_ctag = pkt_info->vlan_ctag;
drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c
973
desc_data->tx.packets = pkt_info->tx_packets;
drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c
974
desc_data->tx.bytes = pkt_info->tx_bytes;
drivers/net/ethernet/synopsys/dwc-xlgmac-net.c
1079
tx_packets += desc_data->tx.packets;
drivers/net/ethernet/synopsys/dwc-xlgmac-net.c
1080
tx_bytes += desc_data->tx.bytes;
drivers/net/ethernet/synopsys/dwc-xlgmac-net.c
1096
if ((ring->tx.queue_stopped == 1) &&
drivers/net/ethernet/synopsys/dwc-xlgmac-net.c
1098
ring->tx.queue_stopped = 0;
drivers/net/ethernet/synopsys/dwc-xlgmac-net.c
134
if (skb_shinfo(skb)->gso_size != ring->tx.cur_mss) {
drivers/net/ethernet/synopsys/dwc-xlgmac-net.c
161
if (skb_vlan_tag_get(skb) != ring->tx.cur_vlan_ctag)
drivers/net/ethernet/synopsys/dwc-xlgmac-net.c
49
ring->tx.queue_stopped = 1;
drivers/net/ethernet/synopsys/dwc-xlgmac-net.c
54
if (ring->tx.xmit_more)
drivers/net/ethernet/synopsys/dwc-xlgmac.h
298
struct xlgmac_tx_desc_data tx;
drivers/net/ethernet/synopsys/dwc-xlgmac.h
352
} tx;
drivers/net/ethernet/tehuti/tn40.c
1620
struct netdev_queue_stats_tx *tx)
drivers/net/ethernet/tehuti/tn40.c
1626
tx->packets = 0;
drivers/net/ethernet/tehuti/tn40.c
1627
tx->bytes = 0;
drivers/net/ethernet/ti/netcp_core.c
1335
config.u.tx.filt_einfo = false;
drivers/net/ethernet/ti/netcp_core.c
1336
config.u.tx.filt_pswords = false;
drivers/net/ethernet/ti/netcp_core.c
1337
config.u.tx.priority = DMA_PRIO_MED_L;
drivers/net/ethernet/toshiba/ps3_gelic_net.c
1616
int tx;
drivers/net/ethernet/toshiba/ps3_gelic_net.c
1620
.tx = GELIC_LV1_VLAN_TX_ETHERNET_0,
drivers/net/ethernet/toshiba/ps3_gelic_net.c
1624
.tx = GELIC_LV1_VLAN_TX_WIRELESS,
drivers/net/ethernet/toshiba/ps3_gelic_net.c
1633
vlan_id_ix[i].tx,
drivers/net/ethernet/toshiba/ps3_gelic_net.c
1639
vlan_id_ix[i].tx, status);
drivers/net/ethernet/toshiba/ps3_gelic_net.c
1640
card->vlan[i].tx = 0;
drivers/net/ethernet/toshiba/ps3_gelic_net.c
1644
card->vlan[i].tx = (u16)v1;
drivers/net/ethernet/toshiba/ps3_gelic_net.c
1656
card->vlan[i].tx = 0;
drivers/net/ethernet/toshiba/ps3_gelic_net.c
1663
i, card->vlan[i].tx, card->vlan[i].rx);
drivers/net/ethernet/toshiba/ps3_gelic_net.c
1666
if (card->vlan[GELIC_PORT_ETHERNET_0].tx) {
drivers/net/ethernet/toshiba/ps3_gelic_net.c
1667
BUG_ON(!card->vlan[GELIC_PORT_WIRELESS].tx);
drivers/net/ethernet/toshiba/ps3_gelic_net.c
1674
card->vlan[GELIC_PORT_WIRELESS].tx = 0;
drivers/net/ethernet/toshiba/ps3_gelic_net.c
792
card->vlan[type].tx);
drivers/net/ethernet/toshiba/ps3_gelic_net.h
264
u16 tx;
drivers/net/ethernet/toshiba/ps3_gelic_wireless.c
2583
if (!card->vlan[GELIC_PORT_WIRELESS].tx)
drivers/net/ethernet/toshiba/ps3_gelic_wireless.c
2625
if (!card->vlan[GELIC_PORT_WIRELESS].tx)
drivers/net/ethernet/tundra/tsi108_eth.c
1392
int tx = data->txtail;
drivers/net/ethernet/tundra/tsi108_eth.c
1394
skb = data->txskbs[tx];
drivers/net/ethernet/tundra/tsi108_eth.c
601
int tx;
drivers/net/ethernet/tundra/tsi108_eth.c
606
tx = data->txtail;
drivers/net/ethernet/tundra/tsi108_eth.c
608
if (data->txring[tx].misc & TSI108_TX_OWN)
drivers/net/ethernet/tundra/tsi108_eth.c
611
skb = data->txskbs[tx];
drivers/net/ethernet/tundra/tsi108_eth.c
613
if (!(data->txring[tx].misc & TSI108_TX_OK))
drivers/net/ethernet/tundra/tsi108_eth.c
615
dev->name, data->txring[tx].misc);
drivers/net/ethernet/tundra/tsi108_eth.c
620
if (data->txring[tx].misc & TSI108_TX_EOF) {
drivers/net/ethernet/tundra/tsi108_eth.c
665
int tx = data->txhead;
drivers/net/ethernet/tundra/tsi108_eth.c
677
if ((tx % TSI108_TX_INT_FREQ == 0) &&
drivers/net/ethernet/tundra/tsi108_eth.c
681
data->txskbs[tx] = skb;
drivers/net/ethernet/tundra/tsi108_eth.c
684
data->txring[tx].buf0 = dma_map_single(&data->pdev->dev,
drivers/net/ethernet/tundra/tsi108_eth.c
687
data->txring[tx].len = skb_headlen(skb);
drivers/net/ethernet/tundra/tsi108_eth.c
692
data->txring[tx].buf0 =
drivers/net/ethernet/tundra/tsi108_eth.c
696
data->txring[tx].len = skb_frag_size(frag);
drivers/net/ethernet/tundra/tsi108_eth.c
710
data->txring[tx].misc = misc | TSI108_TX_OWN;
drivers/net/ethernet/via/via-velocity.c
1417
for (i = 0; i < vptr->tx.numq; i++) {
drivers/net/ethernet/via/via-velocity.c
1418
writel(vptr->tx.pool_dma[i], ®s->TDBaseLo[i]);
drivers/net/ethernet/via/via-velocity.c
1493
pool = dma_alloc_coherent(vptr->dev, tx_ring_size * vptr->tx.numq +
drivers/net/ethernet/via/via-velocity.c
1507
for (i = 0; i < vptr->tx.numq; i++) {
drivers/net/ethernet/via/via-velocity.c
1508
vptr->tx.rings[i] = pool;
drivers/net/ethernet/via/via-velocity.c
1509
vptr->tx.pool_dma[i] = pool_dma;
drivers/net/ethernet/via/via-velocity.c
1665
for (j = 0; j < vptr->tx.numq; j++) {
drivers/net/ethernet/via/via-velocity.c
1667
vptr->tx.infos[j] = kzalloc_objs(struct velocity_td_info,
drivers/net/ethernet/via/via-velocity.c
1669
if (!vptr->tx.infos[j]) {
drivers/net/ethernet/via/via-velocity.c
1671
kfree(vptr->tx.infos[j]);
drivers/net/ethernet/via/via-velocity.c
1675
vptr->tx.tail[j] = vptr->tx.curr[j] = vptr->tx.used[j] = 0;
drivers/net/ethernet/via/via-velocity.c
1689
vptr->options.numtx * sizeof(struct tx_desc) * vptr->tx.numq;
drivers/net/ethernet/via/via-velocity.c
1760
struct velocity_td_info *td_info = &(vptr->tx.infos[q][n]);
drivers/net/ethernet/via/via-velocity.c
1790
for (j = 0; j < vptr->tx.numq; j++) {
drivers/net/ethernet/via/via-velocity.c
1791
if (vptr->tx.infos[j] == NULL)
drivers/net/ethernet/via/via-velocity.c
1796
kfree(vptr->tx.infos[j]);
drivers/net/ethernet/via/via-velocity.c
1797
vptr->tx.infos[j] = NULL;
drivers/net/ethernet/via/via-velocity.c
1916
for (qnum = 0; qnum < vptr->tx.numq; qnum++) {
drivers/net/ethernet/via/via-velocity.c
1917
for (idx = vptr->tx.tail[qnum]; vptr->tx.used[qnum] > 0;
drivers/net/ethernet/via/via-velocity.c
1923
td = &(vptr->tx.rings[qnum][idx]);
drivers/net/ethernet/via/via-velocity.c
1924
tdinfo = &(vptr->tx.infos[qnum][idx]);
drivers/net/ethernet/via/via-velocity.c
1948
vptr->tx.used[qnum]--;
drivers/net/ethernet/via/via-velocity.c
1950
vptr->tx.tail[qnum] = idx;
drivers/net/ethernet/via/via-velocity.c
2304
struct tx_info tx;
drivers/net/ethernet/via/via-velocity.c
2316
tmp_vptr->tx.numq = vptr->tx.numq;
drivers/net/ethernet/via/via-velocity.c
2331
tx = vptr->tx;
drivers/net/ethernet/via/via-velocity.c
2334
vptr->tx = tmp_vptr->tx;
drivers/net/ethernet/via/via-velocity.c
2337
tmp_vptr->tx = tx;
drivers/net/ethernet/via/via-velocity.c
2560
index = vptr->tx.curr[qnum];
drivers/net/ethernet/via/via-velocity.c
2561
td_ptr = &(vptr->tx.rings[qnum][index]);
drivers/net/ethernet/via/via-velocity.c
2562
tdinfo = &(vptr->tx.infos[qnum][index]);
drivers/net/ethernet/via/via-velocity.c
2617
vptr->tx.used[qnum]++;
drivers/net/ethernet/via/via-velocity.c
2618
vptr->tx.curr[qnum] = (index + 1) % vptr->options.numtx;
drivers/net/ethernet/via/via-velocity.c
2623
td_ptr = &(vptr->tx.rings[qnum][prev]);
drivers/net/ethernet/via/via-velocity.c
2661
vptr->tx.numq = info->txqueue;
drivers/net/ethernet/via/via-velocity.c
3220
for (i = 0; i < vptr->tx.numq; i++) {
drivers/net/ethernet/via/via-velocity.c
3221
if (vptr->tx.used[i])
drivers/net/ethernet/via/via-velocity.h
1378
#define AVAIL_TD(p,q) ((p)->options.numtx-((p)->tx.used[(q)]))
drivers/net/ethernet/via/via-velocity.h
1406
} tx;
drivers/net/ethernet/wangxun/libwx/wx_ethtool.c
321
if (wx->q_vector[0]->tx.count && wx->q_vector[0]->rx.count)
drivers/net/ethernet/wangxun/libwx/wx_ethtool.c
379
if (wx->q_vector[0]->tx.count && wx->q_vector[0]->rx.count) {
drivers/net/ethernet/wangxun/libwx/wx_ethtool.c
446
if (wx->q_vector[0]->tx.count && wx->q_vector[0]->rx.count)
drivers/net/ethernet/wangxun/libwx/wx_ethtool.c
451
if (q_vector->tx.count && !q_vector->rx.count)
drivers/net/ethernet/wangxun/libwx/wx_lib.c
1693
new_itr = min(q_vector->rx.itr, q_vector->tx.itr);
drivers/net/ethernet/wangxun/libwx/wx_lib.c
1729
struct wx_ring_container *tx;
drivers/net/ethernet/wangxun/libwx/wx_lib.c
1732
tx = container_of(dim, struct wx_ring_container, dim);
drivers/net/ethernet/wangxun/libwx/wx_lib.c
1735
tx->itr = tx_moder.usec;
drivers/net/ethernet/wangxun/libwx/wx_lib.c
1737
q_vector = container_of(tx, struct wx_q_vector, tx);
drivers/net/ethernet/wangxun/libwx/wx_lib.c
1752
INIT_WORK(&q_vector->tx.dim.work, wx_tx_dim_work);
drivers/net/ethernet/wangxun/libwx/wx_lib.c
1754
q_vector->tx.dim.mode = DIM_CQ_PERIOD_MODE_START_FROM_CQE;
drivers/net/ethernet/wangxun/libwx/wx_lib.c
1769
disable_work_sync(&q_vector->tx.dim.work);
drivers/net/ethernet/wangxun/libwx/wx_lib.c
2146
wx_add_ring(ring, &q_vector->tx);
drivers/net/ethernet/wangxun/libwx/wx_lib.c
2207
wx_for_each_ring(ring, q_vector->tx)
drivers/net/ethernet/wangxun/libwx/wx_lib.c
2353
if (q_vector->rx.ring || q_vector->tx.ring) {
drivers/net/ethernet/wangxun/libwx/wx_lib.c
2378
if (!q_vector->rx.ring && !q_vector->tx.ring)
drivers/net/ethernet/wangxun/libwx/wx_lib.c
2552
wx_for_each_ring(ring, q_vector->tx)
drivers/net/ethernet/wangxun/libwx/wx_lib.c
863
q_vector->tx.total_bytes += total_bytes;
drivers/net/ethernet/wangxun/libwx/wx_lib.c
864
q_vector->tx.total_packets += total_packets;
drivers/net/ethernet/wangxun/libwx/wx_lib.c
906
q_vector->tx.total_packets,
drivers/net/ethernet/wangxun/libwx/wx_lib.c
907
q_vector->tx.total_bytes,
drivers/net/ethernet/wangxun/libwx/wx_lib.c
910
net_dim(&q_vector->tx.dim, &sample);
drivers/net/ethernet/wangxun/libwx/wx_lib.c
934
wx_for_each_ring(ring, q_vector->tx) {
drivers/net/ethernet/wangxun/libwx/wx_type.h
1116
struct wx_ring_container rx, tx;
drivers/net/ethernet/wangxun/libwx/wx_vf_common.c
85
if (q_vector->tx.ring && q_vector->rx.ring)
drivers/net/ethernet/wangxun/libwx/wx_vf_lib.c
64
wx_for_each_ring(ring, q_vector->tx)
drivers/net/ethernet/wangxun/ngbe/ngbe_main.c
269
if (q_vector->tx.ring && q_vector->rx.ring)
drivers/net/ethernet/wangxun/txgbe/txgbe_irq.c
57
if (q_vector->tx.ring && q_vector->rx.ring)
drivers/net/fddi/skfp/fplustm.c
305
smc->hw.fp.tx[QUEUE_S] = queue = &smc->hw.fp.tx_q[QUEUE_S] ;
drivers/net/fddi/skfp/fplustm.c
316
smc->hw.fp.tx[QUEUE_A0] = queue = &smc->hw.fp.tx_q[QUEUE_A0] ;
drivers/net/fddi/skfp/h/fplustm.h
185
struct s_smt_tx_queue *tx[USED_QUEUES] ;
drivers/net/fddi/skfp/hwmtm.c
1546
smc->os.hwm.tx_p = smc->hw.fp.tx[frame_status & QUEUE_A0] ;
drivers/net/fddi/skfp/hwmtm.c
1844
queue = smc->hw.fp.tx[QUEUE_A0] ;
drivers/net/fddi/skfp/hwmtm.c
1946
queue = smc->hw.fp.tx[i] ;
drivers/net/fddi/skfp/hwmtm.c
2037
queue = smc->hw.fp.tx[i] ;
drivers/net/fddi/skfp/hwmtm.c
2061
queue = smc->hw.fp.tx[i] ;
drivers/net/fddi/skfp/hwmtm.c
358
queue = smc->hw.fp.tx[QUEUE_A0] ;
drivers/net/fddi/skfp/hwmtm.c
372
queue = smc->hw.fp.tx[QUEUE_S] ;
drivers/net/fddi/skfp/hwmtm.c
559
phys = repair_txd_ring(smc,smc->hw.fp.tx[QUEUE_A0]) ;
drivers/net/fddi/skfp/hwmtm.c
564
phys = repair_txd_ring(smc,smc->hw.fp.tx[QUEUE_S]) ;
drivers/net/fddi/skfp/rmt.c
313
int tx ;
drivers/net/fddi/skfp/rmt.c
319
if ((tx = sm_mac_get_tx_state(smc)) == 4 || tx == 5) {
drivers/net/fddi/skfp/rmt.c
338
tx, smc->r.bn_flag);
drivers/net/fddi/skfp/rmt.c
391
int tx ;
drivers/net/fddi/skfp/rmt.c
397
if ((tx = sm_mac_get_tx_state(smc)) == 4 || tx == 5) {
drivers/net/fddi/skfp/rmt.c
416
tx, smc->r.bn_flag);
drivers/net/fjes/fjes_hw.c
1091
fjes_hw_setup_epbuf(&hw->ep_shm_info[epidx].tx,
drivers/net/fjes/fjes_hw.c
1146
&hw->ep_shm_info[epidx].tx,
drivers/net/fjes/fjes_hw.c
1161
hw->ep_shm_info[epidx].tx.
drivers/net/fjes/fjes_hw.c
1190
tx.info->v1i.rx_status |=
drivers/net/fjes/fjes_hw.c
250
result = fjes_hw_alloc_epbuf(&buf_pair->tx);
drivers/net/fjes/fjes_hw.c
259
fjes_hw_setup_epbuf(&buf_pair->tx, mac,
drivers/net/fjes/fjes_hw.c
284
fjes_hw_free_epbuf(&hw->ep_shm_info[epidx].tx);
drivers/net/fjes/fjes_hw.c
318
fjes_hw_free_epbuf(&hw->ep_shm_info[epidx].tx);
drivers/net/fjes/fjes_hw.c
522
buf_pair->tx.size,
drivers/net/fjes/fjes_hw.c
527
req_buf->share_buffer.buffer[idx++] = buf_pair->tx.size;
drivers/net/fjes/fjes_hw.c
528
page_count = buf_pair->tx.size / EP_BUFFER_INFO_SIZE;
drivers/net/fjes/fjes_hw.c
530
addr = ((u8 *)(buf_pair->tx.buffer)) +
drivers/net/fjes/fjes_hw.c
763
return (hw->ep_shm_info[src_epid].tx.info->v1i.rx_status &
drivers/net/fjes/fjes_hw.c
813
hw->ep_shm_info[epidx].tx.info->v1i.rx_status |=
drivers/net/fjes/fjes_hw.h
252
} tx, rx;
drivers/net/fjes/fjes_main.c
1038
adapter->hw.ep_shm_info[epidx].tx
drivers/net/fjes/fjes_main.c
1104
info = adapter->hw.ep_shm_info[epid].tx.info;
drivers/net/fjes/fjes_main.c
1148
hw->ep_shm_info[epid].tx.info->v1i.tx_status;
drivers/net/fjes/fjes_main.c
1152
hw->ep_shm_info[epid].tx.info->v1i.tx_status =
drivers/net/fjes/fjes_main.c
1245
fjes_hw_setup_epbuf(&hw->ep_shm_info[epidx].tx,
drivers/net/fjes/fjes_main.c
1290
&hw->ep_shm_info[epidx].tx,
drivers/net/fjes/fjes_main.c
1302
hw->ep_shm_info[epidx].tx.info->v1i.rx_status &=
drivers/net/fjes/fjes_main.c
214
fjes_hw_setup_epbuf(&buf_pair->tx, netdev->dev_addr,
drivers/net/fjes/fjes_main.c
265
hw->ep_shm_info[src_epid].tx.info->v1i.rx_status |=
drivers/net/fjes/fjes_main.c
305
hw->ep_shm_info[src_epid].tx.info->v1i.rx_status |=
drivers/net/fjes/fjes_main.c
451
fjes_hw_setup_epbuf(&buf_pair->tx,
drivers/net/fjes/fjes_main.c
544
.tx.info->v1i.rx_status &=
drivers/net/fjes/fjes_main.c
572
retval = fjes_hw_epbuf_tx_pkt_send(&adapter->hw.ep_shm_info[dest].tx,
drivers/net/fjes/fjes_main.c
577
adapter->hw.ep_shm_info[dest].tx.info->v1i.tx_status =
drivers/net/fjes/fjes_main.c
796
hw->ep_shm_info[epidx].tx.info->v1i.rx_status &=
drivers/net/fjes/fjes_main.c
819
fjes_hw_setup_epbuf(&hw->ep_shm_info[epidx].tx,
drivers/net/fjes/fjes_main.c
823
hw->ep_shm_info[epidx].tx.info->v1i.rx_status |=
drivers/net/fjes/fjes_main.c
856
&adapter->hw.ep_shm_info[epid].tx, vid))
drivers/net/fjes/fjes_main.c
858
&adapter->hw.ep_shm_info[epid].tx, vid);
drivers/net/fjes/fjes_main.c
874
fjes_hw_del_vlan_id(&adapter->hw.ep_shm_info[epid].tx, vid);
drivers/net/fjes/fjes_main.c
973
.tx.info->v1i.rx_status |= FJES_RX_POLL_WORK;
drivers/net/fjes/fjes_trace.h
100
__field(u64, tx)
drivers/net/fjes/fjes_trace.h
106
void *tx, *rx;
drivers/net/fjes/fjes_trace.h
108
tx = (void *)buf_pair->tx.buffer;
drivers/net/fjes/fjes_trace.h
112
__entry->tx_size = buf_pair->tx.size;
drivers/net/fjes/fjes_trace.h
114
__entry->tx = page_to_phys(vmalloc_to_page(tx)) +
drivers/net/fjes/fjes_trace.h
115
offset_in_page(tx);
drivers/net/fjes/fjes_trace.h
120
__entry->length, __entry->epid, __entry->tx, __entry->tx_size,
drivers/net/fjes/fjes_trace.h
285
hw->ep_shm_info[src_epid].tx.info->v1i.rx_status;
drivers/net/fjes/fjes_trace.h
305
__entry->rx_status = hw->ep_shm_info[src_epid].tx.info->v1i.rx_status;
drivers/net/fjes/fjes_trace.h
329
hw->ep_shm_info[src_epid].tx.info->v1i.rx_status;
drivers/net/fjes/fjes_trace.h
350
hw->ep_shm_info[src_epid].tx.info->v1i.rx_status;
drivers/net/hamradio/scc.c
204
static void scc_key_trx (struct scc_channel *scc, char tx);
drivers/net/hamradio/scc.c
901
static void scc_key_trx(struct scc_channel *scc, char tx)
drivers/net/hamradio/scc.c
906
Outb(scc->ctrl + 4, scc->option | (tx? 0x80 : 0));
drivers/net/hamradio/scc.c
911
time_const = (unsigned) (scc->clock / (scc->modem.speed * (tx? 2:64))) - 2;
drivers/net/hamradio/scc.c
915
if (tx)
drivers/net/hamradio/scc.c
923
if (tx)
drivers/net/hamradio/scc.c
959
if (tx)
drivers/net/ieee802154/at86rf230.c
1586
at86rf230_setup_spi_messages(lp, &lp->tx);
drivers/net/ieee802154/at86rf230.c
880
struct at86rf230_state_change *ctx = &lp->tx;
drivers/net/ieee802154/at86rf230.c
96
struct at86rf230_state_change tx;
drivers/net/ipa/data/ipa_data-v3.1.c
129
.tx = {
drivers/net/ipa/data/ipa_data-v3.1.c
83
.tx = {
drivers/net/ipa/data/ipa_data-v3.5.1.c
120
.tx = {
drivers/net/ipa/data/ipa_data-v3.5.1.c
74
.tx = {
drivers/net/ipa/data/ipa_data-v4.11.c
114
.tx = {
drivers/net/ipa/data/ipa_data-v4.11.c
68
.tx = {
drivers/net/ipa/data/ipa_data-v4.2.c
110
.tx = {
drivers/net/ipa/data/ipa_data-v4.2.c
64
.tx = {
drivers/net/ipa/data/ipa_data-v4.5.c
123
.tx = {
drivers/net/ipa/data/ipa_data-v4.5.c
77
.tx = {
drivers/net/ipa/data/ipa_data-v4.7.c
110
.tx = {
drivers/net/ipa/data/ipa_data-v4.7.c
64
.tx = {
drivers/net/ipa/data/ipa_data-v4.9.c
115
.tx = {
drivers/net/ipa/data/ipa_data-v4.9.c
69
.tx = {
drivers/net/ipa/data/ipa_data-v5.0.c
127
.tx = {
drivers/net/ipa/data/ipa_data-v5.0.c
81
.tx = {
drivers/net/ipa/data/ipa_data-v5.5.c
127
.tx = {
drivers/net/ipa/data/ipa_data-v5.5.c
81
.tx = {
drivers/net/ipa/ipa_endpoint.c
1219
val = reg_encode(reg, SEQ_TYPE, endpoint->config.tx.seq_type);
drivers/net/ipa/ipa_endpoint.c
1224
endpoint->config.tx.seq_rep_type);
drivers/net/ipa/ipa_endpoint.c
1286
name = endpoint->config.tx.status_endpoint;
drivers/net/ipa/ipa_endpoint.c
323
if (data->endpoint.config.tx.seq_rep_type) {
drivers/net/ipa/ipa_endpoint.c
331
other_name = data->endpoint.config.tx.status_endpoint;
drivers/net/ipa/ipa_endpoint.h
119
struct ipa_endpoint_tx tx;
drivers/net/ipa/ipa_modem.c
102
ipa_endpoint_disable_one(priv->tx);
drivers/net/ipa/ipa_modem.c
257
ipa_endpoint_suspend_one(priv->tx);
drivers/net/ipa/ipa_modem.c
275
netif_wake_queue(priv->tx->netdev);
drivers/net/ipa/ipa_modem.c
291
ipa_endpoint_resume_one(priv->tx);
drivers/net/ipa/ipa_modem.c
323
priv->tx = ipa->name_map[IPA_ENDPOINT_AP_MODEM_TX];
drivers/net/ipa/ipa_modem.c
327
priv->tx->netdev = netdev;
drivers/net/ipa/ipa_modem.c
336
priv->tx->netdev = NULL;
drivers/net/ipa/ipa_modem.c
380
priv->tx->netdev = NULL;
drivers/net/ipa/ipa_modem.c
46
struct ipa_endpoint *tx;
drivers/net/ipa/ipa_modem.c
64
ret = ipa_endpoint_enable_one(priv->tx);
drivers/net/ipa/ipa_modem.c
79
ipa_endpoint_disable_one(priv->tx);
drivers/net/netdevsim/ethtool.c
29
pause->tx_pause = ns->ethtool.pauseparam.tx;
drivers/net/netdevsim/ethtool.c
41
ns->ethtool.pauseparam.tx = pause->tx_pause;
drivers/net/netdevsim/ipsec.c
266
ipsec->tx++;
drivers/net/netdevsim/ipsec.c
34
ipsec->count, ipsec->tx);
drivers/net/netdevsim/netdev.c
707
struct netdev_queue_stats_tx *tx)
drivers/net/netdevsim/netdev.c
715
tx->packets = !!rtstats.tx_packets;
drivers/net/netdevsim/netdev.c
716
tx->bytes = 0;
drivers/net/netdevsim/netdevsim.h
56
u32 tx;
drivers/net/netdevsim/netdevsim.h
80
bool tx;
drivers/net/netdevsim/psp.c
45
if (!psp_dev_encapsulate(net, skb, pas->tx.spi, pas->version, 0)) {
drivers/net/netdevsim/psp.c
60
generation = pas->tx.key[0];
drivers/net/ovpn/netlink.c
624
atomic64_read(&peer->vpn_stats.tx.bytes)) ||
drivers/net/ovpn/netlink.c
626
atomic64_read(&peer->vpn_stats.tx.packets)) ||
drivers/net/ovpn/netlink.c
634
atomic64_read(&peer->link_stats.tx.bytes)) ||
drivers/net/ovpn/netlink.c
636
atomic64_read(&peer->link_stats.tx.packets)))
drivers/net/ovpn/stats.c
19
atomic64_set(&ps->tx.bytes, 0);
drivers/net/ovpn/stats.c
20
atomic64_set(&ps->tx.packets, 0);
drivers/net/ovpn/stats.h
23
struct ovpn_peer_stat tx;
drivers/net/ovpn/stats.h
44
ovpn_peer_stats_increment(&stats->tx, n);
drivers/net/phy/linkmode.c
89
void linkmode_set_pause(unsigned long *advertisement, bool tx, bool rx)
drivers/net/phy/linkmode.c
93
rx ^ tx);
drivers/net/phy/micrel.c
1177
u16 rx, tx, rx_clk, tx_clk;
drivers/net/phy/micrel.c
1182
tx = TX_ND;
drivers/net/phy/micrel.c
1188
tx = TX_ID;
drivers/net/phy/micrel.c
1194
tx = TX_ND;
drivers/net/phy/micrel.c
1200
tx = TX_ID;
drivers/net/phy/micrel.c
1211
FIELD_PREP(MII_KSZ9031RN_TX_CTL_M, tx));
drivers/net/phy/micrel.c
1224
FIELD_PREP(MII_KSZ9031RN_TXD3, tx) |
drivers/net/phy/micrel.c
1225
FIELD_PREP(MII_KSZ9031RN_TXD2, tx) |
drivers/net/phy/micrel.c
1226
FIELD_PREP(MII_KSZ9031RN_TXD1, tx) |
drivers/net/phy/micrel.c
1227
FIELD_PREP(MII_KSZ9031RN_TXD0, tx));
drivers/net/phy/phy_device.c
2996
void phy_set_sym_pause(struct phy_device *phydev, bool rx, bool tx,
drivers/net/phy/phy_device.c
3001
if (rx && tx && autoneg)
drivers/net/phy/phy_device.c
3020
void phy_set_asym_pause(struct phy_device *phydev, bool rx, bool tx)
drivers/net/phy/phy_device.c
3025
linkmode_set_pause(phydev->advertising, tx, rx);
drivers/net/usb/hso.c
1602
icount->tx = cnow.tx;
drivers/net/virtio_net.c
5623
struct netdev_queue_stats_tx *tx)
drivers/net/virtio_net.c
5654
tx->bytes = 0;
drivers/net/virtio_net.c
5655
tx->packets = 0;
drivers/net/virtio_net.c
5656
tx->stop = 0;
drivers/net/virtio_net.c
5657
tx->wake = 0;
drivers/net/virtio_net.c
5660
tx->hw_drops = 0;
drivers/net/virtio_net.c
5661
tx->hw_drop_errors = 0;
drivers/net/virtio_net.c
5665
tx->csum_none = 0;
drivers/net/virtio_net.c
5666
tx->needs_csum = 0;
drivers/net/virtio_net.c
5670
tx->hw_gso_packets = 0;
drivers/net/virtio_net.c
5671
tx->hw_gso_bytes = 0;
drivers/net/virtio_net.c
5672
tx->hw_gso_wire_packets = 0;
drivers/net/virtio_net.c
5673
tx->hw_gso_wire_bytes = 0;
drivers/net/virtio_net.c
5677
tx->hw_drop_ratelimits = 0;
drivers/net/virtio_net.c
5681
dev->real_num_tx_queues, vi->max_queue_pairs, tx);
drivers/net/wireless/admtek/adm8211.c
1766
.tx = adm8211_tx,
drivers/net/wireless/ath/ar5523/ar5523.c
1368
.tx = ar5523_tx,
drivers/net/wireless/ath/ath10k/mac.c
9473
.tx = ath10k_mac_op_tx,
drivers/net/wireless/ath/ath10k/wmi-tlv.c
1554
ath10k_wmi_pull_pdev_stats_tx(&src->tx, dst);
drivers/net/wireless/ath/ath10k/wmi.c
3054
ath10k_wmi_pull_pdev_stats_tx(&src->tx, dst);
drivers/net/wireless/ath/ath10k/wmi.c
3108
ath10k_wmi_pull_pdev_stats_tx(&src->tx, dst);
drivers/net/wireless/ath/ath10k/wmi.c
3169
ath10k_wmi_pull_pdev_stats_tx(&src->tx, dst);
drivers/net/wireless/ath/ath10k/wmi.c
3246
ath10k_wmi_pull_pdev_stats_tx(&src->tx, dst);
drivers/net/wireless/ath/ath10k/wmi.c
3338
ath10k_wmi_10_4_pull_pdev_stats_tx(&src->tx, dst);
drivers/net/wireless/ath/ath10k/wmi.c
5813
u64 busy, total, tx, rx, rx_bss;
drivers/net/wireless/ath/ath10k/wmi.c
5826
tx = __le64_to_cpu(ev->cycle_tx);
drivers/net/wireless/ath/ath10k/wmi.c
5832
freq, noise_floor, busy, total, tx, rx, rx_bss);
drivers/net/wireless/ath/ath10k/wmi.c
5848
survey->time_tx = div_u64(tx, cc_freq_hz);
drivers/net/wireless/ath/ath10k/wmi.h
4706
struct wmi_pdev_stats_tx tx;
drivers/net/wireless/ath/ath10k/wmi.h
4722
struct wmi_pdev_stats_tx tx;
drivers/net/wireless/ath/ath10k/wmi.h
4735
struct wmi_pdev_stats_tx tx;
drivers/net/wireless/ath/ath10k/wmi.h
4746
struct wmi_10_4_pdev_stats_tx tx;
drivers/net/wireless/ath/ath11k/ahb.c
546
if (ab->hw_params.ring_mask->tx[i] & BIT(j)) {
drivers/net/wireless/ath/ath11k/dp.c
144
grp_mask = &ab->hw_params.ring_mask->tx[0];
drivers/net/wireless/ath/ath11k/dp.c
787
ab->hw_params.ring_mask->tx[grp_id])
drivers/net/wireless/ath/ath11k/hw.c
1214
.tx = {
drivers/net/wireless/ath/ath11k/hw.c
1255
.tx = {
drivers/net/wireless/ath/ath11k/hw.c
1996
.tx = {
drivers/net/wireless/ath/ath11k/hw.c
2037
.tx = {
drivers/net/wireless/ath/ath11k/hw.h
118
u8 tx[ATH11K_EXT_IRQ_GRP_NUM_MAX];
drivers/net/wireless/ath/ath11k/mac.c
10081
.tx = ath11k_mac_op_tx,
drivers/net/wireless/ath/ath11k/mac.c
1063
param.chains[NL80211_BAND_2GHZ].tx = ar->num_tx_chains;
drivers/net/wireless/ath/ath11k/mac.c
1067
param.chains[NL80211_BAND_5GHZ].tx = ar->num_tx_chains;
drivers/net/wireless/ath/ath11k/mac.c
6853
params->chains[NL80211_BAND_2GHZ].tx = ar->num_tx_chains;
drivers/net/wireless/ath/ath11k/mac.c
6857
params->chains[NL80211_BAND_5GHZ].tx = ar->num_tx_chains;
drivers/net/wireless/ath/ath11k/mac.c
6862
params->chains[NL80211_BAND_6GHZ].tx = ar->num_tx_chains;
drivers/net/wireless/ath/ath11k/pcic.c
605
if (ab->hw_params.ring_mask->tx[i] ||
drivers/net/wireless/ath/ath11k/wmi.c
6700
ath11k_wmi_pull_pdev_stats_tx(&src->tx, dst);
drivers/net/wireless/ath/ath11k/wmi.c
768
param->chains[NL80211_BAND_2GHZ].tx;
drivers/net/wireless/ath/ath11k/wmi.c
778
param->chains[NL80211_BAND_5GHZ].tx;
drivers/net/wireless/ath/ath11k/wmi.c
8025
u64 busy, total, tx, rx, rx_bss;
drivers/net/wireless/ath/ath11k/wmi.c
8039
tx = (u64)(bss_ch_info_ev.tx_cycle_count_high) << 32 |
drivers/net/wireless/ath/ath11k/wmi.c
8052
tx, rx, rx_bss);
drivers/net/wireless/ath/ath11k/wmi.c
8078
survey->time_tx = div_u64(tx, cc_freq_hz);
drivers/net/wireless/ath/ath11k/wmi.h
2617
u8 tx;
drivers/net/wireless/ath/ath11k/wmi.h
4860
struct wmi_pdev_stats_tx tx;
drivers/net/wireless/ath/ath12k/ahb.c
582
if (ring_mask->tx[i] &
drivers/net/wireless/ath/ath12k/dp.c
153
grp_mask = &ab->hw_params->ring_mask->tx[0];
drivers/net/wireless/ath/ath12k/hw.h
121
u8 tx[ATH12K_EXT_IRQ_GRP_NUM_MAX];
drivers/net/wireless/ath/ath12k/mac.c
9946
arg->chains[NL80211_BAND_2GHZ].tx = ar->num_tx_chains;
drivers/net/wireless/ath/ath12k/mac.c
9950
arg->chains[NL80211_BAND_5GHZ].tx = ar->num_tx_chains;
drivers/net/wireless/ath/ath12k/mac.c
9955
arg->chains[NL80211_BAND_6GHZ].tx = ar->num_tx_chains;
drivers/net/wireless/ath/ath12k/pci.c
564
if (ab->hw_params->ring_mask->tx[i] ||
drivers/net/wireless/ath/ath12k/wifi7/dp.c
31
if (dp->hw_params->ring_mask->tx[grp_id]) {
drivers/net/wireless/ath/ath12k/wifi7/dp.c
32
i = fls(dp->hw_params->ring_mask->tx[grp_id]) - 1;
drivers/net/wireless/ath/ath12k/wifi7/hw.c
214
.tx = {
drivers/net/wireless/ath/ath12k/wifi7/hw.c
256
.tx = {
drivers/net/wireless/ath/ath12k/wifi7/hw.c
296
.tx = {
drivers/net/wireless/ath/ath12k/wifi7/hw.c
967
.tx = ath12k_wifi7_mac_op_tx,
drivers/net/wireless/ath/ath12k/wmi.c
7612
u64 busy, total, tx, rx, rx_bss;
drivers/net/wireless/ath/ath12k/wmi.c
7626
tx = (u64)(le32_to_cpu(bss_ch_info_ev.tx_cycle_count_high)) << 32 |
drivers/net/wireless/ath/ath12k/wmi.c
7639
tx, rx, rx_bss);
drivers/net/wireless/ath/ath12k/wmi.c
7665
survey->time_tx = div_u64(tx, cc_freq_hz);
drivers/net/wireless/ath/ath12k/wmi.c
8342
ath12k_wmi_pull_pdev_stats_tx(&src->tx, dst);
drivers/net/wireless/ath/ath12k/wmi.c
895
cpu_to_le32(args->chains[NL80211_BAND_2GHZ].tx);
drivers/net/wireless/ath/ath12k/wmi.c
904
cpu_to_le32(args->chains[NL80211_BAND_5GHZ].tx);
drivers/net/wireless/ath/ath12k/wmi.h
2884
u8 tx;
drivers/net/wireless/ath/ath12k/wmi.h
6037
struct ath12k_wmi_pdev_tx_stats_params tx;
drivers/net/wireless/ath/ath5k/mac80211-ops.c
745
u32 *tx, u32 *tx_max, u32 *rx, u32 *rx_max)
drivers/net/wireless/ath/ath5k/mac80211-ops.c
749
*tx = ah->txqs[AR5K_TX_QUEUE_ID_DATA_MIN].txq_max;
drivers/net/wireless/ath/ath5k/mac80211-ops.c
756
static int ath5k_set_ringparam(struct ieee80211_hw *hw, u32 tx, u32 rx)
drivers/net/wireless/ath/ath5k/mac80211-ops.c
766
if (!tx || tx > ATH5K_TXQ_LEN_MAX)
drivers/net/wireless/ath/ath5k/mac80211-ops.c
776
ah->txqs[qnum].txq_max = tx;
drivers/net/wireless/ath/ath5k/mac80211-ops.c
790
.tx = ath5k_tx,
drivers/net/wireless/ath/ath6kl/cfg80211.c
3411
.tx = BIT(IEEE80211_STYPE_ACTION >> 4) |
drivers/net/wireless/ath/ath6kl/cfg80211.c
3417
.tx = BIT(IEEE80211_STYPE_ACTION >> 4) |
drivers/net/wireless/ath/ath6kl/cfg80211.c
3423
.tx = BIT(IEEE80211_STYPE_ACTION >> 4) |
drivers/net/wireless/ath/ath6kl/cfg80211.c
3429
.tx = BIT(IEEE80211_STYPE_ACTION >> 4) |
drivers/net/wireless/ath/ath6kl/htc-ops.h
48
return target->dev->ar->htc_ops->tx(target, packet);
drivers/net/wireless/ath/ath6kl/htc.h
324
struct htc_tx_packet_info tx;
drivers/net/wireless/ath/ath6kl/htc.h
562
int (*tx)(struct htc_target *target, struct htc_packet *packet);
drivers/net/wireless/ath/ath6kl/htc.h
650
packet->info.tx.tag = tag;
drivers/net/wireless/ath/ath6kl/htc_mbox.c
1168
(tag == packet->info.tx.tag))
drivers/net/wireless/ath/ath6kl/htc_mbox.c
1180
packet->endpoint, packet->info.tx.tag);
drivers/net/wireless/ath/ath6kl/htc_mbox.c
2916
.tx = ath6kl_htc_mbox_tx,
drivers/net/wireless/ath/ath6kl/htc_mbox.c
383
bool tx)
drivers/net/wireless/ath/ath6kl/htc_mbox.c
388
buf_list = tx ? &target->free_ctrl_txbuf : &target->free_ctrl_rxbuf;
drivers/net/wireless/ath/ath6kl/htc_mbox.c
401
if (tx)
drivers/net/wireless/ath/ath6kl/htc_mbox.c
419
packet->info.tx.cred_used);
drivers/net/wireless/ath/ath6kl/htc_mbox.c
424
packet->info.tx.cred_used;
drivers/net/wireless/ath/ath6kl/htc_mbox.c
457
packet->info.tx.seqno);
drivers/net/wireless/ath/ath6kl/htc_mbox.c
522
send_len, packet->info.tx.seqno, padded_len,
drivers/net/wireless/ath/ath6kl/htc_mbox.c
632
packet->info.tx.cred_used = req_cred;
drivers/net/wireless/ath/ath6kl/htc_mbox.c
640
packet->info.tx.flags = flags;
drivers/net/wireless/ath/ath6kl/htc_mbox.c
641
packet->info.tx.seqno = endpoint->seqno;
drivers/net/wireless/ath/ath6kl/htc_mbox.c
714
flags = packet->info.tx.flags | HTC_FLAGS_SEND_BUNDLE;
drivers/net/wireless/ath/ath6kl/htc_mbox.c
716
cred_pad, packet->info.tx.seqno);
drivers/net/wireless/ath/ath6kl/htc_mbox.c
727
i, packet, packet->info.tx.seqno, len, rem_scat);
drivers/net/wireless/ath/ath6kl/htc_mbox.c
924
ath6kl_htc_tx_prep_pkt(packet, packet->info.tx.flags,
drivers/net/wireless/ath/ath6kl/htc_mbox.c
925
0, packet->info.tx.seqno);
drivers/net/wireless/ath/ath6kl/htc_pipe.c
163
packet->info.tx.cred_used = credits_required;
drivers/net/wireless/ath/ath6kl/htc_pipe.c
165
packet->info.tx.flags = send_flags;
drivers/net/wireless/ath/ath6kl/htc_pipe.c
166
packet->info.tx.seqno = ep->seqno;
drivers/net/wireless/ath/ath6kl/htc_pipe.c
1708
.tx = ath6kl_htc_pipe_tx,
drivers/net/wireless/ath/ath6kl/htc_pipe.c
192
packet->info.tx.seqno = ep->seqno;
drivers/net/wireless/ath/ath6kl/htc_pipe.c
193
packet->info.tx.flags = 0;
drivers/net/wireless/ath/ath6kl/htc_pipe.c
194
packet->info.tx.cred_used = 0;
drivers/net/wireless/ath/ath6kl/htc_pipe.c
238
packet->info.tx.flags |= HTC_FLAGS_TX_FIXUP_NETBUF;
drivers/net/wireless/ath/ath6kl/htc_pipe.c
241
htc_hdr->flags = packet->info.tx.flags;
drivers/net/wireless/ath/ath6kl/htc_pipe.c
244
htc_hdr->ctrl[1] = (u8) packet->info.tx.seqno;
drivers/net/wireless/ath/ath6kl/htc_pipe.c
271
ep->cred_dist.credits += packet->info.tx.cred_used;
drivers/net/wireless/ath/ath6kl/htc_pipe.c
31
if (packet->info.tx.flags & HTC_FLAGS_TX_FIXUP_NETBUF) {
drivers/net/wireless/ath/ath6kl/htc_pipe.c
33
packet->info.tx.flags &= ~HTC_FLAGS_TX_FIXUP_NETBUF;
drivers/net/wireless/ath/ath6kl/main.c
718
stats->tx_pkt += le32_to_cpu(tgt_stats->stats.tx.pkt);
drivers/net/wireless/ath/ath6kl/main.c
719
stats->tx_byte += le32_to_cpu(tgt_stats->stats.tx.byte);
drivers/net/wireless/ath/ath6kl/main.c
720
stats->tx_ucast_pkt += le32_to_cpu(tgt_stats->stats.tx.ucast_pkt);
drivers/net/wireless/ath/ath6kl/main.c
721
stats->tx_ucast_byte += le32_to_cpu(tgt_stats->stats.tx.ucast_byte);
drivers/net/wireless/ath/ath6kl/main.c
722
stats->tx_mcast_pkt += le32_to_cpu(tgt_stats->stats.tx.mcast_pkt);
drivers/net/wireless/ath/ath6kl/main.c
723
stats->tx_mcast_byte += le32_to_cpu(tgt_stats->stats.tx.mcast_byte);
drivers/net/wireless/ath/ath6kl/main.c
724
stats->tx_bcast_pkt += le32_to_cpu(tgt_stats->stats.tx.bcast_pkt);
drivers/net/wireless/ath/ath6kl/main.c
725
stats->tx_bcast_byte += le32_to_cpu(tgt_stats->stats.tx.bcast_byte);
drivers/net/wireless/ath/ath6kl/main.c
727
le32_to_cpu(tgt_stats->stats.tx.rts_success_cnt);
drivers/net/wireless/ath/ath6kl/main.c
731
le32_to_cpu(tgt_stats->stats.tx.pkt_per_ac[ac]);
drivers/net/wireless/ath/ath6kl/main.c
733
stats->tx_err += le32_to_cpu(tgt_stats->stats.tx.err);
drivers/net/wireless/ath/ath6kl/main.c
734
stats->tx_fail_cnt += le32_to_cpu(tgt_stats->stats.tx.fail_cnt);
drivers/net/wireless/ath/ath6kl/main.c
735
stats->tx_retry_cnt += le32_to_cpu(tgt_stats->stats.tx.retry_cnt);
drivers/net/wireless/ath/ath6kl/main.c
737
le32_to_cpu(tgt_stats->stats.tx.mult_retry_cnt);
drivers/net/wireless/ath/ath6kl/main.c
739
le32_to_cpu(tgt_stats->stats.tx.rts_fail_cnt);
drivers/net/wireless/ath/ath6kl/main.c
741
rate = a_sle32_to_cpu(tgt_stats->stats.tx.ucast_rate);
drivers/net/wireless/ath/ath6kl/trace.h
106
__entry->tx = 1;
drivers/net/wireless/ath/ath6kl/trace.h
108
__entry->tx = 0;
drivers/net/wireless/ath/ath6kl/trace.h
113
__entry->tx ? "tx" : "rx",
drivers/net/wireless/ath/ath6kl/trace.h
127
__field(unsigned int, tx)
drivers/net/wireless/ath/ath6kl/trace.h
147
__entry->tx = 1;
drivers/net/wireless/ath/ath6kl/trace.h
149
__entry->tx = 0;
drivers/net/wireless/ath/ath6kl/trace.h
166
__entry->tx ? "tx" : "rx",
drivers/net/wireless/ath/ath6kl/trace.h
92
__field(unsigned int, tx)
drivers/net/wireless/ath/ath6kl/txrx.c
607
if (packet->info.tx.tag == ATH6KL_CONTROL_PKT_TAG)
drivers/net/wireless/ath/ath6kl/wmi.h
1832
struct tx_stats tx;
drivers/net/wireless/ath/ath9k/ar9003_phy.c
667
void ar9003_hw_set_chain_masks(struct ath_hw *ah, u8 rx, u8 tx)
drivers/net/wireless/ath/ath9k/ar9003_phy.c
676
if ((ah->caps.hw_caps & ATH9K_HW_CAP_APM) && (tx == 0x7))
drivers/net/wireless/ath/ath9k/ar9003_phy.c
677
tx = 3;
drivers/net/wireless/ath/ath9k/ar9003_phy.c
679
REG_WRITE(ah, AR_SELFGEN_MASK, tx);
drivers/net/wireless/ath/ath9k/ath9k.h
1029
struct ath_tx tx;
drivers/net/wireless/ath/ath9k/ath9k.h
85
#define ATH_TXQ_SETUP(sc, i) ((sc)->tx.txqsetup & (1<<i))
drivers/net/wireless/ath/ath9k/beacon.c
50
txq = sc->tx.txq_map[IEEE80211_AC_BE];
drivers/net/wireless/ath/ath9k/channel.c
1011
txctl.txq = sc->tx.txq_map[IEEE80211_AC_VO];
drivers/net/wireless/ath/ath9k/channel.c
1130
txctl.txq = sc->tx.txq_map[IEEE80211_AC_VO];
drivers/net/wireless/ath/ath9k/debug.c
632
txq = sc->tx.txq_map[i];
drivers/net/wireless/ath/ath9k/debug.h
198
#define PR_QNUM(_n) sc->tx.txq_map[_n]->axq_qnum
drivers/net/wireless/ath/ath9k/gpio.c
430
txq = sc->tx.txq_map[IEEE80211_AC_BE];
drivers/net/wireless/ath/ath9k/hif_usb.c
160
spin_lock_irqsave(&hif_dev->tx.tx_lock, flags);
drivers/net/wireless/ath/ath9k/hif_usb.c
161
if (hif_dev->tx.flags & HIF_USB_TX_FLUSH) {
drivers/net/wireless/ath/ath9k/hif_usb.c
162
spin_unlock_irqrestore(&hif_dev->tx.tx_lock, flags);
drivers/net/wireless/ath/ath9k/hif_usb.c
167
spin_unlock_irqrestore(&hif_dev->tx.tx_lock, flags);
drivers/net/wireless/ath/ath9k/hif_usb.c
277
spin_lock(&hif_dev->tx.tx_lock);
drivers/net/wireless/ath/ath9k/hif_usb.c
278
if (hif_dev->tx.flags & HIF_USB_TX_FLUSH) {
drivers/net/wireless/ath/ath9k/hif_usb.c
279
spin_unlock(&hif_dev->tx.tx_lock);
drivers/net/wireless/ath/ath9k/hif_usb.c
283
spin_unlock(&hif_dev->tx.tx_lock);
drivers/net/wireless/ath/ath9k/hif_usb.c
298
spin_lock(&hif_dev->tx.tx_lock);
drivers/net/wireless/ath/ath9k/hif_usb.c
299
list_move_tail(&tx_buf->list, &hif_dev->tx.tx_buf);
drivers/net/wireless/ath/ath9k/hif_usb.c
300
hif_dev->tx.tx_buf_cnt++;
drivers/net/wireless/ath/ath9k/hif_usb.c
301
if (!(hif_dev->tx.flags & HIF_USB_TX_STOP))
drivers/net/wireless/ath/ath9k/hif_usb.c
304
spin_unlock(&hif_dev->tx.tx_lock);
drivers/net/wireless/ath/ath9k/hif_usb.c
317
if (hif_dev->tx.tx_skb_cnt == 0)
drivers/net/wireless/ath/ath9k/hif_usb.c
321
if (list_empty(&hif_dev->tx.tx_buf))
drivers/net/wireless/ath/ath9k/hif_usb.c
324
tx_buf = list_first_entry(&hif_dev->tx.tx_buf, struct tx_buf, list);
drivers/net/wireless/ath/ath9k/hif_usb.c
325
list_move_tail(&tx_buf->list, &hif_dev->tx.tx_pending);
drivers/net/wireless/ath/ath9k/hif_usb.c
326
hif_dev->tx.tx_buf_cnt--;
drivers/net/wireless/ath/ath9k/hif_usb.c
328
tx_skb_cnt = min_t(u16, hif_dev->tx.tx_skb_cnt, MAX_TX_AGGR_NUM);
drivers/net/wireless/ath/ath9k/hif_usb.c
331
nskb = __skb_dequeue(&hif_dev->tx.tx_skb_queue);
drivers/net/wireless/ath/ath9k/hif_usb.c
336
hif_dev->tx.tx_skb_cnt--;
drivers/net/wireless/ath/ath9k/hif_usb.c
367
list_move_tail(&tx_buf->list, &hif_dev->tx.tx_buf);
drivers/net/wireless/ath/ath9k/hif_usb.c
368
hif_dev->tx.tx_buf_cnt++;
drivers/net/wireless/ath/ath9k/hif_usb.c
382
spin_lock_irqsave(&hif_dev->tx.tx_lock, flags);
drivers/net/wireless/ath/ath9k/hif_usb.c
384
if (hif_dev->tx.flags & HIF_USB_TX_STOP) {
drivers/net/wireless/ath/ath9k/hif_usb.c
385
spin_unlock_irqrestore(&hif_dev->tx.tx_lock, flags);
drivers/net/wireless/ath/ath9k/hif_usb.c
390
if (hif_dev->tx.tx_skb_cnt > MAX_TX_BUF_NUM) {
drivers/net/wireless/ath/ath9k/hif_usb.c
391
spin_unlock_irqrestore(&hif_dev->tx.tx_lock, flags);
drivers/net/wireless/ath/ath9k/hif_usb.c
395
spin_unlock_irqrestore(&hif_dev->tx.tx_lock, flags);
drivers/net/wireless/ath/ath9k/hif_usb.c
405
spin_lock_irqsave(&hif_dev->tx.tx_lock, flags);
drivers/net/wireless/ath/ath9k/hif_usb.c
409
__skb_queue_tail(&hif_dev->tx.tx_skb_queue, skb);
drivers/net/wireless/ath/ath9k/hif_usb.c
410
hif_dev->tx.tx_skb_cnt++;
drivers/net/wireless/ath/ath9k/hif_usb.c
414
if ((hif_dev->tx.tx_buf_cnt == MAX_TX_URB_NUM) &&
drivers/net/wireless/ath/ath9k/hif_usb.c
415
(hif_dev->tx.tx_skb_cnt < 2)) {
drivers/net/wireless/ath/ath9k/hif_usb.c
419
spin_unlock_irqrestore(&hif_dev->tx.tx_lock, flags);
drivers/net/wireless/ath/ath9k/hif_usb.c
431
spin_lock_irqsave(&hif_dev->tx.tx_lock, flags);
drivers/net/wireless/ath/ath9k/hif_usb.c
432
hif_dev->tx.flags &= ~HIF_USB_TX_STOP;
drivers/net/wireless/ath/ath9k/hif_usb.c
433
spin_unlock_irqrestore(&hif_dev->tx.tx_lock, flags);
drivers/net/wireless/ath/ath9k/hif_usb.c
442
spin_lock_irqsave(&hif_dev->tx.tx_lock, flags);
drivers/net/wireless/ath/ath9k/hif_usb.c
443
ath9k_skb_queue_complete(hif_dev, &hif_dev->tx.tx_skb_queue, false);
drivers/net/wireless/ath/ath9k/hif_usb.c
444
hif_dev->tx.tx_skb_cnt = 0;
drivers/net/wireless/ath/ath9k/hif_usb.c
445
hif_dev->tx.flags |= HIF_USB_TX_STOP;
drivers/net/wireless/ath/ath9k/hif_usb.c
446
spin_unlock_irqrestore(&hif_dev->tx.tx_lock, flags);
drivers/net/wireless/ath/ath9k/hif_usb.c
449
spin_lock_irqsave(&hif_dev->tx.tx_lock, flags);
drivers/net/wireless/ath/ath9k/hif_usb.c
451
&hif_dev->tx.tx_pending, list) {
drivers/net/wireless/ath/ath9k/hif_usb.c
453
spin_unlock_irqrestore(&hif_dev->tx.tx_lock, flags);
drivers/net/wireless/ath/ath9k/hif_usb.c
459
spin_lock_irqsave(&hif_dev->tx.tx_lock, flags);
drivers/net/wireless/ath/ath9k/hif_usb.c
461
spin_unlock_irqrestore(&hif_dev->tx.tx_lock, flags);
drivers/net/wireless/ath/ath9k/hif_usb.c
507
spin_lock_irqsave(&hif_dev->tx.tx_lock, flags);
drivers/net/wireless/ath/ath9k/hif_usb.c
509
skb_queue_walk_safe(&hif_dev->tx.tx_skb_queue, skb, tmp) {
drivers/net/wireless/ath/ath9k/hif_usb.c
511
__skb_unlink(skb, &hif_dev->tx.tx_skb_queue);
drivers/net/wireless/ath/ath9k/hif_usb.c
514
hif_dev->tx.tx_skb_cnt--;
drivers/net/wireless/ath/ath9k/hif_usb.c
519
spin_unlock_irqrestore(&hif_dev->tx.tx_lock, flags);
drivers/net/wireless/ath/ath9k/hif_usb.c
809
spin_lock_irqsave(&hif_dev->tx.tx_lock, flags);
drivers/net/wireless/ath/ath9k/hif_usb.c
811
&hif_dev->tx.tx_buf, list) {
drivers/net/wireless/ath/ath9k/hif_usb.c
817
spin_unlock_irqrestore(&hif_dev->tx.tx_lock, flags);
drivers/net/wireless/ath/ath9k/hif_usb.c
819
spin_lock_irqsave(&hif_dev->tx.tx_lock, flags);
drivers/net/wireless/ath/ath9k/hif_usb.c
820
hif_dev->tx.flags |= HIF_USB_TX_FLUSH;
drivers/net/wireless/ath/ath9k/hif_usb.c
821
spin_unlock_irqrestore(&hif_dev->tx.tx_lock, flags);
drivers/net/wireless/ath/ath9k/hif_usb.c
823
spin_lock_irqsave(&hif_dev->tx.tx_lock, flags);
drivers/net/wireless/ath/ath9k/hif_usb.c
825
&hif_dev->tx.tx_pending, list) {
drivers/net/wireless/ath/ath9k/hif_usb.c
827
spin_unlock_irqrestore(&hif_dev->tx.tx_lock, flags);
drivers/net/wireless/ath/ath9k/hif_usb.c
833
spin_lock_irqsave(&hif_dev->tx.tx_lock, flags);
drivers/net/wireless/ath/ath9k/hif_usb.c
835
spin_unlock_irqrestore(&hif_dev->tx.tx_lock, flags);
drivers/net/wireless/ath/ath9k/hif_usb.c
845
INIT_LIST_HEAD(&hif_dev->tx.tx_buf);
drivers/net/wireless/ath/ath9k/hif_usb.c
846
INIT_LIST_HEAD(&hif_dev->tx.tx_pending);
drivers/net/wireless/ath/ath9k/hif_usb.c
847
spin_lock_init(&hif_dev->tx.tx_lock);
drivers/net/wireless/ath/ath9k/hif_usb.c
848
__skb_queue_head_init(&hif_dev->tx.tx_skb_queue);
drivers/net/wireless/ath/ath9k/hif_usb.c
867
list_add_tail(&tx_buf->list, &hif_dev->tx.tx_buf);
drivers/net/wireless/ath/ath9k/hif_usb.c
870
hif_dev->tx.tx_buf_cnt = MAX_TX_URB_NUM;
drivers/net/wireless/ath/ath9k/hif_usb.h
123
struct hif_usb_tx tx;
drivers/net/wireless/ath/ath9k/htc.h
504
struct ath9k_htc_tx tx;
drivers/net/wireless/ath/ath9k/htc_drv_beacon.c
180
spin_lock_bh(&priv->tx.tx_lock);
drivers/net/wireless/ath/ath9k/htc_drv_beacon.c
181
priv->tx.queued_cnt++;
drivers/net/wireless/ath/ath9k/htc_drv_beacon.c
182
spin_unlock_bh(&priv->tx.tx_lock);
drivers/net/wireless/ath/ath9k/htc_drv_debug.c
296
spin_lock_bh(&priv->tx.tx_lock);
drivers/net/wireless/ath/ath9k/htc_drv_debug.c
300
MAX_TX_BUF_NUM, priv->tx.tx_slot,
drivers/net/wireless/ath/ath9k/htc_drv_debug.c
301
bitmap_weight(priv->tx.tx_slot, MAX_TX_BUF_NUM));
drivers/net/wireless/ath/ath9k/htc_drv_debug.c
302
spin_unlock_bh(&priv->tx.tx_lock);
drivers/net/wireless/ath/ath9k/htc_drv_debug.c
321
"Mgmt endpoint", skb_queue_len(&priv->tx.mgmt_ep_queue));
drivers/net/wireless/ath/ath9k/htc_drv_debug.c
324
"Cab endpoint", skb_queue_len(&priv->tx.cab_ep_queue));
drivers/net/wireless/ath/ath9k/htc_drv_debug.c
327
"Data BE endpoint", skb_queue_len(&priv->tx.data_be_queue));
drivers/net/wireless/ath/ath9k/htc_drv_debug.c
330
"Data BK endpoint", skb_queue_len(&priv->tx.data_bk_queue));
drivers/net/wireless/ath/ath9k/htc_drv_debug.c
333
"Data VI endpoint", skb_queue_len(&priv->tx.data_vi_queue));
drivers/net/wireless/ath/ath9k/htc_drv_debug.c
336
"Data VO endpoint", skb_queue_len(&priv->tx.data_vo_queue));
drivers/net/wireless/ath/ath9k/htc_drv_debug.c
339
"Failed queue", skb_queue_len(&priv->tx.tx_failed));
drivers/net/wireless/ath/ath9k/htc_drv_debug.c
341
spin_lock_bh(&priv->tx.tx_lock);
drivers/net/wireless/ath/ath9k/htc_drv_debug.c
343
"Queued count", priv->tx.queued_cnt);
drivers/net/wireless/ath/ath9k/htc_drv_debug.c
344
spin_unlock_bh(&priv->tx.tx_lock);
drivers/net/wireless/ath/ath9k/htc_drv_init.c
122
void (*tx) (void *,
drivers/net/wireless/ath/ath9k/htc_drv_init.c
135
req.ep_callbacks.tx = tx;
drivers/net/wireless/ath/ath9k/htc_drv_init.c
645
spin_lock_init(&priv->tx.tx_lock);
drivers/net/wireless/ath/ath9k/htc_drv_init.c
653
timer_setup(&priv->tx.cleanup_timer, ath9k_htc_tx_cleanup_timer, 0);
drivers/net/wireless/ath/ath9k/htc_drv_main.c
1000
timer_delete_sync(&priv->tx.cleanup_timer);
drivers/net/wireless/ath/ath9k/htc_drv_main.c
1694
spin_lock_bh(&priv->tx.tx_lock);
drivers/net/wireless/ath/ath9k/htc_drv_main.c
1696
spin_unlock_bh(&priv->tx.tx_lock);
drivers/net/wireless/ath/ath9k/htc_drv_main.c
1879
.tx = ath9k_htc_tx,
drivers/net/wireless/ath/ath9k/htc_drv_main.c
201
timer_delete_sync(&priv->tx.cleanup_timer);
drivers/net/wireless/ath/ath9k/htc_drv_main.c
233
mod_timer(&priv->tx.cleanup_timer,
drivers/net/wireless/ath/ath9k/htc_drv_main.c
263
timer_delete_sync(&priv->tx.cleanup_timer);
drivers/net/wireless/ath/ath9k/htc_drv_main.c
311
mod_timer(&priv->tx.cleanup_timer,
drivers/net/wireless/ath/ath9k/htc_drv_main.c
735
spin_lock_bh(&priv->tx.tx_lock);
drivers/net/wireless/ath/ath9k/htc_drv_main.c
737
spin_unlock_bh(&priv->tx.tx_lock);
drivers/net/wireless/ath/ath9k/htc_drv_main.c
960
spin_lock_bh(&priv->tx.tx_lock);
drivers/net/wireless/ath/ath9k/htc_drv_main.c
961
priv->tx.flags &= ~ATH9K_HTC_OP_TX_QUEUES_STOP;
drivers/net/wireless/ath/ath9k/htc_drv_main.c
962
spin_unlock_bh(&priv->tx.tx_lock);
drivers/net/wireless/ath/ath9k/htc_drv_main.c
966
mod_timer(&priv->tx.cleanup_timer,
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
137
epid_queue = &priv->tx.mgmt_ep_queue;
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
139
epid_queue = &priv->tx.cab_ep_queue;
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
141
epid_queue = &priv->tx.data_be_queue;
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
143
epid_queue = &priv->tx.data_bk_queue;
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
145
epid_queue = &priv->tx.data_vi_queue;
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
147
epid_queue = &priv->tx.data_vo_queue;
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
397
spin_lock_bh(&priv->tx.tx_lock);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
400
spin_unlock_bh(&priv->tx.tx_lock);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
435
spin_lock_bh(&priv->tx.tx_lock);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
437
spin_unlock_bh(&priv->tx.tx_lock);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
509
spin_lock_bh(&priv->tx.tx_lock);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
510
if (WARN_ON(--priv->tx.queued_cnt < 0))
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
511
priv->tx.queued_cnt = 0;
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
512
spin_unlock_bh(&priv->tx.tx_lock);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
543
spin_lock_bh(&priv->tx.tx_lock);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
544
priv->tx.flags |= ATH9K_HTC_OP_TX_DRAIN;
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
545
spin_unlock_bh(&priv->tx.tx_lock);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
555
ath9k_htc_tx_drainq(priv, &priv->tx.mgmt_ep_queue);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
556
ath9k_htc_tx_drainq(priv, &priv->tx.cab_ep_queue);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
557
ath9k_htc_tx_drainq(priv, &priv->tx.data_be_queue);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
558
ath9k_htc_tx_drainq(priv, &priv->tx.data_bk_queue);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
559
ath9k_htc_tx_drainq(priv, &priv->tx.data_vi_queue);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
560
ath9k_htc_tx_drainq(priv, &priv->tx.data_vo_queue);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
561
ath9k_htc_tx_drainq(priv, &priv->tx.tx_failed);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
573
spin_lock_bh(&priv->tx.tx_lock);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
574
priv->tx.flags &= ~ATH9K_HTC_OP_TX_DRAIN;
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
575
spin_unlock_bh(&priv->tx.tx_lock);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
58
spin_lock_bh(&priv->tx.tx_lock);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
582
spin_lock(&priv->tx.tx_lock);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
583
if (priv->tx.flags & ATH9K_HTC_OP_TX_DRAIN) {
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
584
spin_unlock(&priv->tx.tx_lock);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
587
spin_unlock(&priv->tx.tx_lock);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
589
ath9k_htc_tx_drainq(priv, &priv->tx.tx_failed);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
59
priv->tx.queued_cnt++;
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
60
if ((priv->tx.queued_cnt >= ATH9K_HTC_TX_THRESHOLD) &&
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
61
!(priv->tx.flags & ATH9K_HTC_OP_TX_QUEUES_STOP)) {
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
62
priv->tx.flags |= ATH9K_HTC_OP_TX_QUEUES_STOP;
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
65
spin_unlock_bh(&priv->tx.tx_lock);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
70
spin_lock_bh(&priv->tx.tx_lock);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
702
skb_queue_tail(&priv->tx.tx_failed, skb);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
71
if ((priv->tx.queued_cnt < ATH9K_HTC_TX_THRESHOLD) &&
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
72
(priv->tx.flags & ATH9K_HTC_OP_TX_QUEUES_STOP)) {
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
73
priv->tx.flags &= ~ATH9K_HTC_OP_TX_QUEUES_STOP;
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
76
spin_unlock_bh(&priv->tx.tx_lock);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
764
tx.cleanup_timer);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
795
ath9k_htc_tx_cleanup_queue(priv, &priv->tx.mgmt_ep_queue);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
796
ath9k_htc_tx_cleanup_queue(priv, &priv->tx.cab_ep_queue);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
797
ath9k_htc_tx_cleanup_queue(priv, &priv->tx.data_be_queue);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
798
ath9k_htc_tx_cleanup_queue(priv, &priv->tx.data_bk_queue);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
799
ath9k_htc_tx_cleanup_queue(priv, &priv->tx.data_vi_queue);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
800
ath9k_htc_tx_cleanup_queue(priv, &priv->tx.data_vo_queue);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
805
mod_timer(&priv->tx.cleanup_timer,
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
811
skb_queue_head_init(&priv->tx.mgmt_ep_queue);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
812
skb_queue_head_init(&priv->tx.cab_ep_queue);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
813
skb_queue_head_init(&priv->tx.data_be_queue);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
814
skb_queue_head_init(&priv->tx.data_bk_queue);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
815
skb_queue_head_init(&priv->tx.data_vi_queue);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
816
skb_queue_head_init(&priv->tx.data_vo_queue);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
817
skb_queue_head_init(&priv->tx.tx_failed);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
83
spin_lock_bh(&priv->tx.tx_lock);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
84
slot = find_first_zero_bit(priv->tx.tx_slot, MAX_TX_BUF_NUM);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
86
spin_unlock_bh(&priv->tx.tx_lock);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
89
__set_bit(slot, priv->tx.tx_slot);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
90
spin_unlock_bh(&priv->tx.tx_lock);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
97
spin_lock_bh(&priv->tx.tx_lock);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
98
__clear_bit(slot, priv->tx.tx_slot);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
99
spin_unlock_bh(&priv->tx.tx_lock);
drivers/net/wireless/ath/ath9k/htc_hst.c
361
if (endpoint->ep_callbacks.tx) {
drivers/net/wireless/ath/ath9k/htc_hst.c
362
endpoint->ep_callbacks.tx(endpoint->ep_callbacks.priv,
drivers/net/wireless/ath/ath9k/htc_hst.h
94
void (*tx) (void *, struct sk_buff *, enum htc_endpoint_id, bool txok);
drivers/net/wireless/ath/ath9k/hw.h
1132
void ar9003_hw_set_chain_masks(struct ath_hw *ah, u8 rx, u8 tx);
drivers/net/wireless/ath/ath9k/init.c
1100
ath_tx_cleanupq(sc, &sc->tx.txq[i]);
drivers/net/wireless/ath/ath9k/init.c
413
sc->tx.uapsdq = ath_txq_setup(sc, ATH9K_TX_QUEUE_UAPSD, 0);
drivers/net/wireless/ath/ath9k/init.c
416
sc->tx.txq_map[i] = ath_txq_setup(sc, ATH9K_TX_QUEUE_DATA, i);
drivers/net/wireless/ath/ath9k/init.c
417
sc->tx.txq_map[i]->mac80211_qnum = i;
drivers/net/wireless/ath/ath9k/init.c
789
ath_tx_cleanupq(sc, &sc->tx.txq[i]);
drivers/net/wireless/ath/ath9k/link.c
211
txctl.txq = sc->tx.txq_map[IEEE80211_AC_BE];
drivers/net/wireless/ath/ath9k/link.c
32
txq = sc->tx.txq_map[i];
drivers/net/wireless/ath/ath9k/mac.h
303
} tx;
drivers/net/wireless/ath/ath9k/mac.h
321
#define ds_ctl2 u.tx.ctl2
drivers/net/wireless/ath/ath9k/mac.h
322
#define ds_ctl3 u.tx.ctl3
drivers/net/wireless/ath/ath9k/mac.h
323
#define ds_ctl4 u.tx.ctl4
drivers/net/wireless/ath/ath9k/mac.h
324
#define ds_ctl5 u.tx.ctl5
drivers/net/wireless/ath/ath9k/mac.h
325
#define ds_ctl6 u.tx.ctl6
drivers/net/wireless/ath/ath9k/mac.h
326
#define ds_ctl7 u.tx.ctl7
drivers/net/wireless/ath/ath9k/mac.h
327
#define ds_ctl8 u.tx.ctl8
drivers/net/wireless/ath/ath9k/mac.h
328
#define ds_ctl9 u.tx.ctl9
drivers/net/wireless/ath/ath9k/mac.h
329
#define ds_ctl10 u.tx.ctl10
drivers/net/wireless/ath/ath9k/mac.h
330
#define ds_ctl11 u.tx.ctl11
drivers/net/wireless/ath/ath9k/mac.h
332
#define ds_txstatus0 u.tx.status0
drivers/net/wireless/ath/ath9k/mac.h
333
#define ds_txstatus1 u.tx.status1
drivers/net/wireless/ath/ath9k/mac.h
334
#define ds_txstatus2 u.tx.status2
drivers/net/wireless/ath/ath9k/mac.h
335
#define ds_txstatus3 u.tx.status3
drivers/net/wireless/ath/ath9k/mac.h
336
#define ds_txstatus4 u.tx.status4
drivers/net/wireless/ath/ath9k/mac.h
337
#define ds_txstatus5 u.tx.status5
drivers/net/wireless/ath/ath9k/mac.h
338
#define ds_txstatus6 u.tx.status6
drivers/net/wireless/ath/ath9k/mac.h
339
#define ds_txstatus7 u.tx.status7
drivers/net/wireless/ath/ath9k/mac.h
340
#define ds_txstatus8 u.tx.status8
drivers/net/wireless/ath/ath9k/mac.h
341
#define ds_txstatus9 u.tx.status9
drivers/net/wireless/ath/ath9k/main.c
1727
txq = sc->tx.txq_map[queue];
drivers/net/wireless/ath/ath9k/main.c
2156
npend = ath9k_has_pending_frames(sc, &sc->tx.txq[i],
drivers/net/wireless/ath/ath9k/main.c
2795
.tx = ath9k_tx,
drivers/net/wireless/ath/ath9k/main.c
814
txctl.txq = sc->tx.txq_map[skb_get_queue_mapping(skb)];
drivers/net/wireless/ath/ath9k/main.c
859
txq = &sc->tx.txq[i];
drivers/net/wireless/ath/ath9k/tx99.c
126
txctl.txq = sc->tx.txq_map[IEEE80211_AC_VO];
drivers/net/wireless/ath/ath9k/wmi.c
178
spin_lock_bh(&priv->tx.tx_lock);
drivers/net/wireless/ath/ath9k/wmi.c
179
if (priv->tx.flags & ATH9K_HTC_OP_TX_DRAIN) {
drivers/net/wireless/ath/ath9k/wmi.c
180
spin_unlock_bh(&priv->tx.tx_lock);
drivers/net/wireless/ath/ath9k/wmi.c
183
spin_unlock_bh(&priv->tx.tx_lock);
drivers/net/wireless/ath/ath9k/wmi.c
273
connect.ep_callbacks.tx = ath9k_wmi_ctrl_tx;
drivers/net/wireless/ath/ath9k/xmit.c
1162
cur_ht20 = sc->tx.max_aggr_framelen[queue][MCS_HT20];
drivers/net/wireless/ath/ath9k/xmit.c
1163
cur_ht20_sgi = sc->tx.max_aggr_framelen[queue][MCS_HT20_SGI];
drivers/net/wireless/ath/ath9k/xmit.c
1164
cur_ht40 = sc->tx.max_aggr_framelen[queue][MCS_HT40];
drivers/net/wireless/ath/ath9k/xmit.c
1165
cur_ht40_sgi = sc->tx.max_aggr_framelen[queue][MCS_HT40_SGI];
drivers/net/wireless/ath/ath9k/xmit.c
1435
txq == sc->tx.uapsdq)
drivers/net/wireless/ath/ath9k/xmit.c
1698
struct ath_txq *txq = sc->tx.uapsdq;
drivers/net/wireless/ath/ath9k/xmit.c
1715
ret = ath_tx_get_tid_subframe(sc, sc->tx.uapsdq,
drivers/net/wireless/ath/ath9k/xmit.c
1810
struct ath_txq *txq = &sc->tx.txq[axq_qnum];
drivers/net/wireless/ath/ath9k/xmit.c
1821
sc->tx.txqsetup |= 1<<axq_qnum;
drivers/net/wireless/ath/ath9k/xmit.c
1827
return &sc->tx.txq[axq_qnum];
drivers/net/wireless/ath/ath9k/xmit.c
1837
BUG_ON(sc->tx.txq[qnum].axq_qnum != qnum);
drivers/net/wireless/ath/ath9k/xmit.c
1947
if (!sc->tx.txq[i].axq_depth)
drivers/net/wireless/ath/ath9k/xmit.c
1950
if (ath9k_hw_numtxpending(ah, sc->tx.txq[i].axq_qnum))
drivers/net/wireless/ath/ath9k/xmit.c
1964
txq = &sc->tx.txq[i];
drivers/net/wireless/ath/ath9k/xmit.c
1974
sc->tx.txqsetup &= ~(1<<txq->axq_qnum);
drivers/net/wireless/ath/ath9k/xmit.c
2025
txq = sc->tx.txq_map[i];
drivers/net/wireless/ath/ath9k/xmit.c
219
txq = sc->tx.txq_map[q];
drivers/net/wireless/ath/ath9k/xmit.c
2354
txq = sc->tx.uapsdq;
drivers/net/wireless/ath/ath9k/xmit.c
2362
if (txq == sc->tx.txq_map[q]) {
drivers/net/wireless/ath/ath9k/xmit.c
2550
spin_lock_irqsave(&sc->tx.txbuflock, flags);
drivers/net/wireless/ath/ath9k/xmit.c
2551
list_splice_tail_init(bf_q, &sc->tx.txbuf);
drivers/net/wireless/ath/ath9k/xmit.c
2552
spin_unlock_irqrestore(&sc->tx.txbuflock, flags);
drivers/net/wireless/ath/ath9k/xmit.c
257
if (tid->txq == sc->tx.txq_map[q]) {
drivers/net/wireless/ath/ath9k/xmit.c
2706
ath_tx_processq(sc, &sc->tx.txq[i]);
drivers/net/wireless/ath/ath9k/xmit.c
2749
txq = &sc->tx.txq[ts.qid];
drivers/net/wireless/ath/ath9k/xmit.c
2832
spin_lock_init(&sc->tx.txbuflock);
drivers/net/wireless/ath/ath9k/xmit.c
2834
error = ath_descdma_setup(sc, &sc->tx.txdma, &sc->tx.txbuf,
drivers/net/wireless/ath/ath9k/xmit.c
2873
tid->txq = sc->tx.txq_map[acno];
drivers/net/wireless/ath/ath9k/xmit.c
411
spin_lock_bh(&sc->tx.txbuflock);
drivers/net/wireless/ath/ath9k/xmit.c
413
if (unlikely(list_empty(&sc->tx.txbuf))) {
drivers/net/wireless/ath/ath9k/xmit.c
414
spin_unlock_bh(&sc->tx.txbuflock);
drivers/net/wireless/ath/ath9k/xmit.c
418
bf = list_first_entry(&sc->tx.txbuf, struct ath_buf, list);
drivers/net/wireless/ath/ath9k/xmit.c
421
spin_unlock_bh(&sc->tx.txbuflock);
drivers/net/wireless/ath/ath9k/xmit.c
428
spin_lock_bh(&sc->tx.txbuflock);
drivers/net/wireless/ath/ath9k/xmit.c
429
list_add_tail(&bf->list, &sc->tx.txbuf);
drivers/net/wireless/ath/ath9k/xmit.c
430
spin_unlock_bh(&sc->tx.txbuflock);
drivers/net/wireless/ath/ath9k/xmit.c
836
frmlen = sc->tx.max_aggr_framelen[q][modeidx][rates[i].idx];
drivers/net/wireless/ath/carl9170/main.c
1720
.tx = carl9170_op_tx,
drivers/net/wireless/ath/hw.c
144
u32 cycles, busy, rx, tx;
drivers/net/wireless/ath/hw.c
154
tx = REG_READ(ah, AR_TFCNT);
drivers/net/wireless/ath/hw.c
169
common->cc_ani.tx_frame += tx;
drivers/net/wireless/ath/hw.c
174
common->cc_survey.tx_frame += tx;
drivers/net/wireless/ath/wcn36xx/main.c
1367
.tx = wcn36xx_tx,
drivers/net/wireless/ath/wil6210/cfg80211.c
275
.tx = BIT(IEEE80211_STYPE_ACTION >> 4) |
drivers/net/wireless/ath/wil6210/cfg80211.c
281
.tx = BIT(IEEE80211_STYPE_ACTION >> 4) |
drivers/net/wireless/ath/wil6210/cfg80211.c
296
.tx = BIT(IEEE80211_STYPE_ACTION >> 4) |
drivers/net/wireless/ath/wil6210/cfg80211.c
302
.tx = BIT(IEEE80211_STYPE_ACTION >> 4) |
drivers/net/wireless/ath/wil6210/cfg80211.c
308
.tx = BIT(IEEE80211_STYPE_ACTION >> 4) |
drivers/net/wireless/ath/wil6210/debugfs.c
1066
bool tx;
drivers/net/wireless/ath/wil6210/debugfs.c
1079
tx = ring_idx > 0; /* desc ring 0 is reserved for RX */
drivers/net/wireless/ath/wil6210/debugfs.c
1086
tx = (ring_idx < WIL6210_MAX_TX_RINGS);
drivers/net/wireless/ath/wil6210/debugfs.c
1089
ring = tx ? &wil->ring_tx[ring_idx] : &wil->ring_rx;
drivers/net/wireless/ath/wil6210/debugfs.c
1092
if (tx)
drivers/net/wireless/ath/wil6210/debugfs.c
1100
if (tx)
drivers/net/wireless/ath/wil6210/debugfs.c
1112
d = &ring->va[txdesc_idx].tx.legacy;
drivers/net/wireless/ath/wil6210/debugfs.c
1117
if (tx) {
drivers/net/wireless/ath/wil6210/debugfs.c
1134
if (tx)
drivers/net/wireless/ath/wil6210/debugfs.c
1162
bool tx;
drivers/net/wireless/ath/wil6210/debugfs.c
1172
tx = !sring->is_rx;
drivers/net/wireless/ath/wil6210/debugfs.c
1175
seq_printf(s, "No %cX status ring\n", tx ? 'T' : 'R');
drivers/net/wireless/ath/wil6210/debugfs.c
1181
tx ? 'T' : 'R', status_msg_idx, sring->size);
drivers/net/wireless/ath/wil6210/debugfs.c
1188
tx ? 'T' : 'R', sring_idx, status_msg_idx);
drivers/net/wireless/ath/wil6210/debugfs.c
1192
if (!tx && !wil->use_compressed_rx_status)
drivers/net/wireless/ath/wil6210/debugfs.c
122
&ring->va[i].tx.legacy;
drivers/net/wireless/ath/wil6210/debugfs.c
1452
int tx = atomic_xchg(&wil->isr_count_tx, 0);
drivers/net/wireless/ath/wil6210/debugfs.c
1461
seq_printf(s, "Tx irqs:packets : %8d : %8ld\n", tx, txf - txf_old);
drivers/net/wireless/ath/wil6210/debugfs.c
364
offsetof(struct wil6210_mbox_ctl, tx));
drivers/net/wireless/ath/wil6210/debugfs.c
62
&ring->va[idx].tx.enhanced;
drivers/net/wireless/ath/wil6210/interrupt.c
515
wil_mbox_ring_le2cpus(&wil->mbox_ctl.tx);
drivers/net/wireless/ath/wil6210/interrupt.c
528
if (wil->mbox_ctl.tx.entry_size < min_size) {
drivers/net/wireless/ath/wil6210/interrupt.c
530
wil->mbox_ctl.tx.entry_size);
drivers/net/wireless/ath/wil6210/txrx.c
160
&vring->va[i].tx.legacy;
drivers/net/wireless/ath/wil6210/txrx.c
1818
_hdr_desc = &vring->va[i].tx.legacy;
drivers/net/wireless/ath/wil6210/txrx.c
1884
_desc = &vring->va[i].tx.legacy;
drivers/net/wireless/ath/wil6210/txrx.c
2013
d = (struct vring_tx_desc *)&vring->va[i].tx.legacy;
drivers/net/wireless/ath/wil6210/txrx.c
2014
_desc = &vring->va[i].tx.legacy;
drivers/net/wireless/ath/wil6210/txrx.c
2056
_d = &ring->va[i].tx.legacy;
drivers/net/wireless/ath/wil6210/txrx.c
2096
_d = &ring->va[i].tx.legacy;
drivers/net/wireless/ath/wil6210/txrx.c
216
&vring->va[vring->swtail].tx.legacy;
drivers/net/wireless/ath/wil6210/txrx.c
2171
_d = &ring->va[i].tx.legacy;
drivers/net/wireless/ath/wil6210/txrx.c
2459
_d = &vring->va[lf].tx.legacy;
drivers/net/wireless/ath/wil6210/txrx.c
2471
_d = &vring->va[vring->swtail].tx.legacy;
drivers/net/wireless/ath/wil6210/txrx.h
489
union wil_tx_desc tx;
drivers/net/wireless/ath/wil6210/txrx_edma.c
1231
&ring->va[ring->swtail].tx.enhanced;
drivers/net/wireless/ath/wil6210/txrx_edma.c
1358
&ring->va[i].tx.enhanced;
drivers/net/wireless/ath/wil6210/txrx_edma.c
1533
&ring->va[i].tx.enhanced;
drivers/net/wireless/ath/wil6210/txrx_edma.c
453
&ring->va[ring->swtail].tx.enhanced;
drivers/net/wireless/ath/wil6210/wil6210.h
470
struct wil6210_mbox_ring tx;
drivers/net/wireless/ath/wil6210/wmi.c
650
struct wil6210_mbox_ring *r = &wil->mbox_ctl.tx;
drivers/net/wireless/ath/wil6210/wmi.c
711
offsetof(struct wil6210_mbox_ctl, tx.tail));
drivers/net/wireless/ath/wil6210/wmi.c
741
wil_w(wil, RGF_MBOX + offsetof(struct wil6210_mbox_ctl, tx.head),
drivers/net/wireless/atmel/at76c50x-usb.c
2179
.tx = at76_mac80211_tx,
drivers/net/wireless/broadcom/b43/dma.c
1359
B43_WARN_ON(!ring->tx);
drivers/net/wireless/broadcom/b43/dma.c
1429
B43_WARN_ON(!ring->tx);
drivers/net/wireless/broadcom/b43/dma.c
1671
B43_WARN_ON(ring->tx);
drivers/net/wireless/broadcom/b43/dma.c
1692
B43_WARN_ON(ring->tx);
drivers/net/wireless/broadcom/b43/dma.c
1708
B43_WARN_ON(!ring->tx);
drivers/net/wireless/broadcom/b43/dma.c
1714
B43_WARN_ON(!ring->tx);
drivers/net/wireless/broadcom/b43/dma.c
289
ring->tx ? "TX" : "RX", ring->index);
drivers/net/wireless/broadcom/b43/dma.c
304
B43_WARN_ON(!ring->tx);
drivers/net/wireless/broadcom/b43/dma.c
348
unsigned char *buf, size_t len, int tx)
drivers/net/wireless/broadcom/b43/dma.c
352
if (tx) {
drivers/net/wireless/broadcom/b43/dma.c
365
dma_addr_t addr, size_t len, int tx)
drivers/net/wireless/broadcom/b43/dma.c
367
if (tx) {
drivers/net/wireless/broadcom/b43/dma.c
380
B43_WARN_ON(ring->tx);
drivers/net/wireless/broadcom/b43/dma.c
389
B43_WARN_ON(ring->tx);
drivers/net/wireless/broadcom/b43/dma.c
399
if (ring->tx)
drivers/net/wireless/broadcom/b43/dma.c
599
B43_WARN_ON(ring->tx);
drivers/net/wireless/broadcom/b43/dma.c
681
if (ring->tx) {
drivers/net/wireless/broadcom/b43/dma.c
755
if (ring->tx) {
drivers/net/wireless/broadcom/b43/dma.c
786
B43_WARN_ON(!ring->tx);
drivers/net/wireless/broadcom/b43/dma.c
789
if (ring->tx) {
drivers/net/wireless/broadcom/b43/dma.c
864
ring->tx = true;
drivers/net/wireless/broadcom/b43/dma.h
255
bool tx;
drivers/net/wireless/broadcom/b43/main.c
5179
.tx = b43_op_tx,
drivers/net/wireless/broadcom/b43/phy_lp.c
1839
static int lpphy_rx_iq_cal(struct b43_wldev *dev, bool noise, bool tx,
drivers/net/wireless/broadcom/b43/phy_lp.c
1875
tx = true;
drivers/net/wireless/broadcom/b43/phy_lp.c
1880
lpphy_set_trsw_over(dev, tx, rx);
drivers/net/wireless/broadcom/b43/phy_lp.c
746
static void lpphy_set_trsw_over(struct b43_wldev *dev, bool tx, bool rx)
drivers/net/wireless/broadcom/b43/phy_lp.c
748
u16 trsw = (tx << 1) | rx;
drivers/net/wireless/broadcom/b43/radio_2056.c
10229
B2056_TX0, pts->tx, pts->tx_length);
drivers/net/wireless/broadcom/b43/radio_2056.c
10231
B2056_TX1, pts->tx, pts->tx_length);
drivers/net/wireless/broadcom/b43/radio_2056.c
3011
.tx = prefix##_tx, \
drivers/net/wireless/broadcom/b43/radio_2056.c
32
const struct b2056_inittab_entry *tx;
drivers/net/wireless/broadcom/b43legacy/dma.c
1054
B43legacy_WARN_ON(!ring->tx);
drivers/net/wireless/broadcom/b43legacy/dma.c
1112
B43legacy_WARN_ON(!ring->tx);
drivers/net/wireless/broadcom/b43legacy/dma.c
1323
B43legacy_WARN_ON(ring->tx);
drivers/net/wireless/broadcom/b43legacy/dma.c
1339
B43legacy_WARN_ON(!ring->tx);
drivers/net/wireless/broadcom/b43legacy/dma.c
1345
B43legacy_WARN_ON(!ring->tx);
drivers/net/wireless/broadcom/b43legacy/dma.c
141
ring->tx ? "TX" : "RX",
drivers/net/wireless/broadcom/b43legacy/dma.c
157
B43legacy_WARN_ON(!ring->tx);
drivers/net/wireless/broadcom/b43legacy/dma.c
229
int tx)
drivers/net/wireless/broadcom/b43legacy/dma.c
233
if (tx)
drivers/net/wireless/broadcom/b43legacy/dma.c
249
int tx)
drivers/net/wireless/broadcom/b43legacy/dma.c
251
if (tx)
drivers/net/wireless/broadcom/b43legacy/dma.c
266
B43legacy_WARN_ON(ring->tx);
drivers/net/wireless/broadcom/b43legacy/dma.c
277
B43legacy_WARN_ON(ring->tx);
drivers/net/wireless/broadcom/b43legacy/dma.c
429
B43legacy_WARN_ON(ring->tx);
drivers/net/wireless/broadcom/b43legacy/dma.c
514
if (ring->tx) {
drivers/net/wireless/broadcom/b43legacy/dma.c
550
if (ring->tx) {
drivers/net/wireless/broadcom/b43legacy/dma.c
572
B43legacy_WARN_ON(!ring->tx);
drivers/net/wireless/broadcom/b43legacy/dma.c
575
if (ring->tx)
drivers/net/wireless/broadcom/b43legacy/dma.c
667
ring->tx = true;
drivers/net/wireless/broadcom/b43legacy/dma.c
713
(ring->tx) ? "TX" : "RX", ring->max_used_slots,
drivers/net/wireless/broadcom/b43legacy/dma.h
149
bool tx;
drivers/net/wireless/broadcom/b43legacy/main.c
3537
.tx = b43legacy_op_tx,
drivers/net/wireless/broadcom/b43legacy/phy.c
1291
u16 tx)
drivers/net/wireless/broadcom/b43legacy/phy.c
1300
if (tx == 3)
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
112
u32 tx;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
7397
.tx = 0xffff,
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
7403
.tx = 0xffff,
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
7408
.tx = 0xffff,
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
7418
.tx = 0xffff,
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
7423
.tx = 0xffff,
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
8049
survey->tx = val;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
8220
info->time_tx = survey.tx;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c
310
static void brcmf_p2p_print_actframe(bool tx, void *frame, u32 frame_len)
drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c
324
(tx) ? "TX" : "RX");
drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c
328
(tx) ? "TX" : "RX");
drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c
332
(tx) ? "TX" : "RX");
drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c
336
(tx) ? "TX" : "RX");
drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c
340
(tx) ? "TX" : "RX");
drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c
344
(tx) ? "TX" : "RX");
drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c
348
(tx) ? "TX" : "RX");
drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c
352
(tx) ? "TX" : "RX");
drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c
356
(tx) ? "TX" : "RX");
drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c
360
(tx) ? "TX" : "RX");
drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c
368
(tx) ? "TX" : "RX");
drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c
372
(tx) ? "TX" : "RX");
drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c
376
(tx) ? "TX" : "RX");
drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c
380
(tx) ? "TX" : "RX");
drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c
384
(tx) ? "TX" : "RX");
drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c
392
(tx) ? "TX" : "RX");
drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c
396
(tx) ? "TX" : "RX");
drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c
400
(tx) ? "TX" : "RX");
drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c
404
(tx) ? "TX" : "RX");
drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c
408
(tx) ? "TX" : "RX");
drivers/net/wireless/broadcom/brcm80211/brcmfmac/p2p.c
416
static void brcmf_p2p_print_actframe(bool tx, void *frame, u32 frame_len)
drivers/net/wireless/broadcom/brcm80211/brcmsmac/d11.h
70
struct pio2regs tx;
drivers/net/wireless/broadcom/brcm80211/brcmsmac/d11.h
82
struct pio4regs tx;
drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c
968
.tx = brcms_ops_tx,
drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c
1097
static void wlc_lcnphy_set_trsw_override(struct brcms_phy *pi, bool tx, bool rx)
drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c
1102
(0x1 << 0), (tx ? (0x1 << 1) : 0) | (rx ? (0x1 << 0) : 0));
drivers/net/wireless/intel/iwlegacy/3945-debug.c
307
struct iwl39_stats_tx *tx, *accum_tx, *delta_tx, *max_tx;
drivers/net/wireless/intel/iwlegacy/3945-debug.c
323
tx = &il->_3945.stats.tx;
drivers/net/wireless/intel/iwlegacy/3945-debug.c
324
accum_tx = &il->_3945.accum_stats.tx;
drivers/net/wireless/intel/iwlegacy/3945-debug.c
325
delta_tx = &il->_3945.delta_stats.tx;
drivers/net/wireless/intel/iwlegacy/3945-debug.c
326
max_tx = &il->_3945.max_delta.tx;
drivers/net/wireless/intel/iwlegacy/3945-debug.c
336
le32_to_cpu(tx->preamble_cnt), accum_tx->preamble_cnt,
drivers/net/wireless/intel/iwlegacy/3945-debug.c
341
le32_to_cpu(tx->rx_detected_cnt),
drivers/net/wireless/intel/iwlegacy/3945-debug.c
347
le32_to_cpu(tx->bt_prio_defer_cnt),
drivers/net/wireless/intel/iwlegacy/3945-debug.c
353
le32_to_cpu(tx->bt_prio_kill_cnt),
drivers/net/wireless/intel/iwlegacy/3945-debug.c
359
le32_to_cpu(tx->few_bytes_cnt), accum_tx->few_bytes_cnt,
drivers/net/wireless/intel/iwlegacy/3945-debug.c
364
le32_to_cpu(tx->cts_timeout), accum_tx->cts_timeout,
drivers/net/wireless/intel/iwlegacy/3945-debug.c
369
le32_to_cpu(tx->ack_timeout), accum_tx->ack_timeout,
drivers/net/wireless/intel/iwlegacy/3945-debug.c
374
le32_to_cpu(tx->expected_ack_cnt),
drivers/net/wireless/intel/iwlegacy/3945-debug.c
380
le32_to_cpu(tx->actual_ack_cnt), accum_tx->actual_ack_cnt,
drivers/net/wireless/intel/iwlegacy/3945-mac.c
1445
il->isr_stats.tx++;
drivers/net/wireless/intel/iwlegacy/3945-mac.c
3403
.tx = il3945_mac_tx,
drivers/net/wireless/intel/iwlegacy/3945.c
2423
tx_beacon_cmd->tx.sta_id = il->hw_params.bcast_id;
drivers/net/wireless/intel/iwlegacy/3945.c
2424
tx_beacon_cmd->tx.stop_time.life_time = TX_CMD_LIFE_TIME_INFINITE;
drivers/net/wireless/intel/iwlegacy/3945.c
2431
tx_beacon_cmd->tx.len = cpu_to_le16((u16) frame_size);
drivers/net/wireless/intel/iwlegacy/3945.c
2433
tx_beacon_cmd->tx.rate = rate;
drivers/net/wireless/intel/iwlegacy/3945.c
2434
tx_beacon_cmd->tx.tx_flags =
drivers/net/wireless/intel/iwlegacy/3945.c
2438
tx_beacon_cmd->tx.supp_rates[0] =
drivers/net/wireless/intel/iwlegacy/3945.c
2441
tx_beacon_cmd->tx.supp_rates[1] = (IL_CCK_BASIC_RATES_MASK & 0xF);
drivers/net/wireless/intel/iwlegacy/4965-debug.c
461
struct stats_tx *tx, *accum_tx, *delta_tx, *max_tx;
drivers/net/wireless/intel/iwlegacy/4965-debug.c
476
tx = &il->_4965.stats.tx;
drivers/net/wireless/intel/iwlegacy/4965-debug.c
477
accum_tx = &il->_4965.accum_stats.tx;
drivers/net/wireless/intel/iwlegacy/4965-debug.c
478
delta_tx = &il->_4965.delta_stats.tx;
drivers/net/wireless/intel/iwlegacy/4965-debug.c
479
max_tx = &il->_4965.max_delta.tx;
drivers/net/wireless/intel/iwlegacy/4965-debug.c
485
le32_to_cpu(tx->preamble_cnt), accum_tx->preamble_cnt,
drivers/net/wireless/intel/iwlegacy/4965-debug.c
489
le32_to_cpu(tx->rx_detected_cnt),
drivers/net/wireless/intel/iwlegacy/4965-debug.c
494
le32_to_cpu(tx->bt_prio_defer_cnt),
drivers/net/wireless/intel/iwlegacy/4965-debug.c
499
le32_to_cpu(tx->bt_prio_kill_cnt),
drivers/net/wireless/intel/iwlegacy/4965-debug.c
504
le32_to_cpu(tx->few_bytes_cnt), accum_tx->few_bytes_cnt,
drivers/net/wireless/intel/iwlegacy/4965-debug.c
508
le32_to_cpu(tx->cts_timeout), accum_tx->cts_timeout,
drivers/net/wireless/intel/iwlegacy/4965-debug.c
512
le32_to_cpu(tx->ack_timeout), accum_tx->ack_timeout,
drivers/net/wireless/intel/iwlegacy/4965-debug.c
516
le32_to_cpu(tx->expected_ack_cnt),
drivers/net/wireless/intel/iwlegacy/4965-debug.c
521
le32_to_cpu(tx->actual_ack_cnt), accum_tx->actual_ack_cnt,
drivers/net/wireless/intel/iwlegacy/4965-debug.c
525
le32_to_cpu(tx->dump_msdu_cnt), accum_tx->dump_msdu_cnt,
drivers/net/wireless/intel/iwlegacy/4965-debug.c
530
le32_to_cpu(tx->burst_abort_next_frame_mismatch_cnt),
drivers/net/wireless/intel/iwlegacy/4965-debug.c
537
le32_to_cpu(tx->burst_abort_missing_next_frame_cnt),
drivers/net/wireless/intel/iwlegacy/4965-debug.c
544
le32_to_cpu(tx->cts_timeout_collision),
drivers/net/wireless/intel/iwlegacy/4965-debug.c
551
le32_to_cpu(tx->ack_or_ba_timeout_collision),
drivers/net/wireless/intel/iwlegacy/4965-debug.c
557
le32_to_cpu(tx->agg.ba_timeout), accum_tx->agg.ba_timeout,
drivers/net/wireless/intel/iwlegacy/4965-debug.c
562
le32_to_cpu(tx->agg.ba_reschedule_frames),
drivers/net/wireless/intel/iwlegacy/4965-debug.c
569
le32_to_cpu(tx->agg.scd_query_agg_frame_cnt),
drivers/net/wireless/intel/iwlegacy/4965-debug.c
576
le32_to_cpu(tx->agg.scd_query_no_agg),
drivers/net/wireless/intel/iwlegacy/4965-debug.c
582
le32_to_cpu(tx->agg.scd_query_agg),
drivers/net/wireless/intel/iwlegacy/4965-debug.c
588
le32_to_cpu(tx->agg.scd_query_mismatch),
drivers/net/wireless/intel/iwlegacy/4965-debug.c
594
le32_to_cpu(tx->agg.frame_not_ready),
drivers/net/wireless/intel/iwlegacy/4965-debug.c
600
le32_to_cpu(tx->agg.underrun), accum_tx->agg.underrun,
drivers/net/wireless/intel/iwlegacy/4965-debug.c
604
le32_to_cpu(tx->agg.bt_prio_kill),
drivers/net/wireless/intel/iwlegacy/4965-debug.c
609
le32_to_cpu(tx->agg.rx_ba_rsp_cnt),
drivers/net/wireless/intel/iwlegacy/4965-mac.c
1775
tx_cmd = container_of(&out_cmd->cmd.tx, struct il_tx_cmd, __hdr);
drivers/net/wireless/intel/iwlegacy/4965-mac.c
3812
tx_beacon_cmd->tx.len = cpu_to_le16((u16) frame_size);
drivers/net/wireless/intel/iwlegacy/4965-mac.c
3813
tx_beacon_cmd->tx.sta_id = il->hw_params.bcast_id;
drivers/net/wireless/intel/iwlegacy/4965-mac.c
3814
tx_beacon_cmd->tx.stop_time.life_time = TX_CMD_LIFE_TIME_INFINITE;
drivers/net/wireless/intel/iwlegacy/4965-mac.c
3815
tx_beacon_cmd->tx.tx_flags =
drivers/net/wireless/intel/iwlegacy/4965-mac.c
3829
tx_beacon_cmd->tx.rate_n_flags = cpu_to_le32(rate | rate_flags);
drivers/net/wireless/intel/iwlegacy/4965-mac.c
4486
il->isr_stats.tx++;
drivers/net/wireless/intel/iwlegacy/4965-mac.c
6313
.tx = il4965_mac_tx,
drivers/net/wireless/intel/iwlegacy/commands.h
2678
struct il3945_tx_cmd_hdr tx;
drivers/net/wireless/intel/iwlegacy/commands.h
2686
struct il_tx_cmd_hdr tx;
drivers/net/wireless/intel/iwlegacy/commands.h
2995
struct iwl39_stats_tx tx;
drivers/net/wireless/intel/iwlegacy/commands.h
3002
struct stats_tx tx;
drivers/net/wireless/intel/iwlegacy/common.h
1026
u32 tx;
drivers/net/wireless/intel/iwlegacy/common.h
556
struct il_tx_cmd_hdr tx;
drivers/net/wireless/intel/iwlegacy/debug.c
704
il->isr_stats.tx);
drivers/net/wireless/intel/iwlwifi/dvm/commands.h
2428
struct iwl_tx_cmd_hdr tx;
drivers/net/wireless/intel/iwlwifi/dvm/commands.h
2715
struct statistics_tx tx;
drivers/net/wireless/intel/iwlwifi/dvm/commands.h
2722
struct statistics_tx tx;
drivers/net/wireless/intel/iwlwifi/dvm/commands.h
3790
u8 tx[IWLAGN_MIC_KEY_SIZE];
drivers/net/wireless/intel/iwlwifi/dvm/commands.h
3803
struct iwlagn_p1k_cache tx;
drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c
1108
struct statistics_tx *tx, *accum_tx, *delta_tx, *max_tx;
drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c
1123
tx = &priv->statistics.tx;
drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c
1124
accum_tx = &priv->accum_stats.tx;
drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c
1125
delta_tx = &priv->delta_stats.tx;
drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c
1126
max_tx = &priv->max_delta_stats.tx;
drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c
1133
le32_to_cpu(tx->preamble_cnt),
drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c
1138
le32_to_cpu(tx->rx_detected_cnt),
drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c
1143
le32_to_cpu(tx->bt_prio_defer_cnt),
drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c
1149
le32_to_cpu(tx->bt_prio_kill_cnt),
drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c
1155
le32_to_cpu(tx->few_bytes_cnt),
drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c
1160
le32_to_cpu(tx->cts_timeout), accum_tx->cts_timeout,
drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c
1164
le32_to_cpu(tx->ack_timeout),
drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c
1169
le32_to_cpu(tx->expected_ack_cnt),
drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c
1175
le32_to_cpu(tx->actual_ack_cnt),
drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c
1181
le32_to_cpu(tx->dump_msdu_cnt),
drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c
1187
le32_to_cpu(tx->burst_abort_next_frame_mismatch_cnt),
drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c
1193
le32_to_cpu(tx->burst_abort_missing_next_frame_cnt),
drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c
1199
le32_to_cpu(tx->cts_timeout_collision),
drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c
1205
le32_to_cpu(tx->ack_or_ba_timeout_collision),
drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c
1211
le32_to_cpu(tx->agg.ba_timeout),
drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c
1217
le32_to_cpu(tx->agg.ba_reschedule_frames),
drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c
1223
le32_to_cpu(tx->agg.scd_query_agg_frame_cnt),
drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c
1229
le32_to_cpu(tx->agg.scd_query_no_agg),
drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c
1235
le32_to_cpu(tx->agg.scd_query_agg),
drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c
1241
le32_to_cpu(tx->agg.scd_query_mismatch),
drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c
1247
le32_to_cpu(tx->agg.frame_not_ready),
drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c
1253
le32_to_cpu(tx->agg.underrun),
drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c
1258
le32_to_cpu(tx->agg.bt_prio_kill),
drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c
1264
le32_to_cpu(tx->agg.rx_ba_rsp_cnt),
drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c
1269
if (tx->tx_power.ant_a || tx->tx_power.ant_b || tx->tx_power.ant_c) {
drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c
1273
tx->tx_power.ant_a)
drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c
1276
tx->tx_power.ant_a);
drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c
1278
tx->tx_power.ant_b)
drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c
1281
tx->tx_power.ant_b);
drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c
1283
tx->tx_power.ant_c)
drivers/net/wireless/intel/iwlwifi/dvm/debugfs.c
1286
tx->tx_power.ant_c);
drivers/net/wireless/intel/iwlwifi/dvm/dev.h
765
struct statistics_tx tx;
drivers/net/wireless/intel/iwlwifi/dvm/dev.h
779
struct statistics_tx tx;
drivers/net/wireless/intel/iwlwifi/dvm/lib.c
918
iwlagn_convert_p1k(p1k, data->tkip->tx.p1k);
drivers/net/wireless/intel/iwlwifi/dvm/lib.c
920
memcpy(data->tkip->mic_keys.tx,
drivers/net/wireless/intel/iwlwifi/dvm/mac80211.c
1577
.tx = iwlagn_mac_tx,
drivers/net/wireless/intel/iwlwifi/dvm/main.c
216
tx_beacon_cmd->tx.len = cpu_to_le16((u16)frame_size);
drivers/net/wireless/intel/iwlwifi/dvm/main.c
217
tx_beacon_cmd->tx.sta_id = priv->beacon_ctx->bcast_sta_id;
drivers/net/wireless/intel/iwlwifi/dvm/main.c
218
tx_beacon_cmd->tx.stop_time.life_time = TX_CMD_LIFE_TIME_INFINITE;
drivers/net/wireless/intel/iwlwifi/dvm/main.c
219
tx_beacon_cmd->tx.tx_flags = TX_CMD_FLG_SEQ_CTL_MSK |
drivers/net/wireless/intel/iwlwifi/dvm/main.c
250
tx_beacon_cmd->tx.rate_n_flags =
drivers/net/wireless/intel/iwlwifi/dvm/rx.c
210
struct statistics_tx *tx,
drivers/net/wireless/intel/iwlwifi/dvm/rx.c
307
struct statistics_tx *tx,
drivers/net/wireless/intel/iwlwifi/dvm/rx.c
323
ACCUM(tx);
drivers/net/wireless/intel/iwlwifi/dvm/rx.c
336
struct statistics_tx *tx,
drivers/net/wireless/intel/iwlwifi/dvm/rx.c
356
struct statistics_tx *tx;
drivers/net/wireless/intel/iwlwifi/dvm/rx.c
373
tx = &stats->tx;
drivers/net/wireless/intel/iwlwifi/dvm/rx.c
391
tx = &stats->tx;
drivers/net/wireless/intel/iwlwifi/dvm/rx.c
406
rx_ofdm_ht, rx_cck, tx, bt_activity);
drivers/net/wireless/intel/iwlwifi/dvm/rx.c
408
iwlagn_recover_from_statistics(priv, rx_ofdm, rx_ofdm_ht, tx, stamp);
drivers/net/wireless/intel/iwlwifi/dvm/rx.c
416
memcpy(&priv->statistics.tx, tx, sizeof(*tx));
drivers/net/wireless/intel/iwlwifi/fw/api/d3.h
470
u8 tx[IWL_MIC_KEY_SIZE];
drivers/net/wireless/intel/iwlwifi/fw/api/d3.h
484
struct iwl_p1k_cache tx;
drivers/net/wireless/intel/iwlwifi/fw/api/d3.h
491
struct iwl_p1k_cache tx;
drivers/net/wireless/intel/iwlwifi/fw/api/stats.h
365
struct mvm_statistics_tx_v4 tx;
drivers/net/wireless/intel/iwlwifi/fw/api/stats.h
372
struct mvm_statistics_tx_v4 tx;
drivers/net/wireless/intel/iwlwifi/fw/api/stats.h
380
struct mvm_statistics_tx tx;
drivers/net/wireless/intel/iwlwifi/fw/api/tx.h
756
struct iwl_tx_cmd_v6_params tx;
drivers/net/wireless/intel/iwlwifi/fw/api/tx.h
775
struct iwl_tx_cmd_v6_params tx;
drivers/net/wireless/intel/iwlwifi/mei/trace.h
24
TP_PROTO(const struct iwl_sap_hdr *sap_cmd, bool tx),
drivers/net/wireless/intel/iwlwifi/mei/trace.h
25
TP_ARGS(sap_cmd, tx),
drivers/net/wireless/intel/iwlwifi/mei/trace.h
29
__field(u8, tx)
drivers/net/wireless/intel/iwlwifi/mei/trace.h
37
__entry->tx = tx;
drivers/net/wireless/intel/iwlwifi/mei/trace.h
42
TP_printk("sap_cmd %s: type %d len %d seq %d", __entry->tx ? "Tx" : "Rx",
drivers/net/wireless/intel/iwlwifi/mei/trace.h
47
TP_PROTO(const struct iwl_sap_me_msg_hdr *hdr, bool tx),
drivers/net/wireless/intel/iwlwifi/mei/trace.h
48
TP_ARGS(hdr, tx),
drivers/net/wireless/intel/iwlwifi/mei/trace.h
51
__field(u8, tx)
drivers/net/wireless/intel/iwlwifi/mei/trace.h
57
__entry->tx = tx;
drivers/net/wireless/intel/iwlwifi/mei/trace.h
59
TP_printk("ME message: %s: type %d seq %d", __entry->tx ? "Tx" : "Rx",
drivers/net/wireless/intel/iwlwifi/mld/mac80211.c
2731
.tx = iwl_mld_mac80211_tx,
drivers/net/wireless/intel/iwlwifi/mld/mlo.c
586
total_tx += queue_counter->per_link[link].tx;
drivers/net/wireless/intel/iwlwifi/mld/mlo.c
591
sec_link_tx += queue_counter->per_link[sec_link_id].tx;
drivers/net/wireless/intel/iwlwifi/mld/sta.c
854
u32 count, bool tx)
drivers/net/wireless/intel/iwlwifi/mld/sta.c
899
if (tx)
drivers/net/wireless/intel/iwlwifi/mld/sta.c
900
link_counter->tx += count;
drivers/net/wireless/intel/iwlwifi/mld/sta.c
912
total_mpdus += tx ? queue_counter->per_link[i].tx :
drivers/net/wireless/intel/iwlwifi/mld/sta.h
79
u32 tx;
drivers/net/wireless/intel/iwlwifi/mld/tlc.c
294
u8 tx = u8_get_bits(tx_nss, IEEE80211_EHT_MCS_NSS_TX);
drivers/net/wireless/intel/iwlwifi/mld/tlc.c
299
return min(tx, rx);
drivers/net/wireless/intel/iwlwifi/mld/tlc.c
302
#define MAX_NSS_MCS(mcs_num, rx, tx) \
drivers/net/wireless/intel/iwlwifi/mld/tlc.c
304
(tx)->rx_tx_mcs ##mcs_num## _max_nss)
drivers/net/wireless/intel/iwlwifi/mvm/d3.c
542
iwl_mvm_convert_p1k(p1k, data->tkip.tx.p1k);
drivers/net/wireless/intel/iwlwifi/mvm/d3.c
544
memcpy(data->tkip.mic_keys.tx,
drivers/net/wireless/intel/iwlwifi/mvm/mac-ctxt.c
1024
iwl_mvm_mac_ctxt_set_tx(mvm, vif, beacon, &beacon_cmd.tx);
drivers/net/wireless/intel/iwlwifi/mvm/mac-ctxt.c
1044
iwl_mvm_mac_ctxt_set_tx(mvm, vif, beacon, &beacon_cmd.tx);
drivers/net/wireless/intel/iwlwifi/mvm/mac80211.c
6345
.tx = iwl_mvm_mac_tx,
drivers/net/wireless/intel/iwlwifi/mvm/mld-mac80211.c
1017
.tx = iwl_mvm_mac_tx,
drivers/net/wireless/intel/iwlwifi/mvm/mvm.h
656
} tx;
drivers/net/wireless/intel/iwlwifi/mvm/rs-fw.c
244
u8 tx = u8_get_bits(tx_nss, IEEE80211_EHT_MCS_NSS_TX);
drivers/net/wireless/intel/iwlwifi/mvm/rs-fw.c
249
return min(tx, rx);
drivers/net/wireless/intel/iwlwifi/mvm/rs-fw.c
252
#define MAX_NSS_MCS(mcs_num, rx, tx) \
drivers/net/wireless/intel/iwlwifi/mvm/rs-fw.c
254
(tx)->rx_tx_mcs ##mcs_num## _max_nss)
drivers/net/wireless/intel/iwlwifi/mvm/tx.c
1047
mdata->tx.airtime += airtime;
drivers/net/wireless/intel/iwlwifi/mvm/tx.c
1062
mdata->tx.pkts[ac]++;
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
1004
mdata->tx.pkts[ac];
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
1015
memset(&mdata->tx.pkts, 0, sizeof(mdata->tx.pkts));
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
1026
memset(&mdata->tx.airtime, 0, sizeof(mdata->tx.airtime));
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
1134
memset(&mdata->tx.pkts, 0, sizeof(mdata->tx.pkts));
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
1136
memset(&mdata->tx.airtime, 0, sizeof(mdata->tx.airtime));
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
992
u32 airtime = mdata->rx.airtime + mdata->tx.airtime;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/internal.h
71
u32 tx;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/rx.c
2111
isr_stats->tx++;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/rx.c
2337
isr_stats->tx++;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/rx.c
2346
isr_stats->tx++;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/trans.c
2926
isr_stats->tx);
drivers/net/wireless/intersil/p54/main.c
315
priv->survey_raw.tx = 0;
drivers/net/wireless/intersil/p54/main.c
714
.tx = p54_tx_80211,
drivers/net/wireless/intersil/p54/p54.h
167
void (*tx)(struct ieee80211_hw *dev, struct sk_buff *skb);
drivers/net/wireless/intersil/p54/p54.h
212
u64 tx;
drivers/net/wireless/intersil/p54/p54pci.c
622
priv->common.tx = p54p_tx;
drivers/net/wireless/intersil/p54/p54spi.c
644
priv->common.tx = p54spi_op_tx;
drivers/net/wireless/intersil/p54/p54usb.c
1040
priv->common.tx = p54u_tx_lm87;
drivers/net/wireless/intersil/p54/p54usb.c
1046
priv->common.tx = p54u_tx_net2280;
drivers/net/wireless/intersil/p54/txrx.c
163
priv->tx(priv->hw, skb);
drivers/net/wireless/intersil/p54/txrx.c
523
unsigned int i, rssi, tx, cca, dtime, dtotal, dcca, dtx, drssi, unit;
drivers/net/wireless/intersil/p54/txrx.c
553
tx = le32_to_cpu(stats->sample_tx);
drivers/net/wireless/intersil/p54/txrx.c
560
dtx = tx - priv->survey_raw.cached_tx;
drivers/net/wireless/intersil/p54/txrx.c
578
priv->survey_raw.tx += dtx * unit;
drivers/net/wireless/intersil/p54/txrx.c
579
priv->survey_raw.cached_tx = tx;
drivers/net/wireless/intersil/p54/txrx.c
598
survey->time_tx = priv->survey_raw.tx;
drivers/net/wireless/intersil/p54/txrx.c
599
survey->time_busy = priv->survey_raw.tx +
drivers/net/wireless/marvell/libertas_tf/main.c
480
.tx = lbtf_op_tx,
drivers/net/wireless/marvell/mwifiex/cfg80211.c
1745
.tx = BIT(IEEE80211_STYPE_ACTION >> 4) |
drivers/net/wireless/marvell/mwifiex/cfg80211.c
1751
.tx = BIT(IEEE80211_STYPE_ACTION >> 4) |
drivers/net/wireless/marvell/mwifiex/cfg80211.c
1757
.tx = BIT(IEEE80211_STYPE_ACTION >> 4) |
drivers/net/wireless/marvell/mwifiex/cfg80211.c
1763
.tx = BIT(IEEE80211_STYPE_ACTION >> 4) |
drivers/net/wireless/marvell/mwifiex/cfg80211.c
4736
adapter->mwifiex_mgmt_stypes[NL80211_IFTYPE_AP].tx = 0xffff;
drivers/net/wireless/marvell/mwl8k.c
1689
int tx;
drivers/net/wireless/marvell/mwl8k.c
1701
tx = txq->head;
drivers/net/wireless/marvell/mwl8k.c
1702
tx_desc = txq->txd + tx;
drivers/net/wireless/marvell/mwl8k.c
1713
txq->head = (tx + 1) % MWL8K_TX_DESCS;
drivers/net/wireless/marvell/mwl8k.c
1720
skb = txq->skb[tx];
drivers/net/wireless/marvell/mwl8k.c
1721
txq->skb[tx] = NULL;
drivers/net/wireless/marvell/mwl8k.c
1917
struct mwl8k_tx_desc *tx;
drivers/net/wireless/marvell/mwl8k.c
2104
tx = txq->txd + txq->tail;
drivers/net/wireless/marvell/mwl8k.c
2105
tx->data_rate = txdatarate;
drivers/net/wireless/marvell/mwl8k.c
2106
tx->tx_priority = txpriority;
drivers/net/wireless/marvell/mwl8k.c
2107
tx->qos_control = cpu_to_le16(qos);
drivers/net/wireless/marvell/mwl8k.c
2108
tx->pkt_phys_addr = cpu_to_le32(dma);
drivers/net/wireless/marvell/mwl8k.c
2109
tx->pkt_len = cpu_to_le16(skb->len);
drivers/net/wireless/marvell/mwl8k.c
2110
tx->rate_info = 0;
drivers/net/wireless/marvell/mwl8k.c
2112
tx->peer_id = MWL8K_STA(sta)->peer_id;
drivers/net/wireless/marvell/mwl8k.c
2114
tx->peer_id = 0;
drivers/net/wireless/marvell/mwl8k.c
2117
tx->timestamp = cpu_to_le32(ioread32(priv->regs +
drivers/net/wireless/marvell/mwl8k.c
2120
tx->timestamp = 0;
drivers/net/wireless/marvell/mwl8k.c
2123
tx->status = cpu_to_le32(MWL8K_TXD_STATUS_FW_OWNED | txstatus);
drivers/net/wireless/marvell/mwl8k.c
3613
static int mwl8k_cmd_mimo_config(struct ieee80211_hw *hw, __u8 rx, __u8 tx)
drivers/net/wireless/marvell/mwl8k.c
3626
cmd->tx_antenna_map = tx;
drivers/net/wireless/marvell/mwl8k.c
5689
.tx = mwl8k_tx,
drivers/net/wireless/mediatek/mt76/mt7603/main.c
700
.tx = mt7603_tx,
drivers/net/wireless/mediatek/mt76/mt7615/main.c
1322
.tx = mt7615_tx,
drivers/net/wireless/mediatek/mt76/mt7615/mcu.c
933
bool enable, bool tx)
drivers/net/wireless/mediatek/mt76/mt7615/mcu.c
946
mt76_connac_mcu_sta_ba_tlv(skb, params, enable, tx);
drivers/net/wireless/mediatek/mt76/mt7615/mcu.c
955
mt76_connac_mcu_wtbl_ba_tlv(&dev->mt76, skb, params, enable, tx,
drivers/net/wireless/mediatek/mt76/mt7615/mt7615.h
116
} tx;
drivers/net/wireless/mediatek/mt76/mt7615/sdio.c
68
intr->tx.wtqcr = irq_data->tx.wtqcr;
drivers/net/wireless/mediatek/mt76/mt76_connac_mcu.c
1098
bool enable, bool tx, void *sta_wtbl,
drivers/net/wireless/mediatek/mt76/mt76_connac_mcu.c
1110
if (tx) {
drivers/net/wireless/mediatek/mt76/mt76_connac_mcu.c
1128
if (enable && tx) {
drivers/net/wireless/mediatek/mt76/mt76_connac_mcu.c
1242
bool enable, bool tx)
drivers/net/wireless/mediatek/mt76/mt76_connac_mcu.c
1250
ba->ba_type = tx ? MT_BA_TYPE_ORIGINATOR : MT_BA_TYPE_RECIPIENT;
drivers/net/wireless/mediatek/mt76/mt76_connac_mcu.c
1274
int cmd, bool enable, bool tx)
drivers/net/wireless/mediatek/mt76/mt76_connac_mcu.c
1294
mt76_connac_mcu_wtbl_ba_tlv(dev, skb, params, enable, tx, sta_wtbl,
drivers/net/wireless/mediatek/mt76/mt76_connac_mcu.c
1309
mt76_connac_mcu_sta_ba_tlv(skb, params, enable, tx);
drivers/net/wireless/mediatek/mt76/mt76_connac_mcu.h
1952
bool enable, bool tx, void *sta_wtbl,
drivers/net/wireless/mediatek/mt76/mt76_connac_mcu.h
1956
bool enable, bool tx);
drivers/net/wireless/mediatek/mt76/mt76_connac_mcu.h
1964
int cmd, bool enable, bool tx);
drivers/net/wireless/mediatek/mt76/mt76x0/pci.c
66
.tx = mt76x02_tx,
drivers/net/wireless/mediatek/mt76/mt76x0/usb.c
125
.tx = mt76x02_tx,
drivers/net/wireless/mediatek/mt76/mt76x2/pci_main.c
128
.tx = mt76x02_tx,
drivers/net/wireless/mediatek/mt76/mt76x2/usb_main.c
95
.tx = mt76x02_tx,
drivers/net/wireless/mediatek/mt76/mt7915/debugfs.c
516
bool tx, rx, en;
drivers/net/wireless/mediatek/mt76/mt7915/debugfs.c
526
tx = dev->fw.debug_wm || (dev->fw.debug_bin & BIT(1));
drivers/net/wireless/mediatek/mt76/mt7915/debugfs.c
538
val = en && tx;
drivers/net/wireless/mediatek/mt76/mt7915/main.c
1764
.tx = mt7915_tx,
drivers/net/wireless/mediatek/mt76/mt7921/main.c
1521
.tx = mt792x_tx,
drivers/net/wireless/mediatek/mt76/mt7921/mt7921.h
95
} tx;
drivers/net/wireless/mediatek/mt76/mt7921/pci.c
265
.tx = {
drivers/net/wireless/mediatek/mt76/mt7921/pci.c
516
dev->irq_map->tx.all_complete_mask |
drivers/net/wireless/mediatek/mt76/mt7921/pci_mac.c
98
dev->irq_map->tx.all_complete_mask |
drivers/net/wireless/mediatek/mt76/mt7921/sdio.c
76
intr->tx.wtqcr = irq_data->tx.wtqcr;
drivers/net/wireless/mediatek/mt76/mt7925/main.c
2162
.tx = mt792x_tx,
drivers/net/wireless/mediatek/mt76/mt7925/mcu.c
627
bool enable, bool tx)
drivers/net/wireless/mediatek/mt76/mt7925/mcu.c
644
ba->ba_type = tx ? MT_BA_TYPE_ORIGINATOR : MT_BA_TYPE_RECIPIENT;
drivers/net/wireless/mediatek/mt76/mt7925/pci.c
300
.tx = {
drivers/net/wireless/mediatek/mt76/mt7925/pci.c
551
dev->irq_map->tx.all_complete_mask |
drivers/net/wireless/mediatek/mt76/mt7925/pci_mac.c
120
dev->irq_map->tx.all_complete_mask |
drivers/net/wireless/mediatek/mt76/mt7925/pci_mac.c
96
if (irq_map->tx.all_complete_mask)
drivers/net/wireless/mediatek/mt76/mt792x.h
196
} tx;
drivers/net/wireless/mediatek/mt76/mt792x_dma.c
162
dev->irq_map->tx.all_complete_mask |
drivers/net/wireless/mediatek/mt76/mt792x_dma.c
331
dev->irq_map->tx.all_complete_mask);
drivers/net/wireless/mediatek/mt76/mt792x_dma.c
45
if (intr & dev->irq_map->tx.mcu_complete_mask)
drivers/net/wireless/mediatek/mt76/mt792x_dma.c
46
mask |= dev->irq_map->tx.mcu_complete_mask;
drivers/net/wireless/mediatek/mt76/mt792x_dma.c
62
if (intr & dev->irq_map->tx.all_complete_mask)
drivers/net/wireless/mediatek/mt76/mt7996/debugfs.c
310
bool tx, rx, en;
drivers/net/wireless/mediatek/mt76/mt7996/debugfs.c
320
tx = dev->fw_debug_wm || (dev->fw_debug_bin & BIT(1));
drivers/net/wireless/mediatek/mt76/mt7996/debugfs.c
335
val = en && tx;
drivers/net/wireless/mediatek/mt76/mt7996/main.c
2285
.tx = mt7996_tx,
drivers/net/wireless/mediatek/mt76/mt7996/mcu.c
1194
struct mt76_wcid *wcid, bool enable, bool tx)
drivers/net/wireless/mediatek/mt76/mt7996/mcu.c
1208
ba->ba_type = tx ? MT_BA_TYPE_ORIGINATOR : MT_BA_TYPE_RECIPIENT;
drivers/net/wireless/mediatek/mt76/mt7996/mcu.c
1214
ba->ba_rdd_rro = !tx && enable && mt7996_has_hwrro(dev);
drivers/net/wireless/mediatek/mt76/sdio.h
133
} tx;
drivers/net/wireless/mediatek/mt76/sdio_txrx.c
182
nframes += !!mt76s_refill_sched_quota(dev, intr.tx.wtqcr);
drivers/net/wireless/mediatek/mt7601u/main.c
413
.tx = mt7601u_tx,
drivers/net/wireless/microchip/wilc1000/cfg80211.c
23
.tx = BIT(IEEE80211_STYPE_ACTION >> 4) |
drivers/net/wireless/microchip/wilc1000/cfg80211.c
30
.tx = 0xffff,
drivers/net/wireless/microchip/wilc1000/cfg80211.c
40
.tx = 0xffff,
drivers/net/wireless/purelifi/plfxlc/mac.c
299
struct plfxlc_usb_tx *tx = &usb->tx;
drivers/net/wireless/purelifi/plfxlc/mac.c
302
if (!(tx->station[sidx].flag & STATION_CONNECTED_FLAG))
drivers/net/wireless/purelifi/plfxlc/mac.c
304
if (memcmp(tx->station[sidx].mac, dst_mac, ETH_ALEN))
drivers/net/wireless/purelifi/plfxlc/mac.c
315
if (skb_queue_len(&tx->station[sidx].data_list) > 60)
drivers/net/wireless/purelifi/plfxlc/mac.c
319
if (skb_queue_len(&tx->station[sidx].data_list) > 256)
drivers/net/wireless/purelifi/plfxlc/mac.c
321
skb_queue_tail(&tx->station[sidx].data_list, skb);
drivers/net/wireless/purelifi/plfxlc/mac.c
325
spin_lock_irqsave(&usb->tx.lock, flags);
drivers/net/wireless/purelifi/plfxlc/mac.c
328
spin_unlock_irqrestore(&usb->tx.lock, flags);
drivers/net/wireless/purelifi/plfxlc/mac.c
400
struct plfxlc_usb_tx *tx;
drivers/net/wireless/purelifi/plfxlc/mac.c
445
tx = &mac->chip.usb.tx;
drivers/net/wireless/purelifi/plfxlc/mac.c
448
if (memcmp(&buffer[10], tx->station[sidx].mac, ETH_ALEN))
drivers/net/wireless/purelifi/plfxlc/mac.c
450
if (tx->station[sidx].flag & STATION_CONNECTED_FLAG) {
drivers/net/wireless/purelifi/plfxlc/mac.c
451
tx->station[sidx].flag |= STATION_HEARTBEAT_FLAG;
drivers/net/wireless/purelifi/plfxlc/mac.c
458
if (tx->station[sidx].flag & STATION_CONNECTED_FLAG)
drivers/net/wireless/purelifi/plfxlc/mac.c
460
memcpy(tx->station[sidx].mac, &buffer[10], ETH_ALEN);
drivers/net/wireless/purelifi/plfxlc/mac.c
461
tx->station[sidx].flag |= STATION_CONNECTED_FLAG;
drivers/net/wireless/purelifi/plfxlc/mac.c
462
tx->station[sidx].flag |= STATION_HEARTBEAT_FLAG;
drivers/net/wireless/purelifi/plfxlc/mac.c
686
.tx = plfxlc_op_tx,
drivers/net/wireless/purelifi/plfxlc/usb.c
108
if (tx->submitted_urbs++ < PURELIFI_URB_RETRY_MAX) {
drivers/net/wireless/purelifi/plfxlc/usb.c
110
tx->submitted_urbs++);
drivers/net/wireless/purelifi/plfxlc/usb.c
114
tx->submitted_urbs = 0;
drivers/net/wireless/purelifi/plfxlc/usb.c
135
tx->mac_fifo_full = 1;
drivers/net/wireless/purelifi/plfxlc/usb.c
137
tx->station[sidx].flag |= STATION_FIFO_FULL_FLAG;
drivers/net/wireless/purelifi/plfxlc/usb.c
143
tx->station[sidx].flag &= STATION_ACTIVE_FLAG;
drivers/net/wireless/purelifi/plfxlc/usb.c
306
struct plfxlc_usb_tx *tx = &usb->tx;
drivers/net/wireless/purelifi/plfxlc/usb.c
309
clear_bit(PLF_BIT_ENABLED, &tx->enabled);
drivers/net/wireless/purelifi/plfxlc/usb.c
312
usb_kill_anchored_urbs(&tx->submitted);
drivers/net/wireless/purelifi/plfxlc/usb.c
314
spin_lock_irqsave(&tx->lock, flags);
drivers/net/wireless/purelifi/plfxlc/usb.c
315
WARN_ON(!skb_queue_empty(&tx->submitted_skbs));
drivers/net/wireless/purelifi/plfxlc/usb.c
316
WARN_ON(tx->submitted_urbs != 0);
drivers/net/wireless/purelifi/plfxlc/usb.c
317
tx->submitted_urbs = 0;
drivers/net/wireless/purelifi/plfxlc/usb.c
318
spin_unlock_irqrestore(&tx->lock, flags);
drivers/net/wireless/purelifi/plfxlc/usb.c
327
struct plfxlc_usb_tx *tx = &usb->tx;
drivers/net/wireless/purelifi/plfxlc/usb.c
330
spin_lock_irqsave(&tx->lock, flags);
drivers/net/wireless/purelifi/plfxlc/usb.c
331
set_bit(PLF_BIT_ENABLED, &tx->enabled);
drivers/net/wireless/purelifi/plfxlc/usb.c
332
tx->submitted_urbs = 0;
drivers/net/wireless/purelifi/plfxlc/usb.c
334
tx->stopped = 0;
drivers/net/wireless/purelifi/plfxlc/usb.c
335
spin_unlock_irqrestore(&tx->lock, flags);
drivers/net/wireless/purelifi/plfxlc/usb.c
38
struct plfxlc_usb_tx *tx = &usb->tx;
drivers/net/wireless/purelifi/plfxlc/usb.c
390
struct plfxlc_usb_tx *tx = &usb->tx;
drivers/net/wireless/purelifi/plfxlc/usb.c
392
spin_lock_init(&tx->lock);
drivers/net/wireless/purelifi/plfxlc/usb.c
393
clear_bit(PLF_BIT_ENABLED, &tx->enabled);
drivers/net/wireless/purelifi/plfxlc/usb.c
394
tx->stopped = 0;
drivers/net/wireless/purelifi/plfxlc/usb.c
395
skb_queue_head_init(&tx->submitted_skbs);
drivers/net/wireless/purelifi/plfxlc/usb.c
396
init_usb_anchor(&tx->submitted);
drivers/net/wireless/purelifi/plfxlc/usb.c
43
spin_lock_irqsave(&tx->lock, flags);
drivers/net/wireless/purelifi/plfxlc/usb.c
47
if (!(tx->station[usb->sidx].flag & STATION_CONNECTED_FLAG))
drivers/net/wireless/purelifi/plfxlc/usb.c
49
if (!(tx->station[usb->sidx].flag & STATION_FIFO_FULL_FLAG))
drivers/net/wireless/purelifi/plfxlc/usb.c
50
skb = skb_peek(&tx->station[usb->sidx].data_list);
drivers/net/wireless/purelifi/plfxlc/usb.c
54
skb = skb_dequeue(&tx->station[usb->sidx].data_list);
drivers/net/wireless/purelifi/plfxlc/usb.c
551
struct plfxlc_usb *usb = timer_container_of(usb, t, tx.tx_retry_timer);
drivers/net/wireless/purelifi/plfxlc/usb.c
554
timer_setup(&usb->tx.tx_retry_timer,
drivers/net/wireless/purelifi/plfxlc/usb.c
556
mod_timer(&usb->tx.tx_retry_timer, jiffies + TX_RETRY_BACKOFF_JIFF);
drivers/net/wireless/purelifi/plfxlc/usb.c
562
struct plfxlc_usb_tx *tx = &usb->tx;
drivers/net/wireless/purelifi/plfxlc/usb.c
566
if (!(tx->station[sidx].flag & STATION_CONNECTED_FLAG))
drivers/net/wireless/purelifi/plfxlc/usb.c
568
if (tx->station[sidx].flag & STATION_HEARTBEAT_FLAG) {
drivers/net/wireless/purelifi/plfxlc/usb.c
569
tx->station[sidx].flag ^= STATION_HEARTBEAT_FLAG;
drivers/net/wireless/purelifi/plfxlc/usb.c
57
if (skb_queue_len(&tx->station[usb->sidx].data_list) <= 60)
drivers/net/wireless/purelifi/plfxlc/usb.c
571
eth_zero_addr(tx->station[sidx].mac);
drivers/net/wireless/purelifi/plfxlc/usb.c
572
tx->station[sidx].flag = 0;
drivers/net/wireless/purelifi/plfxlc/usb.c
585
struct plfxlc_usb_tx *tx;
drivers/net/wireless/purelifi/plfxlc/usb.c
60
spin_unlock_irqrestore(&tx->lock, flags);
drivers/net/wireless/purelifi/plfxlc/usb.c
602
tx = &usb->tx;
drivers/net/wireless/purelifi/plfxlc/usb.c
638
tx->mac_fifo_full = 0;
drivers/net/wireless/purelifi/plfxlc/usb.c
639
spin_lock_init(&tx->lock);
drivers/net/wireless/purelifi/plfxlc/usb.c
674
skb_queue_head_init(&tx->station[i].data_list);
drivers/net/wireless/purelifi/plfxlc/usb.c
675
tx->station[i].flag = 0;
drivers/net/wireless/purelifi/plfxlc/usb.c
678
tx->station[STA_BROADCAST_INDEX].flag |= STATION_CONNECTED_FLAG;
drivers/net/wireless/purelifi/plfxlc/usb.c
680
tx->station[STA_BROADCAST_INDEX].mac[i] = 0xFF;
drivers/net/wireless/purelifi/plfxlc/usb.c
682
timer_setup(&tx->tx_retry_timer, slif_data_plane_sap_timer_callb, 0);
drivers/net/wireless/purelifi/plfxlc/usb.c
683
tx->tx_retry_timer.expires = jiffies + TX_RETRY_BACKOFF_JIFF;
drivers/net/wireless/purelifi/plfxlc/usb.c
684
add_timer(&tx->tx_retry_timer);
drivers/net/wireless/purelifi/plfxlc/usb.c
71
struct plfxlc_usb_tx *tx;
drivers/net/wireless/purelifi/plfxlc/usb.c
719
timer_delete_sync(&usb->tx.tx_retry_timer);
drivers/net/wireless/purelifi/plfxlc/usb.c
94
tx = &usb->tx;
drivers/net/wireless/purelifi/plfxlc/usb.h
129
struct plfxlc_usb_tx tx;
drivers/net/wireless/quantenna/qtnfmac/cfg80211.c
56
.tx = BIT(IEEE80211_STYPE_ACTION >> 4) |
drivers/net/wireless/quantenna/qtnfmac/cfg80211.c
63
.tx = BIT(IEEE80211_STYPE_ACTION >> 4) |
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
1494
rt2x00dev->default_ant.tx =
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
1505
if (rt2x00dev->default_ant.tx == ANTENNA_SW_DIVERSITY)
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
1506
rt2x00dev->default_ant.tx = ANTENNA_HW_DIVERSITY;
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
1678
rt2x00dev->tx->cw_min, rt2x00dev->tx->cw_max);
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
1712
.tx = rt2x00mac_tx,
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
401
ant->tx == ANTENNA_SW_DIVERSITY);
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
409
switch (ant->tx) {
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
766
rt2x00_set_field32(®, TXCSR2_TXD_SIZE, rt2x00dev->tx[0].desc_size);
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
767
rt2x00_set_field32(®, TXCSR2_NUM_TXD, rt2x00dev->tx[1].limit);
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
769
rt2x00_set_field32(®, TXCSR2_NUM_PRIO, rt2x00dev->tx[0].limit);
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
772
entry_priv = rt2x00dev->tx[1].entries[0].priv_data;
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
778
entry_priv = rt2x00dev->tx[0].entries[0].priv_data;
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
1653
rt2x00dev->default_ant.tx =
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
2010
.tx = rt2x00mac_tx,
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
408
ant->tx == ANTENNA_SW_DIVERSITY);
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
417
switch (ant->tx) {
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
851
rt2x00_set_field32(®, TXCSR2_TXD_SIZE, rt2x00dev->tx[0].desc_size);
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
852
rt2x00_set_field32(®, TXCSR2_NUM_TXD, rt2x00dev->tx[1].limit);
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
854
rt2x00_set_field32(®, TXCSR2_NUM_PRIO, rt2x00dev->tx[0].limit);
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
857
entry_priv = rt2x00dev->tx[1].entries[0].priv_data;
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
863
entry_priv = rt2x00dev->tx[0].entries[0].priv_data;
drivers/net/wireless/ralink/rt2x00/rt2500usb.c
1461
rt2x00dev->default_ant.tx =
drivers/net/wireless/ralink/rt2x00/rt2500usb.c
1472
if (rt2x00dev->default_ant.tx == ANTENNA_SW_DIVERSITY)
drivers/net/wireless/ralink/rt2x00/rt2500usb.c
1473
rt2x00dev->default_ant.tx = ANTENNA_HW_DIVERSITY;
drivers/net/wireless/ralink/rt2x00/rt2500usb.c
1801
.tx = rt2x00mac_tx,
drivers/net/wireless/ralink/rt2x00/rt2500usb.c
513
ant->tx == ANTENNA_SW_DIVERSITY);
drivers/net/wireless/ralink/rt2x00/rt2500usb.c
523
switch (ant->tx) {
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
11257
rt2x00dev->default_ant.tx = ANTENNA_A;
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
11261
rt2x00dev->default_ant.tx = ANTENNA_A;
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
11266
rt2x00dev->default_ant.tx = ANTENNA_A;
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
11273
rt2x00dev->default_ant.tx = ANTENNA_HW_DIVERSITY; /* Unused */
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
668
entry_priv = rt2x00dev->tx[0].entries[0].priv_data;
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
672
rt2x00dev->tx[0].limit);
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
676
entry_priv = rt2x00dev->tx[1].entries[0].priv_data;
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
680
rt2x00dev->tx[1].limit);
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
684
entry_priv = rt2x00dev->tx[2].entries[0].priv_data;
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
688
rt2x00dev->tx[2].limit);
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
692
entry_priv = rt2x00dev->tx[3].entries[0].priv_data;
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
696
rt2x00dev->tx[3].limit);
drivers/net/wireless/ralink/rt2x00/rt2800pci.c
297
.tx = rt2x00mac_tx,
drivers/net/wireless/ralink/rt2x00/rt2800soc.c
161
.tx = rt2x00mac_tx,
drivers/net/wireless/ralink/rt2x00/rt2800usb.c
636
.tx = rt2x00mac_tx,
drivers/net/wireless/ralink/rt2x00/rt2x00.h
1316
if (queue < rt2x00dev->ops->tx_queues && rt2x00dev->tx)
drivers/net/wireless/ralink/rt2x00/rt2x00.h
1317
return &rt2x00dev->tx[queue];
drivers/net/wireless/ralink/rt2x00/rt2x00.h
1497
u32 *tx, u32 *tx_max, u32 *rx, u32 *rx_max);
drivers/net/wireless/ralink/rt2x00/rt2x00.h
211
enum antenna tx;
drivers/net/wireless/ralink/rt2x00/rt2x00.h
965
struct data_queue *tx;
drivers/net/wireless/ralink/rt2x00/rt2x00config.c
132
if (config.tx == ANTENNA_SW_DIVERSITY) {
drivers/net/wireless/ralink/rt2x00/rt2x00config.c
135
if (def->tx == ANTENNA_SW_DIVERSITY)
drivers/net/wireless/ralink/rt2x00/rt2x00config.c
136
config.tx = ANTENNA_B;
drivers/net/wireless/ralink/rt2x00/rt2x00config.c
138
config.tx = def->tx;
drivers/net/wireless/ralink/rt2x00/rt2x00config.c
140
} else if (config.tx == ANTENNA_SW_DIVERSITY)
drivers/net/wireless/ralink/rt2x00/rt2x00config.c
141
config.tx = active->tx;
drivers/net/wireless/ralink/rt2x00/rt2x00dev.c
1170
rt2x00dev->tx->limit *
drivers/net/wireless/ralink/rt2x00/rt2x00dev.c
1362
if (WARN_ON(!rt2x00dev->tx))
drivers/net/wireless/ralink/rt2x00/rt2x00dev.c
1366
return rt2x00dev->tx[0].winfo_size + rt2x00dev->tx[0].desc_size;
drivers/net/wireless/ralink/rt2x00/rt2x00dev.c
1368
return rt2x00dev->tx[0].winfo_size;
drivers/net/wireless/ralink/rt2x00/rt2x00link.c
103
new_ant.tx = other_antenna;
drivers/net/wireless/ralink/rt2x00/rt2x00link.c
143
new_ant.tx = (new_ant.tx == ANTENNA_A) ? ANTENNA_B : ANTENNA_A;
drivers/net/wireless/ralink/rt2x00/rt2x00mac.c
768
if (tx_ant == 3 && def->tx == ANTENNA_SW_DIVERSITY) {
drivers/net/wireless/ralink/rt2x00/rt2x00mac.c
778
setup.tx = tx_ant;
drivers/net/wireless/ralink/rt2x00/rt2x00mac.c
801
*tx_ant = active->tx;
drivers/net/wireless/ralink/rt2x00/rt2x00mac.c
813
u32 *tx, u32 *tx_max, u32 *rx, u32 *rx_max)
drivers/net/wireless/ralink/rt2x00/rt2x00mac.c
819
*tx += queue->length;
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1255
rt2x00dev->tx = &queue[1];
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1285
rt2x00dev->tx = NULL;
drivers/net/wireless/ralink/rt2x00/rt2x00queue.h
505
&(__dev)->tx[(__dev)->ops->tx_queues]
drivers/net/wireless/ralink/rt2x00/rt2x00queue.h
552
queue_loop(__entry, (__dev)->tx, tx_queue_end(__dev))
drivers/net/wireless/ralink/rt2x00/rt2x00queue.h
563
queue_loop(__entry, (__dev)->tx, queue_end(__dev))
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
604
struct data_queue *queue = rt2x00dev->tx;
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
631
if (!rt2x00dev->rx->usb_endpoint || !rt2x00dev->tx->usb_endpoint) {
drivers/net/wireless/ralink/rt2x00/rt61pci.c
1337
rt2x00dev->tx[0].limit);
drivers/net/wireless/ralink/rt2x00/rt61pci.c
1339
rt2x00dev->tx[1].limit);
drivers/net/wireless/ralink/rt2x00/rt61pci.c
1341
rt2x00dev->tx[2].limit);
drivers/net/wireless/ralink/rt2x00/rt61pci.c
1343
rt2x00dev->tx[3].limit);
drivers/net/wireless/ralink/rt2x00/rt61pci.c
1348
rt2x00dev->tx[0].desc_size / 4);
drivers/net/wireless/ralink/rt2x00/rt61pci.c
1351
entry_priv = rt2x00dev->tx[0].entries[0].priv_data;
drivers/net/wireless/ralink/rt2x00/rt61pci.c
1357
entry_priv = rt2x00dev->tx[1].entries[0].priv_data;
drivers/net/wireless/ralink/rt2x00/rt61pci.c
1363
entry_priv = rt2x00dev->tx[2].entries[0].priv_data;
drivers/net/wireless/ralink/rt2x00/rt61pci.c
1369
entry_priv = rt2x00dev->tx[3].entries[0].priv_data;
drivers/net/wireless/ralink/rt2x00/rt61pci.c
2081
for (i = 0; i < rt2x00dev->tx->limit; i++) {
drivers/net/wireless/ralink/rt2x00/rt61pci.c
2443
rt2x00dev->default_ant.tx =
drivers/net/wireless/ralink/rt2x00/rt61pci.c
2488
rt2x00dev->default_ant.tx =
drivers/net/wireless/ralink/rt2x00/rt61pci.c
2492
rt2x00dev->default_ant.tx = ANTENNA_SW_DIVERSITY;
drivers/net/wireless/ralink/rt2x00/rt61pci.c
2879
.tx = rt2x00mac_tx,
drivers/net/wireless/ralink/rt2x00/rt61pci.c
716
ant->tx == ANTENNA_SW_DIVERSITY);
drivers/net/wireless/ralink/rt2x00/rt73usb.c
1879
rt2x00dev->default_ant.tx =
drivers/net/wireless/ralink/rt2x00/rt73usb.c
2298
.tx = rt2x00mac_tx,
drivers/net/wireless/ralink/rt2x00/rt73usb.c
697
ant->tx == ANTENNA_SW_DIVERSITY);
drivers/net/wireless/realtek/rtl818x/rtl8180/dev.c
1609
.tx = rtl8180_tx,
drivers/net/wireless/realtek/rtl818x/rtl8187/dev.c
1396
.tx = rtl8187_tx,
drivers/net/wireless/realtek/rtl8xxxu/core.c
7544
.tx = rtl8xxxu_tx,
drivers/net/wireless/realtek/rtlwifi/core.c
1886
.tx = rtl_op_tx,
drivers/net/wireless/realtek/rtw88/mac80211.c
952
.tx = rtw_ops_tx,
drivers/net/wireless/realtek/rtw88/rtw8703b.c
1006
tx ? "TX" : "RX", dur);
drivers/net/wireless/realtek/rtw88/rtw8703b.c
1010
tx ? "TX" : "RX", dur);
drivers/net/wireless/realtek/rtw88/rtw8703b.c
982
static void rtw8703b_iqk_one_shot(struct rtw_dev *rtwdev, bool tx)
drivers/net/wireless/realtek/rtw88/rtw8723d.c
496
static void rtw8723d_iqk_one_shot(struct rtw_dev *rtwdev, bool tx,
drivers/net/wireless/realtek/rtw88/rtw8723d.c
499
u32 pts = (tx ? iqk_cfg->val_iqkpts : 0xf9000000);
drivers/net/wireless/realtek/rtw88/rtw8723d.c
508
iqk_cfg->name, tx ? "TX" : "RX",
drivers/net/wireless/realtek/rtw88/rtw8723d.c
511
iqk_cfg->name, tx ? "TX" : "RX",
drivers/net/wireless/realtek/rtw88/rtw8723d.c
520
tx ? "TX" : "RX");
drivers/net/wireless/realtek/rtw89/coex.c
4450
plt.tx = tx_val;
drivers/net/wireless/realtek/rtw89/core.c
345
struct sk_buff *skb, bool tx)
drivers/net/wireless/realtek/rtw89/core.c
347
if (tx) {
drivers/net/wireless/realtek/rtw89/core.c
359
bool accu_dev, bool tx)
drivers/net/wireless/realtek/rtw89/core.c
371
__rtw89_traffic_stats_accu(&rtwdev->stats, skb, tx);
drivers/net/wireless/realtek/rtw89/core.c
374
__rtw89_traffic_stats_accu(&rtwvif->stats, skb, tx);
drivers/net/wireless/realtek/rtw89/core.c
375
__rtw89_traffic_stats_accu(&rtwvif->stats_ps, skb, tx);
drivers/net/wireless/realtek/rtw89/mac.c
6454
val = (plt->tx & RTW89_MAC_AX_PLT_LTE_RX ? B_AX_TX_PLT_GNT_LTE_RX : 0) |
drivers/net/wireless/realtek/rtw89/mac.c
6455
(plt->tx & RTW89_MAC_AX_PLT_GNT_BT_TX ? B_AX_TX_PLT_GNT_BT_TX : 0) |
drivers/net/wireless/realtek/rtw89/mac.c
6456
(plt->tx & RTW89_MAC_AX_PLT_GNT_BT_RX ? B_AX_TX_PLT_GNT_BT_RX : 0) |
drivers/net/wireless/realtek/rtw89/mac.c
6457
(plt->tx & RTW89_MAC_AX_PLT_GNT_WL ? B_AX_TX_PLT_GNT_WL : 0) |
drivers/net/wireless/realtek/rtw89/mac.h
544
u8 tx;
drivers/net/wireless/realtek/rtw89/mac80211.c
1955
.tx = rtw89_ops_tx,
drivers/net/wireless/realtek/rtw89/mac_be.c
2439
val = (plt->tx & RTW89_MAC_AX_PLT_LTE_RX ? B_BE_TX_PLT_GNT_LTE_RX : 0) |
drivers/net/wireless/realtek/rtw89/mac_be.c
2440
(plt->tx & RTW89_MAC_AX_PLT_GNT_BT_TX ? B_BE_TX_PLT_GNT_BT_TX : 0) |
drivers/net/wireless/realtek/rtw89/mac_be.c
2441
(plt->tx & RTW89_MAC_AX_PLT_GNT_BT_RX ? B_BE_TX_PLT_GNT_BT_RX : 0) |
drivers/net/wireless/realtek/rtw89/mac_be.c
2442
(plt->tx & RTW89_MAC_AX_PLT_GNT_WL ? B_BE_TX_PLT_GNT_WL : 0) |
drivers/net/wireless/realtek/rtw89/pci.c
1094
.tx = {
drivers/net/wireless/realtek/rtw89/pci.c
1117
.tx = {
drivers/net/wireless/realtek/rtw89/pci.c
1140
.tx = {
drivers/net/wireless/realtek/rtw89/pci.c
1163
.tx = {
drivers/net/wireless/realtek/rtw89/pci.c
1203
*addr = &info->dma_addr_set->tx[txch];
drivers/net/wireless/realtek/rtw89/pci.c
1237
struct rtw89_pci_tx_ring *tx_ring = &rtwpci->tx.rings[RTW89_TXCH_CH12];
drivers/net/wireless/realtek/rtw89/pci.c
1253
struct rtw89_pci_tx_ring *tx_ring = &rtwpci->tx.rings[txch];
drivers/net/wireless/realtek/rtw89/pci.c
1270
struct rtw89_pci_tx_ring *tx_ring = &rtwpci->tx.rings[txch];
drivers/net/wireless/realtek/rtw89/pci.c
1363
struct rtw89_pci_tx_ring *tx_ring = &rtwpci->tx.rings[txch];
drivers/net/wireless/realtek/rtw89/pci.c
137
struct rtw89_pci_tx_ring *tx_ring = &rtwpci->tx.rings[RTW89_TXCH_CH12];
drivers/net/wireless/realtek/rtw89/pci.c
1383
tx_ring = &rtwpci->tx.rings[txch];
drivers/net/wireless/realtek/rtw89/pci.c
1391
struct rtw89_pci_tx_ring *tx_ring = &rtwpci->tx.rings[txch];
drivers/net/wireless/realtek/rtw89/pci.c
1655
tx_ring = &rtwpci->tx.rings[txch];
drivers/net/wireless/realtek/rtw89/pci.c
1798
tx_ring = &rtwpci->tx.rings[i];
drivers/net/wireless/realtek/rtw89/pci.c
1895
rtw89_pci_release_tx_ring(rtwdev, &rtwpci->tx.rings[txch]);
drivers/net/wireless/realtek/rtw89/pci.c
1971
tx_ring = &rtwpci->tx.rings[i];
drivers/net/wireless/realtek/rtw89/pci.c
1974
dma_addr_set->tx[i].idx;
drivers/net/wireless/realtek/rtw89/pci.c
3429
struct rtw89_pci_dma_pool *bd_pool = &rtwpci->tx.bd_pool;
drivers/net/wireless/realtek/rtw89/pci.c
3437
tx_ring = &rtwpci->tx.rings[i];
drivers/net/wireless/realtek/rtw89/pci.c
3616
struct rtw89_pci_dma_pool *bd_pool = &rtwpci->tx.bd_pool;
drivers/net/wireless/realtek/rtw89/pci.c
3648
tx_ring = &rtwpci->tx.rings[i];
drivers/net/wireless/realtek/rtw89/pci.c
3665
tx_ring = &rtwpci->tx.rings[i];
drivers/net/wireless/realtek/rtw89/pci.c
619
tx_ring = &rtwpci->tx.rings[rpp_info.txch];
drivers/net/wireless/realtek/rtw89/pci.c
63
u32 cur_idx, bool tx)
drivers/net/wireless/realtek/rtw89/pci.c
73
if (tx) {
drivers/net/wireless/realtek/rtw89/pci.h
1320
struct rtw89_pci_ch_dma_addr tx[RTW89_TXCH_NUM];
drivers/net/wireless/realtek/rtw89/pci.h
1626
struct rtw89_pci_tx_rings tx;
drivers/net/wireless/rsi/rsi_91x_mac80211.c
1972
.tx = rsi_mac80211_tx,
drivers/net/wireless/silabs/wfx/main.c
136
.tx = wfx_tx,
drivers/net/wireless/st/cw1200/bh.c
235
int *tx)
drivers/net/wireless/st/cw1200/bh.c
325
*tx = 1;
drivers/net/wireless/st/cw1200/bh.c
415
int rx, tx, term, suspend;
drivers/net/wireless/st/cw1200/bh.c
450
tx = atomic_xchg(&priv->bh_tx, 0);
drivers/net/wireless/st/cw1200/bh.c
454
(rx || tx || term || suspend || priv->bh_error);
drivers/net/wireless/st/cw1200/bh.c
458
rx, tx, term, suspend, priv->bh_error, status);
drivers/net/wireless/st/cw1200/bh.c
472
if (priv->hw_bufs_used && (!rx || !tx)) {
drivers/net/wireless/st/cw1200/bh.c
537
tx += pending_tx;
drivers/net/wireless/st/cw1200/bh.c
545
ret = cw1200_bh_rx_helper(priv, &ctrl_reg, &tx);
drivers/net/wireless/st/cw1200/bh.c
550
ret = cw1200_bh_rx_helper(priv, &ctrl_reg, &tx);
drivers/net/wireless/st/cw1200/bh.c
556
tx:
drivers/net/wireless/st/cw1200/bh.c
557
if (tx) {
drivers/net/wireless/st/cw1200/bh.c
558
tx = 0;
drivers/net/wireless/st/cw1200/bh.c
568
pending_tx = tx;
drivers/net/wireless/st/cw1200/bh.c
575
tx = ret;
drivers/net/wireless/st/cw1200/bh.c
587
if (tx)
drivers/net/wireless/st/cw1200/bh.c
588
goto tx;
drivers/net/wireless/st/cw1200/debug.c
266
d->tx);
drivers/net/wireless/st/cw1200/debug.h
14
int tx;
drivers/net/wireless/st/cw1200/debug.h
35
++priv->debug->tx;
drivers/net/wireless/st/cw1200/main.c
215
.tx = cw1200_tx,
drivers/net/wireless/st/cw1200/queue.c
330
struct wsm_tx **tx,
drivers/net/wireless/st/cw1200/queue.c
348
*tx = (struct wsm_tx *)item->skb->data;
drivers/net/wireless/st/cw1200/queue.c
351
(*tx)->packet_id = item->packet_id;
drivers/net/wireless/st/cw1200/queue.h
84
struct wsm_tx **tx,
drivers/net/wireless/st/cw1200/sta.c
151
int tx = 0;
drivers/net/wireless/st/cw1200/sta.c
174
tx = 1;
drivers/net/wireless/st/cw1200/sta.c
182
tx = 1;
drivers/net/wireless/st/cw1200/sta.c
190
tx = 0;
drivers/net/wireless/st/cw1200/sta.c
193
if (tx) {
drivers/net/wireless/st/cw1200/txrx.c
895
struct ieee80211_tx_info *tx = IEEE80211_SKB_CB(skb);
drivers/net/wireless/st/cw1200/txrx.c
917
tx->flags |= IEEE80211_TX_STAT_ACK;
drivers/net/wireless/st/cw1200/txrx.c
933
if (tx->status.rates[i].count >= tx_count) {
drivers/net/wireless/st/cw1200/txrx.c
934
tx->status.rates[i].count = tx_count;
drivers/net/wireless/st/cw1200/txrx.c
937
tx_count -= tx->status.rates[i].count;
drivers/net/wireless/st/cw1200/txrx.c
938
if (tx->status.rates[i].flags & IEEE80211_TX_RC_MCS)
drivers/net/wireless/st/cw1200/txrx.c
939
tx->status.rates[i].flags |= ht_flags;
drivers/net/wireless/st/cw1200/txrx.c
943
tx->status.rates[i].count = 0;
drivers/net/wireless/st/cw1200/txrx.c
944
tx->status.rates[i].idx = -1;
drivers/net/wireless/st/cw1200/txrx.c
948
if (tx->control.hw_key) {
drivers/net/wireless/st/cw1200/txrx.c
949
skb_trim(skb, skb->len - tx->control.hw_key->icv_len);
drivers/net/wireless/st/cw1200/txrx.c
950
if (tx->control.hw_key->cipher == WLAN_CIPHER_SUITE_TKIP)
drivers/net/wireless/ti/wl1251/acx.h
1063
struct acx_tx_statistics tx;
drivers/net/wireless/ti/wl1251/debugfs.c
101
DEBUGFS_FWSTATS_FILE(tx, internal_desc_overflow, 20, "%u");
drivers/net/wireless/ti/wl1251/debugfs.c
241
DEBUGFS_FWSTATS_DEL(tx, internal_desc_overflow);
drivers/net/wireless/ti/wl1251/debugfs.c
340
DEBUGFS_FWSTATS_ADD(tx, internal_desc_overflow);
drivers/net/wireless/ti/wl1251/main.c
1366
.tx = wl1251_op_tx,
drivers/net/wireless/ti/wl12xx/acx.h
244
struct wl12xx_acx_tx_statistics tx;
drivers/net/wireless/ti/wl12xx/debugfs.c
119
DEBUGFS_FWSTATS_ADD(tx, internal_desc_overflow);
drivers/net/wireless/ti/wl12xx/debugfs.c
19
WL12XX_DEBUGFS_FWSTATS_FILE(tx, internal_desc_overflow, "%u");
drivers/net/wireless/ti/wl12xx/main.c
115
.tx = {
drivers/net/wireless/ti/wl12xx/main.c
1663
thold = wl->conf.tx.fast_link_thold;
drivers/net/wireless/ti/wl12xx/main.c
1665
thold = wl->conf.tx.slow_link_thold;
drivers/net/wireless/ti/wl18xx/acx.h
270
struct wl18xx_acx_tx_stats tx;
drivers/net/wireless/ti/wl18xx/debugfs.c
427
DEBUGFS_FWSTATS_ADD(tx, tx_prepared_descs);
drivers/net/wireless/ti/wl18xx/debugfs.c
428
DEBUGFS_FWSTATS_ADD(tx, tx_cmplt);
drivers/net/wireless/ti/wl18xx/debugfs.c
429
DEBUGFS_FWSTATS_ADD(tx, tx_template_prepared);
drivers/net/wireless/ti/wl18xx/debugfs.c
430
DEBUGFS_FWSTATS_ADD(tx, tx_data_prepared);
drivers/net/wireless/ti/wl18xx/debugfs.c
431
DEBUGFS_FWSTATS_ADD(tx, tx_template_programmed);
drivers/net/wireless/ti/wl18xx/debugfs.c
432
DEBUGFS_FWSTATS_ADD(tx, tx_data_programmed);
drivers/net/wireless/ti/wl18xx/debugfs.c
433
DEBUGFS_FWSTATS_ADD(tx, tx_burst_programmed);
drivers/net/wireless/ti/wl18xx/debugfs.c
434
DEBUGFS_FWSTATS_ADD(tx, tx_starts);
drivers/net/wireless/ti/wl18xx/debugfs.c
435
DEBUGFS_FWSTATS_ADD(tx, tx_stop);
drivers/net/wireless/ti/wl18xx/debugfs.c
436
DEBUGFS_FWSTATS_ADD(tx, tx_start_templates);
drivers/net/wireless/ti/wl18xx/debugfs.c
437
DEBUGFS_FWSTATS_ADD(tx, tx_start_int_templates);
drivers/net/wireless/ti/wl18xx/debugfs.c
438
DEBUGFS_FWSTATS_ADD(tx, tx_start_fw_gen);
drivers/net/wireless/ti/wl18xx/debugfs.c
439
DEBUGFS_FWSTATS_ADD(tx, tx_start_data);
drivers/net/wireless/ti/wl18xx/debugfs.c
440
DEBUGFS_FWSTATS_ADD(tx, tx_start_null_frame);
drivers/net/wireless/ti/wl18xx/debugfs.c
441
DEBUGFS_FWSTATS_ADD(tx, tx_exch);
drivers/net/wireless/ti/wl18xx/debugfs.c
442
DEBUGFS_FWSTATS_ADD(tx, tx_retry_template);
drivers/net/wireless/ti/wl18xx/debugfs.c
443
DEBUGFS_FWSTATS_ADD(tx, tx_retry_data);
drivers/net/wireless/ti/wl18xx/debugfs.c
444
DEBUGFS_FWSTATS_ADD(tx, tx_retry_per_rate);
drivers/net/wireless/ti/wl18xx/debugfs.c
445
DEBUGFS_FWSTATS_ADD(tx, tx_exch_pending);
drivers/net/wireless/ti/wl18xx/debugfs.c
446
DEBUGFS_FWSTATS_ADD(tx, tx_exch_expiry);
drivers/net/wireless/ti/wl18xx/debugfs.c
447
DEBUGFS_FWSTATS_ADD(tx, tx_done_template);
drivers/net/wireless/ti/wl18xx/debugfs.c
448
DEBUGFS_FWSTATS_ADD(tx, tx_done_data);
drivers/net/wireless/ti/wl18xx/debugfs.c
449
DEBUGFS_FWSTATS_ADD(tx, tx_done_int_template);
drivers/net/wireless/ti/wl18xx/debugfs.c
45
WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_prepared_descs, "%u");
drivers/net/wireless/ti/wl18xx/debugfs.c
450
DEBUGFS_FWSTATS_ADD(tx, tx_cfe1);
drivers/net/wireless/ti/wl18xx/debugfs.c
451
DEBUGFS_FWSTATS_ADD(tx, tx_cfe2);
drivers/net/wireless/ti/wl18xx/debugfs.c
452
DEBUGFS_FWSTATS_ADD(tx, frag_called);
drivers/net/wireless/ti/wl18xx/debugfs.c
453
DEBUGFS_FWSTATS_ADD(tx, frag_mpdu_alloc_failed);
drivers/net/wireless/ti/wl18xx/debugfs.c
454
DEBUGFS_FWSTATS_ADD(tx, frag_init_called);
drivers/net/wireless/ti/wl18xx/debugfs.c
455
DEBUGFS_FWSTATS_ADD(tx, frag_in_process_called);
drivers/net/wireless/ti/wl18xx/debugfs.c
456
DEBUGFS_FWSTATS_ADD(tx, frag_tkip_called);
drivers/net/wireless/ti/wl18xx/debugfs.c
457
DEBUGFS_FWSTATS_ADD(tx, frag_key_not_found);
drivers/net/wireless/ti/wl18xx/debugfs.c
458
DEBUGFS_FWSTATS_ADD(tx, frag_need_fragmentation);
drivers/net/wireless/ti/wl18xx/debugfs.c
459
DEBUGFS_FWSTATS_ADD(tx, frag_bad_mblk_num);
drivers/net/wireless/ti/wl18xx/debugfs.c
46
WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_cmplt, "%u");
drivers/net/wireless/ti/wl18xx/debugfs.c
460
DEBUGFS_FWSTATS_ADD(tx, frag_failed);
drivers/net/wireless/ti/wl18xx/debugfs.c
461
DEBUGFS_FWSTATS_ADD(tx, frag_cache_hit);
drivers/net/wireless/ti/wl18xx/debugfs.c
462
DEBUGFS_FWSTATS_ADD(tx, frag_cache_miss);
drivers/net/wireless/ti/wl18xx/debugfs.c
47
WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_template_prepared, "%u");
drivers/net/wireless/ti/wl18xx/debugfs.c
48
WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_data_prepared, "%u");
drivers/net/wireless/ti/wl18xx/debugfs.c
49
WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_template_programmed, "%u");
drivers/net/wireless/ti/wl18xx/debugfs.c
50
WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_data_programmed, "%u");
drivers/net/wireless/ti/wl18xx/debugfs.c
51
WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_burst_programmed, "%u");
drivers/net/wireless/ti/wl18xx/debugfs.c
52
WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_starts, "%u");
drivers/net/wireless/ti/wl18xx/debugfs.c
53
WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_stop, "%u");
drivers/net/wireless/ti/wl18xx/debugfs.c
54
WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_start_templates, "%u");
drivers/net/wireless/ti/wl18xx/debugfs.c
55
WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_start_int_templates, "%u");
drivers/net/wireless/ti/wl18xx/debugfs.c
56
WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_start_fw_gen, "%u");
drivers/net/wireless/ti/wl18xx/debugfs.c
57
WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_start_data, "%u");
drivers/net/wireless/ti/wl18xx/debugfs.c
58
WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_start_null_frame, "%u");
drivers/net/wireless/ti/wl18xx/debugfs.c
59
WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_exch, "%u");
drivers/net/wireless/ti/wl18xx/debugfs.c
60
WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_retry_template, "%u");
drivers/net/wireless/ti/wl18xx/debugfs.c
61
WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_retry_data, "%u");
drivers/net/wireless/ti/wl18xx/debugfs.c
62
WL18XX_DEBUGFS_FWSTATS_FILE_ARRAY(tx, tx_retry_per_rate,
drivers/net/wireless/ti/wl18xx/debugfs.c
64
WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_exch_pending, "%u");
drivers/net/wireless/ti/wl18xx/debugfs.c
65
WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_exch_expiry, "%u");
drivers/net/wireless/ti/wl18xx/debugfs.c
66
WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_done_template, "%u");
drivers/net/wireless/ti/wl18xx/debugfs.c
67
WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_done_data, "%u");
drivers/net/wireless/ti/wl18xx/debugfs.c
68
WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_done_int_template, "%u");
drivers/net/wireless/ti/wl18xx/debugfs.c
69
WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_cfe1, "%u");
drivers/net/wireless/ti/wl18xx/debugfs.c
70
WL18XX_DEBUGFS_FWSTATS_FILE(tx, tx_cfe2, "%u");
drivers/net/wireless/ti/wl18xx/debugfs.c
71
WL18XX_DEBUGFS_FWSTATS_FILE(tx, frag_called, "%u");
drivers/net/wireless/ti/wl18xx/debugfs.c
72
WL18XX_DEBUGFS_FWSTATS_FILE(tx, frag_mpdu_alloc_failed, "%u");
drivers/net/wireless/ti/wl18xx/debugfs.c
73
WL18XX_DEBUGFS_FWSTATS_FILE(tx, frag_init_called, "%u");
drivers/net/wireless/ti/wl18xx/debugfs.c
74
WL18XX_DEBUGFS_FWSTATS_FILE(tx, frag_in_process_called, "%u");
drivers/net/wireless/ti/wl18xx/debugfs.c
75
WL18XX_DEBUGFS_FWSTATS_FILE(tx, frag_tkip_called, "%u");
drivers/net/wireless/ti/wl18xx/debugfs.c
76
WL18XX_DEBUGFS_FWSTATS_FILE(tx, frag_key_not_found, "%u");
drivers/net/wireless/ti/wl18xx/debugfs.c
77
WL18XX_DEBUGFS_FWSTATS_FILE(tx, frag_need_fragmentation, "%u");
drivers/net/wireless/ti/wl18xx/debugfs.c
78
WL18XX_DEBUGFS_FWSTATS_FILE(tx, frag_bad_mblk_num, "%u");
drivers/net/wireless/ti/wl18xx/debugfs.c
79
WL18XX_DEBUGFS_FWSTATS_FILE(tx, frag_failed, "%u");
drivers/net/wireless/ti/wl18xx/debugfs.c
80
WL18XX_DEBUGFS_FWSTATS_FILE(tx, frag_cache_hit, "%u");
drivers/net/wireless/ti/wl18xx/debugfs.c
81
WL18XX_DEBUGFS_FWSTATS_FILE(tx, frag_cache_miss, "%u");
drivers/net/wireless/ti/wl18xx/main.c
256
.tx = {
drivers/net/wireless/ti/wlcore/acx.c
1537
acx->max_tx_retry = cpu_to_le16(wl->conf.tx.max_tx_retries);
drivers/net/wireless/ti/wlcore/acx.c
551
detection->tx_energy_detection = wl->conf.tx.tx_energy_detection;
drivers/net/wireless/ti/wlcore/acx.c
725
struct conf_tx_rate_class *c = &wl->conf.tx.sta_rc_conf;
drivers/net/wireless/ti/wlcore/acx.c
906
frag_threshold = wl->conf.tx.frag_threshold;
drivers/net/wireless/ti/wlcore/acx.c
943
acx->tx_compl_timeout = cpu_to_le16(wl->conf.tx.tx_compl_timeout);
drivers/net/wireless/ti/wlcore/acx.c
944
acx->tx_compl_threshold = cpu_to_le16(wl->conf.tx.tx_compl_threshold);
drivers/net/wireless/ti/wlcore/cmd.c
1042
cmd->short_retry_limit = wl->conf.tx.tmpl_short_retry_limit;
drivers/net/wireless/ti/wlcore/cmd.c
1043
cmd->long_retry_limit = wl->conf.tx.tmpl_long_retry_limit;
drivers/net/wireless/ti/wlcore/cmd.c
663
cmd->ap.aging_period = cpu_to_le16(wl->conf.tx.ap_aging_period);
drivers/net/wireless/ti/wlcore/conf.h
1133
struct conf_tx_settings tx;
drivers/net/wireless/ti/wlcore/event.c
256
u32 num_packets = wl->conf.tx.max_tx_retries;
drivers/net/wireless/ti/wlcore/init.c
468
for (i = 0; i < wl->conf.tx.ac_conf_count; i++) {
drivers/net/wireless/ti/wlcore/init.c
611
BUG_ON(wl->conf.tx.tid_conf_count != wl->conf.tx.ac_conf_count);
drivers/net/wireless/ti/wlcore/init.c
612
for (i = 0; i < wl->conf.tx.tid_conf_count; i++) {
drivers/net/wireless/ti/wlcore/init.c
613
conf_ac = &wl->conf.tx.ac_conf[i];
drivers/net/wireless/ti/wlcore/init.c
620
conf_tid = &wl->conf.tx.tid_conf[i];
drivers/net/wireless/ti/wlcore/main.c
205
msecs_to_jiffies(wl->conf.tx.tx_watchdog_timeout));
drivers/net/wireless/ti/wlcore/main.c
2338
wlvif->bitrate_masks[NL80211_BAND_2GHZ] = wl->conf.tx.basic_rate;
drivers/net/wireless/ti/wlcore/main.c
2339
wlvif->bitrate_masks[NL80211_BAND_5GHZ] = wl->conf.tx.basic_rate_5;
drivers/net/wireless/ti/wlcore/main.c
263
wl->conf.tx.tx_watchdog_timeout);
drivers/net/wireless/ti/wlcore/main.c
274
wl->conf.tx.tx_watchdog_timeout);
drivers/net/wireless/ti/wlcore/main.c
288
wl->conf.tx.tx_watchdog_timeout,
drivers/net/wireless/ti/wlcore/main.c
295
wl->conf.tx.tx_watchdog_timeout);
drivers/net/wireless/ti/wlcore/main.c
6002
.tx = wl1271_op_tx,
drivers/net/wireless/virtual/mac80211_hwsim.c
4164
.tx = mac80211_hwsim_tx, \
drivers/net/wireless/zydas/zd1211rw/zd_mac.c
1353
.tx = zd_op_tx,
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
1023
struct zd_usb_tx *tx = &usb->tx;
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
1025
if (!atomic_read(&tx->enabled)) {
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
1040
skb_queue_tail(&tx->submitted_skbs, skb);
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
1041
usb_anchor_urb(urb, &tx->submitted);
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
1047
skb_unlink(skb, &tx->submitted_skbs);
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
1060
struct zd_usb_tx *tx = &usb->tx;
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
1061
struct sk_buff_head *q = &tx->submitted_skbs;
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
1085
container_of(work, struct zd_usb, tx.watchdog_work.work);
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
1086
struct zd_usb_tx *tx = &usb->tx;
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
1088
if (!atomic_read(&tx->enabled) || !tx->watchdog_enabled)
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
1101
queue_delayed_work(zd_workqueue, &tx->watchdog_work,
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
1107
struct zd_usb_tx *tx = &usb->tx;
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
1109
if (!tx->watchdog_enabled) {
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
1111
queue_delayed_work(zd_workqueue, &tx->watchdog_work,
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
1113
tx->watchdog_enabled = 1;
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
1119
struct zd_usb_tx *tx = &usb->tx;
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
1121
if (tx->watchdog_enabled) {
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
1123
tx->watchdog_enabled = 0;
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
1124
cancel_delayed_work_sync(&tx->watchdog_work);
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
1188
struct zd_usb_tx *tx = &usb->tx;
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
1190
spin_lock_init(&tx->lock);
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
1191
atomic_set(&tx->enabled, 0);
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
1192
tx->stopped = 0;
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
1193
skb_queue_head_init(&tx->submitted_skbs);
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
1194
init_usb_anchor(&tx->submitted);
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
1195
tx->submitted_urbs = 0;
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
1196
tx->watchdog_enabled = 0;
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
1197
INIT_DELAYED_WORK(&tx->watchdog_work, zd_tx_watchdog_handler);
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
883
struct zd_usb_tx *tx = &usb->tx;
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
886
atomic_set(&tx->enabled, 0);
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
889
usb_kill_anchored_urbs(&tx->submitted);
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
891
spin_lock_irqsave(&tx->lock, flags);
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
892
WARN_ON(!skb_queue_empty(&tx->submitted_skbs));
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
893
WARN_ON(tx->submitted_urbs != 0);
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
894
tx->submitted_urbs = 0;
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
895
spin_unlock_irqrestore(&tx->lock, flags);
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
912
struct zd_usb_tx *tx = &usb->tx;
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
914
spin_lock_irqsave(&tx->lock, flags);
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
915
atomic_set(&tx->enabled, 1);
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
916
tx->submitted_urbs = 0;
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
918
tx->stopped = 0;
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
919
spin_unlock_irqrestore(&tx->lock, flags);
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
924
struct zd_usb_tx *tx = &usb->tx;
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
927
spin_lock_irqsave(&tx->lock, flags);
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
928
--tx->submitted_urbs;
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
929
if (tx->stopped && tx->submitted_urbs <= ZD_USB_TX_LOW) {
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
931
tx->stopped = 0;
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
933
spin_unlock_irqrestore(&tx->lock, flags);
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
938
struct zd_usb_tx *tx = &usb->tx;
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
941
spin_lock_irqsave(&tx->lock, flags);
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
942
++tx->submitted_urbs;
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
943
if (!tx->stopped && tx->submitted_urbs > ZD_USB_TX_HIGH) {
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
945
tx->stopped = 1;
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
947
spin_unlock_irqrestore(&tx->lock, flags);
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
963
struct zd_usb_tx *tx;
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
972
tx = &usb->tx;
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
990
skb_unlink(skb, &usb->tx.submitted_skbs);
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
996
usb_anchor_urb(urb, &tx->submitted);
drivers/net/wireless/zydas/zd1211rw/zd_usb.h
209
struct zd_usb_tx tx;
drivers/net/wwan/iosm/iosm_ipc_port.c
44
.tx = ipc_port_ctrl_tx,
drivers/net/wwan/mhi_wwan_ctrl.c
166
.tx = mhi_wwan_ctrl_tx,
drivers/net/wwan/qcom_bam_dmux.c
191
desc = dmaengine_prep_slave_single(dmux->tx, skb_dma->addr,
drivers/net/wwan/qcom_bam_dmux.c
266
dma_async_issue_pending(dmux->tx);
drivers/net/wwan/qcom_bam_dmux.c
367
dma_async_issue_pending(dmux->tx);
drivers/net/wwan/qcom_bam_dmux.c
396
dma_async_issue_pending(dmux->tx);
drivers/net/wwan/qcom_bam_dmux.c
659
if (dmux->tx) {
drivers/net/wwan/qcom_bam_dmux.c
660
dmaengine_terminate_sync(dmux->tx);
drivers/net/wwan/qcom_bam_dmux.c
661
dma_release_channel(dmux->tx);
drivers/net/wwan/qcom_bam_dmux.c
662
dmux->tx = NULL;
drivers/net/wwan/qcom_bam_dmux.c
752
if (dmux->tx)
drivers/net/wwan/qcom_bam_dmux.c
755
dmux->tx = dma_request_chan(dev, "tx");
drivers/net/wwan/qcom_bam_dmux.c
756
if (IS_ERR(dmux->tx)) {
drivers/net/wwan/qcom_bam_dmux.c
757
dev_err(dev, "Failed to request TX DMA channel: %pe\n", dmux->tx);
drivers/net/wwan/qcom_bam_dmux.c
758
dmux->tx = NULL;
drivers/net/wwan/qcom_bam_dmux.c
78
struct dma_chan *rx, *tx;
drivers/net/wwan/rpmsg_wwan_ctrl.c
94
.tx = rpmsg_wwan_ctrl_tx,
drivers/net/wwan/t7xx/t7xx_port_wwan.c
148
.tx = t7xx_port_wwan_tx,
drivers/net/wwan/wwan_core.c
788
ret = port->ops->tx(port, skb);
drivers/net/wwan/wwan_hwsim.c
198
.tx = wwan_hwsim_at_emul_tx,
drivers/net/wwan/wwan_hwsim.c
299
.tx = wwan_hwsim_nmea_emul_tx,
drivers/net/xen-netback/common.h
160
struct xen_netif_tx_back_ring tx;
drivers/net/xen-netback/interface.c
81
rc = RING_HAS_UNCONSUMED_REQUESTS(&queue->tx);
drivers/net/xen-netback/netback.c
1103
queue->tx.req_cons = idx;
drivers/net/xen-netback/netback.c
1430
RING_IDX i = queue->tx.rsp_prod_pvt;
drivers/net/xen-netback/netback.c
1433
resp = RING_GET_RESPONSE(&queue->tx, i);
drivers/net/xen-netback/netback.c
1438
RING_GET_RESPONSE(&queue->tx, ++i)->status = XEN_NETIF_RSP_NULL;
drivers/net/xen-netback/netback.c
1440
queue->tx.rsp_prod_pvt = ++i;
drivers/net/xen-netback/netback.c
1447
RING_PUSH_RESPONSES_AND_CHECK_NOTIFY(&queue->tx, notify);
drivers/net/xen-netback/netback.c
1520
if (likely(RING_HAS_UNCONSUMED_REQUESTS(&queue->tx)))
drivers/net/xen-netback/netback.c
1533
if (queue->tx.sring)
drivers/net/xen-netback/netback.c
1535
queue->tx.sring);
drivers/net/xen-netback/netback.c
1560
BACK_RING_ATTACH(&queue->tx, txs, rsp_prod, XEN_PAGE_SIZE);
drivers/net/xen-netback/netback.c
1563
if (req_prod - rsp_prod > RING_SIZE(&queue->tx))
drivers/net/xen-netback/netback.c
170
RING_FINAL_CHECK_FOR_REQUESTS(&queue->tx, more_to_do);
drivers/net/xen-netback/netback.c
211
RING_IDX cons = queue->tx.req_cons;
drivers/net/xen-netback/netback.c
217
RING_COPY_REQUEST(&queue->tx, cons++, txp);
drivers/net/xen-netback/netback.c
220
queue->tx.req_cons = cons;
drivers/net/xen-netback/netback.c
238
RING_IDX cons = queue->tx.req_cons;
drivers/net/xen-netback/netback.c
286
RING_COPY_REQUEST(&queue->tx, cons + slots, txp);
drivers/net/xen-netback/netback.c
732
RING_IDX cons = queue->tx.req_cons;
drivers/net/xen-netback/netback.c
741
RING_COPY_REQUEST(&queue->tx, cons, &extra);
drivers/net/xen-netback/netback.c
743
queue->tx.req_cons = ++cons;
drivers/net/xen-netback/netback.c
934
if (queue->tx.sring->req_prod - queue->tx.req_cons >
drivers/net/xen-netback/netback.c
939
queue->tx.sring->req_prod, queue->tx.req_cons,
drivers/net/xen-netback/netback.c
945
work_to_do = XEN_RING_NR_UNCONSUMED_REQUESTS(&queue->tx);
drivers/net/xen-netback/netback.c
949
idx = queue->tx.req_cons;
drivers/net/xen-netback/netback.c
951
RING_COPY_REQUEST(&queue->tx, idx, &txreq);
drivers/net/xen-netback/netback.c
961
queue->tx.req_cons = ++idx;
drivers/net/xen-netback/netback.c
969
idx = queue->tx.req_cons;
drivers/net/xen-netback/xenbus.c
29
struct xen_netif_tx_back_ring *tx_ring = &queue->tx;
drivers/net/xen-netfront.c
126
struct xen_netif_tx_front_ring tx;
drivers/net/xen-netfront.c
1845
xennet_end_access(queue->tx_ring_ref, queue->tx.sring);
drivers/net/xen-netfront.c
1850
queue->tx.sring = NULL;
drivers/net/xen-netfront.c
1985
queue->tx.sring = NULL;
drivers/net/xen-netfront.c
1992
XEN_FRONT_RING_INIT(&queue->tx, txs, XEN_PAGE_SIZE);
drivers/net/xen-netfront.c
2017
xenbus_teardown_ring((void **)&queue->tx.sring, 1, &queue->tx_ring_ref);
drivers/net/xen-netfront.c
255
return (queue->tx.req_prod_pvt - queue->tx.rsp_cons) <
drivers/net/xen-netfront.c
402
prod = queue->tx.sring->rsp_prod;
drivers/net/xen-netfront.c
403
if (RING_RESPONSE_PROD_OVERFLOW(&queue->tx, prod)) {
drivers/net/xen-netfront.c
405
prod - queue->tx.rsp_cons);
drivers/net/xen-netfront.c
410
for (cons = queue->tx.rsp_cons; cons != prod; cons++) {
drivers/net/xen-netfront.c
415
RING_COPY_RESPONSE(&queue->tx, cons, &txrsp);
drivers/net/xen-netfront.c
420
if (id >= RING_SIZE(&queue->tx)) {
drivers/net/xen-netfront.c
449
queue->tx.rsp_cons = prod;
drivers/net/xen-netfront.c
451
RING_FINAL_CHECK_FOR_RESPONSES(&queue->tx, more_to_do);
drivers/net/xen-netfront.c
469
struct xen_netif_tx_request *tx; /* Last request on ring page */
drivers/net/xen-netfront.c
479
struct xen_netif_tx_request *tx;
drivers/net/xen-netfront.c
487
tx = RING_GET_REQUEST(&queue->tx, queue->tx.req_prod_pvt++);
drivers/net/xen-netfront.c
504
*tx = info->tx_local;
drivers/net/xen-netfront.c
512
info->tx = tx;
drivers/net/xen-netfront.c
524
return info->tx;
drivers/net/xen-netfront.c
532
info->tx->flags |= XEN_NETTXF_more_data;
drivers/net/xen-netfront.c
632
RING_PUSH_REQUESTS_AND_CHECK_NOTIFY(&queue->tx, notify);
drivers/net/xen-netfront.c
807
RING_GET_REQUEST(&queue->tx, queue->tx.req_prod_pvt++);
drivers/net/xen-netfront.c
841
RING_PUSH_REQUESTS_AND_CHECK_NOTIFY(&queue->tx, notify);
drivers/ntb/test/ntb_perf.c
788
struct dma_async_tx_descriptor *tx;
drivers/ntb/test/ntb_perf.c
826
tx = dmaengine_prep_dma_memcpy(pthr->dma_chan, dst_dma_addr,
drivers/ntb/test/ntb_perf.c
828
if (!tx)
drivers/ntb/test/ntb_perf.c
830
} while (!tx && (try++ < DMA_TRIES));
drivers/ntb/test/ntb_perf.c
832
if (!tx) {
drivers/ntb/test/ntb_perf.c
837
tx->callback = perf_dma_copy_callback;
drivers/ntb/test/ntb_perf.c
838
tx->callback_param = pthr;
drivers/ntb/test/ntb_perf.c
839
dma_set_unmap(tx, unmap);
drivers/ntb/test/ntb_perf.c
841
ret = dma_submit_error(dmaengine_submit(tx));
drivers/pci/controller/dwc/pcie-tegra194.c
1209
msg.tx.data = &req;
drivers/pci/controller/dwc/pcie-tegra194.c
1210
msg.tx.size = sizeof(req);
drivers/pci/controller/dwc/pcie-tegra194.c
1244
msg.tx.data = &req;
drivers/pci/controller/dwc/pcie-tegra194.c
1245
msg.tx.size = sizeof(req);
drivers/pci/endpoint/functions/pci-epf-test.c
160
struct dma_async_tx_descriptor *tx;
drivers/pci/endpoint/functions/pci-epf-test.c
181
tx = dmaengine_prep_slave_single(chan, dma_local, len, dir,
drivers/pci/endpoint/functions/pci-epf-test.c
184
tx = dmaengine_prep_dma_memcpy(chan, dma_dst, dma_src, len,
drivers/pci/endpoint/functions/pci-epf-test.c
188
if (!tx) {
drivers/pci/endpoint/functions/pci-epf-test.c
195
tx->callback = pci_epf_test_dma_callback;
drivers/pci/endpoint/functions/pci-epf-test.c
196
tx->callback_param = epf_test;
drivers/pci/endpoint/functions/pci-epf-test.c
197
epf_test->transfer_cookie = dmaengine_submit(tx);
drivers/peci/controller/peci-aspeed.c
218
if (req->tx.len > ASPEED_PECI_DATA_BUF_SIZE_MAX ||
drivers/peci/controller/peci-aspeed.c
231
FIELD_PREP(ASPEED_PECI_WR_LEN_MASK, req->tx.len) |
drivers/peci/controller/peci-aspeed.c
236
for (i = 0; i < req->tx.len; i += 4) {
drivers/peci/controller/peci-aspeed.c
239
writel(get_unaligned_le32(&req->tx.buf[i]), priv->base + reg);
drivers/peci/controller/peci-aspeed.c
244
print_hex_dump_bytes("TX : ", DUMP_PREFIX_NONE, req->tx.buf, req->tx.len);
drivers/peci/controller/peci-npcm.c
100
regmap_write(priv->regmap, NPCM_PECI_DAT_INOUT(i), req->tx.buf[i + 1]);
drivers/peci/controller/peci-npcm.c
105
addr, req->tx.len, req->rx.len);
drivers/peci/controller/peci-npcm.c
106
print_hex_dump_bytes("TX : ", DUMP_PREFIX_NONE, req->tx.buf, req->tx.len);
drivers/peci/controller/peci-npcm.c
94
regmap_write(priv->regmap, NPCM_PECI_WR_LENGTH, NPCM_PECI_WR_LEN_MASK & req->tx.len);
drivers/peci/controller/peci-npcm.c
96
if (req->tx.len) {
drivers/peci/controller/peci-npcm.c
97
regmap_write(priv->regmap, NPCM_PECI_CMD, req->tx.buf[0]);
drivers/peci/controller/peci-npcm.c
99
for (i = 0; i < (req->tx.len - 1); i++)
drivers/peci/request.c
155
if (WARN_ON(req->tx.len == 0))
drivers/peci/request.c
169
req->tx.buf[1] |= PECI_RETRY_BIT;
drivers/peci/request.c
211
req->tx.len = tx_len;
drivers/peci/request.c
237
req->tx.buf[0] = PECI_GET_DIB_CMD;
drivers/peci/request.c
258
req->tx.buf[0] = PECI_GET_TEMP_CMD;
drivers/peci/request.c
280
req->tx.buf[0] = PECI_RDPKGCFG_CMD;
drivers/peci/request.c
281
req->tx.buf[1] = 0;
drivers/peci/request.c
282
req->tx.buf[2] = index;
drivers/peci/request.c
283
put_unaligned_le16(param, &req->tx.buf[3]);
drivers/peci/request.c
313
req->tx.buf[0] = PECI_RDPCICFGLOCAL_CMD;
drivers/peci/request.c
314
req->tx.buf[1] = 0;
drivers/peci/request.c
315
put_unaligned_le24(pci_addr, &req->tx.buf[2]);
drivers/peci/request.c
341
req->tx.buf[0] = PECI_RDENDPTCFG_CMD;
drivers/peci/request.c
342
req->tx.buf[1] = 0;
drivers/peci/request.c
343
req->tx.buf[2] = msg_type;
drivers/peci/request.c
344
req->tx.buf[3] = 0;
drivers/peci/request.c
345
req->tx.buf[4] = 0;
drivers/peci/request.c
346
req->tx.buf[5] = 0;
drivers/peci/request.c
347
req->tx.buf[6] = PECI_ENDPTCFG_ADDR_TYPE_PCI;
drivers/peci/request.c
348
req->tx.buf[7] = seg; /* PCI Segment */
drivers/peci/request.c
349
put_unaligned_le32(pci_addr, &req->tx.buf[8]);
drivers/peci/request.c
371
req->tx.buf[0] = PECI_RDENDPTCFG_CMD;
drivers/peci/request.c
372
req->tx.buf[1] = 0;
drivers/peci/request.c
373
req->tx.buf[2] = PECI_ENDPTCFG_TYPE_MMIO;
drivers/peci/request.c
374
req->tx.buf[3] = 0; /* Endpoint ID */
drivers/peci/request.c
375
req->tx.buf[4] = 0; /* Reserved */
drivers/peci/request.c
376
req->tx.buf[5] = bar;
drivers/peci/request.c
377
req->tx.buf[6] = addr_type;
drivers/peci/request.c
378
req->tx.buf[7] = seg; /* PCI Segment */
drivers/peci/request.c
379
req->tx.buf[8] = PCI_DEVFN(dev, func);
drivers/peci/request.c
380
req->tx.buf[9] = bus; /* PCI Bus */
drivers/peci/request.c
383
put_unaligned_le32(offset, &req->tx.buf[10]);
drivers/peci/request.c
385
put_unaligned_le64(offset, &req->tx.buf[10]);
drivers/phy/qualcomm/phy-qcom-qmp-combo.c
2275
void __iomem *tx;
drivers/phy/qualcomm/phy-qcom-qmp-combo.c
3818
void __iomem *tx = qmp->dp_tx;
drivers/phy/qualcomm/phy-qcom-qmp-combo.c
3825
qmp_configure_lane(qmp->dev, tx, cfg->dp_tx_tbl, cfg->dp_tx_tbl_num, 1);
drivers/phy/qualcomm/phy-qcom-qmp-combo.c
3862
void __iomem *tx = qmp->tx;
drivers/phy/qualcomm/phy-qcom-qmp-combo.c
3881
qmp_configure_lane(qmp->dev, tx, cfg->tx_tbl, cfg->tx_tbl_num, 1);
drivers/phy/qualcomm/phy-qcom-qmp-combo.c
4624
qmp->tx = devm_of_iomap(dev, np, 0, NULL);
drivers/phy/qualcomm/phy-qcom-qmp-combo.c
4625
if (IS_ERR(qmp->tx))
drivers/phy/qualcomm/phy-qcom-qmp-combo.c
4626
return PTR_ERR(qmp->tx);
drivers/phy/qualcomm/phy-qcom-qmp-combo.c
4714
qmp->tx = base + offs->txa;
drivers/phy/qualcomm/phy-qcom-qmp-pcie-msm8996.c
186
void __iomem *tx;
drivers/phy/qualcomm/phy-qcom-qmp-pcie-msm8996.c
411
void __iomem *tx = qphy->tx;
drivers/phy/qualcomm/phy-qcom-qmp-pcie-msm8996.c
434
qmp_configure_lane(qmp->dev, tx, cfg->tx_tbl, cfg->tx_tbl_num, 1);
drivers/phy/qualcomm/phy-qcom-qmp-pcie-msm8996.c
671
qphy->tx = devm_of_iomap(dev, np, 0, NULL);
drivers/phy/qualcomm/phy-qcom-qmp-pcie-msm8996.c
672
if (IS_ERR(qphy->tx))
drivers/phy/qualcomm/phy-qcom-qmp-pcie-msm8996.c
673
return PTR_ERR(qphy->tx);
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3228
u16 tx;
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3241
const struct qmp_phy_init_tbl *tx;
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3314
void __iomem *tx;
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3397
.tx = 0x0800,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3404
.tx = 0x0200,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3412
.tx = 0x0200,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3420
.tx = 0x0200,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3428
.tx = 0x0200,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3439
.tx = 0x0000,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3449
.tx = 0x0e00,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3459
.tx = 0x0200,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3470
.tx = 0x0000,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3481
.tx = 0x0,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3491
.tx = 0x1000,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3501
.tx = 0x0000,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3512
.tx = 0x0000,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3523
.tx = 0x0000,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3543
.tx = ipq8074_pcie_tx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3568
.tx = ipq8074_pcie_gen3_tx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3597
.tx = ipq6018_pcie_tx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3624
.tx = ipq8074_pcie_gen3_tx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3652
.tx = ipq8074_pcie_gen3_tx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3680
.tx = qcs615_pcie_tx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3704
.tx = sa8775p_qmp_gen4_pcie_tx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3739
.tx = sdm845_qmp_pcie_tx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3766
.tx = sdm845_qhp_pcie_tx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3789
.tx = sm8250_qmp_pcie_tx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3826
.tx = sm8250_qmp_pcie_tx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3836
.tx = sm8250_qmp_gen3x2_pcie_tx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3863
.tx = msm8998_pcie_tx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3888
.tx = sm8550_qmp_gen3x2_pcie_tx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3900
.tx = sar2130p_qmp_gen3x2_pcie_rc_tx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3933
.tx = sc8180x_qmp_pcie_tx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3960
.tx = sc8280xp_qmp_gen3x1_pcie_tx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3993
.tx = sc8280xp_qmp_gen3x2_pcie_tx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
4026
.tx = sc8280xp_qmp_gen3x2_pcie_tx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
4062
.tx = sdx55_qmp_pcie_tx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
4104
.tx = sm8350_qmp_gen3x1_pcie_tx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
4139
.tx = sm8350_qmp_gen3x2_pcie_tx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
4174
.tx = sdx65_qmp_pcie_tx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
4203
.tx = sm8450_qmp_gen3x1_pcie_tx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
4238
.tx = sm8450_qmp_gen4x2_pcie_tx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
4283
.tx = sm8550_qmp_gen3x2_pcie_tx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
4310
.tx = sm8750_qmp_gen3x2_pcie_tx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
4337
.tx = sm8550_qmp_gen4x2_pcie_tx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
4369
.tx = sm8550_qmp_gen4x2_pcie_tx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
4400
.tx = sa8775p_qmp_gen4_pcie_tx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
4448
.tx = sa8775p_qmp_gen4_pcie_tx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
4490
.tx = x1e80100_qmp_gen4x2_pcie_tx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
4520
.tx = x1e80100_qmp_gen4x2_pcie_tx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
4599
.tx = kaanapali_qmp_gen3x2_pcie_tx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
4658
tx3 = qmp->port_b + offs->tx;
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
4669
qmp_configure_lane(qmp->dev, tx3, tbls->tx, tbls->tx_num, 1);
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
4672
qmp_configure_lane(qmp->dev, tx4, tbls->tx, tbls->tx_num, 2);
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
4685
void __iomem *tx = qmp->tx;
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
4706
qmp_configure_lane(qmp->dev, tx, tbls->tx, tbls->tx_num, 1);
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
4710
qmp_configure_lane(qmp->dev, tx2, tbls->tx, tbls->tx_num, 2);
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
5178
qmp->tx = devm_of_iomap(dev, np, 0, NULL);
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
5179
if (IS_ERR(qmp->tx))
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
5180
return PTR_ERR(qmp->tx);
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
5183
qmp->rx = qmp->tx;
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
5295
qmp->tx = base + offs->tx;
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1135
u16 tx;
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1145
const struct qmp_phy_init_tbl *tx;
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1189
void __iomem *tx;
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1312
.tx = 0x400,
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1321
.tx = 0x1000,
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1336
.tx = milos_ufsphy_tx,
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1362
.tx = msm8996_ufsphy_tx,
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1385
.tx = sm8350_ufsphy_tx,
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1397
.tx = sm8350_ufsphy_g4_tx,
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1419
.tx = sc7280_ufsphy_tx,
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1431
.tx = sm8250_ufsphy_hs_g4_tx,
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1453
.tx = sm8350_ufsphy_tx,
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1465
.tx = sm8350_ufsphy_g4_tx,
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1487
.tx = sdm845_ufsphy_tx,
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1514
.tx = sm6115_ufsphy_tx,
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1541
.tx = sdm845_ufsphy_tx,
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1568
.tx = sm8150_ufsphy_tx,
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1580
.tx = sm8150_ufsphy_hs_g4_tx,
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1602
.tx = sm8150_ufsphy_tx,
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1614
.tx = sm8250_ufsphy_hs_g4_tx,
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1636
.tx = sm8350_ufsphy_tx,
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1648
.tx = sm8350_ufsphy_g4_tx,
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1670
.tx = sm8350_ufsphy_tx,
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1682
.tx = sm8350_ufsphy_g4_tx,
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1704
.tx = sm8550_ufsphy_tx,
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1718
.tx = sm8550_ufsphy_g4_tx,
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1740
.tx = sm8550_ufsphy_tx,
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1754
.tx = sm8550_ufsphy_g4_tx,
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1785
.tx = sm8650_ufsphy_tx,
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1817
.tx = sm8750_ufsphy_tx,
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1857
void __iomem *tx = qmp->tx;
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1860
qmp_configure_lane(qmp->dev, tx, tbls->tx, tbls->tx_num, 1);
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1864
qmp_configure_lane(qmp->dev, qmp->tx2, tbls->tx, tbls->tx_num, 2);
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
2137
qmp->tx = devm_of_iomap(dev, np, 0, NULL);
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
2138
if (IS_ERR(qmp->tx))
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
2139
return PTR_ERR(qmp->tx);
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
2185
qmp->tx = base + offs->tx;
drivers/phy/qualcomm/phy-qcom-qmp-usb-legacy.c
1170
qmp->tx = devm_of_iomap(dev, np, 0, NULL);
drivers/phy/qualcomm/phy-qcom-qmp-usb-legacy.c
1171
if (IS_ERR(qmp->tx))
drivers/phy/qualcomm/phy-qcom-qmp-usb-legacy.c
1172
return PTR_ERR(qmp->tx);
drivers/phy/qualcomm/phy-qcom-qmp-usb-legacy.c
1226
qmp->tx = base + offs->tx;
drivers/phy/qualcomm/phy-qcom-qmp-usb-legacy.c
485
u16 tx;
drivers/phy/qualcomm/phy-qcom-qmp-usb-legacy.c
531
void __iomem *tx;
drivers/phy/qualcomm/phy-qcom-qmp-usb-legacy.c
827
void __iomem *tx = qmp->tx;
drivers/phy/qualcomm/phy-qcom-qmp-usb-legacy.c
843
qmp_usb_legacy_configure_lane(tx, cfg->tx_tbl, cfg->tx_tbl_num, 1);
drivers/phy/qualcomm/phy-qcom-qmp-usb.c
1382
u16 tx;
drivers/phy/qualcomm/phy-qcom-qmp-usb.c
1425
void __iomem *tx;
drivers/phy/qualcomm/phy-qcom-qmp-usb.c
1490
.tx = 0x200,
drivers/phy/qualcomm/phy-qcom-qmp-usb.c
1498
.tx = 0x200,
drivers/phy/qualcomm/phy-qcom-qmp-usb.c
1505
.tx = 0x200,
drivers/phy/qualcomm/phy-qcom-qmp-usb.c
1513
.tx = 0x0200,
drivers/phy/qualcomm/phy-qcom-qmp-usb.c
1521
.tx = 0x0e00,
drivers/phy/qualcomm/phy-qcom-qmp-usb.c
1529
.tx = 0x0e00,
drivers/phy/qualcomm/phy-qcom-qmp-usb.c
1537
.tx = 0x0e00,
drivers/phy/qualcomm/phy-qcom-qmp-usb.c
1545
.tx = 0x0e00,
drivers/phy/qualcomm/phy-qcom-qmp-usb.c
1939
void __iomem *tx = qmp->tx;
drivers/phy/qualcomm/phy-qcom-qmp-usb.c
1956
qmp_configure_lane(qmp->dev, tx, cfg->tx_tbl, cfg->tx_tbl_num, 1);
drivers/phy/qualcomm/phy-qcom-qmp-usb.c
2294
qmp->tx = devm_of_iomap(dev, np, 0, NULL);
drivers/phy/qualcomm/phy-qcom-qmp-usb.c
2295
if (IS_ERR(qmp->tx))
drivers/phy/qualcomm/phy-qcom-qmp-usb.c
2296
return PTR_ERR(qmp->tx);
drivers/phy/qualcomm/phy-qcom-qmp-usb.c
2358
qmp->tx = base + offs->tx;
drivers/phy/qualcomm/phy-qcom-qmp-usbc.c
1020
qmp_configure_lane(qmp->dev, qmp->tx, cfg->tx_tbl, cfg->tx_tbl_num, 1);
drivers/phy/qualcomm/phy-qcom-qmp-usbc.c
1257
void __iomem *tx = qmp->dp_tx;
drivers/phy/qualcomm/phy-qcom-qmp-usbc.c
1272
qmp_configure_lane(qmp->dev, tx, cfg->dp_tx_tbl, cfg->dp_tx_tbl_num, 1);
drivers/phy/qualcomm/phy-qcom-qmp-usbc.c
1774
qmp->tx = devm_of_iomap(dev, np, 0, NULL);
drivers/phy/qualcomm/phy-qcom-qmp-usbc.c
1775
if (IS_ERR(qmp->tx))
drivers/phy/qualcomm/phy-qcom-qmp-usbc.c
1776
return PTR_ERR(qmp->tx);
drivers/phy/qualcomm/phy-qcom-qmp-usbc.c
1847
qmp->tx = base + offs->tx;
drivers/phy/qualcomm/phy-qcom-qmp-usbc.c
374
u16 tx;
drivers/phy/qualcomm/phy-qcom-qmp-usbc.c
438
void __iomem *tx;
drivers/phy/qualcomm/phy-qcom-qmp-usbc.c
538
.tx = 0x200,
drivers/phy/qualcomm/phy-qcom-qmp-usbc.c
548
.tx = 0x200,
drivers/phy/qualcomm/phy-qcom-qmp-usbc.c
781
void __iomem *tx = qmp->dp_tx;
drivers/phy/qualcomm/phy-qcom-qmp-usbc.c
812
writel(voltage_swing_cfg, tx + QSERDES_V2_TX_TX_DRV_LVL);
drivers/phy/qualcomm/phy-qcom-qmp-usbc.c
813
writel(pre_emphasis_cfg, tx + QSERDES_V2_TX_TX_EMP_POST1_LVL);
drivers/phy/qualcomm/phy-qcom-qmp-usbc.c
870
void __iomem *tx = qmp->dp_tx;
drivers/phy/qualcomm/phy-qcom-qmp-usbc.c
874
writel(0x2a, tx + QSERDES_V2_TX_TX_DRV_LVL);
drivers/phy/qualcomm/phy-qcom-qmp-usbc.c
875
writel(0x20, tx + QSERDES_V2_TX_TX_EMP_POST1_LVL);
drivers/phy/qualcomm/phy-qcom-qmp-usbc.c
880
writel(0xc4, tx + QSERDES_V2_TX_LANE_MODE_1);
drivers/phy/qualcomm/phy-qcom-qmp-usbc.c
883
writel(0xc6, tx + QSERDES_V2_TX_LANE_MODE_1);
drivers/pinctrl/pinctrl-mcp23s08_spi.c
65
u8 tx[2];
drivers/pinctrl/pinctrl-mcp23s08_spi.c
70
tx[0] = mcp->addr | 0x01;
drivers/pinctrl/pinctrl-mcp23s08_spi.c
71
tx[1] = *((u8 *) reg);
drivers/pinctrl/pinctrl-mcp23s08_spi.c
73
return spi_write_then_read(spi, tx, sizeof(tx), val, val_size);
drivers/pinctrl/renesas/pfc-r8a7778.c
1291
#define CAN_PFC_DATA(name, tx, rx) SH_PFC_MUX2(name, tx, rx)
drivers/pinctrl/renesas/pfc-r8a7778.c
1334
#define SCIF_PFC_DAT(name, tx, rx) SH_PFC_MUX2(name, tx, rx)
drivers/pinctrl/renesas/pfc-r8a7778.c
1366
#define HSPI_PFC_DAT(name, clk, cs, rx, tx) SH_PFC_MUX4(name, clk, cs, rx, tx)
drivers/platform/mellanox/mlxbf-tmfifo.c
1297
ctl = readq(fifo->tx.ctl);
drivers/platform/mellanox/mlxbf-tmfifo.c
1306
writeq(ctl, fifo->tx.ctl);
drivers/platform/mellanox/mlxbf-tmfifo.c
1369
fifo->tx.ctl = fifo->res1 + MLXBF_TMFIFO_TX_CTL_BF3;
drivers/platform/mellanox/mlxbf-tmfifo.c
1370
fifo->tx.sts = fifo->res1 + MLXBF_TMFIFO_TX_STS_BF3;
drivers/platform/mellanox/mlxbf-tmfifo.c
1371
fifo->tx.data = fifo->res0 + MLXBF_TMFIFO_TX_DATA_BF3;
drivers/platform/mellanox/mlxbf-tmfifo.c
1376
fifo->tx.ctl = fifo->res1 + MLXBF_TMFIFO_TX_CTL;
drivers/platform/mellanox/mlxbf-tmfifo.c
1377
fifo->tx.sts = fifo->res1 + MLXBF_TMFIFO_TX_STS;
drivers/platform/mellanox/mlxbf-tmfifo.c
1378
fifo->tx.data = fifo->res1 + MLXBF_TMFIFO_TX_DATA;
drivers/platform/mellanox/mlxbf-tmfifo.c
196
struct mlxbf_tmfifo_io tx;
drivers/platform/mellanox/mlxbf-tmfifo.c
442
int rx, tx;
drivers/platform/mellanox/mlxbf-tmfifo.c
445
tx = !test_and_set_bit(MLXBF_TM_TX_LWM_IRQ, &fifo->pend_events);
drivers/platform/mellanox/mlxbf-tmfifo.c
447
if (rx || tx)
drivers/platform/mellanox/mlxbf-tmfifo.c
532
sts = readq(fifo->tx.sts);
drivers/platform/mellanox/mlxbf-tmfifo.c
568
writeq(*(u64 *)&hdr, fifo->tx.data);
drivers/platform/mellanox/mlxbf-tmfifo.c
585
writeq(data, fifo->tx.data);
drivers/platform/mellanox/mlxbf-tmfifo.c
645
writeq(data, fifo->tx.data);
drivers/platform/mellanox/mlxbf-tmfifo.c
721
writeq(*(u64 *)&hdr, fifo->tx.data);
drivers/platform/mellanox/mlxbf-tmfifo.c
911
writeq(0, vring->fifo->tx.data);
drivers/platform/surface/aggregator/ssh_packet_layer.c
1015
wake_up_all(&packet->ptl->tx.packet_wq);
drivers/platform/surface/aggregator/ssh_packet_layer.c
1022
status = wait_for_completion_interruptible(&ptl->tx.thread_cplt_pkt);
drivers/platform/surface/aggregator/ssh_packet_layer.c
1023
reinit_completion(&ptl->tx.thread_cplt_pkt);
drivers/platform/surface/aggregator/ssh_packet_layer.c
1038
status = wait_for_completion_interruptible_timeout(&ptl->tx.thread_cplt_tx,
drivers/platform/surface/aggregator/ssh_packet_layer.c
1040
reinit_completion(&ptl->tx.thread_cplt_tx);
drivers/platform/surface/aggregator/ssh_packet_layer.c
1088
if (kthread_should_stop() || !atomic_read(&ptl->tx.running))
drivers/platform/surface/aggregator/ssh_packet_layer.c
1103
while (!kthread_should_stop() && atomic_read(&ptl->tx.running)) {
drivers/platform/surface/aggregator/ssh_packet_layer.c
1143
complete(&ptl->tx.thread_cplt_pkt);
drivers/platform/surface/aggregator/ssh_packet_layer.c
1154
atomic_set_release(&ptl->tx.running, 1);
drivers/platform/surface/aggregator/ssh_packet_layer.c
1156
ptl->tx.thread = kthread_run(ssh_ptl_tx_threadfn, ptl, "ssam_serial_hub-tx");
drivers/platform/surface/aggregator/ssh_packet_layer.c
1157
if (IS_ERR(ptl->tx.thread))
drivers/platform/surface/aggregator/ssh_packet_layer.c
1158
return PTR_ERR(ptl->tx.thread);
drivers/platform/surface/aggregator/ssh_packet_layer.c
1173
if (!IS_ERR_OR_NULL(ptl->tx.thread)) {
drivers/platform/surface/aggregator/ssh_packet_layer.c
1175
atomic_set_release(&ptl->tx.running, 0);
drivers/platform/surface/aggregator/ssh_packet_layer.c
1182
complete(&ptl->tx.thread_cplt_pkt);
drivers/platform/surface/aggregator/ssh_packet_layer.c
1183
complete(&ptl->tx.thread_cplt_tx);
drivers/platform/surface/aggregator/ssh_packet_layer.c
1186
status = kthread_stop(ptl->tx.thread);
drivers/platform/surface/aggregator/ssh_packet_layer.c
1187
ptl->tx.thread = NULL;
drivers/platform/surface/aggregator/ssh_packet_layer.c
1239
wait_event(packet->ptl->tx.packet_wq,
drivers/platform/surface/aggregator/ssh_packet_layer.c
2041
ptl->tx.thread = NULL;
drivers/platform/surface/aggregator/ssh_packet_layer.c
2042
atomic_set(&ptl->tx.running, 0);
drivers/platform/surface/aggregator/ssh_packet_layer.c
2043
init_completion(&ptl->tx.thread_cplt_pkt);
drivers/platform/surface/aggregator/ssh_packet_layer.c
2044
init_completion(&ptl->tx.thread_cplt_tx);
drivers/platform/surface/aggregator/ssh_packet_layer.c
2045
init_waitqueue_head(&ptl->tx.packet_wq);
drivers/platform/surface/aggregator/ssh_packet_layer.c
995
wake_up_all(&ptl->tx.packet_wq);
drivers/platform/surface/aggregator/ssh_packet_layer.h
181
complete(&ptl->tx.thread_cplt_tx);
drivers/platform/surface/aggregator/ssh_packet_layer.h
98
} tx;
drivers/platform/surface/aggregator/ssh_request_layer.c
1044
INIT_WORK(&rtl->tx.work, ssh_rtl_tx_work_fn);
drivers/platform/surface/aggregator/ssh_request_layer.c
1234
cancel_work_sync(&rtl->tx.work);
drivers/platform/surface/aggregator/ssh_request_layer.c
313
return schedule_work(&rtl->tx.work);
drivers/platform/surface/aggregator/ssh_request_layer.c
318
struct ssh_rtl *rtl = to_ssh_rtl(work, tx.work);
drivers/platform/surface/aggregator/ssh_request_layer.h
82
} tx;
drivers/pmdomain/tegra/powergate-bpmp.c
101
msg.tx.data = &request;
drivers/pmdomain/tegra/powergate-bpmp.c
102
msg.tx.size = sizeof(request);
drivers/pmdomain/tegra/powergate-bpmp.c
131
msg.tx.data = &request;
drivers/pmdomain/tegra/powergate-bpmp.c
132
msg.tx.size = sizeof(request);
drivers/pmdomain/tegra/powergate-bpmp.c
45
msg.tx.data = &request;
drivers/pmdomain/tegra/powergate-bpmp.c
46
msg.tx.size = sizeof(request);
drivers/pmdomain/tegra/powergate-bpmp.c
73
msg.tx.data = &request;
drivers/pmdomain/tegra/powergate-bpmp.c
74
msg.tx.size = sizeof(request);
drivers/ps3/ps3-vuart.c
148
u64 tx;
drivers/ps3/ps3-vuart.c
151
PARAM_TX_TRIGGER, &tx);
drivers/ps3/ps3-vuart.c
152
trig->tx = tx;
drivers/ps3/ps3-vuart.c
181
trig->tx, trig->rx);
drivers/ps3/ps3-vuart.c
186
int ps3_vuart_set_triggers(struct ps3_system_bus_device *dev, unsigned int tx,
drivers/ps3/ps3-vuart.c
193
PARAM_TX_TRIGGER, tx);
drivers/ps3/ps3-vuart.c
220
tx, rx);
drivers/ps3/vuart.h
75
unsigned long tx;
drivers/ps3/vuart.h
80
int ps3_vuart_set_triggers(struct ps3_system_bus_device *dev, unsigned int tx,
drivers/ptp/ptp_clock.c
122
static int ptp_clock_adjtime(struct posix_clock *pc, struct __kernel_timex *tx)
drivers/ptp/ptp_clock.c
128
if (tx->modes & (ADJ_SETOFFSET | ADJ_FREQUENCY | ADJ_OFFSET) &&
drivers/ptp/ptp_clock.c
136
if (tx->modes & ADJ_SETOFFSET) {
drivers/ptp/ptp_clock.c
141
ts.tv_sec = tx->time.tv_sec;
drivers/ptp/ptp_clock.c
142
ts.tv_nsec = tx->time.tv_usec;
drivers/ptp/ptp_clock.c
144
if (!(tx->modes & ADJ_NANO))
drivers/ptp/ptp_clock.c
161
} else if (tx->modes & ADJ_FREQUENCY) {
drivers/ptp/ptp_clock.c
162
long ppb = scaled_ppm_to_ppb(tx->freq);
drivers/ptp/ptp_clock.c
165
err = ops->adjfine(ops, tx->freq);
drivers/ptp/ptp_clock.c
167
ptp->dialed_frequency = tx->freq;
drivers/ptp/ptp_clock.c
168
} else if (tx->modes & ADJ_OFFSET) {
drivers/ptp/ptp_clock.c
171
s32 offset = tx->offset;
drivers/ptp/ptp_clock.c
173
if (!(tx->modes & ADJ_NANO))
drivers/ptp/ptp_clock.c
181
} else if (tx->modes == 0) {
drivers/ptp/ptp_clock.c
182
tx->freq = ptp->dialed_frequency;
drivers/pwm/pwm-mc33xs2410.c
103
tx[i] = FIELD_PREP(MC33XS2410_FRAME_IN_DATA, flag) |
drivers/pwm/pwm-mc33xs2410.c
72
u16 tx[MC33XS2410_MAX_TRANSFERS];
drivers/pwm/pwm-mc33xs2410.c
79
tx[i] = FIELD_PREP(MC33XS2410_FRAME_IN_DATA, val[i]) |
drivers/pwm/pwm-mc33xs2410.c
83
return spi_write(spi, tx, len * 2);
drivers/pwm/pwm-mc33xs2410.c
89
u16 tx[MC33XS2410_MAX_TRANSFERS];
drivers/pwm/pwm-mc33xs2410.c
92
.tx_buf = tx,
drivers/rapidio/devices/rio_mport_cdev.c
680
struct dma_async_tx_descriptor *tx;
drivers/rapidio/devices/rio_mport_cdev.c
699
tx = prep_dma_xfer(chan, xfer, sgt, nents, dir,
drivers/rapidio/devices/rio_mport_cdev.c
702
if (!tx) {
drivers/rapidio/devices/rio_mport_cdev.c
708
} else if (IS_ERR(tx)) {
drivers/rapidio/devices/rio_mport_cdev.c
709
ret = PTR_ERR(tx);
drivers/rapidio/devices/rio_mport_cdev.c
716
tx->callback = dma_xfer_callback;
drivers/rapidio/devices/rio_mport_cdev.c
717
tx->callback_param = req;
drivers/rapidio/devices/rio_mport_cdev.c
722
cookie = dmaengine_submit(tx);
drivers/reset/tegra/reset-bpmp.c
31
msg.tx.data = &request;
drivers/reset/tegra/reset-bpmp.c
32
msg.tx.size = sizeof(request);
drivers/rpmsg/qcom_glink_native.c
1893
struct qcom_glink_pipe *tx,
drivers/rpmsg/qcom_glink_native.c
1904
glink->tx_pipe = tx;
drivers/rpmsg/qcom_glink_native.h
36
struct qcom_glink_pipe *tx,
drivers/rpmsg/qcom_glink_rpm.c
217
struct glink_rpm_pipe *tx)
drivers/rpmsg/qcom_glink_rpm.c
266
tx->native.length = size;
drivers/rpmsg/qcom_glink_rpm.c
268
tx->tail = msg_ram + offset;
drivers/rpmsg/qcom_glink_rpm.c
269
tx->head = msg_ram + offset + sizeof(u32);
drivers/rpmsg/qcom_glink_rpm.c
270
tx->fifo = msg_ram + offset + 2 * sizeof(u32);
drivers/rpmsg/qcom_glink_rpm.c
275
if (!rx->fifo || !tx->fifo) {
drivers/rpmsg/qcom_glink_trace.h
106
__entry->tx = tx;
drivers/rpmsg/qcom_glink_trace.h
109
__entry->tx ? "tx" : "rx",
drivers/rpmsg/qcom_glink_trace.h
120
TP_PROTO(const char *remote, const char *channel, u16 lcid, u16 rcid, bool tx),
drivers/rpmsg/qcom_glink_trace.h
121
TP_ARGS(remote, channel, lcid, rcid, tx),
drivers/rpmsg/qcom_glink_trace.h
127
__field(bool, tx)
drivers/rpmsg/qcom_glink_trace.h
134
__entry->tx = tx;
drivers/rpmsg/qcom_glink_trace.h
137
__entry->tx ? "tx" : "rx",
drivers/rpmsg/qcom_glink_trace.h
14
TP_PROTO(const char *remote, unsigned int version, unsigned int features, bool tx),
drivers/rpmsg/qcom_glink_trace.h
148
TP_PROTO(const char *remote, const char *channel, u16 lcid, u16 rcid, size_t count, size_t size, u32 liid, bool tx),
drivers/rpmsg/qcom_glink_trace.h
149
TP_ARGS(remote, channel, lcid, rcid, count, size, liid, tx),
drivers/rpmsg/qcom_glink_trace.h
15
TP_ARGS(remote, version, features, tx),
drivers/rpmsg/qcom_glink_trace.h
158
__field(bool, tx)
drivers/rpmsg/qcom_glink_trace.h
168
__entry->tx = tx;
drivers/rpmsg/qcom_glink_trace.h
171
__entry->tx ? "tx" : "rx",
drivers/rpmsg/qcom_glink_trace.h
185
TP_PROTO(const char *remote, const char *channel, u16 lcid, u16 rcid, u32 iid, bool reuse, bool tx),
drivers/rpmsg/qcom_glink_trace.h
186
TP_ARGS(remote, channel, lcid, rcid, iid, reuse, tx),
drivers/rpmsg/qcom_glink_trace.h
194
__field(bool, tx)
drivers/rpmsg/qcom_glink_trace.h
20
__field(bool, tx)
drivers/rpmsg/qcom_glink_trace.h
203
__entry->tx = tx;
drivers/rpmsg/qcom_glink_trace.h
206
__entry->tx ? "tx" : "rx",
drivers/rpmsg/qcom_glink_trace.h
219
TP_PROTO(const char *remote, const char *channel, u16 lcid, u16 rcid, size_t size, bool tx),
drivers/rpmsg/qcom_glink_trace.h
220
TP_ARGS(remote, channel, lcid, rcid, size, tx),
drivers/rpmsg/qcom_glink_trace.h
227
__field(bool, tx)
drivers/rpmsg/qcom_glink_trace.h
235
__entry->tx = tx;
drivers/rpmsg/qcom_glink_trace.h
238
__entry->tx ? "tx" : "rx",
drivers/rpmsg/qcom_glink_trace.h
250
TP_PROTO(const char *remote, const char *channel, u16 lcid, u16 rcid, bool granted, bool tx),
drivers/rpmsg/qcom_glink_trace.h
251
TP_ARGS(remote, channel, lcid, rcid, granted, tx),
drivers/rpmsg/qcom_glink_trace.h
258
__field(bool, tx)
drivers/rpmsg/qcom_glink_trace.h
26
__entry->tx = tx;
drivers/rpmsg/qcom_glink_trace.h
266
__entry->tx = tx;
drivers/rpmsg/qcom_glink_trace.h
269
__entry->tx ? "tx" : "rx",
drivers/rpmsg/qcom_glink_trace.h
281
TP_PROTO(const char *remote, const char *channel, u16 lcid, u16 rcid, u32 iid, u32 chunk_size, u32 left_size, bool cont, bool tx),
drivers/rpmsg/qcom_glink_trace.h
282
TP_ARGS(remote, channel, lcid, rcid, iid, chunk_size, left_size, cont, tx),
drivers/rpmsg/qcom_glink_trace.h
29
__entry->tx ? "tx" : "rx",
drivers/rpmsg/qcom_glink_trace.h
292
__field(bool, tx)
drivers/rpmsg/qcom_glink_trace.h
303
__entry->tx = tx;
drivers/rpmsg/qcom_glink_trace.h
306
__entry->tx ? "tx" : "rx",
drivers/rpmsg/qcom_glink_trace.h
321
TP_PROTO(const char *remote, const char *channel, u16 lcid, u16 rcid, bool tx),
drivers/rpmsg/qcom_glink_trace.h
322
TP_ARGS(remote, channel, lcid, rcid, tx),
drivers/rpmsg/qcom_glink_trace.h
328
__field(bool, tx)
drivers/rpmsg/qcom_glink_trace.h
335
__entry->tx = tx;
drivers/rpmsg/qcom_glink_trace.h
338
__entry->tx ? "tx" : "rx",
drivers/rpmsg/qcom_glink_trace.h
349
TP_PROTO(const char *remote, bool tx),
drivers/rpmsg/qcom_glink_trace.h
350
TP_ARGS(remote, tx),
drivers/rpmsg/qcom_glink_trace.h
353
__field(bool, tx)
drivers/rpmsg/qcom_glink_trace.h
357
__entry->tx = tx;
drivers/rpmsg/qcom_glink_trace.h
360
__entry->tx ? "tx" : "rx",
drivers/rpmsg/qcom_glink_trace.h
368
TP_PROTO(const char *remote, const char *channel, u16 lcid, u16 rcid, unsigned int signals, bool tx),
drivers/rpmsg/qcom_glink_trace.h
369
TP_ARGS(remote, channel, lcid, rcid, signals, tx),
drivers/rpmsg/qcom_glink_trace.h
376
__field(bool, tx)
drivers/rpmsg/qcom_glink_trace.h
384
__entry->tx = tx;
drivers/rpmsg/qcom_glink_trace.h
387
__entry->tx ? "tx" : "rx",
drivers/rpmsg/qcom_glink_trace.h
39
TP_PROTO(const char *remote, unsigned int version, unsigned int features, bool tx),
drivers/rpmsg/qcom_glink_trace.h
40
TP_ARGS(remote, version, features, tx),
drivers/rpmsg/qcom_glink_trace.h
45
__field(bool, tx)
drivers/rpmsg/qcom_glink_trace.h
51
__entry->tx = tx;
drivers/rpmsg/qcom_glink_trace.h
54
__entry->tx ? "tx" : "rx",
drivers/rpmsg/qcom_glink_trace.h
64
TP_PROTO(const char *remote, const char *channel, u16 lcid, u16 rcid, bool tx),
drivers/rpmsg/qcom_glink_trace.h
65
TP_ARGS(remote, channel, lcid, rcid, tx),
drivers/rpmsg/qcom_glink_trace.h
71
__field(bool, tx)
drivers/rpmsg/qcom_glink_trace.h
78
__entry->tx = tx;
drivers/rpmsg/qcom_glink_trace.h
81
__entry->tx ? "tx" : "rx",
drivers/rpmsg/qcom_glink_trace.h
92
TP_PROTO(const char *remote, const char *channel, u16 lcid, u16 rcid, bool tx),
drivers/rpmsg/qcom_glink_trace.h
93
TP_ARGS(remote, channel, lcid, rcid, tx),
drivers/rpmsg/qcom_glink_trace.h
99
__field(bool, tx)
drivers/rpmsg/qcom_smd.c
249
struct smd_channel_info tx;
drivers/rpmsg/qcom_smd.c
271
struct smd_channel_info_word tx;
drivers/rpmsg/qcom_smd.c
311
BUILD_BUG_ON(sizeof(channel->info->tx.param) != sizeof(u8)); \
drivers/rpmsg/qcom_smd.c
313
le32_to_cpu(channel->info_word->tx.param) : \
drivers/rpmsg/qcom_smd.c
314
channel->info->tx.param; \
drivers/rpmsg/qcom_smd.c
319
BUILD_BUG_ON(sizeof(channel->info->tx.param) != sizeof(u32)); \
drivers/rpmsg/qcom_smd.c
321
channel->info_word->tx.param : \
drivers/rpmsg/qcom_smd.c
322
channel->info->tx.param); \
drivers/rpmsg/qcom_smd.c
327
BUILD_BUG_ON(sizeof(channel->info->tx.param) != sizeof(u8)); \
drivers/rpmsg/qcom_smd.c
329
channel->info_word->tx.param = cpu_to_le32(value); \
drivers/rpmsg/qcom_smd.c
331
channel->info->tx.param = value; \
drivers/rpmsg/qcom_smd.c
336
BUILD_BUG_ON(sizeof(channel->info->tx.param) != sizeof(u32)); \
drivers/rpmsg/qcom_smd.c
338
channel->info_word->tx.param = cpu_to_le32(value); \
drivers/rpmsg/qcom_smd.c
340
channel->info->tx.param = cpu_to_le32(value); \
drivers/rtc/rtc-ds1305.c
490
u8 *addr, size_t count, char *tx, char *rx)
drivers/rtc/rtc-ds1305.c
501
x->tx_buf = tx;
drivers/rtc/rtc-mcp795.c
64
u8 tx[2];
drivers/rtc/rtc-mcp795.c
66
tx[0] = MCP795_READ;
drivers/rtc/rtc-mcp795.c
67
tx[1] = addr;
drivers/rtc/rtc-mcp795.c
68
ret = spi_write_then_read(spi, tx, sizeof(tx), buf, count);
drivers/rtc/rtc-mcp795.c
81
u8 tx[257];
drivers/rtc/rtc-mcp795.c
83
tx[0] = MCP795_WRITE;
drivers/rtc/rtc-mcp795.c
84
tx[1] = addr;
drivers/rtc/rtc-mcp795.c
85
memcpy(&tx[2], data, count);
drivers/rtc/rtc-mcp795.c
87
ret = spi_write(spi, tx, 2 + count);
drivers/scsi/cxgbi/libcxgbi.h
240
struct cxgbi_skb_tx_cb tx;
drivers/scsi/cxgbi/libcxgbi.h
251
#define cxgbi_skcb_tx_wr_next(skb) (CXGBI_SKB_CB(skb)->tx.wr_next)
drivers/scsi/cxgbi/libcxgbi.h
252
#define cxgbi_skcb_tx_iscsi_hdrlen(skb) (CXGBI_SKB_CB(skb)->tx.iscsi_hdr_len)
drivers/scsi/cxgbi/libcxgbi.h
253
#define cxgbi_skcb_tx_ulp_mode(skb) (CXGBI_SKB_CB(skb)->tx.ulp_mode)
drivers/scsi/fnic/fnic_main.c
268
stats->tx_frames = vs->tx.tx_unicast_frames_ok;
drivers/scsi/fnic/fnic_main.c
269
stats->tx_words = vs->tx.tx_unicast_bytes_ok / 4;
drivers/scsi/fnic/fnic_main.c
272
stats->error_frames = vs->tx.tx_errors + vs->rx.rx_errors;
drivers/scsi/fnic/fnic_main.c
273
stats->dumped_frames = vs->tx.tx_drops + vs->rx.rx_drop;
drivers/scsi/fnic/vnic_stats.h
52
struct vnic_tx_stats tx;
drivers/scsi/mvsas/mv_chips.h
205
static inline void mvs_start_delivery(struct mvs_info *mvi, u32 tx)
drivers/scsi/mvsas/mv_chips.h
208
mw32(MVS_TX_PROD_IDX, tx);
drivers/scsi/mvsas/mv_init.c
100
if (mvi->tx)
drivers/scsi/mvsas/mv_init.c
102
sizeof(*mvi->tx) * MVS_CHIP_SLOT_SZ,
drivers/scsi/mvsas/mv_init.c
103
mvi->tx, mvi->tx_dma);
drivers/scsi/mvsas/mv_init.c
227
mvi->tx = dma_alloc_coherent(mvi->dev,
drivers/scsi/mvsas/mv_init.c
228
sizeof(*mvi->tx) * MVS_CHIP_SLOT_SZ,
drivers/scsi/mvsas/mv_init.c
230
if (!mvi->tx)
drivers/scsi/mvsas/mv_sas.c
368
slot->tx = mvi->tx_prod;
drivers/scsi/mvsas/mv_sas.c
369
mvi->tx[mvi->tx_prod] = cpu_to_le32((TXQ_CMD_SMP << TXQ_CMD_SHIFT) |
drivers/scsi/mvsas/mv_sas.c
441
slot->tx = mvi->tx_prod;
drivers/scsi/mvsas/mv_sas.c
446
mvi->tx[mvi->tx_prod] = cpu_to_le32(del_q);
drivers/scsi/mvsas/mv_sas.c
569
slot->tx = mvi->tx_prod;
drivers/scsi/mvsas/mv_sas.c
570
mvi->tx[mvi->tx_prod] = cpu_to_le32(TXQ_MODE_I | tag |
drivers/scsi/mvsas/mv_sas.h
121
void (*start_delivery)(struct mvs_info *mvi, u32 tx);
drivers/scsi/mvsas/mv_sas.h
313
u32 tx;
drivers/scsi/mvsas/mv_sas.h
349
__le32 *tx;
drivers/scsi/snic/vnic_stats.h
50
struct vnic_tx_stats tx;
drivers/slimbus/slimbus.h
415
struct slim_msg_txn *tx);
drivers/soc/fsl/qbman/qman_test_stash.c
139
struct qman_fq tx;
drivers/soc/fsl/qbman/qman_test_stash.c
285
if (qman_enqueue(&handler->tx, &dqrr->fd)) {
drivers/soc/fsl/qbman/qman_test_stash.c
302
if (qman_enqueue(&handler->tx, &dqrr->fd)) {
drivers/soc/fsl/qbman/qman_test_stash.c
366
qman_destroy_fq(&handler->tx);
drivers/soc/fsl/qbman/qman_test_stash.c
420
memset(&handler->tx, 0, sizeof(handler->tx));
drivers/soc/fsl/qbman/qman_test_stash.c
422
&handler->tx);
drivers/soc/fsl/qbman/qman_test_stash.c
541
err = qman_enqueue(&special_handler->tx, &fd);
drivers/soc/ti/knav_dma.c
151
if (cfg->u.tx.filt_pswords)
drivers/soc/ti/knav_dma.c
153
if (cfg->u.tx.filt_einfo)
drivers/soc/ti/knav_dma.c
160
writel_relaxed(cfg->u.tx.priority, &chan->reg_tx_sched->prio);
drivers/soc/ti/knav_dma.c
322
chan->cfg.u.tx.filt_einfo,
drivers/soc/ti/knav_dma.c
323
chan->cfg.u.tx.filt_pswords,
drivers/soc/ti/knav_dma.c
324
chan->cfg.u.tx.priority);
drivers/spi/atmel-quadspi.c
778
struct dma_async_tx_descriptor *tx;
drivers/spi/atmel-quadspi.c
782
tx = dmaengine_prep_dma_memcpy(chan, dma_dst, dma_src, len,
drivers/spi/atmel-quadspi.c
784
if (!tx) {
drivers/spi/atmel-quadspi.c
790
tx->callback = atmel_qspi_dma_callback;
drivers/spi/atmel-quadspi.c
791
tx->callback_param = aq;
drivers/spi/atmel-quadspi.c
792
cookie = tx->tx_submit(tx);
drivers/spi/spi-altera-core.c
100
| (hw->tx[hw->count * 2 + 1] << 8));
drivers/spi/spi-altera-core.c
103
txd = (hw->tx[hw->count * 4]
drivers/spi/spi-altera-core.c
104
| (hw->tx[hw->count * 4 + 1] << 8)
drivers/spi/spi-altera-core.c
105
| (hw->tx[hw->count * 4 + 2] << 16)
drivers/spi/spi-altera-core.c
106
| (hw->tx[hw->count * 4 + 3] << 24));
drivers/spi/spi-altera-core.c
148
hw->tx = t->tx_buf;
drivers/spi/spi-altera-core.c
93
if (hw->tx) {
drivers/spi/spi-altera-core.c
96
txd = hw->tx[hw->count];
drivers/spi/spi-altera-core.c
99
txd = (hw->tx[hw->count * 2]
drivers/spi/spi-au1550.c
315
hw->tx = t->tx_buf;
drivers/spi/spi-au1550.c
356
hw->tx = hw->rx;
drivers/spi/spi-au1550.c
365
res = au1xxx_dbdma_put_source(hw->dma_tx_ch, virt_to_phys(hw->tx),
drivers/spi/spi-au1550.c
475
if (hw->tx) { \
drivers/spi/spi-au1550.c
476
fifoword = *(u##size *)hw->tx & (u32)(mask); \
drivers/spi/spi-au1550.c
477
hw->tx += (size) / 8; \
drivers/spi/spi-au1550.c
49
const u8 *tx;
drivers/spi/spi-au1550.c
498
hw->tx = t->tx_buf;
drivers/spi/spi-bcm63xx-hsspi.c
379
const u8 *tx = t->tx_buf;
drivers/spi/spi-bcm63xx-hsspi.c
396
if (tx && rx)
drivers/spi/spi-bcm63xx-hsspi.c
398
else if (tx)
drivers/spi/spi-bcm63xx-hsspi.c
425
if (tx)
drivers/spi/spi-bcm63xx-hsspi.c
426
memcpy_toio(bs->fifo + HSSPI_OPCODE_LEN + bs->prepend_cnt, tx,
drivers/spi/spi-bcm63xx-hsspi.c
500
const u8 *tx = t->tx_buf;
drivers/spi/spi-bcm63xx-hsspi.c
508
if (tx && rx)
drivers/spi/spi-bcm63xx-hsspi.c
510
else if (tx)
drivers/spi/spi-bcm63xx-hsspi.c
535
if (tx) {
drivers/spi/spi-bcm63xx-hsspi.c
536
memcpy_toio(bs->fifo + HSSPI_OPCODE_LEN, tx, curr_step);
drivers/spi/spi-bcm63xx-hsspi.c
537
tx += curr_step;
drivers/spi/spi-bcmbca-hsspi.c
258
const u8 *tx = t->tx_buf;
drivers/spi/spi-bcmbca-hsspi.c
264
if (tx && rx)
drivers/spi/spi-bcmbca-hsspi.c
266
else if (tx)
drivers/spi/spi-bcmbca-hsspi.c
291
if (tx) {
drivers/spi/spi-bcmbca-hsspi.c
292
memcpy_toio(bs->fifo + HSSPI_OPCODE_LEN, tx, curr_step);
drivers/spi/spi-bcmbca-hsspi.c
293
tx += curr_step;
drivers/spi/spi-bitbang.c
122
const u32 *tx = t->tx_buf;
drivers/spi/spi-bitbang.c
129
if (tx)
drivers/spi/spi-bitbang.c
130
word = *tx++;
drivers/spi/spi-bitbang.c
60
const u8 *tx = t->tx_buf;
drivers/spi/spi-bitbang.c
67
if (tx)
drivers/spi/spi-bitbang.c
68
word = *tx++;
drivers/spi/spi-bitbang.c
91
const u16 *tx = t->tx_buf;
drivers/spi/spi-bitbang.c
98
if (tx)
drivers/spi/spi-bitbang.c
99
word = *tx++;
drivers/spi/spi-cadence-quadspi.c
1371
struct dma_async_tx_descriptor *tx;
drivers/spi/spi-cadence-quadspi.c
1387
tx = dmaengine_prep_dma_memcpy(cqspi->rx_chan, dma_dst, dma_src,
drivers/spi/spi-cadence-quadspi.c
1389
if (!tx) {
drivers/spi/spi-cadence-quadspi.c
1395
tx->callback = cqspi_rx_dma_callback;
drivers/spi/spi-cadence-quadspi.c
1396
tx->callback_param = cqspi;
drivers/spi/spi-cadence-quadspi.c
1397
cookie = tx->tx_submit(tx);
drivers/spi/spi-cavium-octeon.c
43
p->regs.tx = 0x10;
drivers/spi/spi-cavium-thunderx.c
49
p->regs.tx = 0x1010;
drivers/spi/spi-cavium.h
13
int tx;
drivers/spi/spi-cavium.h
28
#define OCTEON_SPI_TX(x) (x->regs.tx)
drivers/spi/spi-davinci.c
174
const void *tx;
drivers/spi/spi-davinci.c
216
if (dspi->tx) {
drivers/spi/spi-davinci.c
217
const u8 *tx = dspi->tx;
drivers/spi/spi-davinci.c
219
data = *tx++;
drivers/spi/spi-davinci.c
220
dspi->tx = tx;
drivers/spi/spi-davinci.c
229
if (dspi->tx) {
drivers/spi/spi-davinci.c
230
const u16 *tx = dspi->tx;
drivers/spi/spi-davinci.c
232
data = *tx++;
drivers/spi/spi-davinci.c
233
dspi->tx = tx;
drivers/spi/spi-davinci.c
644
dspi->tx = t->tx_buf;
drivers/spi/spi-dln2.c
108
} tx;
drivers/spi/spi-dln2.c
109
unsigned len = sizeof(tx);
drivers/spi/spi-dln2.c
111
tx.port = dln2->port;
drivers/spi/spi-dln2.c
115
len -= sizeof(tx.wait_for_completion);
drivers/spi/spi-dln2.c
117
tx.wait_for_completion = DLN2_TRANSFERS_WAIT_COMPLETE;
drivers/spi/spi-dln2.c
121
return dln2_transfer_tx(dln2->pdev, cmd, &tx, len);
drivers/spi/spi-dln2.c
137
} tx;
drivers/spi/spi-dln2.c
139
tx.port = dln2->port;
drivers/spi/spi-dln2.c
146
tx.cs = ~cs_mask;
drivers/spi/spi-dln2.c
148
return dln2_transfer_tx(dln2->pdev, DLN2_SPI_SET_SS, &tx, sizeof(tx));
drivers/spi/spi-dln2.c
167
} tx;
drivers/spi/spi-dln2.c
170
tx.port = dln2->port;
drivers/spi/spi-dln2.c
171
tx.cs = cs_mask;
drivers/spi/spi-dln2.c
174
return dln2_transfer_tx(dln2->pdev, cmd, &tx, sizeof(tx));
drivers/spi/spi-dln2.c
189
} tx;
drivers/spi/spi-dln2.c
195
tx.port = dln2->port;
drivers/spi/spi-dln2.c
196
ret = dln2_transfer(dln2->pdev, DLN2_SPI_GET_SS_COUNT, &tx, sizeof(tx),
drivers/spi/spi-dln2.c
215
} tx;
drivers/spi/spi-dln2.c
221
tx.port = dln2->port;
drivers/spi/spi-dln2.c
223
ret = dln2_transfer(dln2->pdev, cmd, &tx, sizeof(tx), &rx, &rx_len);
drivers/spi/spi-dln2.c
265
} __packed tx;
drivers/spi/spi-dln2.c
271
tx.port = dln2->port;
drivers/spi/spi-dln2.c
272
tx.speed = cpu_to_le32(speed);
drivers/spi/spi-dln2.c
274
ret = dln2_transfer(dln2->pdev, DLN2_SPI_SET_FREQUENCY, &tx, sizeof(tx),
drivers/spi/spi-dln2.c
292
} tx;
drivers/spi/spi-dln2.c
294
tx.port = dln2->port;
drivers/spi/spi-dln2.c
295
tx.mode = mode;
drivers/spi/spi-dln2.c
297
return dln2_transfer_tx(dln2->pdev, DLN2_SPI_SET_MODE, &tx, sizeof(tx));
drivers/spi/spi-dln2.c
308
} tx;
drivers/spi/spi-dln2.c
310
tx.port = dln2->port;
drivers/spi/spi-dln2.c
311
tx.bpw = bpw;
drivers/spi/spi-dln2.c
314
&tx, sizeof(tx));
drivers/spi/spi-dln2.c
323
} tx;
drivers/spi/spi-dln2.c
331
tx.port = dln2->port;
drivers/spi/spi-dln2.c
334
&tx, sizeof(tx), rx, &rx_len);
drivers/spi/spi-dln2.c
427
} __packed *tx = dln2->buf;
drivers/spi/spi-dln2.c
430
BUILD_BUG_ON(sizeof(*tx) > DLN2_SPI_BUF_SIZE);
drivers/spi/spi-dln2.c
435
tx->port = dln2->port;
drivers/spi/spi-dln2.c
436
tx->size = cpu_to_le16(data_len);
drivers/spi/spi-dln2.c
437
tx->attr = attr;
drivers/spi/spi-dln2.c
439
dln2_spi_copy_to_buf(tx->buf, data, data_len, dln2->bpw);
drivers/spi/spi-dln2.c
441
tx_len = sizeof(*tx) + data_len - DLN2_SPI_MAX_XFER_SIZE;
drivers/spi/spi-dln2.c
442
return dln2_transfer_tx(dln2->pdev, DLN2_SPI_WRITE, tx, tx_len);
drivers/spi/spi-dln2.c
456
} __packed tx;
drivers/spi/spi-dln2.c
468
tx.port = dln2->port;
drivers/spi/spi-dln2.c
469
tx.size = cpu_to_le16(data_len);
drivers/spi/spi-dln2.c
470
tx.attr = attr;
drivers/spi/spi-dln2.c
472
ret = dln2_transfer(dln2->pdev, DLN2_SPI_READ, &tx, sizeof(tx),
drivers/spi/spi-dln2.c
498
} __packed *tx;
drivers/spi/spi-dln2.c
505
BUILD_BUG_ON(sizeof(*tx) > DLN2_SPI_BUF_SIZE ||
drivers/spi/spi-dln2.c
516
tx = dln2->buf;
drivers/spi/spi-dln2.c
519
tx->port = dln2->port;
drivers/spi/spi-dln2.c
520
tx->size = cpu_to_le16(data_len);
drivers/spi/spi-dln2.c
521
tx->attr = attr;
drivers/spi/spi-dln2.c
523
dln2_spi_copy_to_buf(tx->buf, tx_data, data_len, dln2->bpw);
drivers/spi/spi-dln2.c
525
tx_len = sizeof(*tx) + data_len - DLN2_SPI_MAX_XFER_SIZE;
drivers/spi/spi-dln2.c
528
ret = dln2_transfer(dln2->pdev, DLN2_SPI_READ_WRITE, tx, tx_len,
drivers/spi/spi-dw-core.c
141
if (dws->tx) {
drivers/spi/spi-dw-core.c
143
txw = *(u8 *)(dws->tx);
drivers/spi/spi-dw-core.c
145
txw = *(u16 *)(dws->tx);
drivers/spi/spi-dw-core.c
147
txw = *(u32 *)(dws->tx);
drivers/spi/spi-dw-core.c
149
dws->tx += dws->n_bytes;
drivers/spi/spi-dw-core.c
430
dws->tx = (void *)transfer->tx_buf;
drivers/spi/spi-dw-core.c
546
dws->tx = out;
drivers/spi/spi-dw-core.c
561
if (dws->tx != dws->buf)
drivers/spi/spi-dw-core.c
562
kfree(dws->tx);
drivers/spi/spi-dw-core.c
577
buf = dws->tx;
drivers/spi/spi-dw-core.c
587
len = dws->tx_len - ((void *)buf - dws->tx);
drivers/spi/spi-dw-dma.c
106
dws->dma_addr_widths = tx.dst_addr_widths & rx.src_addr_widths;
drivers/spi/spi-dw-dma.c
113
struct dw_dma_slave dma_tx = { .dst_id = 1 }, *tx = &dma_tx;
drivers/spi/spi-dw-dma.c
137
tx->dma_dev = &dma_dev->dev;
drivers/spi/spi-dw-dma.c
138
dws->txchan = dma_request_channel(mask, dw_spi_dma_chan_filter, tx);
drivers/spi/spi-dw-dma.c
77
struct dma_slave_caps tx, rx;
drivers/spi/spi-dw-dma.c
80
ret = dma_get_slave_caps(dws->txchan, &tx);
drivers/spi/spi-dw-dma.c
88
if (!(tx.directions & BIT(DMA_MEM_TO_DEV) &&
drivers/spi/spi-dw-dma.c
92
if (tx.max_sg_burst > 0 && rx.max_sg_burst > 0)
drivers/spi/spi-dw-dma.c
93
dws->dma_sg_burst = min(tx.max_sg_burst, rx.max_sg_burst);
drivers/spi/spi-dw-dma.c
94
else if (tx.max_sg_burst > 0)
drivers/spi/spi-dw-dma.c
95
dws->dma_sg_burst = tx.max_sg_burst;
drivers/spi/spi-dw.h
172
void *tx;
drivers/spi/spi-ep93xx.c
183
val = ((u16 *)xfer->tx_buf)[espi->tx];
drivers/spi/spi-ep93xx.c
184
espi->tx += 2;
drivers/spi/spi-ep93xx.c
187
val = ((u8 *)xfer->tx_buf)[espi->tx];
drivers/spi/spi-ep93xx.c
188
espi->tx += 1;
drivers/spi/spi-ep93xx.c
234
while (espi->fifo_level < SPI_FIFO_SIZE && espi->tx < xfer->len) {
drivers/spi/spi-ep93xx.c
494
espi->tx = 0;
drivers/spi/spi-ep93xx.c
88
size_t tx;
drivers/spi/spi-fsi.c
226
static int fsi_spi_data_out(u64 *out, const u8 *tx, int len)
drivers/spi/spi-fsi.c
236
out_bytes[8 - (i + 1)] = tx[i];
drivers/spi/spi-fsi.c
310
const u8 *tx = transfer->tx_buf;
drivers/spi/spi-fsi.c
313
nb = fsi_spi_data_out(&out, &tx[sent],
drivers/spi/spi-fsl-cpm.c
130
mspi->tx = dst;
drivers/spi/spi-fsl-cpm.c
135
void *nonconst_tx = (void *)mspi->tx; /* shut up gcc */
drivers/spi/spi-fsl-dspi.c
1203
dspi->tx = transfer->tx_buf;
drivers/spi/spi-fsl-dspi.c
352
const void *tx;
drivers/spi/spi-fsl-dspi.c
390
*txdata = *(u8 *)dspi->tx;
drivers/spi/spi-fsl-dspi.c
393
*txdata = *(u16 *)dspi->tx;
drivers/spi/spi-fsl-dspi.c
396
*txdata = *(u32 *)dspi->tx;
drivers/spi/spi-fsl-dspi.c
399
dspi->tx += dspi->oper_word_size;
drivers/spi/spi-fsl-dspi.c
420
*txdata = (__force u32)cpu_to_be32(*(u32 *)dspi->tx);
drivers/spi/spi-fsl-dspi.c
421
dspi->tx += sizeof(u32);
drivers/spi/spi-fsl-dspi.c
432
*txdata = (__force u32)cpu_to_be16(*(u16 *)dspi->tx);
drivers/spi/spi-fsl-dspi.c
433
dspi->tx += sizeof(u16);
drivers/spi/spi-fsl-dspi.c
444
u16 hi = *(u16 *)dspi->tx;
drivers/spi/spi-fsl-dspi.c
445
u16 lo = *(u16 *)(dspi->tx + 2);
drivers/spi/spi-fsl-dspi.c
448
dspi->tx += sizeof(u32);
drivers/spi/spi-fsl-dspi.c
469
if (dspi->tx)
drivers/spi/spi-fsl-lib.c
43
const type *tx = mpc8xxx_spi->tx; \
drivers/spi/spi-fsl-lib.c
44
if (!tx) \
drivers/spi/spi-fsl-lib.c
46
data = *tx++ << mpc8xxx_spi->tx_shift; \
drivers/spi/spi-fsl-lib.c
47
mpc8xxx_spi->tx = tx; \
drivers/spi/spi-fsl-lib.h
25
const void *tx;
drivers/spi/spi-fsl-lpspi.c
115
void (*tx)(struct fsl_lpspi_data *);
drivers/spi/spi-fsl-lpspi.c
262
fsl_lpspi->tx(fsl_lpspi);
drivers/spi/spi-fsl-lpspi.c
384
struct dma_slave_config rx = {}, tx = {};
drivers/spi/spi-fsl-lpspi.c
402
tx.direction = DMA_MEM_TO_DEV;
drivers/spi/spi-fsl-lpspi.c
403
tx.dst_addr = fsl_lpspi->base_phys + IMX7ULP_TDR;
drivers/spi/spi-fsl-lpspi.c
404
tx.dst_addr_width = buswidth;
drivers/spi/spi-fsl-lpspi.c
405
tx.dst_maxburst = 1;
drivers/spi/spi-fsl-lpspi.c
406
ret = dmaengine_slave_config(controller->dma_tx, &tx);
drivers/spi/spi-fsl-lpspi.c
488
fsl_lpspi->tx = fsl_lpspi_buf_tx_u8;
drivers/spi/spi-fsl-lpspi.c
491
fsl_lpspi->tx = fsl_lpspi_buf_tx_u16;
drivers/spi/spi-fsl-lpspi.c
494
fsl_lpspi->tx = fsl_lpspi_buf_tx_u32;
drivers/spi/spi-fsl-lpspi.c
632
struct sg_table *tx = &transfer->tx_sg, *rx = &transfer->rx_sg;
drivers/spi/spi-fsl-lpspi.c
652
tx->sgl, tx->nents, DMA_MEM_TO_DEV,
drivers/spi/spi-fsl-spi.c
270
mpc8xxx_spi->tx = t->tx_buf;
drivers/spi/spi-geni-qcom.c
102
struct dma_chan *tx;
drivers/spi/spi-geni-qcom.c
228
dmaengine_terminate_sync(mas->tx);
drivers/spi/spi-geni-qcom.c
462
dmaengine_slave_config(mas->tx, &config);
drivers/spi/spi-geni-qcom.c
463
tx_desc = dmaengine_prep_slave_sg(mas->tx, xfer->tx_sg.sgl, xfer->tx_sg.nents,
drivers/spi/spi-geni-qcom.c
480
dma_async_issue_pending(mas->tx);
drivers/spi/spi-geni-qcom.c
554
if (mas->tx) {
drivers/spi/spi-geni-qcom.c
555
dma_release_channel(mas->tx);
drivers/spi/spi-geni-qcom.c
556
mas->tx = NULL;
drivers/spi/spi-geni-qcom.c
564
mas->tx = dma_request_chan(mas->dev, "tx");
drivers/spi/spi-geni-qcom.c
565
if (IS_ERR(mas->tx)) {
drivers/spi/spi-geni-qcom.c
566
ret = dev_err_probe(mas->dev, PTR_ERR(mas->tx),
drivers/spi/spi-geni-qcom.c
588
dma_release_channel(mas->tx);
drivers/spi/spi-geni-qcom.c
590
mas->tx = NULL;
drivers/spi/spi-hisi-kunpeng.c
132
const void *tx;
drivers/spi/spi-hisi-kunpeng.c
255
if (hs->tx) {
drivers/spi/spi-hisi-kunpeng.c
258
txw = *(u8 *)(hs->tx);
drivers/spi/spi-hisi-kunpeng.c
261
txw = *(u16 *)(hs->tx);
drivers/spi/spi-hisi-kunpeng.c
264
txw = *(u32 *)(hs->tx);
drivers/spi/spi-hisi-kunpeng.c
267
hs->tx += hs->n_bytes;
drivers/spi/spi-hisi-kunpeng.c
390
hs->tx = transfer->tx_buf;
drivers/spi/spi-imx.c
126
void (*tx)(struct spi_imx_data *spi_imx);
drivers/spi/spi-imx.c
1278
spi_imx->tx(spi_imx);
drivers/spi/spi-imx.c
1354
spi_imx->tx = spi_imx_buf_tx_swap;
drivers/spi/spi-imx.c
1360
spi_imx->tx = spi_imx_buf_tx_u8;
drivers/spi/spi-imx.c
1363
spi_imx->tx = spi_imx_buf_tx_u16;
drivers/spi/spi-imx.c
1366
spi_imx->tx = spi_imx_buf_tx_u32;
drivers/spi/spi-imx.c
1381
spi_imx->tx = mx53_ecspi_tx_target;
drivers/spi/spi-imx.c
1853
struct dma_slave_config rx = {}, tx = {};
drivers/spi/spi-imx.c
1874
tx.direction = DMA_MEM_TO_DEV;
drivers/spi/spi-imx.c
1875
tx.dst_addr = spi_imx->base_phys + MXC_CSPITXDATA;
drivers/spi/spi-imx.c
1876
tx.dst_addr_width = buswidth;
drivers/spi/spi-imx.c
1877
tx.dst_maxburst = spi_imx->wml;
drivers/spi/spi-imx.c
1878
ret = dmaengine_slave_config(controller->dma_tx, &tx);
drivers/spi/spi-jcore.c
107
const unsigned char *tx;
drivers/spi/spi-jcore.c
115
tx = t->tx_buf;
drivers/spi/spi-jcore.c
123
writel(tx ? *tx++ : 0, data_reg);
drivers/spi/spi-kspi2.c
142
static int kspi2_txrx_byte(struct kspi2 *kspi, u8 tx, u8 *rx)
drivers/spi/spi-kspi2.c
148
iowrite8(tx, kspi->base + KSPI2_DATA_REG);
drivers/spi/spi-kspi2.c
166
u8 tx = 0;
drivers/spi/spi-kspi2.c
173
tx = ((const u8 *)t->tx_buf)[i];
drivers/spi/spi-kspi2.c
175
ret = kspi2_txrx_byte(kspi, tx, &rx);
drivers/spi/spi-lantiq-ssc.c
177
const u8 *tx;
drivers/spi/spi-lantiq-ssc.c
490
tx8 = spi->tx;
drivers/spi/spi-lantiq-ssc.c
493
spi->tx++;
drivers/spi/spi-lantiq-ssc.c
496
tx16 = (u16 *) spi->tx;
drivers/spi/spi-lantiq-ssc.c
499
spi->tx += 2;
drivers/spi/spi-lantiq-ssc.c
502
tx32 = (u32 *) spi->tx;
drivers/spi/spi-lantiq-ssc.c
505
spi->tx += 4;
drivers/spi/spi-lantiq-ssc.c
634
if (spi->tx) {
drivers/spi/spi-lantiq-ssc.c
729
spi->tx = t->tx_buf;
drivers/spi/spi-lantiq-ssc.c
743
if (!spi->tx)
drivers/spi/spi-loongson-core.c
135
const u8 *tx = xfer->tx_buf;
drivers/spi/spi-loongson-core.c
140
ret = loongson_spi_write_read_8bit(spi, &tx, &rx, count);
drivers/spi/spi-loopback-test.c
1007
void *tx, void *rx)
drivers/spi/spi-loopback-test.c
1047
tx, rx,
drivers/spi/spi-loopback-test.c
1072
char *rx = NULL, *tx = NULL;
drivers/spi/spi-loopback-test.c
1088
tx = vmalloc(SPI_TEST_MAX_SIZE_PLUS);
drivers/spi/spi-loopback-test.c
1090
tx = kzalloc(SPI_TEST_MAX_SIZE_PLUS, GFP_KERNEL);
drivers/spi/spi-loopback-test.c
1091
if (!tx) {
drivers/spi/spi-loopback-test.c
1104
ret = test->run_test(spi, test, tx, rx);
drivers/spi/spi-loopback-test.c
1106
ret = spi_test_run_test(spi, test, tx, rx);
drivers/spi/spi-loopback-test.c
1119
kvfree(tx);
drivers/spi/spi-loopback-test.c
604
void *tx, void *rx)
drivers/spi/spi-loopback-test.c
666
void *tx, void *rx)
drivers/spi/spi-loopback-test.c
693
*ptr = tx + off;
drivers/spi/spi-loopback-test.c
793
void *tx, void *rx)
drivers/spi/spi-loopback-test.c
811
(void *)tx, rx);
drivers/spi/spi-loopback-test.c
817
(void *)tx, rx);
drivers/spi/spi-loopback-test.c
832
ret = test->execute_msg(spi, test, tx, rx);
drivers/spi/spi-loopback-test.c
834
ret = spi_test_execute_msg(spi, test, tx, rx);
drivers/spi/spi-loopback-test.c
855
void *tx, void *rx,
drivers/spi/spi-loopback-test.c
920
return _spi_test_run_iter(spi, &test, tx, rx);
drivers/spi/spi-loopback-test.c
934
void *tx, void *rx)
drivers/spi/spi-loopback-test.c
976
ret = spi_test_check_loopback_result(spi, msg, tx, rx);
drivers/spi/spi-lp8841-rtc.c
104
const u8 *tx = t->tx_buf;
drivers/spi/spi-lp8841-rtc.c
109
if (tx) {
drivers/spi/spi-lp8841-rtc.c
113
word = *tx++;
drivers/spi/spi-mem.c
107
static int spi_check_buswidth_req(struct spi_mem *mem, u8 buswidth, bool tx)
drivers/spi/spi-mem.c
116
if ((tx &&
drivers/spi/spi-mem.c
118
(!tx &&
drivers/spi/spi-mem.c
125
if ((tx && (mode & (SPI_TX_QUAD | SPI_TX_OCTAL))) ||
drivers/spi/spi-mem.c
126
(!tx && (mode & (SPI_RX_QUAD | SPI_RX_OCTAL))))
drivers/spi/spi-mem.c
132
if ((tx && (mode & SPI_TX_OCTAL)) ||
drivers/spi/spi-mem.c
133
(!tx && (mode & SPI_RX_OCTAL)))
drivers/spi/spi-omap2-mcspi.c
423
struct dma_async_tx_descriptor *tx;
drivers/spi/spi-omap2-mcspi.c
430
tx = dmaengine_prep_slave_sg(mcspi_dma->dma_tx, xfer->tx_sg.sgl,
drivers/spi/spi-omap2-mcspi.c
434
if (tx) {
drivers/spi/spi-omap2-mcspi.c
435
tx->callback = omap2_mcspi_tx_callback;
drivers/spi/spi-omap2-mcspi.c
436
tx->callback_param = spi;
drivers/spi/spi-omap2-mcspi.c
437
dmaengine_submit(tx);
drivers/spi/spi-omap2-mcspi.c
461
struct dma_async_tx_descriptor *tx;
drivers/spi/spi-omap2-mcspi.c
517
tx = dmaengine_prep_slave_sg(mcspi_dma->dma_rx, sg_out[0],
drivers/spi/spi-omap2-mcspi.c
520
if (tx) {
drivers/spi/spi-omap2-mcspi.c
521
tx->callback = omap2_mcspi_rx_callback;
drivers/spi/spi-omap2-mcspi.c
522
tx->callback_param = spi;
drivers/spi/spi-omap2-mcspi.c
523
dmaengine_submit(tx);
drivers/spi/spi-omap2-mcspi.c
600
const u8 *tx;
drivers/spi/spi-omap2-mcspi.c
633
tx = xfer->tx_buf;
drivers/spi/spi-omap2-mcspi.c
639
if (tx) {
drivers/spi/spi-omap2-mcspi.c
651
if (tx != NULL) {
drivers/spi/spi-omap2-mcspi.c
730
const u8 *tx;
drivers/spi/spi-omap2-mcspi.c
733
tx = xfer->tx_buf;
drivers/spi/spi-omap2-mcspi.c
737
if (tx != NULL) {
drivers/spi/spi-omap2-mcspi.c
744
word_len, *tx);
drivers/spi/spi-omap2-mcspi.c
745
writel_relaxed(*tx++, tx_reg);
drivers/spi/spi-omap2-mcspi.c
754
if (c == 1 && tx == NULL &&
drivers/spi/spi-omap2-mcspi.c
767
} else if (c == 0 && tx == NULL) {
drivers/spi/spi-omap2-mcspi.c
780
const u16 *tx;
drivers/spi/spi-omap2-mcspi.c
783
tx = xfer->tx_buf;
drivers/spi/spi-omap2-mcspi.c
786
if (tx != NULL) {
drivers/spi/spi-omap2-mcspi.c
793
word_len, *tx);
drivers/spi/spi-omap2-mcspi.c
794
writel_relaxed(*tx++, tx_reg);
drivers/spi/spi-omap2-mcspi.c
803
if (c == 2 && tx == NULL &&
drivers/spi/spi-omap2-mcspi.c
816
} else if (c == 0 && tx == NULL) {
drivers/spi/spi-omap2-mcspi.c
829
const u32 *tx;
drivers/spi/spi-omap2-mcspi.c
832
tx = xfer->tx_buf;
drivers/spi/spi-omap2-mcspi.c
835
if (tx != NULL) {
drivers/spi/spi-omap2-mcspi.c
842
word_len, *tx);
drivers/spi/spi-omap2-mcspi.c
843
writel_relaxed(*tx++, tx_reg);
drivers/spi/spi-omap2-mcspi.c
852
if (c == 4 && tx == NULL &&
drivers/spi/spi-omap2-mcspi.c
865
} else if (c == 0 && tx == NULL) {
drivers/spi/spi-orion.c
506
const u8 *tx = xfer->tx_buf;
drivers/spi/spi-orion.c
510
if (orion_spi_write_read_8bit(spi, &tx, &rx) < 0)
drivers/spi/spi-orion.c
516
const u16 *tx = xfer->tx_buf;
drivers/spi/spi-orion.c
520
if (orion_spi_write_read_16bit(spi, &tx, &rx) < 0)
drivers/spi/spi-pic32.c
114
const void *tx;
drivers/spi/spi-pic32.c
165
tx_left = (pic32s->tx_end - pic32s->tx) / n_bytes;
drivers/spi/spi-pic32.c
177
(pic32s->tx_end - pic32s->tx)) / n_bytes;
drivers/spi/spi-pic32.c
209
v = *(__type *)(pic32s->tx); \
drivers/spi/spi-pic32.c
211
pic32s->tx += sizeof(__type); \
drivers/spi/spi-pic32.c
287
if (pic32s->tx_end == pic32s->tx)
drivers/spi/spi-pic32.c
535
pic32s->tx = (const void *)transfer->tx_buf;
drivers/spi/spi-pic32.c
537
pic32s->tx_end = pic32s->tx + transfer->len;
drivers/spi/spi-pl022.c
1178
if (pl022->tx == pl022->tx_end) {
drivers/spi/spi-pl022.c
1227
pl022->tx = (void *)transfer->tx_buf;
drivers/spi/spi-pl022.c
1228
pl022->tx_end = pl022->tx + pl022->cur_transfer->len;
drivers/spi/spi-pl022.c
1232
pl022->tx ? pl022->cur_chip->write : WRITING_NULL;
drivers/spi/spi-pl022.c
1311
while (pl022->tx < pl022->tx_end || pl022->rx < pl022->rx_end) {
drivers/spi/spi-pl022.c
371
void *tx;
drivers/spi/spi-pl022.c
598
__func__, pl022->rx, pl022->rx_end, pl022->tx, pl022->tx_end);
drivers/spi/spi-pl022.c
627
&& (pl022->tx < pl022->tx_end)) {
drivers/spi/spi-pl022.c
633
writew(*(u8 *) (pl022->tx), SSP_DR(pl022->virtbase));
drivers/spi/spi-pl022.c
636
writew((*(u16 *) (pl022->tx)), SSP_DR(pl022->virtbase));
drivers/spi/spi-pl022.c
639
writel(*(u32 *) (pl022->tx), SSP_DR(pl022->virtbase));
drivers/spi/spi-pl022.c
642
pl022->tx += (pl022->cur_chip->n_bytes);
drivers/spi/spi-pl022.c
924
setup_dma_scatter(pl022, pl022->tx,
drivers/spi/spi-ppc4xx.c
125
const unsigned char *tx;
drivers/spi/spi-ppc4xx.c
148
hw->tx = t->tx_buf;
drivers/spi/spi-ppc4xx.c
154
data = hw->tx ? hw->tx[0] : 0;
drivers/spi/spi-ppc4xx.c
305
data = hw->tx ? hw->tx[count] : 0;
drivers/spi/spi-pxa2xx-pci.c
157
tx = c->tx_param;
drivers/spi/spi-pxa2xx-pci.c
158
tx->dma_dev = &dma_dev->dev;
drivers/spi/spi-pxa2xx-pci.c
159
tx->m_master = 0;
drivers/spi/spi-pxa2xx-pci.c
160
tx->p_master = 1;
drivers/spi/spi-pxa2xx-pci.c
195
struct dw_dma_slave *tx, *rx;
drivers/spi/spi-pxa2xx-pci.c
233
tx = c->tx_param;
drivers/spi/spi-pxa2xx-pci.c
234
tx->dma_dev = &dma_dev->dev;
drivers/spi/spi-pxa2xx-pci.c
99
struct dw_dma_slave *tx, *rx;
drivers/spi/spi-pxa2xx.c
1067
drv_data->tx += tx_level;
drivers/spi/spi-pxa2xx.c
486
|| (drv_data->tx == drv_data->tx_end))
drivers/spi/spi-pxa2xx.c
490
drv_data->tx += n_bytes;
drivers/spi/spi-pxa2xx.c
510
|| (drv_data->tx == drv_data->tx_end))
drivers/spi/spi-pxa2xx.c
513
pxa2xx_spi_write(drv_data, SSDR, *(u8 *)(drv_data->tx));
drivers/spi/spi-pxa2xx.c
514
++drv_data->tx;
drivers/spi/spi-pxa2xx.c
532
|| (drv_data->tx == drv_data->tx_end))
drivers/spi/spi-pxa2xx.c
535
pxa2xx_spi_write(drv_data, SSDR, *(u16 *)(drv_data->tx));
drivers/spi/spi-pxa2xx.c
536
drv_data->tx += 2;
drivers/spi/spi-pxa2xx.c
554
|| (drv_data->tx == drv_data->tx_end))
drivers/spi/spi-pxa2xx.c
557
pxa2xx_spi_write(drv_data, SSDR, *(u32 *)(drv_data->tx));
drivers/spi/spi-pxa2xx.c
558
drv_data->tx += 4;
drivers/spi/spi-pxa2xx.c
669
if (drv_data->tx == drv_data->tx_end) {
drivers/spi/spi-pxa2xx.c
963
drv_data->tx = (void *)transfer->tx_buf;
drivers/spi/spi-pxa2xx.c
964
drv_data->tx_end = drv_data->tx + transfer->len;
drivers/spi/spi-pxa2xx.c
977
drv_data->write = drv_data->tx ? u8_writer : null_writer;
drivers/spi/spi-pxa2xx.c
981
drv_data->write = drv_data->tx ? u16_writer : null_writer;
drivers/spi/spi-pxa2xx.c
985
drv_data->write = drv_data->tx ? u32_writer : null_writer;
drivers/spi/spi-pxa2xx.h
64
void *tx;
drivers/spi/spi-rockchip.c
179
const void *tx;
drivers/spi/spi-rockchip.c
311
txw = *(u8 *)rs->tx;
drivers/spi/spi-rockchip.c
313
txw = *(u16 *)rs->tx;
drivers/spi/spi-rockchip.c
316
rs->tx += rs->n_bytes;
drivers/spi/spi-rockchip.c
386
rs->tx = xfer->tx_buf;
drivers/spi/spi-rockchip.c
388
rs->tx_left = rs->tx ? xfer->len / rs->n_bytes : 0;
drivers/spi/spi-rockchip.c
459
rs->tx = xfer->tx_buf;
drivers/spi/spi-rspi.c
510
static int rspi_pio_transfer(struct rspi_data *rspi, const u8 *tx, u8 *rx,
drivers/spi/spi-rspi.c
514
if (tx) {
drivers/spi/spi-rspi.c
515
int ret = rspi_data_out(rspi, *tx++);
drivers/spi/spi-rspi.c
538
static int rspi_dma_transfer(struct rspi_data *rspi, struct sg_table *tx,
drivers/spi/spi-rspi.c
568
if (tx) {
drivers/spi/spi-rspi.c
569
desc_tx = dmaengine_prep_slave_sg(rspi->ctlr->dma_tx, tx->sgl,
drivers/spi/spi-rspi.c
570
tx->nents, DMA_MEM_TO_DEV,
drivers/spi/spi-rspi.c
597
if (tx)
drivers/spi/spi-rspi.c
608
if (tx)
drivers/spi/spi-rspi.c
615
if (tx)
drivers/spi/spi-rspi.c
624
if (tx)
drivers/spi/spi-rspi.c
632
if (tx)
drivers/spi/spi-rspi.c
755
static int qspi_trigger_transfer_out_in(struct rspi_data *rspi, const u8 *tx,
drivers/spi/spi-rspi.c
770
rspi_write_data(rspi, *tx++);
drivers/spi/spi-rspi.c
803
const u8 *tx = xfer->tx_buf;
drivers/spi/spi-rspi.c
822
rspi_write_data(rspi, *tx++);
drivers/spi/spi-sh-hspi.c
148
u32 tx;
drivers/spi/spi-sh-hspi.c
174
tx = 0;
drivers/spi/spi-sh-hspi.c
176
tx = (u32)((u8 *)t->tx_buf)[i];
drivers/spi/spi-sh-hspi.c
178
hspi_write(hspi, SPTBR, tx);
drivers/spi/spi-sh-msiof.c
646
static int sh_msiof_dma_once(struct sh_msiof_spi_priv *p, const void *tx,
drivers/spi/spi-sh-msiof.c
672
if (tx) {
drivers/spi/spi-sh-msiof.c
701
sh_msiof_spi_set_mode_regs(p, tx, rx, 32, words1, words2);
drivers/spi/spi-sh-msiof.c
706
if (tx)
drivers/spi/spi-sh-msiof.c
713
if (tx)
drivers/spi/spi-sh-msiof.c
722
if (tx) {
drivers/spi/spi-sh-msiof.c
763
if (tx)
drivers/spi/spi-stm32.c
1312
dma_conf->dst_addr = spi->phys_addr + spi->cfg->regs->tx.reg;
drivers/spi/spi-stm32.c
246
const struct stm32_spi_reg tx;
drivers/spi/spi-stm32.c
399
.tx = { STM32FX_SPI_DR },
drivers/spi/spi-stm32.c
419
.tx = { STM32H7_SPI_TXDR },
drivers/spi/spi-stm32.c
439
.tx = { STM32H7_SPI_TXDR },
drivers/spi/spi-test.h
109
void *tx, void *rx);
drivers/spi/spi-test.h
114
void *tx, void *rx);
drivers/spi/spi-test.h
78
void *tx, void *rx);
drivers/spi/spi-test.h
80
void *tx, void *rx);
drivers/spi/spi-ti-qspi.c
445
struct dma_async_tx_descriptor *tx;
drivers/spi/spi-ti-qspi.c
449
tx = dmaengine_prep_dma_memcpy(chan, dma_dst, dma_src, len, flags);
drivers/spi/spi-ti-qspi.c
450
if (!tx) {
drivers/spi/spi-ti-qspi.c
455
tx->callback = ti_qspi_dma_callback;
drivers/spi/spi-ti-qspi.c
456
tx->callback_param = qspi;
drivers/spi/spi-ti-qspi.c
457
cookie = tx->tx_submit(tx);
drivers/staging/media/atomisp/pci/sh_css_param_shading.c
144
int tx, src_x0, src_x1;
drivers/staging/media/atomisp/pci/sh_css_param_shading.c
149
tx = out_start_col + j * out_cell_size;
drivers/staging/media/atomisp/pci/sh_css_param_shading.c
151
src_x0 = tx / (int)in_cell_size;
drivers/staging/media/atomisp/pci/sh_css_param_shading.c
153
src_x1 = (tx + out_cell_size) /
drivers/staging/media/atomisp/pci/sh_css_param_shading.c
161
tx = min(clamp(tx, 0, (int)sensor_width - 1),
drivers/staging/media/atomisp/pci/sh_css_param_shading.c
173
dx0 = tx - sx0;
drivers/staging/media/atomisp/pci/sh_css_param_shading.c
174
dx1 = sx1 - tx;
drivers/staging/media/sunxi/cedrus/cedrus_h265.c
368
int i, x, tx, y, ty;
drivers/staging/media/sunxi/cedrus/cedrus_h265.c
375
for (x = 0, tx = 0; tx < pps->num_tile_columns_minus1 + 1; tx++) {
drivers/staging/media/sunxi/cedrus/cedrus_h265.c
376
if (x + pps->column_width_minus1[tx] + 1 > ctb_addr_x)
drivers/staging/media/sunxi/cedrus/cedrus_h265.c
379
x += pps->column_width_minus1[tx] + 1;
drivers/staging/media/sunxi/cedrus/cedrus_h265.c
392
((x + pps->column_width_minus1[tx]) << 0));
drivers/staging/media/sunxi/cedrus/cedrus_h265.c
399
if (tx + 1 >= pps->num_tile_columns_minus1 + 1) {
drivers/staging/media/sunxi/cedrus/cedrus_h265.c
401
tx = 0;
drivers/staging/media/sunxi/cedrus/cedrus_h265.c
404
x += pps->column_width_minus1[tx++] + 1;
drivers/staging/media/sunxi/cedrus/cedrus_h265.c
412
((x + pps->column_width_minus1[tx]) << 0);
drivers/staging/most/net/net.c
199
if (most_start_channel(nd->iface, nd->tx.ch_id, &comp)) {
drivers/staging/most/net/net.c
213
nd->iface->request_netinfo(nd->iface, nd->tx.ch_id, on_netinfo);
drivers/staging/most/net/net.c
226
nd->iface->request_netinfo(nd->iface, nd->tx.ch_id, NULL);
drivers/staging/most/net/net.c
228
most_stop_channel(nd->iface, nd->tx.ch_id, &comp);
drivers/staging/most/net/net.c
240
mbo = most_get_mbo(nd->iface, nd->tx.ch_id, &comp);
drivers/staging/most/net/net.c
297
if (nd && nd->rx.linked && nd->tx.linked)
drivers/staging/most/net/net.c
339
ch = ccfg->direction == MOST_CH_TX ? &nd->tx : &nd->rx;
drivers/staging/most/net/net.c
341
ch = ccfg->direction == MOST_CH_TX ? &nd->tx : &nd->rx;
drivers/staging/most/net/net.c
379
} else if (nd->tx.linked && channel_idx == nd->tx.ch_id) {
drivers/staging/most/net/net.c
380
ch = &nd->tx;
drivers/staging/most/net/net.c
386
if (nd->rx.linked && nd->tx.linked) {
drivers/staging/most/net/net.c
418
if (nd->tx.ch_id != channel_idx)
drivers/staging/most/net/net.c
66
struct net_dev_channel tx;
drivers/staging/nvec/nvec.c
479
if (nvec->tx->pos != nvec->tx->size) {
drivers/staging/nvec/nvec.c
481
nvec->tx->pos = 0;
drivers/staging/nvec/nvec.c
560
nvec->tx = &nvec->tx_scratch;
drivers/staging/nvec/nvec.c
561
list_add_tail(&nvec->tx->node, &nvec->tx_data);
drivers/staging/nvec/nvec.c
563
nvec->tx = list_first_entry(&nvec->tx_data, struct nvec_msg,
drivers/staging/nvec/nvec.c
565
nvec->tx->pos = 0;
drivers/staging/nvec/nvec.c
570
(uint)nvec->tx->size, nvec->tx->data[1]);
drivers/staging/nvec/nvec.c
660
to_send = nvec->tx->data[0];
drivers/staging/nvec/nvec.c
661
nvec->tx->pos = 1;
drivers/staging/nvec/nvec.c
678
} else if (nvec->tx && nvec->tx->pos < nvec->tx->size) {
drivers/staging/nvec/nvec.c
679
to_send = nvec->tx->data[nvec->tx->pos++];
drivers/staging/nvec/nvec.c
683
nvec->tx,
drivers/staging/nvec/nvec.c
684
(uint)(nvec->tx ? nvec->tx->pos : 0),
drivers/staging/nvec/nvec.c
685
(uint)(nvec->tx ? nvec->tx->size : 0));
drivers/staging/nvec/nvec.h
149
struct nvec_msg *tx;
drivers/staging/rtl8723bs/core/rtw_ap.c
1479
void update_beacon(struct adapter *padapter, u8 ie_id, u8 *oui, u8 tx)
drivers/staging/rtl8723bs/core/rtw_ap.c
1548
if (tx) {
drivers/staging/rtl8723bs/core/rtw_pwrctrl.c
180
void traffic_check_for_leave_lps(struct adapter *padapter, u8 tx, u32 tx_packets)
drivers/staging/rtl8723bs/core/rtw_pwrctrl.c
189
if (tx) { /* from tx */
drivers/staging/rtl8723bs/core/rtw_pwrctrl.c
219
rtw_lps_ctrl_wk_cmd(padapter, LPS_CTRL_LEAVE, tx?0:1);
drivers/staging/rtl8723bs/include/rtw_ap.h
13
void update_beacon(struct adapter *padapter, u8 ie_id, u8 *oui, u8 tx);
drivers/staging/rtl8723bs/include/rtw_pwrctrl.h
239
void traffic_check_for_leave_lps(struct adapter *padapter, u8 tx, u32 tx_packets);
drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c
151
.tx = 0xffff,
drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c
155
.tx = 0xffff,
drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c
160
.tx = 0xffff,
drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c
171
.tx = 0xffff,
drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c
181
.tx = 0xffff,
drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c
186
.tx = 0xffff,
drivers/target/iscsi/cxgbit/cxgbit.h
101
#define cxgbit_skcb_tx_extralen(skb) (CXGBIT_SKB_CB(skb)->tx.extra_len)
drivers/target/iscsi/cxgbit/cxgbit.h
85
struct cxgbit_skb_tx_cb tx;
drivers/target/iscsi/cxgbit/cxgbit.h
99
#define cxgbit_skcb_submode(skb) (CXGBIT_SKB_CB(skb)->tx.submode)
drivers/thermal/tegra/tegra-bpmp-thermal.c
154
msg.tx.data = &req;
drivers/thermal/tegra/tegra-bpmp-thermal.c
155
msg.tx.size = sizeof(req);
drivers/thermal/tegra/tegra-bpmp-thermal.c
183
msg.tx.data = &req;
drivers/thermal/tegra/tegra-bpmp-thermal.c
184
msg.tx.size = sizeof(req);
drivers/thermal/tegra/tegra-bpmp-thermal.c
47
msg.tx.data = &req;
drivers/thermal/tegra/tegra-bpmp-thermal.c
48
msg.tx.size = sizeof(req);
drivers/thermal/tegra/tegra-bpmp-thermal.c
88
msg.tx.data = &req;
drivers/thermal/tegra/tegra-bpmp-thermal.c
89
msg.tx.size = sizeof(req);
drivers/thunderbolt/ctl.c
393
res = tb_ring_tx(ctl->tx, &pkg->frame);
drivers/thunderbolt/ctl.c
41
struct tb_ring *tx;
drivers/thunderbolt/ctl.c
674
ctl->tx = tb_ring_alloc_tx(nhi, 0, 10, RING_FLAG_NO_SUSPEND);
drivers/thunderbolt/ctl.c
675
if (!ctl->tx)
drivers/thunderbolt/ctl.c
714
if (ctl->tx)
drivers/thunderbolt/ctl.c
715
tb_ring_free(ctl->tx);
drivers/thunderbolt/ctl.c
734
tb_ring_start(ctl->tx); /* is used to ack hotplug packets, start first */
drivers/thunderbolt/ctl.c
758
tb_ring_stop(ctl->tx);
drivers/tty/amiserial.c
1091
icount->tx = cnow.tx;
drivers/tty/amiserial.c
1393
seq_printf(m, " tx:%d rx:%d", state->icount.tx, state->icount.rx);
drivers/tty/amiserial.c
270
info->icount.tx++;
drivers/tty/amiserial.c
286
info->icount.tx++;
drivers/tty/ipwireless/hardware.c
1105
int tx = 0;
drivers/tty/ipwireless/hardware.c
1183
tx = 1;
drivers/tty/ipwireless/hardware.c
1185
if (tx)
drivers/tty/ipwireless/hardware.c
1191
if (tx || rx)
drivers/tty/mxser.c
1228
icount->tx = cnow.tx;
drivers/tty/mxser.c
1525
port->icount.tx++;
drivers/tty/mxser.c
1543
port->icount.tx++;
drivers/tty/nozomi.c
1701
icount->tx = cnow.tx;
drivers/tty/serial/8250/8250_pci1xxxx.c
495
port->icount.tx++;
drivers/tty/serial/8250/8250_pci1xxxx.c
512
port->icount.tx++;
drivers/tty/serial/amba-pl011.c
1498
uap->port.icount.tx++;
drivers/tty/serial/amba-pl011.c
777
uap->port.icount.tx++;
drivers/tty/serial/ar933x_uart.c
419
up->port.icount.tx++;
drivers/tty/serial/arc_uart.c
164
port->icount.tx++;
drivers/tty/serial/clps711x.c
154
port->icount.tx++;
drivers/tty/serial/cpm_uart.c
671
port->icount.tx++;
drivers/tty/serial/digicolor-usart.c
193
port->icount.tx++;
drivers/tty/serial/dz.c
265
dport->port.icount.tx++;
drivers/tty/serial/fsl_linflexuart.c
182
sport->icount.tx++;
drivers/tty/serial/fsl_lpuart.c
760
sport->port.icount.tx++;
drivers/tty/serial/icom.c
962
icom_port->uart_port.icount.tx += count;
drivers/tty/serial/imx.c
543
sport->port.icount.tx++;
drivers/tty/serial/ip22zilog.c
394
up->port.icount.tx++;
drivers/tty/serial/ip22zilog.c
595
port->icount.tx++;
drivers/tty/serial/max3100.c
157
static int max3100_sr(struct max3100_port *s, u16 tx, u16 *rx)
drivers/tty/serial/max3100.c
168
etx = cpu_to_be16(tx);
drivers/tty/serial/max3100.c
178
dev_dbg(&s->spi->dev, "%04x - %04x\n", tx, *rx);
drivers/tty/serial/max3100.c
241
u16 tx, rx;
drivers/tty/serial/max3100.c
270
tx = 0xffff;
drivers/tty/serial/max3100.c
272
tx = s->port.x_char;
drivers/tty/serial/max3100.c
273
s->port.icount.tx++;
drivers/tty/serial/max3100.c
277
tx = ch;
drivers/tty/serial/max3100.c
279
if (tx != 0xffff) {
drivers/tty/serial/max3100.c
280
max3100_calc_parity(s, &tx);
drivers/tty/serial/max3100.c
281
tx |= MAX3100_WD | (s->rts ? MAX3100_RTS : 0);
drivers/tty/serial/max3100.c
282
max3100_sr(s, tx, &rx);
drivers/tty/serial/max310x.c
756
port->icount.tx++;
drivers/tty/serial/meson_uart.c
156
port->icount.tx++;
drivers/tty/serial/milbeaut_usio.c
105
port->icount.tx++;
drivers/tty/serial/milbeaut_usio.c
86
port->icount.tx++;
drivers/tty/serial/mps2-uart.c
301
MAKE_NAME(-tx), mps_port);
drivers/tty/serial/msm_serial.c
911
port->icount.tx++;
drivers/tty/serial/pch_uart.c
827
port->icount.tx++;
drivers/tty/serial/pch_uart.c
884
port->icount.tx++;
drivers/tty/serial/pic32_uart.c
349
port->icount.tx++;
drivers/tty/serial/pmac_zilog.c
381
uap->port.icount.tx++;
drivers/tty/serial/pmac_zilog.c
606
port->icount.tx++;
drivers/tty/serial/rda-uart.c
346
port->icount.tx++;
drivers/tty/serial/rsci.c
465
port->icount.tx++;
drivers/tty/serial/samsung_tty.c
887
port->icount.tx++;
drivers/tty/serial/sb1250-duart.c
393
sport->port.icount.tx++;
drivers/tty/serial/sc16is7xx.c
660
port->icount.tx++;
drivers/tty/serial/sccnxp.c
447
port->icount.tx++;
drivers/tty/serial/serial_core.c
1281
icount->tx = cnow.tx;
drivers/tty/serial/serial_core.c
2009
uport->icount.tx, uport->icount.rx);
drivers/tty/serial/serial_core.c
701
uport->icount.tx++;
drivers/tty/serial/sh-sci.c
1052
port->icount.tx++;
drivers/tty/serial/sprd_serial.c
348
port->icount.tx++;
drivers/tty/serial/stm32-usart.c
810
port->icount.tx++;
drivers/tty/serial/sunplus-uart.c
207
port->icount.tx++;
drivers/tty/serial/sunsu.c
381
up->port.icount.tx++;
drivers/tty/serial/sunzilog.c
493
up->port.icount.tx++;
drivers/tty/serial/sunzilog.c
700
port->icount.tx++;
drivers/tty/serial/tegra-tcu.c
192
tcu->tx = mbox_request_channel_byname(&tcu->tx_client, "tx");
drivers/tty/serial/tegra-tcu.c
193
if (IS_ERR(tcu->tx)) {
drivers/tty/serial/tegra-tcu.c
194
err = PTR_ERR(tcu->tx);
drivers/tty/serial/tegra-tcu.c
266
mbox_free_channel(tcu->tx);
drivers/tty/serial/tegra-tcu.c
281
mbox_free_channel(tcu->tx);
drivers/tty/serial/tegra-tcu.c
30
struct mbox_chan *tx, *rx;
drivers/tty/serial/tegra-tcu.c
58
mbox_send_message(tcu->tx, msg);
drivers/tty/serial/tegra-tcu.c
59
mbox_flush(tcu->tx, 1000);
drivers/tty/serial/timbuart.c
107
port->icount.tx,
drivers/tty/serial/uartlite.c
201
port->icount.tx++;
drivers/tty/serial/ucc_uart.c
356
port->icount.tx++;
drivers/tty/serial/zs.c
615
zport->port.icount.tx++;
drivers/tty/synclink_gt.c
1049
icount->tx = cnow.tx;
drivers/tty/synclink_gt.c
1212
info->icount.tx, info->icount.rx);
drivers/usb/atm/usbatm.c
603
atomic_inc(&vcc->stats->tx);
drivers/usb/atm/usbatm.c
747
atomic_read(&atm_dev->stats.aal5.tx),
drivers/usb/class/cdc-wdm.c
926
.tx = wdm_wwan_port_tx,
drivers/usb/gadget/function/u_serial.c
263
port->icount.tx += len;
drivers/usb/gadget/function/u_serial.c
906
icount->tx = cnow.tx;
drivers/usb/host/max3421-hcd.c
140
struct max3421_dma_buf *tx;
drivers/usb/host/max3421-hcd.c
1881
max3421_hcd->tx = kmalloc_obj(*max3421_hcd->tx);
drivers/usb/host/max3421-hcd.c
1882
if (!max3421_hcd->tx)
drivers/usb/host/max3421-hcd.c
1917
kfree(max3421_hcd->tx);
drivers/usb/host/max3421-hcd.c
352
max3421_hcd->tx->data[0] =
drivers/usb/host/max3421-hcd.c
356
transfer.tx_buf = max3421_hcd->tx->data;
drivers/usb/host/max3421-hcd.c
378
max3421_hcd->tx->data[0] =
drivers/usb/host/max3421-hcd.c
381
max3421_hcd->tx->data[1] = val;
drivers/usb/host/max3421-hcd.c
383
transfer.tx_buf = max3421_hcd->tx->data;
drivers/usb/host/max3421-hcd.c
402
max3421_hcd->tx->data[0] =
drivers/usb/host/max3421-hcd.c
405
transfer[0].tx_buf = max3421_hcd->tx->data;
drivers/usb/host/max3421-hcd.c
428
max3421_hcd->tx->data[0] =
drivers/usb/host/max3421-hcd.c
432
transfer[0].tx_buf = max3421_hcd->tx->data;
drivers/usb/musb/cppi_dma.h
116
struct cppi_channel tx[4];
drivers/usb/musb/musb_gadget.c
101
request->tx
drivers/usb/musb/musb_gadget.c
109
request->tx
drivers/usb/musb/musb_gadget.c
1170
if (req->tx) {
drivers/usb/musb/musb_gadget.c
1229
request->tx = musb_ep->is_in;
drivers/usb/musb/musb_gadget.c
63
request->tx
drivers/usb/musb/musb_gadget.c
76
request->tx
drivers/usb/musb/musb_gadget.h
58
u8 tx; /* endpoint direction */
drivers/usb/musb/musb_gadget_ep0.c
923
req->tx = ep->is_in;
drivers/usb/musb/musb_trace.h
252
__entry->is_tx = req->tx;
drivers/usb/musb/tusb6010_omap.c
112
chdat->epnum, chdat->tx ? "tx" : "rx");
drivers/usb/musb/tusb6010_omap.c
114
if (chdat->tx)
drivers/usb/musb/tusb6010_omap.c
124
chdat->tx ? "tx" : "rx", remaining);
drivers/usb/musb/tusb6010_omap.c
139
if (chdat->tx) {
drivers/usb/musb/tusb6010_omap.c
158
musb_dma_completion(musb, chdat->epnum, chdat->tx);
drivers/usb/musb/tusb6010_omap.c
168
if (chdat->tx) {
drivers/usb/musb/tusb6010_omap.c
219
if (chdat->tx)
drivers/usb/musb/tusb6010_omap.c
227
chdat->tx ? "tx" : "rx", dma_remaining);
drivers/usb/musb/tusb6010_omap.c
260
if (chdat->tx) {
drivers/usb/musb/tusb6010_omap.c
287
chdat->epnum, chdat->tx ? "tx" : "rx", &dma_addr,
drivers/usb/musb/tusb6010_omap.c
317
chdat->epnum, chdat->tx ? "tx" : "rx",
drivers/usb/musb/tusb6010_omap.c
327
if (chdat->tx) {
drivers/usb/musb/tusb6010_omap.c
34
u8 tx;
drivers/usb/musb/tusb6010_omap.c
344
if (chdat->tx) {
drivers/usb/musb/tusb6010_omap.c
396
if (chdat->tx)
drivers/usb/musb/tusb6010_omap.c
424
u8 tx)
drivers/usb/musb/tusb6010_omap.c
438
dev_dbg(musb->controller, "Not allowing DMA for ep0 %s\n", tx ? "tx" : "rx");
drivers/usb/musb/tusb6010_omap.c
461
if (tx)
drivers/usb/musb/tusb6010_omap.c
462
chdat->tx = 1;
drivers/usb/musb/tusb6010_omap.c
464
chdat->tx = 0;
drivers/usb/musb/tusb6010_omap.c
484
chdat->tx ? "tx" : "rx",
drivers/usb/musb/tusb6010_omap.c
71
if (chdat->tx)
drivers/usb/serial/ftdi_sio.c
2387
port->icount.tx += c;
drivers/usb/serial/ftdi_sio.c
2395
port->icount.tx += count;
drivers/usb/serial/generic.c
561
icount->tx = cnow.tx;
drivers/usb/serial/io_edgeport.c
1299
edge_port->port->icount.tx += count;
drivers/usb/serial/io_edgeport.c
1311
edge_port->port->icount.tx -= count;
drivers/usb/serial/io_ti.c
2062
edge_port->port->icount.tx += count;
drivers/usb/serial/mos7840.c
911
port->icount.tx += transfer_size;
drivers/usb/serial/mos7840.c
912
dev_dbg(&port->dev, "icount.tx is %d:\n", port->icount.tx);
drivers/usb/serial/ti_usb_3410_5052.c
1323
port->icount.tx += count;
drivers/usb/serial/xr_serial.c
24
u16 tx;
drivers/usb/serial/xr_serial.c
574
tx_mask = xr21v141x_txrx_clk_masks[idx].tx;
drivers/video/fbdev/mmp/hw/mmp_ctrl.h
388
#define CFG_TXBITS(tx) (((tx) - 1)<<8) /* 0x1F~0x1 */
drivers/video/fbdev/mmp/hw/mmp_ctrl.h
396
#define CFG_TXBITSTO0(tx) ((tx)<<4)
drivers/video/fbdev/pxa168fb.h
179
#define CFG_TXBITS(tx) ((tx) << 8) /* 0x1F~0x1 */
drivers/video/fbdev/pxa168fb.h
187
#define CFG_TXBITSTO0(tx) ((tx) << 4)
fs/xfs/scrub/attr_repair.c
1342
struct xrep_tempexch *tx)
fs/xfs/scrub/attr_repair.c
1373
return xrep_tempexch_contents(sc, tx);
fs/xfs/scrub/attr_repair.c
1389
return xrep_tempexch_trans_alloc(sc, XFS_ATTR_FORK, &rx->tx);
fs/xfs/scrub/attr_repair.c
1402
error = xrep_tempexch_trans_alloc(sc, XFS_ATTR_FORK, &rx->tx);
fs/xfs/scrub/attr_repair.c
1481
error = xrep_xattr_swap(sc, &rx->tx);
fs/xfs/scrub/attr_repair.c
89
struct xrep_tempexch tx;
fs/xfs/scrub/attr_repair.h
11
int xrep_xattr_swap(struct xfs_scrub *sc, struct xrep_tempexch *tx);
fs/xfs/scrub/dir_repair.c
128
struct xrep_tempexch tx;
fs/xfs/scrub/dir_repair.c
1609
return xrep_tempexch_trans_alloc(sc, XFS_DATA_FORK, &rd->tx);
fs/xfs/scrub/dir_repair.c
1622
error = xrep_tempexch_trans_alloc(sc, XFS_DATA_FORK, &rd->tx);
fs/xfs/scrub/dir_repair.c
1671
rd->pscan.parent_ino, rd->tx.req.resblks);
fs/xfs/scrub/dir_repair.c
1716
return xrep_tempexch_contents(sc, &rd->tx);
fs/xfs/scrub/parent_repair.c
1238
error = xrep_tempexch_trans_alloc(sc, XFS_ATTR_FORK, &rp->tx);
fs/xfs/scrub/parent_repair.c
130
struct xrep_tempexch tx;
fs/xfs/scrub/parent_repair.c
1308
error = xrep_xattr_swap(sc, &rp->tx);
fs/xfs/scrub/symlink_repair.c
329
struct xrep_tempexch *tx = sc->buf;
fs/xfs/scrub/symlink_repair.c
352
return xrep_tempexch_contents(sc, tx);
fs/xfs/scrub/symlink_repair.c
390
struct xrep_tempexch *tx;
fs/xfs/scrub/symlink_repair.c
462
tx = sc->buf;
fs/xfs/scrub/symlink_repair.c
463
error = xrep_tempexch_trans_alloc(sc, XFS_DATA_FORK, tx);
fs/xfs/scrub/tempfile.c
611
struct xrep_tempexch *tx)
fs/xfs/scrub/tempfile.c
613
struct xfs_exchmaps_req *req = &tx->req;
fs/xfs/scrub/tempfile.c
615
memset(tx, 0, sizeof(struct xrep_tempexch));
fs/xfs/scrub/tempfile.c
659
struct xrep_tempexch *tx)
fs/xfs/scrub/tempfile.c
661
struct xfs_exchmaps_req *req = &tx->req;
fs/xfs/scrub/tempfile.c
743
const struct xrep_tempexch *tx)
fs/xfs/scrub/tempfile.c
746
const struct xfs_exchmaps_req *req = &tx->req;
fs/xfs/scrub/tempfile.c
804
struct xrep_tempexch *tx)
fs/xfs/scrub/tempfile.c
812
error = xrep_tempexch_prep_request(sc, whichfork, off, len, tx);
fs/xfs/scrub/tempfile.c
816
error = xfs_exchmaps_estimate(&tx->req);
fs/xfs/scrub/tempfile.c
820
error = xfs_trans_reserve_more(sc->tp, tx->req.resblks, 0);
fs/xfs/scrub/tempfile.c
824
return xrep_tempexch_reserve_quota(sc, tx);
fs/xfs/scrub/tempfile.c
842
struct xrep_tempexch *tx)
fs/xfs/scrub/tempfile.c
851
tx);
fs/xfs/scrub/tempfile.c
855
error = xrep_tempexch_estimate(sc, tx);
fs/xfs/scrub/tempfile.c
863
tx->req.resblks, 0, flags, &sc->tp);
fs/xfs/scrub/tempfile.c
871
return xrep_tempexch_reserve_quota(sc, tx);
fs/xfs/scrub/tempfile.c
882
struct xrep_tempexch *tx)
fs/xfs/scrub/tempfile.c
888
xfs_exchange_mappings(sc->tp, &tx->req);
fs/xfs/scrub/tempfile.c
897
if (tx->req.flags & XFS_EXCHMAPS_SET_SIZES) {
include/acpi/processor.h
200
int tx; /* throttle level */
include/linux/async_tx.h
112
static inline void async_tx_issue_pending(struct dma_async_tx_descriptor *tx)
include/linux/async_tx.h
147
struct dma_async_tx_descriptor *tx,
include/linux/async_tx.h
152
args->depend_tx = tx;
include/linux/async_tx.h
158
void async_tx_submit(struct dma_chan *chan, struct dma_async_tx_descriptor *tx,
include/linux/async_tx.h
202
void async_tx_quiesce(struct dma_async_tx_descriptor **tx);
include/linux/async_tx.h
88
static inline void async_tx_issue_pending(struct dma_async_tx_descriptor *tx)
include/linux/async_tx.h
90
if (likely(tx)) {
include/linux/async_tx.h
91
struct dma_chan *chan = tx->chan;
include/linux/dmaengine.h
1398
void dma_async_tx_descriptor_init(struct dma_async_tx_descriptor *tx,
include/linux/dmaengine.h
1401
static inline void async_tx_ack(struct dma_async_tx_descriptor *tx)
include/linux/dmaengine.h
1403
tx->flags |= DMA_CTRL_ACK;
include/linux/dmaengine.h
1406
static inline void async_tx_clear_ack(struct dma_async_tx_descriptor *tx)
include/linux/dmaengine.h
1408
tx->flags &= ~DMA_CTRL_ACK;
include/linux/dmaengine.h
1411
static inline bool async_tx_test_ack(struct dma_async_tx_descriptor *tx)
include/linux/dmaengine.h
1413
return (tx->flags & DMA_CTRL_ACK) == DMA_CTRL_ACK;
include/linux/dmaengine.h
1416
#define dma_cap_set(tx, mask) __dma_cap_set((tx), &(mask))
include/linux/dmaengine.h
1423
#define dma_cap_clear(tx, mask) __dma_cap_clear((tx), &(mask))
include/linux/dmaengine.h
1436
#define dma_has_cap(tx, mask) __dma_has_cap((tx), &(mask))
include/linux/dmaengine.h
1519
enum dma_status dma_wait_for_async_tx(struct dma_async_tx_descriptor *tx);
include/linux/dmaengine.h
1540
static inline enum dma_status dma_wait_for_async_tx(struct dma_async_tx_descriptor *tx)
include/linux/dmaengine.h
1580
static inline int dmaengine_desc_set_reuse(struct dma_async_tx_descriptor *tx)
include/linux/dmaengine.h
1585
ret = dma_get_slave_caps(tx->chan, &caps);
include/linux/dmaengine.h
1592
tx->flags |= DMA_CTRL_REUSE;
include/linux/dmaengine.h
1596
static inline void dmaengine_desc_clear_reuse(struct dma_async_tx_descriptor *tx)
include/linux/dmaengine.h
1598
tx->flags &= ~DMA_CTRL_REUSE;
include/linux/dmaengine.h
1601
static inline bool dmaengine_desc_test_reuse(struct dma_async_tx_descriptor *tx)
include/linux/dmaengine.h
1603
return (tx->flags & DMA_CTRL_REUSE) == DMA_CTRL_REUSE;
include/linux/dmaengine.h
1625
void dma_run_dependencies(struct dma_async_tx_descriptor *tx);
include/linux/dmaengine.h
619
dma_cookie_t (*tx_submit)(struct dma_async_tx_descriptor *tx);
include/linux/dmaengine.h
620
int (*desc_free)(struct dma_async_tx_descriptor *tx);
include/linux/dmaengine.h
635
static inline void dma_set_unmap(struct dma_async_tx_descriptor *tx,
include/linux/dmaengine.h
639
tx->unmap = unmap;
include/linux/dmaengine.h
646
static inline void dma_set_unmap(struct dma_async_tx_descriptor *tx,
include/linux/dmaengine.h
660
static inline void dma_descriptor_unmap(struct dma_async_tx_descriptor *tx)
include/linux/dmaengine.h
662
if (!tx->unmap)
include/linux/dmaengine.h
665
dmaengine_unmap_put(tx->unmap);
include/linux/dmaengine.h
666
tx->unmap = NULL;
include/linux/iio/imu/adis.h
158
u8 tx[10] __aligned(IIO_DMA_MINALIGN);
include/linux/linkmode.h
83
void linkmode_set_pause(unsigned long *advertisement, bool tx, bool rx);
include/linux/mfd/ipaq-micro.h
114
struct ipaq_micro_txdev tx; /* transmit ISR state */
include/linux/moxtet.h
43
u8 tx[TURRIS_MOX_MAX_MODULES];
include/linux/peci.h
106
} rx, tx;
include/linux/phy.h
2384
void phy_set_sym_pause(struct phy_device *phydev, bool rx, bool tx,
include/linux/phy.h
2386
void phy_set_asym_pause(struct phy_device *phydev, bool rx, bool tx);
include/linux/posix-clock.h
42
int (*clock_adjtime)(struct posix_clock *pc, struct __kernel_timex *tx);
include/linux/serial.h
41
__u32 cts, dsr, rng, dcd, tx, rx;
include/linux/serial_core.h
419
__u32 tx;
include/linux/serial_core.h
859
up->icount.tx += chars;
include/linux/serial_core.h
868
up->icount.tx += chars;
include/linux/serial_core.h
880
up->icount.tx += chars;
include/linux/serial_core.h
923
for (; (for_test) && (tx_ready); (for_post), __port->icount.tx++) { \
include/linux/soc/ti/knav_dma.h
125
struct knav_dma_tx_cfg tx;
include/linux/spi/altera.h
40
const unsigned char *tx;
include/linux/syscalls.h
901
struct __kernel_timex __user *tx);
include/linux/syscalls.h
903
struct old_timex32 __user *tx);
include/linux/wwan.h
67
int (*tx)(struct wwan_port *port, struct sk_buff *skb);
include/net/cfg80211.h
5492
u16 tx, rx;
include/net/mac80211.h
4563
void (*tx)(struct ieee80211_hw *hw,
include/net/mac80211.h
4771
int (*set_ringparam)(struct ieee80211_hw *hw, u32 tx, u32 rx);
include/net/mac80211.h
4773
u32 *tx, u32 *tx_max, u32 *rx, u32 *rx_max);
include/net/netdev_queues.h
109
struct netdev_queue_stats_tx *tx);
include/net/psp/functions.h
107
if (!pas->tx.spi ||
include/net/psp/functions.h
50
if (pas && pas->tx.spi)
include/net/psp/types.h
142
struct psp_key_parsed tx;
include/net/tcp.h
1096
} tx; /* only used for outgoing skbs */
include/net/tls.h
242
struct cipher_context tx;
include/net/xdp_sock.h
65
struct xsk_queue *tx ____cacheline_aligned_in_smp;
include/soc/fsl/qe/immap_qe.h
203
u8 tx[0x400];
include/soc/tegra/bpmp.h
117
} tx;
include/soc/tegra/ivc.h
23
} rx, tx;
include/soc/tegra/ivc.h
97
dma_addr_t rx_phys, const struct iosys_map *tx, dma_addr_t tx_phys,
include/trace/events/afs.h
1668
TP_PROTO(struct afs_server *server, bool tx, struct afs_endpoint_state *estate,
include/trace/events/afs.h
1671
TP_ARGS(server, tx, estate, addr_index, error, abort_code, rtt_us),
include/trace/events/afs.h
1676
__field(bool, tx)
include/trace/events/afs.h
1688
__entry->tx = tx;
include/trace/events/afs.h
1698
__entry->server, __entry->tx ? "tx" : "rx", __entry->estate,
include/trace/events/afs.h
1704
TP_PROTO(struct afs_vlserver *server, bool tx, struct afs_addr_list *alist,
include/trace/events/afs.h
1707
TP_ARGS(server, tx, alist, addr_index, error, abort_code, rtt_us),
include/trace/events/afs.h
1711
__field(bool, tx)
include/trace/events/afs.h
1722
__entry->tx = tx;
include/trace/events/afs.h
1732
__entry->server, __entry->tx ? "tx" : "rx", __entry->addr_index,
include/uapi/linux/atm_eni.h
15
int tx,rx; /* values are in percent and must be > 100 */
include/uapi/linux/atmdev.h
35
__HANDLE_ITEM(tx); /* TX okay */ \
include/uapi/linux/if_xdp.h
68
struct xdp_ring_offset tx;
include/uapi/linux/serial.h
105
int rx, tx;
include/uapi/linux/synclink.h
214
__u32 cts, dsr, rng, dcd, tx, rx;
kernel/time/posix-clock.c
224
static int pc_clock_adjtime(clockid_t id, struct __kernel_timex *tx)
kernel/time/posix-clock.c
233
if (tx->modes && (cd.fp->f_mode & FMODE_WRITE) == 0) {
kernel/time/posix-clock.c
239
err = cd.clk->ops.clock_adjtime(cd.clk, tx);
kernel/time/posix-timers.h
20
int (*clock_adj)(const clockid_t which_clock, struct __kernel_timex *tx);
net/atm/proc.c
48
atomic_read(&stats->tx), atomic_read(&stats->tx_err),
net/can/isotp.c
1000
(size > so->tx.ll_dl - SF_PCI_SZ4 - ae - off)) {
net/can/isotp.c
1005
err = memcpy_from_msg(so->tx.buf, msg, size);
net/can/isotp.c
1032
so->tx.len = size;
net/can/isotp.c
1033
so->tx.idx = 0;
net/can/isotp.c
1043
if (size <= so->tx.ll_dl - SF_PCI_SZ4 - ae - off) {
net/can/isotp.c
1088
so->tx.state = ISOTP_WAIT_FIRST_FC;
net/can/isotp.c
1123
err = wait_event_interruptible(so->wait, so->tx.state == ISOTP_IDLE);
net/can/isotp.c
1136
so->tx.state = ISOTP_IDLE;
net/can/isotp.c
1141
so->tx.state = ISOTP_IDLE;
net/can/isotp.c
1204
while (wait_event_interruptible(so->wait, so->tx.state == ISOTP_IDLE) == 0 &&
net/can/isotp.c
1205
cmpxchg(&so->tx.state, ISOTP_IDLE, ISOTP_SHUTDOWN) != ISOTP_IDLE)
net/can/isotp.c
1209
so->tx.state = ISOTP_SHUTDOWN;
net/can/isotp.c
1472
so->tx.ll_dl = ll.tx_dl;
net/can/isotp.c
1630
if (so->tx.buf != so->tx.sbuf)
net/can/isotp.c
1631
kfree(so->tx.buf);
net/can/isotp.c
1656
so->tx.ll_dl = so->ll.tx_dl;
net/can/isotp.c
1659
so->tx.state = ISOTP_IDLE;
net/can/isotp.c
1662
so->tx.buf = so->tx.sbuf;
net/can/isotp.c
1664
so->tx.buflen = ARRAY_SIZE(so->tx.sbuf);
net/can/isotp.c
167
struct tpcon rx, tx;
net/can/isotp.c
1693
if ((mask & EPOLLWRNORM) && (so->tx.state != ISOTP_IDLE))
net/can/isotp.c
372
if (so->tx.state != ISOTP_WAIT_FC &&
net/can/isotp.c
373
so->tx.state != ISOTP_WAIT_FIRST_FC)
net/can/isotp.c
386
so->tx.state = ISOTP_IDLE;
net/can/isotp.c
392
if (so->tx.state == ISOTP_WAIT_FIRST_FC ||
net/can/isotp.c
416
so->tx.state = ISOTP_WAIT_FC;
net/can/isotp.c
421
so->tx.bs = 0;
net/can/isotp.c
422
so->tx.state = ISOTP_SENDING;
net/can/isotp.c
444
so->tx.state = ISOTP_IDLE;
net/can/isotp.c
680
if ((so->tx.state != ISOTP_IDLE && n_pci_type != N_PCI_FC) ||
net/can/isotp.c
740
int space = so->tx.ll_dl - pcilen;
net/can/isotp.c
741
int num = min_t(int, so->tx.len - so->tx.idx, space);
net/can/isotp.c
761
cf->data[pcilen + i] = so->tx.buf[so->tx.idx++];
net/can/isotp.c
803
cf->data[ae] = N_PCI_CF | so->tx.sn++;
net/can/isotp.c
804
so->tx.sn %= 16;
net/can/isotp.c
805
so->tx.bs++;
net/can/isotp.c
837
cf->len = so->tx.ll_dl;
net/can/isotp.c
842
if (so->tx.len > MAX_12BIT_PDU_SIZE) {
net/can/isotp.c
846
cf->data[ae + 2] = (u8)(so->tx.len >> 24) & 0xFFU;
net/can/isotp.c
847
cf->data[ae + 3] = (u8)(so->tx.len >> 16) & 0xFFU;
net/can/isotp.c
848
cf->data[ae + 4] = (u8)(so->tx.len >> 8) & 0xFFU;
net/can/isotp.c
849
cf->data[ae + 5] = (u8)so->tx.len & 0xFFU;
net/can/isotp.c
853
cf->data[ae] = (u8)(so->tx.len >> 8) | N_PCI_FF;
net/can/isotp.c
854
cf->data[ae + 1] = (u8)so->tx.len & 0xFFU;
net/can/isotp.c
859
for (i = ae + ff_pci_sz; i < so->tx.ll_dl; i++)
net/can/isotp.c
860
cf->data[i] = so->tx.buf[so->tx.idx++];
net/can/isotp.c
862
so->tx.sn = 1;
net/can/isotp.c
881
if (so->tx.idx >= so->tx.len) {
net/can/isotp.c
883
so->tx.state = ISOTP_IDLE;
net/can/isotp.c
888
if (so->txfc.bs && so->tx.bs >= so->txfc.bs) {
net/can/isotp.c
890
so->tx.state = ISOTP_WAIT_FC;
net/can/isotp.c
916
if (so->tx.state == ISOTP_IDLE || so->tx.state == ISOTP_SHUTDOWN)
net/can/isotp.c
927
so->tx.state = ISOTP_IDLE;
net/can/isotp.c
943
if (so->tx.state == ISOTP_SENDING && !so->cfecho)
net/can/isotp.c
963
if (!so->bound || so->tx.state == ISOTP_SHUTDOWN)
net/can/isotp.c
966
while (cmpxchg(&so->tx.state, ISOTP_IDLE, ISOTP_SENDING) != ISOTP_IDLE) {
net/can/isotp.c
971
if (so->tx.state == ISOTP_SHUTDOWN)
net/can/isotp.c
975
err = wait_event_interruptible(so->wait, so->tx.state == ISOTP_IDLE);
net/can/isotp.c
981
if (size > so->tx.buflen && so->tx.buflen < max_pdu_size) {
net/can/isotp.c
985
so->tx.buf = newbuf;
net/can/isotp.c
986
so->tx.buflen = max_pdu_size;
net/can/isotp.c
990
if (!size || size > so->tx.buflen) {
net/can/isotp.c
996
off = (so->tx.ll_dl > CAN_MAX_DLEN) ? 1 : 0;
net/can/j1939/j1939-priv.h
280
unsigned int tx;
net/can/j1939/transport.c
1468
session->pkt.tx = session->pkt.tx_acked;
net/can/j1939/transport.c
1671
session->pkt.tx = 0;
net/can/j1939/transport.c
741
session->pkt.tx = 0;
net/can/j1939/transport.c
790
session->pkt.tx = session->pkt.tx_acked;
net/can/j1939/transport.c
808
se_skb = j1939_session_skb_get_by_offset(session, session->pkt.tx * 7);
net/can/j1939/transport.c
822
while (session->pkt.tx < pkt_end) {
net/can/j1939/transport.c
823
dat[0] = session->pkt.tx - session->pkt.dpo + 1;
net/can/j1939/transport.c
824
offset = (session->pkt.tx * 7) - se_skcb->offset;
net/can/j1939/transport.c
833
se_skb->len , session->pkt.tx);
net/can/j1939/transport.c
856
session->pkt.tx++;
net/can/j1939/transport.c
860
if (session->pkt.tx < session->pkt.total && pdelay) {
net/core/dev.c
11229
struct netdev_queue *tx;
net/core/dev.c
11230
size_t sz = count * sizeof(*tx);
net/core/dev.c
11235
tx = kvzalloc(sz, GFP_KERNEL_ACCOUNT | __GFP_RETRY_MAYFAIL);
net/core/dev.c
11236
if (!tx)
net/core/dev.c
11239
dev->_tx = tx;
net/core/netdev-genl.c
620
netdev_nl_stats_write_tx(struct sk_buff *rsp, struct netdev_queue_stats_tx *tx)
net/core/netdev-genl.c
622
if (netdev_stat_put(rsp, NETDEV_A_QSTATS_TX_PACKETS, tx->packets) ||
net/core/netdev-genl.c
623
netdev_stat_put(rsp, NETDEV_A_QSTATS_TX_BYTES, tx->bytes) ||
net/core/netdev-genl.c
624
netdev_stat_put(rsp, NETDEV_A_QSTATS_TX_HW_DROPS, tx->hw_drops) ||
net/core/netdev-genl.c
625
netdev_stat_put(rsp, NETDEV_A_QSTATS_TX_HW_DROP_ERRORS, tx->hw_drop_errors) ||
net/core/netdev-genl.c
626
netdev_stat_put(rsp, NETDEV_A_QSTATS_TX_CSUM_NONE, tx->csum_none) ||
net/core/netdev-genl.c
627
netdev_stat_put(rsp, NETDEV_A_QSTATS_TX_NEEDS_CSUM, tx->needs_csum) ||
net/core/netdev-genl.c
628
netdev_stat_put(rsp, NETDEV_A_QSTATS_TX_HW_GSO_PACKETS, tx->hw_gso_packets) ||
net/core/netdev-genl.c
629
netdev_stat_put(rsp, NETDEV_A_QSTATS_TX_HW_GSO_BYTES, tx->hw_gso_bytes) ||
net/core/netdev-genl.c
630
netdev_stat_put(rsp, NETDEV_A_QSTATS_TX_HW_GSO_WIRE_PACKETS, tx->hw_gso_wire_packets) ||
net/core/netdev-genl.c
631
netdev_stat_put(rsp, NETDEV_A_QSTATS_TX_HW_GSO_WIRE_BYTES, tx->hw_gso_wire_bytes) ||
net/core/netdev-genl.c
632
netdev_stat_put(rsp, NETDEV_A_QSTATS_TX_HW_DROP_RATELIMITS, tx->hw_drop_ratelimits) ||
net/core/netdev-genl.c
633
netdev_stat_put(rsp, NETDEV_A_QSTATS_TX_STOP, tx->stop) ||
net/core/netdev-genl.c
634
netdev_stat_put(rsp, NETDEV_A_QSTATS_TX_WAKE, tx->wake))
net/core/netdev-genl.c
645
struct netdev_queue_stats_tx tx;
net/core/netdev-genl.c
666
memset(&tx, 0xff, sizeof(tx));
net/core/netdev-genl.c
667
ops->get_queue_stats_tx(netdev, i, &tx);
net/core/netdev-genl.c
668
if (!memchr_inv(&tx, 0xff, sizeof(tx)))
net/core/netdev-genl.c
670
if (netdev_nl_stats_write_tx(rsp, &tx))
net/core/netdev-genl.c
743
struct netdev_queue_stats_tx tx;
net/core/netdev-genl.c
755
memset(&tx, 0xff, sizeof(tx));
net/core/netdev-genl.c
757
ops->get_queue_stats_tx(netdev, i, &tx);
net/core/netdev-genl.c
758
netdev_nl_stats_add(tx_sum, &tx, sizeof(tx));
net/ipv4/tcp_input.c
1815
if (!scb->tx.delivered_mstamp)
net/ipv4/tcp_input.c
1822
rs->prior_delivered_ce = scb->tx.delivered_ce;
net/ipv4/tcp_input.c
1823
rs->prior_delivered = scb->tx.delivered;
net/ipv4/tcp_input.c
1824
rs->prior_mstamp = scb->tx.delivered_mstamp;
net/ipv4/tcp_input.c
1825
rs->is_app_limited = scb->tx.is_app_limited;
net/ipv4/tcp_input.c
1833
scb->tx.first_tx_mstamp);
net/ipv4/tcp_input.c
1841
scb->tx.delivered_mstamp = 0;
net/ipv4/tcp_input.c
1912
if (unlikely(TCP_SKB_CB(prev)->tx.delivered_mstamp))
net/ipv4/tcp_input.c
1913
TCP_SKB_CB(prev)->tx.delivered_mstamp = 0;
net/ipv4/tcp_output.c
1490
TCP_SKB_CB(skb)->tx.first_tx_mstamp = tp->first_tx_mstamp;
net/ipv4/tcp_output.c
1491
TCP_SKB_CB(skb)->tx.delivered_mstamp = tp->delivered_mstamp;
net/ipv4/tcp_output.c
1492
TCP_SKB_CB(skb)->tx.delivered = tp->delivered;
net/ipv4/tcp_output.c
1493
TCP_SKB_CB(skb)->tx.delivered_ce = tp->delivered_ce;
net/ipv4/tcp_output.c
1494
TCP_SKB_CB(skb)->tx.is_app_limited = tp->app_limited ? 1 : 0;
net/ipv4/tcp_output.c
1891
TCP_SKB_CB(buff)->tx = TCP_SKB_CB(skb)->tx;
net/mac80211/agg-rx.c
277
u16 buf_size, bool tx, bool auto_seq,
net/mac80211/agg-rx.c
465
if (tx)
net/mac80211/agg-rx.c
59
u16 initiator, u16 reason, bool tx)
net/mac80211/agg-rx.c
95
if (initiator == WLAN_BACK_RECIPIENT && tx)
net/mac80211/debugfs_sta.c
370
bool start, tx;
net/mac80211/debugfs_sta.c
380
tx = true;
net/mac80211/debugfs_sta.c
382
tx = false;
net/mac80211/debugfs_sta.c
391
if (!tx)
net/mac80211/debugfs_sta.c
404
if (!buf || !tx || !start)
net/mac80211/debugfs_sta.c
412
if (tx) {
net/mac80211/driver-ops.h
38
local->ops->tx(&local->hw, control, skb);
net/mac80211/driver-ops.h
833
u32 tx, u32 rx)
net/mac80211/driver-ops.h
840
trace_drv_set_ringparam(local, tx, rx);
net/mac80211/driver-ops.h
842
ret = local->ops->set_ringparam(&local->hw, tx, rx);
net/mac80211/driver-ops.h
849
u32 *tx, u32 *tx_max, u32 *rx, u32 *rx_max)
net/mac80211/driver-ops.h
854
trace_drv_get_ringparam(local, tx, tx_max, rx, rx_max);
net/mac80211/driver-ops.h
856
local->ops->get_ringparam(&local->hw, tx, tx_max, rx, rx_max);
net/mac80211/ieee80211_i.h
1928
void ieee80211_tx_set_protected(struct ieee80211_tx_data *tx);
net/mac80211/ieee80211_i.h
1981
u8 reason, bool tx);
net/mac80211/ieee80211_i.h
2206
u16 buf_size, bool tx, bool auto_seq,
net/mac80211/key.h
75
struct tkip_ctx tx;
net/mac80211/main.c
672
.tx = 0xffff,
net/mac80211/main.c
679
.tx = 0xffff,
net/mac80211/main.c
697
.tx = 0xffff,
net/mac80211/main.c
708
.tx = 0xffff,
net/mac80211/main.c
718
.tx = 0xffff,
net/mac80211/main.c
723
.tx = 0xffff,
net/mac80211/main.c
733
.tx = 0xffff,
net/mac80211/main.c
739
.tx = 0xffff,
net/mac80211/main.c
749
.tx = 0xffff,
net/mac80211/main.c
800
if (WARN_ON(!ops->tx || !ops->start || !ops->stop || !ops->config ||
net/mac80211/mesh.h
298
bool tx, bool acked);
net/mac80211/mesh_ps.c
544
bool tx, bool acked)
net/mac80211/mesh_ps.c
549
if (tx) {
net/mac80211/mlme.c
10147
bool tx = !req->local_state_change;
net/mac80211/mlme.c
10163
req->reason_code, tx,
net/mac80211/mlme.c
10184
req->reason_code, tx,
net/mac80211/mlme.c
10202
req->reason_code, tx, frame_buf);
net/mac80211/mlme.c
4118
u16 stype, u16 reason, bool tx,
net/mac80211/mlme.c
4140
if (WARN_ON_ONCE(tx && !frame_buf))
net/mac80211/mlme.c
4150
if (tx) {
net/mac80211/mlme.c
4172
tx = tx_link_found;
net/mac80211/mlme.c
4207
if (tx) {
net/mac80211/mlme.c
4623
const u8 *buf, size_t len, bool tx,
net/mac80211/mlme.c
4628
.u.mlme.data = tx ? DEAUTH_TX_EVENT : DEAUTH_RX_EVENT,
net/mac80211/mlme.c
4632
if (tx)
net/mac80211/mlme.c
8395
u8 reason, bool tx)
net/mac80211/mlme.c
8400
tx, frame_buf);
net/mac80211/tkip.c
155
struct tkip_ctx *ctx = &key->u.tkip.tx;
net/mac80211/tkip.c
176
struct tkip_ctx *ctx = &key->u.tkip.tx;
net/mac80211/tkip.c
202
struct tkip_ctx *ctx = &key->u.tkip.tx;
net/mac80211/trace.h
1431
TP_PROTO(struct ieee80211_local *local, u32 tx, u32 rx),
net/mac80211/trace.h
1433
TP_ARGS(local, tx, rx),
net/mac80211/trace.h
1437
__field(u32, tx)
net/mac80211/trace.h
1443
__entry->tx = tx;
net/mac80211/trace.h
1449
LOCAL_PR_ARG, __entry->tx, __entry->rx
net/mac80211/trace.h
1454
TP_PROTO(struct ieee80211_local *local, u32 *tx, u32 *tx_max,
net/mac80211/trace.h
1457
TP_ARGS(local, tx, tx_max, rx, rx_max),
net/mac80211/trace.h
1461
__field(u32, tx)
net/mac80211/trace.h
1469
__entry->tx = *tx;
net/mac80211/trace.h
1478
__entry->tx, __entry->tx_max, __entry->rx, __entry->rx_max
net/mac80211/tx.c
1002
skb_queue_walk(&tx->skbs, skb) {
net/mac80211/tx.c
1008
if (!skb_queue_is_last(&tx->skbs, skb)) {
net/mac80211/tx.c
1030
ieee80211_tx_h_stats(struct ieee80211_tx_data *tx)
net/mac80211/tx.c
1035
if (!tx->sta)
net/mac80211/tx.c
1038
skb_queue_walk(&tx->skbs, skb) {
net/mac80211/tx.c
1040
tx->sta->deflink.tx_stats.bytes[ac] += skb->len;
net/mac80211/tx.c
1043
tx->sta->deflink.tx_stats.packets[ac]++;
net/mac80211/tx.c
1049
ieee80211_tx_h_encrypt(struct ieee80211_tx_data *tx)
net/mac80211/tx.c
1051
if (!tx->key)
net/mac80211/tx.c
1054
switch (tx->key->conf.cipher) {
net/mac80211/tx.c
1057
return ieee80211_crypto_wep_encrypt(tx);
net/mac80211/tx.c
1059
return ieee80211_crypto_tkip_encrypt(tx);
net/mac80211/tx.c
1062
tx, IEEE80211_CCMP_MIC_LEN);
net/mac80211/tx.c
1065
tx, IEEE80211_CCMP_256_MIC_LEN);
net/mac80211/tx.c
1068
tx, IEEE80211_CMAC_128_MIC_LEN);
net/mac80211/tx.c
1071
tx, IEEE80211_CMAC_256_MIC_LEN);
net/mac80211/tx.c
1074
return ieee80211_crypto_aes_gmac_encrypt(tx);
net/mac80211/tx.c
1077
return ieee80211_crypto_gcmp_encrypt(tx);
net/mac80211/tx.c
1084
ieee80211_tx_h_calculate_duration(struct ieee80211_tx_data *tx)
net/mac80211/tx.c
1091
skb_queue_walk(&tx->skbs, skb) {
net/mac80211/tx.c
1095
if (!skb_queue_is_last(&tx->skbs, skb)) {
net/mac80211/tx.c
1096
struct sk_buff *next = skb_queue_next(&tx->skbs, skb);
net/mac80211/tx.c
1103
ieee80211_duration(tx, skb, group_addr, next_len);
net/mac80211/tx.c
1111
static bool ieee80211_tx_prep_agg(struct ieee80211_tx_data *tx,
net/mac80211/tx.c
1128
} else if (!tx->sta->sta.txq[tid]) {
net/mac80211/tx.c
1129
spin_lock(&tx->sta->lock);
net/mac80211/tx.c
1148
tid_tx = rcu_dereference_protected_tid_tx(tx->sta, tid);
net/mac80211/tx.c
1157
clear_sta_flag(tx->sta, WLAN_STA_SP);
net/mac80211/tx.c
1158
ps_dbg(tx->sta->sdata,
net/mac80211/tx.c
1160
tx->sta->sta.addr, tx->sta->sta.aid);
net/mac80211/tx.c
1162
info->control.vif = &tx->sdata->vif;
net/mac80211/tx.c
1169
spin_unlock(&tx->sta->lock);
net/mac80211/tx.c
1172
ieee80211_free_txskb(&tx->local->hw, purge_skb);
net/mac80211/tx.c
1212
struct ieee80211_tx_data *tx,
net/mac80211/tx.c
1221
memset(tx, 0, sizeof(*tx));
net/mac80211/tx.c
1222
tx->skb = skb;
net/mac80211/tx.c
1223
tx->local = local;
net/mac80211/tx.c
1224
tx->sdata = sdata;
net/mac80211/tx.c
1225
__skb_queue_head_init(&tx->skbs);
net/mac80211/tx.c
1238
tx->sta = sta;
net/mac80211/tx.c
1241
tx->sta = rcu_dereference(sdata->u.vlan.sta);
net/mac80211/tx.c
1242
if (!tx->sta && sdata->wdev.use_4addr)
net/mac80211/tx.c
1244
} else if (tx->sdata->control_port_protocol == tx->skb->protocol) {
net/mac80211/tx.c
1245
tx->sta = sta_info_get_bss(sdata, hdr->addr1);
net/mac80211/tx.c
1247
if (!tx->sta && !is_multicast_ether_addr(hdr->addr1)) {
net/mac80211/tx.c
1248
tx->sta = sta_info_get(sdata, hdr->addr1);
net/mac80211/tx.c
1253
if (tx->sta && ieee80211_is_data_qos(hdr->frame_control) &&
net/mac80211/tx.c
1260
tid_tx = rcu_dereference(tx->sta->ampdu_mlme.tid_tx[tid]);
net/mac80211/tx.c
1262
ieee80211_aggr_check(sdata, tx->sta, skb);
net/mac80211/tx.c
1263
tid_tx = rcu_dereference(tx->sta->ampdu_mlme.tid_tx[tid]);
net/mac80211/tx.c
1269
queued = ieee80211_tx_prep_agg(tx, skb, info,
net/mac80211/tx.c
1278
tx->flags &= ~IEEE80211_TX_UNICAST;
net/mac80211/tx.c
1281
tx->flags |= IEEE80211_TX_UNICAST;
net/mac80211/tx.c
1284
if (!(tx->flags & IEEE80211_TX_UNICAST) ||
net/mac80211/tx.c
1290
if (!tx->sta)
net/mac80211/tx.c
1292
else if (test_and_clear_sta_flag(tx->sta, WLAN_STA_CLEAR_PS_FILT)) {
net/mac80211/tx.c
1294
ieee80211_check_fast_xmit(tx->sta);
net/mac80211/tx.c
137
if (tx->sdata->vif.bss_conf.basic_rates & BIT(i))
net/mac80211/tx.c
143
if (tx->sdata->deflink.operating_11g_mode)
net/mac80211/tx.c
176
tx->sdata->vif.bss_conf.use_short_preamble);
net/mac80211/tx.c
1807
static int invoke_tx_handlers_early(struct ieee80211_tx_data *tx)
net/mac80211/tx.c
1813
res = txh(tx); \
net/mac80211/tx.c
1826
tx->sdata->tx_handlers_drop++;
net/mac80211/tx.c
1827
if (tx->skb)
net/mac80211/tx.c
1828
ieee80211_free_txskb(&tx->local->hw, tx->skb);
net/mac80211/tx.c
1830
ieee80211_purge_tx_queue(&tx->local->hw, &tx->skbs);
net/mac80211/tx.c
1833
I802_DEBUG_INC(tx->local->tx_handlers_queued);
net/mac80211/tx.c
1844
static int invoke_tx_handlers_late(struct ieee80211_tx_data *tx)
net/mac80211/tx.c
1846
struct ieee80211_tx_info *info = IEEE80211_SKB_CB(tx->skb);
net/mac80211/tx.c
1849
if (!ieee80211_hw_check(&tx->local->hw, HAS_RATE_CONTROL))
net/mac80211/tx.c
185
tx->sdata->vif.bss_conf.use_short_preamble);
net/mac80211/tx.c
1853
__skb_queue_tail(&tx->skbs, tx->skb);
net/mac80211/tx.c
1854
tx->skb = NULL;
net/mac80211/tx.c
1864
if (!ieee80211_hw_check(&tx->local->hw, HAS_RATE_CONTROL))
net/mac80211/tx.c
1870
tx->sdata->tx_handlers_drop++;
net/mac80211/tx.c
1871
if (tx->skb)
net/mac80211/tx.c
1872
ieee80211_free_txskb(&tx->local->hw, tx->skb);
net/mac80211/tx.c
1874
ieee80211_purge_tx_queue(&tx->local->hw, &tx->skbs);
net/mac80211/tx.c
1877
I802_DEBUG_INC(tx->local->tx_handlers_queued);
net/mac80211/tx.c
1884
static int invoke_tx_handlers(struct ieee80211_tx_data *tx)
net/mac80211/tx.c
1886
int r = invoke_tx_handlers_early(tx);
net/mac80211/tx.c
1890
return invoke_tx_handlers_late(tx);
net/mac80211/tx.c
1899
struct ieee80211_tx_data tx;
net/mac80211/tx.c
1902
if (ieee80211_tx_prepare(sdata, &tx, NULL, skb) == TX_DROP) {
net/mac80211/tx.c
1911
if (invoke_tx_handlers(&tx))
net/mac80211/tx.c
1915
if (tx.sta)
net/mac80211/tx.c
1916
*sta = &tx.sta->sta;
net/mac80211/tx.c
1922
skb2 = __skb_dequeue(&tx.skbs);
net/mac80211/tx.c
1923
if (WARN_ON(skb2 != skb || !skb_queue_empty(&tx.skbs))) {
net/mac80211/tx.c
1925
ieee80211_purge_tx_queue(hw, &tx.skbs);
net/mac80211/tx.c
193
ieee80211_tx_h_dynamic_ps(struct ieee80211_tx_data *tx)
net/mac80211/tx.c
1941
struct ieee80211_tx_data tx;
net/mac80211/tx.c
195
struct ieee80211_local *local = tx->local;
net/mac80211/tx.c
1952
res_prepare = ieee80211_tx_prepare(sdata, &tx, sta, skb);
net/mac80211/tx.c
1956
tx.sdata->tx_handlers_drop++;
net/mac80211/tx.c
1968
if (invoke_tx_handlers_early(&tx))
net/mac80211/tx.c
197
struct ieee80211_tx_info *info = IEEE80211_SKB_CB(tx->skb);
net/mac80211/tx.c
1971
if (ieee80211_queue_skb(local, sdata, tx.sta, tx.skb))
net/mac80211/tx.c
1974
if (!invoke_tx_handlers_late(&tx))
net/mac80211/tx.c
1975
result = __ieee80211_tx(local, &tx.skbs, tx.sta, txpending);
net/mac80211/tx.c
223
if (tx->sdata->vif.type != NL80211_IFTYPE_STATION)
net/mac80211/tx.c
229
ifmgd = &tx->sdata->u.mgd;
net/mac80211/tx.c
246
skb_get_queue_mapping(tx->skb) == IEEE80211_AC_VO)
net/mac80211/tx.c
270
ieee80211_tx_h_check_assoc(struct ieee80211_tx_data *tx)
net/mac80211/tx.c
273
struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)tx->skb->data;
net/mac80211/tx.c
274
struct ieee80211_tx_info *info = IEEE80211_SKB_CB(tx->skb);
net/mac80211/tx.c
280
if (unlikely(test_bit(SCAN_SW_SCANNING, &tx->local->scanning)) &&
net/mac80211/tx.c
281
test_bit(SDATA_STATE_OFFCHANNEL, &tx->sdata->state) &&
net/mac80211/tx.c
297
if (tx->sdata->vif.type == NL80211_IFTYPE_OCB)
net/mac80211/tx.c
300
if (tx->flags & IEEE80211_TX_PS_BUFFERED)
net/mac80211/tx.c
303
if (tx->sta)
net/mac80211/tx.c
304
assoc = test_sta_flag(tx->sta, WLAN_STA_ASSOC);
net/mac80211/tx.c
306
if (likely(tx->flags & IEEE80211_TX_UNICAST)) {
net/mac80211/tx.c
310
sdata_info(tx->sdata,
net/mac80211/tx.c
314
I802_DEBUG_INC(tx->local->tx_handlers_drop_not_assoc);
net/mac80211/tx.c
318
ieee80211_vif_get_num_mcast_if(tx->sdata) == 0)) {
net/mac80211/tx.c
3557
struct ieee80211_tx_data *tx)
net/mac80211/tx.c
3559
struct sk_buff *skb = tx->skb;
net/mac80211/tx.c
3564
if (!ieee80211_hw_check(&tx->local->hw, HAS_RATE_CONTROL) &&
net/mac80211/tx.c
3565
ieee80211_tx_h_rate_ctrl(tx) != TX_CONTINUE)
net/mac80211/tx.c
3681
struct ieee80211_tx_data tx;
net/mac80211/tx.c
3730
__skb_queue_head_init(&tx.skbs);
net/mac80211/tx.c
3732
tx.flags = IEEE80211_TX_UNICAST;
net/mac80211/tx.c
3733
tx.local = local;
net/mac80211/tx.c
3734
tx.sdata = sdata;
net/mac80211/tx.c
3735
tx.sta = sta;
net/mac80211/tx.c
3736
tx.key = fast_tx->key;
net/mac80211/tx.c
3741
tx.skb = skb;
net/mac80211/tx.c
3743
fast_tx->key, &tx);
net/mac80211/tx.c
3744
tx.skb = NULL;
net/mac80211/tx.c
3746
tx.sdata->tx_handlers_drop++;
net/mac80211/tx.c
3754
__skb_queue_tail(&tx.skbs, skb);
net/mac80211/tx.c
3755
ieee80211_tx_frags(local, &sdata->vif, sta, &tx.skbs, false);
net/mac80211/tx.c
381
ieee80211_tx_h_multicast_ps_buf(struct ieee80211_tx_data *tx)
net/mac80211/tx.c
3823
struct ieee80211_tx_data tx;
net/mac80211/tx.c
383
struct ieee80211_tx_info *info = IEEE80211_SKB_CB(tx->skb);
net/mac80211/tx.c
384
struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)tx->skb->data;
net/mac80211/tx.c
3871
memset(&tx, 0, sizeof(tx));
net/mac80211/tx.c
3872
__skb_queue_head_init(&tx.skbs);
net/mac80211/tx.c
3873
tx.local = local;
net/mac80211/tx.c
3874
tx.skb = skb;
net/mac80211/tx.c
3875
tx.sdata = vif_to_sdata(info->control.vif);
net/mac80211/tx.c
3878
tx.sta = container_of(txq->sta, struct sta_info, sta);
net/mac80211/tx.c
3885
!ieee80211_vif_is_mesh(&tx.sdata->vif) &&
net/mac80211/tx.c
3886
tx.sdata->vif.type != NL80211_IFTYPE_OCB &&
net/mac80211/tx.c
3888
!test_sta_flag(tx.sta, WLAN_STA_AUTHORIZED) &&
net/mac80211/tx.c
3891
!ieee80211_is_our_addr(tx.sdata, hdr->addr2,
net/mac80211/tx.c
3904
r = ieee80211_tx_h_select_key(&tx);
net/mac80211/tx.c
3916
r = ieee80211_tx_h_rate_ctrl(&tx);
net/mac80211/tx.c
3930
if (tx.key &&
net/mac80211/tx.c
3931
(tx.key->conf.flags & IEEE80211_KEY_FLAG_GENERATE_IV))
net/mac80211/tx.c
3935
tx.key, &tx);
net/mac80211/tx.c
3941
if (invoke_tx_handlers_late(&tx))
net/mac80211/tx.c
3944
skb = __skb_dequeue(&tx.skbs);
net/mac80211/tx.c
3947
if (!skb_queue_empty(&tx.skbs)) {
net/mac80211/tx.c
3949
skb_queue_splice_tail(&tx.skbs, &txqi->frags);
net/mac80211/tx.c
396
if (tx->sdata->vif.type == NL80211_IFTYPE_AP ||
net/mac80211/tx.c
3962
switch (tx.sdata->vif.type) {
net/mac80211/tx.c
3964
if ((tx.sdata->u.mntr.flags & MONITOR_FLAG_ACTIVE) ||
net/mac80211/tx.c
3966
vif = &tx.sdata->vif;
net/mac80211/tx.c
3969
tx.sdata = rcu_dereference(local->monitor_sdata);
net/mac80211/tx.c
397
tx->sdata->vif.type == NL80211_IFTYPE_AP_VLAN) {
net/mac80211/tx.c
3970
if (tx.sdata &&
net/mac80211/tx.c
3972
vif = &tx.sdata->vif;
net/mac80211/tx.c
398
if (!tx->sdata->bss)
net/mac80211/tx.c
3984
tx.sdata = container_of(tx.sdata->bss,
net/mac80211/tx.c
3988
vif = &tx.sdata->vif;
net/mac80211/tx.c
3995
if (tx.sta &&
net/mac80211/tx.c
4004
ieee80211_sta_update_pending_airtime(local, tx.sta,
net/mac80211/tx.c
401
ps = &tx->sdata->bss->ps;
net/mac80211/tx.c
402
} else if (ieee80211_vif_is_mesh(&tx->sdata->vif)) {
net/mac80211/tx.c
403
ps = &tx->sdata->u.mesh.ps;
net/mac80211/tx.c
416
if (ieee80211_hw_check(&tx->local->hw, QUEUE_CONTROL))
net/mac80211/tx.c
417
info->hw_queue = tx->sdata->vif.cab_queue;
net/mac80211/tx.c
426
if (!ieee80211_hw_check(&tx->local->hw, HOST_BROADCAST_PS_BUFFERING))
net/mac80211/tx.c
43
static __le16 ieee80211_duration(struct ieee80211_tx_data *tx,
net/mac80211/tx.c
430
if (tx->local->total_ps_buffered >= TOTAL_MAX_TX_BUFFER)
net/mac80211/tx.c
431
purge_old_ps_buffers(tx->local);
net/mac80211/tx.c
434
ps_dbg(tx->sdata,
net/mac80211/tx.c
436
ieee80211_free_txskb(&tx->local->hw, skb_dequeue(&ps->bc_buf));
net/mac80211/tx.c
438
tx->local->total_ps_buffered++;
net/mac80211/tx.c
440
skb_queue_tail(&ps->bc_buf, tx->skb);
net/mac80211/tx.c
461
ieee80211_tx_h_unicast_ps_buf(struct ieee80211_tx_data *tx)
net/mac80211/tx.c
463
struct sta_info *sta = tx->sta;
net/mac80211/tx.c
464
struct ieee80211_tx_info *info = IEEE80211_SKB_CB(tx->skb);
net/mac80211/tx.c
465
struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)tx->skb->data;
net/mac80211/tx.c
466
struct ieee80211_local *local = tx->local;
net/mac80211/tx.c
4749
struct ieee80211_tx_data tx = {
net/mac80211/tx.c
475
int ac = skb_get_queue_mapping(tx->skb);
net/mac80211/tx.c
4769
tx.sta = sta_info_get(sdata, hdr->addr1);
net/mac80211/tx.c
4770
tx.skb = skb;
net/mac80211/tx.c
4772
if (ieee80211_tx_h_select_key(&tx) != TX_CONTINUE) {
net/mac80211/tx.c
478
!ieee80211_is_bufferable_mmpdu(tx->skb)) {
net/mac80211/tx.c
485
if (tx->local->total_ps_buffered >= TOTAL_MAX_TX_BUFFER)
net/mac80211/tx.c
486
purge_old_ps_buffers(tx->local);
net/mac80211/tx.c
49
struct ieee80211_local *local = tx->local;
net/mac80211/tx.c
504
ps_dbg(tx->sdata,
net/mac80211/tx.c
509
tx->local->total_ps_buffered++;
net/mac80211/tx.c
512
info->control.vif = &tx->sdata->vif;
net/mac80211/tx.c
515
skb_queue_tail(&sta->ps_tx_buf[ac], tx->skb);
net/mac80211/tx.c
5289
struct ieee80211_tx_data tx;
net/mac80211/tx.c
5292
memset(&tx, 0, sizeof(tx));
net/mac80211/tx.c
5293
tx.key = rcu_dereference(link->default_beacon_key);
net/mac80211/tx.c
5294
if (!tx.key)
net/mac80211/tx.c
5297
if (unlikely(tx.key->flags & KEY_FLAG_TAINTED)) {
net/mac80211/tx.c
5298
tx.key = NULL;
net/mac80211/tx.c
5302
if (!(tx.key->conf.flags & IEEE80211_KEY_FLAG_SW_MGMT_TX) &&
net/mac80211/tx.c
5303
tx.key->flags & KEY_FLAG_UPLOADED_TO_HARDWARE)
net/mac80211/tx.c
5304
IEEE80211_SKB_CB(skb)->control.hw_key = &tx.key->conf;
net/mac80211/tx.c
5306
tx.local = local;
net/mac80211/tx.c
5307
tx.sdata = sdata;
net/mac80211/tx.c
5308
__skb_queue_head_init(&tx.skbs);
net/mac80211/tx.c
5309
__skb_queue_tail(&tx.skbs, skb);
net/mac80211/tx.c
531
ps_dbg(tx->sdata,
net/mac80211/tx.c
5310
res = ieee80211_tx_h_encrypt(&tx);
net/mac80211/tx.c
5311
check_skb = __skb_dequeue(&tx.skbs);
net/mac80211/tx.c
540
ieee80211_tx_h_ps_buf(struct ieee80211_tx_data *tx)
net/mac80211/tx.c
542
if (unlikely(tx->flags & IEEE80211_TX_PS_BUFFERED))
net/mac80211/tx.c
545
if (tx->flags & IEEE80211_TX_UNICAST)
net/mac80211/tx.c
546
return ieee80211_tx_h_unicast_ps_buf(tx);
net/mac80211/tx.c
548
return ieee80211_tx_h_multicast_ps_buf(tx);
net/mac80211/tx.c
55
if (tx->rate.flags & (IEEE80211_TX_RC_MCS | IEEE80211_TX_RC_VHT_MCS))
net/mac80211/tx.c
552
ieee80211_tx_h_check_control_port_protocol(struct ieee80211_tx_data *tx)
net/mac80211/tx.c
554
struct ieee80211_tx_info *info = IEEE80211_SKB_CB(tx->skb);
net/mac80211/tx.c
556
if (unlikely(tx->sdata->control_port_protocol == tx->skb->protocol)) {
net/mac80211/tx.c
557
if (tx->sdata->control_port_no_encrypt)
net/mac80211/tx.c
567
ieee80211_select_link_key(struct ieee80211_tx_data *tx)
net/mac80211/tx.c
569
struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)tx->skb->data;
net/mac80211/tx.c
570
struct ieee80211_tx_info *info = IEEE80211_SKB_CB(tx->skb);
net/mac80211/tx.c
576
link = &tx->sdata->deflink;
net/mac80211/tx.c
578
link = rcu_dereference(tx->sdata->link[link_id]);
net/mac80211/tx.c
583
if (ieee80211_is_group_privacy_action(tx->skb))
net/mac80211/tx.c
587
ieee80211_is_robust_mgmt_frame(tx->skb))
net/mac80211/tx.c
59
if (WARN_ON_ONCE(tx->rate.idx < 0))
net/mac80211/tx.c
596
ieee80211_tx_h_select_key(struct ieee80211_tx_data *tx)
net/mac80211/tx.c
599
struct ieee80211_tx_info *info = IEEE80211_SKB_CB(tx->skb);
net/mac80211/tx.c
600
struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)tx->skb->data;
net/mac80211/tx.c
603
tx->key = NULL;
net/mac80211/tx.c
607
if (tx->sta &&
net/mac80211/tx.c
6078
struct ieee80211_tx_data tx;
net/mac80211/tx.c
608
(key = rcu_dereference(tx->sta->ptk[tx->sta->ptk_idx])))
net/mac80211/tx.c
609
tx->key = key;
net/mac80211/tx.c
610
else if ((key = ieee80211_select_link_key(tx)))
net/mac80211/tx.c
611
tx->key = key;
net/mac80211/tx.c
6127
if (!ieee80211_tx_prepare(sdata, &tx, NULL, skb))
net/mac80211/tx.c
613
(key = rcu_dereference(tx->sdata->default_unicast_key)))
net/mac80211/tx.c
6134
tx.flags |= IEEE80211_TX_PS_BUFFERED;
net/mac80211/tx.c
6137
if (invoke_tx_handlers(&tx))
net/mac80211/tx.c
614
tx->key = key;
net/mac80211/tx.c
616
tx->key = NULL;
net/mac80211/tx.c
619
if (tx->key && tx->key->flags & KEY_FLAG_UPLOADED_TO_HARDWARE)
net/mac80211/tx.c
620
info->control.hw_key = &tx->key->conf;
net/mac80211/tx.c
624
if (tx->key) {
net/mac80211/tx.c
629
switch (tx->key->conf.cipher) {
net/mac80211/tx.c
634
tx->key = NULL;
net/mac80211/tx.c
641
!ieee80211_use_mfp(hdr->frame_control, tx->sta,
net/mac80211/tx.c
642
tx->skb) &&
net/mac80211/tx.c
643
!ieee80211_is_group_privacy_action(tx->skb) &&
net/mac80211/tx.c
645
tx->sta))
net/mac80211/tx.c
646
tx->key = NULL;
net/mac80211/tx.c
648
skip_hw = (tx->key->conf.flags &
net/mac80211/tx.c
657
tx->key = NULL;
net/mac80211/tx.c
66
txrate = &sband->bitrates[tx->rate.idx];
net/mac80211/tx.c
661
if (unlikely(tx->key && tx->key->flags & KEY_FLAG_TAINTED &&
net/mac80211/tx.c
663
tx->skb->protocol != tx->sdata->control_port_protocol)
net/mac80211/tx.c
666
if (!skip_hw && tx->key &&
net/mac80211/tx.c
667
tx->key->flags & KEY_FLAG_UPLOADED_TO_HARDWARE)
net/mac80211/tx.c
668
info->control.hw_key = &tx->key->conf;
net/mac80211/tx.c
669
} else if (ieee80211_is_data_present(hdr->frame_control) && tx->sta &&
net/mac80211/tx.c
670
test_sta_flag(tx->sta, WLAN_STA_USES_ENCRYPTION)) {
net/mac80211/tx.c
678
ieee80211_tx_h_rate_ctrl(struct ieee80211_tx_data *tx)
net/mac80211/tx.c
680
struct ieee80211_tx_info *info = IEEE80211_SKB_CB(tx->skb);
net/mac80211/tx.c
681
struct ieee80211_hdr *hdr = (void *)tx->skb->data;
net/mac80211/tx.c
692
sband = tx->local->hw.wiphy->bands[info->band];
net/mac80211/tx.c
696
len = min_t(u32, tx->skb->len + FCS_LEN,
net/mac80211/tx.c
697
tx->local->hw.wiphy->frag_threshold);
net/mac80211/tx.c
700
txrc.hw = &tx->local->hw;
net/mac80211/tx.c
702
txrc.bss_conf = &tx->sdata->vif.bss_conf;
net/mac80211/tx.c
703
txrc.skb = tx->skb;
net/mac80211/tx.c
709
txrc.rate_idx_mask = tx->sdata->rc_rateidx_mask[info->band];
net/mac80211/tx.c
711
if (tx->sdata->rc_has_mcs_mask[info->band])
net/mac80211/tx.c
713
tx->sdata->rc_rateidx_mcs_mask[info->band];
net/mac80211/tx.c
716
txrc.bss = (tx->sdata->vif.type == NL80211_IFTYPE_AP ||
net/mac80211/tx.c
717
tx->sdata->vif.type == NL80211_IFTYPE_MESH_POINT ||
net/mac80211/tx.c
718
tx->sdata->vif.type == NL80211_IFTYPE_ADHOC ||
net/mac80211/tx.c
719
tx->sdata->vif.type == NL80211_IFTYPE_OCB);
net/mac80211/tx.c
722
if (len > tx->local->hw.wiphy->rts_threshold) {
net/mac80211/tx.c
727
info->control.use_cts_prot = tx->sdata->vif.bss_conf.use_cts_prot;
net/mac80211/tx.c
735
if (tx->sdata->vif.bss_conf.use_short_preamble &&
net/mac80211/tx.c
736
(ieee80211_is_tx_data(tx->skb) ||
net/mac80211/tx.c
737
(tx->sta && test_sta_flag(tx->sta, WLAN_STA_SHORT_PREAMBLE))))
net/mac80211/tx.c
746
if (tx->sta)
net/mac80211/tx.c
747
assoc = test_sta_flag(tx->sta, WLAN_STA_ASSOC);
net/mac80211/tx.c
753
if (WARN(test_bit(SCAN_SW_SCANNING, &tx->local->scanning) && assoc &&
net/mac80211/tx.c
754
!rate_usable_index_exists(sband, &tx->sta->sta),
net/mac80211/tx.c
758
tx->sdata->name,
net/mac80211/tx.c
767
rate_control_get_rate(tx->sdata, tx->sta, &txrc);
net/mac80211/tx.c
769
if (tx->sta && !info->control.skip_table)
net/mac80211/tx.c
770
ratetbl = rcu_dereference(tx->sta->sta.rates);
net/mac80211/tx.c
783
tx->rate = rate;
net/mac80211/tx.c
788
tx->rate = info->control.rates[0];
net/mac80211/tx.c
792
txrc.reported_rate = tx->rate;
net/mac80211/tx.c
793
if (tx->sta && ieee80211_is_tx_data(tx->skb))
net/mac80211/tx.c
794
tx->sta->deflink.tx_stats.last_rate = txrc.reported_rate;
net/mac80211/tx.c
795
} else if (tx->sta)
net/mac80211/tx.c
796
tx->sta->deflink.tx_stats.last_rate = txrc.reported_rate;
net/mac80211/tx.c
823
ieee80211_tx_h_sequence(struct ieee80211_tx_data *tx)
net/mac80211/tx.c
825
struct ieee80211_tx_info *info = IEEE80211_SKB_CB(tx->skb);
net/mac80211/tx.c
826
struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)tx->skb->data;
net/mac80211/tx.c
854
tx->sdata->mld_mcast_seq += 0x10;
net/mac80211/tx.c
855
hdr->seq_ctrl = cpu_to_le16(tx->sdata->mld_mcast_seq);
net/mac80211/tx.c
870
hdr->seq_ctrl = cpu_to_le16(tx->sdata->sequence_number);
net/mac80211/tx.c
871
tx->sdata->sequence_number += 0x10;
net/mac80211/tx.c
872
if (tx->sta)
net/mac80211/tx.c
873
tx->sta->deflink.tx_stats.msdu[IEEE80211_NUM_TIDS]++;
net/mac80211/tx.c
882
if (!tx->sta)
net/mac80211/tx.c
887
tx->sta->deflink.tx_stats.msdu[tid]++;
net/mac80211/tx.c
889
hdr->seq_ctrl = ieee80211_tx_next_seq(tx->sta, tid);
net/mac80211/tx.c
894
static int ieee80211_fragment(struct ieee80211_tx_data *tx,
net/mac80211/tx.c
898
struct ieee80211_local *local = tx->local;
net/mac80211/tx.c
923
__skb_queue_tail(&tx->skbs, tmp);
net/mac80211/tx.c
955
ieee80211_tx_h_fragment(struct ieee80211_tx_data *tx)
net/mac80211/tx.c
957
struct sk_buff *skb = tx->skb;
net/mac80211/tx.c
960
int frag_threshold = tx->local->hw.wiphy->frag_threshold;
net/mac80211/tx.c
965
__skb_queue_tail(&tx->skbs, skb);
net/mac80211/tx.c
966
tx->skb = NULL;
net/mac80211/tx.c
971
if (ieee80211_hw_check(&tx->local->hw, SUPPORTS_TX_FRAG))
net/mac80211/tx.c
996
if (ieee80211_fragment(tx, skb, hdrlen, frag_threshold))
net/mac80211/util.c
105
void ieee80211_tx_set_protected(struct ieee80211_tx_data *tx)
net/mac80211/util.c
110
skb_queue_walk(&tx->skbs, skb) {
net/mac80211/wep.c
271
static int wep_encrypt_skb(struct ieee80211_tx_data *tx, struct sk_buff *skb)
net/mac80211/wep.c
277
if (ieee80211_wep_encrypt(tx->local, skb, tx->key->conf.key,
net/mac80211/wep.c
278
tx->key->conf.keylen,
net/mac80211/wep.c
279
tx->key->conf.keyidx))
net/mac80211/wep.c
283
if (!ieee80211_wep_add_iv(tx->local, skb,
net/mac80211/wep.c
284
tx->key->conf.keylen,
net/mac80211/wep.c
285
tx->key->conf.keyidx))
net/mac80211/wep.c
293
ieee80211_crypto_wep_encrypt(struct ieee80211_tx_data *tx)
net/mac80211/wep.c
297
ieee80211_tx_set_protected(tx);
net/mac80211/wep.c
299
skb_queue_walk(&tx->skbs, skb) {
net/mac80211/wep.c
300
if (wep_encrypt_skb(tx, skb) < 0) {
net/mac80211/wep.c
301
I802_DEBUG_INC(tx->local->tx_handlers_drop_wep);
net/mac80211/wep.h
28
ieee80211_crypto_wep_encrypt(struct ieee80211_tx_data *tx);
net/mac80211/wpa.c
194
static int tkip_encrypt_skb(struct ieee80211_tx_data *tx, struct sk_buff *skb)
net/mac80211/wpa.c
197
struct ieee80211_key *key = tx->key;
net/mac80211/wpa.c
243
return ieee80211_tkip_encrypt_data(&tx->local->wep_tx_ctx,
net/mac80211/wpa.c
249
ieee80211_crypto_tkip_encrypt(struct ieee80211_tx_data *tx)
net/mac80211/wpa.c
253
ieee80211_tx_set_protected(tx);
net/mac80211/wpa.c
255
skb_queue_walk(&tx->skbs, skb) {
net/mac80211/wpa.c
256
if (tkip_encrypt_skb(tx, skb) < 0)
net/mac80211/wpa.c
30
ieee80211_tx_h_michael_mic_add(struct ieee80211_tx_data *tx)
net/mac80211/wpa.c
36
struct sk_buff *skb = tx->skb;
net/mac80211/wpa.c
41
if (!tx->key || tx->key->conf.cipher != WLAN_CIPHER_SUITE_TKIP ||
net/mac80211/wpa.c
426
static int ccmp_encrypt_skb(struct ieee80211_tx_data *tx, struct sk_buff *skb,
net/mac80211/wpa.c
430
struct ieee80211_key *key = tx->key;
net/mac80211/wpa.c
498
ieee80211_crypto_ccmp_encrypt(struct ieee80211_tx_data *tx,
net/mac80211/wpa.c
503
ieee80211_tx_set_protected(tx);
net/mac80211/wpa.c
505
skb_queue_walk(&tx->skbs, skb) {
net/mac80211/wpa.c
506
if (ccmp_encrypt_skb(tx, skb, mic_len) < 0)
net/mac80211/wpa.c
59
ieee80211_hw_check(&tx->local->hw, SUPPORTS_TX_FRAG)) &&
net/mac80211/wpa.c
60
!(tx->key->conf.flags & (IEEE80211_KEY_FLAG_GENERATE_MMIC |
net/mac80211/wpa.c
628
static int gcmp_encrypt_skb(struct ieee80211_tx_data *tx, struct sk_buff *skb)
net/mac80211/wpa.c
631
struct ieee80211_key *key = tx->key;
net/mac80211/wpa.c
699
ieee80211_crypto_gcmp_encrypt(struct ieee80211_tx_data *tx)
net/mac80211/wpa.c
703
ieee80211_tx_set_protected(tx);
net/mac80211/wpa.c
705
skb_queue_walk(&tx->skbs, skb) {
net/mac80211/wpa.c
706
if (gcmp_encrypt_skb(tx, skb) < 0)
net/mac80211/wpa.c
79
if (tx->key->conf.flags & IEEE80211_KEY_FLAG_PUT_MIC_SPACE) {
net/mac80211/wpa.c
833
ieee80211_crypto_aes_cmac_encrypt(struct ieee80211_tx_data *tx,
net/mac80211/wpa.c
838
struct ieee80211_key *key = tx->key;
net/mac80211/wpa.c
844
if (WARN_ON(skb_queue_len(&tx->skbs) != 1))
net/mac80211/wpa.c
847
skb = skb_peek(&tx->skbs);
net/mac80211/wpa.c
85
key = &tx->key->conf.key[NL80211_TKIP_DATA_OFFSET_TX_MIC_KEY];
net/mac80211/wpa.c
940
ieee80211_crypto_aes_gmac_encrypt(struct ieee80211_tx_data *tx)
net/mac80211/wpa.c
944
struct ieee80211_key *key = tx->key;
net/mac80211/wpa.c
951
if (WARN_ON(skb_queue_len(&tx->skbs) != 1))
net/mac80211/wpa.c
954
skb = skb_peek(&tx->skbs);
net/mac80211/wpa.h
15
ieee80211_tx_h_michael_mic_add(struct ieee80211_tx_data *tx);
net/mac80211/wpa.h
20
ieee80211_crypto_tkip_encrypt(struct ieee80211_tx_data *tx);
net/mac80211/wpa.h
25
ieee80211_crypto_ccmp_encrypt(struct ieee80211_tx_data *tx,
net/mac80211/wpa.h
32
ieee80211_crypto_aes_cmac_encrypt(struct ieee80211_tx_data *tx,
net/mac80211/wpa.h
38
ieee80211_crypto_aes_gmac_encrypt(struct ieee80211_tx_data *tx);
net/mac80211/wpa.h
43
ieee80211_crypto_gcmp_encrypt(struct ieee80211_tx_data *tx);
net/nfc/nci/spi.c
180
struct spi_transfer tx, rx;
net/nfc/nci/spi.c
186
memset(&tx, 0, sizeof(struct spi_transfer));
net/nfc/nci/spi.c
189
tx.tx_buf = req;
net/nfc/nci/spi.c
190
tx.len = 2;
net/nfc/nci/spi.c
191
tx.cs_change = 0;
net/nfc/nci/spi.c
192
tx.speed_hz = nspi->xfer_speed_hz;
net/nfc/nci/spi.c
193
spi_message_add_tail(&tx, &m);
net/psp/psp_sock.c
204
if (pas->tx.spi) {
net/psp/psp_sock.c
226
memcpy(&dummy->tx, key, sizeof(*key));
net/psp/psp_sock.c
232
memcpy(&pas->tx, key, sizeof(*key));
net/psp/psp_sock.c
290
if (pas && pas->tx.spi)
net/psp/psp_sock.c
88
if (pas->tx.spi)
net/smc/smc_stats.h
119
SMC_STAT_PAYLOAD_SUB(_smc_stats, SMC_TYPE_D, tx, _len, _rc); \
net/smc/smc_stats.h
121
SMC_STAT_PAYLOAD_SUB(_smc_stats, SMC_TYPE_R, tx, _len, _rc); \
net/smc/smc_stats.h
175
SMC_STAT_RMB_SIZE_SUB(_smc_stats, SMC_TYPE_D, tx, is_add, l); \
net/smc/smc_stats.h
179
SMC_STAT_RMB_SIZE_SUB(_smc_stats, SMC_TYPE_R, tx, is_add, l); \
net/smc/smc_stats.h
192
SMC_STAT_RMB_SUB(_smc_stats, type, SMC_TYPE_D, tx); \
net/smc/smc_stats.h
196
SMC_STAT_RMB_SUB(_smc_stats, type, SMC_TYPE_R, tx); \
net/tipc/crypto.c
1213
struct tipc_crypto *tx, *rx;
net/tipc/crypto.c
1220
tx = tipc_net(rx->net)->crypto_tx;
net/tipc/crypto.c
1230
tipc_aead_users_dec(tx->aead[k], 0);
net/tipc/crypto.c
1232
tx->timer1 = jiffies;
net/tipc/crypto.c
1319
static struct tipc_aead *tipc_crypto_key_pick_tx(struct tipc_crypto *tx,
net/tipc/crypto.c
1326
struct tipc_key key = tx->key;
net/tipc/crypto.c
1340
spin_lock(&tx->lock);
net/tipc/crypto.c
1342
aead = tipc_aead_rcu_ptr(tx->aead[KEY_MASTER], &tx->lock);
net/tipc/crypto.c
1350
aead = tipc_aead_rcu_ptr(tx->aead[k], &tx->lock);
net/tipc/crypto.c
1370
spin_unlock(&tx->lock);
net/tipc/crypto.c
1393
struct tipc_crypto *tx = tipc_net(rx->net)->crypto_tx;
net/tipc/crypto.c
1404
tx->legacy_user = 1;
net/tipc/crypto.c
1413
tx->timer2 = jiffies;
net/tipc/crypto.c
1415
if (tx->key.keys &&
net/tipc/crypto.c
1420
if (queue_delayed_work(tx->wq, &rx->work, delay))
net/tipc/crypto.c
1431
if (tx->key.keys &&
net/tipc/crypto.c
1435
tipc_aead_users_inc(tx->aead[new], INT_MAX);
net/tipc/crypto.c
1437
tipc_aead_users_dec(tx->aead[cur], 0);
net/tipc/crypto.c
1441
tx->timer1 = jiffies;
net/tipc/crypto.c
1444
tx->name, cur, new, rx->name);
net/tipc/crypto.c
1450
struct tipc_crypto *tx = tipc_net(net)->crypto_tx;
net/tipc/crypto.c
1453
spin_lock_bh(&tx->lock);
net/tipc/crypto.c
1454
key = tx->key;
net/tipc/crypto.c
1458
tipc_crypto_key_set_state(tx, key.passive, 0, key.pending);
net/tipc/crypto.c
1459
tipc_crypto_key_detach(tx->aead[key.active], &tx->lock);
net/tipc/crypto.c
1460
spin_unlock_bh(&tx->lock);
net/tipc/crypto.c
1462
pr_warn("%s: key is revoked\n", tx->name);
net/tipc/crypto.c
1554
struct tipc_crypto *tx = tn->crypto_tx;
net/tipc/crypto.c
1559
spin_lock(&tx->lock);
net/tipc/crypto.c
1560
key = tx->key;
net/tipc/crypto.c
1561
if (key.active && tipc_aead_users(tx->aead[key.active]) > 0)
net/tipc/crypto.c
1563
if (!key.pending || tipc_aead_users(tx->aead[key.pending]) <= 0)
net/tipc/crypto.c
1565
if (time_before(jiffies, tx->timer1 + TIPC_TX_LASTING_TIME))
net/tipc/crypto.c
1568
tipc_crypto_key_set_state(tx, key.passive, key.pending, 0);
net/tipc/crypto.c
1570
tipc_crypto_key_detach(tx->aead[key.active], &tx->lock);
net/tipc/crypto.c
1571
this_cpu_inc(tx->stats->stat[STAT_SWITCHES]);
net/tipc/crypto.c
1572
pr_info("%s: key[%d] is activated\n", tx->name, key.pending);
net/tipc/crypto.c
1575
spin_unlock(&tx->lock);
net/tipc/crypto.c
1638
if (time_after(jiffies, tx->timer2 + TIPC_TX_GRACE_PERIOD))
net/tipc/crypto.c
1639
tx->legacy_user = 0;
net/tipc/crypto.c
1694
struct tipc_crypto *tx = tipc_net(net)->crypto_tx;
net/tipc/crypto.c
1695
struct tipc_crypto_stats __percpu *stats = tx->stats;
net/tipc/crypto.c
1697
struct tipc_key key = tx->key;
net/tipc/crypto.c
1705
if (!tx->working)
net/tipc/crypto.c
1711
if (!tx->key_master && !key.active)
net/tipc/crypto.c
1716
pr_debug("%s: probing for key[%d]\n", tx->name,
net/tipc/crypto.c
1726
if (tx->key_master) {
net/tipc/crypto.c
1731
pr_debug("%s: gracing for msg (%d %d)\n", tx->name,
net/tipc/crypto.c
1738
time_before(jiffies, tx->timer2 + TIPC_TX_GRACE_PERIOD)) {
net/tipc/crypto.c
1743
if (likely(!tx->legacy_user))
net/tipc/crypto.c
1760
aead = tipc_aead_get(tx->aead[tx_key]);
net/tipc/crypto.c
1817
struct tipc_crypto *tx = tipc_net(net)->crypto_tx;
net/tipc/crypto.c
1844
aead = tipc_crypto_key_pick_tx(tx, rx, *skb, tx_key);
net/tipc/crypto.c
1857
stats = ((rx) ?: tx)->stats;
net/tipc/crypto.c
2010
struct tipc_crypto *tx = tn->crypto_tx, *rx;
net/tipc/crypto.c
2031
tipc_crypto_key_dump(tx, buf));
net/tipc/crypto.c
2053
stat = per_cpu_ptr(tx->stats, cpu)->stat[i];
net/tipc/crypto.c
2200
int tipc_crypto_key_distr(struct tipc_crypto *tx, u8 key,
net/tipc/crypto.c
2212
aead = tipc_aead_get(tx->aead[key]);
net/tipc/crypto.c
2214
rc = tipc_crypto_key_xmit(tx->net, aead->key,
net/tipc/crypto.c
2289
struct tipc_crypto *tx = tipc_net(rx->net)->crypto_tx;
net/tipc/crypto.c
2342
if (likely(skey && queue_delayed_work(tx->wq, &rx->work, 0)))
net/tipc/crypto.c
2359
struct tipc_crypto *tx = tipc_net(rx->net)->crypto_tx;
net/tipc/crypto.c
2370
key = tx->key.pending ?: tx->key.active;
net/tipc/crypto.c
2371
rc = tipc_crypto_key_distr(tx, key, rx->node);
net/tipc/crypto.c
2374
tx->name, key, tipc_node_get_id_str(rx->node),
net/tipc/crypto.c
2403
if (resched && queue_delayed_work(tx->wq, &rx->work, delay))
net/tipc/crypto.c
2415
void tipc_crypto_rekeying_sched(struct tipc_crypto *tx, bool changed,
net/tipc/crypto.c
2425
tx->rekeying_intv = new_intv;
net/tipc/crypto.c
2426
cancel_delayed_work_sync(&tx->work);
net/tipc/crypto.c
2429
if (tx->rekeying_intv || now) {
net/tipc/crypto.c
2430
delay = (now) ? 0 : tx->rekeying_intv * 60 * 1000;
net/tipc/crypto.c
2431
queue_delayed_work(tx->wq, &tx->work, msecs_to_jiffies(delay));
net/tipc/crypto.c
2447
struct tipc_crypto *tx = container_of(dwork, struct tipc_crypto, work);
net/tipc/crypto.c
2449
struct tipc_key key = tx->key;
net/tipc/crypto.c
2458
aead = rcu_dereference(tx->aead[key.active ?: KEY_MASTER]);
net/tipc/crypto.c
2472
tipc_crypto_key_init(tx, skey, PER_NODE_KEY, false);
net/tipc/crypto.c
2474
rc = tipc_crypto_key_distr(tx, rc, NULL);
net/tipc/crypto.c
2479
pr_warn_ratelimited("%s: rekeying returns %d\n", tx->name, rc);
net/tipc/crypto.c
2483
tipc_crypto_rekeying_sched(tx, false, 0);
net/tipc/crypto.c
286
static struct tipc_aead *tipc_crypto_key_pick_tx(struct tipc_crypto *tx,
net/tipc/crypto.c
847
struct tipc_crypto *tx = aead->crypto;
net/tipc/crypto.c
848
struct net *net = tx->net;
net/tipc/crypto.c
852
this_cpu_inc(tx->stats->stat[STAT_ASYNC_OK]);
net/tipc/crypto.c
863
this_cpu_inc(tx->stats->stat[STAT_ASYNC_NOK]);
net/tipc/crypto.h
171
int tipc_crypto_key_distr(struct tipc_crypto *tx, u8 key,
net/tipc/crypto.h
174
void tipc_crypto_rekeying_sched(struct tipc_crypto *tx, bool changed,
net/tipc/node.c
2964
struct tipc_crypto *tx = tipc_net(net)->crypto_tx, *c = tx;
net/tipc/node.c
3029
} else if (c == tx) {
net/tipc/node.c
3031
if (!master_key && tipc_crypto_key_distr(tx, rc, NULL))
net/tipc/node.c
3035
tipc_crypto_rekeying_sched(tx, rekeying, intv);
net/tls/tls.h
150
int tls_set_sw_offload(struct sock *sk, int tx,
net/tls/tls.h
347
ctx->tx.iv + prot->salt_size, iv_size);
net/tls/tls_device.c
1054
memcpy(&rcd_sn, ctx->tx.rec_seq, sizeof(rcd_sn));
net/tls/tls_device.c
1110
memcpy(ctx->tx.iv + cipher_desc->salt, iv, cipher_desc->iv);
net/tls/tls_device.c
1111
memcpy(ctx->tx.rec_seq, rec_seq, cipher_desc->rec_seq);
net/tls/tls_device.c
240
rcd_sn = tls_ctx->tx.rec_seq;
net/tls/tls_device.c
297
tls_advance_record_sn(sk, prot, &ctx->tx);
net/tls/tls_main.c
1103
static u16 tls_user_config(struct tls_context *ctx, bool tx)
net/tls/tls_main.c
1105
u16 config = tx ? ctx->tx_conf : ctx->rx_conf;
net/tls/tls_main.c
436
int __user *optlen, int tx)
net/tls/tls_main.c
459
if (tx) {
net/tls/tls_main.c
461
cctx = &ctx->tx;
net/tls/tls_main.c
636
unsigned int optlen, int tx)
net/tls/tls_main.c
651
if (tx) {
net/tls/tls_main.c
666
TLS_INC_STATS(sock_net(sk), tx ? LINUX_MIB_TLSTXREKEYERROR
net/tls/tls_main.c
713
if (tx) {
net/tls/tls_main.c
757
if (tx)
net/tls/tls_main.c
766
if (tx) {
net/tls/tls_main.c
778
TLS_INC_STATS(sock_net(sk), tx ? LINUX_MIB_TLSTXREKEYERROR
net/tls/tls_sw.c
2778
int tls_set_sw_offload(struct sock *sk, int tx,
net/tls/tls_sw.c
2798
if (tx) {
net/tls/tls_sw.c
2809
if (tx) {
net/tls/tls_sw.c
2812
cctx = &ctx->tx;
net/tls/tls_sw.c
2867
if (!tx && !new_crypto_info) {
net/tls/tls_sw.c
2889
if (!tx)
net/tls/tls_sw.c
2900
if (tx) {
net/tls/tls_sw.c
558
memcpy(&rec->iv_data[iv_offset], tls_ctx->tx.iv,
net/tls/tls_sw.c
562
tls_ctx->tx.rec_seq);
net/tls/tls_sw.c
603
tls_advance_record_sn(sk, prot, &tls_ctx->tx);
net/tls/tls_sw.c
810
tls_ctx->tx.rec_seq, record_type, prot);
net/wireless/mlme.c
875
if (!(wdev->wiphy->mgmt_stypes[wdev->iftype].tx & BIT(stype >> 4)))
net/wireless/nl80211.c
15884
bool tx)
net/wireless/nl80211.c
15909
if (tx) {
net/wireless/nl80211.c
2169
stypes = mgmt_stypes[ift].tx;
net/wireless/trace.h
1883
TP_PROTO(struct wiphy *wiphy, int ret, u32 tx, u32 rx),
net/wireless/trace.h
1884
TP_ARGS(wiphy, ret, tx, rx),
net/wireless/trace.h
1888
__field(u32, tx)
net/wireless/trace.h
1894
__entry->tx = tx;
net/wireless/trace.h
1898
WIPHY_PR_ARG, __entry->ret, __entry->tx, __entry->rx)
net/wireless/trace.h
1902
TP_PROTO(struct wiphy *wiphy, u32 tx, u32 tx_max,
net/wireless/trace.h
1904
TP_ARGS(wiphy, tx, tx_max, rx, rx_max),
net/wireless/trace.h
1907
__field(u32, tx)
net/wireless/trace.h
1914
__entry->tx = tx;
net/wireless/trace.h
1920
WIPHY_PR_ARG, __entry->tx, __entry->tx_max, __entry->rx,
net/wireless/trace.h
1925
TP_PROTO(struct wiphy *wiphy, int radio_idx, u32 tx, u32 rx),
net/wireless/trace.h
1926
TP_ARGS(wiphy, radio_idx, tx, rx),
net/wireless/trace.h
1930
__field(u32, tx)
net/wireless/trace.h
1936
__entry->tx = tx;
net/wireless/trace.h
1941
__entry->tx, __entry->rx)
net/xdp/xsk.c
1044
if (unlikely(!xs->tx))
net/xdp/xsk.c
1129
else if (xs->tx)
net/xdp/xsk.c
1136
if (xs->tx && xsk_tx_writeable(xs))
net/xdp/xsk.c
1247
xskq_destroy(xs->tx);
net/xdp/xsk.c
1320
if (!xs->rx && !xs->tx) {
net/xdp/xsk.c
1400
if (xs->tx && !xs->pool->tx_descs) {
net/xdp/xsk.c
1502
q = (optname == XDP_TX_RING) ? &xs->tx : &xs->rx;
net/xdp/xsk.c
1506
xs->tx->ring->flags |= XDP_RING_NEED_WAKEUP;
net/xdp/xsk.c
1584
if (!xs->tx ||
net/xdp/xsk.c
1585
budget < TX_BATCH_SIZE || budget > xs->tx->nentries)
net/xdp/xsk.c
1655
stats.tx_ring_empty_descs = xskq_nb_queue_empty_descs(xs->tx);
net/xdp/xsk.c
1660
stats.tx_invalid_descs = xskq_nb_invalid_descs(xs->tx);
net/xdp/xsk.c
1689
&off.tx);
net/xdp/xsk.c
1696
off.tx.flags = offsetof(struct xdp_rxtx_ring,
net/xdp/xsk.c
1707
xsk_enter_rxtx_offsets(&off_v1.tx);
net/xdp/xsk.c
1764
q = READ_ONCE(xs->tx);
net/xdp/xsk.c
310
if (xskq_cons_present_entries(xs->tx) > xs->tx->nentries / 2)
net/xdp/xsk.c
318
__xskq_cons_release(xs->tx);
net/xdp/xsk.c
449
if (!xskq_cons_peek_desc(xs->tx, desc, pool)) {
net/xdp/xsk.c
450
if (xskq_has_descs(xs->tx))
net/xdp/xsk.c
451
xskq_cons_release(xs->tx);
net/xdp/xsk.c
465
xskq_cons_release(xs->tx);
net/xdp/xsk.c
513
nb_pkts = xskq_cons_nb_entries(xs->tx, nb_pkts);
net/xdp/xsk.c
525
nb_pkts = xskq_cons_read_desc_batch(xs->tx, pool, nb_pkts);
net/xdp/xsk.c
527
xs->tx->queue_empty_descs++;
net/xdp/xsk.c
531
__xskq_cons_release(xs->tx);
net/xdp/xsk.c
65
xs->tx->ring->flags |= XDP_RING_NEED_WAKEUP;
net/xdp/xsk.c
677
xdp_sk(skb->sk)->tx->invalid_descs += xsk_get_num_desc(skb);
net/xdp/xsk.c
899
xskq_cons_release(xs->tx);
net/xdp/xsk.c
92
xs->tx->ring->flags &= ~XDP_RING_NEED_WAKEUP;
net/xdp/xsk.c
929
while (xskq_cons_peek_desc(xs->tx, &desc, xs->pool)) {
net/xdp/xsk.c
955
xskq_cons_release(xs->tx);
net/xdp/xsk.c
965
xskq_cons_cancel_n(xs->tx, xsk_get_num_desc(skb));
net/xdp/xsk.c
983
if (xskq_has_descs(xs->tx)) {
net/xdp/xsk.c
986
xskq_cons_release(xs->tx);
net/xdp/xsk.h
22
struct xdp_ring_offset_v1 tx;
net/xdp/xsk_buff_pool.c
15
if (!xs->tx)
net/xdp/xsk_buff_pool.c
25
if (!xs->tx)
net/xdp/xsk_buff_pool.c
45
pool->tx_descs = kvzalloc_objs(*pool->tx_descs, xs->tx->nentries);
net/xdp/xsk_buff_pool.c
69
if (xs->tx)
net/xdp/xsk_diag.c
42
if (!err && xs->tx)
net/xdp/xsk_diag.c
43
err = xsk_diag_put_ring(xs->tx, XDP_DIAG_TX_RING, nlskb);
net/xdp/xsk_diag.c
87
du.n_tx_invalid = xskq_nb_invalid_descs(xs->tx);
net/xdp/xsk_diag.c
88
du.n_tx_ring_empty = xskq_nb_queue_empty_descs(xs->tx);
scripts/kconfig/gconf.c
755
gint tx = (gint) event->x;
scripts/kconfig/gconf.c
758
gtk_tree_view_get_path_at_pos(view, tx, ty, &path, &column, NULL, NULL);
scripts/kconfig/gconf.c
864
gint tx = (gint) event->x;
scripts/kconfig/gconf.c
867
gtk_tree_view_get_path_at_pos(view, tx, ty, &path, NULL, NULL, NULL);
sound/firewire/amdtp-stream.c
1331
packet_count = header_length / s->ctx_data.tx.ctx_header_size;
sound/firewire/amdtp-stream.c
1375
packets = header_length / s->ctx_data.tx.ctx_header_size;
sound/firewire/amdtp-stream.c
1377
ctx_header += (packets - 1) * s->ctx_data.tx.ctx_header_size / sizeof(*ctx_header);
sound/firewire/amdtp-stream.c
1403
packets = header_length / s->ctx_data.tx.ctx_header_size;
sound/firewire/amdtp-stream.c
1413
ctx_header += s->ctx_data.tx.ctx_header_size / sizeof(__be32);
sound/firewire/amdtp-stream.c
1420
size_t length = s->ctx_data.tx.ctx_header_size * offset;
sound/firewire/amdtp-stream.c
1455
count = header_length / s->ctx_data.tx.ctx_header_size;
sound/firewire/amdtp-stream.c
1492
ctx_header += s->ctx_data.tx.ctx_header_size / sizeof(__be32);
sound/firewire/amdtp-stream.c
1498
s->ctx_data.tx.event_starts = true;
sound/firewire/amdtp-stream.c
1509
if (s->ctx_data.tx.event_starts)
sound/firewire/amdtp-stream.c
1594
struct amdtp_stream *tx;
sound/firewire/amdtp-stream.c
1601
tx = rx->ctx_data.rx.replay_target;
sound/firewire/amdtp-stream.c
1602
cached_cycles = calculate_cached_cycle_count(tx, 0);
sound/firewire/amdtp-stream.c
1603
if (cached_cycles > tx->ctx_data.tx.cache.size / 2)
sound/firewire/amdtp-stream.c
1728
s->ctx_data.tx.max_ctx_payload_length = max_ctx_payload_size;
sound/firewire/amdtp-stream.c
1729
s->ctx_data.tx.ctx_header_size = ctx_header_size;
sound/firewire/amdtp-stream.c
1730
s->ctx_data.tx.event_starts = false;
sound/firewire/amdtp-stream.c
1735
s->ctx_data.tx.cache.size = max_t(unsigned int, s->syt_interval * 2,
sound/firewire/amdtp-stream.c
1737
s->ctx_data.tx.cache.pos = 0;
sound/firewire/amdtp-stream.c
1738
s->ctx_data.tx.cache.descs = kzalloc_objs(*s->ctx_data.tx.cache.descs,
sound/firewire/amdtp-stream.c
1739
s->ctx_data.tx.cache.size);
sound/firewire/amdtp-stream.c
1740
if (!s->ctx_data.tx.cache.descs) {
sound/firewire/amdtp-stream.c
1842
kfree(s->ctx_data.tx.cache.descs);
sound/firewire/amdtp-stream.c
1934
kfree(s->ctx_data.tx.cache.descs);
sound/firewire/amdtp-stream.c
2018
struct amdtp_stream *tx = NULL;
sound/firewire/amdtp-stream.c
2024
tx = s;
sound/firewire/amdtp-stream.c
2031
if (!tx) {
sound/firewire/amdtp-stream.c
2035
tx = s;
sound/firewire/amdtp-stream.c
2040
if (!tx)
sound/firewire/amdtp-stream.c
2044
rx->ctx_data.rx.replay_target = tx;
sound/firewire/amdtp-stream.c
496
const unsigned int cache_size = s->ctx_data.tx.cache.size;
sound/firewire/amdtp-stream.c
497
unsigned int cycles = s->ctx_data.tx.cache.pos;
sound/firewire/amdtp-stream.c
509
const unsigned int cache_size = s->ctx_data.tx.cache.size;
sound/firewire/amdtp-stream.c
510
struct seq_desc *cache = s->ctx_data.tx.cache.descs;
sound/firewire/amdtp-stream.c
511
unsigned int cache_pos = s->ctx_data.tx.cache.pos;
sound/firewire/amdtp-stream.c
528
s->ctx_data.tx.cache.pos = cache_pos;
sound/firewire/amdtp-stream.c
546
const struct seq_desc *cache = target->ctx_data.tx.cache.descs;
sound/firewire/amdtp-stream.c
547
const unsigned int cache_size = target->ctx_data.tx.cache.size;
sound/firewire/amdtp-stream.c
573
struct amdtp_stream *tx = s->ctx_data.rx.replay_target;
sound/firewire/amdtp-stream.c
574
const unsigned int cache_size = tx->ctx_data.tx.cache.size;
sound/firewire/amdtp-stream.c
576
unsigned int cached_cycles = calculate_cached_cycle_count(tx, cache_pos);
sound/firewire/amdtp-stream.c
671
params->header_length = s->ctx_data.tx.ctx_header_size;
sound/firewire/amdtp-stream.c
672
params->payload_length = s->ctx_data.tx.max_ctx_payload_length;
sound/firewire/amdtp-stream.c
787
if (*data_blocks > 0 && s->ctx_data.tx.dbc_interval > 0)
sound/firewire/amdtp-stream.c
788
dbc_interval = s->ctx_data.tx.dbc_interval;
sound/firewire/amdtp-stream.c
831
if (payload_length > cip_header_size + s->ctx_data.tx.max_ctx_payload_length) {
sound/firewire/amdtp-stream.c
834
payload_length, cip_header_size + s->ctx_data.tx.max_ctx_payload_length);
sound/firewire/amdtp-stream.c
994
ctx_header += s->ctx_data.tx.ctx_header_size / sizeof(*ctx_header);
sound/firewire/amdtp-stream.h
156
} tx;
sound/firewire/dice/dice-proc.c
172
quadlets = min_t(u32, tx_rx_header.size, sizeof(buf.tx) / 4);
sound/firewire/dice/dice-proc.c
174
if (dice_proc_read_mem(dice, &buf.tx, sections[2] + 2 +
sound/firewire/dice/dice-proc.c
179
snd_iprintf(buffer, " iso channel: %d\n", (int)buf.tx.iso);
sound/firewire/dice/dice-proc.c
181
buf.tx.number_audio);
sound/firewire/dice/dice-proc.c
182
snd_iprintf(buffer, " midi ports: %u\n", buf.tx.number_midi);
sound/firewire/dice/dice-proc.c
183
snd_iprintf(buffer, " speed: S%u\n", 100u << buf.tx.speed);
sound/firewire/dice/dice-proc.c
185
dice_proc_fixup_string(buf.tx.names, TX_NAMES_SIZE);
sound/firewire/dice/dice-proc.c
186
snd_iprintf(buffer, " names: %s\n", buf.tx.names);
sound/firewire/dice/dice-proc.c
190
buf.tx.ac3_caps);
sound/firewire/dice/dice-proc.c
192
buf.tx.ac3_enable);
sound/firewire/dice/dice-proc.c
98
} tx;
sound/firewire/fireworks/fireworks_stream.c
59
efw->tx_stream.ctx_data.tx.dbc_interval = 8;
sound/soc/bcm/bcm2835-i2s.c
154
bool tx, bool rx)
sound/soc/bcm/bcm2835-i2s.c
164
off = tx ? BCM2835_I2S_TXON : 0;
sound/soc/bcm/bcm2835-i2s.c
167
clr = tx ? BCM2835_I2S_TXCLR : 0;
sound/soc/codecs/cs42xx8.c
265
bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK;
sound/soc/codecs/cs42xx8.c
272
if (tx)
sound/soc/codecs/cs42xx8.c
275
rate[tx] = params_rate(params);
sound/soc/codecs/cs42xx8.c
276
rate[!tx] = cs42xx8->rate[!tx];
sound/soc/codecs/cs42xx8.c
278
ratio[tx] = rate[tx] > 0 ? cs42xx8->sysclk / rate[tx] : 0;
sound/soc/codecs/cs42xx8.c
279
ratio[!tx] = rate[!tx] > 0 ? cs42xx8->sysclk / rate[!tx] : 0;
sound/soc/codecs/cs42xx8.c
302
condition1 = ((fm[tx] == CS42XX8_FM_AUTO) ?
sound/soc/codecs/cs42xx8.c
303
(cs42xx8_ratios[i].ratio[0] == ratio[tx] ||
sound/soc/codecs/cs42xx8.c
304
cs42xx8_ratios[i].ratio[1] == ratio[tx] ||
sound/soc/codecs/cs42xx8.c
305
cs42xx8_ratios[i].ratio[2] == ratio[tx]) :
sound/soc/codecs/cs42xx8.c
306
(cs42xx8_ratios[i].ratio[fm[tx]] == ratio[tx])) &&
sound/soc/codecs/cs42xx8.c
310
if (!ratio[tx])
sound/soc/codecs/cs42xx8.c
314
condition2 = ((fm[!tx] == CS42XX8_FM_AUTO) ?
sound/soc/codecs/cs42xx8.c
315
(cs42xx8_ratios[i].ratio[0] == ratio[!tx] ||
sound/soc/codecs/cs42xx8.c
316
cs42xx8_ratios[i].ratio[1] == ratio[!tx] ||
sound/soc/codecs/cs42xx8.c
317
cs42xx8_ratios[i].ratio[2] == ratio[!tx]) :
sound/soc/codecs/cs42xx8.c
318
(cs42xx8_ratios[i].ratio[fm[!tx]] == ratio[!tx]));
sound/soc/codecs/cs42xx8.c
320
if (!ratio[!tx])
sound/soc/codecs/cs42xx8.c
336
cs42xx8->rate[tx] = params_rate(params);
sound/soc/codecs/cs42xx8.c
342
CS42XX8_FUNCMOD_xC_FM_MASK(tx) | mask,
sound/soc/codecs/cs42xx8.c
343
CS42XX8_FUNCMOD_xC_FM(tx, fm[tx]) | val);
sound/soc/codecs/cs42xx8.c
353
bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK;
sound/soc/codecs/cs42xx8.c
356
cs42xx8->rate[tx] = 0;
sound/soc/codecs/cs42xx8.c
359
CS42XX8_FUNCMOD_xC_FM_MASK(tx),
sound/soc/codecs/cs42xx8.c
360
CS42XX8_FUNCMOD_xC_FM(tx, CS42XX8_FM_AUTO));
sound/soc/codecs/hdmi-codec.c
447
bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK;
sound/soc/codecs/hdmi-codec.c
452
if (!((has_playback && tx) || (has_capture && !tx)))
sound/soc/codecs/hdmi-codec.c
468
if (tx && hcp->hcd.ops->get_eld) {
sound/soc/codecs/hdmi-codec.c
496
bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK;
sound/soc/codecs/hdmi-codec.c
500
if (!((has_playback && tx) || (has_capture && !tx)))
sound/soc/codecs/lpass-tx-macro.c
1005
if (tx->bcs_enable) {
sound/soc/codecs/lpass-tx-macro.c
1008
tx->bcs_clk_en = true;
sound/soc/codecs/lpass-tx-macro.c
1013
tx->tx_hpf_work[decimator].hpf_cut_off_freq;
sound/soc/codecs/lpass-tx-macro.c
1017
&tx->tx_hpf_work[decimator].dwork)) {
sound/soc/codecs/lpass-tx-macro.c
1023
if (is_amic_enabled(component, tx, decimator))
sound/soc/codecs/lpass-tx-macro.c
1047
cancel_delayed_work_sync(&tx->tx_mute_dwork[decimator].dwork);
sound/soc/codecs/lpass-tx-macro.c
1056
if (tx->bcs_enable) {
sound/soc/codecs/lpass-tx-macro.c
1063
tx->bcs_clk_en = false;
sound/soc/codecs/lpass-tx-macro.c
1074
struct tx_macro *tx = snd_soc_component_get_drvdata(component);
sound/soc/codecs/lpass-tx-macro.c
1078
ucontrol->value.integer.value[0] = tx->dec_mode[path];
sound/soc/codecs/lpass-tx-macro.c
1090
struct tx_macro *tx = snd_soc_component_get_drvdata(component);
sound/soc/codecs/lpass-tx-macro.c
1092
if (tx->dec_mode[path] == value)
sound/soc/codecs/lpass-tx-macro.c
1095
tx->dec_mode[path] = value;
sound/soc/codecs/lpass-tx-macro.c
1104
struct tx_macro *tx = snd_soc_component_get_drvdata(component);
sound/soc/codecs/lpass-tx-macro.c
1106
ucontrol->value.integer.value[0] = tx->bcs_enable;
sound/soc/codecs/lpass-tx-macro.c
1116
struct tx_macro *tx = snd_soc_component_get_drvdata(component);
sound/soc/codecs/lpass-tx-macro.c
1118
tx->bcs_enable = value;
sound/soc/codecs/lpass-tx-macro.c
1131
struct tx_macro *tx = snd_soc_component_get_drvdata(component);
sound/soc/codecs/lpass-tx-macro.c
1162
for_each_set_bit(decimator, &tx->active_ch_mask[dai->id], TX_MACRO_DEC_MAX)
sound/soc/codecs/lpass-tx-macro.c
1174
struct tx_macro *tx = snd_soc_component_get_drvdata(component);
sound/soc/codecs/lpass-tx-macro.c
1180
*tx_slot = tx->active_ch_mask[dai->id];
sound/soc/codecs/lpass-tx-macro.c
1181
*tx_num = tx->active_ch_cnt[dai->id];
sound/soc/codecs/lpass-tx-macro.c
1192
struct tx_macro *tx = snd_soc_component_get_drvdata(component);
sound/soc/codecs/lpass-tx-macro.c
1196
if (tx->active_decimator[dai->id] == -1)
sound/soc/codecs/lpass-tx-macro.c
1199
decimator = tx->active_decimator[dai->id];
sound/soc/codecs/lpass-tx-macro.c
2085
struct tx_macro *tx = snd_soc_component_get_drvdata(comp);
sound/soc/codecs/lpass-tx-macro.c
2088
if (tx->data->extra_widgets_num) {
sound/soc/codecs/lpass-tx-macro.c
2089
ret = snd_soc_dapm_new_controls(dapm, tx->data->extra_widgets,
sound/soc/codecs/lpass-tx-macro.c
2090
tx->data->extra_widgets_num);
sound/soc/codecs/lpass-tx-macro.c
2092
dev_err(tx->dev, "failed to add extra widgets: %d\n", ret);
sound/soc/codecs/lpass-tx-macro.c
2097
if (tx->data->extra_routes_num) {
sound/soc/codecs/lpass-tx-macro.c
2098
ret = snd_soc_dapm_add_routes(dapm, tx->data->extra_routes,
sound/soc/codecs/lpass-tx-macro.c
2099
tx->data->extra_routes_num);
sound/soc/codecs/lpass-tx-macro.c
2101
dev_err(tx->dev, "failed to add extra routes: %d\n", ret);
sound/soc/codecs/lpass-tx-macro.c
2111
struct tx_macro *tx = snd_soc_component_get_drvdata(comp);
sound/soc/codecs/lpass-tx-macro.c
2118
snd_soc_component_init_regmap(comp, tx->regmap);
sound/soc/codecs/lpass-tx-macro.c
2121
tx->tx_hpf_work[i].tx = tx;
sound/soc/codecs/lpass-tx-macro.c
2122
tx->tx_hpf_work[i].decimator = i;
sound/soc/codecs/lpass-tx-macro.c
2123
INIT_DELAYED_WORK(&tx->tx_hpf_work[i].dwork,
sound/soc/codecs/lpass-tx-macro.c
2128
tx->tx_mute_dwork[i].tx = tx;
sound/soc/codecs/lpass-tx-macro.c
2129
tx->tx_mute_dwork[i].decimator = i;
sound/soc/codecs/lpass-tx-macro.c
2130
INIT_DELAYED_WORK(&tx->tx_mute_dwork[i].dwork,
sound/soc/codecs/lpass-tx-macro.c
2133
tx->component = comp;
sound/soc/codecs/lpass-tx-macro.c
2148
struct tx_macro *tx = to_tx_macro(hw);
sound/soc/codecs/lpass-tx-macro.c
2149
struct regmap *regmap = tx->regmap;
sound/soc/codecs/lpass-tx-macro.c
2152
ret = clk_prepare_enable(tx->mclk);
sound/soc/codecs/lpass-tx-macro.c
2154
dev_err(tx->dev, "failed to enable mclk\n");
sound/soc/codecs/lpass-tx-macro.c
2158
tx_macro_mclk_enable(tx, true);
sound/soc/codecs/lpass-tx-macro.c
2168
struct tx_macro *tx = to_tx_macro(hw);
sound/soc/codecs/lpass-tx-macro.c
2169
struct regmap *regmap = tx->regmap;
sound/soc/codecs/lpass-tx-macro.c
2174
tx_macro_mclk_enable(tx, false);
sound/soc/codecs/lpass-tx-macro.c
2175
clk_disable_unprepare(tx->mclk);
sound/soc/codecs/lpass-tx-macro.c
2180
struct tx_macro *tx = to_tx_macro(hw);
sound/soc/codecs/lpass-tx-macro.c
2183
regmap_read(tx->regmap, CDC_TX_CLK_RST_CTRL_SWR_CONTROL, &val);
sound/soc/codecs/lpass-tx-macro.c
2203
static int tx_macro_register_mclk_output(struct tx_macro *tx)
sound/soc/codecs/lpass-tx-macro.c
2205
struct device *dev = tx->dev;
sound/soc/codecs/lpass-tx-macro.c
2212
if (tx->npl)
sound/soc/codecs/lpass-tx-macro.c
2213
parent_clk_name = __clk_get_name(tx->npl);
sound/soc/codecs/lpass-tx-macro.c
2215
parent_clk_name = __clk_get_name(tx->mclk);
sound/soc/codecs/lpass-tx-macro.c
2222
tx->hw.init = &init;
sound/soc/codecs/lpass-tx-macro.c
2223
hw = &tx->hw;
sound/soc/codecs/lpass-tx-macro.c
2246
struct tx_macro *tx;
sound/soc/codecs/lpass-tx-macro.c
2250
tx = devm_kzalloc(dev, sizeof(*tx), GFP_KERNEL);
sound/soc/codecs/lpass-tx-macro.c
2251
if (!tx)
sound/soc/codecs/lpass-tx-macro.c
2254
tx->data = device_get_match_data(dev);
sound/soc/codecs/lpass-tx-macro.c
2256
tx->macro = devm_clk_get_optional(dev, "macro");
sound/soc/codecs/lpass-tx-macro.c
2257
if (IS_ERR(tx->macro))
sound/soc/codecs/lpass-tx-macro.c
2258
return dev_err_probe(dev, PTR_ERR(tx->macro), "unable to get macro clock\n");
sound/soc/codecs/lpass-tx-macro.c
2260
tx->dcodec = devm_clk_get_optional(dev, "dcodec");
sound/soc/codecs/lpass-tx-macro.c
2261
if (IS_ERR(tx->dcodec))
sound/soc/codecs/lpass-tx-macro.c
2262
return dev_err_probe(dev, PTR_ERR(tx->dcodec), "unable to get dcodec clock\n");
sound/soc/codecs/lpass-tx-macro.c
2264
tx->mclk = devm_clk_get(dev, "mclk");
sound/soc/codecs/lpass-tx-macro.c
2265
if (IS_ERR(tx->mclk))
sound/soc/codecs/lpass-tx-macro.c
2266
return dev_err_probe(dev, PTR_ERR(tx->mclk), "unable to get mclk clock\n");
sound/soc/codecs/lpass-tx-macro.c
2268
if (tx->data->flags & LPASS_MACRO_FLAG_HAS_NPL_CLOCK) {
sound/soc/codecs/lpass-tx-macro.c
2269
tx->npl = devm_clk_get(dev, "npl");
sound/soc/codecs/lpass-tx-macro.c
2270
if (IS_ERR(tx->npl))
sound/soc/codecs/lpass-tx-macro.c
2271
return dev_err_probe(dev, PTR_ERR(tx->npl), "unable to get npl clock\n");
sound/soc/codecs/lpass-tx-macro.c
2274
tx->fsgen = devm_clk_get(dev, "fsgen");
sound/soc/codecs/lpass-tx-macro.c
2275
if (IS_ERR(tx->fsgen))
sound/soc/codecs/lpass-tx-macro.c
2276
return dev_err_probe(dev, PTR_ERR(tx->fsgen), "unable to get fsgen clock\n");
sound/soc/codecs/lpass-tx-macro.c
2278
tx->pds = lpass_macro_pds_init(dev);
sound/soc/codecs/lpass-tx-macro.c
2279
if (IS_ERR(tx->pds))
sound/soc/codecs/lpass-tx-macro.c
2280
return PTR_ERR(tx->pds);
sound/soc/codecs/lpass-tx-macro.c
2302
tx->regmap = devm_regmap_init_mmio(dev, base, &tx_regmap_config);
sound/soc/codecs/lpass-tx-macro.c
2303
if (IS_ERR(tx->regmap)) {
sound/soc/codecs/lpass-tx-macro.c
2304
ret = PTR_ERR(tx->regmap);
sound/soc/codecs/lpass-tx-macro.c
2308
dev_set_drvdata(dev, tx);
sound/soc/codecs/lpass-tx-macro.c
2310
tx->dev = dev;
sound/soc/codecs/lpass-tx-macro.c
2313
tx->active_decimator[TX_MACRO_AIF1_CAP] = -1;
sound/soc/codecs/lpass-tx-macro.c
2314
tx->active_decimator[TX_MACRO_AIF2_CAP] = -1;
sound/soc/codecs/lpass-tx-macro.c
2315
tx->active_decimator[TX_MACRO_AIF3_CAP] = -1;
sound/soc/codecs/lpass-tx-macro.c
2318
clk_set_rate(tx->mclk, MCLK_FREQ);
sound/soc/codecs/lpass-tx-macro.c
2319
clk_set_rate(tx->npl, MCLK_FREQ);
sound/soc/codecs/lpass-tx-macro.c
2321
ret = clk_prepare_enable(tx->macro);
sound/soc/codecs/lpass-tx-macro.c
2325
ret = clk_prepare_enable(tx->dcodec);
sound/soc/codecs/lpass-tx-macro.c
2329
ret = clk_prepare_enable(tx->mclk);
sound/soc/codecs/lpass-tx-macro.c
2333
ret = clk_prepare_enable(tx->npl);
sound/soc/codecs/lpass-tx-macro.c
2337
ret = clk_prepare_enable(tx->fsgen);
sound/soc/codecs/lpass-tx-macro.c
2343
if (tx->data->flags & LPASS_MACRO_FLAG_RESET_SWR)
sound/soc/codecs/lpass-tx-macro.c
2344
regmap_update_bits(tx->regmap, CDC_TX_CLK_RST_CTRL_SWR_CONTROL,
sound/soc/codecs/lpass-tx-macro.c
2347
regmap_update_bits(tx->regmap, CDC_TX_CLK_RST_CTRL_SWR_CONTROL,
sound/soc/codecs/lpass-tx-macro.c
2351
if (tx->data->flags & LPASS_MACRO_FLAG_RESET_SWR)
sound/soc/codecs/lpass-tx-macro.c
2352
regmap_update_bits(tx->regmap, CDC_TX_CLK_RST_CTRL_SWR_CONTROL,
sound/soc/codecs/lpass-tx-macro.c
2367
ret = tx_macro_register_mclk_output(tx);
sound/soc/codecs/lpass-tx-macro.c
2374
clk_disable_unprepare(tx->fsgen);
sound/soc/codecs/lpass-tx-macro.c
2376
clk_disable_unprepare(tx->npl);
sound/soc/codecs/lpass-tx-macro.c
2378
clk_disable_unprepare(tx->mclk);
sound/soc/codecs/lpass-tx-macro.c
2380
clk_disable_unprepare(tx->dcodec);
sound/soc/codecs/lpass-tx-macro.c
2382
clk_disable_unprepare(tx->macro);
sound/soc/codecs/lpass-tx-macro.c
2384
lpass_macro_pds_exit(tx->pds);
sound/soc/codecs/lpass-tx-macro.c
2391
struct tx_macro *tx = dev_get_drvdata(&pdev->dev);
sound/soc/codecs/lpass-tx-macro.c
2393
clk_disable_unprepare(tx->macro);
sound/soc/codecs/lpass-tx-macro.c
2394
clk_disable_unprepare(tx->dcodec);
sound/soc/codecs/lpass-tx-macro.c
2395
clk_disable_unprepare(tx->mclk);
sound/soc/codecs/lpass-tx-macro.c
2396
clk_disable_unprepare(tx->npl);
sound/soc/codecs/lpass-tx-macro.c
2397
clk_disable_unprepare(tx->fsgen);
sound/soc/codecs/lpass-tx-macro.c
2399
lpass_macro_pds_exit(tx->pds);
sound/soc/codecs/lpass-tx-macro.c
2404
struct tx_macro *tx = dev_get_drvdata(dev);
sound/soc/codecs/lpass-tx-macro.c
2406
regcache_cache_only(tx->regmap, true);
sound/soc/codecs/lpass-tx-macro.c
2407
regcache_mark_dirty(tx->regmap);
sound/soc/codecs/lpass-tx-macro.c
2409
clk_disable_unprepare(tx->fsgen);
sound/soc/codecs/lpass-tx-macro.c
2410
clk_disable_unprepare(tx->npl);
sound/soc/codecs/lpass-tx-macro.c
2411
clk_disable_unprepare(tx->mclk);
sound/soc/codecs/lpass-tx-macro.c
2418
struct tx_macro *tx = dev_get_drvdata(dev);
sound/soc/codecs/lpass-tx-macro.c
2421
ret = clk_prepare_enable(tx->mclk);
sound/soc/codecs/lpass-tx-macro.c
2427
ret = clk_prepare_enable(tx->npl);
sound/soc/codecs/lpass-tx-macro.c
2433
ret = clk_prepare_enable(tx->fsgen);
sound/soc/codecs/lpass-tx-macro.c
2439
regcache_cache_only(tx->regmap, false);
sound/soc/codecs/lpass-tx-macro.c
2440
regcache_sync(tx->regmap);
sound/soc/codecs/lpass-tx-macro.c
2444
clk_disable_unprepare(tx->npl);
sound/soc/codecs/lpass-tx-macro.c
2446
clk_disable_unprepare(tx->mclk);
sound/soc/codecs/lpass-tx-macro.c
245
struct tx_macro *tx;
sound/soc/codecs/lpass-tx-macro.c
251
struct tx_macro *tx;
sound/soc/codecs/lpass-tx-macro.c
613
static int tx_macro_mclk_enable(struct tx_macro *tx,
sound/soc/codecs/lpass-tx-macro.c
616
struct regmap *regmap = tx->regmap;
sound/soc/codecs/lpass-tx-macro.c
619
if (tx->tx_mclk_users == 0) {
sound/soc/codecs/lpass-tx-macro.c
631
tx->tx_mclk_users++;
sound/soc/codecs/lpass-tx-macro.c
633
if (tx->tx_mclk_users <= 0) {
sound/soc/codecs/lpass-tx-macro.c
634
dev_err(tx->dev, "clock already disabled\n");
sound/soc/codecs/lpass-tx-macro.c
635
tx->tx_mclk_users = 0;
sound/soc/codecs/lpass-tx-macro.c
638
tx->tx_mclk_users--;
sound/soc/codecs/lpass-tx-macro.c
639
if (tx->tx_mclk_users == 0) {
sound/soc/codecs/lpass-tx-macro.c
651
struct tx_macro *tx, u8 decimator)
sound/soc/codecs/lpass-tx-macro.c
658
if (tx->data->ver > LPASS_VER_9_0_0)
sound/soc/codecs/lpass-tx-macro.c
676
struct tx_macro *tx;
sound/soc/codecs/lpass-tx-macro.c
683
tx = hpf_work->tx;
sound/soc/codecs/lpass-tx-macro.c
684
component = tx->component;
sound/soc/codecs/lpass-tx-macro.c
690
if (is_amic_enabled(component, tx, hpf_work->decimator)) {
sound/soc/codecs/lpass-tx-macro.c
720
struct tx_macro *tx;
sound/soc/codecs/lpass-tx-macro.c
726
tx = tx_mute_dwork->tx;
sound/soc/codecs/lpass-tx-macro.c
727
component = tx->component;
sound/soc/codecs/lpass-tx-macro.c
738
struct tx_macro *tx = snd_soc_component_get_drvdata(component);
sound/soc/codecs/lpass-tx-macro.c
742
tx_macro_mclk_enable(tx, true);
sound/soc/codecs/lpass-tx-macro.c
745
tx_macro_mclk_enable(tx, false);
sound/soc/codecs/lpass-tx-macro.c
756
struct tx_macro *tx, u16 mic_sel_reg,
sound/soc/codecs/lpass-tx-macro.c
778
struct tx_macro *tx, u16 mic_sel_reg,
sound/soc/codecs/lpass-tx-macro.c
806
struct tx_macro *tx = snd_soc_component_get_drvdata(component);
sound/soc/codecs/lpass-tx-macro.c
848
else if (tx->data->ver <= LPASS_VER_9_0_0)
sound/soc/codecs/lpass-tx-macro.c
849
tx_macro_update_smic_sel_v9(component, widget, tx,
sound/soc/codecs/lpass-tx-macro.c
852
tx_macro_update_smic_sel_v9_2(component, widget, tx,
sound/soc/codecs/lpass-tx-macro.c
867
struct tx_macro *tx = snd_soc_component_get_drvdata(component);
sound/soc/codecs/lpass-tx-macro.c
869
if (test_bit(dec_id, &tx->active_ch_mask[dai_id]))
sound/soc/codecs/lpass-tx-macro.c
887
struct tx_macro *tx = snd_soc_component_get_drvdata(component);
sound/soc/codecs/lpass-tx-macro.c
890
if (tx->active_decimator[dai_id] == dec_id)
sound/soc/codecs/lpass-tx-macro.c
893
set_bit(dec_id, &tx->active_ch_mask[dai_id]);
sound/soc/codecs/lpass-tx-macro.c
894
tx->active_ch_cnt[dai_id]++;
sound/soc/codecs/lpass-tx-macro.c
895
tx->active_decimator[dai_id] = dec_id;
sound/soc/codecs/lpass-tx-macro.c
897
if (tx->active_decimator[dai_id] == -1)
sound/soc/codecs/lpass-tx-macro.c
900
tx->active_ch_cnt[dai_id]--;
sound/soc/codecs/lpass-tx-macro.c
901
clear_bit(dec_id, &tx->active_ch_mask[dai_id]);
sound/soc/codecs/lpass-tx-macro.c
902
tx->active_decimator[dai_id] = -1;
sound/soc/codecs/lpass-tx-macro.c
920
struct tx_macro *tx = snd_soc_component_get_drvdata(component);
sound/soc/codecs/lpass-tx-macro.c
946
tx->dec_mode[decimator]);
sound/soc/codecs/lpass-tx-macro.c
954
if (!is_amic_enabled(component, tx, decimator)) {
sound/soc/codecs/lpass-tx-macro.c
962
tx->tx_hpf_work[decimator].hpf_cut_off_freq =
sound/soc/codecs/lpass-tx-macro.c
970
if (is_amic_enabled(component, tx, decimator)) {
sound/soc/codecs/lpass-tx-macro.c
976
&tx->tx_mute_dwork[decimator].dwork,
sound/soc/codecs/lpass-tx-macro.c
978
if (tx->tx_hpf_work[decimator].hpf_cut_off_freq != CF_MIN_3DB_150HZ) {
sound/soc/codecs/lpass-tx-macro.c
980
&tx->tx_hpf_work[decimator].dwork,
sound/soc/codecs/lpass-tx-macro.c
986
if (!is_amic_enabled(component, tx, decimator))
sound/soc/codecs/wcd9335.c
3913
bool tx;
sound/soc/codecs/wcd9335.c
3923
tx = (j >= 16);
sound/soc/codecs/wcd9335.c
3924
port_id = (tx ? j - 16 : j);
sound/soc/codecs/wcd9335.c
3928
if (!tx)
sound/soc/codecs/wcd9335.c
3947
__func__, (tx ? "TX" : "RX"), port_id, val);
sound/soc/codecs/wcd9335.c
3952
__func__, (tx ? "TX" : "RX"), port_id, val);
sound/soc/codecs/wcd9335.c
3956
if (!tx)
sound/soc/codecs/wcd934x.c
2232
bool tx;
sound/soc/codecs/wcd934x.c
2242
tx = false;
sound/soc/codecs/wcd934x.c
2246
tx = true;
sound/soc/codecs/wcd934x.c
2253
if (!tx)
sound/soc/codecs/wcd934x.c
2265
(tx ? "TX" : "RX"), port_id, val);
sound/soc/codecs/wcd934x.c
2270
(tx ? "TX" : "RX"), port_id, val);
sound/soc/codecs/wcd934x.c
2274
if (!tx)
sound/soc/codecs/wcd934x.c
2292
(tx ? "TX" : "RX"), port_id, val);
sound/soc/codecs/wm8960.c
836
bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK;
sound/soc/codecs/wm8960.c
868
if (tx) {
sound/soc/codecs/wm8960.c
881
wm8960->is_stream_in_use[tx] = true;
sound/soc/codecs/wm8960.c
883
if (!wm8960->is_stream_in_use[!tx])
sound/soc/codecs/wm8960.c
894
bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK;
sound/soc/codecs/wm8960.c
896
wm8960->is_stream_in_use[tx] = false;
sound/soc/codecs/zl38060.c
274
bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK;
sound/soc/codecs/zl38060.c
283
if (priv->is_stream_in_use[!tx])
sound/soc/codecs/zl38060.c
311
priv->is_stream_in_use[tx] = true;
sound/soc/codecs/zl38060.c
320
bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK;
sound/soc/codecs/zl38060.c
322
priv->is_stream_in_use[tx] = false;
sound/soc/fsl/fsl-asoc-card.c
176
bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK;
sound/soc/fsl/fsl-asoc-card.c
192
if (!cpu_priv->sysclk_freq[tx] && cpu_priv->sysclk_ratio[tx])
sound/soc/fsl/fsl-asoc-card.c
193
sysclk_freq = priv->sample_rate * cpu_priv->sysclk_ratio[tx];
sound/soc/fsl/fsl-asoc-card.c
195
sysclk_freq = cpu_priv->sysclk_freq[tx];
sound/soc/fsl/fsl-asoc-card.c
198
ret = snd_soc_dai_set_sysclk(snd_soc_rtd_to_cpu(rtd, 0), cpu_priv->sysclk_id[tx],
sound/soc/fsl/fsl-asoc-card.c
200
cpu_priv->sysclk_dir[tx]);
sound/soc/fsl/fsl_asrc_dma.c
134
bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK;
sound/soc/fsl/fsl_asrc_dma.c
151
u8 dir = tx ? OUT : IN;
sound/soc/fsl/fsl_asrc_dma.c
226
tmp_chan = dma_request_chan(dev_be, tx ? "tx" : "rx");
sound/soc/fsl/fsl_asrc_dma.c
295
if (tx && (be_peripheral_type == IMX_DMATYPE_SSI_DUAL ||
sound/soc/fsl/fsl_asrc_dma.c
298
if (!tx && (be_peripheral_type == IMX_DMATYPE_SSI_DUAL ||
sound/soc/fsl/fsl_asrc_dma.c
302
if (tx) {
sound/soc/fsl/fsl_asrc_dma.c
324
bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK;
sound/soc/fsl/fsl_asrc_dma.c
327
u8 dir = tx ? OUT : IN;
sound/soc/fsl/fsl_asrc_dma.c
345
bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK;
sound/soc/fsl/fsl_asrc_dma.c
353
u8 dir = tx ? OUT : IN;
sound/soc/fsl/fsl_esai.c
165
static int fsl_esai_divisor_cal(struct snd_soc_dai *dai, bool tx, u32 ratio,
sound/soc/fsl/fsl_esai.c
234
regmap_update_bits(esai_priv->regmap, REG_ESAI_xCCR(tx),
sound/soc/fsl/fsl_esai.c
243
regmap_update_bits(esai_priv->regmap, REG_ESAI_xCCR(tx),
sound/soc/fsl/fsl_esai.c
264
bool tx = (clk_id <= ESAI_HCKT_EXTAL || esai_priv->synchronous);
sound/soc/fsl/fsl_esai.c
272
in ? "in" : "out", tx ? 'T' : 'R');
sound/soc/fsl/fsl_esai.c
277
if (freq == esai_priv->hck_rate[tx] && dir == esai_priv->hck_dir[tx])
sound/soc/fsl/fsl_esai.c
281
esai_priv->sck_div[tx] = true;
sound/soc/fsl/fsl_esai.c
284
regmap_update_bits(esai_priv->regmap, REG_ESAI_xCCR(tx),
sound/soc/fsl/fsl_esai.c
323
tx ? 'T' : 'R');
sound/soc/fsl/fsl_esai.c
330
ecr |= tx ? ESAI_ECR_ETO : ESAI_ECR_ERO;
sound/soc/fsl/fsl_esai.c
335
tx ? 'T' : 'R');
sound/soc/fsl/fsl_esai.c
339
ret = fsl_esai_divisor_cal(dai, tx, ratio, false, 0);
sound/soc/fsl/fsl_esai.c
343
esai_priv->sck_div[tx] = false;
sound/soc/fsl/fsl_esai.c
346
esai_priv->hck_dir[tx] = dir;
sound/soc/fsl/fsl_esai.c
347
esai_priv->hck_rate[tx] = freq;
sound/soc/fsl/fsl_esai.c
350
tx ? ESAI_ECR_ETI | ESAI_ECR_ETO :
sound/soc/fsl/fsl_esai.c
362
static int fsl_esai_set_bclk(struct snd_soc_dai *dai, bool tx, u32 freq)
sound/soc/fsl/fsl_esai.c
365
u32 hck_rate = esai_priv->hck_rate[tx];
sound/soc/fsl/fsl_esai.c
370
if (esai_priv->consumer_mode || esai_priv->sck_rate[tx] == freq)
sound/soc/fsl/fsl_esai.c
383
tx ? 'T' : 'R');
sound/soc/fsl/fsl_esai.c
388
if (!esai_priv->sck_div[tx] && (ratio > 16 || ratio == 0)) {
sound/soc/fsl/fsl_esai.c
393
ret = fsl_esai_divisor_cal(dai, tx, ratio, true,
sound/soc/fsl/fsl_esai.c
394
esai_priv->sck_div[tx] ? 0 : ratio);
sound/soc/fsl/fsl_esai.c
399
esai_priv->sck_rate[tx] = freq;
sound/soc/fsl/fsl_esai.c
540
bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK;
sound/soc/fsl/fsl_esai.c
554
ret = fsl_esai_set_bclk(dai, esai_priv->synchronous || tx, bclk);
sound/soc/fsl/fsl_esai.c
561
regmap_update_bits(esai_priv->regmap, REG_ESAI_xCR(tx), mask, val);
sound/soc/fsl/fsl_esai.c
563
if (!tx && esai_priv->synchronous)
sound/soc/fsl/fsl_esai.c
567
regmap_update_bits(esai_priv->regmap, REG_ESAI_xCR(tx),
sound/soc/fsl/fsl_esai.c
571
regmap_update_bits(esai_priv->regmap, REG_ESAI_xFCR(tx),
sound/soc/fsl/fsl_esai.c
575
(tx ? ESAI_xFCR_TE_MASK | ESAI_xFCR_TIEN : ESAI_xFCR_RE_MASK);
sound/soc/fsl/fsl_esai.c
577
(tx ? ESAI_xFCR_TE(pins) | ESAI_xFCR_TIEN : ESAI_xFCR_RE(pins));
sound/soc/fsl/fsl_esai.c
579
regmap_update_bits(esai_priv->regmap, REG_ESAI_xFCR(tx), mask, val);
sound/soc/fsl/fsl_esai.c
581
if (tx)
sound/soc/fsl/fsl_esai.c
649
static void fsl_esai_trigger_start(struct fsl_esai *esai_priv, bool tx)
sound/soc/fsl/fsl_esai.c
651
u8 i, channels = esai_priv->channels[tx];
sound/soc/fsl/fsl_esai.c
655
regmap_update_bits(esai_priv->regmap, REG_ESAI_xFCR(tx),
sound/soc/fsl/fsl_esai.c
659
for (i = 0; tx && i < channels; i++)
sound/soc/fsl/fsl_esai.c
674
regmap_update_bits(esai_priv->regmap, REG_ESAI_xCR(tx),
sound/soc/fsl/fsl_esai.c
675
tx ? ESAI_xCR_TE_MASK : ESAI_xCR_RE_MASK,
sound/soc/fsl/fsl_esai.c
676
tx ? ESAI_xCR_TE(pins) : ESAI_xCR_RE(pins));
sound/soc/fsl/fsl_esai.c
677
mask = tx ? esai_priv->tx_mask : esai_priv->rx_mask;
sound/soc/fsl/fsl_esai.c
679
regmap_update_bits(esai_priv->regmap, REG_ESAI_xSMB(tx),
sound/soc/fsl/fsl_esai.c
681
regmap_update_bits(esai_priv->regmap, REG_ESAI_xSMA(tx),
sound/soc/fsl/fsl_esai.c
685
regmap_update_bits(esai_priv->regmap, REG_ESAI_xCR(tx),
sound/soc/fsl/fsl_esai.c
689
static void fsl_esai_trigger_stop(struct fsl_esai *esai_priv, bool tx)
sound/soc/fsl/fsl_esai.c
691
regmap_update_bits(esai_priv->regmap, REG_ESAI_xCR(tx),
sound/soc/fsl/fsl_esai.c
694
regmap_update_bits(esai_priv->regmap, REG_ESAI_xCR(tx),
sound/soc/fsl/fsl_esai.c
695
tx ? ESAI_xCR_TE_MASK : ESAI_xCR_RE_MASK, 0);
sound/soc/fsl/fsl_esai.c
696
regmap_update_bits(esai_priv->regmap, REG_ESAI_xSMA(tx),
sound/soc/fsl/fsl_esai.c
698
regmap_update_bits(esai_priv->regmap, REG_ESAI_xSMB(tx),
sound/soc/fsl/fsl_esai.c
702
regmap_update_bits(esai_priv->regmap, REG_ESAI_xFCR(tx),
sound/soc/fsl/fsl_esai.c
704
regmap_update_bits(esai_priv->regmap, REG_ESAI_xFCR(tx),
sound/soc/fsl/fsl_esai.c
711
bool tx = true, rx = false, enabled[2];
sound/soc/fsl/fsl_esai.c
719
enabled[tx] = tfcr & ESAI_xFCR_xFEN;
sound/soc/fsl/fsl_esai.c
723
fsl_esai_trigger_stop(esai_priv, tx);
sound/soc/fsl/fsl_esai.c
749
if (enabled[tx])
sound/soc/fsl/fsl_esai.c
750
fsl_esai_trigger_start(esai_priv, tx);
sound/soc/fsl/fsl_esai.c
761
bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK;
sound/soc/fsl/fsl_esai.c
764
esai_priv->channels[tx] = substream->runtime->channels;
sound/soc/fsl/fsl_esai.c
771
fsl_esai_trigger_start(esai_priv, tx);
sound/soc/fsl/fsl_esai.c
778
fsl_esai_trigger_stop(esai_priv, tx);
sound/soc/fsl/fsl_esai.h
22
#define REG_ESAI_xFCR(tx) (tx ? REG_ESAI_TFCR : REG_ESAI_RFCR)
sound/soc/fsl/fsl_esai.h
23
#define REG_ESAI_xFSR(tx) (tx ? REG_ESAI_TFSR : REG_ESAI_RFSR)
sound/soc/fsl/fsl_esai.h
41
#define REG_ESAI_xCR(tx) (tx ? REG_ESAI_TCR : REG_ESAI_RCR)
sound/soc/fsl/fsl_esai.h
42
#define REG_ESAI_xCCR(tx) (tx ? REG_ESAI_TCCR : REG_ESAI_RCCR)
sound/soc/fsl/fsl_esai.h
47
#define REG_ESAI_xSMA(tx) (tx ? REG_ESAI_TSMA : REG_ESAI_RSMA)
sound/soc/fsl/fsl_esai.h
48
#define REG_ESAI_xSMB(tx) (tx ? REG_ESAI_TSMB : REG_ESAI_RSMB)
sound/soc/fsl/fsl_sai.c
1344
u32 rx, tx, type;
sound/soc/fsl/fsl_sai.c
1388
ret = of_property_read_u32_index(np, propname, index++, &tx);
sound/soc/fsl/fsl_sai.c
1392
if ((rx & ~soc_dl) || (tx & ~soc_dl)) {
sound/soc/fsl/fsl_sai.c
1398
tx = tx & soc_dl;
sound/soc/fsl/fsl_sai.c
1407
cfg[i].pins[1] = hweight8(tx);
sound/soc/fsl/fsl_sai.c
1408
cfg[i].mask[1] = tx;
sound/soc/fsl/fsl_sai.c
1409
dl_mask = tx;
sound/soc/fsl/fsl_sai.c
1411
cfg[i].next_off[1] = fsl_sai_calc_dl_off(tx);
sound/soc/fsl/fsl_sai.c
171
bool tx = true;
sound/soc/fsl/fsl_sai.c
173
sai->slots[tx] = slots;
sound/soc/fsl/fsl_sai.c
174
sai->slot_width[tx] = slot_width;
sound/soc/fsl/fsl_sai.c
183
bool tx = false;
sound/soc/fsl/fsl_sai.c
185
sai->slots[tx] = slots;
sound/soc/fsl/fsl_sai.c
186
sai->slot_width[tx] = slot_width;
sound/soc/fsl/fsl_sai.c
221
int clk_id, unsigned int freq, bool tx)
sound/soc/fsl/fsl_sai.c
244
regmap_update_bits(sai->regmap, FSL_SAI_xCR2(tx, ofs),
sound/soc/fsl/fsl_sai.c
306
unsigned int fmt, bool tx)
sound/soc/fsl/fsl_sai.c
316
sai->is_dsp_mode[tx] = false;
sound/soc/fsl/fsl_sai.c
345
sai->is_dsp_mode[tx] = true;
sound/soc/fsl/fsl_sai.c
353
sai->is_dsp_mode[tx] = true;
sound/soc/fsl/fsl_sai.c
392
sai->is_consumer_mode[tx] = false;
sound/soc/fsl/fsl_sai.c
395
sai->is_consumer_mode[tx] = true;
sound/soc/fsl/fsl_sai.c
399
sai->is_consumer_mode[tx] = false;
sound/soc/fsl/fsl_sai.c
403
sai->is_consumer_mode[tx] = true;
sound/soc/fsl/fsl_sai.c
409
regmap_update_bits(sai->regmap, FSL_SAI_xCR2(tx, ofs),
sound/soc/fsl/fsl_sai.c
411
regmap_update_bits(sai->regmap, FSL_SAI_xCR4(tx, ofs),
sound/soc/fsl/fsl_sai.c
445
static int fsl_sai_set_bclk(struct snd_soc_dai *dai, bool tx, u32 freq)
sound/soc/fsl/fsl_sai.c
451
int adir = tx ? RX : TX;
sound/soc/fsl/fsl_sai.c
452
int dir = tx ? TX : RX;
sound/soc/fsl/fsl_sai.c
457
if (sai->is_consumer_mode[tx])
sound/soc/fsl/fsl_sai.c
498
sai->mclk_id[tx] = id;
sound/soc/fsl/fsl_sai.c
508
tx ? 'T' : 'R', freq);
sound/soc/fsl/fsl_sai.c
513
sai->mclk_id[tx], savediv, bestdiff);
sound/soc/fsl/fsl_sai.c
526
reg = FSL_SAI_xCR2(!tx, ofs);
sound/soc/fsl/fsl_sai.c
528
reg = FSL_SAI_xCR2(tx, ofs);
sound/soc/fsl/fsl_sai.c
533
FSL_SAI_CR2_MSEL(sai->mclk_id[tx]));
sound/soc/fsl/fsl_sai.c
540
regmap_update_bits(sai->regmap, FSL_SAI_xCR2(tx, ofs),
sound/soc/fsl/fsl_sai.c
543
regmap_update_bits(sai->regmap, FSL_SAI_xCR2(tx, ofs),
sound/soc/fsl/fsl_sai.c
560
bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK;
sound/soc/fsl/fsl_sai.c
571
int adir = tx ? RX : TX;
sound/soc/fsl/fsl_sai.c
576
if (sai->slot_width[tx])
sound/soc/fsl/fsl_sai.c
577
slot_width = sai->slot_width[tx];
sound/soc/fsl/fsl_sai.c
579
if (sai->slots[tx])
sound/soc/fsl/fsl_sai.c
580
slots = sai->slots[tx];
sound/soc/fsl/fsl_sai.c
596
if (dl_cfg[i].type == dl_type && dl_cfg[i].pins[tx] == pins) {
sound/soc/fsl/fsl_sai.c
602
if (hweight8(dl_cfg[dl_cfg_idx].mask[tx]) < pins) {
sound/soc/fsl/fsl_sai.c
620
if (!sai->is_consumer_mode[tx]) {
sound/soc/fsl/fsl_sai.c
621
ret = fsl_sai_set_bclk(cpu_dai, tx, bclk);
sound/soc/fsl/fsl_sai.c
627
ret = clk_prepare_enable(sai->mclk_clk[sai->mclk_id[tx]]);
sound/soc/fsl/fsl_sai.c
635
if (!sai->is_dsp_mode[tx] && !sai->is_pdm_mode)
sound/soc/fsl/fsl_sai.c
652
if (tx)
sound/soc/fsl/fsl_sai.c
662
regmap_update_bits(sai->regmap, FSL_SAI_xCR4(!tx, ofs),
sound/soc/fsl/fsl_sai.c
666
regmap_update_bits(sai->regmap, FSL_SAI_xCR5(!tx, ofs),
sound/soc/fsl/fsl_sai.c
679
if (hweight8(dl_cfg[dl_cfg_idx].mask[tx]) <= 1 || sai->is_multi_fifo_dma)
sound/soc/fsl/fsl_sai.c
680
regmap_update_bits(sai->regmap, FSL_SAI_xCR4(tx, ofs),
sound/soc/fsl/fsl_sai.c
683
regmap_update_bits(sai->regmap, FSL_SAI_xCR4(tx, ofs),
sound/soc/fsl/fsl_sai.c
686
dma_params = tx ? &sai->dma_params_tx : &sai->dma_params_rx;
sound/soc/fsl/fsl_sai.c
687
dma_params->addr = sai->res->start + FSL_SAI_xDR0(tx) +
sound/soc/fsl/fsl_sai.c
688
dl_cfg[dl_cfg_idx].start_off[tx] * 0x4;
sound/soc/fsl/fsl_sai.c
691
sai->audio_config[tx].words_per_fifo = min(slots, channels);
sound/soc/fsl/fsl_sai.c
692
if (tx) {
sound/soc/fsl/fsl_sai.c
693
sai->audio_config[tx].n_fifos_dst = pins;
sound/soc/fsl/fsl_sai.c
694
sai->audio_config[tx].stride_fifos_dst = dl_cfg[dl_cfg_idx].next_off[tx];
sound/soc/fsl/fsl_sai.c
696
sai->audio_config[tx].n_fifos_src = pins;
sound/soc/fsl/fsl_sai.c
697
sai->audio_config[tx].stride_fifos_src = dl_cfg[dl_cfg_idx].next_off[tx];
sound/soc/fsl/fsl_sai.c
699
dma_params->maxburst = sai->audio_config[tx].words_per_fifo * pins;
sound/soc/fsl/fsl_sai.c
700
dma_params->peripheral_config = &sai->audio_config[tx];
sound/soc/fsl/fsl_sai.c
701
dma_params->peripheral_size = sizeof(sai->audio_config[tx]);
sound/soc/fsl/fsl_sai.c
703
watermark = tx ? (sai->soc_data->fifo_depth - dma_params->maxburst) :
sound/soc/fsl/fsl_sai.c
705
regmap_update_bits(sai->regmap, FSL_SAI_xCR1(tx, ofs),
sound/soc/fsl/fsl_sai.c
713
if (hweight8(dl_cfg[dl_cfg_idx].mask[tx] & trce_mask) == pins)
sound/soc/fsl/fsl_sai.c
717
regmap_update_bits(sai->regmap, FSL_SAI_xCR3(tx, ofs),
sound/soc/fsl/fsl_sai.c
719
FSL_SAI_CR3_TRCE((dl_cfg[dl_cfg_idx].mask[tx] & trce_mask)));
sound/soc/fsl/fsl_sai.c
731
!sai->is_consumer_mode[tx])
sound/soc/fsl/fsl_sai.c
732
regmap_update_bits(sai->regmap, FSL_SAI_xCR4(tx, ofs),
sound/soc/fsl/fsl_sai.c
735
regmap_update_bits(sai->regmap, FSL_SAI_xCR4(tx, ofs),
sound/soc/fsl/fsl_sai.c
739
regmap_update_bits(sai->regmap, FSL_SAI_xCR5(tx, ofs),
sound/soc/fsl/fsl_sai.c
745
!sai->is_consumer_mode[tx])
sound/soc/fsl/fsl_sai.c
746
regmap_update_bits(sai->regmap, FSL_SAI_xCR4(tx, ofs),
sound/soc/fsl/fsl_sai.c
749
regmap_write(sai->regmap, FSL_SAI_xMR(tx),
sound/soc/fsl/fsl_sai.c
759
bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK;
sound/soc/fsl/fsl_sai.c
763
regmap_write(sai->regmap, FSL_SAI_xMR(tx), 0);
sound/soc/fsl/fsl_sai.c
765
regmap_update_bits(sai->regmap, FSL_SAI_xCR3(tx, ofs),
sound/soc/fsl/fsl_sai.c
768
if (!sai->is_consumer_mode[tx] &&
sound/soc/fsl/fsl_sai.c
770
clk_disable_unprepare(sai->mclk_clk[sai->mclk_id[tx]]);
sound/soc/fsl/fsl_sai.c
780
bool tx = dir == TX;
sound/soc/fsl/fsl_sai.c
788
regmap_update_bits(sai->regmap, FSL_SAI_xCSR(tx, ofs),
sound/soc/fsl/fsl_sai.c
794
regmap_read(sai->regmap, FSL_SAI_xCSR(tx, ofs), &xcsr);
sound/soc/fsl/fsl_sai.c
797
regmap_update_bits(sai->regmap, FSL_SAI_xCSR(tx, ofs),
sound/soc/fsl/fsl_sai.c
812
regmap_update_bits(sai->regmap, FSL_SAI_xCSR(tx, ofs), FSL_SAI_CSR_SR, FSL_SAI_CSR_SR);
sound/soc/fsl/fsl_sai.c
814
regmap_update_bits(sai->regmap, FSL_SAI_xCSR(tx, ofs), FSL_SAI_CSR_SR, 0);
sound/soc/fsl/fsl_sai.c
823
bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK;
sound/soc/fsl/fsl_sai.c
824
int adir = tx ? RX : TX;
sound/soc/fsl/fsl_sai.c
825
int dir = tx ? TX : RX;
sound/soc/fsl/fsl_sai.c
846
regmap_update_bits(sai->regmap, FSL_SAI_xCSR(tx, ofs),
sound/soc/fsl/fsl_sai.c
849
regmap_update_bits(sai->regmap, FSL_SAI_xCSR(tx, ofs),
sound/soc/fsl/fsl_sai.c
863
regmap_update_bits(sai->regmap, FSL_SAI_xCSR((!tx), ofs),
sound/soc/fsl/fsl_sai.c
866
regmap_update_bits(sai->regmap, FSL_SAI_xCSR(tx, ofs),
sound/soc/fsl/fsl_sai.c
872
regmap_update_bits(sai->regmap, FSL_SAI_xCSR(tx, ofs),
sound/soc/fsl/fsl_sai.c
874
regmap_update_bits(sai->regmap, FSL_SAI_xCSR(tx, ofs),
sound/soc/fsl/fsl_sai.c
878
regmap_read(sai->regmap, FSL_SAI_xCSR(!tx, ofs), &xcsr);
sound/soc/fsl/fsl_sai.c
908
bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK;
sound/soc/fsl/fsl_sai.c
918
tx ? sai->dma_params_tx.maxburst :
sound/soc/fsl/fsl_sai.c
921
if (sai->is_consumer_mode[tx])
sound/soc/fsl/fsl_sai.h
82
#define FSL_SAI_xCSR(tx, ofs) (tx ? FSL_SAI_TCSR(ofs) : FSL_SAI_RCSR(ofs))
sound/soc/fsl/fsl_sai.h
83
#define FSL_SAI_xCR1(tx, ofs) (tx ? FSL_SAI_TCR1(ofs) : FSL_SAI_RCR1(ofs))
sound/soc/fsl/fsl_sai.h
84
#define FSL_SAI_xCR2(tx, ofs) (tx ? FSL_SAI_TCR2(ofs) : FSL_SAI_RCR2(ofs))
sound/soc/fsl/fsl_sai.h
85
#define FSL_SAI_xCR3(tx, ofs) (tx ? FSL_SAI_TCR3(ofs) : FSL_SAI_RCR3(ofs))
sound/soc/fsl/fsl_sai.h
86
#define FSL_SAI_xCR4(tx, ofs) (tx ? FSL_SAI_TCR4(ofs) : FSL_SAI_RCR4(ofs))
sound/soc/fsl/fsl_sai.h
87
#define FSL_SAI_xCR5(tx, ofs) (tx ? FSL_SAI_TCR5(ofs) : FSL_SAI_RCR5(ofs))
sound/soc/fsl/fsl_sai.h
88
#define FSL_SAI_xDR0(tx) (tx ? FSL_SAI_TDR0 : FSL_SAI_RDR0)
sound/soc/fsl/fsl_sai.h
89
#define FSL_SAI_xFR0(tx) (tx ? FSL_SAI_TFR0 : FSL_SAI_RFR0)
sound/soc/fsl/fsl_sai.h
90
#define FSL_SAI_xMR(tx) (tx ? FSL_SAI_TMR : FSL_SAI_RMR)
sound/soc/fsl/fsl_spdif.c
36
#define SIE_INTR_FOR(tx) (tx ? INTR_FOR_PLAYBACK : INTR_FOR_CAPTURE)
sound/soc/fsl/fsl_spdif.c
740
bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK;
sound/soc/fsl/fsl_spdif.c
741
u32 intr = SIE_INTR_FOR(tx);
sound/soc/fsl/fsl_spdif.c
742
u32 dmaen = SCR_DMA_xX_EN(tx);
sound/soc/fsl/fsl_spdif.h
108
#define SCR_DMA_xX_EN(tx) (tx ? SCR_DMA_TX_EN : SCR_DMA_RX_EN)
sound/soc/fsl/fsl_ssi.c
1112
bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK;
sound/soc/fsl/fsl_ssi.c
1125
if (tx && fsl_ssi_is_ac97(ssi))
sound/soc/fsl/fsl_ssi.c
1127
fsl_ssi_config_enable(ssi, tx);
sound/soc/fsl/fsl_ssi.c
1133
fsl_ssi_config_disable(ssi, tx);
sound/soc/fsl/fsl_ssi.c
404
static void fsl_ssi_config_enable(struct fsl_ssi *ssi, bool tx)
sound/soc/fsl/fsl_ssi.c
407
int dir = tx ? TX : RX;
sound/soc/fsl/fsl_ssi.c
412
SSI_SOR_xX_CLR(tx), SSI_SOR_xX_CLR(tx));
sound/soc/fsl/fsl_ssi.c
449
if (ssi->use_dma && tx) {
sound/soc/fsl/fsl_ssi.c
507
static void fsl_ssi_config_disable(struct fsl_ssi *ssi, bool tx)
sound/soc/fsl/fsl_ssi.c
511
int adir = tx ? RX : TX;
sound/soc/fsl/fsl_ssi.c
512
int dir = tx ? TX : RX;
sound/soc/fsl/fsl_ssi.c
565
SSI_SOR_xX_CLR(tx), SSI_SOR_xX_CLR(tx));
sound/soc/fsl/fsl_ssi.c
683
bool tx2, tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK;
sound/soc/fsl/fsl_ssi.c
776
tx2 = tx || ssi->synchronous;
sound/soc/fsl/fsl_ssi.c
808
bool tx2, tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK;
sound/soc/fsl/fsl_ssi.c
862
tx2 = tx || ssi->synchronous;
sound/soc/fsl/fsl_ssi.h
250
#define SSI_SOR_xX_CLR(tx) ((tx) ? SSI_SOR_TX_CLR : SSI_SOR_RX_CLR)
sound/soc/fsl/fsl_ssi.h
33
#define REG_SSI_SxCR(tx) ((tx) ? REG_SSI_STCR : REG_SSI_SRCR)
sound/soc/fsl/fsl_ssi.h
38
#define REG_SSI_SxCCR(tx) ((tx) ? REG_SSI_STCCR : REG_SSI_SRCCR)
sound/soc/fsl/fsl_ssi.h
67
#define REG_SSI_SxMSK(tx) ((tx) ? REG_SSI_STMSK : REG_SSI_SRMSK)
sound/soc/fsl/fsl_xcvr.c
332
static int fsl_xcvr_en_phy_pll(struct fsl_xcvr *xcvr, u32 freq, bool tx)
sound/soc/fsl/fsl_xcvr.c
381
if (tx) { /* TX is enabled for SPDIF only */
sound/soc/fsl/fsl_xcvr.c
427
} else if (!tx) { /* SPDIF / ARC RX mode */
sound/soc/fsl/fsl_xcvr.c
500
bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK;
sound/soc/fsl/fsl_xcvr.c
508
if (xcvr->soc_data->spdif_only && tx) {
sound/soc/fsl/fsl_xcvr.c
519
if (tx) {
sound/soc/fsl/fsl_xcvr.c
555
ret = fsl_xcvr_en_phy_pll(xcvr, FSL_XCVR_SPDIF_RX_FREQ, tx);
sound/soc/fsl/fsl_xcvr.c
564
if (!tx) {
sound/soc/fsl/fsl_xcvr.c
585
m_ctl |= FSL_XCVR_EXT_CTRL_CMDC_RESET(tx);
sound/soc/fsl/fsl_xcvr.c
588
v_ctl |= (tx ? FSL_XCVR_EXT_CTRL_TX_RX_MODE : 0);
sound/soc/fsl/fsl_xcvr.c
625
bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK;
sound/soc/fsl/fsl_xcvr.c
629
dev_err(dai->dev, "%sX busy\n", tx ? "T" : "R");
sound/soc/fsl/fsl_xcvr.c
640
tx ? xcvr->dma_prms_tx.maxburst :
sound/soc/fsl/fsl_xcvr.c
646
if (xcvr->soc_data->spdif_only && tx)
sound/soc/fsl/fsl_xcvr.c
681
bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK;
sound/soc/fsl/fsl_xcvr.c
714
mask |= FSL_XCVR_EXT_CTRL_CMDC_RESET(tx);
sound/soc/fsl/fsl_xcvr.c
715
val |= FSL_XCVR_EXT_CTRL_CMDC_RESET(tx);
sound/soc/fsl/fsl_xcvr.c
729
bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK;
sound/soc/fsl/fsl_xcvr.c
741
FSL_XCVR_EXT_CTRL_DPTH_RESET(tx),
sound/soc/fsl/fsl_xcvr.c
742
FSL_XCVR_EXT_CTRL_DPTH_RESET(tx));
sound/soc/fsl/fsl_xcvr.c
748
if (tx) {
sound/soc/fsl/fsl_xcvr.c
774
FSL_XCVR_EXT_CTRL_DMA_DIS(tx), 0);
sound/soc/fsl/fsl_xcvr.c
789
FSL_XCVR_EXT_CTRL_DPTH_RESET(tx),
sound/soc/fsl/fsl_xcvr.c
802
FSL_XCVR_EXT_CTRL_DMA_DIS(tx),
sound/soc/fsl/fsl_xcvr.c
803
FSL_XCVR_EXT_CTRL_DMA_DIS(tx));
sound/soc/fsl/fsl_xcvr.c
816
if (tx) {
sound/soc/fsl/imx-audmix.c
53
bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK;
sound/soc/fsl/imx-audmix.c
59
fmt |= tx ? SND_SOC_DAIFMT_BP_FP : SND_SOC_DAIFMT_BC_FC;
sound/soc/fsl/imx-audmix.c
60
dir = tx ? SND_SOC_CLOCK_OUT : SND_SOC_CLOCK_IN;
sound/soc/fsl/imx-audmix.c
92
bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK;
sound/soc/fsl/imx-audmix.c
96
if (!tx)
sound/soc/fsl/imx-hdmi.c
37
bool tx = substream->stream == SNDRV_PCM_STREAM_PLAYBACK;
sound/soc/fsl/imx-hdmi.c
45
ret = snd_soc_dai_set_sysclk(cpu_dai, data->cpu_priv.sysclk_id[tx],
sound/soc/fsl/imx-hdmi.c
47
tx ? SND_SOC_CLOCK_OUT : SND_SOC_CLOCK_IN);
sound/soc/intel/atom/sst-atom-controls.c
166
unsigned int is_tx = e->tx;
sound/soc/intel/atom/sst-atom-controls.c
226
unsigned int is_tx = e->tx;
sound/soc/intel/atom/sst-atom-controls.h
838
bool tx;
sound/soc/intel/atom/sst-atom-controls.h
847
(struct sst_enum){ .reg = s_ch_no, .tx = is_tx, .max = 4+1, .texts = xtexts, }
sound/soc/intel/avs/apl.c
225
bool avs_apl_d0ix_toggle(struct avs_dev *adev, struct avs_ipc_msg *tx, bool wake)
sound/soc/intel/avs/avs.h
286
bool avs_apl_d0ix_toggle(struct avs_dev *adev, struct avs_ipc_msg *tx, bool wake);
sound/soc/intel/avs/avs.h
287
bool avs_icl_d0ix_toggle(struct avs_dev *adev, struct avs_ipc_msg *tx, bool wake);
sound/soc/intel/avs/icl.c
114
bool avs_icl_d0ix_toggle(struct avs_dev *adev, struct avs_ipc_msg *tx, bool wake)
sound/soc/intel/avs/icl.c
117
if (tx->glb.set_ppl_state.state == AVS_PPL_STATE_RUNNING)
sound/soc/intel/avs/icl.c
121
return tx->size;
sound/soc/intel/avs/ipc.c
371
static void avs_dsp_send_tx(struct avs_dev *adev, struct avs_ipc_msg *tx, bool read_fwregs)
sound/soc/intel/avs/ipc.c
377
tx->header |= spec->hipc->req_busy_mask;
sound/soc/intel/avs/ipc.c
383
trace_avs_request(tx, sts, lec);
sound/soc/intel/avs/ipc.c
385
if (tx->size)
sound/soc/intel/avs/ipc.c
386
memcpy_toio(avs_downlink_addr(adev), tx->data, tx->size);
sound/soc/intel/avs/ipc.c
387
snd_hdac_adsp_writel(adev, spec->hipc->req_ext_offset, tx->header >> 32);
sound/soc/intel/avs/ipc.c
388
snd_hdac_adsp_writel(adev, spec->hipc->req_offset, tx->header & UINT_MAX);
sound/soc/intel/avs/ipc.c
45
static void avs_dsp_schedule_d0ix(struct avs_dev *adev, struct avs_ipc_msg *tx)
sound/soc/intel/avs/ipc.c
61
static int avs_dsp_wake_d0i0(struct avs_dev *adev, struct avs_ipc_msg *tx)
sound/soc/intel/avs/skl.c
153
static bool avs_skl_d0ix_toggle(struct avs_dev *adev, struct avs_ipc_msg *tx, bool wake)
sound/soc/intel/boards/sof_maxim_common.c
370
unsigned int tx;
sound/soc/intel/boards/sof_maxim_common.c
373
{.tx = 0x01, .rx = 0x3},
sound/soc/intel/boards/sof_maxim_common.c
374
{.tx = 0x02, .rx = 0x3},
sound/soc/intel/boards/sof_maxim_common.c
375
{.tx = 0x04, .rx = 0x3},
sound/soc/intel/boards/sof_maxim_common.c
376
{.tx = 0x08, .rx = 0x3},
sound/soc/intel/boards/sof_maxim_common.c
398
max_98390_tdm_mask[i].tx,
sound/soc/intel/boards/sof_realtek_common.c
127
unsigned int tx;
sound/soc/intel/boards/sof_realtek_common.c
130
{.tx = 0x4, .rx = 0x1},
sound/soc/intel/boards/sof_realtek_common.c
131
{.tx = 0x8, .rx = 0x2},
sound/soc/intel/boards/sof_realtek_common.c
132
{.tx = 0x1, .rx = 0x1},
sound/soc/intel/boards/sof_realtek_common.c
133
{.tx = 0x2, .rx = 0x2},
sound/soc/intel/boards/sof_realtek_common.c
169
ret = snd_soc_dai_set_tdm_slot(codec_dai, rt1011_tdm_mask[i].tx,
sound/soc/intel/boards/sof_realtek_common.c
388
unsigned int tx;
sound/soc/intel/boards/sof_realtek_common.c
391
{.tx = 0x0, .rx = 0x1},
sound/soc/intel/boards/sof_realtek_common.c
392
{.tx = 0x0, .rx = 0x2},
sound/soc/intel/boards/sof_realtek_common.c
435
rt1015_tdm_mask[i].tx,
sound/soc/intel/catpt/ipc.c
56
const struct catpt_ipc_msg *tx)
sound/soc/intel/catpt/ipc.c
58
u32 header = tx->header | CATPT_IPCC_BUSY;
sound/soc/intel/catpt/ipc.c
61
trace_catpt_ipc_payload(tx->data, tx->size);
sound/soc/intel/catpt/ipc.c
63
memcpy_toio(catpt_outbox_addr(cdev), tx->data, tx->size);
sound/soc/mediatek/common/mtk-btcvsd.c
1000
bt->tx->packet_r : bt->rx->packet_w;
sound/soc/mediatek/common/mtk-btcvsd.c
1076
bool lpbk_en = bt->tx->state == BT_SCO_STATE_LOOPBACK;
sound/soc/mediatek/common/mtk-btcvsd.c
1089
mtk_btcvsd_snd_set_state(bt, bt->tx, BT_SCO_STATE_LOOPBACK);
sound/soc/mediatek/common/mtk-btcvsd.c
1092
mtk_btcvsd_snd_set_state(bt, bt->tx, BT_SCO_STATE_RUNNING);
sound/soc/mediatek/common/mtk-btcvsd.c
1104
if (!bt->tx) {
sound/soc/mediatek/common/mtk-btcvsd.c
1109
ucontrol->value.integer.value[0] = bt->tx->mute;
sound/soc/mediatek/common/mtk-btcvsd.c
1119
if (!bt->tx)
sound/soc/mediatek/common/mtk-btcvsd.c
1122
bt->tx->mute = ucontrol->value.integer.value[0];
sound/soc/mediatek/common/mtk-btcvsd.c
1186
if (!bt->tx)
sound/soc/mediatek/common/mtk-btcvsd.c
1189
ucontrol->value.integer.value[0] = bt->tx->rw_cnt ? 1 : 0;
sound/soc/mediatek/common/mtk-btcvsd.c
1199
ucontrol->value.integer.value[0] = bt->tx->timeout;
sound/soc/mediatek/common/mtk-btcvsd.c
1293
btcvsd->tx = devm_kzalloc(btcvsd->dev, sizeof(*btcvsd->tx), GFP_KERNEL);
sound/soc/mediatek/common/mtk-btcvsd.c
1294
if (!btcvsd->tx)
sound/soc/mediatek/common/mtk-btcvsd.c
135
struct mtk_btcvsd_snd_stream *tx;
sound/soc/mediatek/common/mtk-btcvsd.c
1366
mtk_btcvsd_snd_set_state(btcvsd, btcvsd->tx, BT_SCO_STATE_IDLE);
sound/soc/mediatek/common/mtk-btcvsd.c
212
bt->tx->state, bt->rx->state, bt->irq_disabled);
sound/soc/mediatek/common/mtk-btcvsd.c
216
if (bt->tx->state == BT_SCO_STATE_IDLE &&
sound/soc/mediatek/common/mtk-btcvsd.c
234
memset(bt->tx, 0, sizeof(*bt->tx));
sound/soc/mediatek/common/mtk-btcvsd.c
237
bt->tx->packet_size = BTCVSD_TX_PACKET_SIZE;
sound/soc/mediatek/common/mtk-btcvsd.c
238
bt->tx->buf_size = BTCVSD_TX_BUF_SIZE;
sound/soc/mediatek/common/mtk-btcvsd.c
239
bt->tx->timeout = 0;
sound/soc/mediatek/common/mtk-btcvsd.c
240
bt->tx->rw_cnt = 0;
sound/soc/mediatek/common/mtk-btcvsd.c
241
bt->tx->stream = SNDRV_PCM_STREAM_PLAYBACK;
sound/soc/mediatek/common/mtk-btcvsd.c
261
ts->time_stamp_us = bt->tx->time_stamp;
sound/soc/mediatek/common/mtk-btcvsd.c
262
ts->data_count_equi_time = bt->tx->buf_data_equivalent_time;
sound/soc/mediatek/common/mtk-btcvsd.c
327
memset(bt->tx->temp_packet_buf, 170, SCO_PACKET_180);
sound/soc/mediatek/common/mtk-btcvsd.c
329
memcpy(bt->tx->temp_packet_buf,
sound/soc/mediatek/common/mtk-btcvsd.c
334
num_valid_addr = bt->tx->buffer_info.num_valid_addr;
sound/soc/mediatek/common/mtk-btcvsd.c
343
bt->tx->buffer_info.bt_sram_addr[i]);
sound/soc/mediatek/common/mtk-btcvsd.c
345
dst = (void *)bt->tx->buffer_info.bt_sram_addr[i];
sound/soc/mediatek/common/mtk-btcvsd.c
348
bt->tx->temp_packet_buf, dst,
sound/soc/mediatek/common/mtk-btcvsd.c
349
bt->tx->buffer_info.packet_length,
sound/soc/mediatek/common/mtk-btcvsd.c
350
bt->tx->buffer_info.packet_num);
sound/soc/mediatek/common/mtk-btcvsd.c
435
memcpy(bt->tx->temp_packet_buf + (bt->tx->packet_size * i),
sound/soc/mediatek/common/mtk-btcvsd.c
437
(bt->tx->packet_r % SCO_TX_PACKER_BUF_NUM) *
sound/soc/mediatek/common/mtk-btcvsd.c
438
bt->tx->packet_size),
sound/soc/mediatek/common/mtk-btcvsd.c
439
bt->tx->packet_size);
sound/soc/mediatek/common/mtk-btcvsd.c
441
bt->tx->packet_r++;
sound/soc/mediatek/common/mtk-btcvsd.c
447
if (!bt->tx->mute) {
sound/soc/mediatek/common/mtk-btcvsd.c
449
bt->tx->temp_packet_buf, dst,
sound/soc/mediatek/common/mtk-btcvsd.c
454
bt->tx->buffer_info.packet_length = packet_length;
sound/soc/mediatek/common/mtk-btcvsd.c
455
bt->tx->buffer_info.packet_num = packet_num;
sound/soc/mediatek/common/mtk-btcvsd.c
456
for (i = 0; i < bt->tx->buffer_info.num_valid_addr; i++) {
sound/soc/mediatek/common/mtk-btcvsd.c
457
if (bt->tx->buffer_info.bt_sram_addr[i] == ap_addr_tx) {
sound/soc/mediatek/common/mtk-btcvsd.c
466
bt->tx->buffer_info.num_valid_addr++;
sound/soc/mediatek/common/mtk-btcvsd.c
467
next_idx = bt->tx->buffer_info.num_valid_addr - 1;
sound/soc/mediatek/common/mtk-btcvsd.c
468
bt->tx->buffer_info.bt_sram_addr[next_idx] = ap_addr_tx;
sound/soc/mediatek/common/mtk-btcvsd.c
472
bt->tx->buffer_info.num_valid_addr);
sound/soc/mediatek/common/mtk-btcvsd.c
475
if (bt->tx->mute)
sound/soc/mediatek/common/mtk-btcvsd.c
489
bt->tx->state != BT_SCO_STATE_RUNNING &&
sound/soc/mediatek/common/mtk-btcvsd.c
490
bt->tx->state != BT_SCO_STATE_ENDING &&
sound/soc/mediatek/common/mtk-btcvsd.c
491
bt->tx->state != BT_SCO_STATE_LOOPBACK) {
sound/soc/mediatek/common/mtk-btcvsd.c
493
__func__, bt->rx->state, bt->tx->state);
sound/soc/mediatek/common/mtk-btcvsd.c
517
if (bt->tx->state == BT_SCO_STATE_LOOPBACK) {
sound/soc/mediatek/common/mtk-btcvsd.c
542
bt->tx->temp_packet_buf,
sound/soc/mediatek/common/mtk-btcvsd.c
546
bt->tx->temp_packet_buf, dst,
sound/soc/mediatek/common/mtk-btcvsd.c
550
bt->tx->rw_cnt++;
sound/soc/mediatek/common/mtk-btcvsd.c
585
bt->tx->timeout = 0;
sound/soc/mediatek/common/mtk-btcvsd.c
586
if ((bt->tx->state == BT_SCO_STATE_RUNNING ||
sound/soc/mediatek/common/mtk-btcvsd.c
587
bt->tx->state == BT_SCO_STATE_ENDING) &&
sound/soc/mediatek/common/mtk-btcvsd.c
588
bt->tx->trigger_start) {
sound/soc/mediatek/common/mtk-btcvsd.c
589
if (bt->tx->xrun) {
sound/soc/mediatek/common/mtk-btcvsd.c
593
if (bt->tx->packet_w - bt->tx->packet_r >=
sound/soc/mediatek/common/mtk-btcvsd.c
595
bt->tx->xrun = 0;
sound/soc/mediatek/common/mtk-btcvsd.c
601
if ((!bt->tx->xrun &&
sound/soc/mediatek/common/mtk-btcvsd.c
602
(bt->tx->packet_w - bt->tx->packet_r >= buf_cnt_tx)) ||
sound/soc/mediatek/common/mtk-btcvsd.c
603
bt->tx->state == BT_SCO_STATE_ENDING) {
sound/soc/mediatek/common/mtk-btcvsd.c
609
bt->tx->rw_cnt++;
sound/soc/mediatek/common/mtk-btcvsd.c
611
bt->tx->xrun = 1;
sound/soc/mediatek/common/mtk-btcvsd.c
624
if (bt->tx->state == BT_SCO_STATE_RUNNING ||
sound/soc/mediatek/common/mtk-btcvsd.c
625
bt->tx->state == BT_SCO_STATE_ENDING) {
sound/soc/mediatek/common/mtk-btcvsd.c
626
bt->tx->wait_flag = 1;
sound/soc/mediatek/common/mtk-btcvsd.c
628
snd_pcm_period_elapsed(bt->tx->substream);
sound/soc/mediatek/common/mtk-btcvsd.c
782
unsigned int packet_size = bt->tx->packet_size;
sound/soc/mediatek/common/mtk-btcvsd.c
788
bt->tx->time_stamp = sched_clock();
sound/soc/mediatek/common/mtk-btcvsd.c
789
bt->tx->buf_data_equivalent_time =
sound/soc/mediatek/common/mtk-btcvsd.c
790
(unsigned long long)(bt->tx->packet_w - bt->tx->packet_r) *
sound/soc/mediatek/common/mtk-btcvsd.c
794
bt->tx->buf_data_equivalent_time *= 1000;
sound/soc/mediatek/common/mtk-btcvsd.c
799
avail = bt->tx->buf_size -
sound/soc/mediatek/common/mtk-btcvsd.c
800
(bt->tx->packet_w - bt->tx->packet_r) * packet_size;
sound/soc/mediatek/common/mtk-btcvsd.c
802
cur_write_idx = (bt->tx->packet_w % SCO_TX_PACKER_BUF_NUM) *
sound/soc/mediatek/common/mtk-btcvsd.c
807
int ret = wait_for_bt_irq(bt, bt->tx);
sound/soc/mediatek/common/mtk-btcvsd.c
830
cont = bt->tx->buf_size - cur_write_idx;
sound/soc/mediatek/common/mtk-btcvsd.c
842
bt->tx->packet_w += write_size / packet_size;
sound/soc/mediatek/common/mtk-btcvsd.c
854
return bt->tx;
sound/soc/mediatek/common/mtk-btcvsd.c
884
bt->tx->substream = substream;
sound/soc/mediatek/common/mtk-btcvsd.c
913
params_buffer_bytes(hw_params) % bt->tx->packet_size != 0) {
sound/soc/mediatek/common/mtk-btcvsd.c
992
bt_stream = bt->tx;
sound/soc/meson/axg-card.c
15
u32 tx;
sound/soc/meson/axg-card.c
162
u32 tx, rx;
sound/soc/meson/axg-card.c
172
for (i = 0, tx = 0; i < AXG_TDM_NUM_LANES; i++) {
sound/soc/meson/axg-card.c
175
tx = max(tx, be->tx_mask[i]);
sound/soc/meson/axg-card.c
179
if (!tx)
sound/soc/meson/axg-card.c
193
if (!tx && !rx) {
sound/soc/meson/axg-card.c
204
be->slots = fls(max(tx, rx));
sound/soc/meson/axg-card.c
205
} else if (be->slots < fls(max(tx, rx)) || be->slots > 32) {
sound/soc/meson/axg-card.c
237
&codec_mask->tx);
sound/soc/meson/axg-card.c
65
be->codec_masks[i].tx,
sound/soc/meson/axg-tdm-interface.c
43
struct axg_tdm_stream *tx = snd_soc_dai_dma_data_get_playback(dai);
sound/soc/meson/axg-tdm-interface.c
84
if (tx) {
sound/soc/meson/axg-tdm-interface.c
85
tx->mask = tx_mask;
sound/soc/rockchip/rockchip_i2s_tdm.c
220
bool tx = clr & I2S_CLR_TXC;
sound/soc/rockchip/rockchip_i2s_tdm.c
223
if (!(rx || tx))
sound/soc/rockchip/rockchip_i2s_tdm.c
226
if (tx) {
sound/soc/rockchip/rockchip_i2s_tdm.c
247
tx ? "tx" : "", rx ? "rx" : "");
sound/soc/rockchip/rockchip_i2s_tdm.c
248
if (rx && tx)
sound/soc/rockchip/rockchip_i2s_tdm.c
250
else if (tx)
sound/soc/samsung/dma.h
16
const char *tx, const char *rx,
sound/soc/samsung/dmaengine.c
18
const char *tx, const char *rx,
sound/soc/samsung/dmaengine.c
31
pcm_conf->chan_names[SNDRV_PCM_STREAM_PLAYBACK] = tx;
sound/soc/sof/ipc4.c
435
static bool sof_ipc4_tx_payload_for_get_data(struct sof_ipc4_msg *tx)
sound/soc/sof/ipc4.c
443
switch (tx->extension & SOF_IPC4_MOD_EXT_MSG_PARAM_ID_MASK) {
sound/soc/sof/ipc4.c
461
struct sof_ipc4_msg tx = {{ 0 }};
sound/soc/sof/ipc4.c
478
tx.primary = ipc4_msg->primary;
sound/soc/sof/ipc4.c
479
tx.extension = ipc4_msg->extension;
sound/soc/sof/ipc4.c
482
tx.primary |= SOF_IPC4_MSG_TYPE_SET(SOF_IPC4_MOD_LARGE_CONFIG_SET);
sound/soc/sof/ipc4.c
484
tx.primary |= SOF_IPC4_MSG_TYPE_SET(SOF_IPC4_MOD_LARGE_CONFIG_GET);
sound/soc/sof/ipc4.c
486
tx.extension &= ~SOF_IPC4_MOD_EXT_MSG_SIZE_MASK;
sound/soc/sof/ipc4.c
487
tx.extension |= SOF_IPC4_MOD_EXT_MSG_SIZE(payload_bytes);
sound/soc/sof/ipc4.c
489
tx.extension |= SOF_IPC4_MOD_EXT_MSG_FIRST_BLOCK(1);
sound/soc/sof/ipc4.c
491
if (sof_ipc4_tx_payload_for_get_data(&tx)) {
sound/soc/sof/ipc4.c
517
tx.extension |= SOF_IPC4_MOD_EXT_MSG_LAST_BLOCK(1);
sound/soc/sof/ipc4.c
521
tx.extension &= ~SOF_IPC4_MOD_EXT_MSG_FIRST_BLOCK_MASK;
sound/soc/sof/ipc4.c
522
tx.extension &= ~SOF_IPC4_MOD_EXT_MSG_SIZE_MASK;
sound/soc/sof/ipc4.c
523
tx.extension |= SOF_IPC4_MOD_EXT_MSG_SIZE(offset);
sound/soc/sof/ipc4.c
527
tx.data_size = chunk_size;
sound/soc/sof/ipc4.c
528
tx.data_ptr = ipc4_msg->data_ptr + offset;
sound/soc/sof/ipc4.c
540
tx.data_size = tx_size;
sound/soc/sof/ipc4.c
541
tx.data_ptr = tx_payload_for_get;
sound/soc/sof/ipc4.c
544
tx.data_size = 0;
sound/soc/sof/ipc4.c
545
tx.data_ptr = NULL;
sound/soc/sof/ipc4.c
551
ret = ipc4_tx_msg_unlocked(sdev->ipc, &tx, tx_size, &rx, rx_size);
sound/soc/starfive/jh7110_tdm.c
121
struct tdm_chan_cfg tx;
sound/soc/starfive/jh7110_tdm.c
191
sl = max(tdm->rx.sl, tdm->tx.sl);
sound/soc/starfive/jh7110_tdm.c
192
sscale = max(tdm->rx.sscale, tdm->tx.sscale);
sound/soc/starfive/jh7110_tdm.c
201
(tdm->rx.sscale <= 1 || tdm->tx.sscale <= 1) &&
sound/soc/starfive/jh7110_tdm.c
227
datatx = (tdm->tx.ifl << IFL_BIT) |
sound/soc/starfive/jh7110_tdm.c
228
(tdm->tx.wl << WL_BIT) |
sound/soc/starfive/jh7110_tdm.c
229
(tdm->tx.sscale << SSCALE_BIT) |
sound/soc/starfive/jh7110_tdm.c
230
(tdm->tx.sl << SL_BIT) |
sound/soc/starfive/jh7110_tdm.c
231
(tdm->tx.lrj << LRJ_BIT);
sound/soc/starfive/jh7110_tdm.c
377
tdm->tx.wl = chan_wl;
sound/soc/starfive/jh7110_tdm.c
378
tdm->tx.sl = chan_sl;
sound/soc/starfive/jh7110_tdm.c
379
tdm->tx.sscale = chan_nr;
sound/soc/starfive/jh7110_tdm.c
530
tdm->tx.ifl = TDM_FIFO_HALF;
sound/soc/starfive/jh7110_tdm.c
532
tdm->tx.wl = TDM_16BIT_WORD_LEN;
sound/soc/starfive/jh7110_tdm.c
534
tdm->tx.sscale = 2;
sound/soc/starfive/jh7110_tdm.c
536
tdm->tx.lrj = TDM_LEFT_JUSTIFT;
sound/soc/ti/omap-mcbsp.c
416
int tx = (stream == SNDRV_PCM_STREAM_PLAYBACK);
sound/soc/ti/omap-mcbsp.c
417
int rx = !tx;
sound/soc/ti/omap-mcbsp.c
437
tx &= 1;
sound/soc/ti/omap-mcbsp.c
439
MCBSP_WRITE(mcbsp, SPCR2, w | tx);
sound/soc/ti/omap-mcbsp.c
462
w &= ~(tx ? XDISABLE : 0);
sound/soc/ti/omap-mcbsp.c
475
int tx = (stream == SNDRV_PCM_STREAM_PLAYBACK);
sound/soc/ti/omap-mcbsp.c
476
int rx = !tx;
sound/soc/ti/omap-mcbsp.c
481
tx &= 1;
sound/soc/ti/omap-mcbsp.c
484
w |= (tx ? XDISABLE : 0);
sound/soc/ti/omap-mcbsp.c
488
MCBSP_WRITE(mcbsp, SPCR2, w & ~tx);
sound/soc/ti/omap-mcbsp.c
819
int tx = (substream->stream == SNDRV_PCM_STREAM_PLAYBACK);
sound/soc/ti/omap-mcbsp.c
820
int stream1 = tx ? SNDRV_PCM_STREAM_PLAYBACK : SNDRV_PCM_STREAM_CAPTURE;
sound/soc/ti/omap-mcbsp.c
821
int stream2 = tx ? SNDRV_PCM_STREAM_CAPTURE : SNDRV_PCM_STREAM_PLAYBACK;
sound/soc/ti/omap-mcbsp.c
842
int tx = (substream->stream == SNDRV_PCM_STREAM_PLAYBACK);
sound/soc/ti/omap-mcbsp.c
843
int stream1 = tx ? SNDRV_PCM_STREAM_PLAYBACK : SNDRV_PCM_STREAM_CAPTURE;
sound/soc/ti/omap-mcbsp.c
844
int stream2 = tx ? SNDRV_PCM_STREAM_CAPTURE : SNDRV_PCM_STREAM_PLAYBACK;
sound/soc/ti/omap-mcpdm.c
268
int tx = (substream->stream == SNDRV_PCM_STREAM_PLAYBACK);
sound/soc/ti/omap-mcpdm.c
269
int stream1 = tx ? SNDRV_PCM_STREAM_PLAYBACK : SNDRV_PCM_STREAM_CAPTURE;
sound/soc/ti/omap-mcpdm.c
270
int stream2 = tx ? SNDRV_PCM_STREAM_CAPTURE : SNDRV_PCM_STREAM_PLAYBACK;
sound/soc/ti/omap-mcpdm.c
380
int tx = (substream->stream == SNDRV_PCM_STREAM_PLAYBACK);
sound/soc/ti/omap-mcpdm.c
381
int stream1 = tx ? SNDRV_PCM_STREAM_PLAYBACK : SNDRV_PCM_STREAM_CAPTURE;
sound/soc/ti/omap-mcpdm.c
382
int stream2 = tx ? SNDRV_PCM_STREAM_CAPTURE : SNDRV_PCM_STREAM_PLAYBACK;
tools/include/uapi/linux/if_xdp.h
68
struct xdp_ring_offset tx;
tools/net/ynl/samples/ethtool.c
51
if (dev->_present.tx)
tools/net/ynl/samples/ethtool.c
52
printf("tx %d ", dev->tx);
tools/spi/spidev_test.c
119
static void transfer(int fd, uint8_t const *tx, uint8_t const *rx, size_t len)
tools/spi/spidev_test.c
124
.tx_buf = (unsigned long)tx,
tools/spi/spidev_test.c
157
hex_dump(tx, len, 32, "TX");
tools/spi/spidev_test.c
343
uint8_t *tx;
tools/spi/spidev_test.c
346
tx = malloc(size);
tools/spi/spidev_test.c
347
if (!tx)
tools/spi/spidev_test.c
354
size = unescape((char *)tx, str, size);
tools/spi/spidev_test.c
355
transfer(fd, tx, rx, size);
tools/spi/spidev_test.c
357
free(tx);
tools/spi/spidev_test.c
365
uint8_t *tx;
tools/spi/spidev_test.c
375
tx = malloc(sb.st_size);
tools/spi/spidev_test.c
376
if (!tx)
tools/spi/spidev_test.c
383
bytes = read(tx_fd, tx, sb.st_size);
tools/spi/spidev_test.c
387
transfer(fd, tx, rx, sb.st_size);
tools/spi/spidev_test.c
389
free(tx);
tools/spi/spidev_test.c
412
uint8_t *tx;
tools/spi/spidev_test.c
416
tx = malloc(len);
tools/spi/spidev_test.c
417
if (!tx)
tools/spi/spidev_test.c
420
tx[i] = random();
tools/spi/spidev_test.c
426
transfer(fd, tx, rx, len);
tools/spi/spidev_test.c
432
if (memcmp(tx, rx, len)) {
tools/spi/spidev_test.c
434
hex_dump(tx, len, 32, "TX");
tools/spi/spidev_test.c
441
free(tx);
tools/testing/selftests/bpf/prog_tests/flow_dissector_classification.c
449
int fdt = -1, fdr = -1, len, tx = 0, rx = 0, err;
tools/testing/selftests/bpf/prog_tests/flow_dissector_classification.c
467
while (tx < TEST_PACKETS_COUNT) {
tools/testing/selftests/bpf/prog_tests/flow_dissector_classification.c
470
tx++;
tools/testing/selftests/bpf/prog_tests/flow_dissector_classification.c
478
if (rx < tx) {
tools/testing/selftests/bpf/prog_tests/flow_dissector_classification.c
480
while (rx < tx) {
tools/testing/selftests/bpf/prog_tests/test_xsk.c
1191
while (xsk_ring_prod__reserve(&xsk->tx, xsk->batch_size, &idx) < xsk->batch_size) {
tools/testing/selftests/bpf/prog_tests/test_xsk.c
1224
xsk_ring_prod__cancel(&xsk->tx, xsk->batch_size - i);
tools/testing/selftests/bpf/prog_tests/test_xsk.c
1230
struct xdp_desc *tx_desc = xsk_ring_prod__tx_desc(&xsk->tx, idx + i);
tools/testing/selftests/bpf/prog_tests/test_xsk.c
1268
xsk_ring_prod__submit(&xsk->tx, i);
tools/testing/selftests/bpf/prog_tests/test_xsk.c
1479
struct xsk_umem_info *umem, bool tx)
tools/testing/selftests/bpf/prog_tests/test_xsk.c
1484
bool shared = (ifobject->shared_umem && tx) ? true : !!i;
tools/testing/selftests/bpf/prog_tests/test_xsk.c
177
txr = ifobject->tx_on ? &xsk->tx : NULL;
tools/testing/selftests/bpf/prog_tests/test_xsk.c
910
cons = load_value(xsk->tx.consumer);
tools/testing/selftests/bpf/prog_tests/test_xsk.c
911
ready_to_send = load_value(xsk->tx.producer) - cons;
tools/testing/selftests/bpf/prog_tests/test_xsk.c
914
delta = load_value(xsk->tx.consumer) - cons;
tools/testing/selftests/bpf/prog_tests/test_xsk.c
964
if (xsk_ring_prod__needs_wakeup(&xsk->tx)) {
tools/testing/selftests/bpf/prog_tests/test_xsk.h
64
struct xsk_ring_prod tx;
tools/testing/selftests/bpf/prog_tests/xdp_metadata.c
147
ret = xsk_ring_prod__reserve(&xsk->tx, 1, &idx);
tools/testing/selftests/bpf/prog_tests/xdp_metadata.c
151
tx_desc = xsk_ring_prod__tx_desc(&xsk->tx, idx);
tools/testing/selftests/bpf/prog_tests/xdp_metadata.c
194
xsk_ring_prod__submit(&xsk->tx, 1);
tools/testing/selftests/bpf/prog_tests/xdp_metadata.c
54
struct xsk_ring_prod tx;
tools/testing/selftests/bpf/prog_tests/xdp_metadata.c
95
&xsk->tx,
tools/testing/selftests/bpf/prog_tests/xsk.c
46
int configure_ifobj(struct ifobject *tx, struct ifobject *rx)
tools/testing/selftests/bpf/prog_tests/xsk.c
52
tx->ifindex = if_nametoindex(VETH_TX);
tools/testing/selftests/bpf/prog_tests/xsk.c
53
if (!ASSERT_OK_FD(tx->ifindex, "get TX ifindex"))
tools/testing/selftests/bpf/prog_tests/xsk.c
56
tx->shared_umem = false;
tools/testing/selftests/bpf/progs/xdp_features.c
135
xdp_update_stats(struct xdp_md *xdp, bool tx, bool dut)
tools/testing/selftests/bpf/progs/xdp_features.c
139
if (xdp_process_echo_packet(xdp, tx))
tools/testing/selftests/bpf/test_sockmap.c
630
struct msg_stats *s, bool tx,
tools/testing/selftests/bpf/test_sockmap.c
639
if (!tx && opt->check_recved_len)
tools/testing/selftests/bpf/test_sockmap.c
642
err = msg_alloc_iov(&msg, iov_count, iov_alloc_length, data, tx);
tools/testing/selftests/bpf/test_sockmap.c
646
err = msg_alloc_iov(&msg_peek, iov_count, iov_length, data, tx);
tools/testing/selftests/bpf/test_sockmap.c
651
if (tx) {
tools/testing/selftests/bpf/xdp_hw_metadata.c
127
&xsk->tx,
tools/testing/selftests/bpf/xdp_hw_metadata.c
366
ret = xsk_ring_prod__reserve(&xsk->tx, 1, &idx);
tools/testing/selftests/bpf/xdp_hw_metadata.c
372
tx_desc = xsk_ring_prod__tx_desc(&xsk->tx, idx);
tools/testing/selftests/bpf/xdp_hw_metadata.c
390
xsk_ring_prod__cancel(&xsk->tx, 1);
tools/testing/selftests/bpf/xdp_hw_metadata.c
441
xsk_ring_prod__submit(&xsk->tx, 1);
tools/testing/selftests/bpf/xdp_hw_metadata.c
63
struct xsk_ring_prod tx;
tools/testing/selftests/bpf/xsk.c
542
struct xsk_ring_prod *tx,
tools/testing/selftests/bpf/xsk.c
555
if (!umem || !xsk_ptr || !(rx || tx))
tools/testing/selftests/bpf/xsk.c
606
if (tx && !tx_setup_done) {
tools/testing/selftests/bpf/xsk.c
645
if (tx) {
tools/testing/selftests/bpf/xsk.c
646
tx_map = mmap(NULL, off.tx.desc +
tools/testing/selftests/bpf/xsk.c
655
tx->mask = xsk->config.tx_size - 1;
tools/testing/selftests/bpf/xsk.c
656
tx->size = xsk->config.tx_size;
tools/testing/selftests/bpf/xsk.c
657
tx->producer = tx_map + off.tx.producer;
tools/testing/selftests/bpf/xsk.c
658
tx->consumer = tx_map + off.tx.consumer;
tools/testing/selftests/bpf/xsk.c
659
tx->flags = tx_map + off.tx.flags;
tools/testing/selftests/bpf/xsk.c
660
tx->ring = tx_map + off.tx.desc;
tools/testing/selftests/bpf/xsk.c
661
tx->cached_prod = *tx->producer;
tools/testing/selftests/bpf/xsk.c
665
tx->cached_cons = *tx->consumer + xsk->config.tx_size;
tools/testing/selftests/bpf/xsk.c
667
xsk->tx = tx;
tools/testing/selftests/bpf/xsk.c
691
if (tx)
tools/testing/selftests/bpf/xsk.c
692
munmap(tx_map, off.tx.desc +
tools/testing/selftests/bpf/xsk.c
710
struct xsk_ring_cons *rx, struct xsk_ring_prod *tx,
tools/testing/selftests/bpf/xsk.c
717
rx, tx, umem->fill_save,
tools/testing/selftests/bpf/xsk.c
768
if (xsk->tx) {
tools/testing/selftests/bpf/xsk.c
769
munmap(xsk->tx->ring - off.tx.desc,
tools/testing/selftests/bpf/xsk.c
770
off.tx.desc + xsk->config.tx_size * desc_sz);
tools/testing/selftests/bpf/xsk.c
80
struct xsk_ring_prod *tx;
tools/testing/selftests/bpf/xsk.h
228
struct xsk_ring_prod *tx,
tools/testing/selftests/bpf/xsk.h
234
struct xsk_ring_prod *tx,
tools/testing/selftests/bpf/xsk.h
65
static inline struct xdp_desc *xsk_ring_prod__tx_desc(struct xsk_ring_prod *tx,
tools/testing/selftests/bpf/xsk.h
68
struct xdp_desc *descs = (struct xdp_desc *)tx->ring;
tools/testing/selftests/bpf/xsk.h
70
return &descs[idx & tx->mask];
tools/testing/selftests/drivers/net/hw/ncdevmem.c
555
static int check_changing_channels(unsigned int rx, unsigned int tx)
tools/testing/selftests/drivers/net/hw/ncdevmem.c
564
fprintf(stderr, "setting channel count rx:%u tx:%u\n", rx, tx);
tools/testing/selftests/drivers/net/hw/ncdevmem.c
601
if (rx == tx) {
tools/testing/selftests/drivers/net/hw/ncdevmem.c
603
} else if (rx > tx) {
tools/testing/selftests/drivers/net/hw/ncdevmem.c
604
ethtool_channels_set_req_set_combined_count(schan, tx);
tools/testing/selftests/drivers/net/hw/ncdevmem.c
605
ethtool_channels_set_req_set_rx_count(schan, rx - tx);
tools/testing/selftests/drivers/net/hw/ncdevmem.c
608
ethtool_channels_set_req_set_tx_count(schan, tx - rx);
tools/testing/selftests/drivers/net/hw/ncdevmem.c
613
ethtool_channels_set_req_set_tx_count(schan, tx);
tools/testing/selftests/drivers/net/psp_responder.c
36
unsigned char tx;
tools/testing/selftests/drivers/net/psp_responder.c
87
psp_tx_assoc_req_set_version(teq, psp_vers.tx);
tools/testing/selftests/net/psock_tpacket.c
446
struct tpacket3_hdr *tx = next;
tools/testing/selftests/net/psock_tpacket.c
448
tx->tp_snaplen = packet_len;
tools/testing/selftests/net/psock_tpacket.c
449
tx->tp_len = packet_len;
tools/testing/selftests/net/psock_tpacket.c
450
tx->tp_next_offset = 0;
tools/testing/selftests/net/psock_tpacket.c
452
memcpy((uint8_t *)tx + TPACKET3_HDRLEN -
tools/testing/selftests/net/psock_tpacket.c
455
total_bytes += tx->tp_snaplen;
tools/testing/selftests/net/tls.c
2900
__u8 tx[1024], rx[2000];
tools/testing/selftests/net/tls.c
2927
memset(tx, 0, sizeof(tx));
tools/testing/selftests/net/tls.c
2928
ASSERT_EQ(send(cfd, tx, sizeof(tx), 0), sizeof(tx));
tools/testing/selftests/net/tls.c
2941
ASSERT_EQ(total_plaintext_rx, sizeof(tx));
tools/testing/selftests/net/tls.c
2950
__u8 tx[1024], rx[2000];
tools/testing/selftests/net/tls.c
2971
memset(tx, 0, sizeof(tx));
tools/testing/selftests/net/tls.c
2972
ASSERT_EQ(send(cfd, tx, tx_partial, MSG_MORE), tx_partial);
tools/testing/selftests/net/tls.c
2983
ASSERT_EQ(send(cfd, tx + tx_partial, sizeof(tx) - tx_partial, MSG_EOR),
tools/testing/selftests/net/tls.c
2984
sizeof(tx) - tx_partial);
tools/testing/selftests/net/tls.c
2996
ASSERT_EQ(total_plaintext_rx, sizeof(tx));
tools/testing/selftests/ptp/testptp.c
164
struct timex tx;
tools/testing/selftests/ptp/testptp.c
362
memset(&tx, 0, sizeof(tx));
tools/testing/selftests/ptp/testptp.c
363
tx.modes = ADJ_FREQUENCY;
tools/testing/selftests/ptp/testptp.c
364
tx.freq = ppb_to_scaled_ppm(adjfreq);
tools/testing/selftests/ptp/testptp.c
365
if (clock_adjtime(clkid, &tx)) {
tools/testing/selftests/ptp/testptp.c
373
memset(&tx, 0, sizeof(tx));
tools/testing/selftests/ptp/testptp.c
374
tx.modes = ADJ_SETOFFSET | ADJ_NANO;
tools/testing/selftests/ptp/testptp.c
375
tx.time.tv_sec = adjtime;
tools/testing/selftests/ptp/testptp.c
376
tx.time.tv_usec = adjns;
tools/testing/selftests/ptp/testptp.c
377
while (tx.time.tv_usec < 0) {
tools/testing/selftests/ptp/testptp.c
378
tx.time.tv_sec -= 1;
tools/testing/selftests/ptp/testptp.c
379
tx.time.tv_usec += NSEC_PER_SEC;
tools/testing/selftests/ptp/testptp.c
382
if (clock_adjtime(clkid, &tx) < 0) {
tools/testing/selftests/ptp/testptp.c
390
memset(&tx, 0, sizeof(tx));
tools/testing/selftests/ptp/testptp.c
391
tx.modes = ADJ_OFFSET | ADJ_NANO;
tools/testing/selftests/ptp/testptp.c
392
tx.offset = adjphase;
tools/testing/selftests/ptp/testptp.c
394
if (clock_adjtime(clkid, &tx) < 0) {
tools/testing/selftests/ptp/testptp.c
43
static int clock_adjtime(clockid_t id, struct timex *tx)
tools/testing/selftests/ptp/testptp.c
45
return syscall(__NR_clock_adjtime, id, tx);
tools/testing/selftests/timers/change_skew.c
35
struct timex tx;
tools/testing/selftests/timers/change_skew.c
38
tx.modes = ADJ_FREQUENCY;
tools/testing/selftests/timers/change_skew.c
39
tx.freq = ppm << 16;
tools/testing/selftests/timers/change_skew.c
41
ret = adjtimex(&tx);
tools/testing/selftests/timers/change_skew.c
57
struct timex tx;
tools/testing/selftests/timers/change_skew.c
66
tx.modes = ADJ_OFFSET;
tools/testing/selftests/timers/change_skew.c
67
tx.offset = 0;
tools/testing/selftests/timers/change_skew.c
68
ret = adjtimex(&tx);
tools/testing/selftests/timers/change_skew.c
83
tx.modes = ADJ_FREQUENCY;
tools/testing/selftests/timers/change_skew.c
84
tx.offset = 0;
tools/testing/selftests/timers/change_skew.c
85
adjtimex(&tx);
tools/testing/selftests/timers/leap-a-day.c
106
tx.modes = ADJ_STATUS;
tools/testing/selftests/timers/leap-a-day.c
107
tx.status = STA_PLL;
tools/testing/selftests/timers/leap-a-day.c
108
ret = adjtimex(&tx);
tools/testing/selftests/timers/leap-a-day.c
111
tx.modes = ADJ_MAXERROR;
tools/testing/selftests/timers/leap-a-day.c
112
tx.maxerror = 0;
tools/testing/selftests/timers/leap-a-day.c
113
ret = adjtimex(&tx);
tools/testing/selftests/timers/leap-a-day.c
116
tx.modes = ADJ_STATUS;
tools/testing/selftests/timers/leap-a-day.c
117
tx.status = 0;
tools/testing/selftests/timers/leap-a-day.c
118
ret = adjtimex(&tx);
tools/testing/selftests/timers/leap-a-day.c
132
struct timex tx;
tools/testing/selftests/timers/leap-a-day.c
135
tx.modes = 0;
tools/testing/selftests/timers/leap-a-day.c
136
ret = adjtimex(&tx);
tools/testing/selftests/timers/leap-a-day.c
138
if (tx.time.tv_sec < next_leap) {
tools/testing/selftests/timers/leap-a-day.c
142
tx.time.tv_sec,
tools/testing/selftests/timers/leap-a-day.c
143
tx.time.tv_usec,
tools/testing/selftests/timers/leap-a-day.c
144
tx.tai,
tools/testing/selftests/timers/leap-a-day.c
151
tx.time.tv_sec,
tools/testing/selftests/timers/leap-a-day.c
152
tx.time.tv_usec,
tools/testing/selftests/timers/leap-a-day.c
153
tx.tai,
tools/testing/selftests/timers/leap-a-day.c
239
struct timex tx;
tools/testing/selftests/timers/leap-a-day.c
262
tx.modes = ADJ_STATUS;
tools/testing/selftests/timers/leap-a-day.c
264
tx.status = STA_INS;
tools/testing/selftests/timers/leap-a-day.c
266
tx.status = STA_DEL;
tools/testing/selftests/timers/leap-a-day.c
267
ret = adjtimex(&tx);
tools/testing/selftests/timers/leap-a-day.c
275
tx.modes = 0;
tools/testing/selftests/timers/leap-a-day.c
276
ret = adjtimex(&tx);
tools/testing/selftests/timers/leap-a-day.c
277
if (tx.status != STA_INS && tx.status != STA_DEL) {
tools/testing/selftests/timers/leap-a-day.c
315
tx.modes = 0;
tools/testing/selftests/timers/leap-a-day.c
316
ret = adjtimex(&tx);
tools/testing/selftests/timers/leap-a-day.c
317
if (tx.status != STA_INS && tx.status != STA_DEL) {
tools/testing/selftests/timers/leap-a-day.c
319
tx.modes = ADJ_STATUS;
tools/testing/selftests/timers/leap-a-day.c
321
tx.status = STA_INS;
tools/testing/selftests/timers/leap-a-day.c
323
tx.status = STA_DEL;
tools/testing/selftests/timers/leap-a-day.c
324
ret = adjtimex(&tx);
tools/testing/selftests/timers/leap-a-day.c
328
now = tx.time.tv_sec;
tools/testing/selftests/timers/leap-a-day.c
334
tx.modes = 0;
tools/testing/selftests/timers/leap-a-day.c
335
ret = adjtimex(&tx);
tools/testing/selftests/timers/leap-a-day.c
344
ctime_r(&tx.time.tv_sec, buf);
tools/testing/selftests/timers/leap-a-day.c
349
tx.time.tv_usec,
tools/testing/selftests/timers/leap-a-day.c
350
tx.tai,
tools/testing/selftests/timers/leap-a-day.c
353
now = tx.time.tv_sec;
tools/testing/selftests/timers/leap-a-day.c
97
struct timex tx;
tools/testing/selftests/timers/leapcrash.c
100
adjtimex(&tx);
tools/testing/selftests/timers/leapcrash.c
30
struct timex tx;
tools/testing/selftests/timers/leapcrash.c
39
tx.modes = ADJ_STATUS;
tools/testing/selftests/timers/leapcrash.c
40
tx.status = STA_PLL;
tools/testing/selftests/timers/leapcrash.c
41
ret = adjtimex(&tx);
tools/testing/selftests/timers/leapcrash.c
43
tx.modes = ADJ_STATUS;
tools/testing/selftests/timers/leapcrash.c
44
tx.status = 0;
tools/testing/selftests/timers/leapcrash.c
45
ret = adjtimex(&tx);
tools/testing/selftests/timers/leapcrash.c
60
struct timex tx;
tools/testing/selftests/timers/leapcrash.c
92
tx.modes = 0;
tools/testing/selftests/timers/leapcrash.c
93
adjtimex(&tx);
tools/testing/selftests/timers/leapcrash.c
96
while (tx.time.tv_sec < next_leap + 1) {
tools/testing/selftests/timers/leapcrash.c
98
tx.modes = ADJ_STATUS;
tools/testing/selftests/timers/leapcrash.c
99
tx.status = STA_INS;
tools/testing/selftests/timers/set-tai.c
30
struct timex tx;
tools/testing/selftests/timers/set-tai.c
32
memset(&tx, 0, sizeof(tx));
tools/testing/selftests/timers/set-tai.c
34
tx.modes = ADJ_TAI;
tools/testing/selftests/timers/set-tai.c
35
tx.constant = offset;
tools/testing/selftests/timers/set-tai.c
37
return adjtimex(&tx);
tools/testing/selftests/timers/set-tai.c
42
struct timex tx;
tools/testing/selftests/timers/set-tai.c
44
memset(&tx, 0, sizeof(tx));
tools/testing/selftests/timers/set-tai.c
46
adjtimex(&tx);
tools/testing/selftests/timers/set-tai.c
47
return tx.tai;
tools/testing/selftests/timers/skew_consistency.c
41
struct timex tx;
tools/testing/selftests/timers/skew_consistency.c
57
tx.modes = ADJ_FREQUENCY;
tools/testing/selftests/timers/skew_consistency.c
58
tx.freq = ppm << 16;
tools/testing/selftests/timers/skew_consistency.c
59
adjtimex(&tx);
tools/testing/selftests/timers/skew_consistency.c
64
tx.modes = ADJ_FREQUENCY;
tools/testing/selftests/timers/skew_consistency.c
65
tx.offset = 0;
tools/testing/selftests/timers/skew_consistency.c
66
adjtimex(&tx);
tools/testing/selftests/timers/valid-adjtimex.c
113
struct timex tx;
tools/testing/selftests/timers/valid-adjtimex.c
119
memset(&tx, 0, sizeof(struct timex));
tools/testing/selftests/timers/valid-adjtimex.c
125
tx.modes = ADJ_FREQUENCY;
tools/testing/selftests/timers/valid-adjtimex.c
126
tx.freq = valid_freq[i];
tools/testing/selftests/timers/valid-adjtimex.c
128
ret = adjtimex(&tx);
tools/testing/selftests/timers/valid-adjtimex.c
136
tx.modes = 0;
tools/testing/selftests/timers/valid-adjtimex.c
137
ret = adjtimex(&tx);
tools/testing/selftests/timers/valid-adjtimex.c
138
if (tx.freq != valid_freq[i]) {
tools/testing/selftests/timers/valid-adjtimex.c
140
tx.freq, valid_freq[i]);
tools/testing/selftests/timers/valid-adjtimex.c
144
tx.modes = ADJ_FREQUENCY;
tools/testing/selftests/timers/valid-adjtimex.c
145
tx.freq = outofrange_freq[i];
tools/testing/selftests/timers/valid-adjtimex.c
147
ret = adjtimex(&tx);
tools/testing/selftests/timers/valid-adjtimex.c
155
tx.modes = 0;
tools/testing/selftests/timers/valid-adjtimex.c
156
ret = adjtimex(&tx);
tools/testing/selftests/timers/valid-adjtimex.c
157
if (tx.freq == outofrange_freq[i]) {
tools/testing/selftests/timers/valid-adjtimex.c
160
tx.freq);
tools/testing/selftests/timers/valid-adjtimex.c
169
tx.modes = ADJ_FREQUENCY;
tools/testing/selftests/timers/valid-adjtimex.c
170
tx.freq = invalid_freq[i];
tools/testing/selftests/timers/valid-adjtimex.c
171
ret = adjtimex(&tx);
tools/testing/selftests/timers/valid-adjtimex.c
185
tx.modes = ADJ_FREQUENCY;
tools/testing/selftests/timers/valid-adjtimex.c
186
tx.freq = 0;
tools/testing/selftests/timers/valid-adjtimex.c
187
ret = adjtimex(&tx);
tools/testing/selftests/timers/valid-adjtimex.c
38
int clock_adjtime(clockid_t id, struct timex *tx)
tools/testing/selftests/timers/valid-adjtimex.c
40
return syscall(__NR_clock_adjtime, id, tx);
tools/testing/selftests/timers/valid-adjtimex.c
47
struct timex tx;
tools/testing/selftests/timers/valid-adjtimex.c
50
tx.modes = ADJ_STATUS;
tools/testing/selftests/timers/valid-adjtimex.c
51
tx.status = 0;
tools/testing/selftests/timers/valid-adjtimex.c
52
ret = adjtimex(&tx);