arch/arm64/kvm/hyp/nvhe/ffa.c
153
.a2 = hyp_virt_to_phys(hyp_buffers.rx),
arch/arm64/kvm/hyp/nvhe/ffa.c
237
DECLARE_REG(phys_addr_t, rx, ctxt, 2);
arch/arm64/kvm/hyp/nvhe/ffa.c
247
if (!PAGE_ALIGNED(tx) || !PAGE_ALIGNED(rx)) {
arch/arm64/kvm/hyp/nvhe/ffa.c
272
ret = __pkvm_host_share_hyp(hyp_phys_to_pfn(rx));
arch/arm64/kvm/hyp/nvhe/ffa.c
285
rx_virt = hyp_phys_to_virt(rx);
arch/arm64/kvm/hyp/nvhe/ffa.c
293
host_buffers.rx = rx_virt;
arch/arm64/kvm/hyp/nvhe/ffa.c
304
__pkvm_host_unshare_hyp(hyp_phys_to_pfn(rx));
arch/arm64/kvm/hyp/nvhe/ffa.c
333
hyp_unpin_shared_mem(host_buffers.rx, host_buffers.rx + 1);
arch/arm64/kvm/hyp/nvhe/ffa.c
334
WARN_ON(__pkvm_host_unshare_hyp(hyp_virt_to_pfn(host_buffers.rx)));
arch/arm64/kvm/hyp/nvhe/ffa.c
335
host_buffers.rx = NULL;
arch/arm64/kvm/hyp/nvhe/ffa.c
595
buf = hyp_buffers.rx;
arch/arm64/kvm/hyp/nvhe/ffa.c
60
void *rx;
arch/arm64/kvm/hyp/nvhe/ffa.c
624
memcpy(buf, hyp_buffers.rx, fraglen);
arch/arm64/kvm/hyp/nvhe/ffa.c
635
memcpy((void *)buf + fragoff, hyp_buffers.rx, fraglen);
arch/arm64/kvm/hyp/nvhe/ffa.c
822
if (!host_buffers.rx) {
arch/arm64/kvm/hyp/nvhe/ffa.c
860
memcpy(host_buffers.rx, hyp_buffers.rx, copy_sz);
arch/arm64/kvm/hyp/nvhe/ffa.c
937
void *tx, *rx;
arch/arm64/kvm/hyp/nvhe/ffa.c
972
rx = pages;
arch/arm64/kvm/hyp/nvhe/ffa.c
984
.rx = rx,
arch/csky/abiv1/alignment.c
13
static inline uint32_t get_ptreg(struct pt_regs *regs, uint32_t rx)
arch/csky/abiv1/alignment.c
15
return rx == 15 ? regs->lr : *((uint32_t *)&(regs->a0) - 2 + rx);
arch/csky/abiv1/alignment.c
18
static inline void put_ptreg(struct pt_regs *regs, uint32_t rx, uint32_t val)
arch/csky/abiv1/alignment.c
20
if (rx == 15)
arch/csky/abiv1/alignment.c
217
uint32_t rx = 0;
arch/csky/abiv1/alignment.c
23
*((uint32_t *)&(regs->a0) - 2 + rx) = val;
arch/csky/abiv1/alignment.c
253
rx = opcode & 0xf;
arch/csky/abiv1/alignment.c
258
if (rx == 0 || rx == 1 || rz == 0 || rz == 1)
arch/csky/abiv1/alignment.c
263
addr = get_ptreg(regs, rx) + (imm << 1);
arch/csky/abiv1/alignment.c
267
addr = get_ptreg(regs, rx) + (imm << 2);
arch/csky/abiv1/alignment.c
271
addr = get_ptreg(regs, rx) + (imm << 1);
arch/csky/abiv1/alignment.c
275
addr = get_ptreg(regs, rx) + (imm << 2);
arch/csky/abiv1/alignment.c
294
__func__, opcode, rz, rx, imm, addr);
arch/csky/abiv1/inc/abi/entry.h
129
.macro RD_MIR rx
arch/csky/abiv1/inc/abi/entry.h
133
.macro RD_MEH rx
arch/csky/abiv1/inc/abi/entry.h
137
.macro RD_MCIR rx
arch/csky/abiv1/inc/abi/entry.h
141
.macro RD_PGDR rx
arch/csky/abiv1/inc/abi/entry.h
145
.macro WR_MEH rx
arch/csky/abiv1/inc/abi/entry.h
149
.macro WR_MCIR rx
arch/csky/abiv2/inc/abi/entry.h
208
.macro RD_MIR rx
arch/csky/abiv2/inc/abi/entry.h
212
.macro RD_MEH rx
arch/csky/abiv2/inc/abi/entry.h
216
.macro RD_MCIR rx
arch/csky/abiv2/inc/abi/entry.h
220
.macro RD_PGDR rx
arch/csky/abiv2/inc/abi/entry.h
224
.macro RD_PGDR_K rx
arch/csky/abiv2/inc/abi/entry.h
228
.macro WR_MEH rx
arch/csky/abiv2/inc/abi/entry.h
232
.macro WR_MCIR rx
arch/mips/include/asm/ip32/mace.h
239
volatile unsigned long rx;
arch/mips/include/asm/octeon/cvmx-agl-defs.h
1066
uint64_t rx:2;
arch/mips/include/asm/octeon/cvmx-agl-defs.h
1068
uint64_t rx:2;
arch/mips/include/asm/octeon/cvmx-agl-defs.h
1079
uint64_t rx:1;
arch/mips/include/asm/octeon/cvmx-agl-defs.h
1081
uint64_t rx:1;
arch/mips/include/uapi/asm/inst.h
1062
__BITFIELD_FIELD(unsigned int rx : 3,
arch/mips/include/uapi/asm/inst.h
1095
__BITFIELD_FIELD(unsigned int rx : 3,
arch/mips/include/uapi/asm/inst.h
1102
__BITFIELD_FIELD(unsigned int rx : 3,
arch/mips/kernel/branch.c
380
regs->regs[reg16to32[inst.rr.rx]];
arch/mips/kernel/unaligned.c
1294
reg = reg16to32[mips16inst.ri.rx];
arch/mips/kernel/unaligned.c
1309
reg = reg16to32[mips16inst.ri.rx];
arch/mips/kernel/unaligned.c
1313
reg = reg16to32[mips16inst.ri.rx];
arch/s390/boot/physmem_info.c
62
union register_pair rx;
arch/s390/boot/physmem_info.c
66
rx.even = rx1;
arch/s390/boot/physmem_info.c
67
rx.odd = rx2;
arch/s390/boot/physmem_info.c
77
: [rx] "d" (rx.pair)
arch/s390/include/asm/diag.h
352
int (*diag26c)(unsigned long rx, unsigned long rx1, enum diag26c_sc subcode);
arch/s390/include/asm/diag.h
353
int (*diag14)(unsigned long rx, unsigned long ry1, unsigned long subcode);
arch/s390/include/asm/diag.h
355
void (*diag0c)(unsigned long rx);
arch/s390/include/asm/diag.h
363
int _diag26c_amode31(unsigned long rx, unsigned long rx1, enum diag26c_sc subcode);
arch/s390/include/asm/diag.h
364
int _diag14_amode31(unsigned long rx, unsigned long ry1, unsigned long subcode);
arch/s390/include/asm/diag.h
365
void _diag0c_amode31(unsigned long rx);
arch/s390/include/asm/diag.h
80
extern int diag14(unsigned long rx, unsigned long ry1, unsigned long subcode);
arch/s390/kernel/cpcmd.c
32
: [rx] "d" (__pa(cpcmd_buf))
arch/s390/kernel/cpcmd.c
39
union register_pair rx, ry;
arch/s390/kernel/cpcmd.c
42
rx.even = __pa(cpcmd_buf);
arch/s390/kernel/cpcmd.c
43
rx.odd = __pa(response);
arch/s390/kernel/cpcmd.c
50
: [rx] "d" (rx.pair)
arch/s390/kernel/diag/diag.c
176
int diag14(unsigned long rx, unsigned long ry1, unsigned long subcode)
arch/s390/kernel/diag/diag.c
182
rx = virt_to_phys((void *)rx);
arch/s390/kernel/diag/diag.c
188
return diag_amode31_ops.diag14(rx, ry1, subcode);
arch/s390/kernel/uprobes.c
259
union split_register *rx;
arch/s390/kernel/uprobes.c
266
rx = (union split_register *) ®s->gprs[insn->reg];
arch/s390/kernel/uprobes.c
274
rx->u64 = (unsigned long)uptr;
arch/s390/kernel/uprobes.c
281
rc = emu_load_ril((u16 __user *)uptr, &rx->u32[1]);
arch/s390/kernel/uprobes.c
284
rc = emu_load_ril((s16 __user *)uptr, &rx->u64);
arch/s390/kernel/uprobes.c
287
rc = emu_load_ril((s16 __user *)uptr, &rx->u32[1]);
arch/s390/kernel/uprobes.c
290
rc = emu_load_ril((u16 __user *)uptr, &rx->u64);
arch/s390/kernel/uprobes.c
293
rc = emu_load_ril((u64 __user *)uptr, &rx->u64);
arch/s390/kernel/uprobes.c
296
rc = emu_load_ril((s32 __user *)uptr, &rx->u64);
arch/s390/kernel/uprobes.c
299
rc = emu_load_ril((u32 __user *)uptr, &rx->u32[1]);
arch/s390/kernel/uprobes.c
302
rc = emu_load_ril((u32 __user *)uptr, &rx->u64);
arch/s390/kernel/uprobes.c
305
rc = emu_store_ril(regs, (u16 __user *)uptr, &rx->u16[3]);
arch/s390/kernel/uprobes.c
308
rc = emu_store_ril(regs, (u64 __user *)uptr, &rx->u64);
arch/s390/kernel/uprobes.c
311
rc = emu_store_ril(regs, (u32 __user *)uptr, &rx->u32[1]);
arch/s390/kernel/uprobes.c
318
rc = emu_cmp_ril(regs, (s16 __user *)uptr, &rx->s64);
arch/s390/kernel/uprobes.c
321
rc = emu_cmp_ril(regs, (s16 __user *)uptr, &rx->s32[1]);
arch/s390/kernel/uprobes.c
324
rc = emu_cmp_ril(regs, (u16 __user *)uptr, &rx->u64);
arch/s390/kernel/uprobes.c
327
rc = emu_cmp_ril(regs, (u16 __user *)uptr, &rx->u32[1]);
arch/s390/kernel/uprobes.c
330
rc = emu_cmp_ril(regs, (s64 __user *)uptr, &rx->s64);
arch/s390/kernel/uprobes.c
333
rc = emu_cmp_ril(regs, (u64 __user *)uptr, &rx->u64);
arch/s390/kernel/uprobes.c
336
rc = emu_cmp_ril(regs, (s32 __user *)uptr, &rx->s64);
arch/s390/kernel/uprobes.c
339
rc = emu_cmp_ril(regs, (s32 __user *)uptr, &rx->s32[1]);
arch/s390/kernel/uprobes.c
342
rc = emu_cmp_ril(regs, (u32 __user *)uptr, &rx->u64);
arch/s390/kernel/uprobes.c
345
rc = emu_cmp_ril(regs, (u32 __user *)uptr, &rx->u32[1]);
arch/s390/kvm/diag.c
100
rc = read_guest_real(vcpu, vcpu->run->s.regs.gprs[rx], &parm, sizeof(parm));
arch/s390/kvm/diag.c
92
u16 rx = (vcpu->arch.sie_block->ipa & 0xf0) >> 4;
arch/s390/kvm/diag.c
96
vcpu->run->s.regs.gprs[rx]);
arch/s390/kvm/diag.c
98
if (vcpu->run->s.regs.gprs[rx] & 7)
arch/s390/mm/extmem.c
137
unsigned long rx, ry;
arch/s390/mm/extmem.c
140
rx = virt_to_phys(parameter);
arch/s390/mm/extmem.c
147
: CC_OUT(cc, cc), [rx] "+d" (rx), [ry] "+d" (ry)
arch/s390/mm/extmem.c
150
*ret1 = rx;
crypto/ecc.c
1326
u64 rx[2][ECC_MAX_DIGITS];
crypto/ecc.c
1343
vli_set(rx[1], point->x, ndigits);
crypto/ecc.c
1346
xycz_initial_double(rx[1], ry[1], rx[0], ry[0], initial_z, curve);
crypto/ecc.c
1350
xycz_add_c(rx[1 - nb], ry[1 - nb], rx[nb], ry[nb], curve);
crypto/ecc.c
1351
xycz_add(rx[nb], ry[nb], rx[1 - nb], ry[1 - nb], curve);
crypto/ecc.c
1355
xycz_add_c(rx[1 - nb], ry[1 - nb], rx[nb], ry[nb], curve);
crypto/ecc.c
1359
vli_mod_sub(z, rx[1], rx[0], curve_prime, ndigits);
crypto/ecc.c
1371
vli_mod_mult_fast(z, z, rx[1 - nb], curve);
crypto/ecc.c
1374
xycz_add(rx[nb], ry[nb], rx[1 - nb], ry[1 - nb], curve);
crypto/ecc.c
1376
apply_z(rx[0], ry[0], z, curve);
crypto/ecc.c
1378
vli_set(result->x, rx[0], ndigits);
crypto/ecc.c
1412
u64 *rx = result->x;
crypto/ecc.c
1434
vli_set(rx, point->x, ndigits);
crypto/ecc.c
1440
ecc_point_double_jacobian(rx, ry, z, curve);
crypto/ecc.c
1452
vli_mod_sub(tz, rx, tx, curve->p, ndigits);
crypto/ecc.c
1453
xycz_add(tx, ty, rx, ry, curve);
crypto/ecc.c
1458
apply_z(rx, ry, z, curve);
drivers/accel/qaic/sahara.c
159
struct sahara_packet *rx;
drivers/accel/qaic/sahara.c
298
le32_to_cpu(context->rx->length),
drivers/accel/qaic/sahara.c
299
le32_to_cpu(context->rx->hello.version),
drivers/accel/qaic/sahara.c
300
le32_to_cpu(context->rx->hello.version_compat),
drivers/accel/qaic/sahara.c
301
le32_to_cpu(context->rx->hello.max_length),
drivers/accel/qaic/sahara.c
302
le32_to_cpu(context->rx->hello.mode));
drivers/accel/qaic/sahara.c
304
if (le32_to_cpu(context->rx->length) != SAHARA_HELLO_LENGTH) {
drivers/accel/qaic/sahara.c
306
le32_to_cpu(context->rx->length));
drivers/accel/qaic/sahara.c
309
if (le32_to_cpu(context->rx->hello.version) != SAHARA_VERSION) {
drivers/accel/qaic/sahara.c
311
le32_to_cpu(context->rx->hello.version));
drivers/accel/qaic/sahara.c
315
if (le32_to_cpu(context->rx->hello.mode) != SAHARA_MODE_IMAGE_TX_PENDING &&
drivers/accel/qaic/sahara.c
316
le32_to_cpu(context->rx->hello.mode) != SAHARA_MODE_IMAGE_TX_COMPLETE &&
drivers/accel/qaic/sahara.c
317
le32_to_cpu(context->rx->hello.mode) != SAHARA_MODE_MEMORY_DEBUG) {
drivers/accel/qaic/sahara.c
319
le32_to_cpu(context->rx->hello.mode));
drivers/accel/qaic/sahara.c
328
context->tx[0]->hello_resp.mode = context->rx->hello_resp.mode;
drivers/accel/qaic/sahara.c
374
le32_to_cpu(context->rx->length),
drivers/accel/qaic/sahara.c
375
le32_to_cpu(context->rx->read_data.image),
drivers/accel/qaic/sahara.c
376
le32_to_cpu(context->rx->read_data.offset),
drivers/accel/qaic/sahara.c
377
le32_to_cpu(context->rx->read_data.length));
drivers/accel/qaic/sahara.c
379
if (le32_to_cpu(context->rx->length) != SAHARA_READ_DATA_LENGTH) {
drivers/accel/qaic/sahara.c
381
le32_to_cpu(context->rx->length));
drivers/accel/qaic/sahara.c
385
image_id = le32_to_cpu(context->rx->read_data.image);
drivers/accel/qaic/sahara.c
386
data_offset = le32_to_cpu(context->rx->read_data.offset);
drivers/accel/qaic/sahara.c
387
data_len = le32_to_cpu(context->rx->read_data.length);
drivers/accel/qaic/sahara.c
446
le32_to_cpu(context->rx->length),
drivers/accel/qaic/sahara.c
447
le32_to_cpu(context->rx->end_of_image.image),
drivers/accel/qaic/sahara.c
448
le32_to_cpu(context->rx->end_of_image.status));
drivers/accel/qaic/sahara.c
450
if (le32_to_cpu(context->rx->length) != SAHARA_END_OF_IMAGE_LENGTH) {
drivers/accel/qaic/sahara.c
452
le32_to_cpu(context->rx->length));
drivers/accel/qaic/sahara.c
457
le32_to_cpu(context->rx->end_of_image.image) != context->active_image_id) {
drivers/accel/qaic/sahara.c
459
le32_to_cpu(context->rx->end_of_image.image));
drivers/accel/qaic/sahara.c
465
if (le32_to_cpu(context->rx->end_of_image.status))
drivers/accel/qaic/sahara.c
483
le32_to_cpu(context->rx->length),
drivers/accel/qaic/sahara.c
484
le64_to_cpu(context->rx->memory_debug64.table_address),
drivers/accel/qaic/sahara.c
485
le64_to_cpu(context->rx->memory_debug64.table_length));
drivers/accel/qaic/sahara.c
487
if (le32_to_cpu(context->rx->length) != SAHARA_MEM_DEBUG64_LENGTH) {
drivers/accel/qaic/sahara.c
489
le32_to_cpu(context->rx->length));
drivers/accel/qaic/sahara.c
493
context->dump_table_address = le64_to_cpu(context->rx->memory_debug64.table_address);
drivers/accel/qaic/sahara.c
494
context->dump_table_length = le64_to_cpu(context->rx->memory_debug64.table_length);
drivers/accel/qaic/sahara.c
543
switch (le32_to_cpu(context->rx->cmd)) {
drivers/accel/qaic/sahara.c
564
le32_to_cpu(context->rx->cmd));
drivers/accel/qaic/sahara.c
568
ret = mhi_queue_buf(context->mhi_dev, DMA_FROM_DEVICE, context->rx,
drivers/accel/qaic/sahara.c
589
dev_table = (struct sahara_debug_table_entry64 *)(context->rx);
drivers/accel/qaic/sahara.c
688
memcpy(context->mem_dump_freespace, context->rx, context->rx_size);
drivers/accel/qaic/sahara.c
751
le32_to_cpu(context->rx->cmd) == SAHARA_END_OF_IMAGE_CMD) {
drivers/accel/qaic/sahara.c
754
le32_to_cpu(context->rx->end_of_image.status));
drivers/accel/qaic/sahara.c
759
le32_to_cpu(context->rx->cmd) != SAHARA_END_OF_IMAGE_CMD) {
drivers/accel/qaic/sahara.c
762
le32_to_cpu(context->rx->cmd));
drivers/accel/qaic/sahara.c
777
ret = mhi_queue_buf(context->mhi_dev, DMA_FROM_DEVICE, context->rx,
drivers/accel/qaic/sahara.c
807
context->rx = devm_kzalloc(&mhi_dev->dev, SAHARA_PACKET_MAX_SIZE, GFP_KERNEL);
drivers/accel/qaic/sahara.c
808
if (!context->rx)
drivers/accel/qaic/sahara.c
869
ret = mhi_queue_buf(mhi_dev, DMA_FROM_DEVICE, context->rx, SAHARA_PACKET_MAX_SIZE, MHI_EOT);
drivers/accessibility/speakup/main.c
1584
speakup_console[vc_num]->ht.rx[bi] = vc->state.x;
drivers/accessibility/speakup/main.c
1683
spk_x = spk_cx = speakup_console[vc_num]->ht.rx[hc];
drivers/accessibility/speakup/spk_types.h
68
u_long rpos[8], rx[8], ry[8];
drivers/atm/atmtcp.c
230
atomic_inc(&out_vcc->stats->rx);
drivers/atm/atmtcp.c
329
atomic_inc(&out_vcc->stats->rx);
drivers/atm/eni.c
188
if (eni_dev->rx_map[i] && ENI_VCC(eni_dev->rx_map[i])->rx)
drivers/atm/eni.c
2012
if ((mult.tx && mult.tx <= 100) || (mult.rx &&mult.rx <= 100) ||
drivers/atm/eni.c
2013
mult.tx > 65536 || mult.rx > 65536)
drivers/atm/eni.c
2016
if (mult.rx) eni_dev->rx_mult = mult.rx;
drivers/atm/eni.c
2171
if (eni_vcc->rx) {
drivers/atm/eni.c
628
if (ENI_VCC(vcc)->rx(vcc)) return 1;
drivers/atm/eni.c
645
if (ENI_VCC(vcc)->rx(vcc)) return 1;
drivers/atm/eni.c
769
atomic_inc(&vcc->stats->rx);
drivers/atm/eni.c
784
eni_vcc->rx = NULL;
drivers/atm/eni.c
794
eni_vcc->rx = vcc->qos.aal == ATM_AAL5 ? rx_aal5 : rx_aal0;
drivers/atm/eni.c
815
if (!eni_vcc->rx) return 0;
drivers/atm/eni.c
843
if (!eni_vcc->rx) return;
drivers/atm/eni.c
895
eni_vcc->rx = NULL;
drivers/atm/eni.h
54
int (*rx)(struct atm_vcc *vcc); /* RX function, NULL if none */
drivers/atm/fore200e.c
1046
atomic_inc(&vcc->stats->rx);
drivers/atm/he.c
1767
atomic_inc(&vcc->stats->rx);
drivers/atm/idt77252.c
1101
atomic_inc(&vcc->stats->rx);
drivers/atm/idt77252.c
1173
atomic_inc(&vcc->stats->rx);
drivers/atm/idt77252.c
1196
atomic_inc(&vcc->stats->rx);
drivers/atm/idt77252.c
1332
atomic_inc(&vcc->stats->rx);
drivers/atm/iphase.c
1353
atomic_inc(&vcc->stats->rx);
drivers/atm/lanai.c
1366
int m = ((const unsigned char *) lvcc->rx.buf.ptr) + n -
drivers/atm/lanai.c
1367
((const unsigned char *) (lvcc->rx.buf.end));
drivers/atm/lanai.c
1370
memcpy(dest, lvcc->rx.buf.ptr, n - m);
drivers/atm/lanai.c
1371
memcpy(dest + n - m, lvcc->rx.buf.start, m);
drivers/atm/lanai.c
1382
u32 *end = &lvcc->rx.buf.start[endptr * 4];
drivers/atm/lanai.c
1383
int n = ((unsigned long) end) - ((unsigned long) lvcc->rx.buf.ptr);
drivers/atm/lanai.c
1385
n += lanai_buf_size(&lvcc->rx.buf);
drivers/atm/lanai.c
1386
APRINTK(n >= 0 && n < lanai_buf_size(&lvcc->rx.buf) && !(n & 15),
drivers/atm/lanai.c
1388
n, lanai_buf_size(&lvcc->rx.buf));
drivers/atm/lanai.c
1390
if ((x = &end[-2]) < lvcc->rx.buf.start)
drivers/atm/lanai.c
1391
x = &lvcc->rx.buf.end[-2];
drivers/atm/lanai.c
1402
lvcc->rx.atmvcc->dev->number, lvcc->vci, size, n);
drivers/atm/lanai.c
1406
skb = atm_alloc_charge(lvcc->rx.atmvcc, size, GFP_ATOMIC);
drivers/atm/lanai.c
1413
ATM_SKB(skb)->vcc = lvcc->rx.atmvcc;
drivers/atm/lanai.c
1415
lvcc->rx.atmvcc->push(lvcc->rx.atmvcc, skb);
drivers/atm/lanai.c
1416
atomic_inc(&lvcc->rx.atmvcc->stats->rx);
drivers/atm/lanai.c
1418
lvcc->rx.buf.ptr = end;
drivers/atm/lanai.c
1501
return lanai_get_sized_buffer(lanai, &lvcc->rx.buf,
drivers/atm/lanai.c
1645
if (unlikely(lvcc->rx.atmvcc == NULL)) {
drivers/atm/lanai.c
1652
if (unlikely(lvcc->rx.atmvcc->qos.aal != ATM_AAL5)) {
drivers/atm/lanai.c
1657
atomic_inc(&lvcc->rx.atmvcc->stats->rx_err);
drivers/atm/lanai.c
1669
atomic_inc(&lvcc->rx.atmvcc->stats->rx_err);
drivers/atm/lanai.c
1672
(((unsigned long) lvcc->rx.buf.ptr) -
drivers/atm/lanai.c
1673
((unsigned long) lvcc->rx.buf.start)) + 47;
drivers/atm/lanai.c
1675
bytes += lanai_buf_size(&lvcc->rx.buf);
drivers/atm/lanai.c
1681
atomic_inc(&lvcc->rx.atmvcc->stats->rx_err);
drivers/atm/lanai.c
1689
atomic_inc(&lvcc->rx.atmvcc->stats->rx_err);
drivers/atm/lanai.c
1691
lvcc->rx.buf.ptr = &lvcc->rx.buf.start[SERVICE_GET_END(s) * 4];
drivers/atm/lanai.c
1995
lvcc->rx.atmvcc != NULL && lvcc->rx.atmvcc != atmvcc)
drivers/atm/lanai.c
2007
if (vci0 != NULL && vci0->rx.atmvcc != NULL)
drivers/atm/lanai.c
2281
if (lvcc->rx.atmvcc == atmvcc) {
drivers/atm/lanai.c
2287
lanai_buf_deallocate(&lvcc->rx.buf, lanai->pci);
drivers/atm/lanai.c
2288
lvcc->rx.atmvcc = NULL;
drivers/atm/lanai.c
2338
APRINTK(lvcc->rx.atmvcc == NULL, "rx.atmvcc!=NULL, vci=%d\n",
drivers/atm/lanai.c
234
} rx;
drivers/atm/lanai.c
2348
lvcc->rx.atmvcc = atmvcc;
drivers/atm/lanai.c
2376
if (atmvcc == lvcc->rx.atmvcc)
drivers/atm/lanai.c
2508
if (lvcc->rx.atmvcc != NULL) {
drivers/atm/lanai.c
2510
lvcc->rx.atmvcc->qos.aal == ATM_AAL5 ? 5 : 0);
drivers/atm/lanai.c
2511
if (lvcc->rx.atmvcc->qos.aal == ATM_AAL5)
drivers/atm/lanai.c
2515
lanai_buf_size(&lvcc->rx.buf),
drivers/atm/lanai.c
696
if (lvcc->rx.atmvcc->qos.aal == ATM_AAL5) {
drivers/atm/lanai.c
697
dma_addr_t dmaaddr = lvcc->rx.buf.dmaaddr;
drivers/atm/lanai.c
705
RXADDR1_SET_SIZE(lanai_buf_size_cardorder(&lvcc->rx.buf))|
drivers/atm/nicstar.c
1258
if (vcc->qos.rxtp.traffic_class != ATM_NONE && vc->rx)
drivers/atm/nicstar.c
1385
vc->rx = 1;
drivers/atm/nicstar.c
1436
vc->rx = 0;
drivers/atm/nicstar.c
2019
if (!vc->rx) {
drivers/atm/nicstar.c
2062
atomic_inc(&vcc->stats->rx);
drivers/atm/nicstar.c
2180
atomic_inc(&vcc->stats->rx);
drivers/atm/nicstar.c
2198
atomic_inc(&vcc->stats->rx);
drivers/atm/nicstar.c
2217
atomic_inc(&vcc->stats->rx);
drivers/atm/nicstar.c
2323
atomic_inc(&vcc->stats->rx);
drivers/atm/nicstar.h
699
volatile unsigned int rx:1; /* RX vc? */
drivers/atm/solos-pci.c
840
atomic_inc(&vcc->stats->rx);
drivers/bluetooth/virtio_bt.c
222
struct virtio_bluetooth, rx);
drivers/bluetooth/virtio_bt.c
25
struct work_struct rx;
drivers/bluetooth/virtio_bt.c
252
schedule_work(&vbt->rx);
drivers/bluetooth/virtio_bt.c
285
INIT_WORK(&vbt->rx, virtbt_rx_work);
drivers/bluetooth/virtio_bt.c
74
cancel_work_sync(&vbt->rx);
drivers/clk/clk-rpmi.c
150
struct rpmi_get_num_clocks_rx rx, *resp;
drivers/clk/clk-rpmi.c
155
NULL, 0, &rx, sizeof(rx));
drivers/clk/clk-rpmi.c
173
struct rpmi_get_attrs_rx rx, *resp;
drivers/clk/clk-rpmi.c
179
&tx, sizeof(tx), &rx, sizeof(rx));
drivers/clk/clk-rpmi.c
223
struct rpmi_get_supp_rates_rx *rx __free(kfree) =
drivers/clk/clk-rpmi.c
225
if (!rx)
drivers/clk/clk-rpmi.c
229
&tx, sizeof(tx), rx, context->max_msg_data_size);
drivers/clk/clk-rpmi.c
264
rx, context->max_msg_data_size);
drivers/clk/clk-rpmi.c
307
struct rpmi_get_rate_rx rx, *resp;
drivers/clk/clk-rpmi.c
313
&tx, sizeof(tx), &rx, sizeof(rx));
drivers/clk/clk-rpmi.c
369
struct rpmi_set_rate_rx rx, *resp;
drivers/clk/clk-rpmi.c
377
&tx, sizeof(tx), &rx, sizeof(rx));
drivers/clk/clk-rpmi.c
398
struct rpmi_set_config_rx rx, *resp;
drivers/clk/clk-rpmi.c
405
&tx, sizeof(tx), &rx, sizeof(rx));
drivers/clk/clk-rpmi.c
426
struct rpmi_set_config_rx rx;
drivers/clk/clk-rpmi.c
432
&tx, sizeof(tx), &rx, sizeof(rx));
drivers/clk/tegra/clk-bpmp.c
133
msg.rx.data = &response;
drivers/clk/tegra/clk-bpmp.c
134
msg.rx.size = sizeof(response);
drivers/clk/tegra/clk-bpmp.c
157
msg.rx.data = &response;
drivers/clk/tegra/clk-bpmp.c
158
msg.rx.size = sizeof(response);
drivers/clk/tegra/clk-bpmp.c
187
msg.rx.data = &response;
drivers/clk/tegra/clk-bpmp.c
188
msg.rx.size = sizeof(response);
drivers/clk/tegra/clk-bpmp.c
215
msg.rx.data = &response;
drivers/clk/tegra/clk-bpmp.c
216
msg.rx.size = sizeof(response);
drivers/clk/tegra/clk-bpmp.c
238
msg.rx.data = &response;
drivers/clk/tegra/clk-bpmp.c
239
msg.rx.size = sizeof(response);
drivers/clk/tegra/clk-bpmp.c
271
msg.rx.data = &response;
drivers/clk/tegra/clk-bpmp.c
272
msg.rx.size = sizeof(response);
drivers/clk/tegra/clk-bpmp.c
339
msg.rx.data = &response;
drivers/clk/tegra/clk-bpmp.c
340
msg.rx.size = sizeof(response);
drivers/clk/tegra/clk-bpmp.c
363
msg.rx.data = &response;
drivers/clk/tegra/clk-bpmp.c
364
msg.rx.size = sizeof(response);
drivers/clk/tegra/clk-bpmp.c
56
} rx;
drivers/clk/tegra/clk-bpmp.c
83
msg.rx.data = clk->rx.data;
drivers/clk/tegra/clk-bpmp.c
84
msg.rx.size = clk->rx.size;
drivers/clk/tegra/clk-bpmp.c
89
else if (msg.rx.ret < 0)
drivers/cpufreq/tegra186-cpufreq.c
293
if (msg.rx.ret) {
drivers/cpufreq/tegra194-cpufreq.c
632
msg.rx.data = &resp;
drivers/cpufreq/tegra194-cpufreq.c
633
msg.rx.size = sizeof(resp);
drivers/cpufreq/tegra194-cpufreq.c
638
if (msg.rx.ret == -BPMP_EINVAL) {
drivers/cpufreq/tegra194-cpufreq.c
642
if (msg.rx.ret)
drivers/crypto/gemini/sl3516-ce-core.c
114
rdd = &ce->rx[ce->crx];
drivers/crypto/gemini/sl3516-ce-core.c
38
ce->rx = dma_alloc_coherent(ce->dev, sz, &ce->drx, GFP_KERNEL);
drivers/crypto/gemini/sl3516-ce-core.c
39
if (!ce->rx)
drivers/crypto/gemini/sl3516-ce-core.c
49
ce->rx[i].frame_ctrl.bits.own = CE_CPU;
drivers/crypto/gemini/sl3516-ce-core.c
50
ce->rx[i].next_desc.next_descriptor = ce->drx + (i + 1) * sizeof(struct descriptor);
drivers/crypto/gemini/sl3516-ce-core.c
52
ce->rx[MAXDESC - 1].next_desc.next_descriptor = ce->drx;
drivers/crypto/gemini/sl3516-ce-core.c
61
dma_free_coherent(ce->dev, sz, ce->rx, ce->drx);
drivers/crypto/gemini/sl3516-ce-core.c
72
dma_free_coherent(ce->dev, sz, ce->rx, ce->drx);
drivers/crypto/gemini/sl3516-ce.h
240
struct descriptor *rx;
drivers/crypto/intel/qat/qat_common/adf_bank_state.c
109
if (state->ringfstat & BIT(rx))
drivers/crypto/intel/qat/qat_common/adf_bank_state.c
110
ops->write_csr_ring_tail(base, bank, rx, state->rings[rx].tail);
drivers/crypto/intel/qat/qat_common/adf_bank_state.c
77
int rx = tx + tx_rx_gap;
drivers/crypto/intel/qat/qat_common/adf_bank_state.c
94
ops->write_csr_ring_tail(base, bank, rx, state->rings[rx].tail);
drivers/crypto/intel/qat/qat_common/adf_bank_state.c
99
ops->write_csr_ring_head(base, bank, rx, state->rings[rx].head);
drivers/crypto/starfive/jh7110-aes.c
432
dmaengine_slave_config(cryp->rx, &cryp->cfg_out);
drivers/crypto/starfive/jh7110-aes.c
460
out_desc = dmaengine_prep_slave_sg(cryp->rx, dst, 1, DMA_DEV_TO_MEM,
drivers/crypto/starfive/jh7110-aes.c
472
dma_async_issue_pending(cryp->rx);
drivers/crypto/starfive/jh7110-cryp.c
76
cryp->rx = dma_request_chan(cryp->dev, "rx");
drivers/crypto/starfive/jh7110-cryp.c
77
if (IS_ERR(cryp->rx)) {
drivers/crypto/starfive/jh7110-cryp.c
79
return dev_err_probe(cryp->dev, PTR_ERR(cryp->rx),
drivers/crypto/starfive/jh7110-cryp.c
89
dma_release_channel(cryp->rx);
drivers/crypto/starfive/jh7110-cryp.h
188
struct dma_chan *rx;
drivers/dma/timb_dma.c
688
if ((i % 2) == pchan->rx) {
drivers/dma/timb_dma.c
704
td_chan->direction = pchan->rx ? DMA_DEV_TO_MEM :
drivers/dma/timb_dma.c
709
(pchan->rx ? 0 : TIMBDMA_INSTANCE_TX_OFFSET);
drivers/firmware/arm_scmi/base.c
123
strscpy(vendor_id, t->rx.buf, size);
drivers/firmware/arm_scmi/base.c
154
impl_ver = t->rx.buf;
drivers/firmware/arm_scmi/base.c
190
num_ret = t->rx.buf;
drivers/firmware/arm_scmi/base.c
191
list = t->rx.buf + sizeof(*num_ret);
drivers/firmware/arm_scmi/base.c
214
if (t->rx.len < (sizeof(u32) * 2)) {
drivers/firmware/arm_scmi/base.c
216
t->rx.len);
drivers/firmware/arm_scmi/base.c
221
real_list_sz = t->rx.len - sizeof(u32);
drivers/firmware/arm_scmi/base.c
279
agent_info = t->rx.buf;
drivers/firmware/arm_scmi/base.c
80
attr_info = t->rx.buf;
drivers/firmware/arm_scmi/clock.c
203
attr = t->rx.buf;
drivers/firmware/arm_scmi/clock.c
335
perm = get_unaligned_le32(t->rx.buf);
drivers/firmware/arm_scmi/clock.c
362
attr = t->rx.buf;
drivers/firmware/arm_scmi/clock.c
566
*value = get_unaligned_le64(t->rx.buf);
drivers/firmware/arm_scmi/clock.c
608
resp = t->rx.buf;
drivers/firmware/arm_scmi/clock.c
711
*parent_id = get_unaligned_le32(t->rx.buf);
drivers/firmware/arm_scmi/clock.c
817
struct scmi_msg_resp_clock_config_get *resp = t->rx.buf;
drivers/firmware/arm_scmi/clock.c
853
resp = t->rx.buf;
drivers/firmware/arm_scmi/driver.c
1108
xfer->hdr.status, xfer->rx.buf, xfer->rx.len);
drivers/firmware/arm_scmi/driver.c
1112
xfer->hdr.id, xfer->rx.buf, xfer->rx.len, ts);
drivers/firmware/arm_scmi/driver.c
1147
xfer->rx.len = info->desc->max_msg_size;
drivers/firmware/arm_scmi/driver.c
1160
xfer->rx.buf, xfer->rx.len);
drivers/firmware/arm_scmi/driver.c
1306
xfer->rx.buf, xfer->rx.len);
drivers/firmware/arm_scmi/driver.c
1478
xfer->rx.len = info->desc->max_msg_size;
drivers/firmware/arm_scmi/driver.c
1585
xfer->rx.len = rx_size ? : info->desc->max_msg_size;
drivers/firmware/arm_scmi/driver.c
1617
rev_info = t->rx.buf;
drivers/firmware/arm_scmi/driver.c
1699
resp = t->rx.buf;
drivers/firmware/arm_scmi/driver.c
1755
*attributes = get_unaligned_le32(t->rx.buf);
drivers/firmware/arm_scmi/driver.c
1810
i->resp = i->t->rx.buf;
drivers/firmware/arm_scmi/driver.c
1836
st->rx_len = i->t->rx.len;
drivers/firmware/arm_scmi/driver.c
1951
resp = t->rx.buf;
drivers/firmware/arm_scmi/driver.c
2653
xfer->rx.buf = devm_kcalloc(dev, sizeof(u8), desc->max_msg_size,
drivers/firmware/arm_scmi/driver.c
2655
if (!xfer->rx.buf)
drivers/firmware/arm_scmi/driver.c
2658
xfer->tx.buf = xfer->rx.buf;
drivers/firmware/arm_scmi/msg.c
106
xfer->rx.len = min_t(size_t, max_len,
drivers/firmware/arm_scmi/msg.c
110
memcpy(xfer->rx.buf, msg->msg_payload, xfer->rx.len);
drivers/firmware/arm_scmi/msg.c
47
return sizeof(struct scmi_msg_payld) + sizeof(__le32) + xfer->rx.len;
drivers/firmware/arm_scmi/msg.c
88
xfer->rx.len = min_t(size_t, xfer->rx.len,
drivers/firmware/arm_scmi/msg.c
92
memcpy(xfer->rx.buf, &msg->msg_payload[1], xfer->rx.len);
drivers/firmware/arm_scmi/perf.c
207
attr = t->rx.buf;
drivers/firmware/arm_scmi/perf.c
266
attr = t->rx.buf;
drivers/firmware/arm_scmi/perf.c
619
limits = t->rx.buf;
drivers/firmware/arm_scmi/perf.c
759
*level = get_unaligned_le32(t->rx.buf);
drivers/firmware/arm_scmi/pinctrl.c
139
attr = t->rx.buf;
drivers/firmware/arm_scmi/pinctrl.c
197
struct scmi_resp_pinctrl_attributes *rx;
drivers/firmware/arm_scmi/pinctrl.c
208
sizeof(*rx), &t);
drivers/firmware/arm_scmi/pinctrl.c
213
rx = t->rx.buf;
drivers/firmware/arm_scmi/pinctrl.c
220
*n_elems = NUM_ELEMS(rx->attributes);
drivers/firmware/arm_scmi/pinctrl.c
222
strscpy(name, rx->name, SCMI_SHORT_NAME_MAX_SIZE);
drivers/firmware/arm_scmi/pinctrl.c
224
ext_name_flag = !!EXT_NAME_FLAG(rx->attributes);
drivers/firmware/arm_scmi/power.c
124
attr = t->rx.buf;
drivers/firmware/arm_scmi/power.c
189
*state = get_unaligned_le32(t->rx.buf);
drivers/firmware/arm_scmi/power.c
89
attr = t->rx.buf;
drivers/firmware/arm_scmi/powercap.c
156
attributes = get_unaligned_le32(t->rx.buf);
drivers/firmware/arm_scmi/powercap.c
209
resp = t->rx.buf;
drivers/firmware/arm_scmi/powercap.c
329
*power_cap = get_unaligned_le32(t->rx.buf);
drivers/firmware/arm_scmi/powercap.c
392
resp = t->rx.buf;
drivers/firmware/arm_scmi/powercap.c
480
*pai = get_unaligned_le32(t->rx.buf);
drivers/firmware/arm_scmi/powercap.c
571
resp = t->rx.buf;
drivers/firmware/arm_scmi/protocols.h
131
struct scmi_msg rx;
drivers/firmware/arm_scmi/raw_mode.c
1334
msg_size = xfer->rx.len + sizeof(u32);
drivers/firmware/arm_scmi/raw_mode.c
1347
memcpy(++m, xfer->rx.buf, xfer->rx.len);
drivers/firmware/arm_scmi/raw_mode.c
1464
memset(xfer->rx.buf, 0x00, xfer->rx.len);
drivers/firmware/arm_scmi/raw_mode.c
1496
xfer.rx.len = raw->desc->max_msg_size;
drivers/firmware/arm_scmi/raw_mode.c
1497
xfer.rx.buf = kzalloc(xfer.rx.len, GFP_ATOMIC);
drivers/firmware/arm_scmi/raw_mode.c
1498
if (!xfer.rx.buf) {
drivers/firmware/arm_scmi/raw_mode.c
1513
kfree(xfer.rx.buf);
drivers/firmware/arm_scmi/raw_mode.c
244
struct scmi_msg rx;
drivers/firmware/arm_scmi/raw_mode.c
559
xfer->rx.len = raw->desc->max_msg_size;
drivers/firmware/arm_scmi/raw_mode.c
793
ret = scmi_raw_message_receive(rd->raw, rd->rx.buf, rd->rx.len,
drivers/firmware/arm_scmi/raw_mode.c
810
rd->rx.buf, rd->rx_size);
drivers/firmware/arm_scmi/raw_mode.c
915
rd->rx.len = raw->desc->max_msg_size + sizeof(u32);
drivers/firmware/arm_scmi/raw_mode.c
916
rd->rx.buf = kzalloc(rd->rx.len, GFP_KERNEL);
drivers/firmware/arm_scmi/raw_mode.c
917
if (!rd->rx.buf) {
drivers/firmware/arm_scmi/raw_mode.c
925
kfree(rd->rx.buf);
drivers/firmware/arm_scmi/raw_mode.c
944
kfree(rd->rx.buf);
drivers/firmware/arm_scmi/reset.c
127
attr = t->rx.buf;
drivers/firmware/arm_scmi/reset.c
87
attr = get_unaligned_le32(t->rx.buf);
drivers/firmware/arm_scmi/sensors.c
238
attr = t->rx.buf;
drivers/firmware/arm_scmi/sensors.c
796
*sensor_config = get_unaligned_le64(t->rx.buf);
drivers/firmware/arm_scmi/sensors.c
876
resp = t->rx.buf;
drivers/firmware/arm_scmi/sensors.c
887
*value = get_unaligned_le64(t->rx.buf);
drivers/firmware/arm_scmi/sensors.c
949
resp = t->rx.buf;
drivers/firmware/arm_scmi/sensors.c
965
resp_readings = t->rx.buf;
drivers/firmware/arm_scmi/shmem.c
144
xfer->rx.len = min_t(size_t, xfer->rx.len, len > 8 ? len - 8 : 0);
drivers/firmware/arm_scmi/shmem.c
147
copy_fromio(xfer->rx.buf, shmem->msg_payload + 4, xfer->rx.len);
drivers/firmware/arm_scmi/shmem.c
157
xfer->rx.len = min_t(size_t, max_len, len > 4 ? len - 4 : 0);
drivers/firmware/arm_scmi/shmem.c
160
copy_fromio(xfer->rx.buf, shmem->msg_payload, xfer->rx.len);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-bbm.c
102
attr = t->rx.buf;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-bbm.c
284
*value = get_unaligned_le64(t->rx.buf);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-bbm.c
331
*state = get_unaligned_le32(t->rx.buf);
drivers/firmware/arm_scmi/vendors/imx/imx-sm-cpu.c
163
out = t->rx.buf;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-cpu.c
192
attr = t->rx.buf;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-cpu.c
221
out = t->rx.buf;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-lmm.c
207
attr = t->rx.buf;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-lmm.c
90
out = t->rx.buf;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-misc.c
117
attr = t->rx.buf;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-misc.c
260
out = t->rx.buf;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-misc.c
264
*num * sizeof(__le32) > t->rx.len - sizeof(__le32)) {
drivers/firmware/arm_scmi/vendors/imx/imx-sm-misc.c
326
out = t->rx.buf;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-misc.c
352
out = t->rx.buf;
drivers/firmware/arm_scmi/vendors/imx/imx-sm-misc.c
376
out = t->rx.buf;
drivers/firmware/arm_scmi/voltage.c
222
resp_dom = td->rx.buf;
drivers/firmware/arm_scmi/voltage.c
281
*value = get_unaligned_le32(t->rx.buf);
drivers/firmware/arm_scmi/voltage.c
354
resp = t->rx.buf;
drivers/firmware/arm_scmi/voltage.c
87
NUM_VOLTAGE_DOMAINS(get_unaligned_le32(t->rx.buf));
drivers/firmware/samsung/exynos-acpm.c
145
struct acpm_queue rx;
drivers/firmware/samsung/exynos-acpm.c
230
rx_front = readl(achan->rx.front);
drivers/firmware/samsung/exynos-acpm.c
231
i = readl(achan->rx.rear);
drivers/firmware/samsung/exynos-acpm.c
240
base = achan->rx.base;
drivers/firmware/samsung/exynos-acpm.c
283
writel(rx_front, achan->rx.rear);
drivers/firmware/samsung/exynos-acpm.c
347
ret = readl_poll_timeout(achan->rx.front, val, next_tx_front != val, 0,
drivers/firmware/samsung/exynos-acpm.c
476
struct acpm_queue *rx = &achan->rx;
drivers/firmware/samsung/exynos-acpm.c
488
rx->base = base + readl(&chan_shmem->tx_base);
drivers/firmware/samsung/exynos-acpm.c
489
rx->rear = base + readl(&chan_shmem->tx_rear);
drivers/firmware/samsung/exynos-acpm.c
490
rx->front = base + readl(&chan_shmem->tx_front);
drivers/firmware/tegra/bpmp-debugfs.c
120
.rx = {
drivers/firmware/tegra/bpmp-debugfs.c
137
else if (msg.rx.ret < 0)
drivers/firmware/tegra/bpmp-debugfs.c
161
.rx = {
drivers/firmware/tegra/bpmp-debugfs.c
171
else if (msg.rx.ret < 0)
drivers/firmware/tegra/bpmp-debugfs.c
190
.rx = {
drivers/firmware/tegra/bpmp-debugfs.c
215
} else if (msg.rx.ret < 0) {
drivers/firmware/tegra/bpmp-debugfs.c
255
.rx = {
drivers/firmware/tegra/bpmp-debugfs.c
285
} else if (msg.rx.ret < 0) {
drivers/firmware/tegra/bpmp-debugfs.c
318
.rx = {
drivers/firmware/tegra/bpmp-debugfs.c
342
} else if (msg.rx.ret < 0) {
drivers/firmware/tegra/bpmp-debugfs.c
505
.rx = {
drivers/firmware/tegra/bpmp-debugfs.c
515
else if (msg.rx.ret < 0)
drivers/firmware/tegra/bpmp-debugfs.c
564
.rx = {
drivers/firmware/tegra/bpmp-debugfs.c
574
else if (msg.rx.ret < 0)
drivers/firmware/tegra/bpmp-tegra186.c
113
struct iosys_map rx, tx;
drivers/firmware/tegra/bpmp-tegra186.c
126
if (priv->rx.pool) {
drivers/firmware/tegra/bpmp-tegra186.c
127
iosys_map_set_vaddr_iomem(&rx, priv->rx.sram + offset);
drivers/firmware/tegra/bpmp-tegra186.c
130
iosys_map_set_vaddr(&rx, priv->rx.dram + offset);
drivers/firmware/tegra/bpmp-tegra186.c
134
err = tegra_ivc_init(channel->ivc, NULL, &rx, priv->rx.phys + offset, &tx,
drivers/firmware/tegra/bpmp-tegra186.c
188
gen_pool_free(priv->rx.pool, (unsigned long)priv->rx.sram, 4096);
drivers/firmware/tegra/bpmp-tegra186.c
216
priv->rx.phys = res.start + SZ_4K;
drivers/firmware/tegra/bpmp-tegra186.c
226
priv->rx.dram = priv->tx.dram + SZ_4K;
drivers/firmware/tegra/bpmp-tegra186.c
249
priv->rx.pool = of_gen_pool_get(bpmp->dev->of_node, "shmem", 1);
drivers/firmware/tegra/bpmp-tegra186.c
250
if (!priv->rx.pool) {
drivers/firmware/tegra/bpmp-tegra186.c
256
priv->rx.sram = (void __iomem *)gen_pool_dma_alloc(priv->rx.pool, 4096,
drivers/firmware/tegra/bpmp-tegra186.c
257
&priv->rx.phys);
drivers/firmware/tegra/bpmp-tegra186.c
258
if (!priv->rx.sram) {
drivers/firmware/tegra/bpmp-tegra186.c
28
} tx, rx;
drivers/firmware/tegra/bpmp.c
353
return __tegra_bpmp_channel_read(channel, msg->rx.data, msg->rx.size,
drivers/firmware/tegra/bpmp.c
354
&msg->rx.ret);
drivers/firmware/tegra/bpmp.c
394
return tegra_bpmp_channel_read(channel, msg->rx.data, msg->rx.size,
drivers/firmware/tegra/bpmp.c
395
&msg->rx.ret);
drivers/firmware/tegra/bpmp.c
524
.rx = {
drivers/firmware/tegra/bpmp.c
532
if (err || msg.rx.ret)
drivers/firmware/tegra/bpmp.c
572
msg.rx.data = &response;
drivers/firmware/tegra/bpmp.c
573
msg.rx.size = sizeof(response);
drivers/firmware/tegra/bpmp.c
636
.rx = {
drivers/firmware/tegra/bpmp.c
650
if (msg.rx.ret < 0)
drivers/firmware/tegra/bpmp.c
90
(msg->rx.size <= MSG_DATA_MIN_SZ) &&
drivers/firmware/tegra/bpmp.c
92
(msg->rx.size == 0 || msg->rx.data);
drivers/firmware/tegra/ivc.c
103
u32 rx = tegra_ivc_header_read_field(map, rx.count);
drivers/firmware/tegra/ivc.c
115
if (tx - rx > ivc->num_frames)
drivers/firmware/tegra/ivc.c
118
return tx == rx;
drivers/firmware/tegra/ivc.c
124
u32 rx = tegra_ivc_header_read_field(map, rx.count);
drivers/firmware/tegra/ivc.c
130
return tx - rx >= ivc->num_frames;
drivers/firmware/tegra/ivc.c
136
u32 rx = tegra_ivc_header_read_field(map, rx.count);
drivers/firmware/tegra/ivc.c
144
return tx - rx;
drivers/firmware/tegra/ivc.c
161
unsigned int count = tegra_ivc_header_read_field(&ivc->rx.map, rx.count);
drivers/firmware/tegra/ivc.c
163
tegra_ivc_header_write_field(&ivc->rx.map, rx.count, count + 1);
drivers/firmware/tegra/ivc.c
165
if (ivc->rx.position == ivc->num_frames - 1)
drivers/firmware/tegra/ivc.c
166
ivc->rx.position = 0;
drivers/firmware/tegra/ivc.c
168
ivc->rx.position++;
drivers/firmware/tegra/ivc.c
195
if (!tegra_ivc_empty(ivc, &ivc->rx.map))
drivers/firmware/tegra/ivc.c
198
tegra_ivc_invalidate(ivc, ivc->rx.phys + offset);
drivers/firmware/tegra/ivc.c
200
if (tegra_ivc_empty(ivc, &ivc->rx.map))
drivers/firmware/tegra/ivc.c
208
unsigned int offset = offsetof(struct tegra_ivc_header, rx.count);
drivers/firmware/tegra/ivc.c
296
tegra_ivc_invalidate_frame(ivc, ivc->rx.phys, ivc->rx.position, 0,
drivers/firmware/tegra/ivc.c
299
return tegra_ivc_frame_virt(ivc, &ivc->rx.map, ivc->rx.position, map);
drivers/firmware/tegra/ivc.c
305
unsigned int rx = offsetof(struct tegra_ivc_header, rx.count);
drivers/firmware/tegra/ivc.c
320
tegra_ivc_flush(ivc, ivc->rx.phys + rx);
drivers/firmware/tegra/ivc.c
333
tegra_ivc_invalidate(ivc, ivc->rx.phys + tx);
drivers/firmware/tegra/ivc.c
335
if (tegra_ivc_available(ivc, &ivc->rx.map) == ivc->num_frames - 1)
drivers/firmware/tegra/ivc.c
359
unsigned int rx = offsetof(struct tegra_ivc_header, rx.count);
drivers/firmware/tegra/ivc.c
389
tegra_ivc_invalidate(ivc, ivc->tx.phys + rx);
drivers/firmware/tegra/ivc.c
434
tegra_ivc_invalidate(ivc, ivc->rx.phys + offset);
drivers/firmware/tegra/ivc.c
435
rx_state = tegra_ivc_header_read_field(&ivc->rx.map, tx.state);
drivers/firmware/tegra/ivc.c
453
tegra_ivc_header_write_field(&ivc->rx.map, rx.count, 0);
drivers/firmware/tegra/ivc.c
456
ivc->rx.position = 0;
drivers/firmware/tegra/ivc.c
492
tegra_ivc_header_write_field(&ivc->rx.map, rx.count, 0);
drivers/firmware/tegra/ivc.c
495
ivc->rx.position = 0;
drivers/firmware/tegra/ivc.c
574
static int tegra_ivc_check_params(unsigned long rx, unsigned long tx,
drivers/firmware/tegra/ivc.c
579
BUILD_BUG_ON(!IS_ALIGNED(offsetof(struct tegra_ivc_header, rx.count),
drivers/firmware/tegra/ivc.c
598
if (!IS_ALIGNED(rx, TEGRA_IVC_ALIGN)) {
drivers/firmware/tegra/ivc.c
599
pr_err("IVC channel start not aligned: %#lx\n", rx);
drivers/firmware/tegra/ivc.c
608
if (rx < tx) {
drivers/firmware/tegra/ivc.c
609
if (rx + frame_size * num_frames > tx) {
drivers/firmware/tegra/ivc.c
611
rx, frame_size * num_frames, tx);
drivers/firmware/tegra/ivc.c
615
if (tx + frame_size * num_frames > rx) {
drivers/firmware/tegra/ivc.c
617
tx, frame_size * num_frames, rx);
drivers/firmware/tegra/ivc.c
646
int tegra_ivc_init(struct tegra_ivc *ivc, struct device *peer, const struct iosys_map *rx,
drivers/firmware/tegra/ivc.c
665
err = tegra_ivc_check_params(iosys_map_get_address(rx), iosys_map_get_address(tx),
drivers/firmware/tegra/ivc.c
673
ivc->rx.phys = dma_map_single(peer, iosys_map_get_vaddr(rx), queue_size,
drivers/firmware/tegra/ivc.c
675
if (dma_mapping_error(peer, ivc->rx.phys))
drivers/firmware/tegra/ivc.c
68
} rx;
drivers/firmware/tegra/ivc.c
681
dma_unmap_single(peer, ivc->rx.phys, queue_size,
drivers/firmware/tegra/ivc.c
686
ivc->rx.phys = rx_phys;
drivers/firmware/tegra/ivc.c
690
iosys_map_copy(&ivc->rx.map, rx);
drivers/firmware/tegra/ivc.c
703
ivc->rx.position = 0;
drivers/firmware/tegra/ivc.c
715
dma_unmap_single(ivc->peer, ivc->rx.phys, size,
drivers/fpga/machxo2-spi.c
69
struct spi_transfer rx, tx;
drivers/fpga/machxo2-spi.c
73
memset(&rx, 0, sizeof(rx));
drivers/fpga/machxo2-spi.c
77
rx.rx_buf = status;
drivers/fpga/machxo2-spi.c
78
rx.len = 4;
drivers/fpga/machxo2-spi.c
81
spi_message_add_tail(&rx, &msg);
drivers/fpga/microchip-spi.c
47
u8 rx;
drivers/fpga/microchip-spi.c
60
.rx_buf = &priv->rx,
drivers/fpga/microchip-spi.c
65
.rx_buf = &priv->rx,
drivers/fpga/microchip-spi.c
78
status = priv->rx;
drivers/gpu/drm/amd/pm/powerplay/hwmgr/smu8_hwmgr.h
78
#define DDI_POWERGATING_ARG(phyID, lanemask, rx, tx, core) \
drivers/gpu/drm/amd/pm/powerplay/hwmgr/smu8_hwmgr.h
81
((rx) ? DISPLAYPHY_RX_SELECT : 0) | \
drivers/gpu/drm/mcde/mcde_dsi.c
274
u8 *rx = msg->rx_buf;
drivers/gpu/drm/mcde/mcde_dsi.c
286
rx[i] = (rddat >> (i * 8)) & 0xff;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
1138
msgqRxHeader rx;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
1171
cmdq->tx.rxHdrOff = offsetof(typeof(*cmdq), rx.readPtr);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
1177
gsp->cmdq.rptr = &msgq->rx.readPtr;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/rm/r535/gsp.c
1180
gsp->msgq.rptr = &cmdq->rx.readPtr;
drivers/gpu/drm/tegra/dsi.c
1210
u8 *rx = msg->rx_buf;
drivers/gpu/drm/tegra/dsi.c
1231
rx[0] = (value >> 8) & 0xff;
drivers/gpu/drm/tegra/dsi.c
1236
rx[0] = (value >> 8) & 0xff;
drivers/gpu/drm/tegra/dsi.c
1237
rx[1] = (value >> 16) & 0xff;
drivers/gpu/drm/tegra/dsi.c
1259
u8 *rx = msg->rx_buf + j;
drivers/gpu/drm/tegra/dsi.c
1264
rx[j + k] = (value >> (k << 3)) & 0xff;
drivers/gpu/drm/tiny/repaper.c
104
const void *tx, void *rx, size_t len)
drivers/gpu/drm/tiny/repaper.c
128
if (rx) {
drivers/gpu/drm/tiny/repaper.c
142
if (rx && !ret)
drivers/gpu/drm/tiny/repaper.c
143
memcpy(rx, rxbuf, len);
drivers/greybus/connection.c
671
static int _gb_connection_enable(struct gb_connection *connection, bool rx)
drivers/greybus/connection.c
677
if (!(connection->handler && rx))
drivers/greybus/connection.c
700
if (connection->handler && rx)
drivers/hid/hid-led.c
328
union delcom_packet dp = { .rx.cmd = 104 };
drivers/hid/hid-led.c
55
} rx;
drivers/hid/hid-playstation.c
1465
input_report_abs(ds->gamepad, ABS_RX, ds_report->rx);
drivers/hid/hid-playstation.c
238
u8 rx, ry;
drivers/hid/hid-playstation.c
2416
input_report_abs(ds4->gamepad, ABS_RX, ds4_report->rx);
drivers/hid/hid-playstation.c
462
u8 rx, ry;
drivers/hid/hid-wiimote-modules.c
1046
__s8 rx, ry, lx, ly, lt, rt;
drivers/hid/hid-wiimote-modules.c
1115
rx = (ext[0] >> 3) & 0x18;
drivers/hid/hid-wiimote-modules.c
1116
rx |= (ext[1] >> 5) & 0x06;
drivers/hid/hid-wiimote-modules.c
1117
rx |= (ext[2] >> 7) & 0x01;
drivers/hid/hid-wiimote-modules.c
1124
rx <<= 1;
drivers/hid/hid-wiimote-modules.c
1131
input_report_abs(wdata->extension.input, ABS_HAT2X, rx - 0x20);
drivers/hid/hid-wiimote-modules.c
1625
__s16 rx, ry, lx, ly;
drivers/hid/hid-wiimote-modules.c
1669
rx = (ext[2] & 0xff) | ((ext[3] & 0x0f) << 8);
drivers/hid/hid-wiimote-modules.c
1676
rx -= 0x800;
drivers/hid/hid-wiimote-modules.c
1690
if (abs(rx) < 500)
drivers/hid/hid-wiimote-modules.c
1691
wdata->state.calib_pro_sticks[2] = -rx;
drivers/hid/hid-wiimote-modules.c
1699
rx += wdata->state.calib_pro_sticks[2];
drivers/hid/hid-wiimote-modules.c
1704
input_report_abs(wdata->extension.input, ABS_RX, rx);
drivers/hwmon/ad7314.c
41
u16 rx ____cacheline_aligned;
drivers/hwmon/ad7314.c
48
ret = spi_read(chip->spi_dev, (u8 *)&chip->rx, sizeof(chip->rx));
drivers/hwmon/ad7314.c
54
return be16_to_cpu(chip->rx);
drivers/hwmon/sht21.c
163
u8 rx[8];
drivers/hwmon/sht21.c
176
.buf = rx,
drivers/hwmon/sht21.c
186
eic[2] = rx[0];
drivers/hwmon/sht21.c
187
eic[3] = rx[2];
drivers/hwmon/sht21.c
188
eic[4] = rx[4];
drivers/hwmon/sht21.c
189
eic[5] = rx[6];
drivers/hwmon/sht21.c
197
eic[0] = rx[3];
drivers/hwmon/sht21.c
198
eic[1] = rx[4];
drivers/hwmon/sht21.c
199
eic[6] = rx[0];
drivers/hwmon/sht21.c
200
eic[7] = rx[1];
drivers/i2c/busses/i2c-cpm.c
182
struct i2c_msg *pmsg, int num, int tx, int rx)
drivers/i2c/busses/i2c-cpm.c
192
rbdf = cpm->rbase + rx;
drivers/i2c/busses/i2c-cpm.c
197
rb = cpm->rxbuf[rx];
drivers/i2c/busses/i2c-cpm.c
225
if (rx + 1 == CPM_MAXBD)
drivers/i2c/busses/i2c-cpm.c
241
struct i2c_msg *pmsg, int tx, int rx)
drivers/i2c/busses/i2c-cpm.c
250
rbdf = cpm->rbase + rx;
drivers/i2c/busses/i2c-cpm.c
253
rb = cpm->rxbuf[rx];
drivers/i2c/busses/i2c-dln2.c
115
} __packed *rx = dln2->buf;
drivers/i2c/busses/i2c-dln2.c
116
unsigned rx_len = sizeof(*rx);
drivers/i2c/busses/i2c-dln2.c
118
BUILD_BUG_ON(sizeof(*rx) > DLN2_I2C_BUF_SIZE);
drivers/i2c/busses/i2c-dln2.c
127
rx, &rx_len);
drivers/i2c/busses/i2c-dln2.c
130
if (rx_len < sizeof(rx->buf_len) + data_len)
drivers/i2c/busses/i2c-dln2.c
132
if (le16_to_cpu(rx->buf_len) != data_len)
drivers/i2c/busses/i2c-dln2.c
135
memcpy(data, rx->buf, data_len);
drivers/i2c/busses/i2c-imx-lpi2c.c
952
struct dma_slave_config rx = {}, tx = {};
drivers/i2c/busses/i2c-imx-lpi2c.c
966
rx.src_addr = dma->phy_addr + LPI2C_MRDR;
drivers/i2c/busses/i2c-imx-lpi2c.c
967
rx.src_addr_width = DMA_SLAVE_BUSWIDTH_1_BYTE;
drivers/i2c/busses/i2c-imx-lpi2c.c
968
rx.src_maxburst = dma->rx_burst_num;
drivers/i2c/busses/i2c-imx-lpi2c.c
969
rx.direction = DMA_DEV_TO_MEM;
drivers/i2c/busses/i2c-imx-lpi2c.c
970
ret = dmaengine_slave_config(dma->chan_rx, &rx);
drivers/i2c/busses/i2c-tegra-bpmp.c
189
msg.rx.data = response;
drivers/i2c/busses/i2c-tegra-bpmp.c
190
msg.rx.size = sizeof(*response);
drivers/i2c/busses/i2c-tegra-bpmp.c
202
if (msg.rx.ret != 0) {
drivers/i2c/busses/i2c-tegra-bpmp.c
203
if (msg.rx.ret == -BPMP_EAGAIN) {
drivers/i2c/busses/i2c-tegra-bpmp.c
208
if (msg.rx.ret == -BPMP_ETIMEDOUT) {
drivers/i2c/busses/i2c-tegra-bpmp.c
213
if (msg.rx.ret == -BPMP_ENXIO) {
drivers/i2c/busses/i2c-tegra-bpmp.c
218
dev_err(i2c->dev, "transaction failed: %d\n", msg.rx.ret);
drivers/iio/accel/sca3000.c
1006
num_available = st->rx[0];
drivers/iio/accel/sca3000.c
1011
ret = sca3000_read_data(st, SCA3000_REG_RING_OUT_ADDR, st->rx,
drivers/iio/accel/sca3000.c
1023
iio_push_to_buffers(indio_dev, st->rx + i * 3 * 2);
drivers/iio/accel/sca3000.c
1055
val = st->rx[0];
drivers/iio/accel/sca3000.c
1121
ret = !!(st->rx[0] & SCA3000_REG_MODE_FREE_FALL_DETECT);
drivers/iio/accel/sca3000.c
1130
if ((st->rx[0] & SCA3000_REG_MODE_MODE_MASK)
drivers/iio/accel/sca3000.c
1163
if (state && !(st->rx[0] & SCA3000_REG_MODE_FREE_FALL_DETECT))
drivers/iio/accel/sca3000.c
1165
st->rx[0] | SCA3000_REG_MODE_FREE_FALL_DETECT);
drivers/iio/accel/sca3000.c
1167
else if (!state && (st->rx[0] & SCA3000_REG_MODE_FREE_FALL_DETECT))
drivers/iio/accel/sca3000.c
1169
st->rx[0] & ~SCA3000_REG_MODE_FREE_FALL_DETECT);
drivers/iio/accel/sca3000.c
1213
((st->rx[0] & SCA3000_REG_MODE_MODE_MASK)
drivers/iio/accel/sca3000.c
1216
(st->rx[0] & ~SCA3000_REG_MODE_MODE_MASK)
drivers/iio/accel/sca3000.c
1220
((st->rx[0] & SCA3000_REG_MODE_MODE_MASK)
drivers/iio/accel/sca3000.c
1223
st->rx[0] & SCA3000_REG_MODE_MODE_MASK);
drivers/iio/accel/sca3000.c
1288
(st->rx[0] | SCA3000_REG_MODE_RING_BUF_ENABLE));
drivers/iio/accel/sca3000.c
1292
(st->rx[0] & ~SCA3000_REG_MODE_RING_BUF_ENABLE));
drivers/iio/accel/sca3000.c
1321
st->rx[0] | SCA3000_REG_INT_MASK_RING_HALF);
drivers/iio/accel/sca3000.c
1352
st->rx[0] & ~SCA3000_REG_INT_MASK_RING_HALF);
drivers/iio/accel/sca3000.c
1421
(st->rx[0] & SCA3000_MODE_PROT_MASK));
drivers/iio/accel/sca3000.c
1512
(st->rx[0] &
drivers/iio/accel/sca3000.c
168
u8 rx[384] __aligned(IIO_DMA_MINALIGN);
drivers/iio/accel/sca3000.c
292
.rx_buf = st->rx,
drivers/iio/accel/sca3000.c
314
return !(st->rx[0] & SCA3000_LOCKED);
drivers/iio/accel/sca3000.c
413
return st->rx[0];
drivers/iio/accel/sca3000.c
433
st->rx[0] & SCA3000_REG_REVID_MAJOR_MASK,
drivers/iio/accel/sca3000.c
434
st->rx[0] & SCA3000_REG_REVID_MINOR_MASK);
drivers/iio/accel/sca3000.c
575
switch (SCA3000_REG_MODE_MODE_MASK & st->rx[0]) {
drivers/iio/accel/sca3000.c
666
st->rx[0] &= SCA3000_REG_MODE_MODE_MASK;
drivers/iio/accel/sca3000.c
667
switch (st->rx[0]) {
drivers/iio/accel/sca3000.c
703
st->rx[0] &= ~SCA3000_REG_MODE_MODE_MASK;
drivers/iio/accel/sca3000.c
704
st->rx[0] |= (mode & SCA3000_REG_MODE_MODE_MASK);
drivers/iio/accel/sca3000.c
706
return sca3000_write_reg(st, SCA3000_REG_MODE_ADDR, st->rx[0]);
drivers/iio/accel/sca3000.c
733
*val = sign_extend32(be16_to_cpup((__be16 *)st->rx) >>
drivers/iio/accel/sca3000.c
745
*val = (be16_to_cpup((__be16 *)st->rx) >>
drivers/iio/accel/sca3000.c
827
val = st->rx[0];
drivers/iio/accel/sca3000.c
965
u8 *rx,
drivers/iio/accel/sca3000.c
975
.rx_buf = rx,
drivers/iio/adc/ad7280a.c
188
__be32 rx;
drivers/iio/adc/ad7280a.c
232
.rx_buf = &st->rx,
drivers/iio/adc/ad7280a.c
242
*val = be32_to_cpu(st->rx);
drivers/iio/adc/ad7380.c
916
u16 rx;
drivers/iio/adc/ad7380.c
956
.rx_buf = &st->rx,
drivers/iio/adc/ad7380.c
969
*val = FIELD_GET(AD7380_REG_DATA, st->rx);
drivers/iio/adc/max1363.c
796
u8 rx;
drivers/iio/adc/max1363.c
800
st->recv(st->client, &rx, 1);
drivers/iio/adc/max1363.c
801
mask = rx;
drivers/iio/adc/ti-tsc2046.c
171
struct tsc2046_adc_atom *rx;
drivers/iio/adc/ti-tsc2046.c
395
val = tsc2046_adc_get_value(&priv->rx[l->offset + l->skip + i]);
drivers/iio/adc/ti-tsc2046.c
689
priv->rx = devm_kzalloc(&priv->spi->dev, size, GFP_KERNEL);
drivers/iio/adc/ti-tsc2046.c
690
if (!priv->rx)
drivers/iio/adc/ti-tsc2046.c
694
priv->xfer.rx_buf = priv->rx;
drivers/iio/gyro/adxrs450.c
101
.rx_buf = &st->rx,
drivers/iio/gyro/adxrs450.c
102
.len = sizeof(st->rx),
drivers/iio/gyro/adxrs450.c
120
*val = (be32_to_cpu(st->rx) >> 5) & 0xFFFF;
drivers/iio/gyro/adxrs450.c
173
.rx_buf = &st->rx,
drivers/iio/gyro/adxrs450.c
174
.len = sizeof(st->rx),
drivers/iio/gyro/adxrs450.c
187
*val = (be32_to_cpu(st->rx) >> 10) & 0xFFFF;
drivers/iio/gyro/adxrs450.c
207
.rx_buf = &st->rx,
drivers/iio/gyro/adxrs450.c
222
*val = be32_to_cpu(st->rx);
drivers/iio/gyro/adxrs450.c
77
__be32 rx;
drivers/iio/health/afe4403.c
218
u8 rx[3];
drivers/iio/health/afe4403.c
226
ret = spi_write_then_read(afe->spi, ®, 1, rx, sizeof(rx));
drivers/iio/health/afe4403.c
230
*val = get_unaligned_be24(&rx[0]);
drivers/iio/health/afe4403.c
315
u8 rx[3];
drivers/iio/health/afe4403.c
325
rx, sizeof(rx));
drivers/iio/health/afe4403.c
329
afe->buffer[i++] = get_unaligned_be24(&rx[0]);
drivers/iio/imu/adis.c
143
.rx_buf = adis->rx,
drivers/iio/imu/adis.c
149
.rx_buf = adis->rx + 2,
drivers/iio/imu/adis.c
191
*val = get_unaligned_be32(adis->rx);
drivers/iio/imu/adis.c
194
*val = get_unaligned_be16(adis->rx + 2);
drivers/iio/imu/adis_buffer.c
109
adis->xfer[j].rx_buf = &rx[j - 1];
drivers/iio/imu/adis_buffer.c
74
__be16 *tx, *rx;
drivers/iio/imu/adis_buffer.c
95
rx = adis->buffer;
drivers/iio/imu/adis_buffer.c
96
tx = rx + scan_count;
drivers/iio/imu/bno055/bno055_ser_core.c
156
} rx;
drivers/iio/imu/bno055/bno055_ser_core.c
348
priv->rx.databuf_count != priv->expected_data_len) {
drivers/iio/imu/bno055/bno055_ser_core.c
392
switch (priv->rx.state) {
drivers/iio/imu/bno055/bno055_ser_core.c
404
priv->rx.type = buf[0];
drivers/iio/imu/bno055/bno055_ser_core.c
405
priv->rx.state = RX_START;
drivers/iio/imu/bno055/bno055_ser_core.c
408
priv->rx.databuf_count = 0;
drivers/iio/imu/bno055/bno055_ser_core.c
419
if (priv->rx.type == 0xEE) {
drivers/iio/imu/bno055/bno055_ser_core.c
427
priv->rx.state = RX_IDLE;
drivers/iio/imu/bno055/bno055_ser_core.c
432
priv->rx.state = RX_DATA;
drivers/iio/imu/bno055/bno055_ser_core.c
433
priv->rx.expected_len = buf[0];
drivers/iio/imu/bno055/bno055_ser_core.c
444
if (priv->rx.databuf_count + remaining > priv->rx.expected_len) {
drivers/iio/imu/bno055/bno055_ser_core.c
451
priv->rx.state = RX_IDLE;
drivers/iio/imu/bno055/bno055_ser_core.c
469
(priv->rx.databuf_count + remaining <= priv->expected_data_len))
drivers/iio/imu/bno055/bno055_ser_core.c
470
memcpy(priv->response_buf + priv->rx.databuf_count,
drivers/iio/imu/bno055/bno055_ser_core.c
474
priv->rx.databuf_count += remaining;
drivers/iio/imu/bno055/bno055_ser_core.c
480
if (priv->rx.databuf_count == priv->rx.expected_len) {
drivers/iio/imu/bno055/bno055_ser_core.c
482
priv->rx.state = RX_IDLE;
drivers/iio/pressure/mpl115_spi.c
20
u8 rx[4];
drivers/iio/pressure/mpl115_spi.c
43
.rx_buf = buf->rx,
drivers/iio/pressure/mpl115_spi.c
55
return (buf->rx[1] << 8) | buf->rx[3];
drivers/iio/resolver/ad2s1200.c
43
__be16 rx __aligned(IIO_DMA_MINALIGN);
drivers/iio/resolver/ad2s1200.c
81
ret = spi_read(st->sdev, &st->rx, 2);
drivers/iio/resolver/ad2s1200.c
89
*val = be16_to_cpup(&st->rx) >> 4;
drivers/iio/resolver/ad2s1200.c
92
*val = sign_extend32(be16_to_cpup(&st->rx) >> 4, 11);
drivers/iio/resolver/ad2s1210.c
171
u8 rx[2];
drivers/iio/resolver/ad2s1210.c
202
.rx_buf = &st->rx[0],
drivers/iio/resolver/ad2s1210.c
207
.rx_buf = &st->rx[1],
drivers/iio/resolver/ad2s1210.c
248
.rx_buf = &st->rx[0],
drivers/iio/resolver/ad2s1210.c
253
.rx_buf = &st->rx[1],
drivers/iio/resolver/ad2s1210.c
285
&& st->rx[1] & AD2S1210_ADDRESS_DATA)
drivers/iio/resolver/ad2s1210.c
288
*val = st->rx[1];
drivers/iio/resolver/ad2s90.c
27
u8 rx[2] __aligned(IIO_DMA_MINALIGN);
drivers/iio/resolver/ad2s90.c
50
ret = spi_read(st->sdev, st->rx, 2);
drivers/iio/resolver/ad2s90.c
55
*val = (((u16)(st->rx[0])) << 4) | ((st->rx[1] & 0xF0) >> 4);
drivers/infiniband/hw/hfi1/chip.c
7312
u16 tx, rx;
drivers/infiniband/hw/hfi1/chip.c
7326
rx = nibble_to_count(enable_lane_rx);
drivers/infiniband/hw/hfi1/chip.c
7354
enable_lane_tx, tx, enable_lane_rx, rx);
drivers/infiniband/hw/hfi1/chip.c
7356
*rx_width = link_width_to_bits(dd, rx);
drivers/infiniband/hw/hfi1/chip.c
7375
u16 widths, tx, rx;
drivers/infiniband/hw/hfi1/chip.c
7381
rx = (widths >> 8) & 0xf;
drivers/infiniband/hw/hfi1/chip.c
7384
*rx_width = link_width_to_bits(dd, rx);
drivers/infiniband/hw/hfi1/chip.c
7605
u16 tx, rx;
drivers/infiniband/hw/hfi1/chip.c
7631
get_link_widths(ppd->dd, &tx, &rx);
drivers/infiniband/hw/hfi1/chip.c
7633
ppd->link_width_downgrade_rx_active = rx;
drivers/infiniband/hw/hfi1/netdev.h
24
struct hfi1_netdev_rx *rx;
drivers/infiniband/hw/hfi1/netdev_rx.c
121
static int hfi1_netdev_allot_ctxt(struct hfi1_netdev_rx *rx,
drivers/infiniband/hw/hfi1/netdev_rx.c
125
struct hfi1_devdata *dd = rx->dd;
drivers/infiniband/hw/hfi1/netdev_rx.c
133
rc = hfi1_netdev_setup_ctxt(rx, *ctxt);
drivers/infiniband/hw/hfi1/netdev_rx.c
186
static int hfi1_netdev_rxq_init(struct hfi1_netdev_rx *rx)
drivers/infiniband/hw/hfi1/netdev_rx.c
190
struct hfi1_devdata *dd = rx->dd;
drivers/infiniband/hw/hfi1/netdev_rx.c
191
struct net_device *dev = rx->rx_napi;
drivers/infiniband/hw/hfi1/netdev_rx.c
193
rx->num_rx_q = dd->num_netdev_contexts;
drivers/infiniband/hw/hfi1/netdev_rx.c
194
rx->rxq = kcalloc_node(rx->num_rx_q, sizeof(*rx->rxq),
drivers/infiniband/hw/hfi1/netdev_rx.c
197
if (!rx->rxq) {
drivers/infiniband/hw/hfi1/netdev_rx.c
20
static int hfi1_netdev_setup_ctxt(struct hfi1_netdev_rx *rx,
drivers/infiniband/hw/hfi1/netdev_rx.c
202
for (i = 0; i < rx->num_rx_q; i++) {
drivers/infiniband/hw/hfi1/netdev_rx.c
203
struct hfi1_netdev_rxq *rxq = &rx->rxq[i];
drivers/infiniband/hw/hfi1/netdev_rx.c
205
rc = hfi1_netdev_allot_ctxt(rx, &rxq->rcd);
drivers/infiniband/hw/hfi1/netdev_rx.c
210
rxq->rx = rx;
drivers/infiniband/hw/hfi1/netdev_rx.c
230
struct hfi1_netdev_rxq *rxq = &rx->rxq[i];
drivers/infiniband/hw/hfi1/netdev_rx.c
238
kfree(rx->rxq);
drivers/infiniband/hw/hfi1/netdev_rx.c
239
rx->rxq = NULL;
drivers/infiniband/hw/hfi1/netdev_rx.c
24
struct hfi1_devdata *dd = rx->dd;
drivers/infiniband/hw/hfi1/netdev_rx.c
244
static void hfi1_netdev_rxq_deinit(struct hfi1_netdev_rx *rx)
drivers/infiniband/hw/hfi1/netdev_rx.c
247
struct hfi1_devdata *dd = rx->dd;
drivers/infiniband/hw/hfi1/netdev_rx.c
249
for (i = 0; i < rx->num_rx_q; i++) {
drivers/infiniband/hw/hfi1/netdev_rx.c
250
struct hfi1_netdev_rxq *rxq = &rx->rxq[i];
drivers/infiniband/hw/hfi1/netdev_rx.c
258
kfree(rx->rxq);
drivers/infiniband/hw/hfi1/netdev_rx.c
259
rx->rxq = NULL;
drivers/infiniband/hw/hfi1/netdev_rx.c
260
rx->num_rx_q = 0;
drivers/infiniband/hw/hfi1/netdev_rx.c
263
static void enable_queues(struct hfi1_netdev_rx *rx)
drivers/infiniband/hw/hfi1/netdev_rx.c
267
for (i = 0; i < rx->num_rx_q; i++) {
drivers/infiniband/hw/hfi1/netdev_rx.c
268
struct hfi1_netdev_rxq *rxq = &rx->rxq[i];
drivers/infiniband/hw/hfi1/netdev_rx.c
270
dd_dev_info(rx->dd, "enabling queue %d on context %d\n", i,
drivers/infiniband/hw/hfi1/netdev_rx.c
273
hfi1_rcvctrl(rx->dd,
drivers/infiniband/hw/hfi1/netdev_rx.c
279
static void disable_queues(struct hfi1_netdev_rx *rx)
drivers/infiniband/hw/hfi1/netdev_rx.c
283
msix_netdev_synchronize_irq(rx->dd);
drivers/infiniband/hw/hfi1/netdev_rx.c
285
for (i = 0; i < rx->num_rx_q; i++) {
drivers/infiniband/hw/hfi1/netdev_rx.c
286
struct hfi1_netdev_rxq *rxq = &rx->rxq[i];
drivers/infiniband/hw/hfi1/netdev_rx.c
288
dd_dev_info(rx->dd, "disabling queue %d on context %d\n", i,
drivers/infiniband/hw/hfi1/netdev_rx.c
292
hfi1_rcvctrl(rx->dd,
drivers/infiniband/hw/hfi1/netdev_rx.c
309
struct hfi1_netdev_rx *rx = dd->netdev_rx;
drivers/infiniband/hw/hfi1/netdev_rx.c
312
if (atomic_fetch_inc(&rx->netdevs))
drivers/infiniband/hw/hfi1/netdev_rx.c
316
res = hfi1_netdev_rxq_init(rx);
drivers/infiniband/hw/hfi1/netdev_rx.c
329
struct hfi1_netdev_rx *rx = dd->netdev_rx;
drivers/infiniband/hw/hfi1/netdev_rx.c
332
if (atomic_fetch_add_unless(&rx->netdevs, -1, 0) == 1) {
drivers/infiniband/hw/hfi1/netdev_rx.c
334
hfi1_netdev_rxq_deinit(rx);
drivers/infiniband/hw/hfi1/netdev_rx.c
355
struct hfi1_netdev_rx *rx;
drivers/infiniband/hw/hfi1/netdev_rx.c
357
dd_dev_info(dd, "allocating rx size %ld\n", sizeof(*rx));
drivers/infiniband/hw/hfi1/netdev_rx.c
358
rx = kzalloc_node(sizeof(*rx), GFP_KERNEL, dd->node);
drivers/infiniband/hw/hfi1/netdev_rx.c
360
if (!rx)
drivers/infiniband/hw/hfi1/netdev_rx.c
362
rx->dd = dd;
drivers/infiniband/hw/hfi1/netdev_rx.c
363
rx->rx_napi = alloc_netdev_dummy(0);
drivers/infiniband/hw/hfi1/netdev_rx.c
364
if (!rx->rx_napi) {
drivers/infiniband/hw/hfi1/netdev_rx.c
365
kfree(rx);
drivers/infiniband/hw/hfi1/netdev_rx.c
369
xa_init(&rx->dev_tbl);
drivers/infiniband/hw/hfi1/netdev_rx.c
370
atomic_set(&rx->enabled, 0);
drivers/infiniband/hw/hfi1/netdev_rx.c
371
atomic_set(&rx->netdevs, 0);
drivers/infiniband/hw/hfi1/netdev_rx.c
372
dd->netdev_rx = rx;
drivers/infiniband/hw/hfi1/netdev_rx.c
398
struct hfi1_netdev_rx *rx;
drivers/infiniband/hw/hfi1/netdev_rx.c
403
rx = dd->netdev_rx;
drivers/infiniband/hw/hfi1/netdev_rx.c
404
if (atomic_fetch_inc(&rx->enabled))
drivers/infiniband/hw/hfi1/netdev_rx.c
408
enable_queues(rx);
drivers/infiniband/hw/hfi1/netdev_rx.c
414
struct hfi1_netdev_rx *rx;
drivers/infiniband/hw/hfi1/netdev_rx.c
419
rx = dd->netdev_rx;
drivers/infiniband/hw/hfi1/netdev_rx.c
420
if (atomic_dec_if_positive(&rx->enabled))
drivers/infiniband/hw/hfi1/netdev_rx.c
424
disable_queues(rx);
drivers/infiniband/hw/hfi1/netdev_rx.c
440
struct hfi1_netdev_rx *rx = dd->netdev_rx;
drivers/infiniband/hw/hfi1/netdev_rx.c
442
return xa_insert(&rx->dev_tbl, id, data, GFP_NOWAIT);
drivers/infiniband/hw/hfi1/netdev_rx.c
454
struct hfi1_netdev_rx *rx = dd->netdev_rx;
drivers/infiniband/hw/hfi1/netdev_rx.c
456
return xa_erase(&rx->dev_tbl, id);
drivers/infiniband/hw/hfi1/netdev_rx.c
467
struct hfi1_netdev_rx *rx = dd->netdev_rx;
drivers/infiniband/hw/hfi1/netdev_rx.c
469
return xa_load(&rx->dev_tbl, id);
drivers/infiniband/hw/hfi1/netdev_rx.c
480
struct hfi1_netdev_rx *rx = dd->netdev_rx;
drivers/infiniband/hw/hfi1/netdev_rx.c
484
ret = xa_find(&rx->dev_tbl, &index, UINT_MAX, XA_PRESENT);
drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c
288
has_data = pvrdma_idx_ring_has_data(&cq->ring_state->rx,
drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c
293
int tail = pvrdma_idx(&cq->ring_state->rx.prod_tail,
drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c
315
&cq->ring_state->rx.cons_head,
drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c
333
has_data = pvrdma_idx_ring_has_data(&cq->ring_state->rx,
drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c
374
pvrdma_idx_ring_inc(&cq->ring_state->rx.cons_head, cq->ibcq.cqe);
drivers/infiniband/hw/vmw_pvrdma/pvrdma_cq.c
82
has_data = pvrdma_idx_ring_has_data(&cq->ring_state->rx,
drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c
424
struct pvrdma_ring *ring = &dev->async_ring_state->rx;
drivers/infiniband/hw/vmw_pvrdma/pvrdma_main.c
500
struct pvrdma_ring *ring = &dev->cq_ring_state->rx;
drivers/infiniband/hw/vmw_pvrdma/pvrdma_ring.h
60
struct pvrdma_ring rx; /* Rx ring. */
drivers/infiniband/sw/siw/siw.h
470
#define rx_qp(rx) container_of(rx, struct siw_qp, rx_stream)
drivers/infiniband/ulp/ipoib/ipoib_cm.c
117
struct ipoib_cm_rx *rx,
drivers/infiniband/ulp/ipoib/ipoib_cm.c
127
sge[i].addr = rx->rx_ring[id].mapping[i];
drivers/infiniband/ulp/ipoib/ipoib_cm.c
129
ret = ib_post_recv(rx->qp, wr, NULL);
drivers/infiniband/ulp/ipoib/ipoib_cm.c
133
rx->rx_ring[id].mapping);
drivers/infiniband/ulp/ipoib/ipoib_cm.c
134
dev_kfree_skb_any(rx->rx_ring[id].skb);
drivers/infiniband/ulp/ipoib/ipoib_cm.c
135
rx->rx_ring[id].skb = NULL;
drivers/infiniband/ulp/ipoib/ipoib_cm.c
348
struct ipoib_cm_rx *rx)
drivers/infiniband/ulp/ipoib/ipoib_cm.c
358
rx->rx_ring = vzalloc(array_size(ipoib_recvq_size,
drivers/infiniband/ulp/ipoib/ipoib_cm.c
359
sizeof(*rx->rx_ring)));
drivers/infiniband/ulp/ipoib/ipoib_cm.c
360
if (!rx->rx_ring)
drivers/infiniband/ulp/ipoib/ipoib_cm.c
384
if (!ipoib_cm_alloc_rx_skb(dev, rx->rx_ring, i, IPOIB_CM_RX_SG - 1,
drivers/infiniband/ulp/ipoib/ipoib_cm.c
385
rx->rx_ring[i].mapping,
drivers/infiniband/ulp/ipoib/ipoib_cm.c
391
ret = ipoib_cm_post_receive_nonsrq(dev, rx, &t->wr, t->sge, i);
drivers/infiniband/ulp/ipoib/ipoib_cm.c
400
rx->recv_count = ipoib_recvq_size;
drivers/infiniband/ulp/ipoib/ipoib_cm.c
415
ipoib_cm_free_rx_ring(dev, rx->rx_ring);
drivers/infiniband/ulp/ipoib/ipoib_cm.c
907
struct ipoib_cm_rx *rx, *n;
drivers/infiniband/ulp/ipoib/ipoib_cm.c
914
list_for_each_entry_safe(rx, n, &list, list) {
drivers/infiniband/ulp/ipoib/ipoib_cm.c
915
ib_destroy_cm_id(rx->id);
drivers/infiniband/ulp/ipoib/ipoib_cm.c
916
ib_destroy_qp(rx->qp);
drivers/infiniband/ulp/ipoib/ipoib_cm.c
918
ipoib_cm_free_rx_ring(priv->dev, rx->rx_ring);
drivers/infiniband/ulp/ipoib/ipoib_cm.c
923
kfree(rx);
drivers/input/rmi4/rmi_f54.c
222
u8 rx = drv_data->num_rx_electrodes ? : f54->num_rx_electrodes;
drivers/input/rmi4/rmi_f54.c
228
size = rx * tx;
drivers/input/rmi4/rmi_f54.c
235
size = sizeof(u16) * rx * tx;
drivers/input/rmi4/rmi_f54.c
419
u8 rx = drv_data->num_rx_electrodes ? : f54->num_rx_electrodes;
drivers/input/rmi4/rmi_f54.c
435
f->width = rx;
drivers/input/rmi4/rmi_f54.c
663
u8 rx, tx;
drivers/input/rmi4/rmi_f54.c
679
rx = f54->num_rx_electrodes;
drivers/input/rmi4/rmi_f54.c
682
array3_size(tx, rx, sizeof(u16)),
drivers/input/serio/maceps2.c
78
byte = port->rx;
drivers/input/serio/ps2-gpio.c
169
byte = drvdata->rx.byte;
drivers/input/serio/ps2-gpio.c
170
cnt = drvdata->rx.cnt;
drivers/input/serio/ps2-gpio.c
265
drvdata->rx.cnt = cnt;
drivers/input/serio/ps2-gpio.c
266
drvdata->rx.byte = byte;
drivers/input/serio/ps2-gpio.c
84
} rx;
drivers/input/touchscreen/ads7846.c
1094
packet->rx = devm_kzalloc(&ts->spi->dev, size, GFP_KERNEL);
drivers/input/touchscreen/ads7846.c
1095
if (!packet->rx)
drivers/input/touchscreen/ads7846.c
1126
x->rx_buf = packet->rx;
drivers/input/touchscreen/ads7846.c
781
val = ads7846_get_value(&packet->rx[l->offset + b]);
drivers/input/touchscreen/ads7846.c
85
struct ads7846_buf *rx;
drivers/input/touchscreen/atmel_mxt_ts.c
2406
unsigned int i, rx, ry;
drivers/input/touchscreen/atmel_mxt_ts.c
2410
rx = data->xy_switch ? y : x;
drivers/input/touchscreen/atmel_mxt_ts.c
2412
rx = data->invertx ? (data->xsize - 1 - rx) : rx;
drivers/input/touchscreen/atmel_mxt_ts.c
2415
outbuf[i] = mxt_get_debug_value(data, rx, ry);
drivers/ipack/devices/ipoctal.c
118
stats->rx = 0;
drivers/ipack/devices/ipoctal.c
150
icount->rx = channel->stats.rx;
drivers/ipack/devices/ipoctal.h
29
unsigned long rx;
drivers/irqchip/irq-riscv-rpmi-sysmsi.c
104
struct rpmi_sysmsi_set_msi_target_rx rx;
drivers/irqchip/irq-riscv-rpmi-sysmsi.c
113
&tx, sizeof(tx), &rx, sizeof(rx));
drivers/irqchip/irq-riscv-rpmi-sysmsi.c
117
if (rx.status)
drivers/irqchip/irq-riscv-rpmi-sysmsi.c
118
return rpmi_to_linux_error(le32_to_cpu(rx.status));
drivers/irqchip/irq-riscv-rpmi-sysmsi.c
64
struct rpmi_sysmsi_get_attrs_rx rx;
drivers/irqchip/irq-riscv-rpmi-sysmsi.c
69
NULL, 0, &rx, sizeof(rx));
drivers/irqchip/irq-riscv-rpmi-sysmsi.c
73
if (rx.status)
drivers/irqchip/irq-riscv-rpmi-sysmsi.c
74
return rpmi_to_linux_error(le32_to_cpu(rx.status));
drivers/irqchip/irq-riscv-rpmi-sysmsi.c
76
return le32_to_cpu(rx.sys_num_msi);
drivers/irqchip/irq-riscv-rpmi-sysmsi.c
83
struct rpmi_sysmsi_set_msi_state_rx rx;
drivers/irqchip/irq-riscv-rpmi-sysmsi.c
90
&tx, sizeof(tx), &rx, sizeof(rx));
drivers/irqchip/irq-riscv-rpmi-sysmsi.c
94
if (rx.status)
drivers/irqchip/irq-riscv-rpmi-sysmsi.c
95
return rpmi_to_linux_error(le32_to_cpu(rx.status));
drivers/leds/trigger/ledtrig-netdev.c
439
DEFINE_NETDEV_TRIGGER(rx, TRIGGER_NETDEV_RX);
drivers/leds/trigger/ledtrig-tty.c
188
DEFINE_TTY_TRIGGER(rx, TRIGGER_TTY_RX);
drivers/leds/trigger/ledtrig-tty.c
19
int rx, tx;
drivers/leds/trigger/ledtrig-tty.c
270
if (trigger_data->mode_rx && (icount.rx != trigger_data->rx)) {
drivers/leds/trigger/ledtrig-tty.c
271
trigger_data->rx = icount.rx;
drivers/mailbox/bcm-pdc-mailbox.c
1005
dev_dbg(dev, " - base DMA addr of rx ring %pad", &rx.dmabase);
drivers/mailbox/bcm-pdc-mailbox.c
1006
dev_dbg(dev, " - base virtual addr of rx ring %p", rx.vbase);
drivers/mailbox/bcm-pdc-mailbox.c
1009
memcpy(&pdcs->rx_ring_alloc, &rx, sizeof(rx));
drivers/mailbox/bcm-pdc-mailbox.c
987
struct pdc_ring_alloc rx;
drivers/mailbox/bcm-pdc-mailbox.c
997
rx.vbase = dma_pool_zalloc(pdcs->ring_pool, GFP_KERNEL, &rx.dmabase);
drivers/mailbox/bcm-pdc-mailbox.c
998
if (unlikely(!rx.vbase)) {
drivers/mailbox/bcm74110-mailbox.c
130
BCM74110_OFFSET_IO_READL_MACRO(rx, BCM_MBOX_BASE(mbox->rx_chan));
drivers/mailbox/imx-mailbox.c
1009
.rx = imx_mu_specific_rx,
drivers/mailbox/imx-mailbox.c
1020
.rx = imx_mu_specific_rx,
drivers/mailbox/imx-mailbox.c
1031
.rx = imx_mu_specific_rx,
drivers/mailbox/imx-mailbox.c
1042
.rx = imx_mu_generic_rx,
drivers/mailbox/imx-mailbox.c
117
int (*rx)(struct imx_mu_priv *priv, struct imx_mu_con_priv *cp);
drivers/mailbox/imx-mailbox.c
556
priv->dcfg->rx(priv, cp);
drivers/mailbox/imx-mailbox.c
974
.rx = imx_mu_generic_rx,
drivers/mailbox/imx-mailbox.c
985
.rx = imx_mu_generic_rx,
drivers/mailbox/imx-mailbox.c
997
.rx = imx_mu_generic_rx,
drivers/mailbox/riscv-sbi-mpxy-mbox.c
239
void *rx, unsigned long max_rx_len,
drivers/mailbox/riscv-sbi-mpxy-mbox.c
259
if (rx && !sret.error) {
drivers/mailbox/riscv-sbi-mpxy-mbox.c
266
memcpy(rx, mpxy->shmem, rx_bytes);
drivers/media/cec/platform/s5p/s5p_cec.c
114
if (cec->rx != STATE_IDLE)
drivers/media/cec/platform/s5p/s5p_cec.c
116
cec->rx = STATE_BUSY;
drivers/media/cec/platform/s5p/s5p_cec.c
123
cec->rx = STATE_DONE;
drivers/media/cec/platform/s5p/s5p_cec.c
161
switch (cec->rx) {
drivers/media/cec/platform/s5p/s5p_cec.c
164
cec->rx = STATE_IDLE;
drivers/media/cec/platform/s5p/s5p_cec.h
71
enum cec_state rx;
drivers/media/dvb-frontends/dib7000p.c
2009
u8 *tx, *rx;
drivers/media/dvb-frontends/dib7000p.c
2019
rx = kzalloc(2, GFP_KERNEL);
drivers/media/dvb-frontends/dib7000p.c
2020
if (!rx) {
drivers/media/dvb-frontends/dib7000p.c
2026
msg[1].buf = rx;
drivers/media/dvb-frontends/dib7000p.c
2032
if (rx[0] == 0x01 && rx[1] == 0xb3) {
drivers/media/dvb-frontends/dib7000p.c
2041
if (rx[0] == 0x01 && rx[1] == 0xb3) {
drivers/media/dvb-frontends/dib7000p.c
2050
kfree(rx);
drivers/media/i2c/ds90ub960.c
3421
unsigned int rx, tx;
drivers/media/i2c/ds90ub960.c
3423
rx = ub960_pad_to_port(priv, route->sink_pad);
drivers/media/i2c/ds90ub960.c
3424
if (BIT(rx) & handled_mask)
drivers/media/i2c/ds90ub960.c
3429
vc[rx] = cur_vc[tx]++;
drivers/media/i2c/ds90ub960.c
3430
handled_mask |= BIT(rx);
drivers/media/i2c/st-mipid02.c
307
struct v4l2_fwnode_endpoint *ep = &bridge->rx;
drivers/media/i2c/st-mipid02.c
333
struct v4l2_fwnode_endpoint *ep = &bridge->rx;
drivers/media/i2c/st-mipid02.c
383
struct v4l2_fwnode_endpoint *ep = &bridge->rx;
drivers/media/i2c/st-mipid02.c
740
bridge->rx = ep;
drivers/media/i2c/st-mipid02.c
96
struct v4l2_fwnode_endpoint rx;
drivers/media/pci/cobalt/cobalt-irq.c
18
int rx = s->video_channel;
drivers/media/pci/cobalt/cobalt-irq.c
20
COBALT_CVI_FREEWHEEL(s->cobalt, rx);
drivers/media/pci/cobalt/cobalt-irq.c
22
COBALT_CVI_VMR(s->cobalt, rx);
drivers/media/pci/cobalt/cobalt-irq.c
24
COBALT_CVI(s->cobalt, rx);
drivers/media/pci/cobalt/cobalt-irq.c
26
COBALT_CVI_CLK_LOSS(s->cobalt, rx);
drivers/media/pci/cobalt/cobalt-v4l2.c
250
int rx = s->video_channel;
drivers/media/pci/cobalt/cobalt-v4l2.c
252
COBALT_CVI_EVCNT(cobalt, rx);
drivers/media/pci/cobalt/cobalt-v4l2.c
284
int rx = s->video_channel;
drivers/media/pci/cobalt/cobalt-v4l2.c
285
struct m00389_cvi_regmap __iomem *cvi = COBALT_CVI(cobalt, rx);
drivers/media/pci/cobalt/cobalt-v4l2.c
286
struct m00460_evcnt_regmap __iomem *evcnt = COBALT_CVI_EVCNT(cobalt, rx);
drivers/media/pci/cobalt/cobalt-v4l2.c
301
fw = COBALT_CVI_FREEWHEEL(cobalt, rx);
drivers/media/pci/cobalt/cobalt-v4l2.c
302
vmr = COBALT_CVI_VMR(cobalt, rx);
drivers/media/pci/cobalt/cobalt-v4l2.c
303
clkloss = COBALT_CVI_CLK_LOSS(cobalt, rx);
drivers/media/pci/cobalt/cobalt-v4l2.c
351
int rx = s->video_channel;
drivers/media/pci/cobalt/cobalt-v4l2.c
353
COBALT_CVI_EVCNT(cobalt, rx);
drivers/media/pci/cobalt/cobalt-v4l2.c
388
int rx = s->video_channel;
drivers/media/pci/cobalt/cobalt-v4l2.c
410
fw = COBALT_CVI_FREEWHEEL(cobalt, rx);
drivers/media/pci/cobalt/cobalt-v4l2.c
411
vmr = COBALT_CVI_VMR(cobalt, rx);
drivers/media/pci/cobalt/cobalt-v4l2.c
412
clkloss = COBALT_CVI_CLK_LOSS(cobalt, rx);
drivers/media/pci/cobalt/cobalt-v4l2.c
494
int rx = s->video_channel;
drivers/media/pci/cobalt/cobalt-v4l2.c
499
cvi = COBALT_CVI(cobalt, rx);
drivers/media/pci/cobalt/cobalt-v4l2.c
500
vmr = COBALT_CVI_VMR(cobalt, rx);
drivers/media/pci/cobalt/cobalt-v4l2.c
501
fw = COBALT_CVI_FREEWHEEL(cobalt, rx);
drivers/media/pci/cobalt/cobalt-v4l2.c
502
clkloss = COBALT_CVI_CLK_LOSS(cobalt, rx);
drivers/media/pci/cobalt/cobalt-v4l2.c
503
packer = COBALT_CVI_PACKER(cobalt, rx);
drivers/media/pci/cobalt/cobalt-v4l2.c
508
cobalt_info("rx%d: cvi resolution: %dx%d\n", rx,
drivers/media/pci/cobalt/cobalt-v4l2.c
510
cobalt_info("rx%d: cvi control: %s%s%s\n", rx,
drivers/media/pci/cobalt/cobalt-v4l2.c
517
cobalt_info("rx%d: cvi status: %s%s\n", rx,
drivers/media/pci/cobalt/cobalt-v4l2.c
523
cobalt_info("rx%d: Measurements: %s%s%s%s%s%s%s\n", rx,
drivers/media/pci/cobalt/cobalt-v4l2.c
538
cobalt_info("rx%d: irq_status: 0x%02x irq_triggers: 0x%02x\n", rx,
drivers/media/pci/cobalt/cobalt-v4l2.c
541
cobalt_info("rx%d: vsync: %d\n", rx, ioread32(&vmr->vsync_time));
drivers/media/pci/cobalt/cobalt-v4l2.c
542
cobalt_info("rx%d: vbp: %d\n", rx, ioread32(&vmr->vback_porch));
drivers/media/pci/cobalt/cobalt-v4l2.c
543
cobalt_info("rx%d: vact: %d\n", rx, ioread32(&vmr->vactive_area));
drivers/media/pci/cobalt/cobalt-v4l2.c
544
cobalt_info("rx%d: vfb: %d\n", rx, ioread32(&vmr->vfront_porch));
drivers/media/pci/cobalt/cobalt-v4l2.c
545
cobalt_info("rx%d: hsync: %d\n", rx, ioread32(&vmr->hsync_time));
drivers/media/pci/cobalt/cobalt-v4l2.c
546
cobalt_info("rx%d: hbp: %d\n", rx, ioread32(&vmr->hback_porch));
drivers/media/pci/cobalt/cobalt-v4l2.c
547
cobalt_info("rx%d: hact: %d\n", rx, ioread32(&vmr->hactive_area));
drivers/media/pci/cobalt/cobalt-v4l2.c
548
cobalt_info("rx%d: hfb: %d\n", rx, ioread32(&vmr->hfront_porch));
drivers/media/pci/cobalt/cobalt-v4l2.c
549
cobalt_info("rx%d: Freewheeling: %s%s%s\n", rx,
drivers/media/pci/cobalt/cobalt-v4l2.c
557
cobalt_info("rx%d: Clock Loss Detection: %s%s\n", rx,
drivers/media/pci/cobalt/cobalt-v4l2.c
562
cobalt_info("rx%d: Packer: %x\n", rx, ioread32(&packer->control));
drivers/media/pci/saa7134/saa7134-tvaudio.c
465
int max1,max2,carrier,rx,mode,lastmode,default_carrier;
drivers/media/pci/saa7134/saa7134-tvaudio.c
593
rx = tvaudio_getstereo(dev, &tvaudio[audio]);
drivers/media/pci/saa7134/saa7134-tvaudio.c
594
mode = saa7134_tvaudio_rx2mode(rx);
drivers/media/pci/saa7134/saa7134-tvaudio.c
895
int saa7134_tvaudio_rx2mode(u32 rx)
drivers/media/pci/saa7134/saa7134-tvaudio.c
900
if (rx & V4L2_TUNER_SUB_STEREO)
drivers/media/pci/saa7134/saa7134-tvaudio.c
902
else if (rx & V4L2_TUNER_SUB_LANG1)
drivers/media/pci/saa7134/saa7134-tvaudio.c
904
else if (rx & V4L2_TUNER_SUB_LANG2)
drivers/media/pci/saa7134/saa7134-video.c
1398
int rx, mode;
drivers/media/pci/saa7134/saa7134-video.c
1405
rx = saa7134_tvaudio_getstereo(dev);
drivers/media/pci/saa7134/saa7134-video.c
1406
mode = saa7134_tvaudio_rx2mode(rx);
drivers/media/pci/saa7134/saa7134.h
865
int saa7134_tvaudio_rx2mode(u32 rx);
drivers/media/platform/amphion/vpu.h
156
struct vpu_mbox rx;
drivers/media/platform/amphion/vpu_mbox.c
19
struct vpu_mbox *rx = container_of(cl, struct vpu_mbox, cl);
drivers/media/platform/amphion/vpu_mbox.c
20
struct vpu_core *core = container_of(rx, struct vpu_core, rx);
drivers/media/platform/amphion/vpu_mbox.c
64
scnprintf(core->rx.name, sizeof(core->rx.name) - 1, "rx");
drivers/media/platform/amphion/vpu_mbox.c
65
core->rx.block = true;
drivers/media/platform/amphion/vpu_mbox.c
80
ret = vpu_mbox_request_channel(core->dev, &core->rx);
drivers/media/platform/amphion/vpu_mbox.c
95
mbox_free_channel(core->rx.ch);
drivers/media/platform/amphion/vpu_mbox.c
98
core->rx.ch = NULL;
drivers/media/usb/dvb-usb/dib0700.h
60
int dib0700_ctrl_rd(struct dvb_usb_device *d, u8 *tx, u8 txlen, u8 *rx,
drivers/media/usb/dvb-usb/dib0700_core.c
103
debug_dump(rx, rxlen, deb_data);
drivers/media/usb/dvb-usb/dib0700_core.c
71
int dib0700_ctrl_rd(struct dvb_usb_device *d, u8 *tx, u8 txlen, u8 *rx, u8 rxlen)
drivers/media/usb/dvb-usb/dib0700_core.c
96
USB_TYPE_VENDOR | USB_DIR_IN, value, index, rx, rxlen,
drivers/media/usb/dvb-usb/pctv452e.c
519
u8 *b0, *rx;
drivers/media/usb/dvb-usb/pctv452e.c
534
rx = b0 + 5;
drivers/media/usb/dvb-usb/pctv452e.c
549
ret = dvb_usb_generic_rw(d, b0, 5, rx, PCTV_ANSWER_LEN, 0);
drivers/media/usb/dvb-usb/pctv452e.c
556
ret = dvb_usb_generic_rw(d, b0, 5, rx, PCTV_ANSWER_LEN, 0);
drivers/media/usb/dvb-usb/pctv452e.c
570
u8 *b, *rx;
drivers/media/usb/dvb-usb/pctv452e.c
578
rx = b + CMD_BUFFER_SIZE;
drivers/media/usb/dvb-usb/pctv452e.c
589
ret = dvb_usb_generic_rw(d, b, 4, rx, PCTV_ANSWER_LEN, 0);
drivers/media/usb/dvb-usb/pctv452e.c
594
info("%s: read: %2d: %*ph: ", __func__, ret, 3, rx);
drivers/media/usb/dvb-usb/pctv452e.c
595
for (i = 0; (i < rx[3]) && ((i+3) < PCTV_ANSWER_LEN); i++)
drivers/media/usb/dvb-usb/pctv452e.c
596
info(" %02x", rx[i+3]);
drivers/media/usb/dvb-usb/pctv452e.c
601
if ((rx[3] == 9) && (rx[12] & 0x01)) {
drivers/media/usb/dvb-usb/pctv452e.c
603
state->last_rc_key = RC_SCANCODE_RC5(rx[7], rx[6]);
drivers/media/usb/dvb-usb/pctv452e.c
606
__func__, rx[6], rx[7]);
drivers/media/usb/dvb-usb/technisat-usb2.c
112
if (rx != NULL) {
drivers/media/usb/dvb-usb/technisat-usb2.c
147
if (rx != NULL) {
drivers/media/usb/dvb-usb/technisat-usb2.c
148
memcpy(rx, &b[2], rxlen);
drivers/media/usb/dvb-usb/technisat-usb2.c
151
debug_dump(rx, rxlen, deb_i2c);
drivers/media/usb/dvb-usb/technisat-usb2.c
85
u8 device_addr, u8 *tx, u8 txlen, u8 *rx, u8 rxlen)
drivers/media/usb/dvb-usb/ttusb2.c
123
u8 rx[60];/* (64 -4) */
drivers/media/usb/dvb-usb/ttusb2.c
124
ret = ttusb2_msg(d, cmd, data, write_len, rx, read_len);
drivers/media/usb/dvb-usb/ttusb2.c
126
memcpy(data, rx, read_len);
drivers/media/usb/dvb-usb/ttusb2.c
449
u8 rx[9]; /* A CMD_GET_IR_CODE reply is 9 bytes long */
drivers/media/usb/dvb-usb/ttusb2.c
451
ret = ttusb2_msg(d, CMD_GET_IR_CODE, NULL, 0, rx, sizeof(rx));
drivers/media/usb/dvb-usb/ttusb2.c
455
if (rx[8] & 0x01) {
drivers/media/usb/dvb-usb/ttusb2.c
457
st->last_rc_key = RC_SCANCODE_RC5(rx[3], rx[2]);
drivers/media/usb/dvb-usb/ttusb2.c
458
deb_info("%s: cmd=0x%02x sys=0x%02x\n", __func__, rx[2], rx[3]);
drivers/media/usb/dvb-usb/ttusb2.c
459
rc_keydown(d->rc_dev, RC_PROTO_RC5, st->last_rc_key, rx[1]);
drivers/media/usb/hackrf/hackrf.c
241
const bool rx = test_bit(RX_ON, &dev->flags);
drivers/media/usb/hackrf/hackrf.c
264
if (!rx && !tx) {
drivers/media/usb/hackrf/hackrf.c
270
if (rx && test_and_clear_bit(RX_ADC_FREQUENCY, &dev->flags)) {
drivers/media/usb/hackrf/hackrf.c
298
if (rx && test_and_clear_bit(RX_BANDWIDTH, &dev->flags)) {
drivers/media/usb/hackrf/hackrf.c
346
if (rx && test_and_clear_bit(RX_RF_FREQUENCY, &dev->flags)) {
drivers/media/usb/hackrf/hackrf.c
374
if (rx && test_and_clear_bit(RX_RF_GAIN, &dev->flags)) {
drivers/media/usb/hackrf/hackrf.c
398
if (rx && test_and_clear_bit(RX_LNA_GAIN, &dev->flags)) {
drivers/media/usb/hackrf/hackrf.c
409
if (rx && test_and_clear_bit(RX_IF_GAIN, &dev->flags)) {
drivers/memory/tegra/tegra186-emc.c
164
msg.rx.data = &response;
drivers/memory/tegra/tegra186-emc.c
165
msg.rx.size = sizeof(response);
drivers/memory/tegra/tegra186-emc.c
172
if (msg.rx.ret < 0) {
drivers/memory/tegra/tegra186-emc.c
173
dev_err(emc->dev, "EMC DVFS MRQ failed: %d (BPMP error code)\n", msg.rx.ret);
drivers/memory/tegra/tegra234.c
1076
msg.rx.data = &bwmgr_resp;
drivers/memory/tegra/tegra234.c
1077
msg.rx.size = sizeof(bwmgr_resp);
drivers/memory/tegra/tegra234.c
1088
if (msg.rx.ret < 0) {
drivers/memory/tegra/tegra234.c
1090
bwmgr_req.bwmgr_calc_set_req.client_id, msg.rx.ret);
drivers/memory/tegra/tegra264.c
245
msg.rx.data = &bwmgr_resp;
drivers/memory/tegra/tegra264.c
246
msg.rx.size = sizeof(bwmgr_resp);
drivers/memory/tegra/tegra264.c
253
if (msg.rx.ret < 0) {
drivers/memory/tegra/tegra264.c
255
bwmgr_req.bwmgr_calc_set_req.client_id, msg.rx.ret);
drivers/mfd/ipaq-micro.c
141
struct ipaq_micro_rxdev *rx = µ->rx;
drivers/mfd/ipaq-micro.c
143
switch (rx->state) {
drivers/mfd/ipaq-micro.c
146
rx->state = STATE_ID; /* Next byte is the id and len */
drivers/mfd/ipaq-micro.c
149
rx->id = (ch & 0xf0) >> 4;
drivers/mfd/ipaq-micro.c
150
rx->len = (ch & 0x0f);
drivers/mfd/ipaq-micro.c
151
rx->index = 0;
drivers/mfd/ipaq-micro.c
152
rx->chksum = ch;
drivers/mfd/ipaq-micro.c
153
rx->state = (rx->len > 0) ? STATE_DATA : STATE_CHKSUM;
drivers/mfd/ipaq-micro.c
156
rx->chksum += ch;
drivers/mfd/ipaq-micro.c
157
rx->buf[rx->index] = ch;
drivers/mfd/ipaq-micro.c
158
if (++rx->index == rx->len)
drivers/mfd/ipaq-micro.c
159
rx->state = STATE_CHKSUM;
drivers/mfd/ipaq-micro.c
162
if (ch == rx->chksum)
drivers/mfd/ipaq-micro.c
163
micro_rx_msg(micro, rx->id, rx->len, rx->buf);
drivers/mfd/ipaq-micro.c
164
rx->state = STATE_SOF;
drivers/mfd/ipaq-micro.c
297
struct ipaq_micro_rxdev *rx = µ->rx;
drivers/mfd/ipaq-micro.c
304
rx->state = STATE_SOF; /* Reset the state machine */
drivers/mfd/qnap-mcu.c
217
unsigned char rx[QNAP_MCU_RX_BUFFER_SIZE];
drivers/mfd/qnap-mcu.c
222
if (length > sizeof(rx)) {
drivers/mfd/qnap-mcu.c
229
reply->data = rx;
drivers/mfd/qnap-mcu.c
245
if (!qnap_mcu_verify_checksum(rx, reply->received)) {
drivers/mfd/qnap-mcu.c
250
if (qnap_mcu_reply_is_any_error(mcu, rx, reply->received))
drivers/mfd/qnap-mcu.c
253
memcpy(reply_data, rx, reply_data_size);
drivers/mfd/qnap-mcu.c
282
u8 rx[14];
drivers/mfd/qnap-mcu.c
286
ret = qnap_mcu_exec(mcu, cmd, sizeof(cmd), rx, QNAP_MCU_VERSION_LEN + 2);
drivers/mfd/qnap-mcu.c
290
memcpy(mcu->version, &rx[2], QNAP_MCU_VERSION_LEN);
drivers/mfd/timberdale.c
302
.rx = true,
drivers/mfd/timberdale.c
308
.rx = false,
drivers/mfd/timberdale.c
314
.rx = true,
drivers/mfd/timberdale.c
320
.rx = false,
drivers/mfd/timberdale.c
326
.rx = true,
drivers/mfd/timberdale.c
336
.rx = true,
drivers/mfd/timberdale.c
343
.rx = true,
drivers/mfd/timberdale.c
349
.rx = false,
drivers/mmc/core/sdio_uart.c
362
port->icount.rx++;
drivers/mmc/core/sdio_uart.c
56
__u32 rx;
drivers/mmc/core/sdio_uart.c
956
port->icount.tx, port->icount.rx);
drivers/mmc/host/rtsx_pci_sdmmc.c
615
u8 sample_point, bool rx)
drivers/mmc/host/rtsx_pci_sdmmc.c
620
__func__, rx ? "RX" : "TX", sample_point);
drivers/mmc/host/rtsx_pci_sdmmc.c
623
if (rx) {
drivers/net/arcnet/arc-rawmode.c
156
.rx = rx,
drivers/net/arcnet/arcdevice.h
199
void (*rx)(struct net_device *dev, int bufnum,
drivers/net/arcnet/arcnet.c
1175
arc_proto_map[soft->proto]->rx(dev, bufnum, &rxdata.pkt, length);
drivers/net/arcnet/arcnet.c
93
.rx = null_rx,
drivers/net/arcnet/capmode.c
235
.rx = rx,
drivers/net/arcnet/rfc1051.c
40
static void rx(struct net_device *dev, int bufnum,
drivers/net/arcnet/rfc1051.c
51
.rx = rx,
drivers/net/arcnet/rfc1201.c
42
static void rx(struct net_device *dev, int bufnum,
drivers/net/arcnet/rfc1201.c
54
.rx = rx,
drivers/net/can/grcan.c
1185
slot = dma->rx.buf + rd;
drivers/net/can/grcan.c
1213
rd = grcan_ring_add(rd, GRCAN_MSG_SIZE, dma->rx.size);
drivers/net/can/grcan.c
226
struct grcan_dma_buffer rx;
drivers/net/can/grcan.c
938
struct grcan_dma_buffer *large = rsize > tsize ? &dma->rx : &dma->tx;
drivers/net/can/grcan.c
939
struct grcan_dma_buffer *small = rsize > tsize ? &dma->tx : &dma->rx;
drivers/net/can/grcan.c
962
dma->rx.size = rsize;
drivers/net/can/grcan.c
987
grcan_write_reg(®s->rxaddr, priv->dma.rx.handle);
drivers/net/can/grcan.c
988
grcan_write_reg(®s->rxsize, priv->dma.rx.size);
drivers/net/can/mscan/mscan.c
294
can_id = in_be16(®s->rx.idr1_0);
drivers/net/can/mscan/mscan.c
297
can_id = ((can_id << 16) | in_be16(®s->rx.idr3_2));
drivers/net/can/mscan/mscan.c
309
frame->len = can_cc_dlc2len(in_8(®s->rx.dlr) & 0xf);
drivers/net/can/mscan/mscan.c
312
void __iomem *data = ®s->rx.dsr1_0;
drivers/net/can/mscan/mscan.h
198
} rx;
drivers/net/can/spi/mcp251xfd/mcp251xfd-core.c
1461
priv->regs_status.rxif = BIT(priv->rx[0]->fifo_nr);
drivers/net/can/spi/mcp251xfd/mcp251xfd-dump.c
153
const struct mcp251xfd_rx_ring *rx)
drivers/net/can/spi/mcp251xfd/mcp251xfd-dump.c
158
.val = rx->head,
drivers/net/can/spi/mcp251xfd/mcp251xfd-dump.c
161
.val = rx->tail,
drivers/net/can/spi/mcp251xfd/mcp251xfd-dump.c
164
.val = rx->base,
drivers/net/can/spi/mcp251xfd/mcp251xfd-dump.c
167
.val = rx->nr,
drivers/net/can/spi/mcp251xfd/mcp251xfd-dump.c
170
.val = rx->fifo_nr,
drivers/net/can/spi/mcp251xfd/mcp251xfd-dump.c
173
.val = rx->obj_num,
drivers/net/can/spi/mcp251xfd/mcp251xfd-dump.c
176
.val = rx->obj_size,
drivers/net/can/spi/mcp251xfd/mcp251xfd-ram.c
109
num_rx = can_ram_rounddown_pow_of_two(config, &config->rx, 0, num_rx);
drivers/net/can/spi/mcp251xfd/mcp251xfd-ram.c
121
(u32)config->rx.fifo_depth_coalesce_min,
drivers/net/can/spi/mcp251xfd/mcp251xfd-ram.c
125
num_rx = can_ram_rounddown_pow_of_two(config, &config->rx,
drivers/net/can/spi/mcp251xfd/mcp251xfd-ram.c
129
ram_free = config->size - config->rx.size[fd_mode] * num_rx;
drivers/net/can/spi/mcp251xfd/mcp251xfd-ram.c
77
num_rx = ram_free / config->rx.size[fd_mode];
drivers/net/can/spi/mcp251xfd/mcp251xfd-ram.c
79
layout->default_rx = can_ram_rounddown_pow_of_two(config, &config->rx, 0, num_rx);
drivers/net/can/spi/mcp251xfd/mcp251xfd-ram.c
86
num_rx = ram_free / config->rx.size[fd_mode];
drivers/net/can/spi/mcp251xfd/mcp251xfd-ram.c
89
ram_free -= config->rx.size[fd_mode] * config->rx.min;
drivers/net/can/spi/mcp251xfd/mcp251xfd-ram.c
92
layout->max_rx = can_ram_rounddown_pow_of_two(config, &config->rx, 0, num_rx);
drivers/net/can/spi/mcp251xfd/mcp251xfd-ram.h
35
const struct can_ram_obj_config rx;
drivers/net/can/spi/mcp251xfd/mcp251xfd-ring.c
317
priv->regs_status.rxif = BIT(priv->rx[0]->fifo_nr);
drivers/net/can/spi/mcp251xfd/mcp251xfd-ring.c
407
for (i = ARRAY_SIZE(priv->rx) - 1; i >= 0; i--) {
drivers/net/can/spi/mcp251xfd/mcp251xfd-ring.c
408
kfree(priv->rx[i]);
drivers/net/can/spi/mcp251xfd/mcp251xfd-ring.c
409
priv->rx[i] = NULL;
drivers/net/can/spi/mcp251xfd/mcp251xfd-ring.c
417
struct mcp251xfd_rx_ring *ring = priv->rx[0];
drivers/net/can/spi/mcp251xfd/mcp251xfd-ring.c
442
.rx = {
drivers/net/can/spi/mcp251xfd/mcp251xfd-ring.c
518
for (i = 0; i < ARRAY_SIZE(priv->rx) && rem; i++) {
drivers/net/can/spi/mcp251xfd/mcp251xfd-ring.c
540
priv->rx[i] = rx_ring;
drivers/net/can/spi/mcp251xfd/mcp251xfd.h
646
struct mcp251xfd_rx_ring *rx[MCP251XFD_FIFO_RX_NUM];
drivers/net/can/spi/mcp251xfd/mcp251xfd.h
951
for ((n) = 0, (ring) = *((priv)->rx + (n)); \
drivers/net/can/spi/mcp251xfd/mcp251xfd.h
953
(n)++, (ring) = *((priv)->rx + (n)))
drivers/net/can/usb/ems_usb.c
186
u8 rx;
drivers/net/can/usb/esd_usb.c
246
struct esd_usb_rx_msg rx;
drivers/net/can/usb/esd_usb.c
306
u32 id = le32_to_cpu(msg->rx.id) & ESD_USB_IDMASK;
drivers/net/can/usb/esd_usb.c
309
u8 state = msg->rx.ev_can_err_ext.status;
drivers/net/can/usb/esd_usb.c
310
u8 ecc = msg->rx.ev_can_err_ext.ecc;
drivers/net/can/usb/esd_usb.c
312
priv->bec.rxerr = msg->rx.ev_can_err_ext.rec;
drivers/net/can/usb/esd_usb.c
313
priv->bec.txerr = msg->rx.ev_can_err_ext.tec;
drivers/net/can/usb/esd_usb.c
317
msg->rx.dlc, state, ecc,
drivers/net/can/usb/esd_usb.c
410
id = le32_to_cpu(msg->rx.id);
drivers/net/can/usb/esd_usb.c
415
if (msg->rx.dlc & ESD_USB_FD) {
drivers/net/can/usb/esd_usb.c
429
if (msg->rx.dlc & ESD_USB_FD) {
drivers/net/can/usb/esd_usb.c
431
cfd->len = can_fd_dlc2len(msg->rx.dlc);
drivers/net/can/usb/esd_usb.c
433
if ((msg->rx.dlc & ESD_USB_NO_BRS) == 0)
drivers/net/can/usb/esd_usb.c
435
if (msg->rx.dlc & ESD_USB_ESI)
drivers/net/can/usb/esd_usb.c
438
can_frame_set_cc_len(cf, msg->rx.dlc & ~ESD_USB_RTR, priv->can.ctrlmode);
drivers/net/can/usb/esd_usb.c
440
if (msg->rx.dlc & ESD_USB_RTR) {
drivers/net/can/usb/esd_usb.c
449
memcpy(cfd->data, msg->rx.data_fd, len);
drivers/net/can/usb/esd_usb.c
515
if (msg->rx.net >= dev->net_count) {
drivers/net/can/usb/esd_usb.c
520
esd_usb_rx_can_msg(dev->nets[msg->rx.net], msg);
drivers/net/can/usb/nct6694_canfd.c
141
struct nct6694_canfd_frame rx;
drivers/net/can/usb/nct6694_canfd.c
212
struct nct6694_canfd_frame *frame = &priv->rx;
drivers/net/can/usb/peak_usb/pcan_usb_pro.c
534
struct pcan_usb_pro_rxmsg *rx)
drivers/net/can/usb/peak_usb/pcan_usb_pro.c
536
const unsigned int ctrl_idx = (rx->len >> 4) & 0x0f;
drivers/net/can/usb/peak_usb/pcan_usb_pro.c
547
can_frame->can_id = le32_to_cpu(rx->id);
drivers/net/can/usb/peak_usb/pcan_usb_pro.c
548
can_frame->len = rx->len & 0x0f;
drivers/net/can/usb/peak_usb/pcan_usb_pro.c
550
if (rx->flags & PCAN_USBPRO_EXT)
drivers/net/can/usb/peak_usb/pcan_usb_pro.c
553
if (rx->flags & PCAN_USBPRO_RTR) {
drivers/net/can/usb/peak_usb/pcan_usb_pro.c
556
memcpy(can_frame->data, rx->data, can_frame->len);
drivers/net/can/usb/peak_usb/pcan_usb_pro.c
563
peak_usb_get_ts_time(&usb_if->time_ref, le32_to_cpu(rx->ts32),
drivers/net/dsa/microchip/ksz_common.c
2315
stats->rx_bytes = raw->rx + raw->rx_hi - stats->rx_packets * ETH_FCS_LEN;
drivers/net/dsa/microchip/ksz_common.c
81
u64 rx;
drivers/net/dsa/mxl862xx/mxl862xx-api.h
618
u8 rx; /* enum mxl862xx_ss_sp_tag_rx */
drivers/net/dsa/mxl862xx/mxl862xx.c
252
.rx = enable ? MXL862XX_SS_SP_TAG_RX_TAG_NO_INSERT :
drivers/net/ethernet/3com/typhoon.c
1631
struct rx_desc *rx;
drivers/net/ethernet/3com/typhoon.c
1646
rx = (struct rx_desc *) (rxRing->ringBase + rxaddr);
drivers/net/ethernet/3com/typhoon.c
1647
idx = rx->addr;
drivers/net/ethernet/3com/typhoon.c
1654
if (rx->flags & TYPHOON_RX_ERROR) {
drivers/net/ethernet/3com/typhoon.c
1659
pkt_len = le16_to_cpu(rx->frameLen);
drivers/net/ethernet/3com/typhoon.c
1680
csum_bits = rx->rxStatus & (TYPHOON_RX_IP_CHK_GOOD |
drivers/net/ethernet/3com/typhoon.c
1690
if (rx->rxStatus & TYPHOON_RX_VLAN)
drivers/net/ethernet/3com/typhoon.c
1692
ntohl(rx->vlanTag) & 0xffff);
drivers/net/ethernet/amazon/ena/ena_ethtool.c
39
ENA_STAT_ENTRY(stat, rx)
drivers/net/ethernet/amd/sunlance.c
1357
lp->rx = lance_rx_pio;
drivers/net/ethernet/amd/sunlance.c
1369
lp->rx = lance_rx_dvma;
drivers/net/ethernet/amd/sunlance.c
256
void (*rx)(struct net_device *);
drivers/net/ethernet/amd/sunlance.c
831
lp->rx(dev);
drivers/net/ethernet/amd/xgbe/xgbe-desc.c
268
xgbe_set_buffer_data(&rdata->rx.hdr, &ring->rx_hdr_pa,
drivers/net/ethernet/amd/xgbe/xgbe-desc.c
271
xgbe_set_buffer_data(&rdata->rx.hdr, &ring->rx_hdr_pa,
drivers/net/ethernet/amd/xgbe/xgbe-desc.c
276
xgbe_set_buffer_data(&rdata->rx.buf, &ring->rx_buf_pa,
drivers/net/ethernet/amd/xgbe/xgbe-desc.c
386
if (rdata->rx.hdr.pa.pages)
drivers/net/ethernet/amd/xgbe/xgbe-desc.c
387
put_page(rdata->rx.hdr.pa.pages);
drivers/net/ethernet/amd/xgbe/xgbe-desc.c
389
if (rdata->rx.hdr.pa_unmap.pages) {
drivers/net/ethernet/amd/xgbe/xgbe-desc.c
390
dma_unmap_page(pdata->dev, rdata->rx.hdr.pa_unmap.pages_dma,
drivers/net/ethernet/amd/xgbe/xgbe-desc.c
391
rdata->rx.hdr.pa_unmap.pages_len,
drivers/net/ethernet/amd/xgbe/xgbe-desc.c
393
put_page(rdata->rx.hdr.pa_unmap.pages);
drivers/net/ethernet/amd/xgbe/xgbe-desc.c
396
if (rdata->rx.buf.pa.pages)
drivers/net/ethernet/amd/xgbe/xgbe-desc.c
397
put_page(rdata->rx.buf.pa.pages);
drivers/net/ethernet/amd/xgbe/xgbe-desc.c
399
if (rdata->rx.buf.pa_unmap.pages) {
drivers/net/ethernet/amd/xgbe/xgbe-desc.c
400
dma_unmap_page(pdata->dev, rdata->rx.buf.pa_unmap.pages_dma,
drivers/net/ethernet/amd/xgbe/xgbe-desc.c
401
rdata->rx.buf.pa_unmap.pages_len,
drivers/net/ethernet/amd/xgbe/xgbe-desc.c
403
put_page(rdata->rx.buf.pa_unmap.pages);
drivers/net/ethernet/amd/xgbe/xgbe-desc.c
407
memset(&rdata->rx, 0, sizeof(rdata->rx));
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
1506
hdr_dma = rdata->rx.hdr.dma_base + rdata->rx.hdr.dma_off;
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
1507
buf_dma = rdata->rx.buf.dma_base + rdata->rx.buf.dma_off;
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
1874
rdata->rx.hdr_len = XGMAC_GET_BITS_LE(rdesc->desc2,
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
1876
if (rdata->rx.hdr_len)
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
1912
rdata->rx.len = XGMAC_GET_BITS_LE(rdesc->desc3, RX_NORMAL_DESC3, PL);
drivers/net/ethernet/amd/xgbe/xgbe-dev.c
1978
pdata->ext_stats.rxq_bytes[channel->queue_index] += rdata->rx.len;
drivers/net/ethernet/amd/xgbe/xgbe-drv.c
2090
skb = napi_alloc_skb(napi, rdata->rx.hdr.dma_len);
drivers/net/ethernet/amd/xgbe/xgbe-drv.c
2097
dma_sync_single_range_for_cpu(pdata->dev, rdata->rx.hdr.dma_base,
drivers/net/ethernet/amd/xgbe/xgbe-drv.c
2098
rdata->rx.hdr.dma_off,
drivers/net/ethernet/amd/xgbe/xgbe-drv.c
2099
rdata->rx.hdr.dma_len, DMA_FROM_DEVICE);
drivers/net/ethernet/amd/xgbe/xgbe-drv.c
2101
packet = page_address(rdata->rx.hdr.pa.pages) +
drivers/net/ethernet/amd/xgbe/xgbe-drv.c
2102
rdata->rx.hdr.pa.pages_offset;
drivers/net/ethernet/amd/xgbe/xgbe-drv.c
2117
if (rdata->rx.hdr_len)
drivers/net/ethernet/amd/xgbe/xgbe-drv.c
2118
return rdata->rx.hdr_len;
drivers/net/ethernet/amd/xgbe/xgbe-drv.c
2124
return rdata->rx.hdr.dma_len;
drivers/net/ethernet/amd/xgbe/xgbe-drv.c
2129
return min_t(unsigned int, rdata->rx.hdr.dma_len, rdata->rx.len);
drivers/net/ethernet/amd/xgbe/xgbe-drv.c
2138
return rdata->rx.buf.dma_len;
drivers/net/ethernet/amd/xgbe/xgbe-drv.c
2143
return rdata->rx.len - len;
drivers/net/ethernet/amd/xgbe/xgbe-drv.c
2298
if (buf2_len > rdata->rx.buf.dma_len) {
drivers/net/ethernet/amd/xgbe/xgbe-drv.c
2317
rdata->rx.buf.dma_base,
drivers/net/ethernet/amd/xgbe/xgbe-drv.c
2318
rdata->rx.buf.dma_off,
drivers/net/ethernet/amd/xgbe/xgbe-drv.c
2319
rdata->rx.buf.dma_len,
drivers/net/ethernet/amd/xgbe/xgbe-drv.c
2323
rdata->rx.buf.pa.pages,
drivers/net/ethernet/amd/xgbe/xgbe-drv.c
2324
rdata->rx.buf.pa.pages_offset,
drivers/net/ethernet/amd/xgbe/xgbe-drv.c
2326
rdata->rx.buf.dma_len);
drivers/net/ethernet/amd/xgbe/xgbe-drv.c
2327
rdata->rx.buf.pa.pages = NULL;
drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c
555
unsigned int rx, tx;
drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c
580
rx = __rounddown_pow_of_two(ringparam->rx_pending);
drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c
581
if (rx != ringparam->rx_pending)
drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c
584
rx);
drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c
592
if ((rx == pdata->rx_desc_count) &&
drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c
596
pdata->rx_desc_count = rx;
drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c
609
unsigned int rx, tx, combined;
drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c
617
rx = min(pdata->hw_feat.rx_ch_cnt, pdata->rx_max_channel_count);
drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c
618
rx = min(rx, pdata->channel_irq_count);
drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c
623
combined = min(rx, tx);
drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c
626
channels->max_rx = rx ? rx - 1 : 0;
drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c
630
rx = pdata->new_rx_ring_count ? : pdata->rx_ring_count;
drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c
633
combined = min(rx, tx);
drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c
634
rx -= combined;
drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c
638
channels->rx_count = rx;
drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c
654
unsigned int rx, rx_curr, tx, tx_curr, combined;
drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c
662
rx = min(pdata->hw_feat.rx_ch_cnt, pdata->rx_max_channel_count);
drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c
663
rx = min(rx, pdata->channel_irq_count);
drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c
668
combined = min(rx, tx);
drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c
703
if ((channels->combined_count + channels->rx_count) > rx) {
drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c
706
channels->combined_count + channels->rx_count, rx);
drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c
719
rx = channels->combined_count + channels->rx_count;
drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c
725
if ((rx == rx_curr) && (tx == tx_curr))
drivers/net/ethernet/amd/xgbe/xgbe-ethtool.c
728
pdata->new_rx_ring_count = rx;
drivers/net/ethernet/amd/xgbe/xgbe.h
353
struct xgbe_rx_ring_data rx; /* Rx-related data */
drivers/net/ethernet/apm/xgene/xgene_enet_hw.c
601
u32 *rx, u32 *tx)
drivers/net/ethernet/apm/xgene/xgene_enet_hw.c
606
*rx = ICM_DROP_COUNT(count);
drivers/net/ethernet/apm/xgene/xgene_enet_main.h
149
void (*get_drop_cnt)(struct xgene_enet_pdata *pdata, u32 *rx, u32 *tx);
drivers/net/ethernet/apm/xgene/xgene_enet_sgmac.c
87
u32 *rx, u32 *tx)
drivers/net/ethernet/apm/xgene/xgene_enet_sgmac.c
95
*rx = ICM_DROP_COUNT(count);
drivers/net/ethernet/apm/xgene/xgene_enet_xgmac.c
178
u32 *rx, u32 *tx)
drivers/net/ethernet/apm/xgene/xgene_enet_xgmac.c
183
*rx = ICM_DROP_COUNT(count);
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
100
u64_stats_update_begin(&self->stats.rx.syncp);
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
101
self->stats.rx.pg_flips++;
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
102
u64_stats_update_end(&self->stats.rx.syncp);
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
110
u64_stats_update_begin(&self->stats.rx.syncp);
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
111
self->stats.rx.pg_losts++;
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
112
u64_stats_update_end(&self->stats.rx.syncp);
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
116
u64_stats_update_begin(&self->stats.rx.syncp);
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
117
self->stats.rx.pg_reuses++;
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
118
u64_stats_update_end(&self->stats.rx.syncp);
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
125
u64_stats_update_begin(&self->stats.rx.syncp);
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
126
self->stats.rx.alloc_fails++;
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
127
u64_stats_update_end(&self->stats.rx.syncp);
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
238
u64_stats_init(&self->stats.rx.syncp);
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
347
u64_stats_update_begin(&self->stats.rx.syncp);
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
348
++self->stats.rx.errors;
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
349
u64_stats_update_end(&self->stats.rx.syncp);
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
417
u64_stats_update_begin(&rx_ring->stats.rx.syncp);
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
418
++rx_ring->stats.rx.packets;
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
419
rx_ring->stats.rx.bytes += xdp_get_buff_len(xdp);
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
420
u64_stats_update_end(&rx_ring->stats.rx.syncp);
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
438
u64_stats_update_begin(&rx_ring->stats.rx.syncp);
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
439
++rx_ring->stats.rx.xdp_pass;
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
440
u64_stats_update_end(&rx_ring->stats.rx.syncp);
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
450
u64_stats_update_begin(&rx_ring->stats.rx.syncp);
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
451
++rx_ring->stats.rx.xdp_tx;
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
452
u64_stats_update_end(&rx_ring->stats.rx.syncp);
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
459
u64_stats_update_begin(&rx_ring->stats.rx.syncp);
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
460
++rx_ring->stats.rx.xdp_redirect;
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
461
u64_stats_update_end(&rx_ring->stats.rx.syncp);
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
468
u64_stats_update_begin(&rx_ring->stats.rx.syncp);
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
469
++rx_ring->stats.rx.xdp_aborted;
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
470
u64_stats_update_end(&rx_ring->stats.rx.syncp);
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
475
u64_stats_update_begin(&rx_ring->stats.rx.syncp);
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
476
++rx_ring->stats.rx.xdp_drop;
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
477
u64_stats_update_end(&rx_ring->stats.rx.syncp);
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
597
u64_stats_update_begin(&self->stats.rx.syncp);
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
598
++self->stats.rx.errors;
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
599
u64_stats_update_end(&self->stats.rx.syncp);
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
605
u64_stats_update_begin(&self->stats.rx.syncp);
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
606
++self->stats.rx.errors;
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
607
u64_stats_update_end(&self->stats.rx.syncp);
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
618
u64_stats_update_begin(&self->stats.rx.syncp);
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
619
self->stats.rx.skb_alloc_fails++;
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
620
u64_stats_update_end(&self->stats.rx.syncp);
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
691
u64_stats_update_begin(&self->stats.rx.syncp);
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
692
++self->stats.rx.packets;
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
693
self->stats.rx.bytes += skb->len;
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
694
u64_stats_update_end(&self->stats.rx.syncp);
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
770
u64_stats_update_begin(&rx_ring->stats.rx.syncp);
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
771
++rx_ring->stats.rx.errors;
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
772
u64_stats_update_end(&rx_ring->stats.rx.syncp);
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
778
u64_stats_update_begin(&rx_ring->stats.rx.syncp);
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
779
++rx_ring->stats.rx.errors;
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
780
u64_stats_update_end(&rx_ring->stats.rx.syncp);
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
803
u64_stats_update_begin(&rx_ring->stats.rx.syncp);
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
804
++rx_ring->stats.rx.packets;
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
805
rx_ring->stats.rx.bytes += xdp_get_buff_len(&xdp);
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
806
++rx_ring->stats.rx.xdp_aborted;
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
807
u64_stats_update_end(&rx_ring->stats.rx.syncp);
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
949
start = u64_stats_fetch_begin(&self->stats.rx.syncp);
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
950
data[count] = self->stats.rx.packets;
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
951
data[++count] = self->stats.rx.jumbo_packets;
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
952
data[++count] = self->stats.rx.lro_packets;
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
953
data[++count] = self->stats.rx.errors;
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
954
data[++count] = self->stats.rx.alloc_fails;
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
955
data[++count] = self->stats.rx.skb_alloc_fails;
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
956
data[++count] = self->stats.rx.polls;
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
957
data[++count] = self->stats.rx.pg_flips;
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
958
data[++count] = self->stats.rx.pg_reuses;
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
959
data[++count] = self->stats.rx.pg_losts;
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
960
data[++count] = self->stats.rx.xdp_aborted;
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
961
data[++count] = self->stats.rx.xdp_drop;
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
962
data[++count] = self->stats.rx.xdp_pass;
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
963
data[++count] = self->stats.rx.xdp_tx;
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
964
data[++count] = self->stats.rx.xdp_invalid;
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
965
data[++count] = self->stats.rx.xdp_redirect;
drivers/net/ethernet/aquantia/atlantic/aq_ring.c
966
} while (u64_stats_fetch_retry(&self->stats.rx.syncp, start));
drivers/net/ethernet/aquantia/atlantic/aq_ring.h
126
struct aq_ring_stats_rx_s rx;
drivers/net/ethernet/aquantia/atlantic/aq_vec.c
43
u64_stats_update_begin(&ring[AQ_VEC_RX_ID].stats.rx.syncp);
drivers/net/ethernet/aquantia/atlantic/aq_vec.c
44
ring[AQ_VEC_RX_ID].stats.rx.polls++;
drivers/net/ethernet/aquantia/atlantic/aq_vec.c
45
u64_stats_update_end(&ring[AQ_VEC_RX_ID].stats.rx.syncp);
drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c
718
++ring->stats.rx.jumbo_packets;
drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c
992
++ring->stats.rx.lro_packets;
drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c
998
++ring->stats.rx.jumbo_packets;
drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_llh.c
176
u32 rx)
drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_llh.c
210
aq_hw_write_reg_bit(aq_hw, itr_imr_rxren_adr[rx],
drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_llh.c
211
itr_imr_rxren_msk[rx],
drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_llh.c
212
itr_imr_rxren_shift[rx],
drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_llh.c
257
void hw_atl_itr_irq_map_rx_set(struct aq_hw_s *aq_hw, u32 irq_map_rx, u32 rx)
drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_llh.c
291
aq_hw_write_reg_bit(aq_hw, itr_imr_rxr_adr[rx],
drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_llh.c
292
itr_imr_rxr_msk[rx],
drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_llh.c
293
itr_imr_rxr_shift[rx],
drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_llh.h
142
u32 rx);
drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_llh.h
149
void hw_atl_itr_irq_map_rx_set(struct aq_hw_s *aq_hw, u32 irq_map_rx, u32 rx);
drivers/net/ethernet/atheros/ag71xx.c
1185
if (ring->buf[i].rx.rx_buf) {
drivers/net/ethernet/atheros/ag71xx.c
1187
ring->buf[i].rx.dma_addr,
drivers/net/ethernet/atheros/ag71xx.c
1189
skb_free_frag(ring->buf[i].rx.rx_buf);
drivers/net/ethernet/atheros/ag71xx.c
1213
buf->rx.rx_buf = data;
drivers/net/ethernet/atheros/ag71xx.c
1214
buf->rx.dma_addr = dma_map_single(&ag->pdev->dev, data, ag->rx_buf_size,
drivers/net/ethernet/atheros/ag71xx.c
1216
if (dma_mapping_error(&ag->pdev->dev, buf->rx.dma_addr)) {
drivers/net/ethernet/atheros/ag71xx.c
1218
buf->rx.rx_buf = NULL;
drivers/net/ethernet/atheros/ag71xx.c
1221
desc->data = (u32)buf->rx.dma_addr + offset;
drivers/net/ethernet/atheros/ag71xx.c
1281
if (!ring->buf[i].rx.rx_buf &&
drivers/net/ethernet/atheros/ag71xx.c
1302
struct ag71xx_ring *rx = &ag->rx_ring;
drivers/net/ethernet/atheros/ag71xx.c
1305
ring_size = BIT(tx->order) + BIT(rx->order);
drivers/net/ethernet/atheros/ag71xx.c
1321
rx->buf = &tx->buf[tx_size];
drivers/net/ethernet/atheros/ag71xx.c
1322
rx->descs_cpu = ((void *)tx->descs_cpu) + tx_size * AG71XX_DESC_SIZE;
drivers/net/ethernet/atheros/ag71xx.c
1323
rx->descs_dma = tx->descs_dma + tx_size * AG71XX_DESC_SIZE;
drivers/net/ethernet/atheros/ag71xx.c
1332
struct ag71xx_ring *rx = &ag->rx_ring;
drivers/net/ethernet/atheros/ag71xx.c
1335
ring_size = BIT(tx->order) + BIT(rx->order);
drivers/net/ethernet/atheros/ag71xx.c
1344
rx->descs_cpu = NULL;
drivers/net/ethernet/atheros/ag71xx.c
1346
rx->buf = NULL;
drivers/net/ethernet/atheros/ag71xx.c
1640
dma_unmap_single(&ag->pdev->dev, ring->buf[i].rx.dma_addr,
drivers/net/ethernet/atheros/ag71xx.c
1646
skb = napi_build_skb(ring->buf[i].rx.rx_buf, ag71xx_buffer_size(ag));
drivers/net/ethernet/atheros/ag71xx.c
1649
skb_free_frag(ring->buf[i].rx.rx_buf);
drivers/net/ethernet/atheros/ag71xx.c
1662
ring->buf[i].rx.rx_buf = NULL;
drivers/net/ethernet/atheros/ag71xx.c
1692
if (!rx_ring->buf[rx_ring->dirty % rx_ring_size].rx.rx_buf)
drivers/net/ethernet/atheros/ag71xx.c
313
} rx;
drivers/net/ethernet/atheros/ag71xx.c
846
u32 rx, tx;
drivers/net/ethernet/atheros/ag71xx.c
850
rx = ag71xx_rr(ag, AG71XX_REG_RX_CTRL) & RX_CTRL_RXE;
drivers/net/ethernet/atheros/ag71xx.c
852
if (!rx && !tx)
drivers/net/ethernet/broadcom/bcmsysport.c
203
STAT_MIB_RX("rx_64_octets", mib.rx.pkt_cnt.cnt_64),
drivers/net/ethernet/broadcom/bcmsysport.c
204
STAT_MIB_RX("rx_65_127_oct", mib.rx.pkt_cnt.cnt_127),
drivers/net/ethernet/broadcom/bcmsysport.c
205
STAT_MIB_RX("rx_128_255_oct", mib.rx.pkt_cnt.cnt_255),
drivers/net/ethernet/broadcom/bcmsysport.c
206
STAT_MIB_RX("rx_256_511_oct", mib.rx.pkt_cnt.cnt_511),
drivers/net/ethernet/broadcom/bcmsysport.c
207
STAT_MIB_RX("rx_512_1023_oct", mib.rx.pkt_cnt.cnt_1023),
drivers/net/ethernet/broadcom/bcmsysport.c
208
STAT_MIB_RX("rx_1024_1518_oct", mib.rx.pkt_cnt.cnt_1518),
drivers/net/ethernet/broadcom/bcmsysport.c
209
STAT_MIB_RX("rx_vlan_1519_1522_oct", mib.rx.pkt_cnt.cnt_mgv),
drivers/net/ethernet/broadcom/bcmsysport.c
210
STAT_MIB_RX("rx_1522_2047_oct", mib.rx.pkt_cnt.cnt_2047),
drivers/net/ethernet/broadcom/bcmsysport.c
211
STAT_MIB_RX("rx_2048_4095_oct", mib.rx.pkt_cnt.cnt_4095),
drivers/net/ethernet/broadcom/bcmsysport.c
212
STAT_MIB_RX("rx_4096_9216_oct", mib.rx.pkt_cnt.cnt_9216),
drivers/net/ethernet/broadcom/bcmsysport.c
213
STAT_MIB_RX("rx_pkts", mib.rx.pkt),
drivers/net/ethernet/broadcom/bcmsysport.c
214
STAT_MIB_RX("rx_bytes", mib.rx.bytes),
drivers/net/ethernet/broadcom/bcmsysport.c
215
STAT_MIB_RX("rx_multicast", mib.rx.mca),
drivers/net/ethernet/broadcom/bcmsysport.c
216
STAT_MIB_RX("rx_broadcast", mib.rx.bca),
drivers/net/ethernet/broadcom/bcmsysport.c
217
STAT_MIB_RX("rx_fcs", mib.rx.fcs),
drivers/net/ethernet/broadcom/bcmsysport.c
218
STAT_MIB_RX("rx_control", mib.rx.cf),
drivers/net/ethernet/broadcom/bcmsysport.c
219
STAT_MIB_RX("rx_pause", mib.rx.pf),
drivers/net/ethernet/broadcom/bcmsysport.c
220
STAT_MIB_RX("rx_unknown", mib.rx.uo),
drivers/net/ethernet/broadcom/bcmsysport.c
221
STAT_MIB_RX("rx_align", mib.rx.aln),
drivers/net/ethernet/broadcom/bcmsysport.c
222
STAT_MIB_RX("rx_outrange", mib.rx.flr),
drivers/net/ethernet/broadcom/bcmsysport.c
223
STAT_MIB_RX("rx_code", mib.rx.cde),
drivers/net/ethernet/broadcom/bcmsysport.c
224
STAT_MIB_RX("rx_carrier", mib.rx.fcr),
drivers/net/ethernet/broadcom/bcmsysport.c
225
STAT_MIB_RX("rx_oversize", mib.rx.ovr),
drivers/net/ethernet/broadcom/bcmsysport.c
226
STAT_MIB_RX("rx_jabber", mib.rx.jbr),
drivers/net/ethernet/broadcom/bcmsysport.c
227
STAT_MIB_RX("rx_mtu_err", mib.rx.mtue),
drivers/net/ethernet/broadcom/bcmsysport.c
228
STAT_MIB_RX("rx_good_pkts", mib.rx.pok),
drivers/net/ethernet/broadcom/bcmsysport.c
229
STAT_MIB_RX("rx_unicast", mib.rx.uc),
drivers/net/ethernet/broadcom/bcmsysport.c
230
STAT_MIB_RX("rx_ppp", mib.rx.ppp),
drivers/net/ethernet/broadcom/bcmsysport.c
231
STAT_MIB_RX("rx_crc", mib.rx.rcrc),
drivers/net/ethernet/broadcom/bcmsysport.h
559
struct bcm_sysport_rx_counters rx;
drivers/net/ethernet/broadcom/bgmac.c
338
struct bgmac_rx_header *rx;
drivers/net/ethernet/broadcom/bgmac.c
347
rx = buf + BGMAC_RX_BUF_OFFSET;
drivers/net/ethernet/broadcom/bgmac.c
348
rx->len = cpu_to_le16(0xdead);
drivers/net/ethernet/broadcom/bgmac.c
349
rx->flags = cpu_to_le16(0xbeef);
drivers/net/ethernet/broadcom/bgmac.c
402
struct bgmac_rx_header *rx = slot->buf + BGMAC_RX_BUF_OFFSET;
drivers/net/ethernet/broadcom/bgmac.c
406
rx->len = cpu_to_le16(0xdead);
drivers/net/ethernet/broadcom/bgmac.c
407
rx->flags = cpu_to_le16(0xbeef);
drivers/net/ethernet/broadcom/bgmac.c
427
struct bgmac_rx_header *rx = slot->buf + BGMAC_RX_BUF_OFFSET;
drivers/net/ethernet/broadcom/bgmac.c
445
len = le16_to_cpu(rx->len);
drivers/net/ethernet/broadcom/bgmac.c
446
flags = le16_to_cpu(rx->flags);
drivers/net/ethernet/broadcom/bnge/bnge_hwrm_lib.c
392
u16 rx, tx;
drivers/net/ethernet/broadcom/bnge/bnge_hwrm_lib.c
416
rx = hw_resc->resv_rx_rings;
drivers/net/ethernet/broadcom/bnge/bnge_hwrm_lib.c
419
rx >>= 1;
drivers/net/ethernet/broadcom/bnge/bnge_hwrm_lib.c
420
if (cp < (rx + tx)) {
drivers/net/ethernet/broadcom/bnge/bnge_hwrm_lib.c
421
rc = bnge_fix_rings_count(&rx, &tx, cp, false);
drivers/net/ethernet/broadcom/bnge/bnge_hwrm_lib.c
425
rx <<= 1;
drivers/net/ethernet/broadcom/bnge/bnge_hwrm_lib.c
426
hw_resc->resv_rx_rings = rx;
drivers/net/ethernet/broadcom/bnge/bnge_hwrm_lib.c
430
hw_resc->resv_hw_ring_grps = rx;
drivers/net/ethernet/broadcom/bnge/bnge_hwrm_lib.c
452
enables |= hwr->rx ? FUNC_CFG_REQ_ENABLES_NUM_RX_RINGS : 0;
drivers/net/ethernet/broadcom/bnge/bnge_hwrm_lib.c
459
req->num_rx_rings = cpu_to_le16(hwr->rx);
drivers/net/ethernet/broadcom/bnge/bnge_netdev.c
267
int rx = 0, tx = 0;
drivers/net/ethernet/broadcom/bnge/bnge_netdev.c
281
rx = 1;
drivers/net/ethernet/broadcom/bnge/bnge_netdev.c
306
if (!k && rx) {
drivers/net/ethernet/broadcom/bnge/bnge_netdev.c
310
int n, tc = k - rx;
drivers/net/ethernet/broadcom/bnge/bnge_resc.c
137
int bnge_fix_rings_count(u16 *rx, u16 *tx, u16 max, bool shared)
drivers/net/ethernet/broadcom/bnge/bnge_resc.c
139
u16 _rx = *rx, _tx = *tx;
drivers/net/ethernet/broadcom/bnge/bnge_resc.c
142
*rx = min_t(u16, _rx, max);
drivers/net/ethernet/broadcom/bnge/bnge_resc.c
153
*rx = _rx;
drivers/net/ethernet/broadcom/bnge/bnge_resc.c
160
static int bnge_adjust_rings(struct bnge_dev *bd, u16 *rx,
drivers/net/ethernet/broadcom/bnge/bnge_resc.c
168
rc = bnge_fix_rings_count(rx, &tx_chunks, max_nq, sh);
drivers/net/ethernet/broadcom/bnge/bnge_resc.c
176
return bnge_fix_rings_count(rx, tx, max_nq, sh);
drivers/net/ethernet/broadcom/bnge/bnge_resc.c
229
hwr->rx = hw_resc->resv_rx_rings;
drivers/net/ethernet/broadcom/bnge/bnge_resc.c
240
return hwr->tx && hwr->rx && hwr->nq && hwr->grp && hwr->vnic &&
drivers/net/ethernet/broadcom/bnge/bnge_resc.c
248
u16 rx = bd->rx_nr_rings, stat;
drivers/net/ethernet/broadcom/bnge/bnge_resc.c
255
vnic = bnge_get_total_vnics(bd, rx);
drivers/net/ethernet/broadcom/bnge/bnge_resc.c
258
rx <<= 1;
drivers/net/ethernet/broadcom/bnge/bnge_resc.c
260
if (hw_resc->resv_rx_rings != rx || hw_resc->resv_cp_rings != cprs ||
drivers/net/ethernet/broadcom/bnge/bnge_resc.c
296
hwr.rx = bd->rx_nr_rings;
drivers/net/ethernet/broadcom/bnge/bnge_resc.c
299
hwr.cmpl = hwr.rx + hwr.tx;
drivers/net/ethernet/broadcom/bnge/bnge_resc.c
301
hwr.vnic = bnge_get_total_vnics(bd, hwr.rx);
drivers/net/ethernet/broadcom/bnge/bnge_resc.c
304
hwr.rx <<= 1;
drivers/net/ethernet/broadcom/bnge/bnge_resc.c
316
rx_rings = hwr.rx;
drivers/net/ethernet/broadcom/bnge/bnge_resc.c
318
if (hwr.rx >= 2)
drivers/net/ethernet/broadcom/bnge/bnge_resc.c
319
rx_rings = hwr.rx >> 1;
drivers/net/ethernet/broadcom/bnge/bnge_resc.c
333
hwr.rx = rx_rings << 1;
drivers/net/ethernet/broadcom/bnge/bnge_resc.c
466
u16 rx, tx, nq;
drivers/net/ethernet/broadcom/bnge/bnge_resc.c
468
_bnge_get_max_rings(bd, &rx, &tx, &nq);
drivers/net/ethernet/broadcom/bnge/bnge_resc.c
469
*max_rx = rx;
drivers/net/ethernet/broadcom/bnge/bnge_resc.c
471
if (!rx || !tx || !nq)
drivers/net/ethernet/broadcom/bnge/bnge_resc.h
48
u16 rx;
drivers/net/ethernet/broadcom/bnge/bnge_resc.h
69
int bnge_fix_rings_count(u16 *rx, u16 *tx, u16 max, bool shared);
drivers/net/ethernet/broadcom/bnx2.c
7322
bnx2_change_ring_size(struct bnx2 *bp, u32 rx, u32 tx, bool reset_irq)
drivers/net/ethernet/broadcom/bnx2.c
7340
bnx2_set_rx_ring_size(bp, rx);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
1995
int rc, tx, rx;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
1998
rx = BNX2X_NUM_ETH_QUEUES(bp);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
2002
rx++;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
2011
rc = netif_set_real_num_rx_queues(bp->dev, rx);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
2018
tx, rx);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_hsi.h
3741
struct client_init_rx_data rx;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c
12151
u32 rx = 0, tx = 0, i;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c
12158
rx = REG_RD(bp, shmem_base +
drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c
12166
rx = REG_RD(bp, shmem_base +
drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c
12175
phy->rx_preemphasis[i << 1] = ((rx>>16) & 0xffff);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_link.c
12176
phy->rx_preemphasis[(i << 1) + 1] = (rx & 0xffff);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c
8748
__set_bit(BNX2X_Q_FLG_HC, &init_params->rx.flags);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c
8754
__set_bit(BNX2X_Q_FLG_HC_EN, &init_params->rx.flags);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c
8758
init_params->rx.hc_rate = bp->rx_ticks ?
drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c
8764
init_params->rx.fw_sb_id = init_params->tx.fw_sb_id =
drivers/net/ethernet/broadcom/bnx2x/bnx2x_main.c
8771
init_params->rx.sb_cq_index = HC_INDEX_ETH_RX_CQ_CONS;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c
4787
data->rx.tpa_en |= test_bit(BNX2X_Q_FLG_TPA_IPV6, ¶ms->flags) *
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c
4963
&data->rx,
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c
4968
&data->rx);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c
5024
test_bit(BNX2X_Q_FLG_HC, &init->rx.flags)) {
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c
5025
hc_usec = init->rx.hc_rate ? 1000000 / init->rx.hc_rate : 0;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c
5027
bnx2x_update_coalesce_sb_index(bp, init->rx.fw_sb_id,
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c
5028
init->rx.sb_cq_index,
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c
5029
!test_bit(BNX2X_Q_FLG_HC_EN, &init->rx.flags),
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.h
893
} rx;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c
154
init_params->rx.sb_cq_index,
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c
155
init_params->rx.hc_rate,
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c
178
if (test_bit(BNX2X_Q_FLG_HC, &init_p->rx.flags))
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c
179
__set_bit(BNX2X_Q_FLG_HC_EN, &init_p->rx.flags);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c
185
init_p->rx.fw_sb_id = vf_igu_sb(vf, q->sb_idx);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.c
1600
init_p->rx.hc_rate = setup_q->rxq.hc_rate;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.c
1601
init_p->rx.sb_cq_index = setup_q->rxq.sb_index;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.c
1603
&init_p->rx.flags);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
11210
static int __bnxt_trim_rings(struct bnxt *bp, int *rx, int *tx, int max,
drivers/net/ethernet/broadcom/bnxt/bnxt.c
11213
int _rx = *rx, _tx = *tx;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
11216
*rx = min_t(int, _rx, max);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
11228
*rx = _rx;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
11256
static int bnxt_trim_rings(struct bnxt *bp, int *rx, int *tx, int max,
drivers/net/ethernet/broadcom/bnxt/bnxt.c
11264
rc = __bnxt_trim_rings(bp, rx, &tx_cp, max, sh);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
11271
return __bnxt_trim_rings(bp, rx, tx, max, sh);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
11875
cpr->sw_stats->rx.rx_resets++;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
13440
cpr->sw_stats->rx.rx_netpoll_discards +
drivers/net/ethernet/broadcom/bnxt/bnxt.c
13441
cpr->sw_stats->rx.rx_oom_discards;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
13480
u64 *rx = bp->port_stats.sw_stats;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
13485
BNXT_GET_RX_PORT_STATS64(rx, rx_fcs_err_frames);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
13487
BNXT_GET_RX_PORT_STATS64(rx, rx_align_err_frames);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
13489
BNXT_GET_RX_PORT_STATS64(rx, rx_undrsz_frames) +
drivers/net/ethernet/broadcom/bnxt/bnxt.c
13490
BNXT_GET_RX_PORT_STATS64(rx, rx_ovrsz_frames) +
drivers/net/ethernet/broadcom/bnxt/bnxt.c
13491
BNXT_GET_RX_PORT_STATS64(rx, rx_runt_frames);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
13493
BNXT_GET_RX_PORT_STATS64(rx, rx_false_carrier_frames) +
drivers/net/ethernet/broadcom/bnxt/bnxt.c
13494
BNXT_GET_RX_PORT_STATS64(rx, rx_jbr_frames);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
13511
stats->rx_total_l4_csum_errors += sw_stats->rx.rx_l4_csum_errors;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
13512
stats->rx_total_resets += sw_stats->rx.rx_resets;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
13513
stats->rx_total_buf_errors += sw_stats->rx.rx_buf_errors;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
13514
stats->rx_total_oom_discards += sw_stats->rx.rx_oom_discards;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
13515
stats->rx_total_netpoll_discards += sw_stats->rx.rx_netpoll_discards;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
13518
stats->rx_total_hw_gro_packets += sw_stats->rx.rx_hw_gro_packets;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
13519
stats->rx_total_hw_gro_wire_packets += sw_stats->rx.rx_hw_gro_wire_packets;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
14339
cpr->sw_stats->rx.rx_resets++;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
14723
int bnxt_check_rings(struct bnxt *bp, int tx, int rx, bool sh, int tcs,
drivers/net/ethernet/broadcom/bnxt/bnxt.c
14728
int rx_rings = rx;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
14742
hwr.rx = rx_rings;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
14747
hwr.vnic = bnxt_get_total_vnics(bp, rx);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
14750
hwr.cp = sh ? max_t(int, tx_cp, rx) : tx_cp + rx;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
14757
hwr.grp = rx;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
14761
hwr.cp_p5 = hwr.tx + rx;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
15946
stats->alloc_fail = cpr->sw_stats->rx.rx_oom_discards;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
15947
stats->hw_gro_packets = cpr->sw_stats->rx.rx_hw_gro_packets;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
15948
stats->hw_gro_wire_packets = cpr->sw_stats->rx.rx_hw_gro_wire_packets;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
15976
struct netdev_queue_stats_rx *rx,
drivers/net/ethernet/broadcom/bnxt/bnxt.c
15981
rx->packets = bp->net_stats_prev.rx_packets;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
15982
rx->bytes = bp->net_stats_prev.rx_bytes;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
15983
rx->alloc_fail = bp->ring_err_stats_prev.rx_total_oom_discards;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
15984
rx->hw_gro_packets = bp->ring_err_stats_prev.rx_total_hw_gro_packets;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
15985
rx->hw_gro_wire_packets = bp->ring_err_stats_prev.rx_total_hw_gro_wire_packets;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
16478
int rx, tx, cp;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
16480
_bnxt_get_max_rings(bp, &rx, &tx, &cp);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
16481
*max_rx = rx;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
16483
if (!rx || !tx || !cp)
drivers/net/ethernet/broadcom/bnxt/bnxt.c
1918
cpr->sw_stats->rx.rx_oom_discards += 1;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
1929
cpr->sw_stats->rx.rx_oom_discards += 1;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
1944
cpr->sw_stats->rx.rx_oom_discards += 1;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
1957
cpr->sw_stats->rx.rx_oom_discards += 1;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
1992
&cpr->sw_stats->rx);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
2218
bnapi->cp_ring.sw_stats->rx.rx_buf_errors++;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
2341
bnapi->cp_ring.sw_stats->rx.rx_l4_csum_errors++;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
2376
cpr->sw_stats->rx.rx_oom_discards += 1;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
2425
cpr->sw_stats->rx.rx_netpoll_discards += 1;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
4203
int rx = 0, tx = 0;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
4223
rx = 1;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
4247
if (!k && rx) {
drivers/net/ethernet/broadcom/bnxt/bnxt.c
4251
int n, tc = k - rx;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
4900
int rx, tx;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
4902
bnxt_get_max_rings(bp, &rx, &tx, true);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
4903
if (rx > 1) {
drivers/net/ethernet/broadcom/bnxt/bnxt.c
7655
static int __bnxt_trim_rings(struct bnxt *bp, int *rx, int *tx, int max,
drivers/net/ethernet/broadcom/bnxt/bnxt.c
7657
static int bnxt_trim_rings(struct bnxt *bp, int *rx, int *tx, int max,
drivers/net/ethernet/broadcom/bnxt/bnxt.c
7695
int rx = hw_resc->resv_rx_rings;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
7699
rx >>= 1;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
7700
if (cp < (rx + tx)) {
drivers/net/ethernet/broadcom/bnxt/bnxt.c
7701
rc = __bnxt_trim_rings(bp, &rx, &tx, cp, false);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
7705
rx <<= 1;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
7706
hw_resc->resv_rx_rings = rx;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
7710
hw_resc->resv_hw_ring_grps = rx;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
7758
enables |= hwr->rx ? FUNC_CFG_REQ_ENABLES_NUM_RX_RINGS : 0;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
7773
req->num_rx_rings = cpu_to_le16(hwr->rx);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
7799
enables |= hwr->rx ? FUNC_VF_CFG_REQ_ENABLES_NUM_RX_RINGS |
drivers/net/ethernet/broadcom/bnxt/bnxt.c
7816
req->num_rx_rings = cpu_to_le16(hwr->rx);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
7958
hwr->rx = bp->rx_nr_rings;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
7959
hwr->grp = hwr->rx;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
7960
hwr->vnic = bnxt_get_total_vnics(bp, hwr->rx);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
7963
hwr->rx <<= 1;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
7988
if (hw_resc->resv_rx_rings != hwr.rx ||
drivers/net/ethernet/broadcom/bnxt/bnxt.c
8013
hwr->rx = hw_resc->resv_rx_rings;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
8026
return hwr->tx && hwr->rx && hwr->cp && hwr->grp && hwr->vnic &&
drivers/net/ethernet/broadcom/bnxt/bnxt.c
8059
hwr.rx = bp->rx_nr_rings;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
8063
hwr.cp_p5 = hwr.rx + hwr.tx;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
8065
hwr.vnic = bnxt_get_total_vnics(bp, hwr.rx);
drivers/net/ethernet/broadcom/bnxt/bnxt.c
8068
hwr.rx <<= 1;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
8080
rx_rings = hwr.rx;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
8082
if (hwr.rx >= 2) {
drivers/net/ethernet/broadcom/bnxt/bnxt.c
8083
rx_rings = hwr.rx >> 1;
drivers/net/ethernet/broadcom/bnxt/bnxt.c
8103
hwr.rx = rx_rings << 1;
drivers/net/ethernet/broadcom/bnxt/bnxt.h
1145
struct bnxt_rx_sw_stats rx;
drivers/net/ethernet/broadcom/bnxt/bnxt.h
1333
int rx;
drivers/net/ethernet/broadcom/bnxt/bnxt.h
2984
int bnxt_check_rings(struct bnxt *bp, int tx, int rx, bool sh, int tcs,
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
3272
u64 *rx;
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
3277
rx = bp->rx_port_stats_ext.sw_stats;
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
3279
*(rx + BNXT_RX_STATS_EXT_OFFSET(rx_corrected_bits));
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
3285
*(rx + BNXT_RX_STATS_EXT_OFFSET(rx_fec_corrected_blocks));
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
3287
*(rx + BNXT_RX_STATS_EXT_OFFSET(rx_fec_uncorrectable_blocks));
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
3371
u64 *rx, *tx;
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
3376
rx = bp->port_stats.sw_stats;
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
3379
epstat->rx_pause_frames = BNXT_GET_RX_PORT_STATS64(rx, rx_pause_frames);
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
526
int rx, tx, cmn;
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
528
rx = NUM_RING_RX_HW_STATS + NUM_RING_RX_SW_STATS +
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
532
return rx * bp->rx_nr_rings +
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
5534
u64 *rx;
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
5539
rx = bp->rx_port_stats_ext.sw_stats;
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
5541
*(rx + BNXT_RX_STATS_EXT_OFFSET(rx_pcs_symbol_err));
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
5548
u64 *rx, *tx;
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
5553
rx = bp->port_stats.sw_stats;
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
5557
BNXT_GET_RX_PORT_STATS64(rx, rx_good_frames);
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
5561
BNXT_GET_RX_PORT_STATS64(rx, rx_fcs_err_frames);
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
5563
BNXT_GET_RX_PORT_STATS64(rx, rx_align_err_frames);
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
5565
BNXT_GET_RX_PORT_STATS64(rx, rx_oor_len_frames);
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
5572
u64 *rx;
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
5577
rx = bp->port_stats.sw_stats;
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
5579
BNXT_GET_RX_PORT_STATS64(rx, rx_ctrl_frames);
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
5601
u64 *rx, *tx;
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
5606
rx = bp->port_stats.sw_stats;
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
5610
BNXT_GET_RX_PORT_STATS64(rx, rx_jbr_frames);
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
5612
BNXT_GET_RX_PORT_STATS64(rx, rx_ovrsz_frames);
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
5614
BNXT_GET_RX_PORT_STATS64(rx, rx_undrsz_frames);
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
5616
rmon_stats->hist[0] = BNXT_GET_RX_PORT_STATS64(rx, rx_64b_frames);
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
5617
rmon_stats->hist[1] = BNXT_GET_RX_PORT_STATS64(rx, rx_65b_127b_frames);
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
5618
rmon_stats->hist[2] = BNXT_GET_RX_PORT_STATS64(rx, rx_128b_255b_frames);
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
5619
rmon_stats->hist[3] = BNXT_GET_RX_PORT_STATS64(rx, rx_256b_511b_frames);
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
5621
BNXT_GET_RX_PORT_STATS64(rx, rx_512b_1023b_frames);
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
5623
BNXT_GET_RX_PORT_STATS64(rx, rx_1024b_1518b_frames);
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
5625
BNXT_GET_RX_PORT_STATS64(rx, rx_1519b_2047b_frames);
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
5627
BNXT_GET_RX_PORT_STATS64(rx, rx_2048b_4095b_frames);
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
5629
BNXT_GET_RX_PORT_STATS64(rx, rx_4096b_9216b_frames);
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
5631
BNXT_GET_RX_PORT_STATS64(rx, rx_9217b_16383b_frames);
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
5674
u64 *rx;
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
5679
rx = bp->rx_port_stats_ext.sw_stats;
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
5681
*(rx + BNXT_RX_STATS_EXT_OFFSET(link_down_events));
drivers/net/ethernet/broadcom/bnxt/bnxt_ethtool.c
635
sw = (u64 *)&cpr->sw_stats->rx;
drivers/net/ethernet/broadcom/cnic.c
5018
data->rx.bd_page_base.hi = cpu_to_le32(val);
drivers/net/ethernet/broadcom/cnic.c
5022
data->rx.bd_page_base.lo = cpu_to_le32(val);
drivers/net/ethernet/broadcom/cnic.c
5027
data->rx.cqe_page_base.hi = cpu_to_le32(val);
drivers/net/ethernet/broadcom/cnic.c
5031
data->rx.cqe_page_base.lo = cpu_to_le32(val);
drivers/net/ethernet/broadcom/cnic.c
5034
data->rx.client_qzone_id = cl_qzone_id;
drivers/net/ethernet/broadcom/cnic.c
5035
data->rx.rx_sb_index_number = HC_SP_INDEX_ETH_ISCSI_RX_CQ_CONS;
drivers/net/ethernet/broadcom/cnic.c
5036
data->rx.status_block_id = BNX2X_DEF_SB_ID;
drivers/net/ethernet/broadcom/cnic.c
5038
data->rx.cache_line_alignment_log_size = L1_CACHE_SHIFT;
drivers/net/ethernet/broadcom/cnic.c
5040
data->rx.max_bytes_on_bd = cpu_to_le16(cp->l2_single_buf_size);
drivers/net/ethernet/broadcom/cnic.c
5041
data->rx.outer_vlan_removal_enable_flg = 1;
drivers/net/ethernet/broadcom/cnic.c
5042
data->rx.silent_vlan_removal_flg = 1;
drivers/net/ethernet/broadcom/cnic.c
5043
data->rx.silent_vlan_value = 0;
drivers/net/ethernet/broadcom/cnic.c
5044
data->rx.silent_vlan_mask = 0xffff;
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1085
STAT_GENET_MIB_RX("rx_64_octets", mib.rx.pkt_cnt.cnt_64),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1086
STAT_GENET_MIB_RX("rx_65_127_oct", mib.rx.pkt_cnt.cnt_127),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1087
STAT_GENET_MIB_RX("rx_128_255_oct", mib.rx.pkt_cnt.cnt_255),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1088
STAT_GENET_MIB_RX("rx_256_511_oct", mib.rx.pkt_cnt.cnt_511),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1089
STAT_GENET_MIB_RX("rx_512_1023_oct", mib.rx.pkt_cnt.cnt_1023),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1090
STAT_GENET_MIB_RX("rx_1024_1518_oct", mib.rx.pkt_cnt.cnt_1518),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1091
STAT_GENET_MIB_RX("rx_vlan_1519_1522_oct", mib.rx.pkt_cnt.cnt_mgv),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1092
STAT_GENET_MIB_RX("rx_1522_2047_oct", mib.rx.pkt_cnt.cnt_2047),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1093
STAT_GENET_MIB_RX("rx_2048_4095_oct", mib.rx.pkt_cnt.cnt_4095),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1094
STAT_GENET_MIB_RX("rx_4096_9216_oct", mib.rx.pkt_cnt.cnt_9216),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1095
STAT_GENET_MIB_RX("rx_pkts", mib.rx.pkt),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1096
STAT_GENET_MIB_RX("rx_bytes", mib.rx.bytes),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1097
STAT_GENET_MIB_RX("rx_multicast", mib.rx.mca),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1098
STAT_GENET_MIB_RX("rx_broadcast", mib.rx.bca),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1099
STAT_GENET_MIB_RX("rx_fcs", mib.rx.fcs),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1100
STAT_GENET_MIB_RX("rx_control", mib.rx.cf),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1101
STAT_GENET_MIB_RX("rx_pause", mib.rx.pf),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1102
STAT_GENET_MIB_RX("rx_unknown", mib.rx.uo),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1103
STAT_GENET_MIB_RX("rx_align", mib.rx.aln),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1104
STAT_GENET_MIB_RX("rx_outrange", mib.rx.flr),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1105
STAT_GENET_MIB_RX("rx_code", mib.rx.cde),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1106
STAT_GENET_MIB_RX("rx_carrier", mib.rx.fcr),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1107
STAT_GENET_MIB_RX("rx_oversize", mib.rx.ovr),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1108
STAT_GENET_MIB_RX("rx_jabber", mib.rx.jbr),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1109
STAT_GENET_MIB_RX("rx_mtu_err", mib.rx.mtue),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1110
STAT_GENET_MIB_RX("rx_good_pkts", mib.rx.pok),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1111
STAT_GENET_MIB_RX("rx_unicast", mib.rx.uc),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1112
STAT_GENET_MIB_RX("rx_ppp", mib.rx.ppp),
drivers/net/ethernet/broadcom/genet/bcmgenet.c
1113
STAT_GENET_MIB_RX("rx_crc", mib.rx.rcrc),
drivers/net/ethernet/broadcom/genet/bcmgenet.h
142
struct bcmgenet_rx_counters rx;
drivers/net/ethernet/broadcom/genet/bcmgenet.h
738
void bcmgenet_phy_pause_set(struct net_device *dev, bool rx, bool tx);
drivers/net/ethernet/broadcom/genet/bcmmii.c
132
void bcmgenet_phy_pause_set(struct net_device *dev, bool rx, bool tx)
drivers/net/ethernet/broadcom/genet/bcmmii.c
136
linkmode_mod_bit(ETHTOOL_LINK_MODE_Pause_BIT, phydev->advertising, rx);
drivers/net/ethernet/broadcom/genet/bcmmii.c
138
rx | tx);
drivers/net/ethernet/brocade/bna/bfa_cs.h
53
BFA_SM_TABLE(rx, bna_rx, bna_rx_event, bna_fsm_rx_t)
drivers/net/ethernet/brocade/bna/bna.h
110
cbfn(cbarg, rxf->rx); \
drivers/net/ethernet/brocade/bna/bna.h
322
void bna_bfi_rx_enet_start_rsp(struct bna_rx *rx,
drivers/net/ethernet/brocade/bna/bna.h
324
void bna_bfi_rx_enet_stop_rsp(struct bna_rx *rx,
drivers/net/ethernet/brocade/bna/bna.h
349
void bna_rx_destroy(struct bna_rx *rx);
drivers/net/ethernet/brocade/bna/bna.h
350
void bna_rx_enable(struct bna_rx *rx);
drivers/net/ethernet/brocade/bna/bna.h
351
void bna_rx_disable(struct bna_rx *rx, enum bna_cleanup_type type,
drivers/net/ethernet/brocade/bna/bna.h
353
void bna_rx_cleanup_complete(struct bna_rx *rx);
drivers/net/ethernet/brocade/bna/bna.h
354
void bna_rx_coalescing_timeo_set(struct bna_rx *rx, int coalescing_timeo);
drivers/net/ethernet/brocade/bna/bna.h
357
enum bna_cb_status bna_rx_ucast_set(struct bna_rx *rx, const u8 *ucmac);
drivers/net/ethernet/brocade/bna/bna.h
358
enum bna_cb_status bna_rx_ucast_listset(struct bna_rx *rx, int count,
drivers/net/ethernet/brocade/bna/bna.h
360
enum bna_cb_status bna_rx_mcast_add(struct bna_rx *rx, const u8 *mcmac,
drivers/net/ethernet/brocade/bna/bna.h
363
enum bna_cb_status bna_rx_mcast_listset(struct bna_rx *rx, int count,
drivers/net/ethernet/brocade/bna/bna.h
366
bna_rx_mcast_delall(struct bna_rx *rx);
drivers/net/ethernet/brocade/bna/bna.h
368
bna_rx_mode_set(struct bna_rx *rx, enum bna_rxmode rxmode,
drivers/net/ethernet/brocade/bna/bna.h
370
void bna_rx_vlan_add(struct bna_rx *rx, int vlan_id);
drivers/net/ethernet/brocade/bna/bna.h
371
void bna_rx_vlan_del(struct bna_rx *rx, int vlan_id);
drivers/net/ethernet/brocade/bna/bna.h
372
void bna_rx_vlanfilter_enable(struct bna_rx *rx);
drivers/net/ethernet/brocade/bna/bna.h
373
void bna_rx_vlan_strip_enable(struct bna_rx *rx);
drivers/net/ethernet/brocade/bna/bna.h
374
void bna_rx_vlan_strip_disable(struct bna_rx *rx);
drivers/net/ethernet/brocade/bna/bna_enet.c
277
struct bna_rx *rx;
drivers/net/ethernet/brocade/bna/bna_enet.c
281
bna_rx_from_rid(bna, msghdr->enet_id, rx);
drivers/net/ethernet/brocade/bna/bna_enet.c
282
if (rx)
drivers/net/ethernet/brocade/bna/bna_enet.c
283
bna_bfi_rx_enet_start_rsp(rx, msghdr);
drivers/net/ethernet/brocade/bna/bna_enet.c
287
bna_rx_from_rid(bna, msghdr->enet_id, rx);
drivers/net/ethernet/brocade/bna/bna_enet.c
288
if (rx)
drivers/net/ethernet/brocade/bna/bna_enet.c
289
bna_bfi_rx_enet_stop_rsp(rx, msghdr);
drivers/net/ethernet/brocade/bna/bna_enet.c
304
bna_rx_from_rid(bna, msghdr->enet_id, rx);
drivers/net/ethernet/brocade/bna/bna_enet.c
305
if (rx)
drivers/net/ethernet/brocade/bna/bna_enet.c
306
bna_bfi_rxf_cfg_rsp(&rx->rxf, msghdr);
drivers/net/ethernet/brocade/bna/bna_enet.c
310
bna_rx_from_rid(bna, msghdr->enet_id, rx);
drivers/net/ethernet/brocade/bna/bna_enet.c
311
if (rx)
drivers/net/ethernet/brocade/bna/bna_enet.c
312
bna_bfi_rxf_ucast_set_rsp(&rx->rxf, msghdr);
drivers/net/ethernet/brocade/bna/bna_enet.c
316
bna_rx_from_rid(bna, msghdr->enet_id, rx);
drivers/net/ethernet/brocade/bna/bna_enet.c
317
if (rx)
drivers/net/ethernet/brocade/bna/bna_enet.c
318
bna_bfi_rxf_mcast_add_rsp(&rx->rxf, msghdr);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1001
bna_ucam_mod_del_q(rxf->rx->bna));
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1006
bna_ucam_mod_del_q(rxf->rx->bna));
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1039
struct bna *bna = rxf->rx->bna;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1067
struct bna *bna = rxf->rx->bna;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1153
struct bna *bna = rxf->rx->bna;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1169
bna->promisc_rid = rxf->rx->rid;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1179
struct bna *bna = rxf->rx->bna;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1270
#define call_rx_stop_cbfn(rx) \
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1272
if ((rx)->stop_cbfn) { \
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1275
cbfn = (rx)->stop_cbfn; \
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1276
cbarg = (rx)->stop_cbarg; \
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1277
(rx)->stop_cbfn = NULL; \
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1278
(rx)->stop_cbarg = NULL; \
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1279
cbfn(cbarg, rx); \
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1283
#define call_rx_stall_cbfn(rx) \
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1285
if ((rx)->rx_stall_cbfn) \
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1286
(rx)->rx_stall_cbfn((rx)->bna->bnad, (rx)); \
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1301
static void bna_bfi_rx_enet_start(struct bna_rx *rx);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1302
static void bna_rx_enet_stop(struct bna_rx *rx);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1303
static void bna_rx_mod_cb_rx_stopped(void *arg, struct bna_rx *rx);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1326
static void bna_rx_sm_stopped_entry(struct bna_rx *rx)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1328
call_rx_stop_cbfn(rx);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1331
static void bna_rx_sm_stopped(struct bna_rx *rx,
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1336
bfa_fsm_set_state(rx, bna_rx_sm_start_wait);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1340
call_rx_stop_cbfn(rx);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1353
static void bna_rx_sm_start_wait_entry(struct bna_rx *rx)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1355
bna_bfi_rx_enet_start(rx);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1359
bna_rx_sm_stop_wait_entry(struct bna_rx *rx)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1364
bna_rx_sm_stop_wait(struct bna_rx *rx, enum bna_rx_event event)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1369
bfa_fsm_set_state(rx, bna_rx_sm_cleanup_wait);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1370
rx->rx_cleanup_cbfn(rx->bna->bnad, rx);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1374
bna_rx_enet_stop(rx);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1383
static void bna_rx_sm_start_wait(struct bna_rx *rx,
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1388
bfa_fsm_set_state(rx, bna_rx_sm_start_stop_wait);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1392
bfa_fsm_set_state(rx, bna_rx_sm_stopped);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1396
bfa_fsm_set_state(rx, bna_rx_sm_rxf_start_wait);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1405
static void bna_rx_sm_rxf_start_wait_entry(struct bna_rx *rx)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1407
rx->rx_post_cbfn(rx->bna->bnad, rx);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1408
bna_rxf_start(&rx->rxf);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1412
bna_rx_sm_rxf_stop_wait_entry(struct bna_rx *rx)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1417
bna_rx_sm_rxf_stop_wait(struct bna_rx *rx, enum bna_rx_event event)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1421
bfa_fsm_set_state(rx, bna_rx_sm_cleanup_wait);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1422
bna_rxf_fail(&rx->rxf);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1423
call_rx_stall_cbfn(rx);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1424
rx->rx_cleanup_cbfn(rx->bna->bnad, rx);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1428
bna_rxf_stop(&rx->rxf);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1432
bfa_fsm_set_state(rx, bna_rx_sm_stop_wait);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1433
call_rx_stall_cbfn(rx);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1434
bna_rx_enet_stop(rx);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1445
bna_rx_sm_start_stop_wait_entry(struct bna_rx *rx)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1450
bna_rx_sm_start_stop_wait(struct bna_rx *rx, enum bna_rx_event event)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1455
bfa_fsm_set_state(rx, bna_rx_sm_stopped);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1459
bna_rx_enet_stop(rx);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1468
bna_rx_sm_started_entry(struct bna_rx *rx)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1471
int is_regular = (rx->type == BNA_RX_T_REGULAR);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1474
list_for_each_entry(rxp, &rx->rxp_q, qe)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1475
bna_ib_start(rx->bna, &rxp->cq.ib, is_regular);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1477
bna_ethport_cb_rx_started(&rx->bna->ethport);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1481
bna_rx_sm_started(struct bna_rx *rx, enum bna_rx_event event)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1485
bfa_fsm_set_state(rx, bna_rx_sm_rxf_stop_wait);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1486
bna_ethport_cb_rx_stopped(&rx->bna->ethport);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1487
bna_rxf_stop(&rx->rxf);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1491
bfa_fsm_set_state(rx, bna_rx_sm_failed);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1492
bna_ethport_cb_rx_stopped(&rx->bna->ethport);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1493
bna_rxf_fail(&rx->rxf);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1494
call_rx_stall_cbfn(rx);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1495
rx->rx_cleanup_cbfn(rx->bna->bnad, rx);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1504
static void bna_rx_sm_rxf_start_wait(struct bna_rx *rx,
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1509
bfa_fsm_set_state(rx, bna_rx_sm_rxf_stop_wait);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1513
bfa_fsm_set_state(rx, bna_rx_sm_failed);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1514
bna_rxf_fail(&rx->rxf);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1515
call_rx_stall_cbfn(rx);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1516
rx->rx_cleanup_cbfn(rx->bna->bnad, rx);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1520
bfa_fsm_set_state(rx, bna_rx_sm_started);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1530
bna_rx_sm_cleanup_wait_entry(struct bna_rx *rx)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1535
bna_rx_sm_cleanup_wait(struct bna_rx *rx, enum bna_rx_event event)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1544
bfa_fsm_set_state(rx, bna_rx_sm_stopped);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1554
bna_rx_sm_failed_entry(struct bna_rx *rx)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1559
bna_rx_sm_failed(struct bna_rx *rx, enum bna_rx_event event)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1563
bfa_fsm_set_state(rx, bna_rx_sm_quiesce_wait);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1567
bfa_fsm_set_state(rx, bna_rx_sm_cleanup_wait);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1577
bfa_fsm_set_state(rx, bna_rx_sm_stopped);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1586
bna_rx_sm_quiesce_wait_entry(struct bna_rx *rx)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1591
bna_rx_sm_quiesce_wait(struct bna_rx *rx, enum bna_rx_event event)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1595
bfa_fsm_set_state(rx, bna_rx_sm_cleanup_wait);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1599
bfa_fsm_set_state(rx, bna_rx_sm_failed);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1603
bfa_fsm_set_state(rx, bna_rx_sm_start_wait);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1613
bna_bfi_rx_enet_start(struct bna_rx *rx)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1615
struct bfi_enet_rx_cfg_req *cfg_req = &rx->bfi_enet_cmd.cfg_req;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1621
BFI_ENET_H2I_RX_CFG_SET_REQ, 0, rx->rid);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1625
cfg_req->rx_cfg.frame_size = bna_enet_mtu_get(&rx->bna->enet);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1626
cfg_req->num_queue_sets = rx->num_paths;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1627
for (i = 0; i < rx->num_paths; i++) {
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1629
: list_first_entry(&rx->rxp_q, struct bna_rxp, qe);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1655
bna_enet_mtu_get(&rx->bna->enet);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1695
cfg_req->rx_cfg.hds.type = rx->hds_cfg.hdr_type;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1696
cfg_req->rx_cfg.hds.force_offset = rx->hds_cfg.forced_offset;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1697
cfg_req->rx_cfg.hds.max_header_size = rx->hds_cfg.forced_offset;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1707
cfg_req->rx_cfg.strip_vlan = rx->rxf.vlan_strip_status;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1709
bfa_msgq_cmd_set(&rx->msgq_cmd, NULL, NULL,
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1711
bfa_msgq_cmd_post(&rx->bna->msgq, &rx->msgq_cmd);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1715
bna_bfi_rx_enet_stop(struct bna_rx *rx)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1717
struct bfi_enet_req *req = &rx->bfi_enet_cmd.req;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1720
BFI_ENET_H2I_RX_CFG_CLR_REQ, 0, rx->rid);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1723
bfa_msgq_cmd_set(&rx->msgq_cmd, NULL, NULL, sizeof(struct bfi_enet_req),
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1725
bfa_msgq_cmd_post(&rx->bna->msgq, &rx->msgq_cmd);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1729
bna_rx_enet_stop(struct bna_rx *rx)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1734
list_for_each_entry(rxp, &rx->rxp_q, qe)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1735
bna_ib_stop(rx->bna, &rxp->cq.ib);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1737
bna_bfi_rx_enet_stop(rx);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1802
struct bna_rx *rx = NULL;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1806
rx = list_first_entry(&rx_mod->rx_free_q, struct bna_rx, qe);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1808
rx = list_last_entry(&rx_mod->rx_free_q, struct bna_rx, qe);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1811
list_move_tail(&rx->qe, &rx_mod->rx_active_q);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1812
rx->type = type;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1814
return rx;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1818
bna_rx_put(struct bna_rx_mod *rx_mod, struct bna_rx *rx)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1823
if (((struct bna_rx *)qe)->rid < rx->rid)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1826
list_add(&rx->qe, qe);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
184
bfi_msgq_mhdr_set(req->mh, BFI_MC_ENET, req_type, 0, rxf->rx->rid);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
190
bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1930
bna_rx_mod_cb_rx_stopped(void *arg, struct bna_rx *rx)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1948
bna_rx_start(struct bna_rx *rx)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1950
rx->rx_flags |= BNA_RX_F_ENET_STARTED;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1951
if (rx->rx_flags & BNA_RX_F_ENABLED)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1952
bfa_fsm_send_event(rx, RX_E_START);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1956
bna_rx_stop(struct bna_rx *rx)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1958
rx->rx_flags &= ~BNA_RX_F_ENET_STARTED;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1959
if (rx->fsm == bna_rx_sm_stopped)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1960
bna_rx_mod_cb_rx_stopped(&rx->bna->rx_mod, rx);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1962
rx->stop_cbfn = bna_rx_mod_cb_rx_stopped;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1963
rx->stop_cbarg = &rx->bna->rx_mod;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1964
bfa_fsm_send_event(rx, RX_E_STOP);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1969
bna_rx_fail(struct bna_rx *rx)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1972
rx->rx_flags &= ~BNA_RX_F_ENET_STARTED;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1973
bfa_fsm_send_event(rx, RX_E_FAIL);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1979
struct bna_rx *rx;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1985
list_for_each_entry(rx, &rx_mod->rx_active_q, qe)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1986
if (rx->type == type)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1987
bna_rx_start(rx);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
1993
struct bna_rx *rx;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
200
0, rxf->rx->rid);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2002
list_for_each_entry(rx, &rx_mod->rx_active_q, qe)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2003
if (rx->type == type) {
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2005
bna_rx_stop(rx);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2014
struct bna_rx *rx;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2019
list_for_each_entry(rx, &rx_mod->rx_active_q, qe)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2020
bna_rx_fail(rx);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2034
rx_mod->rx = (struct bna_rx *)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2052
rx_ptr = &rx_mod->rx[index];
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
206
bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2086
bna_bfi_rx_enet_start_rsp(struct bna_rx *rx, struct bfi_msgq_mhdr *msghdr)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2088
struct bfi_enet_rx_cfg_rsp *cfg_rsp = &rx->bfi_enet_cmd.cfg_rsp;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2093
bfa_msgq_rsp_copy(&rx->bna->msgq, (u8 *)cfg_rsp,
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2096
rx->hw_id = cfg_rsp->hw_id;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2098
for (i = 0, rxp = list_first_entry(&rx->rxp_q, struct bna_rxp, qe);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2099
i < rx->num_paths; i++, rxp = list_next_entry(rxp, qe)) {
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2104
rx->bna->pcidev.pci_bar_kva
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2108
rx->bna->pcidev.pci_bar_kva
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2113
rx->bna->pcidev.pci_bar_kva
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2126
bfa_fsm_send_event(rx, RX_E_STARTED);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2130
bna_bfi_rx_enet_stop_rsp(struct bna_rx *rx, struct bfi_msgq_mhdr *msghdr)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2132
bfa_fsm_send_event(rx, RX_E_STOPPED);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
216
0, rxf->rx->rid);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
222
bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2257
struct bna_rx *rx;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2307
rx = bna_rx_get(rx_mod, rx_cfg->rx_type);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2308
rx->bna = bna;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2309
rx->rx_flags = 0;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
231
BFI_ENET_H2I_MAC_MCAST_FILTER_REQ, 0, rxf->rx->rid);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2310
INIT_LIST_HEAD(&rx->rxp_q);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2311
rx->stop_cbfn = NULL;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2312
rx->stop_cbarg = NULL;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2313
rx->priv = priv;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2315
rx->rcb_setup_cbfn = rx_cbfn->rcb_setup_cbfn;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2316
rx->rcb_destroy_cbfn = rx_cbfn->rcb_destroy_cbfn;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2317
rx->ccb_setup_cbfn = rx_cbfn->ccb_setup_cbfn;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2318
rx->ccb_destroy_cbfn = rx_cbfn->ccb_destroy_cbfn;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2319
rx->rx_stall_cbfn = rx_cbfn->rx_stall_cbfn;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2321
rx->rx_cleanup_cbfn = rx_cbfn->rx_cleanup_cbfn;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2322
rx->rx_post_cbfn = rx_cbfn->rx_post_cbfn;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2324
if (rx->bna->rx_mod.flags & BNA_RX_MOD_F_ENET_STARTED) {
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2325
switch (rx->type) {
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2327
if (!(rx->bna->rx_mod.flags &
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2329
rx->rx_flags |= BNA_RX_F_ENET_STARTED;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2332
if (rx->bna->rx_mod.flags & BNA_RX_MOD_F_ENET_LOOPBACK)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2333
rx->rx_flags |= BNA_RX_F_ENET_STARTED;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2338
rx->num_paths = rx_cfg->num_paths;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2340
i < rx->num_paths; i++) {
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2342
list_add_tail(&rxp->qe, &rx->rxp_q);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2344
rxp->rx = rx;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2345
rxp->cq.rx = rx;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
237
bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2379
q0->rx = rx;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2400
if (rx->rcb_setup_cbfn)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2401
rx->rcb_setup_cbfn(bnad, q0->rcb);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2406
q1->rx = rx;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2430
if (rx->rcb_setup_cbfn)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2431
rx->rcb_setup_cbfn(bnad, q1->rcb);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
246
BFI_ENET_H2I_RX_PROMISCUOUS_REQ, 0, rxf->rx->rid);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2467
if (rx->ccb_setup_cbfn)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2468
rx->ccb_setup_cbfn(bnad, rxp->cq.ccb);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2471
rx->hds_cfg = rx_cfg->hds_config;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2473
bna_rxf_init(&rx->rxf, rx, rx_cfg, res_info);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2475
bfa_fsm_set_state(rx, bna_rx_sm_stopped);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2477
rx_mod->rid_mask |= BIT(rx->rid);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2479
return rx;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2483
bna_rx_destroy(struct bna_rx *rx)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2485
struct bna_rx_mod *rx_mod = &rx->bna->rx_mod;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2491
bna_rxf_uninit(&rx->rxf);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2493
while (!list_empty(&rx->rxp_q)) {
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2494
rxp = list_first_entry(&rx->rxp_q, struct bna_rxp, qe);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2497
if (rx->rcb_destroy_cbfn)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2498
rx->rcb_destroy_cbfn(rx->bna->bnad, q0->rcb);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2501
q0->rx = NULL;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2505
if (rx->rcb_destroy_cbfn)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2506
rx->rcb_destroy_cbfn(rx->bna->bnad, q1->rcb);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2509
q1->rx = NULL;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2515
if (rx->ccb_destroy_cbfn)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2516
rx->ccb_destroy_cbfn(rx->bna->bnad, rxp->cq.ccb);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2518
rxp->rx = NULL;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
252
bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2523
if (qe == &rx->qe) {
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2524
list_del(&rx->qe);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2528
rx_mod->rid_mask &= ~BIT(rx->rid);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2530
rx->bna = NULL;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2531
rx->priv = NULL;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2532
bna_rx_put(rx_mod, rx);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2536
bna_rx_enable(struct bna_rx *rx)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2538
if (rx->fsm != bna_rx_sm_stopped)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2541
rx->rx_flags |= BNA_RX_F_ENABLED;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2542
if (rx->rx_flags & BNA_RX_F_ENET_STARTED)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2543
bfa_fsm_send_event(rx, RX_E_START);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2547
bna_rx_disable(struct bna_rx *rx, enum bna_cleanup_type type,
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2552
(*cbfn)(rx->bna->bnad, rx);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2554
rx->stop_cbfn = cbfn;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2555
rx->stop_cbarg = rx->bna->bnad;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2557
rx->rx_flags &= ~BNA_RX_F_ENABLED;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2559
bfa_fsm_send_event(rx, RX_E_STOP);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2564
bna_rx_cleanup_complete(struct bna_rx *rx)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2566
bfa_fsm_send_event(rx, RX_E_CLEANUP_DONE);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2570
bna_rx_vlan_strip_enable(struct bna_rx *rx)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2572
struct bna_rxf *rxf = &rx->rxf;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2582
bna_rx_vlan_strip_disable(struct bna_rx *rx)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2584
struct bna_rxf *rxf = &rx->rxf;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2594
bna_rx_mode_set(struct bna_rx *rx, enum bna_rxmode new_mode,
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2597
struct bna_rxf *rxf = &rx->rxf;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2604
if ((rx->bna->promisc_rid != BFI_INVALID_RID) &&
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2605
(rx->bna->promisc_rid != rxf->rx->rid))
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2609
if (rx->bna->default_mode_rid != BFI_INVALID_RID)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2619
if ((rx->bna->default_mode_rid != BFI_INVALID_RID) &&
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2620
(rx->bna->default_mode_rid != rxf->rx->rid)) {
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2625
if (rx->bna->promisc_rid != BFI_INVALID_RID)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
263
BFI_ENET_H2I_RX_VLAN_SET_REQ, 0, rxf->rx->rid);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2651
rxf->cam_fltr_cbarg = rx->bna->bnad;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2662
bna_rx_vlanfilter_enable(struct bna_rx *rx)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2664
struct bna_rxf *rxf = &rx->rxf;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2674
bna_rx_coalescing_timeo_set(struct bna_rx *rx, int coalescing_timeo)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2678
list_for_each_entry(rxp, &rx->rxp_q, qe) {
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
2697
struct bna *bna = ccb->cq->rx->bna;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
277
bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
286
BFI_ENET_H2I_RX_VLAN_STRIP_ENABLE_REQ, 0, rxf->rx->rid);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
292
bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
301
BFI_ENET_H2I_RIT_CFG_REQ, 0, rxf->rx->rid);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
308
bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
318
BFI_ENET_H2I_RSS_CFG_REQ, 0, rxf->rx->rid);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
328
bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
337
BFI_ENET_H2I_RSS_ENABLE_REQ, 0, rxf->rx->rid);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
343
bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
384
mchandle = bna_mcam_mod_handle_get(&rxf->rx->bna->mcam_mod);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
411
bna_mcam_mod_handle_put(&rxf->rx->bna->mcam_mod, mchandle);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
429
list_move_tail(&mac->qe, bna_mcam_mod_del_q(rxf->rx->bna));
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
477
list_move_tail(&mac->qe, bna_mcam_mod_del_q(rxf->rx->bna));
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
561
struct bna_rx *rx = rxf->rx;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
566
list_for_each_entry(rxp, &rx->rxp_q, qe) {
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
609
struct bna_rx *rx,
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
613
rxf->rx = rx;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
661
list_move_tail(&mac->qe, bna_ucam_mod_free_q(rxf->rx->bna));
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
666
bna_ucam_mod_free_q(rxf->rx->bna));
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
673
list_move_tail(&mac->qe, bna_mcam_mod_free_q(rxf->rx->bna));
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
678
if (rxf->rx->bna->promisc_rid == rxf->rx->rid)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
679
rxf->rx->bna->promisc_rid = BFI_INVALID_RID;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
680
if (rxf->rx->bna->default_mode_rid == rxf->rx->rid)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
681
rxf->rx->bna->default_mode_rid = BFI_INVALID_RID;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
686
rxf->rx = NULL;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
690
bna_rx_cb_rxf_started(struct bna_rx *rx)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
692
bfa_fsm_send_event(rx, RX_E_RXF_STARTED);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
699
rxf->start_cbarg = rxf->rx;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
704
bna_rx_cb_rxf_stopped(struct bna_rx *rx)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
706
bfa_fsm_send_event(rx, RX_E_RXF_STOPPED);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
713
rxf->stop_cbarg = rxf->rx;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
724
bna_rx_ucast_set(struct bna_rx *rx, const u8 *ucmac)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
726
struct bna_rxf *rxf = &rx->rxf;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
730
bna_cam_mod_mac_get(bna_ucam_mod_free_q(rxf->rx->bna));
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
738
rxf->cam_fltr_cbarg = rx->bna->bnad;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
746
bna_rx_mcast_add(struct bna_rx *rx, const u8 *addr,
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
749
struct bna_rxf *rxf = &rx->rxf;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
756
cbfn(rx->bna->bnad, rx);
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
760
mac = bna_cam_mod_mac_get(bna_mcam_mod_free_q(rxf->rx->bna));
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
767
rxf->cam_fltr_cbarg = rx->bna->bnad;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
775
bna_rx_ucast_listset(struct bna_rx *rx, int count, const u8 *uclist)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
777
struct bna_ucam_mod *ucam_mod = &rx->bna->ucam_mod;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
778
struct bna_rxf *rxf = &rx->rxf;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
833
bna_rx_mcast_listset(struct bna_rx *rx, int count, const u8 *mclist)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
835
struct bna_mcam_mod *mcam_mod = &rx->bna->mcam_mod;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
836
struct bna_rxf *rxf = &rx->rxf;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
893
bna_rx_mcast_delall(struct bna_rx *rx)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
895
struct bna_rxf *rxf = &rx->rxf;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
903
list_move_tail(&mac->qe, bna_mcam_mod_free_q(rxf->rx->bna));
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
911
del_mac = bna_cam_mod_mac_get(bna_mcam_mod_del_q(rxf->rx->bna));
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
915
list_add_tail(&mac->qe, bna_mcam_mod_free_q(rxf->rx->bna));
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
924
bna_rx_vlan_add(struct bna_rx *rx, int vlan_id)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
926
struct bna_rxf *rxf = &rx->rxf;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
939
bna_rx_vlan_del(struct bna_rx *rx, int vlan_id)
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
941
struct bna_rxf *rxf = &rx->rxf;
drivers/net/ethernet/brocade/bna/bna_tx_rx.c
963
list_move_tail(&mac->qe, bna_ucam_mod_del_q(rxf->rx->bna));
drivers/net/ethernet/brocade/bna/bna_types.h
583
struct bna_rx *rx;
drivers/net/ethernet/brocade/bna/bna_types.h
649
struct bna_rx *rx;
drivers/net/ethernet/brocade/bna/bna_types.h
702
struct bna_rx *rx;
drivers/net/ethernet/brocade/bna/bna_types.h
728
void (*start_cbfn) (struct bna_rx *rx);
drivers/net/ethernet/brocade/bna/bna_types.h
732
void (*stop_cbfn) (struct bna_rx *rx);
drivers/net/ethernet/brocade/bna/bna_types.h
742
void (*cam_fltr_cbfn)(struct bnad *bnad, struct bna_rx *rx);
drivers/net/ethernet/brocade/bna/bna_types.h
780
struct bna_rx *rx;
drivers/net/ethernet/brocade/bna/bna_types.h
823
void (*stop_cbfn)(void *arg, struct bna_rx *rx);
drivers/net/ethernet/brocade/bna/bna_types.h
845
struct bna_rx *rx; /* BFI_MAX_RXQ entries */
drivers/net/ethernet/brocade/bna/bnad.c
1023
(struct bnad_rx_info *)ccb->cq->rx->priv;
drivers/net/ethernet/brocade/bna/bnad.c
1033
(struct bnad_rx_info *)ccb->cq->rx->priv;
drivers/net/ethernet/brocade/bna/bnad.c
1150
bnad_cb_rx_stall(struct bnad *bnad, struct bna_rx *rx)
drivers/net/ethernet/brocade/bna/bnad.c
1152
struct bnad_rx_info *rx_info = rx->priv;
drivers/net/ethernet/brocade/bna/bnad.c
1204
bna_rx_cleanup_complete(rx_info->rx);
drivers/net/ethernet/brocade/bna/bnad.c
1209
bnad_cb_rx_cleanup(struct bnad *bnad, struct bna_rx *rx)
drivers/net/ethernet/brocade/bna/bnad.c
1211
struct bnad_rx_info *rx_info = rx->priv;
drivers/net/ethernet/brocade/bna/bnad.c
1232
bnad_cb_rx_post(struct bnad *bnad, struct bna_rx *rx)
drivers/net/ethernet/brocade/bna/bnad.c
1234
struct bnad_rx_info *rx_info = rx->priv;
drivers/net/ethernet/brocade/bna/bnad.c
1262
bnad_cb_rx_disabled(void *arg, struct bna_rx *rx)
drivers/net/ethernet/brocade/bna/bnad.c
1270
bnad_cb_rx_mcast_add(struct bnad *bnad, struct bna_rx *rx)
drivers/net/ethernet/brocade/bna/bnad.c
1759
if (!rx_info->rx)
drivers/net/ethernet/brocade/bna/bnad.c
2110
if (!bnad->rx_info[rx_id].rx)
drivers/net/ethernet/brocade/bna/bnad.c
2130
if (bnad->rx_info[0].rx && !err) {
drivers/net/ethernet/brocade/bna/bnad.c
2152
if (!rx_info->rx)
drivers/net/ethernet/brocade/bna/bnad.c
2169
bna_rx_disable(rx_info->rx, BNA_HARD_CLEANUP, bnad_cb_rx_disabled);
drivers/net/ethernet/brocade/bna/bnad.c
2179
bna_rx_destroy(rx_info->rx);
drivers/net/ethernet/brocade/bna/bnad.c
2181
rx_info->rx = NULL;
drivers/net/ethernet/brocade/bna/bnad.c
2207
struct bna_rx *rx;
drivers/net/ethernet/brocade/bna/bnad.c
2243
rx = bna_rx_create(&bnad->bna, bnad, rx_config, &rx_cbfn, res_info,
drivers/net/ethernet/brocade/bna/bnad.c
2245
if (!rx) {
drivers/net/ethernet/brocade/bna/bnad.c
2250
rx_info->rx = rx;
drivers/net/ethernet/brocade/bna/bnad.c
2276
bna_rx_vlanfilter_enable(rx);
drivers/net/ethernet/brocade/bna/bnad.c
2282
bna_rx_enable(rx);
drivers/net/ethernet/brocade/bna/bnad.c
2314
if (!rx_info->rx)
drivers/net/ethernet/brocade/bna/bnad.c
2316
bna_rx_coalescing_timeo_set(rx_info->rx,
drivers/net/ethernet/brocade/bna/bnad.c
2333
if (!bnad->rx_info[0].rx)
drivers/net/ethernet/brocade/bna/bnad.c
2336
ret = bna_rx_ucast_set(bnad->rx_info[0].rx, mac_addr);
drivers/net/ethernet/brocade/bna/bnad.c
2354
ret = bna_rx_mcast_add(rx_info->rx, bnad_bcast_addr,
drivers/net/ethernet/brocade/bna/bnad.c
2378
bna_rx_vlan_add(bnad->rx_info[rx_id].rx, vid);
drivers/net/ethernet/brocade/bna/bnad.c
3121
bna_rx_ucast_listset(bnad->rx_info[0].rx, 0, NULL);
drivers/net/ethernet/brocade/bna/bnad.c
3138
ret = bna_rx_ucast_listset(bnad->rx_info[0].rx, entry, mac_list);
drivers/net/ethernet/brocade/bna/bnad.c
3149
bna_rx_ucast_listset(bnad->rx_info[0].rx, 0, NULL);
drivers/net/ethernet/brocade/bna/bnad.c
3178
ret = bna_rx_mcast_listset(bnad->rx_info[0].rx, mc_count + 1, mac_list);
drivers/net/ethernet/brocade/bna/bnad.c
3188
bna_rx_mcast_delall(bnad->rx_info[0].rx);
drivers/net/ethernet/brocade/bna/bnad.c
3200
if (bnad->rx_info[0].rx == NULL) {
drivers/net/ethernet/brocade/bna/bnad.c
3227
bna_rx_mode_set(bnad->rx_info[0].rx, new_mode, mode_mask);
drivers/net/ethernet/brocade/bna/bnad.c
3310
if (!bnad->rx_info[0].rx)
drivers/net/ethernet/brocade/bna/bnad.c
3316
bna_rx_vlan_add(bnad->rx_info[0].rx, vid);
drivers/net/ethernet/brocade/bna/bnad.c
3331
if (!bnad->rx_info[0].rx)
drivers/net/ethernet/brocade/bna/bnad.c
3338
bna_rx_vlan_del(bnad->rx_info[0].rx, vid);
drivers/net/ethernet/brocade/bna/bnad.c
3357
bna_rx_vlan_strip_enable(bnad->rx_info[0].rx);
drivers/net/ethernet/brocade/bna/bnad.c
3359
bna_rx_vlan_strip_disable(bnad->rx_info[0].rx);
drivers/net/ethernet/brocade/bna/bnad.c
3390
if (!rx_info->rx)
drivers/net/ethernet/brocade/bna/bnad.c
836
if (!rx_info->rx)
drivers/net/ethernet/brocade/bna/bnad.h
208
struct bna_rx *rx; /* 1:1 between rx_info & rx */
drivers/net/ethernet/brocade/bna/bnad_ethtool.c
459
if (!bnad->rx_info[i].rx)
drivers/net/ethernet/brocade/bna/bnad_ethtool.c
467
if (!err && bnad->rx_info[0].rx) {
drivers/net/ethernet/brocade/bna/bnad_ethtool.c
630
if (!bnad->rx_info[i].rx)
drivers/net/ethernet/brocade/bna/bnad_ethtool.c
638
if (!bnad->rx_info[i].rx)
drivers/net/ethernet/brocade/bna/bnad_ethtool.c
684
if (!bnad->rx_info[i].rx)
drivers/net/ethernet/brocade/bna/bnad_ethtool.c
711
if (!bnad->rx_info[i].rx)
drivers/net/ethernet/brocade/bna/bnad_ethtool.c
736
if (!bnad->rx_info[i].rx)
drivers/net/ethernet/cadence/macb_main.c
2603
void *tx, *rx;
drivers/net/ethernet/cadence/macb_main.c
2621
rx = dma_alloc_coherent(dev, size, &rx_dma, GFP_KERNEL);
drivers/net/ethernet/cadence/macb_main.c
2622
if (!rx || upper_32_bits(rx_dma) != upper_32_bits(rx_dma + size - 1))
drivers/net/ethernet/cadence/macb_main.c
2625
size, bp->num_queues, (unsigned long)rx_dma, rx);
drivers/net/ethernet/cadence/macb_main.c
2631
queue->rx_ring = rx + macb_rx_ring_size_per_queue(bp) * q;
drivers/net/ethernet/calxeda/xgmac.c
644
static int xgmac_set_flow_ctrl(struct xgmac_priv *priv, int rx, int tx)
drivers/net/ethernet/calxeda/xgmac.c
649
priv->rx_pause = rx;
drivers/net/ethernet/calxeda/xgmac.c
652
if (rx || tx) {
drivers/net/ethernet/calxeda/xgmac.c
653
if (rx)
drivers/net/ethernet/cavium/thunder/nic.h
503
u8 rx;
drivers/net/ethernet/cavium/thunder/nic_main.c
241
mbx.bgx_stats.rx = bgx->rx;
drivers/net/ethernet/cavium/thunder/nic_main.c
243
if (bgx->rx)
drivers/net/ethernet/cavium/thunder/nicvf_main.c
1630
mbx.bgx_stats.rx = 1;
drivers/net/ethernet/cavium/thunder/nicvf_main.c
1641
mbx.bgx_stats.rx = 0;
drivers/net/ethernet/cavium/thunder/nicvf_main.c
194
if (bgx->rx)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
2680
#define R3(fmt_spec, s, v) S3(fmt_spec, s, rx[i].v)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
2681
#define R(s, v) S3("u", s, rx[i].v)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
2686
const struct sge_eth_rxq *rx = &s->ethrxq[base_qset];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
2693
rx[i].rspq.netdev ? rx[i].rspq.netdev->name : "N/A");
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
2713
S3("u", "Intr delay:", qtimer_val(adap, &rx[i].rspq));
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
2714
S3("u", "Intr pktcnt:", s->counter_val[rx[i].rspq.pktcnt_idx]);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
2747
const struct sge_eth_rxq *rx;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
2761
rx = &s->mirror_rxq[j][r * 4];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
2766
rx[i].rspq.netdev ? rx[i].rspq.netdev->name : "N/A");
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
2772
S3("u", "Intr delay:", qtimer_val(adap, &rx[i].rspq));
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
2773
S3("u", "Intr pktcnt:", s->counter_val[rx[i].rspq.pktcnt_idx]);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
2809
const struct sge_ofld_rxq *rx = &s->eohw_rxq[base_qset];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
2816
rx[i].rspq.netdev ? rx[i].rspq.netdev->name : "N/A");
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
2827
S3("u", "Intr delay:", qtimer_val(adap, &rx[i].rspq));
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
2828
S3("u", "Intr pktcnt:", s->counter_val[rx[i].rspq.pktcnt_idx]);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
2830
S3("u", "FL size:", rx->fl.size ? rx->fl.size - 8 : 0);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
2936
const struct sge_ofld_rxq *rx;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
2939
rx = &urxq_info->uldrxq[r * 4];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
2944
rx[i].rspq.netdev ? rx[i].rspq.netdev->name : "N/A");
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
2950
S3("u", "Intr delay:", qtimer_val(adap, &rx[i].rspq));
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
2951
S3("u", "Intr pktcnt:", s->counter_val[rx[i].rspq.pktcnt_idx]);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
2964
const struct sge_ofld_rxq *rx;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
2969
rx = &urxq_info->uldrxq[ciq_idx];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
2974
rx[i].rspq.netdev ? rx[i].rspq.netdev->name : "N/A");
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
2980
S3("u", "Intr delay:", qtimer_val(adap, &rx[i].rspq));
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
2981
S3("u", "Intr pktcnt:", s->counter_val[rx[i].rspq.pktcnt_idx]);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
2988
const struct sge_ofld_rxq *rx;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
2991
rx = &urxq_info->uldrxq[r * 4];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3000
S3("u", "Intr delay:", qtimer_val(adap, &rx[i].rspq));
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3001
S3("u", "Intr pktcnt:", s->counter_val[rx[i].rspq.pktcnt_idx]);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3014
const struct sge_ofld_rxq *rx;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3017
rx = &urxq_info->uldrxq[r * 4];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3026
S3("u", "Intr delay:", qtimer_val(adap, &rx[i].rspq));
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3027
S3("u", "Intr pktcnt:", s->counter_val[rx[i].rspq.pktcnt_idx]);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3040
const struct sge_ofld_rxq *rx;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3043
rx = &urxq_info->uldrxq[r * 4];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3052
S3("u", "Intr delay:", qtimer_val(adap, &rx[i].rspq));
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3053
S3("u", "Intr pktcnt:", s->counter_val[rx[i].rspq.pktcnt_idx]);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3066
const struct sge_ofld_rxq *rx;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3072
rx = &urxq_info->uldrxq[r * 4];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3086
S3("u", "Intr delay:", qtimer_val(adap, &rx[i].rspq));
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3087
S3("u", "Intr pktcnt:", s->counter_val[rx[i].rspq.pktcnt_idx]);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
284
const struct sge_eth_rxq *rx = &adap->sge.ethrxq[p->first_qset];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
292
for (i = 0; i < p->nqsets; i++, rx++, tx++) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
296
s->rx_csum += rx->stats.rx_cso;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
297
s->vlan_ex += rx->stats.vlan_ex;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
299
s->gro_pkts += rx->stats.lro_pkts;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_ethtool.c
300
s->gro_merged += rx->stats.lro_merged;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
3473
struct sge_eth_rxq *rx = &adap->sge.ethrxq[pi->first_qset];
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
3475
for (i = pi->nqsets; i; i--, rx++)
drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c
3476
t4_sge_intr_msix(0, &rx->rspq);
drivers/net/ethernet/cirrus/ep93xx_eth.c
312
int rx;
drivers/net/ethernet/cirrus/ep93xx_eth.c
314
rx = ep93xx_rx(dev, budget);
drivers/net/ethernet/cirrus/ep93xx_eth.c
315
if (rx < budget && napi_complete_done(napi, rx)) {
drivers/net/ethernet/cirrus/ep93xx_eth.c
321
if (rx) {
drivers/net/ethernet/cirrus/ep93xx_eth.c
322
wrw(ep, REG_RXDENQ, rx);
drivers/net/ethernet/cirrus/ep93xx_eth.c
323
wrw(ep, REG_RXSTSENQ, rx);
drivers/net/ethernet/cirrus/ep93xx_eth.c
326
return rx;
drivers/net/ethernet/cisco/enic/enic_ethtool.c
336
*(data++) = ((u64 *)&vstats->rx)[enic_rx_stats[i].index];
drivers/net/ethernet/cisco/enic/enic_main.c
918
net_stats->rx_packets = stats->rx.rx_frames_ok;
drivers/net/ethernet/cisco/enic/enic_main.c
919
net_stats->rx_bytes = stats->rx.rx_bytes_ok;
drivers/net/ethernet/cisco/enic/enic_main.c
920
net_stats->rx_errors = stats->rx.rx_errors;
drivers/net/ethernet/cisco/enic/enic_main.c
921
net_stats->multicast = stats->rx.rx_multicast_frames_ok;
drivers/net/ethernet/cisco/enic/enic_main.c
933
net_stats->rx_dropped = stats->rx.rx_no_bufs + stats->rx.rx_drop;
drivers/net/ethernet/cisco/enic/vnic_stats.h
59
struct vnic_rx_stats rx;
drivers/net/ethernet/cortina/gemini.c
1446
struct gmac_rxdesc *rx = NULL;
drivers/net/ethernet/cortina/gemini.c
1471
rx = port->rxq_ring + r;
drivers/net/ethernet/cortina/gemini.c
1472
word0 = rx->word0;
drivers/net/ethernet/cortina/gemini.c
1473
word1 = rx->word1;
drivers/net/ethernet/cortina/gemini.c
1474
mapping = rx->word2.buf_adr;
drivers/net/ethernet/cortina/gemini.c
1475
word3 = rx->word3;
drivers/net/ethernet/cortina/gemini.c
273
static void gmac_set_flow_control(struct net_device *netdev, bool tx, bool rx)
drivers/net/ethernet/cortina/gemini.c
285
if (rx)
drivers/net/ethernet/dec/tulip/interrupt.c
529
int rx = 0;
drivers/net/ethernet/dec/tulip/interrupt.c
579
rx += tulip_rx(dev);
drivers/net/ethernet/dec/tulip/interrupt.c
739
if (tx > maxtx || rx > maxrx || oi > maxoi) {
drivers/net/ethernet/dec/tulip/interrupt.c
742
csr5, tp->nir, tx, rx, oi);
drivers/net/ethernet/dec/tulip/interrupt.c
793
tp->nir, tp->cur_rx, tp->ttimer, rx);
drivers/net/ethernet/engleder/tsnep.h
149
struct tsnep_rx *rx;
drivers/net/ethernet/engleder/tsnep.h
196
struct tsnep_rx rx[TSNEP_MAX_QUEUES];
drivers/net/ethernet/engleder/tsnep_ethtool.c
181
tsnep_stats.rx_packets += adapter->rx[i].packets;
drivers/net/ethernet/engleder/tsnep_ethtool.c
182
tsnep_stats.rx_bytes += adapter->rx[i].bytes;
drivers/net/ethernet/engleder/tsnep_ethtool.c
183
tsnep_stats.rx_dropped += adapter->rx[i].dropped;
drivers/net/ethernet/engleder/tsnep_ethtool.c
184
tsnep_stats.rx_multicast += adapter->rx[i].multicast;
drivers/net/ethernet/engleder/tsnep_ethtool.c
185
tsnep_stats.rx_alloc_failed += adapter->rx[i].alloc_failed;
drivers/net/ethernet/engleder/tsnep_ethtool.c
204
tsnep_rx_queue_stats.rx_packets = adapter->rx[i].packets;
drivers/net/ethernet/engleder/tsnep_ethtool.c
205
tsnep_rx_queue_stats.rx_bytes = adapter->rx[i].bytes;
drivers/net/ethernet/engleder/tsnep_ethtool.c
206
tsnep_rx_queue_stats.rx_dropped = adapter->rx[i].dropped;
drivers/net/ethernet/engleder/tsnep_ethtool.c
207
tsnep_rx_queue_stats.rx_multicast = adapter->rx[i].multicast;
drivers/net/ethernet/engleder/tsnep_ethtool.c
209
adapter->rx[i].alloc_failed;
drivers/net/ethernet/engleder/tsnep_ethtool.c
355
if (adapter->queue[i].rx) {
drivers/net/ethernet/engleder/tsnep_ethtool.c
396
if (adapter->queue[i].rx)
drivers/net/ethernet/engleder/tsnep_ethtool.c
452
if (queue_with_tx && !queue_with_tx->rx) {
drivers/net/ethernet/engleder/tsnep_hw.h
229
__le64 rx;
drivers/net/ethernet/engleder/tsnep_main.c
1001
entry = &rx->entry[TSNEP_RING_ENTRIES_PER_PAGE * i + j];
drivers/net/ethernet/engleder/tsnep_main.c
1003
(((u8 *)rx->page[i]) + TSNEP_DESC_SIZE * j);
drivers/net/ethernet/engleder/tsnep_main.c
1006
entry->desc_dma = rx->page_dma[i] + TSNEP_DESC_SIZE * j;
drivers/net/ethernet/engleder/tsnep_main.c
1018
rx->page_pool = page_pool_create(&pp_params);
drivers/net/ethernet/engleder/tsnep_main.c
1019
if (IS_ERR(rx->page_pool)) {
drivers/net/ethernet/engleder/tsnep_main.c
1020
retval = PTR_ERR(rx->page_pool);
drivers/net/ethernet/engleder/tsnep_main.c
1021
rx->page_pool = NULL;
drivers/net/ethernet/engleder/tsnep_main.c
1026
entry = &rx->entry[i];
drivers/net/ethernet/engleder/tsnep_main.c
1027
next_entry = &rx->entry[(i + 1) & TSNEP_RING_MASK];
drivers/net/ethernet/engleder/tsnep_main.c
1034
tsnep_rx_ring_cleanup(rx);
drivers/net/ethernet/engleder/tsnep_main.c
1038
static void tsnep_rx_init(struct tsnep_rx *rx)
drivers/net/ethernet/engleder/tsnep_main.c
1042
dma = rx->entry[0].desc_dma | TSNEP_RESET_OWNER_COUNTER;
drivers/net/ethernet/engleder/tsnep_main.c
1043
iowrite32(DMA_ADDR_LOW(dma), rx->addr + TSNEP_RX_DESC_ADDR_LOW);
drivers/net/ethernet/engleder/tsnep_main.c
1044
iowrite32(DMA_ADDR_HIGH(dma), rx->addr + TSNEP_RX_DESC_ADDR_HIGH);
drivers/net/ethernet/engleder/tsnep_main.c
1045
rx->write = 0;
drivers/net/ethernet/engleder/tsnep_main.c
1046
rx->read = 0;
drivers/net/ethernet/engleder/tsnep_main.c
1047
rx->owner_counter = 1;
drivers/net/ethernet/engleder/tsnep_main.c
1048
rx->increment_owner_counter = TSNEP_RING_SIZE - 1;
drivers/net/ethernet/engleder/tsnep_main.c
1051
static void tsnep_rx_enable(struct tsnep_rx *rx)
drivers/net/ethernet/engleder/tsnep_main.c
1056
iowrite32(TSNEP_CONTROL_RX_ENABLE, rx->addr + TSNEP_CONTROL);
drivers/net/ethernet/engleder/tsnep_main.c
1059
static void tsnep_rx_disable(struct tsnep_rx *rx)
drivers/net/ethernet/engleder/tsnep_main.c
1063
iowrite32(TSNEP_CONTROL_RX_DISABLE, rx->addr + TSNEP_CONTROL);
drivers/net/ethernet/engleder/tsnep_main.c
1064
readx_poll_timeout(ioread32, rx->addr + TSNEP_CONTROL, val,
drivers/net/ethernet/engleder/tsnep_main.c
1069
static int tsnep_rx_desc_available(struct tsnep_rx *rx)
drivers/net/ethernet/engleder/tsnep_main.c
1071
if (rx->read <= rx->write)
drivers/net/ethernet/engleder/tsnep_main.c
1072
return TSNEP_RING_SIZE - rx->write + rx->read - 1;
drivers/net/ethernet/engleder/tsnep_main.c
1074
return rx->read - rx->write - 1;
drivers/net/ethernet/engleder/tsnep_main.c
1077
static void tsnep_rx_free_page_buffer(struct tsnep_rx *rx)
drivers/net/ethernet/engleder/tsnep_main.c
1084
page = rx->page_buffer;
drivers/net/ethernet/engleder/tsnep_main.c
1086
page_pool_put_full_page(rx->page_pool, *page, false);
drivers/net/ethernet/engleder/tsnep_main.c
1092
static int tsnep_rx_alloc_page_buffer(struct tsnep_rx *rx)
drivers/net/ethernet/engleder/tsnep_main.c
1100
rx->page_buffer[i] = page_pool_dev_alloc_pages(rx->page_pool);
drivers/net/ethernet/engleder/tsnep_main.c
1101
if (!rx->page_buffer[i]) {
drivers/net/ethernet/engleder/tsnep_main.c
1102
tsnep_rx_free_page_buffer(rx);
drivers/net/ethernet/engleder/tsnep_main.c
1111
static void tsnep_rx_set_page(struct tsnep_rx *rx, struct tsnep_rx_entry *entry,
drivers/net/ethernet/engleder/tsnep_main.c
1117
entry->desc->rx = __cpu_to_le64(entry->dma + TSNEP_RX_OFFSET);
drivers/net/ethernet/engleder/tsnep_main.c
1120
static int tsnep_rx_alloc_buffer(struct tsnep_rx *rx, int index)
drivers/net/ethernet/engleder/tsnep_main.c
1122
struct tsnep_rx_entry *entry = &rx->entry[index];
drivers/net/ethernet/engleder/tsnep_main.c
1125
page = page_pool_dev_alloc_pages(rx->page_pool);
drivers/net/ethernet/engleder/tsnep_main.c
1128
tsnep_rx_set_page(rx, entry, page);
drivers/net/ethernet/engleder/tsnep_main.c
1133
static void tsnep_rx_reuse_buffer(struct tsnep_rx *rx, int index)
drivers/net/ethernet/engleder/tsnep_main.c
1135
struct tsnep_rx_entry *entry = &rx->entry[index];
drivers/net/ethernet/engleder/tsnep_main.c
1136
struct tsnep_rx_entry *read = &rx->entry[rx->read];
drivers/net/ethernet/engleder/tsnep_main.c
1138
tsnep_rx_set_page(rx, entry, read->page);
drivers/net/ethernet/engleder/tsnep_main.c
1142
static void tsnep_rx_activate(struct tsnep_rx *rx, int index)
drivers/net/ethernet/engleder/tsnep_main.c
1144
struct tsnep_rx_entry *entry = &rx->entry[index];
drivers/net/ethernet/engleder/tsnep_main.c
1149
if (index == rx->increment_owner_counter) {
drivers/net/ethernet/engleder/tsnep_main.c
1150
rx->owner_counter++;
drivers/net/ethernet/engleder/tsnep_main.c
1151
if (rx->owner_counter == 4)
drivers/net/ethernet/engleder/tsnep_main.c
1152
rx->owner_counter = 1;
drivers/net/ethernet/engleder/tsnep_main.c
1153
rx->increment_owner_counter--;
drivers/net/ethernet/engleder/tsnep_main.c
1154
if (rx->increment_owner_counter < 0)
drivers/net/ethernet/engleder/tsnep_main.c
1155
rx->increment_owner_counter = TSNEP_RING_SIZE - 1;
drivers/net/ethernet/engleder/tsnep_main.c
1158
(rx->owner_counter << TSNEP_DESC_OWNER_COUNTER_SHIFT) &
drivers/net/ethernet/engleder/tsnep_main.c
1169
static int tsnep_rx_alloc(struct tsnep_rx *rx, int count, bool reuse)
drivers/net/ethernet/engleder/tsnep_main.c
1175
index = (rx->write + i) & TSNEP_RING_MASK;
drivers/net/ethernet/engleder/tsnep_main.c
1177
if (unlikely(tsnep_rx_alloc_buffer(rx, index))) {
drivers/net/ethernet/engleder/tsnep_main.c
1178
rx->alloc_failed++;
drivers/net/ethernet/engleder/tsnep_main.c
1183
tsnep_rx_reuse_buffer(rx, index);
drivers/net/ethernet/engleder/tsnep_main.c
1188
tsnep_rx_activate(rx, index);
drivers/net/ethernet/engleder/tsnep_main.c
1192
rx->write = (rx->write + i) & TSNEP_RING_MASK;
drivers/net/ethernet/engleder/tsnep_main.c
1197
static int tsnep_rx_refill(struct tsnep_rx *rx, int count, bool reuse)
drivers/net/ethernet/engleder/tsnep_main.c
1201
desc_refilled = tsnep_rx_alloc(rx, count, reuse);
drivers/net/ethernet/engleder/tsnep_main.c
1203
tsnep_rx_enable(rx);
drivers/net/ethernet/engleder/tsnep_main.c
1208
static void tsnep_rx_set_xdp(struct tsnep_rx *rx, struct tsnep_rx_entry *entry,
drivers/net/ethernet/engleder/tsnep_main.c
1214
entry->desc->rx = __cpu_to_le64(entry->dma);
drivers/net/ethernet/engleder/tsnep_main.c
1217
static void tsnep_rx_reuse_buffer_zc(struct tsnep_rx *rx, int index)
drivers/net/ethernet/engleder/tsnep_main.c
1219
struct tsnep_rx_entry *entry = &rx->entry[index];
drivers/net/ethernet/engleder/tsnep_main.c
1220
struct tsnep_rx_entry *read = &rx->entry[rx->read];
drivers/net/ethernet/engleder/tsnep_main.c
1222
tsnep_rx_set_xdp(rx, entry, read->xdp);
drivers/net/ethernet/engleder/tsnep_main.c
1226
static int tsnep_rx_alloc_zc(struct tsnep_rx *rx, int count, bool reuse)
drivers/net/ethernet/engleder/tsnep_main.c
1231
allocated = xsk_buff_alloc_batch(rx->xsk_pool, rx->xdp_batch, count);
drivers/net/ethernet/engleder/tsnep_main.c
1233
int index = (rx->write + i) & TSNEP_RING_MASK;
drivers/net/ethernet/engleder/tsnep_main.c
1234
struct tsnep_rx_entry *entry = &rx->entry[index];
drivers/net/ethernet/engleder/tsnep_main.c
1236
tsnep_rx_set_xdp(rx, entry, rx->xdp_batch[i]);
drivers/net/ethernet/engleder/tsnep_main.c
1237
tsnep_rx_activate(rx, index);
drivers/net/ethernet/engleder/tsnep_main.c
1240
rx->alloc_failed++;
drivers/net/ethernet/engleder/tsnep_main.c
1243
tsnep_rx_reuse_buffer_zc(rx, rx->write);
drivers/net/ethernet/engleder/tsnep_main.c
1244
tsnep_rx_activate(rx, rx->write);
drivers/net/ethernet/engleder/tsnep_main.c
1249
rx->write = (rx->write + i) & TSNEP_RING_MASK;
drivers/net/ethernet/engleder/tsnep_main.c
1254
static void tsnep_rx_free_zc(struct tsnep_rx *rx)
drivers/net/ethernet/engleder/tsnep_main.c
1259
struct tsnep_rx_entry *entry = &rx->entry[i];
drivers/net/ethernet/engleder/tsnep_main.c
1267
static int tsnep_rx_refill_zc(struct tsnep_rx *rx, int count, bool reuse)
drivers/net/ethernet/engleder/tsnep_main.c
1271
desc_refilled = tsnep_rx_alloc_zc(rx, count, reuse);
drivers/net/ethernet/engleder/tsnep_main.c
1273
tsnep_rx_enable(rx);
drivers/net/ethernet/engleder/tsnep_main.c
1278
static void tsnep_xsk_rx_need_wakeup(struct tsnep_rx *rx, int desc_available)
drivers/net/ethernet/engleder/tsnep_main.c
1281
xsk_set_rx_need_wakeup(rx->xsk_pool);
drivers/net/ethernet/engleder/tsnep_main.c
1283
xsk_clear_rx_need_wakeup(rx->xsk_pool);
drivers/net/ethernet/engleder/tsnep_main.c
1286
static bool tsnep_xdp_run_prog(struct tsnep_rx *rx, struct bpf_prog *prog,
drivers/net/ethernet/engleder/tsnep_main.c
1301
if (!tsnep_xdp_xmit_back(rx->adapter, xdp, tx_nq, tx, false))
drivers/net/ethernet/engleder/tsnep_main.c
1306
if (xdp_do_redirect(rx->adapter->netdev, xdp, prog) < 0)
drivers/net/ethernet/engleder/tsnep_main.c
1311
bpf_warn_invalid_xdp_action(rx->adapter->netdev, prog, act);
drivers/net/ethernet/engleder/tsnep_main.c
1315
trace_xdp_exception(rx->adapter->netdev, prog, act);
drivers/net/ethernet/engleder/tsnep_main.c
1324
page_pool_put_page(rx->page_pool, virt_to_head_page(xdp->data),
drivers/net/ethernet/engleder/tsnep_main.c
1330
static bool tsnep_xdp_run_prog_zc(struct tsnep_rx *rx, struct bpf_prog *prog,
drivers/net/ethernet/engleder/tsnep_main.c
1341
if (xdp_do_redirect(rx->adapter->netdev, xdp, prog) < 0)
drivers/net/ethernet/engleder/tsnep_main.c
1351
if (!tsnep_xdp_xmit_back(rx->adapter, xdp, tx_nq, tx, true))
drivers/net/ethernet/engleder/tsnep_main.c
1356
bpf_warn_invalid_xdp_action(rx->adapter->netdev, prog, act);
drivers/net/ethernet/engleder/tsnep_main.c
1360
trace_xdp_exception(rx->adapter->netdev, prog, act);
drivers/net/ethernet/engleder/tsnep_main.c
1381
static struct sk_buff *tsnep_build_skb(struct tsnep_rx *rx, struct page *page,
drivers/net/ethernet/engleder/tsnep_main.c
1394
if (rx->adapter->hwtstamp_config.rx_filter == HWTSTAMP_FILTER_ALL) {
drivers/net/ethernet/engleder/tsnep_main.c
1406
skb_record_rx_queue(skb, rx->queue_index);
drivers/net/ethernet/engleder/tsnep_main.c
1407
skb->protocol = eth_type_trans(skb, rx->adapter->netdev);
drivers/net/ethernet/engleder/tsnep_main.c
1412
static void tsnep_rx_page(struct tsnep_rx *rx, struct napi_struct *napi,
drivers/net/ethernet/engleder/tsnep_main.c
1417
skb = tsnep_build_skb(rx, page, length);
drivers/net/ethernet/engleder/tsnep_main.c
1421
rx->packets++;
drivers/net/ethernet/engleder/tsnep_main.c
1422
rx->bytes += length;
drivers/net/ethernet/engleder/tsnep_main.c
1424
rx->multicast++;
drivers/net/ethernet/engleder/tsnep_main.c
1428
page_pool_recycle_direct(rx->page_pool, page);
drivers/net/ethernet/engleder/tsnep_main.c
1430
rx->dropped++;
drivers/net/ethernet/engleder/tsnep_main.c
1434
static int tsnep_rx_poll(struct tsnep_rx *rx, struct napi_struct *napi,
drivers/net/ethernet/engleder/tsnep_main.c
1437
struct device *dmadev = rx->adapter->dmadev;
drivers/net/ethernet/engleder/tsnep_main.c
1449
desc_available = tsnep_rx_desc_available(rx);
drivers/net/ethernet/engleder/tsnep_main.c
1450
dma_dir = page_pool_get_dma_dir(rx->page_pool);
drivers/net/ethernet/engleder/tsnep_main.c
1451
prog = READ_ONCE(rx->adapter->xdp_prog);
drivers/net/ethernet/engleder/tsnep_main.c
1453
tx_nq = netdev_get_tx_queue(rx->adapter->netdev,
drivers/net/ethernet/engleder/tsnep_main.c
1454
rx->tx_queue_index);
drivers/net/ethernet/engleder/tsnep_main.c
1455
tx = &rx->adapter->tx[rx->tx_queue_index];
drivers/net/ethernet/engleder/tsnep_main.c
1457
xdp_init_buff(&xdp, PAGE_SIZE, &rx->xdp_rxq);
drivers/net/ethernet/engleder/tsnep_main.c
1460
while (likely(done < budget) && (rx->read != rx->write)) {
drivers/net/ethernet/engleder/tsnep_main.c
1461
entry = &rx->entry[rx->read];
drivers/net/ethernet/engleder/tsnep_main.c
1471
desc_available -= tsnep_rx_refill(rx, desc_available,
drivers/net/ethernet/engleder/tsnep_main.c
1478
rx->read = (rx->read + 1) & TSNEP_RING_MASK;
drivers/net/ethernet/engleder/tsnep_main.c
1481
rx->dropped++;
drivers/net/ethernet/engleder/tsnep_main.c
1505
rx->read = (rx->read + 1) & TSNEP_RING_MASK;
drivers/net/ethernet/engleder/tsnep_main.c
1515
consume = tsnep_xdp_run_prog(rx, prog, &xdp,
drivers/net/ethernet/engleder/tsnep_main.c
1518
rx->packets++;
drivers/net/ethernet/engleder/tsnep_main.c
1519
rx->bytes += length;
drivers/net/ethernet/engleder/tsnep_main.c
1527
tsnep_rx_page(rx, napi, entry->page, length);
drivers/net/ethernet/engleder/tsnep_main.c
1532
tsnep_finalize_xdp(rx->adapter, xdp_status, tx_nq, tx);
drivers/net/ethernet/engleder/tsnep_main.c
1535
tsnep_rx_refill(rx, desc_available, false);
drivers/net/ethernet/engleder/tsnep_main.c
1540
static int tsnep_rx_poll_zc(struct tsnep_rx *rx, struct napi_struct *napi,
drivers/net/ethernet/engleder/tsnep_main.c
1553
desc_available = tsnep_rx_desc_available(rx);
drivers/net/ethernet/engleder/tsnep_main.c
1554
prog = READ_ONCE(rx->adapter->xdp_prog);
drivers/net/ethernet/engleder/tsnep_main.c
1556
tx_nq = netdev_get_tx_queue(rx->adapter->netdev,
drivers/net/ethernet/engleder/tsnep_main.c
1557
rx->tx_queue_index);
drivers/net/ethernet/engleder/tsnep_main.c
1558
tx = &rx->adapter->tx[rx->tx_queue_index];
drivers/net/ethernet/engleder/tsnep_main.c
1561
while (likely(done < budget) && (rx->read != rx->write)) {
drivers/net/ethernet/engleder/tsnep_main.c
1562
entry = &rx->entry[rx->read];
drivers/net/ethernet/engleder/tsnep_main.c
1572
desc_available -= tsnep_rx_refill_zc(rx, desc_available,
drivers/net/ethernet/engleder/tsnep_main.c
1579
rx->read = (rx->read + 1) & TSNEP_RING_MASK;
drivers/net/ethernet/engleder/tsnep_main.c
1582
rx->dropped++;
drivers/net/ethernet/engleder/tsnep_main.c
1606
rx->read = (rx->read + 1) & TSNEP_RING_MASK;
drivers/net/ethernet/engleder/tsnep_main.c
1615
consume = tsnep_xdp_run_prog_zc(rx, prog, entry->xdp,
drivers/net/ethernet/engleder/tsnep_main.c
1618
rx->packets++;
drivers/net/ethernet/engleder/tsnep_main.c
1619
rx->bytes += length;
drivers/net/ethernet/engleder/tsnep_main.c
1627
page = page_pool_dev_alloc_pages(rx->page_pool);
drivers/net/ethernet/engleder/tsnep_main.c
1632
tsnep_rx_page(rx, napi, page, length);
drivers/net/ethernet/engleder/tsnep_main.c
1634
rx->dropped++;
drivers/net/ethernet/engleder/tsnep_main.c
1641
tsnep_finalize_xdp(rx->adapter, xdp_status, tx_nq, tx);
drivers/net/ethernet/engleder/tsnep_main.c
1644
desc_available -= tsnep_rx_refill_zc(rx, desc_available, false);
drivers/net/ethernet/engleder/tsnep_main.c
1646
if (xsk_uses_need_wakeup(rx->xsk_pool)) {
drivers/net/ethernet/engleder/tsnep_main.c
1647
tsnep_xsk_rx_need_wakeup(rx, desc_available);
drivers/net/ethernet/engleder/tsnep_main.c
1655
static bool tsnep_rx_pending(struct tsnep_rx *rx)
drivers/net/ethernet/engleder/tsnep_main.c
1659
if (rx->read != rx->write) {
drivers/net/ethernet/engleder/tsnep_main.c
1660
entry = &rx->entry[rx->read];
drivers/net/ethernet/engleder/tsnep_main.c
1670
static int tsnep_rx_open(struct tsnep_rx *rx)
drivers/net/ethernet/engleder/tsnep_main.c
1675
retval = tsnep_rx_ring_create(rx);
drivers/net/ethernet/engleder/tsnep_main.c
1679
tsnep_rx_init(rx);
drivers/net/ethernet/engleder/tsnep_main.c
1681
desc_available = tsnep_rx_desc_available(rx);
drivers/net/ethernet/engleder/tsnep_main.c
1682
if (rx->xsk_pool)
drivers/net/ethernet/engleder/tsnep_main.c
1683
retval = tsnep_rx_alloc_zc(rx, desc_available, false);
drivers/net/ethernet/engleder/tsnep_main.c
1685
retval = tsnep_rx_alloc(rx, desc_available, false);
drivers/net/ethernet/engleder/tsnep_main.c
1695
if (rx->xsk_pool) {
drivers/net/ethernet/engleder/tsnep_main.c
1696
retval = tsnep_rx_alloc_page_buffer(rx);
drivers/net/ethernet/engleder/tsnep_main.c
1704
tsnep_rx_ring_cleanup(rx);
drivers/net/ethernet/engleder/tsnep_main.c
1708
static void tsnep_rx_close(struct tsnep_rx *rx)
drivers/net/ethernet/engleder/tsnep_main.c
1710
if (rx->xsk_pool)
drivers/net/ethernet/engleder/tsnep_main.c
1711
tsnep_rx_free_page_buffer(rx);
drivers/net/ethernet/engleder/tsnep_main.c
1713
tsnep_rx_ring_cleanup(rx);
drivers/net/ethernet/engleder/tsnep_main.c
1716
static void tsnep_rx_reopen(struct tsnep_rx *rx)
drivers/net/ethernet/engleder/tsnep_main.c
1718
struct page **page = rx->page_buffer;
drivers/net/ethernet/engleder/tsnep_main.c
1721
tsnep_rx_init(rx);
drivers/net/ethernet/engleder/tsnep_main.c
1724
struct tsnep_rx_entry *entry = &rx->entry[i];
drivers/net/ethernet/engleder/tsnep_main.c
1734
tsnep_rx_set_page(rx, entry, *page);
drivers/net/ethernet/engleder/tsnep_main.c
1735
tsnep_rx_activate(rx, rx->write);
drivers/net/ethernet/engleder/tsnep_main.c
1736
rx->write++;
drivers/net/ethernet/engleder/tsnep_main.c
1744
static void tsnep_rx_reopen_xsk(struct tsnep_rx *rx)
drivers/net/ethernet/engleder/tsnep_main.c
1746
struct page **page = rx->page_buffer;
drivers/net/ethernet/engleder/tsnep_main.c
1750
tsnep_rx_init(rx);
drivers/net/ethernet/engleder/tsnep_main.c
1756
allocated = xsk_buff_alloc_batch(rx->xsk_pool, rx->xdp_batch,
drivers/net/ethernet/engleder/tsnep_main.c
1760
struct tsnep_rx_entry *entry = &rx->entry[i];
drivers/net/ethernet/engleder/tsnep_main.c
1779
tsnep_rx_set_xdp(rx, entry,
drivers/net/ethernet/engleder/tsnep_main.c
1780
rx->xdp_batch[allocated - 1]);
drivers/net/ethernet/engleder/tsnep_main.c
1781
tsnep_rx_activate(rx, rx->write);
drivers/net/ethernet/engleder/tsnep_main.c
1782
rx->write++;
drivers/net/ethernet/engleder/tsnep_main.c
1792
if (xsk_uses_need_wakeup(rx->xsk_pool))
drivers/net/ethernet/engleder/tsnep_main.c
1793
tsnep_xsk_rx_need_wakeup(rx, tsnep_rx_desc_available(rx));
drivers/net/ethernet/engleder/tsnep_main.c
1801
if (queue->rx && tsnep_rx_pending(queue->rx))
drivers/net/ethernet/engleder/tsnep_main.c
1821
if (queue->rx) {
drivers/net/ethernet/engleder/tsnep_main.c
1822
done = queue->rx->xsk_pool ?
drivers/net/ethernet/engleder/tsnep_main.c
1823
tsnep_rx_poll_zc(queue->rx, napi, budget) :
drivers/net/ethernet/engleder/tsnep_main.c
1824
tsnep_rx_poll(queue->rx, napi, budget);
drivers/net/ethernet/engleder/tsnep_main.c
1861
if (queue->tx && queue->rx)
drivers/net/ethernet/engleder/tsnep_main.c
1863
name, queue->rx->queue_index);
drivers/net/ethernet/engleder/tsnep_main.c
1869
name, queue->rx->queue_index);
drivers/net/ethernet/engleder/tsnep_main.c
1901
struct tsnep_rx *rx = queue->rx;
drivers/net/ethernet/engleder/tsnep_main.c
1905
if (rx) {
drivers/net/ethernet/engleder/tsnep_main.c
1906
if (xdp_rxq_info_is_reg(&rx->xdp_rxq))
drivers/net/ethernet/engleder/tsnep_main.c
1907
xdp_rxq_info_unreg(&rx->xdp_rxq);
drivers/net/ethernet/engleder/tsnep_main.c
1908
if (xdp_rxq_info_is_reg(&rx->xdp_rxq_zc))
drivers/net/ethernet/engleder/tsnep_main.c
1909
xdp_rxq_info_unreg(&rx->xdp_rxq_zc);
drivers/net/ethernet/engleder/tsnep_main.c
1918
struct tsnep_rx *rx = queue->rx;
drivers/net/ethernet/engleder/tsnep_main.c
1924
if (rx) {
drivers/net/ethernet/engleder/tsnep_main.c
1927
rx->tx_queue_index = tx->queue_index;
drivers/net/ethernet/engleder/tsnep_main.c
1928
else if (rx->queue_index < adapter->num_tx_queues)
drivers/net/ethernet/engleder/tsnep_main.c
1929
rx->tx_queue_index = rx->queue_index;
drivers/net/ethernet/engleder/tsnep_main.c
1931
rx->tx_queue_index = 0;
drivers/net/ethernet/engleder/tsnep_main.c
1937
retval = xdp_rxq_info_reg(&rx->xdp_rxq, adapter->netdev,
drivers/net/ethernet/engleder/tsnep_main.c
1938
rx->queue_index, queue->napi.napi_id);
drivers/net/ethernet/engleder/tsnep_main.c
1941
retval = xdp_rxq_info_reg_mem_model(&rx->xdp_rxq,
drivers/net/ethernet/engleder/tsnep_main.c
1943
rx->page_pool);
drivers/net/ethernet/engleder/tsnep_main.c
1946
retval = xdp_rxq_info_reg(&rx->xdp_rxq_zc, adapter->netdev,
drivers/net/ethernet/engleder/tsnep_main.c
1947
rx->queue_index, queue->napi.napi_id);
drivers/net/ethernet/engleder/tsnep_main.c
1950
retval = xdp_rxq_info_reg_mem_model(&rx->xdp_rxq_zc,
drivers/net/ethernet/engleder/tsnep_main.c
1955
if (rx->xsk_pool)
drivers/net/ethernet/engleder/tsnep_main.c
1956
xsk_pool_set_rxq_info(rx->xsk_pool, &rx->xdp_rxq_zc);
drivers/net/ethernet/engleder/tsnep_main.c
1988
if (queue->rx) {
drivers/net/ethernet/engleder/tsnep_main.c
1989
netif_queue_set_napi(adapter->netdev, queue->rx->queue_index,
drivers/net/ethernet/engleder/tsnep_main.c
1991
tsnep_rx_enable(queue->rx);
drivers/net/ethernet/engleder/tsnep_main.c
1999
if (queue->rx)
drivers/net/ethernet/engleder/tsnep_main.c
2000
netif_queue_set_napi(adapter->netdev, queue->rx->queue_index,
drivers/net/ethernet/engleder/tsnep_main.c
2015
if (queue->rx)
drivers/net/ethernet/engleder/tsnep_main.c
2016
tsnep_rx_disable(queue->rx);
drivers/net/ethernet/engleder/tsnep_main.c
2030
if (adapter->queue[i].rx) {
drivers/net/ethernet/engleder/tsnep_main.c
2031
retval = tsnep_rx_open(adapter->queue[i].rx);
drivers/net/ethernet/engleder/tsnep_main.c
2066
if (adapter->queue[i].rx)
drivers/net/ethernet/engleder/tsnep_main.c
2067
tsnep_rx_close(adapter->queue[i].rx);
drivers/net/ethernet/engleder/tsnep_main.c
2087
if (adapter->queue[i].rx)
drivers/net/ethernet/engleder/tsnep_main.c
2088
tsnep_rx_close(adapter->queue[i].rx);
drivers/net/ethernet/engleder/tsnep_main.c
2105
queue->rx->page_buffer = kzalloc_objs(*queue->rx->page_buffer,
drivers/net/ethernet/engleder/tsnep_main.c
2107
if (!queue->rx->page_buffer)
drivers/net/ethernet/engleder/tsnep_main.c
2109
queue->rx->xdp_batch = kzalloc_objs(*queue->rx->xdp_batch,
drivers/net/ethernet/engleder/tsnep_main.c
2111
if (!queue->rx->xdp_batch) {
drivers/net/ethernet/engleder/tsnep_main.c
2112
kfree(queue->rx->page_buffer);
drivers/net/ethernet/engleder/tsnep_main.c
2113
queue->rx->page_buffer = NULL;
drivers/net/ethernet/engleder/tsnep_main.c
2118
xsk_pool_set_rxq_info(pool, &queue->rx->xdp_rxq_zc);
drivers/net/ethernet/engleder/tsnep_main.c
2124
queue->rx->xsk_pool = pool;
drivers/net/ethernet/engleder/tsnep_main.c
2127
tsnep_rx_reopen_xsk(queue->rx);
drivers/net/ethernet/engleder/tsnep_main.c
2141
tsnep_rx_free_zc(queue->rx);
drivers/net/ethernet/engleder/tsnep_main.c
2143
queue->rx->xsk_pool = NULL;
drivers/net/ethernet/engleder/tsnep_main.c
2147
tsnep_rx_reopen(queue->rx);
drivers/net/ethernet/engleder/tsnep_main.c
2151
kfree(queue->rx->xdp_batch);
drivers/net/ethernet/engleder/tsnep_main.c
2152
queue->rx->xdp_batch = NULL;
drivers/net/ethernet/engleder/tsnep_main.c
2153
kfree(queue->rx->page_buffer);
drivers/net/ethernet/engleder/tsnep_main.c
2154
queue->rx->page_buffer = NULL;
drivers/net/ethernet/engleder/tsnep_main.c
2199
stats->rx_packets += adapter->rx[i].packets;
drivers/net/ethernet/engleder/tsnep_main.c
2200
stats->rx_bytes += adapter->rx[i].bytes;
drivers/net/ethernet/engleder/tsnep_main.c
2201
stats->rx_dropped += adapter->rx[i].dropped;
drivers/net/ethernet/engleder/tsnep_main.c
2202
stats->multicast += adapter->rx[i].multicast;
drivers/net/ethernet/engleder/tsnep_main.c
2508
adapter->queue[0].rx = &adapter->rx[0];
drivers/net/ethernet/engleder/tsnep_main.c
2509
adapter->queue[0].rx->adapter = adapter;
drivers/net/ethernet/engleder/tsnep_main.c
2510
adapter->queue[0].rx->addr = adapter->addr + TSNEP_QUEUE(0);
drivers/net/ethernet/engleder/tsnep_main.c
2511
adapter->queue[0].rx->queue_index = 0;
drivers/net/ethernet/engleder/tsnep_main.c
2539
adapter->queue[i].rx = &adapter->rx[i];
drivers/net/ethernet/engleder/tsnep_main.c
2540
adapter->queue[i].rx->adapter = adapter;
drivers/net/ethernet/engleder/tsnep_main.c
2541
adapter->queue[i].rx->addr = adapter->addr + TSNEP_QUEUE(i);
drivers/net/ethernet/engleder/tsnep_main.c
2542
adapter->queue[i].rx->queue_index = i;
drivers/net/ethernet/engleder/tsnep_main.c
951
static void tsnep_rx_ring_cleanup(struct tsnep_rx *rx)
drivers/net/ethernet/engleder/tsnep_main.c
953
struct device *dmadev = rx->adapter->dmadev;
drivers/net/ethernet/engleder/tsnep_main.c
958
entry = &rx->entry[i];
drivers/net/ethernet/engleder/tsnep_main.c
959
if (!rx->xsk_pool && entry->page)
drivers/net/ethernet/engleder/tsnep_main.c
960
page_pool_put_full_page(rx->page_pool, entry->page,
drivers/net/ethernet/engleder/tsnep_main.c
962
if (rx->xsk_pool && entry->xdp)
drivers/net/ethernet/engleder/tsnep_main.c
968
if (rx->page_pool)
drivers/net/ethernet/engleder/tsnep_main.c
969
page_pool_destroy(rx->page_pool);
drivers/net/ethernet/engleder/tsnep_main.c
971
memset(rx->entry, 0, sizeof(rx->entry));
drivers/net/ethernet/engleder/tsnep_main.c
974
if (rx->page[i]) {
drivers/net/ethernet/engleder/tsnep_main.c
975
dma_free_coherent(dmadev, PAGE_SIZE, rx->page[i],
drivers/net/ethernet/engleder/tsnep_main.c
976
rx->page_dma[i]);
drivers/net/ethernet/engleder/tsnep_main.c
977
rx->page[i] = NULL;
drivers/net/ethernet/engleder/tsnep_main.c
978
rx->page_dma[i] = 0;
drivers/net/ethernet/engleder/tsnep_main.c
983
static int tsnep_rx_ring_create(struct tsnep_rx *rx)
drivers/net/ethernet/engleder/tsnep_main.c
985
struct device *dmadev = rx->adapter->dmadev;
drivers/net/ethernet/engleder/tsnep_main.c
993
rx->page[i] =
drivers/net/ethernet/engleder/tsnep_main.c
994
dma_alloc_coherent(dmadev, PAGE_SIZE, &rx->page_dma[i],
drivers/net/ethernet/engleder/tsnep_main.c
996
if (!rx->page[i]) {
drivers/net/ethernet/engleder/tsnep_xdp.c
32
if (queue->rx->queue_index != queue_id ||
drivers/net/ethernet/faraday/ftmac100.c
914
int rx = 0;
drivers/net/ethernet/faraday/ftmac100.c
929
retry = ftmac100_rx_packet(priv, &rx);
drivers/net/ethernet/faraday/ftmac100.c
930
} while (retry && rx < budget);
drivers/net/ethernet/faraday/ftmac100.c
932
if (retry && rx == budget)
drivers/net/ethernet/faraday/ftmac100.c
979
return rx;
drivers/net/ethernet/freescale/fman/fman_port.c
1399
iowrite32be(NIA_ENG_HWK, &port->bmi_regs->rx.fmbm_rfpne);
drivers/net/ethernet/freescale/fman/fman_port.c
1403
&port->bmi_regs->rx.fmbm_rfpne);
drivers/net/ethernet/freescale/fman/fman_port.c
1574
bmi_cfg_reg = &port->bmi_regs->rx.fmbm_rcfg;
drivers/net/ethernet/freescale/fman/fman_port.c
1575
bmi_status_reg = &port->bmi_regs->rx.fmbm_rst;
drivers/net/ethernet/freescale/fman/fman_port.c
1650
bmi_cfg_reg = &port->bmi_regs->rx.fmbm_rcfg;
drivers/net/ethernet/freescale/fman/fman_port.c
277
struct fman_port_rx_bmi_regs rx;
drivers/net/ethernet/freescale/fman/fman_port.c
459
struct fman_port_rx_bmi_regs __iomem *regs = &port->bmi_regs->rx;
drivers/net/ethernet/freescale/fman/fman_port.c
756
bp_reg = port->bmi_regs->rx.fmbm_ebmpi;
drivers/net/ethernet/freescale/fman/fman_port.c
757
bp_depl_reg = &port->bmi_regs->rx.fmbm_mpd;
drivers/net/ethernet/google/gve/gve.h
1253
void gve_rx_write_doorbell(struct gve_priv *priv, struct gve_rx_ring *rx);
drivers/net/ethernet/google/gve/gve.h
1255
bool gve_rx_work_pending(struct gve_rx_ring *rx);
drivers/net/ethernet/google/gve/gve.h
1258
struct gve_rx_ring *rx,
drivers/net/ethernet/google/gve/gve.h
1260
void gve_rx_free_ring_gqi(struct gve_priv *priv, struct gve_rx_ring *rx,
drivers/net/ethernet/google/gve/gve.h
1278
struct gve_rx_buf_state_dqo *gve_alloc_buf_state(struct gve_rx_ring *rx);
drivers/net/ethernet/google/gve/gve.h
1279
bool gve_buf_state_is_allocated(struct gve_rx_ring *rx,
drivers/net/ethernet/google/gve/gve.h
1281
void gve_free_buf_state(struct gve_rx_ring *rx,
drivers/net/ethernet/google/gve/gve.h
1283
struct gve_rx_buf_state_dqo *gve_dequeue_buf_state(struct gve_rx_ring *rx,
drivers/net/ethernet/google/gve/gve.h
1285
void gve_enqueue_buf_state(struct gve_rx_ring *rx, struct gve_index_list *list,
drivers/net/ethernet/google/gve/gve.h
1287
struct gve_rx_buf_state_dqo *gve_get_recycled_buf_state(struct gve_rx_ring *rx);
drivers/net/ethernet/google/gve/gve.h
1288
void gve_try_recycle_buf(struct gve_priv *priv, struct gve_rx_ring *rx,
drivers/net/ethernet/google/gve/gve.h
1290
void gve_free_to_page_pool(struct gve_rx_ring *rx,
drivers/net/ethernet/google/gve/gve.h
1293
int gve_alloc_qpl_page_dqo(struct gve_rx_ring *rx,
drivers/net/ethernet/google/gve/gve.h
1296
void gve_reuse_buffer(struct gve_rx_ring *rx,
drivers/net/ethernet/google/gve/gve.h
1298
void gve_free_buffer(struct gve_rx_ring *rx,
drivers/net/ethernet/google/gve/gve.h
1300
int gve_alloc_buffer(struct gve_rx_ring *rx, struct gve_rx_desc_dqo *desc);
drivers/net/ethernet/google/gve/gve.h
1302
struct gve_rx_ring *rx,
drivers/net/ethernet/google/gve/gve.h
674
struct gve_rx_ring *rx; /* rx rings on this block */
drivers/net/ethernet/google/gve/gve.h
738
struct gve_rx_ring *rx;
drivers/net/ethernet/google/gve/gve.h
800
struct gve_rx_ring *rx; /* array of rx_cfg.num_queues */
drivers/net/ethernet/google/gve/gve_adminq.c
757
struct gve_rx_ring *rx = &priv->rx[queue_index];
drivers/net/ethernet/google/gve/gve_adminq.c
763
.ntfy_id = cpu_to_be32(rx->ntfy_id),
drivers/net/ethernet/google/gve/gve_adminq.c
764
.queue_resources_addr = cpu_to_be64(rx->q_resources_bus),
drivers/net/ethernet/google/gve/gve_adminq.c
766
.packet_buffer_size = cpu_to_be16(rx->packet_buffer_size),
drivers/net/ethernet/google/gve/gve_adminq.c
771
GVE_RAW_ADDRESSING_QPL_ID : rx->data.qpl->id;
drivers/net/ethernet/google/gve/gve_adminq.c
774
cpu_to_be64(rx->desc.bus);
drivers/net/ethernet/google/gve/gve_adminq.c
776
cpu_to_be64(rx->data.data_bus);
drivers/net/ethernet/google/gve/gve_adminq.c
785
qpl_id = rx->dqo.qpl->id;
drivers/net/ethernet/google/gve/gve_adminq.c
788
cpu_to_be64(rx->dqo.complq.bus);
drivers/net/ethernet/google/gve/gve_adminq.c
790
cpu_to_be64(rx->dqo.bufq.bus);
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
104
buf_state = gve_dequeue_buf_state(rx, &rx->dqo.recycled_buf_states);
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
108
if (unlikely(rx->dqo.used_buf_states.head == -1))
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
117
buf_state = gve_dequeue_buf_state(rx, &rx->dqo.used_buf_states);
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
119
rx->dqo.used_buf_states_cnt--;
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
123
gve_enqueue_buf_state(rx, &rx->dqo.used_buf_states, buf_state);
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
129
int gve_alloc_qpl_page_dqo(struct gve_rx_ring *rx,
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
132
struct gve_priv *priv = rx->gve;
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
135
idx = rx->dqo.next_qpl_page_idx;
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
141
buf_state->page_info.page = rx->dqo.qpl->pages[idx];
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
142
buf_state->addr = rx->dqo.qpl->page_buses[idx];
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
143
rx->dqo.next_qpl_page_idx++;
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
147
buf_state->page_info.buf_size = rx->packet_buffer_truesize;
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
148
buf_state->page_info.pad = rx->rx_headroom;
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
16
struct gve_rx_buf_state_dqo *gve_alloc_buf_state(struct gve_rx_ring *rx)
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
168
void gve_try_recycle_buf(struct gve_priv *priv, struct gve_rx_ring *rx,
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
171
const u16 data_buffer_size = rx->packet_buffer_truesize;
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
202
gve_enqueue_buf_state(rx, &rx->dqo.recycled_buf_states, buf_state);
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
206
gve_enqueue_buf_state(rx, &rx->dqo.used_buf_states, buf_state);
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
207
rx->dqo.used_buf_states_cnt++;
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
21
buffer_id = rx->dqo.free_buf_states;
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
210
void gve_free_to_page_pool(struct gve_rx_ring *rx,
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
223
static int gve_alloc_from_page_pool(struct gve_rx_ring *rx,
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
228
buf_state->page_info.buf_size = rx->packet_buffer_truesize;
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
229
netmem = page_pool_alloc_netmem(rx->dqo.page_pool,
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
240
buf_state->page_info.pad = rx->dqo.page_pool->p.offset;
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
246
struct gve_rx_ring *rx,
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
249
u32 ntfy_id = gve_rx_idx_to_ntfy(priv, rx->q_num);
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
25
buf_state = &rx->dqo.buf_states[buffer_id];
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
265
pp.queue_idx = rx->q_num;
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
271
void gve_free_buffer(struct gve_rx_ring *rx,
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
274
if (rx->dqo.page_pool) {
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
275
gve_free_to_page_pool(rx, buf_state, true);
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
276
gve_free_buf_state(rx, buf_state);
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
278
gve_enqueue_buf_state(rx, &rx->dqo.recycled_buf_states,
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
28
rx->dqo.free_buf_states = buf_state->next;
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
283
void gve_reuse_buffer(struct gve_rx_ring *rx,
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
286
if (rx->dqo.page_pool) {
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
288
gve_free_buf_state(rx, buf_state);
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
291
gve_try_recycle_buf(rx->gve, rx, buf_state);
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
295
int gve_alloc_buffer(struct gve_rx_ring *rx, struct gve_rx_desc_dqo *desc)
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
299
if (rx->xsk_pool) {
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
300
buf_state = gve_alloc_buf_state(rx);
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
304
buf_state->xsk_buff = xsk_buff_alloc(rx->xsk_pool);
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
306
xsk_set_rx_need_wakeup(rx->xsk_pool);
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
307
gve_free_buf_state(rx, buf_state);
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
311
xsk_clear_rx_need_wakeup(rx->xsk_pool);
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
312
desc->buf_id = cpu_to_le16(buf_state - rx->dqo.buf_states);
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
316
} else if (rx->dqo.page_pool) {
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
317
buf_state = gve_alloc_buf_state(rx);
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
321
if (gve_alloc_from_page_pool(rx, buf_state))
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
324
buf_state = gve_get_recycled_buf_state(rx);
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
326
buf_state = gve_alloc_buf_state(rx);
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
330
if (unlikely(gve_alloc_qpl_page_dqo(rx, buf_state)))
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
334
desc->buf_id = cpu_to_le16(buf_state - rx->dqo.buf_states);
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
342
gve_free_buf_state(rx, buf_state);
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
40
bool gve_buf_state_is_allocated(struct gve_rx_ring *rx,
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
43
s16 buffer_id = buf_state - rx->dqo.buf_states;
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
48
void gve_free_buf_state(struct gve_rx_ring *rx,
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
51
s16 buffer_id = buf_state - rx->dqo.buf_states;
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
53
buf_state->next = rx->dqo.free_buf_states;
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
54
rx->dqo.free_buf_states = buffer_id;
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
57
struct gve_rx_buf_state_dqo *gve_dequeue_buf_state(struct gve_rx_ring *rx,
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
67
buf_state = &rx->dqo.buf_states[buffer_id];
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
80
void gve_enqueue_buf_state(struct gve_rx_ring *rx, struct gve_index_list *list,
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
83
s16 buffer_id = buf_state - rx->dqo.buf_states;
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
93
rx->dqo.buf_states[tail].next = buffer_id;
drivers/net/ethernet/google/gve/gve_buffer_mgmt_dqo.c
98
struct gve_rx_buf_state_dqo *gve_get_recycled_buf_state(struct gve_rx_ring *rx)
drivers/net/ethernet/google/gve/gve_dqo.h
52
struct gve_rx_ring *rx,
drivers/net/ethernet/google/gve/gve_dqo.h
54
void gve_rx_free_ring_dqo(struct gve_priv *priv, struct gve_rx_ring *rx,
drivers/net/ethernet/google/gve/gve_dqo.h
64
void gve_rx_post_buffers_dqo(struct gve_rx_ring *rx);
drivers/net/ethernet/google/gve/gve_ethtool.c
204
if (priv->rx) {
drivers/net/ethernet/google/gve/gve_ethtool.c
206
struct gve_rx_ring *rx = &priv->rx[ring];
drivers/net/ethernet/google/gve/gve_ethtool.c
209
u64_stats_fetch_begin(&priv->rx[ring].statss);
drivers/net/ethernet/google/gve/gve_ethtool.c
210
tmp_rx_pkts = rx->rpackets;
drivers/net/ethernet/google/gve/gve_ethtool.c
211
tmp_rx_hsplit_pkt = rx->rx_hsplit_pkt;
drivers/net/ethernet/google/gve/gve_ethtool.c
212
tmp_rx_bytes = rx->rbytes;
drivers/net/ethernet/google/gve/gve_ethtool.c
213
tmp_rx_skb_alloc_fail = rx->rx_skb_alloc_fail;
drivers/net/ethernet/google/gve/gve_ethtool.c
214
tmp_rx_buf_alloc_fail = rx->rx_buf_alloc_fail;
drivers/net/ethernet/google/gve/gve_ethtool.c
216
rx->rx_desc_err_dropped_pkt;
drivers/net/ethernet/google/gve/gve_ethtool.c
218
rx->rx_hsplit_unsplit_pkt;
drivers/net/ethernet/google/gve/gve_ethtool.c
219
tmp_xdp_tx_errors = rx->xdp_tx_errors;
drivers/net/ethernet/google/gve/gve_ethtool.c
221
rx->xdp_redirect_errors;
drivers/net/ethernet/google/gve/gve_ethtool.c
222
} while (u64_stats_fetch_retry(&priv->rx[ring].statss,
drivers/net/ethernet/google/gve/gve_ethtool.c
322
if (priv->rx) {
drivers/net/ethernet/google/gve/gve_ethtool.c
324
struct gve_rx_ring *rx = &priv->rx[ring];
drivers/net/ethernet/google/gve/gve_ethtool.c
326
data[i++] = rx->fill_cnt;
drivers/net/ethernet/google/gve/gve_ethtool.c
327
data[i++] = rx->cnt;
drivers/net/ethernet/google/gve/gve_ethtool.c
328
data[i++] = rx->fill_cnt - rx->cnt;
drivers/net/ethernet/google/gve/gve_ethtool.c
331
u64_stats_fetch_begin(&priv->rx[ring].statss);
drivers/net/ethernet/google/gve/gve_ethtool.c
332
tmp_rx_bytes = rx->rbytes;
drivers/net/ethernet/google/gve/gve_ethtool.c
333
tmp_rx_hsplit_bytes = rx->rx_hsplit_bytes;
drivers/net/ethernet/google/gve/gve_ethtool.c
334
tmp_rx_skb_alloc_fail = rx->rx_skb_alloc_fail;
drivers/net/ethernet/google/gve/gve_ethtool.c
335
tmp_rx_buf_alloc_fail = rx->rx_buf_alloc_fail;
drivers/net/ethernet/google/gve/gve_ethtool.c
337
rx->rx_desc_err_dropped_pkt;
drivers/net/ethernet/google/gve/gve_ethtool.c
338
tmp_xdp_tx_errors = rx->xdp_tx_errors;
drivers/net/ethernet/google/gve/gve_ethtool.c
340
rx->xdp_redirect_errors;
drivers/net/ethernet/google/gve/gve_ethtool.c
341
} while (u64_stats_fetch_retry(&priv->rx[ring].statss,
drivers/net/ethernet/google/gve/gve_ethtool.c
345
data[i++] = rx->rx_cont_packet_cnt;
drivers/net/ethernet/google/gve/gve_ethtool.c
346
data[i++] = rx->rx_frag_flip_cnt;
drivers/net/ethernet/google/gve/gve_ethtool.c
347
data[i++] = rx->rx_frag_copy_cnt;
drivers/net/ethernet/google/gve/gve_ethtool.c
348
data[i++] = rx->rx_frag_alloc_cnt;
drivers/net/ethernet/google/gve/gve_ethtool.c
354
data[i++] = rx->rx_copybreak_pkt;
drivers/net/ethernet/google/gve/gve_ethtool.c
355
data[i++] = rx->rx_copied_pkt;
drivers/net/ethernet/google/gve/gve_ethtool.c
371
start = u64_stats_fetch_begin(&priv->rx[ring].statss);
drivers/net/ethernet/google/gve/gve_ethtool.c
373
data[i + j] = rx->xdp_actions[j];
drivers/net/ethernet/google/gve/gve_ethtool.c
374
data[i + j++] = rx->xdp_tx_errors;
drivers/net/ethernet/google/gve/gve_ethtool.c
375
data[i + j++] = rx->xdp_redirect_errors;
drivers/net/ethernet/google/gve/gve_ethtool.c
376
data[i + j++] = rx->xdp_alloc_fails;
drivers/net/ethernet/google/gve/gve_ethtool.c
377
} while (u64_stats_fetch_retry(&priv->rx[ring].statss,
drivers/net/ethernet/google/gve/gve_main.c
117
if (priv->rx) {
drivers/net/ethernet/google/gve/gve_main.c
1174
struct gve_rx_ring *rx;
drivers/net/ethernet/google/gve/gve_main.c
1176
if (!priv->rx)
drivers/net/ethernet/google/gve/gve_main.c
1179
rx = &priv->rx[qid];
drivers/net/ethernet/google/gve/gve_main.c
1180
rx->xsk_pool = NULL;
drivers/net/ethernet/google/gve/gve_main.c
1181
if (xdp_rxq_info_is_reg(&rx->xdp_rxq))
drivers/net/ethernet/google/gve/gve_main.c
1182
xdp_rxq_info_unreg_mem_model(&rx->xdp_rxq);
drivers/net/ethernet/google/gve/gve_main.c
1192
struct gve_rx_ring *rx;
drivers/net/ethernet/google/gve/gve_main.c
1196
rx = &priv->rx[qid];
drivers/net/ethernet/google/gve/gve_main.c
1197
err = xdp_rxq_info_reg_mem_model(&rx->xdp_rxq,
drivers/net/ethernet/google/gve/gve_main.c
1204
rx->xsk_pool = pool;
drivers/net/ethernet/google/gve/gve_main.c
121
u64_stats_fetch_begin(&priv->rx[ring].statss);
drivers/net/ethernet/google/gve/gve_main.c
1216
if (!priv->tx_cfg.num_xdp_queues || !priv->rx)
drivers/net/ethernet/google/gve/gve_main.c
122
packets = priv->rx[ring].rpackets;
drivers/net/ethernet/google/gve/gve_main.c
1220
struct gve_rx_ring *rx = &priv->rx[i];
drivers/net/ethernet/google/gve/gve_main.c
1222
if (xdp_rxq_info_is_reg(&rx->xdp_rxq))
drivers/net/ethernet/google/gve/gve_main.c
1223
xdp_rxq_info_unreg(&rx->xdp_rxq);
drivers/net/ethernet/google/gve/gve_main.c
123
bytes = priv->rx[ring].rbytes;
drivers/net/ethernet/google/gve/gve_main.c
124
} while (u64_stats_fetch_retry(&priv->rx[ring].statss,
drivers/net/ethernet/google/gve/gve_main.c
1240
struct gve_rx_ring *rx;
drivers/net/ethernet/google/gve/gve_main.c
1250
rx = &priv->rx[i];
drivers/net/ethernet/google/gve/gve_main.c
1251
napi = &priv->ntfy_blocks[rx->ntfy_id].napi;
drivers/net/ethernet/google/gve/gve_main.c
1253
err = xdp_rxq_info_reg(&rx->xdp_rxq, dev, i,
drivers/net/ethernet/google/gve/gve_main.c
1262
err = xdp_rxq_info_reg_mem_model(&rx->xdp_rxq,
drivers/net/ethernet/google/gve/gve_main.c
1266
err = xdp_rxq_info_reg_mem_model(&rx->xdp_rxq,
drivers/net/ethernet/google/gve/gve_main.c
1268
rx->dqo.page_pool);
drivers/net/ethernet/google/gve/gve_main.c
1285
page_frag_cache_drain(&priv->rx[i].page_cache);
drivers/net/ethernet/google/gve/gve_main.c
1297
cfg->rx = priv->rx;
drivers/net/ethernet/google/gve/gve_main.c
1337
if (!priv->rx)
drivers/net/ethernet/google/gve/gve_main.c
1352
priv->rx = NULL;
drivers/net/ethernet/google/gve/gve_main.c
1367
priv->rx = rx_alloc_cfg->rx;
drivers/net/ethernet/google/gve/gve_main.c
1669
napi_rx = &priv->ntfy_blocks[priv->rx[qid].ntfy_id].napi;
drivers/net/ethernet/google/gve/gve_main.c
1682
if (gve_rx_work_pending(&priv->rx[qid]))
drivers/net/ethernet/google/gve/gve_main.c
2307
if (priv->rx) {
drivers/net/ethernet/google/gve/gve_main.c
2311
.value = cpu_to_be64(priv->rx[idx].desc.seqno),
drivers/net/ethernet/google/gve/gve_main.c
2316
.value = cpu_to_be64(priv->rx[idx].fill_cnt),
drivers/net/ethernet/google/gve/gve_main.c
2572
if (!priv->rx)
drivers/net/ethernet/google/gve/gve_main.c
2600
*gve_per_q_mem = priv->rx[idx];
drivers/net/ethernet/google/gve/gve_main.c
2601
memset(&priv->rx[idx], 0, sizeof(priv->rx[idx]));
drivers/net/ethernet/google/gve/gve_main.c
2629
if (!priv->rx)
drivers/net/ethernet/google/gve/gve_main.c
2651
if (!priv->rx)
drivers/net/ethernet/google/gve/gve_main.c
2655
priv->rx[idx] = *gve_per_q_mem;
drivers/net/ethernet/google/gve/gve_main.c
2675
gve_rx_write_doorbell(priv, &priv->rx[idx]);
drivers/net/ethernet/google/gve/gve_main.c
2677
gve_rx_post_buffers_dqo(&priv->rx[idx]);
drivers/net/ethernet/google/gve/gve_main.c
2691
memset(&priv->rx[idx], 0, sizeof(priv->rx[idx]));
drivers/net/ethernet/google/gve/gve_main.c
2707
struct gve_rx_ring *rx = &priv->rx[idx];
drivers/net/ethernet/google/gve/gve_main.c
2711
start = u64_stats_fetch_begin(&rx->statss);
drivers/net/ethernet/google/gve/gve_main.c
2712
rx_stats->packets = rx->rpackets;
drivers/net/ethernet/google/gve/gve_main.c
2713
rx_stats->bytes = rx->rbytes;
drivers/net/ethernet/google/gve/gve_main.c
2714
rx_stats->alloc_fail = rx->rx_skb_alloc_fail +
drivers/net/ethernet/google/gve/gve_main.c
2715
rx->rx_buf_alloc_fail;
drivers/net/ethernet/google/gve/gve_main.c
2716
} while (u64_stats_fetch_retry(&rx->statss, start));
drivers/net/ethernet/google/gve/gve_main.c
2734
struct netdev_queue_stats_rx *rx,
drivers/net/ethernet/google/gve/gve_main.c
2737
rx->packets = 0;
drivers/net/ethernet/google/gve/gve_main.c
2738
rx->bytes = 0;
drivers/net/ethernet/google/gve/gve_main.c
2739
rx->alloc_fail = 0;
drivers/net/ethernet/google/gve/gve_main.c
373
if (block->rx) {
drivers/net/ethernet/google/gve/gve_main.c
401
if (block->rx)
drivers/net/ethernet/google/gve/gve_main.c
402
reschedule |= gve_rx_work_pending(block->rx);
drivers/net/ethernet/google/gve/gve_main.c
428
if (block->rx) {
drivers/net/ethernet/google/gve/gve_main.c
829
struct gve_rx_ring *rx = &priv->rx[idx];
drivers/net/ethernet/google/gve/gve_main.c
832
return rx->data.qpl;
drivers/net/ethernet/google/gve/gve_main.c
834
return rx->dqo.qpl;
drivers/net/ethernet/google/gve/gve_main.c
924
gve_rx_write_doorbell(priv, &priv->rx[i]);
drivers/net/ethernet/google/gve/gve_main.c
928
gve_rx_post_buffers_dqo(&priv->rx[i]);
drivers/net/ethernet/google/gve/gve_main.c
960
u64_stats_init(&priv->rx[i].statss);
drivers/net/ethernet/google/gve/gve_rx.c
1003
static int gve_clean_rx_done(struct gve_rx_ring *rx, int budget,
drivers/net/ethernet/google/gve/gve_rx.c
1006
u64 xdp_redirects = rx->xdp_actions[XDP_REDIRECT];
drivers/net/ethernet/google/gve/gve_rx.c
1007
u64 xdp_txs = rx->xdp_actions[XDP_TX];
drivers/net/ethernet/google/gve/gve_rx.c
1008
struct gve_rx_ctx *ctx = &rx->ctx;
drivers/net/ethernet/google/gve/gve_rx.c
1009
struct gve_priv *priv = rx->gve;
drivers/net/ethernet/google/gve/gve_rx.c
1012
u32 idx = rx->cnt & rx->mask;
drivers/net/ethernet/google/gve/gve_rx.c
1015
struct gve_rx_desc *desc = &rx->desc.desc_ring[idx];
drivers/net/ethernet/google/gve/gve_rx.c
1018
while ((GVE_SEQNO(desc->flags_seq) == rx->desc.seqno) &&
drivers/net/ethernet/google/gve/gve_rx.c
102
void gve_rx_free_ring_gqi(struct gve_priv *priv, struct gve_rx_ring *rx,
drivers/net/ethernet/google/gve/gve_rx.c
1020
next_desc = &rx->desc.desc_ring[(idx + 1) & rx->mask];
drivers/net/ethernet/google/gve/gve_rx.c
1023
gve_rx(rx, feat, desc, idx, &cnts);
drivers/net/ethernet/google/gve/gve_rx.c
1025
rx->cnt++;
drivers/net/ethernet/google/gve/gve_rx.c
1026
idx = rx->cnt & rx->mask;
drivers/net/ethernet/google/gve/gve_rx.c
1027
desc = &rx->desc.desc_ring[idx];
drivers/net/ethernet/google/gve/gve_rx.c
1028
rx->desc.seqno = gve_next_seqno(rx->desc.seqno);
drivers/net/ethernet/google/gve/gve_rx.c
1034
struct napi_struct *napi = &priv->ntfy_blocks[rx->ntfy_id].napi;
drivers/net/ethernet/google/gve/gve_rx.c
1037
gve_rx_ctx_clear(&rx->ctx);
drivers/net/ethernet/google/gve/gve_rx.c
1039
GVE_SEQNO(desc->flags_seq), rx->desc.seqno);
drivers/net/ethernet/google/gve/gve_rx.c
1040
gve_schedule_reset(rx->gve);
drivers/net/ethernet/google/gve/gve_rx.c
1043
if (!work_done && rx->fill_cnt - rx->cnt > rx->db_threshold)
drivers/net/ethernet/google/gve/gve_rx.c
1047
u64_stats_update_begin(&rx->statss);
drivers/net/ethernet/google/gve/gve_rx.c
1048
rx->rpackets += cnts.ok_pkt_cnt;
drivers/net/ethernet/google/gve/gve_rx.c
1049
rx->rbytes += cnts.ok_pkt_bytes;
drivers/net/ethernet/google/gve/gve_rx.c
1050
rx->rx_cont_packet_cnt += cnts.cont_pkt_cnt;
drivers/net/ethernet/google/gve/gve_rx.c
1051
rx->rx_desc_err_dropped_pkt += cnts.desc_err_pkt_cnt;
drivers/net/ethernet/google/gve/gve_rx.c
1052
u64_stats_update_end(&rx->statss);
drivers/net/ethernet/google/gve/gve_rx.c
1055
if (xdp_txs != rx->xdp_actions[XDP_TX])
drivers/net/ethernet/google/gve/gve_rx.c
1056
gve_xdp_tx_flush(priv, rx->q_num);
drivers/net/ethernet/google/gve/gve_rx.c
1058
if (xdp_redirects != rx->xdp_actions[XDP_REDIRECT])
drivers/net/ethernet/google/gve/gve_rx.c
106
u32 slots = rx->mask + 1;
drivers/net/ethernet/google/gve/gve_rx.c
1062
if (!rx->data.raw_addressing) {
drivers/net/ethernet/google/gve/gve_rx.c
1064
rx->fill_cnt += work_done;
drivers/net/ethernet/google/gve/gve_rx.c
1065
} else if (rx->fill_cnt - rx->cnt <= rx->db_threshold) {
drivers/net/ethernet/google/gve/gve_rx.c
1069
if (!gve_rx_refill_buffers(priv, rx))
drivers/net/ethernet/google/gve/gve_rx.c
107
int idx = rx->q_num;
drivers/net/ethernet/google/gve/gve_rx.c
1075
if (rx->fill_cnt - rx->cnt <= rx->db_threshold) {
drivers/net/ethernet/google/gve/gve_rx.c
1076
gve_rx_write_doorbell(priv, rx);
drivers/net/ethernet/google/gve/gve_rx.c
1081
gve_rx_write_doorbell(priv, rx);
drivers/net/ethernet/google/gve/gve_rx.c
1087
struct gve_rx_ring *rx = block->rx;
drivers/net/ethernet/google/gve/gve_rx.c
1094
work_done = gve_clean_rx_done(rx, budget, feat);
drivers/net/ethernet/google/gve/gve_rx.c
111
if (rx->desc.desc_ring) {
drivers/net/ethernet/google/gve/gve_rx.c
113
dma_free_coherent(dev, bytes, rx->desc.desc_ring, rx->desc.bus);
drivers/net/ethernet/google/gve/gve_rx.c
114
rx->desc.desc_ring = NULL;
drivers/net/ethernet/google/gve/gve_rx.c
117
if (rx->q_resources) {
drivers/net/ethernet/google/gve/gve_rx.c
118
dma_free_coherent(dev, sizeof(*rx->q_resources),
drivers/net/ethernet/google/gve/gve_rx.c
119
rx->q_resources, rx->q_resources_bus);
drivers/net/ethernet/google/gve/gve_rx.c
120
rx->q_resources = NULL;
drivers/net/ethernet/google/gve/gve_rx.c
123
gve_rx_unfill_pages(priv, rx, cfg);
drivers/net/ethernet/google/gve/gve_rx.c
125
if (rx->data.data_ring) {
drivers/net/ethernet/google/gve/gve_rx.c
126
bytes = sizeof(*rx->data.data_ring) * slots;
drivers/net/ethernet/google/gve/gve_rx.c
127
dma_free_coherent(dev, bytes, rx->data.data_ring,
drivers/net/ethernet/google/gve/gve_rx.c
128
rx->data.data_bus);
drivers/net/ethernet/google/gve/gve_rx.c
129
rx->data.data_ring = NULL;
drivers/net/ethernet/google/gve/gve_rx.c
132
kvfree(rx->qpl_copy_pool);
drivers/net/ethernet/google/gve/gve_rx.c
133
rx->qpl_copy_pool = NULL;
drivers/net/ethernet/google/gve/gve_rx.c
135
if (rx->data.qpl) {
drivers/net/ethernet/google/gve/gve_rx.c
137
gve_free_queue_page_list(priv, rx->data.qpl, qpl_id);
drivers/net/ethernet/google/gve/gve_rx.c
138
rx->data.qpl = NULL;
drivers/net/ethernet/google/gve/gve_rx.c
144
static void gve_setup_rx_buffer(struct gve_rx_ring *rx,
drivers/net/ethernet/google/gve/gve_rx.c
152
page_info->buf_size = rx->packet_buffer_size;
drivers/net/ethernet/google/gve/gve_rx.c
162
struct gve_rx_ring *rx)
drivers/net/ethernet/google/gve/gve_rx.c
171
u64_stats_update_begin(&rx->statss);
drivers/net/ethernet/google/gve/gve_rx.c
172
rx->rx_buf_alloc_fail++;
drivers/net/ethernet/google/gve/gve_rx.c
173
u64_stats_update_end(&rx->statss);
drivers/net/ethernet/google/gve/gve_rx.c
177
gve_setup_rx_buffer(rx, page_info, dma, page, &data_slot->addr);
drivers/net/ethernet/google/gve/gve_rx.c
181
static int gve_rx_prefill_pages(struct gve_rx_ring *rx,
drivers/net/ethernet/google/gve/gve_rx.c
184
struct gve_priv *priv = rx->gve;
drivers/net/ethernet/google/gve/gve_rx.c
193
slots = rx->mask + 1;
drivers/net/ethernet/google/gve/gve_rx.c
195
rx->data.page_info = kvcalloc_node(slots, sizeof(*rx->data.page_info),
drivers/net/ethernet/google/gve/gve_rx.c
197
if (!rx->data.page_info)
drivers/net/ethernet/google/gve/gve_rx.c
201
if (!rx->data.raw_addressing) {
drivers/net/ethernet/google/gve/gve_rx.c
202
struct page *page = rx->data.qpl->pages[i];
drivers/net/ethernet/google/gve/gve_rx.c
205
gve_setup_rx_buffer(rx, &rx->data.page_info[i], addr,
drivers/net/ethernet/google/gve/gve_rx.c
207
&rx->data.data_ring[i].qpl_offset);
drivers/net/ethernet/google/gve/gve_rx.c
211
&rx->data.page_info[i],
drivers/net/ethernet/google/gve/gve_rx.c
212
&rx->data.data_ring[i], rx);
drivers/net/ethernet/google/gve/gve_rx.c
217
if (!rx->data.raw_addressing) {
drivers/net/ethernet/google/gve/gve_rx.c
218
for (j = 0; j < rx->qpl_copy_pool_mask + 1; j++) {
drivers/net/ethernet/google/gve/gve_rx.c
227
rx->qpl_copy_pool[j].page = page;
drivers/net/ethernet/google/gve/gve_rx.c
228
rx->qpl_copy_pool[j].page_offset = 0;
drivers/net/ethernet/google/gve/gve_rx.c
229
rx->qpl_copy_pool[j].page_address = page_address(page);
drivers/net/ethernet/google/gve/gve_rx.c
230
rx->qpl_copy_pool[j].buf_size = rx->packet_buffer_size;
drivers/net/ethernet/google/gve/gve_rx.c
234
rx->qpl_copy_pool[j].pagecnt_bias = INT_MAX;
drivers/net/ethernet/google/gve/gve_rx.c
243
page_ref_sub(rx->qpl_copy_pool[j].page,
drivers/net/ethernet/google/gve/gve_rx.c
244
rx->qpl_copy_pool[j].pagecnt_bias - 1);
drivers/net/ethernet/google/gve/gve_rx.c
245
put_page(rx->qpl_copy_pool[j].page);
drivers/net/ethernet/google/gve/gve_rx.c
252
page_ref_sub(rx->data.page_info[i].page,
drivers/net/ethernet/google/gve/gve_rx.c
253
rx->data.page_info[i].pagecnt_bias - 1);
drivers/net/ethernet/google/gve/gve_rx.c
260
&rx->data.page_info[i],
drivers/net/ethernet/google/gve/gve_rx.c
261
&rx->data.data_ring[i]);
drivers/net/ethernet/google/gve/gve_rx.c
27
struct gve_rx_ring *rx,
drivers/net/ethernet/google/gve/gve_rx.c
275
struct gve_rx_ring *rx,
drivers/net/ethernet/google/gve/gve_rx.c
288
memset(rx, 0, sizeof(*rx));
drivers/net/ethernet/google/gve/gve_rx.c
290
rx->gve = priv;
drivers/net/ethernet/google/gve/gve_rx.c
291
rx->q_num = idx;
drivers/net/ethernet/google/gve/gve_rx.c
292
rx->packet_buffer_size = cfg->packet_buffer_size;
drivers/net/ethernet/google/gve/gve_rx.c
294
rx->mask = slots - 1;
drivers/net/ethernet/google/gve/gve_rx.c
295
rx->data.raw_addressing = cfg->raw_addressing;
drivers/net/ethernet/google/gve/gve_rx.c
298
bytes = sizeof(*rx->data.data_ring) * slots;
drivers/net/ethernet/google/gve/gve_rx.c
299
rx->data.data_ring = dma_alloc_coherent(hdev, bytes,
drivers/net/ethernet/google/gve/gve_rx.c
30
u32 slots = rx->mask + 1;
drivers/net/ethernet/google/gve/gve_rx.c
300
&rx->data.data_bus,
drivers/net/ethernet/google/gve/gve_rx.c
302
if (!rx->data.data_ring)
drivers/net/ethernet/google/gve/gve_rx.c
305
rx->qpl_copy_pool_mask = min_t(u32, U32_MAX, slots * 2) - 1;
drivers/net/ethernet/google/gve/gve_rx.c
306
rx->qpl_copy_pool_head = 0;
drivers/net/ethernet/google/gve/gve_rx.c
307
rx->qpl_copy_pool = kvcalloc_node(rx->qpl_copy_pool_mask + 1,
drivers/net/ethernet/google/gve/gve_rx.c
308
sizeof(rx->qpl_copy_pool[0]),
drivers/net/ethernet/google/gve/gve_rx.c
310
if (!rx->qpl_copy_pool) {
drivers/net/ethernet/google/gve/gve_rx.c
315
if (!rx->data.raw_addressing) {
drivers/net/ethernet/google/gve/gve_rx.c
316
qpl_id = gve_get_rx_qpl_id(cfg->qcfg_tx, rx->q_num);
drivers/net/ethernet/google/gve/gve_rx.c
319
rx->data.qpl = gve_alloc_queue_page_list(priv, qpl_id,
drivers/net/ethernet/google/gve/gve_rx.c
321
if (!rx->data.qpl) {
drivers/net/ethernet/google/gve/gve_rx.c
327
filled_pages = gve_rx_prefill_pages(rx, cfg);
drivers/net/ethernet/google/gve/gve_rx.c
33
if (!rx->data.page_info)
drivers/net/ethernet/google/gve/gve_rx.c
332
rx->fill_cnt = filled_pages;
drivers/net/ethernet/google/gve/gve_rx.c
337
rx->q_resources =
drivers/net/ethernet/google/gve/gve_rx.c
339
sizeof(*rx->q_resources),
drivers/net/ethernet/google/gve/gve_rx.c
340
&rx->q_resources_bus,
drivers/net/ethernet/google/gve/gve_rx.c
342
if (!rx->q_resources) {
drivers/net/ethernet/google/gve/gve_rx.c
347
(unsigned long)rx->data.data_bus);
drivers/net/ethernet/google/gve/gve_rx.c
351
rx->desc.desc_ring = dma_alloc_coherent(hdev, bytes, &rx->desc.bus,
drivers/net/ethernet/google/gve/gve_rx.c
353
if (!rx->desc.desc_ring) {
drivers/net/ethernet/google/gve/gve_rx.c
357
rx->db_threshold = slots / 2;
drivers/net/ethernet/google/gve/gve_rx.c
358
gve_rx_init_ring_state_gqi(rx);
drivers/net/ethernet/google/gve/gve_rx.c
36
if (rx->data.raw_addressing) {
drivers/net/ethernet/google/gve/gve_rx.c
360
gve_rx_ctx_clear(&rx->ctx);
drivers/net/ethernet/google/gve/gve_rx.c
365
dma_free_coherent(hdev, sizeof(*rx->q_resources),
drivers/net/ethernet/google/gve/gve_rx.c
366
rx->q_resources, rx->q_resources_bus);
drivers/net/ethernet/google/gve/gve_rx.c
367
rx->q_resources = NULL;
drivers/net/ethernet/google/gve/gve_rx.c
369
gve_rx_unfill_pages(priv, rx, cfg);
drivers/net/ethernet/google/gve/gve_rx.c
371
if (!rx->data.raw_addressing) {
drivers/net/ethernet/google/gve/gve_rx.c
372
gve_free_queue_page_list(priv, rx->data.qpl, qpl_id);
drivers/net/ethernet/google/gve/gve_rx.c
373
rx->data.qpl = NULL;
drivers/net/ethernet/google/gve/gve_rx.c
376
kvfree(rx->qpl_copy_pool);
drivers/net/ethernet/google/gve/gve_rx.c
377
rx->qpl_copy_pool = NULL;
drivers/net/ethernet/google/gve/gve_rx.c
379
bytes = sizeof(*rx->data.data_ring) * slots;
drivers/net/ethernet/google/gve/gve_rx.c
38
gve_rx_free_buffer(&priv->pdev->dev, &rx->data.page_info[i],
drivers/net/ethernet/google/gve/gve_rx.c
380
dma_free_coherent(hdev, bytes, rx->data.data_ring, rx->data.data_bus);
drivers/net/ethernet/google/gve/gve_rx.c
381
rx->data.data_ring = NULL;
drivers/net/ethernet/google/gve/gve_rx.c
389
struct gve_rx_ring *rx;
drivers/net/ethernet/google/gve/gve_rx.c
39
&rx->data.data_ring[i]);
drivers/net/ethernet/google/gve/gve_rx.c
393
rx = kvzalloc_objs(struct gve_rx_ring, cfg->qcfg_rx->max_queues);
drivers/net/ethernet/google/gve/gve_rx.c
394
if (!rx)
drivers/net/ethernet/google/gve/gve_rx.c
398
err = gve_rx_alloc_ring_gqi(priv, cfg, &rx[i], i);
drivers/net/ethernet/google/gve/gve_rx.c
407
cfg->rx = rx;
drivers/net/ethernet/google/gve/gve_rx.c
412
gve_rx_free_ring_gqi(priv, &rx[j], cfg);
drivers/net/ethernet/google/gve/gve_rx.c
413
kvfree(rx);
drivers/net/ethernet/google/gve/gve_rx.c
42
page_ref_sub(rx->data.page_info[i].page,
drivers/net/ethernet/google/gve/gve_rx.c
420
struct gve_rx_ring *rx = cfg->rx;
drivers/net/ethernet/google/gve/gve_rx.c
423
if (!rx)
drivers/net/ethernet/google/gve/gve_rx.c
427
gve_rx_free_ring_gqi(priv, &rx[i], cfg);
drivers/net/ethernet/google/gve/gve_rx.c
429
kvfree(rx);
drivers/net/ethernet/google/gve/gve_rx.c
43
rx->data.page_info[i].pagecnt_bias - 1);
drivers/net/ethernet/google/gve/gve_rx.c
430
cfg->rx = NULL;
drivers/net/ethernet/google/gve/gve_rx.c
433
void gve_rx_write_doorbell(struct gve_priv *priv, struct gve_rx_ring *rx)
drivers/net/ethernet/google/gve/gve_rx.c
435
u32 db_idx = be32_to_cpu(rx->q_resources->db_index);
drivers/net/ethernet/google/gve/gve_rx.c
437
iowrite32be(rx->fill_cnt, &priv->db_bar2[db_idx]);
drivers/net/ethernet/google/gve/gve_rx.c
45
for (i = 0; i < rx->qpl_copy_pool_mask + 1; i++) {
drivers/net/ethernet/google/gve/gve_rx.c
46
page_ref_sub(rx->qpl_copy_pool[i].page,
drivers/net/ethernet/google/gve/gve_rx.c
47
rx->qpl_copy_pool[i].pagecnt_bias - 1);
drivers/net/ethernet/google/gve/gve_rx.c
48
put_page(rx->qpl_copy_pool[i].page);
drivers/net/ethernet/google/gve/gve_rx.c
51
kvfree(rx->data.page_info);
drivers/net/ethernet/google/gve/gve_rx.c
52
rx->data.page_info = NULL;
drivers/net/ethernet/google/gve/gve_rx.c
537
static struct sk_buff *gve_rx_copy_to_pool(struct gve_rx_ring *rx,
drivers/net/ethernet/google/gve/gve_rx.c
541
u32 pool_idx = rx->qpl_copy_pool_head & rx->qpl_copy_pool_mask;
drivers/net/ethernet/google/gve/gve_rx.c
544
struct gve_rx_ctx *ctx = &rx->ctx;
drivers/net/ethernet/google/gve/gve_rx.c
549
copy_page_info = &rx->qpl_copy_pool[pool_idx];
drivers/net/ethernet/google/gve/gve_rx.c
554
gve_schedule_reset(rx->gve);
drivers/net/ethernet/google/gve/gve_rx.c
568
rx->qpl_copy_pool_head++;
drivers/net/ethernet/google/gve/gve_rx.c
584
u64_stats_update_begin(&rx->statss);
drivers/net/ethernet/google/gve/gve_rx.c
585
rx->rx_frag_copy_cnt++;
drivers/net/ethernet/google/gve/gve_rx.c
586
rx->rx_frag_alloc_cnt++;
drivers/net/ethernet/google/gve/gve_rx.c
587
u64_stats_update_end(&rx->statss);
drivers/net/ethernet/google/gve/gve_rx.c
609
rx->qpl_copy_pool_head++;
drivers/net/ethernet/google/gve/gve_rx.c
610
prefetch(rx->qpl_copy_pool[rx->qpl_copy_pool_head & rx->qpl_copy_pool_mask].page);
drivers/net/ethernet/google/gve/gve_rx.c
615
u64_stats_update_begin(&rx->statss);
drivers/net/ethernet/google/gve/gve_rx.c
616
rx->rx_frag_copy_cnt++;
drivers/net/ethernet/google/gve/gve_rx.c
617
u64_stats_update_end(&rx->statss);
drivers/net/ethernet/google/gve/gve_rx.c
624
struct gve_rx_ring *rx, struct gve_rx_slot_page_info *page_info,
drivers/net/ethernet/google/gve/gve_rx.c
628
struct gve_rx_ctx *ctx = &rx->ctx;
drivers/net/ethernet/google/gve/gve_rx.c
64
static void gve_rx_init_ring_state_gqi(struct gve_rx_ring *rx)
drivers/net/ethernet/google/gve/gve_rx.c
646
skb = gve_rx_copy_to_pool(rx, page_info, len, napi);
drivers/net/ethernet/google/gve/gve_rx.c
651
static struct sk_buff *gve_rx_skb(struct gve_priv *priv, struct gve_rx_ring *rx,
drivers/net/ethernet/google/gve/gve_rx.c
657
struct gve_rx_ctx *ctx = &rx->ctx;
drivers/net/ethernet/google/gve/gve_rx.c
66
rx->desc.seqno = 1;
drivers/net/ethernet/google/gve/gve_rx.c
664
u64_stats_update_begin(&rx->statss);
drivers/net/ethernet/google/gve/gve_rx.c
665
rx->rx_copied_pkt++;
drivers/net/ethernet/google/gve/gve_rx.c
666
rx->rx_frag_copy_cnt++;
drivers/net/ethernet/google/gve/gve_rx.c
667
rx->rx_copybreak_pkt++;
drivers/net/ethernet/google/gve/gve_rx.c
668
u64_stats_update_end(&rx->statss);
drivers/net/ethernet/google/gve/gve_rx.c
67
rx->cnt = 0;
drivers/net/ethernet/google/gve/gve_rx.c
679
u64_stats_update_begin(&rx->statss);
drivers/net/ethernet/google/gve/gve_rx.c
68
gve_rx_ctx_clear(&rx->ctx);
drivers/net/ethernet/google/gve/gve_rx.c
680
rx->rx_frag_flip_cnt++;
drivers/net/ethernet/google/gve/gve_rx.c
681
u64_stats_update_end(&rx->statss);
drivers/net/ethernet/google/gve/gve_rx.c
684
if (rx->data.raw_addressing) {
drivers/net/ethernet/google/gve/gve_rx.c
690
skb = gve_rx_qpl(&priv->pdev->dev, netdev, rx,
drivers/net/ethernet/google/gve/gve_rx.c
698
struct gve_rx_ring *rx,
drivers/net/ethernet/google/gve/gve_rx.c
705
if (rx->xsk_pool->frame_len < len)
drivers/net/ethernet/google/gve/gve_rx.c
707
xdp = xsk_buff_alloc(rx->xsk_pool);
drivers/net/ethernet/google/gve/gve_rx.c
709
u64_stats_update_begin(&rx->statss);
drivers/net/ethernet/google/gve/gve_rx.c
710
rx->xdp_alloc_fails++;
drivers/net/ethernet/google/gve/gve_rx.c
711
u64_stats_update_end(&rx->statss);
drivers/net/ethernet/google/gve/gve_rx.c
722
static int gve_xdp_redirect(struct net_device *dev, struct gve_rx_ring *rx,
drivers/net/ethernet/google/gve/gve_rx.c
73
struct gve_rx_ring *rx = &priv->rx[idx];
drivers/net/ethernet/google/gve/gve_rx.c
731
if (rx->xsk_pool)
drivers/net/ethernet/google/gve/gve_rx.c
732
return gve_xsk_pool_redirect(dev, rx, orig->data,
drivers/net/ethernet/google/gve/gve_rx.c
737
frame = page_frag_alloc(&rx->page_cache, total_len, GFP_ATOMIC);
drivers/net/ethernet/google/gve/gve_rx.c
739
u64_stats_update_begin(&rx->statss);
drivers/net/ethernet/google/gve/gve_rx.c
740
rx->xdp_alloc_fails++;
drivers/net/ethernet/google/gve/gve_rx.c
741
u64_stats_update_end(&rx->statss);
drivers/net/ethernet/google/gve/gve_rx.c
744
xdp_init_buff(&new, total_len, &rx->xdp_rxq);
drivers/net/ethernet/google/gve/gve_rx.c
755
static void gve_xdp_done(struct gve_priv *priv, struct gve_rx_ring *rx,
drivers/net/ethernet/google/gve/gve_rx.c
769
tx_qid = gve_xdp_tx_queue_id(priv, rx->q_num);
drivers/net/ethernet/google/gve/gve_rx.c
777
u64_stats_update_begin(&rx->statss);
drivers/net/ethernet/google/gve/gve_rx.c
778
rx->xdp_tx_errors++;
drivers/net/ethernet/google/gve/gve_rx.c
779
u64_stats_update_end(&rx->statss);
drivers/net/ethernet/google/gve/gve_rx.c
78
if (rx->desc.desc_ring) {
drivers/net/ethernet/google/gve/gve_rx.c
783
err = gve_xdp_redirect(priv->dev, rx, xdp, xprog);
drivers/net/ethernet/google/gve/gve_rx.c
786
u64_stats_update_begin(&rx->statss);
drivers/net/ethernet/google/gve/gve_rx.c
787
rx->xdp_redirect_errors++;
drivers/net/ethernet/google/gve/gve_rx.c
788
u64_stats_update_end(&rx->statss);
drivers/net/ethernet/google/gve/gve_rx.c
79
size = slots * sizeof(rx->desc.desc_ring[0]);
drivers/net/ethernet/google/gve/gve_rx.c
792
u64_stats_update_begin(&rx->statss);
drivers/net/ethernet/google/gve/gve_rx.c
794
rx->xdp_actions[xdp_act]++;
drivers/net/ethernet/google/gve/gve_rx.c
795
u64_stats_update_end(&rx->statss);
drivers/net/ethernet/google/gve/gve_rx.c
799
static void gve_rx(struct gve_rx_ring *rx, netdev_features_t feat,
drivers/net/ethernet/google/gve/gve_rx.c
80
memset(rx->desc.desc_ring, 0, size);
drivers/net/ethernet/google/gve/gve_rx.c
806
struct gve_rx_ctx *ctx = &rx->ctx;
drivers/net/ethernet/google/gve/gve_rx.c
808
struct gve_priv *priv = rx->gve;
drivers/net/ethernet/google/gve/gve_rx.c
816
struct napi_struct *napi = &priv->ntfy_blocks[rx->ntfy_id].napi;
drivers/net/ethernet/google/gve/gve_rx.c
831
if (unlikely(frag_size > rx->packet_buffer_size)) {
drivers/net/ethernet/google/gve/gve_rx.c
833
frag_size, rx->packet_buffer_size);
drivers/net/ethernet/google/gve/gve_rx.c
836
gve_schedule_reset(rx->gve);
drivers/net/ethernet/google/gve/gve_rx.c
84
if (rx->q_resources)
drivers/net/ethernet/google/gve/gve_rx.c
841
page_info = &rx->data.page_info[(idx + 2) & rx->mask];
drivers/net/ethernet/google/gve/gve_rx.c
847
page_info = &rx->data.page_info[idx];
drivers/net/ethernet/google/gve/gve_rx.c
848
data_slot = &rx->data.data_ring[idx];
drivers/net/ethernet/google/gve/gve_rx.c
849
page_bus = (rx->data.raw_addressing) ?
drivers/net/ethernet/google/gve/gve_rx.c
85
memset(rx->q_resources, 0, sizeof(*rx->q_resources));
drivers/net/ethernet/google/gve/gve_rx.c
851
rx->data.qpl->page_buses[idx];
drivers/net/ethernet/google/gve/gve_rx.c
863
xdp_init_buff(&xdp, page_info->buf_size, &rx->xdp_rxq);
drivers/net/ethernet/google/gve/gve_rx.c
87
gve_rx_init_ring_state_gqi(rx);
drivers/net/ethernet/google/gve/gve_rx.c
870
gve_xdp_done(priv, rx, &xdp, xprog, xdp_act);
drivers/net/ethernet/google/gve/gve_rx.c
878
u64_stats_update_begin(&rx->statss);
drivers/net/ethernet/google/gve/gve_rx.c
879
rx->xdp_actions[XDP_PASS]++;
drivers/net/ethernet/google/gve/gve_rx.c
880
u64_stats_update_end(&rx->statss);
drivers/net/ethernet/google/gve/gve_rx.c
883
skb = gve_rx_skb(priv, rx, page_info, napi, len,
drivers/net/ethernet/google/gve/gve_rx.c
886
u64_stats_update_begin(&rx->statss);
drivers/net/ethernet/google/gve/gve_rx.c
887
rx->rx_skb_alloc_fail++;
drivers/net/ethernet/google/gve/gve_rx.c
888
u64_stats_update_end(&rx->statss);
drivers/net/ethernet/google/gve/gve_rx.c
914
skb_record_rx_queue(skb, rx->q_num);
drivers/net/ethernet/google/gve/gve_rx.c
936
bool gve_rx_work_pending(struct gve_rx_ring *rx)
drivers/net/ethernet/google/gve/gve_rx.c
942
next_idx = rx->cnt & rx->mask;
drivers/net/ethernet/google/gve/gve_rx.c
943
desc = rx->desc.desc_ring + next_idx;
drivers/net/ethernet/google/gve/gve_rx.c
947
return (GVE_SEQNO(flags_seq) == rx->desc.seqno);
drivers/net/ethernet/google/gve/gve_rx.c
950
static bool gve_rx_refill_buffers(struct gve_priv *priv, struct gve_rx_ring *rx)
drivers/net/ethernet/google/gve/gve_rx.c
952
int refill_target = rx->mask + 1;
drivers/net/ethernet/google/gve/gve_rx.c
953
u32 fill_cnt = rx->fill_cnt;
drivers/net/ethernet/google/gve/gve_rx.c
955
while (fill_cnt - rx->cnt < refill_target) {
drivers/net/ethernet/google/gve/gve_rx.c
957
u32 idx = fill_cnt & rx->mask;
drivers/net/ethernet/google/gve/gve_rx.c
959
page_info = &rx->data.page_info[idx];
drivers/net/ethernet/google/gve/gve_rx.c
965
&rx->data.data_ring[idx];
drivers/net/ethernet/google/gve/gve_rx.c
980
if (!rx->data.raw_addressing)
drivers/net/ethernet/google/gve/gve_rx.c
987
&rx->data.data_ring[idx];
drivers/net/ethernet/google/gve/gve_rx.c
992
data_slot, rx)) {
drivers/net/ethernet/google/gve/gve_rx.c
999
rx->fill_cnt = fill_cnt;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
100
if (rx->dqo.page_pool)
drivers/net/ethernet/google/gve/gve_rx_dqo.c
1009
struct gve_rx_ring *rx;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
101
gve_free_to_page_pool(rx, bs, false);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
1020
rx = block->rx;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
1021
priv = rx->gve;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
1022
complq = &rx->dqo.complq;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
1024
xdp_redirects = rx->xdp_actions[XDP_REDIRECT];
drivers/net/ethernet/google/gve/gve_rx_dqo.c
1025
xdp_txs = rx->xdp_actions[XDP_TX];
drivers/net/ethernet/google/gve/gve_rx_dqo.c
1043
err = gve_rx_dqo(napi, rx, compl_desc, complq->head, rx->q_num);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
1045
gve_rx_free_skb(napi, rx);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
1046
u64_stats_update_begin(&rx->statss);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
1048
rx->rx_skb_alloc_fail++;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
1050
rx->rx_desc_err_dropped_pkt++;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
1051
u64_stats_update_end(&rx->statss);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
1064
struct gve_rx_buf_queue_dqo *bufq = &rx->dqo.bufq;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
107
gve_rx_init_ring_state_dqo(rx, buffer_queue_slots,
drivers/net/ethernet/google/gve/gve_rx_dqo.c
1070
rx->cnt++;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
1072
if (!rx->ctx.skb_head)
drivers/net/ethernet/google/gve/gve_rx_dqo.c
1079
pkt_bytes = rx->ctx.skb_head->len;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
1083
if (skb_headlen(rx->ctx.skb_head))
drivers/net/ethernet/google/gve/gve_rx_dqo.c
1087
if (gve_rx_complete_skb(rx, napi, compl_desc, feat) != 0) {
drivers/net/ethernet/google/gve/gve_rx_dqo.c
1088
gve_rx_free_skb(napi, rx);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
1089
u64_stats_update_begin(&rx->statss);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
1090
rx->rx_desc_err_dropped_pkt++;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
1091
u64_stats_update_end(&rx->statss);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
1096
rx->ctx.skb_head = NULL;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
1097
rx->ctx.skb_tail = NULL;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
1100
if (xdp_txs != rx->xdp_actions[XDP_TX])
drivers/net/ethernet/google/gve/gve_rx_dqo.c
1101
gve_xdp_tx_flush_dqo(priv, rx->q_num);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
1103
if (xdp_redirects != rx->xdp_actions[XDP_REDIRECT])
drivers/net/ethernet/google/gve/gve_rx_dqo.c
1106
gve_rx_post_buffers_dqo(rx);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
1108
u64_stats_update_begin(&rx->statss);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
1109
rx->rpackets += work_done;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
1110
rx->rbytes += bytes;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
1111
u64_stats_update_end(&rx->statss);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
114
struct gve_rx_ring *rx = &priv->rx[idx];
drivers/net/ethernet/google/gve/gve_rx_dqo.c
119
if (rx->dqo.page_pool)
drivers/net/ethernet/google/gve/gve_rx_dqo.c
120
page_pool_disable_direct_recycling(rx->dqo.page_pool);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
126
void gve_rx_free_ring_dqo(struct gve_priv *priv, struct gve_rx_ring *rx,
drivers/net/ethernet/google/gve/gve_rx_dqo.c
132
int idx = rx->q_num;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
137
completion_queue_slots = rx->dqo.complq.mask + 1;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
138
buffer_queue_slots = rx->dqo.bufq.mask + 1;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
140
if (rx->q_resources) {
drivers/net/ethernet/google/gve/gve_rx_dqo.c
141
dma_free_coherent(hdev, sizeof(*rx->q_resources),
drivers/net/ethernet/google/gve/gve_rx_dqo.c
142
rx->q_resources, rx->q_resources_bus);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
143
rx->q_resources = NULL;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
146
for (i = 0; i < rx->dqo.num_buf_states; i++) {
drivers/net/ethernet/google/gve/gve_rx_dqo.c
147
struct gve_rx_buf_state_dqo *bs = &rx->dqo.buf_states[i];
drivers/net/ethernet/google/gve/gve_rx_dqo.c
149
if (rx->dqo.page_pool)
drivers/net/ethernet/google/gve/gve_rx_dqo.c
150
gve_free_to_page_pool(rx, bs, false);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
153
if (gve_buf_state_is_allocated(rx, bs) && bs->xsk_buff) {
drivers/net/ethernet/google/gve/gve_rx_dqo.c
159
if (rx->dqo.qpl) {
drivers/net/ethernet/google/gve/gve_rx_dqo.c
160
qpl_id = gve_get_rx_qpl_id(cfg->qcfg_tx, rx->q_num);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
161
gve_free_queue_page_list(priv, rx->dqo.qpl, qpl_id);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
162
rx->dqo.qpl = NULL;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
165
if (rx->dqo.bufq.desc_ring) {
drivers/net/ethernet/google/gve/gve_rx_dqo.c
166
size = sizeof(rx->dqo.bufq.desc_ring[0]) * buffer_queue_slots;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
167
dma_free_coherent(hdev, size, rx->dqo.bufq.desc_ring,
drivers/net/ethernet/google/gve/gve_rx_dqo.c
168
rx->dqo.bufq.bus);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
169
rx->dqo.bufq.desc_ring = NULL;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
172
if (rx->dqo.complq.desc_ring) {
drivers/net/ethernet/google/gve/gve_rx_dqo.c
173
size = sizeof(rx->dqo.complq.desc_ring[0]) *
drivers/net/ethernet/google/gve/gve_rx_dqo.c
175
dma_free_coherent(hdev, size, rx->dqo.complq.desc_ring,
drivers/net/ethernet/google/gve/gve_rx_dqo.c
176
rx->dqo.complq.bus);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
177
rx->dqo.complq.desc_ring = NULL;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
180
kvfree(rx->dqo.buf_states);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
181
rx->dqo.buf_states = NULL;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
183
if (rx->dqo.page_pool) {
drivers/net/ethernet/google/gve/gve_rx_dqo.c
184
page_pool_destroy(rx->dqo.page_pool);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
185
rx->dqo.page_pool = NULL;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
188
gve_rx_free_hdr_bufs(priv, rx);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
193
static int gve_rx_alloc_hdr_bufs(struct gve_priv *priv, struct gve_rx_ring *rx,
drivers/net/ethernet/google/gve/gve_rx_dqo.c
198
rx->dqo.hdr_bufs.data = dma_alloc_coherent(hdev, priv->header_buf_size * buf_count,
drivers/net/ethernet/google/gve/gve_rx_dqo.c
199
&rx->dqo.hdr_bufs.addr, GFP_KERNEL);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
200
if (!rx->dqo.hdr_bufs.data)
drivers/net/ethernet/google/gve/gve_rx_dqo.c
21
static void gve_rx_free_hdr_bufs(struct gve_priv *priv, struct gve_rx_ring *rx)
drivers/net/ethernet/google/gve/gve_rx_dqo.c
216
struct gve_rx_ring *rx,
drivers/net/ethernet/google/gve/gve_rx_dqo.c
230
memset(rx, 0, sizeof(*rx));
drivers/net/ethernet/google/gve/gve_rx_dqo.c
231
rx->gve = priv;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
232
rx->q_num = idx;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
233
rx->packet_buffer_size = cfg->packet_buffer_size;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
236
rx->packet_buffer_truesize = GVE_XDP_RX_BUFFER_SIZE_DQO;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
237
rx->rx_headroom = XDP_PACKET_HEADROOM;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
239
rx->packet_buffer_truesize = rx->packet_buffer_size;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
24
int buf_count = rx->dqo.bufq.mask + 1;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
240
rx->rx_headroom = 0;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
248
rx->dqo.num_buf_states = cfg->raw_addressing ? buffer_queue_slots :
drivers/net/ethernet/google/gve/gve_rx_dqo.c
250
rx->dqo.buf_states = kvcalloc_node(rx->dqo.num_buf_states,
drivers/net/ethernet/google/gve/gve_rx_dqo.c
251
sizeof(rx->dqo.buf_states[0]),
drivers/net/ethernet/google/gve/gve_rx_dqo.c
253
if (!rx->dqo.buf_states)
drivers/net/ethernet/google/gve/gve_rx_dqo.c
258
if (gve_rx_alloc_hdr_bufs(priv, rx, buffer_queue_slots))
drivers/net/ethernet/google/gve/gve_rx_dqo.c
26
if (rx->dqo.hdr_bufs.data) {
drivers/net/ethernet/google/gve/gve_rx_dqo.c
262
size = sizeof(rx->dqo.complq.desc_ring[0]) *
drivers/net/ethernet/google/gve/gve_rx_dqo.c
264
rx->dqo.complq.desc_ring =
drivers/net/ethernet/google/gve/gve_rx_dqo.c
265
dma_alloc_coherent(hdev, size, &rx->dqo.complq.bus, GFP_KERNEL);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
266
if (!rx->dqo.complq.desc_ring)
drivers/net/ethernet/google/gve/gve_rx_dqo.c
270
size = sizeof(rx->dqo.bufq.desc_ring[0]) * buffer_queue_slots;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
271
rx->dqo.bufq.desc_ring =
drivers/net/ethernet/google/gve/gve_rx_dqo.c
272
dma_alloc_coherent(hdev, size, &rx->dqo.bufq.bus, GFP_KERNEL);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
273
if (!rx->dqo.bufq.desc_ring)
drivers/net/ethernet/google/gve/gve_rx_dqo.c
277
pool = gve_rx_create_page_pool(priv, rx, cfg->xdp);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
28
rx->dqo.hdr_bufs.data, rx->dqo.hdr_bufs.addr);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
281
rx->dqo.page_pool = pool;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
283
qpl_id = gve_get_rx_qpl_id(cfg->qcfg_tx, rx->q_num);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
286
rx->dqo.qpl = gve_alloc_queue_page_list(priv, qpl_id,
drivers/net/ethernet/google/gve/gve_rx_dqo.c
288
if (!rx->dqo.qpl)
drivers/net/ethernet/google/gve/gve_rx_dqo.c
29
rx->dqo.hdr_bufs.data = NULL;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
290
rx->dqo.next_qpl_page_idx = 0;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
293
rx->q_resources = dma_alloc_coherent(hdev, sizeof(*rx->q_resources),
drivers/net/ethernet/google/gve/gve_rx_dqo.c
294
&rx->q_resources_bus, GFP_KERNEL);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
295
if (!rx->q_resources)
drivers/net/ethernet/google/gve/gve_rx_dqo.c
298
gve_rx_init_ring_state_dqo(rx, buffer_queue_slots,
drivers/net/ethernet/google/gve/gve_rx_dqo.c
304
gve_rx_free_ring_dqo(priv, rx, cfg);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
310
const struct gve_rx_ring *rx = &priv->rx[queue_idx];
drivers/net/ethernet/google/gve/gve_rx_dqo.c
311
u64 index = be32_to_cpu(rx->q_resources->db_index);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
313
iowrite32(rx->dqo.bufq.tail, &priv->db_bar2[index]);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
319
struct gve_rx_ring *rx;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
323
rx = kvzalloc_objs(struct gve_rx_ring, cfg->qcfg_rx->max_queues);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
324
if (!rx)
drivers/net/ethernet/google/gve/gve_rx_dqo.c
328
err = gve_rx_alloc_ring_dqo(priv, cfg, &rx[i], i);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
33
static void gve_rx_init_ring_state_dqo(struct gve_rx_ring *rx,
drivers/net/ethernet/google/gve/gve_rx_dqo.c
337
cfg->rx = rx;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
342
gve_rx_free_ring_dqo(priv, &rx[i], cfg);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
343
kvfree(rx);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
350
struct gve_rx_ring *rx = cfg->rx;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
353
if (!rx)
drivers/net/ethernet/google/gve/gve_rx_dqo.c
357
gve_rx_free_ring_dqo(priv, &rx[i], cfg);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
359
kvfree(rx);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
360
cfg->rx = NULL;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
363
void gve_rx_post_buffers_dqo(struct gve_rx_ring *rx)
drivers/net/ethernet/google/gve/gve_rx_dqo.c
365
struct gve_rx_compl_queue_dqo *complq = &rx->dqo.complq;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
366
struct gve_rx_buf_queue_dqo *bufq = &rx->dqo.bufq;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
367
struct gve_priv *priv = rx->gve;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
379
if (unlikely(gve_alloc_buffer(rx, desc))) {
drivers/net/ethernet/google/gve/gve_rx_dqo.c
380
u64_stats_update_begin(&rx->statss);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
381
rx->rx_buf_alloc_fail++;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
382
u64_stats_update_end(&rx->statss);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
386
if (rx->dqo.hdr_bufs.data)
drivers/net/ethernet/google/gve/gve_rx_dqo.c
388
cpu_to_le64(rx->dqo.hdr_bufs.addr +
drivers/net/ethernet/google/gve/gve_rx_dqo.c
396
gve_rx_write_doorbell_dqo(priv, rx->q_num);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
399
rx->fill_cnt += num_posted;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
40
rx->dqo.bufq.mask = buffer_queue_slots - 1;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
41
rx->dqo.bufq.head = 0;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
42
rx->dqo.bufq.tail = 0;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
45
rx->dqo.complq.num_free_slots = completion_queue_slots;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
46
rx->dqo.complq.mask = completion_queue_slots - 1;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
47
rx->dqo.complq.cur_gen_bit = 0;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
472
static void gve_rx_skb_hwtstamp(struct gve_rx_ring *rx,
drivers/net/ethernet/google/gve/gve_rx_dqo.c
475
struct sk_buff *skb = rx->ctx.skb_head;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
479
gve_rx_get_hwtstamp(rx->gve, le32_to_cpu(desc->ts));
drivers/net/ethernet/google/gve/gve_rx_dqo.c
48
rx->dqo.complq.head = 0;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
497
static void gve_rx_free_skb(struct napi_struct *napi, struct gve_rx_ring *rx)
drivers/net/ethernet/google/gve/gve_rx_dqo.c
499
if (!rx->ctx.skb_head)
drivers/net/ethernet/google/gve/gve_rx_dqo.c
502
if (rx->ctx.skb_head == napi->skb)
drivers/net/ethernet/google/gve/gve_rx_dqo.c
504
dev_kfree_skb_any(rx->ctx.skb_head);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
505
rx->ctx.skb_head = NULL;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
506
rx->ctx.skb_tail = NULL;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
509
static bool gve_rx_should_trigger_copy_ondemand(struct gve_rx_ring *rx)
drivers/net/ethernet/google/gve/gve_rx_dqo.c
51
rx->ctx.skb_head = NULL;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
511
if (!rx->dqo.qpl)
drivers/net/ethernet/google/gve/gve_rx_dqo.c
513
if (rx->dqo.used_buf_states_cnt <
drivers/net/ethernet/google/gve/gve_rx_dqo.c
514
(rx->dqo.num_buf_states -
drivers/net/ethernet/google/gve/gve_rx_dqo.c
52
rx->ctx.skb_tail = NULL;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
520
static int gve_rx_copy_ondemand(struct gve_rx_ring *rx,
drivers/net/ethernet/google/gve/gve_rx_dqo.c
524
struct page *page = alloc_pages_node(rx->gve->numa_node, GFP_ATOMIC, 0);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
534
num_frags = skb_shinfo(rx->ctx.skb_tail)->nr_frags;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
535
skb_add_rx_frag(rx->ctx.skb_tail, num_frags, page,
drivers/net/ethernet/google/gve/gve_rx_dqo.c
538
u64_stats_update_begin(&rx->statss);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
539
rx->rx_frag_alloc_cnt++;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
540
u64_stats_update_end(&rx->statss);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
542
gve_enqueue_buf_state(rx, &rx->dqo.recycled_buf_states, buf_state);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
546
static void gve_skb_add_rx_frag(struct gve_rx_ring *rx,
drivers/net/ethernet/google/gve/gve_rx_dqo.c
55
if (rx->dqo.buf_states) {
drivers/net/ethernet/google/gve/gve_rx_dqo.c
550
if (rx->dqo.page_pool) {
drivers/net/ethernet/google/gve/gve_rx_dqo.c
551
skb_add_rx_frag_netmem(rx->ctx.skb_tail, num_frags,
drivers/net/ethernet/google/gve/gve_rx_dqo.c
557
skb_add_rx_frag(rx->ctx.skb_tail, num_frags,
drivers/net/ethernet/google/gve/gve_rx_dqo.c
56
for (i = 0; i < rx->dqo.num_buf_states - 1; i++)
drivers/net/ethernet/google/gve/gve_rx_dqo.c
57
rx->dqo.buf_states[i].next = i + 1;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
570
u16 buf_len, struct gve_rx_ring *rx,
drivers/net/ethernet/google/gve/gve_rx_dqo.c
573
int num_frags = skb_shinfo(rx->ctx.skb_tail)->nr_frags;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
58
rx->dqo.buf_states[rx->dqo.num_buf_states - 1].next = -1;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
582
if (rx->dqo.page_pool)
drivers/net/ethernet/google/gve/gve_rx_dqo.c
585
if (rx->ctx.skb_tail == rx->ctx.skb_head)
drivers/net/ethernet/google/gve/gve_rx_dqo.c
586
skb_shinfo(rx->ctx.skb_head)->frag_list = skb;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
588
rx->ctx.skb_tail->next = skb;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
589
rx->ctx.skb_tail = skb;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
592
if (rx->ctx.skb_tail != rx->ctx.skb_head) {
drivers/net/ethernet/google/gve/gve_rx_dqo.c
593
rx->ctx.skb_head->len += buf_len;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
594
rx->ctx.skb_head->data_len += buf_len;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
595
rx->ctx.skb_head->truesize += buf_state->page_info.buf_size;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
599
if (gve_rx_should_trigger_copy_ondemand(rx))
drivers/net/ethernet/google/gve/gve_rx_dqo.c
600
return gve_rx_copy_ondemand(rx, buf_state, buf_len);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
602
gve_skb_add_rx_frag(rx, buf_state, num_frags, buf_len);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
603
gve_reuse_buffer(rx, buf_state);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
607
static int gve_xdp_tx_dqo(struct gve_priv *priv, struct gve_rx_ring *rx,
drivers/net/ethernet/google/gve/gve_rx_dqo.c
61
rx->dqo.free_buf_states = 0;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
617
if (rx->xsk_pool)
drivers/net/ethernet/google/gve/gve_rx_dqo.c
62
rx->dqo.recycled_buf_states.head = -1;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
622
tx_qid = gve_xdp_tx_queue_id(priv, rx->q_num);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
63
rx->dqo.recycled_buf_states.tail = -1;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
631
static void gve_xsk_done_dqo(struct gve_priv *priv, struct gve_rx_ring *rx,
drivers/net/ethernet/google/gve/gve_rx_dqo.c
64
rx->dqo.used_buf_states.head = -1;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
642
if (unlikely(gve_xdp_tx_dqo(priv, rx, xdp)))
drivers/net/ethernet/google/gve/gve_rx_dqo.c
65
rx->dqo.used_buf_states.tail = -1;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
651
u64_stats_update_begin(&rx->statss);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
653
rx->xdp_actions[xdp_act]++;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
654
u64_stats_update_end(&rx->statss);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
658
u64_stats_update_begin(&rx->statss);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
660
rx->xdp_tx_errors++;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
662
rx->xdp_redirect_errors++;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
663
u64_stats_update_end(&rx->statss);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
666
static void gve_xdp_done_dqo(struct gve_priv *priv, struct gve_rx_ring *rx,
drivers/net/ethernet/google/gve/gve_rx_dqo.c
676
gve_free_buffer(rx, buf_state);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
679
err = gve_xdp_tx_dqo(priv, rx, xdp);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
682
gve_reuse_buffer(rx, buf_state);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
688
gve_reuse_buffer(rx, buf_state);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
691
u64_stats_update_begin(&rx->statss);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
693
rx->xdp_actions[xdp_act]++;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
694
u64_stats_update_end(&rx->statss);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
697
u64_stats_update_begin(&rx->statss);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
699
rx->xdp_tx_errors++;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
70
struct gve_rx_ring *rx = &priv->rx[idx];
drivers/net/ethernet/google/gve/gve_rx_dqo.c
701
rx->xdp_redirect_errors++;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
702
u64_stats_update_end(&rx->statss);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
703
gve_free_buffer(rx, buf_state);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
707
static int gve_rx_xsk_dqo(struct napi_struct *napi, struct gve_rx_ring *rx,
drivers/net/ethernet/google/gve/gve_rx_dqo.c
714
struct gve_priv *priv = rx->gve;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
729
gve_xsk_done_dqo(priv, rx, xdp, xprog, xdp_act);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
730
gve_free_buf_state(rx, buf_state);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
736
rx->ctx.skb_head = gve_rx_copy_data(priv->dev, napi,
drivers/net/ethernet/google/gve/gve_rx_dqo.c
738
if (unlikely(!rx->ctx.skb_head)) {
drivers/net/ethernet/google/gve/gve_rx_dqo.c
740
gve_free_buf_state(rx, buf_state);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
743
rx->ctx.skb_tail = rx->ctx.skb_head;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
747
gve_free_buf_state(rx, buf_state);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
750
u64_stats_update_begin(&rx->statss);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
751
rx->xdp_actions[XDP_PASS]++;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
752
u64_stats_update_end(&rx->statss);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
756
static void gve_dma_sync(struct gve_priv *priv, struct gve_rx_ring *rx,
drivers/net/ethernet/google/gve/gve_rx_dqo.c
761
if (rx->dqo.page_pool) {
drivers/net/ethernet/google/gve/gve_rx_dqo.c
762
page_pool_dma_sync_netmem_for_cpu(rx->dqo.page_pool,
drivers/net/ethernet/google/gve/gve_rx_dqo.c
778
static int gve_rx_dqo(struct napi_struct *napi, struct gve_rx_ring *rx,
drivers/net/ethernet/google/gve/gve_rx_dqo.c
78
if (rx->dqo.bufq.desc_ring) {
drivers/net/ethernet/google/gve/gve_rx_dqo.c
787
struct gve_priv *priv = rx->gve;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
79
size = sizeof(rx->dqo.bufq.desc_ring[0]) *
drivers/net/ethernet/google/gve/gve_rx_dqo.c
792
if (unlikely(buffer_id >= rx->dqo.num_buf_states)) {
drivers/net/ethernet/google/gve/gve_rx_dqo.c
797
buf_state = &rx->dqo.buf_states[buffer_id];
drivers/net/ethernet/google/gve/gve_rx_dqo.c
798
if (unlikely(!gve_buf_state_is_allocated(rx, buf_state))) {
drivers/net/ethernet/google/gve/gve_rx_dqo.c
805
gve_free_buffer(rx, buf_state);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
81
memset(rx->dqo.bufq.desc_ring, 0, size);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
814
return gve_rx_xsk_dqo(napi, rx, compl_desc, buf_state, xprog);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
819
if (rx->dqo.page_pool) {
drivers/net/ethernet/google/gve/gve_rx_dqo.c
831
rx->ctx.skb_head = gve_rx_copy_data(priv->dev, napi,
drivers/net/ethernet/google/gve/gve_rx_dqo.c
832
rx->dqo.hdr_bufs.data +
drivers/net/ethernet/google/gve/gve_rx_dqo.c
835
if (unlikely(!rx->ctx.skb_head))
drivers/net/ethernet/google/gve/gve_rx_dqo.c
837
rx->ctx.skb_tail = rx->ctx.skb_head;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
839
if (rx->dqo.page_pool)
drivers/net/ethernet/google/gve/gve_rx_dqo.c
840
skb_mark_for_recycle(rx->ctx.skb_head);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
844
u64_stats_update_begin(&rx->statss);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
845
rx->rx_hsplit_pkt++;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
846
rx->rx_hsplit_unsplit_pkt += unsplit;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
847
rx->rx_hsplit_bytes += hdr_len;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
848
u64_stats_update_end(&rx->statss);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
849
} else if (!rx->ctx.skb_head && rx->dqo.page_pool &&
drivers/net/ethernet/google/gve/gve_rx_dqo.c
85
if (rx->dqo.complq.desc_ring) {
drivers/net/ethernet/google/gve/gve_rx_dqo.c
86
size = sizeof(rx->dqo.complq.desc_ring[0]) *
drivers/net/ethernet/google/gve/gve_rx_dqo.c
860
gve_dma_sync(priv, rx, buf_state, buf_len);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
863
if (rx->ctx.skb_head) {
drivers/net/ethernet/google/gve/gve_rx_dqo.c
864
if (unlikely(gve_rx_append_frags(napi, buf_state, buf_len, rx,
drivers/net/ethernet/google/gve/gve_rx_dqo.c
877
&rx->xdp_rxq);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
88
memset(rx->dqo.complq.desc_ring, 0, size);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
891
gve_xdp_done_dqo(priv, rx, &gve_xdp.xdp, xprog, xdp_act,
drivers/net/ethernet/google/gve/gve_rx_dqo.c
896
u64_stats_update_begin(&rx->statss);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
897
rx->xdp_actions[XDP_PASS]++;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
898
u64_stats_update_end(&rx->statss);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
902
!(rx->dqo.page_pool &&
drivers/net/ethernet/google/gve/gve_rx_dqo.c
904
rx->ctx.skb_head = gve_rx_copy(priv->dev, napi,
drivers/net/ethernet/google/gve/gve_rx_dqo.c
906
if (unlikely(!rx->ctx.skb_head))
drivers/net/ethernet/google/gve/gve_rx_dqo.c
908
rx->ctx.skb_tail = rx->ctx.skb_head;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
910
u64_stats_update_begin(&rx->statss);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
911
rx->rx_copied_pkt++;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
912
rx->rx_copybreak_pkt++;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
913
u64_stats_update_end(&rx->statss);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
915
gve_free_buffer(rx, buf_state);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
919
rx->ctx.skb_head = napi_get_frags(napi);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
92
if (rx->q_resources)
drivers/net/ethernet/google/gve/gve_rx_dqo.c
920
if (unlikely(!rx->ctx.skb_head))
drivers/net/ethernet/google/gve/gve_rx_dqo.c
922
rx->ctx.skb_tail = rx->ctx.skb_head;
drivers/net/ethernet/google/gve/gve_rx_dqo.c
924
if (gve_rx_should_trigger_copy_ondemand(rx)) {
drivers/net/ethernet/google/gve/gve_rx_dqo.c
925
if (gve_rx_copy_ondemand(rx, buf_state, buf_len) < 0)
drivers/net/ethernet/google/gve/gve_rx_dqo.c
93
memset(rx->q_resources, 0, sizeof(*rx->q_resources));
drivers/net/ethernet/google/gve/gve_rx_dqo.c
930
if (rx->dqo.page_pool)
drivers/net/ethernet/google/gve/gve_rx_dqo.c
931
skb_mark_for_recycle(rx->ctx.skb_head);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
933
gve_skb_add_rx_frag(rx, buf_state, 0, buf_len);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
934
gve_reuse_buffer(rx, buf_state);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
938
gve_free_buffer(rx, buf_state);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
96
if (rx->dqo.buf_states) {
drivers/net/ethernet/google/gve/gve_rx_dqo.c
968
static int gve_rx_complete_skb(struct gve_rx_ring *rx, struct napi_struct *napi,
drivers/net/ethernet/google/gve/gve_rx_dqo.c
97
for (i = 0; i < rx->dqo.num_buf_states; i++) {
drivers/net/ethernet/google/gve/gve_rx_dqo.c
973
rx->gve->ptype_lut_dqo->ptypes[desc->packet_type];
drivers/net/ethernet/google/gve/gve_rx_dqo.c
976
skb_record_rx_queue(rx->ctx.skb_head, rx->q_num);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
979
gve_rx_skb_hash(rx->ctx.skb_head, desc, ptype);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
98
struct gve_rx_buf_state_dqo *bs = &rx->dqo.buf_states[i];
drivers/net/ethernet/google/gve/gve_rx_dqo.c
982
gve_rx_skb_csum(rx->ctx.skb_head, desc, ptype);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
984
if (rx->gve->ts_config.rx_filter == HWTSTAMP_FILTER_ALL)
drivers/net/ethernet/google/gve/gve_rx_dqo.c
985
gve_rx_skb_hwtstamp(rx, desc);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
991
err = gve_rx_complete_rsc(rx->ctx.skb_head, desc, ptype);
drivers/net/ethernet/google/gve/gve_rx_dqo.c
996
if (skb_headlen(rx->ctx.skb_head) == 0)
drivers/net/ethernet/google/gve/gve_rx_dqo.c
999
napi_gro_receive(napi, rx->ctx.skb_head);
drivers/net/ethernet/google/gve/gve_tx.c
976
struct gve_rx_ring *rx = rx_block->rx;
drivers/net/ethernet/google/gve/gve_tx.c
977
struct gve_priv *priv = rx->gve;
drivers/net/ethernet/google/gve/gve_tx.c
981
tx = &priv->tx[gve_xdp_tx_queue_id(priv, rx->q_num)];
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1499
struct gve_rx_ring *rx = rx_block->rx;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1500
struct gve_priv *priv = rx->gve;
drivers/net/ethernet/google/gve/gve_tx_dqo.c
1503
tx = &priv->tx[gve_xdp_tx_queue_id(priv, rx->q_num)];
drivers/net/ethernet/google/gve/gve_utils.c
46
return block->rx != NULL;
drivers/net/ethernet/google/gve/gve_utils.c
54
block->rx = NULL;
drivers/net/ethernet/google/gve/gve_utils.c
61
struct gve_rx_ring *rx = &priv->rx[queue_idx];
drivers/net/ethernet/google/gve/gve_utils.c
63
block->rx = rx;
drivers/net/ethernet/google/gve/gve_utils.c
64
rx->ntfy_id = ntfy_idx;
drivers/net/ethernet/hisilicon/hip04_eth.c
584
int rx = 0;
drivers/net/ethernet/hisilicon/hip04_eth.c
622
rx++;
drivers/net/ethernet/hisilicon/hip04_eth.c
638
if (rx >= budget) {
drivers/net/ethernet/hisilicon/hip04_eth.c
652
napi_complete_done(napi, rx);
drivers/net/ethernet/hisilicon/hip04_eth.c
655
if (rx < budget && tx_remaining)
drivers/net/ethernet/hisilicon/hip04_eth.c
658
return rx;
drivers/net/ethernet/hisilicon/hix5hd2_gmac.c
331
static void hix5hd2_set_desc_depth(struct hix5hd2_priv *priv, int rx, int tx)
drivers/net/ethernet/hisilicon/hix5hd2_gmac.c
334
writel_relaxed(rx << 3, priv->base + RX_FQ_DEPTH);
drivers/net/ethernet/hisilicon/hix5hd2_gmac.c
338
writel_relaxed(rx << 3, priv->base + RX_BQ_DEPTH);
drivers/net/ethernet/hisilicon/hns/hnae.h
209
} rx;
drivers/net/ethernet/hisilicon/hns/hnae.h
647
ring->desc[i].rx.ipoff_bnum_pid_flag = 0;
drivers/net/ethernet/hisilicon/hns/hnae.h
655
ring->desc[i].rx.ipoff_bnum_pid_flag = 0;
drivers/net/ethernet/hisilicon/hns/hns_dsaf_gmac.c
103
*rx = dsaf_get_bit(porten, GMAC_PORT_RX_EN_B);
drivers/net/ethernet/hisilicon/hns/hns_dsaf_gmac.c
494
u32 rx;
drivers/net/ethernet/hisilicon/hns/hns_dsaf_gmac.c
506
hns_gmac_get_en(mac_drv, &rx, &tx);
drivers/net/ethernet/hisilicon/hns/hns_dsaf_gmac.c
507
mac_info->port_en = rx && tx;
drivers/net/ethernet/hisilicon/hns/hns_dsaf_gmac.c
96
static void hns_gmac_get_en(void *mac_drv, u32 *rx, u32 *tx)
drivers/net/ethernet/hisilicon/hns/hns_enet.c
446
size = le16_to_cpu(desc->rx.size);
drivers/net/ethernet/hisilicon/hns/hns_enet.c
594
length = le16_to_cpu(desc->rx.pkt_len);
drivers/net/ethernet/hisilicon/hns/hns_enet.c
595
bnum_flag = le32_to_cpu(desc->rx.ipoff_bnum_pid_flag);
drivers/net/ethernet/hisilicon/hns/hns_enet.c
650
bnum_flag = le32_to_cpu(desc->rx.ipoff_bnum_pid_flag);
drivers/net/ethernet/hisilicon/hns/hns_enet.c
660
if (unlikely((!desc->rx.pkt_len) ||
drivers/net/ethernet/hisilicon/hns3/hns3_debugfs.c
617
idx, le32_to_cpu(desc->rx.l234_info),
drivers/net/ethernet/hisilicon/hns3/hns3_debugfs.c
618
le16_to_cpu(desc->rx.pkt_len), le16_to_cpu(desc->rx.size),
drivers/net/ethernet/hisilicon/hns3/hns3_debugfs.c
619
le32_to_cpu(desc->rx.rss_hash), le16_to_cpu(desc->rx.fd_id),
drivers/net/ethernet/hisilicon/hns3/hns3_debugfs.c
620
le16_to_cpu(desc->rx.vlan_tag),
drivers/net/ethernet/hisilicon/hns3/hns3_debugfs.c
621
le16_to_cpu(desc->rx.o_dm_vlan_id_fb),
drivers/net/ethernet/hisilicon/hns3/hns3_debugfs.c
622
le16_to_cpu(desc->rx.ot_vlan_tag),
drivers/net/ethernet/hisilicon/hns3/hns3_debugfs.c
623
le32_to_cpu(desc->rx.bd_base_info));
drivers/net/ethernet/hisilicon/hns3/hns3_debugfs.c
626
u32 ol_info = le32_to_cpu(desc->rx.ol_info);
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
3609
ring->desc[i].rx.bd_base_info = 0;
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
3618
ring->desc[i].rx.bd_base_info = 0;
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
3778
int size = le16_to_cpu(desc->rx.size);
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
3805
int size = le16_to_cpu(desc->rx.size);
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
4025
*vlan_tag = le16_to_cpu(desc->rx.ot_vlan_tag);
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
4027
*vlan_tag = le16_to_cpu(desc->rx.vlan_tag);
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
4047
*vlan_tag = le16_to_cpu(desc->rx.ot_vlan_tag);
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
4054
*vlan_tag = le16_to_cpu(desc->rx.vlan_tag);
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
4059
*vlan_tag = le16_to_cpu(desc->rx.ot_vlan_tag);
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
4061
*vlan_tag = le16_to_cpu(desc->rx.vlan_tag);
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
4071
ring->desc[ring->next_to_clean].rx.bd_base_info &=
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
4145
bd_base_info = le32_to_cpu(desc->rx.bd_base_info);
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
4175
head_skb->data_len += le16_to_cpu(desc->rx.size);
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
4176
head_skb->len += le16_to_cpu(desc->rx.size);
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
4320
bd_base_info = le32_to_cpu(desc->rx.bd_base_info);
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
4321
l234info = le32_to_cpu(desc->rx.l234_info);
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
4322
ol_info = le32_to_cpu(desc->rx.ol_info);
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
4329
if (unlikely(!desc->rx.pkt_len || (l234info & (BIT(HNS3_RXD_TRUNCAT_B) |
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
4368
hns3_set_rx_skb_rss_type(ring, skb, le32_to_cpu(desc->rx.rss_hash),
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
4388
bd_base_info = le32_to_cpu(desc->rx.bd_base_info);
drivers/net/ethernet/hisilicon/hns3/hns3_enet.c
4394
length = le16_to_cpu(desc->rx.size);
drivers/net/ethernet/hisilicon/hns3/hns3_enet.h
313
} rx;
drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_tm.c
162
int hclge_mac_pause_en_cfg(struct hclge_dev *hdev, bool tx, bool rx)
drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_tm.c
169
(rx ? HCLGE_RX_MAC_PAUSE_EN_MSK : 0));
drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_tm.h
246
int hclge_mac_pause_en_cfg(struct hclge_dev *hdev, bool tx, bool rx);
drivers/net/ethernet/ibm/ehea/ehea_main.c
885
int rx = 0;
drivers/net/ethernet/ibm/ehea/ehea_main.c
888
rx += ehea_proc_rwqes(dev, pr, budget - rx);
drivers/net/ethernet/ibm/ehea/ehea_main.c
890
while (rx != budget) {
drivers/net/ethernet/ibm/ehea/ehea_main.c
901
return rx;
drivers/net/ethernet/ibm/ehea/ehea_main.c
904
return rx;
drivers/net/ethernet/ibm/ehea/ehea_main.c
907
rx += ehea_proc_rwqes(dev, pr, budget - rx);
drivers/net/ethernet/ibm/ehea/ehea_main.c
910
return rx;
drivers/net/ethernet/intel/e100.c
1900
static inline void e100_start_receiver(struct nic *nic, struct rx *rx)
drivers/net/ethernet/intel/e100.c
1906
if (!rx) rx = nic->rxs;
drivers/net/ethernet/intel/e100.c
1909
if (rx->skb) {
drivers/net/ethernet/intel/e100.c
1910
e100_exec_cmd(nic, ruc_start, rx->dma_addr);
drivers/net/ethernet/intel/e100.c
1916
static int e100_rx_alloc_skb(struct nic *nic, struct rx *rx)
drivers/net/ethernet/intel/e100.c
1918
if (!(rx->skb = netdev_alloc_skb_ip_align(nic->netdev, RFD_BUF_LEN)))
drivers/net/ethernet/intel/e100.c
1922
skb_copy_to_linear_data(rx->skb, &nic->blank_rfd, sizeof(struct rfd));
drivers/net/ethernet/intel/e100.c
1923
rx->dma_addr = dma_map_single(&nic->pdev->dev, rx->skb->data,
drivers/net/ethernet/intel/e100.c
1926
if (dma_mapping_error(&nic->pdev->dev, rx->dma_addr)) {
drivers/net/ethernet/intel/e100.c
1927
dev_kfree_skb_any(rx->skb);
drivers/net/ethernet/intel/e100.c
1928
rx->skb = NULL;
drivers/net/ethernet/intel/e100.c
1929
rx->dma_addr = 0;
drivers/net/ethernet/intel/e100.c
1936
if (rx->prev->skb) {
drivers/net/ethernet/intel/e100.c
1937
struct rfd *prev_rfd = (struct rfd *)rx->prev->skb->data;
drivers/net/ethernet/intel/e100.c
1938
put_unaligned_le32(rx->dma_addr, &prev_rfd->link);
drivers/net/ethernet/intel/e100.c
1940
rx->prev->dma_addr,
drivers/net/ethernet/intel/e100.c
1948
static int e100_rx_indicate(struct nic *nic, struct rx *rx,
drivers/net/ethernet/intel/e100.c
1952
struct sk_buff *skb = rx->skb;
drivers/net/ethernet/intel/e100.c
1961
dma_sync_single_for_cpu(&nic->pdev->dev, rx->dma_addr,
drivers/net/ethernet/intel/e100.c
1981
dma_sync_single_for_device(&nic->pdev->dev, rx->dma_addr,
drivers/net/ethernet/intel/e100.c
1995
dma_unmap_single(&nic->pdev->dev, rx->dma_addr, RFD_BUF_LEN,
drivers/net/ethernet/intel/e100.c
2042
rx->skb = NULL;
drivers/net/ethernet/intel/e100.c
2050
struct rx *rx;
drivers/net/ethernet/intel/e100.c
2052
struct rx *old_before_last_rx, *new_before_last_rx;
drivers/net/ethernet/intel/e100.c
2056
for (rx = nic->rx_to_clean; rx->skb; rx = nic->rx_to_clean = rx->next) {
drivers/net/ethernet/intel/e100.c
2057
err = e100_rx_indicate(nic, rx, work_done, work_to_do);
drivers/net/ethernet/intel/e100.c
2077
for (rx = nic->rx_to_use; !rx->skb; rx = nic->rx_to_use = rx->next) {
drivers/net/ethernet/intel/e100.c
2078
if (unlikely(e100_rx_alloc_skb(nic, rx)))
drivers/net/ethernet/intel/e100.c
2129
struct rx *rx;
drivers/net/ethernet/intel/e100.c
2135
for (rx = nic->rxs, i = 0; i < count; rx++, i++) {
drivers/net/ethernet/intel/e100.c
2136
if (rx->skb) {
drivers/net/ethernet/intel/e100.c
2138
rx->dma_addr, RFD_BUF_LEN,
drivers/net/ethernet/intel/e100.c
2140
dev_kfree_skb(rx->skb);
drivers/net/ethernet/intel/e100.c
2152
struct rx *rx;
drivers/net/ethernet/intel/e100.c
2159
if (!(nic->rxs = kzalloc_objs(struct rx, count)))
drivers/net/ethernet/intel/e100.c
2162
for (rx = nic->rxs, i = 0; i < count; rx++, i++) {
drivers/net/ethernet/intel/e100.c
2163
rx->next = (i + 1 < count) ? rx + 1 : nic->rxs;
drivers/net/ethernet/intel/e100.c
2164
rx->prev = (i == 0) ? nic->rxs + count - 1 : rx - 1;
drivers/net/ethernet/intel/e100.c
2165
if (e100_rx_alloc_skb(nic, rx)) {
drivers/net/ethernet/intel/e100.c
2177
rx = nic->rxs->prev->prev;
drivers/net/ethernet/intel/e100.c
2178
before_last = (struct rfd *)rx->skb->data;
drivers/net/ethernet/intel/e100.c
2181
dma_sync_single_for_device(&nic->pdev->dev, rx->dma_addr,
drivers/net/ethernet/intel/e100.c
416
struct rx *next, *prev;
drivers/net/ethernet/intel/e100.c
543
struct rx *rxs ____cacheline_aligned;
drivers/net/ethernet/intel/e100.c
544
struct rx *rx_to_use;
drivers/net/ethernet/intel/e100.c
545
struct rx *rx_to_clean;
drivers/net/ethernet/intel/fm10k/fm10k.h
189
struct fm10k_ring_container rx, tx;
drivers/net/ethernet/intel/fm10k/fm10k_debugfs.c
120
if (ring < q_vector->rx.ring)
drivers/net/ethernet/intel/fm10k/fm10k_debugfs.c
176
for (i = 0; i < q_vector->rx.count; i++) {
drivers/net/ethernet/intel/fm10k/fm10k_debugfs.c
177
struct fm10k_ring *ring = &q_vector->rx.ring[i];
drivers/net/ethernet/intel/fm10k/fm10k_ethtool.c
688
qv->rx.itr = rx_itr;
drivers/net/ethernet/intel/fm10k/fm10k_main.c
1408
fm10k_update_itr(&q_vector->rx);
drivers/net/ethernet/intel/fm10k/fm10k_main.c
1414
itr |= (q_vector->rx.itr & FM10K_ITR_MAX) << FM10K_ITR_INTERVAL1_SHIFT;
drivers/net/ethernet/intel/fm10k/fm10k_main.c
1440
if (q_vector->rx.count > 1)
drivers/net/ethernet/intel/fm10k/fm10k_main.c
1441
per_ring_budget = max(budget / q_vector->rx.count, 1);
drivers/net/ethernet/intel/fm10k/fm10k_main.c
1445
fm10k_for_each_ring(ring, q_vector->rx) {
drivers/net/ethernet/intel/fm10k/fm10k_main.c
1646
q_vector->rx.ring = ring;
drivers/net/ethernet/intel/fm10k/fm10k_main.c
1647
q_vector->rx.itr = interface->rx_itr;
drivers/net/ethernet/intel/fm10k/fm10k_main.c
1648
q_vector->rx.itr_scale = interface->hw.mac.itr_scale;
drivers/net/ethernet/intel/fm10k/fm10k_main.c
1649
q_vector->rx.count = rxr_count;
drivers/net/ethernet/intel/fm10k/fm10k_main.c
1699
fm10k_for_each_ring(ring, q_vector->rx)
drivers/net/ethernet/intel/fm10k/fm10k_main.c
627
q_vector->rx.total_packets += total_packets;
drivers/net/ethernet/intel/fm10k/fm10k_main.c
628
q_vector->rx.total_bytes += total_bytes;
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1063
mbx->max_size = mbx->rx.size - 1;
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1085
mbx->rx.tail = 0;
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1086
mbx->rx.head = 0;
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1166
if (size > mbx->rx.size) {
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1167
mbx->max_size = mbx->rx.size - 1;
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1439
if (!mbx->rx.buffer)
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1609
fm10k_fifo_init(&mbx->rx, &mbx->buffer[FM10K_MBX_TX_BUFFER_SIZE],
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
1697
if (!mbx->rx.buffer)
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
216
u32 *tail = mbx->rx.buffer + fm10k_fifo_tail_offset(&mbx->rx, 0);
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
2168
fm10k_fifo_init(&mbx->rx, &mbx->buffer[FM10K_MBX_TX_BUFFER_SIZE],
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
298
struct fm10k_mbx_fifo *fifo = &mbx->rx;
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
426
struct fm10k_mbx_fifo *fifo = &mbx->rx;
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
475
struct fm10k_mbx_fifo *fifo = &mbx->rx;
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
640
struct fm10k_mbx_fifo *fifo = &mbx->rx;
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
664
u16 msg_size = fm10k_fifo_head_len(&mbx->rx);
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
666
return msg_size && (fm10k_fifo_used(&mbx->rx) >= msg_size);
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
705
struct fm10k_mbx_fifo *fifo = &mbx->rx;
drivers/net/ethernet/intel/fm10k/fm10k_mbx.c
842
FM10K_MSG_HDR_FIELD_SET(mbx->rx.size - 1, CONNECT_SIZE);
drivers/net/ethernet/intel/fm10k/fm10k_mbx.h
256
struct fm10k_mbx_fifo rx;
drivers/net/ethernet/intel/fm10k/fm10k_pci.c
1187
if (q_vector->rx.count || q_vector->tx.count)
drivers/net/ethernet/intel/fm10k/fm10k_pci.c
1734
if (!q_vector->tx.count && !q_vector->rx.count)
drivers/net/ethernet/intel/fm10k/fm10k_pci.c
1768
if (q_vector->tx.count && q_vector->rx.count) {
drivers/net/ethernet/intel/fm10k/fm10k_pci.c
1772
} else if (q_vector->rx.count) {
drivers/net/ethernet/intel/fm10k/fm10k_pci.c
1818
if (!q_vector->tx.count && !q_vector->rx.count)
drivers/net/ethernet/intel/fm10k/fm10k_pci.c
722
if (!qv->tx.count && !qv->rx.count)
drivers/net/ethernet/intel/i40e/i40e.h
955
struct i40e_ring_container rx;
drivers/net/ethernet/intel/i40e/i40e_ethtool.c
2988
q_vector->rx.target_itr = ITR_TO_REG(rx_ring->itr_setting);
drivers/net/ethernet/intel/i40e/i40e_main.c
13390
if (q_vector->rx.ring || q_vector->tx.ring) {
drivers/net/ethernet/intel/i40e/i40e_main.c
3894
q_vector->rx.next_update = jiffies + 1;
drivers/net/ethernet/intel/i40e/i40e_main.c
3895
q_vector->rx.target_itr =
drivers/net/ethernet/intel/i40e/i40e_main.c
3898
q_vector->rx.target_itr >> 1);
drivers/net/ethernet/intel/i40e/i40e_main.c
3899
q_vector->rx.current_itr = q_vector->rx.target_itr;
drivers/net/ethernet/intel/i40e/i40e_main.c
4014
q_vector->rx.next_update = jiffies + 1;
drivers/net/ethernet/intel/i40e/i40e_main.c
4015
q_vector->rx.target_itr = ITR_TO_REG(vsi->rx_rings[0]->itr_setting);
drivers/net/ethernet/intel/i40e/i40e_main.c
4016
wr32(hw, I40E_PFINT_ITR0(I40E_RX_ITR), q_vector->rx.target_itr >> 1);
drivers/net/ethernet/intel/i40e/i40e_main.c
4017
q_vector->rx.current_itr = q_vector->rx.target_itr;
drivers/net/ethernet/intel/i40e/i40e_main.c
4084
if (!q_vector->tx.ring && !q_vector->rx.ring)
drivers/net/ethernet/intel/i40e/i40e_main.c
4142
if (q_vector->tx.ring && q_vector->rx.ring) {
drivers/net/ethernet/intel/i40e/i40e_main.c
4146
} else if (q_vector->rx.ring) {
drivers/net/ethernet/intel/i40e/i40e_main.c
4552
rx_ring->next = q_vector->rx.ring;
drivers/net/ethernet/intel/i40e/i40e_main.c
4553
q_vector->rx.ring = rx_ring;
drivers/net/ethernet/intel/i40e/i40e_main.c
4554
q_vector->rx.count++;
drivers/net/ethernet/intel/i40e/i40e_main.c
4589
q_vector->rx.count = 0;
drivers/net/ethernet/intel/i40e/i40e_main.c
4591
q_vector->rx.ring = NULL;
drivers/net/ethernet/intel/i40e/i40e_main.c
5109
i40e_for_each_ring(ring, q_vector->rx)
drivers/net/ethernet/intel/i40e/i40e_main.c
5196
if (q_vector->rx.ring || q_vector->tx.ring)
drivers/net/ethernet/intel/i40e/i40e_main.c
5215
if (q_vector->rx.ring || q_vector->tx.ring)
drivers/net/ethernet/intel/i40e/i40e_txrx.c
1116
return &q_vector->rx == rc;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
1216
(q_vector->rx.target_itr & I40E_ITR_MASK) ==
drivers/net/ethernet/intel/i40e/i40e_txrx.c
1244
itr = min(q_vector->tx.current_itr, q_vector->rx.current_itr);
drivers/net/ethernet/intel/i40e/i40e_txrx.c
2336
rx_ring->q_vector->rx.total_packets += total_rx_packets;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
2337
rx_ring->q_vector->rx.total_bytes += total_rx_bytes;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
2680
i40e_update_itr(q_vector, &q_vector->rx);
drivers/net/ethernet/intel/i40e/i40e_txrx.c
2690
if (q_vector->rx.target_itr < q_vector->rx.current_itr) {
drivers/net/ethernet/intel/i40e/i40e_txrx.c
2693
interval = q_vector->rx.target_itr;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
2694
q_vector->rx.current_itr = q_vector->rx.target_itr;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
2697
((q_vector->rx.target_itr - q_vector->rx.current_itr) <
drivers/net/ethernet/intel/i40e/i40e_txrx.c
2706
} else if (q_vector->rx.current_itr != q_vector->rx.target_itr) {
drivers/net/ethernet/intel/i40e/i40e_txrx.c
2709
interval = q_vector->rx.target_itr;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
2710
q_vector->rx.current_itr = q_vector->rx.target_itr;
drivers/net/ethernet/intel/i40e/i40e_txrx.c
2794
i40e_for_each_ring(ring, q_vector->rx) {
drivers/net/ethernet/intel/iavf/iavf.h
108
struct iavf_ring_container rx;
drivers/net/ethernet/intel/iavf/iavf_ethtool.c
664
q_vector->rx.target_itr = ITR_TO_REG(rx_ring->itr_setting);
drivers/net/ethernet/intel/iavf/iavf_main.c
414
if (!q_vector->tx.ring && !q_vector->rx.ring)
drivers/net/ethernet/intel/iavf/iavf_main.c
436
rx_ring->next = q_vector->rx.ring;
drivers/net/ethernet/intel/iavf/iavf_main.c
438
q_vector->rx.ring = rx_ring;
drivers/net/ethernet/intel/iavf/iavf_main.c
439
q_vector->rx.count++;
drivers/net/ethernet/intel/iavf/iavf_main.c
440
q_vector->rx.next_update = jiffies + 1;
drivers/net/ethernet/intel/iavf/iavf_main.c
441
q_vector->rx.target_itr = ITR_TO_REG(rx_ring->itr_setting);
drivers/net/ethernet/intel/iavf/iavf_main.c
444
q_vector->rx.current_itr >> 1);
drivers/net/ethernet/intel/iavf/iavf_main.c
445
q_vector->rx.current_itr = q_vector->rx.target_itr;
drivers/net/ethernet/intel/iavf/iavf_main.c
530
if (q_vector->tx.ring && q_vector->rx.ring) {
drivers/net/ethernet/intel/iavf/iavf_main.c
534
} else if (q_vector->rx.ring) {
drivers/net/ethernet/intel/iavf/iavf_txrx.c
1483
rx_ring->q_vector->rx.total_packets += total_rx_packets;
drivers/net/ethernet/intel/iavf/iavf_txrx.c
1484
rx_ring->q_vector->rx.total_bytes += total_rx_bytes;
drivers/net/ethernet/intel/iavf/iavf_txrx.c
1544
iavf_update_itr(q_vector, &q_vector->rx);
drivers/net/ethernet/intel/iavf/iavf_txrx.c
1554
if (q_vector->rx.target_itr < q_vector->rx.current_itr) {
drivers/net/ethernet/intel/iavf/iavf_txrx.c
1557
q_vector->rx.target_itr);
drivers/net/ethernet/intel/iavf/iavf_txrx.c
1558
q_vector->rx.current_itr = q_vector->rx.target_itr;
drivers/net/ethernet/intel/iavf/iavf_txrx.c
1561
((q_vector->rx.target_itr - q_vector->rx.current_itr) <
drivers/net/ethernet/intel/iavf/iavf_txrx.c
1570
} else if (q_vector->rx.current_itr != q_vector->rx.target_itr) {
drivers/net/ethernet/intel/iavf/iavf_txrx.c
1573
q_vector->rx.target_itr);
drivers/net/ethernet/intel/iavf/iavf_txrx.c
1574
q_vector->rx.current_itr = q_vector->rx.target_itr;
drivers/net/ethernet/intel/iavf/iavf_txrx.c
1633
iavf_for_each_ring(ring, q_vector->rx) {
drivers/net/ethernet/intel/iavf/iavf_txrx.c
398
return &q_vector->rx == rc;
drivers/net/ethernet/intel/iavf/iavf_txrx.c
529
(q_vector->rx.target_itr & IAVF_ITR_MASK) ==
drivers/net/ethernet/intel/iavf/iavf_txrx.c
557
itr = min(q_vector->tx.current_itr, q_vector->rx.current_itr);
drivers/net/ethernet/intel/ice/ice.h
483
struct ice_ring_container rx;
drivers/net/ethernet/intel/ice/ice_base.c
117
q_vector->rx.itr_setting = ICE_DFLT_RX_ITR;
drivers/net/ethernet/intel/ice/ice_base.c
119
q_vector->rx.itr_mode = ITR_DYNAMIC;
drivers/net/ethernet/intel/ice/ice_base.c
121
q_vector->rx.type = ICE_RX_CONTAINER;
drivers/net/ethernet/intel/ice/ice_base.c
1219
ice_write_itr(&q_vector->rx, q_vector->rx.itr_setting);
drivers/net/ethernet/intel/ice/ice_base.c
196
ice_for_each_rx_ring(rx_ring, vsi->q_vectors[v_idx]->rx)
drivers/net/ethernet/intel/ice/ice_base.c
964
q_vector->rx.rx_ring = NULL;
drivers/net/ethernet/intel/ice/ice_base.c
965
q_vector->rx.itr_idx = ICE_RX_ITR;
drivers/net/ethernet/intel/ice/ice_base.c
972
rx_ring->next = q_vector->rx.rx_ring;
drivers/net/ethernet/intel/ice/ice_base.c
973
q_vector->rx.rx_ring = rx_ring;
drivers/net/ethernet/intel/ice/ice_ethtool.c
4079
&vsi->rx_rings[q_num]->q_vector->rx))
drivers/net/ethernet/intel/ice/ice_ethtool.c
4086
&vsi->rx_rings[q_num]->q_vector->rx))
drivers/net/ethernet/intel/ice/ice_ethtool.c
4247
&vsi->rx_rings[q_num]->q_vector->rx,
drivers/net/ethernet/intel/ice/ice_ethtool.c
4257
&vsi->rx_rings[q_num]->q_vector->rx,
drivers/net/ethernet/intel/ice/ice_ethtool.c
4315
ice_print_if_odd_usecs(netdev, q_vector->rx.itr_setting,
drivers/net/ethernet/intel/ice/ice_ethtool.c
698
static const int rx = ICE_AQC_OP_CODE_RX_EQU;
drivers/net/ethernet/intel/ice/ice_ethtool.c
709
{ ICE_AQC_RX_EQU_PRE2, rx, &ptr->rx_equ_pre2 },
drivers/net/ethernet/intel/ice/ice_ethtool.c
710
{ ICE_AQC_RX_EQU_PRE1, rx, &ptr->rx_equ_pre1 },
drivers/net/ethernet/intel/ice/ice_ethtool.c
711
{ ICE_AQC_RX_EQU_POST1, rx, &ptr->rx_equ_post1 },
drivers/net/ethernet/intel/ice/ice_ethtool.c
712
{ ICE_AQC_RX_EQU_BFLF, rx, &ptr->rx_equ_bflf },
drivers/net/ethernet/intel/ice/ice_ethtool.c
713
{ ICE_AQC_RX_EQU_BFHF, rx, &ptr->rx_equ_bfhf },
drivers/net/ethernet/intel/ice/ice_ethtool.c
714
{ ICE_AQC_RX_EQU_CTLE_GAINHF, rx, &ptr->rx_equ_ctle_gainhf },
drivers/net/ethernet/intel/ice/ice_ethtool.c
715
{ ICE_AQC_RX_EQU_CTLE_GAINLF, rx, &ptr->rx_equ_ctle_gainlf },
drivers/net/ethernet/intel/ice/ice_ethtool.c
716
{ ICE_AQC_RX_EQU_CTLE_GAINDC, rx, &ptr->rx_equ_ctle_gaindc },
drivers/net/ethernet/intel/ice/ice_ethtool.c
717
{ ICE_AQC_RX_EQU_CTLE_BW, rx, &ptr->rx_equ_ctle_bw },
drivers/net/ethernet/intel/ice/ice_ethtool.c
718
{ ICE_AQC_RX_EQU_DFE_GAIN, rx, &ptr->rx_equ_dfe_gain },
drivers/net/ethernet/intel/ice/ice_ethtool.c
719
{ ICE_AQC_RX_EQU_DFE_GAIN2, rx, &ptr->rx_equ_dfe_gain_2 },
drivers/net/ethernet/intel/ice/ice_ethtool.c
720
{ ICE_AQC_RX_EQU_DFE_2, rx, &ptr->rx_equ_dfe_2 },
drivers/net/ethernet/intel/ice/ice_ethtool.c
721
{ ICE_AQC_RX_EQU_DFE_3, rx, &ptr->rx_equ_dfe_3 },
drivers/net/ethernet/intel/ice/ice_ethtool.c
722
{ ICE_AQC_RX_EQU_DFE_4, rx, &ptr->rx_equ_dfe_4 },
drivers/net/ethernet/intel/ice/ice_ethtool.c
723
{ ICE_AQC_RX_EQU_DFE_5, rx, &ptr->rx_equ_dfe_5 },
drivers/net/ethernet/intel/ice/ice_ethtool.c
724
{ ICE_AQC_RX_EQU_DFE_6, rx, &ptr->rx_equ_dfe_6 },
drivers/net/ethernet/intel/ice/ice_ethtool.c
725
{ ICE_AQC_RX_EQU_DFE_7, rx, &ptr->rx_equ_dfe_7 },
drivers/net/ethernet/intel/ice/ice_ethtool.c
726
{ ICE_AQC_RX_EQU_DFE_8, rx, &ptr->rx_equ_dfe_8 },
drivers/net/ethernet/intel/ice/ice_ethtool.c
727
{ ICE_AQC_RX_EQU_DFE_9, rx, &ptr->rx_equ_dfe_9 },
drivers/net/ethernet/intel/ice/ice_ethtool.c
728
{ ICE_AQC_RX_EQU_DFE_10, rx, &ptr->rx_equ_dfe_10 },
drivers/net/ethernet/intel/ice/ice_ethtool.c
729
{ ICE_AQC_RX_EQU_DFE_11, rx, &ptr->rx_equ_dfe_11 },
drivers/net/ethernet/intel/ice/ice_ethtool.c
730
{ ICE_AQC_RX_EQU_DFE_12, rx, &ptr->rx_equ_dfe_12 },
drivers/net/ethernet/intel/ice/ice_lib.c
1362
q_vector->rx.rx_ring = NULL;
drivers/net/ethernet/intel/ice/ice_lib.c
1895
if (ITR_IS_DYNAMIC(&q_vector->tx) || ITR_IS_DYNAMIC(&q_vector->rx)) {
drivers/net/ethernet/intel/ice/ice_lib.c
1947
q_vector->rx.itr_idx);
drivers/net/ethernet/intel/ice/ice_lib.c
2625
ice_write_itr(&q_vector->rx, 0);
drivers/net/ethernet/intel/ice/ice_lib.c
2917
coalesce[i].itr_rx = q_vector->rx.itr_settings;
drivers/net/ethernet/intel/ice/ice_lib.c
2972
rc = &vsi->q_vectors[i]->rx;
drivers/net/ethernet/intel/ice/ice_lib.c
2976
rc = &vsi->q_vectors[i]->rx;
drivers/net/ethernet/intel/ice/ice_lib.c
3005
rc = &vsi->q_vectors[i]->rx;
drivers/net/ethernet/intel/ice/ice_lib.c
491
ice_clean_ctrl_rx_irq(q_vector->rx.rx_ring);
drivers/net/ethernet/intel/ice/ice_lib.c
506
if (!q_vector->tx.tx_ring && !q_vector->rx.rx_ring)
drivers/net/ethernet/intel/ice/ice_main.c
2549
if (q_vector->tx.tx_ring && q_vector->rx.rx_ring) {
drivers/net/ethernet/intel/ice/ice_main.c
2553
} else if (q_vector->rx.rx_ring) {
drivers/net/ethernet/intel/ice/ice_main.c
6702
ice_trace(rx_dim_work, container_of(rc, struct ice_q_vector, rx), dim);
drivers/net/ethernet/intel/ice/ice_main.c
6736
rc = &q_vector->rx;
drivers/net/ethernet/intel/ice/ice_main.c
6766
if (q_vector->rx.rx_ring || q_vector->tx.tx_ring)
drivers/net/ethernet/intel/ice/ice_main.c
7265
if (q_vector->rx.rx_ring || q_vector->tx.tx_ring)
drivers/net/ethernet/intel/ice/ice_main.c
7269
cancel_work_sync(&q_vector->rx.dim.work);
drivers/net/ethernet/intel/ice/ice_main.c
8716
rc = &rx_q_vector->rx;
drivers/net/ethernet/intel/ice/ice_trace.h
68
__string(devname, q_vector->rx.rx_ring->netdev->name)),
drivers/net/ethernet/intel/ice/ice_trace.h
76
__entry->q_vector->rx.rx_ring->q_index,
drivers/net/ethernet/intel/ice/ice_txrx.c
1144
struct ice_ring_container *rx = &q_vector->rx;
drivers/net/ethernet/intel/ice/ice_txrx.c
1153
if (ITR_IS_DYNAMIC(rx)) {
drivers/net/ethernet/intel/ice/ice_txrx.c
1156
__ice_update_sample(q_vector, rx, &dim_sample, false);
drivers/net/ethernet/intel/ice/ice_txrx.c
1157
net_dim(&rx->dim, &dim_sample);
drivers/net/ethernet/intel/ice/ice_txrx.c
1309
ice_for_each_rx_ring(rx_ring, q_vector->rx) {
drivers/net/ethernet/intel/ice/ice_txrx.h
146
struct_group(rx,
drivers/net/ethernet/intel/ice/ice_xsk.c
95
ice_cfg_rxq_interrupt(vsi, _qid, reg_idx, q_vector->rx.itr_idx);
drivers/net/ethernet/intel/ice/virt/queues.c
481
q_vector->rx.itr_idx = map->rxitr_idx;
drivers/net/ethernet/intel/ice/virt/queues.c
485
q_vector->rx.itr_idx);
drivers/net/ethernet/intel/idpf/idpf_singleq_txrx.c
1050
rx_desc = &rx_q->rx[ntc];
drivers/net/ethernet/intel/idpf/idpf_singleq_txrx.c
1128
struct idpf_rx_queue *rxq = q_vec->rx[i];
drivers/net/ethernet/intel/idpf/idpf_txrx.c
1142
qs->qs[num++].rxq = qv->rx[i];
drivers/net/ethernet/intel/idpf/idpf_txrx.c
1160
u32 idx = xdp_txq_offset + qv->rx[i]->idx;
drivers/net/ethernet/intel/idpf/idpf_txrx.c
3551
rx_desc = &rxq->rx[ntc].flex_adv_nic_3_wb;
drivers/net/ethernet/intel/idpf/idpf_txrx.c
3833
kfree(q_vector->rx);
drivers/net/ethernet/intel/idpf/idpf_txrx.c
3834
q_vector->rx = NULL;
drivers/net/ethernet/intel/idpf/idpf_txrx.c
3847
netif_queue_set_napi(dev, q_vector->rx[i]->idx,
drivers/net/ethernet/intel/idpf/idpf_txrx.c
3996
struct idpf_rx_queue *rxq = q_vector->rx[i];
drivers/net/ethernet/intel/idpf/idpf_txrx.c
4105
else if (!tx && !q_vector->rx)
drivers/net/ethernet/intel/idpf/idpf_txrx.c
4295
struct idpf_rx_queue *rxq = q_vec->rx[i];
drivers/net/ethernet/intel/idpf/idpf_txrx.c
4403
q->q_vector->rx[q_index] = q;
drivers/net/ethernet/intel/idpf/idpf_txrx.c
4596
q_vector->rx = kzalloc_objs(*q_vector->rx, rxqs_per_vector);
drivers/net/ethernet/intel/idpf/idpf_txrx.c
4597
if (!q_vector->rx)
drivers/net/ethernet/intel/idpf/idpf_txrx.h
403
struct idpf_rx_queue **rx;
drivers/net/ethernet/intel/idpf/idpf_txrx.h
520
union virtchnl2_rx_desc *rx;
drivers/net/ethernet/intel/idpf/xsk.c
493
rx_desc = &rxq->rx[ntc].flex_adv_nic_3_wb;
drivers/net/ethernet/intel/igb/igb.h
377
struct igb_ring_container rx, tx;
drivers/net/ethernet/intel/igb/igb_ethtool.c
2232
if (q_vector->rx.ring)
drivers/net/ethernet/intel/igb/igb_main.c
1012
if (q_vector->rx.ring)
drivers/net/ethernet/intel/igb/igb_main.c
1013
adapter->rx_ring[q_vector->rx.ring->queue_index] = NULL;
drivers/net/ethernet/intel/igb/igb_main.c
1272
igb_add_ring(ring, &q_vector->rx);
drivers/net/ethernet/intel/igb/igb_main.c
2107
if (q_vector->rx.ring)
drivers/net/ethernet/intel/igb/igb_main.c
2109
q_vector->rx.ring->queue_index,
drivers/net/ethernet/intel/igb/igb_main.c
5762
if (!q_vector->rx.ring)
drivers/net/ethernet/intel/igb/igb_main.c
5765
rx_ring = adapter->rx_ring[q_vector->rx.ring->queue_index];
drivers/net/ethernet/intel/igb/igb_main.c
5840
packets = q_vector->rx.total_packets;
drivers/net/ethernet/intel/igb/igb_main.c
5842
avg_wire_size = q_vector->rx.total_bytes / packets;
drivers/net/ethernet/intel/igb/igb_main.c
5867
((q_vector->rx.ring && adapter->rx_itr_setting == 3) ||
drivers/net/ethernet/intel/igb/igb_main.c
5868
(!q_vector->rx.ring && adapter->tx_itr_setting == 3)))
drivers/net/ethernet/intel/igb/igb_main.c
5877
q_vector->rx.total_bytes = 0;
drivers/net/ethernet/intel/igb/igb_main.c
5878
q_vector->rx.total_packets = 0;
drivers/net/ethernet/intel/igb/igb_main.c
5965
igb_update_itr(q_vector, &q_vector->rx);
drivers/net/ethernet/intel/igb/igb_main.c
5967
current_itr = max(q_vector->rx.itr, q_vector->tx.itr);
drivers/net/ethernet/intel/igb/igb_main.c
5971
((q_vector->rx.ring && adapter->rx_itr_setting == 3) ||
drivers/net/ethernet/intel/igb/igb_main.c
5972
(!q_vector->rx.ring && adapter->tx_itr_setting == 3)))
drivers/net/ethernet/intel/igb/igb_main.c
7219
if (q_vector->rx.ring)
drivers/net/ethernet/intel/igb/igb_main.c
7220
igb_update_rx_dca(adapter, q_vector->rx.ring, cpu);
drivers/net/ethernet/intel/igb/igb_main.c
766
if (q_vector->rx.ring)
drivers/net/ethernet/intel/igb/igb_main.c
767
rx_queue = q_vector->rx.ring->reg_idx;
drivers/net/ethernet/intel/igb/igb_main.c
8258
if ((q_vector->rx.ring && (adapter->rx_itr_setting & 3)) ||
drivers/net/ethernet/intel/igb/igb_main.c
8259
(!q_vector->rx.ring && (adapter->tx_itr_setting & 3))) {
drivers/net/ethernet/intel/igb/igb_main.c
8295
if (q_vector->rx.ring) {
drivers/net/ethernet/intel/igb/igb_main.c
8298
xsk_pool = READ_ONCE(q_vector->rx.ring->xsk_pool);
drivers/net/ethernet/intel/igb/igb_main.c
9006
struct igb_ring *ring = q_vector->rx.ring;
drivers/net/ethernet/intel/igb/igb_main.c
9013
q_vector->rx.total_packets += packets;
drivers/net/ethernet/intel/igb/igb_main.c
9014
q_vector->rx.total_bytes += bytes;
drivers/net/ethernet/intel/igb/igb_main.c
9021
struct igb_ring *rx_ring = q_vector->rx.ring;
drivers/net/ethernet/intel/igb/igb_main.c
933
if (q_vector->rx.ring && q_vector->tx.ring)
drivers/net/ethernet/intel/igb/igb_main.c
935
q_vector->rx.ring->queue_index);
drivers/net/ethernet/intel/igb/igb_main.c
939
else if (q_vector->rx.ring)
drivers/net/ethernet/intel/igb/igb_main.c
941
q_vector->rx.ring->queue_index);
drivers/net/ethernet/intel/igb/igb_xsk.c
346
struct igb_ring *rx_ring = q_vector->rx.ring;
drivers/net/ethernet/intel/igc/igc.h
617
struct igc_ring_container rx, tx;
drivers/net/ethernet/intel/igc/igc_ethtool.c
949
if (q_vector->rx.ring)
drivers/net/ethernet/intel/igc/igc_main.c
2594
struct igc_ring *ring = q_vector->rx.ring;
drivers/net/ethernet/intel/igc/igc_main.c
2601
q_vector->rx.total_packets += packets;
drivers/net/ethernet/intel/igc/igc_main.c
2602
q_vector->rx.total_bytes += bytes;
drivers/net/ethernet/intel/igc/igc_main.c
2609
struct igc_ring *rx_ring = q_vector->rx.ring;
drivers/net/ethernet/intel/igc/igc_main.c
2775
struct igc_ring *ring = q_vector->rx.ring;
drivers/net/ethernet/intel/igc/igc_main.c
2804
struct igc_ring *ring = q_vector->rx.ring;
drivers/net/ethernet/intel/igc/igc_main.c
4231
if (q_vector->rx.ring)
drivers/net/ethernet/intel/igc/igc_main.c
4232
rx_queue = q_vector->rx.ring->reg_idx;
drivers/net/ethernet/intel/igc/igc_main.c
4409
if (q_vector->rx.ring)
drivers/net/ethernet/intel/igc/igc_main.c
4410
adapter->rx_ring[q_vector->rx.ring->queue_index] = NULL;
drivers/net/ethernet/intel/igc/igc_main.c
4542
igc_update_itr(q_vector, &q_vector->rx);
drivers/net/ethernet/intel/igc/igc_main.c
4544
current_itr = max(q_vector->rx.itr, q_vector->tx.itr);
drivers/net/ethernet/intel/igc/igc_main.c
4548
((q_vector->rx.ring && adapter->rx_itr_setting == 3) ||
drivers/net/ethernet/intel/igc/igc_main.c
4549
(!q_vector->rx.ring && adapter->tx_itr_setting == 3)))
drivers/net/ethernet/intel/igc/igc_main.c
4706
packets = q_vector->rx.total_packets;
drivers/net/ethernet/intel/igc/igc_main.c
4708
avg_wire_size = q_vector->rx.total_bytes / packets;
drivers/net/ethernet/intel/igc/igc_main.c
4733
((q_vector->rx.ring && adapter->rx_itr_setting == 3) ||
drivers/net/ethernet/intel/igc/igc_main.c
4734
(!q_vector->rx.ring && adapter->tx_itr_setting == 3)))
drivers/net/ethernet/intel/igc/igc_main.c
4743
q_vector->rx.total_bytes = 0;
drivers/net/ethernet/intel/igc/igc_main.c
4744
q_vector->rx.total_packets = 0;
drivers/net/ethernet/intel/igc/igc_main.c
4754
if ((q_vector->rx.ring && (adapter->rx_itr_setting & 3)) ||
drivers/net/ethernet/intel/igc/igc_main.c
4755
(!q_vector->rx.ring && (adapter->tx_itr_setting & 3))) {
drivers/net/ethernet/intel/igc/igc_main.c
4809
struct igc_ring *rx_ring = q_vector->rx.ring;
drivers/net/ethernet/intel/igc/igc_main.c
4934
igc_add_ring(ring, &q_vector->rx);
drivers/net/ethernet/intel/igc/igc_main.c
5097
if (q_vector->rx.ring)
drivers/net/ethernet/intel/igc/igc_main.c
5099
q_vector->rx.ring->queue_index,
drivers/net/ethernet/intel/igc/igc_main.c
5727
if (q_vector->rx.ring && q_vector->tx.ring)
drivers/net/ethernet/intel/igc/igc_main.c
5729
q_vector->rx.ring->queue_index);
drivers/net/ethernet/intel/igc/igc_main.c
5733
else if (q_vector->rx.ring)
drivers/net/ethernet/intel/igc/igc_main.c
5735
q_vector->rx.ring->queue_index);
drivers/net/ethernet/intel/igc/igc_main.c
5996
if (!q_vector->rx.ring)
drivers/net/ethernet/intel/igc/igc_main.c
5999
rx_ring = adapter->rx_ring[q_vector->rx.ring->queue_index];
drivers/net/ethernet/intel/ixgbe/ixgbe.h
511
struct ixgbe_ring_container rx, tx;
drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c
29
int rx = DCB_RX_CONFIG;
drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c
68
if (dst->path[rx].prio_type != src->path[rx].prio_type) {
drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c
69
dst->path[rx].prio_type = src->path[rx].prio_type;
drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c
73
if (dst->path[rx].bwg_id != src->path[rx].bwg_id) {
drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c
74
dst->path[rx].bwg_id = src->path[rx].bwg_id;
drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c
78
if (dst->path[rx].bwg_percent != src->path[rx].bwg_percent) {
drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c
79
dst->path[rx].bwg_percent = src->path[rx].bwg_percent;
drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c
83
if (dst->path[rx].up_to_tc_bitmap !=
drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c
84
src->path[rx].up_to_tc_bitmap) {
drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c
85
dst->path[rx].up_to_tc_bitmap =
drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c
86
src->path[rx].up_to_tc_bitmap;
drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c
97
if (dcfg->bw_percentage[rx][j] != scfg->bw_percentage[rx][j]) {
drivers/net/ethernet/intel/ixgbe/ixgbe_dcb_nl.c
98
dcfg->bw_percentage[rx][j] = scfg->bw_percentage[rx][j];
drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c
2538
if (adapter->q_vector[0]->tx.count && adapter->q_vector[0]->rx.count)
drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c
2591
if (adapter->q_vector[0]->tx.count && adapter->q_vector[0]->rx.count) {
drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c
2625
if (adapter->q_vector[0]->tx.count && adapter->q_vector[0]->rx.count)
drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c
2645
if (q_vector->tx.count && !q_vector->rx.count)
drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c
103
*rx = tc << 3; /* 0, 8, 16, 24, 32, 40, 48, 56 */
drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c
1030
ixgbe_for_each_ring(ring, q_vector->rx)
drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c
117
*rx = tc << 4;
drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c
130
*rx = tc << 5;
drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c
888
q_vector->rx.itr = IXGBE_ITR_ADAPTIVE_MAX_USECS |
drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c
91
unsigned int *tx, unsigned int *rx)
drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c
97
*rx = 0;
drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c
972
ixgbe_add_ring(ring, &q_vector->rx);
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
1197
q_vector->rx.total_bytes += bytes;
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
1198
q_vector->rx.total_packets += pkts;
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
1563
ixgbe_for_each_ring(ring, q_vector->rx)
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
2675
ixgbe_for_each_ring(ring, q_vector->rx)
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
2945
ixgbe_update_itr(q_vector, &q_vector->rx);
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
2948
new_itr = min(q_vector->rx.itr, q_vector->tx.itr);
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
3564
if (q_vector->rx.ring || q_vector->tx.ring)
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
3606
if (q_vector->rx.count > 1)
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
3607
per_ring_budget = max(budget/q_vector->rx.count, 1);
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
3611
ixgbe_for_each_ring(ring, q_vector->rx) {
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
3656
if (q_vector->tx.ring && q_vector->rx.ring) {
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
3660
} else if (q_vector->rx.ring) {
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
3831
if (!q_vector->rx.ring && !q_vector->tx.ring)
drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
7997
if (qv->rx.ring || qv->tx.ring)
drivers/net/ethernet/intel/ixgbevf/ethtool.c
803
if (adapter->q_vector[0]->tx.count && adapter->q_vector[0]->rx.count)
drivers/net/ethernet/intel/ixgbevf/ethtool.c
827
adapter->q_vector[0]->rx.count && ec->tx_coalesce_usecs)
drivers/net/ethernet/intel/ixgbevf/ethtool.c
858
if (q_vector->tx.count && !q_vector->rx.count)
drivers/net/ethernet/intel/ixgbevf/ixgbevf.h
239
struct ixgbevf_ring_container rx, tx;
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
1257
q_vector->rx.total_packets += total_rx_packets;
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
1258
q_vector->rx.total_bytes += total_rx_bytes;
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
1291
if (q_vector->rx.count > 1)
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
1292
per_ring_budget = max(budget/q_vector->rx.count, 1);
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
1296
ixgbevf_for_each_ring(ring, q_vector->rx) {
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
1365
ixgbevf_for_each_ring(ring, q_vector->rx)
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
1371
if (q_vector->tx.ring && !q_vector->rx.ring) {
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
1472
ixgbevf_update_itr(q_vector, &q_vector->rx);
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
1474
current_itr = max(q_vector->rx.itr, q_vector->tx.itr);
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
1527
if (q_vector->rx.ring || q_vector->tx.ring)
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
1551
if (q_vector->tx.ring && q_vector->rx.ring) {
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
1555
} else if (q_vector->rx.ring) {
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
1637
if (!adapter->q_vector[i]->rx.ring &&
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
2849
ixgbevf_add_ring(ring, &q_vector->rx);
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
2891
ixgbevf_for_each_ring(ring, q_vector->rx)
drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c
3269
if (qv->rx.ring || qv->tx.ring)
drivers/net/ethernet/lantiq_xrx200.c
287
int rx = 0;
drivers/net/ethernet/lantiq_xrx200.c
290
while (rx < budget) {
drivers/net/ethernet/lantiq_xrx200.c
299
rx++;
drivers/net/ethernet/lantiq_xrx200.c
305
if (rx < budget) {
drivers/net/ethernet/lantiq_xrx200.c
306
if (napi_complete_done(&ch->napi, rx))
drivers/net/ethernet/lantiq_xrx200.c
310
return rx;
drivers/net/ethernet/marvell/mv643xx_eth.c
510
int rx;
drivers/net/ethernet/marvell/mv643xx_eth.c
512
rx = 0;
drivers/net/ethernet/marvell/mv643xx_eth.c
513
while (rx < budget && rxq->rx_desc_count) {
drivers/net/ethernet/marvell/mv643xx_eth.c
536
rx++;
drivers/net/ethernet/marvell/mv643xx_eth.c
592
if (rx < budget)
drivers/net/ethernet/marvell/mv643xx_eth.c
595
return rx;
drivers/net/ethernet/marvell/mv643xx_eth.c
606
int rx;
drivers/net/ethernet/marvell/mv643xx_eth.c
623
rx = rxq->rx_used_desc++;
drivers/net/ethernet/marvell/mv643xx_eth.c
627
rx_desc = rxq->rx_desc_area + rx;
drivers/net/ethernet/marvell/mv643xx_eth.c
634
rxq->rx_skb[rx] = skb;
drivers/net/ethernet/marvell/octeon_ep/octep_ctrl_net.h
144
struct octep_ctrl_net_h2f_req_cmd_state rx;
drivers/net/ethernet/marvell/octeon_ep/octep_ctrl_net.h
201
struct octep_ctrl_net_h2f_resp_cmd_state rx;
drivers/net/ethernet/marvell/octeontx2/af/mbox.h
1192
} rx;
drivers/net/ethernet/marvell/octeontx2/af/mbox.h
1491
} rx;
drivers/net/ethernet/marvell/octeontx2/af/mcs.c
1569
err = mcs_alloc_struct_mem(mcs, &mcs->rx);
drivers/net/ethernet/marvell/octeontx2/af/mcs.c
298
map = &mcs->rx;
drivers/net/ethernet/marvell/octeontx2/af/mcs.c
501
__set_bit(flow_id, mcs->rx.flow_ids.bmap);
drivers/net/ethernet/marvell/octeontx2/af/mcs.c
514
__set_bit(secy_id, mcs->rx.secy.bmap);
drivers/net/ethernet/marvell/octeontx2/af/mcs.c
550
map = &mcs->rx;
drivers/net/ethernet/marvell/octeontx2/af/mcs.c
590
map = (req->dir == MCS_RX) ? &mcs->rx : &mcs->tx;
drivers/net/ethernet/marvell/octeontx2/af/mcs.c
718
map = &mcs->rx;
drivers/net/ethernet/marvell/octeontx2/af/mcs.c
778
map = &mcs->rx;
drivers/net/ethernet/marvell/octeontx2/af/mcs.c
895
event.pcifunc = mcs->rx.sa2pf_map[event.sa_id];
drivers/net/ethernet/marvell/octeontx2/af/mcs.h
140
struct mcs_rsrc_map rx;
drivers/net/ethernet/marvell/octeontx2/af/mcs_rvu_if.c
644
map = &mcs->rx;
drivers/net/ethernet/marvell/octeontx2/af/mcs_rvu_if.c
694
map = &mcs->rx;
drivers/net/ethernet/marvell/octeontx2/af/mcs_rvu_if.c
776
map = (req->dir == MCS_RX) ? &mcs->rx : &mcs->tx;
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
265
map = &mcs->rx.sa;
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
331
map = &mcs->rx.sc;
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
371
map = &mcs->rx.flow_ids;
drivers/net/ethernet/marvell/octeontx2/af/rvu_debugfs.c
448
map = &mcs->rx.secy;
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
2966
if (req->rx.vtag_type > NIX_AF_LFX_RX_VTAG_TYPE7 ||
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
2971
if (req->rx.vtag_type == NIX_AF_LFX_RX_VTAG_TYPE7)
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
2974
if (req->rx.capture_vtag)
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
2976
if (req->rx.strip_vtag)
drivers/net/ethernet/marvell/octeontx2/af/rvu_nix.c
2980
NIX_AF_LFX_RX_VTAG_TYPEX(nixlf, req->rx.vtag_type), regval);
drivers/net/ethernet/marvell/octeontx2/af/rvu_rep.c
164
rsp->rx.octs = RVU_LF_RX_STATS(RX_OCTS);
drivers/net/ethernet/marvell/octeontx2/af/rvu_rep.c
165
rsp->rx.ucast = RVU_LF_RX_STATS(RX_UCAST);
drivers/net/ethernet/marvell/octeontx2/af/rvu_rep.c
166
rsp->rx.bcast = RVU_LF_RX_STATS(RX_BCAST);
drivers/net/ethernet/marvell/octeontx2/af/rvu_rep.c
167
rsp->rx.mcast = RVU_LF_RX_STATS(RX_MCAST);
drivers/net/ethernet/marvell/octeontx2/af/rvu_rep.c
168
rsp->rx.drop = RVU_LF_RX_STATS(RX_DROP);
drivers/net/ethernet/marvell/octeontx2/af/rvu_rep.c
169
rsp->rx.err = RVU_LF_RX_STATS(RX_ERR);
drivers/net/ethernet/marvell/octeontx2/af/rvu_rep.c
170
rsp->rx.drop_octs = RVU_LF_RX_STATS(RX_DROP_OCTS);
drivers/net/ethernet/marvell/octeontx2/af/rvu_rep.c
171
rsp->rx.drop_mcast = RVU_LF_RX_STATS(RX_DRP_MCAST);
drivers/net/ethernet/marvell/octeontx2/af/rvu_rep.c
172
rsp->rx.drop_bcast = RVU_LF_RX_STATS(RX_DRP_BCAST);
drivers/net/ethernet/marvell/octeontx2/af/rvu_rep.c
227
req.rx.vtag_type = NIX_AF_LFX_RX_VTAG_TYPE0;
drivers/net/ethernet/marvell/octeontx2/af/rvu_rep.c
228
req.rx.strip_vtag = true;
drivers/net/ethernet/marvell/octeontx2/af/rvu_rep.c
229
req.rx.capture_vtag = false;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
1489
req->rx.vtag_type = NIX_AF_LFX_RX_VTAG_TYPE0;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
1490
req->rx.strip_vtag = enable;
drivers/net/ethernet/marvell/octeontx2/nic/otx2_flows.c
1491
req->rx.capture_vtag = enable;
drivers/net/ethernet/marvell/octeontx2/nic/rep.c
367
stats->rx_bytes = rsp->rx.octs;
drivers/net/ethernet/marvell/octeontx2/nic/rep.c
368
stats->rx_frames = rsp->rx.ucast + rsp->rx.bcast +
drivers/net/ethernet/marvell/octeontx2/nic/rep.c
369
rsp->rx.mcast;
drivers/net/ethernet/marvell/octeontx2/nic/rep.c
370
stats->rx_drops = rsp->rx.drop;
drivers/net/ethernet/marvell/octeontx2/nic/rep.c
371
stats->rx_mcast_frames = rsp->rx.mcast;
drivers/net/ethernet/marvell/sky2.c
2225
static enum flow_control sky2_flow(int rx, int tx)
drivers/net/ethernet/marvell/sky2.c
2227
if (rx)
drivers/net/ethernet/mediatek/mtk_eth_soc.c
1836
rxd = ring->dma + idx * eth->soc->rx.desc_size;
drivers/net/ethernet/mediatek/mtk_eth_soc.c
2204
rxd = ring->dma + idx * eth->soc->rx.desc_size;
drivers/net/ethernet/mediatek/mtk_eth_soc.c
2343
if (*rxdcsum & eth->soc->rx.dma_l4_valid)
drivers/net/ethernet/mediatek/mtk_eth_soc.c
2612
mtk_w32(eth, eth->soc->rx.irq_done_mask,
drivers/net/ethernet/mediatek/mtk_eth_soc.c
2628
eth->soc->rx.irq_done_mask);
drivers/net/ethernet/mediatek/mtk_eth_soc.c
2631
mtk_rx_irq_enable(eth, eth->soc->rx.irq_done_mask);
drivers/net/ethernet/mediatek/mtk_eth_soc.c
2789
rx_dma_size = soc->rx.dma_size;
drivers/net/ethernet/mediatek/mtk_eth_soc.c
2811
rx_dma_size * eth->soc->rx.desc_size,
drivers/net/ethernet/mediatek/mtk_eth_soc.c
2822
rxd = ring->dma + i * eth->soc->rx.desc_size;
drivers/net/ethernet/mediatek/mtk_eth_soc.c
2913
rxd = ring->dma + i * eth->soc->rx.desc_size;
drivers/net/ethernet/mediatek/mtk_eth_soc.c
2929
mtk_dma_ring_free(eth, ring->dma_size * eth->soc->rx.desc_size,
drivers/net/ethernet/mediatek/mtk_eth_soc.c
3401
mtk_rx_irq_disable(eth, eth->soc->rx.irq_done_mask);
drivers/net/ethernet/mediatek/mtk_eth_soc.c
3427
eth->soc->rx.irq_done_mask) {
drivers/net/ethernet/mediatek/mtk_eth_soc.c
3429
eth->soc->rx.irq_done_mask)
drivers/net/ethernet/mediatek/mtk_eth_soc.c
3447
mtk_rx_irq_disable(eth, eth->soc->rx.irq_done_mask);
drivers/net/ethernet/mediatek/mtk_eth_soc.c
3450
mtk_rx_irq_enable(eth, eth->soc->rx.irq_done_mask);
drivers/net/ethernet/mediatek/mtk_eth_soc.c
3624
mtk_rx_irq_enable(eth, soc->rx.irq_done_mask);
drivers/net/ethernet/mediatek/mtk_eth_soc.c
3709
mtk_rx_irq_disable(eth, eth->soc->rx.irq_done_mask);
drivers/net/ethernet/mediatek/mtk_eth_soc.c
4198
mtk_w32(eth, eth->soc->rx.irq_done_mask, reg_map->pdma.int_grp + 4);
drivers/net/ethernet/mediatek/mtk_eth_soc.c
4200
mtk_w32(eth, eth->soc->rx.irq_done_mask, reg_map->qdma.int_grp + 4);
drivers/net/ethernet/mediatek/mtk_eth_soc.c
5365
.rx = {
drivers/net/ethernet/mediatek/mtk_eth_soc.c
5393
.rx = {
drivers/net/ethernet/mediatek/mtk_eth_soc.c
5423
.rx = {
drivers/net/ethernet/mediatek/mtk_eth_soc.c
5452
.rx = {
drivers/net/ethernet/mediatek/mtk_eth_soc.c
5478
.rx = {
drivers/net/ethernet/mediatek/mtk_eth_soc.c
5508
.rx = {
drivers/net/ethernet/mediatek/mtk_eth_soc.c
5538
.rx = {
drivers/net/ethernet/mediatek/mtk_eth_soc.c
5568
.rx = {
drivers/net/ethernet/mediatek/mtk_eth_soc.c
5591
.rx = {
drivers/net/ethernet/mediatek/mtk_eth_soc.h
1237
} rx;
drivers/net/ethernet/mediatek/mtk_eth_soc.h
363
#define RX_DMA_PREP_PLEN0(x) (((x) & eth->soc->rx.dma_max_len) << eth->soc->rx.dma_len_offset)
drivers/net/ethernet/mediatek/mtk_eth_soc.h
364
#define RX_DMA_GET_PLEN0(x) (((x) >> eth->soc->rx.dma_len_offset) & eth->soc->rx.dma_max_len)
drivers/net/ethernet/mediatek/mtk_star_emac.c
439
bool rx, bool tx)
drivers/net/ethernet/mediatek/mtk_star_emac.c
447
if (rx)
drivers/net/ethernet/mediatek/mtk_star_emac.c
454
bool rx, bool tx)
drivers/net/ethernet/mediatek/mtk_star_emac.c
462
if (rx)
drivers/net/ethernet/mediatek/mtk_star_emac.c
774
bool rx, tx;
drivers/net/ethernet/mediatek/mtk_star_emac.c
776
rx = (intr_status & MTK_STAR_BIT_INT_STS_FNRC) &&
drivers/net/ethernet/mediatek/mtk_star_emac.c
781
if (rx || tx) {
drivers/net/ethernet/mediatek/mtk_star_emac.c
784
mtk_star_disable_dma_irq(priv, rx, tx);
drivers/net/ethernet/mediatek/mtk_star_emac.c
787
if (rx)
drivers/net/ethernet/mediatek/mtk_wed_wo.c
135
bool rx)
drivers/net/ethernet/mediatek/mtk_wed_wo.c
137
enum dma_data_direction dir = rx ? DMA_FROM_DEVICE : DMA_TO_DEVICE;
drivers/net/ethernet/mediatek/mtk_wed_wo.c
162
if (rx) {
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
3130
struct netdev_queue_stats_rx *rx,
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
3141
rx->packets = 0;
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
3142
rx->bytes = 0;
drivers/net/ethernet/mellanox/mlx4/en_netdev.c
3143
rx->alloc_fail = 0;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1000
rx->sa_sel.rule = rule;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1045
struct mlx5e_ipsec_rx *rx, u32 family)
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1053
ipsec_rx_create_attr_set(ipsec, rx, family, &attr);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1055
err = ipsec_rx_status_pass_dest_get(ipsec, rx, &attr, &dest[0]);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1064
rx->ft.status = ft;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1066
err = ipsec_rx_sa_selector_create(ipsec, rx, &attr);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1072
rx->allow_tunnel_mode =
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1073
mlx5_eswitch_block_encap(mdev, rx == ipsec->rx_esw);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1075
if (rx->allow_tunnel_mode)
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1082
rx->ft.sa = ft;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1084
ipsec_rx_sa_miss_dest_get(ipsec, rx, &attr, &dest[0], &miss_dest);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1085
err = ipsec_miss_create(mdev, rx->ft.sa, &rx->sa, &miss_dest);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1089
err = ipsec_rx_policy_create(ipsec, rx, &attr, &dest[0]);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1094
dest[1].counter = rx->fc->cnt;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1095
err = mlx5_ipsec_rx_status_create(ipsec, rx, dest);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1100
if (rx != ipsec->rx_esw)
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1101
ipsec_rx_ft_connect(ipsec, rx, &attr);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1105
ipsec_rx_policy_destroy(rx);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1107
mlx5_del_flow_rules(rx->sa.rule);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1108
mlx5_destroy_flow_group(rx->sa.group);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1110
mlx5_destroy_flow_table(rx->ft.sa);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1111
if (rx->allow_tunnel_mode)
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1114
ipsec_rx_sa_selector_destroy(mdev, rx);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1116
mlx5_destroy_flow_table(rx->ft.status);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1123
struct mlx5e_ipsec_rx *rx, u32 family)
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1127
if (rx->ft.refcnt)
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1134
err = rx_create(mdev, ipsec, rx, family);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1141
rx->ft.refcnt++;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1145
static void rx_put(struct mlx5e_ipsec *ipsec, struct mlx5e_ipsec_rx *rx,
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1148
if (--rx->ft.refcnt)
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1151
rx_destroy(ipsec->mdev, ipsec, rx, family);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1159
struct mlx5e_ipsec_rx *rx = ipsec_rx(ipsec, family, type);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1162
mutex_lock(&rx->ft.mutex);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1163
err = rx_get(mdev, ipsec, rx, family);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1164
mutex_unlock(&rx->ft.mutex);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1168
return rx;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1175
struct mlx5e_ipsec_rx *rx = ipsec_rx(ipsec, family, type);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1179
mutex_lock(&rx->ft.mutex);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1180
err = rx_get(mdev, ipsec, rx, family);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1184
ft = rx->chains ? ipsec_chains_get_table(rx->chains, prio) : rx->ft.pol;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1190
mutex_unlock(&rx->ft.mutex);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1194
rx_put(ipsec, rx, family);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1196
mutex_unlock(&rx->ft.mutex);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1202
struct mlx5e_ipsec_rx *rx = ipsec_rx(ipsec, family, type);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1204
mutex_lock(&rx->ft.mutex);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1205
rx_put(ipsec, rx, family);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1206
mutex_unlock(&rx->ft.mutex);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1211
struct mlx5e_ipsec_rx *rx = ipsec_rx(ipsec, family, type);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1213
mutex_lock(&rx->ft.mutex);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1214
if (rx->chains)
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1215
ipsec_chains_put_table(rx->chains, prio);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1217
rx_put(ipsec, rx, family);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1218
mutex_unlock(&rx->ft.mutex);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
165
struct mlx5e_ipsec_rx *rx)
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
167
mlx5_del_flow_rules(rx->status_checks.all.rule);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
168
mlx5_fc_destroy(ipsec->mdev, rx->status_checks.all.fc);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
169
mlx5_destroy_flow_group(rx->status_checks.drop_all_group);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
173
struct mlx5e_ipsec_rx *rx)
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
175
mlx5_del_flow_rules(rx->status_checks.packet_offload_pass_rule);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
176
mlx5_del_flow_rules(rx->status_checks.crypto_offload_pass_rule);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
180
struct mlx5e_ipsec_rx *rx,
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
185
if (rx == ipsec->rx_esw) {
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1954
struct mlx5e_ipsec_rx *rx,
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1979
ipsec_rx_rule_add_match_obj(sa_entry, rx, spec);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1985
dest[0].ft = rx->ft.sa_sel;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1987
dest[1].counter = rx->fc->cnt;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
1989
rule = mlx5_add_flow_rules(rx->ft.status, spec, &flow_act, dest, 2);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
199
struct mlx5e_ipsec_rx *rx)
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
2009
dest[0].ft = rx->ft.pol;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
2011
rule = mlx5_add_flow_rules(rx->ft.sa_sel, spec, &flow_act, &dest[0], 1);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
202
struct mlx5_flow_table *ft = rx->ft.status;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
2041
struct mlx5e_ipsec_rx *rx;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
2045
rx = rx_ft_get(mdev, ipsec, attrs->addrs.family, attrs->type);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
2046
if (IS_ERR(rx))
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
2047
return PTR_ERR(rx);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
2066
if (rx != ipsec->rx_esw)
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
2102
dest[0].ft = rx->ft.status;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
2105
rule = mlx5_add_flow_rules(rx->ft.sa, spec, &flow_act, dest, 2);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
2113
err = rx_add_rule_sa_selector(sa_entry, rx, &attrs->upspec);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
2119
err = rx_add_rule_drop_replay(sa_entry, rx);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
2123
err = rx_add_rule_drop_auth_trailer(sa_entry, rx);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
228
if (rx == ipsec->rx_esw)
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
234
ipsec_rx_rule_add_match_obj(sa_entry, rx, spec);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
2347
struct mlx5e_ipsec_rx *rx;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
2355
rx = ipsec_rx(pol_entry->ipsec, attrs->addrs.family, attrs->type);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
2378
dest[dstn].counter = rx->fc->drop;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
2388
if (rx == ipsec->rx_esw && rx->chains)
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
2390
ipsec_rx_default_dest_get(ipsec, rx, &dest[dstn]);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
278
static int rx_add_rule_drop_replay(struct mlx5e_ipsec_sa_entry *sa_entry, struct mlx5e_ipsec_rx *rx)
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
281
struct mlx5_flow_table *ft = rx->ft.status;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
2864
struct mlx5e_ipsec_rx *rx = ipsec_rx(ipsec, x->props.family,
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
2867
from_fdb = (rx == ipsec->rx_esw);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
306
if (rx == ipsec->rx_esw)
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
312
ipsec_rx_rule_add_match_obj(sa_entry, rx, spec);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
335
struct mlx5e_ipsec_rx *rx)
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
338
struct mlx5_flow_table *ft = rx->ft.status;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
377
if (rx == ipsec->rx_esw)
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
387
rx->status_checks.drop_all_group = g;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
388
rx->status_checks.all.rule = rule;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
389
rx->status_checks.all.fc = flow_counter;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
406
struct mlx5e_ipsec_rx *rx)
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
409
struct mlx5_flow_table *ft = rx->ft.status;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
440
rx->status_checks.pass_group = fg;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
448
struct mlx5e_ipsec_rx *rx,
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
469
if (rx == ipsec->rx_esw)
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
475
rule = mlx5_add_flow_rules(rx->ft.status, spec, &flow_act, dest, 2);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
492
struct mlx5e_ipsec_rx *rx)
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
494
ipsec_rx_status_pass_destroy(ipsec, rx);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
495
mlx5_destroy_flow_group(rx->status_checks.pass_group);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
496
ipsec_rx_status_drop_destroy(ipsec, rx);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
500
struct mlx5e_ipsec_rx *rx,
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
507
err = ipsec_rx_status_drop_all_create(ipsec, rx);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
511
err = ipsec_rx_status_pass_group_create(ipsec, rx);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
515
rule = ipsec_rx_status_pass_create(ipsec, rx, dest,
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
521
rx->status_checks.crypto_offload_pass_rule = rule;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
524
pol_dest[0].ft = rx->ft.pol;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
526
rule = ipsec_rx_status_pass_create(ipsec, rx, pol_dest,
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
532
rx->status_checks.packet_offload_pass_rule = rule;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
537
mlx5_del_flow_rules(rx->status_checks.crypto_offload_pass_rule);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
539
mlx5_destroy_flow_group(rx->status_checks.pass_group);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
541
ipsec_rx_status_drop_destroy(ipsec, rx);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
603
static void ipsec_rx_update_default_dest(struct mlx5e_ipsec_rx *rx,
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
607
mlx5_modify_rule_destination(rx->pol_miss_rule, new_dest, old_dest);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
608
mlx5_modify_rule_destination(rx->status_checks.crypto_offload_pass_rule,
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
614
struct mlx5e_ipsec_rx *rx = ipsec_rx(ipsec, family, XFRM_DEV_OFFLOAD_PACKET);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
626
ipsec_rx_update_default_dest(rx, &old_dest, &new_dest);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
631
struct mlx5e_ipsec_rx *rx = ipsec_rx(ipsec, family, XFRM_DEV_OFFLOAD_PACKET);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
638
ipsec_rx_update_default_dest(rx, &old_dest, &new_dest);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
688
struct mlx5e_ipsec_rx *rx, u32 family)
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
692
if (rx->ttc_rules_added)
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
697
static void ipsec_rx_policy_destroy(struct mlx5e_ipsec_rx *rx)
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
699
if (rx->chains) {
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
700
ipsec_chains_destroy(rx->chains);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
702
mlx5_del_flow_rules(rx->pol.rule);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
703
mlx5_destroy_flow_group(rx->pol.group);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
704
mlx5_destroy_flow_table(rx->ft.pol);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
707
if (rx->pol_miss_rule) {
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
708
mlx5_del_flow_rules(rx->pol_miss_rule);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
709
mlx5_destroy_flow_table(rx->pol_miss_ft);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
714
struct mlx5e_ipsec_rx *rx)
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
716
mlx5_del_flow_rules(rx->sa_sel.rule);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
717
mlx5_fc_destroy(mdev, rx->sa_sel.fc);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
718
rx->sa_sel.fc = NULL;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
719
mlx5_destroy_flow_group(rx->sa_sel.group);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
720
mlx5_destroy_flow_table(rx->ft.sa_sel);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
724
struct mlx5e_ipsec_rx *rx, u32 family)
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
727
if (rx != ipsec->rx_esw)
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
728
ipsec_rx_ft_disconnect(ipsec, rx, family);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
730
mlx5_del_flow_rules(rx->sa.rule);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
731
mlx5_destroy_flow_group(rx->sa.group);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
732
mlx5_destroy_flow_table(rx->ft.sa);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
733
if (rx->allow_tunnel_mode)
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
735
mlx5_ipsec_rx_status_destroy(ipsec, rx);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
736
mlx5_destroy_flow_table(rx->ft.status);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
738
ipsec_rx_sa_selector_destroy(mdev, rx);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
740
ipsec_rx_policy_destroy(rx);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
745
if (rx == ipsec->rx_esw)
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
752
struct mlx5e_ipsec_rx *rx,
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
756
if (rx == ipsec->rx_esw) {
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
775
struct mlx5e_ipsec_rx *rx,
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
782
if (rx == ipsec->rx_esw)
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
802
struct mlx5e_ipsec_rx *rx,
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
807
if (rx == ipsec->rx_esw)
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
816
struct mlx5e_ipsec_rx *rx,
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
820
dest->ft = rx->pol_miss_ft;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
824
struct mlx5e_ipsec_rx *rx,
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
831
dest.ft = rx->ft.sa;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
837
rx->ttc_rules_added = !mlx5_ttc_create_ipsec_rules(ttc, inner_ttc);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
841
struct mlx5e_ipsec_rx *rx,
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
851
if (rx == ipsec->rx_esw) {
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
855
rx->pol_miss_ft = dest->ft;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
874
rx->pol_miss_ft = ft;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
875
rx->pol_miss_rule = rule;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
885
struct mlx5e_ipsec_rx *rx,
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
894
err = ipsec_rx_chains_create_miss(ipsec, rx, attr, dest);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
898
ipsec_rx_default_dest_get(ipsec, rx, &default_dest);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
901
rx->chains = ipsec_chains_create(mdev,
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
906
&rx->ft.pol);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
907
if (IS_ERR(rx->chains))
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
908
err = PTR_ERR(rx->chains);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
916
rx->ft.pol = ft;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
918
err = ipsec_miss_create(mdev, rx->ft.pol, &rx->pol,
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
921
mlx5_destroy_flow_table(rx->ft.pol);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
928
if (rx->pol_miss_rule) {
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
929
mlx5_del_flow_rules(rx->pol_miss_rule);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
930
mlx5_destroy_flow_table(rx->pol_miss_ft);
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
936
struct mlx5e_ipsec_rx *rx,
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
997
rx->ft.sa_sel = ft;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
998
rx->sa_sel.group = fg;
drivers/net/ethernet/mellanox/mlx5/core/en_accel/ipsec_fs.c
999
rx->sa_sel.fc = fc;
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
5460
struct netdev_queue_stats_rx *rx,
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
5468
rx->packets = 0;
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
5469
rx->bytes = 0;
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
5470
rx->alloc_fail = 0;
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
5477
rx->packets += rx_i.packets;
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
5478
rx->bytes += rx_i.bytes;
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
5479
rx->alloc_fail += rx_i.alloc_fail;
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
5489
rx->packets += rq_stats->packets;
drivers/net/ethernet/mellanox/mlx5/core/en_main.c
5490
rx->bytes += rq_stats->bytes;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_action.c
730
chunk = is_rx_rule ? dest_tbl->tbl->rx.s_anchor->chunk :
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_dbg.c
551
struct mlx5dr_rule_rx_tx *rx = &rule->rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_dbg.c
568
if (rx->nic_matcher) {
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_dbg.c
569
ret = dr_dump_rule_rx_tx(file, buff, rx, true, rule_id, format_ver);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_dbg.c
747
struct mlx5dr_matcher_rx_tx *rx = &matcher->rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_dbg.c
771
if (rx->nic_tbl) {
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_dbg.c
772
ret = dr_dump_matcher_rx_tx(file, buff, true, rx, matcher_id);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_dbg.c
834
struct mlx5dr_table_rx_tx *rx = &table->rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_dbg.c
850
if (rx->nic_dmn) {
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_dbg.c
851
ret = dr_dump_table_rx_tx(file, buff, true, rx,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_domain.c
412
if (!DR_DOMAIN_SW_STEERING_SUPPORTED(dmn, rx))
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_domain.c
416
dmn->info.rx.type = DR_DOMAIN_NIC_TYPE_RX;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_domain.c
417
dmn->info.rx.default_icm_addr = dmn->info.caps.nic_rx_drop_address;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_domain.c
418
dmn->info.rx.drop_icm_addr = dmn->info.caps.nic_rx_drop_address;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_domain.c
436
dmn->info.rx.type = DR_DOMAIN_NIC_TYPE_RX;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_domain.c
442
dmn->info.rx.default_icm_addr = vport_cap->icm_address_rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_domain.c
443
dmn->info.rx.drop_icm_addr = dmn->info.caps.esw_rx_drop_address;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_domain.c
477
mutex_init(&dmn->info.rx.mutex);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_domain.c
533
mutex_destroy(&dmn->info.rx.mutex);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
413
bool inner, rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
418
rx = nic_dmn->type == DR_DOMAIN_NIC_TYPE_RX;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
451
rx && mask.misc.source_port) {
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
467
&mask, inner, rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
471
&mask, inner, rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
475
&mask, inner, rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
481
&mask, dmn, inner, rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
487
&mask, inner, rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
492
&mask, inner, rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
496
&mask, inner, rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
501
&mask, inner, rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
505
&mask, inner, rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
509
&mask, inner, rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
513
&mask, inner, rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
518
&mask, inner, rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
523
&mask, inner, rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
526
&mask, inner, rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
530
inner, rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
534
inner, rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
539
inner, rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
544
inner, rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
548
&mask, inner, rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
551
&mask, inner, rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
556
&mask, inner, rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
560
&mask, inner, rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
565
inner, rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
569
inner, rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
574
inner, rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
578
&mask, inner, rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
590
&mask, inner, rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
595
&mask, inner, rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
600
&mask, inner, rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
604
&mask, inner, rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
609
&mask, inner, rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
613
&mask, inner, rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
617
&mask, inner, rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
621
&mask, inner, rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
626
&mask, inner, rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
631
&mask, inner, rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
635
&mask, inner, rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
640
inner, rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
644
inner, rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
650
&mask, false, rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
654
&mask, false, rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
660
mlx5dr_ste_build_empty_always_hit(&sb[idx++], rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
796
dr_matcher_uninit_nic(&matcher->rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
806
dr_matcher_uninit_nic(&matcher->rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
882
ret = dr_matcher_init_nic(matcher, &matcher->rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
893
dr_matcher_uninit_nic(&matcher->rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
954
matcher->rx.nic_tbl = &tbl->rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
955
ret = dr_matcher_init_nic(matcher, &matcher->rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
962
matcher->rx.nic_tbl = &tbl->rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_rule.c
1046
dr_rule_destroy_rule_nic(rule, &rule->rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_rule.c
1059
dr_rule_destroy_rule_nic(rule, &rule->rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_rule.c
1090
bool rx = nic_type == DR_DOMAIN_NIC_TYPE_RX;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_rule.c
1096
if (rx && value->misc.source_port != MLX5_VPORT_UPLINK)
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_rule.c
1099
if (!rx && value->misc.source_port == MLX5_VPORT_UPLINK)
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_rule.c
1103
if (rx && flow_source == MLX5_FLOW_CONTEXT_FLOW_SOURCE_LOCAL_VPORT)
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_rule.c
1106
if (!rx && flow_source == MLX5_FLOW_CONTEXT_FLOW_SOURCE_UPLINK)
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_rule.c
1269
ret = dr_rule_create_rule_nic(rule, &rule->rx, param,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_rule.c
1282
dr_rule_destroy_rule_nic(rule, &rule->rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_rule.c
1315
rule->rx.nic_matcher = &matcher->rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_rule.c
1316
ret = dr_rule_create_rule_nic(rule, &rule->rx, ¶m,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_rule.c
1325
rule->rx.nic_matcher = &matcher->rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1130
bool inner, bool rx)
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1132
sb->rx = rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1140
bool inner, bool rx)
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1142
sb->rx = rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1150
bool inner, bool rx)
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1152
sb->rx = rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1160
bool inner, bool rx)
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1162
sb->rx = rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1170
bool inner, bool rx)
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1172
sb->rx = rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1180
bool inner, bool rx)
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1182
sb->rx = rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1189
struct mlx5dr_match_param *mask, bool inner, bool rx)
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1191
sb->rx = rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1199
bool inner, bool rx)
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1201
sb->rx = rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1209
bool inner, bool rx)
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1211
sb->rx = rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1223
void mlx5dr_ste_build_empty_always_hit(struct mlx5dr_ste_build *sb, bool rx)
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1225
sb->rx = rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1234
bool inner, bool rx)
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1236
sb->rx = rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1244
bool inner, bool rx)
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1246
sb->rx = rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1255
bool inner, bool rx)
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1257
sb->rx = rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1267
bool inner, bool rx)
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1269
sb->rx = rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1279
bool inner, bool rx)
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1281
sb->rx = rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1290
bool inner, bool rx)
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1292
sb->rx = rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1300
bool inner, bool rx)
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1302
sb->rx = rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1310
bool inner, bool rx)
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1312
sb->rx = rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1320
bool inner, bool rx)
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1322
sb->rx = rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1331
bool inner, bool rx)
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1333
sb->rx = rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1343
bool inner, bool rx)
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1348
sb->rx = rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1357
bool inner, bool rx)
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1359
sb->rx = rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1368
bool inner, bool rx)
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1370
sb->rx = rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1380
bool inner, bool rx)
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1382
sb->rx = rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1391
bool inner, bool rx)
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1393
sb->rx = rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1401
bool inner, bool rx)
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1403
sb->rx = rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1412
bool inner, bool rx)
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1417
sb->rx = rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1426
bool inner, bool rx)
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1428
sb->rx = rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1436
bool inner, bool rx)
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1438
sb->rx = rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1446
bool inner, bool rx)
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1448
sb->rx = rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1024
sb->lu_type = DR_STE_CALC_LU_TYPE(ETHL2_SRC, sb->rx, sb->inner);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1061
sb->lu_type = DR_STE_CALC_LU_TYPE(ETHL2_DST, sb->rx, sb->inner);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1170
sb->lu_type = DR_STE_CALC_LU_TYPE(ETHL3_IPV4_MISC, sb->rx, sb->inner);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1212
sb->lu_type = DR_STE_CALC_LU_TYPE(ETHL4, sb->rx, sb->inner);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1238
sb->lu_type = DR_STE_CALC_LU_TYPE(MPLS_FIRST, sb->rx, sb->inner);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1519
sb->lu_type = DR_STE_CALC_LU_TYPE(ETHL4_MISC, sb->rx, sb->inner);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
37
#define DR_STE_CALC_LU_TYPE(lookup_type, rx, inner) \
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
39
(rx) ? DR_STE_V0_LU_TYPE_##lookup_type##_D : \
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
788
sb->lu_type = DR_STE_CALC_LU_TYPE(ETHL2_SRC_DST, sb->rx, sb->inner);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
814
sb->lu_type = DR_STE_CALC_LU_TYPE(ETHL3_IPV6_DST, sb->rx, sb->inner);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
840
sb->lu_type = DR_STE_CALC_LU_TYPE(ETHL3_IPV6_SRC, sb->rx, sb->inner);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
877
sb->lu_type = DR_STE_CALC_LU_TYPE(ETHL3_IPV4_5_TUPLE, sb->rx, sb->inner);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_table.c
104
dr_table_uninit_nic(&tbl->rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_table.c
163
ret = dr_table_init_nic(tbl->dmn, &tbl->rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_table.c
174
dr_table_uninit_nic(&tbl->rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_table.c
189
tbl->rx.nic_dmn = &tbl->dmn->info.rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_table.c
190
ret = dr_table_init_nic(tbl->dmn, &tbl->rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_table.c
199
tbl->rx.nic_dmn = &tbl->dmn->info.rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_table.c
229
if (tbl->rx.s_anchor)
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_table.c
230
icm_addr_rx = mlx5dr_icm_pool_get_chunk_icm_addr(tbl->rx.s_anchor->chunk);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_table.c
28
action->dest_tbl->tbl->rx.s_anchor->chunk :
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_table.c
58
ret = dr_table_set_miss_action_nic(tbl->dmn, &tbl->rx, action);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_table.c
94
dr_table_uninit_nic(&tbl->rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
1158
struct mlx5dr_rule_rx_tx rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
1199
mlx5dr_domain_nic_lock(&dmn->info.rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
1206
mlx5dr_domain_nic_unlock(&dmn->info.rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
216
u8 rx:1;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
394
bool inner, bool rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
398
bool inner, bool rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
402
bool inner, bool rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
406
bool inner, bool rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
410
bool inner, bool rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
414
bool inner, bool rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
418
bool inner, bool rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
422
bool inner, bool rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
426
bool inner, bool rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
430
bool inner, bool rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
434
bool inner, bool rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
438
bool inner, bool rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
443
bool inner, bool rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
448
bool inner, bool rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
453
bool inner, bool rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
457
bool inner, bool rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
461
bool inner, bool rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
466
bool inner, bool rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
471
bool inner, bool rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
475
bool inner, bool rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
480
bool inner, bool rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
485
bool inner, bool rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
489
bool inner, bool rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
493
bool inner, bool rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
497
bool inner, bool rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
501
bool inner, bool rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
506
bool inner, bool rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
510
bool inner, bool rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
514
bool inner, bool rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
515
void mlx5dr_ste_build_empty_always_hit(struct mlx5dr_ste_build *sb, bool rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
928
struct mlx5dr_domain_rx_tx rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
968
struct mlx5dr_table_rx_tx rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
998
struct mlx5dr_matcher_rx_tx rx;
drivers/net/ethernet/meta/fbnic/fbnic_ethtool.c
163
memset(clone->rx, 0, sizeof(clone->rx));
drivers/net/ethernet/meta/fbnic/fbnic_ethtool.c
216
for (i = 0; i < ARRAY_SIZE(orig->rx); i++)
drivers/net/ethernet/meta/fbnic/fbnic_ethtool.c
217
swap(clone->rx[i], orig->rx[i]);
drivers/net/ethernet/meta/fbnic/fbnic_netdev.c
475
struct fbnic_ring *rxr = fbn->rx[i];
drivers/net/ethernet/meta/fbnic/fbnic_netdev.c
486
rx_length = stats->rx.length_errors;
drivers/net/ethernet/meta/fbnic/fbnic_netdev.c
562
struct netdev_queue_stats_rx *rx)
drivers/net/ethernet/meta/fbnic/fbnic_netdev.c
566
struct fbnic_ring *rxr = fbn->rx[idx];
drivers/net/ethernet/meta/fbnic/fbnic_netdev.c
584
alloc_fail = stats->rx.alloc_failed;
drivers/net/ethernet/meta/fbnic/fbnic_netdev.c
585
csum_complete = stats->rx.csum_complete;
drivers/net/ethernet/meta/fbnic/fbnic_netdev.c
586
csum_none = stats->rx.csum_none;
drivers/net/ethernet/meta/fbnic/fbnic_netdev.c
603
rx->bytes = bytes;
drivers/net/ethernet/meta/fbnic/fbnic_netdev.c
604
rx->packets = packets;
drivers/net/ethernet/meta/fbnic/fbnic_netdev.c
605
rx->alloc_fail = alloc_fail;
drivers/net/ethernet/meta/fbnic/fbnic_netdev.c
606
rx->csum_complete = csum_complete;
drivers/net/ethernet/meta/fbnic/fbnic_netdev.c
607
rx->csum_none = csum_none;
drivers/net/ethernet/meta/fbnic/fbnic_netdev.c
612
rx->hw_drop_overruns = fbd->hw_stats.hw_q[idx].rde_pkt_cq_drop.value +
drivers/net/ethernet/meta/fbnic/fbnic_netdev.c
614
rx->hw_drops = fbd->hw_stats.hw_q[idx].rde_pkt_err.value +
drivers/net/ethernet/meta/fbnic/fbnic_netdev.c
615
rx->hw_drop_overruns;
drivers/net/ethernet/meta/fbnic/fbnic_netdev.c
666
struct netdev_queue_stats_rx *rx,
drivers/net/ethernet/meta/fbnic/fbnic_netdev.c
678
rx->bytes = fbn->rx_stats.bytes;
drivers/net/ethernet/meta/fbnic/fbnic_netdev.c
679
rx->packets = fbn->rx_stats.packets;
drivers/net/ethernet/meta/fbnic/fbnic_netdev.c
680
rx->alloc_fail = fbn->rx_stats.rx.alloc_failed +
drivers/net/ethernet/meta/fbnic/fbnic_netdev.c
682
rx->csum_complete = fbn->rx_stats.rx.csum_complete;
drivers/net/ethernet/meta/fbnic/fbnic_netdev.c
683
rx->csum_none = fbn->rx_stats.rx.csum_none;
drivers/net/ethernet/meta/fbnic/fbnic_netdev.c
693
unsigned int tx, unsigned int rx)
drivers/net/ethernet/meta/fbnic/fbnic_netdev.c
703
rx = min(rx, max_napis);
drivers/net/ethernet/meta/fbnic/fbnic_netdev.c
704
fbn->num_rx_queues = rx;
drivers/net/ethernet/meta/fbnic/fbnic_netdev.c
706
fbn->num_napi = max(tx, rx);
drivers/net/ethernet/meta/fbnic/fbnic_netdev.h
24
struct fbnic_ring *rx[FBNIC_MAX_RXQS];
drivers/net/ethernet/meta/fbnic/fbnic_netdev.h
90
unsigned int tx, unsigned int rx);
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
1327
rcq->stats.rx.alloc_failed += alloc_failed;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
1328
rcq->stats.rx.csum_complete += csum_complete;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
1329
rcq->stats.rx.csum_none += csum_none;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
1330
rcq->stats.rx.length_errors += length_errors;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
1411
fbn->rx_stats.rx.alloc_failed += stats->rx.alloc_failed;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
1412
fbn->rx_stats.rx.csum_complete += stats->rx.csum_complete;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
1413
fbn->rx_stats.rx.csum_none += stats->rx.csum_none;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
1414
fbn->rx_stats.rx.length_errors += stats->rx.length_errors;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
1416
BUILD_BUG_ON(sizeof(fbn->rx_stats.rx) / 8 != 4);
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
1498
WARN_ON(fbn->rx[rxr->q_idx] && fbn->rx[rxr->q_idx] != rxr);
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
1499
fbn->rx[rxr->q_idx] = NULL;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
1741
fbn->rx[rxq_idx] = &qt->cmpl;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
2339
}, rx[] = {
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
2363
false, fbd, rx, ARRAY_SIZE(rx));
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
2365
fbnic_idle_dump(fbd, rx, ARRAY_SIZE(rx), "Rx", err);
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
2370
fbnic_wait_queue_idle(struct fbnic_net *fbn, bool rx, unsigned int idx)
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
2385
regs = rx ? rx_regs : tx_regs;
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
2386
reg_cnt = rx ? ARRAY_SIZE(rx_regs) : ARRAY_SIZE(tx_regs);
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
2398
rx ? "Rx" : "Tx", idx, regs[i] + off, i,
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
2843
real = container_of(fbn->rx[idx], struct fbnic_q_triad, cmpl);
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
2894
real = container_of(fbn->rx[idx], struct fbnic_q_triad, cmpl);
drivers/net/ethernet/meta/fbnic/fbnic_txrx.c
2916
real = container_of(fbn->rx[idx], struct fbnic_q_triad, cmpl);
drivers/net/ethernet/meta/fbnic/fbnic_txrx.h
94
} rx;
drivers/net/ethernet/micrel/ks8851_spi.c
196
__le16 rx = 0;
drivers/net/ethernet/micrel/ks8851_spi.c
198
ks8851_rdreg(ks, MK_OP(reg & 2 ? 0xC : 0x3, reg), (u8 *)&rx, 2);
drivers/net/ethernet/micrel/ks8851_spi.c
199
return le16_to_cpu(rx);
drivers/net/ethernet/micrel/ksz884x.c
1518
status.rx.hw_owned = 0;
drivers/net/ethernet/micrel/ksz884x.c
1548
desc->sw.buf.rx.buf_size = len;
drivers/net/ethernet/micrel/ksz884x.c
2836
static void set_flow_ctrl(struct ksz_hw *hw, int rx, int tx)
drivers/net/ethernet/micrel/ksz884x.c
2843
if (rx)
drivers/net/ethernet/micrel/ksz884x.c
2862
int rx;
drivers/net/ethernet/micrel/ksz884x.c
2868
rx = tx = 0;
drivers/net/ethernet/micrel/ksz884x.c
2870
rx = tx = 1;
drivers/net/ethernet/micrel/ksz884x.c
2873
rx = tx = 1;
drivers/net/ethernet/micrel/ksz884x.c
2883
rx = 1;
drivers/net/ethernet/micrel/ksz884x.c
2886
set_flow_ctrl(hw, rx, tx);
drivers/net/ethernet/micrel/ksz884x.c
3471
previous->sw.buf.rx.end_of_ring = 1;
drivers/net/ethernet/micrel/ksz884x.c
4617
packet_len = status.rx.frame_len - 4;
drivers/net/ethernet/micrel/ksz884x.c
4671
if (status.rx.hw_owned)
drivers/net/ethernet/micrel/ksz884x.c
4675
if (status.rx.last_desc && status.rx.first_desc) {
drivers/net/ethernet/micrel/ksz884x.c
4707
if (status.rx.hw_owned)
drivers/net/ethernet/micrel/ksz884x.c
4712
int p = HW_TO_DEV_PORT(status.rx.src_port);
drivers/net/ethernet/micrel/ksz884x.c
4720
if (status.rx.last_desc && status.rx.first_desc) {
drivers/net/ethernet/micrel/ksz884x.c
4752
if (status.rx.hw_owned)
drivers/net/ethernet/micrel/ksz884x.c
4757
int p = HW_TO_DEV_PORT(status.rx.src_port);
drivers/net/ethernet/micrel/ksz884x.c
4765
if (status.rx.last_desc && status.rx.first_desc) {
drivers/net/ethernet/micrel/ksz884x.c
4771
if (!status.rx.error || (status.data &
drivers/net/ethernet/micrel/ksz884x.c
854
struct ksz_desc_rx_stat rx;
drivers/net/ethernet/micrel/ksz884x.c
860
struct ksz_desc_rx_buf rx;
drivers/net/ethernet/microchip/enc28j60.c
104
spi_message_add_tail(&rx, &msg);
drivers/net/ethernet/microchip/enc28j60.c
93
struct spi_transfer rx = {
drivers/net/ethernet/microchip/lan743x_ethtool.c
861
for (i = 0; i < ARRAY_SIZE(adapter->rx); i++)
drivers/net/ethernet/microchip/lan743x_ethtool.c
862
data[data_index++] = (u64)(adapter->rx[i].frame_count);
drivers/net/ethernet/microchip/lan743x_main.c
1694
channel_number = adapter->rx[index].channel_number;
drivers/net/ethernet/microchip/lan743x_main.c
2304
static int lan743x_rx_next_index(struct lan743x_rx *rx, int index)
drivers/net/ethernet/microchip/lan743x_main.c
2306
return ((++index) % rx->ring_size);
drivers/net/ethernet/microchip/lan743x_main.c
2309
static void lan743x_rx_update_tail(struct lan743x_rx *rx, int index)
drivers/net/ethernet/microchip/lan743x_main.c
2313
lan743x_csr_write(rx->adapter, RX_TAIL(rx->channel_number),
drivers/net/ethernet/microchip/lan743x_main.c
2317
static int lan743x_rx_init_ring_element(struct lan743x_rx *rx, int index,
drivers/net/ethernet/microchip/lan743x_main.c
2320
struct net_device *netdev = rx->adapter->netdev;
drivers/net/ethernet/microchip/lan743x_main.c
2321
struct device *dev = &rx->adapter->pdev->dev;
drivers/net/ethernet/microchip/lan743x_main.c
2330
descriptor = &rx->ring_cpu_ptr[index];
drivers/net/ethernet/microchip/lan743x_main.c
2331
buffer_info = &rx->buffer_info[index];
drivers/net/ethernet/microchip/lan743x_main.c
2369
lan743x_rx_update_tail(rx, index);
drivers/net/ethernet/microchip/lan743x_main.c
2374
static void lan743x_rx_reuse_ring_element(struct lan743x_rx *rx, int index)
drivers/net/ethernet/microchip/lan743x_main.c
2379
descriptor = &rx->ring_cpu_ptr[index];
drivers/net/ethernet/microchip/lan743x_main.c
2380
buffer_info = &rx->buffer_info[index];
drivers/net/ethernet/microchip/lan743x_main.c
2388
lan743x_rx_update_tail(rx, index);
drivers/net/ethernet/microchip/lan743x_main.c
2391
static void lan743x_rx_release_ring_element(struct lan743x_rx *rx, int index)
drivers/net/ethernet/microchip/lan743x_main.c
2396
descriptor = &rx->ring_cpu_ptr[index];
drivers/net/ethernet/microchip/lan743x_main.c
2397
buffer_info = &rx->buffer_info[index];
drivers/net/ethernet/microchip/lan743x_main.c
2402
dma_unmap_single(&rx->adapter->pdev->dev,
drivers/net/ethernet/microchip/lan743x_main.c
2432
static int lan743x_rx_process_buffer(struct lan743x_rx *rx)
drivers/net/ethernet/microchip/lan743x_main.c
2434
int current_head_index = le32_to_cpu(*rx->head_cpu_ptr);
drivers/net/ethernet/microchip/lan743x_main.c
2436
struct net_device *netdev = rx->adapter->netdev;
drivers/net/ethernet/microchip/lan743x_main.c
2445
if (current_head_index < 0 || current_head_index >= rx->ring_size)
drivers/net/ethernet/microchip/lan743x_main.c
2448
if (rx->last_head < 0 || rx->last_head >= rx->ring_size)
drivers/net/ethernet/microchip/lan743x_main.c
2451
if (rx->last_head == current_head_index)
drivers/net/ethernet/microchip/lan743x_main.c
2454
descriptor = &rx->ring_cpu_ptr[rx->last_head];
drivers/net/ethernet/microchip/lan743x_main.c
2457
buffer_info = &rx->buffer_info[rx->last_head];
drivers/net/ethernet/microchip/lan743x_main.c
2464
int index = lan743x_rx_next_index(rx, rx->last_head);
drivers/net/ethernet/microchip/lan743x_main.c
2469
desc_ext = &rx->ring_cpu_ptr[index];
drivers/net/ethernet/microchip/lan743x_main.c
2498
if (lan743x_rx_init_ring_element(rx, rx->last_head, GFP_ATOMIC)) {
drivers/net/ethernet/microchip/lan743x_main.c
2503
lan743x_rx_reuse_ring_element(rx, rx->last_head);
drivers/net/ethernet/microchip/lan743x_main.c
2505
dev_kfree_skb_irq(rx->skb_head);
drivers/net/ethernet/microchip/lan743x_main.c
2506
rx->skb_head = NULL;
drivers/net/ethernet/microchip/lan743x_main.c
2514
if (rx->skb_head)
drivers/net/ethernet/microchip/lan743x_main.c
2515
dev_kfree_skb_irq(rx->skb_head);
drivers/net/ethernet/microchip/lan743x_main.c
2516
rx->skb_head = skb;
drivers/net/ethernet/microchip/lan743x_main.c
2517
} else if (rx->skb_head) {
drivers/net/ethernet/microchip/lan743x_main.c
2519
if (skb_shinfo(rx->skb_head)->frag_list)
drivers/net/ethernet/microchip/lan743x_main.c
2520
rx->skb_tail->next = skb;
drivers/net/ethernet/microchip/lan743x_main.c
2522
skb_shinfo(rx->skb_head)->frag_list = skb;
drivers/net/ethernet/microchip/lan743x_main.c
2523
rx->skb_tail = skb;
drivers/net/ethernet/microchip/lan743x_main.c
2524
rx->skb_head->len += skb->len;
drivers/net/ethernet/microchip/lan743x_main.c
2525
rx->skb_head->data_len += skb->len;
drivers/net/ethernet/microchip/lan743x_main.c
2526
rx->skb_head->truesize += skb->truesize;
drivers/net/ethernet/microchip/lan743x_main.c
2543
if (rx->skb_head)
drivers/net/ethernet/microchip/lan743x_main.c
2544
skb_hwtstamps(rx->skb_head)->hwtstamp =
drivers/net/ethernet/microchip/lan743x_main.c
2546
lan743x_rx_reuse_ring_element(rx, extension_index);
drivers/net/ethernet/microchip/lan743x_main.c
2547
rx->last_head = extension_index;
drivers/net/ethernet/microchip/lan743x_main.c
2551
if (is_last && rx->skb_head)
drivers/net/ethernet/microchip/lan743x_main.c
2552
rx->skb_head = lan743x_rx_trim_skb(rx->skb_head, frame_length);
drivers/net/ethernet/microchip/lan743x_main.c
2554
if (is_last && rx->skb_head) {
drivers/net/ethernet/microchip/lan743x_main.c
2555
rx->skb_head->protocol = eth_type_trans(rx->skb_head,
drivers/net/ethernet/microchip/lan743x_main.c
2556
rx->adapter->netdev);
drivers/net/ethernet/microchip/lan743x_main.c
2557
if (rx->adapter->netdev->features & NETIF_F_RXCSUM) {
drivers/net/ethernet/microchip/lan743x_main.c
2562
rx->skb_head->len);
drivers/net/ethernet/microchip/lan743x_main.c
2563
napi_gro_receive(&rx->napi, rx->skb_head);
drivers/net/ethernet/microchip/lan743x_main.c
2564
rx->skb_head = NULL;
drivers/net/ethernet/microchip/lan743x_main.c
2569
rx->last_tail = rx->last_head;
drivers/net/ethernet/microchip/lan743x_main.c
2570
rx->last_head = lan743x_rx_next_index(rx, rx->last_head);
drivers/net/ethernet/microchip/lan743x_main.c
2578
struct lan743x_rx *rx = container_of(napi, struct lan743x_rx, napi);
drivers/net/ethernet/microchip/lan743x_main.c
2579
struct lan743x_adapter *adapter = rx->adapter;
drivers/net/ethernet/microchip/lan743x_main.c
2584
if (rx->vector_flags & LAN743X_VECTOR_FLAG_SOURCE_STATUS_W2C) {
drivers/net/ethernet/microchip/lan743x_main.c
2587
DMAC_INT_BIT_RXFRM_(rx->channel_number));
drivers/net/ethernet/microchip/lan743x_main.c
2590
result = lan743x_rx_process_buffer(rx);
drivers/net/ethernet/microchip/lan743x_main.c
2594
rx->frame_count += count;
drivers/net/ethernet/microchip/lan743x_main.c
2602
if (rx->vector_flags & LAN743X_VECTOR_FLAG_VECTOR_ENABLE_AUTO_SET)
drivers/net/ethernet/microchip/lan743x_main.c
2604
if (rx->vector_flags & LAN743X_VECTOR_FLAG_SOURCE_ENABLE_AUTO_SET) {
drivers/net/ethernet/microchip/lan743x_main.c
2608
INT_BIT_DMA_RX_(rx->channel_number));
drivers/net/ethernet/microchip/lan743x_main.c
2612
lan743x_csr_write(adapter, RX_TAIL(rx->channel_number),
drivers/net/ethernet/microchip/lan743x_main.c
2613
rx_tail_flags | rx->last_tail);
drivers/net/ethernet/microchip/lan743x_main.c
2618
static void lan743x_rx_ring_cleanup(struct lan743x_rx *rx)
drivers/net/ethernet/microchip/lan743x_main.c
262
struct lan743x_rx *rx = context;
drivers/net/ethernet/microchip/lan743x_main.c
2620
if (rx->buffer_info && rx->ring_cpu_ptr) {
drivers/net/ethernet/microchip/lan743x_main.c
2623
for (index = 0; index < rx->ring_size; index++)
drivers/net/ethernet/microchip/lan743x_main.c
2624
lan743x_rx_release_ring_element(rx, index);
drivers/net/ethernet/microchip/lan743x_main.c
2627
if (rx->head_cpu_ptr) {
drivers/net/ethernet/microchip/lan743x_main.c
2628
dma_free_coherent(&rx->adapter->pdev->dev,
drivers/net/ethernet/microchip/lan743x_main.c
2629
sizeof(*rx->head_cpu_ptr), rx->head_cpu_ptr,
drivers/net/ethernet/microchip/lan743x_main.c
263
struct lan743x_adapter *adapter = rx->adapter;
drivers/net/ethernet/microchip/lan743x_main.c
2630
rx->head_dma_ptr);
drivers/net/ethernet/microchip/lan743x_main.c
2631
rx->head_cpu_ptr = NULL;
drivers/net/ethernet/microchip/lan743x_main.c
2632
rx->head_dma_ptr = 0;
drivers/net/ethernet/microchip/lan743x_main.c
2635
kfree(rx->buffer_info);
drivers/net/ethernet/microchip/lan743x_main.c
2636
rx->buffer_info = NULL;
drivers/net/ethernet/microchip/lan743x_main.c
2638
if (rx->ring_cpu_ptr) {
drivers/net/ethernet/microchip/lan743x_main.c
2639
dma_free_coherent(&rx->adapter->pdev->dev,
drivers/net/ethernet/microchip/lan743x_main.c
2640
rx->ring_allocation_size, rx->ring_cpu_ptr,
drivers/net/ethernet/microchip/lan743x_main.c
2641
rx->ring_dma_ptr);
drivers/net/ethernet/microchip/lan743x_main.c
2642
rx->ring_allocation_size = 0;
drivers/net/ethernet/microchip/lan743x_main.c
2643
rx->ring_cpu_ptr = NULL;
drivers/net/ethernet/microchip/lan743x_main.c
2644
rx->ring_dma_ptr = 0;
drivers/net/ethernet/microchip/lan743x_main.c
2647
rx->ring_size = 0;
drivers/net/ethernet/microchip/lan743x_main.c
2648
rx->last_head = 0;
drivers/net/ethernet/microchip/lan743x_main.c
2651
static int lan743x_rx_ring_init(struct lan743x_rx *rx)
drivers/net/ethernet/microchip/lan743x_main.c
2659
rx->ring_size = LAN743X_RX_RING_SIZE;
drivers/net/ethernet/microchip/lan743x_main.c
2660
if (rx->ring_size <= 1) {
drivers/net/ethernet/microchip/lan743x_main.c
2664
if (rx->ring_size & ~RX_CFG_B_RX_RING_LEN_MASK_) {
drivers/net/ethernet/microchip/lan743x_main.c
2668
if (dma_set_mask_and_coherent(&rx->adapter->pdev->dev,
drivers/net/ethernet/microchip/lan743x_main.c
2670
dev_warn(&rx->adapter->pdev->dev,
drivers/net/ethernet/microchip/lan743x_main.c
2675
ring_allocation_size = ALIGN(rx->ring_size *
drivers/net/ethernet/microchip/lan743x_main.c
2679
cpu_ptr = dma_alloc_coherent(&rx->adapter->pdev->dev,
drivers/net/ethernet/microchip/lan743x_main.c
268
INT_BIT_DMA_RX_(rx->channel_number));
drivers/net/ethernet/microchip/lan743x_main.c
2685
rx->ring_allocation_size = ring_allocation_size;
drivers/net/ethernet/microchip/lan743x_main.c
2686
rx->ring_cpu_ptr = (struct lan743x_rx_descriptor *)cpu_ptr;
drivers/net/ethernet/microchip/lan743x_main.c
2687
rx->ring_dma_ptr = dma_ptr;
drivers/net/ethernet/microchip/lan743x_main.c
2689
cpu_ptr = kzalloc_objs(*rx->buffer_info, rx->ring_size);
drivers/net/ethernet/microchip/lan743x_main.c
2694
rx->buffer_info = (struct lan743x_rx_buffer_info *)cpu_ptr;
drivers/net/ethernet/microchip/lan743x_main.c
2696
cpu_ptr = dma_alloc_coherent(&rx->adapter->pdev->dev,
drivers/net/ethernet/microchip/lan743x_main.c
2697
sizeof(*rx->head_cpu_ptr), &dma_ptr,
drivers/net/ethernet/microchip/lan743x_main.c
2704
rx->head_cpu_ptr = cpu_ptr;
drivers/net/ethernet/microchip/lan743x_main.c
2705
rx->head_dma_ptr = dma_ptr;
drivers/net/ethernet/microchip/lan743x_main.c
2706
if (rx->head_dma_ptr & 0x3) {
drivers/net/ethernet/microchip/lan743x_main.c
271
if (int_sts & INT_BIT_DMA_RX_(rx->channel_number)) {
drivers/net/ethernet/microchip/lan743x_main.c
2711
rx->last_head = 0;
drivers/net/ethernet/microchip/lan743x_main.c
2712
for (index = 0; index < rx->ring_size; index++) {
drivers/net/ethernet/microchip/lan743x_main.c
2713
ret = lan743x_rx_init_ring_element(rx, index, GFP_KERNEL);
drivers/net/ethernet/microchip/lan743x_main.c
272
u32 rx_frame_bit = DMAC_INT_BIT_RXFRM_(rx->channel_number);
drivers/net/ethernet/microchip/lan743x_main.c
2720
netif_warn(rx->adapter, ifup, rx->adapter->netdev,
drivers/net/ethernet/microchip/lan743x_main.c
2723
lan743x_rx_ring_cleanup(rx);
drivers/net/ethernet/microchip/lan743x_main.c
2727
static void lan743x_rx_close(struct lan743x_rx *rx)
drivers/net/ethernet/microchip/lan743x_main.c
2729
struct lan743x_adapter *adapter = rx->adapter;
drivers/net/ethernet/microchip/lan743x_main.c
2732
FCT_RX_CTL_DIS_(rx->channel_number));
drivers/net/ethernet/microchip/lan743x_main.c
2734
FCT_RX_CTL_EN_(rx->channel_number),
drivers/net/ethernet/microchip/lan743x_main.c
2738
DMAC_CMD_STOP_R_(rx->channel_number));
drivers/net/ethernet/microchip/lan743x_main.c
2739
lan743x_dmac_rx_wait_till_stopped(adapter, rx->channel_number);
drivers/net/ethernet/microchip/lan743x_main.c
2742
DMAC_INT_BIT_RXFRM_(rx->channel_number));
drivers/net/ethernet/microchip/lan743x_main.c
2744
INT_BIT_DMA_RX_(rx->channel_number));
drivers/net/ethernet/microchip/lan743x_main.c
2745
napi_disable(&rx->napi);
drivers/net/ethernet/microchip/lan743x_main.c
2747
netif_napi_del(&rx->napi);
drivers/net/ethernet/microchip/lan743x_main.c
2749
lan743x_rx_ring_cleanup(rx);
drivers/net/ethernet/microchip/lan743x_main.c
2752
static int lan743x_rx_open(struct lan743x_rx *rx)
drivers/net/ethernet/microchip/lan743x_main.c
2754
struct lan743x_adapter *adapter = rx->adapter;
drivers/net/ethernet/microchip/lan743x_main.c
2758
rx->frame_count = 0;
drivers/net/ethernet/microchip/lan743x_main.c
2759
ret = lan743x_rx_ring_init(rx);
drivers/net/ethernet/microchip/lan743x_main.c
2763
netif_napi_add(adapter->netdev, &rx->napi, lan743x_rx_napi_poll);
drivers/net/ethernet/microchip/lan743x_main.c
2766
DMAC_CMD_RX_SWR_(rx->channel_number));
drivers/net/ethernet/microchip/lan743x_main.c
2768
DMAC_CMD_RX_SWR_(rx->channel_number),
drivers/net/ethernet/microchip/lan743x_main.c
2773
RX_BASE_ADDRH(rx->channel_number),
drivers/net/ethernet/microchip/lan743x_main.c
2774
DMA_ADDR_HIGH32(rx->ring_dma_ptr));
drivers/net/ethernet/microchip/lan743x_main.c
2776
RX_BASE_ADDRL(rx->channel_number),
drivers/net/ethernet/microchip/lan743x_main.c
2777
DMA_ADDR_LOW32(rx->ring_dma_ptr));
drivers/net/ethernet/microchip/lan743x_main.c
2781
RX_HEAD_WRITEBACK_ADDRH(rx->channel_number),
drivers/net/ethernet/microchip/lan743x_main.c
2782
DMA_ADDR_HIGH32(rx->head_dma_ptr));
drivers/net/ethernet/microchip/lan743x_main.c
2784
RX_HEAD_WRITEBACK_ADDRL(rx->channel_number),
drivers/net/ethernet/microchip/lan743x_main.c
2785
DMA_ADDR_LOW32(rx->head_dma_ptr));
drivers/net/ethernet/microchip/lan743x_main.c
2796
RX_CFG_A(rx->channel_number), data);
drivers/net/ethernet/microchip/lan743x_main.c
2799
data = lan743x_csr_read(adapter, RX_CFG_B(rx->channel_number));
drivers/net/ethernet/microchip/lan743x_main.c
2806
data |= ((rx->ring_size) & RX_CFG_B_RX_RING_LEN_MASK_);
drivers/net/ethernet/microchip/lan743x_main.c
2810
lan743x_csr_write(adapter, RX_CFG_B(rx->channel_number), data);
drivers/net/ethernet/microchip/lan743x_main.c
2811
rx->vector_flags = lan743x_intr_get_vector_flags(adapter,
drivers/net/ethernet/microchip/lan743x_main.c
2813
(rx->channel_number));
drivers/net/ethernet/microchip/lan743x_main.c
2817
if (rx->vector_flags & LAN743X_VECTOR_FLAG_SOURCE_ENABLE_AUTO_CLEAR)
drivers/net/ethernet/microchip/lan743x_main.c
2819
if (rx->vector_flags & LAN743X_VECTOR_FLAG_SOURCE_STATUS_AUTO_CLEAR)
drivers/net/ethernet/microchip/lan743x_main.c
2821
if (rx->vector_flags & LAN743X_VECTOR_FLAG_SOURCE_STATUS_R2C)
drivers/net/ethernet/microchip/lan743x_main.c
2823
if (rx->vector_flags & LAN743X_VECTOR_FLAG_SOURCE_ENABLE_R2C)
drivers/net/ethernet/microchip/lan743x_main.c
2825
lan743x_csr_write(adapter, RX_CFG_C(rx->channel_number), data);
drivers/net/ethernet/microchip/lan743x_main.c
2827
rx->last_tail = ((u32)(rx->ring_size - 1));
drivers/net/ethernet/microchip/lan743x_main.c
2828
lan743x_csr_write(adapter, RX_TAIL(rx->channel_number),
drivers/net/ethernet/microchip/lan743x_main.c
2829
rx->last_tail);
drivers/net/ethernet/microchip/lan743x_main.c
2830
rx->last_head = lan743x_csr_read(adapter, RX_HEAD(rx->channel_number));
drivers/net/ethernet/microchip/lan743x_main.c
2831
if (rx->last_head) {
drivers/net/ethernet/microchip/lan743x_main.c
2836
napi_enable(&rx->napi);
drivers/net/ethernet/microchip/lan743x_main.c
2839
INT_BIT_DMA_RX_(rx->channel_number));
drivers/net/ethernet/microchip/lan743x_main.c
2841
DMAC_INT_BIT_RXFRM_(rx->channel_number));
drivers/net/ethernet/microchip/lan743x_main.c
2843
DMAC_INT_BIT_RXFRM_(rx->channel_number));
drivers/net/ethernet/microchip/lan743x_main.c
2845
DMAC_CMD_START_R_(rx->channel_number));
drivers/net/ethernet/microchip/lan743x_main.c
2849
FCT_RX_CTL_RESET_(rx->channel_number));
drivers/net/ethernet/microchip/lan743x_main.c
2851
FCT_RX_CTL_RESET_(rx->channel_number),
drivers/net/ethernet/microchip/lan743x_main.c
2853
lan743x_csr_write(adapter, FCT_FLOW(rx->channel_number),
drivers/net/ethernet/microchip/lan743x_main.c
2860
FCT_RX_CTL_EN_(rx->channel_number));
drivers/net/ethernet/microchip/lan743x_main.c
2864
netif_napi_del(&rx->napi);
drivers/net/ethernet/microchip/lan743x_main.c
2865
lan743x_rx_ring_cleanup(rx);
drivers/net/ethernet/microchip/lan743x_main.c
289
napi_schedule(&rx->napi);
drivers/net/ethernet/microchip/lan743x_main.c
297
INT_BIT_DMA_RX_(rx->channel_number));
drivers/net/ethernet/microchip/lan743x_main.c
312
lan743x_rx_isr(&adapter->rx[channel],
drivers/net/ethernet/microchip/lan743x_main.c
3244
lan743x_rx_close(&adapter->rx[index]);
drivers/net/ethernet/microchip/lan743x_main.c
3282
ret = lan743x_rx_open(&adapter->rx[index]);
drivers/net/ethernet/microchip/lan743x_main.c
3316
if (adapter->rx[index].ring_cpu_ptr)
drivers/net/ethernet/microchip/lan743x_main.c
3317
lan743x_rx_close(&adapter->rx[index]);
drivers/net/ethernet/microchip/lan743x_main.c
3547
adapter->rx[index].adapter = adapter;
drivers/net/ethernet/microchip/lan743x_main.c
3548
adapter->rx[index].channel_number = index;
drivers/net/ethernet/microchip/lan743x_main.c
732
&adapter->rx[index]);
drivers/net/ethernet/microchip/lan743x_main.h
1070
struct lan743x_rx rx[LAN743X_USED_RX_CHANNELS];
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
104
rx->page_pool);
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
107
return PTR_ERR_OR_ZERO(rx->page_pool);
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
110
static int lan966x_fdma_rx_alloc(struct lan966x_rx *rx)
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
112
struct lan966x *lan966x = rx->lan966x;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
113
struct fdma *fdma = &rx->fdma;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
116
if (lan966x_fdma_rx_alloc_page_pool(rx))
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
117
return PTR_ERR(rx->page_pool);
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
129
static void lan966x_fdma_rx_start(struct lan966x_rx *rx)
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
13
struct lan966x_rx *rx = &lan966x->rx;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
131
struct lan966x *lan966x = rx->lan966x;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
132
struct fdma *fdma = &rx->fdma;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
16
page = page_pool_dev_alloc_pages(rx->page_pool);
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
168
static void lan966x_fdma_rx_disable(struct lan966x_rx *rx)
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
170
struct lan966x *lan966x = rx->lan966x;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
171
struct fdma *fdma = &rx->fdma;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
188
static void lan966x_fdma_rx_reload(struct lan966x_rx *rx)
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
190
struct lan966x *lan966x = rx->lan966x;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
192
lan_rmw(FDMA_CH_RELOAD_CH_RELOAD_SET(BIT(rx->fdma.channel_id)),
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
20
rx->page[dcb][db] = page;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
331
struct lan966x_rx *rx = &lan966x->rx;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
375
page_pool_recycle_direct(rx->page_pool,
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
390
static int lan966x_fdma_rx_check_frame(struct lan966x_rx *rx, u64 *src_port)
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
392
struct lan966x *lan966x = rx->lan966x;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
393
struct fdma *fdma = &rx->fdma;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
399
page = rx->page[fdma->dcb_index][fdma->db_index];
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
420
static struct sk_buff *lan966x_fdma_rx_get_frame(struct lan966x_rx *rx,
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
423
struct lan966x *lan966x = rx->lan966x;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
424
struct fdma *fdma = &rx->fdma;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
432
page = rx->page[fdma->dcb_index][fdma->db_index];
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
468
page_pool_recycle_direct(rx->page_pool, page);
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
476
struct lan966x_rx *rx = &lan966x->rx;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
478
struct fdma *fdma = &rx->fdma;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
494
switch (lan966x_fdma_rx_check_frame(rx, &src_port)) {
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
498
lan966x_fdma_rx_free_page(rx);
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
508
lan966x_fdma_rx_free_page(rx);
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
51
static void lan966x_fdma_rx_free_pages(struct lan966x_rx *rx)
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
513
skb = lan966x_fdma_rx_get_frame(rx, src_port);
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
53
struct fdma *fdma = &rx->fdma;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
531
lan966x_fdma_rx_reload(rx);
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
58
page_pool_put_full_page(rx->page_pool,
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
59
rx->page[i][j], false);
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
63
static void lan966x_fdma_rx_free_page(struct lan966x_rx *rx)
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
65
struct fdma *fdma = &rx->fdma;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
68
page = rx->page[fdma->dcb_index][fdma->db_index];
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
72
page_pool_recycle_direct(rx->page_pool, page);
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
75
static int lan966x_fdma_rx_alloc_page_pool(struct lan966x_rx *rx)
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
77
struct lan966x *lan966x = rx->lan966x;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
79
.order = rx->page_order,
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
81
.pool_size = rx->fdma.n_dcbs,
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
816
memcpy(&fdma_rx_old, &lan966x->rx.fdma, sizeof(struct fdma));
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
817
page_pool = lan966x->rx.page_pool;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
823
lan966x_fdma_rx_disable(&lan966x->rx);
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
824
lan966x_fdma_rx_free_pages(&lan966x->rx);
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
825
lan966x->rx.page_order = round_up(new_mtu, PAGE_SIZE) / PAGE_SIZE - 1;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
826
lan966x->rx.max_mtu = new_mtu;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
827
err = lan966x_fdma_rx_alloc(&lan966x->rx);
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
830
lan966x_fdma_rx_start(&lan966x->rx);
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
841
lan966x->rx.page_pool = page_pool;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
842
memcpy(&lan966x->rx.fdma, &fdma_rx_old, sizeof(struct fdma));
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
843
lan966x_fdma_rx_start(&lan966x->rx);
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
86
.max_len = rx->max_mtu -
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
892
if (max_mtu == lan966x->rx.max_mtu)
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
93
rx->page_pool = page_pool_create(&pp_params);
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
931
lan966x->rx.lan966x = lan966x;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
932
lan966x->rx.fdma.channel_id = FDMA_XTR_CHANNEL;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
933
lan966x->rx.fdma.n_dcbs = FDMA_DCB_MAX;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
934
lan966x->rx.fdma.n_dbs = FDMA_RX_DCB_MAX_DBS;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
935
lan966x->rx.fdma.priv = lan966x;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
936
lan966x->rx.fdma.size = fdma_get_size(&lan966x->rx.fdma);
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
937
lan966x->rx.fdma.db_size = PAGE_SIZE << lan966x->rx.page_order;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
938
lan966x->rx.fdma.ops.nextptr_cb = &fdma_nextptr_cb;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
939
lan966x->rx.fdma.ops.dataptr_cb = &lan966x_fdma_rx_dataptr_cb;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
940
lan966x->rx.max_mtu = lan966x_fdma_get_max_frame(lan966x);
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
947
lan966x->tx.fdma.db_size = PAGE_SIZE << lan966x->rx.page_order;
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
951
err = lan966x_fdma_rx_alloc(&lan966x->rx);
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
957
fdma_free_coherent(lan966x->dev, &lan966x->rx.fdma);
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
961
lan966x_fdma_rx_start(&lan966x->rx);
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
971
lan966x_fdma_rx_disable(&lan966x->rx);
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
977
lan966x_fdma_rx_free_pages(&lan966x->rx);
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
978
fdma_free_coherent(lan966x->dev, &lan966x->rx.fdma);
drivers/net/ethernet/microchip/lan966x/lan966x_fdma.c
979
page_pool_destroy(lan966x->rx.page_pool);
drivers/net/ethernet/microchip/lan966x/lan966x_main.h
330
struct lan966x_rx rx;
drivers/net/ethernet/microchip/lan966x/lan966x_xdp.c
82
xdp_init_buff(&xdp, PAGE_SIZE << lan966x->rx.page_order,
drivers/net/ethernet/microchip/sparx5/lan969x/lan969x_fdma.c
101
struct sparx5_rx *rx)
drivers/net/ethernet/microchip/sparx5/lan969x/lan969x_fdma.c
104
struct fdma *fdma = &rx->fdma;
drivers/net/ethernet/microchip/sparx5/lan969x/lan969x_fdma.c
112
page = rx->page[fdma->dcb_index][fdma->db_index];
drivers/net/ethernet/microchip/sparx5/lan969x/lan969x_fdma.c
145
page_pool_recycle_direct(rx->page_pool, page);
drivers/net/ethernet/microchip/sparx5/lan969x/lan969x_fdma.c
152
struct sparx5_rx *rx = &sparx5->rx;
drivers/net/ethernet/microchip/sparx5/lan969x/lan969x_fdma.c
153
struct fdma *fdma = &rx->fdma;
drivers/net/ethernet/microchip/sparx5/lan969x/lan969x_fdma.c
168
rx->page_pool = page_pool_create(&pp_params);
drivers/net/ethernet/microchip/sparx5/lan969x/lan969x_fdma.c
169
if (IS_ERR(rx->page_pool))
drivers/net/ethernet/microchip/sparx5/lan969x/lan969x_fdma.c
170
return PTR_ERR(rx->page_pool);
drivers/net/ethernet/microchip/sparx5/lan969x/lan969x_fdma.c
208
struct fdma *fdma = &sparx5->rx.fdma;
drivers/net/ethernet/microchip/sparx5/lan969x/lan969x_fdma.c
224
sparx5->rx.ndev = port->ndev;
drivers/net/ethernet/microchip/sparx5/lan969x/lan969x_fdma.c
246
struct sparx5_rx *rx = container_of(napi, struct sparx5_rx, napi);
drivers/net/ethernet/microchip/sparx5/lan969x/lan969x_fdma.c
247
struct sparx5 *sparx5 = container_of(rx, struct sparx5, rx);
drivers/net/ethernet/microchip/sparx5/lan969x/lan969x_fdma.c
249
struct fdma *fdma = &rx->fdma;
drivers/net/ethernet/microchip/sparx5/lan969x/lan969x_fdma.c
261
skb = lan969x_fdma_rx_get_frame(sparx5, rx);
drivers/net/ethernet/microchip/sparx5/lan969x/lan969x_fdma.c
265
napi_gro_receive(&rx->napi, skb);
drivers/net/ethernet/microchip/sparx5/lan969x/lan969x_fdma.c
28
struct sparx5_rx *rx = &FDMA_PRIV(fdma)->rx;
drivers/net/ethernet/microchip/sparx5/lan969x/lan969x_fdma.c
31
page = page_pool_dev_alloc_pages(rx->page_pool);
drivers/net/ethernet/microchip/sparx5/lan969x/lan969x_fdma.c
35
rx->page[dcb][db] = page;
drivers/net/ethernet/microchip/sparx5/lan969x/lan969x_fdma.c
357
struct sparx5_rx *rx = &sparx5->rx;
drivers/net/ethernet/microchip/sparx5/lan969x/lan969x_fdma.c
379
fdma_free_coherent(sparx5->dev, &rx->fdma);
drivers/net/ethernet/microchip/sparx5/lan969x/lan969x_fdma.c
394
struct sparx5_rx *rx = &sparx5->rx;
drivers/net/ethernet/microchip/sparx5/lan969x/lan969x_fdma.c
399
fdma_free_coherent(sparx5->dev, &rx->fdma);
drivers/net/ethernet/microchip/sparx5/lan969x/lan969x_fdma.c
400
lan969x_fdma_free_pages(rx);
drivers/net/ethernet/microchip/sparx5/lan969x/lan969x_fdma.c
401
page_pool_destroy(rx->page_pool);
drivers/net/ethernet/microchip/sparx5/lan969x/lan969x_fdma.c
89
static void lan969x_fdma_free_pages(struct sparx5_rx *rx)
drivers/net/ethernet/microchip/sparx5/lan969x/lan969x_fdma.c
91
struct fdma *fdma = &rx->fdma;
drivers/net/ethernet/microchip/sparx5/lan969x/lan969x_fdma.c
95
page_pool_put_full_page(rx->page_pool,
drivers/net/ethernet/microchip/sparx5/lan969x/lan969x_fdma.c
96
rx->page[i][j], false);
drivers/net/ethernet/microchip/sparx5/sparx5_fdma.c
139
static bool sparx5_fdma_rx_get_frame(struct sparx5 *sparx5, struct sparx5_rx *rx)
drivers/net/ethernet/microchip/sparx5/sparx5_fdma.c
141
struct fdma *fdma = &rx->fdma;
drivers/net/ethernet/microchip/sparx5/sparx5_fdma.c
151
skb = rx->skb[fdma->dcb_index][fdma->db_index];
drivers/net/ethernet/microchip/sparx5/sparx5_fdma.c
178
rx->packets++;
drivers/net/ethernet/microchip/sparx5/sparx5_fdma.c
185
struct sparx5_rx *rx = container_of(napi, struct sparx5_rx, napi);
drivers/net/ethernet/microchip/sparx5/sparx5_fdma.c
186
struct sparx5 *sparx5 = container_of(rx, struct sparx5, rx);
drivers/net/ethernet/microchip/sparx5/sparx5_fdma.c
187
struct fdma *fdma = &rx->fdma;
drivers/net/ethernet/microchip/sparx5/sparx5_fdma.c
190
while (counter < weight && sparx5_fdma_rx_get_frame(sparx5, rx)) {
drivers/net/ethernet/microchip/sparx5/sparx5_fdma.c
203
napi_complete_done(&rx->napi, counter);
drivers/net/ethernet/microchip/sparx5/sparx5_fdma.c
245
struct sparx5_rx *rx = &sparx5->rx;
drivers/net/ethernet/microchip/sparx5/sparx5_fdma.c
246
struct fdma *fdma = &rx->fdma;
drivers/net/ethernet/microchip/sparx5/sparx5_fdma.c
276
struct sparx5_rx *rx, int channel)
drivers/net/ethernet/microchip/sparx5/sparx5_fdma.c
278
struct fdma *fdma = &rx->fdma;
drivers/net/ethernet/microchip/sparx5/sparx5_fdma.c
286
fdma->size = fdma_get_size(&sparx5->rx.fdma);
drivers/net/ethernet/microchip/sparx5/sparx5_fdma.c
294
rx->ndev = port->ndev;
drivers/net/ethernet/microchip/sparx5/sparx5_fdma.c
326
napi_schedule(&sparx5->rx.napi);
drivers/net/ethernet/microchip/sparx5/sparx5_fdma.c
37
struct sparx5_rx *rx = &sparx5->rx;
drivers/net/ethernet/microchip/sparx5/sparx5_fdma.c
40
skb = __netdev_alloc_skb(rx->ndev, fdma->db_size, GFP_ATOMIC);
drivers/net/ethernet/microchip/sparx5/sparx5_fdma.c
420
sparx5_fdma_rx_init(sparx5, &sparx5->rx, FDMA_XTR_CHANNEL);
drivers/net/ethernet/microchip/sparx5/sparx5_fdma.c
438
fdma_free_phys(&sparx5->rx.fdma);
drivers/net/ethernet/microchip/sparx5/sparx5_fdma.c
452
struct sparx5_rx *rx = &sparx5->rx;
drivers/net/ethernet/microchip/sparx5/sparx5_fdma.c
455
netif_napi_add_weight(rx->ndev,
drivers/net/ethernet/microchip/sparx5/sparx5_fdma.c
456
&rx->napi,
drivers/net/ethernet/microchip/sparx5/sparx5_fdma.c
46
rx->skb[dcb][db] = skb;
drivers/net/ethernet/microchip/sparx5/sparx5_fdma.c
460
napi_enable(&rx->napi);
drivers/net/ethernet/microchip/sparx5/sparx5_fdma.c
462
sparx5_fdma_rx_activate(sparx5, rx);
drivers/net/ethernet/microchip/sparx5/sparx5_fdma.c
470
struct sparx5_rx *rx = &sparx5->rx;
drivers/net/ethernet/microchip/sparx5/sparx5_fdma.c
474
napi_disable(&rx->napi);
drivers/net/ethernet/microchip/sparx5/sparx5_fdma.c
477
sparx5_fdma_rx_deactivate(sparx5, rx);
drivers/net/ethernet/microchip/sparx5/sparx5_fdma.c
51
static void sparx5_fdma_rx_activate(struct sparx5 *sparx5, struct sparx5_rx *rx)
drivers/net/ethernet/microchip/sparx5/sparx5_fdma.c
53
struct fdma *fdma = &rx->fdma;
drivers/net/ethernet/microchip/sparx5/sparx5_fdma.c
85
static void sparx5_fdma_rx_deactivate(struct sparx5 *sparx5, struct sparx5_rx *rx)
drivers/net/ethernet/microchip/sparx5/sparx5_fdma.c
87
struct fdma *fdma = &rx->fdma;
drivers/net/ethernet/microchip/sparx5/sparx5_main.h
419
struct sparx5_rx rx;
drivers/net/ethernet/microsoft/mana/mana_en.c
1356
enum TRI_STATE rx,
drivers/net/ethernet/microsoft/mana/mana_en.c
1380
req->rx_enable = rx;
drivers/net/ethernet/microsoft/mana/mana_en.c
2892
int mana_config_rss(struct mana_port_context *apc, enum TRI_STATE rx,
drivers/net/ethernet/microsoft/mana/mana_en.c
2906
err = mana_cfg_vport_steering(apc, rx, true, update_hash, update_tab);
drivers/net/ethernet/moxa/moxart_ether.c
215
int rx = 0;
drivers/net/ethernet/moxa/moxart_ether.c
217
while (rx < budget) {
drivers/net/ethernet/moxa/moxart_ether.c
254
rx++;
drivers/net/ethernet/moxa/moxart_ether.c
269
if (rx < budget)
drivers/net/ethernet/moxa/moxart_ether.c
270
napi_complete_done(napi, rx);
drivers/net/ethernet/moxa/moxart_ether.c
275
return rx;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1190
myri10ge_alloc_rx_pages(struct myri10ge_priv *mgp, struct myri10ge_rx_buf *rx,
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1200
if (unlikely(rx->watchdog_needed && !watchdog))
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1204
while (rx->fill_cnt != (rx->cnt + rx->mask + 1)) {
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1205
idx = rx->fill_cnt & rx->mask;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1206
if (rx->page_offset + bytes <= MYRI10GE_ALLOC_SIZE) {
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1208
get_page(rx->page);
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1215
if (rx->fill_cnt - rx->cnt < 16)
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1216
rx->watchdog_needed = 1;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1225
if (rx->fill_cnt - rx->cnt < 16)
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1226
rx->watchdog_needed = 1;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1230
rx->page = page;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1231
rx->page_offset = 0;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1232
rx->bus = bus;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1235
rx->info[idx].page = rx->page;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1236
rx->info[idx].page_offset = rx->page_offset;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1239
dma_unmap_addr_set(&rx->info[idx], bus, rx->bus);
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1240
rx->shadow[idx].addr_low =
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1241
htonl(MYRI10GE_LOWPART_TO_U32(rx->bus) + rx->page_offset);
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1242
rx->shadow[idx].addr_high =
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1243
htonl(MYRI10GE_HIGHPART_TO_U32(rx->bus));
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1246
rx->page_offset += SKB_DATA_ALIGN(bytes);
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1250
end_offset = rx->page_offset + bytes - 1;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1251
if ((unsigned)(rx->page_offset ^ end_offset) > 4095)
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1252
rx->page_offset = end_offset & ~4095;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1254
rx->fill_cnt++;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1258
myri10ge_submit_8rx(&rx->lanai[idx - 7],
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1259
&rx->shadow[idx - 7]);
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1322
struct myri10ge_rx_buf *rx;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1329
rx = &ss->rx_small;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1332
rx = &ss->rx_big;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1337
idx = rx->cnt & rx->mask;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1338
va = page_address(rx->info[idx].page) + rx->info[idx].page_offset;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1345
myri10ge_unmap_rx_page(pdev, &rx->info[idx], bytes);
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1346
put_page(rx->info[idx].page);
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1347
rx->cnt++;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1348
idx = rx->cnt & rx->mask;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1356
myri10ge_unmap_rx_page(pdev, &rx->info[idx], bytes);
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1357
skb_fill_page_desc(skb, i, rx->info[idx].page,
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1358
rx->info[idx].page_offset,
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1361
rx->cnt++;
drivers/net/ethernet/myricom/myri10ge/myri10ge.c
1362
idx = rx->cnt & rx->mask;
drivers/net/ethernet/netronome/nfp/nfp_net_debugfs.c
131
struct dentry *queues, *tx, *rx, *xdp;
drivers/net/ethernet/netronome/nfp/nfp_net_debugfs.c
147
rx = debugfs_create_dir("rx", queues);
drivers/net/ethernet/netronome/nfp/nfp_net_debugfs.c
153
debugfs_create_file(name, 0400, rx,
drivers/net/ethernet/pasemi/pasemi_mac.c
107
return mac->rx;
drivers/net/ethernet/pasemi/pasemi_mac.c
1103
write_iob_reg(PAS_IOB_DMA_RXCH_CFG(mac->rx->chan.chno),
drivers/net/ethernet/pasemi/pasemi_mac.c
1110
PAS_MAC_IPC_CHNL_DCHNO(mac->rx->chan.chno) |
drivers/net/ethernet/pasemi/pasemi_mac.c
1111
PAS_MAC_IPC_CHNL_BCH(mac->rx->chan.chno));
drivers/net/ethernet/pasemi/pasemi_mac.c
1182
ret = request_irq(mac->rx->chan.irq, pasemi_mac_rx_intr, 0,
drivers/net/ethernet/pasemi/pasemi_mac.c
1183
mac->rx_irq_name, mac->rx);
drivers/net/ethernet/pasemi/pasemi_mac.c
1186
mac->rx->chan.irq, ret);
drivers/net/ethernet/pasemi/pasemi_mac.c
1322
free_irq(mac->rx->chan.irq, mac->rx);
drivers/net/ethernet/pasemi/pasemi_mac.c
1589
disable_irq(mac->rx->chan.irq);
drivers/net/ethernet/pasemi/pasemi_mac.c
1590
pasemi_mac_rx_intr(mac->rx->chan.irq, mac->rx);
drivers/net/ethernet/pasemi/pasemi_mac.c
1591
enable_irq(mac->rx->chan.irq);
drivers/net/ethernet/pasemi/pasemi_mac.c
1808
pasemi_dma_free_chan(&mac->rx->chan);
drivers/net/ethernet/pasemi/pasemi_mac.c
432
mac->rx = ring;
drivers/net/ethernet/pasemi/pasemi_mac.c
542
struct pasemi_mac_rxring *rx = rx_ring(mac);
drivers/net/ethernet/pasemi/pasemi_mac.c
547
info = &RX_DESC_INFO(rx, i);
drivers/net/ethernet/pasemi/pasemi_mac.c
558
RX_BUFF(rx, i) = 0;
drivers/net/ethernet/pasemi/pasemi_mac.c
570
mac->rx = NULL;
drivers/net/ethernet/pasemi/pasemi_mac.c
577
struct pasemi_mac_rxring *rx = rx_ring(mac);
drivers/net/ethernet/pasemi/pasemi_mac.c
585
struct pasemi_mac_buffer *info = &RX_DESC_INFO(rx, fill);
drivers/net/ethernet/pasemi/pasemi_mac.c
586
u64 *buff = &RX_BUFF(rx, fill);
drivers/net/ethernet/pasemi/pasemi_mac.c
624
struct pasemi_mac_rxring *rx = rx_ring(mac);
drivers/net/ethernet/pasemi/pasemi_mac.c
630
pcnt = *rx->chan.status & PAS_STATUS_PCNT_M;
drivers/net/ethernet/pasemi/pasemi_mac.c
634
if (*rx->chan.status & PAS_STATUS_TIMER)
drivers/net/ethernet/pasemi/pasemi_mac.c
637
write_iob_reg(PAS_IOB_DMA_RXCH_RESET(mac->rx->chan.chno), reg);
drivers/net/ethernet/pasemi/pasemi_mac.c
689
static int pasemi_mac_clean_rx(struct pasemi_mac_rxring *rx,
drivers/net/ethernet/pasemi/pasemi_mac.c
692
const struct pasemi_dmachan *chan = &rx->chan;
drivers/net/ethernet/pasemi/pasemi_mac.c
693
struct pasemi_mac *mac = rx->mac;
drivers/net/ethernet/pasemi/pasemi_mac.c
706
spin_lock(&rx->lock);
drivers/net/ethernet/pasemi/pasemi_mac.c
708
n = rx->next_to_clean;
drivers/net/ethernet/pasemi/pasemi_mac.c
710
prefetch(&RX_DESC(rx, n));
drivers/net/ethernet/pasemi/pasemi_mac.c
713
macrx = RX_DESC(rx, n);
drivers/net/ethernet/pasemi/pasemi_mac.c
714
prefetch(&RX_DESC(rx, n+4));
drivers/net/ethernet/pasemi/pasemi_mac.c
727
eval = (RX_DESC(rx, n+1) & XCT_RXRES_8B_EVAL_M) >>
drivers/net/ethernet/pasemi/pasemi_mac.c
731
dma = (RX_DESC(rx, n+2) & XCT_PTR_ADDR_M);
drivers/net/ethernet/pasemi/pasemi_mac.c
732
info = &RX_DESC_INFO(rx, buf_index);
drivers/net/ethernet/pasemi/pasemi_mac.c
773
RX_DESC(rx, n) = 0;
drivers/net/ethernet/pasemi/pasemi_mac.c
774
RX_DESC(rx, n+1) = 0;
drivers/net/ethernet/pasemi/pasemi_mac.c
779
RX_BUFF(rx, buf_index) = 0;
drivers/net/ethernet/pasemi/pasemi_mac.c
796
write_dma_reg(PAS_DMA_RXCHAN_INCR(mac->rx->chan.chno), count << 1);
drivers/net/ethernet/pasemi/pasemi_mac.h
77
struct pasemi_mac_rxring *rx;
drivers/net/ethernet/pasemi/pasemi_mac.h
96
#define RX_DESC(rx, num) ((rx)->chan.ring_virt[(num) & (RX_RING_SIZE-1)])
drivers/net/ethernet/pasemi/pasemi_mac.h
97
#define RX_DESC_INFO(rx, num) ((rx)->ring_info[(num) & (RX_RING_SIZE-1)])
drivers/net/ethernet/pasemi/pasemi_mac.h
98
#define RX_BUFF(rx, num) ((rx)->buffers[(num) & (RX_RING_SIZE-1)])
drivers/net/ethernet/pasemi/pasemi_mac_ethtool.c
81
ering->rx_pending = RING_USED(mac->rx)/4;
drivers/net/ethernet/qlogic/qed/qed_hsi.h
3814
u8 rx;
drivers/net/ethernet/qlogic/qed/qed_l2.c
1260
p_ramrod->filter_cmd_hdr.rx = p_filter_cmd->is_rx_filter ? 1 : 0;
drivers/net/ethernet/qlogic/qed/qed_l2.c
69
u8 rx = 0, tx = 0;
drivers/net/ethernet/qlogic/qed/qed_l2.c
71
qed_vf_get_num_rxqs(p_hwfn, &rx);
drivers/net/ethernet/qlogic/qed/qed_l2.c
74
p_l2_info->queues = max_t(u8, rx, tx);
drivers/net/ethernet/qlogic/qlcnic/qlcnic.h
1503
struct __qlcnic_esw_statistics rx;
drivers/net/ethernet/qlogic/qlcnic/qlcnic_ethtool.c
1387
QLCNIC_QUERY_RX_COUNTER, &port_stats.rx);
drivers/net/ethernet/qlogic/qlcnic/qlcnic_ethtool.c
1391
data = qlcnic_fill_stats(data, &port_stats.rx, QLCNIC_ESW_STATS);
drivers/net/ethernet/qlogic/qlcnic/qlcnic_sysfs.c
792
&port_stats.rx);
drivers/net/ethernet/qlogic/qlcnic/qlcnic_sysfs.c
827
&esw_stats.rx);
drivers/net/ethernet/realtek/8139cp.c
468
int rx = 0;
drivers/net/ethernet/realtek/8139cp.c
472
while (rx < budget) {
drivers/net/ethernet/realtek/8139cp.c
538
rx++;
drivers/net/ethernet/realtek/8139cp.c
557
if (rx < budget && napi_complete_done(napi, rx)) {
drivers/net/ethernet/realtek/8139cp.c
565
return rx;
drivers/net/ethernet/realtek/r8169_leds.c
40
bool rx, tx;
drivers/net/ethernet/realtek/r8169_leds.c
47
rx = flags & BIT(TRIGGER_NETDEV_RX);
drivers/net/ethernet/realtek/r8169_leds.c
50
return rx == tx;
drivers/net/ethernet/sfc/efx.c
646
struct netdev_queue_stats_rx *rx,
drivers/net/ethernet/sfc/efx.c
654
rx->packets = 0;
drivers/net/ethernet/sfc/efx.c
655
rx->bytes = 0;
drivers/net/ethernet/sfc/efx.c
656
rx->hw_drops = 0;
drivers/net/ethernet/sfc/efx.c
657
rx->hw_drop_overruns = 0;
drivers/net/ethernet/sfc/efx.c
669
rx->packets += rx_queue->rx_packets;
drivers/net/ethernet/sfc/efx.c
670
rx->bytes += rx_queue->rx_bytes;
drivers/net/ethernet/sfc/efx.c
671
rx->hw_drops += efx_get_queue_stat_rx_hw_drops(channel);
drivers/net/ethernet/sfc/efx.c
672
rx->hw_drop_overruns += channel->n_rx_nodesc_trunc;
drivers/net/ethernet/sfc/efx.c
674
rx->packets += rx_queue->old_rx_packets;
drivers/net/ethernet/sfc/efx.c
675
rx->bytes += rx_queue->old_rx_bytes;
drivers/net/ethernet/sfc/efx.c
676
rx->hw_drops += channel->old_n_rx_hw_drops;
drivers/net/ethernet/sfc/efx.c
677
rx->hw_drop_overruns += channel->old_n_rx_hw_drop_overruns;
drivers/net/ethernet/socionext/sni_ave.c
1282
priv->rx.desc = kzalloc_objs(*priv->rx.desc, priv->rx.ndesc);
drivers/net/ethernet/socionext/sni_ave.c
1283
if (!priv->rx.desc) {
drivers/net/ethernet/socionext/sni_ave.c
1301
priv->rx.proc_idx = 0;
drivers/net/ethernet/socionext/sni_ave.c
1302
priv->rx.done_idx = 0;
drivers/net/ethernet/socionext/sni_ave.c
1303
for (entry = 0; entry < priv->rx.ndesc; entry++) {
drivers/net/ethernet/socionext/sni_ave.c
1308
(((priv->rx.ndesc * priv->desc_size) << 16) & AVE_RXDC0_SIZE),
drivers/net/ethernet/socionext/sni_ave.c
1379
for (entry = 0; entry < priv->rx.ndesc; entry++) {
drivers/net/ethernet/socionext/sni_ave.c
1380
if (!priv->rx.desc[entry].skbs)
drivers/net/ethernet/socionext/sni_ave.c
1383
ave_dma_unmap(ndev, &priv->rx.desc[entry], DMA_FROM_DEVICE);
drivers/net/ethernet/socionext/sni_ave.c
1384
dev_kfree_skb_any(priv->rx.desc[entry].skbs);
drivers/net/ethernet/socionext/sni_ave.c
1385
priv->rx.desc[entry].skbs = NULL;
drivers/net/ethernet/socionext/sni_ave.c
1387
priv->rx.proc_idx = 0;
drivers/net/ethernet/socionext/sni_ave.c
1388
priv->rx.done_idx = 0;
drivers/net/ethernet/socionext/sni_ave.c
1391
kfree(priv->rx.desc);
drivers/net/ethernet/socionext/sni_ave.c
1620
priv->rx.daddr = AVE_RXDM_64;
drivers/net/ethernet/socionext/sni_ave.c
1625
priv->rx.daddr = AVE_RXDM_32;
drivers/net/ethernet/socionext/sni_ave.c
1633
priv->rx.ndesc = AVE_NR_RXDESC;
drivers/net/ethernet/socionext/sni_ave.c
278
struct ave_desc_info rx;
drivers/net/ethernet/socionext/sni_ave.c
303
addr = ((id == AVE_DESCID_TX) ? priv->tx.daddr : priv->rx.daddr)
drivers/net/ethernet/socionext/sni_ave.c
321
addr = ((id == AVE_DESCID_TX) ? priv->tx.daddr : priv->rx.daddr)
drivers/net/ethernet/socionext/sni_ave.c
586
skb = priv->rx.desc[entry].skbs;
drivers/net/ethernet/socionext/sni_ave.c
608
ret = ave_dma_map(ndev, &priv->rx.desc[entry],
drivers/net/ethernet/socionext/sni_ave.c
617
priv->rx.desc[entry].skbs = skb;
drivers/net/ethernet/socionext/sni_ave.c
750
proc_idx = priv->rx.proc_idx;
drivers/net/ethernet/socionext/sni_ave.c
751
done_idx = priv->rx.done_idx;
drivers/net/ethernet/socionext/sni_ave.c
752
ndesc = priv->rx.ndesc;
drivers/net/ethernet/socionext/sni_ave.c
775
skb = priv->rx.desc[proc_idx].skbs;
drivers/net/ethernet/socionext/sni_ave.c
776
priv->rx.desc[proc_idx].skbs = NULL;
drivers/net/ethernet/socionext/sni_ave.c
778
ave_dma_unmap(ndev, &priv->rx.desc[proc_idx], DMA_FROM_DEVICE);
drivers/net/ethernet/socionext/sni_ave.c
795
priv->rx.proc_idx = proc_idx;
drivers/net/ethernet/socionext/sni_ave.c
810
priv->rx.done_idx = done_idx;
drivers/net/ethernet/socionext/sni_ave.c
902
ave_rx_receive(ndev, priv->rx.ndesc);
drivers/net/ethernet/stmicro/stmmac/dwmac-rk.c
249
#define DELAY_ENABLE(soc, tx, rx) \
drivers/net/ethernet/stmicro/stmmac/dwmac-rk.c
251
((rx) ? soc##_GMAC_RXCLK_DLY_ENABLE : soc##_GMAC_RXCLK_DLY_DISABLE))
drivers/net/ethernet/stmicro/stmmac/dwmac-sun8i.c
350
bool rx, bool tx)
drivers/net/ethernet/stmicro/stmmac/dwmac-sun8i.c
354
if (rx)
drivers/net/ethernet/stmicro/stmmac/dwmac-sun8i.c
364
bool rx, bool tx)
drivers/net/ethernet/stmicro/stmmac/dwmac-sun8i.c
368
if (rx)
drivers/net/ethernet/stmicro/stmmac/dwmac4_descs.c
413
static void dwmac4_display_ring(void *head, unsigned int size, bool rx,
drivers/net/ethernet/stmicro/stmmac/dwmac4_descs.c
419
pr_info("%s descriptor ring:\n", rx ? "RX" : "TX");
drivers/net/ethernet/stmicro/stmmac/dwmac4_dma.h
176
u32 chan, bool rx, bool tx);
drivers/net/ethernet/stmicro/stmmac/dwmac4_dma.h
178
u32 chan, bool rx, bool tx);
drivers/net/ethernet/stmicro/stmmac/dwmac4_dma.h
180
u32 chan, bool rx, bool tx);
drivers/net/ethernet/stmicro/stmmac/dwmac4_dma.h
182
u32 chan, bool rx, bool tx);
drivers/net/ethernet/stmicro/stmmac/dwmac4_lib.c
113
u32 chan, bool rx, bool tx)
drivers/net/ethernet/stmicro/stmmac/dwmac4_lib.c
118
if (rx)
drivers/net/ethernet/stmicro/stmmac/dwmac4_lib.c
127
u32 chan, bool rx, bool tx)
drivers/net/ethernet/stmicro/stmmac/dwmac4_lib.c
132
if (rx)
drivers/net/ethernet/stmicro/stmmac/dwmac4_lib.c
141
u32 chan, bool rx, bool tx)
drivers/net/ethernet/stmicro/stmmac/dwmac4_lib.c
146
if (rx)
drivers/net/ethernet/stmicro/stmmac/dwmac4_lib.c
155
u32 chan, bool rx, bool tx)
drivers/net/ethernet/stmicro/stmmac/dwmac4_lib.c
160
if (rx)
drivers/net/ethernet/stmicro/stmmac/dwmac_dma.h
148
u32 chan, bool rx, bool tx);
drivers/net/ethernet/stmicro/stmmac/dwmac_dma.h
150
u32 chan, bool rx, bool tx);
drivers/net/ethernet/stmicro/stmmac/dwmac_lib.c
42
u32 chan, bool rx, bool tx)
drivers/net/ethernet/stmicro/stmmac/dwmac_lib.c
46
if (rx)
drivers/net/ethernet/stmicro/stmmac/dwmac_lib.c
55
u32 chan, bool rx, bool tx)
drivers/net/ethernet/stmicro/stmmac/dwmac_lib.c
59
if (rx)
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c
21
u32 tx, rx;
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c
24
rx = readl(ioaddr + XGMAC_RX_CONFIG);
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c
27
writel(rx | XGMAC_CORE_INIT_RX, ioaddr + XGMAC_RX_CONFIG);
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c
54
u32 rx = readl(ioaddr + XGMAC_RX_CONFIG);
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c
58
rx |= XGMAC_CONFIG_RE;
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c
61
rx &= ~XGMAC_CONFIG_RE;
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_core.c
65
writel(rx, ioaddr + XGMAC_RX_CONFIG);
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_dma.c
223
bool rx, bool tx)
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_dma.c
227
if (rx)
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_dma.c
237
bool rx, bool tx)
drivers/net/ethernet/stmicro/stmmac/dwxgmac2_dma.c
241
if (rx)
drivers/net/ethernet/stmicro/stmmac/enh_desc.c
412
static void enh_desc_display_ring(void *head, unsigned int size, bool rx,
drivers/net/ethernet/stmicro/stmmac/enh_desc.c
419
pr_info("Extended %s descriptor ring:\n", rx ? "RX" : "TX");
drivers/net/ethernet/stmicro/stmmac/hwif.h
206
u32 chan, bool rx, bool tx);
drivers/net/ethernet/stmicro/stmmac/hwif.h
208
u32 chan, bool rx, bool tx);
drivers/net/ethernet/stmicro/stmmac/hwif.h
89
void (*display_ring)(void *head, unsigned int size, bool rx,
drivers/net/ethernet/stmicro/stmmac/norm_desc.c
258
static void ndesc_display_ring(void *head, unsigned int size, bool rx,
drivers/net/ethernet/stmicro/stmmac/norm_desc.c
265
pr_info("%s descriptor ring:\n", rx ? "RX" : "TX");
drivers/net/ethernet/sun/niu.c
9445
int rx)
drivers/net/ethernet/sun/niu.c
9453
arr = (rx ? p->rxchan_per_port : p->txchan_per_port);
drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c
407
xlgmac_set_buffer_data(&desc_data->rx.hdr, &ring->rx_hdr_pa,
drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c
41
if (desc_data->rx.hdr.pa.pages)
drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c
411
xlgmac_set_buffer_data(&desc_data->rx.buf, &ring->rx_buf_pa,
drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c
42
put_page(desc_data->rx.hdr.pa.pages);
drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c
44
if (desc_data->rx.hdr.pa_unmap.pages) {
drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c
45
dma_unmap_page(pdata->dev, desc_data->rx.hdr.pa_unmap.pages_dma,
drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c
46
desc_data->rx.hdr.pa_unmap.pages_len,
drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c
48
put_page(desc_data->rx.hdr.pa_unmap.pages);
drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c
51
if (desc_data->rx.buf.pa.pages)
drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c
52
put_page(desc_data->rx.buf.pa.pages);
drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c
54
if (desc_data->rx.buf.pa_unmap.pages) {
drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c
55
dma_unmap_page(pdata->dev, desc_data->rx.buf.pa_unmap.pages_dma,
drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c
56
desc_data->rx.buf.pa_unmap.pages_len,
drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c
58
put_page(desc_data->rx.buf.pa_unmap.pages);
drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c
62
memset(&desc_data->rx, 0, sizeof(desc_data->rx));
drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c
1110
hdr_dma = desc_data->rx.hdr.dma_base + desc_data->rx.hdr.dma_off;
drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c
1111
buf_dma = desc_data->rx.buf.dma_base + desc_data->rx.buf.dma_off;
drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c
2704
desc_data->rx.hdr_len = XLGMAC_GET_REG_BITS_LE(dma_desc->desc2,
drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c
2707
if (desc_data->rx.hdr_len)
drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c
2739
desc_data->rx.len = XLGMAC_GET_REG_BITS_LE(dma_desc->desc3,
drivers/net/ethernet/synopsys/dwc-xlgmac-net.c
1003
dma_sync_single_range_for_cpu(pdata->dev, desc_data->rx.hdr.dma_base,
drivers/net/ethernet/synopsys/dwc-xlgmac-net.c
1004
desc_data->rx.hdr.dma_off,
drivers/net/ethernet/synopsys/dwc-xlgmac-net.c
1005
desc_data->rx.hdr.dma_len,
drivers/net/ethernet/synopsys/dwc-xlgmac-net.c
1008
packet = page_address(desc_data->rx.hdr.pa.pages) +
drivers/net/ethernet/synopsys/dwc-xlgmac-net.c
1009
desc_data->rx.hdr.pa.pages_offset;
drivers/net/ethernet/synopsys/dwc-xlgmac-net.c
1010
copy_len = (desc_data->rx.hdr_len) ? desc_data->rx.hdr_len : len;
drivers/net/ethernet/synopsys/dwc-xlgmac-net.c
1011
copy_len = min(desc_data->rx.hdr.dma_len, copy_len);
drivers/net/ethernet/synopsys/dwc-xlgmac-net.c
1019
desc_data->rx.buf.dma_base,
drivers/net/ethernet/synopsys/dwc-xlgmac-net.c
1020
desc_data->rx.buf.dma_off,
drivers/net/ethernet/synopsys/dwc-xlgmac-net.c
1021
desc_data->rx.buf.dma_len,
drivers/net/ethernet/synopsys/dwc-xlgmac-net.c
1025
desc_data->rx.buf.pa.pages,
drivers/net/ethernet/synopsys/dwc-xlgmac-net.c
1026
desc_data->rx.buf.pa.pages_offset,
drivers/net/ethernet/synopsys/dwc-xlgmac-net.c
1027
len, desc_data->rx.buf.dma_len);
drivers/net/ethernet/synopsys/dwc-xlgmac-net.c
1028
desc_data->rx.buf.pa.pages = NULL;
drivers/net/ethernet/synopsys/dwc-xlgmac-net.c
1188
dma_desc_len = desc_data->rx.len - len;
drivers/net/ethernet/synopsys/dwc-xlgmac-net.c
1199
desc_data->rx.buf.dma_base,
drivers/net/ethernet/synopsys/dwc-xlgmac-net.c
1200
desc_data->rx.buf.dma_off,
drivers/net/ethernet/synopsys/dwc-xlgmac-net.c
1201
desc_data->rx.buf.dma_len,
drivers/net/ethernet/synopsys/dwc-xlgmac-net.c
1206
desc_data->rx.buf.pa.pages,
drivers/net/ethernet/synopsys/dwc-xlgmac-net.c
1207
desc_data->rx.buf.pa.pages_offset,
drivers/net/ethernet/synopsys/dwc-xlgmac-net.c
1209
desc_data->rx.buf.dma_len);
drivers/net/ethernet/synopsys/dwc-xlgmac-net.c
1210
desc_data->rx.buf.pa.pages = NULL;
drivers/net/ethernet/synopsys/dwc-xlgmac-net.c
996
skb = napi_alloc_skb(napi, desc_data->rx.hdr.dma_len);
drivers/net/ethernet/synopsys/dwc-xlgmac.h
299
struct xlgmac_rx_desc_data rx;
drivers/net/ethernet/tehuti/tn40.c
1619
struct netdev_queue_stats_rx *rx,
drivers/net/ethernet/tehuti/tn40.c
1622
rx->packets = 0;
drivers/net/ethernet/tehuti/tn40.c
1623
rx->bytes = 0;
drivers/net/ethernet/tehuti/tn40.c
1624
rx->alloc_fail = 0;
drivers/net/ethernet/ti/cpsw_ethtool.c
519
static int cpsw_update_channels_res(struct cpsw_priv *priv, int ch_num, int rx,
drivers/net/ethernet/ti/cpsw_ethtool.c
528
if (rx) {
drivers/net/ethernet/ti/cpsw_ethtool.c
539
vch = rx ? *ch : 7 - *ch;
drivers/net/ethernet/ti/cpsw_ethtool.c
540
vec[*ch].ch = cpdma_chan_create(cpsw->dma, vch, handler, rx);
drivers/net/ethernet/ti/cpsw_ethtool.c
551
(rx ? "rx" : "tx"));
drivers/net/ethernet/ti/cpsw_ethtool.c
563
(rx ? "rx" : "tx"));
drivers/net/ethernet/ti/davinci_cpdma.c
680
int rx, int desc_num,
drivers/net/ethernet/ti/davinci_cpdma.c
691
if (rx) {
drivers/net/ethernet/ti/netcp_core.c
1664
config.u.rx.einfo_present = true;
drivers/net/ethernet/ti/netcp_core.c
1665
config.u.rx.psinfo_present = true;
drivers/net/ethernet/ti/netcp_core.c
1666
config.u.rx.err_mode = DMA_DROP;
drivers/net/ethernet/ti/netcp_core.c
1667
config.u.rx.desc_type = DMA_DESC_HOST;
drivers/net/ethernet/ti/netcp_core.c
1668
config.u.rx.psinfo_at_sop = false;
drivers/net/ethernet/ti/netcp_core.c
1669
config.u.rx.sop_offset = NETCP_SOP_OFFSET;
drivers/net/ethernet/ti/netcp_core.c
1670
config.u.rx.dst_q = netcp->rx_queue_id;
drivers/net/ethernet/ti/netcp_core.c
1671
config.u.rx.thresh = DMA_THRESH_NONE;
drivers/net/ethernet/ti/netcp_core.c
1676
config.u.rx.fdq[i] = last_fdq;
drivers/net/ethernet/toshiba/ps3_gelic_net.c
1011
if (card->vlan[i].rx == vid) {
drivers/net/ethernet/toshiba/ps3_gelic_net.c
1617
int rx;
drivers/net/ethernet/toshiba/ps3_gelic_net.c
1621
.rx = GELIC_LV1_VLAN_RX_ETHERNET_0
drivers/net/ethernet/toshiba/ps3_gelic_net.c
1625
.rx = GELIC_LV1_VLAN_RX_WIRELESS
drivers/net/ethernet/toshiba/ps3_gelic_net.c
1641
card->vlan[i].rx = 0;
drivers/net/ethernet/toshiba/ps3_gelic_net.c
1649
vlan_id_ix[i].rx,
drivers/net/ethernet/toshiba/ps3_gelic_net.c
1655
vlan_id_ix[i].rx, status);
drivers/net/ethernet/toshiba/ps3_gelic_net.c
1657
card->vlan[i].rx = 0;
drivers/net/ethernet/toshiba/ps3_gelic_net.c
1660
card->vlan[i].rx = (u16)v1;
drivers/net/ethernet/toshiba/ps3_gelic_net.c
1663
i, card->vlan[i].tx, card->vlan[i].rx);
drivers/net/ethernet/toshiba/ps3_gelic_net.c
1675
card->vlan[GELIC_PORT_WIRELESS].rx = 0;
drivers/net/ethernet/toshiba/ps3_gelic_net.h
265
u16 rx;
drivers/net/ethernet/tundra/tsi108_eth.c
1405
int rx = data->rxtail;
drivers/net/ethernet/tundra/tsi108_eth.c
1408
skb = data->rxskbs[rx];
drivers/net/ethernet/tundra/tsi108_eth.c
735
int rx = data->rxtail;
drivers/net/ethernet/tundra/tsi108_eth.c
738
if (data->rxring[rx].misc & TSI108_RX_OWN)
drivers/net/ethernet/tundra/tsi108_eth.c
741
skb = data->rxskbs[rx];
drivers/net/ethernet/tundra/tsi108_eth.c
746
if (data->rxring[rx].misc & TSI108_RX_BAD) {
drivers/net/ethernet/tundra/tsi108_eth.c
749
if (data->rxring[rx].misc & TSI108_RX_CRC)
drivers/net/ethernet/tundra/tsi108_eth.c
751
if (data->rxring[rx].misc & TSI108_RX_OVER)
drivers/net/ethernet/tundra/tsi108_eth.c
762
dev->name, data->rxring[rx].len);
drivers/net/ethernet/tundra/tsi108_eth.c
763
for (i = 0; i < data->rxring[rx].len; i++)
drivers/net/ethernet/tundra/tsi108_eth.c
768
skb_put(skb, data->rxring[rx].len);
drivers/net/ethernet/tundra/tsi108_eth.c
782
int rx = data->rxhead;
drivers/net/ethernet/tundra/tsi108_eth.c
786
data->rxskbs[rx] = skb;
drivers/net/ethernet/tundra/tsi108_eth.c
790
data->rxring[rx].buf0 = dma_map_single(&data->pdev->dev,
drivers/net/ethernet/tundra/tsi108_eth.c
799
data->rxring[rx].blen = TSI108_RX_SKB_SIZE;
drivers/net/ethernet/tundra/tsi108_eth.c
800
data->rxring[rx].misc = TSI108_RX_OWN | TSI108_RX_INT;
drivers/net/ethernet/vertexcom/mse102x.c
319
__be16 rx = 0;
drivers/net/ethernet/vertexcom/mse102x.c
324
if (mse102x_rx_cmd_spi(mse, (u8 *)&rx)) {
drivers/net/ethernet/vertexcom/mse102x.c
329
cmd_resp = be16_to_cpu(rx);
drivers/net/ethernet/vertexcom/mse102x.c
385
__be16 rx = 0;
drivers/net/ethernet/vertexcom/mse102x.c
395
ret = mse102x_rx_cmd_spi(mse, (u8 *)&rx);
drivers/net/ethernet/vertexcom/mse102x.c
396
cmd_resp = be16_to_cpu(rx);
drivers/net/ethernet/via/via-velocity.c
1410
writel(vptr->rx.pool_dma, ®s->RDBaseLo);
drivers/net/ethernet/via/via-velocity.c
1455
if (vptr->rx.filled < 4)
drivers/net/ethernet/via/via-velocity.c
1460
unusable = vptr->rx.filled & 0x0003;
drivers/net/ethernet/via/via-velocity.c
1461
dirty = vptr->rx.dirty - unusable;
drivers/net/ethernet/via/via-velocity.c
1462
for (avail = vptr->rx.filled & 0xfffc; avail; avail--) {
drivers/net/ethernet/via/via-velocity.c
1464
vptr->rx.ring[dirty].rdesc0.len |= OWNED_BY_NIC;
drivers/net/ethernet/via/via-velocity.c
1467
writew(vptr->rx.filled & 0xfffc, ®s->RBRDU);
drivers/net/ethernet/via/via-velocity.c
1468
vptr->rx.filled = unusable;
drivers/net/ethernet/via/via-velocity.c
1501
vptr->rx.ring = pool;
drivers/net/ethernet/via/via-velocity.c
1502
vptr->rx.pool_dma = pool_dma;
drivers/net/ethernet/via/via-velocity.c
1519
vptr->rx.buf_sz = (mtu <= ETH_DATA_LEN) ? PKT_BUF_SZ : mtu + 32;
drivers/net/ethernet/via/via-velocity.c
1534
struct rx_desc *rd = &(vptr->rx.ring[idx]);
drivers/net/ethernet/via/via-velocity.c
1535
struct velocity_rd_info *rd_info = &(vptr->rx.info[idx]);
drivers/net/ethernet/via/via-velocity.c
1537
rd_info->skb = netdev_alloc_skb(vptr->netdev, vptr->rx.buf_sz + 64);
drivers/net/ethernet/via/via-velocity.c
1548
vptr->rx.buf_sz, DMA_FROM_DEVICE);
drivers/net/ethernet/via/via-velocity.c
1555
rd->size = cpu_to_le16(vptr->rx.buf_sz) | RX_INTEN;
drivers/net/ethernet/via/via-velocity.c
1564
int dirty = vptr->rx.dirty, done = 0;
drivers/net/ethernet/via/via-velocity.c
1567
struct rx_desc *rd = vptr->rx.ring + dirty;
drivers/net/ethernet/via/via-velocity.c
1573
if (!vptr->rx.info[dirty].skb) {
drivers/net/ethernet/via/via-velocity.c
1579
} while (dirty != vptr->rx.curr);
drivers/net/ethernet/via/via-velocity.c
1582
vptr->rx.dirty = dirty;
drivers/net/ethernet/via/via-velocity.c
1583
vptr->rx.filled += done;
drivers/net/ethernet/via/via-velocity.c
1600
if (vptr->rx.info == NULL)
drivers/net/ethernet/via/via-velocity.c
1604
struct velocity_rd_info *rd_info = &(vptr->rx.info[i]);
drivers/net/ethernet/via/via-velocity.c
1605
struct rx_desc *rd = vptr->rx.ring + i;
drivers/net/ethernet/via/via-velocity.c
1611
dma_unmap_single(vptr->dev, rd_info->skb_dma, vptr->rx.buf_sz,
drivers/net/ethernet/via/via-velocity.c
1619
kfree(vptr->rx.info);
drivers/net/ethernet/via/via-velocity.c
1620
vptr->rx.info = NULL;
drivers/net/ethernet/via/via-velocity.c
1634
vptr->rx.info = kzalloc_objs(struct velocity_rd_info,
drivers/net/ethernet/via/via-velocity.c
1636
if (!vptr->rx.info)
drivers/net/ethernet/via/via-velocity.c
1691
dma_free_coherent(vptr->dev, size, vptr->rx.ring, vptr->rx.pool_dma);
drivers/net/ethernet/via/via-velocity.c
2049
struct velocity_rd_info *rd_info = &(vptr->rx.info[idx]);
drivers/net/ethernet/via/via-velocity.c
2050
struct rx_desc *rd = &(vptr->rx.ring[idx]);
drivers/net/ethernet/via/via-velocity.c
2067
vptr->rx.buf_sz, DMA_FROM_DEVICE);
drivers/net/ethernet/via/via-velocity.c
2074
dma_unmap_single(vptr->dev, rd_info->skb_dma, vptr->rx.buf_sz,
drivers/net/ethernet/via/via-velocity.c
2078
vptr->rx.buf_sz, DMA_FROM_DEVICE);
drivers/net/ethernet/via/via-velocity.c
2109
int rd_curr = vptr->rx.curr;
drivers/net/ethernet/via/via-velocity.c
2113
struct rx_desc *rd = vptr->rx.ring + rd_curr;
drivers/net/ethernet/via/via-velocity.c
2115
if (!vptr->rx.info[rd_curr].skb)
drivers/net/ethernet/via/via-velocity.c
2146
vptr->rx.curr = rd_curr;
drivers/net/ethernet/via/via-velocity.c
2303
struct rx_info rx;
drivers/net/ethernet/via/via-velocity.c
2330
rx = vptr->rx;
drivers/net/ethernet/via/via-velocity.c
2333
vptr->rx = tmp_vptr->rx;
drivers/net/ethernet/via/via-velocity.c
2336
tmp_vptr->rx = rx;
drivers/net/ethernet/via/via-velocity.c
555
vptr->rx.dirty = vptr->rx.filled = vptr->rx.curr = 0;
drivers/net/ethernet/via/via-velocity.c
577
vptr->rx.ring[i].rdesc0.len |= OWNED_BY_NIC;
drivers/net/ethernet/via/via-velocity.c
580
writel(vptr->rx.pool_dma, ®s->RDBaseLo);
drivers/net/ethernet/via/via-velocity.h
1417
} rx;
drivers/net/ethernet/wangxun/libwx/wx_ethtool.c
321
if (wx->q_vector[0]->tx.count && wx->q_vector[0]->rx.count)
drivers/net/ethernet/wangxun/libwx/wx_ethtool.c
379
if (wx->q_vector[0]->tx.count && wx->q_vector[0]->rx.count) {
drivers/net/ethernet/wangxun/libwx/wx_ethtool.c
446
if (wx->q_vector[0]->tx.count && wx->q_vector[0]->rx.count)
drivers/net/ethernet/wangxun/libwx/wx_ethtool.c
451
if (q_vector->tx.count && !q_vector->rx.count)
drivers/net/ethernet/wangxun/libwx/wx_lib.c
1693
new_itr = min(q_vector->rx.itr, q_vector->tx.itr);
drivers/net/ethernet/wangxun/libwx/wx_lib.c
1711
struct wx_ring_container *rx;
drivers/net/ethernet/wangxun/libwx/wx_lib.c
1714
rx = container_of(dim, struct wx_ring_container, dim);
drivers/net/ethernet/wangxun/libwx/wx_lib.c
1717
rx->itr = rx_moder.usec;
drivers/net/ethernet/wangxun/libwx/wx_lib.c
1719
q_vector = container_of(rx, struct wx_q_vector, rx);
drivers/net/ethernet/wangxun/libwx/wx_lib.c
1751
INIT_WORK(&q_vector->rx.dim.work, wx_rx_dim_work);
drivers/net/ethernet/wangxun/libwx/wx_lib.c
1753
q_vector->rx.dim.mode = DIM_CQ_PERIOD_MODE_START_FROM_CQE;
drivers/net/ethernet/wangxun/libwx/wx_lib.c
1768
disable_work_sync(&q_vector->rx.dim.work);
drivers/net/ethernet/wangxun/libwx/wx_lib.c
2173
wx_add_ring(ring, &q_vector->rx);
drivers/net/ethernet/wangxun/libwx/wx_lib.c
2210
wx_for_each_ring(ring, q_vector->rx)
drivers/net/ethernet/wangxun/libwx/wx_lib.c
2353
if (q_vector->rx.ring || q_vector->tx.ring) {
drivers/net/ethernet/wangxun/libwx/wx_lib.c
2378
if (!q_vector->rx.ring && !q_vector->tx.ring)
drivers/net/ethernet/wangxun/libwx/wx_lib.c
2549
wx_for_each_ring(ring, q_vector->rx)
drivers/net/ethernet/wangxun/libwx/wx_lib.c
734
q_vector->rx.total_packets += total_rx_packets;
drivers/net/ethernet/wangxun/libwx/wx_lib.c
735
q_vector->rx.total_bytes += total_rx_bytes;
drivers/net/ethernet/wangxun/libwx/wx_lib.c
894
q_vector->rx.total_packets,
drivers/net/ethernet/wangxun/libwx/wx_lib.c
895
q_vector->rx.total_bytes,
drivers/net/ethernet/wangxun/libwx/wx_lib.c
898
net_dim(&q_vector->rx.dim, &sample);
drivers/net/ethernet/wangxun/libwx/wx_lib.c
946
if (q_vector->rx.count > 1)
drivers/net/ethernet/wangxun/libwx/wx_lib.c
947
per_ring_budget = max(budget / q_vector->rx.count, 1);
drivers/net/ethernet/wangxun/libwx/wx_lib.c
951
wx_for_each_ring(ring, q_vector->rx) {
drivers/net/ethernet/wangxun/libwx/wx_type.h
1116
struct wx_ring_container rx, tx;
drivers/net/ethernet/wangxun/libwx/wx_vf_common.c
85
if (q_vector->tx.ring && q_vector->rx.ring)
drivers/net/ethernet/wangxun/libwx/wx_vf_lib.c
61
wx_for_each_ring(ring, q_vector->rx)
drivers/net/ethernet/wangxun/ngbe/ngbe_main.c
269
if (q_vector->tx.ring && q_vector->rx.ring)
drivers/net/ethernet/wangxun/txgbe/txgbe_irq.c
57
if (q_vector->tx.ring && q_vector->rx.ring)
drivers/net/fddi/skfp/fplustm.c
275
smc->hw.fp.rx[QUEUE_R1] = queue = &smc->hw.fp.rx_q[QUEUE_R1] ;
drivers/net/fddi/skfp/fplustm.c
282
smc->hw.fp.rx[QUEUE_R2] = queue = &smc->hw.fp.rx_q[QUEUE_R2] ;
drivers/net/fddi/skfp/h/fplustm.h
186
struct s_smt_rx_queue *rx[USED_QUEUES] ;
drivers/net/fddi/skfp/hwmtm.c
1052
queue = smc->hw.fp.rx[QUEUE_R1] ;
drivers/net/fddi/skfp/hwmtm.c
1469
queue = smc->hw.fp.rx[QUEUE_R1] ;
drivers/net/fddi/skfp/hwmtm.c
394
queue = smc->hw.fp.rx[QUEUE_R1] ;
drivers/net/fddi/skfp/hwmtm.c
573
phys = repair_rxd_ring(smc,smc->hw.fp.rx[QUEUE_R1]) ;
drivers/net/fjes/fjes_hw.c
254
result = fjes_hw_alloc_epbuf(&buf_pair->rx);
drivers/net/fjes/fjes_hw.c
261
fjes_hw_setup_epbuf(&buf_pair->rx, mac,
drivers/net/fjes/fjes_hw.c
285
fjes_hw_free_epbuf(&hw->ep_shm_info[epidx].rx);
drivers/net/fjes/fjes_hw.c
319
fjes_hw_free_epbuf(&hw->ep_shm_info[epidx].rx);
drivers/net/fjes/fjes_hw.c
523
buf_pair->rx.size);
drivers/net/fjes/fjes_hw.c
537
req_buf->share_buffer.buffer[idx++] = buf_pair->rx.size;
drivers/net/fjes/fjes_hw.c
538
page_count = buf_pair->rx.size / EP_BUFFER_INFO_SIZE;
drivers/net/fjes/fjes_hw.c
540
addr = ((u8 *)(buf_pair->rx.buffer)) +
drivers/net/fjes/fjes_hw.c
833
info = hw->ep_shm_info[epidx].rx.info;
drivers/net/fjes/fjes_hw.h
252
} tx, rx;
drivers/net/fjes/fjes_main.c
1166
!(hw->ep_shm_info[epid].rx.info->v1i.rx_status &
drivers/net/fjes/fjes_main.c
1211
stop_req_done = hw->ep_shm_info[epidx].rx.info->v1i.rx_status &
drivers/net/fjes/fjes_main.c
318
if (hw->ep_shm_info[src_epid].rx.info->v1i.rx_status &
drivers/net/fjes/fjes_main.c
662
&adapter->hw.ep_shm_info[dest_epid].rx, 0)) {
drivers/net/fjes/fjes_main.c
672
&adapter->hw.ep_shm_info[dest_epid].rx,
drivers/net/fjes/fjes_main.c
684
&adapter->hw.ep_shm_info[dest_epid].rx,
drivers/net/fjes/fjes_main.c
924
&hw->ep_shm_info[cur_epid].rx))
drivers/net/fjes/fjes_main.c
942
&adapter->hw.ep_shm_info[*cur_epid].rx, psize);
drivers/net/fjes/fjes_main.c
949
fjes_hw_epbuf_rx_curpkt_drop(&adapter->hw.ep_shm_info[cur_epid].rx);
drivers/net/fjes/fjes_trace.h
102
__field(u64, rx)
drivers/net/fjes/fjes_trace.h
106
void *tx, *rx;
drivers/net/fjes/fjes_trace.h
109
rx = (void *)buf_pair->rx.buffer;
drivers/net/fjes/fjes_trace.h
113
__entry->rx_size = buf_pair->rx.size;
drivers/net/fjes/fjes_trace.h
116
__entry->rx = page_to_phys(vmalloc_to_page(rx)) +
drivers/net/fjes/fjes_trace.h
117
offset_in_page(rx);
drivers/net/fjes/fjes_trace.h
121
__entry->rx, __entry->rx_size)
drivers/net/ipa/data/ipa_data-v3.1.c
104
.rx = {
drivers/net/ipa/data/ipa_data-v3.1.c
153
.rx = {
drivers/net/ipa/data/ipa_data-v3.5.1.c
145
.rx = {
drivers/net/ipa/data/ipa_data-v3.5.1.c
95
.rx = {
drivers/net/ipa/data/ipa_data-v4.11.c
138
.rx = {
drivers/net/ipa/data/ipa_data-v4.11.c
89
.rx = {
drivers/net/ipa/data/ipa_data-v4.2.c
135
.rx = {
drivers/net/ipa/data/ipa_data-v4.2.c
85
.rx = {
drivers/net/ipa/data/ipa_data-v4.5.c
147
.rx = {
drivers/net/ipa/data/ipa_data-v4.5.c
98
.rx = {
drivers/net/ipa/data/ipa_data-v4.7.c
134
.rx = {
drivers/net/ipa/data/ipa_data-v4.7.c
85
.rx = {
drivers/net/ipa/data/ipa_data-v4.9.c
139
.rx = {
drivers/net/ipa/data/ipa_data-v4.9.c
90
.rx = {
drivers/net/ipa/data/ipa_data-v5.0.c
102
.rx = {
drivers/net/ipa/data/ipa_data-v5.0.c
151
.rx = {
drivers/net/ipa/data/ipa_data-v5.5.c
102
.rx = {
drivers/net/ipa/data/ipa_data-v5.5.c
151
.rx = {
drivers/net/ipa/ipa_endpoint.c
1010
rx_config = &endpoint->config.rx;
drivers/net/ipa/ipa_endpoint.c
1309
buffer_size = endpoint->config.rx.buffer_size;
drivers/net/ipa/ipa_endpoint.c
1426
u32 buffer_size = endpoint->config.rx.buffer_size;
drivers/net/ipa/ipa_endpoint.c
1537
u32 buffer_size = endpoint->config.rx.buffer_size;
drivers/net/ipa/ipa_endpoint.c
1570
align = endpoint->config.rx.pad_align ? : 1;
drivers/net/ipa/ipa_endpoint.c
1803
if (endpoint->config.rx.holb_drop)
drivers/net/ipa/ipa_endpoint.c
263
rx_config = &data->endpoint.config.rx;
drivers/net/ipa/ipa_endpoint.c
831
u32 pad_align = endpoint->config.rx.pad_align;
drivers/net/ipa/ipa_endpoint.h
120
struct ipa_endpoint_rx rx;
drivers/net/ipa/ipa_modem.c
101
ipa_endpoint_disable_one(priv->rx);
drivers/net/ipa/ipa_modem.c
256
ipa_endpoint_suspend_one(priv->rx);
drivers/net/ipa/ipa_modem.c
292
ipa_endpoint_resume_one(priv->rx);
drivers/net/ipa/ipa_modem.c
324
priv->rx = ipa->name_map[IPA_ENDPOINT_AP_MODEM_RX];
drivers/net/ipa/ipa_modem.c
328
priv->rx->netdev = netdev;
drivers/net/ipa/ipa_modem.c
335
priv->rx->netdev = NULL;
drivers/net/ipa/ipa_modem.c
379
priv->rx->netdev = NULL;
drivers/net/ipa/ipa_modem.c
47
struct ipa_endpoint *rx;
drivers/net/ipa/ipa_modem.c
68
ret = ipa_endpoint_enable_one(priv->rx);
drivers/net/netdevsim/ethtool.c
28
pause->rx_pause = ns->ethtool.pauseparam.rx;
drivers/net/netdevsim/ethtool.c
40
ns->ethtool.pauseparam.rx = pause->rx_pause;
drivers/net/netdevsim/ipsec.c
183
sa.rx = true;
drivers/net/netdevsim/ipsec.c
45
i, (sap->rx ? 'r' : 't'), &sap->ipaddr);
drivers/net/netdevsim/ipsec.c
49
i, (sap->rx ? 'r' : 't'), &sap->ipaddr[3]);
drivers/net/netdevsim/netdev.c
706
struct netdev_queue_stats_rx *rx,
drivers/net/netdevsim/netdev.c
713
rx->packets = !!rtstats.rx_packets;
drivers/net/netdevsim/netdev.c
714
rx->bytes = 0;
drivers/net/netdevsim/netdevsim.h
49
bool rx;
drivers/net/netdevsim/netdevsim.h
79
bool rx;
drivers/net/ovpn/netlink.c
619
atomic64_read(&peer->vpn_stats.rx.bytes)) ||
drivers/net/ovpn/netlink.c
621
atomic64_read(&peer->vpn_stats.rx.packets)) ||
drivers/net/ovpn/netlink.c
629
atomic64_read(&peer->link_stats.rx.bytes)) ||
drivers/net/ovpn/netlink.c
631
atomic64_read(&peer->link_stats.rx.packets)) ||
drivers/net/ovpn/stats.c
16
atomic64_set(&ps->rx.bytes, 0);
drivers/net/ovpn/stats.c
17
atomic64_set(&ps->rx.packets, 0);
drivers/net/ovpn/stats.h
22
struct ovpn_peer_stat rx;
drivers/net/ovpn/stats.h
38
ovpn_peer_stats_increment(&stats->rx, n);
drivers/net/phy/linkmode.c
89
void linkmode_set_pause(unsigned long *advertisement, bool tx, bool rx)
drivers/net/phy/linkmode.c
91
linkmode_mod_bit(ETHTOOL_LINK_MODE_Pause_BIT, advertisement, rx);
drivers/net/phy/linkmode.c
93
rx ^ tx);
drivers/net/phy/micrel.c
1177
u16 rx, tx, rx_clk, tx_clk;
drivers/net/phy/micrel.c
1184
rx = RX_ND;
drivers/net/phy/micrel.c
1190
rx = RX_ID;
drivers/net/phy/micrel.c
1196
rx = RX_ID;
drivers/net/phy/micrel.c
1202
rx = RX_ND;
drivers/net/phy/micrel.c
1210
FIELD_PREP(MII_KSZ9031RN_RX_CTL_M, rx) |
drivers/net/phy/micrel.c
1216
FIELD_PREP(MII_KSZ9031RN_RXD3, rx) |
drivers/net/phy/micrel.c
1217
FIELD_PREP(MII_KSZ9031RN_RXD2, rx) |
drivers/net/phy/micrel.c
1218
FIELD_PREP(MII_KSZ9031RN_RXD1, rx) |
drivers/net/phy/micrel.c
1219
FIELD_PREP(MII_KSZ9031RN_RXD0, rx));
drivers/net/phy/phy_device.c
2996
void phy_set_sym_pause(struct phy_device *phydev, bool rx, bool tx,
drivers/net/phy/phy_device.c
3001
if (rx && tx && autoneg)
drivers/net/phy/phy_device.c
3020
void phy_set_asym_pause(struct phy_device *phydev, bool rx, bool tx)
drivers/net/phy/phy_device.c
3025
linkmode_set_pause(phydev->advertising, tx, rx);
drivers/net/usb/asix.h
210
struct asix_rx_fixup_info *rx);
drivers/net/usb/asix_common.c
115
static void reset_asix_rx_fixup_info(struct asix_rx_fixup_info *rx)
drivers/net/usb/asix_common.c
122
if (rx->ax_skb) {
drivers/net/usb/asix_common.c
124
kfree_skb(rx->ax_skb);
drivers/net/usb/asix_common.c
125
rx->ax_skb = NULL;
drivers/net/usb/asix_common.c
131
rx->remaining = 0;
drivers/net/usb/asix_common.c
132
rx->split_head = false;
drivers/net/usb/asix_common.c
133
rx->header = 0;
drivers/net/usb/asix_common.c
137
struct asix_rx_fixup_info *rx)
drivers/net/usb/asix_common.c
151
if (rx->remaining && (rx->remaining + sizeof(u32) <= skb->len)) {
drivers/net/usb/asix_common.c
152
offset = ((rx->remaining + 1) & 0xfffe);
drivers/net/usb/asix_common.c
153
rx->header = get_unaligned_le32(skb->data + offset);
drivers/net/usb/asix_common.c
156
size = (u16)(rx->header & 0x7ff);
drivers/net/usb/asix_common.c
157
if (size != ((~rx->header >> 16) & 0x7ff)) {
drivers/net/usb/asix_common.c
159
rx->remaining);
drivers/net/usb/asix_common.c
160
reset_asix_rx_fixup_info(rx);
drivers/net/usb/asix_common.c
167
if (!rx->remaining) {
drivers/net/usb/asix_common.c
169
rx->header = get_unaligned_le16(
drivers/net/usb/asix_common.c
171
rx->split_head = true;
drivers/net/usb/asix_common.c
176
if (rx->split_head == true) {
drivers/net/usb/asix_common.c
177
rx->header |= (get_unaligned_le16(
drivers/net/usb/asix_common.c
179
rx->split_head = false;
drivers/net/usb/asix_common.c
182
rx->header = get_unaligned_le32(skb->data +
drivers/net/usb/asix_common.c
188
size = (u16)(rx->header & 0x7ff);
drivers/net/usb/asix_common.c
189
if (size != ((~rx->header >> 16) & 0x7ff)) {
drivers/net/usb/asix_common.c
191
rx->header, offset);
drivers/net/usb/asix_common.c
192
reset_asix_rx_fixup_info(rx);
drivers/net/usb/asix_common.c
198
reset_asix_rx_fixup_info(rx);
drivers/net/usb/asix_common.c
207
rx->ax_skb = netdev_alloc_skb_ip_align(dev->net, size);
drivers/net/usb/asix_common.c
209
rx->remaining = size;
drivers/net/usb/asix_common.c
212
if (rx->remaining > skb->len - offset) {
drivers/net/usb/asix_common.c
214
rx->remaining -= copy_length;
drivers/net/usb/asix_common.c
216
copy_length = rx->remaining;
drivers/net/usb/asix_common.c
217
rx->remaining = 0;
drivers/net/usb/asix_common.c
220
if (rx->ax_skb) {
drivers/net/usb/asix_common.c
221
skb_put_data(rx->ax_skb, skb->data + offset,
drivers/net/usb/asix_common.c
223
if (!rx->remaining) {
drivers/net/usb/asix_common.c
224
usbnet_skb_return(dev, rx->ax_skb);
drivers/net/usb/asix_common.c
225
rx->ax_skb = NULL;
drivers/net/usb/asix_common.c
235
reset_asix_rx_fixup_info(rx);
drivers/net/usb/asix_common.c
245
struct asix_rx_fixup_info *rx = &dp->rx_fixup_info;
drivers/net/usb/asix_common.c
247
return asix_rx_fixup_internal(dev, skb, rx);
drivers/net/usb/asix_common.c
252
struct asix_rx_fixup_info *rx;
drivers/net/usb/asix_common.c
257
rx = &dp->rx_fixup_info;
drivers/net/usb/asix_common.c
259
if (rx->ax_skb) {
drivers/net/usb/asix_common.c
260
kfree_skb(rx->ax_skb);
drivers/net/usb/asix_common.c
261
rx->ax_skb = NULL;
drivers/net/usb/ax88172a.c
342
struct asix_rx_fixup_info *rx = &dp->rx_fixup_info;
drivers/net/usb/ax88172a.c
344
return asix_rx_fixup_internal(dev, skb, rx);
drivers/net/usb/catc.c
635
u8 rx = RxEnable | RxPolarity | RxMultiCast;
drivers/net/usb/catc.c
645
rx |= (!catc->is_f5u011) ? RxPromisc : AltRxPromisc;
drivers/net/usb/catc.c
661
catc_set_reg_async(catc, RxUnit, rx);
drivers/net/usb/catc.c
665
if (catc->rxmode[0] != rx) {
drivers/net/usb/catc.c
666
catc->rxmode[0] = rx;
drivers/net/usb/hso.c
1601
icount->rx = cnow.rx;
drivers/net/veth.c
456
struct veth_stats rx;
drivers/net/veth.c
461
veth_stats_rx(&rx, dev);
drivers/net/veth.c
462
tot->tx_dropped += rx.xdp_tx_err;
drivers/net/veth.c
463
tot->rx_dropped = rx.rx_drops + rx.peer_tq_xdp_xmit_err;
drivers/net/veth.c
464
tot->rx_bytes += rx.xdp_bytes;
drivers/net/veth.c
465
tot->rx_packets += rx.xdp_packets;
drivers/net/veth.c
476
veth_stats_rx(&rx, peer);
drivers/net/veth.c
477
tot->tx_dropped += rx.peer_tq_xdp_xmit_err;
drivers/net/veth.c
478
tot->rx_dropped += rx.xdp_tx_err;
drivers/net/veth.c
479
tot->tx_bytes += rx.xdp_bytes;
drivers/net/veth.c
480
tot->tx_packets += rx.xdp_packets;
drivers/net/virtio_net.c
5622
struct netdev_queue_stats_rx *rx,
drivers/net/virtio_net.c
5630
rx->bytes = 0;
drivers/net/virtio_net.c
5631
rx->packets = 0;
drivers/net/virtio_net.c
5634
rx->hw_drops = 0;
drivers/net/virtio_net.c
5635
rx->hw_drop_overruns = 0;
drivers/net/virtio_net.c
5639
rx->csum_unnecessary = 0;
drivers/net/virtio_net.c
5640
rx->csum_none = 0;
drivers/net/virtio_net.c
5641
rx->csum_bad = 0;
drivers/net/virtio_net.c
5645
rx->hw_gro_packets = 0;
drivers/net/virtio_net.c
5646
rx->hw_gro_bytes = 0;
drivers/net/virtio_net.c
5647
rx->hw_gro_wire_packets = 0;
drivers/net/virtio_net.c
5648
rx->hw_gro_wire_bytes = 0;
drivers/net/virtio_net.c
5652
rx->hw_drop_ratelimits = 0;
drivers/net/virtio_net.c
5680
dev->real_num_rx_queues, vi->max_queue_pairs, rx,
drivers/net/wireless/ath/ath10k/htt.h
2027
struct htt_rx_fragment_indication *rx,
drivers/net/wireless/ath/ath10k/htt.h
2069
struct htt_rx_fragment_indication *rx,
drivers/net/wireless/ath/ath10k/htt.h
2075
return htt->rx_ops->htt_rx_proc_rx_frag_ind(htt, rx, skb);
drivers/net/wireless/ath/ath10k/htt_rx.c
2399
struct htt_rx_indication_hl *rx)
drivers/net/wireless/ath/ath10k/htt_rx.c
2415
if (!(rx->fw_desc.flags & FW_RX_DESC_FLAGS_FIRST_MSDU))
drivers/net/wireless/ath/ath10k/htt_rx.c
2418
num_mpdu_ranges = MS(__le32_to_cpu(rx->hdr.info1),
drivers/net/wireless/ath/ath10k/htt_rx.c
2421
rx_desc = (struct htt_hl_rx_desc *)&rx->mpdu_ranges[num_mpdu_ranges];
drivers/net/wireless/ath/ath10k/htt_rx.c
2427
tid = MS(rx->hdr.info0, HTT_RX_INDICATION_INFO0_EXT_TID);
drivers/net/wireless/ath/ath10k/htt_rx.c
2456
struct htt_rx_indication_hl *rx,
drivers/net/wireless/ath/ath10k/htt_rx.c
2479
peer_id = __le16_to_cpu(rx->hdr.peer_id);
drivers/net/wireless/ath/ath10k/htt_rx.c
2480
tid = MS(rx->hdr.info0, HTT_RX_INDICATION_INFO0_EXT_TID);
drivers/net/wireless/ath/ath10k/htt_rx.c
2491
num_mpdu_ranges = MS(__le32_to_cpu(rx->hdr.info1),
drivers/net/wireless/ath/ath10k/htt_rx.c
2493
mpdu_ranges = htt_rx_ind_get_mpdu_ranges_hl(rx);
drivers/net/wireless/ath/ath10k/htt_rx.c
2494
fw_desc = &rx->fw_desc;
drivers/net/wireless/ath/ath10k/htt_rx.c
2520
rx_desc = (struct htt_hl_rx_desc *)&rx->mpdu_ranges[num_mpdu_ranges];
drivers/net/wireless/ath/ath10k/htt_rx.c
2529
first_msdu = rx->fw_desc.flags & FW_RX_DESC_FLAGS_FIRST_MSDU;
drivers/net/wireless/ath/ath10k/htt_rx.c
2535
pn_invalid = ath10k_htt_rx_pn_check_replay_hl(ar, peer, rx);
drivers/net/wireless/ath/ath10k/htt_rx.c
2545
tot_hdr_len = sizeof(struct htt_resp_hdr) + sizeof(rx->hdr) +
drivers/net/wireless/ath/ath10k/htt_rx.c
2546
sizeof(rx->ppdu) + sizeof(rx->prefix) +
drivers/net/wireless/ath/ath10k/htt_rx.c
2547
sizeof(rx->fw_desc) +
drivers/net/wireless/ath/ath10k/htt_rx.c
2558
if (rx->ppdu.combined_rssi == 0) {
drivers/net/wireless/ath/ath10k/htt_rx.c
2564
rx->ppdu.combined_rssi;
drivers/net/wireless/ath/ath10k/htt_rx.c
2582
if (rx->fw_desc.flags & FW_RX_DESC_FLAGS_LAST_MSDU)
drivers/net/wireless/ath/ath10k/htt_rx.c
2760
struct htt_rx_fragment_indication *rx,
drivers/net/wireless/ath/ath10k/htt_rx.c
2783
peer_id = __le16_to_cpu(rx->peer_id);
drivers/net/wireless/ath/ath10k/htt_rx.c
2918
struct htt_rx_indication *rx)
drivers/net/wireless/ath/ath10k/htt_rx.c
2927
num_mpdu_ranges = MS(__le32_to_cpu(rx->hdr.info1),
drivers/net/wireless/ath/ath10k/htt_rx.c
2929
peer_id = __le16_to_cpu(rx->hdr.peer_id);
drivers/net/wireless/ath/ath10k/htt_rx.c
2930
tid = MS(rx->hdr.info0, HTT_RX_INDICATION_INFO0_EXT_TID);
drivers/net/wireless/ath/ath10k/htt_rx.c
2932
mpdu_ranges = htt_rx_ind_get_mpdu_ranges(rx);
drivers/net/wireless/ath/ath10k/htt_rx.c
2935
rx, struct_size(rx, mpdu_ranges, num_mpdu_ranges));
drivers/net/wireless/ath/ath10k/htt_rx.c
3216
struct htt_rx_offload_msdu *rx;
drivers/net/wireless/ath/ath10k/htt_rx.c
3225
rx = (void *)msdu->data;
drivers/net/wireless/ath/ath10k/htt_rx.c
3227
skb_put(msdu, sizeof(*rx));
drivers/net/wireless/ath/ath10k/htt_rx.c
3228
skb_pull(msdu, sizeof(*rx));
drivers/net/wireless/ath/ath10k/htt_rx.c
3230
if (skb_tailroom(msdu) < __le16_to_cpu(rx->msdu_len)) {
drivers/net/wireless/ath/ath10k/htt_rx.c
3236
skb_put(msdu, __le16_to_cpu(rx->msdu_len));
drivers/net/wireless/ath/ath10k/htt_rx.c
3256
ath10k_htt_rx_h_channel(ar, status, NULL, rx->vdev_id);
drivers/net/wireless/ath/ath10k/wmi-ops.h
15
void (*rx)(struct ath10k *ar, struct sk_buff *skb);
drivers/net/wireless/ath/ath10k/wmi-ops.h
240
if (WARN_ON_ONCE(!ar->wmi.ops->rx))
drivers/net/wireless/ath/ath10k/wmi-ops.h
243
ar->wmi.ops->rx(ar, skb);
drivers/net/wireless/ath/ath10k/wmi-tlv.c
1555
ath10k_wmi_pull_pdev_stats_rx(&src->rx, dst);
drivers/net/wireless/ath/ath10k/wmi-tlv.c
4529
.rx = ath10k_wmi_tlv_op_rx,
drivers/net/wireless/ath/ath10k/wmi.c
3055
ath10k_wmi_pull_pdev_stats_rx(&src->rx, dst);
drivers/net/wireless/ath/ath10k/wmi.c
3109
ath10k_wmi_pull_pdev_stats_rx(&src->rx, dst);
drivers/net/wireless/ath/ath10k/wmi.c
3170
ath10k_wmi_pull_pdev_stats_rx(&src->rx, dst);
drivers/net/wireless/ath/ath10k/wmi.c
3247
ath10k_wmi_pull_pdev_stats_rx(&src->rx, dst);
drivers/net/wireless/ath/ath10k/wmi.c
3339
ath10k_wmi_pull_pdev_stats_rx(&src->rx, dst);
drivers/net/wireless/ath/ath10k/wmi.c
5813
u64 busy, total, tx, rx, rx_bss;
drivers/net/wireless/ath/ath10k/wmi.c
5827
rx = __le64_to_cpu(ev->cycle_rx);
drivers/net/wireless/ath/ath10k/wmi.c
5832
freq, noise_floor, busy, total, tx, rx, rx_bss);
drivers/net/wireless/ath/ath10k/wmi.c
9150
.rx = ath10k_wmi_op_rx,
drivers/net/wireless/ath/ath10k/wmi.c
9222
.rx = ath10k_wmi_10_1_op_rx,
drivers/net/wireless/ath/ath10k/wmi.c
9294
.rx = ath10k_wmi_10_2_op_rx,
drivers/net/wireless/ath/ath10k/wmi.c
9364
.rx = ath10k_wmi_10_2_op_rx,
drivers/net/wireless/ath/ath10k/wmi.c
9440
.rx = ath10k_wmi_10_4_op_rx,
drivers/net/wireless/ath/ath10k/wmi.h
4707
struct wmi_pdev_stats_rx rx;
drivers/net/wireless/ath/ath10k/wmi.h
4723
struct wmi_pdev_stats_rx rx;
drivers/net/wireless/ath/ath10k/wmi.h
4737
struct wmi_pdev_stats_rx rx;
drivers/net/wireless/ath/ath10k/wmi.h
4747
struct wmi_pdev_stats_rx rx;
drivers/net/wireless/ath/ath11k/ahb.c
551
if (ab->hw_params.ring_mask->rx[i] & BIT(j)) {
drivers/net/wireless/ath/ath11k/dp.c
151
grp_mask = &ab->hw_params.ring_mask->rx[0];
drivers/net/wireless/ath/ath11k/dp.c
810
if (ab->hw_params.ring_mask->rx[grp_id]) {
drivers/net/wireless/ath/ath11k/dp.c
811
i = fls(ab->hw_params.ring_mask->rx[grp_id]) - 1;
drivers/net/wireless/ath/ath11k/hw.c
1225
.rx = {
drivers/net/wireless/ath/ath11k/hw.c
1264
.rx = {
drivers/net/wireless/ath/ath11k/hw.c
2007
.rx = {
drivers/net/wireless/ath/ath11k/hw.c
2048
.rx = {
drivers/net/wireless/ath/ath11k/hw.h
120
u8 rx[ATH11K_EXT_IRQ_GRP_NUM_MAX];
drivers/net/wireless/ath/ath11k/mac.c
1064
param.chains[NL80211_BAND_2GHZ].rx = ar->num_rx_chains;
drivers/net/wireless/ath/ath11k/mac.c
1068
param.chains[NL80211_BAND_5GHZ].rx = ar->num_rx_chains;
drivers/net/wireless/ath/ath11k/mac.c
6854
params->chains[NL80211_BAND_2GHZ].rx = ar->num_rx_chains;
drivers/net/wireless/ath/ath11k/mac.c
6858
params->chains[NL80211_BAND_5GHZ].rx = ar->num_rx_chains;
drivers/net/wireless/ath/ath11k/mac.c
6863
params->chains[NL80211_BAND_6GHZ].rx = ar->num_rx_chains;
drivers/net/wireless/ath/ath11k/pcic.c
606
ab->hw_params.ring_mask->rx[i] ||
drivers/net/wireless/ath/ath11k/wmi.c
6701
ath11k_wmi_pull_pdev_stats_rx(&src->rx, dst);
drivers/net/wireless/ath/ath11k/wmi.c
770
param->chains[NL80211_BAND_2GHZ].rx;
drivers/net/wireless/ath/ath11k/wmi.c
780
param->chains[NL80211_BAND_5GHZ].rx;
drivers/net/wireless/ath/ath11k/wmi.c
8025
u64 busy, total, tx, rx, rx_bss;
drivers/net/wireless/ath/ath11k/wmi.c
8042
rx = (u64)(bss_ch_info_ev.rx_cycle_count_high) << 32 |
drivers/net/wireless/ath/ath11k/wmi.c
8052
tx, rx, rx_bss);
drivers/net/wireless/ath/ath11k/wmi.h
2618
u8 rx;
drivers/net/wireless/ath/ath11k/wmi.h
4861
struct wmi_pdev_stats_rx rx;
drivers/net/wireless/ath/ath12k/ahb.c
588
if (ring_mask->rx[i] & BIT(j)) {
drivers/net/wireless/ath/ath12k/dp.c
160
grp_mask = &ab->hw_params->ring_mask->rx[0];
drivers/net/wireless/ath/ath12k/hw.h
124
u8 rx[ATH12K_EXT_IRQ_GRP_NUM_MAX];
drivers/net/wireless/ath/ath12k/mac.c
9947
arg->chains[NL80211_BAND_2GHZ].rx = ar->num_rx_chains;
drivers/net/wireless/ath/ath12k/mac.c
9951
arg->chains[NL80211_BAND_5GHZ].rx = ar->num_rx_chains;
drivers/net/wireless/ath/ath12k/mac.c
9956
arg->chains[NL80211_BAND_6GHZ].rx = ar->num_rx_chains;
drivers/net/wireless/ath/ath12k/pci.c
565
ab->hw_params->ring_mask->rx[i] ||
drivers/net/wireless/ath/ath12k/wifi7/dp.c
53
if (dp->hw_params->ring_mask->rx[grp_id]) {
drivers/net/wireless/ath/ath12k/wifi7/dp.c
54
i = fls(dp->hw_params->ring_mask->rx[grp_id]) - 1;
drivers/net/wireless/ath/ath12k/wifi7/hw.c
227
.rx = {
drivers/net/wireless/ath/ath12k/wifi7/hw.c
266
.rx = {
drivers/net/wireless/ath/ath12k/wifi7/hw.c
309
.rx = {
drivers/net/wireless/ath/ath12k/wmi.c
7612
u64 busy, total, tx, rx, rx_bss;
drivers/net/wireless/ath/ath12k/wmi.c
7629
rx = (u64)(le32_to_cpu(bss_ch_info_ev.rx_cycle_count_high)) << 32 |
drivers/net/wireless/ath/ath12k/wmi.c
7639
tx, rx, rx_bss);
drivers/net/wireless/ath/ath12k/wmi.c
8343
ath12k_wmi_pull_pdev_stats_rx(&src->rx, dst);
drivers/net/wireless/ath/ath12k/wmi.c
897
cpu_to_le32(args->chains[NL80211_BAND_2GHZ].rx);
drivers/net/wireless/ath/ath12k/wmi.c
906
cpu_to_le32(args->chains[NL80211_BAND_5GHZ].rx);
drivers/net/wireless/ath/ath12k/wmi.h
2885
u8 rx;
drivers/net/wireless/ath/ath12k/wmi.h
6038
struct ath12k_wmi_pdev_rx_stats_params rx;
drivers/net/wireless/ath/ath5k/mac80211-ops.c
745
u32 *tx, u32 *tx_max, u32 *rx, u32 *rx_max)
drivers/net/wireless/ath/ath5k/mac80211-ops.c
752
*rx = *rx_max = ATH_RXBUF;
drivers/net/wireless/ath/ath5k/mac80211-ops.c
756
static int ath5k_set_ringparam(struct ieee80211_hw *hw, u32 tx, u32 rx)
drivers/net/wireless/ath/ath5k/mac80211-ops.c
762
if (rx != ATH_RXBUF)
drivers/net/wireless/ath/ath6kl/cfg80211.c
3413
.rx = BIT(IEEE80211_STYPE_ACTION >> 4) |
drivers/net/wireless/ath/ath6kl/cfg80211.c
3419
.rx = BIT(IEEE80211_STYPE_ACTION >> 4) |
drivers/net/wireless/ath/ath6kl/cfg80211.c
3425
.rx = BIT(IEEE80211_STYPE_ACTION >> 4) |
drivers/net/wireless/ath/ath6kl/cfg80211.c
3431
.rx = BIT(IEEE80211_STYPE_ACTION >> 4) |
drivers/net/wireless/ath/ath6kl/htc.h
325
struct htc_rx_packet_info rx;
drivers/net/wireless/ath/ath6kl/htc.h
347
void (*rx) (struct htc_target *, struct htc_packet *);
drivers/net/wireless/ath/ath6kl/htc_mbox.c
1281
if (packet->info.rx.rx_flags & HTC_RX_PKT_NO_RECYCLE) {
drivers/net/wireless/ath/ath6kl/htc_mbox.c
1284
ep->ep_cb.rx(ep->target, packet);
drivers/net/wireless/ath/ath6kl/htc_mbox.c
1317
packet, packet->info.rx.exp_hdr,
drivers/net/wireless/ath/ath6kl/htc_mbox.c
1343
packet->info.rx.indicat_flags |=
drivers/net/wireless/ath/ath6kl/htc_mbox.c
1434
packet->info.rx.rx_flags = 0;
drivers/net/wireless/ath/ath6kl/htc_mbox.c
1435
packet->info.rx.indicat_flags = 0;
drivers/net/wireless/ath/ath6kl/htc_mbox.c
1444
packet->info.rx.rx_flags |= HTC_RX_PKT_NO_RECYCLE;
drivers/net/wireless/ath/ath6kl/htc_mbox.c
1455
packet->info.rx.rx_flags |= HTC_RX_PKT_REFRESH_HDR;
drivers/net/wireless/ath/ath6kl/htc_mbox.c
1456
packet->info.rx.exp_hdr = 0xFFFFFFFF;
drivers/net/wireless/ath/ath6kl/htc_mbox.c
1459
packet->info.rx.exp_hdr = *lk_ahds;
drivers/net/wireless/ath/ath6kl/htc_mbox.c
1822
if (packet->info.rx.rx_flags & HTC_RX_PKT_REFRESH_HDR) {
drivers/net/wireless/ath/ath6kl/htc_mbox.c
1828
packet->info.rx.exp_hdr = lk_ahd;
drivers/net/wireless/ath/ath6kl/htc_mbox.c
1852
if (lk_ahd != packet->info.rx.exp_hdr) {
drivers/net/wireless/ath/ath6kl/htc_mbox.c
1854
__func__, packet, packet->info.rx.rx_flags);
drivers/net/wireless/ath/ath6kl/htc_mbox.c
1856
"", &packet->info.rx.exp_hdr, 4);
drivers/net/wireless/ath/ath6kl/htc_mbox.c
1872
if (packet->info.rx.rx_flags & HTC_RX_PKT_IGNORE_LOOKAHEAD) {
drivers/net/wireless/ath/ath6kl/htc_mbox.c
1906
endpoint->ep_cb.rx(endpoint->target, packet);
drivers/net/wireless/ath/ath6kl/htc_mbox.c
1971
packet->info.rx.rx_flags |=
drivers/net/wireless/ath/ath6kl/htc_mbox.c
1978
packet->info.rx.rx_flags |= HTC_RX_PKT_PART_OF_BUNDLE;
drivers/net/wireless/ath/ath6kl/htc_mbox.c
2038
packet->info.rx.indicat_flags |=
drivers/net/wireless/ath/ath6kl/htc_mbox.c
2043
if (packet->info.rx.rx_flags & HTC_RX_PKT_PART_OF_BUNDLE)
drivers/net/wireless/ath/ath6kl/htc_mbox.c
2100
packet->info.rx.rx_flags |=
drivers/net/wireless/ath/ath6kl/htc_mbox.c
2279
packet->info.rx.rx_flags = 0;
drivers/net/wireless/ath/ath6kl/htc_mbox.c
2280
packet->info.rx.exp_hdr = look_ahead.word;
drivers/net/wireless/ath/ath6kl/htc_mbox.c
2709
connect.ep_cb.rx = htc_ctrl_rx;
drivers/net/wireless/ath/ath6kl/htc_pipe.c
1597
connect.ep_cb.rx = htc_rxctrl_complete;
drivers/net/wireless/ath/ath6kl/htc_pipe.c
923
ep->ep_cb.rx(ep->target, packet);
drivers/net/wireless/ath/ath6kl/init.c
332
connect.ep_cb.rx = ath6kl_rx;
drivers/net/wireless/ath/ath6kl/main.c
744
stats->rx_pkt += le32_to_cpu(tgt_stats->stats.rx.pkt);
drivers/net/wireless/ath/ath6kl/main.c
745
stats->rx_byte += le32_to_cpu(tgt_stats->stats.rx.byte);
drivers/net/wireless/ath/ath6kl/main.c
746
stats->rx_ucast_pkt += le32_to_cpu(tgt_stats->stats.rx.ucast_pkt);
drivers/net/wireless/ath/ath6kl/main.c
747
stats->rx_ucast_byte += le32_to_cpu(tgt_stats->stats.rx.ucast_byte);
drivers/net/wireless/ath/ath6kl/main.c
748
stats->rx_mcast_pkt += le32_to_cpu(tgt_stats->stats.rx.mcast_pkt);
drivers/net/wireless/ath/ath6kl/main.c
749
stats->rx_mcast_byte += le32_to_cpu(tgt_stats->stats.rx.mcast_byte);
drivers/net/wireless/ath/ath6kl/main.c
750
stats->rx_bcast_pkt += le32_to_cpu(tgt_stats->stats.rx.bcast_pkt);
drivers/net/wireless/ath/ath6kl/main.c
751
stats->rx_bcast_byte += le32_to_cpu(tgt_stats->stats.rx.bcast_byte);
drivers/net/wireless/ath/ath6kl/main.c
752
stats->rx_frgment_pkt += le32_to_cpu(tgt_stats->stats.rx.frgment_pkt);
drivers/net/wireless/ath/ath6kl/main.c
753
stats->rx_err += le32_to_cpu(tgt_stats->stats.rx.err);
drivers/net/wireless/ath/ath6kl/main.c
754
stats->rx_crc_err += le32_to_cpu(tgt_stats->stats.rx.crc_err);
drivers/net/wireless/ath/ath6kl/main.c
756
le32_to_cpu(tgt_stats->stats.rx.key_cache_miss);
drivers/net/wireless/ath/ath6kl/main.c
757
stats->rx_decrypt_err += le32_to_cpu(tgt_stats->stats.rx.decrypt_err);
drivers/net/wireless/ath/ath6kl/main.c
758
stats->rx_dupl_frame += le32_to_cpu(tgt_stats->stats.rx.dupl_frame);
drivers/net/wireless/ath/ath6kl/main.c
760
rate = a_sle32_to_cpu(tgt_stats->stats.rx.ucast_rate);
drivers/net/wireless/ath/ath6kl/wmi.h
1833
struct rx_stats rx;
drivers/net/wireless/ath/ath9k/ar9002_mac.c
415
memset(&ads->u.rx, 0, sizeof(ads->u.rx));
drivers/net/wireless/ath/ath9k/ar9003_phy.c
667
void ar9003_hw_set_chain_masks(struct ath_hw *ah, u8 rx, u8 tx)
drivers/net/wireless/ath/ath9k/ar9003_phy.c
673
REG_WRITE(ah, AR_PHY_RX_CHAINMASK, rx);
drivers/net/wireless/ath/ath9k/ar9003_phy.c
674
REG_WRITE(ah, AR_PHY_CAL_CHAINMASK, rx);
drivers/net/wireless/ath/ath9k/ath9k.h
1028
struct ath_rx rx;
drivers/net/wireless/ath/ath9k/gpio.c
181
(sc->rx.num_pkts < ATH_BTCOEX_STOMP_FTP_THRESH);
drivers/net/wireless/ath/ath9k/gpio.c
185
sc->rx.num_pkts = 0;
drivers/net/wireless/ath/ath9k/htc.h
174
__be32 rx;
drivers/net/wireless/ath/ath9k/htc.h
503
struct ath9k_htc_rx rx;
drivers/net/wireless/ath/ath9k/htc_drv_debug.c
42
be32_to_cpu(cmd_rsp.rx));
drivers/net/wireless/ath/ath9k/htc_drv_init.c
134
req.ep_callbacks.rx = ath9k_htc_rxep;
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
1091
spin_lock_irqsave(&priv->rx.rxbuflock, flags);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
1092
list_for_each_entry(tmp_buf, &priv->rx.rxbuf, list) {
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
1100
spin_unlock_irqrestore(&priv->rx.rxbuflock, flags);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
1120
spin_unlock_irqrestore(&priv->rx.rxbuflock, flags);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
1124
spin_lock_irqsave(&priv->rx.rxbuflock, flags);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
1128
list_move_tail(&rxbuf->list, &priv->rx.rxbuf);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
1130
spin_unlock_irqrestore(&priv->rx.rxbuflock, flags);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
1145
if (!data_race(priv->rx.initialized))
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
1148
spin_lock_irqsave(&priv->rx.rxbuflock, flags);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
1149
list_for_each_entry(tmp_buf, &priv->rx.rxbuf, list) {
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
1155
spin_unlock_irqrestore(&priv->rx.rxbuflock, flags);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
1162
spin_lock_irqsave(&priv->rx.rxbuflock, flags);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
1165
spin_unlock_irqrestore(&priv->rx.rxbuflock, flags);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
1179
list_for_each_entry_safe(rxbuf, tbuf, &priv->rx.rxbuf, list) {
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
1191
INIT_LIST_HEAD(&priv->rx.rxbuf);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
1192
spin_lock_init(&priv->rx.rxbuflock);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
1200
list_add_tail(&rxbuf->list, &priv->rx.rxbuf);
drivers/net/wireless/ath/ath9k/htc_drv_txrx.c
1205
priv->rx.initialized = true;
drivers/net/wireless/ath/ath9k/htc_hst.c
487
if (endpoint->ep_callbacks.rx)
drivers/net/wireless/ath/ath9k/htc_hst.c
488
endpoint->ep_callbacks.rx(endpoint->ep_callbacks.priv,
drivers/net/wireless/ath/ath9k/htc_hst.h
95
void (*rx) (void *, struct sk_buff *, enum htc_endpoint_id);
drivers/net/wireless/ath/ath9k/hw.h
1132
void ar9003_hw_set_chain_masks(struct ath_hw *ah, u8 rx, u8 tx);
drivers/net/wireless/ath/ath9k/mac.c
534
ads.u.rx = adsp->u.rx;
drivers/net/wireless/ath/ath9k/mac.h
314
} rx;
drivers/net/wireless/ath/ath9k/mac.h
343
#define ds_rxstatus0 u.rx.status0
drivers/net/wireless/ath/ath9k/mac.h
344
#define ds_rxstatus1 u.rx.status1
drivers/net/wireless/ath/ath9k/mac.h
345
#define ds_rxstatus2 u.rx.status2
drivers/net/wireless/ath/ath9k/mac.h
346
#define ds_rxstatus3 u.rx.status3
drivers/net/wireless/ath/ath9k/mac.h
347
#define ds_rxstatus4 u.rx.status4
drivers/net/wireless/ath/ath9k/mac.h
348
#define ds_rxstatus5 u.rx.status5
drivers/net/wireless/ath/ath9k/mac.h
349
#define ds_rxstatus6 u.rx.status6
drivers/net/wireless/ath/ath9k/mac.h
350
#define ds_rxstatus7 u.rx.status7
drivers/net/wireless/ath/ath9k/mac.h
351
#define ds_rxstatus8 u.rx.status8
drivers/net/wireless/ath/ath9k/main.c
953
if (sc->rx.frag) {
drivers/net/wireless/ath/ath9k/main.c
954
dev_kfree_skb_any(sc->rx.frag);
drivers/net/wireless/ath/ath9k/main.c
955
sc->rx.frag = NULL;
drivers/net/wireless/ath/ath9k/recv.c
1002
sc->rx.ampdu_ref++;
drivers/net/wireless/ath/ath9k/recv.c
1108
if (sc->rx.frag)
drivers/net/wireless/ath/ath9k/recv.c
1109
hdr_skb = sc->rx.frag;
drivers/net/wireless/ath/ath9k/recv.c
1164
if (sc->rx.frag) {
drivers/net/wireless/ath/ath9k/recv.c
1166
dev_kfree_skb_any(sc->rx.frag);
drivers/net/wireless/ath/ath9k/recv.c
1171
sc->rx.frag = skb;
drivers/net/wireless/ath/ath9k/recv.c
1175
if (sc->rx.frag) {
drivers/net/wireless/ath/ath9k/recv.c
118
rx_edma = &sc->rx.rx_edma[qtype];
drivers/net/wireless/ath/ath9k/recv.c
1184
sc->rx.frag = NULL;
drivers/net/wireless/ath/ath9k/recv.c
1215
if (sc->rx.frag) {
drivers/net/wireless/ath/ath9k/recv.c
1216
dev_kfree_skb_any(sc->rx.frag);
drivers/net/wireless/ath/ath9k/recv.c
1217
sc->rx.frag = NULL;
drivers/net/wireless/ath/ath9k/recv.c
122
bf = list_first_entry(&sc->rx.rxbuf, struct ath_rxbuf, list);
drivers/net/wireless/ath/ath9k/recv.c
1220
list_add_tail(&bf->list, &sc->rx.rxbuf);
drivers/net/wireless/ath/ath9k/recv.c
144
if (list_empty(&sc->rx.rxbuf)) {
drivers/net/wireless/ath/ath9k/recv.c
149
list_for_each_entry_safe(bf, tbf, &sc->rx.rxbuf, list)
drivers/net/wireless/ath/ath9k/recv.c
162
rx_edma = &sc->rx.rx_edma[qtype];
drivers/net/wireless/ath/ath9k/recv.c
167
list_add_tail(&bf->list, &sc->rx.rxbuf);
drivers/net/wireless/ath/ath9k/recv.c
180
list_for_each_entry(bf, &sc->rx.rxbuf, list) {
drivers/net/wireless/ath/ath9k/recv.c
210
ath_rx_edma_init_queue(&sc->rx.rx_edma[ATH9K_RX_QUEUE_LP],
drivers/net/wireless/ath/ath9k/recv.c
212
ath_rx_edma_init_queue(&sc->rx.rx_edma[ATH9K_RX_QUEUE_HP],
drivers/net/wireless/ath/ath9k/recv.c
220
INIT_LIST_HEAD(&sc->rx.rxbuf);
drivers/net/wireless/ath/ath9k/recv.c
246
list_add_tail(&bf->list, &sc->rx.rxbuf);
drivers/net/wireless/ath/ath9k/recv.c
291
error = ath_descdma_setup(sc, &sc->rx.rxdma, &sc->rx.rxbuf,
drivers/net/wireless/ath/ath9k/recv.c
300
list_for_each_entry(bf, &sc->rx.rxbuf, list) {
drivers/net/wireless/ath/ath9k/recv.c
323
sc->rx.rxlink = NULL;
drivers/net/wireless/ath/ath9k/recv.c
343
list_for_each_entry(bf, &sc->rx.rxbuf, list) {
drivers/net/wireless/ath/ath9k/recv.c
451
if (list_empty(&sc->rx.rxbuf))
drivers/net/wireless/ath/ath9k/recv.c
454
sc->rx.buf_hold = NULL;
drivers/net/wireless/ath/ath9k/recv.c
455
sc->rx.rxlink = NULL;
drivers/net/wireless/ath/ath9k/recv.c
456
list_for_each_entry_safe(bf, tbf, &sc->rx.rxbuf, list) {
drivers/net/wireless/ath/ath9k/recv.c
461
if (list_empty(&sc->rx.rxbuf))
drivers/net/wireless/ath/ath9k/recv.c
464
bf = list_first_entry(&sc->rx.rxbuf, struct ath_rxbuf, list);
drivers/net/wireless/ath/ath9k/recv.c
494
sc->rx.rxlink = NULL;
drivers/net/wireless/ath/ath9k/recv.c
63
if (sc->rx.rxlink)
drivers/net/wireless/ath/ath9k/recv.c
632
struct ath_rx_edma *rx_edma = &sc->rx.rx_edma[qtype];
drivers/net/wireless/ath/ath9k/recv.c
64
*sc->rx.rxlink = bf->bf_daddr;
drivers/net/wireless/ath/ath9k/recv.c
660
list_add_tail(&bf->list, &sc->rx.rxbuf);
drivers/net/wireless/ath/ath9k/recv.c
669
list_add_tail(&bf->list, &sc->rx.rxbuf);
drivers/net/wireless/ath/ath9k/recv.c
68
sc->rx.rxlink = &ds->ds_link;
drivers/net/wireless/ath/ath9k/recv.c
704
if (list_empty(&sc->rx.rxbuf)) {
drivers/net/wireless/ath/ath9k/recv.c
705
sc->rx.rxlink = NULL;
drivers/net/wireless/ath/ath9k/recv.c
709
bf = list_first_entry(&sc->rx.rxbuf, struct ath_rxbuf, list);
drivers/net/wireless/ath/ath9k/recv.c
710
if (bf == sc->rx.buf_hold)
drivers/net/wireless/ath/ath9k/recv.c
733
if (list_is_last(&bf->list, &sc->rx.rxbuf)) {
drivers/net/wireless/ath/ath9k/recv.c
734
sc->rx.rxlink = NULL;
drivers/net/wireless/ath/ath9k/recv.c
74
if (sc->rx.buf_hold)
drivers/net/wireless/ath/ath9k/recv.c
75
ath_rx_buf_link(sc, sc->rx.buf_hold, flush);
drivers/net/wireless/ath/ath9k/recv.c
77
sc->rx.buf_hold = bf;
drivers/net/wireless/ath/ath9k/recv.c
818
bool discard_current = sc->rx.discard_next;
drivers/net/wireless/ath/ath9k/recv.c
828
sc->rx.discard_next = false;
drivers/net/wireless/ath/ath9k/recv.c
84
sc->rx.defant = antenna;
drivers/net/wireless/ath/ath9k/recv.c
85
sc->rx.rxotherant = 0;
drivers/net/wireless/ath/ath9k/recv.c
943
sc->rx.num_pkts++;
drivers/net/wireless/ath/ath9k/recv.c
949
sc->rx.discard_next = rx_stats->rs_more;
drivers/net/wireless/ath/ath9k/recv.c
977
if (sc->rx.defant != rs->rs_antenna) {
drivers/net/wireless/ath/ath9k/recv.c
978
if (++sc->rx.rxotherant >= 3)
drivers/net/wireless/ath/ath9k/recv.c
981
sc->rx.rxotherant = 0;
drivers/net/wireless/ath/ath9k/recv.c
998
rxs->ampdu_reference = sc->rx.ampdu_ref;
drivers/net/wireless/ath/ath9k/wmi.c
274
connect.ep_callbacks.rx = ath9k_wmi_ctrl_rx;
drivers/net/wireless/ath/hw.c
144
u32 cycles, busy, rx, tx;
drivers/net/wireless/ath/hw.c
153
rx = REG_READ(ah, AR_RFCNT);
drivers/net/wireless/ath/hw.c
168
common->cc_ani.rx_frame += rx;
drivers/net/wireless/ath/hw.c
173
common->cc_survey.rx_frame += rx;
drivers/net/wireless/ath/wil6210/cfg80211.c
277
.rx = BIT(IEEE80211_STYPE_ACTION >> 4) |
drivers/net/wireless/ath/wil6210/cfg80211.c
287
.rx = BIT(IEEE80211_STYPE_ACTION >> 4) |
drivers/net/wireless/ath/wil6210/cfg80211.c
298
.rx = BIT(IEEE80211_STYPE_ACTION >> 4) |
drivers/net/wireless/ath/wil6210/cfg80211.c
304
.rx = BIT(IEEE80211_STYPE_ACTION >> 4) |
drivers/net/wireless/ath/wil6210/cfg80211.c
310
.rx = BIT(IEEE80211_STYPE_ACTION >> 4) |
drivers/net/wireless/ath/wil6210/debugfs.c
1122
&ring->va[txdesc_idx].rx.enhanced;
drivers/net/wireless/ath/wil6210/debugfs.c
1451
int rx = atomic_xchg(&wil->isr_count_rx, 0);
drivers/net/wireless/ath/wil6210/debugfs.c
1460
seq_printf(s, "Rx irqs:packets : %8d : %8ld\n", rx, rxf - rxf_old);
drivers/net/wireless/ath/wil6210/debugfs.c
366
offsetof(struct wil6210_mbox_ctl, rx));
drivers/net/wireless/ath/wil6210/debugfs.c
52
&ring->va[idx].rx.enhanced;
drivers/net/wireless/ath/wil6210/interrupt.c
514
wil_mbox_ring_le2cpus(&wil->mbox_ctl.rx);
drivers/net/wireless/ath/wil6210/interrupt.c
523
if (wil->mbox_ctl.rx.entry_size < min_size) {
drivers/net/wireless/ath/wil6210/interrupt.c
525
wil->mbox_ctl.rx.entry_size);
drivers/net/wireless/ath/wil6210/txrx.c
234
&vring->va[vring->swhead].rx.legacy;
drivers/net/wireless/ath/wil6210/txrx.c
262
volatile struct vring_rx_desc *_d = &vring->va[i].rx.legacy;
drivers/net/wireless/ath/wil6210/txrx.c
356
_d = (struct vring_rx_desc *)&ring->va[ring->swhead].rx.legacy;
drivers/net/wireless/ath/wil6210/txrx.c
464
_d = &vring->va[i].rx.legacy;
drivers/net/wireless/ath/wil6210/txrx.h
490
union wil_rx_desc rx;
drivers/net/wireless/ath/wil6210/txrx_edma.c
165
&ring->va[i].rx.enhanced;
drivers/net/wireless/ath/wil6210/wil6210.h
471
struct wil6210_mbox_ring rx;
drivers/net/wireless/ath/wil6210/wmi.c
1909
struct wil6210_mbox_ring *r = &wil->mbox_ctl.rx;
drivers/net/wireless/ath/wil6210/wmi.c
1933
offsetof(struct wil6210_mbox_ctl, rx.head));
drivers/net/wireless/ath/wil6210/wmi.c
2017
offsetof(struct wil6210_mbox_ctl, rx.tail), r->tail);
drivers/net/wireless/ath/wil6210/wmi.c
3425
struct wil6210_mbox_ring *r = &wil->mbox_ctl.rx;
drivers/net/wireless/ath/wil6210/wmi.c
3444
offsetof(struct wil6210_mbox_ctl, rx.head));
drivers/net/wireless/broadcom/b43/phy_lp.c
1840
bool rx, bool pa, struct lpphy_tx_gains *gains)
drivers/net/wireless/broadcom/b43/phy_lp.c
1876
rx = false;
drivers/net/wireless/broadcom/b43/phy_lp.c
1880
lpphy_set_trsw_over(dev, tx, rx);
drivers/net/wireless/broadcom/b43/phy_lp.c
746
static void lpphy_set_trsw_over(struct b43_wldev *dev, bool tx, bool rx)
drivers/net/wireless/broadcom/b43/phy_lp.c
748
u16 trsw = (tx << 1) | rx;
drivers/net/wireless/broadcom/b43/radio_2056.c
10233
B2056_RX0, pts->rx, pts->rx_length);
drivers/net/wireless/broadcom/b43/radio_2056.c
10235
B2056_RX1, pts->rx, pts->rx_length);
drivers/net/wireless/broadcom/b43/radio_2056.c
3013
.rx = prefix##_rx, \
drivers/net/wireless/broadcom/b43/radio_2056.c
34
const struct b2056_inittab_entry *rx;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
7398
.rx = BIT(IEEE80211_STYPE_ACTION >> 4) |
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
7404
.rx = BIT(IEEE80211_STYPE_ACTION >> 4) |
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
7409
.rx = BIT(IEEE80211_STYPE_ASSOC_REQ >> 4) |
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
7419
.rx = BIT(IEEE80211_STYPE_ACTION >> 4) |
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
7424
.rx = BIT(IEEE80211_STYPE_ASSOC_REQ >> 4) |
drivers/net/wireless/broadcom/brcm80211/brcmsmac/d11.h
71
struct pio2regs rx;
drivers/net/wireless/broadcom/brcm80211/brcmsmac/d11.h
83
struct pio4regs rx;
drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c
1097
static void wlc_lcnphy_set_trsw_override(struct brcms_phy *pi, bool tx, bool rx)
drivers/net/wireless/broadcom/brcm80211/brcmsmac/phy/phy_lcn.c
1102
(0x1 << 0), (tx ? (0x1 << 1) : 0) | (rx ? (0x1 << 0) : 0));
drivers/net/wireless/intel/iwlegacy/3945-debug.c
63
ofdm = &il->_3945.stats.rx.ofdm;
drivers/net/wireless/intel/iwlegacy/3945-debug.c
64
cck = &il->_3945.stats.rx.cck;
drivers/net/wireless/intel/iwlegacy/3945-debug.c
65
general = &il->_3945.stats.rx.general;
drivers/net/wireless/intel/iwlegacy/3945-debug.c
66
accum_ofdm = &il->_3945.accum_stats.rx.ofdm;
drivers/net/wireless/intel/iwlegacy/3945-debug.c
67
accum_cck = &il->_3945.accum_stats.rx.cck;
drivers/net/wireless/intel/iwlegacy/3945-debug.c
68
accum_general = &il->_3945.accum_stats.rx.general;
drivers/net/wireless/intel/iwlegacy/3945-debug.c
69
delta_ofdm = &il->_3945.delta_stats.rx.ofdm;
drivers/net/wireless/intel/iwlegacy/3945-debug.c
70
delta_cck = &il->_3945.delta_stats.rx.cck;
drivers/net/wireless/intel/iwlegacy/3945-debug.c
71
delta_general = &il->_3945.delta_stats.rx.general;
drivers/net/wireless/intel/iwlegacy/3945-debug.c
72
max_ofdm = &il->_3945.max_delta.rx.ofdm;
drivers/net/wireless/intel/iwlegacy/3945-debug.c
73
max_cck = &il->_3945.max_delta.rx.cck;
drivers/net/wireless/intel/iwlegacy/3945-debug.c
74
max_general = &il->_3945.max_delta.rx.general;
drivers/net/wireless/intel/iwlegacy/3945-mac.c
1439
il->isr_stats.rx++;
drivers/net/wireless/intel/iwlegacy/4965-calib.c
511
rx_info = &(((struct il_notif_stats *)resp)->rx.general);
drivers/net/wireless/intel/iwlegacy/4965-calib.c
512
ofdm = &(((struct il_notif_stats *)resp)->rx.ofdm);
drivers/net/wireless/intel/iwlegacy/4965-calib.c
513
cck = &(((struct il_notif_stats *)resp)->rx.cck);
drivers/net/wireless/intel/iwlegacy/4965-calib.c
814
rx_info = &(((struct il_notif_stats *)stat_resp)->rx.general);
drivers/net/wireless/intel/iwlegacy/4965-debug.c
72
ofdm = &il->_4965.stats.rx.ofdm;
drivers/net/wireless/intel/iwlegacy/4965-debug.c
73
cck = &il->_4965.stats.rx.cck;
drivers/net/wireless/intel/iwlegacy/4965-debug.c
74
general = &il->_4965.stats.rx.general;
drivers/net/wireless/intel/iwlegacy/4965-debug.c
75
ht = &il->_4965.stats.rx.ofdm_ht;
drivers/net/wireless/intel/iwlegacy/4965-debug.c
76
accum_ofdm = &il->_4965.accum_stats.rx.ofdm;
drivers/net/wireless/intel/iwlegacy/4965-debug.c
77
accum_cck = &il->_4965.accum_stats.rx.cck;
drivers/net/wireless/intel/iwlegacy/4965-debug.c
78
accum_general = &il->_4965.accum_stats.rx.general;
drivers/net/wireless/intel/iwlegacy/4965-debug.c
79
accum_ht = &il->_4965.accum_stats.rx.ofdm_ht;
drivers/net/wireless/intel/iwlegacy/4965-debug.c
80
delta_ofdm = &il->_4965.delta_stats.rx.ofdm;
drivers/net/wireless/intel/iwlegacy/4965-debug.c
81
delta_cck = &il->_4965.delta_stats.rx.cck;
drivers/net/wireless/intel/iwlegacy/4965-debug.c
82
delta_general = &il->_4965.delta_stats.rx.general;
drivers/net/wireless/intel/iwlegacy/4965-debug.c
83
delta_ht = &il->_4965.delta_stats.rx.ofdm_ht;
drivers/net/wireless/intel/iwlegacy/4965-debug.c
84
max_ofdm = &il->_4965.max_delta.rx.ofdm;
drivers/net/wireless/intel/iwlegacy/4965-debug.c
85
max_cck = &il->_4965.max_delta.rx.cck;
drivers/net/wireless/intel/iwlegacy/4965-debug.c
86
max_general = &il->_4965.max_delta.rx.general;
drivers/net/wireless/intel/iwlegacy/4965-debug.c
87
max_ht = &il->_4965.max_delta.rx.ofdm_ht;
drivers/net/wireless/intel/iwlegacy/4965-mac.c
1282
rx_info = &(il->_4965.stats.rx.general);
drivers/net/wireless/intel/iwlegacy/4965-mac.c
4479
il->isr_stats.rx++;
drivers/net/wireless/intel/iwlegacy/commands.h
2994
struct iwl39_stats_rx rx;
drivers/net/wireless/intel/iwlegacy/commands.h
3001
struct stats_rx rx;
drivers/net/wireless/intel/iwlegacy/common.h
1024
u32 rx;
drivers/net/wireless/intel/iwlegacy/debug.c
692
il->isr_stats.rx);
drivers/net/wireless/intel/iwlwifi/dvm/commands.h
2714
struct statistics_rx rx;
drivers/net/wireless/intel/iwlwifi/dvm/commands.h
2721
struct statistics_rx_bt rx;
drivers/net/wireless/intel/iwlwifi/dvm/main.c
2141
.rx = iwl_rx_dispatch,
drivers/net/wireless/intel/iwlwifi/dvm/rx.c
369
rx_non_phy = &stats->rx.general.common;
drivers/net/wireless/intel/iwlwifi/dvm/rx.c
370
rx_ofdm = &stats->rx.ofdm;
drivers/net/wireless/intel/iwlwifi/dvm/rx.c
371
rx_ofdm_ht = &stats->rx.ofdm_ht;
drivers/net/wireless/intel/iwlwifi/dvm/rx.c
372
rx_cck = &stats->rx.cck;
drivers/net/wireless/intel/iwlwifi/dvm/rx.c
378
priv->statistics.num_bt_kills = stats->rx.general.num_bt_kills;
drivers/net/wireless/intel/iwlwifi/dvm/rx.c
380
le32_to_cpu(stats->rx.general.num_bt_kills));
drivers/net/wireless/intel/iwlwifi/dvm/rx.c
387
rx_non_phy = &stats->rx.general;
drivers/net/wireless/intel/iwlwifi/dvm/rx.c
388
rx_ofdm = &stats->rx.ofdm;
drivers/net/wireless/intel/iwlwifi/dvm/rx.c
389
rx_ofdm_ht = &stats->rx.ofdm_ht;
drivers/net/wireless/intel/iwlwifi/dvm/rx.c
390
rx_cck = &stats->rx.cck;
drivers/net/wireless/intel/iwlwifi/fw/api/stats.h
364
struct mvm_statistics_rx_v3 rx;
drivers/net/wireless/intel/iwlwifi/fw/api/stats.h
371
struct mvm_statistics_rx_v3 rx;
drivers/net/wireless/intel/iwlwifi/fw/api/stats.h
379
struct mvm_statistics_rx rx;
drivers/net/wireless/intel/iwlwifi/iwl-op-mode.h
159
void (*rx)(struct iwl_op_mode *op_mode, struct napi_struct *napi,
drivers/net/wireless/intel/iwlwifi/iwl-op-mode.h
208
return op_mode->ops->rx(op_mode, napi, rxb);
drivers/net/wireless/intel/iwlwifi/mld/mld.c
760
.rx = iwl_mld_rx,
drivers/net/wireless/intel/iwlwifi/mld/mlo.c
587
total_rx += queue_counter->per_link[link].rx;
drivers/net/wireless/intel/iwlwifi/mld/mlo.c
592
sec_link_rx += queue_counter->per_link[sec_link_id].rx;
drivers/net/wireless/intel/iwlwifi/mld/sta.c
902
link_counter->rx += count;
drivers/net/wireless/intel/iwlwifi/mld/sta.c
913
queue_counter->per_link[i].rx;
drivers/net/wireless/intel/iwlwifi/mld/sta.h
80
u32 rx;
drivers/net/wireless/intel/iwlwifi/mld/tlc.c
295
u8 rx = u8_get_bits(rx_nss, IEEE80211_EHT_MCS_NSS_RX);
drivers/net/wireless/intel/iwlwifi/mld/tlc.c
299
return min(tx, rx);
drivers/net/wireless/intel/iwlwifi/mld/tlc.c
302
#define MAX_NSS_MCS(mcs_num, rx, tx) \
drivers/net/wireless/intel/iwlwifi/mld/tlc.c
303
iwl_mld_get_eht_max_nss((rx)->rx_tx_mcs ##mcs_num## _max_nss, \
drivers/net/wireless/intel/iwlwifi/mvm/mvm.h
661
} rx;
drivers/net/wireless/intel/iwlwifi/mvm/ops.c
2114
.rx = iwl_mvm_rx,
drivers/net/wireless/intel/iwlwifi/mvm/ops.c
2140
.rx = iwl_mvm_rx_mq,
drivers/net/wireless/intel/iwlwifi/mvm/rs-fw.c
245
u8 rx = u8_get_bits(rx_nss, IEEE80211_EHT_MCS_NSS_RX);
drivers/net/wireless/intel/iwlwifi/mvm/rs-fw.c
249
return min(tx, rx);
drivers/net/wireless/intel/iwlwifi/mvm/rs-fw.c
252
#define MAX_NSS_MCS(mcs_num, rx, tx) \
drivers/net/wireless/intel/iwlwifi/mvm/rs-fw.c
253
rs_fw_eht_max_nss((rx)->rx_tx_mcs ##mcs_num## _max_nss, \
drivers/net/wireless/intel/iwlwifi/mvm/rx.c
1110
data.mac_id = stats->rx.general.mac_id;
drivers/net/wireless/intel/iwlwifi/mvm/rx.c
1114
mvm->rx_stats_v3 = stats->rx;
drivers/net/wireless/intel/iwlwifi/mvm/rx.c
1132
data.mac_id = stats->rx.general.mac_id;
drivers/net/wireless/intel/iwlwifi/mvm/rx.c
1136
mvm->rx_stats = stats->rx;
drivers/net/wireless/intel/iwlwifi/mvm/rx.c
229
mdata->rx.pkts[ac]++;
drivers/net/wireless/intel/iwlwifi/mvm/rx.c
232
if (mdata->rx.last_ampdu_ref != mvm->ampdu_ref) {
drivers/net/wireless/intel/iwlwifi/mvm/rx.c
233
mdata->rx.last_ampdu_ref = mvm->ampdu_ref;
drivers/net/wireless/intel/iwlwifi/mvm/rx.c
234
mdata->rx.airtime += le16_to_cpu(phy_info->frame_time);
drivers/net/wireless/intel/iwlwifi/mvm/rx.c
748
mdata->rx.airtime += airtime;
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
1003
vo_vi_pkts += mdata->rx.pkts[ac] +
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
1014
memset(&mdata->rx.pkts, 0, sizeof(mdata->rx.pkts));
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
1025
memset(&mdata->rx.airtime, 0, sizeof(mdata->rx.airtime));
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
1133
memset(&mdata->rx.pkts, 0, sizeof(mdata->rx.pkts));
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
1135
memset(&mdata->rx.airtime, 0, sizeof(mdata->rx.airtime));
drivers/net/wireless/intel/iwlwifi/mvm/utils.c
992
u32 airtime = mdata->rx.airtime + mdata->tx.airtime;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/internal.h
70
u32 rx;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/rx.c
2097
isr_stats->rx++;
drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/trans.c
2923
"Rx command responses:\t\t %u\n", isr_stats->rx);
drivers/net/wireless/marvell/mwifiex/cfg80211.c
1747
.rx = BIT(IEEE80211_STYPE_ACTION >> 4) |
drivers/net/wireless/marvell/mwifiex/cfg80211.c
1753
.rx = BIT(IEEE80211_STYPE_ACTION >> 4) |
drivers/net/wireless/marvell/mwifiex/cfg80211.c
1759
.rx = BIT(IEEE80211_STYPE_ACTION >> 4) |
drivers/net/wireless/marvell/mwifiex/cfg80211.c
1765
.rx = BIT(IEEE80211_STYPE_ACTION >> 4) |
drivers/net/wireless/marvell/mwifiex/cfg80211.c
4737
adapter->mwifiex_mgmt_stypes[NL80211_IFTYPE_AP].rx =
drivers/net/wireless/marvell/mwl8k.c
1221
int rx;
drivers/net/wireless/marvell/mwl8k.c
1236
rx = rxq->tail++;
drivers/net/wireless/marvell/mwl8k.c
1239
rxq->buf[rx].skb = skb;
drivers/net/wireless/marvell/mwl8k.c
1240
dma_unmap_addr_set(&rxq->buf[rx], dma, addr);
drivers/net/wireless/marvell/mwl8k.c
1242
rxd = rxq->rxd + (rx * priv->rxd_ops->rxd_size);
drivers/net/wireless/marvell/mwl8k.c
3613
static int mwl8k_cmd_mimo_config(struct ieee80211_hw *hw, __u8 rx, __u8 tx)
drivers/net/wireless/marvell/mwl8k.c
3625
cmd->rx_antenna_map = rx;
drivers/net/wireless/mediatek/mt76/mt7615/mt7615.h
120
} rx;
drivers/net/wireless/mediatek/mt76/mt7615/sdio.c
69
intr->rx.num = irq_data->rx.num;
drivers/net/wireless/mediatek/mt76/mt7615/sdio.c
71
intr->rx.len[i] = irq_data->rx.len[i];
drivers/net/wireless/mediatek/mt76/mt7615/testmode.c
327
void *rx, *rssi;
drivers/net/wireless/mediatek/mt76/mt7615/testmode.c
330
rx = nla_nest_start(msg, MT76_TM_STATS_ATTR_LAST_RX);
drivers/net/wireless/mediatek/mt76/mt7615/testmode.c
331
if (!rx)
drivers/net/wireless/mediatek/mt76/mt7615/testmode.c
367
nla_nest_end(msg, rx);
drivers/net/wireless/mediatek/mt76/mt76_connac_mcu.c
550
struct wtbl_rx *rx;
drivers/net/wireless/mediatek/mt76/mt76_connac_mcu.c
578
tlv = mt76_connac_mcu_add_nested_tlv(skb, WTBL_RX, sizeof(*rx),
drivers/net/wireless/mediatek/mt76/mt76_connac_mcu.c
581
rx = (struct wtbl_rx *)tlv;
drivers/net/wireless/mediatek/mt76/mt76_connac_mcu.c
582
rx->rca1 = sta ? vif->type != NL80211_IFTYPE_AP : 1;
drivers/net/wireless/mediatek/mt76/mt76_connac_mcu.c
583
rx->rca2 = 1;
drivers/net/wireless/mediatek/mt76/mt76_connac_mcu.c
584
rx->rv = 1;
drivers/net/wireless/mediatek/mt76/mt76x0/eeprom.c
113
struct mt76x02_rx_freq_cal *caldata = &dev->cal.rx;
drivers/net/wireless/mediatek/mt76/mt76x0/eeprom.c
88
dev->cal.rx.temp_offset = mt76x02_sign_extend(val, 8);
drivers/net/wireless/mediatek/mt76/mt76x0/eeprom.c
90
dev->cal.rx.temp_offset = -10;
drivers/net/wireless/mediatek/mt76/mt76x0/eeprom.c
95
struct mt76x02_rx_freq_cal *caldata = &dev->cal.rx;
drivers/net/wireless/mediatek/mt76/mt76x0/phy.c
1038
val = (35 * (val - dev->cal.rx.temp_offset)) / 10 + 25;
drivers/net/wireless/mediatek/mt76/mt76x0/phy.c
1191
min_t(u8, dev->cal.rx.freq_offset, 0xbf));
drivers/net/wireless/mediatek/mt76/mt76x0/phy.c
416
gain -= dev->cal.rx.lna_gain * 2;
drivers/net/wireless/mediatek/mt76/mt76x02.h
38
struct mt76x02_rx_freq_cal rx;
drivers/net/wireless/mediatek/mt76/mt76x02_mac.c
763
struct mt76x02_rx_freq_cal *cal = &dev->cal.rx;
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c
185
s8 *dest = dev->cal.rx.high_gain;
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c
200
s8 *dest = dev->cal.rx.rssi_offset;
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c
274
dev->cal.rx.mcu_gain = (lna_2g & 0xff);
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c
275
dev->cal.rx.mcu_gain |= (lna_5g[0] & 0xff) << 8;
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c
276
dev->cal.rx.mcu_gain |= (lna_5g[1] & 0xff) << 16;
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c
277
dev->cal.rx.mcu_gain |= (lna_5g[2] & 0xff) << 24;
drivers/net/wireless/mediatek/mt76/mt76x2/eeprom.c
288
dev->cal.rx.lna_gain = mt76x02_sign_extend(lna, 8);
drivers/net/wireless/mediatek/mt76/mt76x2/pci_phy.c
204
mt76x2_mcu_init_gain(dev, channel, dev->cal.rx.mcu_gain, true);
drivers/net/wireless/mediatek/mt76/mt76x2/phy.c
35
s8 *gain_adj = dev->cal.rx.high_gain;
drivers/net/wireless/mediatek/mt76/mt76x2/usb_phy.c
141
mt76x2_mcu_init_gain(dev, channel, dev->cal.rx.mcu_gain, true);
drivers/net/wireless/mediatek/mt76/mt7915/debugfs.c
516
bool tx, rx, en;
drivers/net/wireless/mediatek/mt76/mt7915/debugfs.c
527
rx = dev->fw.debug_wm || (dev->fw.debug_bin & BIT(2));
drivers/net/wireless/mediatek/mt76/mt7915/debugfs.c
536
val = en && rx;
drivers/net/wireless/mediatek/mt76/mt7915/testmode.c
728
void *rx, *rssi;
drivers/net/wireless/mediatek/mt76/mt7915/testmode.c
733
rx = nla_nest_start(msg, MT76_TM_STATS_ATTR_LAST_RX);
drivers/net/wireless/mediatek/mt76/mt7915/testmode.c
734
if (!rx)
drivers/net/wireless/mediatek/mt76/mt7915/testmode.c
773
nla_nest_end(msg, rx);
drivers/net/wireless/mediatek/mt76/mt7921/mt7921.h
100
} rx;
drivers/net/wireless/mediatek/mt76/mt7921/pci.c
269
.rx = {
drivers/net/wireless/mediatek/mt76/mt7921/sdio.c
70
if (irq_data->rx.num[0] > 16 ||
drivers/net/wireless/mediatek/mt76/mt7921/sdio.c
71
irq_data->rx.num[1] > 128)
drivers/net/wireless/mediatek/mt76/mt7921/sdio.c
77
intr->rx.num = irq_data->rx.num;
drivers/net/wireless/mediatek/mt76/mt7921/sdio.c
80
intr->rx.len[0] = irq_data->rx.len0;
drivers/net/wireless/mediatek/mt76/mt7921/sdio.c
82
intr->rx.len[1] = irq_data->rx.len1;
drivers/net/wireless/mediatek/mt76/mt7925/pci.c
304
.rx = {
drivers/net/wireless/mediatek/mt76/mt7925/pci_mac.c
90
if (irq_map->rx.data_complete_mask)
drivers/net/wireless/mediatek/mt76/mt7925/pci_mac.c
92
if (irq_map->rx.wm_complete_mask)
drivers/net/wireless/mediatek/mt76/mt7925/pci_mac.c
94
if (irq_map->rx.wm2_complete_mask)
drivers/net/wireless/mediatek/mt76/mt792x.h
201
} rx;
drivers/net/wireless/mediatek/mt76/mt792x_dma.c
163
dev->irq_map->rx.data_complete_mask |
drivers/net/wireless/mediatek/mt76/mt792x_dma.c
164
dev->irq_map->rx.wm2_complete_mask |
drivers/net/wireless/mediatek/mt76/mt792x_dma.c
165
dev->irq_map->rx.wm_complete_mask |
drivers/net/wireless/mediatek/mt76/mt792x_dma.c
42
mask |= intr & (irq_map->rx.data_complete_mask |
drivers/net/wireless/mediatek/mt76/mt792x_dma.c
43
irq_map->rx.wm_complete_mask |
drivers/net/wireless/mediatek/mt76/mt792x_dma.c
44
irq_map->rx.wm2_complete_mask);
drivers/net/wireless/mediatek/mt76/mt792x_dma.c
55
mask |= irq_map->rx.data_complete_mask;
drivers/net/wireless/mediatek/mt76/mt792x_dma.c
56
intr |= irq_map->rx.data_complete_mask;
drivers/net/wireless/mediatek/mt76/mt792x_dma.c
65
if (intr & irq_map->rx.wm_complete_mask)
drivers/net/wireless/mediatek/mt76/mt792x_dma.c
68
if (intr & irq_map->rx.wm2_complete_mask)
drivers/net/wireless/mediatek/mt76/mt792x_dma.c
71
if (intr & irq_map->rx.data_complete_mask)
drivers/net/wireless/mediatek/mt76/mt792x_dma.c
82
mt76_connac_irq_enable(mdev, irq_map->rx.data_complete_mask);
drivers/net/wireless/mediatek/mt76/mt792x_dma.c
84
mt76_connac_irq_enable(mdev, irq_map->rx.wm2_complete_mask);
drivers/net/wireless/mediatek/mt76/mt792x_dma.c
86
mt76_connac_irq_enable(mdev, irq_map->rx.wm_complete_mask);
drivers/net/wireless/mediatek/mt76/mt7996/debugfs.c
310
bool tx, rx, en;
drivers/net/wireless/mediatek/mt76/mt7996/debugfs.c
321
rx = dev->fw_debug_wm || (dev->fw_debug_bin & BIT(2));
drivers/net/wireless/mediatek/mt76/mt7996/debugfs.c
333
val = en && rx;
drivers/net/wireless/mediatek/mt76/sdio.h
137
} rx;
drivers/net/wireless/mediatek/mt76/sdio_txrx.c
123
while (i < intr->rx.num[qid] && buf < end) {
drivers/net/wireless/mediatek/mt76/sdio_txrx.c
94
for (i = 0; i < intr->rx.num[qid]; i++)
drivers/net/wireless/mediatek/mt76/sdio_txrx.c
95
len += round_up(intr->rx.len[qid][i] + 4, 4);
drivers/net/wireless/microchip/wilc1000/cfg80211.c
25
.rx = BIT(IEEE80211_STYPE_ACTION >> 4) |
drivers/net/wireless/microchip/wilc1000/cfg80211.c
31
.rx = BIT(IEEE80211_STYPE_ASSOC_REQ >> 4) |
drivers/net/wireless/microchip/wilc1000/cfg80211.c
41
.rx = BIT(IEEE80211_STYPE_ACTION >> 4) |
drivers/net/wireless/purelifi/plfxlc/usb.c
202
struct plfxlc_usb_rx *rx = &usb->rx;
drivers/net/wireless/purelifi/plfxlc/usb.c
217
spin_lock_irq(&rx->lock);
drivers/net/wireless/purelifi/plfxlc/usb.c
221
if (rx->urbs) {
drivers/net/wireless/purelifi/plfxlc/usb.c
222
spin_unlock_irq(&rx->lock);
drivers/net/wireless/purelifi/plfxlc/usb.c
226
rx->urbs = urbs;
drivers/net/wireless/purelifi/plfxlc/usb.c
227
rx->urbs_count = RX_URBS_COUNT;
drivers/net/wireless/purelifi/plfxlc/usb.c
228
spin_unlock_irq(&rx->lock);
drivers/net/wireless/purelifi/plfxlc/usb.c
241
spin_lock_irq(&rx->lock);
drivers/net/wireless/purelifi/plfxlc/usb.c
242
rx->urbs = NULL;
drivers/net/wireless/purelifi/plfxlc/usb.c
243
rx->urbs_count = 0;
drivers/net/wireless/purelifi/plfxlc/usb.c
244
spin_unlock_irq(&rx->lock);
drivers/net/wireless/purelifi/plfxlc/usb.c
256
struct plfxlc_usb_rx *rx = &usb->rx;
drivers/net/wireless/purelifi/plfxlc/usb.c
259
mutex_lock(&rx->setup_mutex);
drivers/net/wireless/purelifi/plfxlc/usb.c
264
mutex_unlock(&rx->setup_mutex);
drivers/net/wireless/purelifi/plfxlc/usb.c
271
struct plfxlc_usb_rx *rx = &usb->rx;
drivers/net/wireless/purelifi/plfxlc/usb.c
277
spin_lock_irqsave(&rx->lock, flags);
drivers/net/wireless/purelifi/plfxlc/usb.c
278
urbs = rx->urbs;
drivers/net/wireless/purelifi/plfxlc/usb.c
279
count = rx->urbs_count;
drivers/net/wireless/purelifi/plfxlc/usb.c
280
spin_unlock_irqrestore(&rx->lock, flags);
drivers/net/wireless/purelifi/plfxlc/usb.c
290
rx->urbs = NULL;
drivers/net/wireless/purelifi/plfxlc/usb.c
291
rx->urbs_count = 0;
drivers/net/wireless/purelifi/plfxlc/usb.c
296
struct plfxlc_usb_rx *rx = &usb->rx;
drivers/net/wireless/purelifi/plfxlc/usb.c
298
mutex_lock(&rx->setup_mutex);
drivers/net/wireless/purelifi/plfxlc/usb.c
301
mutex_unlock(&rx->setup_mutex);
drivers/net/wireless/purelifi/plfxlc/usb.c
374
struct plfxlc_usb_rx *rx = &usb->rx;
drivers/net/wireless/purelifi/plfxlc/usb.c
376
spin_lock_init(&rx->lock);
drivers/net/wireless/purelifi/plfxlc/usb.c
377
mutex_init(&rx->setup_mutex);
drivers/net/wireless/purelifi/plfxlc/usb.c
380
rx->usb_packet_size = 512;
drivers/net/wireless/purelifi/plfxlc/usb.c
382
rx->usb_packet_size = 64;
drivers/net/wireless/purelifi/plfxlc/usb.c
384
if (rx->fragment_length != 0)
drivers/net/wireless/purelifi/plfxlc/usb.h
128
struct plfxlc_usb_rx rx;
drivers/net/wireless/quantenna/qtnfmac/cfg80211.c
58
.rx = BIT(IEEE80211_STYPE_ACTION >> 4) |
drivers/net/wireless/quantenna/qtnfmac/cfg80211.c
65
.rx = BIT(IEEE80211_STYPE_ACTION >> 4) |
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
1496
rt2x00dev->default_ant.rx =
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
1507
if (rt2x00dev->default_ant.rx == ANTENNA_SW_DIVERSITY)
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
1508
rt2x00dev->default_ant.rx = ANTENNA_HW_DIVERSITY;
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
400
BUG_ON(ant->rx == ANTENNA_SW_DIVERSITY ||
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
425
switch (ant->rx) {
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
797
rt2x00_set_field32(®, RXCSR1_RXD_SIZE, rt2x00dev->rx->desc_size);
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
798
rt2x00_set_field32(®, RXCSR1_NUM_RXD, rt2x00dev->rx->limit);
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
801
entry_priv = rt2x00dev->rx->entries[0].priv_data;
drivers/net/wireless/ralink/rt2x00/rt2400pci.c
827
(rt2x00dev->rx->data_size / 128));
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
1655
rt2x00dev->default_ant.rx =
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
407
BUG_ON(ant->rx == ANTENNA_SW_DIVERSITY ||
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
434
switch (ant->rx) {
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
882
rt2x00_set_field32(®, RXCSR1_RXD_SIZE, rt2x00dev->rx->desc_size);
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
883
rt2x00_set_field32(®, RXCSR1_NUM_RXD, rt2x00dev->rx->limit);
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
886
entry_priv = rt2x00dev->rx->entries[0].priv_data;
drivers/net/wireless/ralink/rt2x00/rt2500pci.c
912
rt2x00dev->rx->data_size / 128);
drivers/net/wireless/ralink/rt2x00/rt2500usb.c
1463
rt2x00dev->default_ant.rx =
drivers/net/wireless/ralink/rt2x00/rt2500usb.c
1474
if (rt2x00dev->default_ant.rx == ANTENNA_SW_DIVERSITY)
drivers/net/wireless/ralink/rt2x00/rt2500usb.c
1475
rt2x00dev->default_ant.rx = ANTENNA_HW_DIVERSITY;
drivers/net/wireless/ralink/rt2x00/rt2500usb.c
512
BUG_ON(ant->rx == ANTENNA_SW_DIVERSITY ||
drivers/net/wireless/ralink/rt2x00/rt2500usb.c
545
switch (ant->rx) {
drivers/net/wireless/ralink/rt2x00/rt2500usb.c
858
rt2x00dev->rx->data_size);
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
11258
rt2x00dev->default_ant.rx = ANTENNA_A;
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
11262
rt2x00dev->default_ant.rx = ANTENNA_B;
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
11267
rt2x00dev->default_ant.rx = ANTENNA_A;
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
11274
rt2x00dev->default_ant.rx = ANTENNA_HW_DIVERSITY; /* Unused */
drivers/net/wireless/ralink/rt2x00/rt2800lib.c
2368
rt2x00dev->default_ant.rx);
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
710
entry_priv = rt2x00dev->rx->entries[0].priv_data;
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
714
rt2x00dev->rx[0].limit);
drivers/net/wireless/ralink/rt2x00/rt2800mmio.c
716
rt2x00dev->rx[0].limit - 1);
drivers/net/wireless/ralink/rt2x00/rt2800usb.c
311
((rt2x00dev->rx->limit * DATA_FRAME_SIZE)
drivers/net/wireless/ralink/rt2x00/rt2x00.h
1497
u32 *tx, u32 *tx_max, u32 *rx, u32 *rx_max);
drivers/net/wireless/ralink/rt2x00/rt2x00.h
210
enum antenna rx;
drivers/net/wireless/ralink/rt2x00/rt2x00.h
964
struct data_queue *rx;
drivers/net/wireless/ralink/rt2x00/rt2x00config.c
120
if (config.rx == ANTENNA_SW_DIVERSITY) {
drivers/net/wireless/ralink/rt2x00/rt2x00config.c
123
if (def->rx == ANTENNA_SW_DIVERSITY)
drivers/net/wireless/ralink/rt2x00/rt2x00config.c
124
config.rx = ANTENNA_B;
drivers/net/wireless/ralink/rt2x00/rt2x00config.c
126
config.rx = def->rx;
drivers/net/wireless/ralink/rt2x00/rt2x00config.c
128
} else if (config.rx == ANTENNA_SW_DIVERSITY)
drivers/net/wireless/ralink/rt2x00/rt2x00config.c
129
config.rx = active->rx;
drivers/net/wireless/ralink/rt2x00/rt2x00config.c
148
rt2x00queue_stop_queue(rt2x00dev->rx);
drivers/net/wireless/ralink/rt2x00/rt2x00config.c
162
rt2x00queue_start_queue(rt2x00dev->rx);
drivers/net/wireless/ralink/rt2x00/rt2x00dev.c
857
rx_status->antenna = rt2x00dev->link.ant.active.rx;
drivers/net/wireless/ralink/rt2x00/rt2x00link.c
100
new_ant.rx = other_antenna;
drivers/net/wireless/ralink/rt2x00/rt2x00link.c
140
new_ant.rx = (new_ant.rx == ANTENNA_A) ? ANTENNA_B : ANTENNA_A;
drivers/net/wireless/ralink/rt2x00/rt2x00link.c
97
other_antenna = (ant->active.rx == ANTENNA_A) ? ANTENNA_B : ANTENNA_A;
drivers/net/wireless/ralink/rt2x00/rt2x00mac.c
326
rt2x00queue_stop_queue(rt2x00dev->rx);
drivers/net/wireless/ralink/rt2x00/rt2x00mac.c
349
rt2x00queue_start_queue(rt2x00dev->rx);
drivers/net/wireless/ralink/rt2x00/rt2x00mac.c
773
if (rx_ant == 3 && def->rx == ANTENNA_SW_DIVERSITY) {
drivers/net/wireless/ralink/rt2x00/rt2x00mac.c
779
setup.rx = rx_ant;
drivers/net/wireless/ralink/rt2x00/rt2x00mac.c
806
*rx_ant = active->rx;
drivers/net/wireless/ralink/rt2x00/rt2x00mac.c
813
u32 *tx, u32 *tx_max, u32 *rx, u32 *rx_max)
drivers/net/wireless/ralink/rt2x00/rt2x00mac.c
823
*rx = rt2x00dev->rx->length;
drivers/net/wireless/ralink/rt2x00/rt2x00mac.c
824
*rx_max = rt2x00dev->rx->limit;
drivers/net/wireless/ralink/rt2x00/rt2x00mmio.c
51
struct data_queue *queue = rt2x00dev->rx;
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1034
rt2x00queue_start_queue(rt2x00dev->rx);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1053
rt2x00queue_stop_queue(rt2x00dev->rx);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1064
rt2x00queue_flush_queue(rt2x00dev->rx, drop);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1166
status = rt2x00queue_alloc_entries(rt2x00dev->rx);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1186
status = rt2x00queue_alloc_rxskbs(rt2x00dev->rx);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1204
rt2x00queue_free_skbs(rt2x00dev->rx);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1254
rt2x00dev->rx = queue;
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1268
rt2x00queue_init(rt2x00dev, rt2x00dev->rx, QID_RX);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1283
kfree(rt2x00dev->rx);
drivers/net/wireless/ralink/rt2x00/rt2x00queue.c
1284
rt2x00dev->rx = NULL;
drivers/net/wireless/ralink/rt2x00/rt2x00queue.h
494
&(__dev)->rx[(__dev)->data_queues]
drivers/net/wireless/ralink/rt2x00/rt2x00queue.h
541
queue_loop(__entry, (__dev)->rx, queue_end(__dev))
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
352
while (!rt2x00queue_empty(rt2x00dev->rx)) {
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
353
entry = rt2x00queue_get_entry(rt2x00dev->rx, Q_INDEX_DONE);
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
618
rt2x00usb_assign_endpoint(rt2x00dev->rx, ep_desc);
drivers/net/wireless/ralink/rt2x00/rt2x00usb.c
631
if (!rt2x00dev->rx->usb_endpoint || !rt2x00dev->tx->usb_endpoint) {
drivers/net/wireless/ralink/rt2x00/rt61pci.c
1376
rt2x00_set_field32(®, RX_RING_CSR_RING_SIZE, rt2x00dev->rx->limit);
drivers/net/wireless/ralink/rt2x00/rt61pci.c
1378
rt2x00dev->rx->desc_size / 4);
drivers/net/wireless/ralink/rt2x00/rt61pci.c
1382
entry_priv = rt2x00dev->rx->entries[0].priv_data;
drivers/net/wireless/ralink/rt2x00/rt61pci.c
2445
rt2x00dev->default_ant.rx =
drivers/net/wireless/ralink/rt2x00/rt61pci.c
2486
rt2x00dev->default_ant.rx =
drivers/net/wireless/ralink/rt2x00/rt61pci.c
2494
rt2x00dev->default_ant.rx = ANTENNA_SW_DIVERSITY;
drivers/net/wireless/ralink/rt2x00/rt61pci.c
549
switch (ant->rx) {
drivers/net/wireless/ralink/rt2x00/rt61pci.c
597
switch (ant->rx) {
drivers/net/wireless/ralink/rt2x00/rt61pci.c
647
switch (ant->rx) {
drivers/net/wireless/ralink/rt2x00/rt61pci.c
715
BUG_ON(ant->rx == ANTENNA_SW_DIVERSITY ||
drivers/net/wireless/ralink/rt2x00/rt73usb.c
1881
rt2x00dev->default_ant.rx =
drivers/net/wireless/ralink/rt2x00/rt73usb.c
584
switch (ant->rx) {
drivers/net/wireless/ralink/rt2x00/rt73usb.c
633
switch (ant->rx) {
drivers/net/wireless/ralink/rt2x00/rt73usb.c
696
BUG_ON(ant->rx == ANTENNA_SW_DIVERSITY ||
drivers/net/wireless/realtek/rtw88/pci.c
1166
bool rx = false;
drivers/net/wireless/realtek/rtw88/pci.c
1187
rx = true;
drivers/net/wireless/realtek/rtw88/pci.c
1194
rtw_pci_enable_interrupt(rtwdev, rtwpci, rx);
drivers/net/wireless/realtek/rtw89/coex.c
4451
plt.rx = rx_val;
drivers/net/wireless/realtek/rtw89/mac.c
6458
(plt->rx & RTW89_MAC_AX_PLT_LTE_RX ? B_AX_RX_PLT_GNT_LTE_RX : 0) |
drivers/net/wireless/realtek/rtw89/mac.c
6459
(plt->rx & RTW89_MAC_AX_PLT_GNT_BT_TX ? B_AX_RX_PLT_GNT_BT_TX : 0) |
drivers/net/wireless/realtek/rtw89/mac.c
6460
(plt->rx & RTW89_MAC_AX_PLT_GNT_BT_RX ? B_AX_RX_PLT_GNT_BT_RX : 0) |
drivers/net/wireless/realtek/rtw89/mac.c
6461
(plt->rx & RTW89_MAC_AX_PLT_GNT_WL ? B_AX_RX_PLT_GNT_WL : 0) |
drivers/net/wireless/realtek/rtw89/mac.h
545
u8 rx;
drivers/net/wireless/realtek/rtw89/mac_be.c
2443
(plt->rx & RTW89_MAC_AX_PLT_LTE_RX ? B_BE_RX_PLT_GNT_LTE_RX : 0) |
drivers/net/wireless/realtek/rtw89/mac_be.c
2444
(plt->rx & RTW89_MAC_AX_PLT_GNT_BT_TX ? B_BE_RX_PLT_GNT_BT_TX : 0) |
drivers/net/wireless/realtek/rtw89/mac_be.c
2445
(plt->rx & RTW89_MAC_AX_PLT_GNT_BT_RX ? B_BE_RX_PLT_GNT_BT_RX : 0) |
drivers/net/wireless/realtek/rtw89/mac_be.c
2446
(plt->rx & RTW89_MAC_AX_PLT_GNT_WL ? B_BE_RX_PLT_GNT_WL : 0) |
drivers/net/wireless/realtek/rtw89/pci.c
1109
.rx = {
drivers/net/wireless/realtek/rtw89/pci.c
1132
.rx = {
drivers/net/wireless/realtek/rtw89/pci.c
1155
.rx = {
drivers/net/wireless/realtek/rtw89/pci.c
1178
.rx = {
drivers/net/wireless/realtek/rtw89/pci.c
1217
*addr = &info->dma_addr_set->rx[rxch];
drivers/net/wireless/realtek/rtw89/pci.c
1278
rx_ring = &rtwpci->rx.rings[RTW89_RXCH_RPQ];
drivers/net/wireless/realtek/rtw89/pci.c
1833
rx_ring = &rtwpci->rx.rings[i];
drivers/net/wireless/realtek/rtw89/pci.c
1978
rx_ring = &rtwpci->rx.rings[i];
drivers/net/wireless/realtek/rtw89/pci.c
1981
dma_addr_set->rx[i].idx;
drivers/net/wireless/realtek/rtw89/pci.c
3475
struct rtw89_pci_dma_pool *bd_pool = &rtwpci->rx.bd_pool;
drivers/net/wireless/realtek/rtw89/pci.c
3480
rx_ring = &rtwpci->rx.rings[i];
drivers/net/wireless/realtek/rtw89/pci.c
3750
struct rtw89_pci_dma_pool *bd_pool = &rtwpci->rx.bd_pool;
drivers/net/wireless/realtek/rtw89/pci.c
3775
rx_ring = &rtwpci->rx.rings[i];
drivers/net/wireless/realtek/rtw89/pci.c
3794
rx_ring = &rtwpci->rx.rings[i];
drivers/net/wireless/realtek/rtw89/pci.c
443
rx_ring = &rtwpci->rx.rings[RTW89_RXCH_RXQ];
drivers/net/wireless/realtek/rtw89/pci.c
727
rx_ring = &rtwpci->rx.rings[RTW89_RXCH_RPQ];
drivers/net/wireless/realtek/rtw89/pci.c
757
rx_ring = &rtwpci->rx.rings[i];
drivers/net/wireless/realtek/rtw89/pci.h
1321
struct rtw89_pci_ch_dma_addr rx[RTW89_RXCH_NUM];
drivers/net/wireless/realtek/rtw89/pci.h
1627
struct rtw89_pci_rx_rings rx;
drivers/net/wireless/realtek/rtw89/pci_be.c
187
rx_ring = &rtwpci->rx.rings[RTW89_RXCH_RXQ];
drivers/net/wireless/realtek/rtw89/pci_be.c
190
rx_ring = &rtwpci->rx.rings[RTW89_RXCH_RPQ];
drivers/net/wireless/st/cw1200/bh.c
415
int rx, tx, term, suspend;
drivers/net/wireless/st/cw1200/bh.c
449
rx = atomic_xchg(&priv->bh_rx, 0);
drivers/net/wireless/st/cw1200/bh.c
454
(rx || tx || term || suspend || priv->bh_error);
drivers/net/wireless/st/cw1200/bh.c
458
rx, tx, term, suspend, priv->bh_error, status);
drivers/net/wireless/st/cw1200/bh.c
472
if (priv->hw_bufs_used && (!rx || !tx)) {
drivers/net/wireless/st/cw1200/bh.c
476
rx = 1;
drivers/net/wireless/st/cw1200/bh.c
536
rx:
drivers/net/wireless/st/cw1200/bh.c
586
goto rx;
drivers/net/wireless/st/cw1200/debug.c
272
d->rx);
drivers/net/wireless/st/cw1200/debug.h
16
int rx;
drivers/net/wireless/st/cw1200/debug.h
52
++priv->debug->rx;
drivers/net/wireless/st/cw1200/wsm.c
865
struct wsm_rx rx;
drivers/net/wireless/st/cw1200/wsm.c
870
rx.status = WSM_GET32(buf);
drivers/net/wireless/st/cw1200/wsm.c
871
rx.channel_number = WSM_GET16(buf);
drivers/net/wireless/st/cw1200/wsm.c
872
rx.rx_rate = WSM_GET8(buf);
drivers/net/wireless/st/cw1200/wsm.c
873
rx.rcpi_rssi = WSM_GET8(buf);
drivers/net/wireless/st/cw1200/wsm.c
874
rx.flags = WSM_GET32(buf);
drivers/net/wireless/st/cw1200/wsm.c
881
if (!rx.rcpi_rssi &&
drivers/net/wireless/st/cw1200/wsm.c
890
rx.rcpi_rssi = rx.rcpi_rssi / 2 - 110;
drivers/net/wireless/st/cw1200/wsm.c
895
if (!rx.status && ieee80211_is_deauth(fctl)) {
drivers/net/wireless/st/cw1200/wsm.c
905
cw1200_rx_cb(priv, &rx, link_id, skb_p);
drivers/net/wireless/ti/wl1251/acx.h
1064
struct acx_rx_statistics rx;
drivers/net/wireless/ti/wl1251/debugfs.c
103
DEBUGFS_FWSTATS_FILE(rx, out_of_mem, 20, "%u");
drivers/net/wireless/ti/wl1251/debugfs.c
104
DEBUGFS_FWSTATS_FILE(rx, hdr_overflow, 20, "%u");
drivers/net/wireless/ti/wl1251/debugfs.c
105
DEBUGFS_FWSTATS_FILE(rx, hw_stuck, 20, "%u");
drivers/net/wireless/ti/wl1251/debugfs.c
106
DEBUGFS_FWSTATS_FILE(rx, dropped, 20, "%u");
drivers/net/wireless/ti/wl1251/debugfs.c
107
DEBUGFS_FWSTATS_FILE(rx, fcs_err, 20, "%u");
drivers/net/wireless/ti/wl1251/debugfs.c
108
DEBUGFS_FWSTATS_FILE(rx, xfr_hint_trig, 20, "%u");
drivers/net/wireless/ti/wl1251/debugfs.c
109
DEBUGFS_FWSTATS_FILE(rx, path_reset, 20, "%u");
drivers/net/wireless/ti/wl1251/debugfs.c
110
DEBUGFS_FWSTATS_FILE(rx, reset_counter, 20, "%u");
drivers/net/wireless/ti/wl1251/debugfs.c
243
DEBUGFS_FWSTATS_DEL(rx, out_of_mem);
drivers/net/wireless/ti/wl1251/debugfs.c
244
DEBUGFS_FWSTATS_DEL(rx, hdr_overflow);
drivers/net/wireless/ti/wl1251/debugfs.c
245
DEBUGFS_FWSTATS_DEL(rx, hw_stuck);
drivers/net/wireless/ti/wl1251/debugfs.c
246
DEBUGFS_FWSTATS_DEL(rx, dropped);
drivers/net/wireless/ti/wl1251/debugfs.c
247
DEBUGFS_FWSTATS_DEL(rx, fcs_err);
drivers/net/wireless/ti/wl1251/debugfs.c
248
DEBUGFS_FWSTATS_DEL(rx, xfr_hint_trig);
drivers/net/wireless/ti/wl1251/debugfs.c
249
DEBUGFS_FWSTATS_DEL(rx, path_reset);
drivers/net/wireless/ti/wl1251/debugfs.c
250
DEBUGFS_FWSTATS_DEL(rx, reset_counter);
drivers/net/wireless/ti/wl1251/debugfs.c
342
DEBUGFS_FWSTATS_ADD(rx, out_of_mem);
drivers/net/wireless/ti/wl1251/debugfs.c
343
DEBUGFS_FWSTATS_ADD(rx, hdr_overflow);
drivers/net/wireless/ti/wl1251/debugfs.c
344
DEBUGFS_FWSTATS_ADD(rx, hw_stuck);
drivers/net/wireless/ti/wl1251/debugfs.c
345
DEBUGFS_FWSTATS_ADD(rx, dropped);
drivers/net/wireless/ti/wl1251/debugfs.c
346
DEBUGFS_FWSTATS_ADD(rx, fcs_err);
drivers/net/wireless/ti/wl1251/debugfs.c
347
DEBUGFS_FWSTATS_ADD(rx, xfr_hint_trig);
drivers/net/wireless/ti/wl1251/debugfs.c
348
DEBUGFS_FWSTATS_ADD(rx, path_reset);
drivers/net/wireless/ti/wl1251/debugfs.c
349
DEBUGFS_FWSTATS_ADD(rx, reset_counter);
drivers/net/wireless/ti/wl12xx/acx.h
245
struct wl12xx_acx_rx_statistics rx;
drivers/net/wireless/ti/wl12xx/debugfs.c
121
DEBUGFS_FWSTATS_ADD(rx, out_of_mem);
drivers/net/wireless/ti/wl12xx/debugfs.c
122
DEBUGFS_FWSTATS_ADD(rx, hdr_overflow);
drivers/net/wireless/ti/wl12xx/debugfs.c
123
DEBUGFS_FWSTATS_ADD(rx, hw_stuck);
drivers/net/wireless/ti/wl12xx/debugfs.c
124
DEBUGFS_FWSTATS_ADD(rx, dropped);
drivers/net/wireless/ti/wl12xx/debugfs.c
125
DEBUGFS_FWSTATS_ADD(rx, fcs_err);
drivers/net/wireless/ti/wl12xx/debugfs.c
126
DEBUGFS_FWSTATS_ADD(rx, xfr_hint_trig);
drivers/net/wireless/ti/wl12xx/debugfs.c
127
DEBUGFS_FWSTATS_ADD(rx, path_reset);
drivers/net/wireless/ti/wl12xx/debugfs.c
128
DEBUGFS_FWSTATS_ADD(rx, reset_counter);
drivers/net/wireless/ti/wl12xx/debugfs.c
21
WL12XX_DEBUGFS_FWSTATS_FILE(rx, out_of_mem, "%u");
drivers/net/wireless/ti/wl12xx/debugfs.c
22
WL12XX_DEBUGFS_FWSTATS_FILE(rx, hdr_overflow, "%u");
drivers/net/wireless/ti/wl12xx/debugfs.c
23
WL12XX_DEBUGFS_FWSTATS_FILE(rx, hw_stuck, "%u");
drivers/net/wireless/ti/wl12xx/debugfs.c
24
WL12XX_DEBUGFS_FWSTATS_FILE(rx, dropped, "%u");
drivers/net/wireless/ti/wl12xx/debugfs.c
25
WL12XX_DEBUGFS_FWSTATS_FILE(rx, fcs_err, "%u");
drivers/net/wireless/ti/wl12xx/debugfs.c
26
WL12XX_DEBUGFS_FWSTATS_FILE(rx, xfr_hint_trig, "%u");
drivers/net/wireless/ti/wl12xx/debugfs.c
27
WL12XX_DEBUGFS_FWSTATS_FILE(rx, path_reset, "%u");
drivers/net/wireless/ti/wl12xx/debugfs.c
28
WL12XX_DEBUGFS_FWSTATS_FILE(rx, reset_counter, "%u");
drivers/net/wireless/ti/wl12xx/main.c
103
.rx = {
drivers/net/wireless/ti/wl18xx/acx.h
271
struct wl18xx_acx_rx_stats rx;
drivers/net/wireless/ti/wl18xx/debugfs.c
100
WL18XX_DEBUGFS_FWSTATS_FILE(rx, defrag_tkip_called, "%u");
drivers/net/wireless/ti/wl18xx/debugfs.c
101
WL18XX_DEBUGFS_FWSTATS_FILE(rx, defrag_need_defrag, "%u");
drivers/net/wireless/ti/wl18xx/debugfs.c
102
WL18XX_DEBUGFS_FWSTATS_FILE(rx, defrag_decrypt_failed, "%u");
drivers/net/wireless/ti/wl18xx/debugfs.c
103
WL18XX_DEBUGFS_FWSTATS_FILE(rx, decrypt_key_not_found, "%u");
drivers/net/wireless/ti/wl18xx/debugfs.c
104
WL18XX_DEBUGFS_FWSTATS_FILE(rx, defrag_need_decrypt, "%u");
drivers/net/wireless/ti/wl18xx/debugfs.c
105
WL18XX_DEBUGFS_FWSTATS_FILE(rx, rx_tkip_replays, "%u");
drivers/net/wireless/ti/wl18xx/debugfs.c
106
WL18XX_DEBUGFS_FWSTATS_FILE(rx, rx_xfr, "%u");
drivers/net/wireless/ti/wl18xx/debugfs.c
464
DEBUGFS_FWSTATS_ADD(rx, rx_beacon_early_term);
drivers/net/wireless/ti/wl18xx/debugfs.c
465
DEBUGFS_FWSTATS_ADD(rx, rx_out_of_mpdu_nodes);
drivers/net/wireless/ti/wl18xx/debugfs.c
466
DEBUGFS_FWSTATS_ADD(rx, rx_hdr_overflow);
drivers/net/wireless/ti/wl18xx/debugfs.c
467
DEBUGFS_FWSTATS_ADD(rx, rx_dropped_frame);
drivers/net/wireless/ti/wl18xx/debugfs.c
468
DEBUGFS_FWSTATS_ADD(rx, rx_done);
drivers/net/wireless/ti/wl18xx/debugfs.c
469
DEBUGFS_FWSTATS_ADD(rx, rx_defrag);
drivers/net/wireless/ti/wl18xx/debugfs.c
470
DEBUGFS_FWSTATS_ADD(rx, rx_defrag_end);
drivers/net/wireless/ti/wl18xx/debugfs.c
471
DEBUGFS_FWSTATS_ADD(rx, rx_cmplt);
drivers/net/wireless/ti/wl18xx/debugfs.c
472
DEBUGFS_FWSTATS_ADD(rx, rx_pre_complt);
drivers/net/wireless/ti/wl18xx/debugfs.c
473
DEBUGFS_FWSTATS_ADD(rx, rx_cmplt_task);
drivers/net/wireless/ti/wl18xx/debugfs.c
474
DEBUGFS_FWSTATS_ADD(rx, rx_phy_hdr);
drivers/net/wireless/ti/wl18xx/debugfs.c
475
DEBUGFS_FWSTATS_ADD(rx, rx_timeout);
drivers/net/wireless/ti/wl18xx/debugfs.c
476
DEBUGFS_FWSTATS_ADD(rx, rx_rts_timeout);
drivers/net/wireless/ti/wl18xx/debugfs.c
477
DEBUGFS_FWSTATS_ADD(rx, rx_timeout_wa);
drivers/net/wireless/ti/wl18xx/debugfs.c
478
DEBUGFS_FWSTATS_ADD(rx, defrag_called);
drivers/net/wireless/ti/wl18xx/debugfs.c
479
DEBUGFS_FWSTATS_ADD(rx, defrag_init_called);
drivers/net/wireless/ti/wl18xx/debugfs.c
480
DEBUGFS_FWSTATS_ADD(rx, defrag_in_process_called);
drivers/net/wireless/ti/wl18xx/debugfs.c
481
DEBUGFS_FWSTATS_ADD(rx, defrag_tkip_called);
drivers/net/wireless/ti/wl18xx/debugfs.c
482
DEBUGFS_FWSTATS_ADD(rx, defrag_need_defrag);
drivers/net/wireless/ti/wl18xx/debugfs.c
483
DEBUGFS_FWSTATS_ADD(rx, defrag_decrypt_failed);
drivers/net/wireless/ti/wl18xx/debugfs.c
484
DEBUGFS_FWSTATS_ADD(rx, decrypt_key_not_found);
drivers/net/wireless/ti/wl18xx/debugfs.c
485
DEBUGFS_FWSTATS_ADD(rx, defrag_need_decrypt);
drivers/net/wireless/ti/wl18xx/debugfs.c
486
DEBUGFS_FWSTATS_ADD(rx, rx_tkip_replays);
drivers/net/wireless/ti/wl18xx/debugfs.c
487
DEBUGFS_FWSTATS_ADD(rx, rx_xfr);
drivers/net/wireless/ti/wl18xx/debugfs.c
83
WL18XX_DEBUGFS_FWSTATS_FILE(rx, rx_beacon_early_term, "%u");
drivers/net/wireless/ti/wl18xx/debugfs.c
84
WL18XX_DEBUGFS_FWSTATS_FILE(rx, rx_out_of_mpdu_nodes, "%u");
drivers/net/wireless/ti/wl18xx/debugfs.c
85
WL18XX_DEBUGFS_FWSTATS_FILE(rx, rx_hdr_overflow, "%u");
drivers/net/wireless/ti/wl18xx/debugfs.c
86
WL18XX_DEBUGFS_FWSTATS_FILE(rx, rx_dropped_frame, "%u");
drivers/net/wireless/ti/wl18xx/debugfs.c
87
WL18XX_DEBUGFS_FWSTATS_FILE(rx, rx_done, "%u");
drivers/net/wireless/ti/wl18xx/debugfs.c
88
WL18XX_DEBUGFS_FWSTATS_FILE(rx, rx_defrag, "%u");
drivers/net/wireless/ti/wl18xx/debugfs.c
89
WL18XX_DEBUGFS_FWSTATS_FILE(rx, rx_defrag_end, "%u");
drivers/net/wireless/ti/wl18xx/debugfs.c
90
WL18XX_DEBUGFS_FWSTATS_FILE(rx, rx_cmplt, "%u");
drivers/net/wireless/ti/wl18xx/debugfs.c
91
WL18XX_DEBUGFS_FWSTATS_FILE(rx, rx_pre_complt, "%u");
drivers/net/wireless/ti/wl18xx/debugfs.c
92
WL18XX_DEBUGFS_FWSTATS_FILE(rx, rx_cmplt_task, "%u");
drivers/net/wireless/ti/wl18xx/debugfs.c
93
WL18XX_DEBUGFS_FWSTATS_FILE(rx, rx_phy_hdr, "%u");
drivers/net/wireless/ti/wl18xx/debugfs.c
94
WL18XX_DEBUGFS_FWSTATS_FILE(rx, rx_timeout, "%u");
drivers/net/wireless/ti/wl18xx/debugfs.c
95
WL18XX_DEBUGFS_FWSTATS_FILE(rx, rx_rts_timeout, "%u");
drivers/net/wireless/ti/wl18xx/debugfs.c
96
WL18XX_DEBUGFS_FWSTATS_FILE(rx, rx_timeout_wa, "%u");
drivers/net/wireless/ti/wl18xx/debugfs.c
97
WL18XX_DEBUGFS_FWSTATS_FILE(rx, defrag_called, "%u");
drivers/net/wireless/ti/wl18xx/debugfs.c
98
WL18XX_DEBUGFS_FWSTATS_FILE(rx, defrag_init_called, "%u");
drivers/net/wireless/ti/wl18xx/debugfs.c
99
WL18XX_DEBUGFS_FWSTATS_FILE(rx, defrag_in_process_called, "%u");
drivers/net/wireless/ti/wl18xx/main.c
244
.rx = {
drivers/net/wireless/ti/wlcore/acx.c
1040
rx_conf->threshold = cpu_to_le16(wl->conf.rx.irq_pkt_threshold);
drivers/net/wireless/ti/wlcore/acx.c
1041
rx_conf->timeout = cpu_to_le16(wl->conf.rx.irq_timeout);
drivers/net/wireless/ti/wlcore/acx.c
1042
rx_conf->mblk_threshold = cpu_to_le16(wl->conf.rx.irq_blk_threshold);
drivers/net/wireless/ti/wlcore/acx.c
1043
rx_conf->queue_type = wl->conf.rx.queue_type;
drivers/net/wireless/ti/wlcore/acx.c
171
acx->lifetime = cpu_to_le32(wl->conf.rx.rx_msdu_life_time);
drivers/net/wireless/ti/wlcore/acx.c
260
rx_timeout->ps_poll_timeout = cpu_to_le16(wl->conf.rx.ps_poll_timeout);
drivers/net/wireless/ti/wlcore/acx.c
261
rx_timeout->upsd_timeout = cpu_to_le16(wl->conf.rx.upsd_timeout);
drivers/net/wireless/ti/wlcore/acx.c
287
rts_threshold = wl->conf.rx.rts_threshold;
drivers/net/wireless/ti/wlcore/acx.c
550
detection->rx_cca_threshold = cpu_to_le16(wl->conf.rx.rx_cca_threshold);
drivers/net/wireless/ti/wlcore/conf.h
1132
struct conf_rx_settings rx;
drivers/net/wireless/ti/wlcore/debugfs.c
166
WL12XX_CONF_DEBUGFS(irq_pkt_threshold, rx, 0, 65535,
drivers/net/wireless/ti/wlcore/debugfs.c
168
WL12XX_CONF_DEBUGFS(irq_blk_threshold, rx, 0, 65535,
drivers/net/wireless/ti/wlcore/debugfs.c
170
WL12XX_CONF_DEBUGFS(irq_timeout, rx, 0, 100,
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
1131
container_of(work, struct zd_usb, rx.idle_work.work);
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
1145
struct zd_usb *usb = from_tasklet(usb, t, rx.reset_timer_tasklet);
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
1152
struct zd_usb_rx *rx = &usb->rx;
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
1154
mod_delayed_work(zd_workqueue, &rx->idle_work, ZD_RX_IDLE_INTERVAL);
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
1170
struct zd_usb_rx *rx = &usb->rx;
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
1172
spin_lock_init(&rx->lock);
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
1173
mutex_init(&rx->setup_mutex);
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
1175
rx->usb_packet_size = 512;
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
1177
rx->usb_packet_size = 64;
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
1179
ZD_ASSERT(rx->fragment_length == 0);
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
1180
INIT_DELAYED_WORK(&rx->idle_work, zd_rx_idle_timer_handler);
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
1181
rx->reset_timer_tasklet.func = (void (*))
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
1183
rx->reset_timer_tasklet.data = (unsigned long)&rx->reset_timer_tasklet;
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
652
struct zd_usb_rx *rx;
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
676
rx = &usb->rx;
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
678
tasklet_schedule(&rx->reset_timer_tasklet);
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
680
if (length%rx->usb_packet_size > rx->usb_packet_size-4) {
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
683
ZD_ASSERT(length <= ARRAY_SIZE(rx->fragment));
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
684
spin_lock_irqsave(&rx->lock, flags);
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
685
memcpy(rx->fragment, buffer, length);
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
686
rx->fragment_length = length;
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
687
spin_unlock_irqrestore(&rx->lock, flags);
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
691
spin_lock_irqsave(&rx->lock, flags);
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
692
if (rx->fragment_length > 0) {
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
694
ZD_ASSERT(length + rx->fragment_length <=
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
695
ARRAY_SIZE(rx->fragment));
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
697
memcpy(rx->fragment+rx->fragment_length, buffer, length);
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
698
handle_rx_packet(usb, rx->fragment,
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
699
rx->fragment_length + length);
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
700
rx->fragment_length = 0;
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
701
spin_unlock_irqrestore(&rx->lock, flags);
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
703
spin_unlock_irqrestore(&rx->lock, flags);
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
749
struct zd_usb_rx *rx = &usb->rx;
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
765
spin_lock_irq(&rx->lock);
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
766
if (rx->urbs) {
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
767
spin_unlock_irq(&rx->lock);
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
771
rx->urbs = urbs;
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
772
rx->urbs_count = RX_URBS_COUNT;
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
773
spin_unlock_irq(&rx->lock);
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
786
spin_lock_irq(&rx->lock);
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
787
rx->urbs = NULL;
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
788
rx->urbs_count = 0;
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
789
spin_unlock_irq(&rx->lock);
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
802
struct zd_usb_rx *rx = &usb->rx;
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
804
mutex_lock(&rx->setup_mutex);
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
806
mutex_unlock(&rx->setup_mutex);
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
819
struct zd_usb_rx *rx = &usb->rx;
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
821
spin_lock_irqsave(&rx->lock, flags);
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
822
urbs = rx->urbs;
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
823
count = rx->urbs_count;
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
824
spin_unlock_irqrestore(&rx->lock, flags);
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
834
spin_lock_irqsave(&rx->lock, flags);
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
835
rx->urbs = NULL;
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
836
rx->urbs_count = 0;
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
837
spin_unlock_irqrestore(&rx->lock, flags);
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
842
struct zd_usb_rx *rx = &usb->rx;
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
844
mutex_lock(&rx->setup_mutex);
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
846
mutex_unlock(&rx->setup_mutex);
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
848
tasklet_kill(&rx->reset_timer_tasklet);
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
849
cancel_delayed_work_sync(&rx->idle_work);
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
855
struct zd_usb_rx *rx = &usb->rx;
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
858
mutex_lock(&rx->setup_mutex);
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
860
spin_lock_irqsave(&rx->lock, flags);
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
861
do_reset = rx->urbs != NULL;
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
862
spin_unlock_irqrestore(&rx->lock, flags);
drivers/net/wireless/zydas/zd1211rw/zd_usb.c
869
mutex_unlock(&rx->setup_mutex);
drivers/net/wireless/zydas/zd1211rw/zd_usb.h
208
struct zd_usb_rx rx;
drivers/net/wwan/qcom_bam_dmux.c
473
desc = dmaengine_prep_slave_single(dmux->rx, skb_dma->addr,
drivers/net/wwan/qcom_bam_dmux.c
611
dma_async_issue_pending(dmux->rx);
drivers/net/wwan/qcom_bam_dmux.c
623
dmux->rx = dma_request_chan(dev, "rx");
drivers/net/wwan/qcom_bam_dmux.c
624
if (IS_ERR(dmux->rx)) {
drivers/net/wwan/qcom_bam_dmux.c
625
dev_err(dev, "Failed to request RX DMA channel: %pe\n", dmux->rx);
drivers/net/wwan/qcom_bam_dmux.c
626
dmux->rx = NULL;
drivers/net/wwan/qcom_bam_dmux.c
629
dmaengine_slave_config(dmux->rx, &dma_rx_conf);
drivers/net/wwan/qcom_bam_dmux.c
635
dma_async_issue_pending(dmux->rx);
drivers/net/wwan/qcom_bam_dmux.c
665
if (dmux->rx) {
drivers/net/wwan/qcom_bam_dmux.c
666
dmaengine_terminate_sync(dmux->rx);
drivers/net/wwan/qcom_bam_dmux.c
667
dma_release_channel(dmux->rx);
drivers/net/wwan/qcom_bam_dmux.c
668
dmux->rx = NULL;
drivers/net/wwan/qcom_bam_dmux.c
746
if (!dmux->rx) {
drivers/net/wwan/qcom_bam_dmux.c
78
struct dma_chan *rx, *tx;
drivers/net/wwan/qcom_bam_dmux.c
876
if (!wait_event_timeout(dmux->pc_wait, !dmux->rx, BAM_DMUX_REMOTE_TIMEOUT))
drivers/net/xen-netback/common.h
197
struct xen_netif_rx_back_ring rx;
drivers/net/xen-netback/netback.c
1536
if (queue->rx.sring)
drivers/net/xen-netback/netback.c
1538
queue->rx.sring);
drivers/net/xen-netback/netback.c
1575
BACK_RING_ATTACH(&queue->rx, rxs, rsp_prod, XEN_PAGE_SIZE);
drivers/net/xen-netback/netback.c
1578
if (req_prod - rsp_prod > RING_SIZE(&queue->rx))
drivers/net/xen-netback/rx.c
177
rsp = RING_GET_RESPONSE(&queue->rx,
drivers/net/xen-netback/rx.c
186
RING_PUSH_RESPONSES_AND_CHECK_NOTIFY(&queue->rx, notify);
drivers/net/xen-netback/rx.c
226
queue->rx_copy.idx[queue->rx_copy.num] = queue->rx.req_cons;
drivers/net/xen-netback/rx.c
332
queue->rx.rsp_prod_pvt = queue->rx.req_cons;
drivers/net/xen-netback/rx.c
473
req = RING_GET_REQUEST(&queue->rx, queue->rx.req_cons);
drivers/net/xen-netback/rx.c
474
rsp = RING_GET_RESPONSE(&queue->rx, queue->rx.req_cons);
drivers/net/xen-netback/rx.c
482
queue->rx.req_cons++;
drivers/net/xen-netback/rx.c
514
prod = queue->rx.sring->req_prod;
drivers/net/xen-netback/rx.c
515
cons = queue->rx.req_cons;
drivers/net/xen-netback/rx.c
68
prod = queue->rx.sring->req_prod;
drivers/net/xen-netback/rx.c
69
cons = queue->rx.req_cons;
drivers/net/xen-netback/rx.c
74
queue->rx.sring->req_event = prod + 1;
drivers/net/xen-netback/rx.c
80
} while (queue->rx.sring->req_prod != prod);
drivers/net/xen-netback/xenbus.c
30
struct xen_netif_rx_back_ring *rx_ring = &queue->rx;
drivers/net/xen-netfront.c
1026
struct xen_netif_rx_response *rx = &rinfo->rx, rx_local;
drivers/net/xen-netfront.c
1027
int max = XEN_NETIF_NR_SLOTS_MIN + (rx->status <= RX_COPY_THRESHOLD);
drivers/net/xen-netfront.c
1028
RING_IDX cons = queue->rx.rsp_cons;
drivers/net/xen-netfront.c
1039
if (rx->flags & XEN_NETRXF_extra_info) {
drivers/net/xen-netfront.c
1046
rx->offset = xdp->u.xdp.headroom;
drivers/net/xen-netfront.c
1049
cons = queue->rx.rsp_cons;
drivers/net/xen-netfront.c
1061
rx->id);
drivers/net/xen-netfront.c
1066
if (unlikely(rx->status < 0 ||
drivers/net/xen-netfront.c
1067
rx->offset + rx->status > XEN_PAGE_SIZE)) {
drivers/net/xen-netfront.c
1070
rx->offset, rx->status);
drivers/net/xen-netfront.c
1089
if (!(rx->flags & XEN_NETRXF_more_data)) {
drivers/net/xen-netfront.c
1093
rx, xdp_prog, &xdp, need_xdp_flush);
drivers/net/xen-netfront.c
1106
if (!(rx->flags & XEN_NETRXF_more_data))
drivers/net/xen-netfront.c
1116
RING_COPY_RESPONSE(&queue->rx, cons + slots, &rx_local);
drivers/net/xen-netfront.c
1117
rx = &rx_local;
drivers/net/xen-netfront.c
1168
RING_IDX cons = queue->rx.rsp_cons;
drivers/net/xen-netfront.c
1172
struct xen_netif_rx_response rx;
drivers/net/xen-netfront.c
1175
RING_COPY_RESPONSE(&queue->rx, ++cons, &rx);
drivers/net/xen-netfront.c
1192
rx.offset, rx.status, PAGE_SIZE);
drivers/net/xen-netfront.c
1269
struct xen_netif_rx_response *rx = &rinfo.rx;
drivers/net/xen-netfront.c
1285
rp = queue->rx.sring->rsp_prod;
drivers/net/xen-netfront.c
1286
if (RING_RESPONSE_PROD_OVERFLOW(&queue->rx, rp)) {
drivers/net/xen-netfront.c
1288
rp - queue->rx.rsp_cons);
drivers/net/xen-netfront.c
1295
i = queue->rx.rsp_cons;
drivers/net/xen-netfront.c
1298
RING_COPY_RESPONSE(&queue->rx, i, rx);
drivers/net/xen-netfront.c
1313
i = queue->rx.rsp_cons;
drivers/net/xen-netfront.c
1326
queue->rx.rsp_cons +
drivers/net/xen-netfront.c
1332
NETFRONT_SKB_CB(skb)->pull_to = rx->status;
drivers/net/xen-netfront.c
1336
skb_frag_off_set(&skb_shinfo(skb)->frags[0], rx->offset);
drivers/net/xen-netfront.c
1337
skb_frag_size_set(&skb_shinfo(skb)->frags[0], rx->status);
drivers/net/xen-netfront.c
1338
skb->data_len = rx->status;
drivers/net/xen-netfront.c
1339
skb->len += rx->status;
drivers/net/xen-netfront.c
1344
if (rx->flags & XEN_NETRXF_csum_blank)
drivers/net/xen-netfront.c
1346
else if (rx->flags & XEN_NETRXF_data_validated)
drivers/net/xen-netfront.c
1351
i = queue->rx.rsp_cons + 1;
drivers/net/xen-netfront.c
1369
RING_FINAL_CHECK_FOR_RESPONSES(&queue->rx, more_to_do);
drivers/net/xen-netfront.c
144
struct xen_netif_rx_front_ring rx;
drivers/net/xen-netfront.c
1548
work_queued = XEN_RING_NR_UNCONSUMED_RESPONSES(&queue->rx);
drivers/net/xen-netfront.c
1846
xennet_end_access(queue->rx_ring_ref, queue->rx.sring);
drivers/net/xen-netfront.c
1851
queue->rx.sring = NULL;
drivers/net/xen-netfront.c
187
struct xen_netif_rx_response rx;
drivers/net/xen-netfront.c
1984
queue->rx.sring = NULL;
drivers/net/xen-netfront.c
1999
XEN_FRONT_RING_INIT(&queue->rx, rxs, XEN_PAGE_SIZE);
drivers/net/xen-netfront.c
2016
xenbus_teardown_ring((void **)&queue->rx.sring, 1, &queue->rx_ring_ref);
drivers/net/xen-netfront.c
301
RING_IDX req_prod = queue->rx.req_prod_pvt;
drivers/net/xen-netfront.c
308
for (req_prod = queue->rx.req_prod_pvt;
drivers/net/xen-netfront.c
309
req_prod - queue->rx.rsp_cons < NET_RX_RING_SIZE;
drivers/net/xen-netfront.c
334
req = RING_GET_REQUEST(&queue->rx, req_prod);
drivers/net/xen-netfront.c
343
queue->rx.req_prod_pvt = req_prod;
drivers/net/xen-netfront.c
350
if (req_prod - queue->rx.rsp_cons < NET_RX_SLOTS_MIN ||
drivers/net/xen-netfront.c
356
RING_PUSH_REQUESTS_AND_CHECK_NOTIFY(&queue->rx, notify);
drivers/net/xen-netfront.c
378
queue->rx.sring->rsp_event = queue->rx.rsp_cons + 1;
drivers/net/xen-netfront.c
379
if (RING_HAS_UNCONSUMED_RESPONSES(&queue->rx))
drivers/net/xen-netfront.c
907
queue->rx.rsp_cons = val;
drivers/net/xen-netfront.c
908
queue->rx_rsp_unconsumed = XEN_RING_NR_UNCONSUMED_RESPONSES(&queue->rx);
drivers/net/xen-netfront.c
915
int new = xennet_rxidx(queue->rx.req_prod_pvt);
drivers/net/xen-netfront.c
920
RING_GET_REQUEST(&queue->rx, queue->rx.req_prod_pvt)->id = new;
drivers/net/xen-netfront.c
921
RING_GET_REQUEST(&queue->rx, queue->rx.req_prod_pvt)->gref = ref;
drivers/net/xen-netfront.c
922
queue->rx.req_prod_pvt++;
drivers/net/xen-netfront.c
932
RING_IDX cons = queue->rx.rsp_cons;
drivers/net/xen-netfront.c
946
RING_COPY_RESPONSE(&queue->rx, ++cons, &extra);
drivers/net/xen-netfront.c
968
struct xen_netif_rx_response *rx, struct bpf_prog *prog,
drivers/net/xen-netfront.c
972
u32 len = rx->status;
drivers/pci/controller/dwc/pcie-tegra194.c
1211
msg.rx.data = &resp;
drivers/pci/controller/dwc/pcie-tegra194.c
1212
msg.rx.size = sizeof(resp);
drivers/pci/controller/dwc/pcie-tegra194.c
1217
if (msg.rx.ret)
drivers/pci/controller/dwc/pcie-tegra194.c
1246
msg.rx.data = &resp;
drivers/pci/controller/dwc/pcie-tegra194.c
1247
msg.rx.size = sizeof(resp);
drivers/pci/controller/dwc/pcie-tegra194.c
1252
if (msg.rx.ret)
drivers/peci/controller/peci-aspeed.c
219
req->rx.len > ASPEED_PECI_DATA_BUF_SIZE_MAX)
drivers/peci/controller/peci-aspeed.c
232
FIELD_PREP(ASPEED_PECI_RD_LEN_MASK, req->rx.len);
drivers/peci/controller/peci-aspeed.c
276
for (i = 0; i < req->rx.len; i += 4) {
drivers/peci/controller/peci-aspeed.c
280
put_unaligned_le32(rx_data, &req->rx.buf[i]);
drivers/peci/controller/peci-aspeed.c
284
print_hex_dump_bytes("RX : ", DUMP_PREFIX_NONE, req->rx.buf, req->rx.len);
drivers/peci/controller/peci-npcm.c
105
addr, req->tx.len, req->rx.len);
drivers/peci/controller/peci-npcm.c
134
for (i = 0; i < req->rx.len; i++) {
drivers/peci/controller/peci-npcm.c
136
req->rx.buf[i] = (u8)msg_rd;
drivers/peci/controller/peci-npcm.c
142
print_hex_dump_bytes("RX : ", DUMP_PREFIX_NONE, req->rx.buf, req->rx.len);
drivers/peci/controller/peci-npcm.c
93
regmap_write(priv->regmap, NPCM_PECI_RD_LENGTH, NPCM_PECI_WR_LEN_MASK & req->rx.len);
drivers/peci/request.c
212
req->rx.len = rx_len;
drivers/peci/request.c
398
return req->rx.buf[1];
drivers/peci/request.c
404
return get_unaligned_le16(&req->rx.buf[1]);
drivers/peci/request.c
410
return get_unaligned_le32(&req->rx.buf[1]);
drivers/peci/request.c
416
return get_unaligned_le64(&req->rx.buf[1]);
drivers/peci/request.c
422
return get_unaligned_le64(&req->rx.buf[0]);
drivers/peci/request.c
428
return get_unaligned_le16(&req->rx.buf[0]);
drivers/peci/request.c
90
return req->rx.buf[0];
drivers/phy/qualcomm/phy-qcom-qmp-combo.c
2276
void __iomem *rx;
drivers/phy/qualcomm/phy-qcom-qmp-combo.c
3863
void __iomem *rx = qmp->rx;
drivers/phy/qualcomm/phy-qcom-qmp-combo.c
3884
qmp_configure_lane(qmp->dev, rx, cfg->rx_tbl, cfg->rx_tbl_num, 1);
drivers/phy/qualcomm/phy-qcom-qmp-combo.c
4628
qmp->rx = devm_of_iomap(dev, np, 1, NULL);
drivers/phy/qualcomm/phy-qcom-qmp-combo.c
4629
if (IS_ERR(qmp->rx))
drivers/phy/qualcomm/phy-qcom-qmp-combo.c
4630
return PTR_ERR(qmp->rx);
drivers/phy/qualcomm/phy-qcom-qmp-combo.c
4715
qmp->rx = base + offs->rxa;
drivers/phy/qualcomm/phy-qcom-qmp-pcie-msm8996.c
187
void __iomem *rx;
drivers/phy/qualcomm/phy-qcom-qmp-pcie-msm8996.c
412
void __iomem *rx = qphy->rx;
drivers/phy/qualcomm/phy-qcom-qmp-pcie-msm8996.c
435
qmp_configure_lane(qmp->dev, rx, cfg->rx_tbl, cfg->rx_tbl_num, 1);
drivers/phy/qualcomm/phy-qcom-qmp-pcie-msm8996.c
675
qphy->rx = devm_of_iomap(dev, np, 1, NULL);
drivers/phy/qualcomm/phy-qcom-qmp-pcie-msm8996.c
676
if (IS_ERR(qphy->rx))
drivers/phy/qualcomm/phy-qcom-qmp-pcie-msm8996.c
677
return PTR_ERR(qphy->rx);
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3229
u16 rx;
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3243
const struct qmp_phy_init_tbl *rx;
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3315
void __iomem *rx;
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3405
.rx = 0x0400,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3413
.rx = 0x0400,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3421
.rx = 0x0400,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3429
.rx = 0x0400,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3440
.rx = 0x0200,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3450
.rx = 0x1000,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3460
.rx = 0x0400,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3471
.rx = 0x0200,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3482
.rx = 0x0200,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3492
.rx = 0x1200,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3502
.rx = 0x0200,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3513
.rx = 0x0200,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3524
.rx = 0x0200,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3545
.rx = ipq8074_pcie_rx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3570
.rx = ipq8074_pcie_gen3_rx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3599
.rx = ipq6018_pcie_rx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3626
.rx = ipq9574_pcie_rx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3654
.rx = ipq9574_pcie_rx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3682
.rx = qcs615_pcie_rx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3706
.rx = qcs8300_qmp_gen4x2_pcie_rx_alt_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3741
.rx = sdm845_qmp_pcie_rx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3791
.rx = sm8250_qmp_pcie_rx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3801
.rx = sm8250_qmp_gen3x1_pcie_rx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3828
.rx = sm8250_qmp_pcie_rx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3838
.rx = sm8250_qmp_gen3x2_pcie_rx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3865
.rx = msm8998_pcie_rx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3890
.rx = sm8550_qmp_gen3x2_pcie_rx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3935
.rx = sc8180x_qmp_pcie_rx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3962
.rx = sc8280xp_qmp_gen3x1_pcie_rx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
3995
.rx = sc8280xp_qmp_gen3x2_pcie_rx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
4028
.rx = sc8280xp_qmp_gen3x2_pcie_rx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
4064
.rx = sdx55_qmp_pcie_rx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
4106
.rx = sm8450_qmp_gen3_pcie_rx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
4117
.rx = sm8350_qmp_gen3x1_pcie_rc_rx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
4141
.rx = sm8450_qmp_gen3_pcie_rx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
4150
.rx = sm8350_qmp_gen3x2_pcie_rc_rx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
4176
.rx = sdx65_qmp_pcie_rx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
4205
.rx = sm8450_qmp_gen3_pcie_rx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
4216
.rx = sm8450_qmp_gen3x1_pcie_rc_rx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
4240
.rx = sm8450_qmp_gen4x2_pcie_rx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
4285
.rx = sm8550_qmp_gen3x2_pcie_rx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
4312
.rx = sm8750_qmp_gen3x2_pcie_rx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
4339
.rx = sm8550_qmp_gen4x2_pcie_rx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
4371
.rx = sm8650_qmp_gen4x2_pcie_rx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
4402
.rx = sa8775p_qmp_gen4x2_pcie_rx_alt_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
4450
.rx = sa8775p_qmp_gen4x4_pcie_rx_alt_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
4492
.rx = x1e80100_qmp_gen4x2_pcie_rx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
4522
.rx = x1e80100_qmp_gen4x2_pcie_rx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
4552
.rx = x1e80100_qmp_gen4x8_pcie_rx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
4601
.rx = kaanapali_qmp_gen3x2_pcie_rx_tbl,
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
4659
rx3 = qmp->port_b + offs->rx;
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
4670
qmp_configure_lane(qmp->dev, rx3, tbls->rx, tbls->rx_num, 1);
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
4673
qmp_configure_lane(qmp->dev, rx4, tbls->rx, tbls->rx_num, 2);
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
4686
void __iomem *rx = qmp->rx;
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
4707
qmp_configure_lane(qmp->dev, rx, tbls->rx, tbls->rx_num, 1);
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
4711
qmp_configure_lane(qmp->dev, rx2, tbls->rx, tbls->rx_num, 2);
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
5183
qmp->rx = qmp->tx;
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
5185
qmp->rx = devm_of_iomap(dev, np, 1, NULL);
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
5186
if (IS_ERR(qmp->rx))
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
5187
return PTR_ERR(qmp->rx);
drivers/phy/qualcomm/phy-qcom-qmp-pcie.c
5296
qmp->rx = base + offs->rx;
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1136
u16 rx;
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1147
const struct qmp_phy_init_tbl *rx;
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1190
void __iomem *rx;
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1313
.rx = 0x600,
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1322
.rx = 0x1200,
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1338
.rx = milos_ufsphy_rx,
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1364
.rx = msm8996_ufsphy_rx,
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1387
.rx = sm8350_ufsphy_rx,
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1399
.rx = sm8350_ufsphy_g4_rx,
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1421
.rx = sc7280_ufsphy_rx,
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1433
.rx = sc7280_ufsphy_hs_g4_rx,
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1455
.rx = sm8350_ufsphy_rx,
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1467
.rx = sm8350_ufsphy_g4_rx,
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1489
.rx = sdm845_ufsphy_rx,
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1516
.rx = sm6115_ufsphy_rx,
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1543
.rx = sm7150_ufsphy_rx,
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1570
.rx = sm8150_ufsphy_rx,
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1582
.rx = sm8150_ufsphy_hs_g4_rx,
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1604
.rx = sm8150_ufsphy_rx,
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1616
.rx = sm8250_ufsphy_hs_g4_rx,
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1638
.rx = sm8350_ufsphy_rx,
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1650
.rx = sm8350_ufsphy_g4_rx,
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1672
.rx = sm8350_ufsphy_rx,
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1684
.rx = sm8350_ufsphy_g4_rx,
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1706
.rx = sm8550_ufsphy_rx,
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1720
.rx = sm8550_ufsphy_g4_rx,
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1742
.rx = sm8550_ufsphy_rx,
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1756
.rx = sm8550_ufsphy_g4_rx,
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1765
.rx = sm8550_ufsphy_g5_rx,
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1787
.rx = sm8650_ufsphy_rx,
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1819
.rx = sm8750_ufsphy_rx,
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1858
void __iomem *rx = qmp->rx;
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1861
qmp_configure_lane(qmp->dev, rx, tbls->rx, tbls->rx_num, 1);
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
1865
qmp_configure_lane(qmp->dev, qmp->rx2, tbls->rx, tbls->rx_num, 2);
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
2141
qmp->rx = devm_of_iomap(dev, np, 1, NULL);
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
2142
if (IS_ERR(qmp->rx))
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
2143
return PTR_ERR(qmp->rx);
drivers/phy/qualcomm/phy-qcom-qmp-ufs.c
2186
qmp->rx = base + offs->rx;
drivers/phy/qualcomm/phy-qcom-qmp-usb-legacy.c
1174
qmp->rx = devm_of_iomap(dev, np, 1, NULL);
drivers/phy/qualcomm/phy-qcom-qmp-usb-legacy.c
1175
if (IS_ERR(qmp->rx))
drivers/phy/qualcomm/phy-qcom-qmp-usb-legacy.c
1176
return PTR_ERR(qmp->rx);
drivers/phy/qualcomm/phy-qcom-qmp-usb-legacy.c
1227
qmp->rx = base + offs->rx;
drivers/phy/qualcomm/phy-qcom-qmp-usb-legacy.c
486
u16 rx;
drivers/phy/qualcomm/phy-qcom-qmp-usb-legacy.c
532
void __iomem *rx;
drivers/phy/qualcomm/phy-qcom-qmp-usb-legacy.c
828
void __iomem *rx = qmp->rx;
drivers/phy/qualcomm/phy-qcom-qmp-usb-legacy.c
844
qmp_usb_legacy_configure_lane(rx, cfg->rx_tbl, cfg->rx_tbl_num, 1);
drivers/phy/qualcomm/phy-qcom-qmp-usb.c
1383
u16 rx;
drivers/phy/qualcomm/phy-qcom-qmp-usb.c
1426
void __iomem *rx;
drivers/phy/qualcomm/phy-qcom-qmp-usb.c
1491
.rx = 0x400,
drivers/phy/qualcomm/phy-qcom-qmp-usb.c
1499
.rx = 0x400,
drivers/phy/qualcomm/phy-qcom-qmp-usb.c
1506
.rx = 0x400,
drivers/phy/qualcomm/phy-qcom-qmp-usb.c
1514
.rx = 0x0400,
drivers/phy/qualcomm/phy-qcom-qmp-usb.c
1522
.rx = 0x1000,
drivers/phy/qualcomm/phy-qcom-qmp-usb.c
1530
.rx = 0x1000,
drivers/phy/qualcomm/phy-qcom-qmp-usb.c
1538
.rx = 0x1000,
drivers/phy/qualcomm/phy-qcom-qmp-usb.c
1546
.rx = 0x1000,
drivers/phy/qualcomm/phy-qcom-qmp-usb.c
1940
void __iomem *rx = qmp->rx;
drivers/phy/qualcomm/phy-qcom-qmp-usb.c
1957
qmp_configure_lane(qmp->dev, rx, cfg->rx_tbl, cfg->rx_tbl_num, 1);
drivers/phy/qualcomm/phy-qcom-qmp-usb.c
2298
qmp->rx = devm_of_iomap(dev, np, 1, NULL);
drivers/phy/qualcomm/phy-qcom-qmp-usb.c
2299
if (IS_ERR(qmp->rx))
drivers/phy/qualcomm/phy-qcom-qmp-usb.c
2300
return PTR_ERR(qmp->rx);
drivers/phy/qualcomm/phy-qcom-qmp-usb.c
2359
qmp->rx = base + offs->rx;
drivers/phy/qualcomm/phy-qcom-qmp-usbc.c
1021
qmp_configure_lane(qmp->dev, qmp->rx, cfg->rx_tbl, cfg->rx_tbl_num, 1);
drivers/phy/qualcomm/phy-qcom-qmp-usbc.c
1778
qmp->rx = devm_of_iomap(dev, np, 1, NULL);
drivers/phy/qualcomm/phy-qcom-qmp-usbc.c
1779
if (IS_ERR(qmp->rx))
drivers/phy/qualcomm/phy-qcom-qmp-usbc.c
1780
return PTR_ERR(qmp->rx);
drivers/phy/qualcomm/phy-qcom-qmp-usbc.c
1848
qmp->rx = base + offs->rx;
drivers/phy/qualcomm/phy-qcom-qmp-usbc.c
375
u16 rx;
drivers/phy/qualcomm/phy-qcom-qmp-usbc.c
439
void __iomem *rx;
drivers/phy/qualcomm/phy-qcom-qmp-usbc.c
539
.rx = 0x400,
drivers/phy/qualcomm/phy-qcom-qmp-usbc.c
549
.rx = 0x400,
drivers/pinctrl/renesas/pfc-r8a7778.c
1291
#define CAN_PFC_DATA(name, tx, rx) SH_PFC_MUX2(name, tx, rx)
drivers/pinctrl/renesas/pfc-r8a7778.c
1334
#define SCIF_PFC_DAT(name, tx, rx) SH_PFC_MUX2(name, tx, rx)
drivers/pinctrl/renesas/pfc-r8a7778.c
1366
#define HSPI_PFC_DAT(name, clk, cs, rx, tx) SH_PFC_MUX4(name, clk, cs, rx, tx)
drivers/platform/mellanox/mlxbf-tmfifo.c
1309
ctl = readq(fifo->rx.ctl);
drivers/platform/mellanox/mlxbf-tmfifo.c
1316
writeq(ctl, fifo->rx.ctl);
drivers/platform/mellanox/mlxbf-tmfifo.c
1366
fifo->rx.ctl = fifo->res1 + MLXBF_TMFIFO_RX_CTL_BF3;
drivers/platform/mellanox/mlxbf-tmfifo.c
1367
fifo->rx.sts = fifo->res1 + MLXBF_TMFIFO_RX_STS_BF3;
drivers/platform/mellanox/mlxbf-tmfifo.c
1368
fifo->rx.data = fifo->res0 + MLXBF_TMFIFO_RX_DATA_BF3;
drivers/platform/mellanox/mlxbf-tmfifo.c
1373
fifo->rx.ctl = fifo->res0 + MLXBF_TMFIFO_RX_CTL;
drivers/platform/mellanox/mlxbf-tmfifo.c
1374
fifo->rx.sts = fifo->res0 + MLXBF_TMFIFO_RX_STS;
drivers/platform/mellanox/mlxbf-tmfifo.c
1375
fifo->rx.data = fifo->res0 + MLXBF_TMFIFO_RX_DATA;
drivers/platform/mellanox/mlxbf-tmfifo.c
195
struct mlxbf_tmfifo_io rx;
drivers/platform/mellanox/mlxbf-tmfifo.c
442
int rx, tx;
drivers/platform/mellanox/mlxbf-tmfifo.c
444
rx = !test_and_set_bit(MLXBF_TM_RX_HWM_IRQ, &fifo->pend_events);
drivers/platform/mellanox/mlxbf-tmfifo.c
447
if (rx || tx)
drivers/platform/mellanox/mlxbf-tmfifo.c
515
sts = readq(fifo->rx.sts);
drivers/platform/mellanox/mlxbf-tmfifo.c
616
data = readq(fifo->rx.data);
drivers/platform/mellanox/mlxbf-tmfifo.c
668
*(u64 *)&hdr = readq(fifo->rx.data);
drivers/platform/surface/aggregator/ssh_packet_layer.c
1623
for (i = 0; i < ARRAY_SIZE(ptl->rx.blocked.seqs); i++) {
drivers/platform/surface/aggregator/ssh_packet_layer.c
1624
if (likely(ptl->rx.blocked.seqs[i] != frame->seq))
drivers/platform/surface/aggregator/ssh_packet_layer.c
1632
ptl->rx.blocked.seqs[ptl->rx.blocked.offset] = frame->seq;
drivers/platform/surface/aggregator/ssh_packet_layer.c
1633
ptl->rx.blocked.offset = (ptl->rx.blocked.offset + 1)
drivers/platform/surface/aggregator/ssh_packet_layer.c
1634
% ARRAY_SIZE(ptl->rx.blocked.seqs);
drivers/platform/surface/aggregator/ssh_packet_layer.c
1803
wait_event_interruptible(ptl->rx.wq,
drivers/platform/surface/aggregator/ssh_packet_layer.c
1804
!kfifo_is_empty(&ptl->rx.fifo) ||
drivers/platform/surface/aggregator/ssh_packet_layer.c
1810
n = sshp_buf_read_from_fifo(&ptl->rx.buf, &ptl->rx.fifo);
drivers/platform/surface/aggregator/ssh_packet_layer.c
1814
ptl->rx.buf.ptr + ptl->rx.buf.len - n,
drivers/platform/surface/aggregator/ssh_packet_layer.c
1818
while (offs < ptl->rx.buf.len) {
drivers/platform/surface/aggregator/ssh_packet_layer.c
1819
sshp_buf_span_from(&ptl->rx.buf, offs, &span);
drivers/platform/surface/aggregator/ssh_packet_layer.c
1828
sshp_buf_drop(&ptl->rx.buf, offs);
drivers/platform/surface/aggregator/ssh_packet_layer.c
1836
wake_up(&ptl->rx.wq);
drivers/platform/surface/aggregator/ssh_packet_layer.c
1847
if (ptl->rx.thread)
drivers/platform/surface/aggregator/ssh_packet_layer.c
1850
ptl->rx.thread = kthread_run(ssh_ptl_rx_threadfn, ptl,
drivers/platform/surface/aggregator/ssh_packet_layer.c
1852
if (IS_ERR(ptl->rx.thread))
drivers/platform/surface/aggregator/ssh_packet_layer.c
1853
return PTR_ERR(ptl->rx.thread);
drivers/platform/surface/aggregator/ssh_packet_layer.c
1868
if (ptl->rx.thread) {
drivers/platform/surface/aggregator/ssh_packet_layer.c
1869
status = kthread_stop(ptl->rx.thread);
drivers/platform/surface/aggregator/ssh_packet_layer.c
1870
ptl->rx.thread = NULL;
drivers/platform/surface/aggregator/ssh_packet_layer.c
1897
used = kfifo_in(&ptl->rx.fifo, buf, n);
drivers/platform/surface/aggregator/ssh_packet_layer.c
2047
ptl->rx.thread = NULL;
drivers/platform/surface/aggregator/ssh_packet_layer.c
2048
init_waitqueue_head(&ptl->rx.wq);
drivers/platform/surface/aggregator/ssh_packet_layer.c
2058
for (i = 0; i < ARRAY_SIZE(ptl->rx.blocked.seqs); i++)
drivers/platform/surface/aggregator/ssh_packet_layer.c
2059
ptl->rx.blocked.seqs[i] = U16_MAX;
drivers/platform/surface/aggregator/ssh_packet_layer.c
2060
ptl->rx.blocked.offset = 0;
drivers/platform/surface/aggregator/ssh_packet_layer.c
2062
status = kfifo_alloc(&ptl->rx.fifo, SSH_PTL_RX_FIFO_LEN, GFP_KERNEL);
drivers/platform/surface/aggregator/ssh_packet_layer.c
2066
status = sshp_buf_alloc(&ptl->rx.buf, SSH_PTL_RX_BUF_LEN, GFP_KERNEL);
drivers/platform/surface/aggregator/ssh_packet_layer.c
2068
kfifo_free(&ptl->rx.fifo);
drivers/platform/surface/aggregator/ssh_packet_layer.c
2084
kfifo_free(&ptl->rx.fifo);
drivers/platform/surface/aggregator/ssh_packet_layer.c
2085
sshp_buf_free(&ptl->rx.buf);
drivers/platform/surface/aggregator/ssh_packet_layer.h
110
} rx;
drivers/pmdomain/tegra/powergate-bpmp.c
103
msg.rx.data = &response;
drivers/pmdomain/tegra/powergate-bpmp.c
104
msg.rx.size = sizeof(response);
drivers/pmdomain/tegra/powergate-bpmp.c
109
else if (msg.rx.ret < 0)
drivers/pmdomain/tegra/powergate-bpmp.c
133
msg.rx.data = &response;
drivers/pmdomain/tegra/powergate-bpmp.c
134
msg.rx.size = sizeof(response);
drivers/pmdomain/tegra/powergate-bpmp.c
137
if (err < 0 || msg.rx.ret < 0)
drivers/pmdomain/tegra/powergate-bpmp.c
51
else if (msg.rx.ret < 0)
drivers/pmdomain/tegra/powergate-bpmp.c
75
msg.rx.data = &response;
drivers/pmdomain/tegra/powergate-bpmp.c
76
msg.rx.size = sizeof(response);
drivers/pmdomain/tegra/powergate-bpmp.c
81
else if (msg.rx.ret < 0)
drivers/ps3/ps3-vuart.c
178
trig->rx = size - val;
drivers/ps3/ps3-vuart.c
181
trig->tx, trig->rx);
drivers/ps3/ps3-vuart.c
187
unsigned int rx)
drivers/ps3/ps3-vuart.c
211
PARAM_RX_TRIGGER, size - rx);
drivers/ps3/ps3-vuart.c
220
tx, rx);
drivers/ps3/vuart.h
74
unsigned long rx;
drivers/ps3/vuart.h
81
unsigned int rx);
drivers/pwm/pwm-mc33xs2410.c
111
val[i - 1] = FIELD_GET(MC33XS2410_FRAME_OUT_DATA, rx[i]);
drivers/pwm/pwm-mc33xs2410.c
90
u16 rx[MC33XS2410_MAX_TRANSFERS];
drivers/pwm/pwm-mc33xs2410.c
93
.rx_buf = rx,
drivers/reset/tegra/reset-bpmp.c
37
if (msg.rx.ret)
drivers/rpmsg/qcom_glink_native.c
1892
struct qcom_glink_pipe *rx,
drivers/rpmsg/qcom_glink_native.c
1905
glink->rx_pipe = rx;
drivers/rpmsg/qcom_glink_native.h
35
struct qcom_glink_pipe *rx,
drivers/rpmsg/qcom_glink_rpm.c
216
struct glink_rpm_pipe *rx,
drivers/rpmsg/qcom_glink_rpm.c
259
rx->native.length = size;
drivers/rpmsg/qcom_glink_rpm.c
261
rx->tail = msg_ram + offset;
drivers/rpmsg/qcom_glink_rpm.c
262
rx->head = msg_ram + offset + sizeof(u32);
drivers/rpmsg/qcom_glink_rpm.c
263
rx->fifo = msg_ram + offset + 2 * sizeof(u32);
drivers/rpmsg/qcom_glink_rpm.c
275
if (!rx->fifo || !tx->fifo) {
drivers/rpmsg/qcom_smd.c
250
struct smd_channel_info rx;
drivers/rpmsg/qcom_smd.c
272
struct smd_channel_info_word rx;
drivers/rpmsg/qcom_smd.c
277
BUILD_BUG_ON(sizeof(channel->info->rx.param) != sizeof(u8)); \
drivers/rpmsg/qcom_smd.c
279
le32_to_cpu(channel->info_word->rx.param) : \
drivers/rpmsg/qcom_smd.c
280
channel->info->rx.param; \
drivers/rpmsg/qcom_smd.c
285
BUILD_BUG_ON(sizeof(channel->info->rx.param) != sizeof(u32)); \
drivers/rpmsg/qcom_smd.c
287
channel->info_word->rx.param : \
drivers/rpmsg/qcom_smd.c
288
channel->info->rx.param); \
drivers/rpmsg/qcom_smd.c
293
BUILD_BUG_ON(sizeof(channel->info->rx.param) != sizeof(u8)); \
drivers/rpmsg/qcom_smd.c
295
channel->info_word->rx.param = cpu_to_le32(value); \
drivers/rpmsg/qcom_smd.c
297
channel->info->rx.param = value; \
drivers/rpmsg/qcom_smd.c
302
BUILD_BUG_ON(sizeof(channel->info->rx.param) != sizeof(u32)); \
drivers/rpmsg/qcom_smd.c
304
channel->info_word->rx.param = cpu_to_le32(value); \
drivers/rpmsg/qcom_smd.c
306
channel->info->rx.param = cpu_to_le32(value); \
drivers/rtc/rtc-ds1305.c
490
u8 *addr, size_t count, char *tx, char *rx)
drivers/rtc/rtc-ds1305.c
502
x->rx_buf = rx;
drivers/s390/block/dasd_diag.c
71
union register_pair rx = { .even = (unsigned long)iob, };
drivers/s390/block/dasd_diag.c
85
: CC_OUT(cc, cc), [rx] "+d" (rx.pair),
drivers/s390/block/dasd_diag.c
90
return cc | rx.odd;
drivers/s390/net/qeth_core.h
246
} rx;
drivers/s390/net/qeth_core.h
853
struct qeth_rx rx;
drivers/s390/net/qeth_core_main.c
2994
memset(&card->rx, 0, sizeof(struct qeth_rx));
drivers/s390/net/qeth_core_main.c
5506
&l3_hdr->next_hop.rx.src_mac, skb->len);
drivers/s390/net/qeth_core_main.c
5518
l3_hdr->next_hop.rx.vlan_id;
drivers/s390/net/qeth_core_main.c
5755
if (qeth_extract_skb(card, buf, &card->rx.buf_element,
drivers/s390/net/qeth_core_main.c
5756
&card->rx.e_offset)) {
drivers/s390/net/qeth_core_main.c
5770
struct qeth_rx *ctx = &card->rx;
drivers/s390/net/qeth_core_main.c
5779
if (!card->rx.b_count) {
drivers/s390/net/qeth_core_main.c
5780
card->rx.qdio_err = 0;
drivers/s390/net/qeth_core_main.c
5781
card->rx.b_count =
drivers/s390/net/qeth_core_main.c
5783
&card->rx.b_index,
drivers/s390/net/qeth_core_main.c
5784
&card->rx.qdio_err);
drivers/s390/net/qeth_core_main.c
5785
if (card->rx.b_count <= 0) {
drivers/s390/net/qeth_core_main.c
5786
card->rx.b_count = 0;
drivers/s390/net/qeth_core_main.c
5792
buffer = &card->qdio.in_q->bufs[card->rx.b_index];
drivers/s390/net/qeth_core_main.c
5793
if (!(card->rx.qdio_err &&
drivers/s390/net/qeth_core_main.c
5795
card->rx.qdio_err, "qinerr")))
drivers/s390/net/qeth_core_main.c
5808
card->rx.b_count--;
drivers/s390/net/qeth_core_main.c
5814
card->rx.b_index = QDIO_BUFNR(card->rx.b_index + 1);
drivers/s390/net/qeth_core_main.c
5815
card->rx.buf_element = 0;
drivers/s390/net/qeth_core_main.c
5816
card->rx.e_offset = 0;
drivers/s390/net/qeth_core_main.c
5862
struct qeth_rx *ctx = &card->rx;
drivers/scsi/aacraid/aacraid.h
1139
#define rx_readb(AEP, CSR) readb(&((AEP)->regs.rx->CSR))
drivers/scsi/aacraid/aacraid.h
1140
#define rx_readl(AEP, CSR) readl(&((AEP)->regs.rx->CSR))
drivers/scsi/aacraid/aacraid.h
1141
#define rx_writeb(AEP, CSR, value) writeb(value, &((AEP)->regs.rx->CSR))
drivers/scsi/aacraid/aacraid.h
1142
#define rx_writel(AEP, CSR, value) writel(value, &((AEP)->regs.rx->CSR))
drivers/scsi/aacraid/aacraid.h
1622
struct rx_registers __iomem *rx;
drivers/scsi/aacraid/nark.c
34
iounmap(dev->regs.rx);
drivers/scsi/aacraid/nark.c
35
dev->regs.rx = NULL;
drivers/scsi/aacraid/nark.c
41
dev->regs.rx = ioremap((u64)pci_resource_start(dev->pdev, 0) |
drivers/scsi/aacraid/nark.c
45
if (dev->regs.rx == NULL)
drivers/scsi/aacraid/nark.c
49
iounmap(dev->regs.rx);
drivers/scsi/aacraid/nark.c
50
dev->regs.rx = NULL;
drivers/scsi/aacraid/rx.c
463
iounmap(dev->regs.rx);
drivers/scsi/aacraid/rx.c
466
dev->base = dev->regs.rx = ioremap(dev->base_start, size);
drivers/scsi/aacraid/rx.c
469
dev->IndexRegs = &dev->regs.rx->IndexRegs;
drivers/scsi/cxgbi/libcxgbi.h
239
struct cxgbi_skb_rx_cb rx;
drivers/scsi/cxgbi/libcxgbi.h
249
#define cxgbi_skcb_rx_ddigest(skb) (CXGBI_SKB_CB(skb)->rx.ddigest)
drivers/scsi/cxgbi/libcxgbi.h
250
#define cxgbi_skcb_rx_pdulen(skb) (CXGBI_SKB_CB(skb)->rx.pdulen)
drivers/scsi/fnic/fnic_main.c
270
stats->rx_frames = vs->rx.rx_unicast_frames_ok;
drivers/scsi/fnic/fnic_main.c
271
stats->rx_words = vs->rx.rx_unicast_bytes_ok / 4;
drivers/scsi/fnic/fnic_main.c
272
stats->error_frames = vs->tx.tx_errors + vs->rx.rx_errors;
drivers/scsi/fnic/fnic_main.c
273
stats->dumped_frames = vs->tx.tx_drops + vs->rx.rx_drop;
drivers/scsi/fnic/fnic_main.c
274
stats->invalid_crc_count = vs->rx.rx_crc_errors;
drivers/scsi/fnic/vnic_stats.h
53
struct vnic_rx_stats rx;
drivers/scsi/mvsas/mv_init.c
107
if (mvi->rx)
drivers/scsi/mvsas/mv_init.c
109
sizeof(*mvi->rx) * (MVS_RX_RING_SZ + 1),
drivers/scsi/mvsas/mv_init.c
110
mvi->rx, mvi->rx_dma);
drivers/scsi/mvsas/mv_init.c
237
mvi->rx = dma_alloc_coherent(mvi->dev,
drivers/scsi/mvsas/mv_init.c
238
sizeof(*mvi->rx) * (MVS_RX_RING_SZ + 1),
drivers/scsi/mvsas/mv_init.c
240
if (!mvi->rx)
drivers/scsi/mvsas/mv_init.c
242
mvi->rx[0] = cpu_to_le32(0xfff);
drivers/scsi/mvsas/mv_sas.c
1875
mvi->rx_cons = le32_to_cpu(mvi->rx[0]);
drivers/scsi/mvsas/mv_sas.c
1892
rx_desc = le32_to_cpu(mvi->rx[rx_prod_idx + 1]);
drivers/scsi/mvsas/mv_sas.h
356
__le32 *rx;
drivers/scsi/snic/vnic_stats.h
51
struct vnic_rx_stats rx;
drivers/soc/apple/mailbox.c
198
mbox->rx(mbox, msg, mbox->cookie);
drivers/soc/apple/mailbox.h
35
void (*rx)(struct apple_mbox *mbox, struct apple_mbox_msg msg, void *cookie);
drivers/soc/apple/rtkit.c
695
rtk->mbox->rx = apple_rtkit_rx;
drivers/soc/fsl/qbman/qman_test_stash.c
137
struct qman_fq rx;
drivers/soc/fsl/qbman/qman_test_stash.c
354
if (qman_retire_fq(&handler->rx, &flags) ||
drivers/soc/fsl/qbman/qman_test_stash.c
360
if (qman_oos_fq(&handler->rx)) {
drivers/soc/fsl/qbman/qman_test_stash.c
365
qman_destroy_fq(&handler->rx);
drivers/soc/fsl/qbman/qman_test_stash.c
398
memset(&handler->rx, 0, sizeof(handler->rx));
drivers/soc/fsl/qbman/qman_test_stash.c
400
handler->rx.cb.dqrr = special_dqrr;
drivers/soc/fsl/qbman/qman_test_stash.c
402
handler->rx.cb.dqrr = normal_dqrr;
drivers/soc/fsl/qbman/qman_test_stash.c
403
err = qman_create_fq(handler->fqid_rx, 0, &handler->rx);
drivers/soc/fsl/qbman/qman_test_stash.c
413
err = qman_init_fq(&handler->rx, QMAN_INITFQ_FLAG_SCHED |
drivers/soc/ti/knav_dma.c
165
if (cfg->u.rx.einfo_present)
drivers/soc/ti/knav_dma.c
167
if (cfg->u.rx.psinfo_present)
drivers/soc/ti/knav_dma.c
169
if (cfg->u.rx.err_mode == DMA_RETRY)
drivers/soc/ti/knav_dma.c
171
v |= (cfg->u.rx.desc_type & DESC_TYPE_MASK) << DESC_TYPE_SHIFT;
drivers/soc/ti/knav_dma.c
172
if (cfg->u.rx.psinfo_at_sop)
drivers/soc/ti/knav_dma.c
174
v |= (cfg->u.rx.sop_offset & CHAN_SOP_OFF_MASK)
drivers/soc/ti/knav_dma.c
176
v |= cfg->u.rx.dst_q & CHAN_QNUM_MASK;
drivers/soc/ti/knav_dma.c
182
v = cfg->u.rx.fdq[0] << 16;
drivers/soc/ti/knav_dma.c
183
v |= cfg->u.rx.fdq[1] & CHAN_QNUM_MASK;
drivers/soc/ti/knav_dma.c
186
v = cfg->u.rx.fdq[2] << 16;
drivers/soc/ti/knav_dma.c
187
v |= cfg->u.rx.fdq[3] & CHAN_QNUM_MASK;
drivers/soc/ti/knav_dma.c
327
chan->cfg.u.rx.einfo_present,
drivers/soc/ti/knav_dma.c
328
chan->cfg.u.rx.psinfo_present,
drivers/soc/ti/knav_dma.c
329
chan->cfg.u.rx.desc_type);
drivers/soc/ti/knav_dma.c
331
chan->cfg.u.rx.dst_q,
drivers/soc/ti/knav_dma.c
332
chan->cfg.u.rx.thresh);
drivers/soc/ti/knav_dma.c
334
seq_printf(s, "[%d]", chan->cfg.u.rx.fdq[i]);
drivers/spi/spi-altera-core.c
120
if (hw->rx) {
drivers/spi/spi-altera-core.c
123
hw->rx[hw->count] = rxd;
drivers/spi/spi-altera-core.c
126
hw->rx[hw->count * 2] = rxd;
drivers/spi/spi-altera-core.c
127
hw->rx[hw->count * 2 + 1] = rxd >> 8;
drivers/spi/spi-altera-core.c
130
hw->rx[hw->count * 4] = rxd;
drivers/spi/spi-altera-core.c
131
hw->rx[hw->count * 4 + 1] = rxd >> 8;
drivers/spi/spi-altera-core.c
132
hw->rx[hw->count * 4 + 2] = rxd >> 16;
drivers/spi/spi-altera-core.c
133
hw->rx[hw->count * 4 + 3] = rxd >> 24;
drivers/spi/spi-altera-core.c
149
hw->rx = t->rx_buf;
drivers/spi/spi-au1550.c
316
hw->rx = t->rx_buf;
drivers/spi/spi-au1550.c
347
hw->rx = hw->dma_rx_tmpbuf;
drivers/spi/spi-au1550.c
356
hw->tx = hw->rx;
drivers/spi/spi-au1550.c
360
res = au1xxx_dbdma_put_dest(hw->dma_rx_ch, virt_to_phys(hw->rx),
drivers/spi/spi-au1550.c
464
if (hw->rx) { \
drivers/spi/spi-au1550.c
465
*(u##size *)hw->rx = (u##size)fifoword; \
drivers/spi/spi-au1550.c
466
hw->rx += (size) / 8; \
drivers/spi/spi-au1550.c
499
hw->rx = t->rx_buf;
drivers/spi/spi-au1550.c
50
u8 *rx;
drivers/spi/spi-bcm63xx-hsspi.c
380
u8 *rx = t->rx_buf;
drivers/spi/spi-bcm63xx-hsspi.c
396
if (tx && rx)
drivers/spi/spi-bcm63xx-hsspi.c
400
else if (rx)
drivers/spi/spi-bcm63xx-hsspi.c
444
if (rx)
drivers/spi/spi-bcm63xx-hsspi.c
445
memcpy_fromio(rx, bs->fifo, t->len);
drivers/spi/spi-bcm63xx-hsspi.c
501
u8 *rx = t->rx_buf;
drivers/spi/spi-bcm63xx-hsspi.c
508
if (tx && rx)
drivers/spi/spi-bcm63xx-hsspi.c
512
else if (rx)
drivers/spi/spi-bcm63xx-hsspi.c
556
if (rx) {
drivers/spi/spi-bcm63xx-hsspi.c
557
memcpy_fromio(rx, bs->fifo, curr_step);
drivers/spi/spi-bcm63xx-hsspi.c
558
rx += curr_step;
drivers/spi/spi-bcmbca-hsspi.c
259
u8 *rx = t->rx_buf;
drivers/spi/spi-bcmbca-hsspi.c
264
if (tx && rx)
drivers/spi/spi-bcmbca-hsspi.c
268
else if (rx)
drivers/spi/spi-bcmbca-hsspi.c
320
if (rx) {
drivers/spi/spi-bcmbca-hsspi.c
321
memcpy_fromio(rx, bs->fifo, curr_step);
drivers/spi/spi-bcmbca-hsspi.c
322
rx += curr_step;
drivers/spi/spi-bitbang.c
103
if (rx)
drivers/spi/spi-bitbang.c
104
*rx++ = word;
drivers/spi/spi-bitbang.c
123
u32 *rx = t->rx_buf;
drivers/spi/spi-bitbang.c
134
if (rx)
drivers/spi/spi-bitbang.c
135
*rx++ = word;
drivers/spi/spi-bitbang.c
61
u8 *rx = t->rx_buf;
drivers/spi/spi-bitbang.c
72
if (rx)
drivers/spi/spi-bitbang.c
73
*rx++ = word;
drivers/spi/spi-bitbang.c
92
u16 *rx = t->rx_buf;
drivers/spi/spi-davinci.c
175
void *rx;
drivers/spi/spi-davinci.c
196
if (dspi->rx) {
drivers/spi/spi-davinci.c
197
u8 *rx = dspi->rx;
drivers/spi/spi-davinci.c
198
*rx++ = (u8)data;
drivers/spi/spi-davinci.c
199
dspi->rx = rx;
drivers/spi/spi-davinci.c
205
if (dspi->rx) {
drivers/spi/spi-davinci.c
206
u16 *rx = dspi->rx;
drivers/spi/spi-davinci.c
207
*rx++ = (u16)data;
drivers/spi/spi-davinci.c
208
dspi->rx = rx;
drivers/spi/spi-davinci.c
645
dspi->rx = t->rx_buf;
drivers/spi/spi-dln2.c
192
} rx;
drivers/spi/spi-dln2.c
193
unsigned rx_len = sizeof(rx);
drivers/spi/spi-dln2.c
197
&rx, &rx_len);
drivers/spi/spi-dln2.c
200
if (rx_len < sizeof(rx))
drivers/spi/spi-dln2.c
203
*cs_num = le16_to_cpu(rx.cs_count);
drivers/spi/spi-dln2.c
218
} rx;
drivers/spi/spi-dln2.c
219
unsigned rx_len = sizeof(rx);
drivers/spi/spi-dln2.c
223
ret = dln2_transfer(dln2->pdev, cmd, &tx, sizeof(tx), &rx, &rx_len);
drivers/spi/spi-dln2.c
226
if (rx_len < sizeof(rx))
drivers/spi/spi-dln2.c
229
*freq = le32_to_cpu(rx.speed);
drivers/spi/spi-dln2.c
268
} rx;
drivers/spi/spi-dln2.c
269
int rx_len = sizeof(rx);
drivers/spi/spi-dln2.c
275
&rx, &rx_len);
drivers/spi/spi-dln2.c
278
if (rx_len < sizeof(rx))
drivers/spi/spi-dln2.c
327
} *rx = dln2->buf;
drivers/spi/spi-dln2.c
328
unsigned rx_len = sizeof(*rx);
drivers/spi/spi-dln2.c
334
&tx, sizeof(tx), rx, &rx_len);
drivers/spi/spi-dln2.c
337
if (rx_len < sizeof(*rx))
drivers/spi/spi-dln2.c
339
if (rx->count > ARRAY_SIZE(rx->frame_sizes))
drivers/spi/spi-dln2.c
343
for (i = 0; i < rx->count; i++)
drivers/spi/spi-dln2.c
344
*bpw_mask |= BIT(rx->frame_sizes[i] - 1);
drivers/spi/spi-dln2.c
460
} __packed *rx = dln2->buf;
drivers/spi/spi-dln2.c
461
unsigned rx_len = sizeof(*rx);
drivers/spi/spi-dln2.c
463
BUILD_BUG_ON(sizeof(*rx) > DLN2_SPI_BUF_SIZE);
drivers/spi/spi-dln2.c
473
rx, &rx_len);
drivers/spi/spi-dln2.c
476
if (rx_len < sizeof(rx->size) + data_len)
drivers/spi/spi-dln2.c
478
if (le16_to_cpu(rx->size) != data_len)
drivers/spi/spi-dln2.c
481
dln2_spi_copy_from_buf(data, rx->buf, data_len, dln2->bpw);
drivers/spi/spi-dln2.c
502
} __packed *rx;
drivers/spi/spi-dln2.c
506
sizeof(*rx) > DLN2_SPI_BUF_SIZE);
drivers/spi/spi-dln2.c
517
rx = dln2->buf;
drivers/spi/spi-dln2.c
526
rx_len = sizeof(*rx);
drivers/spi/spi-dln2.c
529
rx, &rx_len);
drivers/spi/spi-dln2.c
532
if (rx_len < sizeof(rx->size) + data_len)
drivers/spi/spi-dln2.c
534
if (le16_to_cpu(rx->size) != data_len)
drivers/spi/spi-dln2.c
537
dln2_spi_copy_from_buf(rx_data, rx->buf, data_len, dln2->bpw);
drivers/spi/spi-dw-core.c
163
if (dws->rx) {
drivers/spi/spi-dw-core.c
165
*(u8 *)(dws->rx) = rxw;
drivers/spi/spi-dw-core.c
167
*(u16 *)(dws->rx) = rxw;
drivers/spi/spi-dw-core.c
169
*(u32 *)(dws->rx) = rxw;
drivers/spi/spi-dw-core.c
171
dws->rx += dws->n_bytes;
drivers/spi/spi-dw-core.c
432
dws->rx = transfer->rx_buf;
drivers/spi/spi-dw-core.c
549
dws->rx = op->data.buf.in;
drivers/spi/spi-dw-core.c
552
dws->rx = NULL;
drivers/spi/spi-dw-core.c
606
buf = dws->rx;
drivers/spi/spi-dw-dma.c
106
dws->dma_addr_widths = tx.dst_addr_widths & rx.src_addr_widths;
drivers/spi/spi-dw-dma.c
114
struct dw_dma_slave dma_rx = { .src_id = 0 }, *rx = &dma_rx;
drivers/spi/spi-dw-dma.c
131
rx->dma_dev = &dma_dev->dev;
drivers/spi/spi-dw-dma.c
132
dws->rxchan = dma_request_channel(mask, dw_spi_dma_chan_filter, rx);
drivers/spi/spi-dw-dma.c
77
struct dma_slave_caps tx, rx;
drivers/spi/spi-dw-dma.c
84
ret = dma_get_slave_caps(dws->rxchan, &rx);
drivers/spi/spi-dw-dma.c
89
rx.directions & BIT(DMA_DEV_TO_MEM)))
drivers/spi/spi-dw-dma.c
92
if (tx.max_sg_burst > 0 && rx.max_sg_burst > 0)
drivers/spi/spi-dw-dma.c
93
dws->dma_sg_burst = min(tx.max_sg_burst, rx.max_sg_burst);
drivers/spi/spi-dw-dma.c
96
else if (rx.max_sg_burst > 0)
drivers/spi/spi-dw-dma.c
97
dws->dma_sg_burst = rx.max_sg_burst;
drivers/spi/spi-dw.h
174
void *rx;
drivers/spi/spi-ep93xx.c
202
((u16 *)xfer->rx_buf)[espi->rx] = val;
drivers/spi/spi-ep93xx.c
203
espi->rx += 2;
drivers/spi/spi-ep93xx.c
206
((u8 *)xfer->rx_buf)[espi->rx] = val;
drivers/spi/spi-ep93xx.c
207
espi->rx += 1;
drivers/spi/spi-ep93xx.c
239
if (espi->rx == xfer->len)
drivers/spi/spi-ep93xx.c
493
espi->rx = 0;
drivers/spi/spi-ep93xx.c
89
size_t rx;
drivers/spi/spi-fsi.c
215
static int fsi_spi_data_in(u64 in, u8 *rx, int len)
drivers/spi/spi-fsi.c
221
rx[i] = (u8)(in >> (8 * ((num_bytes - 1) - i)));
drivers/spi/spi-fsi.c
336
u8 *rx = transfer->rx_buf;
drivers/spi/spi-fsi.c
354
recv += fsi_spi_data_in(in, &rx[recv],
drivers/spi/spi-fsl-cpm.c
148
mspi->rx_dma = dma_map_single(dev, mspi->rx, t->len,
drivers/spi/spi-fsl-dspi.c
1204
dspi->rx = transfer->rx_buf;
drivers/spi/spi-fsl-dspi.c
353
void *rx;
drivers/spi/spi-fsl-dspi.c
406
*(u8 *)dspi->rx = rxdata;
drivers/spi/spi-fsl-dspi.c
409
*(u16 *)dspi->rx = rxdata;
drivers/spi/spi-fsl-dspi.c
412
*(u32 *)dspi->rx = rxdata;
drivers/spi/spi-fsl-dspi.c
415
dspi->rx += dspi->oper_word_size;
drivers/spi/spi-fsl-dspi.c
426
*(u32 *)dspi->rx = be32_to_cpu((__force __be32)rxdata);
drivers/spi/spi-fsl-dspi.c
427
dspi->rx += sizeof(u32);
drivers/spi/spi-fsl-dspi.c
438
*(u16 *)dspi->rx = be16_to_cpu((__force __be16)rxdata);
drivers/spi/spi-fsl-dspi.c
439
dspi->rx += sizeof(u16);
drivers/spi/spi-fsl-dspi.c
456
*(u16 *)dspi->rx = lo;
drivers/spi/spi-fsl-dspi.c
457
*(u16 *)(dspi->rx + 2) = hi;
drivers/spi/spi-fsl-dspi.c
458
dspi->rx += sizeof(u32);
drivers/spi/spi-fsl-dspi.c
478
if (!dspi->rx)
drivers/spi/spi-fsl-dspi.c
550
if (dspi->rx) {
drivers/spi/spi-fsl-lib.c
33
type *rx = mpc8xxx_spi->rx; \
drivers/spi/spi-fsl-lib.c
34
*rx++ = (type)(data >> mpc8xxx_spi->rx_shift); \
drivers/spi/spi-fsl-lib.c
35
mpc8xxx_spi->rx = rx; \
drivers/spi/spi-fsl-lib.h
26
void *rx;
drivers/spi/spi-fsl-lpspi.c
116
void (*rx)(struct fsl_lpspi_data *);
drivers/spi/spi-fsl-lpspi.c
281
fsl_lpspi->rx(fsl_lpspi);
drivers/spi/spi-fsl-lpspi.c
384
struct dma_slave_config rx = {}, tx = {};
drivers/spi/spi-fsl-lpspi.c
413
rx.direction = DMA_DEV_TO_MEM;
drivers/spi/spi-fsl-lpspi.c
414
rx.src_addr = fsl_lpspi->base_phys + IMX7ULP_RDR;
drivers/spi/spi-fsl-lpspi.c
415
rx.src_addr_width = buswidth;
drivers/spi/spi-fsl-lpspi.c
416
rx.src_maxburst = 1;
drivers/spi/spi-fsl-lpspi.c
417
ret = dmaengine_slave_config(controller->dma_rx, &rx);
drivers/spi/spi-fsl-lpspi.c
487
fsl_lpspi->rx = fsl_lpspi_buf_rx_u8;
drivers/spi/spi-fsl-lpspi.c
490
fsl_lpspi->rx = fsl_lpspi_buf_rx_u16;
drivers/spi/spi-fsl-lpspi.c
493
fsl_lpspi->rx = fsl_lpspi_buf_rx_u32;
drivers/spi/spi-fsl-lpspi.c
632
struct sg_table *tx = &transfer->tx_sg, *rx = &transfer->rx_sg;
drivers/spi/spi-fsl-lpspi.c
640
rx->sgl, rx->nents, DMA_DEV_TO_MEM,
drivers/spi/spi-fsl-spi.c
271
mpc8xxx_spi->rx = t->rx_buf;
drivers/spi/spi-fsl-spi.c
433
if (mspi->rx)
drivers/spi/spi-geni-qcom.c
103
struct dma_chan *rx;
drivers/spi/spi-geni-qcom.c
229
dmaengine_terminate_sync(mas->rx);
drivers/spi/spi-geni-qcom.c
449
dmaengine_slave_config(mas->rx, &config);
drivers/spi/spi-geni-qcom.c
450
rx_desc = dmaengine_prep_slave_sg(mas->rx, xfer->rx_sg.sgl, xfer->rx_sg.nents,
drivers/spi/spi-geni-qcom.c
478
dma_async_issue_pending(mas->rx);
drivers/spi/spi-geni-qcom.c
549
if (mas->rx) {
drivers/spi/spi-geni-qcom.c
550
dma_release_channel(mas->rx);
drivers/spi/spi-geni-qcom.c
551
mas->rx = NULL;
drivers/spi/spi-geni-qcom.c
571
mas->rx = dma_request_chan(mas->dev, "rx");
drivers/spi/spi-geni-qcom.c
572
if (IS_ERR(mas->rx)) {
drivers/spi/spi-geni-qcom.c
573
ret = dev_err_probe(mas->dev, PTR_ERR(mas->rx),
drivers/spi/spi-geni-qcom.c
587
mas->rx = NULL;
drivers/spi/spi-hisi-kunpeng.c
134
void *rx;
drivers/spi/spi-hisi-kunpeng.c
230
if (hs->rx) {
drivers/spi/spi-hisi-kunpeng.c
233
*(u8 *)(hs->rx) = rxw;
drivers/spi/spi-hisi-kunpeng.c
236
*(u16 *)(hs->rx) = rxw;
drivers/spi/spi-hisi-kunpeng.c
239
*(u32 *)(hs->rx) = rxw;
drivers/spi/spi-hisi-kunpeng.c
242
hs->rx += hs->n_bytes;
drivers/spi/spi-hisi-kunpeng.c
392
hs->rx = transfer->rx_buf;
drivers/spi/spi-imx.c
127
void (*rx)(struct spi_imx_data *spi_imx);
drivers/spi/spi-imx.c
1292
spi_imx->rx(spi_imx);
drivers/spi/spi-imx.c
1353
spi_imx->rx = spi_imx_buf_rx_swap;
drivers/spi/spi-imx.c
1359
spi_imx->rx = spi_imx_buf_rx_u8;
drivers/spi/spi-imx.c
1362
spi_imx->rx = spi_imx_buf_rx_u16;
drivers/spi/spi-imx.c
1365
spi_imx->rx = spi_imx_buf_rx_u32;
drivers/spi/spi-imx.c
1380
spi_imx->rx = mx53_ecspi_rx_target;
drivers/spi/spi-imx.c
1853
struct dma_slave_config rx = {}, tx = {};
drivers/spi/spi-imx.c
1884
rx.direction = DMA_DEV_TO_MEM;
drivers/spi/spi-imx.c
1885
rx.src_addr = spi_imx->base_phys + MXC_CSPIRXDATA;
drivers/spi/spi-imx.c
1886
rx.src_addr_width = buswidth;
drivers/spi/spi-imx.c
1887
rx.src_maxburst = spi_imx->wml;
drivers/spi/spi-imx.c
1888
ret = dmaengine_slave_config(controller->dma_rx, &rx);
drivers/spi/spi-imx.c
2042
spi_imx->rx(spi_imx);
drivers/spi/spi-jcore.c
108
unsigned char *rx;
drivers/spi/spi-jcore.c
116
rx = t->rx_buf;
drivers/spi/spi-jcore.c
129
if (rx)
drivers/spi/spi-jcore.c
130
*rx++ = readl(data_reg);
drivers/spi/spi-kspi2.c
142
static int kspi2_txrx_byte(struct kspi2 *kspi, u8 tx, u8 *rx)
drivers/spi/spi-kspi2.c
158
if (rx)
drivers/spi/spi-kspi2.c
159
*rx = ioread8(kspi->base + KSPI2_DATA_REG);
drivers/spi/spi-kspi2.c
167
u8 rx;
drivers/spi/spi-kspi2.c
175
ret = kspi2_txrx_byte(kspi, tx, &rx);
drivers/spi/spi-kspi2.c
180
((u8 *)t->rx_buf)[i] = rx;
drivers/spi/spi-lantiq-ssc.c
178
u8 *rx;
drivers/spi/spi-lantiq-ssc.c
539
rx8 = spi->rx;
drivers/spi/spi-lantiq-ssc.c
542
spi->rx++;
drivers/spi/spi-lantiq-ssc.c
545
rx16 = (u16 *) spi->rx;
drivers/spi/spi-lantiq-ssc.c
548
spi->rx += 2;
drivers/spi/spi-lantiq-ssc.c
551
rx32 = (u32 *) spi->rx;
drivers/spi/spi-lantiq-ssc.c
554
spi->rx += 4;
drivers/spi/spi-lantiq-ssc.c
586
rx8 = spi->rx;
drivers/spi/spi-lantiq-ssc.c
593
spi->rx++;
drivers/spi/spi-lantiq-ssc.c
597
rx32 = (u32 *) spi->rx;
drivers/spi/spi-lantiq-ssc.c
601
spi->rx += 4;
drivers/spi/spi-lantiq-ssc.c
635
if (spi->rx && spi->rx_todo)
drivers/spi/spi-lantiq-ssc.c
642
} else if (spi->rx) {
drivers/spi/spi-lantiq-ssc.c
730
spi->rx = t->rx_buf;
drivers/spi/spi-lantiq-ssc.c
739
if (spi->rx) {
drivers/spi/spi-loongson-core.c
136
u8 *rx = xfer->rx_buf;
drivers/spi/spi-loongson-core.c
140
ret = loongson_spi_write_read_8bit(spi, &tx, &rx, count);
drivers/spi/spi-loopback-test.c
1007
void *tx, void *rx)
drivers/spi/spi-loopback-test.c
1047
tx, rx,
drivers/spi/spi-loopback-test.c
1072
char *rx = NULL, *tx = NULL;
drivers/spi/spi-loopback-test.c
1080
rx = vmalloc(SPI_TEST_MAX_SIZE_PLUS);
drivers/spi/spi-loopback-test.c
1082
rx = kzalloc(SPI_TEST_MAX_SIZE_PLUS, GFP_KERNEL);
drivers/spi/spi-loopback-test.c
1083
if (!rx)
drivers/spi/spi-loopback-test.c
1104
ret = test->run_test(spi, test, tx, rx);
drivers/spi/spi-loopback-test.c
1106
ret = spi_test_run_test(spi, test, tx, rx);
drivers/spi/spi-loopback-test.c
1121
kvfree(rx);
drivers/spi/spi-loopback-test.c
509
void *rx)
drivers/spi/spi-loopback-test.c
525
rx, SPI_TEST_MAX_SIZE_PLUS)) {
drivers/spi/spi-loopback-test.c
541
for (addr = rx; addr < (u8 *)rx + SPI_TEST_MAX_SIZE_PLUS; addr++) {
drivers/spi/spi-loopback-test.c
604
void *tx, void *rx)
drivers/spi/spi-loopback-test.c
613
ret = spi_check_rx_ranges(spi, msg, rx);
drivers/spi/spi-loopback-test.c
666
void *tx, void *rx)
drivers/spi/spi-loopback-test.c
685
*ptr = rx + off;
drivers/spi/spi-loopback-test.c
793
void *tx, void *rx)
drivers/spi/spi-loopback-test.c
803
memset(rx, SPI_TEST_PATTERN_DO_NOT_WRITE, SPI_TEST_MAX_SIZE_PLUS);
drivers/spi/spi-loopback-test.c
811
(void *)tx, rx);
drivers/spi/spi-loopback-test.c
817
(void *)tx, rx);
drivers/spi/spi-loopback-test.c
832
ret = test->execute_msg(spi, test, tx, rx);
drivers/spi/spi-loopback-test.c
834
ret = spi_test_execute_msg(spi, test, tx, rx);
drivers/spi/spi-loopback-test.c
855
void *tx, void *rx,
drivers/spi/spi-loopback-test.c
920
return _spi_test_run_iter(spi, &test, tx, rx);
drivers/spi/spi-loopback-test.c
934
void *tx, void *rx)
drivers/spi/spi-loopback-test.c
976
ret = spi_test_check_loopback_result(spi, msg, tx, rx);
drivers/spi/spi-lp8841-rtc.c
105
u8 *rx = t->rx_buf;
drivers/spi/spi-lp8841-rtc.c
118
} else if (rx) {
drivers/spi/spi-lp8841-rtc.c
124
*rx++ = word;
drivers/spi/spi-mt7621.c
193
int rx = min(rx_len, 32);
drivers/spi/spi-mt7621.c
197
val |= (rx * 8) << 12;
drivers/spi/spi-mt7621.c
208
for (i = 0; i < rx; i++) {
drivers/spi/spi-omap2-mcspi.c
599
u8 *rx;
drivers/spi/spi-omap2-mcspi.c
632
rx = xfer->rx_buf;
drivers/spi/spi-omap2-mcspi.c
648
if (rx != NULL)
drivers/spi/spi-omap2-mcspi.c
679
if (rx == NULL) {
drivers/spi/spi-omap2-mcspi.c
729
u8 *rx;
drivers/spi/spi-omap2-mcspi.c
732
rx = xfer->rx_buf;
drivers/spi/spi-omap2-mcspi.c
747
if (rx != NULL) {
drivers/spi/spi-omap2-mcspi.c
757
*rx++ = readl_relaxed(rx_reg);
drivers/spi/spi-omap2-mcspi.c
759
word_len, *(rx - 1));
drivers/spi/spi-omap2-mcspi.c
771
*rx++ = readl_relaxed(rx_reg);
drivers/spi/spi-omap2-mcspi.c
773
word_len, *(rx - 1));
drivers/spi/spi-omap2-mcspi.c
779
u16 *rx;
drivers/spi/spi-omap2-mcspi.c
782
rx = xfer->rx_buf;
drivers/spi/spi-omap2-mcspi.c
796
if (rx != NULL) {
drivers/spi/spi-omap2-mcspi.c
806
*rx++ = readl_relaxed(rx_reg);
drivers/spi/spi-omap2-mcspi.c
808
word_len, *(rx - 1));
drivers/spi/spi-omap2-mcspi.c
820
*rx++ = readl_relaxed(rx_reg);
drivers/spi/spi-omap2-mcspi.c
822
word_len, *(rx - 1));
drivers/spi/spi-omap2-mcspi.c
828
u32 *rx;
drivers/spi/spi-omap2-mcspi.c
831
rx = xfer->rx_buf;
drivers/spi/spi-omap2-mcspi.c
845
if (rx != NULL) {
drivers/spi/spi-omap2-mcspi.c
855
*rx++ = readl_relaxed(rx_reg);
drivers/spi/spi-omap2-mcspi.c
857
word_len, *(rx - 1));
drivers/spi/spi-omap2-mcspi.c
869
*rx++ = readl_relaxed(rx_reg);
drivers/spi/spi-omap2-mcspi.c
871
word_len, *(rx - 1));
drivers/spi/spi-orion.c
507
u8 *rx = xfer->rx_buf;
drivers/spi/spi-orion.c
510
if (orion_spi_write_read_8bit(spi, &tx, &rx) < 0)
drivers/spi/spi-orion.c
517
u16 *rx = xfer->rx_buf;
drivers/spi/spi-orion.c
520
if (orion_spi_write_read_16bit(spi, &tx, &rx) < 0)
drivers/spi/spi-pic32.c
116
const void *rx;
drivers/spi/spi-pic32.c
176
rxtx_gap = ((pic32s->rx_end - pic32s->rx) -
drivers/spi/spi-pic32.c
184
u32 rx_left = (pic32s->rx_end - pic32s->rx) / n_bytes;
drivers/spi/spi-pic32.c
197
*(__type *)(pic32s->rx) = v; \
drivers/spi/spi-pic32.c
198
pic32s->rx += sizeof(__type); \
drivers/spi/spi-pic32.c
268
if (pic32s->rx_end == pic32s->rx) {
drivers/spi/spi-pic32.c
536
pic32s->rx = (const void *)transfer->rx_buf;
drivers/spi/spi-pic32.c
538
pic32s->rx_end = pic32s->rx + transfer->len;
drivers/spi/spi-pl022.c
1190
if (pl022->rx >= pl022->rx_end) {
drivers/spi/spi-pl022.c
1194
if (unlikely(pl022->rx > pl022->rx_end)) {
drivers/spi/spi-pl022.c
1198
(u32) (pl022->rx - pl022->rx_end));
drivers/spi/spi-pl022.c
1229
pl022->rx = (void *)transfer->rx_buf;
drivers/spi/spi-pl022.c
1230
pl022->rx_end = pl022->rx + pl022->cur_transfer->len;
drivers/spi/spi-pl022.c
1233
pl022->read = pl022->rx ? pl022->cur_chip->read : READING_NULL;
drivers/spi/spi-pl022.c
1311
while (pl022->tx < pl022->tx_end || pl022->rx < pl022->rx_end) {
drivers/spi/spi-pl022.c
373
void *rx;
drivers/spi/spi-pl022.c
598
__func__, pl022->rx, pl022->rx_end, pl022->tx, pl022->tx_end);
drivers/spi/spi-pl022.c
602
&& (pl022->rx < pl022->rx_end)) {
drivers/spi/spi-pl022.c
608
*(u8 *) (pl022->rx) =
drivers/spi/spi-pl022.c
612
*(u16 *) (pl022->rx) =
drivers/spi/spi-pl022.c
616
*(u32 *) (pl022->rx) =
drivers/spi/spi-pl022.c
620
pl022->rx += (pl022->cur_chip->n_bytes);
drivers/spi/spi-pl022.c
651
&& (pl022->rx < pl022->rx_end)) {
drivers/spi/spi-pl022.c
657
*(u8 *) (pl022->rx) =
drivers/spi/spi-pl022.c
661
*(u16 *) (pl022->rx) =
drivers/spi/spi-pl022.c
665
*(u32 *) (pl022->rx) =
drivers/spi/spi-pl022.c
669
pl022->rx += (pl022->cur_chip->n_bytes);
drivers/spi/spi-pl022.c
922
setup_dma_scatter(pl022, pl022->rx,
drivers/spi/spi-ppc4xx.c
126
unsigned char *rx;
drivers/spi/spi-ppc4xx.c
149
hw->rx = t->rx_buf;
drivers/spi/spi-ppc4xx.c
299
if (hw->rx)
drivers/spi/spi-ppc4xx.c
300
hw->rx[count] = data;
drivers/spi/spi-pxa2xx-pci.c
162
rx = c->rx_param;
drivers/spi/spi-pxa2xx-pci.c
163
rx->dma_dev = &dma_dev->dev;
drivers/spi/spi-pxa2xx-pci.c
164
rx->m_master = 0;
drivers/spi/spi-pxa2xx-pci.c
165
rx->p_master = 1;
drivers/spi/spi-pxa2xx-pci.c
195
struct dw_dma_slave *tx, *rx;
drivers/spi/spi-pxa2xx-pci.c
236
rx = c->rx_param;
drivers/spi/spi-pxa2xx-pci.c
237
rx->dma_dev = &dma_dev->dev;
drivers/spi/spi-pxa2xx-pci.c
99
struct dw_dma_slave *tx, *rx;
drivers/spi/spi-pxa2xx.c
499
while (read_SSSR_bits(drv_data, SSSR_RNE) && drv_data->rx < drv_data->rx_end) {
drivers/spi/spi-pxa2xx.c
501
drv_data->rx += n_bytes;
drivers/spi/spi-pxa2xx.c
504
return drv_data->rx == drv_data->rx_end;
drivers/spi/spi-pxa2xx.c
521
while (read_SSSR_bits(drv_data, SSSR_RNE) && drv_data->rx < drv_data->rx_end) {
drivers/spi/spi-pxa2xx.c
522
*(u8 *)(drv_data->rx) = pxa2xx_spi_read(drv_data, SSDR);
drivers/spi/spi-pxa2xx.c
523
++drv_data->rx;
drivers/spi/spi-pxa2xx.c
526
return drv_data->rx == drv_data->rx_end;
drivers/spi/spi-pxa2xx.c
543
while (read_SSSR_bits(drv_data, SSSR_RNE) && drv_data->rx < drv_data->rx_end) {
drivers/spi/spi-pxa2xx.c
544
*(u16 *)(drv_data->rx) = pxa2xx_spi_read(drv_data, SSDR);
drivers/spi/spi-pxa2xx.c
545
drv_data->rx += 2;
drivers/spi/spi-pxa2xx.c
548
return drv_data->rx == drv_data->rx_end;
drivers/spi/spi-pxa2xx.c
565
while (read_SSSR_bits(drv_data, SSSR_RNE) && drv_data->rx < drv_data->rx_end) {
drivers/spi/spi-pxa2xx.c
566
*(u32 *)(drv_data->rx) = pxa2xx_spi_read(drv_data, SSDR);
drivers/spi/spi-pxa2xx.c
567
drv_data->rx += 4;
drivers/spi/spi-pxa2xx.c
570
return drv_data->rx == drv_data->rx_end;
drivers/spi/spi-pxa2xx.c
685
bytes_left = drv_data->rx_end - drv_data->rx;
drivers/spi/spi-pxa2xx.c
965
drv_data->rx = transfer->rx_buf;
drivers/spi/spi-pxa2xx.c
966
drv_data->rx_end = drv_data->rx + transfer->len;
drivers/spi/spi-pxa2xx.c
976
drv_data->read = drv_data->rx ? u8_reader : null_reader;
drivers/spi/spi-pxa2xx.c
980
drv_data->read = drv_data->rx ? u16_reader : null_reader;
drivers/spi/spi-pxa2xx.c
984
drv_data->read = drv_data->rx ? u32_reader : null_reader;
drivers/spi/spi-pxa2xx.h
66
void *rx;
drivers/spi/spi-rockchip.c
180
void *rx;
drivers/spi/spi-rockchip.c
343
if (!rs->rx)
drivers/spi/spi-rockchip.c
347
*(u8 *)rs->rx = (u8)rxw;
drivers/spi/spi-rockchip.c
349
*(u16 *)rs->rx = (u16)rxw;
drivers/spi/spi-rockchip.c
350
rs->rx += rs->n_bytes;
drivers/spi/spi-rockchip.c
387
rs->rx = xfer->rx_buf;
drivers/spi/spi-rockchip.c
460
rs->rx = xfer->rx_buf;
drivers/spi/spi-rockchip.c
634
rs->rx = rs->xfer->rx_buf;
drivers/spi/spi-rockchip.c
641
rs->rx += rs->xfer->len - rs->n_bytes * state.residue;
drivers/spi/spi-rockchip.c
646
if (rs->rx) {
drivers/spi/spi-rockchip.c
652
*(u8 *)rs->rx = (u8)rxw;
drivers/spi/spi-rockchip.c
654
*(u16 *)rs->rx = (u16)rxw;
drivers/spi/spi-rockchip.c
655
rs->rx += rs->n_bytes;
drivers/spi/spi-rockchip.c
657
rs->xfer->len = (unsigned int)(rs->rx - rs->xfer->rx_buf);
drivers/spi/spi-rspi.c
510
static int rspi_pio_transfer(struct rspi_data *rspi, const u8 *tx, u8 *rx,
drivers/spi/spi-rspi.c
519
if (rx) {
drivers/spi/spi-rspi.c
523
*rx++ = ret;
drivers/spi/spi-rspi.c
539
struct sg_table *rx)
drivers/spi/spi-rspi.c
548
if (rx) {
drivers/spi/spi-rspi.c
549
desc_rx = dmaengine_prep_slave_sg(rspi->ctlr->dma_rx, rx->sgl,
drivers/spi/spi-rspi.c
550
rx->nents, DMA_DEV_TO_MEM,
drivers/spi/spi-rspi.c
577
if (rx) {
drivers/spi/spi-rspi.c
599
if (rx && rspi->rx_irq != other_irq)
drivers/spi/spi-rspi.c
606
if (rx)
drivers/spi/spi-rspi.c
617
if (rx)
drivers/spi/spi-rspi.c
626
if (rx)
drivers/spi/spi-rspi.c
634
if (rx && rspi->rx_irq != other_irq)
drivers/spi/spi-rspi.c
640
if (rx)
drivers/spi/spi-rspi.c
756
u8 *rx, unsigned int len)
drivers/spi/spi-rspi.c
778
*rx++ = rspi_read_data(rspi);
drivers/spi/spi-rspi.c
835
u8 *rx = xfer->rx_buf;
drivers/spi/spi-rspi.c
854
*rx++ = rspi_read_data(rspi);
drivers/spi/spi-sh-hspi.c
149
u32 rx;
drivers/spi/spi-sh-hspi.c
185
rx = hspi_read(hspi, SPRBR);
drivers/spi/spi-sh-hspi.c
187
((u8 *)t->rx_buf)[i] = (u8)rx;
drivers/spi/spi-sh-msiof.c
647
void *rx, unsigned int len,
drivers/spi/spi-sh-msiof.c
657
if (rx) {
drivers/spi/spi-sh-msiof.c
701
sh_msiof_spi_set_mode_regs(p, tx, rx, 32, words1, words2);
drivers/spi/spi-sh-msiof.c
711
if (rx)
drivers/spi/spi-sh-msiof.c
716
ret = sh_msiof_spi_start(p, rx);
drivers/spi/spi-sh-msiof.c
729
if (rx) {
drivers/spi/spi-sh-msiof.c
747
ret = sh_msiof_spi_stop(p, rx);
drivers/spi/spi-sh-msiof.c
753
if (rx)
drivers/spi/spi-sh-msiof.c
761
sh_msiof_spi_stop(p, rx);
drivers/spi/spi-sh-msiof.c
766
if (rx)
drivers/spi/spi-stm32.c
1305
dma_conf->src_addr = spi->phys_addr + spi->cfg->regs->rx.reg;
drivers/spi/spi-stm32.c
245
const struct stm32_spi_reg rx;
drivers/spi/spi-stm32.c
398
.rx = { STM32FX_SPI_DR },
drivers/spi/spi-stm32.c
418
.rx = { STM32H7_SPI_RXDR },
drivers/spi/spi-stm32.c
438
.rx = { STM32H7_SPI_RXDR },
drivers/spi/spi-test.h
109
void *tx, void *rx);
drivers/spi/spi-test.h
114
void *tx, void *rx);
drivers/spi/spi-test.h
78
void *tx, void *rx);
drivers/spi/spi-test.h
80
void *tx, void *rx);
drivers/spi/spi-ti-qspi.c
314
u32 rx;
drivers/spi/spi-ti-qspi.c
373
rx = readl(qspi->base + QSPI_SPI_DATA_REG_3);
drivers/spi/spi-ti-qspi.c
374
*rxp++ = be32_to_cpu(rx);
drivers/spi/spi-ti-qspi.c
375
rx = readl(qspi->base + QSPI_SPI_DATA_REG_2);
drivers/spi/spi-ti-qspi.c
376
*rxp++ = be32_to_cpu(rx);
drivers/spi/spi-ti-qspi.c
377
rx = readl(qspi->base + QSPI_SPI_DATA_REG_1);
drivers/spi/spi-ti-qspi.c
378
*rxp++ = be32_to_cpu(rx);
drivers/spi/spi-ti-qspi.c
379
rx = readl(qspi->base + QSPI_SPI_DATA_REG);
drivers/spi/spi-ti-qspi.c
380
*rxp++ = be32_to_cpu(rx);
drivers/spi/spi-ti-qspi.c
383
rx = readl(qspi->base + QSPI_SPI_DATA_REG);
drivers/spi/spi-ti-qspi.c
385
*rxp++ = rx >> (rx_wlen - 8);
drivers/spi/spi-ti-qspi.c
387
*rxp++ = rx >> (rx_wlen - 16);
drivers/spi/spi-ti-qspi.c
389
*rxp++ = rx >> (rx_wlen - 24);
drivers/spi/spi-ti-qspi.c
391
*rxp++ = rx;
drivers/staging/most/net/net.c
193
if (most_start_channel(nd->iface, nd->rx.ch_id, &comp)) {
drivers/staging/most/net/net.c
201
most_stop_channel(nd->iface, nd->rx.ch_id, &comp);
drivers/staging/most/net/net.c
227
most_stop_channel(nd->iface, nd->rx.ch_id, &comp);
drivers/staging/most/net/net.c
297
if (nd && nd->rx.linked && nd->tx.linked)
drivers/staging/most/net/net.c
339
ch = ccfg->direction == MOST_CH_TX ? &nd->tx : &nd->rx;
drivers/staging/most/net/net.c
341
ch = ccfg->direction == MOST_CH_TX ? &nd->tx : &nd->rx;
drivers/staging/most/net/net.c
377
if (nd->rx.linked && channel_idx == nd->rx.ch_id) {
drivers/staging/most/net/net.c
378
ch = &nd->rx;
drivers/staging/most/net/net.c
386
if (nd->rx.linked && nd->tx.linked) {
drivers/staging/most/net/net.c
443
if (nd->rx.ch_id != mbo->hdm_channel_id) {
drivers/staging/most/net/net.c
65
struct net_dev_channel rx;
drivers/staging/nvec/nvec.c
496
if (nvec->rx->pos != nvec_msg_size(nvec->rx)) {
drivers/staging/nvec/nvec.c
498
(uint)nvec_msg_size(nvec->rx),
drivers/staging/nvec/nvec.c
499
(uint)nvec->rx->pos);
drivers/staging/nvec/nvec.c
501
nvec_msg_free(nvec, nvec->rx);
drivers/staging/nvec/nvec.c
505
if (nvec->rx->data[0] == NVEC_BAT)
drivers/staging/nvec/nvec.c
517
list_add_tail(&nvec->rx->node, &nvec->rx_data);
drivers/staging/nvec/nvec.c
523
if (!nvec_msg_is_event(nvec->rx))
drivers/staging/nvec/nvec.c
638
nvec->rx = nvec_msg_alloc(nvec, NVEC_MSG_RX);
drivers/staging/nvec/nvec.c
640
if (unlikely(!nvec->rx)) {
drivers/staging/nvec/nvec.c
644
nvec->rx->data[0] = received;
drivers/staging/nvec/nvec.c
645
nvec->rx->pos = 1;
drivers/staging/nvec/nvec.c
651
if (nvec->rx->data[0] != 0x01) {
drivers/staging/nvec/nvec.c
657
nvec_msg_free(nvec, nvec->rx);
drivers/staging/nvec/nvec.c
666
nvec->rx->data[1] = received;
drivers/staging/nvec/nvec.c
667
nvec->rx->pos = 2;
drivers/staging/nvec/nvec.c
694
else if (nvec->rx && nvec->rx->pos < NVEC_MSG_SIZE)
drivers/staging/nvec/nvec.c
695
nvec->rx->data[nvec->rx->pos++] = received;
drivers/staging/nvec/nvec.c
699
nvec->rx, nvec->rx ? nvec->rx->pos : 0,
drivers/staging/nvec/nvec.c
830
nvec->rx = &nvec->msg_pool[0];
drivers/staging/nvec/nvec.h
147
struct nvec_msg *rx;
drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c
152
.rx = BIT(IEEE80211_STYPE_ACTION >> 4)
drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c
156
.rx = BIT(IEEE80211_STYPE_ACTION >> 4) |
drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c
161
.rx = BIT(IEEE80211_STYPE_ASSOC_REQ >> 4) |
drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c
172
.rx = BIT(IEEE80211_STYPE_ASSOC_REQ >> 4) |
drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c
182
.rx = BIT(IEEE80211_STYPE_ACTION >> 4) |
drivers/staging/rtl8723bs/os_dep/ioctl_cfg80211.c
187
.rx = BIT(IEEE80211_STYPE_ASSOC_REQ >> 4) |
drivers/target/iscsi/cxgbit/cxgbit.h
102
#define cxgbit_skcb_rx_opcode(skb) (CXGBIT_SKB_CB(skb)->rx.opcode)
drivers/target/iscsi/cxgbit/cxgbit.h
103
#define cxgbit_skcb_rx_backlog_fn(skb) (CXGBIT_SKB_CB(skb)->rx.backlog_fn)
drivers/target/iscsi/cxgbit/cxgbit.h
104
#define cxgbit_rx_pdu_cb(skb) (CXGBIT_SKB_CB(skb)->rx.pdu_cb)
drivers/target/iscsi/cxgbit/cxgbit.h
86
struct cxgbit_skb_rx_cb rx;
drivers/thermal/tegra/tegra-bpmp-thermal.c
156
msg.rx.data = &reply;
drivers/thermal/tegra/tegra-bpmp-thermal.c
157
msg.rx.size = sizeof(reply);
drivers/thermal/tegra/tegra-bpmp-thermal.c
162
if (msg.rx.ret)
drivers/thermal/tegra/tegra-bpmp-thermal.c
185
msg.rx.data = &reply;
drivers/thermal/tegra/tegra-bpmp-thermal.c
186
msg.rx.size = sizeof(reply);
drivers/thermal/tegra/tegra-bpmp-thermal.c
192
if (msg.rx.ret == 0) {
drivers/thermal/tegra/tegra-bpmp-thermal.c
195
} else if (msg.rx.ret == -BPMP_ENODEV) {
drivers/thermal/tegra/tegra-bpmp-thermal.c
49
msg.rx.data = &reply;
drivers/thermal/tegra/tegra-bpmp-thermal.c
50
msg.rx.size = sizeof(reply);
drivers/thermal/tegra/tegra-bpmp-thermal.c
55
if (msg.rx.ret == -BPMP_EFAULT)
drivers/thermal/tegra/tegra-bpmp-thermal.c
57
if (msg.rx.ret)
drivers/thermal/tegra/tegra-bpmp-thermal.c
94
if (msg.rx.ret)
drivers/thunderbolt/ctl.c
411
tb_ring_rx(pkg->ctl->rx, &pkg->frame); /*
drivers/thunderbolt/ctl.c
42
struct tb_ring *rx;
drivers/thunderbolt/ctl.c
461
goto rx;
drivers/thunderbolt/ctl.c
477
goto rx;
drivers/thunderbolt/ctl.c
482
goto rx;
drivers/thunderbolt/ctl.c
492
goto rx;
drivers/thunderbolt/ctl.c
497
goto rx;
drivers/thunderbolt/ctl.c
520
rx:
drivers/thunderbolt/ctl.c
678
ctl->rx = tb_ring_alloc_rx(nhi, 0, 10, RING_FLAG_NO_SUSPEND, 0, 0xffff,
drivers/thunderbolt/ctl.c
680
if (!ctl->rx)
drivers/thunderbolt/ctl.c
712
if (ctl->rx)
drivers/thunderbolt/ctl.c
713
tb_ring_free(ctl->rx);
drivers/thunderbolt/ctl.c
735
tb_ring_start(ctl->rx);
drivers/thunderbolt/ctl.c
757
tb_ring_stop(ctl->rx);
drivers/tty/amiserial.c
1090
icount->rx = cnow.rx;
drivers/tty/amiserial.c
1393
seq_printf(m, " tx:%d rx:%d", state->icount.tx, state->icount.rx);
drivers/tty/amiserial.c
199
icount->rx++;
drivers/tty/ipwireless/hardware.c
1106
int rx = 0;
drivers/tty/ipwireless/hardware.c
1157
rx = 1;
drivers/tty/ipwireless/hardware.c
1175
rx = 1;
drivers/tty/ipwireless/hardware.c
1191
if (tx || rx)
drivers/tty/mxser.c
1227
icount->rx = cnow.rx;
drivers/tty/nozomi.c
1700
icount->rx = cnow.rx;
drivers/tty/serial/21285.c
127
port->icount.rx++;
drivers/tty/serial/8250/8250_bcm7271.c
480
up->icount.rx += length;
drivers/tty/serial/8250/8250_dma.c
61
p->port.icount.rx += count;
drivers/tty/serial/8250/8250_mtk.c
125
up->port.icount.rx += copied;
drivers/tty/serial/8250/8250_omap.c
952
p->port.icount.rx += ret;
drivers/tty/serial/8250/8250_pci1xxxx.c
433
port->icount.rx += buff_index;
drivers/tty/serial/8250/8250_port.c
1603
port->icount.rx++;
drivers/tty/serial/altera_jtaguart.c
117
port->icount.rx++;
drivers/tty/serial/altera_uart.c
210
port->icount.rx++;
drivers/tty/serial/amba-pl010.c
123
port->icount.rx++;
drivers/tty/serial/amba-pl011.c
335
uap->port.icount.rx++;
drivers/tty/serial/amba-pl011.c
892
uap->port.icount.rx += dma_count;
drivers/tty/serial/apbuart.c
81
port->icount.rx++;
drivers/tty/serial/ar933x_uart.c
378
up->port.icount.rx++;
drivers/tty/serial/arc_uart.c
228
port->icount.rx++;
drivers/tty/serial/atmel_serial.c
1154
port->icount.rx += count;
drivers/tty/serial/atmel_serial.c
1166
port->icount.rx += count;
drivers/tty/serial/atmel_serial.c
1523
port->icount.rx++;
drivers/tty/serial/atmel_serial.c
1632
port->icount.rx += count;
drivers/tty/serial/bcm63xx_uart.c
261
port->icount.rx++;
drivers/tty/serial/clps711x.c
110
port->icount.rx++;
drivers/tty/serial/cpm_uart.c
276
port->icount.rx++;
drivers/tty/serial/digicolor-usart.c
147
port->icount.rx++;
drivers/tty/serial/dz.c
196
icount->rx++;
drivers/tty/serial/esp32_acm.c
116
u32 rx = esp32s3_acm_read(port, USB_SERIAL_JTAG_EP1_REG);
drivers/tty/serial/esp32_acm.c
118
++port->icount.rx;
drivers/tty/serial/esp32_acm.c
119
tty_insert_flip_char(tty_port, rx, TTY_NORMAL);
drivers/tty/serial/esp32_uart.c
227
u32 rx = esp32_uart_read(port, UART_FIFO_REG);
drivers/tty/serial/esp32_uart.c
229
if (!rx &&
drivers/tty/serial/esp32_uart.c
235
if (uart_handle_sysrq_char(port, (unsigned char)rx))
drivers/tty/serial/esp32_uart.c
237
tty_insert_flip_char(tty_port, rx, TTY_NORMAL);
drivers/tty/serial/esp32_uart.c
238
++port->icount.rx;
drivers/tty/serial/fsl_linflexuart.c
231
unsigned char rx;
drivers/tty/serial/fsl_linflexuart.c
238
rx = readb(sport->membase + BDRM);
drivers/tty/serial/fsl_linflexuart.c
241
sport->icount.rx++;
drivers/tty/serial/fsl_linflexuart.c
248
if (!rx) {
drivers/tty/serial/fsl_linflexuart.c
264
if (uart_handle_sysrq_char(sport, (unsigned char)rx))
drivers/tty/serial/fsl_linflexuart.c
266
tty_insert_flip_char(port, rx, flg);
drivers/tty/serial/fsl_lpuart.c
1027
rx &= 0x7F;
drivers/tty/serial/fsl_lpuart.c
1029
if (tty_insert_flip_char(port, rx, flg) == 0)
drivers/tty/serial/fsl_lpuart.c
1220
sport->port.icount.rx += copied;
drivers/tty/serial/fsl_lpuart.c
1234
sport->port.icount.rx += copied;
drivers/tty/serial/fsl_lpuart.c
887
u8 rx, sr;
drivers/tty/serial/fsl_lpuart.c
893
sport->port.icount.rx++;
drivers/tty/serial/fsl_lpuart.c
899
rx = readb(sport->port.membase + UARTDR);
drivers/tty/serial/fsl_lpuart.c
901
if (uart_prepare_sysrq_char(&sport->port, rx))
drivers/tty/serial/fsl_lpuart.c
932
if (tty_insert_flip_char(port, rx, flg) == 0)
drivers/tty/serial/fsl_lpuart.c
964
u32 rx, sr;
drivers/tty/serial/fsl_lpuart.c
971
sport->port.icount.rx++;
drivers/tty/serial/fsl_lpuart.c
977
rx = lpuart32_read(&sport->port, UARTDATA);
drivers/tty/serial/fsl_lpuart.c
978
rx &= UARTDATA_MASK;
drivers/tty/serial/fsl_lpuart.c
984
is_break = (sr & UARTSTAT_FE) && !rx;
drivers/tty/serial/fsl_lpuart.c
989
if (uart_prepare_sysrq_char(&sport->port, rx))
drivers/tty/serial/icom.c
1006
icount->rx += count;
drivers/tty/serial/imx.c
1256
sport->port.icount.rx += w_bytes;
drivers/tty/serial/imx.c
868
u32 usr2, rx;
drivers/tty/serial/imx.c
875
while ((rx = imx_uart_readl(sport, URXD0)) & URXD_CHARRDY) {
drivers/tty/serial/imx.c
877
sport->port.icount.rx++;
drivers/tty/serial/imx.c
879
if (unlikely(rx & URXD_ERR)) {
drivers/tty/serial/imx.c
880
if (rx & URXD_BRK) {
drivers/tty/serial/imx.c
885
else if (rx & URXD_PRERR)
drivers/tty/serial/imx.c
887
else if (rx & URXD_FRMERR)
drivers/tty/serial/imx.c
889
if (rx & URXD_OVRRUN)
drivers/tty/serial/imx.c
892
if (rx & sport->port.ignore_status_mask)
drivers/tty/serial/imx.c
895
rx &= (sport->port.read_status_mask | 0xFF);
drivers/tty/serial/imx.c
897
if (rx & URXD_BRK)
drivers/tty/serial/imx.c
899
else if (rx & URXD_PRERR)
drivers/tty/serial/imx.c
901
else if (rx & URXD_FRMERR)
drivers/tty/serial/imx.c
903
if (rx & URXD_OVRRUN)
drivers/tty/serial/imx.c
907
} else if (uart_handle_sysrq_char(&sport->port, (unsigned char)rx)) {
drivers/tty/serial/imx.c
914
if (tty_insert_flip_char(port, rx, flg) == 0)
drivers/tty/serial/ip22zilog.c
278
up->port.icount.rx++;
drivers/tty/serial/lantiq.c
184
port->icount.rx++;
drivers/tty/serial/liteuart.c
109
port->icount.rx++;
drivers/tty/serial/lpc32xx_hs.c
254
port->icount.rx++;
drivers/tty/serial/ma35d1_serial.c
241
up->port.icount.rx++;
drivers/tty/serial/max3100.c
157
static int max3100_sr(struct max3100_port *s, u16 tx, u16 *rx)
drivers/tty/serial/max3100.c
176
*rx = be16_to_cpu(erx);
drivers/tty/serial/max3100.c
177
s->tx_empty = (*rx & MAX3100_T) > 0;
drivers/tty/serial/max3100.c
178
dev_dbg(&s->spi->dev, "%04x - %04x\n", tx, *rx);
drivers/tty/serial/max3100.c
182
static int max3100_handlerx_unlocked(struct max3100_port *s, u16 rx)
drivers/tty/serial/max3100.c
188
if (rx & MAX3100_R && s->rx_enabled) {
drivers/tty/serial/max3100.c
190
ch = rx & (s->parity & MAX3100_7BIT ? 0x7f : 0xff);
drivers/tty/serial/max3100.c
191
if (rx & MAX3100_RAFE) {
drivers/tty/serial/max3100.c
197
if (max3100_check_parity(s, rx)) {
drivers/tty/serial/max3100.c
198
s->port.icount.rx++;
drivers/tty/serial/max3100.c
206
s->port.icount.rx++;
drivers/tty/serial/max3100.c
214
cts = (rx & MAX3100_CTS) > 0;
drivers/tty/serial/max3100.c
223
static int max3100_handlerx(struct max3100_port *s, u16 rx)
drivers/tty/serial/max3100.c
229
ret = max3100_handlerx_unlocked(s, rx);
drivers/tty/serial/max3100.c
241
u16 tx, rx;
drivers/tty/serial/max3100.c
257
max3100_sr(s, MAX3100_WC | conf, &rx);
drivers/tty/serial/max3100.c
259
max3100_sr(s, 0x4001, &rx);
drivers/tty/serial/max3100.c
262
(s->rts ? MAX3100_RTS : 0), &rx);
drivers/tty/serial/max3100.c
263
rxchars += max3100_handlerx(s, rx);
drivers/tty/serial/max3100.c
266
max3100_sr(s, MAX3100_RD, &rx);
drivers/tty/serial/max3100.c
267
rxchars += max3100_handlerx(s, rx);
drivers/tty/serial/max3100.c
269
if (rx & MAX3100_T) {
drivers/tty/serial/max3100.c
282
max3100_sr(s, tx, &rx);
drivers/tty/serial/max3100.c
283
rxchars += max3100_handlerx(s, rx);
drivers/tty/serial/max3100.c
296
((rx & MAX3100_R) ||
drivers/tty/serial/max3100.c
527
u16 rx;
drivers/tty/serial/max3100.c
546
max3100_sr(s, MAX3100_WC | MAX3100_SHDN, &rx);
drivers/tty/serial/max3100.c
689
u16 rx;
drivers/tty/serial/max3100.c
739
max3100_sr(max3100s[i], MAX3100_WC | MAX3100_SHDN, &rx);
drivers/tty/serial/max3100.c
779
u16 rx;
drivers/tty/serial/max3100.c
789
max3100_sr(s, MAX3100_WC | MAX3100_SHDN, &rx);
drivers/tty/serial/max310x.c
674
port->icount.rx += rxlen;
drivers/tty/serial/max310x.c
709
port->icount.rx++;
drivers/tty/serial/mcf.c
289
port->icount.rx++;
drivers/tty/serial/men_z135_uart.c
282
port->icount.rx += copied;
drivers/tty/serial/meson_uart.c
185
port->icount.rx++;
drivers/tty/serial/milbeaut_usio.c
166
port->icount.rx++;
drivers/tty/serial/mpc52xx_uart.c
1385
port->icount.rx++;
drivers/tty/serial/mps2-uart.c
176
port->icount.rx++;
drivers/tty/serial/mps2-uart.c
294
MAKE_NAME(-rx), mps_port);
drivers/tty/serial/msm_serial.c
169
} rx;
drivers/tty/serial/msm_serial.c
260
mapped = dma->rx.count;
drivers/tty/serial/msm_serial.c
261
dma->rx.count = 0;
drivers/tty/serial/msm_serial.c
282
dma_unmap_single(dev, dma->rx.phys, mapped, dma->dir);
drivers/tty/serial/msm_serial.c
302
kfree(dma->rx.virt);
drivers/tty/serial/msm_serial.c
374
dma->rx.virt = kzalloc(UARTDM_RX_SIZE, GFP_KERNEL);
drivers/tty/serial/msm_serial.c
375
if (!dma->rx.virt)
drivers/tty/serial/msm_serial.c
402
kfree(dma->rx.virt);
drivers/tty/serial/msm_serial.c
572
if (!dma->rx.count)
drivers/tty/serial/msm_serial.c
587
port->icount.rx += count;
drivers/tty/serial/msm_serial.c
589
dma->rx.count = 0;
drivers/tty/serial/msm_serial.c
591
dma_unmap_single(port->dev, dma->rx.phys, UARTDM_RX_SIZE, dma->dir);
drivers/tty/serial/msm_serial.c
596
if (msm_port->break_detected && dma->rx.virt[i] == 0) {
drivers/tty/serial/msm_serial.c
607
sysrq = uart_prepare_sysrq_char(port, dma->rx.virt[i]);
drivers/tty/serial/msm_serial.c
609
tty_insert_flip_char(tport, dma->rx.virt[i], flag);
drivers/tty/serial/msm_serial.c
633
dma->rx.phys = dma_map_single(uart->dev, dma->rx.virt,
drivers/tty/serial/msm_serial.c
635
ret = dma_mapping_error(uart->dev, dma->rx.phys);
drivers/tty/serial/msm_serial.c
639
dma->desc = dmaengine_prep_slave_single(dma->chan, dma->rx.phys,
drivers/tty/serial/msm_serial.c
667
dma->rx.count = UARTDM_RX_SIZE;
drivers/tty/serial/msm_serial.c
687
dma_unmap_single(uart->dev, dma->rx.phys, UARTDM_RX_SIZE, dma->dir);
drivers/tty/serial/msm_serial.c
751
port->icount.rx += count;
drivers/tty/serial/msm_serial.c
829
port->icount.rx++;
drivers/tty/serial/msm_serial.c
970
if (dma->rx.count) {
drivers/tty/serial/mux.c
210
__u32 start_count = port->icount.rx;
drivers/tty/serial/mux.c
221
port->icount.rx++;
drivers/tty/serial/mux.c
235
if (start_count != port->icount.rx)
drivers/tty/serial/mvebu-uart.c
272
port->icount.rx++;
drivers/tty/serial/mxs-auart.c
618
s->port.icount.rx++;
drivers/tty/serial/omap-serial.c
454
up->port.icount.rx++;
drivers/tty/serial/omap-serial.c
509
up->port.icount.rx++;
drivers/tty/serial/owl-uart.c
216
port->icount.rx++;
drivers/tty/serial/pch_uart.c
625
port->icount.rx += room;
drivers/tty/serial/pic32_uart.c
303
port->icount.rx++;
drivers/tty/serial/pmac_zilog.c
261
uap->port.icount.rx++;
drivers/tty/serial/pxa.c
109
up->port.icount.rx++;
drivers/tty/serial/qcom_geni_serial.c
595
uport->icount.rx++;
drivers/tty/serial/qcom_geni_serial.c
629
uport->icount.rx += ret;
drivers/tty/serial/rda-uart.c
391
port->icount.rx++;
drivers/tty/serial/rp2.c
421
up->port.icount.rx++;
drivers/tty/serial/rsci.c
534
port->icount.rx += count;
drivers/tty/serial/sa1100.c
191
sport->port.icount.rx++;
drivers/tty/serial/samsung_tty.c
523
ourport->port.icount.rx += count;
drivers/tty/serial/samsung_tty.c
807
port->icount.rx++;
drivers/tty/serial/sb1250-duart.c
348
icount->rx++;
drivers/tty/serial/sc16is7xx.c
608
port->icount.rx++;
drivers/tty/serial/sccnxp.c
396
port->icount.rx++;
drivers/tty/serial/serial-tegra.c
662
tup->uport.icount.rx++;
drivers/tty/serial/serial-tegra.c
684
tup->uport.icount.rx += count;
drivers/tty/serial/serial_core.c
1280
icount->rx = cnow.rx;
drivers/tty/serial/serial_core.c
2009
uport->icount.tx, uport->icount.rx);
drivers/tty/serial/serial_txx9.c
258
up->icount.rx++;
drivers/tty/serial/sh-sci.c
1132
port->icount.rx += count;
drivers/tty/serial/sh-sci.c
1458
port->icount.rx += copied;
drivers/tty/serial/sifive.c
416
ssp->port.icount.rx++;
drivers/tty/serial/sprd_serial.c
412
port->icount.rx += sp->rx_dma.trans_len;
drivers/tty/serial/sprd_serial.c
632
port->icount.rx++;
drivers/tty/serial/st-asc.c
271
port->icount.rx++;
drivers/tty/serial/stm32-usart.c
409
port->icount.rx++;
drivers/tty/serial/stm32-usart.c
466
port->icount.rx += dma_count;
drivers/tty/serial/sunhv.c
106
port->icount.rx++;
drivers/tty/serial/sunhv.c
163
port->icount.rx += bytes_read;
drivers/tty/serial/sunplus-uart.c
241
port->icount.rx++;
drivers/tty/serial/sunsab.c
171
up->port.icount.rx++;
drivers/tty/serial/sunsu.c
307
up->port.icount.rx++;
drivers/tty/serial/sunzilog.c
366
up->port.icount.rx++;
drivers/tty/serial/tegra-tcu.c
247
tcu->rx = mbox_request_channel_byname(&tcu->rx_client, "rx");
drivers/tty/serial/tegra-tcu.c
248
if (IS_ERR(tcu->rx)) {
drivers/tty/serial/tegra-tcu.c
249
err = PTR_ERR(tcu->rx);
drivers/tty/serial/tegra-tcu.c
278
mbox_free_channel(tcu->rx);
drivers/tty/serial/tegra-tcu.c
30
struct mbox_chan *tx, *rx;
drivers/tty/serial/tegra-utc.c
185
tup->port.icount.rx++;
drivers/tty/serial/timbuart.c
86
port->icount.rx++;
drivers/tty/serial/timbuart.c
93
__func__, port->icount.rx);
drivers/tty/serial/uartlite.c
152
port->icount.rx++;
drivers/tty/serial/ucc_uart.c
489
port->icount.rx++;
drivers/tty/serial/vt8500_serial.c
181
port->icount.rx++;
drivers/tty/serial/xilinx_uartps.c
253
port->icount.rx++;
drivers/tty/serial/zs.c
561
icount->rx++;
drivers/tty/synclink_gt.c
1048
icount->rx = cnow.rx;
drivers/tty/synclink_gt.c
1212
info->icount.tx, info->icount.rx);
drivers/tty/synclink_gt.c
1778
icount->rx++;
drivers/usb/atm/usbatm.c
403
atomic_inc(&vcc->stats->rx);
drivers/usb/atm/usbatm.c
749
atomic_read(&atm_dev->stats.aal5.rx),
drivers/usb/gadget/function/f_phonet.c
325
spin_lock_irqsave(&fp->rx.lock, flags);
drivers/usb/gadget/function/f_phonet.c
326
skb = fp->rx.skb;
drivers/usb/gadget/function/f_phonet.c
328
skb = fp->rx.skb = netdev_alloc_skb(dev, 12);
drivers/usb/gadget/function/f_phonet.c
330
fp->rx.skb = NULL;
drivers/usb/gadget/function/f_phonet.c
331
spin_unlock_irqrestore(&fp->rx.lock, flags);
drivers/usb/gadget/function/f_phonet.c
391
if (fp->rx.skb) {
drivers/usb/gadget/function/f_phonet.c
392
dev_kfree_skb_irq(fp->rx.skb);
drivers/usb/gadget/function/f_phonet.c
393
fp->rx.skb = NULL;
drivers/usb/gadget/function/f_phonet.c
46
} rx;
drivers/usb/gadget/function/f_phonet.c
686
spin_lock_init(&fp->rx.lock);
drivers/usb/gadget/function/u_serial.c
415
port->icount.rx += size;
drivers/usb/gadget/function/u_serial.c
905
icount->rx = cnow.rx;
drivers/usb/host/max3421-hcd.c
141
struct max3421_dma_buf *rx;
drivers/usb/host/max3421-hcd.c
1884
max3421_hcd->rx = kmalloc_obj(*max3421_hcd->rx);
drivers/usb/host/max3421-hcd.c
1885
if (!max3421_hcd->rx)
drivers/usb/host/max3421-hcd.c
1918
kfree(max3421_hcd->rx);
drivers/usb/host/max3421-hcd.c
357
transfer.rx_buf = max3421_hcd->rx->data;
drivers/usb/host/max3421-hcd.c
363
return max3421_hcd->rx->data[1];
drivers/usb/musb/cppi_dma.h
117
struct cppi_channel rx[4];
drivers/usb/serial/ftdi_sio.c
2483
port->icount.rx += len - 2;
drivers/usb/serial/generic.c
562
icount->rx = cnow.rx;
drivers/usb/serial/io_edgeport.c
1721
edge_port->port->icount.rx += rxLen;
drivers/usb/serial/io_ti.c
1774
edge_port->port->icount.rx += urb->actual_length;
drivers/usb/serial/mos7840.c
440
port->icount.rx += urb->actual_length;
drivers/usb/serial/mos7840.c
441
dev_dbg(&port->dev, "icount.rx is %d:\n", port->icount.rx);
drivers/usb/serial/ti_usb_3410_5052.c
1219
port->icount.rx += urb->actual_length;
drivers/video/backlight/ili922x.c
191
static int ili922x_read(struct spi_device *spi, u8 reg, u16 *rx)
drivers/video/backlight/ili922x.c
235
*rx = (rbuf[1 + send_bytes] << 8) + rbuf[2 + send_bytes];
drivers/video/backlight/ili922x.c
307
u16 rx;
drivers/video/backlight/ili922x.c
312
ili922x_read(spi, reg, &rx);
drivers/video/backlight/ili922x.c
313
dev_dbg(&spi->dev, "reg @ 0x%02X: 0x%04X\n", reg, rx);
drivers/video/fbdev/mmp/hw/mmp_ctrl.h
386
#define CFG_RXBITS(rx) (((rx) - 1)<<16) /* 0x1F~0x1 */
drivers/video/fbdev/mmp/hw/mmp_ctrl.h
394
#define CFG_RXBITSTO0(rx) ((rx)<<5)
drivers/video/fbdev/pxa168fb.h
177
#define CFG_RXBITS(rx) ((rx) << 16) /* 0x1F~0x1 */
drivers/video/fbdev/pxa168fb.h
185
#define CFG_RXBITSTO0(rx) ((rx) << 5)
fs/jffs2/scan.c
333
struct jffs2_raw_xattr *rx, uint32_t ofs,
fs/jffs2/scan.c
340
crc = crc32(0, rx, sizeof(struct jffs2_raw_xattr) - 4);
fs/jffs2/scan.c
341
if (crc != je32_to_cpu(rx->node_crc)) {
fs/jffs2/scan.c
343
ofs, je32_to_cpu(rx->node_crc), crc);
fs/jffs2/scan.c
344
if ((err = jffs2_scan_dirty_space(c, jeb, je32_to_cpu(rx->totlen))))
fs/jffs2/scan.c
349
xid = je32_to_cpu(rx->xid);
fs/jffs2/scan.c
350
version = je32_to_cpu(rx->version);
fs/jffs2/scan.c
353
+ rx->name_len + 1 + je16_to_cpu(rx->value_len));
fs/jffs2/scan.c
354
if (totlen != je32_to_cpu(rx->totlen)) {
fs/jffs2/scan.c
356
ofs, je32_to_cpu(rx->totlen), totlen);
fs/jffs2/scan.c
357
if ((err = jffs2_scan_dirty_space(c, jeb, je32_to_cpu(rx->totlen))))
fs/jffs2/scan.c
373
xd->xprefix = rx->xprefix;
fs/jffs2/scan.c
374
xd->name_len = rx->name_len;
fs/jffs2/scan.c
375
xd->value_len = je16_to_cpu(rx->value_len);
fs/jffs2/scan.c
376
xd->data_crc = je32_to_cpu(rx->data_crc);
fs/jffs2/scan.c
382
jffs2_sum_add_xattr_mem(s, rx, ofs - jeb->offset);
fs/jffs2/summary.c
158
int jffs2_sum_add_xattr_mem(struct jffs2_summary *s, struct jffs2_raw_xattr *rx, uint32_t ofs)
fs/jffs2/summary.c
166
temp->nodetype = rx->nodetype;
fs/jffs2/summary.c
167
temp->xid = rx->xid;
fs/jffs2/summary.c
168
temp->version = rx->version;
fs/jffs2/summary.c
170
temp->totlen = rx->totlen;
fs/jffs2/summary.h
187
int jffs2_sum_add_xattr_mem(struct jffs2_summary *s, struct jffs2_raw_xattr *rx, uint32_t ofs);
fs/jffs2/xattr.c
137
struct jffs2_raw_xattr rx;
fs/jffs2/xattr.c
148
rc = jffs2_flash_read(c, offset, sizeof(rx), &readlen, (char *)&rx);
fs/jffs2/xattr.c
149
if (rc || readlen != sizeof(rx)) {
fs/jffs2/xattr.c
151
rc, sizeof(rx), readlen, offset);
fs/jffs2/xattr.c
154
crc = crc32(0, &rx, sizeof(rx) - 4);
fs/jffs2/xattr.c
155
if (crc != je32_to_cpu(rx.node_crc)) {
fs/jffs2/xattr.c
157
offset, je32_to_cpu(rx.hdr_crc), crc);
fs/jffs2/xattr.c
161
totlen = PAD(sizeof(rx) + rx.name_len + 1 + je16_to_cpu(rx.value_len));
fs/jffs2/xattr.c
162
if (je16_to_cpu(rx.magic) != JFFS2_MAGIC_BITMASK
fs/jffs2/xattr.c
163
|| je16_to_cpu(rx.nodetype) != JFFS2_NODETYPE_XATTR
fs/jffs2/xattr.c
164
|| je32_to_cpu(rx.totlen) != totlen
fs/jffs2/xattr.c
165
|| je32_to_cpu(rx.xid) != xd->xid
fs/jffs2/xattr.c
166
|| je32_to_cpu(rx.version) != xd->version) {
fs/jffs2/xattr.c
169
offset, je16_to_cpu(rx.magic), JFFS2_MAGIC_BITMASK,
fs/jffs2/xattr.c
170
je16_to_cpu(rx.nodetype), JFFS2_NODETYPE_XATTR,
fs/jffs2/xattr.c
171
je32_to_cpu(rx.totlen), totlen,
fs/jffs2/xattr.c
172
je32_to_cpu(rx.xid), xd->xid,
fs/jffs2/xattr.c
173
je32_to_cpu(rx.version), xd->version);
fs/jffs2/xattr.c
177
xd->xprefix = rx.xprefix;
fs/jffs2/xattr.c
178
xd->name_len = rx.name_len;
fs/jffs2/xattr.c
179
xd->value_len = je16_to_cpu(rx.value_len);
fs/jffs2/xattr.c
180
xd->data_crc = je32_to_cpu(rx.data_crc);
fs/jffs2/xattr.c
287
struct jffs2_raw_xattr rx;
fs/jffs2/xattr.c
296
vecs[0].iov_base = ℞
fs/jffs2/xattr.c
297
vecs[0].iov_len = sizeof(rx);
fs/jffs2/xattr.c
303
memset(&rx, 0, sizeof(rx));
fs/jffs2/xattr.c
304
rx.magic = cpu_to_je16(JFFS2_MAGIC_BITMASK);
fs/jffs2/xattr.c
305
rx.nodetype = cpu_to_je16(JFFS2_NODETYPE_XATTR);
fs/jffs2/xattr.c
306
rx.totlen = cpu_to_je32(PAD(totlen));
fs/jffs2/xattr.c
307
rx.hdr_crc = cpu_to_je32(crc32(0, &rx, sizeof(struct jffs2_unknown_node) - 4));
fs/jffs2/xattr.c
309
rx.xid = cpu_to_je32(xd->xid);
fs/jffs2/xattr.c
310
rx.version = cpu_to_je32(++xd->version);
fs/jffs2/xattr.c
311
rx.xprefix = xd->xprefix;
fs/jffs2/xattr.c
312
rx.name_len = xd->name_len;
fs/jffs2/xattr.c
313
rx.value_len = cpu_to_je16(xd->value_len);
fs/jffs2/xattr.c
314
rx.data_crc = cpu_to_je32(crc32(0, vecs[1].iov_base, vecs[1].iov_len));
fs/jffs2/xattr.c
315
rx.node_crc = cpu_to_je32(crc32(0, &rx, sizeof(struct jffs2_raw_xattr) - 4));
fs/xfs/scrub/attr_repair.c
1062
struct xrep_xattr *rx)
fs/xfs/scrub/attr_repair.c
1064
struct xfs_inode *ip = rx->sc->ip;
fs/xfs/scrub/attr_repair.c
1068
if (rx->sc->ip->i_af.if_format == XFS_DINODE_FMT_LOCAL) {
fs/xfs/scrub/attr_repair.c
1069
error = xrep_xattr_recover_sf(rx);
fs/xfs/scrub/attr_repair.c
1073
return xrep_xattr_flush_stashed(rx);
fs/xfs/scrub/attr_repair.c
1084
error = xfs_trans_roll(&rx->sc->tp);
fs/xfs/scrub/attr_repair.c
1088
error = xfs_iread_extents(rx->sc->tp, ip, XFS_ATTR_FORK);
fs/xfs/scrub/attr_repair.c
1092
error = xrep_xattr_recover(rx);
fs/xfs/scrub/attr_repair.c
1096
return xrep_xattr_flush_stashed(rx);
fs/xfs/scrub/attr_repair.c
1106
struct xrep_xattr *rx,
fs/xfs/scrub/attr_repair.c
1110
struct xfs_scrub *sc = rx->sc;
fs/xfs/scrub/attr_repair.c
1120
&pptr->pptr_rec, &rx->pptr_args);
fs/xfs/scrub/attr_repair.c
1129
&pptr->pptr_rec, &rx->pptr_args);
fs/xfs/scrub/attr_repair.c
1148
struct xrep_xattr *rx)
fs/xfs/scrub/attr_repair.c
1153
mutex_lock(&rx->lock);
fs/xfs/scrub/attr_repair.c
1154
foreach_xfarray_idx(rx->pptr_recs, array_cur) {
fs/xfs/scrub/attr_repair.c
1157
error = xfarray_load(rx->pptr_recs, array_cur, &pptr);
fs/xfs/scrub/attr_repair.c
1161
error = xfblob_loadname(rx->pptr_names, pptr.name_cookie,
fs/xfs/scrub/attr_repair.c
1162
&rx->xname, pptr.namelen);
fs/xfs/scrub/attr_repair.c
1165
mutex_unlock(&rx->lock);
fs/xfs/scrub/attr_repair.c
1167
error = xrep_xattr_replay_pptr_update(rx, &rx->xname, &pptr);
fs/xfs/scrub/attr_repair.c
1171
mutex_lock(&rx->lock);
fs/xfs/scrub/attr_repair.c
1175
xfarray_truncate(rx->pptr_recs);
fs/xfs/scrub/attr_repair.c
1176
xfblob_truncate(rx->pptr_names);
fs/xfs/scrub/attr_repair.c
1177
mutex_unlock(&rx->lock);
fs/xfs/scrub/attr_repair.c
1180
mutex_unlock(&rx->lock);
fs/xfs/scrub/attr_repair.c
1190
struct xrep_xattr *rx,
fs/xfs/scrub/attr_repair.c
1200
trace_xrep_xattr_stash_parentadd(rx->sc->tempip, dp, name);
fs/xfs/scrub/attr_repair.c
1203
error = xfblob_storename(rx->pptr_names, &pptr.name_cookie, name);
fs/xfs/scrub/attr_repair.c
1207
return xfarray_append(rx->pptr_recs, &pptr);
fs/xfs/scrub/attr_repair.c
1216
struct xrep_xattr *rx,
fs/xfs/scrub/attr_repair.c
1226
trace_xrep_xattr_stash_parentremove(rx->sc->tempip, dp, name);
fs/xfs/scrub/attr_repair.c
1229
error = xfblob_storename(rx->pptr_names, &pptr.name_cookie, name);
fs/xfs/scrub/attr_repair.c
1233
return xfarray_append(rx->pptr_recs, &pptr);
fs/xfs/scrub/attr_repair.c
1247
struct xrep_xattr *rx;
fs/xfs/scrub/attr_repair.c
1251
rx = container_of(nb, struct xrep_xattr, dhook.dirent_hook.nb);
fs/xfs/scrub/attr_repair.c
1252
sc = rx->sc;
fs/xfs/scrub/attr_repair.c
1262
mutex_lock(&rx->lock);
fs/xfs/scrub/attr_repair.c
1264
error = xrep_xattr_stash_parentadd(rx, p->name, p->dp);
fs/xfs/scrub/attr_repair.c
1266
error = xrep_xattr_stash_parentremove(rx, p->name, p->dp);
fs/xfs/scrub/attr_repair.c
1268
rx->live_update_aborted = true;
fs/xfs/scrub/attr_repair.c
1269
mutex_unlock(&rx->lock);
fs/xfs/scrub/attr_repair.c
1383
struct xrep_xattr *rx)
fs/xfs/scrub/attr_repair.c
1385
struct xfs_scrub *sc = rx->sc;
fs/xfs/scrub/attr_repair.c
1389
return xrep_tempexch_trans_alloc(sc, XFS_ATTR_FORK, &rx->tx);
fs/xfs/scrub/attr_repair.c
1398
error = xrep_xattr_replay_pptr_updates(rx);
fs/xfs/scrub/attr_repair.c
1402
error = xrep_tempexch_trans_alloc(sc, XFS_ATTR_FORK, &rx->tx);
fs/xfs/scrub/attr_repair.c
1406
if (xfarray_length(rx->pptr_recs) == 0)
fs/xfs/scrub/attr_repair.c
1421
struct xrep_xattr *rx)
fs/xfs/scrub/attr_repair.c
1423
struct xfs_scrub *sc = rx->sc;
fs/xfs/scrub/attr_repair.c
1430
if (rx->attrs_found == 0) {
fs/xfs/scrub/attr_repair.c
1463
error = xrep_tempfile_iolock_polled(rx->sc);
fs/xfs/scrub/attr_repair.c
1472
error = xrep_xattr_finalize_tempfile(rx);
fs/xfs/scrub/attr_repair.c
1481
error = xrep_xattr_swap(sc, &rx->tx);
fs/xfs/scrub/attr_repair.c
1511
struct xrep_xattr *rx)
fs/xfs/scrub/attr_repair.c
1513
if (xfs_has_parent(rx->sc->mp))
fs/xfs/scrub/attr_repair.c
1514
xfs_dir_hook_del(rx->sc->mp, &rx->dhook);
fs/xfs/scrub/attr_repair.c
1515
if (rx->pptr_names)
fs/xfs/scrub/attr_repair.c
1516
xfblob_destroy(rx->pptr_names);
fs/xfs/scrub/attr_repair.c
1517
if (rx->pptr_recs)
fs/xfs/scrub/attr_repair.c
1518
xfarray_destroy(rx->pptr_recs);
fs/xfs/scrub/attr_repair.c
1519
if (rx->xattr_blobs)
fs/xfs/scrub/attr_repair.c
1520
xfblob_destroy(rx->xattr_blobs);
fs/xfs/scrub/attr_repair.c
1521
if (rx->xattr_records)
fs/xfs/scrub/attr_repair.c
1522
xfarray_destroy(rx->xattr_records);
fs/xfs/scrub/attr_repair.c
1523
mutex_destroy(&rx->lock);
fs/xfs/scrub/attr_repair.c
1524
kfree(rx);
fs/xfs/scrub/attr_repair.c
1533
struct xrep_xattr *rx;
fs/xfs/scrub/attr_repair.c
1537
rx = kzalloc_obj(struct xrep_xattr, XCHK_GFP_FLAGS);
fs/xfs/scrub/attr_repair.c
1538
if (!rx)
fs/xfs/scrub/attr_repair.c
1540
rx->sc = sc;
fs/xfs/scrub/attr_repair.c
1541
rx->can_flush = true;
fs/xfs/scrub/attr_repair.c
1542
rx->xname.name = rx->namebuf;
fs/xfs/scrub/attr_repair.c
1544
mutex_init(&rx->lock);
fs/xfs/scrub/attr_repair.c
1552
error = xchk_setup_xattr_buf(rx->sc, max_len);
fs/xfs/scrub/attr_repair.c
1560
&rx->xattr_records);
fs/xfs/scrub/attr_repair.c
1564
error = xfblob_create("xattr names", &rx->xattr_blobs);
fs/xfs/scrub/attr_repair.c
1573
&rx->pptr_recs);
fs/xfs/scrub/attr_repair.c
1578
&rx->pptr_names);
fs/xfs/scrub/attr_repair.c
1582
xfs_dir_hook_setup(&rx->dhook, xrep_xattr_live_dirent_update);
fs/xfs/scrub/attr_repair.c
1583
error = xfs_dir_hook_add(sc->mp, &rx->dhook);
fs/xfs/scrub/attr_repair.c
1588
*rxp = rx;
fs/xfs/scrub/attr_repair.c
1591
xfblob_destroy(rx->pptr_names);
fs/xfs/scrub/attr_repair.c
1593
xfarray_destroy(rx->pptr_recs);
fs/xfs/scrub/attr_repair.c
1595
xfblob_destroy(rx->xattr_blobs);
fs/xfs/scrub/attr_repair.c
1597
xfarray_destroy(rx->xattr_records);
fs/xfs/scrub/attr_repair.c
1599
mutex_destroy(&rx->lock);
fs/xfs/scrub/attr_repair.c
1600
kfree(rx);
fs/xfs/scrub/attr_repair.c
1616
struct xrep_xattr *rx = NULL;
fs/xfs/scrub/attr_repair.c
1629
error = xrep_xattr_setup_scan(sc, &rx);
fs/xfs/scrub/attr_repair.c
1635
error = xrep_xattr_salvage_attributes(rx);
fs/xfs/scrub/attr_repair.c
1639
if (rx->live_update_aborted) {
fs/xfs/scrub/attr_repair.c
1648
error = xrep_xattr_rebuild_tree(rx);
fs/xfs/scrub/attr_repair.c
165
struct xrep_xattr *rx,
fs/xfs/scrub/attr_repair.c
1653
xrep_xattr_teardown(rx);
fs/xfs/scrub/attr_repair.c
181
return xfs_parent_valuecheck(rx->sc->mp, value, valuelen);
fs/xfs/scrub/attr_repair.c
189
struct xrep_xattr *rx,
fs/xfs/scrub/attr_repair.c
203
if (xchk_should_terminate(rx->sc, &error))
fs/xfs/scrub/attr_repair.c
213
trace_xrep_xattr_salvage_pptr(rx->sc->ip, flags, name,
fs/xfs/scrub/attr_repair.c
222
trace_xrep_xattr_salvage_rec(rx->sc->ip, flags, name,
fs/xfs/scrub/attr_repair.c
226
error = xfblob_store(rx->xattr_blobs, &key.name_cookie, name,
fs/xfs/scrub/attr_repair.c
231
error = xfblob_store(rx->xattr_blobs, &key.value_cookie, value,
fs/xfs/scrub/attr_repair.c
236
error = xfarray_append(rx->xattr_records, &key);
fs/xfs/scrub/attr_repair.c
240
rx->attrs_found++;
fs/xfs/scrub/attr_repair.c
250
struct xrep_xattr *rx,
fs/xfs/scrub/attr_repair.c
254
struct xfs_scrub *sc = rx->sc;
fs/xfs/scrub/attr_repair.c
267
if (!xrep_xattr_want_salvage(rx, sfe->flags, sfe->nameval,
fs/xfs/scrub/attr_repair.c
271
return xrep_xattr_salvage_key(rx, sfe->flags, sfe->nameval,
fs/xfs/scrub/attr_repair.c
281
struct xrep_xattr *rx,
fs/xfs/scrub/attr_repair.c
287
struct xchk_xattr_buf *ab = rx->sc->buf;
fs/xfs/scrub/attr_repair.c
301
if (!xrep_xattr_want_salvage(rx, ent->flags, lentry->nameval,
fs/xfs/scrub/attr_repair.c
304
if (!xchk_xattr_set_map(rx->sc, ab->usedmap, nameidx, namesize))
fs/xfs/scrub/attr_repair.c
308
return xrep_xattr_salvage_key(rx, ent->flags, lentry->nameval,
fs/xfs/scrub/attr_repair.c
318
struct xrep_xattr *rx,
fs/xfs/scrub/attr_repair.c
326
struct xchk_xattr_buf *ab = rx->sc->buf;
fs/xfs/scrub/attr_repair.c
328
.trans = rx->sc->tp,
fs/xfs/scrub/attr_repair.c
329
.dp = rx->sc->ip,
fs/xfs/scrub/attr_repair.c
331
.geo = rx->sc->mp->m_attr_geo,
fs/xfs/scrub/attr_repair.c
332
.owner = rx->sc->ip->i_ino,
fs/xfs/scrub/attr_repair.c
350
!xrep_xattr_want_salvage(rx, ent->flags, rentry->name,
fs/xfs/scrub/attr_repair.c
353
if (!xchk_xattr_set_map(rx->sc, ab->usedmap, nameidx, namesize))
fs/xfs/scrub/attr_repair.c
360
error = xchk_setup_xattr_buf(rx->sc, args.valuelen);
fs/xfs/scrub/attr_repair.c
377
error = xrep_xattr_salvage_key(rx, ent->flags, rentry->name,
fs/xfs/scrub/attr_repair.c
389
struct xrep_xattr *rx,
fs/xfs/scrub/attr_repair.c
393
struct xfs_scrub *sc = rx->sc;
fs/xfs/scrub/attr_repair.c
400
struct xchk_xattr_buf *ab = rx->sc->buf;
fs/xfs/scrub/attr_repair.c
436
error = xrep_xattr_salvage_local_attr(rx, ent, nameidx,
fs/xfs/scrub/attr_repair.c
440
error = xrep_xattr_salvage_remote_attr(rx, ent, nameidx,
fs/xfs/scrub/attr_repair.c
453
struct xrep_xattr *rx)
fs/xfs/scrub/attr_repair.c
455
struct xfs_scrub *sc = rx->sc;
fs/xfs/scrub/attr_repair.c
465
ifp = xfs_ifork_ptr(rx->sc->ip, XFS_ATTR_FORK);
fs/xfs/scrub/attr_repair.c
491
error = xrep_xattr_salvage_sf_attr(rx, hdr, sfe);
fs/xfs/scrub/attr_repair.c
568
struct xrep_xattr *rx,
fs/xfs/scrub/attr_repair.c
578
error = xrep_xattr_find_buf(rx->sc->mp, fsbno, max_len, true, &bp);
fs/xfs/scrub/attr_repair.c
582
*actual_len = XFS_BB_TO_FSB(rx->sc->mp, bp->b_length);
fs/xfs/scrub/attr_repair.c
584
trace_xrep_xattr_recover_leafblock(rx->sc->ip, dabno,
fs/xfs/scrub/attr_repair.c
593
xfs_attr3_leaf_header_check(bp, rx->sc->ip->i_ino) == NULL)
fs/xfs/scrub/attr_repair.c
594
error = xrep_xattr_recover_leaf(rx, bp);
fs/xfs/scrub/attr_repair.c
612
struct xrep_xattr *rx,
fs/xfs/scrub/attr_repair.c
616
.dp = rx->sc->tempip,
fs/xfs/scrub/attr_repair.c
620
.owner = rx->sc->ip->i_ino,
fs/xfs/scrub/attr_repair.c
621
.geo = rx->sc->mp->m_attr_geo,
fs/xfs/scrub/attr_repair.c
625
struct xchk_xattr_buf *ab = rx->sc->buf;
fs/xfs/scrub/attr_repair.c
641
error = xfblob_load(rx->xattr_blobs, key->name_cookie, ab->name,
fs/xfs/scrub/attr_repair.c
646
error = xfblob_free(rx->xattr_blobs, key->name_cookie);
fs/xfs/scrub/attr_repair.c
650
error = xfblob_load(rx->xattr_blobs, key->value_cookie, args.value,
fs/xfs/scrub/attr_repair.c
655
error = xfblob_free(rx->xattr_blobs, key->value_cookie);
fs/xfs/scrub/attr_repair.c
662
trace_xrep_xattr_insert_pptr(rx->sc->tempip, key->flags,
fs/xfs/scrub/attr_repair.c
667
trace_xrep_xattr_insert_rec(rx->sc->tempip, key->flags,
fs/xfs/scrub/attr_repair.c
690
struct xrep_xattr *rx)
fs/xfs/scrub/attr_repair.c
714
error = xrep_trans_commit(rx->sc);
fs/xfs/scrub/attr_repair.c
717
xchk_iunlock(rx->sc, XFS_ILOCK_EXCL);
fs/xfs/scrub/attr_repair.c
725
error = xrep_tempfile_iolock_polled(rx->sc);
fs/xfs/scrub/attr_repair.c
730
foreach_xfarray_idx(rx->xattr_records, array_cur) {
fs/xfs/scrub/attr_repair.c
733
error = xfarray_load(rx->xattr_records, array_cur, &key);
fs/xfs/scrub/attr_repair.c
737
error = xrep_xattr_insert_rec(rx, &key);
fs/xfs/scrub/attr_repair.c
743
xfarray_truncate(rx->xattr_records);
fs/xfs/scrub/attr_repair.c
744
xfblob_truncate(rx->xattr_blobs);
fs/xfs/scrub/attr_repair.c
746
xrep_tempfile_iounlock(rx->sc);
fs/xfs/scrub/attr_repair.c
749
error = xchk_trans_alloc(rx->sc, 0);
fs/xfs/scrub/attr_repair.c
752
xchk_ilock(rx->sc, XFS_ILOCK_EXCL);
fs/xfs/scrub/attr_repair.c
759
struct xrep_xattr *rx)
fs/xfs/scrub/attr_repair.c
763
if (!rx->can_flush)
fs/xfs/scrub/attr_repair.c
766
bytes = xfarray_bytes(rx->xattr_records) +
fs/xfs/scrub/attr_repair.c
767
xfblob_bytes(rx->xattr_blobs);
fs/xfs/scrub/attr_repair.c
777
struct xrep_xattr *rx)
fs/xfs/scrub/attr_repair.c
781
ASSERT(rx->can_flush);
fs/xfs/scrub/attr_repair.c
783
if (!xfs_has_parent(rx->sc->mp))
fs/xfs/scrub/attr_repair.c
786
xfs_assert_ilocked(rx->sc->ip, XFS_ILOCK_EXCL);
fs/xfs/scrub/attr_repair.c
788
mutex_lock(&rx->lock);
fs/xfs/scrub/attr_repair.c
789
ret = xfarray_bytes(rx->pptr_recs) > 0;
fs/xfs/scrub/attr_repair.c
790
mutex_unlock(&rx->lock);
fs/xfs/scrub/attr_repair.c
803
struct xrep_xattr *rx)
fs/xfs/scrub/attr_repair.c
805
struct xfs_scrub *sc = rx->sc;
fs/xfs/scrub/attr_repair.c
824
xchk_iunlock(rx->sc, XFS_ILOCK_EXCL);
fs/xfs/scrub/attr_repair.c
866
mutex_lock(&rx->lock);
fs/xfs/scrub/attr_repair.c
867
xfarray_truncate(rx->pptr_recs);
fs/xfs/scrub/attr_repair.c
868
xfblob_truncate(rx->pptr_names);
fs/xfs/scrub/attr_repair.c
869
mutex_unlock(&rx->lock);
fs/xfs/scrub/attr_repair.c
871
rx->can_flush = false;
fs/xfs/scrub/attr_repair.c
872
rx->attrs_found = 0;
fs/xfs/scrub/attr_repair.c
874
ASSERT(xfarray_bytes(rx->xattr_records) == 0);
fs/xfs/scrub/attr_repair.c
875
ASSERT(xfblob_bytes(rx->xattr_blobs) == 0);
fs/xfs/scrub/attr_repair.c
882
struct xrep_xattr *rx)
fs/xfs/scrub/attr_repair.c
885
struct xfs_scrub *sc = rx->sc;
fs/xfs/scrub/attr_repair.c
917
if (xchk_should_terminate(rx->sc, &error))
fs/xfs/scrub/attr_repair.c
922
error = xrep_xattr_recover_block(rx, dabno,
fs/xfs/scrub/attr_repair.c
928
if (xrep_xattr_want_flush_stashed(rx)) {
fs/xfs/scrub/attr_repair.c
929
error = xrep_xattr_flush_stashed(rx);
fs/xfs/scrub/attr_repair.c
933
if (xrep_xattr_saw_pptr_conflict(rx)) {
fs/xfs/scrub/attr_repair.c
934
error = xrep_xattr_full_reset(rx);
include/linux/can/can-ml.h
62
struct hlist_head rx[RX_MAX];
include/linux/iio/imu/adis.h
159
u8 rx[4];
include/linux/linkmode.h
83
void linkmode_set_pause(unsigned long *advertisement, bool tx, bool rx);
include/linux/mfd/ipaq-micro.h
115
struct ipaq_micro_rxdev rx; /* receive ISR state */
include/linux/mfd/wcd934x/registers.h
325
#define WCD934X_CDC_RX_PATH_CTL(rx) (0xb41 + rx * 0x14)
include/linux/mfd/wcd934x/registers.h
337
#define WCD934X_CDC_RX_PATH_MIX_CTL(rx) (0xb46 + rx * 0x14)
include/linux/peci.h
106
} rx, tx;
include/linux/phy.h
2384
void phy_set_sym_pause(struct phy_device *phydev, bool rx, bool tx,
include/linux/phy.h
2386
void phy_set_asym_pause(struct phy_device *phydev, bool rx, bool tx);
include/linux/serial.h
41
__u32 cts, dsr, rng, dcd, tx, rx;
include/linux/serial_core.h
418
__u32 rx;
include/linux/soc/ti/knav_dma.h
126
struct knav_dma_rx_cfg rx;
include/linux/spi/altera.h
41
unsigned char *rx;
include/linux/timb_dma.h
26
bool rx;
include/net/cfg80211.h
5492
u16 tx, rx;
include/net/mac80211.h
4771
int (*set_ringparam)(struct ieee80211_hw *hw, u32 tx, u32 rx);
include/net/mac80211.h
4773
u32 *tx, u32 *tx_max, u32 *rx, u32 *rx_max);
include/net/mana/mana.h
569
int mana_config_rss(struct mana_port_context *ac, enum TRI_STATE rx,
include/net/netdev_queues.h
108
struct netdev_queue_stats_rx *rx,
include/net/psp/functions.h
85
return pse && pas->rx.spi == pse->spi &&
include/net/psp/types.h
143
struct psp_key_parsed rx;
include/net/tls.h
243
struct cipher_context rx;
include/net/xdp_sock.h
51
struct xsk_queue *rx ____cacheline_aligned_in_smp;
include/soc/fsl/qe/immap_qe.h
204
u8 rx[0x400];
include/soc/tegra/bpmp.h
123
} rx;
include/soc/tegra/ivc.h
23
} rx, tx;
include/soc/tegra/ivc.h
96
int tegra_ivc_init(struct tegra_ivc *ivc, struct device *peer, const struct iosys_map *rx,
include/uapi/linux/atm_eni.h
15
int tx,rx; /* values are in percent and must be > 100 */
include/uapi/linux/atmdev.h
37
__HANDLE_ITEM(rx); /* RX okay */ \
include/uapi/linux/if_xdp.h
67
struct xdp_ring_offset rx;
include/uapi/linux/serial.h
105
int rx, tx;
include/uapi/linux/synclink.h
214
__u32 cts, dsr, rng, dcd, tx, rx;
net/atm/proc.c
49
atomic_read(&stats->rx), atomic_read(&stats->rx_err),
net/can/af_can.c
375
return &dev_rcv_lists->rx[RX_ERR];
net/can/af_can.c
391
return &dev_rcv_lists->rx[RX_INV];
net/can/af_can.c
395
return &dev_rcv_lists->rx[RX_ALL];
net/can/af_can.c
410
return &dev_rcv_lists->rx[RX_FIL];
net/can/af_can.c
591
hlist_for_each_entry_rcu(rcv, &dev_rcv_lists->rx[RX_ERR], list) {
net/can/af_can.c
601
hlist_for_each_entry_rcu(rcv, &dev_rcv_lists->rx[RX_ALL], list) {
net/can/af_can.c
607
hlist_for_each_entry_rcu(rcv, &dev_rcv_lists->rx[RX_FIL], list) {
net/can/af_can.c
615
hlist_for_each_entry_rcu(rcv, &dev_rcv_lists->rx[RX_INV], list) {
net/can/isotp.c
1210
so->rx.state = ISOTP_IDLE;
net/can/isotp.c
1627
if (so->rx.buf != so->rx.sbuf)
net/can/isotp.c
1628
kfree(so->rx.buf);
net/can/isotp.c
1658
so->rx.state = ISOTP_IDLE;
net/can/isotp.c
1661
so->rx.buf = so->rx.sbuf;
net/can/isotp.c
1663
so->rx.buflen = ARRAY_SIZE(so->rx.sbuf);
net/can/isotp.c
167
struct tpcon rx, tx;
net/can/isotp.c
199
if (so->rx.state == ISOTP_WAIT_DATA) {
net/can/isotp.c
208
so->rx.state = ISOTP_IDLE;
net/can/isotp.c
272
so->rx.bs = 0;
net/can/isotp.c
457
so->rx.state = ISOTP_IDLE;
net/can/isotp.c
491
so->rx.state = ISOTP_IDLE;
net/can/isotp.c
494
so->rx.ll_dl = padlen(cf->len);
net/can/isotp.c
497
if (cf->len != so->rx.ll_dl)
net/can/isotp.c
501
so->rx.len = (cf->data[ae] & 0x0F) << 8;
net/can/isotp.c
502
so->rx.len += cf->data[ae + 1];
net/can/isotp.c
505
if (so->rx.len) {
net/can/isotp.c
509
so->rx.len = cf->data[ae + 2] << 24;
net/can/isotp.c
510
so->rx.len += cf->data[ae + 3] << 16;
net/can/isotp.c
511
so->rx.len += cf->data[ae + 4] << 8;
net/can/isotp.c
512
so->rx.len += cf->data[ae + 5];
net/can/isotp.c
517
off = (so->rx.ll_dl > CAN_MAX_DLEN) ? 1 : 0;
net/can/isotp.c
519
if (so->rx.len + ae + off + ff_pci_sz < so->rx.ll_dl)
net/can/isotp.c
523
if (so->rx.len > so->rx.buflen && so->rx.buflen < max_pdu_size) {
net/can/isotp.c
527
so->rx.buf = newbuf;
net/can/isotp.c
528
so->rx.buflen = max_pdu_size;
net/can/isotp.c
532
if (so->rx.len > so->rx.buflen) {
net/can/isotp.c
539
so->rx.idx = 0;
net/can/isotp.c
540
for (i = ae + ff_pci_sz; i < so->rx.ll_dl; i++)
net/can/isotp.c
541
so->rx.buf[so->rx.idx++] = cf->data[i];
net/can/isotp.c
544
so->rx.sn = 1;
net/can/isotp.c
545
so->rx.state = ISOTP_WAIT_DATA;
net/can/isotp.c
563
if (so->rx.state != ISOTP_WAIT_DATA)
net/can/isotp.c
578
if (cf->len > so->rx.ll_dl)
net/can/isotp.c
582
if (cf->len < so->rx.ll_dl) {
net/can/isotp.c
584
if (so->rx.len - so->rx.idx > so->rx.ll_dl - ae - N_PCI_SZ)
net/can/isotp.c
588
if ((cf->data[ae] & 0x0F) != so->rx.sn) {
net/can/isotp.c
595
so->rx.state = ISOTP_IDLE;
net/can/isotp.c
598
so->rx.sn++;
net/can/isotp.c
599
so->rx.sn %= 16;
net/can/isotp.c
602
so->rx.buf[so->rx.idx++] = cf->data[i];
net/can/isotp.c
603
if (so->rx.idx >= so->rx.len)
net/can/isotp.c
607
if (so->rx.idx >= so->rx.len) {
net/can/isotp.c
609
so->rx.state = ISOTP_IDLE;
net/can/isotp.c
620
nskb = alloc_skb(so->rx.len, gfp_any());
net/can/isotp.c
624
memcpy(skb_put(nskb, so->rx.len), so->rx.buf,
net/can/isotp.c
625
so->rx.len);
net/can/isotp.c
634
if (!so->rxfc.bs || ++so->rx.bs < so->rxfc.bs) {
net/can/isotp.c
681
(so->rx.state != ISOTP_IDLE && n_pci_type == N_PCI_FC))
net/can/j1939/j1939-priv.h
283
unsigned int rx;
net/can/j1939/transport.c
1029
if (session->pkt.rx >= session->pkt.total) {
net/can/j1939/transport.c
1031
} else if (session->pkt.rx >= session->pkt.last) {
net/can/j1939/transport.c
1670
session->pkt.rx = 0;
net/can/j1939/transport.c
1891
(session->pkt.rx + 1) > session->pkt.total) {
net/can/j1939/transport.c
1930
if (packet == session->pkt.rx)
net/can/j1939/transport.c
1931
session->pkt.rx++;
net/can/j1939/transport.c
1935
if (session->pkt.rx >= session->pkt.total)
net/can/j1939/transport.c
1941
if (session->pkt.rx >= session->pkt.last)
net/can/j1939/transport.c
927
len = session->pkt.total - session->pkt.rx;
net/can/j1939/transport.c
932
pkt = session->pkt.rx + 1;
net/can/j1939/transport.c
941
dat[2] = session->pkt.rx + 1;
net/can/proc.c
306
if (!hlist_empty(&dev_rcv_lists->rx[idx])) {
net/can/proc.c
308
can_print_rcvlist(m, &dev_rcv_lists->rx[idx], dev);
net/core/dev.c
11162
struct netdev_rx_queue *rx;
net/core/dev.c
11163
size_t sz = count * sizeof(*rx);
net/core/dev.c
11168
rx = kvzalloc(sz, GFP_KERNEL_ACCOUNT | __GFP_RETRY_MAYFAIL);
net/core/dev.c
11169
if (!rx)
net/core/dev.c
11172
dev->_rx = rx;
net/core/dev.c
11175
rx[i].dev = dev;
net/core/dev.c
11178
err = xdp_rxq_info_reg(&rx[i].xdp_rxq, dev, i, 0);
net/core/dev.c
11187
xdp_rxq_info_unreg(&rx[i].xdp_rxq);
net/core/netdev-genl.c
599
netdev_nl_stats_write_rx(struct sk_buff *rsp, struct netdev_queue_stats_rx *rx)
net/core/netdev-genl.c
601
if (netdev_stat_put(rsp, NETDEV_A_QSTATS_RX_PACKETS, rx->packets) ||
net/core/netdev-genl.c
602
netdev_stat_put(rsp, NETDEV_A_QSTATS_RX_BYTES, rx->bytes) ||
net/core/netdev-genl.c
603
netdev_stat_put(rsp, NETDEV_A_QSTATS_RX_ALLOC_FAIL, rx->alloc_fail) ||
net/core/netdev-genl.c
604
netdev_stat_put(rsp, NETDEV_A_QSTATS_RX_HW_DROPS, rx->hw_drops) ||
net/core/netdev-genl.c
605
netdev_stat_put(rsp, NETDEV_A_QSTATS_RX_HW_DROP_OVERRUNS, rx->hw_drop_overruns) ||
net/core/netdev-genl.c
606
netdev_stat_put(rsp, NETDEV_A_QSTATS_RX_CSUM_COMPLETE, rx->csum_complete) ||
net/core/netdev-genl.c
607
netdev_stat_put(rsp, NETDEV_A_QSTATS_RX_CSUM_UNNECESSARY, rx->csum_unnecessary) ||
net/core/netdev-genl.c
608
netdev_stat_put(rsp, NETDEV_A_QSTATS_RX_CSUM_NONE, rx->csum_none) ||
net/core/netdev-genl.c
609
netdev_stat_put(rsp, NETDEV_A_QSTATS_RX_CSUM_BAD, rx->csum_bad) ||
net/core/netdev-genl.c
610
netdev_stat_put(rsp, NETDEV_A_QSTATS_RX_HW_GRO_PACKETS, rx->hw_gro_packets) ||
net/core/netdev-genl.c
611
netdev_stat_put(rsp, NETDEV_A_QSTATS_RX_HW_GRO_BYTES, rx->hw_gro_bytes) ||
net/core/netdev-genl.c
612
netdev_stat_put(rsp, NETDEV_A_QSTATS_RX_HW_GRO_WIRE_PACKETS, rx->hw_gro_wire_packets) ||
net/core/netdev-genl.c
613
netdev_stat_put(rsp, NETDEV_A_QSTATS_RX_HW_GRO_WIRE_BYTES, rx->hw_gro_wire_bytes) ||
net/core/netdev-genl.c
614
netdev_stat_put(rsp, NETDEV_A_QSTATS_RX_HW_DROP_RATELIMITS, rx->hw_drop_ratelimits))
net/core/netdev-genl.c
644
struct netdev_queue_stats_rx rx;
net/core/netdev-genl.c
658
memset(&rx, 0xff, sizeof(rx));
net/core/netdev-genl.c
659
ops->get_queue_stats_rx(netdev, i, &rx);
net/core/netdev-genl.c
660
if (!memchr_inv(&rx, 0xff, sizeof(rx)))
net/core/netdev-genl.c
662
if (netdev_nl_stats_write_rx(rsp, &rx))
net/core/netdev-genl.c
742
struct netdev_queue_stats_rx rx;
net/core/netdev-genl.c
749
memset(&rx, 0xff, sizeof(rx));
net/core/netdev-genl.c
751
ops->get_queue_stats_rx(netdev, i, &rx);
net/core/netdev-genl.c
752
netdev_nl_stats_add(rx_sum, &rx, sizeof(rx));
net/dsa/tag_yt921x.c
102
rx = ntohs(tag[2]);
net/dsa/tag_yt921x.c
103
if (unlikely((rx & YT921X_TAG_PORT_EN) == 0)) {
net/dsa/tag_yt921x.c
105
"Unexpected rx tag 0x%04x\n", rx);
net/dsa/tag_yt921x.c
109
port = FIELD_GET(YT921X_TAG_RX_PORT_M, rx);
net/dsa/tag_yt921x.c
117
skb->priority = FIELD_GET(YT921X_TAG_PRIO_M, rx);
net/dsa/tag_yt921x.c
119
if (!(rx & YT921X_TAG_CODE_EN)) {
net/dsa/tag_yt921x.c
123
u16 code = FIELD_GET(YT921X_TAG_CODE_M, rx);
net/dsa/tag_yt921x.c
87
u16 rx;
net/mac80211/debugfs_key.c
171
key->u.tkip.rx[i].iv32,
net/mac80211/debugfs_key.c
172
key->u.tkip.rx[i].iv16);
net/mac80211/driver-ops.h
833
u32 tx, u32 rx)
net/mac80211/driver-ops.h
840
trace_drv_set_ringparam(local, tx, rx);
net/mac80211/driver-ops.h
842
ret = local->ops->set_ringparam(&local->hw, tx, rx);
net/mac80211/driver-ops.h
849
u32 *tx, u32 *tx_max, u32 *rx, u32 *rx_max)
net/mac80211/driver-ops.h
854
trace_drv_get_ringparam(local, tx, tx_max, rx, rx_max);
net/mac80211/driver-ops.h
856
local->ops->get_ringparam(&local->hw, tx, tx_max, rx, rx_max);
net/mac80211/ieee80211_i.h
2890
ieee80211_drop_unencrypted_mgmt(struct ieee80211_rx_data *rx);
net/mac80211/key.c
1266
seq->tkip.iv32 = key->u.tkip.rx[tid].iv32;
net/mac80211/key.c
1267
seq->tkip.iv16 = key->u.tkip.rx[tid].iv16;
net/mac80211/key.c
1319
key->u.tkip.rx[tid].iv32 = seq->tkip.iv32;
net/mac80211/key.c
1320
key->u.tkip.rx[tid].iv16 = seq->tkip.iv16;
net/mac80211/key.c
631
key->u.tkip.rx[i].iv32 =
net/mac80211/key.c
633
key->u.tkip.rx[i].iv16 =
net/mac80211/key.h
78
struct tkip_ctx_rx rx[IEEE80211_NUM_TIDS];
net/mac80211/main.c
673
.rx = BIT(IEEE80211_STYPE_ACTION >> 4) |
net/mac80211/main.c
692
.rx = BIT(IEEE80211_STYPE_ACTION >> 4) |
net/mac80211/main.c
698
.rx = BIT(IEEE80211_STYPE_ASSOC_REQ >> 4) |
net/mac80211/main.c
709
.rx = BIT(IEEE80211_STYPE_ASSOC_REQ >> 4) |
net/mac80211/main.c
719
.rx = BIT(IEEE80211_STYPE_ACTION >> 4) |
net/mac80211/main.c
724
.rx = BIT(IEEE80211_STYPE_ASSOC_REQ >> 4) |
net/mac80211/main.c
734
.rx = BIT(IEEE80211_STYPE_ACTION >> 4) |
net/mac80211/main.c
744
.rx = BIT(IEEE80211_STYPE_ACTION >> 4) |
net/mac80211/main.c
750
.rx = BIT(IEEE80211_STYPE_ACTION >> 4) |
net/mac80211/rx.c
1019
rx->seqno_idx = seqno_idx;
net/mac80211/rx.c
1020
rx->security_idx = security_idx;
net/mac80211/rx.c
1023
rx->skb->priority = (tid > 7) ? 0 : tid;
net/mac80211/rx.c
1051
static void ieee80211_verify_alignment(struct ieee80211_rx_data *rx)
net/mac80211/rx.c
1054
WARN_ON_ONCE((unsigned long)rx->skb->data & 1);
net/mac80211/rx.c
1131
static ieee80211_rx_result ieee80211_rx_mesh_check(struct ieee80211_rx_data *rx)
net/mac80211/rx.c
1133
struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)rx->skb->data;
net/mac80211/rx.c
1134
char *dev_addr = rx->sdata->vif.addr;
net/mac80211/rx.c
1155
if (!rx->sta || sta_plink_state(rx->sta) != NL80211_PLINK_ESTAB) {
net/mac80211/rx.c
1165
if (rx->skb->len < IEEE80211_MIN_ACTION_SIZE)
net/mac80211/rx.c
1445
static void ieee80211_rx_reorder_ampdu(struct ieee80211_rx_data *rx,
net/mac80211/rx.c
1448
struct sk_buff *skb = rx->skb;
net/mac80211/rx.c
1450
struct sta_info *sta = rx->sta;
net/mac80211/rx.c
1474
!test_bit(tid, rx->sta->ampdu_mlme.agg_session_valid) &&
net/mac80211/rx.c
1475
!test_and_set_bit(tid, rx->sta->ampdu_mlme.unexpected_agg))
net/mac80211/rx.c
1476
ieee80211_send_delba(rx->sdata, rx->sta->sta.addr, tid,
net/mac80211/rx.c
1499
ieee80211_queue_skb_to_iface(rx->sdata, rx->link_id, NULL, skb);
net/mac80211/rx.c
1510
if (ieee80211_sta_manage_reorder_buf(rx->sdata, tid_agg_rx, skb,
net/mac80211/rx.c
1519
ieee80211_rx_h_check_dup(struct ieee80211_rx_data *rx)
net/mac80211/rx.c
1521
struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)rx->skb->data;
net/mac80211/rx.c
1522
struct ieee80211_rx_status *status = IEEE80211_SKB_RXCB(rx->skb);
net/mac80211/rx.c
1532
if (rx->skb->len < 24)
net/mac80211/rx.c
1539
if (!rx->sta)
net/mac80211/rx.c
1543
struct ieee80211_sub_if_data *sdata = rx->sdata;
net/mac80211/rx.c
1562
rx->sta->last_seq_ctrl[rx->seqno_idx] == hdr->seq_ctrl)) {
net/mac80211/rx.c
1563
I802_DEBUG_INC(rx->local->dot11FrameDuplicateCount);
net/mac80211/rx.c
1564
rx->link_sta->rx_stats.num_duplicates++;
net/mac80211/rx.c
1567
rx->sta->last_seq_ctrl[rx->seqno_idx] = hdr->seq_ctrl;
net/mac80211/rx.c
1574
ieee80211_rx_h_check(struct ieee80211_rx_data *rx)
net/mac80211/rx.c
1576
struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)rx->skb->data;
net/mac80211/rx.c
1587
if (ieee80211_vif_is_mesh(&rx->sdata->vif))
net/mac80211/rx.c
1588
return ieee80211_rx_mesh_check(rx);
net/mac80211/rx.c
1592
rx->sdata->vif.type != NL80211_IFTYPE_ADHOC &&
net/mac80211/rx.c
1593
rx->sdata->vif.type != NL80211_IFTYPE_OCB &&
net/mac80211/rx.c
1594
(!rx->sta || !test_sta_flag(rx->sta, WLAN_STA_ASSOC)))) {
net/mac80211/rx.c
1600
if (rx->sta && rx->sdata->vif.type == NL80211_IFTYPE_STATION &&
net/mac80211/rx.c
1607
if (rx->skb->len < hdrlen + 8)
net/mac80211/rx.c
1610
skb_copy_bits(rx->skb, hdrlen + 6, ðertype, 2);
net/mac80211/rx.c
1611
if (ethertype == rx->sdata->control_port_protocol)
net/mac80211/rx.c
1615
if (rx->sdata->vif.type == NL80211_IFTYPE_AP &&
net/mac80211/rx.c
1616
cfg80211_rx_spurious_frame(rx->sdata->dev, hdr->addr2,
net/mac80211/rx.c
1617
rx->link_id, GFP_ATOMIC))
net/mac80211/rx.c
1628
ieee80211_rx_h_check_more_data(struct ieee80211_rx_data *rx)
net/mac80211/rx.c
1634
local = rx->local;
net/mac80211/rx.c
1635
skb = rx->skb;
net/mac80211/rx.c
1655
ieee80211_send_pspoll(local, rx->sdata);
net/mac80211/rx.c
1785
ieee80211_rx_h_uapsd_and_pspoll(struct ieee80211_rx_data *rx)
net/mac80211/rx.c
1787
struct ieee80211_sub_if_data *sdata = rx->sdata;
net/mac80211/rx.c
1788
struct ieee80211_hdr *hdr = (void *)rx->skb->data;
net/mac80211/rx.c
1789
struct ieee80211_rx_status *status = IEEE80211_SKB_RXCB(rx->skb);
net/mac80211/rx.c
1791
if (!rx->sta)
net/mac80211/rx.c
1811
if (!test_sta_flag(rx->sta, WLAN_STA_PS_STA))
net/mac80211/rx.c
1815
ieee80211_sta_pspoll(&rx->sta->sta);
net/mac80211/rx.c
1819
dev_kfree_skb(rx->skb);
net/mac80211/rx.c
1829
ieee80211_sta_uapsd_trigger(&rx->sta->sta, tid);
net/mac80211/rx.c
1836
ieee80211_rx_h_sta_process(struct ieee80211_rx_data *rx)
net/mac80211/rx.c
1838
struct sta_info *sta = rx->sta;
net/mac80211/rx.c
1839
struct link_sta_info *link_sta = rx->link_sta;
net/mac80211/rx.c
1840
struct sk_buff *skb = rx->skb;
net/mac80211/rx.c
1856
if (rx->sdata->vif.type == NL80211_IFTYPE_ADHOC) {
net/mac80211/rx.c
1857
u8 *bssid = ieee80211_get_bssid(hdr, rx->skb->len,
net/mac80211/rx.c
1859
if (ether_addr_equal(bssid, rx->sdata->u.ibss.bssid) &&
net/mac80211/rx.c
1867
} else if (rx->sdata->vif.type == NL80211_IFTYPE_OCB) {
net/mac80211/rx.c
1883
u64_stats_add(&link_sta->rx_stats.bytes, rx->skb->len);
net/mac80211/rx.c
1920
(rx->sdata->vif.type == NL80211_IFTYPE_AP ||
net/mac80211/rx.c
1921
rx->sdata->vif.type == NL80211_IFTYPE_AP_VLAN)) {
net/mac80211/rx.c
1932
if (ieee80211_vif_is_mesh(&rx->sdata->vif))
net/mac80211/rx.c
1940
I802_DEBUG_INC(rx->local->rx_handlers_drop_nullfunc);
net/mac80211/rx.c
1949
(rx->sdata->vif.type == NL80211_IFTYPE_AP ||
net/mac80211/rx.c
1950
(rx->sdata->vif.type == NL80211_IFTYPE_AP_VLAN &&
net/mac80211/rx.c
1951
!rx->sdata->u.vlan.sta))) {
net/mac80211/rx.c
1954
rx->sdata->dev, sta->sta.addr,
net/mac80211/rx.c
1955
rx->link_id, GFP_ATOMIC);
net/mac80211/rx.c
1963
dev_kfree_skb(rx->skb);
net/mac80211/rx.c
1971
ieee80211_rx_get_bigtk(struct ieee80211_rx_data *rx, int idx)
net/mac80211/rx.c
1992
if (rx->link_sta)
net/mac80211/rx.c
1993
key = rcu_dereference(rx->link_sta->gtk[idx]);
net/mac80211/rx.c
1995
key = rcu_dereference(rx->link->gtk[idx]);
net/mac80211/rx.c
1996
if (!key && rx->link_sta)
net/mac80211/rx.c
1997
key = rcu_dereference(rx->link_sta->gtk[idx2]);
net/mac80211/rx.c
1999
key = rcu_dereference(rx->link->gtk[idx2]);
net/mac80211/rx.c
2005
ieee80211_rx_h_decrypt(struct ieee80211_rx_data *rx)
net/mac80211/rx.c
2007
struct sk_buff *skb = rx->skb;
net/mac80211/rx.c
2049
rx->key = NULL;
net/mac80211/rx.c
2052
if (rx->sta) {
net/mac80211/rx.c
2053
int keyid = rx->sta->ptk_idx;
net/mac80211/rx.c
2054
sta_ptk = rcu_dereference(rx->sta->ptk[keyid]);
net/mac80211/rx.c
2058
keyid = ieee80211_get_keyid(rx->skb);
net/mac80211/rx.c
2063
ptk_idx = rcu_dereference(rx->sta->ptk[keyid]);
net/mac80211/rx.c
2068
mmie_keyidx = ieee80211_get_mmie_keyidx(rx->skb);
net/mac80211/rx.c
2071
rx->key = ptk_idx ? ptk_idx : sta_ptk;
net/mac80211/rx.c
2087
if (rx->sdata->dev)
net/mac80211/rx.c
2088
cfg80211_rx_unprot_mlme_mgmt(rx->sdata->dev,
net/mac80211/rx.c
2094
rx->key = ieee80211_rx_get_bigtk(rx, mmie_keyidx);
net/mac80211/rx.c
2095
if (!rx->key)
net/mac80211/rx.c
2106
if (rx->link_sta) {
net/mac80211/rx.c
2108
test_sta_flag(rx->sta, WLAN_STA_MFP))
net/mac80211/rx.c
2111
rx->key = rcu_dereference(rx->link_sta->gtk[mmie_keyidx]);
net/mac80211/rx.c
2113
if (!rx->key)
net/mac80211/rx.c
2114
rx->key = rcu_dereference(rx->link->gtk[mmie_keyidx]);
net/mac80211/rx.c
2126
key = ieee80211_rx_get_bigtk(rx, -1);
net/mac80211/rx.c
2129
key = rcu_dereference(rx->link->default_mgmt_key);
net/mac80211/rx.c
2131
if (rx->link_sta) {
net/mac80211/rx.c
2133
key = rcu_dereference(rx->link_sta->gtk[i]);
net/mac80211/rx.c
2140
key = rcu_dereference(rx->link->gtk[i]);
net/mac80211/rx.c
2147
rx->key = key;
net/mac80211/rx.c
2163
keyidx = ieee80211_get_keyid(rx->skb);
net/mac80211/rx.c
2169
if (is_multicast_ether_addr(hdr->addr1) && rx->link_sta)
net/mac80211/rx.c
2170
rx->key = rcu_dereference(rx->link_sta->gtk[keyidx]);
net/mac80211/rx.c
2173
if (!rx->key) {
net/mac80211/rx.c
2175
rx->key = rcu_dereference(rx->link->gtk[keyidx]);
net/mac80211/rx.c
2176
if (!rx->key)
net/mac80211/rx.c
2177
rx->key = rcu_dereference(rx->sdata->keys[keyidx]);
net/mac80211/rx.c
2184
if (rx->key &&
net/mac80211/rx.c
2185
rx->key->conf.cipher != WLAN_CIPHER_SUITE_WEP40 &&
net/mac80211/rx.c
2186
rx->key->conf.cipher != WLAN_CIPHER_SUITE_WEP104 &&
net/mac80211/rx.c
2188
rx->key = NULL;
net/mac80211/rx.c
2192
if (rx->key) {
net/mac80211/rx.c
2193
if (unlikely(rx->key->flags & KEY_FLAG_TAINTED))
net/mac80211/rx.c
2201
switch (rx->key->conf.cipher) {
net/mac80211/rx.c
2204
result = ieee80211_crypto_wep_decrypt(rx);
net/mac80211/rx.c
2207
result = ieee80211_crypto_tkip_decrypt(rx);
net/mac80211/rx.c
2211
rx, IEEE80211_CCMP_MIC_LEN);
net/mac80211/rx.c
2215
rx, IEEE80211_CCMP_256_MIC_LEN);
net/mac80211/rx.c
2219
rx, IEEE80211_CMAC_128_MIC_LEN);
net/mac80211/rx.c
2223
rx, IEEE80211_CMAC_256_MIC_LEN);
net/mac80211/rx.c
2227
result = ieee80211_crypto_aes_gmac_decrypt(rx);
net/mac80211/rx.c
2231
result = ieee80211_crypto_gcmp_decrypt(rx);
net/mac80211/rx.c
2243
rx->sdata->dev))
net/mac80211/rx.c
2244
cfg80211_rx_unprot_mlme_mgmt(rx->sdata->dev,
net/mac80211/rx.c
2336
static bool requires_sequential_pn(struct ieee80211_rx_data *rx, __le16 fc)
net/mac80211/rx.c
2338
return rx->key &&
net/mac80211/rx.c
2339
(rx->key->conf.cipher == WLAN_CIPHER_SUITE_CCMP ||
net/mac80211/rx.c
2340
rx->key->conf.cipher == WLAN_CIPHER_SUITE_CCMP_256 ||
net/mac80211/rx.c
2341
rx->key->conf.cipher == WLAN_CIPHER_SUITE_GCMP ||
net/mac80211/rx.c
2342
rx->key->conf.cipher == WLAN_CIPHER_SUITE_GCMP_256) &&
net/mac80211/rx.c
2347
ieee80211_rx_h_defragment(struct ieee80211_rx_data *rx)
net/mac80211/rx.c
2349
struct ieee80211_fragment_cache *cache = &rx->sdata->frags;
net/mac80211/rx.c
2356
struct ieee80211_rx_status *status = IEEE80211_SKB_RXCB(rx->skb);
net/mac80211/rx.c
2358
hdr = (struct ieee80211_hdr *)rx->skb->data;
net/mac80211/rx.c
2367
if (rx->sta)
net/mac80211/rx.c
2368
cache = &rx->sta->frags;
net/mac80211/rx.c
2376
I802_DEBUG_INC(rx->local->rx_handlers_fragments);
net/mac80211/rx.c
2378
if (skb_linearize(rx->skb))
net/mac80211/rx.c
2386
hdr = (struct ieee80211_hdr *)rx->skb->data;
net/mac80211/rx.c
2392
rx->seqno_idx, &(rx->skb));
net/mac80211/rx.c
2393
if (requires_sequential_pn(rx, fc)) {
net/mac80211/rx.c
2394
int queue = rx->security_idx;
net/mac80211/rx.c
2401
entry->key_color = rx->key->color;
net/mac80211/rx.c
2403
rx->key->u.ccmp.rx_pn[queue],
net/mac80211/rx.c
2409
BUILD_BUG_ON(sizeof(rx->key->u.ccmp.rx_pn[queue]) !=
net/mac80211/rx.c
2410
sizeof(rx->key->u.gcmp.rx_pn[queue]));
net/mac80211/rx.c
2413
} else if (rx->key &&
net/mac80211/rx.c
2417
entry->key_color = rx->key->color;
net/mac80211/rx.c
2426
rx->seqno_idx, hdr);
net/mac80211/rx.c
2428
I802_DEBUG_INC(rx->local->rx_handlers_drop_defrag);
net/mac80211/rx.c
2441
if (!requires_sequential_pn(rx, fc))
net/mac80211/rx.c
2445
if (entry->key_color != rx->key->color)
net/mac80211/rx.c
2455
rpn = rx->ccm_gcm.pn;
net/mac80211/rx.c
2460
(!rx->key ||
net/mac80211/rx.c
2463
rx->key->color != entry->key_color)) {
net/mac80211/rx.c
2469
} else if (entry->is_protected && rx->key &&
net/mac80211/rx.c
2470
entry->key_color != rx->key->color &&
net/mac80211/rx.c
2475
skb_pull(rx->skb, ieee80211_hdrlen(fc));
net/mac80211/rx.c
2476
__skb_queue_tail(&entry->skb_list, rx->skb);
net/mac80211/rx.c
2478
entry->extra_len += rx->skb->len;
net/mac80211/rx.c
2480
rx->skb = NULL;
net/mac80211/rx.c
2484
rx->skb = __skb_dequeue(&entry->skb_list);
net/mac80211/rx.c
2485
if (skb_tailroom(rx->skb) < entry->extra_len) {
net/mac80211/rx.c
2486
I802_DEBUG_INC(rx->local->rx_expand_skb_head_defrag);
net/mac80211/rx.c
2487
if (unlikely(pskb_expand_head(rx->skb, 0, entry->extra_len,
net/mac80211/rx.c
2489
I802_DEBUG_INC(rx->local->rx_handlers_drop_defrag);
net/mac80211/rx.c
2495
skb_put_data(rx->skb, skb->data, skb->len);
net/mac80211/rx.c
2500
ieee80211_led_rx(rx->local);
net/mac80211/rx.c
2501
if (rx->sta)
net/mac80211/rx.c
2502
rx->link_sta->rx_stats.packets++;
net/mac80211/rx.c
2506
static int ieee80211_802_1x_port_control(struct ieee80211_rx_data *rx)
net/mac80211/rx.c
2508
if (unlikely(!rx->sta || !test_sta_flag(rx->sta, WLAN_STA_AUTHORIZED)))
net/mac80211/rx.c
2514
static int ieee80211_drop_unencrypted(struct ieee80211_rx_data *rx, __le16 fc)
net/mac80211/rx.c
2516
struct sk_buff *skb = rx->skb;
net/mac80211/rx.c
2529
ieee80211_is_data(fc) && rx->key))
net/mac80211/rx.c
2536
ieee80211_drop_unencrypted_mgmt(struct ieee80211_rx_data *rx)
net/mac80211/rx.c
2538
struct ieee80211_rx_status *status = IEEE80211_SKB_RXCB(rx->skb);
net/mac80211/rx.c
2539
struct ieee80211_mgmt *mgmt = (void *)rx->skb->data;
net/mac80211/rx.c
2554
if (rx->sta && test_sta_flag(rx->sta, WLAN_STA_MFP)) {
net/mac80211/rx.c
2556
ieee80211_is_unicast_robust_mgmt_frame(rx->skb))) {
net/mac80211/rx.c
2563
if (!rx->key)
net/mac80211/rx.c
2566
cfg80211_rx_unprot_mlme_mgmt(rx->sdata->dev,
net/mac80211/rx.c
2567
rx->skb->data,
net/mac80211/rx.c
2568
rx->skb->len);
net/mac80211/rx.c
2573
if (unlikely(ieee80211_is_multicast_robust_mgmt_frame(rx->skb) &&
net/mac80211/rx.c
2574
ieee80211_get_mmie_keyidx(rx->skb) < 0)) {
net/mac80211/rx.c
2577
cfg80211_rx_unprot_mlme_mgmt(rx->sdata->dev,
net/mac80211/rx.c
2578
rx->skb->data,
net/mac80211/rx.c
2579
rx->skb->len);
net/mac80211/rx.c
2582
if (unlikely(ieee80211_is_beacon(fc) && rx->key &&
net/mac80211/rx.c
2583
ieee80211_get_mmie_keyidx(rx->skb) < 0)) {
net/mac80211/rx.c
2584
cfg80211_rx_unprot_mlme_mgmt(rx->sdata->dev,
net/mac80211/rx.c
2585
rx->skb->data,
net/mac80211/rx.c
2586
rx->skb->len);
net/mac80211/rx.c
2593
if (unlikely(ieee80211_is_action(fc) && !rx->key &&
net/mac80211/rx.c
2594
ieee80211_is_robust_mgmt_frame(rx->skb)))
net/mac80211/rx.c
2599
ieee80211_is_protected_dual_of_public_action(rx->skb))
net/mac80211/rx.c
2608
ieee80211_is_robust_mgmt_frame(rx->skb) &&
net/mac80211/rx.c
2609
(!rx->sta || !test_sta_flag(rx->sta, WLAN_STA_ASSOC)))
net/mac80211/rx.c
2617
ieee80211_require_encrypted_assoc(fc, rx->sta))
net/mac80211/rx.c
2625
__ieee80211_data_to_8023(struct ieee80211_rx_data *rx, bool *port_control)
net/mac80211/rx.c
2627
struct ieee80211_sub_if_data *sdata = rx->sdata;
net/mac80211/rx.c
2628
struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)rx->skb->data;
net/mac80211/rx.c
2650
ret = ieee80211_data_to_8023(rx->skb, sdata->vif.addr, sdata->vif.type);
net/mac80211/rx.c
2654
ehdr = (struct ethhdr *) rx->skb->data;
net/mac80211/rx.c
2655
if (ehdr->h_proto == rx->sdata->control_port_protocol)
net/mac80211/rx.c
2695
static bool ieee80211_frame_allowed(struct ieee80211_rx_data *rx, __le16 fc)
net/mac80211/rx.c
2699
struct ethhdr *ehdr = (struct ethhdr *) rx->skb->data;
net/mac80211/rx.c
2706
if (unlikely(ehdr->h_proto == rx->sdata->control_port_protocol))
net/mac80211/rx.c
2707
return ieee80211_is_our_addr(rx->sdata, ehdr->h_dest, NULL) ||
net/mac80211/rx.c
2710
if (ieee80211_802_1x_port_control(rx) ||
net/mac80211/rx.c
2711
ieee80211_drop_unencrypted(rx, fc))
net/mac80211/rx.c
2718
struct ieee80211_rx_data *rx)
net/mac80211/rx.c
2720
struct ieee80211_sub_if_data *sdata = rx->sdata;
net/mac80211/rx.c
2730
cfg80211_rx_control_port(dev, skb, noencrypt, rx->link_id);
net/mac80211/rx.c
2757
if (rx->list)
net/mac80211/rx.c
2758
list_add_tail(&skb->list, rx->list);
net/mac80211/rx.c
2768
ieee80211_deliver_skb(struct ieee80211_rx_data *rx)
net/mac80211/rx.c
2770
struct ieee80211_sub_if_data *sdata = rx->sdata;
net/mac80211/rx.c
2773
struct ethhdr *ehdr = (struct ethhdr *) rx->skb->data;
net/mac80211/rx.c
2776
skb = rx->skb;
net/mac80211/rx.c
2781
if (rx->sta) {
net/mac80211/rx.c
2787
u64_stats_update_begin(&rx->link_sta->rx_stats.syncp);
net/mac80211/rx.c
2788
u64_stats_inc(&rx->link_sta->rx_stats.msdu[rx->seqno_idx]);
net/mac80211/rx.c
2789
u64_stats_update_end(&rx->link_sta->rx_stats.syncp);
net/mac80211/rx.c
2795
ehdr->h_proto != rx->sdata->control_port_protocol &&
net/mac80211/rx.c
2851
ieee80211_deliver_skb_to_local_stack(skb, rx);
net/mac80211/rx.c
3119
__ieee80211_rx_h_amsdu(struct ieee80211_rx_data *rx, u8 data_offset)
net/mac80211/rx.c
3121
struct net_device *dev = rx->sdata->dev;
net/mac80211/rx.c
3122
struct sk_buff *skb = rx->skb;
net/mac80211/rx.c
3132
} else switch (rx->sdata->vif.type) {
net/mac80211/rx.c
3138
if (!test_sta_flag(rx->sta, WLAN_STA_TDLS_PEER))
net/mac80211/rx.c
3153
rx->sdata->vif.addr,
net/mac80211/rx.c
3154
rx->sdata->vif.type,
net/mac80211/rx.c
3158
if (rx->sta->amsdu_mesh_control < 0) {
net/mac80211/rx.c
3175
rx->sta->amsdu_mesh_control = valid;
net/mac80211/rx.c
3179
rx->sdata->vif.type,
net/mac80211/rx.c
3180
rx->local->hw.extra_tx_headroom,
net/mac80211/rx.c
3182
rx->sta->amsdu_mesh_control);
net/mac80211/rx.c
3185
rx->skb = __skb_dequeue(&frame_list);
net/mac80211/rx.c
3187
switch (ieee80211_rx_mesh_data(rx->sdata, rx->sta, rx->skb)) {
net/mac80211/rx.c
3191
if (ieee80211_frame_allowed(rx, fc)) {
net/mac80211/rx.c
3192
ieee80211_deliver_skb(rx);
net/mac80211/rx.c
3197
dev_kfree_skb(rx->skb);
net/mac80211/rx.c
3205
ieee80211_rx_h_amsdu(struct ieee80211_rx_data *rx)
net/mac80211/rx.c
3207
struct sk_buff *skb = rx->skb;
net/mac80211/rx.c
3222
switch (rx->sdata->vif.type) {
net/mac80211/rx.c
3224
if (!rx->sdata->u.vlan.sta)
net/mac80211/rx.c
3228
if (!rx->sdata->u.mgd.use_4addr)
net/mac80211/rx.c
3238
if (is_multicast_ether_addr(hdr->addr1) || !rx->sta)
net/mac80211/rx.c
3241
if (rx->key) {
net/mac80211/rx.c
3248
switch (rx->key->conf.cipher) {
net/mac80211/rx.c
3258
return __ieee80211_rx_h_amsdu(rx, 0);
net/mac80211/rx.c
3262
ieee80211_rx_h_data(struct ieee80211_rx_data *rx)
net/mac80211/rx.c
3264
struct ieee80211_sub_if_data *sdata = rx->sdata;
net/mac80211/rx.c
3265
struct ieee80211_local *local = rx->local;
net/mac80211/rx.c
3267
struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)rx->skb->data;
net/mac80211/rx.c
3281
if (rx->sta &&
net/mac80211/rx.c
3282
!test_and_set_sta_flag(rx->sta, WLAN_STA_4ADDR_EVENT))
net/mac80211/rx.c
3284
rx->sdata->dev, rx->sta->sta.addr, rx->link_id,
net/mac80211/rx.c
3289
res = __ieee80211_data_to_8023(rx, &port_control);
net/mac80211/rx.c
3293
res = ieee80211_rx_mesh_data(rx->sdata, rx->sta, rx->skb);
net/mac80211/rx.c
3297
if (!ieee80211_frame_allowed(rx, fc))
net/mac80211/rx.c
3301
if (unlikely(((struct ethhdr *)rx->skb->data)->h_proto ==
net/mac80211/rx.c
3303
struct ieee80211_tdls_data *tf = (void *)rx->skb->data;
net/mac80211/rx.c
3305
if (pskb_may_pull(rx->skb,
net/mac80211/rx.c
3311
rx->skb->protocol = cpu_to_be16(ETH_P_TDLS);
net/mac80211/rx.c
3312
__ieee80211_queue_skb_to_iface(sdata, rx->link_id,
net/mac80211/rx.c
3313
rx->sta, rx->skb);
net/mac80211/rx.c
3318
if (rx->sdata->vif.type == NL80211_IFTYPE_AP_VLAN &&
net/mac80211/rx.c
3323
rx->sdata = sdata;
net/mac80211/rx.c
3326
rx->skb->dev = dev;
net/mac80211/rx.c
3331
((struct ethhdr *)rx->skb->data)->h_dest) &&
net/mac80211/rx.c
3337
ieee80211_deliver_skb(rx);
net/mac80211/rx.c
3343
ieee80211_rx_h_ctrl(struct ieee80211_rx_data *rx, struct sk_buff_head *frames)
net/mac80211/rx.c
3345
struct sk_buff *skb = rx->skb;
net/mac80211/rx.c
3362
if (!rx->sta)
net/mac80211/rx.c
3371
if (!test_bit(tid, rx->sta->ampdu_mlme.agg_session_valid) &&
net/mac80211/rx.c
3372
!test_and_set_bit(tid, rx->sta->ampdu_mlme.unexpected_agg))
net/mac80211/rx.c
3373
ieee80211_send_delba(rx->sdata, rx->sta->sta.addr, tid,
net/mac80211/rx.c
3377
tid_agg_rx = rcu_dereference(rx->sta->ampdu_mlme.tid_rx[tid]);
net/mac80211/rx.c
3384
event.u.ba.sta = &rx->sta->sta;
net/mac80211/rx.c
3393
ieee80211_release_reorder_frames(rx->sdata, tid_agg_rx,
net/mac80211/rx.c
3397
drv_event_callback(rx->local, rx->sdata, &event);
net/mac80211/rx.c
3452
ieee80211_rx_check_bss_color_collision(struct ieee80211_rx_data *rx)
net/mac80211/rx.c
3454
struct ieee80211_mgmt *mgmt = (void *)rx->skb->data;
net/mac80211/rx.c
3459
if (!wiphy_ext_feature_isset(rx->local->hw.wiphy,
net/mac80211/rx.c
3463
if (ieee80211_hw_check(&rx->local->hw, DETECTS_COLOR_COLLISION))
net/mac80211/rx.c
3466
bss_conf = rx->link->conf;
net/mac80211/rx.c
3471
baselen = mgmt->u.beacon.variable - rx->skb->data;
net/mac80211/rx.c
3472
if (baselen > rx->skb->len)
net/mac80211/rx.c
3477
rx->skb->len - baselen);
net/mac80211/rx.c
3491
ieee80211_obss_color_collision_notify(&rx->sdata->vif,
net/mac80211/rx.c
3498
ieee80211_rx_h_mgmt_check(struct ieee80211_rx_data *rx)
net/mac80211/rx.c
3500
struct ieee80211_mgmt *mgmt = (struct ieee80211_mgmt *) rx->skb->data;
net/mac80211/rx.c
3501
struct ieee80211_rx_status *status = IEEE80211_SKB_RXCB(rx->skb);
net/mac80211/rx.c
3511
if (rx->skb->len < 24)
net/mac80211/rx.c
3519
rx->skb->len < IEEE80211_MIN_ACTION_SIZE)
net/mac80211/rx.c
3527
if (rx->sdata->vif.type == NL80211_IFTYPE_AP &&
net/mac80211/rx.c
3529
!(rx->flags & IEEE80211_RX_BEACON_REPORTED)) {
net/mac80211/rx.c
3533
ieee80211_rx_check_bss_color_collision(rx);
net/mac80211/rx.c
3535
if (ieee80211_hw_check(&rx->local->hw, SIGNAL_DBM) &&
net/mac80211/rx.c
3539
cfg80211_report_obss_beacon_khz(rx->local->hw.wiphy,
net/mac80211/rx.c
3540
rx->skb->data, rx->skb->len,
net/mac80211/rx.c
3543
rx->flags |= IEEE80211_RX_BEACON_REPORTED;
net/mac80211/rx.c
3546
return ieee80211_drop_unencrypted_mgmt(rx);
net/mac80211/rx.c
3550
ieee80211_process_rx_twt_action(struct ieee80211_rx_data *rx)
net/mac80211/rx.c
3552
struct ieee80211_mgmt *mgmt = (struct ieee80211_mgmt *)rx->skb->data;
net/mac80211/rx.c
3553
struct ieee80211_sub_if_data *sdata = rx->sdata;
net/mac80211/rx.c
3559
if (!rx->local->ops->add_twt_setup)
net/mac80211/rx.c
3565
if (!rx->sta)
net/mac80211/rx.c
3572
if (rx->skb->len < IEEE80211_MIN_ACTION_SIZE +
net/mac80211/rx.c
3582
if (rx->skb->len < IEEE80211_MIN_ACTION_SIZE +
net/mac80211/rx.c
3590
if (rx->skb->len < IEEE80211_MIN_ACTION_SIZE + 2)
net/mac80211/rx.c
3602
ieee80211_rx_h_action(struct ieee80211_rx_data *rx)
net/mac80211/rx.c
3604
struct ieee80211_local *local = rx->local;
net/mac80211/rx.c
3605
struct ieee80211_sub_if_data *sdata = rx->sdata;
net/mac80211/rx.c
3606
struct ieee80211_mgmt *mgmt = (struct ieee80211_mgmt *) rx->skb->data;
net/mac80211/rx.c
3607
struct ieee80211_rx_status *status = IEEE80211_SKB_RXCB(rx->skb);
net/mac80211/rx.c
3608
int len = rx->skb->len;
net/mac80211/rx.c
3613
if (!rx->sta && mgmt->u.action.category != WLAN_CATEGORY_PUBLIC &&
net/mac80211/rx.c
3623
if (!rx->link_sta->pub->ht_cap.ht_supported &&
net/mac80211/rx.c
3624
!rx->link_sta->pub->he_cap.has_he)
net/mac80211/rx.c
3664
if (rx->link_sta->pub->smps_mode == smps_mode)
net/mac80211/rx.c
3666
rx->link_sta->pub->smps_mode = smps_mode;
net/mac80211/rx.c
3671
sband = rx->local->hw.wiphy->bands[status->band];
net/mac80211/rx.c
3673
rate_control_rate_update(local, sband, rx->link_sta,
net/mac80211/rx.c
3676
rx->sta->addr,
net/mac80211/rx.c
3689
if (!(rx->link_sta->pub->ht_cap.cap &
net/mac80211/rx.c
3706
if (!rx->sta)
net/mac80211/rx.c
3865
if (ieee80211_process_rx_twt_action(rx))
net/mac80211/rx.c
3953
if (rx->sta)
net/mac80211/rx.c
3954
rx->link_sta->rx_stats.packets++;
net/mac80211/rx.c
3955
dev_kfree_skb(rx->skb);
net/mac80211/rx.c
3959
ieee80211_queue_skb_to_iface(sdata, rx->link_id, rx->sta, rx->skb);
net/mac80211/rx.c
3964
ieee80211_rx_h_userspace_mgmt(struct ieee80211_rx_data *rx)
net/mac80211/rx.c
3966
struct ieee80211_rx_status *status = IEEE80211_SKB_RXCB(rx->skb);
net/mac80211/rx.c
3969
.buf = rx->skb->data,
net/mac80211/rx.c
3970
.len = rx->skb->len,
net/mac80211/rx.c
3971
.link_id = rx->link_id,
net/mac80211/rx.c
3972
.have_link_id = rx->link_id >= 0,
net/mac80211/rx.c
3986
if (ieee80211_hw_check(&rx->local->hw, SIGNAL_DBM) &&
net/mac80211/rx.c
3990
if (ieee80211_is_timing_measurement(rx->skb) ||
net/mac80211/rx.c
3991
ieee80211_is_ftm(rx->skb)) {
net/mac80211/rx.c
3992
info.rx_tstamp = ktime_to_ns(skb_hwtstamps(rx->skb)->hwtstamp);
net/mac80211/rx.c
3996
if (cfg80211_rx_mgmt_ext(&rx->sdata->wdev, &info)) {
net/mac80211/rx.c
3997
if (rx->sta)
net/mac80211/rx.c
3998
rx->link_sta->rx_stats.packets++;
net/mac80211/rx.c
3999
dev_kfree_skb(rx->skb);
net/mac80211/rx.c
4007
ieee80211_rx_h_action_post_userspace(struct ieee80211_rx_data *rx)
net/mac80211/rx.c
4009
struct ieee80211_sub_if_data *sdata = rx->sdata;
net/mac80211/rx.c
4010
struct ieee80211_mgmt *mgmt = (struct ieee80211_mgmt *) rx->skb->data;
net/mac80211/rx.c
4011
int len = rx->skb->len;
net/mac80211/rx.c
4035
if (rx->sta)
net/mac80211/rx.c
4036
rx->link_sta->rx_stats.packets++;
net/mac80211/rx.c
4037
dev_kfree_skb(rx->skb);
net/mac80211/rx.c
4042
ieee80211_rx_h_action_return(struct ieee80211_rx_data *rx)
net/mac80211/rx.c
4044
struct ieee80211_local *local = rx->local;
net/mac80211/rx.c
4045
struct ieee80211_mgmt *mgmt = (struct ieee80211_mgmt *) rx->skb->data;
net/mac80211/rx.c
4047
struct ieee80211_sub_if_data *sdata = rx->sdata;
net/mac80211/rx.c
4048
struct ieee80211_rx_status *status = IEEE80211_SKB_RXCB(rx->skb);
net/mac80211/rx.c
4074
nskb = skb_copy_expand(rx->skb, local->hw.extra_tx_headroom, 0,
net/mac80211/rx.c
4081
memcpy(nmgmt->sa, rx->sdata->vif.addr, ETH_ALEN);
net/mac80211/rx.c
4085
if (rx->sdata->vif.type == NL80211_IFTYPE_P2P_DEVICE) {
net/mac80211/rx.c
4096
__ieee80211_tx_skb_tid_band(rx->sdata, nskb, 7, -1,
net/mac80211/rx.c
4104
ieee80211_rx_h_ext(struct ieee80211_rx_data *rx)
net/mac80211/rx.c
4106
struct ieee80211_sub_if_data *sdata = rx->sdata;
net/mac80211/rx.c
4107
struct ieee80211_hdr *hdr = (void *)rx->skb->data;
net/mac80211/rx.c
4116
ieee80211_queue_skb_to_iface(sdata, rx->link_id, rx->sta, rx->skb);
net/mac80211/rx.c
4122
ieee80211_rx_h_mgmt(struct ieee80211_rx_data *rx)
net/mac80211/rx.c
4124
struct ieee80211_sub_if_data *sdata = rx->sdata;
net/mac80211/rx.c
4125
struct ieee80211_mgmt *mgmt = (void *)rx->skb->data;
net/mac80211/rx.c
4173
ieee80211_queue_skb_to_iface(sdata, rx->link_id, rx->sta, rx->skb);
net/mac80211/rx.c
4178
static void ieee80211_rx_handlers_result(struct ieee80211_rx_data *rx,
net/mac80211/rx.c
4182
I802_DEBUG_INC(rx->sdata->local->rx_handlers_queued);
net/mac80211/rx.c
4187
I802_DEBUG_INC(rx->sdata->local->rx_handlers_drop);
net/mac80211/rx.c
4188
if (rx->sta)
net/mac80211/rx.c
4189
rx->link_sta->rx_stats.dropped++;
net/mac80211/rx.c
4192
kfree_skb_reason(rx->skb, (__force u32)res);
net/mac80211/rx.c
4195
static void ieee80211_rx_handlers(struct ieee80211_rx_data *rx,
net/mac80211/rx.c
4203
res = rxh(rx); \
net/mac80211/rx.c
4214
spin_lock_bh(&rx->local->rx_path_lock);
net/mac80211/rx.c
4222
rx->skb = skb;
net/mac80211/rx.c
4224
if (WARN_ON_ONCE(!rx->link)) {
net/mac80211/rx.c
4240
res = ieee80211_rx_h_ctrl(rx, frames);
net/mac80211/rx.c
4253
ieee80211_rx_handlers_result(rx, res);
net/mac80211/rx.c
4258
spin_unlock_bh(&rx->local->rx_path_lock);
net/mac80211/rx.c
4261
static void ieee80211_invoke_rx_handlers(struct ieee80211_rx_data *rx)
net/mac80211/rx.c
4270
res = rxh(rx); \
net/mac80211/rx.c
4278
ieee80211_rx_reorder_ampdu(rx, &reorder_release);
net/mac80211/rx.c
4280
ieee80211_rx_handlers(rx, &reorder_release);
net/mac80211/rx.c
4284
ieee80211_rx_handlers_result(rx, res);
net/mac80211/rx.c
4295
static bool ieee80211_rx_data_set_link(struct ieee80211_rx_data *rx,
net/mac80211/rx.c
4298
rx->link_id = link_id;
net/mac80211/rx.c
4299
rx->link = rcu_dereference(rx->sdata->link[link_id]);
net/mac80211/rx.c
4301
if (!rx->sta)
net/mac80211/rx.c
4302
return rx->link;
net/mac80211/rx.c
4304
if (!ieee80211_rx_is_valid_sta_link_id(&rx->sta->sta, link_id))
net/mac80211/rx.c
4307
rx->link_sta = rcu_dereference(rx->sta->link[link_id]);
net/mac80211/rx.c
4309
return rx->link && rx->link_sta;
net/mac80211/rx.c
4312
static bool ieee80211_rx_data_set_sta(struct ieee80211_rx_data *rx,
net/mac80211/rx.c
4315
rx->link_id = link_id;
net/mac80211/rx.c
4316
rx->sta = sta;
net/mac80211/rx.c
4319
rx->local = sta->sdata->local;
net/mac80211/rx.c
4320
if (!rx->sdata)
net/mac80211/rx.c
4321
rx->sdata = sta->sdata;
net/mac80211/rx.c
4322
rx->link_sta = &sta->deflink;
net/mac80211/rx.c
4324
rx->link_sta = NULL;
net/mac80211/rx.c
4328
if (ieee80211_vif_is_mld(&rx->sdata->vif) &&
net/mac80211/rx.c
4330
rx->link =
net/mac80211/rx.c
4331
rcu_dereference(rx->sdata->link[sta->deflink.link_id]);
net/mac80211/rx.c
4333
rx->link = &rx->sdata->deflink;
net/mac80211/rx.c
4334
} else if (!ieee80211_rx_data_set_link(rx, link_id)) {
net/mac80211/rx.c
4348
struct ieee80211_rx_data rx = {
net/mac80211/rx.c
4360
if (!ieee80211_rx_data_set_sta(&rx, sta, link_id))
net/mac80211/rx.c
4379
drv_event_callback(rx.local, rx.sdata, &event);
net/mac80211/rx.c
4382
ieee80211_rx_handlers(&rx, &frames);
net/mac80211/rx.c
4393
struct ieee80211_rx_data rx = {
net/mac80211/rx.c
4412
if (!ieee80211_rx_data_set_sta(&rx, sta, -1))
net/mac80211/rx.c
4462
ieee80211_rx_handlers(&rx, &frames);
net/mac80211/rx.c
4477
static bool ieee80211_accept_frame(struct ieee80211_rx_data *rx)
net/mac80211/rx.c
4479
struct ieee80211_sub_if_data *sdata = rx->sdata;
net/mac80211/rx.c
4480
struct sk_buff *skb = rx->skb;
net/mac80211/rx.c
4492
ieee80211_is_robust_mgmt_frame(skb) && !rx->sta)
net/mac80211/rx.c
4496
return ieee80211_is_our_addr(sdata, hdr->addr1, &rx->link_id);
net/mac80211/rx.c
4511
if (!rx->sta) {
net/mac80211/rx.c
4535
if (!rx->sta) {
net/mac80211/rx.c
4555
&rx->link_id);
net/mac80211/rx.c
4567
&rx->link_id))
net/mac80211/rx.c
4821
static void ieee80211_rx_8023(struct ieee80211_rx_data *rx,
net/mac80211/rx.c
4826
struct ieee80211_rx_status *status = IEEE80211_SKB_RXCB(rx->skb);
net/mac80211/rx.c
4827
struct sta_info *sta = rx->sta;
net/mac80211/rx.c
4829
struct sk_buff *skb = rx->skb;
net/mac80211/rx.c
4833
if (rx->link_id >= 0) {
net/mac80211/rx.c
4834
link_sta = rcu_dereference(sta->link[rx->link_id]);
net/mac80211/rx.c
4836
dev_kfree_skb(rx->skb);
net/mac80211/rx.c
4889
u64_stats_inc(&stats->msdu[rx->seqno_idx]);
net/mac80211/rx.c
4898
sta_info_get(rx->sdata, da)) {
net/mac80211/rx.c
4922
ieee80211_deliver_skb_to_local_stack(skb, rx);
net/mac80211/rx.c
4925
static bool ieee80211_invoke_fast_rx(struct ieee80211_rx_data *rx,
net/mac80211/rx.c
4928
struct sk_buff *skb = rx->skb;
net/mac80211/rx.c
4993
if (!ieee80211_vif_is_mesh(&rx->sdata->vif) &&
net/mac80211/rx.c
5016
stats = this_cpu_ptr(rx->link_sta->pcpu_rx_stats);
net/mac80211/rx.c
5018
stats = &rx->link_sta->rx_stats;
net/mac80211/rx.c
5020
if (rx->key && !(status->flag & RX_FLAG_MIC_STRIPPED) &&
net/mac80211/rx.c
5024
if (rx->key && !ieee80211_has_protected(hdr->frame_control))
net/mac80211/rx.c
5028
if (__ieee80211_rx_h_amsdu(rx, snap_offs - hdrlen) !=
net/mac80211/rx.c
5038
if (ieee80211_vif_is_mesh(&rx->sdata->vif)) {
net/mac80211/rx.c
5051
res = ieee80211_rx_mesh_data(rx->sdata, rx->sta, rx->skb);
net/mac80211/rx.c
5063
ieee80211_rx_8023(rx, fast_rx, orig_len);
net/mac80211/rx.c
5079
static bool ieee80211_prepare_and_rx_handle(struct ieee80211_rx_data *rx,
net/mac80211/rx.c
5082
struct ieee80211_local *local = rx->local;
net/mac80211/rx.c
5083
struct ieee80211_sub_if_data *sdata = rx->sdata;
net/mac80211/rx.c
5085
struct link_sta_info *link_sta = rx->link_sta;
net/mac80211/rx.c
5086
struct ieee80211_link_data *link = rx->link;
net/mac80211/rx.c
5088
rx->skb = skb;
net/mac80211/rx.c
5097
if (consume && rx->sta) {
net/mac80211/rx.c
5100
fast_rx = rcu_dereference(rx->sta->fast_rx);
net/mac80211/rx.c
5101
if (fast_rx && ieee80211_invoke_fast_rx(rx, fast_rx))
net/mac80211/rx.c
5105
if (!ieee80211_accept_frame(rx))
net/mac80211/rx.c
5111
rx->skb = skb_copy(skb, GFP_ATOMIC);
net/mac80211/rx.c
5112
if (!rx->skb) {
net/mac80211/rx.c
5123
shwt = skb_hwtstamps(rx->skb);
net/mac80211/rx.c
5127
hdr = (struct ieee80211_hdr *)rx->skb->data;
net/mac80211/rx.c
5130
if (unlikely(rx->sta && rx->sta->sta.mlo) &&
net/mac80211/rx.c
5136
ether_addr_copy(hdr->addr1, rx->sdata->vif.addr);
net/mac80211/rx.c
5138
ether_addr_copy(hdr->addr2, rx->sta->addr);
net/mac80211/rx.c
5143
ether_addr_copy(hdr->addr3, rx->sta->addr);
net/mac80211/rx.c
5145
ether_addr_copy(hdr->addr3, rx->sdata->vif.addr);
net/mac80211/rx.c
5150
ieee80211_invoke_rx_handlers(rx);
net/mac80211/rx.c
5162
struct ieee80211_rx_data rx;
net/mac80211/rx.c
5166
memset(&rx, 0, sizeof(rx));
net/mac80211/rx.c
5167
rx.skb = skb;
net/mac80211/rx.c
5168
rx.local = local;
net/mac80211/rx.c
5169
rx.list = list;
net/mac80211/rx.c
5170
rx.link_id = -1;
net/mac80211/rx.c
5192
if (!ieee80211_rx_data_set_sta(&rx, sta, link_id))
net/mac80211/rx.c
5195
fast_rx = rcu_dereference(rx.sta->fast_rx);
net/mac80211/rx.c
5199
ieee80211_rx_8023(&rx, fast_rx, skb->len);
net/mac80211/rx.c
5206
static bool ieee80211_rx_for_interface(struct ieee80211_rx_data *rx,
net/mac80211/rx.c
5220
link_sta = link_sta_info_get_bss(rx->sdata, hdr->addr2);
net/mac80211/rx.c
5227
sta = sta_info_get_bss(rx->sdata, hdr->addr2);
net/mac80211/rx.c
5230
} else if (ieee80211_vif_is_mld(&rx->sdata->vif) &&
net/mac80211/rx.c
5235
for_each_link_data_rcu(rx->sdata, link) {
net/mac80211/rx.c
5248
if (!ieee80211_rx_data_set_sta(rx, sta, link_id))
net/mac80211/rx.c
5251
return ieee80211_prepare_and_rx_handle(rx, skb, consume);
net/mac80211/rx.c
5268
struct ieee80211_rx_data rx;
net/mac80211/rx.c
5274
memset(&rx, 0, sizeof(rx));
net/mac80211/rx.c
5275
rx.skb = skb;
net/mac80211/rx.c
5276
rx.local = local;
net/mac80211/rx.c
5277
rx.list = list;
net/mac80211/rx.c
5278
rx.link_id = -1;
net/mac80211/rx.c
5299
ieee80211_parse_qos(&rx);
net/mac80211/rx.c
5300
ieee80211_verify_alignment(&rx);
net/mac80211/rx.c
5316
if (!ieee80211_rx_data_set_sta(&rx, sta, link_id))
net/mac80211/rx.c
5330
link_sta = link_sta_info_get_bss(rx.sdata,
net/mac80211/rx.c
5335
ieee80211_rx_data_set_link(&rx, link_sta->link_id);
net/mac80211/rx.c
5338
if (ieee80211_prepare_and_rx_handle(&rx, skb, true))
net/mac80211/rx.c
5351
rx.sdata = prev_sta->sdata;
net/mac80211/rx.c
5355
link_sta = link_sta_info_get_bss(rx.sdata,
net/mac80211/rx.c
5363
if (!ieee80211_rx_data_set_sta(&rx, prev_sta, link_id))
net/mac80211/rx.c
5366
ieee80211_prepare_and_rx_handle(&rx, skb, false);
net/mac80211/rx.c
5372
rx.sdata = prev_sta->sdata;
net/mac80211/rx.c
5376
link_sta = link_sta_info_get_bss(rx.sdata,
net/mac80211/rx.c
5384
if (!ieee80211_rx_data_set_sta(&rx, prev_sta, link_id))
net/mac80211/rx.c
5387
if (ieee80211_prepare_and_rx_handle(&rx, skb, true))
net/mac80211/rx.c
5414
rx.sdata = prev;
net/mac80211/rx.c
5415
ieee80211_rx_for_interface(&rx, skb, false);
net/mac80211/rx.c
5421
rx.sdata = prev;
net/mac80211/rx.c
5423
if (ieee80211_rx_for_interface(&rx, skb, true))
net/mac80211/rx.c
984
static void ieee80211_parse_qos(struct ieee80211_rx_data *rx)
net/mac80211/rx.c
986
struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)rx->skb->data;
net/mac80211/rx.c
987
struct ieee80211_rx_status *status = IEEE80211_SKB_RXCB(rx->skb);
net/mac80211/tests/mfp.c
216
struct ieee80211_rx_data rx = {
net/mac80211/tests/mfp.c
241
rx.skb = kunit_zalloc_skb(test, 128, GFP_KERNEL);
net/mac80211/tests/mfp.c
242
KUNIT_ASSERT_NOT_NULL(test, rx.skb);
net/mac80211/tests/mfp.c
243
status = IEEE80211_SKB_RXCB(rx.skb);
net/mac80211/tests/mfp.c
255
skb_put_data(rx.skb, &hdr, sizeof(hdr));
net/mac80211/tests/mfp.c
259
skb_put_u8(rx.skb, params->category);
net/mac80211/tests/mfp.c
260
skb_put_u8(rx.skb, params->action);
net/mac80211/tests/mfp.c
266
skb_put_data(rx.skb, &reason, sizeof(reason));
net/mac80211/tests/mfp.c
272
(__force u32)ieee80211_drop_unencrypted_mgmt(&rx),
net/mac80211/tkip.c
250
struct tkip_ctx_rx *rx_ctx = &key->u.tkip.rx[queue];
net/mac80211/trace.h
1431
TP_PROTO(struct ieee80211_local *local, u32 tx, u32 rx),
net/mac80211/trace.h
1433
TP_ARGS(local, tx, rx),
net/mac80211/trace.h
1438
__field(u32, rx)
net/mac80211/trace.h
1444
__entry->rx = rx;
net/mac80211/trace.h
1449
LOCAL_PR_ARG, __entry->tx, __entry->rx
net/mac80211/trace.h
1455
u32 *rx, u32 *rx_max),
net/mac80211/trace.h
1457
TP_ARGS(local, tx, tx_max, rx, rx_max),
net/mac80211/trace.h
1463
__field(u32, rx)
net/mac80211/trace.h
1471
__entry->rx = *rx;
net/mac80211/trace.h
1478
__entry->tx, __entry->tx_max, __entry->rx, __entry->rx_max
net/mac80211/wep.c
242
ieee80211_crypto_wep_decrypt(struct ieee80211_rx_data *rx)
net/mac80211/wep.c
244
struct sk_buff *skb = rx->skb;
net/mac80211/wep.c
253
if (skb_linearize(rx->skb))
net/mac80211/wep.c
255
if (ieee80211_wep_decrypt(rx->local, rx->skb, rx->key))
net/mac80211/wep.c
258
if (!pskb_may_pull(rx->skb, ieee80211_hdrlen(fc) +
net/mac80211/wep.c
261
ieee80211_wep_remove_iv(rx->local, rx->skb, rx->key);
net/mac80211/wep.c
264
pskb_trim(rx->skb, rx->skb->len - IEEE80211_WEP_ICV_LEN))
net/mac80211/wep.h
26
ieee80211_crypto_wep_decrypt(struct ieee80211_rx_data *rx);
net/mac80211/wpa.c
101
struct sk_buff *skb = rx->skb;
net/mac80211/wpa.c
122
if (!(status->flag & RX_FLAG_IV_STRIPPED) && rx->key &&
net/mac80211/wpa.c
123
rx->key->conf.cipher == WLAN_CIPHER_SUITE_TKIP)
net/mac80211/wpa.c
134
if (!rx->key || rx->key->conf.cipher != WLAN_CIPHER_SUITE_TKIP ||
net/mac80211/wpa.c
138
if (rx->sdata->vif.type == NL80211_IFTYPE_AP && rx->key->conf.keyidx) {
net/mac80211/wpa.c
155
if (skb_linearize(rx->skb))
net/mac80211/wpa.c
161
key = &rx->key->conf.key[NL80211_TKIP_DATA_OFFSET_RX_MIC_KEY];
net/mac80211/wpa.c
171
rx->key->u.tkip.rx[rx->security_idx].iv32 = rx->tkip.iv32;
net/mac80211/wpa.c
172
rx->key->u.tkip.rx[rx->security_idx].iv16 = rx->tkip.iv16;
net/mac80211/wpa.c
177
rx->key->u.tkip.mic_failures++;
net/mac80211/wpa.c
185
cfg80211_michael_mic_failure(rx->sdata->dev, hdr->addr2,
net/mac80211/wpa.c
189
rx->key ? rx->key->conf.keyidx : -1,
net/mac80211/wpa.c
265
ieee80211_crypto_tkip_decrypt(struct ieee80211_rx_data *rx)
net/mac80211/wpa.c
267
struct ieee80211_hdr *hdr = (struct ieee80211_hdr *) rx->skb->data;
net/mac80211/wpa.c
269
struct ieee80211_key *key = rx->key;
net/mac80211/wpa.c
270
struct sk_buff *skb = rx->skb;
net/mac80211/wpa.c
278
if (!rx->sta || skb->len - hdrlen < 12)
net/mac80211/wpa.c
282
if (skb_linearize(rx->skb))
net/mac80211/wpa.c
294
res = ieee80211_tkip_decrypt_data(&rx->local->wep_rx_ctx,
net/mac80211/wpa.c
296
skb->len - hdrlen, rx->sta->sta.addr,
net/mac80211/wpa.c
297
hdr->addr1, hwaccel, rx->security_idx,
net/mac80211/wpa.c
298
&rx->tkip.iv32,
net/mac80211/wpa.c
299
&rx->tkip.iv16);
net/mac80211/wpa.c
515
ieee80211_crypto_ccmp_decrypt(struct ieee80211_rx_data *rx,
net/mac80211/wpa.c
518
struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)rx->skb->data;
net/mac80211/wpa.c
520
struct ieee80211_key *key = rx->key;
net/mac80211/wpa.c
521
struct sk_buff *skb = rx->skb;
net/mac80211/wpa.c
531
!ieee80211_require_encrypted_assoc(hdr->frame_control, rx->sta))
net/mac80211/wpa.c
535
if (!pskb_may_pull(rx->skb, hdrlen + IEEE80211_CCMP_HDR_LEN))
net/mac80211/wpa.c
540
if (skb_linearize(rx->skb))
net/mac80211/wpa.c
545
hdr = (void *)rx->skb->data;
net/mac80211/wpa.c
548
if (!rx->sta || data_len < 0)
net/mac80211/wpa.c
556
queue = rx->security_idx;
net/mac80211/wpa.c
583
memcpy(rx->ccm_gcm.pn, pn, IEEE80211_CCMP_PN_LEN);
net/mac80211/wpa.c
714
ieee80211_crypto_gcmp_decrypt(struct ieee80211_rx_data *rx)
net/mac80211/wpa.c
716
struct ieee80211_hdr *hdr = (struct ieee80211_hdr *)rx->skb->data;
net/mac80211/wpa.c
718
struct ieee80211_key *key = rx->key;
net/mac80211/wpa.c
719
struct sk_buff *skb = rx->skb;
net/mac80211/wpa.c
728
!ieee80211_require_encrypted_assoc(hdr->frame_control, rx->sta))
net/mac80211/wpa.c
732
if (!pskb_may_pull(rx->skb, hdrlen + IEEE80211_GCMP_HDR_LEN))
net/mac80211/wpa.c
737
if (skb_linearize(rx->skb))
net/mac80211/wpa.c
742
hdr = (void *)rx->skb->data;
net/mac80211/wpa.c
745
if (!rx->sta || data_len < 0)
net/mac80211/wpa.c
753
queue = rx->security_idx;
net/mac80211/wpa.c
781
memcpy(rx->ccm_gcm.pn, pn, IEEE80211_CCMP_PN_LEN);
net/mac80211/wpa.c
884
ieee80211_crypto_aes_cmac_decrypt(struct ieee80211_rx_data *rx,
net/mac80211/wpa.c
887
struct sk_buff *skb = rx->skb;
net/mac80211/wpa.c
889
struct ieee80211_key *key = rx->key;
net/mac80211/wpa.c
95
ieee80211_rx_h_michael_mic_verify(struct ieee80211_rx_data *rx)
net/mac80211/wpa.c
993
ieee80211_crypto_aes_gmac_decrypt(struct ieee80211_rx_data *rx)
net/mac80211/wpa.c
995
struct sk_buff *skb = rx->skb;
net/mac80211/wpa.c
997
struct ieee80211_key *key = rx->key;
net/mac80211/wpa.h
17
ieee80211_rx_h_michael_mic_verify(struct ieee80211_rx_data *rx);
net/mac80211/wpa.h
22
ieee80211_crypto_tkip_decrypt(struct ieee80211_rx_data *rx);
net/mac80211/wpa.h
28
ieee80211_crypto_ccmp_decrypt(struct ieee80211_rx_data *rx,
net/mac80211/wpa.h
35
ieee80211_crypto_aes_cmac_decrypt(struct ieee80211_rx_data *rx,
net/mac80211/wpa.h
40
ieee80211_crypto_aes_gmac_decrypt(struct ieee80211_rx_data *rx);
net/mac80211/wpa.h
45
ieee80211_crypto_gcmp_decrypt(struct ieee80211_rx_data *rx);
net/mctp/test/route-test.c
1502
const struct mctp_bind_lookup_test *rx;
net/mctp/test/route-test.c
1510
rx = test->param_value;
net/mctp/test/route-test.c
1512
__mctp_route_test_init(test, &dev, &dst, &sock_ty0, rx->net);
net/mctp/test/route-test.c
1520
if (rx->expect &&
net/mctp/test/route-test.c
1521
strcmp(rx->expect, lookup_binds[i].name) == 0) {
net/mctp/test/route-test.c
1526
KUNIT_ASSERT_EQ(test, !!sock_expect, !!rx->expect);
net/mctp/test/route-test.c
1529
skb_pkt = mctp_test_create_skb_data(&rx->hdr, &rx->ty);
net/mctp/test/route-test.c
1534
if (rx->expect) {
net/mctp/test/route-test.c
1548
rx->expect);
net/nfc/nci/spi.c
180
struct spi_transfer tx, rx;
net/nfc/nci/spi.c
195
memset(&rx, 0, sizeof(struct spi_transfer));
net/nfc/nci/spi.c
196
rx.rx_buf = resp_hdr;
net/nfc/nci/spi.c
197
rx.len = 2;
net/nfc/nci/spi.c
198
rx.cs_change = 1;
net/nfc/nci/spi.c
199
rx.speed_hz = nspi->xfer_speed_hz;
net/nfc/nci/spi.c
200
spi_message_add_tail(&rx, &m);
net/nfc/nci/spi.c
218
memset(&rx, 0, sizeof(struct spi_transfer));
net/nfc/nci/spi.c
219
rx.rx_buf = skb_put(skb, rx_len);
net/nfc/nci/spi.c
220
rx.len = rx_len;
net/nfc/nci/spi.c
221
rx.cs_change = 0;
net/nfc/nci/spi.c
222
rx.delay.value = nspi->xfer_udelay;
net/nfc/nci/spi.c
223
rx.delay.unit = SPI_DELAY_UNIT_USECS;
net/nfc/nci/spi.c
224
rx.speed_hz = nspi->xfer_speed_hz;
net/nfc/nci/spi.c
225
spi_message_add_tail(&rx, &m);
net/psp/psp_sock.c
138
memcpy(&pas->rx, key, sizeof(*key));
net/rxrpc/af_rxrpc.c
108
if (rx->family != AF_INET6)
net/rxrpc/af_rxrpc.c
134
struct rxrpc_sock *rx = rxrpc_sk(sock->sk);
net/rxrpc/af_rxrpc.c
138
_enter("%p,%p,%d", rx, saddr, len);
net/rxrpc/af_rxrpc.c
140
ret = rxrpc_validate_address(rx, srx, len);
net/rxrpc/af_rxrpc.c
145
lock_sock(&rx->sk);
net/rxrpc/af_rxrpc.c
147
switch (rx->sk.sk_state) {
net/rxrpc/af_rxrpc.c
149
rx->srx = *srx;
net/rxrpc/af_rxrpc.c
150
local = rxrpc_lookup_local(sock_net(&rx->sk), &rx->srx);
net/rxrpc/af_rxrpc.c
160
rx->local = local;
net/rxrpc/af_rxrpc.c
161
local->service = rx;
net/rxrpc/af_rxrpc.c
164
rx->sk.sk_state = RXRPC_SERVER_BOUND;
net/rxrpc/af_rxrpc.c
166
rx->local = local;
net/rxrpc/af_rxrpc.c
167
rx->sk.sk_state = RXRPC_CLIENT_BOUND;
net/rxrpc/af_rxrpc.c
176
if (service_id == rx->srx.srx_service)
net/rxrpc/af_rxrpc.c
179
srx->srx_service = rx->srx.srx_service;
net/rxrpc/af_rxrpc.c
180
if (memcmp(srx, &rx->srx, sizeof(*srx)) != 0)
net/rxrpc/af_rxrpc.c
182
rx->second_service = service_id;
net/rxrpc/af_rxrpc.c
183
rx->sk.sk_state = RXRPC_SERVER_BOUND2;
net/rxrpc/af_rxrpc.c
191
release_sock(&rx->sk);
net/rxrpc/af_rxrpc.c
201
release_sock(&rx->sk);
net/rxrpc/af_rxrpc.c
213
struct rxrpc_sock *rx = rxrpc_sk(sk);
net/rxrpc/af_rxrpc.c
217
_enter("%p,%d", rx, backlog);
net/rxrpc/af_rxrpc.c
219
lock_sock(&rx->sk);
net/rxrpc/af_rxrpc.c
221
switch (rx->sk.sk_state) {
net/rxrpc/af_rxrpc.c
227
ASSERT(rx->local != NULL);
net/rxrpc/af_rxrpc.c
236
ret = rxrpc_service_prealloc(rx, GFP_KERNEL);
net/rxrpc/af_rxrpc.c
238
rx->sk.sk_state = RXRPC_SERVER_LISTENING;
net/rxrpc/af_rxrpc.c
244
rx->sk.sk_state = RXRPC_SERVER_LISTEN_DISABLED;
net/rxrpc/af_rxrpc.c
246
rxrpc_discard_prealloc(rx);
net/rxrpc/af_rxrpc.c
256
release_sock(&rx->sk);
net/rxrpc/af_rxrpc.c
277
struct rxrpc_sock *rx = rxrpc_sk(sock->sk);
net/rxrpc/af_rxrpc.c
280
ret = rxrpc_validate_address(rx, srx, sizeof(*srx));
net/rxrpc/af_rxrpc.c
284
peer = rxrpc_lookup_peer(rx->local, srx, gfp);
net/rxrpc/af_rxrpc.c
355
struct rxrpc_sock *rx = rxrpc_sk(sock->sk);
net/rxrpc/af_rxrpc.c
359
if (WARN_ON_ONCE(peer->local != rx->local))
net/rxrpc/af_rxrpc.c
362
lock_sock(&rx->sk);
net/rxrpc/af_rxrpc.c
365
key = rx->key;
net/rxrpc/af_rxrpc.c
377
cp.local = rx->local;
net/rxrpc/af_rxrpc.c
380
cp.security_level = rx->min_sec_level;
net/rxrpc/af_rxrpc.c
384
call = rxrpc_new_client_call(rx, &cp, &p, gfp, debug_id);
net/rxrpc/af_rxrpc.c
475
struct rxrpc_sock *rx = rxrpc_sk(sock->sk);
net/rxrpc/af_rxrpc.c
477
rx->app_ops = app_ops;
net/rxrpc/af_rxrpc.c
490
struct rxrpc_sock *rx = rxrpc_sk(sock->sk);
net/rxrpc/af_rxrpc.c
493
_enter("%p,%p,%d,%d", rx, addr, addr_len, flags);
net/rxrpc/af_rxrpc.c
495
ret = rxrpc_validate_address(rx, srx, addr_len);
net/rxrpc/af_rxrpc.c
501
lock_sock(&rx->sk);
net/rxrpc/af_rxrpc.c
504
if (test_bit(RXRPC_SOCK_CONNECTED, &rx->flags))
net/rxrpc/af_rxrpc.c
507
switch (rx->sk.sk_state) {
net/rxrpc/af_rxrpc.c
509
rx->sk.sk_state = RXRPC_CLIENT_UNBOUND;
net/rxrpc/af_rxrpc.c
519
rx->connect_srx = *srx;
net/rxrpc/af_rxrpc.c
520
set_bit(RXRPC_SOCK_CONNECTED, &rx->flags);
net/rxrpc/af_rxrpc.c
524
release_sock(&rx->sk);
net/rxrpc/af_rxrpc.c
540
struct rxrpc_sock *rx = rxrpc_sk(sock->sk);
net/rxrpc/af_rxrpc.c
543
_enter(",{%d},,%zu", rx->sk.sk_state, len);
net/rxrpc/af_rxrpc.c
549
ret = rxrpc_validate_address(rx, m->msg_name, m->msg_namelen);
net/rxrpc/af_rxrpc.c
556
lock_sock(&rx->sk);
net/rxrpc/af_rxrpc.c
558
switch (rx->sk.sk_state) {
net/rxrpc/af_rxrpc.c
561
rx->srx.srx_family = AF_RXRPC;
net/rxrpc/af_rxrpc.c
562
rx->srx.srx_service = 0;
net/rxrpc/af_rxrpc.c
563
rx->srx.transport_type = SOCK_DGRAM;
net/rxrpc/af_rxrpc.c
564
rx->srx.transport.family = rx->family;
net/rxrpc/af_rxrpc.c
565
switch (rx->family) {
net/rxrpc/af_rxrpc.c
567
rx->srx.transport_len = sizeof(struct sockaddr_in);
net/rxrpc/af_rxrpc.c
571
rx->srx.transport_len = sizeof(struct sockaddr_in6);
net/rxrpc/af_rxrpc.c
578
local = rxrpc_lookup_local(sock_net(sock->sk), &rx->srx);
net/rxrpc/af_rxrpc.c
584
rx->local = local;
net/rxrpc/af_rxrpc.c
585
rx->sk.sk_state = RXRPC_CLIENT_BOUND;
net/rxrpc/af_rxrpc.c
590
test_bit(RXRPC_SOCK_CONNECTED, &rx->flags)) {
net/rxrpc/af_rxrpc.c
591
m->msg_name = &rx->connect_srx;
net/rxrpc/af_rxrpc.c
592
m->msg_namelen = sizeof(rx->connect_srx);
net/rxrpc/af_rxrpc.c
598
ret = rxrpc_sendmsg_oob(rx, m, len);
net/rxrpc/af_rxrpc.c
600
ret = rxrpc_do_sendmsg(rx, m, len);
net/rxrpc/af_rxrpc.c
609
release_sock(&rx->sk);
net/rxrpc/af_rxrpc.c
634
struct rxrpc_sock *rx = rxrpc_sk(sock->sk);
net/rxrpc/af_rxrpc.c
641
lock_sock(&rx->sk);
net/rxrpc/af_rxrpc.c
651
if (rx->sk.sk_state != RXRPC_UNBOUND)
net/rxrpc/af_rxrpc.c
653
rx->exclusive = true;
net/rxrpc/af_rxrpc.c
658
if (rx->key)
net/rxrpc/af_rxrpc.c
661
if (rx->sk.sk_state != RXRPC_UNBOUND)
net/rxrpc/af_rxrpc.c
663
ret = rxrpc_request_key(rx, optval, optlen);
net/rxrpc/af_rxrpc.c
668
if (rx->key)
net/rxrpc/af_rxrpc.c
671
if (rx->sk.sk_state != RXRPC_UNBOUND)
net/rxrpc/af_rxrpc.c
673
ret = rxrpc_server_keyring(rx, optval, optlen);
net/rxrpc/af_rxrpc.c
681
if (rx->sk.sk_state != RXRPC_UNBOUND)
net/rxrpc/af_rxrpc.c
691
rx->min_sec_level = min_sec_level;
net/rxrpc/af_rxrpc.c
697
rx->service_upgrade.from != 0)
net/rxrpc/af_rxrpc.c
700
if (rx->sk.sk_state != RXRPC_SERVER_BOUND2)
net/rxrpc/af_rxrpc.c
707
if ((service_upgrade[0] != rx->srx.srx_service ||
net/rxrpc/af_rxrpc.c
708
service_upgrade[1] != rx->second_service) &&
net/rxrpc/af_rxrpc.c
709
(service_upgrade[0] != rx->second_service ||
net/rxrpc/af_rxrpc.c
710
service_upgrade[1] != rx->srx.srx_service))
net/rxrpc/af_rxrpc.c
712
rx->service_upgrade.from = service_upgrade[0];
net/rxrpc/af_rxrpc.c
713
rx->service_upgrade.to = service_upgrade[1];
net/rxrpc/af_rxrpc.c
721
if (rx->sk.sk_state != RXRPC_UNBOUND)
net/rxrpc/af_rxrpc.c
731
set_bit(RXRPC_SOCK_MANAGE_RESPONSE, &rx->flags);
net/rxrpc/af_rxrpc.c
733
clear_bit(RXRPC_SOCK_MANAGE_RESPONSE, &rx->flags);
net/rxrpc/af_rxrpc.c
744
release_sock(&rx->sk);
net/rxrpc/af_rxrpc.c
76
static int rxrpc_validate_address(struct rxrpc_sock *rx,
net/rxrpc/af_rxrpc.c
783
struct rxrpc_sock *rx = rxrpc_sk(sk);
net/rxrpc/af_rxrpc.c
791
if (!list_empty(&rx->recvmsg_q))
net/rxrpc/af_rxrpc.c
810
struct rxrpc_sock *rx;
net/rxrpc/af_rxrpc.c
837
rx = rxrpc_sk(sk);
net/rxrpc/af_rxrpc.c
838
rx->family = protocol;
net/rxrpc/af_rxrpc.c
839
rx->calls = RB_ROOT;
net/rxrpc/af_rxrpc.c
841
spin_lock_init(&rx->incoming_lock);
net/rxrpc/af_rxrpc.c
842
skb_queue_head_init(&rx->recvmsg_oobq);
net/rxrpc/af_rxrpc.c
843
rx->pending_oobq = RB_ROOT;
net/rxrpc/af_rxrpc.c
844
INIT_LIST_HEAD(&rx->sock_calls);
net/rxrpc/af_rxrpc.c
845
INIT_LIST_HEAD(&rx->to_be_accepted);
net/rxrpc/af_rxrpc.c
846
INIT_LIST_HEAD(&rx->recvmsg_q);
net/rxrpc/af_rxrpc.c
847
spin_lock_init(&rx->recvmsg_lock);
net/rxrpc/af_rxrpc.c
848
rwlock_init(&rx->call_lock);
net/rxrpc/af_rxrpc.c
849
memset(&rx->srx, 0, sizeof(rx->srx));
net/rxrpc/af_rxrpc.c
851
rxnet = rxrpc_net(sock_net(&rx->sk));
net/rxrpc/af_rxrpc.c
854
_leave(" = 0 [%p]", rx);
net/rxrpc/af_rxrpc.c
864
struct rxrpc_sock *rx = rxrpc_sk(sk);
net/rxrpc/af_rxrpc.c
877
spin_lock_irq(&rx->recvmsg_lock);
net/rxrpc/af_rxrpc.c
880
spin_unlock_irq(&rx->recvmsg_lock);
net/rxrpc/af_rxrpc.c
885
rxrpc_discard_prealloc(rx);
net/rxrpc/af_rxrpc.c
896
struct rxrpc_sock *rx = rxrpc_sk(sk);
net/rxrpc/af_rxrpc.c
899
while ((skb = skb_dequeue(&rx->recvmsg_oobq)))
net/rxrpc/af_rxrpc.c
901
while (!RB_EMPTY_ROOT(&rx->pending_oobq)) {
net/rxrpc/af_rxrpc.c
902
skb = rb_entry(rx->pending_oobq.rb_node, struct sk_buff, rbnode);
net/rxrpc/af_rxrpc.c
903
rb_erase(&skb->rbnode, &rx->pending_oobq);
net/rxrpc/af_rxrpc.c
933
struct rxrpc_sock *rx = rxrpc_sk(sk);
net/rxrpc/af_rxrpc.c
950
rx->local->service_closed = true;
net/rxrpc/af_rxrpc.c
954
spin_lock_irq(&rx->recvmsg_lock);
net/rxrpc/af_rxrpc.c
956
spin_unlock_irq(&rx->recvmsg_lock);
net/rxrpc/af_rxrpc.c
958
if (rx->local && rx->local->service == rx) {
net/rxrpc/af_rxrpc.c
959
write_lock(&rx->local->services_lock);
net/rxrpc/af_rxrpc.c
960
rx->local->service = NULL;
net/rxrpc/af_rxrpc.c
961
write_unlock(&rx->local->services_lock);
net/rxrpc/af_rxrpc.c
965
rxrpc_discard_prealloc(rx);
net/rxrpc/af_rxrpc.c
966
rxrpc_release_calls_on_socket(rx);
net/rxrpc/af_rxrpc.c
971
rxrpc_unuse_local(rx->local, rxrpc_local_unuse_release_sock);
net/rxrpc/af_rxrpc.c
972
rxrpc_put_local(rx->local, rxrpc_local_put_release_sock);
net/rxrpc/af_rxrpc.c
973
rx->local = NULL;
net/rxrpc/af_rxrpc.c
974
key_put(rx->key);
net/rxrpc/af_rxrpc.c
975
rx->key = NULL;
net/rxrpc/af_rxrpc.c
976
key_put(rx->securities);
net/rxrpc/af_rxrpc.c
977
rx->securities = NULL;
net/rxrpc/af_rxrpc.c
98
if (rx->family != AF_INET &&
net/rxrpc/af_rxrpc.c
99
rx->family != AF_INET6)
net/rxrpc/ar-internal.h
1060
__releases(&rx->sk.sk_lock)
net/rxrpc/ar-internal.h
1357
void rxrpc_add_pending_oob(struct rxrpc_sock *rx, struct sk_buff *skb);
net/rxrpc/ar-internal.h
1358
int rxrpc_sendmsg_oob(struct rxrpc_sock *rx, struct msghdr *msg, size_t len);
net/rxrpc/call_accept.c
107
write_lock(&rx->call_lock);
net/rxrpc/call_accept.c
110
pp = &rx->calls.rb_node;
net/rxrpc/call_accept.c
125
if (rx->app_ops &&
net/rxrpc/call_accept.c
126
rx->app_ops->user_attach_call) {
net/rxrpc/call_accept.c
128
rx->app_ops->user_attach_call(call, user_call_ID);
net/rxrpc/call_accept.c
133
rb_insert_color(&call->sock_node, &rx->calls);
net/rxrpc/call_accept.c
136
list_add(&call->sock_link, &rx->sock_calls);
net/rxrpc/call_accept.c
138
write_unlock(&rx->call_lock);
net/rxrpc/call_accept.c
151
write_unlock(&rx->call_lock);
net/rxrpc/call_accept.c
162
int rxrpc_service_prealloc(struct rxrpc_sock *rx, gfp_t gfp)
net/rxrpc/call_accept.c
164
struct rxrpc_backlog *b = rx->backlog;
net/rxrpc/call_accept.c
170
rx->backlog = b;
net/rxrpc/call_accept.c
179
void rxrpc_discard_prealloc(struct rxrpc_sock *rx)
net/rxrpc/call_accept.c
181
struct rxrpc_backlog *b = rx->backlog;
net/rxrpc/call_accept.c
182
struct rxrpc_net *rxnet = rxrpc_net(sock_net(&rx->sk));
net/rxrpc/call_accept.c
187
rx->backlog = NULL;
net/rxrpc/call_accept.c
192
spin_lock_irq(&rx->incoming_lock);
net/rxrpc/call_accept.c
193
spin_unlock_irq(&rx->incoming_lock);
net/rxrpc/call_accept.c
223
rcu_assign_pointer(call->socket, rx);
net/rxrpc/call_accept.c
224
if (rx->app_ops &&
net/rxrpc/call_accept.c
225
rx->app_ops->discard_new_call) {
net/rxrpc/call_accept.c
227
rx->app_ops->discard_new_call(call, call->user_call_ID);
net/rxrpc/call_accept.c
233
rxrpc_release_call(rx, call);
net/rxrpc/call_accept.c
245
static struct rxrpc_call *rxrpc_alloc_incoming_call(struct rxrpc_sock *rx,
net/rxrpc/call_accept.c
253
struct rxrpc_backlog *b = rx->backlog;
net/rxrpc/call_accept.c
300
rxrpc_new_incoming_connection(rx, conn, sec, skb);
net/rxrpc/call_accept.c
34
static int rxrpc_service_prealloc_one(struct rxrpc_sock *rx,
net/rxrpc/call_accept.c
344
struct rxrpc_sock *rx;
net/rxrpc/call_accept.c
358
rx = local->service;
net/rxrpc/call_accept.c
359
if (!rx || (sp->hdr.serviceId != rx->srx.srx_service &&
net/rxrpc/call_accept.c
360
sp->hdr.serviceId != rx->second_service)
net/rxrpc/call_accept.c
369
sec = rxrpc_get_incoming_security(rx, skb);
net/rxrpc/call_accept.c
374
spin_lock(&rx->incoming_lock);
net/rxrpc/call_accept.c
375
if (rx->sk.sk_state == RXRPC_SERVER_LISTEN_DISABLED ||
net/rxrpc/call_accept.c
376
rx->sk.sk_state == RXRPC_CLOSE) {
net/rxrpc/call_accept.c
382
call = rxrpc_alloc_incoming_call(rx, local, peer, conn, sec, peer_srx,
net/rxrpc/call_accept.c
393
rxrpc_incoming_call(rx, call, skb);
net/rxrpc/call_accept.c
396
if (rx->app_ops &&
net/rxrpc/call_accept.c
397
rx->app_ops->notify_new_call)
net/rxrpc/call_accept.c
398
rx->app_ops->notify_new_call(&rx->sk, call, call->user_call_ID);
net/rxrpc/call_accept.c
408
spin_unlock(&rx->incoming_lock);
net/rxrpc/call_accept.c
41
struct rxrpc_net *rxnet = rxrpc_net(sock_net(&rx->sk));
net/rxrpc/call_accept.c
432
spin_unlock(&rx->incoming_lock);
net/rxrpc/call_accept.c
444
int rxrpc_user_charge_accept(struct rxrpc_sock *rx, unsigned long user_call_ID)
net/rxrpc/call_accept.c
446
struct rxrpc_backlog *b = rx->backlog;
net/rxrpc/call_accept.c
448
if (rx->sk.sk_state == RXRPC_CLOSE)
net/rxrpc/call_accept.c
451
return rxrpc_service_prealloc_one(rx, b, NULL, user_call_ID, GFP_KERNEL,
net/rxrpc/call_accept.c
47
max = rx->sk.sk_max_ack_backlog;
net/rxrpc/call_accept.c
474
struct rxrpc_sock *rx = rxrpc_sk(sock->sk);
net/rxrpc/call_accept.c
475
struct rxrpc_backlog *b = rx->backlog;
net/rxrpc/call_accept.c
48
tmp = rx->sk.sk_ack_backlog;
net/rxrpc/call_accept.c
480
return rxrpc_service_prealloc_one(rx, b, notify_rx, user_call_ID,
net/rxrpc/call_accept.c
73
peer = rxrpc_alloc_peer(rx->local, gfp, rxrpc_peer_new_prealloc);
net/rxrpc/call_accept.c
97
call = rxrpc_alloc_call(rx, gfp, debug_id);
net/rxrpc/call_object.c
107
read_unlock(&rx->call_lock);
net/rxrpc/call_object.c
113
read_unlock(&rx->call_lock);
net/rxrpc/call_object.c
121
struct rxrpc_call *rxrpc_alloc_call(struct rxrpc_sock *rx, gfp_t gfp,
net/rxrpc/call_object.c
125
struct rxrpc_net *rxnet = rxrpc_net(sock_net(&rx->sk));
net/rxrpc/call_object.c
136
if (rx->sk.sk_kern_sock)
net/rxrpc/call_object.c
188
static struct rxrpc_call *rxrpc_alloc_client_call(struct rxrpc_sock *rx,
net/rxrpc/call_object.c
200
call = rxrpc_alloc_call(rx, gfp, debug_id);
net/rxrpc/call_object.c
320
struct rxrpc_call *rxrpc_new_client_call(struct rxrpc_sock *rx,
net/rxrpc/call_object.c
325
__releases(&rx->sk.sk_lock)
net/rxrpc/call_object.c
334
_enter("%p,%lx", rx, p->user_call_ID);
net/rxrpc/call_object.c
337
release_sock(&rx->sk);
net/rxrpc/call_object.c
343
release_sock(&rx->sk);
net/rxrpc/call_object.c
347
call = rxrpc_alloc_client_call(rx, cp, p, gfp, debug_id);
net/rxrpc/call_object.c
349
release_sock(&rx->sk);
net/rxrpc/call_object.c
361
write_lock(&rx->call_lock);
net/rxrpc/call_object.c
363
pp = &rx->calls.rb_node;
net/rxrpc/call_object.c
377
rcu_assign_pointer(call->socket, rx);
net/rxrpc/call_object.c
382
rb_insert_color(&call->sock_node, &rx->calls);
net/rxrpc/call_object.c
383
list_add(&call->sock_link, &rx->sock_calls);
net/rxrpc/call_object.c
385
write_unlock(&rx->call_lock);
net/rxrpc/call_object.c
393
release_sock(&rx->sk);
net/rxrpc/call_object.c
411
write_unlock(&rx->call_lock);
net/rxrpc/call_object.c
412
release_sock(&rx->sk);
net/rxrpc/call_object.c
438
void rxrpc_incoming_call(struct rxrpc_sock *rx,
net/rxrpc/call_object.c
448
rcu_assign_pointer(call->socket, rx);
net/rxrpc/call_object.c
561
void rxrpc_release_call(struct rxrpc_sock *rx, struct rxrpc_call *call)
net/rxrpc/call_object.c
581
spin_lock_irq(&rx->recvmsg_lock);
net/rxrpc/call_object.c
582
spin_unlock_irq(&rx->recvmsg_lock);
net/rxrpc/call_object.c
584
write_lock(&rx->call_lock);
net/rxrpc/call_object.c
587
rb_erase(&call->sock_node, &rx->calls);
net/rxrpc/call_object.c
593
write_unlock(&rx->call_lock);
net/rxrpc/call_object.c
606
void rxrpc_release_calls_on_socket(struct rxrpc_sock *rx)
net/rxrpc/call_object.c
610
_enter("%p", rx);
net/rxrpc/call_object.c
612
while (!list_empty(&rx->to_be_accepted)) {
net/rxrpc/call_object.c
613
call = list_entry(rx->to_be_accepted.next,
net/rxrpc/call_object.c
621
while (!list_empty(&rx->sock_calls)) {
net/rxrpc/call_object.c
622
call = list_entry(rx->sock_calls.next,
net/rxrpc/call_object.c
627
rxrpc_release_call(rx, call);
net/rxrpc/call_object.c
631
while ((call = list_first_entry_or_null(&rx->recvmsg_q,
net/rxrpc/call_object.c
85
struct rxrpc_call *rxrpc_find_call_by_user_ID(struct rxrpc_sock *rx,
net/rxrpc/call_object.c
91
_enter("%p,%lx", rx, user_call_ID);
net/rxrpc/call_object.c
93
read_lock(&rx->call_lock);
net/rxrpc/call_object.c
95
p = rx->calls.rb_node;
net/rxrpc/conn_event.c
416
struct rxrpc_sock *rx;
net/rxrpc/conn_event.c
432
rx = rcu_dereference(call->socket);
net/rxrpc/conn_event.c
433
if (!rx) {
net/rxrpc/conn_event.c
439
if (test_bit(RXRPC_SOCK_MANAGE_RESPONSE, &rx->flags))
net/rxrpc/conn_service.c
148
void rxrpc_new_incoming_connection(struct rxrpc_sock *rx,
net/rxrpc/conn_service.c
174
conn->service_id == rx->service_upgrade.from)
net/rxrpc/conn_service.c
175
conn->service_id = rx->service_upgrade.to;
net/rxrpc/key.c
612
int rxrpc_request_key(struct rxrpc_sock *rx, sockptr_t optval, int optlen)
net/rxrpc/key.c
619
if (optlen <= 0 || optlen > PAGE_SIZE - 1 || rx->securities)
net/rxrpc/key.c
626
key = request_key_net(&key_type_rxrpc, description, sock_net(&rx->sk), NULL);
net/rxrpc/key.c
633
rx->key = key;
net/rxrpc/oob.c
105
rb_insert_color(&skb->rbnode, &rx->pending_oobq);
net/rxrpc/oob.c
173
static int rxrpc_respond_to_oob(struct rxrpc_sock *rx,
net/rxrpc/oob.c
182
skb = rxrpc_find_pending_oob(rx, p->oob_id);
net/rxrpc/oob.c
184
rb_erase(&skb->rbnode, &rx->pending_oobq);
net/rxrpc/oob.c
185
release_sock(&rx->sk);
net/rxrpc/oob.c
222
int rxrpc_sendmsg_oob(struct rxrpc_sock *rx, struct msghdr *msg, size_t len)
net/rxrpc/oob.c
234
return rxrpc_respond_to_oob(rx, &p, msg);
net/rxrpc/oob.c
236
release_sock(&rx->sk);
net/rxrpc/oob.c
248
release_sock(&rx->sk);
net/rxrpc/oob.c
302
struct rxrpc_sock *rx = rxrpc_sk(sock->sk);
net/rxrpc/oob.c
305
oob = skb_dequeue(&rx->recvmsg_oobq);
net/rxrpc/oob.c
38
struct rxrpc_sock *rx;
net/rxrpc/oob.c
43
rx = rcu_dereference(call->socket);
net/rxrpc/oob.c
44
if (rx) {
net/rxrpc/oob.c
45
sk = &rx->sk;
net/rxrpc/oob.c
46
spin_lock_irq(&rx->recvmsg_lock);
net/rxrpc/oob.c
49
skb->skb_mstamp_ns = rx->oob_id_counter++;
net/rxrpc/oob.c
51
skb_queue_tail(&rx->recvmsg_oobq, skb);
net/rxrpc/oob.c
54
if (rx->app_ops)
net/rxrpc/oob.c
55
rx->app_ops->notify_oob(sk, skb);
net/rxrpc/oob.c
58
spin_unlock_irq(&rx->recvmsg_lock);
net/rxrpc/oob.c
59
if (!rx->app_ops && !sock_flag(sk, SOCK_DEAD))
net/rxrpc/oob.c
69
static struct sk_buff *rxrpc_find_pending_oob(struct rxrpc_sock *rx, u64 oob_id)
net/rxrpc/oob.c
74
p = rx->pending_oobq.rb_node;
net/rxrpc/oob.c
95
void rxrpc_add_pending_oob(struct rxrpc_sock *rx, struct sk_buff *skb)
net/rxrpc/oob.c
97
struct rb_node **pp = &rx->pending_oobq.rb_node, *p = NULL;
net/rxrpc/recvmsg.c
201
struct rxrpc_sock *rx = rxrpc_sk(sock->sk);
net/rxrpc/recvmsg.c
206
skb = skb_peek(&rx->recvmsg_oobq);
net/rxrpc/recvmsg.c
229
skb_unlink(skb, &rx->recvmsg_oobq);
net/rxrpc/recvmsg.c
231
rxrpc_add_pending_oob(rx, skb);
net/rxrpc/recvmsg.c
248
struct rxrpc_sock *rx = rxrpc_sk(sock->sk);
net/rxrpc/recvmsg.c
25
struct rxrpc_sock *rx;
net/rxrpc/recvmsg.c
339
if (!rx->app_ops &&
net/rxrpc/recvmsg.c
340
!skb_queue_empty_lockless(&rx->recvmsg_oobq)) {
net/rxrpc/recvmsg.c
370
struct rxrpc_sock *rx = rxrpc_sk(sock->sk);
net/rxrpc/recvmsg.c
384
timeo = sock_rcvtimeo(&rx->sk, flags & MSG_DONTWAIT);
net/rxrpc/recvmsg.c
387
lock_sock(&rx->sk);
net/rxrpc/recvmsg.c
39
rx = rcu_dereference(call->socket);
net/rxrpc/recvmsg.c
390
if (RB_EMPTY_ROOT(&rx->calls) &&
net/rxrpc/recvmsg.c
391
list_empty(&rx->recvmsg_q) &&
net/rxrpc/recvmsg.c
392
skb_queue_empty_lockless(&rx->recvmsg_oobq) &&
net/rxrpc/recvmsg.c
393
rx->sk.sk_state != RXRPC_SERVER_LISTENING) {
net/rxrpc/recvmsg.c
394
release_sock(&rx->sk);
net/rxrpc/recvmsg.c
398
if (list_empty(&rx->recvmsg_q)) {
net/rxrpc/recvmsg.c
40
sk = &rx->sk;
net/rxrpc/recvmsg.c
405
release_sock(&rx->sk);
net/rxrpc/recvmsg.c
408
prepare_to_wait_exclusive(sk_sleep(&rx->sk), &wait,
net/rxrpc/recvmsg.c
41
if (rx && sk->sk_state < RXRPC_CLOSE) {
net/rxrpc/recvmsg.c
410
ret = sock_error(&rx->sk);
net/rxrpc/recvmsg.c
414
if (list_empty(&rx->recvmsg_q) &&
net/rxrpc/recvmsg.c
415
skb_queue_empty_lockless(&rx->recvmsg_oobq)) {
net/rxrpc/recvmsg.c
421
finish_wait(sk_sleep(&rx->sk), &wait);
net/rxrpc/recvmsg.c
426
if (!skb_queue_empty_lockless(&rx->recvmsg_oobq)) {
net/rxrpc/recvmsg.c
428
release_sock(&rx->sk);
net/rxrpc/recvmsg.c
439
spin_lock_irq(&rx->recvmsg_lock);
net/rxrpc/recvmsg.c
440
l = rx->recvmsg_q.next;
net/rxrpc/recvmsg.c
445
skb_queue_empty(&rx->recvmsg_oobq)) {
net/rxrpc/recvmsg.c
447
spin_unlock_irq(&rx->recvmsg_lock);
net/rxrpc/recvmsg.c
448
release_sock(&rx->sk);
net/rxrpc/recvmsg.c
458
spin_unlock_irq(&rx->recvmsg_lock);
net/rxrpc/recvmsg.c
459
release_sock(&rx->sk);
net/rxrpc/recvmsg.c
468
spin_unlock_irq(&rx->recvmsg_lock);
net/rxrpc/recvmsg.c
47
spin_lock_irq(&rx->recvmsg_lock);
net/rxrpc/recvmsg.c
485
release_sock(&rx->sk);
net/rxrpc/recvmsg.c
50
list_add_tail(&call->recvmsg_link, &rx->recvmsg_q);
net/rxrpc/recvmsg.c
52
spin_unlock_irq(&rx->recvmsg_lock);
net/rxrpc/recvmsg.c
533
rxrpc_release_call(rx, call);
net/rxrpc/recvmsg.c
552
spin_lock_irq(&rx->recvmsg_lock);
net/rxrpc/recvmsg.c
554
list_add(&call->recvmsg_link, &rx->recvmsg_q);
net/rxrpc/recvmsg.c
556
spin_unlock_irq(&rx->recvmsg_lock);
net/rxrpc/recvmsg.c
557
} else if (list_is_first(&call->recvmsg_link, &rx->recvmsg_q)) {
net/rxrpc/recvmsg.c
558
spin_unlock_irq(&rx->recvmsg_lock);
net/rxrpc/recvmsg.c
561
list_move(&call->recvmsg_link, &rx->recvmsg_q);
net/rxrpc/recvmsg.c
562
spin_unlock_irq(&rx->recvmsg_lock);
net/rxrpc/recvmsg.c
570
release_sock(&rx->sk);
net/rxrpc/recvmsg.c
578
finish_wait(sk_sleep(&rx->sk), &wait);
net/rxrpc/security.c
133
const struct rxrpc_security *rxrpc_get_incoming_security(struct rxrpc_sock *rx,
net/rxrpc/security.c
149
!rx->securities) {
net/rxrpc/security.c
166
struct rxrpc_sock *rx;
net/rxrpc/security.c
186
rx = conn->local->service;
net/rxrpc/security.c
187
if (!rx)
net/rxrpc/security.c
191
kref = keyring_search(make_key_ref(rx->securities, 1UL),
net/rxrpc/sendmsg.c
107
static int rxrpc_wait_for_tx_window_intr(struct rxrpc_sock *rx,
net/rxrpc/sendmsg.c
131
static int rxrpc_wait_for_tx_window_waitall(struct rxrpc_sock *rx,
net/rxrpc/sendmsg.c
171
static int rxrpc_wait_for_tx_window_nonintr(struct rxrpc_sock *rx,
net/rxrpc/sendmsg.c
192
static int rxrpc_wait_for_tx_window(struct rxrpc_sock *rx,
net/rxrpc/sendmsg.c
208
ret = rxrpc_wait_for_tx_window_waitall(rx, call);
net/rxrpc/sendmsg.c
210
ret = rxrpc_wait_for_tx_window_intr(rx, call, timeo);
net/rxrpc/sendmsg.c
215
ret = rxrpc_wait_for_tx_window_nonintr(rx, call, timeo);
net/rxrpc/sendmsg.c
229
static void rxrpc_notify_end_tx(struct rxrpc_sock *rx, struct rxrpc_call *call,
net/rxrpc/sendmsg.c
233
notify_end_tx(&rx->sk, call, call->user_call_ID);
net/rxrpc/sendmsg.c
241
static void rxrpc_queue_packet(struct rxrpc_sock *rx, struct rxrpc_call *call,
net/rxrpc/sendmsg.c
270
rxrpc_notify_end_tx(rx, call, notify_end_tx);
net/rxrpc/sendmsg.c
320
static int rxrpc_send_data(struct rxrpc_sock *rx,
net/rxrpc/sendmsg.c
327
struct sock *sk = &rx->sk;
net/rxrpc/sendmsg.c
448
rxrpc_queue_packet(rx, call, txb, notify_end_tx);
net/rxrpc/sendmsg.c
483
ret = rxrpc_wait_for_tx_window(rx, call, &timeo,
net/rxrpc/sendmsg.c
608
rxrpc_new_client_call_for_sendmsg(struct rxrpc_sock *rx, struct msghdr *msg,
net/rxrpc/sendmsg.c
610
__releases(&rx->sk.sk_lock)
net/rxrpc/sendmsg.c
623
release_sock(&rx->sk);
net/rxrpc/sendmsg.c
627
peer = rxrpc_lookup_peer(rx->local, srx, GFP_KERNEL);
net/rxrpc/sendmsg.c
629
release_sock(&rx->sk);
net/rxrpc/sendmsg.c
633
key = rx->key;
net/rxrpc/sendmsg.c
634
if (key && !rx->key->payload.data[0])
net/rxrpc/sendmsg.c
638
cp.local = rx->local;
net/rxrpc/sendmsg.c
640
cp.key = rx->key;
net/rxrpc/sendmsg.c
641
cp.security_level = rx->min_sec_level;
net/rxrpc/sendmsg.c
642
cp.exclusive = rx->exclusive | p->exclusive;
net/rxrpc/sendmsg.c
645
call = rxrpc_new_client_call(rx, &cp, &p->call, GFP_KERNEL,
net/rxrpc/sendmsg.c
659
int rxrpc_do_sendmsg(struct rxrpc_sock *rx, struct msghdr *msg, size_t len)
net/rxrpc/sendmsg.c
684
if (rx->sk.sk_state != RXRPC_SERVER_LISTENING)
net/rxrpc/sendmsg.c
686
ret = rxrpc_user_charge_accept(rx, p.call.user_call_ID);
net/rxrpc/sendmsg.c
690
call = rxrpc_find_call_by_user_ID(rx, p.call.user_call_ID);
net/rxrpc/sendmsg.c
695
call = rxrpc_new_client_call_for_sendmsg(rx, msg, &p);
net/rxrpc/sendmsg.c
721
release_sock(&rx->sk);
net/rxrpc/sendmsg.c
771
ret = rxrpc_send_data(rx, call, msg, len, NULL, &dropped_lock);
net/rxrpc/sendmsg.c
787
release_sock(&rx->sk);
net/rxrpc/server_key.c
121
int rxrpc_server_keyring(struct rxrpc_sock *rx, sockptr_t optval, int optlen)
net/rxrpc/server_key.c
142
rx->securities = key;
net/rxrpc/server_key.c
160
struct rxrpc_sock *rx = rxrpc_sk(sk);
net/rxrpc/server_key.c
164
if (rx->securities)
net/rxrpc/server_key.c
166
else if (rx->sk.sk_state != RXRPC_UNBOUND)
net/rxrpc/server_key.c
169
rx->securities = key_get(keyring);
net/rxrpc/server_key.c
201
struct rxrpc_sock *rx = rxrpc_sk(sk);
net/rxrpc/server_key.c
205
ret = !!test_bit(RXRPC_SOCK_MANAGE_RESPONSE, &rx->flags);
net/rxrpc/server_key.c
207
set_bit(RXRPC_SOCK_MANAGE_RESPONSE, &rx->flags);
net/rxrpc/server_key.c
209
clear_bit(RXRPC_SOCK_MANAGE_RESPONSE, &rx->flags);
net/smc/smc_stats.h
134
SMC_STAT_PAYLOAD_SUB(_smc_stats, SMC_TYPE_D, rx, _len, _rc); \
net/smc/smc_stats.h
136
SMC_STAT_PAYLOAD_SUB(_smc_stats, SMC_TYPE_R, rx, _len, _rc); \
net/smc/smc_stats.h
173
SMC_STAT_RMB_SIZE_SUB(_smc_stats, SMC_TYPE_D, rx, is_add, l); \
net/smc/smc_stats.h
177
SMC_STAT_RMB_SIZE_SUB(_smc_stats, SMC_TYPE_R, rx, is_add, l); \
net/smc/smc_stats.h
190
SMC_STAT_RMB_SUB(_smc_stats, type, SMC_TYPE_D, rx); \
net/smc/smc_stats.h
194
SMC_STAT_RMB_SUB(_smc_stats, type, SMC_TYPE_R, rx); \
net/tipc/crypto.c
1213
struct tipc_crypto *tx, *rx;
net/tipc/crypto.c
1219
rx = c;
net/tipc/crypto.c
1220
tx = tipc_net(rx->net)->crypto_tx;
net/tipc/crypto.c
1221
if (cancel_delayed_work(&rx->work)) {
net/tipc/crypto.c
1222
kfree_sensitive(rx->skey);
net/tipc/crypto.c
1223
rx->skey = NULL;
net/tipc/crypto.c
1224
atomic_xchg(&rx->key_distr, 0);
net/tipc/crypto.c
1225
tipc_node_put(rx->node);
net/tipc/crypto.c
1228
k = atomic_xchg(&rx->peer_rx_active, 0);
net/tipc/crypto.c
1257
static bool tipc_crypto_key_try_align(struct tipc_crypto *rx, u8 new_pending)
net/tipc/crypto.c
1265
spin_lock(&rx->lock);
net/tipc/crypto.c
1266
key = rx->key;
net/tipc/crypto.c
1275
if (tipc_aead_users(rx->aead[key.pending]) > 0)
net/tipc/crypto.c
1279
tmp1 = tipc_aead_rcu_ptr(rx->aead[key.pending], &rx->lock);
net/tipc/crypto.c
1282
rcu_assign_pointer(rx->aead[key.pending], NULL);
net/tipc/crypto.c
1286
tmp2 = rcu_replace_pointer(rx->aead[key.passive], tmp2, lockdep_is_held(&rx->lock));
net/tipc/crypto.c
1292
tipc_crypto_key_set_state(rx, new_passive, 0, new_pending);
net/tipc/crypto.c
1293
rcu_assign_pointer(rx->aead[new_pending], tmp1);
net/tipc/crypto.c
1295
rcu_assign_pointer(rx->aead[new_passive], tmp2);
net/tipc/crypto.c
1298
pr_info_ratelimited("%s: key[%d] -> key[%d]\n", rx->name, key.pending,
net/tipc/crypto.c
1302
spin_unlock(&rx->lock);
net/tipc/crypto.c
1320
struct tipc_crypto *rx,
net/tipc/crypto.c
1335
skb_cb->tx_clone_ctx.rx = rx;
net/tipc/crypto.c
1390
static void tipc_crypto_key_synch(struct tipc_crypto *rx, struct sk_buff *skb)
net/tipc/crypto.c
1393
struct tipc_crypto *tx = tipc_net(rx->net)->crypto_tx;
net/tipc/crypto.c
1395
u32 self = tipc_own_addr(rx->net);
net/tipc/crypto.c
1402
rx->key_master = ehdr->master_key;
net/tipc/crypto.c
1403
if (!rx->key_master)
net/tipc/crypto.c
1416
!atomic_cmpxchg(&rx->key_distr, 0, KEY_DISTR_SCHED)) {
net/tipc/crypto.c
1420
if (queue_delayed_work(tx->wq, &rx->work, delay))
net/tipc/crypto.c
1421
tipc_node_get(rx->node);
net/tipc/crypto.c
1425
atomic_xchg(&rx->key_distr, 0);
net/tipc/crypto.c
1429
cur = atomic_read(&rx->peer_rx_active);
net/tipc/crypto.c
1433
atomic_cmpxchg(&rx->peer_rx_active, cur, new) == cur) {
net/tipc/crypto.c
1439
atomic64_set(&rx->sndnxt, 0);
net/tipc/crypto.c
1444
tx->name, cur, new, rx->name);
net/tipc/crypto.c
1551
void tipc_crypto_timeout(struct tipc_crypto *rx)
net/tipc/crypto.c
1553
struct tipc_net *tn = tipc_net(rx->net);
net/tipc/crypto.c
1578
spin_lock(&rx->lock);
net/tipc/crypto.c
1579
key = rx->key;
net/tipc/crypto.c
1580
if (!key.pending || tipc_aead_users(rx->aead[key.pending]) <= 0)
net/tipc/crypto.c
1586
rx->timer2 = jiffies;
net/tipc/crypto.c
1587
tipc_crypto_key_set_state(rx, key.passive, key.active, 0);
net/tipc/crypto.c
1588
this_cpu_inc(rx->stats->stat[STAT_SWITCHES]);
net/tipc/crypto.c
1589
pr_info("%s: key[%d] is activated\n", rx->name, key.pending);
net/tipc/crypto.c
1594
if (!key.pending || tipc_aead_users(rx->aead[key.pending]) > -10)
net/tipc/crypto.c
1597
tipc_crypto_key_set_state(rx, key.passive, key.active, 0);
net/tipc/crypto.c
1598
tipc_crypto_key_detach(rx->aead[key.pending], &rx->lock);
net/tipc/crypto.c
1599
pr_debug("%s: key[%d] is removed\n", rx->name, key.pending);
net/tipc/crypto.c
1606
if (time_before(jiffies, rx->timer1 + TIPC_RX_ACTIVE_LIM) &&
net/tipc/crypto.c
1607
tipc_aead_users(rx->aead[key.active]) > 0)
net/tipc/crypto.c
1614
rx->timer2 = jiffies;
net/tipc/crypto.c
1615
tipc_crypto_key_set_state(rx, key.passive, 0, key.pending);
net/tipc/crypto.c
1616
tipc_aead_users_set(rx->aead[key.pending], 0);
net/tipc/crypto.c
1617
pr_debug("%s: key[%d] is deactivated\n", rx->name, key.active);
net/tipc/crypto.c
1624
if (time_before(jiffies, rx->timer2 + TIPC_RX_PASSIVE_LIM) &&
net/tipc/crypto.c
1625
tipc_aead_users(rx->aead[key.passive]) > -10)
net/tipc/crypto.c
1628
tipc_crypto_key_set_state(rx, 0, key.active, key.pending);
net/tipc/crypto.c
1629
tipc_crypto_key_detach(rx->aead[key.passive], &rx->lock);
net/tipc/crypto.c
1630
pr_debug("%s: key[%d] is freed\n", rx->name, key.passive);
net/tipc/crypto.c
1633
spin_unlock(&rx->lock);
net/tipc/crypto.c
1647
tipc_crypto_do_cmd(rx->net, cmd);
net/tipc/crypto.c
1814
int tipc_crypto_rcv(struct net *net, struct tipc_crypto *rx,
net/tipc/crypto.c
1829
if (unlikely(!rx || tx_key == KEY_MASTER))
net/tipc/crypto.c
1833
key = rx->key;
net/tipc/crypto.c
1839
if (tipc_crypto_key_try_align(rx, tx_key))
net/tipc/crypto.c
1844
aead = tipc_crypto_key_pick_tx(tx, rx, *skb, tx_key);
net/tipc/crypto.c
1852
aead = tipc_aead_get(rx->aead[tx_key]);
net/tipc/crypto.c
1857
stats = ((rx) ?: tx)->stats;
net/tipc/crypto.c
1872
if (rx) {
net/tipc/crypto.c
1878
rx->nokey = !(rx->skey ||
net/tipc/crypto.c
1879
rcu_access_pointer(rx->aead[n]));
net/tipc/crypto.c
1881
rx->name, rx->nokey,
net/tipc/crypto.c
1882
tx_key, rx->key.keys);
net/tipc/crypto.c
1883
tipc_node_put(rx->node);
net/tipc/crypto.c
1902
struct tipc_crypto *rx = aead->crypto;
net/tipc/crypto.c
1909
rx = skb_cb->tx_clone_ctx.rx;
net/tipc/crypto.c
1911
(rx) ? tipc_node_get_id_str(rx->node) : "-", err, aead,
net/tipc/crypto.c
1928
if (!rx) {
net/tipc/crypto.c
1932
rx = tipc_node_crypto_rx(n);
net/tipc/crypto.c
1933
if (unlikely(!rx))
net/tipc/crypto.c
1943
if (tipc_crypto_key_attach(rx, tmp, ehdr->tx_key, false) < 0) {
net/tipc/crypto.c
1960
rx->timer1 = jiffies;
net/tipc/crypto.c
1967
if (rx->key.passive && ehdr->tx_key == rx->key.passive)
net/tipc/crypto.c
1968
rx->timer2 = jiffies;
net/tipc/crypto.c
1982
tipc_crypto_key_synch(rx, *skb);
net/tipc/crypto.c
2003
if (rx)
net/tipc/crypto.c
2004
tipc_node_put(rx->node);
net/tipc/crypto.c
2010
struct tipc_crypto *tx = tn->crypto_tx, *rx;
net/tipc/crypto.c
2035
rx = tipc_node_crypto_rx_by_list(p);
net/tipc/crypto.c
2036
pr_info("RX(%7.7s)\n%s", tipc_node_get_id_str(rx->node),
net/tipc/crypto.c
2037
tipc_crypto_key_dump(rx, buf));
net/tipc/crypto.c
2062
rx = tipc_node_crypto_rx_by_list(p);
net/tipc/crypto.c
2064
tipc_node_get_id_str(rx->node));
net/tipc/crypto.c
2067
stat = per_cpu_ptr(rx->stats, cpu)->stat[i];
net/tipc/crypto.c
2166
struct tipc_crypto *rx;
net/tipc/crypto.c
2173
rx = tipc_node_crypto_rx_by_addr(net, msg_prevnode(hdr));
net/tipc/crypto.c
2174
if (unlikely(!rx))
net/tipc/crypto.c
2179
if (tipc_crypto_key_rcv(rx, hdr))
net/tipc/crypto.c
2186
tipc_node_put(rx->node);
net/tipc/crypto.c
2287
static bool tipc_crypto_key_rcv(struct tipc_crypto *rx, struct tipc_msg *hdr)
net/tipc/crypto.c
2289
struct tipc_crypto *tx = tipc_net(rx->net)->crypto_tx;
net/tipc/crypto.c
2298
pr_debug("%s: message data size is too small\n", rx->name);
net/tipc/crypto.c
2307
pr_debug("%s: invalid MSG_CRYPTO key size\n", rx->name);
net/tipc/crypto.c
2311
spin_lock(&rx->lock);
net/tipc/crypto.c
2312
if (unlikely(rx->skey || (key_gen == rx->key_gen && rx->key.keys))) {
net/tipc/crypto.c
2313
pr_err("%s: key existed <%p>, gen %d vs %d\n", rx->name,
net/tipc/crypto.c
2314
rx->skey, key_gen, rx->key_gen);
net/tipc/crypto.c
2321
pr_err("%s: unable to allocate memory for skey\n", rx->name);
net/tipc/crypto.c
2331
rx->key_gen = key_gen;
net/tipc/crypto.c
2332
rx->skey_mode = msg_key_mode(hdr);
net/tipc/crypto.c
2333
rx->skey = skey;
net/tipc/crypto.c
2334
rx->nokey = 0;
net/tipc/crypto.c
2338
spin_unlock(&rx->lock);
net/tipc/crypto.c
2342
if (likely(skey && queue_delayed_work(tx->wq, &rx->work, 0)))
net/tipc/crypto.c
2358
struct tipc_crypto *rx = container_of(dwork, struct tipc_crypto, work);
net/tipc/crypto.c
2359
struct tipc_crypto *tx = tipc_net(rx->net)->crypto_tx;
net/tipc/crypto.c
2366
if (atomic_cmpxchg(&rx->key_distr,
net/tipc/crypto.c
2371
rc = tipc_crypto_key_distr(tx, key, rx->node);
net/tipc/crypto.c
2374
tx->name, key, tipc_node_get_id_str(rx->node),
net/tipc/crypto.c
2380
atomic_cmpxchg(&rx->key_distr, KEY_DISTR_COMPL, 0);
net/tipc/crypto.c
2384
if (rx->skey) {
net/tipc/crypto.c
2385
rc = tipc_crypto_key_init(rx, rx->skey, rx->skey_mode, false);
net/tipc/crypto.c
2388
rx->name, rc);
net/tipc/crypto.c
2397
kfree_sensitive(rx->skey);
net/tipc/crypto.c
2398
rx->skey = NULL;
net/tipc/crypto.c
2403
if (resched && queue_delayed_work(tx->wq, &rx->work, delay))
net/tipc/crypto.c
2406
tipc_node_put(rx->node);
net/tipc/crypto.c
285
static bool tipc_crypto_key_try_align(struct tipc_crypto *rx, u8 new_pending);
net/tipc/crypto.c
287
struct tipc_crypto *rx,
net/tipc/crypto.c
290
static void tipc_crypto_key_synch(struct tipc_crypto *rx, struct sk_buff *skb);
net/tipc/crypto.c
305
static bool tipc_crypto_key_rcv(struct tipc_crypto *rx, struct tipc_msg *hdr);
net/tipc/crypto.h
162
void tipc_crypto_timeout(struct tipc_crypto *rx);
net/tipc/crypto.h
166
int tipc_crypto_rcv(struct net *net, struct tipc_crypto *rx,
net/tipc/msg.h
120
struct tipc_crypto *rx;
net/tls/tls_device.c
788
memcpy(rcd_sn, tls_ctx->rx.rec_seq, prot->rec_seq_size);
net/tls/tls_device.c
874
memcpy(rcd_sn, tls_ctx->rx.rec_seq, prot->rec_seq_size);
net/tls/tls_device.c
991
tls_ctx->rx.rec_seq, rxm->full_len,
net/tls/tls_main.c
464
cctx = &ctx->rx;
net/tls/tls_sw.c
1590
memcpy(&dctx->iv[iv_offset], tls_ctx->rx.iv,
net/tls/tls_sw.c
1598
memcpy(&dctx->iv[iv_offset], tls_ctx->rx.iv, prot->salt_size);
net/tls/tls_sw.c
1600
tls_xor_iv_with_seq(prot, &dctx->iv[iv_offset], tls_ctx->rx.rec_seq);
net/tls/tls_sw.c
1605
tls_ctx->rx.rec_seq, tlm->control, prot);
net/tls/tls_sw.c
1819
tls_advance_record_sn(sk, prot, &tls_ctx->rx);
net/tls/tls_sw.c
2817
cctx = &ctx->rx;
net/wireless/mlme.c
685
if (!(wdev->wiphy->mgmt_stypes[wdev->iftype].rx & BIT(mgmt_type))) {
net/wireless/mlme.c
978
if (!(stypes->rx & BIT(stype))) {
net/wireless/nl80211.c
2192
stypes = mgmt_stypes[ift].rx;
net/wireless/trace.h
1883
TP_PROTO(struct wiphy *wiphy, int ret, u32 tx, u32 rx),
net/wireless/trace.h
1884
TP_ARGS(wiphy, ret, tx, rx),
net/wireless/trace.h
1889
__field(u32, rx)
net/wireless/trace.h
1895
__entry->rx = rx;
net/wireless/trace.h
1898
WIPHY_PR_ARG, __entry->ret, __entry->tx, __entry->rx)
net/wireless/trace.h
1903
u32 rx, u32 rx_max),
net/wireless/trace.h
1904
TP_ARGS(wiphy, tx, tx_max, rx, rx_max),
net/wireless/trace.h
1909
__field(u32, rx)
net/wireless/trace.h
1916
__entry->rx = rx;
net/wireless/trace.h
1920
WIPHY_PR_ARG, __entry->tx, __entry->tx_max, __entry->rx,
net/wireless/trace.h
1925
TP_PROTO(struct wiphy *wiphy, int radio_idx, u32 tx, u32 rx),
net/wireless/trace.h
1926
TP_ARGS(wiphy, radio_idx, tx, rx),
net/wireless/trace.h
1931
__field(u32, rx)
net/wireless/trace.h
1937
__entry->rx = rx;
net/wireless/trace.h
1941
__entry->tx, __entry->rx)
net/xdp/xsk.c
1083
if (unlikely(!xs->rx))
net/xdp/xsk.c
1134
if (xs->rx && !xskq_prod_is_empty(xs->rx))
net/xdp/xsk.c
1246
xskq_destroy(xs->rx);
net/xdp/xsk.c
1320
if (!xs->rx && !xs->tx) {
net/xdp/xsk.c
1502
q = (optname == XDP_TX_RING) ? &xs->tx : &xs->rx;
net/xdp/xsk.c
153
err = xskq_prod_reserve_desc(xs->rx, addr, len, flags);
net/xdp/xsk.c
1659
stats.rx_invalid_descs = xskq_nb_invalid_descs(xs->rx);
net/xdp/xsk.c
1687
&off.rx);
net/xdp/xsk.c
1694
off.rx.flags = offsetof(struct xdp_rxtx_ring,
net/xdp/xsk.c
1706
xsk_enter_rxtx_offsets(&off_v1.rx);
net/xdp/xsk.c
1762
q = READ_ONCE(xs->rx);
net/xdp/xsk.c
182
if (xskq_prod_nb_free(xs->rx, num_desc) < num_desc) {
net/xdp/xsk.c
278
if (xskq_prod_nb_free(xs->rx, num_desc) < num_desc) {
net/xdp/xsk.c
351
xskq_prod_submit(xs->rx);
net/xdp/xsk.h
21
struct xdp_ring_offset_v1 rx;
net/xdp/xsk_diag.c
40
if (xs->rx)
net/xdp/xsk_diag.c
41
err = xsk_diag_put_ring(xs->rx, XDP_DIAG_RX_RING, nlskb);
net/xdp/xsk_diag.c
84
du.n_rx_invalid = xskq_nb_invalid_descs(xs->rx);
net/xfrm/xfrm_state.c
1651
struct xfrm_state *rx = NULL, *x = NULL;
net/xfrm/xfrm_state.c
1665
rx = x;
net/xfrm/xfrm_state.c
1670
if (rx)
net/xfrm/xfrm_state.c
1671
xfrm_state_hold(rx);
net/xfrm/xfrm_state.c
1675
return rx;
samples/bpf/xdp_sample_user.c
1110
print_always(FMT_COLUMNl, RX(out->totals.rx));
samples/bpf/xdp_sample_user.c
114
__u64 rx;
samples/bpf/xdp_sample_user.c
1328
if (sample_out.totals.rx) {
samples/bpf/xdp_sample_user.c
1329
double pkts = sample_out.totals.rx;
samples/bpf/xdp_sample_user.c
1332
sample_out.totals.rx);
samples/bpf/xdp_sample_user.c
1428
sample_out.totals.rx += out->totals.rx;
samples/bpf/xdp_sample_user.c
59
#define RX(rx) rx, "rx/s"
samples/bpf/xdp_sample_user.c
685
out->totals.rx += pps;
sound/firewire/amdtp-am824.c
94
s->ctx_data.rx.fdf = AMDTP_FDF_AM824 | s->sfc;
sound/firewire/amdtp-stream.c
1018
struct seq_desc *seq_descs = s->ctx_data.rx.seq.descs;
sound/firewire/amdtp-stream.c
1019
unsigned int seq_size = s->ctx_data.rx.seq.size;
sound/firewire/amdtp-stream.c
1020
unsigned int seq_pos = s->ctx_data.rx.seq.pos;
sound/firewire/amdtp-stream.c
1057
s->ctx_data.rx.seq.pos = seq_pos;
sound/firewire/amdtp-stream.c
1163
unsigned int event_count = s->ctx_data.rx.event_count;
sound/firewire/amdtp-stream.c
1223
s->ctx_data.rx.event_count = event_count;
sound/firewire/amdtp-stream.c
1303
s->ctx_data.rx.cache_pos = 0;
sound/firewire/amdtp-stream.c
1591
struct amdtp_stream *rx;
sound/firewire/amdtp-stream.c
1593
list_for_each_entry(rx, &d->streams, list) {
sound/firewire/amdtp-stream.c
1597
if (rx->direction != AMDTP_OUT_STREAM)
sound/firewire/amdtp-stream.c
1601
tx = rx->ctx_data.rx.replay_target;
sound/firewire/amdtp-stream.c
1759
s->ctx_data.rx.seq.descs = kzalloc_objs(*s->ctx_data.rx.seq.descs,
sound/firewire/amdtp-stream.c
1761
if (!s->ctx_data.rx.seq.descs) {
sound/firewire/amdtp-stream.c
1765
s->ctx_data.rx.seq.size = queue_size;
sound/firewire/amdtp-stream.c
1766
s->ctx_data.rx.seq.pos = 0;
sound/firewire/amdtp-stream.c
1769
s->ctx_data.rx.data_block_state = entry->data_block;
sound/firewire/amdtp-stream.c
1770
s->ctx_data.rx.syt_offset_state = entry->syt_offset;
sound/firewire/amdtp-stream.c
1771
s->ctx_data.rx.last_syt_offset = TICKS_PER_CYCLE;
sound/firewire/amdtp-stream.c
1773
s->ctx_data.rx.event_count = 0;
sound/firewire/amdtp-stream.c
1839
kfree(s->ctx_data.rx.seq.descs);
sound/firewire/amdtp-stream.c
1931
kfree(s->ctx_data.rx.seq.descs);
sound/firewire/amdtp-stream.c
2012
struct amdtp_stream *rx;
sound/firewire/amdtp-stream.c
2015
list_for_each_entry(rx, &d->streams, list) {
sound/firewire/amdtp-stream.c
2016
if (rx->direction == AMDTP_OUT_STREAM) {
sound/firewire/amdtp-stream.c
2044
rx->ctx_data.rx.replay_target = tx;
sound/firewire/amdtp-stream.c
373
unsigned int state = s->ctx_data.rx.data_block_state;
sound/firewire/amdtp-stream.c
407
s->ctx_data.rx.data_block_state = state;
sound/firewire/amdtp-stream.c
453
unsigned int last = s->ctx_data.rx.last_syt_offset;
sound/firewire/amdtp-stream.c
454
unsigned int state = s->ctx_data.rx.syt_offset_state;
sound/firewire/amdtp-stream.c
465
s->ctx_data.rx.last_syt_offset = last;
sound/firewire/amdtp-stream.c
466
s->ctx_data.rx.syt_offset_state = state;
sound/firewire/amdtp-stream.c
545
struct amdtp_stream *target = s->ctx_data.rx.replay_target;
sound/firewire/amdtp-stream.c
548
unsigned int cache_pos = s->ctx_data.rx.cache_pos;
sound/firewire/amdtp-stream.c
557
s->ctx_data.rx.cache_pos = cache_pos;
sound/firewire/amdtp-stream.c
567
if (!d->replay.enable || !s->ctx_data.rx.replay_target) {
sound/firewire/amdtp-stream.c
573
struct amdtp_stream *tx = s->ctx_data.rx.replay_target;
sound/firewire/amdtp-stream.c
575
const unsigned int cache_pos = s->ctx_data.rx.cache_pos;
sound/firewire/amdtp-stream.c
686
((s->ctx_data.rx.fdf << CIP_FDF_SHIFT) & CIP_FDF_MASK) |
sound/firewire/amdtp-stream.h
177
} rx;
sound/firewire/dice/dice-proc.c
107
} rx;
sound/firewire/dice/dice-proc.c
198
quadlets = min_t(u32, tx_rx_header.size, sizeof(buf.rx) / 4);
sound/firewire/dice/dice-proc.c
200
if (dice_proc_read_mem(dice, &buf.rx, sections[4] + 2 +
sound/firewire/dice/dice-proc.c
205
snd_iprintf(buffer, " iso channel: %d\n", (int)buf.rx.iso);
sound/firewire/dice/dice-proc.c
206
snd_iprintf(buffer, " sequence start: %u\n", buf.rx.seq_start);
sound/firewire/dice/dice-proc.c
208
buf.rx.number_audio);
sound/firewire/dice/dice-proc.c
209
snd_iprintf(buffer, " midi ports: %u\n", buf.rx.number_midi);
sound/firewire/dice/dice-proc.c
211
dice_proc_fixup_string(buf.rx.names, RX_NAMES_SIZE);
sound/firewire/dice/dice-proc.c
212
snd_iprintf(buffer, " names: %s\n", buf.rx.names);
sound/firewire/dice/dice-proc.c
216
buf.rx.ac3_caps);
sound/firewire/dice/dice-proc.c
218
buf.rx.ac3_enable);
sound/firewire/digi00x/amdtp-dot.c
130
s->ctx_data.rx.fdf = AMDTP_FDF_AM824 | s->sfc;
sound/firewire/motu/amdtp-motu.c
478
s->ctx_data.rx.fdf = MOTU_FDF_AM824;
sound/firewire/tascam/amdtp-tascam.c
244
s->ctx_data.rx.fdf = 0x00;
sound/soc/bcm/bcm2835-i2s.c
154
bool tx, bool rx)
sound/soc/bcm/bcm2835-i2s.c
165
off |= rx ? BCM2835_I2S_RXON : 0;
sound/soc/bcm/bcm2835-i2s.c
168
clr |= rx ? BCM2835_I2S_RXCLR : 0;
sound/soc/codecs/lpass-rx-macro.c
1387
struct rx_macro *rx = dev_get_drvdata(dev);
sound/soc/codecs/lpass-rx-macro.c
1604
switch (rx->codec_version) {
sound/soc/codecs/lpass-rx-macro.c
162
#define CDC_RX_RXn_RX_PATH_CTL(rx, n) (0x0400 + rx->rxn_reg_stride * n)
sound/soc/codecs/lpass-rx-macro.c
1687
struct rx_macro *rx = snd_soc_component_get_drvdata(component);
sound/soc/codecs/lpass-rx-macro.c
1694
if (e->reg == CDC_RX_RXn_RX_PATH_CFG1(rx, 0))
sound/soc/codecs/lpass-rx-macro.c
1695
look_ahead_dly_reg = CDC_RX_RXn_RX_PATH_CFG0(rx, 0);
sound/soc/codecs/lpass-rx-macro.c
1696
else if (e->reg == CDC_RX_RXn_RX_PATH_CFG1(rx, 1))
sound/soc/codecs/lpass-rx-macro.c
1697
look_ahead_dly_reg = CDC_RX_RXn_RX_PATH_CFG0(rx, 1);
sound/soc/codecs/lpass-rx-macro.c
170
#define CDC_RX_RXn_RX_PATH_CFG0(rx, n) (0x0404 + rx->rxn_reg_stride * n)
sound/soc/codecs/lpass-rx-macro.c
1732
struct rx_macro *rx = snd_soc_component_get_drvdata(component);
sound/soc/codecs/lpass-rx-macro.c
1734
for_each_set_bit(port, &rx->active_ch_mask[dai->id], RX_MACRO_PORTS_MAX) {
sound/soc/codecs/lpass-rx-macro.c
1755
int_fs_reg = CDC_RX_RXn_RX_PATH_CTL(rx, j);
sound/soc/codecs/lpass-rx-macro.c
177
#define CDC_RX_RXn_RX_PATH_CFG1(rx, n) (0x0408 + rx->rxn_reg_stride * n)
sound/soc/codecs/lpass-rx-macro.c
1777
struct rx_macro *rx = snd_soc_component_get_drvdata(component);
sound/soc/codecs/lpass-rx-macro.c
1779
for_each_set_bit(port, &rx->active_ch_mask[dai->id], RX_MACRO_PORTS_MAX) {
sound/soc/codecs/lpass-rx-macro.c
1788
int_fs_reg = CDC_RX_RXn_RX_PATH_MIX_CTL(rx, j);
sound/soc/codecs/lpass-rx-macro.c
181
#define CDC_RX_RXn_RX_PATH_CFG2(rx, n) (0x040C + rx->rxn_reg_stride * n)
sound/soc/codecs/lpass-rx-macro.c
1823
struct rx_macro *rx = snd_soc_component_get_drvdata(component);
sound/soc/codecs/lpass-rx-macro.c
1834
rx->bit_width[dai->id] = params_width(params);
sound/soc/codecs/lpass-rx-macro.c
184
#define CDC_RX_RXn_RX_PATH_CFG3(rx, n) (0x0410 + rx->rxn_reg_stride * n)
sound/soc/codecs/lpass-rx-macro.c
1847
struct rx_macro *rx = snd_soc_component_get_drvdata(component);
sound/soc/codecs/lpass-rx-macro.c
1855
for_each_set_bit(temp, &rx->active_ch_mask[dai->id],
sound/soc/codecs/lpass-rx-macro.c
1876
*rx_num = rx->active_ch_cnt[dai->id];
sound/soc/codecs/lpass-rx-macro.c
188
#define CDC_RX_RXn_RX_VOL_CTL(rx, n) (0x0414 + rx->rxn_reg_stride * n)
sound/soc/codecs/lpass-rx-macro.c
190
#define CDC_RX_RXn_RX_PATH_MIX_CTL(rx, n) (0x0418 + rx->rxn_reg_stride * n)
sound/soc/codecs/lpass-rx-macro.c
1907
struct rx_macro *rx = snd_soc_component_get_drvdata(component);
sound/soc/codecs/lpass-rx-macro.c
1916
reg = CDC_RX_RXn_RX_PATH_CTL(rx, j);
sound/soc/codecs/lpass-rx-macro.c
1917
mix_reg = CDC_RX_RXn_RX_PATH_MIX_CTL(rx, j);
sound/soc/codecs/lpass-rx-macro.c
1923
if (rx->main_clk_users[j] > 0) {
sound/soc/codecs/lpass-rx-macro.c
1933
for_each_set_bit(port, &rx->active_ch_mask[dai->id], RX_MACRO_PORTS_MAX) {
sound/soc/codecs/lpass-rx-macro.c
1943
if (rx->main_clk_users[j] > 0) {
sound/soc/codecs/lpass-rx-macro.c
197
#define CDC_RX_RXn_RX_VOL_MIX_CTL(rx, n) (0x0420 + rx->rxn_reg_stride * n)
sound/soc/codecs/lpass-rx-macro.c
202
#define CDC_RX_RXn_RX_PATH_SEC3(rx, n) (0x042c + rx->rxn_reg_stride * n)
sound/soc/codecs/lpass-rx-macro.c
2033
static void rx_macro_mclk_enable(struct rx_macro *rx, bool mclk_enable)
sound/soc/codecs/lpass-rx-macro.c
2035
struct regmap *regmap = rx->regmap;
sound/soc/codecs/lpass-rx-macro.c
2038
if (rx->rx_mclk_users == 0) {
sound/soc/codecs/lpass-rx-macro.c
205
#define CDC_RX_RXn_RX_PATH_SEC7(rx, n) \
sound/soc/codecs/lpass-rx-macro.c
2052
rx->rx_mclk_users++;
sound/soc/codecs/lpass-rx-macro.c
2054
if (rx->rx_mclk_users <= 0) {
sound/soc/codecs/lpass-rx-macro.c
2055
dev_err(rx->dev, "%s: clock already disabled\n", __func__);
sound/soc/codecs/lpass-rx-macro.c
2056
rx->rx_mclk_users = 0;
sound/soc/codecs/lpass-rx-macro.c
2059
rx->rx_mclk_users--;
sound/soc/codecs/lpass-rx-macro.c
206
(0x0434 + (rx->rxn_reg_stride * n) + ((n > 1) ? rx->rxn_reg_stride2 : 0))
sound/soc/codecs/lpass-rx-macro.c
2060
if (rx->rx_mclk_users == 0) {
sound/soc/codecs/lpass-rx-macro.c
2077
struct rx_macro *rx = snd_soc_component_get_drvdata(component);
sound/soc/codecs/lpass-rx-macro.c
2082
rx_macro_mclk_enable(rx, true);
sound/soc/codecs/lpass-rx-macro.c
2085
rx_macro_mclk_enable(rx, false);
sound/soc/codecs/lpass-rx-macro.c
211
#define CDC_RX_RXn_RX_PATH_DSM_CTL(rx, n) \
sound/soc/codecs/lpass-rx-macro.c
212
(0x0440 + (rx->rxn_reg_stride * n) + ((n > 1) ? rx->rxn_reg_stride2 : 0))
sound/soc/codecs/lpass-rx-macro.c
2138
struct rx_macro *rx = snd_soc_component_get_drvdata(component);
sound/soc/codecs/lpass-rx-macro.c
2141
reg = CDC_RX_RXn_RX_PATH_CTL(rx, w->shift);
sound/soc/codecs/lpass-rx-macro.c
2142
gain_reg = CDC_RX_RXn_RX_VOL_CTL(rx, w->shift);
sound/soc/codecs/lpass-rx-macro.c
2165
struct rx_macro *rx,
sound/soc/codecs/lpass-rx-macro.c
2174
pcm_rate = snd_soc_component_read(component, CDC_RX_RXn_RX_PATH_CTL(rx, comp)) & 0x0F;
sound/soc/codecs/lpass-rx-macro.c
2185
snd_soc_component_update_bits(component, CDC_RX_RXn_RX_PATH_CFG3(rx, comp),
sound/soc/codecs/lpass-rx-macro.c
2189
snd_soc_component_update_bits(component, CDC_RX_RXn_RX_PATH_CFG3(rx, comp),
sound/soc/codecs/lpass-rx-macro.c
2191
if (!rx->comp_enabled[comp])
sound/soc/codecs/lpass-rx-macro.c
2202
snd_soc_component_write_field(component, CDC_RX_RXn_RX_PATH_CFG0(rx, comp),
sound/soc/codecs/lpass-rx-macro.c
2209
snd_soc_component_write_field(component, CDC_RX_RXn_RX_PATH_CFG0(rx, comp),
sound/soc/codecs/lpass-rx-macro.c
2221
struct rx_macro *rx,
sound/soc/codecs/lpass-rx-macro.c
2232
if (!rx->comp_enabled[comp])
sound/soc/codecs/lpass-rx-macro.c
2246
hph_pwr_mode = rx->hph_pwr_mode;
sound/soc/codecs/lpass-rx-macro.c
2262
struct rx_macro *rx, bool enable)
sound/soc/codecs/lpass-rx-macro.c
2265
if (rx->softclip_clk_users == 0)
sound/soc/codecs/lpass-rx-macro.c
2268
rx->softclip_clk_users++;
sound/soc/codecs/lpass-rx-macro.c
2270
rx->softclip_clk_users--;
sound/soc/codecs/lpass-rx-macro.c
2271
if (rx->softclip_clk_users == 0)
sound/soc/codecs/lpass-rx-macro.c
2278
struct rx_macro *rx, int event)
sound/soc/codecs/lpass-rx-macro.c
2281
if (!rx->is_softclip_on)
sound/soc/codecs/lpass-rx-macro.c
2286
rx_macro_enable_softclip_clk(component, rx, true);
sound/soc/codecs/lpass-rx-macro.c
2295
rx_macro_enable_softclip_clk(component, rx, false);
sound/soc/codecs/lpass-rx-macro.c
2302
struct rx_macro *rx, int event)
sound/soc/codecs/lpass-rx-macro.c
2306
if (!rx->is_aux_hpf_on)
sound/soc/codecs/lpass-rx-macro.c
2308
CDC_RX_RXn_RX_PATH_CFG1(rx, 2), 0x04, 0x00);
sound/soc/codecs/lpass-rx-macro.c
2314
CDC_RX_RXn_RX_PATH_CFG1(rx, 2), 0x04, 0x04);
sound/soc/codecs/lpass-rx-macro.c
2320
static inline void rx_macro_enable_clsh_block(struct rx_macro *rx, bool enable)
sound/soc/codecs/lpass-rx-macro.c
2322
if ((enable && ++rx->clsh_users == 1) || (!enable && --rx->clsh_users == 0))
sound/soc/codecs/lpass-rx-macro.c
2323
snd_soc_component_update_bits(rx->component, CDC_RX_CLSH_CRC,
sound/soc/codecs/lpass-rx-macro.c
2325
if (rx->clsh_users < 0)
sound/soc/codecs/lpass-rx-macro.c
2326
rx->clsh_users = 0;
sound/soc/codecs/lpass-rx-macro.c
2330
struct rx_macro *rx,
sound/soc/codecs/lpass-rx-macro.c
2334
rx_macro_enable_clsh_block(rx, false);
sound/soc/codecs/lpass-rx-macro.c
2341
rx_macro_enable_clsh_block(rx, true);
sound/soc/codecs/lpass-rx-macro.c
2354
if (rx->is_ear_mode_on)
sound/soc/codecs/lpass-rx-macro.c
2366
CDC_RX_RXn_RX_PATH_CFG0(rx, 0),
sound/soc/codecs/lpass-rx-macro.c
2370
if (rx->is_ear_mode_on)
sound/soc/codecs/lpass-rx-macro.c
2382
CDC_RX_RXn_RX_PATH_CFG0(rx, 1),
sound/soc/codecs/lpass-rx-macro.c
2387
CDC_RX_RXn_RX_PATH_CFG0(rx, 2),
sound/soc/codecs/lpass-rx-macro.c
2390
CDC_RX_RXn_RX_PATH_CFG0(rx, 2),
sound/soc/codecs/lpass-rx-macro.c
2401
struct rx_macro *rx = snd_soc_component_get_drvdata(component);
sound/soc/codecs/lpass-rx-macro.c
2406
hd2_scale_reg = CDC_RX_RXn_RX_PATH_SEC3(rx, 0);
sound/soc/codecs/lpass-rx-macro.c
2407
hd2_enable_reg = CDC_RX_RXn_RX_PATH_CFG0(rx, 0);
sound/soc/codecs/lpass-rx-macro.c
2410
hd2_scale_reg = CDC_RX_RXn_RX_PATH_SEC3(rx, 1);
sound/soc/codecs/lpass-rx-macro.c
2411
hd2_enable_reg = CDC_RX_RXn_RX_PATH_CFG0(rx, 1);
sound/soc/codecs/lpass-rx-macro.c
2435
struct rx_macro *rx = snd_soc_component_get_drvdata(component);
sound/soc/codecs/lpass-rx-macro.c
2437
ucontrol->value.integer.value[0] = rx->comp_enabled[comp];
sound/soc/codecs/lpass-rx-macro.c
2447
struct rx_macro *rx = snd_soc_component_get_drvdata(component);
sound/soc/codecs/lpass-rx-macro.c
2449
rx->comp_enabled[comp] = value;
sound/soc/codecs/lpass-rx-macro.c
2459
struct rx_macro *rx = snd_soc_component_get_drvdata(component);
sound/soc/codecs/lpass-rx-macro.c
2462
rx->rx_port_value[widget->shift];
sound/soc/codecs/lpass-rx-macro.c
2476
struct rx_macro *rx = snd_soc_component_get_drvdata(component);
sound/soc/codecs/lpass-rx-macro.c
2478
aif_rst = rx->rx_port_value[widget->shift];
sound/soc/codecs/lpass-rx-macro.c
2487
rx->rx_port_value[widget->shift] = rx_port_value;
sound/soc/codecs/lpass-rx-macro.c
2496
if (rx->active_ch_cnt[dai_id]) {
sound/soc/codecs/lpass-rx-macro.c
2497
clear_bit(widget->shift, &rx->active_ch_mask[dai_id]);
sound/soc/codecs/lpass-rx-macro.c
2498
rx->active_ch_cnt[dai_id]--;
sound/soc/codecs/lpass-rx-macro.c
2507
set_bit(widget->shift, &rx->active_ch_mask[dai_id]);
sound/soc/codecs/lpass-rx-macro.c
2508
rx->active_ch_cnt[dai_id]++;
sound/soc/codecs/lpass-rx-macro.c
2547
struct rx_macro *rx = snd_soc_component_get_drvdata(component);
sound/soc/codecs/lpass-rx-macro.c
2549
ucontrol->value.integer.value[0] = rx->is_ear_mode_on;
sound/soc/codecs/lpass-rx-macro.c
2557
struct rx_macro *rx = snd_soc_component_get_drvdata(component);
sound/soc/codecs/lpass-rx-macro.c
2559
rx->is_ear_mode_on = (!ucontrol->value.integer.value[0] ? false : true);
sound/soc/codecs/lpass-rx-macro.c
2567
struct rx_macro *rx = snd_soc_component_get_drvdata(component);
sound/soc/codecs/lpass-rx-macro.c
2569
ucontrol->value.integer.value[0] = rx->hph_hd2_mode;
sound/soc/codecs/lpass-rx-macro.c
2577
struct rx_macro *rx = snd_soc_component_get_drvdata(component);
sound/soc/codecs/lpass-rx-macro.c
2579
rx->hph_hd2_mode = ucontrol->value.integer.value[0];
sound/soc/codecs/lpass-rx-macro.c
2587
struct rx_macro *rx = snd_soc_component_get_drvdata(component);
sound/soc/codecs/lpass-rx-macro.c
2589
ucontrol->value.enumerated.item[0] = rx->hph_pwr_mode;
sound/soc/codecs/lpass-rx-macro.c
2597
struct rx_macro *rx = snd_soc_component_get_drvdata(component);
sound/soc/codecs/lpass-rx-macro.c
2599
rx->hph_pwr_mode = ucontrol->value.enumerated.item[0];
sound/soc/codecs/lpass-rx-macro.c
2607
struct rx_macro *rx = snd_soc_component_get_drvdata(component);
sound/soc/codecs/lpass-rx-macro.c
2609
ucontrol->value.integer.value[0] = rx->is_softclip_on;
sound/soc/codecs/lpass-rx-macro.c
2618
struct rx_macro *rx = snd_soc_component_get_drvdata(component);
sound/soc/codecs/lpass-rx-macro.c
2620
rx->is_softclip_on = ucontrol->value.integer.value[0];
sound/soc/codecs/lpass-rx-macro.c
2629
struct rx_macro *rx = snd_soc_component_get_drvdata(component);
sound/soc/codecs/lpass-rx-macro.c
2631
ucontrol->value.integer.value[0] = rx->is_aux_hpf_on;
sound/soc/codecs/lpass-rx-macro.c
2640
struct rx_macro *rx = snd_soc_component_get_drvdata(component);
sound/soc/codecs/lpass-rx-macro.c
2642
rx->is_aux_hpf_on = ucontrol->value.integer.value[0];
sound/soc/codecs/lpass-rx-macro.c
2648
struct rx_macro *rx,
sound/soc/codecs/lpass-rx-macro.c
2669
if (rx->is_ear_mode_on)
sound/soc/codecs/lpass-rx-macro.c
2671
CDC_RX_RXn_RX_PATH_CFG1(rx, 0),
sound/soc/codecs/lpass-rx-macro.c
2681
if (rx->hph_pwr_mode)
sound/soc/codecs/lpass-rx-macro.c
2688
CDC_RX_RXn_RX_PATH_CFG1(rx, 0),
sound/soc/codecs/lpass-rx-macro.c
2703
struct rx_macro *rx = snd_soc_component_get_drvdata(component);
sound/soc/codecs/lpass-rx-macro.c
2705
main_reg = CDC_RX_RXn_RX_PATH_CTL(rx, interp_idx);
sound/soc/codecs/lpass-rx-macro.c
2706
dsm_reg = CDC_RX_RXn_RX_PATH_DSM_CTL(rx, interp_idx);
sound/soc/codecs/lpass-rx-macro.c
2707
rx_cfg2_reg = CDC_RX_RXn_RX_PATH_CFG2(rx, interp_idx);
sound/soc/codecs/lpass-rx-macro.c
2710
if (rx->main_clk_users[interp_idx] == 0) {
sound/soc/codecs/lpass-rx-macro.c
2718
rx_macro_load_compander_coeff(component, rx, interp_idx, event);
sound/soc/codecs/lpass-rx-macro.c
2719
if (rx->hph_hd2_mode)
sound/soc/codecs/lpass-rx-macro.c
2721
rx_macro_hphdelay_lutbypass(component, rx, interp_idx, event);
sound/soc/codecs/lpass-rx-macro.c
2722
rx_macro_config_compander(component, rx, interp_idx, event);
sound/soc/codecs/lpass-rx-macro.c
2724
rx_macro_config_softclip(component, rx, event);
sound/soc/codecs/lpass-rx-macro.c
2725
rx_macro_config_aux_hpf(component, rx, event);
sound/soc/codecs/lpass-rx-macro.c
2727
rx_macro_config_classh(component, rx, interp_idx, event);
sound/soc/codecs/lpass-rx-macro.c
2729
rx->main_clk_users[interp_idx]++;
sound/soc/codecs/lpass-rx-macro.c
2733
rx->main_clk_users[interp_idx]--;
sound/soc/codecs/lpass-rx-macro.c
2734
if (rx->main_clk_users[interp_idx] <= 0) {
sound/soc/codecs/lpass-rx-macro.c
2735
rx->main_clk_users[interp_idx] = 0;
sound/soc/codecs/lpass-rx-macro.c
2755
rx_macro_config_classh(component, rx, interp_idx, event);
sound/soc/codecs/lpass-rx-macro.c
2756
rx_macro_config_compander(component, rx, interp_idx, event);
sound/soc/codecs/lpass-rx-macro.c
2758
rx_macro_config_softclip(component, rx, event);
sound/soc/codecs/lpass-rx-macro.c
2759
rx_macro_config_aux_hpf(component, rx, event);
sound/soc/codecs/lpass-rx-macro.c
2761
rx_macro_hphdelay_lutbypass(component, rx, interp_idx, event);
sound/soc/codecs/lpass-rx-macro.c
2762
if (rx->hph_hd2_mode)
sound/soc/codecs/lpass-rx-macro.c
2767
return rx->main_clk_users[interp_idx];
sound/soc/codecs/lpass-rx-macro.c
2774
struct rx_macro *rx = snd_soc_component_get_drvdata(component);
sound/soc/codecs/lpass-rx-macro.c
2777
gain_reg = CDC_RX_RXn_RX_VOL_MIX_CTL(rx, w->shift);
sound/soc/codecs/lpass-rx-macro.c
2778
mix_reg = CDC_RX_RXn_RX_PATH_MIX_CTL(rx, w->shift);
sound/soc/codecs/lpass-rx-macro.c
2809
struct rx_macro *rx = snd_soc_component_get_drvdata(component);
sound/soc/codecs/lpass-rx-macro.c
2814
snd_soc_component_write_field(component, CDC_RX_RXn_RX_PATH_CFG1(rx, w->shift),
sound/soc/codecs/lpass-rx-macro.c
2816
snd_soc_component_write_field(component, CDC_RX_RXn_RX_PATH_CTL(rx, w->shift),
sound/soc/codecs/lpass-rx-macro.c
2820
snd_soc_component_write_field(component, CDC_RX_RXn_RX_PATH_CFG1(rx, w->shift),
sound/soc/codecs/lpass-rx-macro.c
3609
struct rx_macro *rx = snd_soc_component_get_drvdata(component);
sound/soc/codecs/lpass-rx-macro.c
3615
snd_soc_component_init_regmap(component, rx->regmap);
sound/soc/codecs/lpass-rx-macro.c
3617
snd_soc_component_update_bits(component, CDC_RX_RXn_RX_PATH_SEC7(rx, 0),
sound/soc/codecs/lpass-rx-macro.c
3620
snd_soc_component_update_bits(component, CDC_RX_RXn_RX_PATH_SEC7(rx, 1),
sound/soc/codecs/lpass-rx-macro.c
3623
snd_soc_component_update_bits(component, CDC_RX_RXn_RX_PATH_SEC7(rx, 2),
sound/soc/codecs/lpass-rx-macro.c
3626
snd_soc_component_update_bits(component, CDC_RX_RXn_RX_PATH_CFG3(rx, 0),
sound/soc/codecs/lpass-rx-macro.c
3629
snd_soc_component_update_bits(component, CDC_RX_RXn_RX_PATH_CFG3(rx, 1),
sound/soc/codecs/lpass-rx-macro.c
3632
snd_soc_component_update_bits(component, CDC_RX_RXn_RX_PATH_CFG3(rx, 2),
sound/soc/codecs/lpass-rx-macro.c
3636
switch (rx->codec_version) {
sound/soc/codecs/lpass-rx-macro.c
3660
rx->component = component;
sound/soc/codecs/lpass-rx-macro.c
3671
struct rx_macro *rx = to_rx_macro(hw);
sound/soc/codecs/lpass-rx-macro.c
3674
ret = clk_prepare_enable(rx->mclk);
sound/soc/codecs/lpass-rx-macro.c
3676
dev_err(rx->dev, "unable to prepare mclk\n");
sound/soc/codecs/lpass-rx-macro.c
3680
rx_macro_mclk_enable(rx, true);
sound/soc/codecs/lpass-rx-macro.c
3682
regmap_update_bits(rx->regmap, CDC_RX_CLK_RST_CTRL_SWR_CONTROL,
sound/soc/codecs/lpass-rx-macro.c
3690
struct rx_macro *rx = to_rx_macro(hw);
sound/soc/codecs/lpass-rx-macro.c
3692
regmap_update_bits(rx->regmap, CDC_RX_CLK_RST_CTRL_SWR_CONTROL,
sound/soc/codecs/lpass-rx-macro.c
3695
rx_macro_mclk_enable(rx, false);
sound/soc/codecs/lpass-rx-macro.c
3696
clk_disable_unprepare(rx->mclk);
sound/soc/codecs/lpass-rx-macro.c
3701
struct rx_macro *rx = to_rx_macro(hw);
sound/soc/codecs/lpass-rx-macro.c
3704
regmap_read(rx->regmap, CDC_RX_CLK_RST_CTRL_SWR_CONTROL, &val);
sound/soc/codecs/lpass-rx-macro.c
3724
static int rx_macro_register_mclk_output(struct rx_macro *rx)
sound/soc/codecs/lpass-rx-macro.c
3726
struct device *dev = rx->dev;
sound/soc/codecs/lpass-rx-macro.c
3733
if (rx->npl)
sound/soc/codecs/lpass-rx-macro.c
3734
parent_clk_name = __clk_get_name(rx->npl);
sound/soc/codecs/lpass-rx-macro.c
3736
parent_clk_name = __clk_get_name(rx->mclk);
sound/soc/codecs/lpass-rx-macro.c
3743
rx->hw.init = &init;
sound/soc/codecs/lpass-rx-macro.c
3744
hw = &rx->hw;
sound/soc/codecs/lpass-rx-macro.c
3745
ret = devm_clk_hw_register(rx->dev, hw);
sound/soc/codecs/lpass-rx-macro.c
3767
struct rx_macro *rx;
sound/soc/codecs/lpass-rx-macro.c
3773
rx = devm_kzalloc(dev, sizeof(*rx), GFP_KERNEL);
sound/soc/codecs/lpass-rx-macro.c
3774
if (!rx)
sound/soc/codecs/lpass-rx-macro.c
3777
rx->macro = devm_clk_get_optional(dev, "macro");
sound/soc/codecs/lpass-rx-macro.c
3778
if (IS_ERR(rx->macro))
sound/soc/codecs/lpass-rx-macro.c
3779
return dev_err_probe(dev, PTR_ERR(rx->macro), "unable to get macro clock\n");
sound/soc/codecs/lpass-rx-macro.c
3781
rx->dcodec = devm_clk_get_optional(dev, "dcodec");
sound/soc/codecs/lpass-rx-macro.c
3782
if (IS_ERR(rx->dcodec))
sound/soc/codecs/lpass-rx-macro.c
3783
return dev_err_probe(dev, PTR_ERR(rx->dcodec), "unable to get dcodec clock\n");
sound/soc/codecs/lpass-rx-macro.c
3785
rx->mclk = devm_clk_get(dev, "mclk");
sound/soc/codecs/lpass-rx-macro.c
3786
if (IS_ERR(rx->mclk))
sound/soc/codecs/lpass-rx-macro.c
3787
return dev_err_probe(dev, PTR_ERR(rx->mclk), "unable to get mclk clock\n");
sound/soc/codecs/lpass-rx-macro.c
3790
rx->npl = devm_clk_get(dev, "npl");
sound/soc/codecs/lpass-rx-macro.c
3791
if (IS_ERR(rx->npl))
sound/soc/codecs/lpass-rx-macro.c
3792
return dev_err_probe(dev, PTR_ERR(rx->npl), "unable to get npl clock\n");
sound/soc/codecs/lpass-rx-macro.c
3795
rx->fsgen = devm_clk_get(dev, "fsgen");
sound/soc/codecs/lpass-rx-macro.c
3796
if (IS_ERR(rx->fsgen))
sound/soc/codecs/lpass-rx-macro.c
3797
return dev_err_probe(dev, PTR_ERR(rx->fsgen), "unable to get fsgen clock\n");
sound/soc/codecs/lpass-rx-macro.c
3799
rx->pds = lpass_macro_pds_init(dev);
sound/soc/codecs/lpass-rx-macro.c
3800
if (IS_ERR(rx->pds))
sound/soc/codecs/lpass-rx-macro.c
3801
return PTR_ERR(rx->pds);
sound/soc/codecs/lpass-rx-macro.c
3803
ret = devm_add_action_or_reset(dev, lpass_macro_pds_exit_action, rx->pds);
sound/soc/codecs/lpass-rx-macro.c
3811
rx->codec_version = lpass_macro_get_codec_version();
sound/soc/codecs/lpass-rx-macro.c
3814
switch (rx->codec_version) {
sound/soc/codecs/lpass-rx-macro.c
3820
rx->rxn_reg_stride = 0x80;
sound/soc/codecs/lpass-rx-macro.c
3821
rx->rxn_reg_stride2 = 0xc;
sound/soc/codecs/lpass-rx-macro.c
3834
rx->rxn_reg_stride = 0xc0;
sound/soc/codecs/lpass-rx-macro.c
3835
rx->rxn_reg_stride2 = 0x0;
sound/soc/codecs/lpass-rx-macro.c
3845
dev_err(dev, "Unsupported Codec version (%d)\n", rx->codec_version);
sound/soc/codecs/lpass-rx-macro.c
3858
rx->regmap = devm_regmap_init_mmio(dev, base, reg_config);
sound/soc/codecs/lpass-rx-macro.c
3859
if (IS_ERR(rx->regmap))
sound/soc/codecs/lpass-rx-macro.c
3860
return PTR_ERR(rx->regmap);
sound/soc/codecs/lpass-rx-macro.c
3862
dev_set_drvdata(dev, rx);
sound/soc/codecs/lpass-rx-macro.c
3864
rx->dev = dev;
sound/soc/codecs/lpass-rx-macro.c
3867
clk_set_rate(rx->mclk, MCLK_FREQ);
sound/soc/codecs/lpass-rx-macro.c
3868
clk_set_rate(rx->npl, MCLK_FREQ);
sound/soc/codecs/lpass-rx-macro.c
3870
ret = clk_prepare_enable(rx->macro);
sound/soc/codecs/lpass-rx-macro.c
3874
ret = clk_prepare_enable(rx->dcodec);
sound/soc/codecs/lpass-rx-macro.c
3878
ret = clk_prepare_enable(rx->mclk);
sound/soc/codecs/lpass-rx-macro.c
3882
ret = clk_prepare_enable(rx->npl);
sound/soc/codecs/lpass-rx-macro.c
3886
ret = clk_prepare_enable(rx->fsgen);
sound/soc/codecs/lpass-rx-macro.c
3891
regmap_update_bits(rx->regmap, CDC_RX_CLK_RST_CTRL_SWR_CONTROL,
sound/soc/codecs/lpass-rx-macro.c
3895
regmap_update_bits(rx->regmap, CDC_RX_CLK_RST_CTRL_SWR_CONTROL,
sound/soc/codecs/lpass-rx-macro.c
3898
regmap_update_bits(rx->regmap, CDC_RX_CLK_RST_CTRL_SWR_CONTROL,
sound/soc/codecs/lpass-rx-macro.c
3914
ret = rx_macro_register_mclk_output(rx);
sound/soc/codecs/lpass-rx-macro.c
3921
clk_disable_unprepare(rx->fsgen);
sound/soc/codecs/lpass-rx-macro.c
3923
clk_disable_unprepare(rx->npl);
sound/soc/codecs/lpass-rx-macro.c
3925
clk_disable_unprepare(rx->mclk);
sound/soc/codecs/lpass-rx-macro.c
3927
clk_disable_unprepare(rx->dcodec);
sound/soc/codecs/lpass-rx-macro.c
3929
clk_disable_unprepare(rx->macro);
sound/soc/codecs/lpass-rx-macro.c
3936
struct rx_macro *rx = dev_get_drvdata(&pdev->dev);
sound/soc/codecs/lpass-rx-macro.c
3938
clk_disable_unprepare(rx->mclk);
sound/soc/codecs/lpass-rx-macro.c
3939
clk_disable_unprepare(rx->npl);
sound/soc/codecs/lpass-rx-macro.c
3940
clk_disable_unprepare(rx->fsgen);
sound/soc/codecs/lpass-rx-macro.c
3941
clk_disable_unprepare(rx->macro);
sound/soc/codecs/lpass-rx-macro.c
3942
clk_disable_unprepare(rx->dcodec);
sound/soc/codecs/lpass-rx-macro.c
3971
struct rx_macro *rx = dev_get_drvdata(dev);
sound/soc/codecs/lpass-rx-macro.c
3973
regcache_cache_only(rx->regmap, true);
sound/soc/codecs/lpass-rx-macro.c
3974
regcache_mark_dirty(rx->regmap);
sound/soc/codecs/lpass-rx-macro.c
3976
clk_disable_unprepare(rx->fsgen);
sound/soc/codecs/lpass-rx-macro.c
3977
clk_disable_unprepare(rx->npl);
sound/soc/codecs/lpass-rx-macro.c
3978
clk_disable_unprepare(rx->mclk);
sound/soc/codecs/lpass-rx-macro.c
3985
struct rx_macro *rx = dev_get_drvdata(dev);
sound/soc/codecs/lpass-rx-macro.c
3988
ret = clk_prepare_enable(rx->mclk);
sound/soc/codecs/lpass-rx-macro.c
3994
ret = clk_prepare_enable(rx->npl);
sound/soc/codecs/lpass-rx-macro.c
4000
ret = clk_prepare_enable(rx->fsgen);
sound/soc/codecs/lpass-rx-macro.c
4005
regcache_cache_only(rx->regmap, false);
sound/soc/codecs/lpass-rx-macro.c
4006
regcache_sync(rx->regmap);
sound/soc/codecs/lpass-rx-macro.c
4010
clk_disable_unprepare(rx->npl);
sound/soc/codecs/lpass-rx-macro.c
4012
clk_disable_unprepare(rx->mclk);
sound/soc/codecs/wcd9335.h
443
#define WCD9335_CDC_RX_PATH_CTL(rx) WCD9335_REG(0x0b, (0x041 + rx * 0x14))
sound/soc/codecs/wcd9335.h
450
#define WCD9335_CDC_RX_PATH_MIX_CTL(rx) WCD9335_REG(0x0b, (0x46 + rx * 0x14))
sound/soc/fsl/fsl_esai.c
711
bool tx = true, rx = false, enabled[2];
sound/soc/fsl/fsl_esai.c
720
enabled[rx] = rfcr & ESAI_xFCR_xFEN;
sound/soc/fsl/fsl_esai.c
724
fsl_esai_trigger_stop(esai_priv, rx);
sound/soc/fsl/fsl_esai.c
751
if (enabled[rx])
sound/soc/fsl/fsl_esai.c
752
fsl_esai_trigger_start(esai_priv, rx);
sound/soc/fsl/fsl_sai.c
1344
u32 rx, tx, type;
sound/soc/fsl/fsl_sai.c
1384
ret = of_property_read_u32_index(np, propname, index++, &rx);
sound/soc/fsl/fsl_sai.c
1392
if ((rx & ~soc_dl) || (tx & ~soc_dl)) {
sound/soc/fsl/fsl_sai.c
1397
rx = rx & soc_dl;
sound/soc/fsl/fsl_sai.c
1401
cfg[i].pins[0] = hweight8(rx);
sound/soc/fsl/fsl_sai.c
1402
cfg[i].mask[0] = rx;
sound/soc/fsl/fsl_sai.c
1403
dl_mask = rx;
sound/soc/fsl/fsl_sai.c
1405
cfg[i].next_off[0] = fsl_sai_calc_dl_off(rx);
sound/soc/intel/avs/avs.h
221
struct avs_ipc_msg rx;
sound/soc/intel/avs/ipc.c
195
ipc->rx.header = header;
sound/soc/intel/avs/ipc.c
201
ipc->rx.size = min_t(u32, AVS_MAILBOX_SIZE,
sound/soc/intel/avs/ipc.c
204
memcpy_fromio(ipc->rx.data, avs_uplink_addr(adev), ipc->rx.size);
sound/soc/intel/avs/ipc.c
205
trace_avs_msg_payload(ipc->rx.data, ipc->rx.size);
sound/soc/intel/avs/ipc.c
363
ipc->rx.header = 0;
sound/soc/intel/avs/ipc.c
364
ipc->rx.size = reply ? reply->size : 0;
sound/soc/intel/avs/ipc.c
418
ret = ipc->rx.rsp.status;
sound/soc/intel/avs/ipc.c
433
reply->header = ipc->rx.header;
sound/soc/intel/avs/ipc.c
434
reply->size = ipc->rx.size;
sound/soc/intel/avs/ipc.c
435
if (reply->data && ipc->rx.size)
sound/soc/intel/avs/ipc.c
436
memcpy(reply->data, ipc->rx.data, reply->size);
sound/soc/intel/avs/ipc.c
560
ipc->rx.data = devm_kzalloc(dev, AVS_MAILBOX_SIZE, GFP_KERNEL);
sound/soc/intel/avs/ipc.c
561
if (!ipc->rx.data)
sound/soc/intel/boards/sof_cirrus_common.c
128
ARRAY_SIZE(cs35l41_channel_map[i].rx),
sound/soc/intel/boards/sof_cirrus_common.c
129
(unsigned int *)cs35l41_channel_map[i].rx);
sound/soc/intel/boards/sof_cirrus_common.c
84
unsigned int rx[2];
sound/soc/intel/boards/sof_cirrus_common.c
86
{.rx = {0, 1}}, /* WL */
sound/soc/intel/boards/sof_cirrus_common.c
87
{.rx = {1, 0}}, /* WR */
sound/soc/intel/boards/sof_cirrus_common.c
88
{.rx = {0, 1}}, /* TL */
sound/soc/intel/boards/sof_cirrus_common.c
89
{.rx = {1, 0}}, /* TR */
sound/soc/intel/boards/sof_maxim_common.c
162
if (fls(max_98373_tdm_mask[i].rx) > tdm_slots) {
sound/soc/intel/boards/sof_maxim_common.c
164
fls(max_98373_tdm_mask[i].rx), tdm_slots);
sound/soc/intel/boards/sof_maxim_common.c
169
tx_mask, max_98373_tdm_mask[i].rx,
sound/soc/intel/boards/sof_maxim_common.c
173
max_98373_tdm_mask[i].rx,
sound/soc/intel/boards/sof_maxim_common.c
371
unsigned int rx;
sound/soc/intel/boards/sof_maxim_common.c
373
{.tx = 0x01, .rx = 0x3},
sound/soc/intel/boards/sof_maxim_common.c
374
{.tx = 0x02, .rx = 0x3},
sound/soc/intel/boards/sof_maxim_common.c
375
{.tx = 0x04, .rx = 0x3},
sound/soc/intel/boards/sof_maxim_common.c
376
{.tx = 0x08, .rx = 0x3},
sound/soc/intel/boards/sof_maxim_common.c
399
max_98390_tdm_mask[i].rx,
sound/soc/intel/boards/sof_maxim_common.c
82
unsigned int rx;
sound/soc/intel/boards/sof_maxim_common.c
84
{.rx = 0x3},
sound/soc/intel/boards/sof_maxim_common.c
85
{.rx = 0x3},
sound/soc/intel/boards/sof_realtek_common.c
128
unsigned int rx;
sound/soc/intel/boards/sof_realtek_common.c
130
{.tx = 0x4, .rx = 0x1},
sound/soc/intel/boards/sof_realtek_common.c
131
{.tx = 0x8, .rx = 0x2},
sound/soc/intel/boards/sof_realtek_common.c
132
{.tx = 0x1, .rx = 0x1},
sound/soc/intel/boards/sof_realtek_common.c
133
{.tx = 0x2, .rx = 0x2},
sound/soc/intel/boards/sof_realtek_common.c
170
rt1011_tdm_mask[i].rx, 4,
sound/soc/intel/boards/sof_realtek_common.c
389
unsigned int rx;
sound/soc/intel/boards/sof_realtek_common.c
391
{.tx = 0x0, .rx = 0x1},
sound/soc/intel/boards/sof_realtek_common.c
392
{.tx = 0x0, .rx = 0x2},
sound/soc/intel/boards/sof_realtek_common.c
436
rt1015_tdm_mask[i].rx,
sound/soc/intel/catpt/core.h
37
struct catpt_ipc_msg rx;
sound/soc/intel/catpt/ipc.c
113
ret = ipc->rx.rsp.status;
sound/soc/intel/catpt/ipc.c
117
reply->header = ipc->rx.header;
sound/soc/intel/catpt/ipc.c
120
memcpy(reply->data, ipc->rx.data, reply->size);
sound/soc/intel/catpt/ipc.c
189
ipc->rx.header = header;
sound/soc/intel/catpt/ipc.c
190
if (ipc->rx.rsp.status != CATPT_REPLY_SUCCESS)
sound/soc/intel/catpt/ipc.c
193
memcpy_fromio(ipc->rx.data, catpt_outbox_addr(cdev), ipc->rx.size);
sound/soc/intel/catpt/ipc.c
194
trace_catpt_ipc_payload(ipc->rx.data, ipc->rx.size);
sound/soc/intel/catpt/ipc.c
34
ipc->rx.data = devm_kzalloc(ipc->dev, config->outbox_size, GFP_KERNEL);
sound/soc/intel/catpt/ipc.c
35
if (!ipc->rx.data)
sound/soc/intel/catpt/ipc.c
49
ipc->rx.header = 0;
sound/soc/intel/catpt/ipc.c
50
ipc->rx.size = reply ? reply->size : 0;
sound/soc/intel/catpt/ipc.c
76
if (ipc->rx.rsp.status != CATPT_REPLY_PENDING)
sound/soc/mediatek/common/mtk-btcvsd.c
1000
bt->tx->packet_r : bt->rx->packet_w;
sound/soc/mediatek/common/mtk-btcvsd.c
1090
mtk_btcvsd_snd_set_state(bt, bt->rx, BT_SCO_STATE_LOOPBACK);
sound/soc/mediatek/common/mtk-btcvsd.c
1093
mtk_btcvsd_snd_set_state(bt, bt->rx, BT_SCO_STATE_RUNNING);
sound/soc/mediatek/common/mtk-btcvsd.c
1132
if (!bt->rx)
sound/soc/mediatek/common/mtk-btcvsd.c
1135
ucontrol->value.integer.value[0] = bt->rx->rw_cnt ? 1 : 0;
sound/soc/mediatek/common/mtk-btcvsd.c
1145
if (!bt->rx)
sound/soc/mediatek/common/mtk-btcvsd.c
1148
ucontrol->value.integer.value[0] = bt->rx->timeout;
sound/soc/mediatek/common/mtk-btcvsd.c
1149
bt->rx->timeout = 0;
sound/soc/mediatek/common/mtk-btcvsd.c
1289
btcvsd->rx = devm_kzalloc(btcvsd->dev, sizeof(*btcvsd->rx), GFP_KERNEL);
sound/soc/mediatek/common/mtk-btcvsd.c
1290
if (!btcvsd->rx)
sound/soc/mediatek/common/mtk-btcvsd.c
136
struct mtk_btcvsd_snd_stream *rx;
sound/soc/mediatek/common/mtk-btcvsd.c
1367
mtk_btcvsd_snd_set_state(btcvsd, btcvsd->rx, BT_SCO_STATE_IDLE);
sound/soc/mediatek/common/mtk-btcvsd.c
212
bt->tx->state, bt->rx->state, bt->irq_disabled);
sound/soc/mediatek/common/mtk-btcvsd.c
217
bt->rx->state == BT_SCO_STATE_IDLE) {
sound/soc/mediatek/common/mtk-btcvsd.c
247
memset(bt->rx, 0, sizeof(*bt->rx));
sound/soc/mediatek/common/mtk-btcvsd.c
250
bt->rx->packet_size = BTCVSD_RX_PACKET_SIZE;
sound/soc/mediatek/common/mtk-btcvsd.c
251
bt->rx->buf_size = BTCVSD_RX_BUF_SIZE;
sound/soc/mediatek/common/mtk-btcvsd.c
252
bt->rx->timeout = 0;
sound/soc/mediatek/common/mtk-btcvsd.c
253
bt->rx->rw_cnt = 0;
sound/soc/mediatek/common/mtk-btcvsd.c
254
bt->rx->stream = SNDRV_PCM_STREAM_CAPTURE;
sound/soc/mediatek/common/mtk-btcvsd.c
268
ts->time_stamp_us = bt->rx->time_stamp;
sound/soc/mediatek/common/mtk-btcvsd.c
269
ts->data_count_equi_time = bt->rx->buf_data_equivalent_time;
sound/soc/mediatek/common/mtk-btcvsd.c
385
bt->rx->temp_packet_buf, packet_length,
sound/soc/mediatek/common/mtk-btcvsd.c
390
packet_buf_ofs = (bt->rx->packet_w & SCO_RX_PACKET_MASK) *
sound/soc/mediatek/common/mtk-btcvsd.c
391
bt->rx->packet_size;
sound/soc/mediatek/common/mtk-btcvsd.c
393
bt->rx->temp_packet_buf + (SCO_RX_PLC_SIZE * i),
sound/soc/mediatek/common/mtk-btcvsd.c
404
bt->rx->packet_w++;
sound/soc/mediatek/common/mtk-btcvsd.c
487
if (bt->rx->state != BT_SCO_STATE_RUNNING &&
sound/soc/mediatek/common/mtk-btcvsd.c
488
bt->rx->state != BT_SCO_STATE_ENDING &&
sound/soc/mediatek/common/mtk-btcvsd.c
493
__func__, bt->rx->state, bt->tx->state);
sound/soc/mediatek/common/mtk-btcvsd.c
549
bt->rx->rw_cnt++;
sound/soc/mediatek/common/mtk-btcvsd.c
553
if (bt->rx->state == BT_SCO_STATE_RUNNING ||
sound/soc/mediatek/common/mtk-btcvsd.c
554
bt->rx->state == BT_SCO_STATE_ENDING) {
sound/soc/mediatek/common/mtk-btcvsd.c
555
if (bt->rx->xrun) {
sound/soc/mediatek/common/mtk-btcvsd.c
556
if (bt->rx->packet_w - bt->rx->packet_r <=
sound/soc/mediatek/common/mtk-btcvsd.c
562
bt->rx->xrun = 0;
sound/soc/mediatek/common/mtk-btcvsd.c
568
if (!bt->rx->xrun &&
sound/soc/mediatek/common/mtk-btcvsd.c
569
(bt->rx->packet_w - bt->rx->packet_r <=
sound/soc/mediatek/common/mtk-btcvsd.c
577
bt->rx->rw_cnt++;
sound/soc/mediatek/common/mtk-btcvsd.c
579
bt->rx->xrun = 1;
sound/soc/mediatek/common/mtk-btcvsd.c
618
if (bt->rx->state == BT_SCO_STATE_RUNNING ||
sound/soc/mediatek/common/mtk-btcvsd.c
619
bt->rx->state == BT_SCO_STATE_ENDING) {
sound/soc/mediatek/common/mtk-btcvsd.c
620
bt->rx->wait_flag = 1;
sound/soc/mediatek/common/mtk-btcvsd.c
622
snd_pcm_period_elapsed(bt->rx->substream);
sound/soc/mediatek/common/mtk-btcvsd.c
705
unsigned int packet_size = bt->rx->packet_size;
sound/soc/mediatek/common/mtk-btcvsd.c
710
avail = (bt->rx->packet_w - bt->rx->packet_r) * packet_size;
sound/soc/mediatek/common/mtk-btcvsd.c
712
cur_read_idx = (bt->rx->packet_r & SCO_RX_PACKET_MASK) *
sound/soc/mediatek/common/mtk-btcvsd.c
717
int ret = wait_for_bt_irq(bt, bt->rx);
sound/soc/mediatek/common/mtk-btcvsd.c
741
cont = bt->rx->buf_size - cur_read_idx;
sound/soc/mediatek/common/mtk-btcvsd.c
753
bt->rx->packet_r += read_size / packet_size;
sound/soc/mediatek/common/mtk-btcvsd.c
764
bt->rx->time_stamp = sched_clock();
sound/soc/mediatek/common/mtk-btcvsd.c
765
bt->rx->buf_data_equivalent_time =
sound/soc/mediatek/common/mtk-btcvsd.c
766
(unsigned long long)(bt->rx->packet_w - bt->rx->packet_r) *
sound/soc/mediatek/common/mtk-btcvsd.c
768
bt->rx->buf_data_equivalent_time += read_count * SCO_RX_PLC_SIZE *
sound/soc/mediatek/common/mtk-btcvsd.c
771
bt->rx->buf_data_equivalent_time *= 1000;
sound/soc/mediatek/common/mtk-btcvsd.c
856
return bt->rx;
sound/soc/mediatek/common/mtk-btcvsd.c
887
bt->rx->substream = substream;
sound/soc/mediatek/common/mtk-btcvsd.c
995
bt_stream = bt->rx;
sound/soc/meson/axg-card.c
16
u32 rx;
sound/soc/meson/axg-card.c
162
u32 tx, rx;
sound/soc/meson/axg-card.c
182
for (i = 0, rx = 0; i < AXG_TDM_NUM_LANES; i++) {
sound/soc/meson/axg-card.c
185
rx = max(rx, be->rx_mask[i]);
sound/soc/meson/axg-card.c
189
if (!rx)
sound/soc/meson/axg-card.c
193
if (!tx && !rx) {
sound/soc/meson/axg-card.c
204
be->slots = fls(max(tx, rx));
sound/soc/meson/axg-card.c
205
} else if (be->slots < fls(max(tx, rx)) || be->slots > 32) {
sound/soc/meson/axg-card.c
235
&codec_mask->rx);
sound/soc/meson/axg-card.c
66
be->codec_masks[i].rx,
sound/soc/meson/axg-tdm-interface.c
44
struct axg_tdm_stream *rx = snd_soc_dai_dma_data_get_capture(dai);
sound/soc/meson/axg-tdm-interface.c
90
if (rx) {
sound/soc/meson/axg-tdm-interface.c
91
rx->mask = rx_mask;
sound/soc/rockchip/rockchip_i2s_tdm.c
221
bool rx = clr & I2S_CLR_RXC;
sound/soc/rockchip/rockchip_i2s_tdm.c
223
if (!(rx || tx))
sound/soc/rockchip/rockchip_i2s_tdm.c
230
if (rx) {
sound/soc/rockchip/rockchip_i2s_tdm.c
247
tx ? "tx" : "", rx ? "rx" : "");
sound/soc/rockchip/rockchip_i2s_tdm.c
248
if (rx && tx)
sound/soc/rockchip/rockchip_i2s_tdm.c
252
else if (rx)
sound/soc/samsung/dma.h
16
const char *tx, const char *rx,
sound/soc/samsung/dmaengine.c
18
const char *tx, const char *rx,
sound/soc/samsung/dmaengine.c
32
pcm_conf->chan_names[SNDRV_PCM_STREAM_CAPTURE] = rx;
sound/soc/sof/ipc4.c
462
struct sof_ipc4_msg rx = {{ 0 }};
sound/soc/sof/ipc4.c
533
rx.primary = 0;
sound/soc/sof/ipc4.c
534
rx.extension = 0;
sound/soc/sof/ipc4.c
535
rx.data_size = chunk_size;
sound/soc/sof/ipc4.c
536
rx.data_ptr = ipc4_msg->data_ptr + offset;
sound/soc/sof/ipc4.c
551
ret = ipc4_tx_msg_unlocked(sdev->ipc, &tx, tx_size, &rx, rx_size);
sound/soc/sof/ipc4.c
559
if (!set && rx.extension & SOF_IPC4_MOD_EXT_MSG_FIRST_BLOCK_MASK) {
sound/soc/sof/ipc4.c
561
rx_size = rx.extension & SOF_IPC4_MOD_EXT_MSG_SIZE_MASK;
sound/soc/starfive/jh7110_tdm.c
122
struct tdm_chan_cfg rx;
sound/soc/starfive/jh7110_tdm.c
191
sl = max(tdm->rx.sl, tdm->tx.sl);
sound/soc/starfive/jh7110_tdm.c
192
sscale = max(tdm->rx.sscale, tdm->tx.sscale);
sound/soc/starfive/jh7110_tdm.c
201
(tdm->rx.sscale <= 1 || tdm->tx.sscale <= 1) &&
sound/soc/starfive/jh7110_tdm.c
221
datarx = (tdm->rx.ifl << IFL_BIT) |
sound/soc/starfive/jh7110_tdm.c
222
(tdm->rx.wl << WL_BIT) |
sound/soc/starfive/jh7110_tdm.c
223
(tdm->rx.sscale << SSCALE_BIT) |
sound/soc/starfive/jh7110_tdm.c
224
(tdm->rx.sl << SL_BIT) |
sound/soc/starfive/jh7110_tdm.c
225
(tdm->rx.lrj << LRJ_BIT);
sound/soc/starfive/jh7110_tdm.c
383
tdm->rx.wl = chan_wl;
sound/soc/starfive/jh7110_tdm.c
384
tdm->rx.sl = chan_sl;
sound/soc/starfive/jh7110_tdm.c
385
tdm->rx.sscale = chan_nr;
sound/soc/starfive/jh7110_tdm.c
529
tdm->rx.ifl = TDM_FIFO_HALF;
sound/soc/starfive/jh7110_tdm.c
531
tdm->rx.wl = TDM_16BIT_WORD_LEN;
sound/soc/starfive/jh7110_tdm.c
533
tdm->rx.sscale = 2;
sound/soc/starfive/jh7110_tdm.c
535
tdm->rx.lrj = TDM_LEFT_JUSTIFT;
sound/soc/ti/omap-mcbsp.c
417
int rx = !tx;
sound/soc/ti/omap-mcbsp.c
441
rx &= 1;
sound/soc/ti/omap-mcbsp.c
443
MCBSP_WRITE(mcbsp, SPCR1, w | rx);
sound/soc/ti/omap-mcbsp.c
465
w &= ~(rx ? RDISABLE : 0);
sound/soc/ti/omap-mcbsp.c
476
int rx = !tx;
sound/soc/ti/omap-mcbsp.c
491
rx &= 1;
sound/soc/ti/omap-mcbsp.c
494
w |= (rx ? RDISABLE : 0);
sound/soc/ti/omap-mcbsp.c
498
MCBSP_WRITE(mcbsp, SPCR1, w & ~rx);
tools/include/uapi/linux/if_xdp.h
67
struct xdp_ring_offset rx;
tools/net/ynl/samples/ethtool.c
49
if (dev->_present.rx)
tools/net/ynl/samples/ethtool.c
50
printf("rx %d ", dev->rx);
tools/spi/spidev_test.c
119
static void transfer(int fd, uint8_t const *tx, uint8_t const *rx, size_t len)
tools/spi/spidev_test.c
125
.rx_buf = (unsigned long)rx,
tools/spi/spidev_test.c
164
ret = write(out_fd, rx, len);
tools/spi/spidev_test.c
172
hex_dump(rx, len, 32, "RX");
tools/spi/spidev_test.c
344
uint8_t *rx;
tools/spi/spidev_test.c
350
rx = malloc(size);
tools/spi/spidev_test.c
351
if (!rx)
tools/spi/spidev_test.c
355
transfer(fd, tx, rx, size);
tools/spi/spidev_test.c
356
free(rx);
tools/spi/spidev_test.c
366
uint8_t *rx;
tools/spi/spidev_test.c
379
rx = malloc(sb.st_size);
tools/spi/spidev_test.c
380
if (!rx)
tools/spi/spidev_test.c
387
transfer(fd, tx, rx, sb.st_size);
tools/spi/spidev_test.c
388
free(rx);
tools/spi/spidev_test.c
413
uint8_t *rx;
tools/spi/spidev_test.c
422
rx = malloc(len);
tools/spi/spidev_test.c
423
if (!rx)
tools/spi/spidev_test.c
426
transfer(fd, tx, rx, len);
tools/spi/spidev_test.c
432
if (memcmp(tx, rx, len)) {
tools/spi/spidev_test.c
435
hex_dump(rx, len, 32, "RX");
tools/spi/spidev_test.c
440
free(rx);
tools/testing/selftests/bpf/prog_tests/flow_dissector_classification.c
449
int fdt = -1, fdr = -1, len, tx = 0, rx = 0, err;
tools/testing/selftests/bpf/prog_tests/flow_dissector_classification.c
474
rx += err;
tools/testing/selftests/bpf/prog_tests/flow_dissector_classification.c
478
if (rx < tx) {
tools/testing/selftests/bpf/prog_tests/flow_dissector_classification.c
480
while (rx < tx) {
tools/testing/selftests/bpf/prog_tests/flow_dissector_classification.c
490
rx += err;
tools/testing/selftests/bpf/prog_tests/flow_dissector_classification.c
497
return rx;
tools/testing/selftests/bpf/prog_tests/test_xsk.c
1024
rcvd = xsk_ring_cons__peek(&xsk->rx, xsk->batch_size, &idx_rx);
tools/testing/selftests/bpf/prog_tests/test_xsk.c
1041
const struct xdp_desc *desc = xsk_ring_cons__rx_desc(&xsk->rx, idx_rx++);
tools/testing/selftests/bpf/prog_tests/test_xsk.c
1087
xsk_ring_cons__cancel(&xsk->rx, nb_frags);
tools/testing/selftests/bpf/prog_tests/test_xsk.c
1100
xsk_ring_cons__release(&xsk->rx, frags_processed);
tools/testing/selftests/bpf/prog_tests/test_xsk.c
178
rxr = ifobject->rx_on ? &xsk->rx : NULL;
tools/testing/selftests/bpf/prog_tests/test_xsk.h
63
struct xsk_ring_cons rx;
tools/testing/selftests/bpf/prog_tests/xdp_metadata.c
284
ret = xsk_ring_cons__peek(&xsk->rx, 1, &idx);
tools/testing/selftests/bpf/prog_tests/xdp_metadata.c
288
rx_desc = xsk_ring_cons__rx_desc(&xsk->rx, idx);
tools/testing/selftests/bpf/prog_tests/xdp_metadata.c
331
xsk_ring_cons__release(&xsk->rx, 1);
tools/testing/selftests/bpf/prog_tests/xdp_metadata.c
449
bpf_program__fd(bpf_obj->progs.rx),
tools/testing/selftests/bpf/prog_tests/xdp_metadata.c
55
struct xsk_ring_cons rx;
tools/testing/selftests/bpf/prog_tests/xdp_metadata.c
94
&xsk->rx,
tools/testing/selftests/bpf/prog_tests/xsk.c
46
int configure_ifobj(struct ifobject *tx, struct ifobject *rx)
tools/testing/selftests/bpf/prog_tests/xsk.c
48
rx->ifindex = if_nametoindex(VETH_RX);
tools/testing/selftests/bpf/prog_tests/xsk.c
49
if (!ASSERT_OK_FD(rx->ifindex, "get RX ifindex"))
tools/testing/selftests/bpf/prog_tests/xsk.c
57
rx->shared_umem = false;
tools/testing/selftests/bpf/progs/xdp_hw_metadata.c
28
int rx(struct xdp_md *ctx)
tools/testing/selftests/bpf/progs/xdp_metadata.c
38
int rx(struct xdp_md *ctx)
tools/testing/selftests/bpf/xdp_hw_metadata.c
126
&xsk->rx,
tools/testing/selftests/bpf/xdp_hw_metadata.c
495
ret = xsk_ring_cons__peek(&xsk->rx, 1, &idx);
tools/testing/selftests/bpf/xdp_hw_metadata.c
500
rx_desc = xsk_ring_cons__rx_desc(&xsk->rx, idx);
tools/testing/selftests/bpf/xdp_hw_metadata.c
534
xsk_ring_cons__release(&xsk->rx, 1);
tools/testing/selftests/bpf/xdp_hw_metadata.c
617
opts.old_prog_fd = bpf_program__fd(bpf_obj->progs.rx);
tools/testing/selftests/bpf/xdp_hw_metadata.c
64
struct xsk_ring_cons rx;
tools/testing/selftests/bpf/xdp_hw_metadata.c
876
bpf_program__fd(bpf_obj->progs.rx),
tools/testing/selftests/bpf/xsk.c
541
struct xsk_ring_cons *rx,
tools/testing/selftests/bpf/xsk.c
555
if (!umem || !xsk_ptr || !(rx || tx))
tools/testing/selftests/bpf/xsk.c
595
if (rx && !rx_setup_done) {
tools/testing/selftests/bpf/xsk.c
624
if (rx) {
tools/testing/selftests/bpf/xsk.c
625
rx_map = mmap(NULL, off.rx.desc +
tools/testing/selftests/bpf/xsk.c
634
rx->mask = xsk->config.rx_size - 1;
tools/testing/selftests/bpf/xsk.c
635
rx->size = xsk->config.rx_size;
tools/testing/selftests/bpf/xsk.c
636
rx->producer = rx_map + off.rx.producer;
tools/testing/selftests/bpf/xsk.c
637
rx->consumer = rx_map + off.rx.consumer;
tools/testing/selftests/bpf/xsk.c
638
rx->flags = rx_map + off.rx.flags;
tools/testing/selftests/bpf/xsk.c
639
rx->ring = rx_map + off.rx.desc;
tools/testing/selftests/bpf/xsk.c
640
rx->cached_prod = *rx->producer;
tools/testing/selftests/bpf/xsk.c
641
rx->cached_cons = *rx->consumer;
tools/testing/selftests/bpf/xsk.c
643
xsk->rx = rx;
tools/testing/selftests/bpf/xsk.c
695
if (rx)
tools/testing/selftests/bpf/xsk.c
696
munmap(rx_map, off.rx.desc +
tools/testing/selftests/bpf/xsk.c
710
struct xsk_ring_cons *rx, struct xsk_ring_prod *tx,
tools/testing/selftests/bpf/xsk.c
717
rx, tx, umem->fill_save,
tools/testing/selftests/bpf/xsk.c
764
if (xsk->rx) {
tools/testing/selftests/bpf/xsk.c
765
munmap(xsk->rx->ring - off.rx.desc,
tools/testing/selftests/bpf/xsk.c
766
off.rx.desc + xsk->config.rx_size * desc_sz);
tools/testing/selftests/bpf/xsk.c
79
struct xsk_ring_cons *rx;
tools/testing/selftests/bpf/xsk.h
227
struct xsk_ring_cons *rx,
tools/testing/selftests/bpf/xsk.h
233
struct xsk_ring_cons *rx,
tools/testing/selftests/bpf/xsk.h
74
xsk_ring_cons__rx_desc(const struct xsk_ring_cons *rx, __u32 idx)
tools/testing/selftests/bpf/xsk.h
76
const struct xdp_desc *descs = (const struct xdp_desc *)rx->ring;
tools/testing/selftests/bpf/xsk.h
78
return &descs[idx & rx->mask];
tools/testing/selftests/drivers/net/hw/ncdevmem.c
555
static int check_changing_channels(unsigned int rx, unsigned int tx)
tools/testing/selftests/drivers/net/hw/ncdevmem.c
564
fprintf(stderr, "setting channel count rx:%u tx:%u\n", rx, tx);
tools/testing/selftests/drivers/net/hw/ncdevmem.c
601
if (rx == tx) {
tools/testing/selftests/drivers/net/hw/ncdevmem.c
602
ethtool_channels_set_req_set_combined_count(schan, rx);
tools/testing/selftests/drivers/net/hw/ncdevmem.c
603
} else if (rx > tx) {
tools/testing/selftests/drivers/net/hw/ncdevmem.c
605
ethtool_channels_set_req_set_rx_count(schan, rx - tx);
tools/testing/selftests/drivers/net/hw/ncdevmem.c
607
ethtool_channels_set_req_set_combined_count(schan, rx);
tools/testing/selftests/drivers/net/hw/ncdevmem.c
608
ethtool_channels_set_req_set_tx_count(schan, tx - rx);
tools/testing/selftests/drivers/net/hw/ncdevmem.c
612
ethtool_channels_set_req_set_rx_count(schan, rx);
tools/testing/selftests/drivers/net/psp_responder.c
37
unsigned char rx;
tools/testing/selftests/drivers/net/psp_responder.c
57
psp_rx_assoc_req_set_version(req, psp_vers.rx);
tools/testing/selftests/net/tls.c
2900
__u8 tx[1024], rx[2000];
tools/testing/selftests/net/tls.c
2931
ret = recv(fd, rx, sizeof(rx), 0);
tools/testing/selftests/net/tls.c
2938
total_plaintext_rx = parse_tls_records(_metadata, rx, ret, overhead,
tools/testing/selftests/net/tls.c
2950
__u8 tx[1024], rx[2000];
tools/testing/selftests/net/tls.c
2987
ret = recv(fd, rx, sizeof(rx), 0);
tools/testing/selftests/net/tls.c
2994
total_plaintext_rx = parse_tls_records(_metadata, rx, ret, overhead,