crypto/heimdal/appl/ftp/ftp/security.c
217
int rx = 0;
crypto/heimdal/appl/ftp/ftp/security.c
229
rx += len;
crypto/heimdal/appl/ftp/ftp/security.c
239
if(rx)
crypto/heimdal/appl/ftp/ftp/security.c
241
return rx;
crypto/heimdal/appl/ftp/ftp/security.c
245
rx += len;
crypto/heimdal/appl/ftp/ftp/security.c
248
return rx;
crypto/heimdal/appl/ftp/ftpd/security.c
217
int rx = 0;
crypto/heimdal/appl/ftp/ftpd/security.c
229
rx += len;
crypto/heimdal/appl/ftp/ftpd/security.c
239
if(rx)
crypto/heimdal/appl/ftp/ftpd/security.c
241
return rx;
crypto/heimdal/appl/ftp/ftpd/security.c
245
rx += len;
crypto/heimdal/appl/ftp/ftpd/security.c
248
return rx;
crypto/libecc/src/sig/bip0340.c
1022
fp_uninit(&rx);
crypto/libecc/src/sig/bip0340.c
1037
fp rx;
crypto/libecc/src/sig/bip0340.c
1059
rx.magic = WORD(0);
crypto/libecc/src/sig/bip0340.c
1169
ret = fp_init(&rx, pub_key->params->ec_curve.a.ctx); EG(ret, err);
crypto/libecc/src/sig/bip0340.c
1170
ret = fp_import_from_buf(&rx, &sig[0], p_len); EG(ret, err);
crypto/libecc/src/sig/bip0340.c
1175
dbg_nn_print("r", &(rx.fp_val));
crypto/libecc/src/sig/bip0340.c
1192
ret = fp_copy(&(R->X), &rx); EG(ret, err);
crypto/libecc/src/sig/bip0340.c
1193
ret = aff_pt_y_from_x(&(R->Y), &(R->Z), &rx, shortw_curve); EG(ret, err);
crypto/libecc/src/sig/bip0340.c
1291
fp_uninit(&rx);
crypto/libecc/src/sig/bip0340.c
391
fp *rx;
crypto/libecc/src/sig/bip0340.c
414
rx = &(ctx->verify_data.bip0340.r);
crypto/libecc/src/sig/bip0340.c
427
ret = fp_init(rx, ctx->pub_key->params->ec_curve.a.ctx); EG(ret, err);
crypto/libecc/src/sig/bip0340.c
428
ret = fp_import_from_buf(rx, &sig[0], p_len); EG(ret, err);
crypto/libecc/src/sig/bip0340.c
433
dbg_nn_print("r", &(rx->fp_val));
crypto/libecc/src/sig/bip0340.c
448
PTR_NULLIFY(rx);
crypto/libecc/src/sig/bip0340.c
817
fp rx;
crypto/libecc/src/sig/bip0340.c
837
rx.magic = WORD(0);
crypto/libecc/src/sig/bip0340.c
917
ret = fp_init(&rx, pub_key->params->ec_curve.a.ctx); EG(ret, err);
crypto/libecc/src/sig/bip0340.c
918
ret = fp_import_from_buf(&rx, &sig[0], p_len); EG(ret, err);
crypto/libecc/src/sig/bip0340.c
923
dbg_nn_print("r", &(rx.fp_val));
crypto/libecc/src/sig/bip0340.c
937
ret = fp_copy(&(R->X), &rx); EG(ret, err);
crypto/libecc/src/sig/bip0340.c
938
ret = aff_pt_y_from_x(&(R->Y), &(R->Z), &rx, shortw_curve); EG(ret, err);
crypto/libecc/src/sig/ecfsdsa.c
424
fp rx, ry;
crypto/libecc/src/sig/ecfsdsa.c
427
rx.magic = ry.magic = WORD(0);
crypto/libecc/src/sig/ecfsdsa.c
453
ret = fp_init(&rx, ctx->pub_key->params->ec_curve.a.ctx); EG(ret, err);
crypto/libecc/src/sig/ecfsdsa.c
454
ret = fp_import_from_buf(&rx, r, p_len); EG(ret, err);
crypto/libecc/src/sig/ecfsdsa.c
459
ret = is_on_shortw_curve(&rx, &ry, &(ctx->pub_key->params->ec_curve), &on_curve); EG(ret, err);
crypto/libecc/src/sig/ecfsdsa.c
487
fp_uninit(&rx);
crypto/libecc/src/sig/ecrdsa.c
206
nn s, rx, ke, k, r, e;
crypto/libecc/src/sig/ecrdsa.c
214
s.magic = rx.magic = ke.magic = WORD(0);
crypto/libecc/src/sig/ecrdsa.c
325
ret = nn_mod_mul(&rx, &r, x, q); EG(ret, err);
crypto/libecc/src/sig/ecrdsa.c
327
ret = nn_mod_add(&s, &rx, &ke, q); EG(ret, err);
crypto/libecc/src/sig/ecrdsa.c
353
nn_uninit(&rx);
crypto/libecc/src/sig/fuzzing_ecrdsa.c
100
tmp.magic = s.magic = rx.magic = ke.magic = WORD(0);
crypto/libecc/src/sig/fuzzing_ecrdsa.c
221
ret = nn_mod_mul(&rx, &r, x, q); EG(ret, err);
crypto/libecc/src/sig/fuzzing_ecrdsa.c
223
ret = nn_mod_add(&s, &rx, &ke, q); EG(ret, err);
crypto/libecc/src/sig/fuzzing_ecrdsa.c
250
nn_uninit(&rx);
crypto/libecc/src/sig/fuzzing_ecrdsa.c
93
nn tmp, s, rx, ke, k, r, e;
crypto/openssl/doc/designs/ddd/ddd-01-conn-blocking.c
174
l = rx(b, buf, sizeof(buf));
crypto/openssl/doc/designs/ddd/ddd-02-conn-nonblocking-threads.c
311
l = rx(conn, rx_buf, sizeof(rx_buf));
crypto/openssl/doc/designs/ddd/ddd-02-conn-nonblocking.c
400
l = rx(conn, rx_buf, sizeof(rx_buf));
crypto/openssl/doc/designs/ddd/ddd-03-fd-blocking.c
200
l = rx(ssl, buf, sizeof(buf));
crypto/openssl/doc/designs/ddd/ddd-04-fd-nonblocking.c
415
l = rx(conn, rx_buf, sizeof(rx_buf));
crypto/openssl/doc/designs/ddd/ddd-05-mem-nonblocking.c
438
l = rx(conn, rx_buf, sizeof(rx_buf));
lib/libifconfig/libifconfig_sfp.c
412
ss->channel[0].rx = get_sff_channel(ii, SFF_8472_DIAG, SFF_8472_RX_POWER);
lib/libifconfig/libifconfig_sfp.c
451
ss->channel[chan].rx =
lib/libifconfig/libifconfig_sfp.h
64
uint16_t rx; /**< channel receive power, LSB 0.1uW */
lib/msun/src/catrig.c
152
do_hard_work(double x, double y, double *rx, int *B_is_usable, double *B,
lib/msun/src/catrig.c
181
*rx = sqrt(x);
lib/msun/src/catrig.c
188
*rx = log1p(Am1 + sqrt(Am1 * (A + 1)));
lib/msun/src/catrig.c
194
*rx = x / sqrt((1 - y) * (1 + y));
lib/msun/src/catrig.c
199
*rx = log1p((y - 1) + sqrt((y - 1) * (y + 1)));
lib/msun/src/catrig.c
202
*rx = log(A + sqrt(A * A - 1));
lib/msun/src/catrig.c
276
double x, y, ax, ay, rx, ry, B, sqrt_A2my2, new_y;
lib/msun/src/catrig.c
322
do_hard_work(ax, ay, &rx, &B_is_usable, &B, &sqrt_A2my2, &new_y);
lib/msun/src/catrig.c
327
return (CMPLX(copysign(rx, x), copysign(ry, y)));
lib/msun/src/catrig.c
357
double x, y, ax, ay, rx, ry, B, sqrt_A2mx2, new_x;
lib/msun/src/catrig.c
390
rx = fabs(cimag(w));
lib/msun/src/catrig.c
394
return (CMPLX(rx, ry));
lib/msun/src/catrig.c
410
rx = acos(B);
lib/msun/src/catrig.c
412
rx = acos(-B);
lib/msun/src/catrig.c
415
rx = atan2(sqrt_A2mx2, new_x);
lib/msun/src/catrig.c
417
rx = atan2(sqrt_A2mx2, -new_x);
lib/msun/src/catrig.c
421
return (CMPLX(rx, ry));
lib/msun/src/catrig.c
432
double rx, ry;
lib/msun/src/catrig.c
435
rx = creal(w);
lib/msun/src/catrig.c
438
if (isnan(rx) && isnan(ry))
lib/msun/src/catrig.c
439
return (CMPLX(ry, rx));
lib/msun/src/catrig.c
442
if (isnan(rx))
lib/msun/src/catrig.c
443
return (CMPLX(fabs(ry), rx));
lib/msun/src/catrig.c
447
return (CMPLX(fabs(ry), copysign(rx, cimag(z))));
lib/msun/src/catrig.c
573
double x, y, ax, ay, rx, ry;
lib/msun/src/catrig.c
619
rx = (m_ln2 - log(ay)) / 2;
lib/msun/src/catrig.c
621
rx = log1p(4 * ax / sum_squares(ax - 1, ay)) / 4;
lib/msun/src/catrig.c
630
return (CMPLX(copysign(rx, x), copysign(ry, y)));
lib/msun/src/catrigf.c
102
*rx = sqrtf(x);
lib/msun/src/catrigf.c
105
*rx = log1pf(Am1 + sqrtf(Am1 * (A + 1)));
lib/msun/src/catrigf.c
107
*rx = x / sqrtf((1 - y) * (1 + y));
lib/msun/src/catrigf.c
109
*rx = log1pf((y - 1) + sqrtf((y - 1) * (y + 1)));
lib/msun/src/catrigf.c
112
*rx = logf(A + sqrtf(A * A - 1));
lib/msun/src/catrigf.c
147
float x, y, ax, ay, rx, ry, B, sqrt_A2my2, new_y;
lib/msun/src/catrigf.c
183
do_hard_work(ax, ay, &rx, &B_is_usable, &B, &sqrt_A2my2, &new_y);
lib/msun/src/catrigf.c
188
return (CMPLXF(copysignf(rx, x), copysignf(ry, y)));
lib/msun/src/catrigf.c
202
float x, y, ax, ay, rx, ry, B, sqrt_A2mx2, new_x;
lib/msun/src/catrigf.c
226
rx = fabsf(cimagf(w));
lib/msun/src/catrigf.c
230
return (CMPLXF(rx, ry));
lib/msun/src/catrigf.c
244
rx = acosf(B);
lib/msun/src/catrigf.c
246
rx = acosf(-B);
lib/msun/src/catrigf.c
249
rx = atan2f(sqrt_A2mx2, new_x);
lib/msun/src/catrigf.c
251
rx = atan2f(sqrt_A2mx2, -new_x);
lib/msun/src/catrigf.c
255
return (CMPLXF(rx, ry));
lib/msun/src/catrigf.c
262
float rx, ry;
lib/msun/src/catrigf.c
265
rx = crealf(w);
lib/msun/src/catrigf.c
267
if (isnan(rx) && isnan(ry))
lib/msun/src/catrigf.c
268
return (CMPLXF(ry, rx));
lib/msun/src/catrigf.c
269
if (isnan(rx))
lib/msun/src/catrigf.c
270
return (CMPLXF(fabsf(ry), rx));
lib/msun/src/catrigf.c
273
return (CMPLXF(fabsf(ry), copysignf(rx, cimagf(z))));
lib/msun/src/catrigf.c
340
float x, y, ax, ay, rx, ry;
lib/msun/src/catrigf.c
372
rx = (m_ln2 - logf(ay)) / 2;
lib/msun/src/catrigf.c
374
rx = log1pf(4 * ax / sum_squares(ax - 1, ay)) / 4;
lib/msun/src/catrigf.c
383
return (CMPLXF(copysignf(rx, x), copysignf(ry, y)));
lib/msun/src/catrigf.c
87
do_hard_work(float x, float y, float *rx, int *B_is_usable, float *B,
lib/msun/src/catrigl.c
106
do_hard_work(long double x, long double y, long double *rx, int *B_is_usable,
lib/msun/src/catrigl.c
121
*rx = sqrtl(x);
lib/msun/src/catrigl.c
124
*rx = log1pl(Am1 + sqrtl(Am1 * (A + 1)));
lib/msun/src/catrigl.c
126
*rx = x / sqrtl((1 - y) * (1 + y));
lib/msun/src/catrigl.c
128
*rx = log1pl((y - 1) + sqrtl((y - 1) * (y + 1)));
lib/msun/src/catrigl.c
131
*rx = logl(A + sqrtl(A * A - 1));
lib/msun/src/catrigl.c
166
long double x, y, ax, ay, rx, ry, B, sqrt_A2my2, new_y;
lib/msun/src/catrigl.c
202
do_hard_work(ax, ay, &rx, &B_is_usable, &B, &sqrt_A2my2, &new_y);
lib/msun/src/catrigl.c
207
return (CMPLXL(copysignl(rx, x), copysignl(ry, y)));
lib/msun/src/catrigl.c
222
long double x, y, ax, ay, rx, ry, B, sqrt_A2mx2, new_x;
lib/msun/src/catrigl.c
246
rx = fabsl(cimagl(w));
lib/msun/src/catrigl.c
250
return (CMPLXL(rx, ry));
lib/msun/src/catrigl.c
264
rx = acosl(B);
lib/msun/src/catrigl.c
266
rx = acosl(-B);
lib/msun/src/catrigl.c
269
rx = atan2l(sqrt_A2mx2, new_x);
lib/msun/src/catrigl.c
271
rx = atan2l(sqrt_A2mx2, -new_x);
lib/msun/src/catrigl.c
275
return (CMPLXL(rx, ry));
lib/msun/src/catrigl.c
282
long double rx, ry;
lib/msun/src/catrigl.c
285
rx = creall(w);
lib/msun/src/catrigl.c
287
if (isnan(rx) && isnan(ry))
lib/msun/src/catrigl.c
288
return (CMPLXL(ry, rx));
lib/msun/src/catrigl.c
289
if (isnan(rx))
lib/msun/src/catrigl.c
290
return (CMPLXL(fabsl(ry), rx));
lib/msun/src/catrigl.c
293
return (CMPLXL(fabsl(ry), copysignl(rx, cimagl(z))));
lib/msun/src/catrigl.c
361
long double x, y, ax, ay, rx, ry;
lib/msun/src/catrigl.c
393
rx = (m_ln2 - logl(ay)) / 2;
lib/msun/src/catrigl.c
395
rx = log1pl(4 * ax / sum_squares(ax - 1, ay)) / 4;
lib/msun/src/catrigl.c
404
return (CMPLXL(copysignl(rx, x), copysignl(ry, y)));
lib/msun/src/s_csqrt.c
103
rx = scale * fabs(b) / (2 * t);
lib/msun/src/s_csqrt.c
107
return (CMPLX(rx, ry));
lib/msun/src/s_csqrt.c
42
double a, b, rx, ry, scale, t;
lib/msun/src/s_csqrt.c
99
rx = scale * t;
lib/msun/src/s_csqrtl.c
114
rx = scale * t;
lib/msun/src/s_csqrtl.c
118
rx = scale * fabsl(b) / (2 * t);
lib/msun/src/s_csqrtl.c
122
return (CMPLXL(rx, ry));
lib/msun/src/s_csqrtl.c
57
long double a, b, rx, ry, scale, t;
sbin/ifconfig/sfp.c
101
uint16_t rx = status.channel[chan].rx;
sbin/ifconfig/sfp.c
105
chan + 1, power_mW(rx), power_dBm(rx), bias_mA(tx));
sys/arm/allwinner/if_awg.c
1088
bus_dmamap_sync(sc->rx.desc_tag, sc->rx.desc_map,
sys/arm/allwinner/if_awg.c
1091
for (i = sc->rx.cur; ; i = RX_NEXT(i)) {
sys/arm/allwinner/if_awg.c
1092
val = le32toh(sc->rx.desc_ring[i].status);
sys/arm/allwinner/if_awg.c
1097
sc->rx.cur = i;
sys/arm/allwinner/if_awg.c
1098
bus_dmamap_sync(sc->rx.desc_tag, sc->rx.desc_map,
sys/arm/allwinner/if_awg.c
1203
bus_dmamap_sync(sc->rx.desc_tag, sc->rx.desc_map,
sys/arm/allwinner/if_awg.c
1206
for (index = sc->rx.cur; ; index = RX_NEXT(index)) {
sys/arm/allwinner/if_awg.c
1207
status = le32toh(sc->rx.desc_ring[index].status);
sys/arm/allwinner/if_awg.c
1220
m = sc->rx.buf_map[index].mbuf;
sys/arm/allwinner/if_awg.c
1264
if (index != sc->rx.cur) {
sys/arm/allwinner/if_awg.c
1265
bus_dmamap_sync(sc->rx.desc_tag, sc->rx.desc_map,
sys/arm/allwinner/if_awg.c
1275
sc->rx.cur = index;
sys/arm/allwinner/if_awg.c
208
struct awg_rxring rx;
sys/arm/allwinner/if_awg.c
476
uint32_t tx, rx;
sys/arm/allwinner/if_awg.c
481
rx = RD4(sc, EMAC_RX_CTL_0);
sys/arm/allwinner/if_awg.c
484
rx |= RX_EN | CHECK_CRC;
sys/arm/allwinner/if_awg.c
487
rx &= ~(RX_EN | CHECK_CRC);
sys/arm/allwinner/if_awg.c
491
WR4(sc, EMAC_RX_CTL_0, rx);
sys/arm/allwinner/if_awg.c
733
sc->rx.desc_ring[index].addr = htole32((uint32_t)paddr);
sys/arm/allwinner/if_awg.c
734
sc->rx.desc_ring[index].size = htole32(size);
sys/arm/allwinner/if_awg.c
735
sc->rx.desc_ring[index].status = htole32(status);
sys/arm/allwinner/if_awg.c
742
sc->rx.desc_ring[index].status = htole32(RX_DESC_CTL);
sys/arm/allwinner/if_awg.c
760
if (bus_dmamap_load_mbuf_sg(sc->rx.buf_tag, sc->rx.buf_spare_map,
sys/arm/allwinner/if_awg.c
766
if (sc->rx.buf_map[index].mbuf != NULL) {
sys/arm/allwinner/if_awg.c
767
bus_dmamap_sync(sc->rx.buf_tag, sc->rx.buf_map[index].map,
sys/arm/allwinner/if_awg.c
769
bus_dmamap_unload(sc->rx.buf_tag, sc->rx.buf_map[index].map);
sys/arm/allwinner/if_awg.c
771
map = sc->rx.buf_map[index].map;
sys/arm/allwinner/if_awg.c
772
sc->rx.buf_map[index].map = sc->rx.buf_spare_map;
sys/arm/allwinner/if_awg.c
773
sc->rx.buf_spare_map = map;
sys/arm/allwinner/if_awg.c
774
bus_dmamap_sync(sc->rx.buf_tag, sc->rx.buf_map[index].map,
sys/arm/allwinner/if_awg.c
777
sc->rx.buf_map[index].mbuf = m;
sys/arm/allwinner/if_awg.c
872
&sc->rx.desc_tag);
sys/arm/allwinner/if_awg.c
878
error = bus_dmamem_alloc(sc->rx.desc_tag, (void **)&sc->rx.desc_ring,
sys/arm/allwinner/if_awg.c
879
BUS_DMA_COHERENT | BUS_DMA_WAITOK | BUS_DMA_ZERO, &sc->rx.desc_map);
sys/arm/allwinner/if_awg.c
885
error = bus_dmamap_load(sc->rx.desc_tag, sc->rx.desc_map,
sys/arm/allwinner/if_awg.c
886
sc->rx.desc_ring, RX_DESC_SIZE, awg_dmamap_cb,
sys/arm/allwinner/if_awg.c
887
&sc->rx.desc_ring_paddr, 0);
sys/arm/allwinner/if_awg.c
903
&sc->rx.buf_tag);
sys/arm/allwinner/if_awg.c
909
error = bus_dmamap_create(sc->rx.buf_tag, 0, &sc->rx.buf_spare_map);
sys/arm/allwinner/if_awg.c
917
sc->rx.desc_ring[i].next =
sys/arm/allwinner/if_awg.c
918
htole32(sc->rx.desc_ring_paddr + DESC_OFF(RX_NEXT(i)));
sys/arm/allwinner/if_awg.c
920
error = bus_dmamap_create(sc->rx.buf_tag, 0,
sys/arm/allwinner/if_awg.c
921
&sc->rx.buf_map[i].map);
sys/arm/allwinner/if_awg.c
926
sc->rx.buf_map[i].mbuf = NULL;
sys/arm/allwinner/if_awg.c
933
bus_dmamap_sync(sc->rx.desc_tag, sc->rx.desc_map,
sys/arm/allwinner/if_awg.c
938
WR4(sc, EMAC_RX_DMA_LIST, sc->rx.desc_ring_paddr);
sys/arm/ti/cpsw/if_cpsw.c
1163
if (!sc->swsc->rx.running && !sc->swsc->tx.running) {
sys/arm/ti/cpsw/if_cpsw.c
1238
sc->rx.teardown = 1;
sys/arm/ti/cpsw/if_cpsw.c
1241
while (sc->rx.running) {
sys/arm/ti/cpsw/if_cpsw.c
1249
if (!sc->rx.running)
sys/arm/ti/cpsw/if_cpsw.c
1578
if (sc->rx.teardown) {
sys/arm/ti/cpsw/if_cpsw.c
1579
sc->rx.running = 0;
sys/arm/ti/cpsw/if_cpsw.c
1580
sc->rx.teardown = 0;
sys/arm/ti/cpsw/if_cpsw.c
1581
cpsw_write_cp(sc, &sc->rx, 0xfffffffc);
sys/arm/ti/cpsw/if_cpsw.c
1616
while ((slot = STAILQ_FIRST(&sc->rx.active)) != NULL) {
sys/arm/ti/cpsw/if_cpsw.c
1630
STAILQ_REMOVE_HEAD(&sc->rx.active, next);
sys/arm/ti/cpsw/if_cpsw.c
1631
STAILQ_INSERT_TAIL(&sc->rx.avail, slot, next);
sys/arm/ti/cpsw/if_cpsw.c
1642
sc->rx.running = 0;
sys/arm/ti/cpsw/if_cpsw.c
1643
sc->rx.teardown = 0;
sys/arm/ti/cpsw/if_cpsw.c
1670
if (nsegs > sc->rx.longest_chain)
sys/arm/ti/cpsw/if_cpsw.c
1671
sc->rx.longest_chain = nsegs;
sys/arm/ti/cpsw/if_cpsw.c
1686
if (STAILQ_FIRST(&sc->rx.active) != NULL &&
sys/arm/ti/cpsw/if_cpsw.c
1689
cpsw_write_hdp_slot(sc, &sc->rx,
sys/arm/ti/cpsw/if_cpsw.c
1690
STAILQ_FIRST(&sc->rx.active));
sys/arm/ti/cpsw/if_cpsw.c
1691
sc->rx.queue_restart++;
sys/arm/ti/cpsw/if_cpsw.c
1713
cpsw_write_cp_slot(sc, &sc->rx, last);
sys/arm/ti/cpsw/if_cpsw.c
1714
sc->rx.queue_removes += removed;
sys/arm/ti/cpsw/if_cpsw.c
1715
sc->rx.avail_queue_len += removed;
sys/arm/ti/cpsw/if_cpsw.c
1716
sc->rx.active_queue_len -= removed;
sys/arm/ti/cpsw/if_cpsw.c
1717
if (sc->rx.avail_queue_len > sc->rx.max_avail_queue_len)
sys/arm/ti/cpsw/if_cpsw.c
1718
sc->rx.max_avail_queue_len = sc->rx.avail_queue_len;
sys/arm/ti/cpsw/if_cpsw.c
1735
last_old_slot = STAILQ_LAST(&sc->rx.active, cpsw_slot, next);
sys/arm/ti/cpsw/if_cpsw.c
1736
while ((slot = STAILQ_FIRST(&sc->rx.avail)) != NULL) {
sys/arm/ti/cpsw/if_cpsw.c
1781
STAILQ_REMOVE_HEAD(&sc->rx.avail, next);
sys/arm/ti/cpsw/if_cpsw.c
1782
STAILQ_INSERT_TAIL(&sc->rx.active, slot, next);
sys/arm/ti/cpsw/if_cpsw.c
1793
cpsw_write_hdp_slot(sc, &sc->rx, first_new_slot);
sys/arm/ti/cpsw/if_cpsw.c
1798
sc->rx.queue_adds += added;
sys/arm/ti/cpsw/if_cpsw.c
1799
sc->rx.avail_queue_len -= added;
sys/arm/ti/cpsw/if_cpsw.c
1800
sc->rx.active_queue_len += added;
sys/arm/ti/cpsw/if_cpsw.c
1802
if (sc->rx.active_queue_len > sc->rx.max_active_queue_len)
sys/arm/ti/cpsw/if_cpsw.c
1803
sc->rx.max_active_queue_len = sc->rx.active_queue_len;
sys/arm/ti/cpsw/if_cpsw.c
2152
cpsw_dump_queue(sc, &sc->rx.active);
sys/arm/ti/cpsw/if_cpsw.c
2781
cpsw_add_queue_sysctls(ctx, node, &sc->rx);
sys/arm/ti/cpsw/if_cpsw.c
341
mtx_assert(&(sc)->rx.lock, MA_NOTOWNED); \
sys/arm/ti/cpsw/if_cpsw.c
350
mtx_lock(&(sc)->rx.lock); \
sys/arm/ti/cpsw/if_cpsw.c
353
#define CPSW_RX_UNLOCK(sc) mtx_unlock(&(sc)->rx.lock)
sys/arm/ti/cpsw/if_cpsw.c
354
#define CPSW_RX_LOCK_ASSERT(sc) mtx_assert(&(sc)->rx.lock, MA_OWNED)
sys/arm/ti/cpsw/if_cpsw.c
669
slot = STAILQ_FIRST(&sc->rx.active);
sys/arm/ti/cpsw/if_cpsw.c
671
cpsw_write_hdp_slot(sc, &sc->rx, slot);
sys/arm/ti/cpsw/if_cpsw.c
673
cpsw_write_4(sc, CPSW_CPDMA_RX_FREEBUFFER(0), sc->rx.active_queue_len);
sys/arm/ti/cpsw/if_cpsw.c
677
sc->rx.running = 1;
sys/arm/ti/cpsw/if_cpsw.c
823
mtx_init(&sc->rx.lock, device_get_nameunit(dev),
sys/arm/ti/cpsw/if_cpsw.c
872
STAILQ_INIT(&sc->rx.avail);
sys/arm/ti/cpsw/if_cpsw.c
873
STAILQ_INIT(&sc->rx.active);
sys/arm/ti/cpsw/if_cpsw.c
879
cpsw_add_slots(sc, &sc->rx, -1)) {
sys/arm/ti/cpsw/if_cpsw.c
885
sc->tx.queue_slots, sc->rx.queue_slots);
sys/arm/ti/cpsw/if_cpsw.c
888
sc->rx.hdp_offset = CPSW_CPDMA_RX_HDP(0);
sys/arm/ti/cpsw/if_cpsw.c
960
mtx_destroy(&sc->rx.lock);
sys/arm/ti/cpsw/if_cpswvar.h
98
struct cpsw_queue rx, tx;
sys/compat/linuxkpi/common/include/net/cfg80211.h
133
uint16_t rx;
sys/dev/aq/aq_hw_llh.c
220
itr_irq_map_en_rx_set(struct aq_hw *aq_hw, uint32_t irq_map_en_rx, uint32_t rx)
sys/dev/aq/aq_hw_llh.c
254
AQ_WRITE_REG_BIT(aq_hw, itr_imr_rxren_adr[rx], itr_imr_rxren_msk[rx],
sys/dev/aq/aq_hw_llh.c
255
itr_imr_rxren_shift[rx], irq_map_en_rx);
sys/dev/aq/aq_hw_llh.c
298
itr_irq_map_rx_set(struct aq_hw *aq_hw, uint32_t irq_map_rx, uint32_t rx)
sys/dev/aq/aq_hw_llh.c
332
AQ_WRITE_REG_BIT(aq_hw, itr_imr_rxr_adr[rx], itr_imr_rxr_msk[rx],
sys/dev/aq/aq_hw_llh.c
333
itr_imr_rxr_shift[rx], irq_map_rx);
sys/dev/aq/aq_hw_llh.h
229
uint32_t rx);
sys/dev/aq/aq_hw_llh.h
236
void itr_irq_map_rx_set(struct aq_hw *aq_hw, uint32_t irq_map_rx, uint32_t rx);
sys/dev/ath/ath_hal/ah_desc.h
249
struct ath_rx_status rx;/* recv status */
sys/dev/ath/ath_hal/ah_desc.h
254
#define ds_rxstat ds_us.rx
sys/dev/ath/ath_hal/ar5212/ar5212desc.h
44
} rx;
sys/dev/ath/ath_hal/ar5212/ar5212desc.h
54
#define ds_rxstatus0 u.rx.status0
sys/dev/ath/ath_hal/ar5212/ar5212desc.h
55
#define ds_rxstatus1 u.rx.status1
sys/dev/ath/ath_hal/ar5416/ar5416desc.h
72
struct ar5416_rx_desc rx;
sys/dev/ath/ath_hal/ar5416/ar5416desc.h
89
#define ds_rxstatus0 u.rx.status0
sys/dev/ath/ath_hal/ar5416/ar5416desc.h
90
#define ds_rxstatus1 u.rx.status1
sys/dev/ath/ath_hal/ar5416/ar5416desc.h
91
#define ds_rxstatus2 u.rx.status2
sys/dev/ath/ath_hal/ar5416/ar5416desc.h
92
#define ds_rxstatus3 u.rx.status3
sys/dev/ath/ath_hal/ar5416/ar5416desc.h
93
#define ds_rxstatus4 u.rx.status4
sys/dev/ath/ath_hal/ar5416/ar5416desc.h
94
#define ds_rxstatus5 u.rx.status5
sys/dev/ath/ath_hal/ar5416/ar5416desc.h
95
#define ds_rxstatus6 u.rx.status6
sys/dev/ath/ath_hal/ar5416/ar5416desc.h
96
#define ds_rxstatus7 u.rx.status7
sys/dev/ath/ath_hal/ar5416/ar5416desc.h
97
#define ds_rxstatus8 u.rx.status8
sys/dev/axgbe/xgbe-dev.c
1391
rdata->rx.hdr_len = XGMAC_GET_BITS_LE(rdesc->desc2,
sys/dev/axgbe/xgbe-dev.c
1393
if (rdata->rx.hdr_len)
sys/dev/axgbe/xgbe-dev.c
1439
rdata->rx.len = XGMAC_GET_BITS_LE(rdesc->desc3, RX_NORMAL_DESC3, PL);
sys/dev/axgbe/xgbe-sysctl.c
1036
rx = rounddown_pow_of_two(sys_op->rx_pending);
sys/dev/axgbe/xgbe-sysctl.c
1037
if (rx != sys_op->rx_pending)
sys/dev/axgbe/xgbe-sysctl.c
1039
rx);
sys/dev/axgbe/xgbe-sysctl.c
1046
if ((rx == pdata->rx_desc_count) &&
sys/dev/axgbe/xgbe-sysctl.c
1050
pdata->rx_desc_count = rx;
sys/dev/axgbe/xgbe-sysctl.c
1071
unsigned int rx, tx, combined;
sys/dev/axgbe/xgbe-sysctl.c
1083
rx = min(pdata->hw_feat.rx_ch_cnt, pdata->rx_max_channel_count);
sys/dev/axgbe/xgbe-sysctl.c
1084
rx = min(rx, pdata->channel_irq_count);
sys/dev/axgbe/xgbe-sysctl.c
1089
combined = min(rx, tx);
sys/dev/axgbe/xgbe-sysctl.c
1092
sys_op->max_rx = rx ? rx - 1 : 0;
sys/dev/axgbe/xgbe-sysctl.c
1096
rx = pdata->rx_ring_count;
sys/dev/axgbe/xgbe-sysctl.c
1099
combined = min(rx, tx);
sys/dev/axgbe/xgbe-sysctl.c
1100
rx -= combined;
sys/dev/axgbe/xgbe-sysctl.c
1104
sys_op->rx_count = rx;
sys/dev/axgbe/xgbe-sysctl.c
969
unsigned int rx, tx;
sys/dev/axgbe/xgbe-txrx.c
632
if (rdata->rx.hdr_len) {
sys/dev/axgbe/xgbe-txrx.c
633
axgbe_printf(1, "%s: hdr_len %d\n", __func__, rdata->rx.hdr_len);
sys/dev/axgbe/xgbe-txrx.c
634
return (rdata->rx.hdr_len);
sys/dev/axgbe/xgbe-txrx.c
655
axgbe_printf(1, "%s: pkt_len %d buf_size %d\n", __func__, rdata->rx.len,
sys/dev/axgbe/xgbe-txrx.c
659
ret = min_t(unsigned int, 256, rdata->rx.len);
sys/dev/axgbe/xgbe-txrx.c
661
ret = rdata->rx.len;
sys/dev/axgbe/xgbe-txrx.c
681
return ((rdata->rx.len != 0)? (rdata->rx.len - len) : 0);
sys/dev/axgbe/xgbe.h
457
struct xgbe_rx_ring_data rx; /* Rx-related data */
sys/dev/bnxt/bnxt_en/bnxt.h
278
(link_info->last_flow_ctrl.rx != link_info->flow_ctrl.rx) || \
sys/dev/bnxt/bnxt_en/bnxt.h
388
bool rx;
sys/dev/bnxt/bnxt_en/bnxt_hwrm.c
1410
if (link_info->flow_ctrl.rx)
sys/dev/bnxt/bnxt_en/bnxt_hwrm.c
1419
if (link_info->flow_ctrl.rx)
sys/dev/bnxt/bnxt_en/bnxt_hwrm.c
2900
link_info->flow_ctrl.rx = false;
sys/dev/bnxt/bnxt_en/bnxt_hwrm.c
2911
link_info->flow_ctrl.rx = true;
sys/dev/bnxt/bnxt_en/bnxt_hwrm.c
2916
link_info->flow_ctrl.rx = true;
sys/dev/bnxt/bnxt_en/bnxt_sysctl.c
1724
BNXT_FLOW_CTRL_FN(bnxt_flow_ctrl_rx, rx)
sys/dev/bnxt/bnxt_en/if_bnxt.c
3316
if (link_info->flow_ctrl.rx)
sys/dev/bnxt/bnxt_en/if_bnxt.c
4825
if (link_info->flow_ctrl.tx & link_info->flow_ctrl.rx)
sys/dev/bnxt/bnxt_en/if_bnxt.c
4829
else if (link_info->flow_ctrl.rx)
sys/dev/bnxt/bnxt_en/if_bnxt.c
4868
link_info->last_flow_ctrl.rx = link_info->flow_ctrl.rx;
sys/dev/bnxt/bnxt_re/stats.c
282
ext_d->rx.atomic_req = s->rx_atomic_req;
sys/dev/bnxt/bnxt_re/stats.c
283
ext_d->rx.read_req = s->rx_read_req;
sys/dev/bnxt/bnxt_re/stats.c
284
ext_d->rx.read_resp = s->rx_read_res;
sys/dev/bnxt/bnxt_re/stats.c
285
ext_d->rx.write_req = s->rx_write_req;
sys/dev/bnxt/bnxt_re/stats.c
286
ext_d->rx.send_req = s->rx_send_req;
sys/dev/bnxt/bnxt_re/stats.c
524
stats->value[BNXT_RE_RX_ATOMIC_REQ] = ext_s->rx.atomic_req;
sys/dev/bnxt/bnxt_re/stats.c
527
stats->value[BNXT_RE_RX_READ_REQ] = ext_s->rx.read_req;
sys/dev/bnxt/bnxt_re/stats.c
528
stats->value[BNXT_RE_RX_READ_RESP] = ext_s->rx.read_resp;
sys/dev/bnxt/bnxt_re/stats.c
530
stats->value[BNXT_RE_RX_WRITE_REQ] = ext_s->rx.write_req;
sys/dev/bnxt/bnxt_re/stats.c
532
stats->value[BNXT_RE_RX_SEND_REQ] = ext_s->rx.send_req;
sys/dev/bnxt/bnxt_re/stats.h
76
struct bnxt_re_ext_cntr rx;
sys/dev/bwn/if_bwn.c
2635
bwn_pio_setupqueue_rx(mac, &pio->rx, 0);
sys/dev/bwn/if_bwn.c
2813
bwn_dma_setup(dma->rx);
sys/dev/bwn/if_bwn.c
4797
bwn_dma_ringfree(&dma->rx);
sys/dev/bwn/if_bwn.c
5073
int i, tx = 0, rx = 0;
sys/dev/bwn/if_bwn.c
5132
bwn_dma_rx(mac->mac_method.dma.rx);
sys/dev/bwn/if_bwn.c
5133
rx = 1;
sys/dev/bwn/if_bwn.c
5136
rx = bwn_pio_rx(&mac->mac_method.pio.rx);
sys/dev/bwn/if_bwn.c
5154
if (tx && rx) {
sys/dev/bwn/if_bwn.c
5161
} else if (rx) {
sys/dev/bwn/if_bwn.c
5163
} else if (rx == 0) {
sys/dev/bwn/if_bwn.c
7273
dma->rx = bwn_dma_ringsetup(mac, 0, 0);
sys/dev/bwn/if_bwn.c
7274
if (!dma->rx)
sys/dev/bwn/if_bwn.c
7348
bwn_dma_ringstop(&dma->rx);
sys/dev/bwn/if_bwn_phy_lp.c
2278
bwn_phy_lp_set_trsw_over(struct bwn_mac *mac, uint8_t tx, uint8_t rx)
sys/dev/bwn/if_bwn_phy_lp.c
2280
uint16_t trsw = (tx << 1) | rx;
sys/dev/bwn/if_bwnvar.h
702
struct bwn_dma_ring *rx;
sys/dev/bwn/if_bwnvar.h
735
struct bwn_pio_rxqueue rx;
sys/dev/bxe/bxe.c
11169
bxe_set_bit(ECORE_Q_FLG_HC, &init_params->rx.flags);
sys/dev/bxe/bxe.c
11172
bxe_set_bit(ECORE_Q_FLG_HC_EN, &init_params->rx.flags);
sys/dev/bxe/bxe.c
11176
init_params->rx.hc_rate =
sys/dev/bxe/bxe.c
11182
init_params->rx.fw_sb_id = init_params->tx.fw_sb_id = fp->fw_sb_id;
sys/dev/bxe/bxe.c
11185
init_params->rx.sb_cq_index = HC_INDEX_ETH_RX_CQ_CONS;
sys/dev/bxe/bxe_elink.c
12961
uint32_t rx = 0, tx = 0, i;
sys/dev/bxe/bxe_elink.c
12968
rx = REG_RD(sc, shmem_base +
sys/dev/bxe/bxe_elink.c
12976
rx = REG_RD(sc, shmem_base +
sys/dev/bxe/bxe_elink.c
12985
phy->rx_preemphasis[i << 1] = ((rx>>16) & 0xffff);
sys/dev/bxe/bxe_elink.c
12986
phy->rx_preemphasis[(i << 1) + 1] = (rx & 0xffff);
sys/dev/bxe/ecore_hsi.h
6757
struct client_init_rx_data rx /* client init rx data */;
sys/dev/bxe/ecore_sp.c
4772
data->rx.tpa_en |= ECORE_TEST_BIT(ECORE_Q_FLG_TPA_IPV6,
sys/dev/bxe/ecore_sp.c
4954
&data->rx,
sys/dev/bxe/ecore_sp.c
4959
&data->rx);
sys/dev/bxe/ecore_sp.c
5015
ECORE_TEST_BIT(ECORE_Q_FLG_HC, &init->rx.flags)) {
sys/dev/bxe/ecore_sp.c
5016
hc_usec = init->rx.hc_rate ? 1000000 / init->rx.hc_rate : 0;
sys/dev/bxe/ecore_sp.c
5018
ECORE_UPDATE_COALESCE_SB_INDEX(sc, init->rx.fw_sb_id,
sys/dev/bxe/ecore_sp.c
5019
init->rx.sb_cq_index,
sys/dev/bxe/ecore_sp.c
5020
!ECORE_TEST_BIT(ECORE_Q_FLG_HC_EN, &init->rx.flags),
sys/dev/bxe/ecore_sp.h
1344
} rx;
sys/dev/dwc/if_dwc_rk.c
177
uint32_t tx, rx;
sys/dev/dwc/if_dwc_rk.c
184
rx = ((reg >> MAC_CON0_GMAC2IO_RX_DL_CFG_SHIFT) & MAC_CON0_GMAC2IO_RX_DL_CFG_MASK);
sys/dev/dwc/if_dwc_rk.c
190
rx, ((reg & MAC_CON1_GMAC2IO_GMAC_RXCLK_DLY_ENA) ? "enabled" : "disabled"));
sys/dev/dwc/if_dwc_rk.c
294
uint32_t reg, tx, rx;
sys/dev/dwc/if_dwc_rk.c
301
rx = ((reg >> SOC_CON6_RX_DL_CFG_SHIFT) & SOC_CON6_RX_DL_CFG_MASK);
sys/dev/dwc/if_dwc_rk.c
306
rx, ((reg & SOC_CON6_GMAC_RXCLK_DLY_ENA) ? "enabled" : "disabled"));
sys/dev/dwc/if_dwc_rk.c
487
uint32_t rx, tx;
sys/dev/dwc/if_dwc_rk.c
506
if (OF_getencprop(node, "rx_delay", &rx, sizeof(rx)) <= 0)
sys/dev/dwc/if_dwc_rk.c
507
rx = 0x10;
sys/dev/dwc/if_dwc_rk.c
509
sc->rx_delay = rx;
sys/dev/enic/if_enic.c
1284
&stats->rx.rx_frames_ok, "RX Frames OK");
sys/dev/enic/if_enic.c
1286
&stats->rx.rx_frames_total, "RX frames total");
sys/dev/enic/if_enic.c
1288
&stats->rx.rx_unicast_frames_ok, "RX unicast frames ok");
sys/dev/enic/if_enic.c
1290
&stats->rx.rx_multicast_frames_ok, "RX multicast Frames ok");
sys/dev/enic/if_enic.c
1292
&stats->rx.rx_broadcast_frames_ok, "RX broadcast frames ok");
sys/dev/enic/if_enic.c
1294
&stats->rx.rx_bytes_ok, "RX bytes ok");
sys/dev/enic/if_enic.c
1296
&stats->rx.rx_unicast_bytes_ok, "RX unicast bytes ok");
sys/dev/enic/if_enic.c
1298
&stats->rx.rx_multicast_bytes_ok, "RX multicast bytes ok");
sys/dev/enic/if_enic.c
1300
&stats->rx.rx_broadcast_bytes_ok, "RX broadcast bytes ok");
sys/dev/enic/if_enic.c
1302
&stats->rx.rx_drop, "RX drop");
sys/dev/enic/if_enic.c
1304
&stats->rx.rx_errors, "RX errors");
sys/dev/enic/if_enic.c
1306
&stats->rx.rx_rss, "RX rss");
sys/dev/enic/if_enic.c
1308
&stats->rx.rx_crc_errors, "RX crc errors");
sys/dev/enic/if_enic.c
1310
&stats->rx.rx_frames_64, "RX frames 64");
sys/dev/enic/if_enic.c
1312
&stats->rx.rx_frames_127, "RX frames 127");
sys/dev/enic/if_enic.c
1314
&stats->rx.rx_frames_255, "RX frames 255");
sys/dev/enic/if_enic.c
1316
&stats->rx.rx_frames_511, "RX frames 511");
sys/dev/enic/if_enic.c
1318
&stats->rx.rx_frames_1023, "RX frames 1023");
sys/dev/enic/if_enic.c
1320
&stats->rx.rx_frames_1518, "RX frames 1518");
sys/dev/enic/if_enic.c
1322
&stats->rx.rx_frames_to_max, "RX frames to max");
sys/dev/enic/vnic_stats.h
53
struct vnic_rx_stats rx;
sys/dev/eqos/if_eqos.c
1084
&sc->rx.desc_tag);
sys/dev/eqos/if_eqos.c
1090
error = bus_dmamem_alloc(sc->rx.desc_tag,
sys/dev/eqos/if_eqos.c
1091
(void **)&sc->rx.desc_ring,
sys/dev/eqos/if_eqos.c
1093
&sc->rx.desc_map);
sys/dev/eqos/if_eqos.c
1100
error = bus_dmamap_load(sc->rx.desc_tag, sc->rx.desc_map,
sys/dev/eqos/if_eqos.c
1101
sc->rx.desc_ring, RX_DESC_SIZE,
sys/dev/eqos/if_eqos.c
1102
eqos_get1paddr, &sc->rx.desc_ring_paddr,
sys/dev/eqos/if_eqos.c
1120
&sc->rx.buf_tag);
sys/dev/eqos/if_eqos.c
1127
if ((error = bus_dmamap_create(sc->rx.buf_tag, BUS_DMA_COHERENT,
sys/dev/eqos/if_eqos.c
1128
&sc->rx.buf_map[i].map))) {
sys/dev/eqos/if_eqos.c
1144
sc->tx.desc_ring_paddr, sc->rx.desc_ring_paddr);
sys/dev/eqos/if_eqos.c
1299
if (sc->rx.desc_tag) {
sys/dev/eqos/if_eqos.c
1300
if (sc->rx.desc_map) {
sys/dev/eqos/if_eqos.c
1301
bus_dmamap_unload(sc->rx.desc_tag, sc->rx.desc_map);
sys/dev/eqos/if_eqos.c
1302
bus_dmamem_free(sc->rx.desc_tag, sc->rx.desc_ring,
sys/dev/eqos/if_eqos.c
1303
sc->rx.desc_map);
sys/dev/eqos/if_eqos.c
1305
bus_dma_tag_destroy(sc->rx.desc_tag);
sys/dev/eqos/if_eqos.c
1307
if (sc->rx.buf_tag) {
sys/dev/eqos/if_eqos.c
1309
m_free(sc->rx.buf_map[i].mbuf);
sys/dev/eqos/if_eqos.c
1310
bus_dmamap_destroy(sc->rx.buf_tag,
sys/dev/eqos/if_eqos.c
1311
sc->rx.buf_map[i].map);
sys/dev/eqos/if_eqos.c
1313
bus_dma_tag_destroy(sc->rx.buf_tag);
sys/dev/eqos/if_eqos.c
330
sc->rx.desc_ring[index].des0 = htole32((uint32_t)paddr);
sys/dev/eqos/if_eqos.c
331
sc->rx.desc_ring[index].des1 = htole32((uint32_t)(paddr >> 32));
sys/dev/eqos/if_eqos.c
332
sc->rx.desc_ring[index].des2 = htole32(0);
sys/dev/eqos/if_eqos.c
333
bus_dmamap_sync(sc->rx.desc_tag, sc->rx.desc_map, BUS_DMASYNC_PREWRITE);
sys/dev/eqos/if_eqos.c
334
sc->rx.desc_ring[index].des3 =
sys/dev/eqos/if_eqos.c
346
error = bus_dmamap_load_mbuf_sg(sc->rx.buf_tag,
sys/dev/eqos/if_eqos.c
347
sc->rx.buf_map[index].map, m, &seg, &nsegs, 0);
sys/dev/eqos/if_eqos.c
351
bus_dmamap_sync(sc->rx.buf_tag, sc->rx.buf_map[index].map,
sys/dev/eqos/if_eqos.c
354
sc->rx.buf_map[index].mbuf = m;
sys/dev/eqos/if_eqos.c
482
(uint32_t)(sc->rx.desc_ring_paddr >> 32));
sys/dev/eqos/if_eqos.c
484
(uint32_t)sc->rx.desc_ring_paddr);
sys/dev/eqos/if_eqos.c
488
(uint32_t)sc->rx.desc_ring_paddr + DESC_OFFSET(RX_DESC_COUNT));
sys/dev/eqos/if_eqos.c
701
rdes3 = le32toh(sc->rx.desc_ring[sc->rx.head].des3);
sys/dev/eqos/if_eqos.c
708
bus_dmamap_sync(sc->rx.buf_tag,
sys/dev/eqos/if_eqos.c
709
sc->rx.buf_map[sc->rx.head].map, BUS_DMASYNC_POSTREAD);
sys/dev/eqos/if_eqos.c
710
bus_dmamap_unload(sc->rx.buf_tag,
sys/dev/eqos/if_eqos.c
711
sc->rx.buf_map[sc->rx.head].map);
sys/dev/eqos/if_eqos.c
715
m = sc->rx.buf_map[sc->rx.head].mbuf;
sys/dev/eqos/if_eqos.c
730
if ((error = eqos_setup_rxbuf(sc, sc->rx.head, m)))
sys/dev/eqos/if_eqos.c
738
(uint32_t)sc->rx.desc_ring_paddr + DESC_OFFSET(sc->rx.head));
sys/dev/eqos/if_eqos.c
740
sc->rx.head = RX_NEXT(sc->rx.head);
sys/dev/eqos/if_eqos_var.h
101
struct eqos_ring rx;
sys/dev/firewire/firewire.h
54
} tx, rx;
sys/dev/firewire/fwdev.c
604
err = fwdev_allocbuf(fc, ir, &d->bufreq.rx);
sys/dev/firewire/fwdev.c
627
bzero(&ibufreq->rx, sizeof(ibufreq->rx));
sys/dev/firewire/fwdev.c
629
ibufreq->rx.nchunk = ir->bnchunk;
sys/dev/firewire/fwdev.c
630
ibufreq->rx.npacket = ir->bnpacket;
sys/dev/firewire/fwdev.c
631
ibufreq->rx.psize = ir->psize;
sys/dev/gve/gve.h
600
struct gve_rx_ring *rx;
sys/dev/gve/gve.h
741
void gve_rx_prefill_buffers_dqo(struct gve_rx_ring *rx);
sys/dev/gve/gve_adminq.c
145
be16toh((*dev_op_modify_ring)->min_ring_size.rx),
sys/dev/gve/gve_adminq.c
288
struct gve_rx_ring *rx = &priv->rx[queue_index];
sys/dev/gve/gve_adminq.c
289
struct gve_dma_handle *qres_dma = &rx->com.q_resources_mem;
sys/dev/gve/gve_adminq.c
296
.ntfy_id = htobe32(rx->com.ntfy_id),
sys/dev/gve/gve_adminq.c
303
htobe64(rx->desc_ring_mem.bus_addr);
sys/dev/gve/gve_adminq.c
305
htobe64(rx->data_ring_mem.bus_addr);
sys/dev/gve/gve_adminq.c
309
htobe32((rx->com.qpl)->id);
sys/dev/gve/gve_adminq.c
316
htobe64(rx->dqo.compl_ring_mem.bus_addr);
sys/dev/gve/gve_adminq.c
318
htobe64(rx->desc_ring_mem.bus_addr);
sys/dev/gve/gve_adminq.c
435
priv->max_rx_desc_cnt = be16toh(dev_op_modify_ring->max_ring_size.rx);
sys/dev/gve/gve_adminq.h
157
__be16 rx;
sys/dev/gve/gve_main.c
702
free(priv->rx, M_GVE);
sys/dev/gve/gve_main.c
703
priv->rx = NULL;
sys/dev/gve/gve_main.c
711
priv->rx = malloc(sizeof(struct gve_rx_ring) * priv->rx_cfg.max_queues,
sys/dev/gve/gve_qpl.c
164
struct gve_rx_ring *rx;
sys/dev/gve/gve_qpl.c
187
rx = &priv->rx[i];
sys/dev/gve/gve_qpl.c
188
com = &rx->com;
sys/dev/gve/gve_qpl.c
209
struct gve_rx_ring *rx;
sys/dev/gve/gve_qpl.c
226
rx = &priv->rx[i];
sys/dev/gve/gve_qpl.c
227
com = &rx->com;
sys/dev/gve/gve_rx.c
106
struct gve_rx_ring *rx = &priv->rx[i];
sys/dev/gve/gve_rx.c
107
struct gve_ring_com *com = &rx->com;
sys/dev/gve/gve_rx.c
112
CACHE_LINE_SIZE, &rx->desc_ring_mem);
sys/dev/gve/gve_rx.c
119
rx->mask = priv->rx_pages_per_qpl - 1;
sys/dev/gve/gve_rx.c
120
rx->desc_ring = rx->desc_ring_mem.cpu_addr;
sys/dev/gve/gve_rx.c
131
rx->page_info = malloc(priv->rx_desc_cnt * sizeof(*rx->page_info),
sys/dev/gve/gve_rx.c
136
CACHE_LINE_SIZE, &rx->data_ring_mem);
sys/dev/gve/gve_rx.c
142
rx->data_ring = rx->data_ring_mem.cpu_addr;
sys/dev/gve/gve_rx.c
144
gve_prefill_rx_slots(rx);
sys/dev/gve/gve_rx.c
155
struct gve_rx_ring *rx = &priv->rx[i];
sys/dev/gve/gve_rx.c
156
struct gve_ring_com *com = &rx->com;
sys/dev/gve/gve_rx.c
162
gve_alloc_counters((counter_u64_t *)&rx->stats, NUM_RX_STATS);
sys/dev/gve/gve_rx.c
193
KASSERT(priv->rx != NULL, ("priv->rx is NULL!"));
sys/dev/gve/gve_rx.c
217
gve_rx_clear_data_ring(struct gve_rx_ring *rx)
sys/dev/gve/gve_rx.c
219
struct gve_priv *priv = rx->com.priv;
sys/dev/gve/gve_rx.c
232
rx->data_ring[i].qpl_offset = htobe64(PAGE_SIZE * i +
sys/dev/gve/gve_rx.c
233
rx->page_info[i].page_offset);
sys/dev/gve/gve_rx.c
234
rx->fill_cnt++;
sys/dev/gve/gve_rx.c
237
bus_dmamap_sync(rx->data_ring_mem.tag, rx->data_ring_mem.map,
sys/dev/gve/gve_rx.c
242
gve_rx_clear_desc_ring(struct gve_rx_ring *rx)
sys/dev/gve/gve_rx.c
244
struct gve_priv *priv = rx->com.priv;
sys/dev/gve/gve_rx.c
248
rx->desc_ring[i] = (struct gve_rx_desc){};
sys/dev/gve/gve_rx.c
250
bus_dmamap_sync(rx->desc_ring_mem.tag, rx->desc_ring_mem.map,
sys/dev/gve/gve_rx.c
257
struct gve_rx_ring *rx = &priv->rx[i];
sys/dev/gve/gve_rx.c
264
rx->seq_no = 1;
sys/dev/gve/gve_rx.c
265
rx->cnt = 0;
sys/dev/gve/gve_rx.c
266
rx->fill_cnt = 0;
sys/dev/gve/gve_rx.c
267
rx->mask = priv->rx_desc_cnt - 1;
sys/dev/gve/gve_rx.c
269
gve_rx_clear_desc_ring(rx);
sys/dev/gve/gve_rx.c
270
gve_rx_clear_data_ring(rx);
sys/dev/gve/gve_rx.c
276
struct gve_rx_ring *rx = &priv->rx[i];
sys/dev/gve/gve_rx.c
277
struct gve_ring_com *com = &rx->com;
sys/dev/gve/gve_rx.c
280
if (tcp_lro_init(&rx->lro) != 0)
sys/dev/gve/gve_rx.c
282
rx->lro.ifp = priv->ifp;
sys/dev/gve/gve_rx.c
286
NET_TASK_INIT(&com->cleanup_task, 0, gve_rx_cleanup_tq, rx);
sys/dev/gve/gve_rx.c
288
NET_TASK_INIT(&com->cleanup_task, 0, gve_rx_cleanup_tq_dqo, rx);
sys/dev/gve/gve_rx.c
297
gve_db_bar_write_4(priv, com->db_offset, rx->fill_cnt);
sys/dev/gve/gve_rx.c
299
gve_rx_prefill_buffers_dqo(rx);
sys/dev/gve/gve_rx.c
306
struct gve_rx_ring *rx;
sys/dev/gve/gve_rx.c
324
rx = &priv->rx[i];
sys/dev/gve/gve_rx.c
325
com = &rx->com;
sys/dev/gve/gve_rx.c
344
struct gve_rx_ring *rx = &priv->rx[i];
sys/dev/gve/gve_rx.c
345
struct gve_ring_com *com = &rx->com;
sys/dev/gve/gve_rx.c
353
tcp_lro_free(&rx->lro);
sys/dev/gve/gve_rx.c
354
rx->ctx = (struct gve_rx_ctx){};
sys/dev/gve/gve_rx.c
379
struct gve_rx_ring *rx = arg;
sys/dev/gve/gve_rx.c
38
struct gve_rx_ring *rx = &priv->rx[i];
sys/dev/gve/gve_rx.c
380
struct gve_priv *priv = rx->com.priv;
sys/dev/gve/gve_rx.c
381
struct gve_ring_com *com = &rx->com;
sys/dev/gve/gve_rx.c
387
taskqueue_enqueue(rx->com.cleanup_tq, &rx->com.cleanup_task);
sys/dev/gve/gve_rx.c
39
struct gve_ring_com *com = &rx->com;
sys/dev/gve/gve_rx.c
41
if (rx->page_info != NULL) {
sys/dev/gve/gve_rx.c
42
free(rx->page_info, M_GVE);
sys/dev/gve/gve_rx.c
424
gve_rx_create_mbuf(struct gve_priv *priv, struct gve_rx_ring *rx,
sys/dev/gve/gve_rx.c
428
struct gve_rx_ctx *ctx = &rx->ctx;
sys/dev/gve/gve_rx.c
43
rx->page_info = NULL;
sys/dev/gve/gve_rx.c
443
counter_u64_add_protected(rx->stats.rx_copybreak_cnt, 1);
sys/dev/gve/gve_rx.c
46
if (rx->data_ring != NULL) {
sys/dev/gve/gve_rx.c
47
gve_dma_free_coherent(&rx->data_ring_mem);
sys/dev/gve/gve_rx.c
48
rx->data_ring = NULL;
sys/dev/gve/gve_rx.c
494
counter_u64_add_protected(rx->stats.rx_frag_flip_cnt, 1);
sys/dev/gve/gve_rx.c
507
counter_u64_add_protected(rx->stats.rx_frag_copy_cnt, 1);
sys/dev/gve/gve_rx.c
51
if (rx->desc_ring != NULL) {
sys/dev/gve/gve_rx.c
52
gve_dma_free_coherent(&rx->desc_ring_mem);
sys/dev/gve/gve_rx.c
529
gve_rx(struct gve_priv *priv, struct gve_rx_ring *rx, struct gve_rx_desc *desc,
sys/dev/gve/gve_rx.c
53
rx->desc_ring = NULL;
sys/dev/gve/gve_rx.c
535
struct gve_rx_ctx *ctx = &rx->ctx;
sys/dev/gve/gve_rx.c
551
counter_u64_add_protected(rx->stats.rx_dropped_pkt_desc_err, 1);
sys/dev/gve/gve_rx.c
552
counter_u64_add_protected(rx->stats.rx_dropped_pkt, 1);
sys/dev/gve/gve_rx.c
558
page_info = &rx->page_info[idx];
sys/dev/gve/gve_rx.c
559
data_slot = &rx->data_ring[idx];
sys/dev/gve/gve_rx.c
560
page_dma_handle = &(rx->com.qpl->dmas[idx]);
sys/dev/gve/gve_rx.c
568
mbuf = gve_rx_create_mbuf(priv, rx, page_info, len, data_slot,
sys/dev/gve/gve_rx.c
573
counter_u64_add_protected(rx->stats.rx_dropped_pkt_mbuf_alloc_fail, 1);
sys/dev/gve/gve_rx.c
574
counter_u64_add_protected(rx->stats.rx_dropped_pkt, 1);
sys/dev/gve/gve_rx.c
606
(rx->lro.lro_cnt != 0) && /* LRO resources exist */
sys/dev/gve/gve_rx.c
607
(tcp_lro_rx(&rx->lro, mbuf, 0) == 0))
sys/dev/gve/gve_rx.c
614
counter_u64_add_protected(rx->stats.rbytes, ctx->total_size);
sys/dev/gve/gve_rx.c
615
counter_u64_add_protected(rx->stats.rpackets, 1);
sys/dev/gve/gve_rx.c
622
rx->ctx = (struct gve_rx_ctx){};
sys/dev/gve/gve_rx.c
626
gve_rx_work_pending(struct gve_rx_ring *rx)
sys/dev/gve/gve_rx.c
632
next_idx = rx->cnt & rx->mask;
sys/dev/gve/gve_rx.c
633
desc = rx->desc_ring + next_idx;
sys/dev/gve/gve_rx.c
637
return (GVE_SEQNO(flags_seq) == rx->seq_no);
sys/dev/gve/gve_rx.c
647
gve_rx_cleanup(struct gve_priv *priv, struct gve_rx_ring *rx, int budget)
sys/dev/gve/gve_rx.c
649
uint32_t idx = rx->cnt & rx->mask;
sys/dev/gve/gve_rx.c
65
struct gve_rx_ring *rx = &priv->rx[i];
sys/dev/gve/gve_rx.c
651
struct gve_rx_ctx *ctx = &rx->ctx;
sys/dev/gve/gve_rx.c
656
bus_dmamap_sync(rx->desc_ring_mem.tag, rx->desc_ring_mem.map,
sys/dev/gve/gve_rx.c
658
desc = &rx->desc_ring[idx];
sys/dev/gve/gve_rx.c
66
struct gve_ring_com *com = &rx->com;
sys/dev/gve/gve_rx.c
661
(GVE_SEQNO(desc->flags_seq) == rx->seq_no)) {
sys/dev/gve/gve_rx.c
663
gve_rx(priv, rx, desc, idx);
sys/dev/gve/gve_rx.c
665
rx->cnt++;
sys/dev/gve/gve_rx.c
666
idx = rx->cnt & rx->mask;
sys/dev/gve/gve_rx.c
667
desc = &rx->desc_ring[idx];
sys/dev/gve/gve_rx.c
668
rx->seq_no = gve_next_seqno(rx->seq_no);
sys/dev/gve/gve_rx.c
675
rx->ctx = (struct gve_rx_ctx){};
sys/dev/gve/gve_rx.c
678
GVE_SEQNO(desc->flags_seq), rx->seq_no);
sys/dev/gve/gve_rx.c
683
tcp_lro_flush_all(&rx->lro);
sys/dev/gve/gve_rx.c
685
bus_dmamap_sync(rx->data_ring_mem.tag, rx->data_ring_mem.map,
sys/dev/gve/gve_rx.c
689
rx->fill_cnt += work_done;
sys/dev/gve/gve_rx.c
69
gve_free_counters((counter_u64_t *)&rx->stats, NUM_RX_STATS);
sys/dev/gve/gve_rx.c
690
gve_db_bar_write_4(priv, rx->com.db_offset, rx->fill_cnt);
sys/dev/gve/gve_rx.c
696
struct gve_rx_ring *rx = arg;
sys/dev/gve/gve_rx.c
697
struct gve_priv *priv = rx->com.priv;
sys/dev/gve/gve_rx.c
702
gve_rx_cleanup(priv, rx, /*budget=*/128);
sys/dev/gve/gve_rx.c
704
gve_db_bar_write_4(priv, rx->com.irq_db_offset,
sys/dev/gve/gve_rx.c
714
if (gve_rx_work_pending(rx)) {
sys/dev/gve/gve_rx.c
715
gve_db_bar_write_4(priv, rx->com.irq_db_offset, GVE_IRQ_MASK);
sys/dev/gve/gve_rx.c
716
taskqueue_enqueue(rx->com.cleanup_tq, &rx->com.cleanup_task);
sys/dev/gve/gve_rx.c
83
gve_prefill_rx_slots(struct gve_rx_ring *rx)
sys/dev/gve/gve_rx.c
85
struct gve_ring_com *com = &rx->com;
sys/dev/gve/gve_rx.c
90
rx->data_ring[i].qpl_offset = htobe64(PAGE_SIZE * i);
sys/dev/gve/gve_rx.c
91
rx->page_info[i].page_offset = 0;
sys/dev/gve/gve_rx.c
92
rx->page_info[i].page_address = com->qpl->dmas[i].cpu_addr;
sys/dev/gve/gve_rx.c
93
rx->page_info[i].page = com->qpl->pages[i];
sys/dev/gve/gve_rx.c
99
bus_dmamap_sync(rx->data_ring_mem.tag, rx->data_ring_mem.map,
sys/dev/gve/gve_rx_dqo.c
100
struct gve_rx_ring *rx = &priv->rx[i];
sys/dev/gve/gve_rx_dqo.c
1000
rx->cnt++;
sys/dev/gve/gve_rx_dqo.c
1001
rx->dqo.tail = (rx->dqo.tail + 1) & rx->dqo.mask;
sys/dev/gve/gve_rx_dqo.c
1002
rx->dqo.cur_gen_bit ^= (rx->dqo.tail == 0);
sys/dev/gve/gve_rx_dqo.c
1005
gve_rx_dqo_qpl(priv, rx, compl_desc, &work_done);
sys/dev/gve/gve_rx_dqo.c
1007
gve_rx_dqo(priv, rx, compl_desc, &work_done);
sys/dev/gve/gve_rx_dqo.c
1011
tcp_lro_flush_all(&rx->lro);
sys/dev/gve/gve_rx_dqo.c
1013
gve_rx_post_buffers_dqo(rx, M_NOWAIT);
sys/dev/gve/gve_rx_dqo.c
1015
gve_rx_maybe_extract_from_used_bufs(rx, /*just_one=*/false);
sys/dev/gve/gve_rx_dqo.c
1022
struct gve_rx_ring *rx = arg;
sys/dev/gve/gve_rx_dqo.c
1023
struct gve_priv *priv = rx->com.priv;
sys/dev/gve/gve_rx_dqo.c
1028
if (gve_rx_cleanup_dqo(priv, rx, /*budget=*/64)) {
sys/dev/gve/gve_rx_dqo.c
1029
taskqueue_enqueue(rx->com.cleanup_tq, &rx->com.cleanup_task);
sys/dev/gve/gve_rx_dqo.c
1033
gve_db_bar_dqo_write_4(priv, rx->com.irq_db_offset,
sys/dev/gve/gve_rx_dqo.c
106
CACHE_LINE_SIZE, &rx->desc_ring_mem);
sys/dev/gve/gve_rx_dqo.c
112
rx->dqo.desc_ring = rx->desc_ring_mem.cpu_addr;
sys/dev/gve/gve_rx_dqo.c
113
rx->dqo.mask = priv->rx_desc_cnt - 1;
sys/dev/gve/gve_rx_dqo.c
117
CACHE_LINE_SIZE, &rx->dqo.compl_ring_mem);
sys/dev/gve/gve_rx_dqo.c
123
rx->dqo.compl_ring = rx->dqo.compl_ring_mem.cpu_addr;
sys/dev/gve/gve_rx_dqo.c
124
rx->dqo.mask = priv->rx_desc_cnt - 1;
sys/dev/gve/gve_rx_dqo.c
126
rx->dqo.buf_cnt = gve_is_qpl(priv) ? GVE_RX_NUM_QPL_PAGES_DQO :
sys/dev/gve/gve_rx_dqo.c
128
rx->dqo.bufs = malloc(rx->dqo.buf_cnt * sizeof(struct gve_rx_buf_dqo),
sys/dev/gve/gve_rx_dqo.c
132
rx->com.qpl = gve_alloc_qpl(priv, i + priv->tx_cfg.max_queues,
sys/dev/gve/gve_rx_dqo.c
134
if (rx->com.qpl == NULL) {
sys/dev/gve/gve_rx_dqo.c
157
&rx->dqo.buf_dmatag);
sys/dev/gve/gve_rx_dqo.c
165
for (j = 0; j < rx->dqo.buf_cnt; j++) {
sys/dev/gve/gve_rx_dqo.c
166
err = bus_dmamap_create(rx->dqo.buf_dmatag, 0,
sys/dev/gve/gve_rx_dqo.c
167
&rx->dqo.bufs[j].dmamap);
sys/dev/gve/gve_rx_dqo.c
174
rx->dqo.bufs[j].mapped = true;
sys/dev/gve/gve_rx_dqo.c
185
gve_rx_clear_desc_ring_dqo(struct gve_rx_ring *rx)
sys/dev/gve/gve_rx_dqo.c
187
struct gve_ring_com *com = &rx->com;
sys/dev/gve/gve_rx_dqo.c
193
rx->dqo.desc_ring[i] = (struct gve_rx_desc_dqo){};
sys/dev/gve/gve_rx_dqo.c
195
bus_dmamap_sync(rx->desc_ring_mem.tag, rx->desc_ring_mem.map,
sys/dev/gve/gve_rx_dqo.c
200
gve_rx_clear_compl_ring_dqo(struct gve_rx_ring *rx)
sys/dev/gve/gve_rx_dqo.c
202
struct gve_ring_com *com = &rx->com;
sys/dev/gve/gve_rx_dqo.c
206
rx->dqo.compl_ring[i] = (struct gve_rx_compl_desc_dqo){};
sys/dev/gve/gve_rx_dqo.c
208
bus_dmamap_sync(rx->dqo.compl_ring_mem.tag, rx->dqo.compl_ring_mem.map,
sys/dev/gve/gve_rx_dqo.c
215
struct gve_rx_ring *rx = &priv->rx[i];
sys/dev/gve/gve_rx_dqo.c
218
rx->fill_cnt = 0;
sys/dev/gve/gve_rx_dqo.c
219
rx->cnt = 0;
sys/dev/gve/gve_rx_dqo.c
220
rx->dqo.mask = priv->rx_desc_cnt - 1;
sys/dev/gve/gve_rx_dqo.c
221
rx->dqo.head = 0;
sys/dev/gve/gve_rx_dqo.c
222
rx->dqo.tail = 0;
sys/dev/gve/gve_rx_dqo.c
223
rx->dqo.cur_gen_bit = 0;
sys/dev/gve/gve_rx_dqo.c
225
gve_rx_clear_desc_ring_dqo(rx);
sys/dev/gve/gve_rx_dqo.c
226
gve_rx_clear_compl_ring_dqo(rx);
sys/dev/gve/gve_rx_dqo.c
228
gve_free_rx_mbufs_dqo(rx);
sys/dev/gve/gve_rx_dqo.c
231
SLIST_INIT(&rx->dqo.free_bufs);
sys/dev/gve/gve_rx_dqo.c
232
STAILQ_INIT(&rx->dqo.used_bufs);
sys/dev/gve/gve_rx_dqo.c
234
for (j = 0; j < rx->dqo.buf_cnt; j++) {
sys/dev/gve/gve_rx_dqo.c
235
struct gve_rx_buf_dqo *buf = &rx->dqo.bufs[j];
sys/dev/gve/gve_rx_dqo.c
237
vm_page_t page = rx->com.qpl->pages[buf - rx->dqo.bufs];
sys/dev/gve/gve_rx_dqo.c
245
SLIST_INSERT_HEAD(&rx->dqo.free_bufs,
sys/dev/gve/gve_rx_dqo.c
248
STAILQ_INSERT_TAIL(&rx->dqo.used_bufs,
sys/dev/gve/gve_rx_dqo.c
255
SLIST_INIT(&rx->dqo.free_bufs);
sys/dev/gve/gve_rx_dqo.c
256
for (j = 0; j < rx->dqo.buf_cnt; j++)
sys/dev/gve/gve_rx_dqo.c
257
SLIST_INSERT_HEAD(&rx->dqo.free_bufs,
sys/dev/gve/gve_rx_dqo.c
258
&rx->dqo.bufs[j], slist_entry);
sys/dev/gve/gve_rx_dqo.c
265
struct gve_rx_ring *rx = arg;
sys/dev/gve/gve_rx_dqo.c
266
struct gve_priv *priv = rx->com.priv;
sys/dev/gve/gve_rx_dqo.c
267
struct gve_ring_com *com = &rx->com;
sys/dev/gve/gve_rx_dqo.c
278
gve_rx_advance_head_dqo(struct gve_rx_ring *rx)
sys/dev/gve/gve_rx_dqo.c
280
rx->dqo.head = (rx->dqo.head + 1) & rx->dqo.mask;
sys/dev/gve/gve_rx_dqo.c
281
rx->fill_cnt++; /* rx->fill_cnt is just a sysctl counter */
sys/dev/gve/gve_rx_dqo.c
283
if ((rx->dqo.head & (GVE_RX_BUF_THRESH_DQO - 1)) == 0) {
sys/dev/gve/gve_rx_dqo.c
284
bus_dmamap_sync(rx->desc_ring_mem.tag, rx->desc_ring_mem.map,
sys/dev/gve/gve_rx_dqo.c
286
gve_db_bar_dqo_write_4(rx->com.priv, rx->com.db_offset,
sys/dev/gve/gve_rx_dqo.c
287
rx->dqo.head);
sys/dev/gve/gve_rx_dqo.c
292
gve_rx_post_buf_dqo(struct gve_rx_ring *rx, struct gve_rx_buf_dqo *buf)
sys/dev/gve/gve_rx_dqo.c
296
bus_dmamap_sync(rx->dqo.buf_dmatag, buf->dmamap,
sys/dev/gve/gve_rx_dqo.c
299
desc = &rx->dqo.desc_ring[rx->dqo.head];
sys/dev/gve/gve_rx_dqo.c
300
desc->buf_id = htole16(buf - rx->dqo.bufs);
sys/dev/gve/gve_rx_dqo.c
303
gve_rx_advance_head_dqo(rx);
sys/dev/gve/gve_rx_dqo.c
307
gve_rx_post_new_mbuf_dqo(struct gve_rx_ring *rx, int how)
sys/dev/gve/gve_rx_dqo.c
314
buf = SLIST_FIRST(&rx->dqo.free_bufs);
sys/dev/gve/gve_rx_dqo.c
316
device_printf(rx->com.priv->dev,
sys/dev/gve/gve_rx_dqo.c
320
SLIST_REMOVE_HEAD(&rx->dqo.free_bufs, slist_entry);
sys/dev/gve/gve_rx_dqo.c
322
bus_size_t segment_size = gve_rx_dqo_mbuf_segment_size(rx->com.priv);
sys/dev/gve/gve_rx_dqo.c
327
counter_u64_add_protected(rx->stats.rx_mbuf_mclget_null, 1);
sys/dev/gve/gve_rx_dqo.c
333
err = bus_dmamap_load_mbuf_sg(rx->dqo.buf_dmatag, buf->dmamap,
sys/dev/gve/gve_rx_dqo.c
338
counter_u64_add_protected(rx->stats.rx_mbuf_dmamap_err, 1);
sys/dev/gve/gve_rx_dqo.c
344
gve_rx_post_buf_dqo(rx, buf);
sys/dev/gve/gve_rx_dqo.c
351
SLIST_INSERT_HEAD(&rx->dqo.free_bufs, buf, slist_entry);
sys/dev/gve/gve_rx_dqo.c
356
gve_get_page_dma_handle(struct gve_rx_ring *rx, struct gve_rx_buf_dqo *buf)
sys/dev/gve/gve_rx_dqo.c
358
return (&(rx->com.qpl->dmas[buf - rx->dqo.bufs]));
sys/dev/gve/gve_rx_dqo.c
36
gve_free_rx_mbufs_dqo(struct gve_rx_ring *rx)
sys/dev/gve/gve_rx_dqo.c
362
gve_rx_post_qpl_buf_dqo(struct gve_rx_ring *rx, struct gve_rx_buf_dqo *buf,
sys/dev/gve/gve_rx_dqo.c
365
struct gve_rx_desc_dqo *desc = &rx->dqo.desc_ring[rx->dqo.head];
sys/dev/gve/gve_rx_dqo.c
369
composed_id.buf_id = buf - rx->dqo.bufs;
sys/dev/gve/gve_rx_dqo.c
373
page_dma_handle = gve_get_page_dma_handle(rx, buf);
sys/dev/gve/gve_rx_dqo.c
377
frag_num * rx->com.priv->rx_buf_size_dqo);
sys/dev/gve/gve_rx_dqo.c
380
gve_rx_advance_head_dqo(rx);
sys/dev/gve/gve_rx_dqo.c
384
gve_rx_maybe_extract_from_used_bufs(struct gve_rx_ring *rx, bool just_one)
sys/dev/gve/gve_rx_dqo.c
392
buf = STAILQ_FIRST(&rx->dqo.used_bufs);
sys/dev/gve/gve_rx_dqo.c
396
page = rx->com.qpl->pages[buf - rx->dqo.bufs];
sys/dev/gve/gve_rx_dqo.c
404
STAILQ_REMOVE_HEAD(&rx->dqo.used_bufs,
sys/dev/gve/gve_rx_dqo.c
409
STAILQ_REMOVE_HEAD(&rx->dqo.used_bufs,
sys/dev/gve/gve_rx_dqo.c
41
if (gve_is_qpl(rx->com.priv))
sys/dev/gve/gve_rx_dqo.c
411
SLIST_INSERT_HEAD(&rx->dqo.free_bufs,
sys/dev/gve/gve_rx_dqo.c
418
STAILQ_INSERT_HEAD(&rx->dqo.used_bufs,
sys/dev/gve/gve_rx_dqo.c
423
gve_rx_post_new_dqo_qpl_buf(struct gve_rx_ring *rx)
sys/dev/gve/gve_rx_dqo.c
427
buf = SLIST_FIRST(&rx->dqo.free_bufs);
sys/dev/gve/gve_rx_dqo.c
429
gve_rx_maybe_extract_from_used_bufs(rx, /*just_one=*/true);
sys/dev/gve/gve_rx_dqo.c
430
buf = SLIST_FIRST(&rx->dqo.free_bufs);
sys/dev/gve/gve_rx_dqo.c
435
gve_rx_post_qpl_buf_dqo(rx, buf, buf->next_idx);
sys/dev/gve/gve_rx_dqo.c
436
if (buf->next_idx == gve_get_dq_num_frags_in_page(rx->com.priv) - 1)
sys/dev/gve/gve_rx_dqo.c
44
for (i = 0; i < rx->dqo.buf_cnt; i++) {
sys/dev/gve/gve_rx_dqo.c
448
SLIST_REMOVE_HEAD(&rx->dqo.free_bufs, slist_entry);
sys/dev/gve/gve_rx_dqo.c
45
buf = &rx->dqo.bufs[i];
sys/dev/gve/gve_rx_dqo.c
453
gve_rx_post_buffers_dqo(struct gve_rx_ring *rx, int how)
sys/dev/gve/gve_rx_dqo.c
460
num_pending_bufs = (rx->dqo.head - rx->dqo.tail) & rx->dqo.mask;
sys/dev/gve/gve_rx_dqo.c
461
num_to_post = rx->dqo.mask - num_pending_bufs;
sys/dev/gve/gve_rx_dqo.c
464
if (gve_is_qpl(rx->com.priv))
sys/dev/gve/gve_rx_dqo.c
465
err = gve_rx_post_new_dqo_qpl_buf(rx);
sys/dev/gve/gve_rx_dqo.c
467
err = gve_rx_post_new_mbuf_dqo(rx, how);
sys/dev/gve/gve_rx_dqo.c
474
gve_rx_prefill_buffers_dqo(struct gve_rx_ring *rx)
sys/dev/gve/gve_rx_dqo.c
476
gve_rx_post_buffers_dqo(rx, M_WAITOK);
sys/dev/gve/gve_rx_dqo.c
49
bus_dmamap_sync(rx->dqo.buf_dmatag, buf->dmamap,
sys/dev/gve/gve_rx_dqo.c
51
bus_dmamap_unload(rx->dqo.buf_dmatag, buf->dmamap);
sys/dev/gve/gve_rx_dqo.c
553
gve_rx_input_mbuf_dqo(struct gve_rx_ring *rx,
sys/dev/gve/gve_rx_dqo.c
556
struct mbuf *mbuf = rx->ctx.mbuf_head;
sys/dev/gve/gve_rx_dqo.c
557
if_t ifp = rx->com.priv->ifp;
sys/dev/gve/gve_rx_dqo.c
562
ptype = &rx->com.priv->ptype_lut_dqo->ptypes[compl_desc->packet_type];
sys/dev/gve/gve_rx_dqo.c
568
mbuf->m_pkthdr.len = rx->ctx.total_size;
sys/dev/gve/gve_rx_dqo.c
570
if (((if_getcapenable(rx->com.priv->ifp) & IFCAP_LRO) != 0) &&
sys/dev/gve/gve_rx_dqo.c
572
(rx->lro.lro_cnt != 0) &&
sys/dev/gve/gve_rx_dqo.c
573
(tcp_lro_rx(&rx->lro, mbuf, 0) == 0))
sys/dev/gve/gve_rx_dqo.c
580
counter_u64_add_protected(rx->stats.rbytes, rx->ctx.total_size);
sys/dev/gve/gve_rx_dqo.c
581
counter_u64_add_protected(rx->stats.rpackets, 1);
sys/dev/gve/gve_rx_dqo.c
584
rx->ctx = (struct gve_rx_ctx){};
sys/dev/gve/gve_rx_dqo.c
588
gve_rx_copybreak_dqo(struct gve_rx_ring *rx, void *va,
sys/dev/gve/gve_rx_dqo.c
598
counter_u64_add_protected(rx->stats.rx_copybreak_cnt, 1);
sys/dev/gve/gve_rx_dqo.c
60
struct gve_rx_ring *rx = &priv->rx[i];
sys/dev/gve/gve_rx_dqo.c
604
rx->ctx.mbuf_head = mbuf;
sys/dev/gve/gve_rx_dqo.c
605
rx->ctx.mbuf_tail = mbuf;
sys/dev/gve/gve_rx_dqo.c
606
rx->ctx.total_size += frag_len;
sys/dev/gve/gve_rx_dqo.c
608
gve_rx_input_mbuf_dqo(rx, compl_desc);
sys/dev/gve/gve_rx_dqo.c
61
struct gve_ring_com *com = &rx->com;
sys/dev/gve/gve_rx_dqo.c
613
gve_rx_dqo(struct gve_priv *priv, struct gve_rx_ring *rx,
sys/dev/gve/gve_rx_dqo.c
618
struct gve_rx_ctx *ctx = &rx->ctx;
sys/dev/gve/gve_rx_dqo.c
626
if (__predict_false(buf_id >= rx->dqo.buf_cnt)) {
sys/dev/gve/gve_rx_dqo.c
628
buf_id, rx->com.id);
sys/dev/gve/gve_rx_dqo.c
632
buf = &rx->dqo.bufs[buf_id];
sys/dev/gve/gve_rx_dqo.c
635
buf_id, rx->com.id);
sys/dev/gve/gve_rx_dqo.c
64
if (rx->dqo.compl_ring != NULL) {
sys/dev/gve/gve_rx_dqo.c
645
counter_u64_add_protected(rx->stats.rx_dropped_pkt_desc_err, 1);
sys/dev/gve/gve_rx_dqo.c
65
gve_dma_free_coherent(&rx->dqo.compl_ring_mem);
sys/dev/gve/gve_rx_dqo.c
650
bus_dmamap_sync(rx->dqo.buf_dmatag, buf->dmamap,
sys/dev/gve/gve_rx_dqo.c
655
err = gve_rx_copybreak_dqo(rx, mtod(buf->mbuf, char*),
sys/dev/gve/gve_rx_dqo.c
66
rx->dqo.compl_ring = NULL;
sys/dev/gve/gve_rx_dqo.c
660
gve_rx_post_buf_dqo(rx, buf);
sys/dev/gve/gve_rx_dqo.c
671
num_pending_bufs = (rx->dqo.head - rx->dqo.tail) & rx->dqo.mask;
sys/dev/gve/gve_rx_dqo.c
684
err = gve_rx_post_new_mbuf_dqo(rx, M_NOWAIT);
sys/dev/gve/gve_rx_dqo.c
689
rx->stats.rx_dropped_pkt_mbuf_alloc_fail, 1);
sys/dev/gve/gve_rx_dqo.c
69
if (rx->dqo.desc_ring != NULL) {
sys/dev/gve/gve_rx_dqo.c
70
gve_dma_free_coherent(&rx->desc_ring_mem);
sys/dev/gve/gve_rx_dqo.c
709
bus_dmamap_unload(rx->dqo.buf_dmatag, buf->dmamap);
sys/dev/gve/gve_rx_dqo.c
71
rx->dqo.desc_ring = NULL;
sys/dev/gve/gve_rx_dqo.c
712
SLIST_INSERT_HEAD(&rx->dqo.free_bufs, buf, slist_entry);
sys/dev/gve/gve_rx_dqo.c
715
gve_rx_input_mbuf_dqo(rx, compl_desc);
sys/dev/gve/gve_rx_dqo.c
723
rx->ctx = (struct gve_rx_ctx){};
sys/dev/gve/gve_rx_dqo.c
727
gve_rx_post_buf_dqo(rx, buf);
sys/dev/gve/gve_rx_dqo.c
735
counter_u64_add_protected(rx->stats.rx_dropped_pkt, 1);
sys/dev/gve/gve_rx_dqo.c
738
rx->ctx = (struct gve_rx_ctx){};
sys/dev/gve/gve_rx_dqo.c
74
if (rx->dqo.bufs != NULL) {
sys/dev/gve/gve_rx_dqo.c
742
gve_get_cpu_addr_for_qpl_buf(struct gve_rx_ring *rx,
sys/dev/gve/gve_rx_dqo.c
745
int page_idx = buf - rx->dqo.bufs;
sys/dev/gve/gve_rx_dqo.c
746
void *va = rx->com.qpl->dmas[page_idx].cpu_addr;
sys/dev/gve/gve_rx_dqo.c
748
va = (char *)va + (buf_frag_num * rx->com.priv->rx_buf_size_dqo);
sys/dev/gve/gve_rx_dqo.c
75
gve_free_rx_mbufs_dqo(rx);
sys/dev/gve/gve_rx_dqo.c
753
gve_rx_add_clmbuf_to_ctx(struct gve_rx_ring *rx,
sys/dev/gve/gve_rx_dqo.c
757
void *va = gve_get_cpu_addr_for_qpl_buf(rx, buf, buf_frag_num);
sys/dev/gve/gve_rx_dqo.c
759
bus_size_t segment_size = gve_rx_dqo_mbuf_segment_size(rx->com.priv);
sys/dev/gve/gve_rx_dqo.c
77
if (!gve_is_qpl(priv) && rx->dqo.buf_dmatag) {
sys/dev/gve/gve_rx_dqo.c
78
for (j = 0; j < rx->dqo.buf_cnt; j++)
sys/dev/gve/gve_rx_dqo.c
780
counter_u64_add_protected(rx->stats.rx_frag_copy_cnt, 1);
sys/dev/gve/gve_rx_dqo.c
786
gve_rx_add_extmbuf_to_ctx(struct gve_rx_ring *rx,
sys/dev/gve/gve_rx_dqo.c
79
if (rx->dqo.bufs[j].mapped)
sys/dev/gve/gve_rx_dqo.c
80
bus_dmamap_destroy(rx->dqo.buf_dmatag,
sys/dev/gve/gve_rx_dqo.c
81
rx->dqo.bufs[j].dmamap);
sys/dev/gve/gve_rx_dqo.c
813
page_idx = buf - rx->dqo.bufs;
sys/dev/gve/gve_rx_dqo.c
814
page = rx->com.qpl->pages[page_idx];
sys/dev/gve/gve_rx_dqo.c
815
page_addr = rx->com.qpl->dmas[page_idx].cpu_addr;
sys/dev/gve/gve_rx_dqo.c
816
va = (char *)page_addr + (buf_frag_num * rx->com.priv->rx_buf_size_dqo);
sys/dev/gve/gve_rx_dqo.c
825
counter_u64_add_protected(rx->stats.rx_frag_flip_cnt, 1);
sys/dev/gve/gve_rx_dqo.c
835
gve_rx_dqo_qpl(struct gve_priv *priv, struct gve_rx_ring *rx,
sys/dev/gve/gve_rx_dqo.c
84
free(rx->dqo.bufs, M_GVE);
sys/dev/gve/gve_rx_dqo.c
842
struct gve_rx_ctx *ctx = &rx->ctx;
sys/dev/gve/gve_rx_dqo.c
85
rx->dqo.bufs = NULL;
sys/dev/gve/gve_rx_dqo.c
854
if (__predict_false(buf_id >= rx->dqo.buf_cnt)) {
sys/dev/gve/gve_rx_dqo.c
856
buf_id, rx->com.id);
sys/dev/gve/gve_rx_dqo.c
860
buf = &rx->dqo.bufs[buf_id];
sys/dev/gve/gve_rx_dqo.c
865
buf_id, rx->com.id, buf_frag_num, buf->num_nic_frags);
sys/dev/gve/gve_rx_dqo.c
877
counter_u64_add_protected(rx->stats.rx_dropped_pkt_desc_err, 1);
sys/dev/gve/gve_rx_dqo.c
88
if (!gve_is_qpl(priv) && rx->dqo.buf_dmatag)
sys/dev/gve/gve_rx_dqo.c
882
page_dma_handle = gve_get_page_dma_handle(rx, buf);
sys/dev/gve/gve_rx_dqo.c
888
void *va = gve_get_cpu_addr_for_qpl_buf(rx, buf, buf_frag_num);
sys/dev/gve/gve_rx_dqo.c
89
bus_dma_tag_destroy(rx->dqo.buf_dmatag);
sys/dev/gve/gve_rx_dqo.c
890
err = gve_rx_copybreak_dqo(rx, va, compl_desc, frag_len);
sys/dev/gve/gve_rx_dqo.c
894
gve_rx_post_qpl_buf_dqo(rx, buf, buf_frag_num);
sys/dev/gve/gve_rx_dqo.c
898
num_pending_bufs = (rx->dqo.head - rx->dqo.tail) & rx->dqo.mask;
sys/dev/gve/gve_rx_dqo.c
899
err = gve_rx_post_new_dqo_qpl_buf(rx);
sys/dev/gve/gve_rx_dqo.c
908
err = gve_rx_add_clmbuf_to_ctx(rx, ctx, buf,
sys/dev/gve/gve_rx_dqo.c
913
rx->stats.rx_dropped_pkt_buf_post_fail, 1);
sys/dev/gve/gve_rx_dqo.c
917
gve_rx_post_qpl_buf_dqo(rx, buf, buf_frag_num);
sys/dev/gve/gve_rx_dqo.c
919
err = gve_rx_add_extmbuf_to_ctx(rx, ctx, buf,
sys/dev/gve/gve_rx_dqo.c
924
rx->stats.rx_dropped_pkt_mbuf_alloc_fail, 1);
sys/dev/gve/gve_rx_dqo.c
940
STAILQ_INSERT_TAIL(&rx->dqo.used_bufs, buf, stailq_entry);
sys/dev/gve/gve_rx_dqo.c
943
gve_rx_input_mbuf_dqo(rx, compl_desc);
sys/dev/gve/gve_rx_dqo.c
951
rx->ctx = (struct gve_rx_ctx){};
sys/dev/gve/gve_rx_dqo.c
955
gve_rx_post_qpl_buf_dqo(rx, buf, buf_frag_num);
sys/dev/gve/gve_rx_dqo.c
963
counter_u64_add_protected(rx->stats.rx_dropped_pkt, 1);
sys/dev/gve/gve_rx_dqo.c
966
rx->ctx = (struct gve_rx_ctx){};
sys/dev/gve/gve_rx_dqo.c
983
gve_rx_cleanup_dqo(struct gve_priv *priv, struct gve_rx_ring *rx, int budget)
sys/dev/gve/gve_rx_dqo.c
991
bus_dmamap_sync(rx->dqo.compl_ring_mem.tag,
sys/dev/gve/gve_rx_dqo.c
992
rx->dqo.compl_ring_mem.map,
sys/dev/gve/gve_rx_dqo.c
995
compl_desc = &rx->dqo.compl_ring[rx->dqo.tail];
sys/dev/gve/gve_rx_dqo.c
997
rx->dqo.cur_gen_bit)
sys/dev/gve/gve_sysctl.c
207
gve_setup_rxq_sysctl(ctx, child, &priv->rx[i]);
sys/dev/gve/gve_sysctl.c
493
rxqstats = &priv->rx[i].stats;
sys/dev/gve/gve_utils.c
281
struct gve_rx_ring *rx;
sys/dev/gve/gve_utils.c
331
rx = &priv->rx[j];
sys/dev/gve/gve_utils.c
332
com = &rx->com;
sys/dev/gve/gve_utils.c
346
&priv->rx[j], &irq->cookie);
sys/dev/gve/gve_utils.c
408
struct gve_rx_ring *rx;
sys/dev/gve/gve_utils.c
421
rx = &priv->rx[idx];
sys/dev/gve/gve_utils.c
423
gve_db_bar_write_4(priv, rx->com.irq_db_offset, 0);
sys/dev/gve/gve_utils.c
425
gve_db_bar_dqo_write_4(priv, rx->com.irq_db_offset,
sys/dev/gve/gve_utils.c
438
struct gve_rx_ring *rx = &priv->rx[idx];
sys/dev/gve/gve_utils.c
439
gve_db_bar_write_4(priv, rx->com.irq_db_offset, GVE_IRQ_MASK);
sys/dev/ice/ice_switch.h
267
u8 rx; /* true means LOOKUP_RX otherwise LOOKUP_TX */
sys/dev/iwm/if_iwm.c
3124
sc->sc_noise = iwm_get_noise(sc, &stats->rx.general);
sys/dev/iwm/if_iwmreg.h
3720
struct iwm_statistics_rx rx;
sys/dev/iwn/if_iwn.c
3426
memcpy(&lstats->rx.ofdm, &stats_bt->rx_bt.ofdm,
sys/dev/iwn/if_iwn.c
3428
memcpy(&lstats->rx.cck, &stats_bt->rx_bt.cck,
sys/dev/iwn/if_iwn.c
3430
memcpy(&lstats->rx.general, &stats_bt->rx_bt.general_bt.common,
sys/dev/iwn/if_iwn.c
3432
memcpy(&lstats->rx.ht, &stats_bt->rx_bt.ht,
sys/dev/iwn/if_iwn.c
3504
sc->noise = iwn_get_noise(&lstats->rx.general);
sys/dev/iwn/if_iwn.c
3508
if (le32toh(lstats->rx.general.flags) != 1) {
sys/dev/iwn/if_iwn.c
3515
iwn_collect_noise(sc, &lstats->rx.general);
sys/dev/iwn/if_iwn.c
3517
iwn_tune_sensitivity(sc, &lstats->rx);
sys/dev/iwn/if_iwn.c
3538
calib->bad_plcp_cck = le32toh(rs->rx.cck.bad_plcp);
sys/dev/iwn/if_iwn.c
3539
calib->fa_cck = le32toh(rs->rx.cck.fa);
sys/dev/iwn/if_iwn.c
3540
calib->bad_plcp_ht = le32toh(rs->rx.ht.bad_plcp);
sys/dev/iwn/if_iwn.c
3541
calib->bad_plcp_ofdm = le32toh(rs->rx.ofdm.bad_plcp);
sys/dev/iwn/if_iwn.c
3542
calib->fa_ofdm = le32toh(rs->rx.ofdm.fa);
sys/dev/iwn/if_iwn.c
6398
delta_cck = le32toh(rs->rx.cck.bad_plcp) - calib->bad_plcp_cck;
sys/dev/iwn/if_iwn.c
6399
delta_ofdm = le32toh(rs->rx.ofdm.bad_plcp) - calib->bad_plcp_ofdm;
sys/dev/iwn/if_iwn.c
6400
delta_ht = le32toh(rs->rx.ht.bad_plcp) - calib->bad_plcp_ht;
sys/dev/iwn/if_iwnreg.h
1719
struct iwn_rx_stats rx;
sys/dev/iwx/if_iwx.c
9149
sc->sc_noise = iwx_get_noise(&stats->rx.general);
sys/dev/iwx/if_iwxreg.h
4107
struct iwx_statistics_rx rx;
sys/dev/mana/mana.h
559
int mana_config_rss(struct mana_port_context *ac, enum TRI_STATE rx,
sys/dev/mana/mana_en.c
1146
enum TRI_STATE rx,
sys/dev/mana/mana_en.c
1167
req->rx_enable = rx;
sys/dev/mana/mana_en.c
2624
int mana_config_rss(struct mana_port_context *apc, enum TRI_STATE rx,
sys/dev/mana/mana_en.c
2638
err = mana_cfg_vport_steering(apc, rx, true, update_hash, update_tab);
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1380
struct mlx5e_ipsec_rx *rx;
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1383
rx = (attrs->family == AF_INET) ? ipsec->rx_ipv4 : ipsec->rx_ipv6;
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1384
ft = rx->chains ? ipsec_chains_get_table(rx->chains, attrs->prio) : rx->ft.pol;
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1401
dest[dstn].counter_id = mlx5_fc_id(rx->fc->drop);
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1411
dest[dstn].ft = rx->ft.sa;
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1458
if (rx->chains != NULL)
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1459
ipsec_chains_put_table(rx->chains, attrs->prio);
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1546
struct mlx5e_ipsec_rx *rx,
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1584
rule = mlx5_add_flow_rules(rx->ft.status, spec, &flow_act, dest, 2);
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1592
rx->status.rule = rule;
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1593
rx->status.modify_hdr = modify_hdr;
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1614
static void ipsec_fs_rx_catchall_rules_destroy(struct mlx5_core_dev *mdev, struct mlx5e_ipsec_rx *rx)
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1616
mutex_lock(&rx->ft.mutex);
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1617
mlx5_del_flow_rules(&rx->sa.rule);
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1618
mlx5_destroy_flow_group(rx->sa.group);
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1619
if (rx->chains == NULL) {
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1620
mlx5_del_flow_rules(&rx->pol.rule);
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1621
mlx5_destroy_flow_group(rx->pol.group);
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1623
mlx5_del_flow_rules(&rx->status.rule);
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1624
mlx5_modify_header_dealloc(mdev, rx->status.modify_hdr);
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1625
ipsec_fs_rx_roce_rules_destroy(&rx->roce);
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1626
mutex_unlock(&rx->ft.mutex);
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1648
static void ipsec_fs_rx_table_destroy(struct mlx5_core_dev *mdev, struct mlx5e_ipsec_rx *rx)
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1650
if (rx->chains) {
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1651
ipsec_chains_destroy(rx->chains);
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1653
mlx5_del_flow_rules(&rx->pol.rule);
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1654
mlx5_destroy_flow_table(rx->ft.pol);
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1656
mlx5_destroy_flow_table(rx->ft.sa);
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1657
mlx5_destroy_flow_table(rx->ft.status);
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1658
ipsec_fs_rx_roce_table_destroy(&rx->roce);
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1720
static int ipsec_roce_rx_rules(struct mlx5e_ipsec_rx *rx, struct mlx5_flow_destination *defdst,
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1731
if (!rx->roce.ns_rdma)
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1747
g = mlx5_create_flow_group(rx->roce.ft, in);
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1753
rx->roce.g = g;
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1759
g = mlx5_create_flow_group(rx->roce.ft, in);
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1766
rx->roce.roce_miss.group = g;
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1768
err = ipsec_roce_rx_rule_setup(defdst, &rx->roce, mdev);
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1776
mlx5_destroy_flow_group(rx->roce.roce_miss.group);
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1778
mlx5_destroy_flow_group(rx->roce.g);
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1785
struct mlx5e_ipsec_rx *rx,
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1792
mutex_lock(&rx->ft.mutex);
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1794
err = ipsec_roce_rx_rules(rx, defdst, mdev);
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1800
if (rx->roce.ft)
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1801
dest[0].ft = rx->roce.ft;
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1806
dest[1].counter_id = mlx5_fc_id(rx->fc->cnt);
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1807
err = ipsec_status_rule(mdev, rx, dest);
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1811
if (!rx->chains) {
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1813
err = ipsec_miss_create(mdev, rx->ft.pol, &rx->pol, defdst);
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1822
err = ipsec_miss_create(mdev, rx->ft.sa, &rx->sa, defdst);
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1826
mutex_unlock(&rx->ft.mutex);
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1830
if (!rx->chains) {
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1831
mlx5_del_flow_rules(&rx->pol.rule);
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1832
mlx5_destroy_flow_group(rx->pol.group);
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1835
mlx5_del_flow_rules(&rx->status.rule);
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1836
mlx5_modify_header_dealloc(mdev, rx->status.modify_hdr);
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1838
ipsec_fs_rx_roce_rules_destroy(&rx->roce);
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1840
mutex_unlock(&rx->ft.mutex);
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1844
static int ipsec_fs_rx_roce_tables_create(struct mlx5e_ipsec_rx *rx,
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1851
if (!rx->roce.ns_rdma)
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1856
ft = mlx5_create_flow_table(rx->ns, &ft_attr);
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1861
rx->roce.ft = ft;
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1865
ft = mlx5_create_flow_table(rx->roce.ns_rdma, &ft_attr);
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1870
rx->roce.ft_rdma = ft;
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1874
mlx5_destroy_flow_table(rx->roce.ft);
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1875
rx->roce.ft = NULL;
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1963
static int ipsec_fs_rx_table_create(struct mlx5_core_dev *mdev, struct mlx5e_ipsec_rx *rx,
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1966
struct mlx5_flow_namespace *ns = rx->ns;
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1970
mutex_lock(&rx->ft.mutex);
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1978
rx->ft.sa = ft;
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1986
rx->ft.status = ft;
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1989
rx->chains = ipsec_chains_create(mdev, rx->ft.sa,
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1991
rx_init_level, &rx->ft.pol);
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1992
if (IS_ERR(rx->chains)) {
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
1993
err = PTR_ERR(rx->chains);
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
2002
rx->ft.pol = ft;
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
2006
err = ipsec_fs_rx_roce_tables_create(rx, rx_init_level + 3,
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
2014
mlx5_destroy_flow_table(rx->ft.pol);
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
2016
mlx5_destroy_flow_table(rx->ft.status);
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
2018
mlx5_destroy_flow_table(rx->ft.sa);
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
2020
mutex_unlock(&rx->ft.mutex);
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
2105
struct mlx5e_ipsec_rx *rx;
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
2107
rx = (pol_entry->attrs.family == AF_INET)
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
2110
if (rx->chains)
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
2111
ipsec_chains_put_table(rx->chains,
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
528
struct mlx5e_ipsec_rx *rx;
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
532
rx = (attrs->family == AF_INET) ? ipsec->rx_ipv4 : ipsec->rx_ipv6;
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
569
dest[0].ft = rx->ft.status;
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
589
rule = mlx5_add_flow_rules(rx->ft.sa, spec, &flow_act, dest, 2);
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_fs.c
601
rule = mlx5_add_flow_rules(rx->ft.sa, spec, &flow_act, dest, 2);
sys/dev/mxge/if_mxge.c
2327
mxge_rx_ring_t *rx = &ss->rx_small;
sys/dev/mxge/if_mxge.c
2332
rx->alloc_fail++;
sys/dev/mxge/if_mxge.c
2337
err = bus_dmamap_load_mbuf_sg(rx->dmat, map, m,
sys/dev/mxge/if_mxge.c
2343
rx->info[idx].m = m;
sys/dev/mxge/if_mxge.c
2344
rx->shadow[idx].addr_low =
sys/dev/mxge/if_mxge.c
2346
rx->shadow[idx].addr_high =
sys/dev/mxge/if_mxge.c
2351
mxge_submit_8rx(&rx->lanai[idx - 7], &rx->shadow[idx - 7]);
sys/dev/mxge/if_mxge.c
2360
mxge_rx_ring_t *rx = &ss->rx_big;
sys/dev/mxge/if_mxge.c
2363
m = m_getjcl(M_NOWAIT, MT_DATA, M_PKTHDR, rx->cl_size);
sys/dev/mxge/if_mxge.c
2365
rx->alloc_fail++;
sys/dev/mxge/if_mxge.c
2369
m->m_len = rx->mlen;
sys/dev/mxge/if_mxge.c
2370
err = bus_dmamap_load_mbuf_sg(rx->dmat, map, m,
sys/dev/mxge/if_mxge.c
2376
rx->info[idx].m = m;
sys/dev/mxge/if_mxge.c
2377
rx->shadow[idx].addr_low =
sys/dev/mxge/if_mxge.c
2379
rx->shadow[idx].addr_high =
sys/dev/mxge/if_mxge.c
2383
for (i = 0; i < rx->nbufs; i++) {
sys/dev/mxge/if_mxge.c
2385
mxge_submit_8rx(&rx->lanai[idx - 7],
sys/dev/mxge/if_mxge.c
2386
&rx->shadow[idx - 7]);
sys/dev/mxge/if_mxge.c
2560
mxge_rx_ring_t *rx;
sys/dev/mxge/if_mxge.c
2566
rx = &ss->rx_big;
sys/dev/mxge/if_mxge.c
2567
idx = rx->cnt & rx->mask;
sys/dev/mxge/if_mxge.c
2568
rx->cnt += rx->nbufs;
sys/dev/mxge/if_mxge.c
2570
m = rx->info[idx].m;
sys/dev/mxge/if_mxge.c
2572
if (mxge_get_buf_big(ss, rx->extra_map, idx)) {
sys/dev/mxge/if_mxge.c
2579
old_map = rx->info[idx].map;
sys/dev/mxge/if_mxge.c
2580
bus_dmamap_sync(rx->dmat, old_map, BUS_DMASYNC_POSTREAD);
sys/dev/mxge/if_mxge.c
2581
bus_dmamap_unload(rx->dmat, old_map);
sys/dev/mxge/if_mxge.c
2584
rx->info[idx].map = rx->extra_map;
sys/dev/mxge/if_mxge.c
2585
rx->extra_map = old_map;
sys/dev/mxge/if_mxge.c
2628
mxge_rx_ring_t *rx;
sys/dev/mxge/if_mxge.c
2634
rx = &ss->rx_small;
sys/dev/mxge/if_mxge.c
2635
idx = rx->cnt & rx->mask;
sys/dev/mxge/if_mxge.c
2636
rx->cnt++;
sys/dev/mxge/if_mxge.c
2638
m = rx->info[idx].m;
sys/dev/mxge/if_mxge.c
2640
if (mxge_get_buf_small(ss, rx->extra_map, idx)) {
sys/dev/mxge/if_mxge.c
2647
old_map = rx->info[idx].map;
sys/dev/mxge/if_mxge.c
2648
bus_dmamap_sync(rx->dmat, old_map, BUS_DMASYNC_POSTREAD);
sys/dev/mxge/if_mxge.c
2649
bus_dmamap_unload(rx->dmat, old_map);
sys/dev/mxge/if_mxge.c
2652
rx->info[idx].map = rx->extra_map;
sys/dev/mxge/if_mxge.c
2653
rx->extra_map = old_map;
sys/dev/neta/if_mvneta.c
1010
struct mvneta_rx_ring *rx;
sys/dev/neta/if_mvneta.c
1016
rx = MVNETA_RX_RING(sc, q);
sys/dev/neta/if_mvneta.c
1018
rx->queue_status = MVNETA_QUEUE_DISABLED;
sys/dev/neta/if_mvneta.c
1203
struct mvneta_rx_ring *rx;
sys/dev/neta/if_mvneta.c
1211
rx = MVNETA_RX_RING(sc, q);
sys/dev/neta/if_mvneta.c
1212
mtx_init(&rx->ring_mtx, "mvneta_rx", NULL, MTX_DEF);
sys/dev/neta/if_mvneta.c
1215
(void**)&(rx->desc),
sys/dev/neta/if_mvneta.c
1217
&rx->desc_map);
sys/dev/neta/if_mvneta.c
1218
if (error != 0 || rx->desc == NULL)
sys/dev/neta/if_mvneta.c
1220
error = bus_dmamap_load(sc->rx_dtag, rx->desc_map,
sys/dev/neta/if_mvneta.c
1221
rx->desc,
sys/dev/neta/if_mvneta.c
1223
mvneta_dmamap_cb, &rx->desc_pa, BUS_DMA_NOWAIT);
sys/dev/neta/if_mvneta.c
1234
rxbuf = &rx->rxbuf[i];
sys/dev/neta/if_mvneta.c
1346
struct mvneta_rx_ring *rx;
sys/dev/neta/if_mvneta.c
1353
rx = MVNETA_RX_RING(sc, q);
sys/dev/neta/if_mvneta.c
1355
if (rx->desc_pa != 0)
sys/dev/neta/if_mvneta.c
1356
bus_dmamap_unload(sc->rx_dtag, rx->desc_map);
sys/dev/neta/if_mvneta.c
1358
kva = (void *)rx->desc;
sys/dev/neta/if_mvneta.c
1360
bus_dmamem_free(sc->rx_dtag, rx->desc, rx->desc_map);
sys/dev/neta/if_mvneta.c
1362
lro = &rx->lro;
sys/dev/neta/if_mvneta.c
1365
if (mtx_name(&rx->ring_mtx) != NULL)
sys/dev/neta/if_mvneta.c
1366
mtx_destroy(&rx->ring_mtx);
sys/dev/neta/if_mvneta.c
1368
memset(rx, 0, sizeof(*rx));
sys/dev/neta/if_mvneta.c
1374
struct mvneta_rx_ring *rx;
sys/dev/neta/if_mvneta.c
1381
rx = MVNETA_RX_RING(sc, q);
sys/dev/neta/if_mvneta.c
1382
rx->dma = rx->cpu = 0;
sys/dev/neta/if_mvneta.c
1383
rx->queue_th_received = MVNETA_RXTH_COUNT;
sys/dev/neta/if_mvneta.c
1384
rx->queue_th_time = (sc->clk_freq / 1000) / 10; /* 0.1 [ms] */
sys/dev/neta/if_mvneta.c
1387
rx->lro_enabled = FALSE;
sys/dev/neta/if_mvneta.c
1389
lro = &rx->lro;
sys/dev/neta/if_mvneta.c
1394
rx->lro_enabled = TRUE;
sys/dev/neta/if_mvneta.c
1470
struct mvneta_rx_ring *rx;
sys/dev/neta/if_mvneta.c
1474
rx = MVNETA_RX_RING(sc, q);
sys/dev/neta/if_mvneta.c
1479
rxbuf = &rx->rxbuf[i];
sys/dev/neta/if_mvneta.c
1482
rx->dma = rx->cpu = 0;
sys/dev/neta/if_mvneta.c
1492
struct mvneta_rx_ring *rx;
sys/dev/neta/if_mvneta.c
1497
rx = MVNETA_RX_RING(sc, q);
sys/dev/neta/if_mvneta.c
1498
DASSERT(rx->desc_pa != 0);
sys/dev/neta/if_mvneta.c
1501
MVNETA_WRITE(sc, MVNETA_PRXDQA(q), rx->desc_pa);
sys/dev/neta/if_mvneta.c
1520
DASSERT(MVNETA_READ(sc, MVNETA_PRXDQA(q)) == rx->desc_pa);
sys/dev/neta/if_mvneta.c
1552
struct mvneta_rx_ring *rx;
sys/dev/neta/if_mvneta.c
1556
rx = MVNETA_RX_RING(sc, q);
sys/dev/neta/if_mvneta.c
1560
reg = MVNETA_PRXDQTH_ODT(rx->queue_th_received);
sys/dev/neta/if_mvneta.c
1563
reg = MVNETA_PRXITTH_RITT(rx->queue_th_time);
sys/dev/neta/if_mvneta.c
1576
rx->queue_status = MVNETA_QUEUE_WORKING;
sys/dev/neta/if_mvneta.c
1753
struct mvneta_rx_ring *rx;
sys/dev/neta/if_mvneta.c
1787
rx = MVNETA_RX_RING(sc, q);
sys/dev/neta/if_mvneta.c
1788
if (rx->needs_refill == TRUE) {
sys/dev/neta/if_mvneta.c
2044
struct mvneta_rx_ring *rx;
sys/dev/neta/if_mvneta.c
2100
rx = MVNETA_RX_RING(sc, q);
sys/dev/neta/if_mvneta.c
2101
rx->lro_enabled = !rx->lro_enabled;
sys/dev/neta/if_mvneta.c
2999
struct mvneta_rx_ring *rx;
sys/dev/neta/if_mvneta.c
3009
rx = MVNETA_RX_RING(sc, q);
sys/dev/neta/if_mvneta.c
3012
if (__predict_false(rx->queue_status == MVNETA_QUEUE_DISABLED))
sys/dev/neta/if_mvneta.c
3015
bus_dmamap_sync(sc->rx_dtag, rx->desc_map,
sys/dev/neta/if_mvneta.c
3020
ndma = rx_counter_adv(rx->dma, 1);
sys/dev/neta/if_mvneta.c
3021
mvneta_prefetch(&rx->desc[ndma]);
sys/dev/neta/if_mvneta.c
3022
mvneta_prefetch(&rx->rxbuf[ndma]);
sys/dev/neta/if_mvneta.c
3025
r = &rx->desc[rx->dma];
sys/dev/neta/if_mvneta.c
3026
rxbuf = &rx->rxbuf[rx->dma];
sys/dev/neta/if_mvneta.c
3050
pktbuf = (uint8_t *)rx->rxbuf_virt_addr[rx->dma] + MVNETA_PACKET_OFFSET +
sys/dev/neta/if_mvneta.c
3063
rx->dma = ndma;
sys/dev/neta/if_mvneta.c
3065
if (__predict_false(rx->lro_enabled &&
sys/dev/neta/if_mvneta.c
3071
if (rx->lro.lro_cnt != 0) {
sys/dev/neta/if_mvneta.c
3072
if (tcp_lro_rx(&rx->lro, m, 0) == 0)
sys/dev/neta/if_mvneta.c
3084
if(__predict_false(rx->queue_status == MVNETA_QUEUE_DISABLED))
sys/dev/neta/if_mvneta.c
3096
rx->dma = ndma;
sys/dev/neta/if_mvneta.c
3116
tcp_lro_flush_all(&rx->lro);
sys/dev/neta/if_mvneta.c
3131
struct mvneta_rx_ring *rx;
sys/dev/neta/if_mvneta.c
3141
rx = MVNETA_RX_RING(sc, q);
sys/dev/neta/if_mvneta.c
3153
rxbuf = &rx->rxbuf[rx->cpu];
sys/dev/neta/if_mvneta.c
3171
r = &rx->desc[rx->cpu];
sys/dev/neta/if_mvneta.c
3173
rx->rxbuf_virt_addr[rx->cpu] = m->m_data;
sys/dev/neta/if_mvneta.c
3175
rx->cpu = rx_counter_adv(rx->cpu, 1);
sys/dev/neta/if_mvneta.c
3179
rx->needs_refill = TRUE;
sys/dev/neta/if_mvneta.c
3183
rx->needs_refill = FALSE;
sys/dev/neta/if_mvneta.c
3184
bus_dmamap_sync(sc->rx_dtag, rx->desc_map, BUS_DMASYNC_PREREAD|BUS_DMASYNC_PREWRITE);
sys/dev/neta/if_mvneta.c
3359
struct mvneta_rx_ring *rx;
sys/dev/neta/if_mvneta.c
3364
rx = NULL;
sys/dev/neta/if_mvneta.c
3380
rx = MVNETA_RX_RING(sc, arg->queue);
sys/dev/neta/if_mvneta.c
3381
time_mvtclk = rx->queue_th_time;
sys/dev/neta/if_mvneta.c
3400
rx->queue_th_time = time_mvtclk;
sys/dev/neta/if_mvneta.c
3401
reg = MVNETA_PRXITTH_RITT(rx->queue_th_time);
sys/dev/netmap/netmap_freebsd.c
455
nm_os_generic_find_num_desc(if_t ifp, unsigned int *tx, unsigned int *rx)
sys/dev/netmap/netmap_kern.h
2173
int nm_os_generic_find_num_desc(if_t ifp, u_int *tx, u_int *rx);
sys/dev/oce/oce_if.h
494
struct oce_rx_stats rx;
sys/dev/oce/oce_sysctl.c
1112
bzero(&adapter_stats->rx, sizeof(struct oce_rx_stats));
sys/dev/oce/oce_sysctl.c
1115
adapter_stats->rx.t_rx_pkts += sc->rq[i]->rx_stats.rx_pkts;
sys/dev/oce/oce_sysctl.c
1116
adapter_stats->rx.t_rx_bytes += sc->rq[i]->rx_stats.rx_bytes;
sys/dev/oce/oce_sysctl.c
1117
adapter_stats->rx.t_rx_frags += sc->rq[i]->rx_stats.rx_frags;
sys/dev/oce/oce_sysctl.c
1118
adapter_stats->rx.t_rx_mcast_pkts +=
sys/dev/oce/oce_sysctl.c
1120
adapter_stats->rx.t_rx_ucast_pkts +=
sys/dev/oce/oce_sysctl.c
1122
adapter_stats->rx.t_rxcp_errs += sc-> rq[i]->rx_stats.rxcp_err;
sys/dev/oce/oce_sysctl.c
712
CTLFLAG_RD, &stats->rx.t_rx_pkts,
sys/dev/oce/oce_sysctl.c
715
CTLFLAG_RD, &stats->rx.t_rx_bytes,
sys/dev/oce/oce_sysctl.c
718
CTLFLAG_RD, &stats->rx.t_rx_frags, 0,
sys/dev/oce/oce_sysctl.c
721
CTLFLAG_RD, &stats->rx.t_rx_mcast_pkts, 0,
sys/dev/oce/oce_sysctl.c
724
CTLFLAG_RD, &stats->rx.t_rx_ucast_pkts, 0,
sys/dev/oce/oce_sysctl.c
727
CTLFLAG_RD, &stats->rx.t_rxcp_errs, 0,
sys/dev/oce/oce_sysctl.c
928
CTLFLAG_RD, &stats->rx.t_rx_pkts,
sys/dev/oce/oce_sysctl.c
931
CTLFLAG_RD, &stats->rx.t_rx_bytes,
sys/dev/oce/oce_sysctl.c
934
CTLFLAG_RD, &stats->rx.t_rx_frags, 0,
sys/dev/oce/oce_sysctl.c
937
CTLFLAG_RD, &stats->rx.t_rx_mcast_pkts, 0,
sys/dev/oce/oce_sysctl.c
940
CTLFLAG_RD, &stats->rx.t_rx_ucast_pkts, 0,
sys/dev/oce/oce_sysctl.c
943
CTLFLAG_RD, &stats->rx.t_rxcp_errs, 0,
sys/dev/qat/include/common/adf_accel_devices.h
577
unsigned int rx;
sys/dev/qat/qat_common/adf_freebsd_pfvf_ctrs_dbg.c
29
value = pfvf_counters->rx;
sys/dev/qlnx/qlnxe/ecore_hsi_eth.h
1220
u8 rx /* If set, apply these commands to the RX path */;
sys/dev/qlnx/qlnxe/ecore_l2.c
1406
p_ramrod->filter_cmd_hdr.rx = p_filter_cmd->is_rx_filter ? 1 : 0;
sys/dev/qlnx/qlnxe/ecore_l2.c
87
u8 rx = 0, tx = 0;
sys/dev/qlnx/qlnxe/ecore_l2.c
89
ecore_vf_get_num_rxqs(p_hwfn, &rx);
sys/dev/qlnx/qlnxe/ecore_l2.c
92
p_l2_info->queues = (u32)OSAL_MAX_T(u8, rx, tx);
sys/dev/ral/rt2560.c
2537
uint8_t rx;
sys/dev/ral/rt2560.c
2539
rx = rt2560_bbp_read(sc, RT2560_BBP_RX) & ~RT2560_BBP_ANTMASK;
sys/dev/ral/rt2560.c
2541
rx |= RT2560_BBP_ANTA;
sys/dev/ral/rt2560.c
2543
rx |= RT2560_BBP_ANTB;
sys/dev/ral/rt2560.c
2545
rx |= RT2560_BBP_DIVERSITY;
sys/dev/ral/rt2560.c
2549
rx &= ~RT2560_BBP_FLIPIQ;
sys/dev/ral/rt2560.c
2551
rt2560_bbp_write(sc, RT2560_BBP_RX, rx);
sys/dev/rge/if_rge.c
1782
struct rge_rxq *rx = &q->q_rx.rge_rxq[i];
sys/dev/rge/if_rge.c
1785
if (rx->rxq_mbuf != NULL) {
sys/dev/rge/if_rge.c
1786
if (rx->rxq_dmamap != NULL) {
sys/dev/rge/if_rge.c
1788
rx->rxq_dmamap, BUS_DMASYNC_POSTREAD);
sys/dev/rge/if_rge.c
1790
rx->rxq_dmamap);
sys/dev/rge/if_rge.c
1792
m_free(rx->rxq_mbuf);
sys/dev/rge/if_rge.c
1793
rx->rxq_mbuf = NULL;
sys/dev/rge/if_rge.c
1797
if (rx->rxq_dmamap != NULL) {
sys/dev/rge/if_rge.c
1798
bus_dmamap_destroy(sc->sc_dmat_rx_buf, rx->rxq_dmamap);
sys/dev/rge/if_rge.c
1799
rx->rxq_dmamap = NULL;
sys/dev/rge/if_rge.c
2081
int i, mlen, rx = 0;
sys/dev/rge/if_rge.c
2134
rx = 1;
sys/dev/rge/if_rge.c
2266
if (!rx)
sys/dev/rtwn/rtl8188e/r88e_calib.c
146
rx[0] = MS(rtwn_bb_read(sc, R92C_RX_POWER_IQK_BEFORE(0)),
sys/dev/rtwn/rtl8188e/r88e_calib.c
148
rx[1] = MS(status, R92C_POWER_IQK_RESULT);
sys/dev/rtwn/rtl8188e/r88e_calib.c
149
if (rx[0] == 0x132 || rx[1] == 0x036)
sys/dev/rtwn/rtl8188e/r88e_calib.c
157
uint16_t rx[2], struct r88e_iq_cal_reg_vals *vals)
sys/dev/rtwn/rtl8188e/r88e_calib.c
235
ret = r88e_iq_calib_chain(sc, tx, rx);
sys/dev/rtwn/rtl8188e/r88e_calib.c
241
rx[0] = 0xff;
sys/dev/rtwn/rtl8188e/r88e_calib.c
242
rx[1] = 0xff;
sys/dev/rtwn/rtl8188e/r88e_calib.c
246
rx[0] = 0xff;
sys/dev/rtwn/rtl8188e/r88e_calib.c
247
rx[1] = 0xff;
sys/dev/rtwn/rtl8188e/r88e_calib.c
256
"rx[1] 0x%x\n", __func__, n, tx[0], tx[1], rx[0], rx[1]);
sys/dev/rtwn/rtl8188e/r88e_calib.c
310
uint16_t rx[2])
sys/dev/rtwn/rtl8188e/r88e_calib.c
339
if (rx[0] == 0xff || rx[1] == 0xff)
sys/dev/rtwn/rtl8188e/r88e_calib.c
343
rx[0] & 0x3ff);
sys/dev/rtwn/rtl8188e/r88e_calib.c
345
(rx[1] & 0x3f) << 10);
sys/dev/rtwn/rtl8188e/r88e_calib.c
347
(rx[1] & 0x3c0) << 22);
sys/dev/rtwn/rtl8188e/r88e_calib.c
355
uint16_t tx[RTWN_IQ_CAL_NRUN][2], rx[RTWN_IQ_CAL_NRUN][2];
sys/dev/rtwn/rtl8188e/r88e_calib.c
363
r88e_iq_calib_run(sc, n, tx[n], rx[n], &vals);
sys/dev/rtwn/rtl8188e/r88e_calib.c
370
rx[n - 1], tx[n], rx[n]);
sys/dev/rtwn/rtl8188e/r88e_calib.c
376
r88e_iq_calib_write_results(sc, tx[n], rx[n]);
sys/dev/rtwn/rtl8188e/r88e_calib.c
66
r88e_iq_calib_chain(struct rtwn_softc *sc, uint16_t tx[2], uint16_t rx[2])
sys/dev/rtwn/rtl8192c/pci/r92ce_calib.c
123
rx[0] = MS(rtwn_bb_read(sc, R92C_RX_POWER_IQK_BEFORE(chain)),
sys/dev/rtwn/rtl8192c/pci/r92ce_calib.c
125
rx[1] = MS(rtwn_bb_read(sc, R92C_RX_POWER_IQK_AFTER(chain)),
sys/dev/rtwn/rtl8192c/pci/r92ce_calib.c
127
if (rx[0] == 0x132 || rx[1] == 0x036)
sys/dev/rtwn/rtl8192c/pci/r92ce_calib.c
135
uint16_t rx[2][2], struct r92ce_iq_cal_reg_vals *vals)
sys/dev/rtwn/rtl8192c/pci/r92ce_calib.c
226
tx[chain], rx[chain]);
sys/dev/rtwn/rtl8192c/pci/r92ce_calib.c
233
rx[chain][0] = 0xff;
sys/dev/rtwn/rtl8192c/pci/r92ce_calib.c
234
rx[chain][1] = 0xff;
sys/dev/rtwn/rtl8192c/pci/r92ce_calib.c
239
rx[chain][0] = 0xff;
sys/dev/rtwn/rtl8192c/pci/r92ce_calib.c
240
rx[chain][1] = 0xff;
sys/dev/rtwn/rtl8192c/pci/r92ce_calib.c
251
tx[chain][0], tx[chain][1], rx[chain][0], rx[chain][1]);
sys/dev/rtwn/rtl8192c/pci/r92ce_calib.c
312
uint16_t rx[2], int chain)
sys/dev/rtwn/rtl8192c/pci/r92ce_calib.c
341
if (rx[0] == 0xff || rx[1] == 0xff)
sys/dev/rtwn/rtl8192c/pci/r92ce_calib.c
345
rx[0] & 0x3ff);
sys/dev/rtwn/rtl8192c/pci/r92ce_calib.c
347
(rx[1] & 0x3f) << 10);
sys/dev/rtwn/rtl8192c/pci/r92ce_calib.c
351
(rx[1] & 0x3c0) << 22);
sys/dev/rtwn/rtl8192c/pci/r92ce_calib.c
354
(rx[1] & 0x3c0) << 6);
sys/dev/rtwn/rtl8192c/pci/r92ce_calib.c
363
uint16_t tx[RTWN_IQ_CAL_NRUN][2][2], rx[RTWN_IQ_CAL_NRUN][2][2];
sys/dev/rtwn/rtl8192c/pci/r92ce_calib.c
368
r92ce_iq_calib_run(sc, n, tx[n], rx[n], &vals);
sys/dev/rtwn/rtl8192c/pci/r92ce_calib.c
375
rx[n - 1], tx[n], rx[n]);
sys/dev/rtwn/rtl8192c/pci/r92ce_calib.c
381
r92ce_iq_calib_write_results(sc, tx[n][0], rx[n][0], 0);
sys/dev/rtwn/rtl8192c/pci/r92ce_calib.c
383
r92ce_iq_calib_write_results(sc, tx[n][1], rx[n][1], 1);
sys/dev/rtwn/rtl8192c/pci/r92ce_calib.c
72
uint16_t rx[2])
sys/dev/rtwn/rtl8192c/r92c_calib.c
123
rx[0] = MS(rtwn_bb_read(sc, R92C_RX_POWER_IQK_BEFORE(chain)),
sys/dev/rtwn/rtl8192c/r92c_calib.c
125
rx[1] = MS(rtwn_bb_read(sc, R92C_RX_POWER_IQK_AFTER(chain)),
sys/dev/rtwn/rtl8192c/r92c_calib.c
127
if (rx[0] == 0x132 || rx[1] == 0x036)
sys/dev/rtwn/rtl8192c/r92c_calib.c
135
uint16_t rx[2][2], struct r92c_iq_cal_reg_vals *vals)
sys/dev/rtwn/rtl8192c/r92c_calib.c
238
tx[chain], rx[chain]);
sys/dev/rtwn/rtl8192c/r92c_calib.c
245
rx[chain][0] = 0xff;
sys/dev/rtwn/rtl8192c/r92c_calib.c
246
rx[chain][1] = 0xff;
sys/dev/rtwn/rtl8192c/r92c_calib.c
251
rx[chain][0] = 0xff;
sys/dev/rtwn/rtl8192c/r92c_calib.c
252
rx[chain][1] = 0xff;
sys/dev/rtwn/rtl8192c/r92c_calib.c
263
tx[chain][0], tx[chain][1], rx[chain][0], rx[chain][1]);
sys/dev/rtwn/rtl8192c/r92c_calib.c
331
uint16_t rx[2], int chain)
sys/dev/rtwn/rtl8192c/r92c_calib.c
360
if (rx[0] == 0xff || rx[1] == 0xff)
sys/dev/rtwn/rtl8192c/r92c_calib.c
364
rx[0] & 0x3ff);
sys/dev/rtwn/rtl8192c/r92c_calib.c
366
(rx[1] & 0x3f) << 10);
sys/dev/rtwn/rtl8192c/r92c_calib.c
370
(rx[1] & 0x3c0) << 22);
sys/dev/rtwn/rtl8192c/r92c_calib.c
373
(rx[1] & 0x3c0) << 6);
sys/dev/rtwn/rtl8192c/r92c_calib.c
382
uint16_t tx[RTWN_IQ_CAL_NRUN][2][2], rx[RTWN_IQ_CAL_NRUN][2][2];
sys/dev/rtwn/rtl8192c/r92c_calib.c
387
r92c_iq_calib_run(sc, n, tx[n], rx[n], &vals);
sys/dev/rtwn/rtl8192c/r92c_calib.c
394
rx[n - 1], tx[n], rx[n]);
sys/dev/rtwn/rtl8192c/r92c_calib.c
400
r92c_iq_calib_write_results(sc, tx[n][0], rx[n][0], 0);
sys/dev/rtwn/rtl8192c/r92c_calib.c
402
r92c_iq_calib_write_results(sc, tx[n][1], rx[n][1], 1);
sys/dev/rtwn/rtl8192c/r92c_calib.c
72
uint16_t rx[2])
sys/dev/thunderbolt/nhi.c
1012
if (rx != NULL) {
sys/dev/thunderbolt/nhi.c
1013
while (rx->cb != NULL) {
sys/dev/thunderbolt/nhi.c
1014
if ((rx->pdf < 0) || (rx->pdf > 15))
sys/dev/thunderbolt/nhi.c
1016
slot = &trkr->rxpdf[rx->pdf];
sys/dev/thunderbolt/nhi.c
1019
rx++;
sys/dev/thunderbolt/nhi.c
1118
rxd->rx.flags = RX_BUFFER_DESC_RS | RX_BUFFER_DESC_IE;
sys/dev/thunderbolt/nhi.c
669
desc = &rp->rx_ring[ci].rx;
sys/dev/thunderbolt/nhi.c
935
struct nhi_dispatch *rx)
sys/dev/thunderbolt/nhi.c
969
if (rx != NULL) {
sys/dev/thunderbolt/nhi.c
970
while (rx->cb != NULL) {
sys/dev/thunderbolt/nhi.c
971
if ((rx->pdf < 0) || (rx->pdf > 15))
sys/dev/thunderbolt/nhi.c
973
slot = &trkr->rxpdf[rx->pdf];
sys/dev/thunderbolt/nhi.c
979
slot->cb = rx->cb;
sys/dev/thunderbolt/nhi.c
980
slot->context = rx->context;
sys/dev/thunderbolt/nhi.c
982
"Registered RX callback for PDF %d\n", rx->pdf);
sys/dev/thunderbolt/nhi.c
983
rx++;
sys/dev/thunderbolt/nhi.c
992
struct nhi_dispatch *rx)
sys/dev/thunderbolt/nhi_reg.h
314
struct nhi_rx_buffer_desc rx;
sys/dev/thunderbolt/router.c
195
struct nhi_dispatch rx[] = { { PDF_READ, router_response_intr, sc },
sys/dev/thunderbolt/router.c
200
return (nhi_register_pdf(sc->ring0, tx, rx));
sys/dev/tpm/tpm_spibus.c
101
uint8_t rx[8] = {0};
sys/dev/tpm/tpm_spibus.c
114
spic.rx_cmd = rx;
sys/dev/tpm/tpm_spibus.c
65
uint8_t rx[4] = {0};
sys/dev/tpm/tpm_spibus.c
68
if (size > sizeof(rx))
sys/dev/tpm/tpm_spibus.c
78
spic.rx_cmd = rx;
sys/dev/tpm/tpm_spibus.c
84
if (!(rx[3] & 0x1)) {
sys/dev/tpm/tpm_spibus.c
91
memcpy(buf, &rx[0], size);
sys/dev/uart/uart_dev_mu.c
431
int rx;
sys/dev/uart/uart_dev_mu.c
439
rx = xc & 0xff;
sys/dev/uart/uart_dev_mu.c
444
uart_rx_put(sc, rx);
sys/dev/uart/uart_dev_mvebu.c
499
int rx, er;
sys/dev/uart/uart_dev_mvebu.c
511
rx = xc & 0xff;
sys/dev/uart/uart_dev_mvebu.c
522
uart_rx_put(sc, rx | er);
sys/dev/uart/uart_dev_pl011.c
582
int rx;
sys/dev/uart/uart_dev_pl011.c
597
rx = xc & 0xff;
sys/dev/uart/uart_dev_pl011.c
600
rx |= UART_STAT_FRAMERR;
sys/dev/uart/uart_dev_pl011.c
602
rx |= UART_STAT_PARERR;
sys/dev/uart/uart_dev_pl011.c
604
uart_rx_put(sc, rx);
sys/dev/usb/wlan/if_ural.c
1996
uint8_t rx;
sys/dev/usb/wlan/if_ural.c
1998
rx = ural_bbp_read(sc, RAL_BBP_RX) & ~RAL_BBP_ANTMASK;
sys/dev/usb/wlan/if_ural.c
2000
rx |= RAL_BBP_ANTA;
sys/dev/usb/wlan/if_ural.c
2002
rx |= RAL_BBP_ANTB;
sys/dev/usb/wlan/if_ural.c
2004
rx |= RAL_BBP_DIVERSITY;
sys/dev/usb/wlan/if_ural.c
2008
rx &= ~RAL_BBP_FLIPIQ;
sys/dev/usb/wlan/if_ural.c
2010
ural_bbp_write(sc, RAL_BBP_RX, rx);
sys/dev/usb/wlan/if_urtw.c
3964
struct urtw_8187b_rxhdr *rx;
sys/dev/usb/wlan/if_urtw.c
3966
if (actlen < sizeof(*rx) + IEEE80211_ACK_LEN)
sys/dev/usb/wlan/if_urtw.c
3969
rx = (struct urtw_8187b_rxhdr *)(data->buf +
sys/dev/usb/wlan/if_urtw.c
3971
flen = le32toh(rx->flag) & 0xfff;
sys/dev/usb/wlan/if_urtw.c
3972
if (flen > actlen - sizeof(*rx))
sys/dev/usb/wlan/if_urtw.c
3975
rate = (le32toh(rx->flag) >> URTW_RX_FLAG_RXRATE_SHIFT) & 0xf;
sys/dev/usb/wlan/if_urtw.c
3977
rssi = rx->rssi & URTW_RX_RSSI_MASK;
sys/dev/usb/wlan/if_urtw.c
3978
noise = rx->noise;
sys/dev/usb/wlan/if_urtw.c
3981
mactime = rx->mactime;
sys/dev/usb/wlan/if_urtw.c
3983
struct urtw_8187l_rxhdr *rx;
sys/dev/usb/wlan/if_urtw.c
3985
if (actlen < sizeof(*rx) + IEEE80211_ACK_LEN)
sys/dev/usb/wlan/if_urtw.c
3988
rx = (struct urtw_8187l_rxhdr *)(data->buf +
sys/dev/usb/wlan/if_urtw.c
3990
flen = le32toh(rx->flag) & 0xfff;
sys/dev/usb/wlan/if_urtw.c
3991
if (flen > actlen - sizeof(*rx))
sys/dev/usb/wlan/if_urtw.c
3994
rate = (le32toh(rx->flag) >> URTW_RX_FLAG_RXRATE_SHIFT) & 0xf;
sys/dev/usb/wlan/if_urtw.c
3996
rssi = rx->rssi & URTW_RX_8187L_RSSI_MASK;
sys/dev/usb/wlan/if_urtw.c
3997
noise = rx->noise;
sys/dev/usb/wlan/if_urtw.c
4000
mactime = rx->mactime;
sys/dev/vnic/nic.h
456
uint8_t rx;
sys/dev/vnic/nic_main.c
534
mbx.bgx_stats.rx = bgx->rx;
sys/dev/vnic/nic_main.c
536
if (bgx->rx != 0) {
sys/dev/vnic/nicvf_main.c
891
if (bgx->rx)
sys/dev/wpi/if_wpireg.h
797
struct wpi_rx_stats rx;
sys/dev/xen/netfront/netfront.c
1245
struct netif_rx_response *rx = &rinfo.rx;
sys/dev/xen/netfront/netfront.c
1269
memcpy(rx, RING_GET_RESPONSE(&rxq->ring, i), sizeof(*rx));
sys/dev/xen/netfront/netfront.c
1283
if (rx->flags & NETRXF_data_validated) {
sys/dev/xen/netfront/netfront.c
1297
if ((rx->flags & NETRXF_extra_info) != 0 &&
sys/dev/xen/netfront/netfront.c
1490
struct netif_rx_response *rx = &rinfo->rx;
sys/dev/xen/netfront/netfront.c
1500
if (rx->flags & NETRXF_extra_info) {
sys/dev/xen/netfront/netfront.c
1512
rx->status, rx->offset, frags);
sys/dev/xen/netfront/netfront.c
1514
if (__predict_false(rx->status < 0 ||
sys/dev/xen/netfront/netfront.c
1515
rx->offset + rx->status > PAGE_SIZE)) {
sys/dev/xen/netfront/netfront.c
1530
printf("%s: Bad rx response id %d.\n", __func__, rx->id);
sys/dev/xen/netfront/netfront.c
1544
m->m_len = rx->status;
sys/dev/xen/netfront/netfront.c
1545
m->m_data += rx->offset;
sys/dev/xen/netfront/netfront.c
1546
m0->m_pkthdr.len += rx->status;
sys/dev/xen/netfront/netfront.c
1549
if (!(rx->flags & NETRXF_more_data))
sys/dev/xen/netfront/netfront.c
1566
rx = RING_GET_RESPONSE(&rxq->ring, *cons + frags);
sys/dev/xen/netfront/netfront.c
240
struct netif_rx_response rx;
sys/net80211/ieee80211_freebsd.c
632
struct ieee80211_rx_params *rx;
sys/net80211/ieee80211_freebsd.c
639
rx = (struct ieee80211_rx_params *)(mtag + 1);
sys/net80211/ieee80211_freebsd.c
640
memcpy(&rx->params, rxs, sizeof(*rxs));
sys/net80211/ieee80211_freebsd.c
649
struct ieee80211_rx_params *rx;
sys/net80211/ieee80211_freebsd.c
655
rx = (struct ieee80211_rx_params *)(mtag + 1);
sys/net80211/ieee80211_freebsd.c
656
memcpy(rxs, &rx->params, sizeof(*rxs));
sys/net80211/ieee80211_freebsd.c
664
struct ieee80211_rx_params *rx;
sys/net80211/ieee80211_freebsd.c
670
rx = (struct ieee80211_rx_params *)(mtag + 1);
sys/net80211/ieee80211_freebsd.c
671
return (&rx->params);
sys/net80211/ieee80211_input.c
106
struct ieee80211_rx_stats rx;
sys/net80211/ieee80211_input.c
108
rx.r_flags = IEEE80211_R_NF | IEEE80211_R_RSSI;
sys/net80211/ieee80211_input.c
109
rx.c_nf = nf;
sys/net80211/ieee80211_input.c
110
rx.c_rssi = rssi;
sys/net80211/ieee80211_input.c
112
if (!ieee80211_add_rx_params(m, &rx))
sys/net80211/ieee80211_input.c
63
ieee80211_process_mimo(struct ieee80211_node *ni, struct ieee80211_rx_stats *rx)
sys/net80211/ieee80211_input.c
68
if ((rx->r_flags & (IEEE80211_R_C_CHAIN | IEEE80211_R_C_NF | IEEE80211_R_C_RSSI)) !=
sys/net80211/ieee80211_input.c
73
for (i = 0; i < MIN(rx->c_chain, IEEE80211_MAX_CHAINS); i++) {
sys/net80211/ieee80211_input.c
74
IEEE80211_RSSI_LPF(ni->ni_mimo_rssi_ctl[i], rx->c_rssi_ctl[i]);
sys/net80211/ieee80211_input.c
75
IEEE80211_RSSI_LPF(ni->ni_mimo_rssi_ext[i], rx->c_rssi_ext[i]);
sys/net80211/ieee80211_input.c
79
for(i = 0; i < MIN(rx->c_chain, IEEE80211_MAX_CHAINS); i++) {
sys/net80211/ieee80211_input.c
80
ni->ni_mimo_noise_ctl[i] = rx->c_nf_ctl[i];
sys/net80211/ieee80211_input.c
81
ni->ni_mimo_noise_ext[i] = rx->c_nf_ext[i];
sys/net80211/ieee80211_input.c
83
ni->ni_mimo_chains = rx->c_chain;
sys/net80211/ieee80211_vht.c
191
int i, tx, rx;
sys/net80211/ieee80211_vht.c
210
rx = (ic->ic_vht_cap.supp_mcs.rx_mcs_map >> (2*i)) & 0x3;
sys/net80211/ieee80211_vht.c
211
if (tx == 3 && rx == 3)
sys/net80211/ieee80211_vht.c
214
i + 1, vht_mcs_to_num(tx), vht_mcs_to_num(rx));
sys/netpfil/ipfilter/netinet/ip_rpcb_pxy.c
1000
rpcb_xact_t *rx;
sys/netpfil/ipfilter/netinet/ip_rpcb_pxy.c
1005
for (rx = rs->rs_rxlist; rx != NULL; rx = rx->rx_next)
sys/netpfil/ipfilter/netinet/ip_rpcb_pxy.c
1006
if (rx->rx_xid == xid)
sys/netpfil/ipfilter/netinet/ip_rpcb_pxy.c
1009
return (rx);
sys/netpfil/ipfilter/netinet/ip_rpcb_pxy.c
1024
ipf_p_rpcb_deref(rpcb_session_t *rs __unused, rpcb_xact_t *rx)
sys/netpfil/ipfilter/netinet/ip_rpcb_pxy.c
1026
if (rx == NULL)
sys/netpfil/ipfilter/netinet/ip_rpcb_pxy.c
1029
if (--rx->rx_ref != 0)
sys/netpfil/ipfilter/netinet/ip_rpcb_pxy.c
1032
if (rx->rx_next != NULL)
sys/netpfil/ipfilter/netinet/ip_rpcb_pxy.c
1033
rx->rx_next->rx_pnext = rx->rx_pnext;
sys/netpfil/ipfilter/netinet/ip_rpcb_pxy.c
1035
*rx->rx_pnext = rx->rx_next;
sys/netpfil/ipfilter/netinet/ip_rpcb_pxy.c
1037
KFREE(rx);
sys/netpfil/ipfilter/netinet/ip_rpcb_pxy.c
271
rpcb_xact_t *rx;
sys/netpfil/ipfilter/netinet/ip_rpcb_pxy.c
282
rx = NULL;
sys/netpfil/ipfilter/netinet/ip_rpcb_pxy.c
300
rx = NULL; /* XXX gcc */
sys/netpfil/ipfilter/netinet/ip_rpcb_pxy.c
303
rv = ipf_p_rpcb_decoderep(fin, nat, rs, rm, &rx);
sys/netpfil/ipfilter/netinet/ip_rpcb_pxy.c
308
if (rx != NULL) {
sys/netpfil/ipfilter/netinet/ip_rpcb_pxy.c
310
ipf_p_rpcb_deref(rs, rx);
sys/netpfil/ipfilter/netinet/ip_rpcb_pxy.c
327
if (rx->rx_type == RPCB_RES_STRING)
sys/netpfil/ipfilter/netinet/ip_rpcb_pxy.c
329
else if (rx->rx_type == RPCB_RES_LIST)
sys/netpfil/ipfilter/netinet/ip_rpcb_pxy.c
338
if (rx != NULL) {
sys/netpfil/ipfilter/netinet/ip_rpcb_pxy.c
346
ipf_p_rpcb_deref(rs, rx);
sys/netpfil/ipfilter/netinet/ip_rpcb_pxy.c
347
ipf_p_rpcb_deref(rs, rx);
sys/netpfil/ipfilter/netinet/ip_rpcb_pxy.c
405
rpcb_xact_t rx;
sys/netpfil/ipfilter/netinet/ip_rpcb_pxy.c
411
bzero((char *)&rx, sizeof(rx));
sys/netpfil/ipfilter/netinet/ip_rpcb_pxy.c
415
rx.rx_xid = B(p++); /* Record this message's XID. */
sys/netpfil/ipfilter/netinet/ip_rpcb_pxy.c
452
rx.rx_type = RPCB_RES_PMAP;
sys/netpfil/ipfilter/netinet/ip_rpcb_pxy.c
453
rx.rx_proto = xdr;
sys/netpfil/ipfilter/netinet/ip_rpcb_pxy.c
461
rx.rx_type = RPCB_RES_STRING;
sys/netpfil/ipfilter/netinet/ip_rpcb_pxy.c
462
rx.rx_proto = (u_int)fin->fin_p;
sys/netpfil/ipfilter/netinet/ip_rpcb_pxy.c
467
rx.rx_type = RPCB_RES_LIST;
sys/netpfil/ipfilter/netinet/ip_rpcb_pxy.c
494
if (ipf_p_rpcb_insert(rs, &rx) != 0) {
sys/netpfil/ipfilter/netinet/ip_rpcb_pxy.c
551
ipf_p_rpcb_insert(rpcb_session_t *rs, rpcb_xact_t *rx)
sys/netpfil/ipfilter/netinet/ip_rpcb_pxy.c
555
rxp = ipf_p_rpcb_lookup(rs, rx->rx_xid);
sys/netpfil/ipfilter/netinet/ip_rpcb_pxy.c
568
bcopy((char *)rx, (char *)rxp, sizeof(*rx));
sys/netpfil/ipfilter/netinet/ip_rpcb_pxy.c
824
rpcb_xact_t *rx;
sys/netpfil/ipfilter/netinet/ip_rpcb_pxy.c
831
bzero((char *)&rx, sizeof(rx));
sys/netpfil/ipfilter/netinet/ip_rpcb_pxy.c
839
if ((rx = ipf_p_rpcb_lookup(rs, xdr)) == NULL) {
sys/netpfil/ipfilter/netinet/ip_rpcb_pxy.c
843
++rx->rx_ref; /* per thread reference */
sys/netpfil/ipfilter/netinet/ip_rpcb_pxy.c
846
*rxp = rx;
sys/netpfil/ipfilter/netinet/ip_rpcb_pxy.c
874
switch(rx->rx_type)
sys/netpfil/ipfilter/netinet/ip_rpcb_pxy.c
893
if (ipf_p_rpcb_getnat(fin, nat, rx->rx_proto, (u_int)xdr) != 0)
sys/netpfil/ipfilter/netinet/ip_rpcb_pxy.c
919
if (ipf_p_rpcb_getnat(fin, nat, rx->rx_proto,
sys/netpfil/ipfilter/netinet/ip_rpcb_pxy.c
984
IPF_PANIC(1, ("illegal rx_type %d", rx->rx_type));
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
120
struct ipoib_cm_rx *rx,
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
130
rx_req = (struct ipoib_rx_buf *)&rx->rx_ring[id];
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
139
ret = ib_post_recv(rx->qp, wr, &bad_wr);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
143
m_freem(rx->rx_ring[id].mb);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
144
rx->rx_ring[id].mb = NULL;
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
301
struct ib_cm_id *cm_id, struct ipoib_cm_rx *rx)
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
310
rx->rx_ring = kzalloc(ipoib_recvq_size * sizeof *rx->rx_ring, GFP_KERNEL);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
311
if (!rx->rx_ring) {
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
317
memset(rx->rx_ring, 0, ipoib_recvq_size * sizeof *rx->rx_ring);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
340
if (!ipoib_cm_alloc_rx_mb(priv, &rx->rx_ring[i])) {
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
345
ret = ipoib_cm_post_receive_nonsrq(priv, rx, &t->wr, t->sge, i);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
354
rx->recv_count = ipoib_recvq_size;
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
367
ipoib_cm_free_rx_ring(priv, rx->rx_ring);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
766
struct ipoib_cm_rx *rx, *n;
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
773
list_for_each_entry_safe(rx, n, &list, list) {
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
774
ib_destroy_cm_id(rx->id);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
775
ib_destroy_qp(rx->qp);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
777
ipoib_cm_free_rx_ring(priv, rx->rx_ring);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
782
kfree(rx);
sys/powerpc/pseries/phyp_llan.c
124
static int llan_add_rxbuf(struct llan_softc *sc, struct llan_xfer *rx);
sys/powerpc/pseries/phyp_llan.c
323
llan_add_rxbuf(struct llan_softc *sc, struct llan_xfer *rx)
sys/powerpc/pseries/phyp_llan.c
336
if (rx->rx_mbuf != NULL) {
sys/powerpc/pseries/phyp_llan.c
337
bus_dmamap_sync(sc->rxbuf_dma_tag, rx->rx_dmamap,
sys/powerpc/pseries/phyp_llan.c
339
bus_dmamap_unload(sc->rxbuf_dma_tag, rx->rx_dmamap);
sys/powerpc/pseries/phyp_llan.c
343
m_copyback(m, 0, 8, (void *)&rx);
sys/powerpc/pseries/phyp_llan.c
345
error = bus_dmamap_load_mbuf_sg(sc->rxbuf_dma_tag, rx->rx_dmamap, m,
sys/powerpc/pseries/phyp_llan.c
349
"cannot load RX DMA map %p, error = %d\n", rx, error);
sys/powerpc/pseries/phyp_llan.c
357
rx->rx_mbuf = m;
sys/powerpc/pseries/phyp_llan.c
359
bus_dmamap_sync(sc->rxbuf_dma_tag, rx->rx_dmamap, BUS_DMASYNC_PREREAD);
sys/powerpc/pseries/phyp_llan.c
361
rx->rx_bufdesc = LLAN_BUFDESC_VALID;
sys/powerpc/pseries/phyp_llan.c
362
rx->rx_bufdesc |= (((uint64_t)segs[0].ds_len) << 32);
sys/powerpc/pseries/phyp_llan.c
363
rx->rx_bufdesc |= segs[0].ds_addr;
sys/powerpc/pseries/phyp_llan.c
364
error = phyp_hcall(H_ADD_LOGICAL_LAN_BUFFER, sc->unit, rx->rx_bufdesc);
sys/powerpc/pseries/phyp_llan.c
367
rx->rx_mbuf = NULL;
sys/powerpc/pseries/phyp_llan.c
378
struct llan_xfer *rx;
sys/powerpc/pseries/phyp_llan.c
386
rx = (struct llan_xfer *)sc->rx_buf[sc->rx_dma_slot].handle;
sys/powerpc/pseries/phyp_llan.c
387
m = rx->rx_mbuf;
sys/powerpc/pseries/phyp_llan.c
392
if (llan_add_rxbuf(sc, rx) != 0) {
sys/x86/x86/local_apic.c
2193
uint64_t rx;
sys/x86/x86/local_apic.c
2199
for (rx = 0; delay == -1 || rx < lapic_ipi_wait_mult * delay; rx++) {
sys/x86/x86/local_apic.c
554
uint64_t r, r1, r2, rx;
sys/x86/x86/local_apic.c
676
for (rx = 0; rx < LOOPS; rx++) {
tools/tools/ath/athsurvey/athsurvey.c
105
rx = (float) hs->samples[i].rx_busy * 100.0 /
tools/tools/ath/athsurvey/athsurvey.c
131
if (rx > max_rx)
tools/tools/ath/athsurvey/athsurvey.c
132
max_rx = rx;
tools/tools/ath/athsurvey/athsurvey.c
143
if (rx < min_rx)
tools/tools/ath/athsurvey/athsurvey.c
144
min_rx = rx;
tools/tools/ath/athsurvey/athsurvey.c
77
float tx = 0.0, rx = 0.0, cc = 0.0, cext = 0.0;
tools/tools/iwn/iwnstats/main.c
252
iwn_stats_phy_print(&s->rx.ofdm, "ofdm");
tools/tools/iwn/iwnstats/main.c
253
iwn_stats_phy_print(&s->rx.cck, "cck");
tools/tools/iwn/iwnstats/main.c
254
iwn_stats_ht_phy_print(&s->rx.ht);
tools/tools/iwn/iwnstats/main.c
255
iwn_stats_rx_general_print(&s->rx.general);
tools/tools/net80211/w00t/ap/ap.c
48
int rx;
tools/tools/net80211/w00t/ap/ap.c
657
rc = sniff(p->rx, buf, sizeof(buf));
tools/tools/net80211/w00t/ap/ap.c
833
FD_SET(p->rx, &fds);
tools/tools/net80211/w00t/ap/ap.c
835
max = p->rx > p->tap ? p->rx : p->tap;
tools/tools/net80211/w00t/ap/ap.c
841
if (FD_ISSET(p->rx, &fds))
tools/tools/net80211/w00t/ap/ap.c
901
if ((p.rx = open_rx(iface)) == -1)
tools/tools/net80211/w00t/assoc/assoc.c
430
rc = sniff(p->rx, buf, sizeof(buf));
tools/tools/net80211/w00t/assoc/assoc.c
491
rc = sniff(p->rx, buf, sizeof(buf));
tools/tools/net80211/w00t/assoc/assoc.c
545
rc = sniff(p->rx, buf, sizeof(buf));
tools/tools/net80211/w00t/assoc/assoc.c
58
int rx;
tools/tools/net80211/w00t/assoc/assoc.c
591
rc = sniff(p->rx, buf, sizeof(buf));
tools/tools/net80211/w00t/assoc/assoc.c
772
if (open_rxtx(iface, &p.rx, &p.tx) == -1)
tools/tools/net80211/w00t/assoc/assoc.c
809
FD_SET(p.rx, &fds);
tools/tools/net80211/w00t/assoc/assoc.c
816
rc = select(p.rx+1, &fds, NULL,
tools/tools/net80211/w00t/assoc/assoc.c
893
FD_SET(p.rx, &fds);
tools/tools/net80211/w00t/assoc/assoc.c
895
max = (p.rx > p.tap) ? p.rx : p.tap;
tools/tools/net80211/w00t/assoc/assoc.c
905
if (FD_ISSET(p.rx, &fds)) {
tools/tools/net80211/w00t/expand/expand.c
335
rc = sniff(p->rx, buf, buflen);
tools/tools/net80211/w00t/expand/expand.c
400
FD_SET(p->rx, &fds);
tools/tools/net80211/w00t/expand/expand.c
402
if (select(p->rx+1, &fds, NULL, NULL, to) == -1)
tools/tools/net80211/w00t/expand/expand.c
405
if (FD_ISSET(p->rx, &fds))
tools/tools/net80211/w00t/expand/expand.c
451
if ((p.rx = open_rx(iface)) == -1)
tools/tools/net80211/w00t/expand/expand.c
59
int rx;
tools/tools/net80211/w00t/libw00t/w00t.c
135
int open_rxtx(char *iface, int *rx, int *tx)
tools/tools/net80211/w00t/libw00t/w00t.c
137
*rx = open_bpf(iface, DLT_IEEE802_11_RADIO);
tools/tools/net80211/w00t/libw00t/w00t.c
138
*tx = *rx;
tools/tools/net80211/w00t/libw00t/w00t.c
140
return *rx;
tools/tools/net80211/w00t/libw00t/w00t.h
36
int open_rxtx(char *iface, int *rx, int *tx);
tools/tools/net80211/w00t/prga/prga.c
166
rc = sniff(p->rx, buf, sizeof(buf));
tools/tools/net80211/w00t/prga/prga.c
366
FD_SET(p->rx, &fds);
tools/tools/net80211/w00t/prga/prga.c
367
if (select(p->rx+1, &fds, NULL, NULL, &now) == -1)
tools/tools/net80211/w00t/prga/prga.c
370
if (!FD_ISSET(p->rx, &fds))
tools/tools/net80211/w00t/prga/prga.c
374
rc = sniff(p->rx, buf, sizeof(buf));
tools/tools/net80211/w00t/prga/prga.c
427
FD_SET(p->rx, &fds);
tools/tools/net80211/w00t/prga/prga.c
428
if (select(p->rx+1, &fds, NULL, NULL, &now) == -1)
tools/tools/net80211/w00t/prga/prga.c
431
if (!FD_ISSET(p->rx, &fds))
tools/tools/net80211/w00t/prga/prga.c
435
rc = sniff(p->rx, buf, sizeof(buf));
tools/tools/net80211/w00t/prga/prga.c
53
int rx;
tools/tools/net80211/w00t/prga/prga.c
635
if ((p.rx = open_rx(iface)) == -1)
tools/tools/net80211/w00t/redir/redir.c
395
rc = sniff(p->rx, buf, buflen);
tools/tools/net80211/w00t/redir/redir.c
602
FD_SET(p->rx, &fds);
tools/tools/net80211/w00t/redir/redir.c
604
max = (p->rx > p->s) ? p->rx : p->s;
tools/tools/net80211/w00t/redir/redir.c
609
if (FD_ISSET(p->rx, &fds))
tools/tools/net80211/w00t/redir/redir.c
61
int rx;
tools/tools/net80211/w00t/redir/redir.c
689
if ((p.rx = open_rx(iface)) == -1)
tools/tools/netmap/bridge.c
367
struct netmap_ring *rx = NETMAP_RXRING(pa->nifp, pa->cur_rx_ring);
tools/tools/netmap/bridge.c
369
rx->head, rx->cur, rx->tail);
tools/tools/netmap/bridge.c
372
struct netmap_ring *rx = NETMAP_RXRING(pb->nifp, pb->cur_rx_ring);
tools/tools/netmap/bridge.c
374
rx->head, rx->cur, rx->tail);
tools/tools/netmap/nmreplay.c
1263
(long long)(q0->rx - olda.rx), (long long)(q0->tx - olda.tx),
tools/tools/netmap/nmreplay.c
614
uint64_t rx ALIGN_CACHE; /* rx counter */
tools/tools/netmap/nmreplay.c
904
q->rx++;
tools/tools/netmap/pkt-gen.c
1482
int rx = 0;
tools/tools/netmap/pkt-gen.c
1519
rx++;
tools/tools/netmap/pkt-gen.c
1559
if (rx < m && ts.tv_sec <= 3 && !targ->cancel)
tools/tools/netmap/pkt-gen.c
1894
u_int head, rx, n;
tools/tools/netmap/pkt-gen.c
1905
for (rx = 0; rx < limit; rx++) {
usr.sbin/bhyve/net_backend_netmap.c
143
priv->rx = NETMAP_RXRING(priv->nmd->nifp, 0);
usr.sbin/bhyve/net_backend_netmap.c
255
struct netmap_ring *ring = priv->rx;
usr.sbin/bhyve/net_backend_netmap.c
284
ring = priv->rx;
usr.sbin/bhyve/net_backend_netmap.c
52
struct netmap_ring *rx;
usr.sbin/fwcontrol/fwcontrol.c
678
bufreq.rx.nchunk = 8;
usr.sbin/fwcontrol/fwcontrol.c
679
bufreq.rx.npacket = RECV_NUM_PACKET;
usr.sbin/fwcontrol/fwcontrol.c
680
bufreq.rx.psize = RECV_PACKET_SZ;
usr.sbin/fwcontrol/fwdv.c
116
bufreq.rx.nchunk = NCHUNK;
usr.sbin/fwcontrol/fwdv.c
117
bufreq.rx.npacket = NPACKET_R;
usr.sbin/fwcontrol/fwdv.c
118
bufreq.rx.psize = PSIZE;
usr.sbin/fwcontrol/fwdv.c
274
bufreq.rx.nchunk = 0;
usr.sbin/fwcontrol/fwdv.c
275
bufreq.rx.npacket = 0;
usr.sbin/fwcontrol/fwdv.c
276
bufreq.rx.psize = 0;
usr.sbin/fwcontrol/fwmpegts.c
171
bufreq.rx.nchunk = NCHUNK;
usr.sbin/fwcontrol/fwmpegts.c
172
bufreq.rx.npacket = NPACKET_R;
usr.sbin/fwcontrol/fwmpegts.c
173
bufreq.rx.psize = PSIZE;