Symbol: mb
bin/pax/ar_io.c
1009
mb.mt_op = MTBSF;
bin/pax/ar_io.c
1010
mb.mt_count = 1;
bin/pax/ar_io.c
1011
if (ioctl(arfd, MTIOCTOP, &mb) < 0) {
bin/pax/ar_io.c
1020
mb.mt_op = MTBSR;
bin/pax/ar_io.c
1021
mb.mt_count = 1;
bin/pax/ar_io.c
1022
if (ioctl(arfd, MTIOCTOP, &mb) < 0) {
bin/pax/ar_io.c
1041
mb.mt_op = MTBSF;
bin/pax/ar_io.c
1042
mb.mt_count = 1;
bin/pax/ar_io.c
1043
if (ioctl(arfd, MTIOCTOP, &mb) < 0) {
bin/pax/ar_io.c
1072
mb.mt_op = MTBSR;
bin/pax/ar_io.c
1073
mb.mt_count = padsz/phyblk;
bin/pax/ar_io.c
1074
if (ioctl(arfd, MTIOCTOP, &mb) < 0) {
bin/pax/ar_io.c
1076
mb.mt_count);
bin/pax/ar_io.c
174
artyp = ioctl(arfd, MTIOCGET, &mb) ? ISCHR : ISTAPE;
bin/pax/ar_io.c
708
struct mtop mb;
bin/pax/ar_io.c
740
mb.mt_op = MTFSR;
bin/pax/ar_io.c
741
mb.mt_count = 1;
bin/pax/ar_io.c
742
if (ioctl(arfd, MTIOCTOP, &mb) < 0)
bin/pax/ar_io.c
851
struct mtop mb;
bin/pax/ar_io.c
954
mb.mt_op = MTBSR;
bin/pax/ar_io.c
955
mb.mt_count = sksz/phyblk;
bin/pax/ar_io.c
956
if (ioctl(arfd, MTIOCTOP, &mb) < 0) {
bin/pax/ar_io.c
958
mb.mt_count);
bin/pax/ar_io.c
96
struct mtget mb;
bin/pax/ar_io.c
985
struct mtop mb;
libexec/rtld-elf/powerpc/reloc.c
579
mb(); /* Order jmptab update before next changes */
libexec/rtld-elf/rtld.c
355
#define LD_UTRACE(e, h, mb, ms, r, n) \
libexec/rtld-elf/rtld.c
358
ld_utrace_log(e, h, mb, ms, r, n); \
sbin/camcontrol/camcontrol.c
1316
u_int mb;
sbin/camcontrol/camcontrol.c
1390
mb = speed / 1000;
sbin/camcontrol/camcontrol.c
1391
if (mb > 0) {
sbin/camcontrol/camcontrol.c
1394
mb, speed % 1000);
sbin/ipf/ipftest/ip_fil.c
125
mb_t *mb;
sbin/ipf/ipftest/ip_fil.c
129
mb = (mb_t *)m;
sbin/ipf/ipftest/ip_fil.c
130
ip = MTOD(mb, ip_t *);
sbin/ipf/ipftest/ipftest.c
216
m = &mb;
sbin/ipf/ipftest/ipftest.c
263
printpacket(dir, &mb);
sbin/ipf/ipftest/ipftest.c
278
while ((m != NULL) && (m != &mb)) {
sbin/ipf/ipftest/ipftest.c
291
m = &mb;
sbin/ipf/ipftest/ipftest.c
68
mb_t mb, *m, *n;
sbin/ipf/ipftest/ipftest.c
71
m = &mb;
sbin/ipf/ipsend/resend.c
120
eh = (ether_header_t *)mb.mb_buf;
sbin/ipf/ipsend/resend.c
69
mb_t mb;
sbin/ipf/ipsend/resend.c
82
ip = (struct ip *)mb.mb_buf;
sbin/ipf/ipsend/resend.c
98
while ((i = (*r->r_readip)(&mb, NULL, NULL)) > 0)
sbin/ipf/libipf/ipft_hx.c
109
mb->mb_flags |= M_MCAST;
sbin/ipf/libipf/ipft_hx.c
113
mb->mb_flags |= M_BCAST;
sbin/ipf/libipf/ipft_hx.c
117
mb->mb_flags |= M_MBCAST;
sbin/ipf/libipf/ipft_hx.c
56
hex_readip(mb_t *mb, char **ifn, int *dir)
sbin/ipf/libipf/ipft_hx.c
63
buf = (char *)mb->mb_buf;
sbin/ipf/libipf/ipft_hx.c
76
mb->mb_len = (char *)ip - buf;
sbin/ipf/libipf/ipft_hx.c
77
return (mb->mb_len);
sbin/ipf/libipf/ipft_pc.c
199
ipcap_readip(mb_t *mb, char **ifn, int *dir)
sbin/ipf/libipf/ipft_pc.c
213
buf = (char *)mb->mb_buf;
sbin/ipf/libipf/ipft_pc.c
214
cnt = sizeof(mb->mb_buf);
sbin/ipf/libipf/ipft_pc.c
243
mb->mb_len = n;
sbin/ipf/libipf/ipft_tx.c
112
text_readip(mb_t *mb, char **ifn, int *dir)
sbin/ipf/libipf/ipft_tx.c
119
buf = (char *)mb->mb_buf;
sbin/ipf/libipf/ipft_tx.c
139
mb->mb_len = ntohs(((ip6_t *)ip)->ip6_plen) +
sbin/ipf/libipf/ipft_tx.c
142
mb->mb_len = 0;
sbin/ipf/libipf/ipft_tx.c
145
mb->mb_len = ntohs(ip->ip_len);
sbin/ipf/libipf/ipft_tx.c
147
return (mb->mb_len);
sbin/ipfw/ipfw2.c
1470
int mb = /* mask length */
sbin/ipfw/ipfw2.c
1473
if (mb == 32 && g_co.do_resolv)
sbin/ipfw/ipfw2.c
1478
else if (mb == 0) /* any */
sbin/ipfw/ipfw2.c
1483
if (mb < 0) {
sbin/ipfw/ipfw2.c
1486
} else if (mb < 32)
sbin/ipfw/ipfw2.c
1487
bprintf(bp, "/%d", mb);
sbin/ipfw/ipv6.c
110
mb = (cmd->o.opcode == O_IP6_SRC ||
sbin/ipfw/ipv6.c
114
if (mb == 128 && g_co.do_resolv)
sbin/ipfw/ipv6.c
120
else if (mb == 0) /* any */
sbin/ipfw/ipv6.c
127
if (mb < 0) /* mask not contiguous */
sbin/ipfw/ipv6.c
130
else if (mb < 128)
sbin/ipfw/ipv6.c
131
bprintf(bp, "/%d", mb);
sbin/ipfw/ipv6.c
90
int len, mb;
sys/amd64/amd64/pmap.c
8301
pmap_copy_pages(vm_page_t ma[], vm_offset_t a_offset, vm_page_t mb[],
sys/amd64/amd64/pmap.c
8314
pages[1] = mb[b_offset >> PAGE_SHIFT];
sys/arm/annapurna/alpine/alpine_machdep_mp.c
219
mb();
sys/arm/annapurna/alpine/alpine_machdep_mp.c
227
mb();
sys/arm/annapurna/alpine/alpine_machdep_mp.c
232
mb();
sys/arm/arm/pmap-v6.c
5962
pmap_copy_pages(vm_page_t ma[], vm_offset_t a_offset, vm_page_t mb[],
sys/arm/arm/pmap-v6.c
5985
b_pg = mb[b_offset >> PAGE_SHIFT];
sys/arm/qualcomm/qcom_cpu_kpssv2.c
141
mb();
sys/arm/qualcomm/qcom_cpu_kpssv2.c
150
mb();
sys/arm/qualcomm/qcom_cpu_kpssv2.c
164
mb();
sys/arm/qualcomm/qcom_cpu_kpssv2.c
174
mb();
sys/arm/qualcomm/qcom_cpu_kpssv2.c
182
mb();
sys/arm/qualcomm/qcom_cpu_kpssv2.c
190
mb();
sys/arm/qualcomm/qcom_cpu_kpssv2.c
198
mb();
sys/arm64/arm64/pmap.c
6890
pmap_copy_pages(vm_page_t ma[], vm_offset_t a_offset, vm_page_t mb[],
sys/arm64/arm64/pmap.c
6904
m_b = mb[b_offset >> PAGE_SHIFT];
sys/cam/ata/ata_xpt.c
2123
u_int speed, mb;
sys/cam/ata/ata_xpt.c
2130
mb = speed / 1000;
sys/cam/ata/ata_xpt.c
2131
if (mb > 0)
sys/cam/ata/ata_xpt.c
2134
mb, speed % 1000);
sys/cam/ctl/ctl_ha.c
671
struct mbuf *mb, *newmb;
sys/cam/ctl/ctl_ha.c
694
mb = newmb;
sys/cam/ctl/ctl_ha.c
695
memcpy(mtodo(mb, 0), &hdr, sizeof(hdr));
sys/cam/ctl/ctl_ha.c
696
mb->m_len += sizeof(hdr);
sys/cam/ctl/ctl_ha.c
698
for (; mb != NULL && off < len; mb = mb->m_next) {
sys/cam/ctl/ctl_ha.c
699
copylen = min(M_TRAILINGSPACE(mb), len - off);
sys/cam/ctl/ctl_ha.c
700
memcpy(mtodo(mb, mb->m_len), (const char *)addr + off, copylen);
sys/cam/ctl/ctl_ha.c
701
mb->m_len += copylen;
sys/cam/ctl/ctl_ha.c
709
for (; mb != NULL && off < len2; mb = mb->m_next) {
sys/cam/ctl/ctl_ha.c
710
copylen = min(M_TRAILINGSPACE(mb), len2 - off);
sys/cam/ctl/ctl_ha.c
711
memcpy(mtodo(mb, mb->m_len), (const char *)addr2 + off, copylen);
sys/cam/ctl/ctl_ha.c
712
mb->m_len += copylen;
sys/cam/scsi/scsi_xpt.c
3171
u_int speed, freq, mb;
sys/cam/scsi/scsi_xpt.c
3178
mb = speed / 1000;
sys/cam/scsi/scsi_xpt.c
3179
if (mb > 0)
sys/cam/scsi/scsi_xpt.c
3182
mb, speed % 1000);
sys/compat/linuxkpi/common/include/asm/barrier.h
46
#define smp_mb() mb()
sys/compat/linuxkpi/common/src/linux_netdev.c
273
mb();
sys/compat/linuxkpi/common/src/linux_netdev.c
287
mb();
sys/dev/aac/aac.c
118
static int aac_sa_get_mailbox(struct aac_softc *sc, int mb);
sys/dev/aac/aac.c
140
static int aac_rx_get_mailbox(struct aac_softc *sc, int mb);
sys/dev/aac/aac.c
167
static int aac_rkt_get_mailbox(struct aac_softc *sc, int mb);
sys/dev/aac/aac.c
2528
aac_sa_get_mailbox(struct aac_softc *sc, int mb)
sys/dev/aac/aac.c
2532
return(AAC_MEM1_GETREG4(sc, AAC_SA_MAILBOX + (mb * 4)));
sys/dev/aac/aac.c
2536
aac_rx_get_mailbox(struct aac_softc *sc, int mb)
sys/dev/aac/aac.c
2540
return(AAC_MEM1_GETREG4(sc, AAC_RX_MAILBOX + (mb * 4)));
sys/dev/aac/aac.c
2544
aac_rkt_get_mailbox(struct aac_softc *sc, int mb)
sys/dev/aac/aac.c
2548
return(AAC_MEM1_GETREG4(sc, AAC_RKT_MAILBOX + (mb * 4)));
sys/dev/aac/aacvar.h
244
int (*aif_get_mailbox)(struct aac_softc *sc, int mb);
sys/dev/aac/aacvar.h
263
#define AAC_GET_MAILBOX(sc, mb) ((sc)->aac_if->aif_get_mailbox((sc), \
sys/dev/aac/aacvar.h
264
(mb)))
sys/dev/aacraid/aacraid.c
117
static int aac_src_get_mailbox(struct aac_softc *sc, int mb);
sys/dev/aacraid/aacraid.c
140
static int aac_srcv_get_mailbox(struct aac_softc *sc, int mb);
sys/dev/aacraid/aacraid.c
2288
aac_src_get_mailbox(struct aac_softc *sc, int mb)
sys/dev/aacraid/aacraid.c
2292
return(AAC_MEM0_GETREG4(sc, AAC_SRC_MAILBOX + (mb * 4)));
sys/dev/aacraid/aacraid.c
2296
aac_srcv_get_mailbox(struct aac_softc *sc, int mb)
sys/dev/aacraid/aacraid.c
2300
return(AAC_MEM0_GETREG4(sc, AAC_SRCV_MAILBOX + (mb * 4)));
sys/dev/aacraid/aacraid_var.h
256
int (*aif_get_mailbox)(struct aac_softc *sc, int mb);
sys/dev/aacraid/aacraid_var.h
273
#define AAC_GET_MAILBOX(sc, mb) ((sc)->aac_if.aif_get_mailbox((sc), \
sys/dev/aacraid/aacraid_var.h
274
(mb)))
sys/dev/bce/if_bce.c
6751
mb();
sys/dev/bxe/bxe.c
10271
mb();
sys/dev/bxe/bxe.c
10290
mb();
sys/dev/bxe/bxe.c
10337
mb();
sys/dev/bxe/bxe.c
10354
mb();
sys/dev/bxe/bxe.c
10399
mb();
sys/dev/bxe/bxe.c
10419
mb();
sys/dev/bxe/bxe.c
10472
mb();
sys/dev/bxe/bxe.c
10974
mb();
sys/dev/bxe/bxe.c
10980
mb();
sys/dev/bxe/bxe.c
11026
mb();
sys/dev/bxe/bxe.c
11040
mb();
sys/dev/bxe/bxe.c
11059
mb();
sys/dev/bxe/bxe.c
11063
mb();
sys/dev/bxe/bxe.c
11152
mb();
sys/dev/bxe/bxe.c
12275
mb();
sys/dev/bxe/bxe.c
12412
mb();
sys/dev/bxe/bxe.c
12453
mb();
sys/dev/bxe/bxe.c
12522
mb();
sys/dev/bxe/bxe.c
16529
mb();
sys/dev/bxe/bxe.c
16537
mb();
sys/dev/bxe/bxe.c
2344
mb();
sys/dev/bxe/bxe.c
2643
mb(); /* status block fields can change */
sys/dev/bxe/bxe.c
2660
mb(); /* status block fields can change */
sys/dev/bxe/bxe.c
3508
mb();
sys/dev/bxe/bxe.c
3928
mb();
sys/dev/bxe/bxe.c
3936
mb();
sys/dev/bxe/bxe.c
4254
mb();
sys/dev/bxe/bxe.c
4272
mb();
sys/dev/bxe/bxe.c
5496
mb();
sys/dev/bxe/bxe.c
6857
mb();
sys/dev/bxe/bxe.c
7174
mb();
sys/dev/bxe/bxe.c
7790
mb();
sys/dev/bxe/bxe.c
8330
mb(); /* status block is written to by the chip */
sys/dev/bxe/bxe.c
8342
mb();
sys/dev/bxe/bxe.c
8605
mb();
sys/dev/bxe/bxe.h
2291
mb(); /* status block is written to by the chip */
sys/dev/bxe/bxe.h
2319
mb();
sys/dev/bxe/bxe.h
2346
mb();
sys/dev/bxe/bxe.h
2381
mb();
sys/dev/bxe/bxe.h
2394
mb();
sys/dev/bxe/bxe.h
2401
mb();
sys/dev/bxe/ecore_sp.c
216
mb();
sys/dev/bxe/ecore_sp.c
248
mb();
sys/dev/bxe/ecore_sp.h
106
#define ECORE_SMP_MB_AFTER_CLEAR_BIT() mb()
sys/dev/bxe/ecore_sp.h
107
#define ECORE_SMP_MB_BEFORE_CLEAR_BIT() mb()
sys/dev/bxe/ecore_sp.h
108
#define ECORE_SMP_MB() mb()
sys/dev/cxgb/cxgb_osdep.h
98
#define smp_mb() mb()
sys/dev/cxgb/cxgb_sge.c
2979
mb(); /* commit Tx queue processed updates */
sys/dev/cxgbe/adapter.h
630
struct mbuf *mb[15];
sys/dev/cxgbe/cxgbei/icl_cxgbei.c
240
icl_cxgbei_mbuf_done(struct mbuf *mb)
sys/dev/cxgbe/cxgbei/icl_cxgbei.c
243
struct icl_cxgbei_pdu *icp = (struct icl_cxgbei_pdu *)mb->m_ext.ext_arg1;
sys/dev/cxgbe/t4_sge.c
3205
m_freem(txp->mb[i]);
sys/dev/cxgbe/t4_sge.c
3264
ETHER_BPF_MTAP(ifp, txp->mb[i]);
sys/dev/cxgbe/t4_sge.c
3273
tx_len16_to_desc(mbuf_len16(txp->mb[0])));
sys/dev/cxgbe/t4_sge.c
3276
txp->mb[0]);
sys/dev/cxgbe/t4_sge.c
3278
n = write_txpkt_wr(sc, txq, txp->mb[0],
sys/dev/cxgbe/t4_sge.c
3367
ETHER_BPF_MTAP(ifp, txp->mb[i]);
sys/dev/cxgbe/t4_sge.c
3376
tx_len16_to_desc(mbuf_len16(txp->mb[0])));
sys/dev/cxgbe/t4_sge.c
3378
n = write_txpkt_vm_wr(sc, txq, txp->mb[0]);
sys/dev/cxgbe/t4_sge.c
3380
n = write_txpkt_wr(sc, txq, txp->mb[0], avail);
sys/dev/cxgbe/t4_sge.c
4813
MPASS(nitems(txp->mb) >= sc->params.max_pkts_per_eth_tx_pkts_wr);
sys/dev/cxgbe/t4_sge.c
4814
txq->txp.max_npkt = min(nitems(txp->mb),
sys/dev/cxgbe/t4_sge.c
5937
txp->mb[txp->npkt++] = m;
sys/dev/cxgbe/t4_sge.c
5948
txp->mb[0] = m;
sys/dev/cxgbe/t4_sge.c
5990
txp->mb[0] = m;
sys/dev/cxgbe/t4_sge.c
6017
txp->mb[txp->npkt++] = m;
sys/dev/cxgbe/t4_sge.c
6066
m = txp->mb[i];
sys/dev/cxgbe/t4_sge.c
6136
txsd->m = txp->mb[0];
sys/dev/cxgbe/t4_sge.c
6158
MPASS(txp->mb[0] != NULL);
sys/dev/cxgbe/t4_sge.c
6179
m = txp->mb[i];
sys/dev/cxgbe/t4_sge.c
6217
txsd->m = txp->mb[0];
sys/dev/drm2/drm_os_freebsd.h
170
#define DRM_MEMORYBARRIER() mb()
sys/dev/drm2/drm_os_freebsd.h
173
#define smp_mb__before_atomic_inc() mb()
sys/dev/drm2/drm_os_freebsd.h
174
#define smp_mb__after_atomic_inc() mb()
sys/dev/drm2/ttm/ttm_bo.c
545
mb();
sys/dev/drm2/ttm/ttm_bo_util.c
357
mb();
sys/dev/e1000/if_em.c
2953
mb();
sys/dev/ena/ena.c
3271
mb();
sys/dev/ena/ena_datapath.c
1076
mb();
sys/dev/ena/ena_datapath.c
321
mb();
sys/dev/ena/ena_datapath.c
89
mb();
sys/dev/eqos/if_eqos.c
282
struct mbuf *mb;
sys/dev/eqos/if_eqos.c
286
if (!(mb = m_defrag(m, M_NOWAIT)))
sys/dev/eqos/if_eqos.c
288
m = mb;
sys/dev/etherswitch/felix/felix.c
857
mb();
sys/dev/etherswitch/felix/felix.c
903
mb();
sys/dev/hyperv/pcib/vmbus_pcib.c
1323
mb();
sys/dev/hyperv/pcib/vmbus_pcib.c
1341
mb();
sys/dev/hyperv/pcib/vmbus_pcib.c
1387
mb();
sys/dev/hyperv/vmbus/vmbus.c
677
mb();
sys/dev/hyperv/vmbus/vmbus_br.c
133
mb();
sys/dev/hyperv/vmbus/vmbus_br.c
159
mb();
sys/dev/hyperv/vmbus/vmbus_br.c
203
mb();
sys/dev/hyperv/vmbus/vmbus_br.c
210
mb();
sys/dev/hyperv/vmbus/vmbus_br.c
252
mb();
sys/dev/hyperv/vmbus/vmbus_br.c
279
mb();
sys/dev/hyperv/vmbus/x86/vmbus_x86.c
103
mb();
sys/dev/irdma/osdep.h
101
#define irdma_mb() mb()
sys/dev/iscsi/icl_soft.c
1225
struct mbuf *mb, *newmb;
sys/dev/iscsi/icl_soft.c
1248
for (mb = newmb; mb != NULL; mb = mb->m_next) {
sys/dev/iscsi/icl_soft.c
1249
copylen = min(M_TRAILINGSPACE(mb), len - off);
sys/dev/iscsi/icl_soft.c
1250
memcpy(mtod(mb, char *), (const char *)addr + off, copylen);
sys/dev/iscsi/icl_soft.c
1251
mb->m_len = copylen;
sys/dev/iscsi/icl_soft.c
235
icl_soft_mbuf_done(struct mbuf *mb)
sys/dev/iscsi/icl_soft.c
237
struct icl_soft_pdu *isp = (struct icl_soft_pdu *)mb->m_ext.ext_arg1;
sys/dev/liquidio/base/lio_iq.h
62
struct mbuf *mb;
sys/dev/liquidio/base/lio_request_manager.c
216
m_freem(finfo->mb);
sys/dev/liquidio/lio_main.c
1514
m_freem(finfo->mb);
sys/dev/liquidio/lio_main.c
1536
m_freem(finfo->mb);
sys/dev/liquidio/lio_network.h
160
struct mbuf *mb = NULL;
sys/dev/liquidio/lio_network.h
162
mb = m_getjcl(M_NOWAIT, MT_DATA, M_PKTHDR, size);
sys/dev/liquidio/lio_network.h
163
if (mb != NULL)
sys/dev/liquidio/lio_network.h
164
mb->m_pkthdr.len = mb->m_len = size;
sys/dev/liquidio/lio_network.h
166
return ((void *)mb);
sys/dev/liquidio/lio_rxtx.c
122
finfo->mb = m_head;
sys/dev/mana/mana_en.c
1531
mb();
sys/dev/mge/if_mge.c
1303
struct mbuf *mb;
sys/dev/mge/if_mge.c
1325
mb = m_devget(dw->buffer->m_data,
sys/dev/mge/if_mge.c
1329
if (mb == NULL)
sys/dev/mge/if_mge.c
1333
mb->m_len -= 2;
sys/dev/mge/if_mge.c
1334
mb->m_pkthdr.len -= 2;
sys/dev/mge/if_mge.c
1335
mb->m_data += 2;
sys/dev/mge/if_mge.c
1337
mb->m_pkthdr.rcvif = ifp;
sys/dev/mge/if_mge.c
1339
mge_offload_process_frame(ifp, mb, status,
sys/dev/mge/if_mge.c
1343
if_input(ifp, mb);
sys/dev/mlx4/mlx4_core/mlx4_eq.c
103
mb();
sys/dev/mlx4/mlx4_en/en.h
241
struct mbuf *mb;
sys/dev/mlx4/mlx4_en/en.h
809
u16 mlx4_en_select_queue(if_t dev, struct mbuf *mb);
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
103
mb = mb_head;
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
106
return (mb);
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
115
struct mbuf *mb;
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
124
mb = mlx4_en_alloc_mbuf(ring);
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
125
if (unlikely(mb == NULL))
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
129
m_adj(mb, MLX4_NET_IP_ALIGN);
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
133
mb, ring->spare.segs, &nsegs, BUS_DMA_NOWAIT);
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
135
m_freem(mb);
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
140
ring->spare.mbuf = mb;
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
160
mb = mlx4_en_alloc_mbuf(ring);
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
161
if (unlikely(mb == NULL))
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
165
m_adj(mb, MLX4_NET_IP_ALIGN);
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
168
mb, segs, &nsegs, BUS_DMA_NOWAIT);
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
170
m_freem(mb);
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
188
mb_list->mbuf = mb;
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
586
static void validate_loopback(struct mlx4_en_priv *priv, struct mbuf *mb)
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
592
if (*(mb->m_data + offset) != (unsigned char) (i & 0xff))
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
599
m_freem(mb);
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
630
struct mbuf *mb;
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
634
(mb = m_gethdr(M_NOWAIT, MT_DATA)) != NULL) {
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
637
mb->m_pkthdr.len = mb->m_len = length;
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
640
mb->m_data += MLX4_NET_IP_ALIGN;
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
645
bcopy(mtod(mb_list->mbuf, caddr_t), mtod(mb, caddr_t), length);
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
647
return (mb);
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
651
mb = mb_list->mbuf;
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
658
if (unlikely(length > mb->m_pkthdr.len))
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
659
length = mb->m_pkthdr.len;
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
663
mb->m_len = mb->m_pkthdr.len = length;
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
665
mb->m_pkthdr.len = length;
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
666
for (mb_head = mb; mb != NULL; mb = mb->m_next) {
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
667
if (mb->m_len > length)
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
668
mb->m_len = length;
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
669
length -= mb->m_len;
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
671
if (likely(mb->m_next != NULL)) {
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
673
m_freem(mb->m_next);
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
674
mb->m_next = NULL;
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
680
mb = mb_head;
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
682
return (mb);
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
745
struct mbuf *mb;
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
76
struct mbuf *mb;
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
786
mb = mlx4_en_rx_mb(priv, ring, rx_desc, mb_list, length);
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
787
if (unlikely(!mb)) {
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
79
mb = m_getjcl(M_NOWAIT, MT_DATA, M_PKTHDR, ring->rx_mb_size);
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
796
validate_loopback(priv, mb);
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
80
if (likely(mb != NULL))
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
801
mb->m_pkthdr.flowid = be32_to_cpu(cqe->immed_rss_invalid);
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
802
M_HASHTYPE_SET(mb, mlx4_en_rss_hash(cqe->status, udp_rss));
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
803
mb->m_pkthdr.rcvif = dev;
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
806
mb->m_pkthdr.ether_vtag = be16_to_cpu(cqe->sl_vid);
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
807
mb->m_flags |= M_VLANTAG;
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
81
mb->m_pkthdr.len = mb->m_len = ring->rx_mb_size;
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
814
mb->m_pkthdr.csum_flags =
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
817
mb->m_pkthdr.csum_data = htons(0xffff);
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
828
tcp_lro_rx(&ring->lro, mb, 0) == 0)
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
83
mb = m_getjcl(M_NOWAIT, MT_DATA, M_PKTHDR, MLX4_EN_MAX_RX_BYTES);
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
836
mb->m_pkthdr.csum_flags = 0;
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
84
if (likely(mb != NULL)) {
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
841
if_input(dev, mb);
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
85
struct mbuf *mb_head = mb;
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
88
mb->m_len = MLX4_EN_MAX_RX_BYTES;
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
89
mb->m_pkthdr.len = MLX4_EN_MAX_RX_BYTES;
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
94
mb = (mb->m_next = m_getjcl(M_NOWAIT, MT_DATA, 0, MLX4_EN_MAX_RX_BYTES));
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
95
if (unlikely(mb == NULL)) {
sys/dev/mlx4/mlx4_en/mlx4_en_rx.c
99
mb->m_len = MLX4_EN_MAX_RX_BYTES;
sys/dev/mlx4/mlx4_en/mlx4_en_tx.c
254
struct mbuf *mb, int len, __be32 owner_bit)
sys/dev/mlx4/mlx4_en/mlx4_en_tx.c
259
m_copydata(mb, 0, len, inl + 4);
sys/dev/mlx4/mlx4_en/mlx4_en_tx.c
294
struct mbuf *mb;
sys/dev/mlx4/mlx4_en/mlx4_en_tx.c
297
mb = tx_info->mb;
sys/dev/mlx4/mlx4_en/mlx4_en_tx.c
299
if (mb == NULL)
sys/dev/mlx4/mlx4_en/mlx4_en_tx.c
306
m_freem(mb);
sys/dev/mlx4/mlx4_en/mlx4_en_tx.c
485
mlx4_en_get_inline_hdr_size(struct mlx4_en_tx_ring *ring, struct mbuf *mb)
sys/dev/mlx4/mlx4_en/mlx4_en_tx.c
490
retval = MIN(ring->inline_thold, mb->m_len);
sys/dev/mlx4/mlx4_en/mlx4_en_tx.c
494
retval = MIN(ring->inline_thold, mb->m_pkthdr.len);
sys/dev/mlx4/mlx4_en/mlx4_en_tx.c
499
mlx4_en_get_header_size(struct mbuf *mb)
sys/dev/mlx4/mlx4_en/mlx4_en_tx.c
509
eh = mtod(mb, struct ether_vlan_header *);
sys/dev/mlx4/mlx4_en/mlx4_en_tx.c
510
if (mb->m_len < ETHER_HDR_LEN)
sys/dev/mlx4/mlx4_en/mlx4_en_tx.c
519
if (mb->m_len < eth_hdr_len)
sys/dev/mlx4/mlx4_en/mlx4_en_tx.c
523
ip = (struct ip *)(mb->m_data + eth_hdr_len);
sys/dev/mlx4/mlx4_en/mlx4_en_tx.c
524
if (mb->m_len < eth_hdr_len + sizeof(*ip))
sys/dev/mlx4/mlx4_en/mlx4_en_tx.c
532
ip6 = (struct ip6_hdr *)(mb->m_data + eth_hdr_len);
sys/dev/mlx4/mlx4_en/mlx4_en_tx.c
533
if (mb->m_len < eth_hdr_len + sizeof(*ip6))
sys/dev/mlx4/mlx4_en/mlx4_en_tx.c
542
if (mb->m_len < eth_hdr_len + sizeof(*th))
sys/dev/mlx4/mlx4_en/mlx4_en_tx.c
544
th = (struct tcphdr *)(mb->m_data + eth_hdr_len);
sys/dev/mlx4/mlx4_en/mlx4_en_tx.c
547
if (mb->m_len < eth_hdr_len)
sys/dev/mlx4/mlx4_en/mlx4_en_tx.c
554
struct mbuf *mb, int len, __be32 owner_bit)
sys/dev/mlx4/mlx4_en/mlx4_en_tx.c
560
m_copydata(mb, 0, len, inl + 4);
sys/dev/mlx4/mlx4_en/mlx4_en_tx.c
564
m_copydata(mb, 0, len, inl + 4);
sys/dev/mlx4/mlx4_en/mlx4_en_tx.c
567
m_copydata(mb, 0, spc, inl + 4);
sys/dev/mlx4/mlx4_en/mlx4_en_tx.c
568
m_copydata(mb, spc, len - spc, inl + 8 + spc);
sys/dev/mlx4/mlx4_en/mlx4_en_tx.c
607
u16 mlx4_en_select_queue(if_t dev, struct mbuf *mb)
sys/dev/mlx4/mlx4_en/mlx4_en_tx.c
616
if (mb->m_flags & M_VLANTAG) {
sys/dev/mlx4/mlx4_en/mlx4_en_tx.c
617
u32 vlan_tag = mb->m_pkthdr.ether_vtag;
sys/dev/mlx4/mlx4_en/mlx4_en_tx.c
621
queue_index = m_ether_tcpip_hash(MBUF_HASHFLAG_L3 | MBUF_HASHFLAG_L4, mb, hashrandom);
sys/dev/mlx4/mlx4_en/mlx4_en_tx.c
645
struct mbuf *mb = *mbp;
sys/dev/mlx4/mlx4_en/mlx4_en_tx.c
679
AVG_PERF_COUNTER(priv->pstats.tx_pktsz_avg, mb->m_pkthdr.len);
sys/dev/mlx4/mlx4_en/mlx4_en_tx.c
692
ETHER_BPF_MTAP(ifp, mb);
sys/dev/mlx4/mlx4_en/mlx4_en_tx.c
697
if (mb->m_pkthdr.csum_flags & (CSUM_IP | CSUM_TSO))
sys/dev/mlx4/mlx4_en/mlx4_en_tx.c
700
if (mb->m_pkthdr.csum_flags & (CSUM_TCP | CSUM_UDP |
sys/dev/mlx4/mlx4_en/mlx4_en_tx.c
711
if (mb->m_flags & M_VLANTAG) {
sys/dev/mlx4/mlx4_en/mlx4_en_tx.c
712
tx_desc->ctrl.vlan_tag = cpu_to_be16(mb->m_pkthdr.ether_vtag);
sys/dev/mlx4/mlx4_en/mlx4_en_tx.c
725
m_copydata(mb, 0, 2, __DEVOLATILE(void *, &tx_desc->ctrl.srcrb_flags16[0]));
sys/dev/mlx4/mlx4_en/mlx4_en_tx.c
726
m_copydata(mb, 2, 4, __DEVOLATILE(void *, &tx_desc->ctrl.imm));
sys/dev/mlx4/mlx4_en/mlx4_en_tx.c
733
if (mb->m_pkthdr.csum_flags & CSUM_TSO) {
sys/dev/mlx4/mlx4_en/mlx4_en_tx.c
735
u32 mss = mb->m_pkthdr.tso_segsz;
sys/dev/mlx4/mlx4_en/mlx4_en_tx.c
740
ihs = mlx4_en_get_header_size(mb);
sys/dev/mlx4/mlx4_en/mlx4_en_tx.c
747
payload_len = mb->m_pkthdr.len - ihs;
sys/dev/mlx4/mlx4_en/mlx4_en_tx.c
759
mb, ihs, owner_bit);
sys/dev/mlx4/mlx4_en/mlx4_en_tx.c
763
ihs = mlx4_en_get_inline_hdr_size(ring, mb);
sys/dev/mlx4/mlx4_en/mlx4_en_tx.c
765
mb->m_pkthdr.len, ETHER_MIN_LEN - ETHER_CRC_LEN);
sys/dev/mlx4/mlx4_en/mlx4_en_tx.c
771
mb, ihs, owner_bit);
sys/dev/mlx4/mlx4_en/mlx4_en_tx.c
773
m_adj(mb, ihs);
sys/dev/mlx4/mlx4_en/mlx4_en_tx.c
776
mb, segs, &nr_segs, BUS_DMA_NOWAIT);
sys/dev/mlx4/mlx4_en/mlx4_en_tx.c
780
m = m_defrag(mb, M_NOWAIT);
sys/dev/mlx4/mlx4_en/mlx4_en_tx.c
785
mb = m;
sys/dev/mlx4/mlx4_en/mlx4_en_tx.c
788
mb, segs, &nr_segs, BUS_DMA_NOWAIT);
sys/dev/mlx4/mlx4_en/mlx4_en_tx.c
803
m_freem(mb);
sys/dev/mlx4/mlx4_en/mlx4_en_tx.c
804
mb = NULL;
sys/dev/mlx4/mlx4_en/mlx4_en_tx.c
845
tx_info->mb = mb;
sys/dev/mlx4/mlx4_en/mlx4_en_tx.c
925
m_freem(mb);
sys/dev/mlx4/mlx4_en/mlx4_en_tx.c
930
mlx4_en_transmit_locked(if_t ifp, int tx_ind, struct mbuf *mb)
sys/dev/mlx4/mlx4_en/mlx4_en_tx.c
938
m_freem(mb);
sys/dev/mlx4/mlx4_en/mlx4_en_tx.c
942
if (mlx4_en_xmit(priv, tx_ind, &mb) != 0) {
sys/dev/mlx4/mlx4_en/mlx4_en_tx.c
944
m_freem(mb);
sys/dev/mlx5/mlx5_accel/ipsec.h
241
void mlx5e_accel_ipsec_handle_tx_wqe(struct mbuf *mb, struct mlx5e_tx_wqe *wqe,
sys/dev/mlx5/mlx5_accel/ipsec.h
250
mlx5e_accel_ipsec_handle_tx(struct mbuf *mb, struct mlx5e_tx_wqe *wqe)
sys/dev/mlx5/mlx5_accel/ipsec.h
254
tag = (struct ipsec_accel_out_tag *)m_tag_find(mb,
sys/dev/mlx5/mlx5_accel/ipsec.h
257
mlx5e_accel_ipsec_handle_tx_wqe(mb, wqe, tag);
sys/dev/mlx5/mlx5_accel/ipsec.h
264
void mlx5e_accel_ipsec_handle_rx_cqe(if_t ifp, struct mbuf *mb,
sys/dev/mlx5/mlx5_accel/ipsec.h
273
mlx5e_accel_ipsec_handle_rx(if_t ifp, struct mbuf *mb, struct mlx5_cqe64 *cqe,
sys/dev/mlx5/mlx5_accel/ipsec.h
279
mlx5e_accel_ipsec_handle_rx_cqe(ifp, mb, cqe, mr);
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_rxtx.c
63
mlx5e_accel_ipsec_handle_rx_cqe(if_t ifp, struct mbuf *mb,
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_rxtx.c
75
m_tag_prepend(mb, &mtag->tag);
sys/dev/mlx5/mlx5_accel/mlx5_ipsec_rxtx.c
82
mlx5e_accel_ipsec_handle_tx_wqe(struct mbuf *mb, struct mlx5e_tx_wqe *wqe,
sys/dev/mlx5/mlx5_core/mlx5_eq.c
218
mb();
sys/dev/mlx5/mlx5_en/mlx5_en_hw_tls.c
717
sbtls_recover_record(struct mbuf *mb, int wait, uint32_t tcp_old, uint32_t *ptcp_seq, bool *pis_start)
sys/dev/mlx5/mlx5_en/mlx5_en_hw_tls.c
724
if (mb->m_next == NULL ||
sys/dev/mlx5/mlx5_en/mlx5_en_hw_tls.c
725
(mb->m_next->m_flags & (M_EXTPG | M_EXT)) != (M_EXTPG | M_EXT)) {
sys/dev/mlx5/mlx5_en/mlx5_en_hw_tls.c
731
offset = mtod(mb->m_next, uintptr_t);
sys/dev/mlx5/mlx5_en/mlx5_en_hw_tls.c
754
mb_dupcl(mr, mb->m_next);
sys/dev/mlx5/mlx5_en/mlx5_en_hw_tls.c
786
mlx5e_sq_tls_populate(struct mbuf *mb, uint64_t *pseq)
sys/dev/mlx5/mlx5_en/mlx5_en_hw_tls.c
789
for (; mb != NULL; mb = mb->m_next) {
sys/dev/mlx5/mlx5_en/mlx5_en_hw_tls.c
790
if (!(mb->m_flags & M_EXTPG))
sys/dev/mlx5/mlx5_en/mlx5_en_hw_tls.c
792
*pseq = mb->m_epg_seqno;
sys/dev/mlx5/mlx5_en/mlx5_en_hw_tls.c
804
struct mbuf *mb = *ppmb;
sys/dev/mlx5/mlx5_en/mlx5_en_hw_tls.c
809
if ((mb->m_pkthdr.csum_flags & CSUM_SND_TAG) == 0)
sys/dev/mlx5/mlx5_en/mlx5_en_hw_tls.c
812
ptag = mb->m_pkthdr.snd_tag;
sys/dev/mlx5/mlx5_en/mlx5_en_hw_tls.c
823
header_size = mlx5e_get_full_header_size(mb, &th);
sys/dev/mlx5/mlx5_en/mlx5_en_hw_tls.c
830
if (header_size == mb->m_pkthdr.len ||
sys/dev/mlx5/mlx5_en/mlx5_en_hw_tls.c
831
mlx5e_sq_tls_populate(mb, &rcd_sn) == 0) {
sys/dev/mlx5/mlx5_en/mlx5_en_hw_tls.c
857
r_mb = sbtls_recover_record(mb, M_NOWAIT, ptls_tag->expected_seq, &tcp_seq, &is_start);
sys/dev/mlx5/mlx5_en/mlx5_en_hw_tls.c
903
MLX5E_TLS_STAT_INC(ptls_tag, tx_bytes, mb->m_pkthdr.len);
sys/dev/mlx5/mlx5_en/mlx5_en_hw_tls.c
905
ptls_tag->expected_seq += mb->m_pkthdr.len - header_size;
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
102
m_freem(mb);
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
150
mlx5e_lro_update_hdr(struct mbuf *mb, struct mlx5_cqe64 *cqe)
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
164
eh = mtod(mb, struct ether_header *);
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
226
ip4->ip_sum = in_cksum_skip(mb, (ip4->ip_hl << 2) +
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
251
tcp_csum = ~in6_cksum_partial_l2(mb, IPPROTO_TCP,
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
326
struct mbuf *mb, struct mlx5e_rq_mbuf *mr, u32 cqe_bcnt)
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
336
mlx5e_lro_update_hdr(mb, cqe);
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
341
mb->m_pkthdr.len = cqe_bcnt;
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
342
for (mb_head = mb; mb != NULL; mb = mb->m_next) {
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
343
if (mb->m_len > cqe_bcnt)
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
344
mb->m_len = cqe_bcnt;
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
345
cqe_bcnt -= mb->m_len;
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
347
if (likely(mb->m_next != NULL)) {
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
349
m_freem(mb->m_next);
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
350
mb->m_next = NULL;
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
356
mb = mb_head;
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
360
mb->m_pkthdr.flowid = be32_to_cpu(cqe->rss_hash_result);
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
366
M_HASHTYPE_SET(mb, M_HASHTYPE_RSS_TCP_IPV4);
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
369
M_HASHTYPE_SET(mb, M_HASHTYPE_RSS_UDP_IPV4);
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
372
M_HASHTYPE_SET(mb, M_HASHTYPE_RSS_IPV4);
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
376
M_HASHTYPE_SET(mb, M_HASHTYPE_RSS_TCP_IPV6);
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
379
M_HASHTYPE_SET(mb, M_HASHTYPE_RSS_UDP_IPV6);
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
382
M_HASHTYPE_SET(mb, M_HASHTYPE_RSS_IPV6);
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
385
M_HASHTYPE_SET(mb, M_HASHTYPE_OPAQUE_HASH);
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
39
struct mbuf *mb;
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
390
M_HASHTYPE_SETINNER(mb);
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
393
mb->m_pkthdr.flowid = rq->ix;
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
394
M_HASHTYPE_SET(mb, M_HASHTYPE_OPAQUE);
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
396
mb->m_pkthdr.rcvif = ifp;
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
397
mb->m_pkthdr.leaf_rcvif = ifp;
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
407
mb->m_pkthdr.csum_flags |=
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
411
mb->m_pkthdr.csum_data = htons(0xffff);
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
414
mb->m_pkthdr.csum_flags |=
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
424
mb->m_pkthdr.csum_flags =
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
427
mb->m_pkthdr.csum_data = htons(0xffff);
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
433
mb->m_pkthdr.ether_vtag = be16_to_cpu(cqe->vlan_info);
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
434
mb->m_flags |= M_VLANTAG;
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
446
mb->m_flags |= M_TSTMP_HPREC;
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
449
mb->m_pkthdr.rcv_tstmp = tstmp;
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
450
mb->m_flags |= M_TSTMP;
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
456
mb->m_pkthdr.csum_flags |= CSUM_TLS_DECRYPTED;
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
466
mlx5e_accel_ipsec_handle_rx(ifp, mb, cqe, mr);
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
48
mb_head = mb = m_getjcl(M_NOWAIT, MT_DATA, M_PKTHDR, rq->wqe_sz);
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
49
if (unlikely(mb == NULL))
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
52
mb->m_len = rq->wqe_sz;
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
53
mb->m_pkthdr.len = rq->wqe_sz;
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
553
struct mbuf *mb;
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
56
mb = mb->m_next = m_getjcl(M_NOWAIT, MT_DATA, 0, rq->wqe_sz);
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
57
if (unlikely(mb == NULL)) {
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
584
rq->mbuf[wqe_counter].data, seglen, rq->ifp, &mb);
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
61
mb->m_len = rq->wqe_sz;
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
615
(mb = m_gethdr(M_NOWAIT, MT_DATA)) != NULL) {
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
617
mb->m_len = MHLEN - MLX5E_NET_IP_ALIGN;
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
619
mb->m_data += MLX5E_NET_IP_ALIGN;
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
621
bcopy(rq->mbuf[wqe_counter].data, mtod(mb, caddr_t),
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
624
mb = rq->mbuf[wqe_counter].mbuf;
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
631
mlx5e_build_rx_mbuf(cqe, rq, mb, &rq->mbuf[wqe_counter],
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
636
mb->m_pkthdr.numa_domain = if_getnumadomain(rq->ifp);
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
640
tcp_lro_queue_mbuf(&rq->lro, mb);
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
642
if (mb->m_pkthdr.csum_flags == 0 ||
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
645
tcp_lro_rx(&rq->lro, mb, 0) != 0) {
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
646
if_input(rq->ifp, mb);
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
65
mb = mb_head;
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
673
struct mbuf *mb = m_gethdr(M_NOWAIT, MT_DATA);
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
675
if (mb != NULL) {
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
677
mb->m_pkthdr.len = mb->m_len = 15;
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
678
memset(mb->m_data, 255, 14);
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
679
mb->m_data[14] = rq->ix;
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
68
m_adj(mb, MLX5E_NET_IP_ALIGN);
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
680
mb->m_pkthdr.rcvif = rq->ifp;
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
681
mb->m_pkthdr.leaf_rcvif = rq->ifp;
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
682
if_input(rq->ifp, mb);
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
74
mb, segs, &nsegs, BUS_DMA_NOWAIT);
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
94
rq->mbuf[ix].mbuf = mb;
sys/dev/mlx5/mlx5_en/mlx5_en_rx.c
95
rq->mbuf[ix].data = mb->m_data;
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
1017
m_freem(mb);
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
1035
struct mbuf *mb;
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
1072
mb = sq->mbuf[ci].mbuf;
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
1077
if (unlikely(mb == NULL)) {
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
1086
m_freem(mb);
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
1105
mlx5e_xmit_locked(if_t ifp, struct mlx5e_sq *sq, struct mbuf *mb)
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
1111
m_freem(mb);
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
1116
if (mlx5e_sq_xmit(sq, &mb) != 0) {
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
1118
m_freem(mb);
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
1141
mlx5e_xmit(if_t ifp, struct mbuf *mb)
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
1146
if (mb->m_pkthdr.csum_flags & CSUM_SND_TAG) {
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
1147
MPASS(mb->m_pkthdr.snd_tag->ifp == ifp);
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
1148
sq = mlx5e_select_queue_by_send_tag(ifp, mb);
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
1154
sq = mlx5e_select_queue(ifp, mb);
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
1157
m_freem(mb);
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
1165
ret = mlx5e_xmit_locked(ifp, sq, mb);
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
137
mlx5e_select_queue(if_t ifp, struct mbuf *mb)
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
145
if (mb->m_flags & M_VLANTAG) {
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
146
tc = (mb->m_pkthdr.ether_vtag >> 13);
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
156
if (M_HASHTYPE_GET(mb) != M_HASHTYPE_NONE) {
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
160
if (rss_hash2bucket(mb->m_pkthdr.flowid,
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
161
M_HASHTYPE_GET(mb), &temp) == 0)
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
165
ch = (mb->m_pkthdr.flowid % 128) % ch;
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
168
MBUF_HASHFLAG_L4, mb, mlx5e_hash_value) % ch;
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
179
mlx5e_get_l2_header_size(struct mlx5e_sq *sq, struct mbuf *mb)
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
185
eh = mtod(mb, struct ether_vlan_header *);
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
186
if (unlikely(mb->m_len < ETHER_HDR_LEN)) {
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
189
if (unlikely(mb->m_len < (ETHER_HDR_LEN + ETHER_VLAN_ENCAP_LEN)))
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
216
if (mb->m_pkthdr.len < min_inline)
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
221
return (MIN(mb->m_pkthdr.len, sq->max_inline));
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
237
mlx5e_get_full_header_size(const struct mbuf *mb, const struct tcphdr **ppth)
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
247
eh = mtod(mb, const struct ether_vlan_header *);
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
248
if (unlikely(mb->m_len < ETHER_HDR_LEN))
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
251
if (unlikely(mb->m_len < ETHER_HDR_LEN + ETHER_VLAN_ENCAP_LEN))
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
262
ip = (const struct ip *)(mb->m_data + eth_hdr_len);
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
263
if (unlikely(mb->m_len < eth_hdr_len + sizeof(*ip)))
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
280
ip6 = (const struct ip6_hdr *)(mb->m_data + eth_hdr_len);
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
281
if (unlikely(mb->m_len < eth_hdr_len + sizeof(*ip6)))
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
299
if (unlikely(mb->m_len < eth_hdr_len + sizeof(*th))) {
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
300
const struct mbuf *m_th = mb->m_next;
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
301
if (unlikely(mb->m_len != eth_hdr_len ||
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
306
th = (const struct tcphdr *)(mb->m_data + eth_hdr_len);
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
316
if (unlikely(mb->m_pkthdr.len < eth_hdr_len))
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
331
mlx5e_parse_mbuf_chain(const struct mbuf **mb, int *poffset, int eth_hdr_len,
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
334
if (unlikely(mb[0]->m_len == eth_hdr_len)) {
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
336
if (unlikely((mb[0] = mb[0]->m_next) == NULL))
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
339
if (unlikely(mb[0]->m_len < eth_hdr_len - poffset[0] + min_len))
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
341
return (mb[0]->m_data + eth_hdr_len - poffset[0]);
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
353
mlx5e_get_vxlan_header_size(const struct mbuf *mb, struct mlx5e_tx_wqe *wqe,
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
370
pkt_hdr_len = mb->m_pkthdr.len;
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
371
has_outer_vlan_tag = (mb->m_flags & M_VLANTAG) != 0;
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
374
eh = mtod(mb, const struct ether_vlan_header *);
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
375
if (unlikely(mb->m_len < ETHER_HDR_LEN))
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
379
if (unlikely(mb->m_len < ETHER_HDR_LEN + ETHER_VLAN_ENCAP_LEN))
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
390
ip4 = mlx5e_parse_mbuf_chain(&mb, &offset, eth_hdr_len,
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
401
udp = mlx5e_parse_mbuf_chain(&mb, &offset, eth_hdr_len,
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
410
ip6 = mlx5e_parse_mbuf_chain(&mb, &offset, eth_hdr_len,
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
420
udp = mlx5e_parse_mbuf_chain(&mb, &offset, eth_hdr_len,
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
439
if (unlikely(mlx5e_parse_mbuf_chain(&mb, &offset, eth_hdr_len,
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
445
eh = mlx5e_parse_mbuf_chain(&mb, &offset, eth_hdr_len, ETHER_HDR_LEN);
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
449
if (unlikely(mb->m_len < eth_hdr_len - offset + ETHER_HDR_LEN +
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
462
ip4 = mlx5e_parse_mbuf_chain(&mb, &offset, eth_hdr_len,
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
473
ip6 = mlx5e_parse_mbuf_chain(&mb, &offset, eth_hdr_len,
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
495
udp = mlx5e_parse_mbuf_chain(&mb, &offset, eth_hdr_len,
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
505
th = mlx5e_parse_mbuf_chain(&mb, &offset, eth_hdr_len,
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
559
struct mbuf *mb;
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
565
mb = *mbp;
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
570
sq->mbuf[pi].num_bytes = mb->m_pkthdr.len;
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
575
mb, segs, &nsegs, BUS_DMA_NOWAIT);
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
580
mb = m_defrag(*mbp, M_NOWAIT);
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
581
if (mb == NULL) {
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
582
mb = *mbp;
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
587
mb, segs, &nsegs, BUS_DMA_NOWAIT);
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
605
m_freem(mb);
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
614
m_freem(mb);
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
667
sq->mbuf[pi].mbuf = mb;
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
680
m_freem(mb);
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
695
struct mbuf *mb;
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
727
mb = *mbp;
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
746
mb = *mbp;
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
748
mlx5e_accel_ipsec_handle_tx(mb, wqe);
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
752
ETHER_BPF_MTAP(ifp, mb);
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
754
if (mb->m_pkthdr.csum_flags & (CSUM_IP | CSUM_TSO)) {
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
757
if (mb->m_pkthdr.csum_flags & (CSUM_TCP | CSUM_UDP | CSUM_UDP_IPV6 | CSUM_TCP_IPV6 | CSUM_TSO)) {
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
763
if (mb->m_pkthdr.csum_flags & CSUM_TSO) {
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
765
u32 mss = mb->m_pkthdr.tso_segsz;
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
771
args.ihs = mlx5e_get_full_header_size(mb, NULL);
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
776
payload_len = mb->m_pkthdr.len - args.ihs;
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
786
} else if (mb->m_pkthdr.csum_flags & CSUM_ENCAP_VXLAN) {
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
788
if (mb->m_pkthdr.csum_flags & (CSUM_INNER_IP_TSO |
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
791
u32 mss = mb->m_pkthdr.tso_segsz;
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
798
args.ihs = mlx5e_get_vxlan_header_size(mb, wqe,
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
810
payload_len = mb->m_pkthdr.len - args.ihs;
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
826
if (mb->m_pkthdr.csum_flags &
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
834
} else if (mb->m_pkthdr.csum_flags & CSUM_INNER_IP) {
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
844
args.ihs = mlx5e_get_vxlan_header_size(mb, wqe,
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
853
mb->m_pkthdr.len, ETHER_MIN_LEN - ETHER_CRC_LEN);
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
862
args.ihs = mlx5e_get_full_header_size(mb, NULL);
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
864
args.ihs = mlx5e_get_l2_header_size(sq, mb);
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
867
args.ihs = mlx5e_get_l2_header_size(sq, mb);
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
872
if ((mb->m_flags & M_VLANTAG) != 0 &&
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
876
wqe->eth.vlan_hdr = htons(mb->m_pkthdr.ether_vtag);
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
878
} else if ((mb->m_flags & M_VLANTAG) == 0 &&
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
884
args.ihs = mlx5e_get_l2_header_size(sq, mb);
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
890
mb->m_pkthdr.len, ETHER_MIN_LEN - ETHER_CRC_LEN);
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
895
} else if ((mb->m_flags & M_VLANTAG) != 0) {
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
90
mlx5e_select_queue_by_send_tag(if_t ifp, struct mbuf *mb)
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
901
if (mb->m_pkthdr.csum_flags & (CSUM_TSO | CSUM_ENCAP_VXLAN)) {
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
910
m_copydata(mb, 0, ETHER_HDR_LEN, (caddr_t)eh);
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
911
m_adj(mb, ETHER_HDR_LEN);
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
915
eh->evl_tag = htons(mb->m_pkthdr.ether_vtag);
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
917
m_copydata(mb, 0, args.ihs - ETHER_HDR_LEN, (caddr_t)(eh + 1));
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
918
m_adj(mb, args.ihs - ETHER_HDR_LEN);
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
925
if (unlikely(mb->m_pkthdr.csum_flags & (CSUM_TSO |
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
932
m_copydata(mb, 0, args.ihs, wqe->eth.inline_hdr_start);
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
933
m_adj(mb, args.ihs);
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
945
mb, segs, &nsegs, BUS_DMA_NOWAIT);
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
95
mb_tag = mb->m_pkthdr.snd_tag;
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
950
mb = m_defrag(*mbp, M_NOWAIT);
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
951
if (mb == NULL) {
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
952
mb = *mbp;
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
957
mb, segs, &nsegs, BUS_DMA_NOWAIT);
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
970
m_freem(mb);
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
971
mb = NULL;
sys/dev/mlx5/mlx5_en/mlx5_en_tx.c
998
sq->mbuf[pi].mbuf = mb;
sys/dev/mpt/mpt.h
1131
void mpt_print_db(uint32_t mb);
sys/dev/mpt/mpt_debug.c
321
mpt_state(u_int32_t mb)
sys/dev/mpt/mpt_debug.c
325
switch (MPT_STATE(mb)) {
sys/dev/mpt/mpt_debug.c
350
mpt_print_db(u_int32_t mb)
sys/dev/mpt/mpt_debug.c
354
mb, mpt_state(mb), mpt_who(MPT_WHO(mb)));
sys/dev/mthca/mthca_eq.c
196
mb();
sys/dev/netmap/netmap.c
2671
mb(); /* make sure previous writes are visible to all CPUs */
sys/dev/netmap/netmap.c
3168
mb(); /* make sure following reads are not from cache */
sys/dev/netmap/netmap.c
3686
mb(); /* make sure following reads are not from cache */
sys/dev/netmap/netmap_kloop.c
599
mb(); /* make sure following reads are not from cache */
sys/dev/netmap/netmap_kloop.c
924
mb(); /* make sure following reads are not from cache */
sys/dev/netmap/netmap_monitor.c
164
mb();
sys/dev/netmap/netmap_monitor.c
673
mb();
sys/dev/netmap/netmap_monitor.c
791
mb();
sys/dev/netmap/netmap_pipe.c
232
mb(); /* make sure the slots are updated before publishing them */
sys/dev/netmap/netmap_pipe.c
276
mb(); /* make sure the slots are updated before publishing them */
sys/dev/nvme/nvme_test.c
154
mb();
sys/dev/nvme/nvme_test.c
230
mb();
sys/dev/oce/oce_mbox.c
371
struct oce_bmbx *mb = NULL;
sys/dev/oce/oce_mbox.c
378
mb = OCE_DMAPTR(&sc->bsmbx, struct oce_bmbx);
sys/dev/oce/oce_mbox.c
379
mb_mbx = &mb->mbx;
sys/dev/oce/oce_mbox.c
395
mb_cqe = &mb->cqe;
sys/dev/oce/oce_mbox.c
81
struct oce_bmbx *mb;
sys/dev/oce/oce_mbox.c
97
mb = OCE_DMAPTR(&sc->bsmbx, struct oce_bmbx);
sys/dev/oce/oce_mbox.c
98
mbx = &mb->mbx;
sys/dev/qlnx/qlnxe/bcm_osal.h
358
#define OSAL_SMP_MB(dev) mb()
sys/dev/sfxge/sfxge_tx.c
248
mb();
sys/dev/sfxge/sfxge_tx.c
386
mb();
sys/dev/sfxge/sfxge_tx.c
390
mb();
sys/dev/smartpqi/smartpqi_helper.c
89
int lockupcode = PCI_MEM_GET32(softs, &softs->ioa_reg->mb[7], LEGACY_SIS_SRCV_OFFSET_MAILBOX_7);
sys/dev/smartpqi/smartpqi_sis.c
166
pqisrc_send_sis_cmd(pqisrc_softstate_t *softs, uint32_t *mb)
sys/dev/smartpqi/smartpqi_sis.c
179
PCI_MEM_PUT32(softs, &softs->ioa_reg->mb[i],
sys/dev/smartpqi/smartpqi_sis.c
180
LEGACY_SIS_SRCV_MAILBOX+i*4, LE_32(mb[i]));
sys/dev/smartpqi/smartpqi_sis.c
202
DBG_ERR("Sync command %x, timedout\n", mb[0]);
sys/dev/smartpqi/smartpqi_sis.c
207
mb[0] = LE_32(PCI_MEM_GET32(softs, &softs->ioa_reg->mb[0], LEGACY_SIS_SRCV_MAILBOX));
sys/dev/smartpqi/smartpqi_sis.c
209
if (mb[0] != SIS_CMD_STATUS_SUCCESS) {
sys/dev/smartpqi/smartpqi_sis.c
211
mb[0]);
sys/dev/smartpqi/smartpqi_sis.c
218
mb[i] = LE_32(PCI_MEM_GET32(softs, &softs->ioa_reg->mb[i], LEGACY_SIS_SRCV_MAILBOX+i*4));
sys/dev/smartpqi/smartpqi_sis.c
234
uint32_t mb[6] = {0};
sys/dev/smartpqi/smartpqi_sis.c
238
mb[0] = SIS_CMD_GET_ADAPTER_PROPERTIES;
sys/dev/smartpqi/smartpqi_sis.c
239
ret = pqisrc_send_sis_cmd(softs, mb);
sys/dev/smartpqi/smartpqi_sis.c
242
mb[1], mb[4]);
sys/dev/smartpqi/smartpqi_sis.c
243
*prop = mb[1];
sys/dev/smartpqi/smartpqi_sis.c
244
*ext_prop = mb[4];
sys/dev/smartpqi/smartpqi_sis.c
256
uint32_t mb[6] = {0};
sys/dev/smartpqi/smartpqi_sis.c
260
mb[0] = SIS_CMD_GET_COMM_PREFERRED_SETTINGS;
sys/dev/smartpqi/smartpqi_sis.c
261
ret = pqisrc_send_sis_cmd(softs, mb);
sys/dev/smartpqi/smartpqi_sis.c
264
softs->pref_settings.max_cmd_size = mb[1] >> 16;
sys/dev/smartpqi/smartpqi_sis.c
266
softs->pref_settings.max_fib_size = mb[1] & 0x0000FFFF;
sys/dev/smartpqi/smartpqi_sis.c
281
uint32_t mb[6] = {0};
sys/dev/smartpqi/smartpqi_sis.c
285
mb[0] = SIS_CMD_GET_PQI_CAPABILITIES;
sys/dev/smartpqi/smartpqi_sis.c
286
ret = pqisrc_send_sis_cmd(softs, mb);
sys/dev/smartpqi/smartpqi_sis.c
288
softs->pqi_cap.max_sg_elem = mb[1];
sys/dev/smartpqi/smartpqi_sis.c
289
softs->pqi_cap.max_transfer_size = mb[2];
sys/dev/smartpqi/smartpqi_sis.c
290
softs->pqi_cap.max_outstanding_io = mb[3];
sys/dev/smartpqi/smartpqi_sis.c
306
softs->pqi_cap.conf_tab_off = mb[4];
sys/dev/smartpqi/smartpqi_sis.c
307
softs->pqi_cap.conf_tab_sz = mb[5];
sys/dev/smartpqi/smartpqi_sis.c
337
uint32_t mb[6] = {0};
sys/dev/smartpqi/smartpqi_sis.c
381
mb[0] = SIS_CMD_INIT_BASE_STRUCT_ADDRESS;
sys/dev/smartpqi/smartpqi_sis.c
382
mb[1] = DMA_PHYS_LOW(&init_struct_mem);
sys/dev/smartpqi/smartpqi_sis.c
383
mb[2] = DMA_PHYS_HIGH(&init_struct_mem);
sys/dev/smartpqi/smartpqi_sis.c
384
mb[3] = init_struct_mem.size;
sys/dev/smartpqi/smartpqi_sis.c
386
ret = pqisrc_send_sis_cmd(softs, mb);
sys/dev/smartpqi/smartpqi_structures.h
135
uint32_t mb[8]; /* 1000h */
sys/dev/usb/net/if_ure.c
675
struct mbuf *m, *mb;
sys/dev/usb/net/if_ure.c
687
for (mb = m; len > 0; mb = mb->m_next) {
sys/dev/usb/net/if_ure.c
688
tlen = MIN(len, M_TRAILINGSPACE(mb));
sys/dev/usb/net/if_ure.c
690
usbd_copy_out(pc, offset, mtod(mb, uint8_t *), tlen);
sys/dev/usb/net/if_ure.c
691
mb->m_len = tlen;
sys/dev/virtio/virtqueue.c
569
mb();
sys/dev/virtio/virtqueue.c
967
mb();
sys/dev/vmware/pvscsi/pvscsi.c
1050
mb();
sys/dev/vmware/pvscsi/pvscsi.c
1112
mb();
sys/dev/vmware/pvscsi/pvscsi.c
1247
mb();
sys/dev/vmware/vmxnet3/if_vmx.c
2532
mb();
sys/dev/xen/balloon/balloon.c
120
#define MB2PAGES(mb) ((mb) << (20 - PAGE_SHIFT))
sys/dev/xen/console/xen_console.c
334
mb();
sys/dev/xen/netfront/netfront.c
1414
mb();
sys/dev/xen/xenstore/xenstore.c
600
mb();
sys/fs/nfs/nfs_commonsubs.c
362
struct mbuf *mb;
sys/fs/nfs/nfs_commonsubs.c
399
NFSMCLGET(mb, M_WAITOK);
sys/fs/nfs/nfs_commonsubs.c
401
NFSMGET(mb);
sys/fs/nfs/nfs_commonsubs.c
402
mb->m_len = 0;
sys/fs/nfs/nfs_commonsubs.c
403
nd->nd_mreq = nd->nd_mb = mb;
sys/fs/nfs/nfs_commonsubs.c
404
nd->nd_bpos = mtod(mb, char *);
sys/fs/nfsserver/nfs_nfsdport.c
7348
nfsm_trimtrailing(struct nfsrv_descript *nd, struct mbuf *mb, char *bpos,
sys/fs/nfsserver/nfs_nfsdport.c
7354
if (mb->m_next != NULL) {
sys/fs/nfsserver/nfs_nfsdport.c
7355
m_freem(mb->m_next);
sys/fs/nfsserver/nfs_nfsdport.c
7356
mb->m_next = NULL;
sys/fs/nfsserver/nfs_nfsdport.c
7358
if ((mb->m_flags & M_EXTPG) != 0) {
sys/fs/nfsserver/nfs_nfsdport.c
7359
KASSERT(bextpg >= 0 && bextpg < mb->m_epg_npgs,
sys/fs/nfsserver/nfs_nfsdport.c
7362
PHYS_TO_DMAP(mb->m_epg_pa[bextpg]) + PAGE_SIZE - bextpgsiz,
sys/fs/nfsserver/nfs_nfsdport.c
7366
for (i = mb->m_epg_npgs - 1; i > bextpg; i--) {
sys/fs/nfsserver/nfs_nfsdport.c
7367
pg = PHYS_TO_VM_PAGE(mb->m_epg_pa[i]);
sys/fs/nfsserver/nfs_nfsdport.c
7371
mb->m_epg_npgs = bextpg + 1;
sys/fs/nfsserver/nfs_nfsdport.c
7373
fullpgsiz = PAGE_SIZE - mb->m_epg_1st_off;
sys/fs/nfsserver/nfs_nfsdport.c
7376
mb->m_epg_last_len = fullpgsiz - bextpgsiz;
sys/fs/nfsserver/nfs_nfsdport.c
7377
mb->m_len = m_epg_pagelen(mb, 0, mb->m_epg_1st_off);
sys/fs/nfsserver/nfs_nfsdport.c
7378
for (i = 1; i < mb->m_epg_npgs; i++)
sys/fs/nfsserver/nfs_nfsdport.c
7379
mb->m_len += m_epg_pagelen(mb, i, 0);
sys/fs/nfsserver/nfs_nfsdport.c
7383
mb->m_len = bpos - mtod(mb, char *);
sys/fs/nfsserver/nfs_nfsdport.c
7384
nd->nd_mb = mb;
sys/fs/nfsserver/nfs_nfsdsocket.c
1215
mb = nd->nd_mb;
sys/fs/nfsserver/nfs_nfsdsocket.c
1366
nfsm_trimtrailing(nd, mb, bpos, bextpg, bextpgsiz);
sys/fs/nfsserver/nfs_nfsdsocket.c
726
struct mbuf *mb, *md;
sys/i386/i386/pmap.c
4673
vm_page_t mb[], vm_offset_t b_offset, int xfersize)
sys/i386/i386/pmap.c
4695
b_pg = mb[b_offset >> PAGE_SHIFT];
sys/i386/i386/pmap_base.c
659
pmap_copy_pages(vm_page_t ma[], vm_offset_t a_offset, vm_page_t mb[],
sys/i386/i386/pmap_base.c
663
pmap_methods_ptr->pm_copy_pages(ma, a_offset, mb, b_offset, xfersize);
sys/kern/kern_mbuf.c
1470
struct mbuf *mb;
sys/kern/kern_mbuf.c
1487
mb = m_getjcl(M_NOWAIT, type, (flags & M_PKTHDR),
sys/kern/kern_mbuf.c
1490
mb = NULL;
sys/kern/kern_mbuf.c
1491
if (mb == NULL) {
sys/kern/kern_mbuf.c
1493
mb = m_getcl(how, type, (flags & M_PKTHDR));
sys/kern/kern_mbuf.c
1495
mb = m_gethdr(how, type);
sys/kern/kern_mbuf.c
1497
mb = m_get(how, type);
sys/kern/kern_mbuf.c
1503
if (__predict_false(mb == NULL)) {
sys/kern/kern_mbuf.c
1510
progress += M_SIZE(mb);
sys/kern/kern_mbuf.c
1511
mc_append(mc, mb);
sys/kern/kern_mbuf.c
1573
m_extadd(struct mbuf *mb, char *buf, u_int size, m_ext_free_t freef,
sys/kern/kern_mbuf.c
1579
mb->m_flags |= (M_EXT | flags);
sys/kern/kern_mbuf.c
1580
mb->m_ext.ext_buf = buf;
sys/kern/kern_mbuf.c
1581
mb->m_data = mb->m_ext.ext_buf;
sys/kern/kern_mbuf.c
1582
mb->m_ext.ext_size = size;
sys/kern/kern_mbuf.c
1583
mb->m_ext.ext_free = freef;
sys/kern/kern_mbuf.c
1584
mb->m_ext.ext_arg1 = arg1;
sys/kern/kern_mbuf.c
1585
mb->m_ext.ext_arg2 = arg2;
sys/kern/kern_mbuf.c
1586
mb->m_ext.ext_type = type;
sys/kern/kern_mbuf.c
1589
mb->m_ext.ext_count = 1;
sys/kern/kern_mbuf.c
1590
mb->m_ext.ext_flags = EXT_FLAG_EMBREF;
sys/kern/kern_mbuf.c
1592
mb->m_ext.ext_flags = 0;
sys/kern/kern_mbuf.c
1600
m_freem(struct mbuf *mb)
sys/kern/kern_mbuf.c
1603
MBUF_PROBE1(m__freem, mb);
sys/kern/kern_mbuf.c
1604
while (mb != NULL)
sys/kern/kern_mbuf.c
1605
mb = m_free(mb);
sys/kern/kern_mbuf.c
1631
m_free_raw(struct mbuf *mb)
sys/kern/kern_mbuf.c
1634
uma_zfree(zone_mbuf, mb);
sys/kern/subr_smp.c
420
struct monitorbuf *mb;
sys/kern/subr_smp.c
427
mb = &pcpu_find(id)->pc_monitorbuf;
sys/kern/subr_smp.c
428
atomic_store_int(&mb->stop_state,
sys/kern/subr_stats.c
1000
} else if (hash == tpllist[i]->mb->tplhash) {
sys/kern/subr_stats.c
1019
if (buf != NULL && len > strlen(tpllist[tpl_id]->mb->tplname))
sys/kern/subr_stats.c
1020
strlcpy(buf, tpllist[tpl_id]->mb->tplname, len);
sys/kern/subr_stats.c
1184
tpl->mb = tpl_mb;
sys/kern/subr_stats.c
1262
tpl_mb = tpllist[tpl_id]->mb;
sys/kern/subr_stats.c
127
struct metablob *mb;
sys/kern/subr_stats.c
1476
sb->tplhash = tpllist[tpl_id]->mb->tplhash;
sys/kern/subr_stats.c
2047
tpl_mb = sctx->tpl ? sctx->tpl->mb : NULL;
sys/kern/subr_stats.c
2109
tpl_mb = sctx->tpl ? sctx->tpl->mb : NULL;
sys/kern/subr_stats.c
3721
tpllist[i]->mb->tplname, tpllist[i]->mb->tplhash);
sys/kern/subr_stats.c
3806
tpllist[rates[i].tpl_slot_id]->mb->tplname,
sys/kern/subr_stats.c
3807
tpllist[rates[i].tpl_slot_id]->mb->tplhash,
sys/kern/subr_stats.c
437
tpl->mb->tplhash = hash32_str(tpl->mb->tplname, 0);
sys/kern/subr_stats.c
439
if (tpl->mb->voi_meta[voi_id].name != NULL)
sys/kern/subr_stats.c
440
tpl->mb->tplhash = hash32_str(
sys/kern/subr_stats.c
441
tpl->mb->voi_meta[voi_id].name, tpl->mb->tplhash);
sys/kern/subr_stats.c
443
tpl->mb->tplhash = hash32_buf(tpl->sb, tpl->sb->cursz,
sys/kern/subr_stats.c
444
tpl->mb->tplhash);
sys/kern/subr_stats.c
993
if (strlen(name) == strlen(tpllist[i]->mb->tplname) &&
sys/kern/subr_stats.c
994
strncmp(name, tpllist[i]->mb->tplname,
sys/kern/subr_stats.c
996
tpllist[i]->mb->tplhash)) {
sys/kern/uipc_ktls.c
2426
ktls_mbuf_crypto_state(struct mbuf *mb, int offset, int len)
sys/kern/uipc_ktls.c
2431
for (; mb != NULL; mb = mb->m_next) {
sys/kern/uipc_ktls.c
2432
if (offset < mb->m_len)
sys/kern/uipc_ktls.c
2434
offset -= mb->m_len;
sys/kern/uipc_ktls.c
2438
for (; mb != NULL; mb = mb->m_next) {
sys/kern/uipc_ktls.c
2439
m_flags_ored |= mb->m_flags;
sys/kern/uipc_ktls.c
2440
m_flags_anded &= mb->m_flags;
sys/kern/uipc_ktls.c
2442
if (offset <= mb->m_len)
sys/kern/uipc_ktls.c
2444
offset -= mb->m_len;
sys/kern/uipc_ktls.c
2446
MPASS(mb != NULL || offset == 0);
sys/kern/uipc_mbuf.c
1862
struct mbuf *m, *mb, *prev;
sys/kern/uipc_mbuf.c
1890
mb = mb_alloc_ext_pgs(how, mb_free_mext_pgs, 0);
sys/kern/uipc_mbuf.c
1891
if (mb == NULL)
sys/kern/uipc_mbuf.c
1893
mb->m_epg_flags = EPG_FLAG_ANON;
sys/kern/uipc_mbuf.c
1894
return (mb);
sys/kern/uipc_mbuf.c
1902
mb = mb_alloc_ext_pgs(how, mb_free_mext_pgs, 0);
sys/kern/uipc_mbuf.c
1903
if (mb == NULL)
sys/kern/uipc_mbuf.c
1906
m = mb;
sys/kern/uipc_mbuf.c
1908
prev->m_next = mb;
sys/kern/uipc_mbuf.c
1909
prev = mb;
sys/kern/uipc_mbuf.c
1910
mb->m_epg_flags = EPG_FLAG_ANON;
sys/kern/uipc_mbuf.c
1923
mb->m_epg_pa[i] = VM_PAGE_TO_PHYS(pg_array[i]);
sys/kern/uipc_mbuf.c
1924
mb->m_epg_npgs++;
sys/kern/uipc_mbuf.c
1926
mb->m_epg_last_len = length - PAGE_SIZE * (mb->m_epg_npgs - 1);
sys/kern/uipc_mbuf.c
1927
MBUF_EXT_PGS_ASSERT_SANITY(mb);
sys/kern/uipc_mbuf.c
1932
mb->m_len = length;
sys/kern/uipc_mbuf.c
1933
mb->m_ext.ext_size += PAGE_SIZE * mb->m_epg_npgs;
sys/kern/uipc_mbuf.c
1996
struct mbuf *mb;
sys/kern/uipc_mbuf.c
2027
STAILQ_FOREACH(mb, &mc->mc_q, m_stailq) {
sys/kern/uipc_mbuf.c
2030
mlen = min(M_TRAILINGSPACE(mb), total - mc->mc_len);
sys/kern/uipc_mbuf.c
2031
error = uiomove(mtod(mb, void *), mlen, uio);
sys/kern/uipc_mbuf.c
2037
mb->m_len = mlen;
sys/kern/uipc_sockbuf.c
1727
sbsndptr_adv(struct sockbuf *sb, struct mbuf *mb, uint32_t len)
sys/kern/uipc_sockbuf.c
1735
if (mb != sb->sb_sndptr) {
sys/kern/uipc_sockbuf.c
1739
m = mb;
sys/kern/uipc_usrreq.c
2077
for (struct mbuf *mb = f; mb != NULL; mb = mb->m_next) {
sys/kern/uipc_usrreq.c
2078
if (mb->m_type == MT_DATA)
sys/kern/uipc_usrreq.c
2079
dcc += mb->m_len;
sys/kern/uipc_usrreq.c
2081
dctl += mb->m_len;
sys/kern/uipc_usrreq.c
2083
if (mb->m_flags & M_EXT)
sys/kern/uipc_usrreq.c
2084
dmbcnt += mb->m_ext.ext_size;
sys/kern/uipc_usrreq.c
2557
struct mbuf *mb;
sys/kern/uipc_usrreq.c
2565
mb = NULL;
sys/kern/uipc_usrreq.c
2569
STAILQ_FOREACH(mb, &sb->uxst_mbq, m_stailq) {
sys/kern/uipc_usrreq.c
2570
if (mb == m) {
sys/kern/uipc_usrreq.c
2578
return (mb != NULL);
sys/net/bpf.c
2303
struct mbuf mb;
sys/net/bpf.c
2320
mb.m_flags = 0;
sys/net/bpf.c
2321
mb.m_next = m;
sys/net/bpf.c
2322
mb.m_data = data;
sys/net/bpf.c
2323
mb.m_len = dlen;
sys/net/bpf.c
2333
slen = bpf_filter(d->bd_rfilter, (u_char *)&mb, pktlen, 0);
sys/net/bpf.c
2343
catchpacket(d, (u_char *)&mb, pktlen, slen,
sys/net/debugnet.c
378
debugnet_handle_rx_msg(struct debugnet_pcb *pcb, struct mbuf **mb)
sys/net/debugnet.c
386
m = *mb;
sys/net/debugnet.c
396
*mb = m;
sys/net/debugnet.c
445
debugnet_handle_ack(struct debugnet_pcb *pcb, struct mbuf **mb, uint16_t sport)
sys/net/debugnet.c
451
m = *mb;
sys/net/debugnet.c
456
*mb = m;
sys/net/debugnet.c
483
debugnet_handle_udp(struct debugnet_pcb *pcb, struct mbuf **mb)
sys/net/debugnet.c
491
m = *mb;
sys/net/debugnet.c
500
*mb = m;
sys/net/debugnet.c
527
debugnet_handle_ack(pcb, mb, sport);
sys/net/debugnet.c
540
debugnet_handle_rx_msg(pcb, mb);
sys/net/debugnet_inet.c
120
*mb = m;
sys/net/debugnet_inet.c
216
debugnet_handle_udp(pcb, mb);
sys/net/debugnet_inet.c
277
debugnet_handle_arp(struct debugnet_pcb *pcb, struct mbuf **mb)
sys/net/debugnet_inet.c
288
m = *mb;
sys/net/debugnet_inet.c
292
*mb = m;
sys/net/debugnet_inet.c
313
*mb = m;
sys/net/debugnet_inet.c
390
*mb = NULL;
sys/net/debugnet_inet.c
81
debugnet_handle_ip(struct debugnet_pcb *pcb, struct mbuf **mb)
sys/net/debugnet_inet.c
91
m = *mb;
sys/net/debugnet_inet.c
98
*mb = m;
sys/net/if_ethersubr.c
1255
struct mbuf mv, mb;
sys/net/if_ethersubr.c
1281
mb.m_next = &mv;
sys/net/if_ethersubr.c
1282
mb.m_data = data;
sys/net/if_ethersubr.c
1283
mb.m_len = dlen;
sys/net/if_ethersubr.c
1284
bpf_mtap(bp, &mb);
sys/net/if_infiniband.c
127
infiniband_bpf_mtap(struct ifnet *ifp, struct mbuf *mb)
sys/net/if_infiniband.c
135
M_ASSERTVALID(mb);
sys/net/if_infiniband.c
136
if (mb->m_len < sizeof(*ibh))
sys/net/if_infiniband.c
139
ibh = mtod(mb, struct infiniband_header *);
sys/net/if_infiniband.c
143
mb->m_data += sizeof(*ibh);
sys/net/if_infiniband.c
144
mb->m_len -= sizeof(*ibh);
sys/net/if_infiniband.c
145
mb->m_pkthdr.len -= sizeof(*ibh);
sys/net/if_infiniband.c
146
bpf_mtap2(ifp->if_bpf, &eh, sizeof(eh), mb);
sys/net/if_infiniband.c
147
mb->m_data -= sizeof(*ibh);
sys/net/if_infiniband.c
148
mb->m_len += sizeof(*ibh);
sys/net/if_infiniband.c
149
mb->m_pkthdr.len += sizeof(*ibh);
sys/net/if_infiniband.c
164
struct mbuf *mb, const struct route *ro, struct infiniband_header *ibh,
sys/net/if_infiniband.c
182
mb->m_flags &= ~M_MCAST;
sys/net/if_infiniband.c
183
mb->m_flags |= M_BCAST;
sys/net/if_tuntap.c
1790
struct mbuf *mb;
sys/net/if_tuntap.c
1792
IFQ_POLL_NOLOCK(&TUN2IFP(tp)->if_snd, mb);
sys/net/if_tuntap.c
1793
for (*(int *)data = 0; mb != NULL; mb = mb->m_next)
sys/net/if_tuntap.c
1794
*(int *)data += mb->m_len;
sys/netgraph/ng_tty.c
409
struct mbuf *m, *mb;
sys/netgraph/ng_tty.c
427
for (mb = m; mb != NULL; mb = mb->m_next) {
sys/netgraph/ng_tty.c
428
length = min(M_TRAILINGSPACE(mb), len - total);
sys/netgraph/ng_tty.c
430
memcpy(mtod(mb, char *), (const char *)buf + total, length);
sys/netgraph/ng_tty.c
431
mb->m_len = length;
sys/netinet/in_pcb.c
3276
in_pcboutput_txrtlmt_locked(struct inpcb *inp, struct ifnet *ifp, struct mbuf *mb, uint32_t max_pacing_rate)
sys/netinet/in_pcb.c
3310
if (M_HASHTYPE_GET(mb) == M_HASHTYPE_NONE) {
sys/netinet/in_pcb.c
3313
error = in_pcbattach_txrtlmt(inp, ifp, M_HASHTYPE_GET(mb),
sys/netinet/in_pcb.c
3314
mb->m_pkthdr.flowid, max_pacing_rate, &inp->inp_snd_tag);
sys/netinet/in_pcb.c
3331
in_pcboutput_txrtlmt(struct inpcb *inp, struct ifnet *ifp, struct mbuf *mb)
sys/netinet/in_pcb.c
3366
in_pcboutput_txrtlmt_locked(inp, ifp, mb, max_pacing_rate);
sys/netinet/sctp_uio.h
927
struct sctp_mbuf_log mb;
sys/netinet/sctputil.c
248
sctp_clog.x.mb.mp = m;
sys/netinet/sctputil.c
249
sctp_clog.x.mb.mbuf_flags = (uint8_t)(SCTP_BUF_GET_FLAGS(m));
sys/netinet/sctputil.c
250
sctp_clog.x.mb.size = (uint16_t)(SCTP_BUF_LEN(m));
sys/netinet/sctputil.c
251
sctp_clog.x.mb.data = SCTP_BUF_AT(m, 0);
sys/netinet/sctputil.c
253
sctp_clog.x.mb.ext = SCTP_BUF_EXTEND_BASE(m);
sys/netinet/sctputil.c
254
sctp_clog.x.mb.refcnt = (uint8_t)(SCTP_BUF_EXTEND_REFCNT(m));
sys/netinet/sctputil.c
256
sctp_clog.x.mb.ext = 0;
sys/netinet/sctputil.c
257
sctp_clog.x.mb.refcnt = 0;
sys/netinet/sctputil.c
6486
sctp_m_freem(struct mbuf *mb)
sys/netinet/sctputil.c
6488
while (mb != NULL)
sys/netinet/sctputil.c
6489
mb = sctp_m_free(mb);
sys/netinet/tcp_lro.c
1231
struct mbuf *mb;
sys/netinet/tcp_lro.c
1234
mb = lc->lro_mbuf_data[x].mb;
sys/netinet/tcp_lro.c
1248
if (tcp_lro_rx_common(lc, mb, 0, false) != 0) {
sys/netinet/tcp_lro.c
1252
(*lc->ifp->if_input)(lc->ifp, mb);
sys/netinet/tcp_lro.c
1450
tcp_lro_queue_mbuf(struct lro_ctrl *lc, struct mbuf *mb)
sys/netinet/tcp_lro.c
1457
m_freem(mb);
sys/netinet/tcp_lro.c
1464
(*lc->ifp->if_input) (lc->ifp, mb);
sys/netinet/tcp_lro.c
1471
((mb->m_flags & M_TSTMP) == 0)) {
sys/netinet/tcp_lro.c
1474
mb->m_pkthdr.rcv_tstmp = bintime2ns(&lc->lro_last_queue_time);
sys/netinet/tcp_lro.c
1475
mb->m_flags |= M_TSTMP_LRO;
sys/netinet/tcp_lro.c
1480
if (M_HASHTYPE_ISHASH(mb))
sys/netinet/tcp_lro.c
1482
(((uint64_t)M_HASHTYPE_GET(mb)) << 56) |
sys/netinet/tcp_lro.c
1483
(((uint64_t)mb->m_pkthdr.flowid) << 24);
sys/netinet/tcp_lro.c
1486
lc->lro_mbuf_data[lc->lro_mbuf_count].mb = mb;
sys/netinet/tcp_lro.c
513
m_freem(lc->lro_mbuf_data[x].mb);
sys/netinet/tcp_lro.h
157
struct mbuf *mb;
sys/netinet/tcp_output.c
1019
struct mbuf *mb;
sys/netinet/tcp_output.c
1077
mb = sbsndptr_noadv(&so->so_snd, off, &moff);
sys/netinet/tcp_output.c
1079
m_copydata(mb, moff, len,
sys/netinet/tcp_output.c
1082
sbsndptr_adv(&so->so_snd, mb, len);
sys/netinet/tcp_output.c
1092
m->m_next = tcp_m_copym(mb, moff,
sys/netinet/tcp_stacks/bbr.c
11819
struct mbuf *mb;
sys/netinet/tcp_stacks/bbr.c
13133
mb = sbsndptr_noadv(sb, sb_offset, &moff);
sys/netinet/tcp_stacks/bbr.c
13135
m_copydata(mb, moff, (int)len,
sys/netinet/tcp_stacks/bbr.c
13138
sbsndptr_adv(sb, mb, len);
sys/netinet/tcp_stacks/bbr.c
13164
mb, moff, &len,
sys/netinet/tcp_stacks/bbr.c
13657
cts, mb, &abandon, rsm, 0, sb);
sys/netinet/tcp_stacks/bbr.c
5866
struct mbuf *mb, int32_t * abandon, struct bbr_sendmap *hintrsm, uint32_t delay_calc,
sys/netinet/tcp_stacks/rack.c
19553
struct mbuf *mb;
sys/netinet/tcp_stacks/rack.c
21405
mb = sbsndptr_noadv(sb, sb_offset, &moff);
sys/netinet/tcp_stacks/rack.c
21406
s_mb = mb;
sys/netinet/tcp_stacks/rack.c
21409
m_copydata(mb, moff, (int)len,
sys/netinet/tcp_stacks/rack.c
21417
sbsndptr_adv(sb, mb, len);
sys/netinet/tcp_stacks/rack.c
21434
mb, moff, &len,
sys/netpfil/ipfw/test/main.c
168
gnet_stats_enq(struct cfg_s *c, struct mbuf *mb)
sys/netpfil/ipfw/test/main.c
171
struct dn_queue *_q = FI2Q(c, mb->flow_id);
sys/netpfil/ipfw/test/main.c
180
gnet_stats_deq(struct cfg_s *c, struct mbuf *mb)
sys/netpfil/ipfw/test/main.c
183
struct dn_queue *_q = FI2Q(c, mb->flow_id);
sys/netpfil/ipfw/test/main.c
184
int len = mb->m_pkthdr.len;
sys/netpfil/ipfw/test/main.c
195
if (c->q_wfi[mb->flow_id] < wfi)
sys/netpfil/ipfw/test/main.c
196
c->q_wfi[mb->flow_id] = wfi;
sys/netsmb/smb_crypt.c
204
struct mbuf *mb;
sys/netsmb/smb_crypt.c
251
for (mb = mbp->mb_top; mb != NULL; mb = mb->m_next)
sys/netsmb/smb_crypt.c
252
MD5Update(&md5, mtod(mb, void *), mb->m_len);
sys/netsmb/smb_crypt.c
270
struct mbuf *mb;
sys/netsmb/smb_crypt.c
286
mb = mdp->md_top;
sys/netsmb/smb_crypt.c
287
KASSERT(mb->m_len >= SMB_HDRLEN, ("forgot to m_pullup"));
sys/netsmb/smb_crypt.c
290
MD5Update(&md5, mtod(mb, void *), 14);
sys/netsmb/smb_crypt.c
294
MD5Update(&md5, mtod(mb, u_char *) + 22, mb->m_len - 22);
sys/netsmb/smb_crypt.c
295
for (mb = mb->m_next; mb != NULL; mb = mb->m_next)
sys/netsmb/smb_crypt.c
296
MD5Update(&md5, mtod(mb, void *), mb->m_len);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib.h
201
struct mbuf *mb;
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib.h
206
struct mbuf *mb;
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib.h
211
struct mbuf *mb;
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib.h
216
struct mbuf *mb;
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib.h
454
void ipoib_send(struct ipoib_dev_priv *priv, struct mbuf *mb,
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib.h
480
void ipoib_mcast_send(struct ipoib_dev_priv *priv, void *mgid, struct mbuf *mb);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib.h
528
void ipoib_dma_mb(struct ipoib_dev_priv *priv, struct mbuf *mb, unsigned int length);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib.h
581
void ipoib_cm_send(struct ipoib_dev_priv *priv, struct mbuf *mb, struct ipoib_cm_tx *tx);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib.h
590
void ipoib_cm_mb_too_long(struct ipoib_dev_priv *priv, struct mbuf *mb,
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib.h
636
void ipoib_cm_send(struct ipoib_dev_priv *priv, struct mbuf *mb, struct ipoib_cm_tx *tx)
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib.h
683
static inline void ipoib_cm_mb_too_long(struct ipoib_dev_priv *priv, struct mbuf *mb,
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib.h
686
m_freem(mb);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
100
for (m = rx_req->mb, i = 0; m != NULL; m = m->m_next, i++) {
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
1097
m_freem(tx_req->mb);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
112
m_freem(priv->cm.srq_ring[id].mb);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
113
priv->cm.srq_ring[id].mb = NULL;
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
1264
struct mbuf *mb;
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
1276
IF_DEQUEUE(&priv->cm.mb_queue, mb);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
1277
if (mb == NULL)
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
1281
proto = htons(*mtod(mb, uint16_t *));
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
1282
m_adj(mb, IPOIB_ENCAP_LEN);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
1286
icmp_error(mb, ICMP_UNREACH, ICMP_UNREACH_NEEDFRAG, 0, mtu);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
1291
icmp6_error(mb, ICMP6_PACKET_TOO_BIG, 0, mtu);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
1295
m_freem(mb);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
1307
ipoib_cm_mb_too_long(struct ipoib_dev_priv *priv, struct mbuf *mb, unsigned int mtu)
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
131
for (m = rx_req->mb, i = 0; m != NULL; m = m->m_next, i++) {
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
1311
IF_ENQUEUE(&priv->cm.mb_queue, mb);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
143
m_freem(rx->rx_ring[id].mb);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
144
rx->rx_ring[id].mb = NULL;
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
163
if (rx_ring[i].mb) {
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
165
m_freem(rx_ring[i].mb);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
482
struct mbuf *mb, *newmb;
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
509
mb = rx_ring[wr_id].mb;
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
557
ipoib_dma_mb(priv, mb, wc->byte_len);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
559
mb->m_pkthdr.rcvif = dev;
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
561
M_PREPEND(mb, sizeof(struct ipoib_pseudoheader), M_NOWAIT);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
562
if (likely(mb != NULL)) {
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
566
if_inc_counter(dev, IFCOUNTER_IBYTES, mb->m_pkthdr.len);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
569
ibh = mtod(mb, struct ipoib_header *);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
573
if_input(dev, mb);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
602
struct mbuf *mb = tx_req->mb;
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
607
for (m = mb, i = 0; m != NULL; m = m->m_next, i++) {
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
618
void ipoib_cm_send(struct ipoib_dev_priv *priv, struct mbuf *mb, struct ipoib_cm_tx *tx)
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
628
m_adj(mb, sizeof(struct ipoib_pseudoheader));
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
629
if (unlikely(mb->m_pkthdr.len > tx->mtu)) {
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
631
mb->m_pkthdr.len, tx->mtu);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
633
ipoib_cm_mb_too_long(priv, mb, IPOIB_CM_MTU(tx->mtu));
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
638
tx->tx_head, mb->m_pkthdr.len, tx->qp->qp_num);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
649
tx_req->mb = mb;
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
653
if (tx_req->mb)
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
654
m_freem(tx_req->mb);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
662
m_freem(mb);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
700
m_freem(tx_req->mb);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
854
struct mbuf *mb;
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
897
_IF_DEQUEUE(&p->path->queue, mb);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
898
if (mb == NULL)
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
900
_IF_ENQUEUE(&mbqueue, mb);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
907
_IF_DEQUEUE(&mbqueue, mb);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
908
if (mb == NULL)
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
910
mb->m_pkthdr.rcvif = dev;
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_cm.c
911
if (if_transmit(dev, mb))
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
105
ipoib_dma_mb(struct ipoib_dev_priv *priv, struct mbuf *mb, unsigned int length)
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
108
m_adj(mb, -(mb->m_pkthdr.len - length));
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
115
struct mbuf *mb, *m;
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
118
rx_req->mb = NULL;
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
119
mb = m_getm2(NULL, align + size, M_NOWAIT, MT_DATA, M_PKTHDR);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
120
if (mb == NULL)
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
122
for (i = 0, m = mb; m != NULL; m = m->m_next, i++) {
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
128
mb->m_pkthdr.len += m->m_len;
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
136
rx_req->mb = mb;
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
137
return (mb);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
139
for (j = 0, m = mb; j < i; m = m->m_next, j++)
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
142
m_freem(mb);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
156
for (m = rx_req->mb, i = 0; m != NULL; m = m->m_next, i++) {
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
167
m_freem(priv->rx_ring[id].mb);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
168
priv->rx_ring[id].mb = NULL;
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
206
struct mbuf *mb;
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
217
mb = priv->rx_ring[wr_id].mb;
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
226
if (mb) {
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
228
m_freem(mb);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
229
priv->rx_ring[wr_id].mb = NULL;
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
256
ipoib_dma_mb(priv, mb, wc->byte_len);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
259
if_inc_counter(dev, IFCOUNTER_IBYTES, mb->m_pkthdr.len);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
260
mb->m_pkthdr.rcvif = dev;
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
261
m_adj(mb, sizeof(struct ib_grh) - INFINIBAND_ALEN);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
262
eh = mtod(mb, struct ipoib_header *);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
266
mb->m_pkthdr.csum_flags = CSUM_IP_CHECKED | CSUM_IP_VALID;
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
268
if_input(dev, mb);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
278
struct mbuf *mb = tx_req->mb;
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
284
for (m = mb, p = NULL, i = 0; m != NULL; p = m, m = m->m_next, i++) {
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
295
tx_req->mb = mb = m_defrag(mb, M_NOWAIT);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
296
if (mb == NULL)
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
298
for (m = mb, i = 0; m != NULL; m = m->m_next, i++);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
303
for (m = mb, i = 0; m != NULL; m = m->m_next, i++) {
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
315
for (m = mb, i = 0; i < end; m = m->m_next, i++)
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
324
struct mbuf *mb = tx_req->mb;
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
329
for (m = mb, i = 0; m != NULL; m = m->m_next, i++)
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
354
m_freem(tx_req->mb);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
454
struct mbuf *mb = tx_req->mb;
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
459
for (m = mb, i = 0; m != NULL; m = m->m_next, i++) {
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
480
ipoib_send(struct ipoib_dev_priv *priv, struct mbuf *mb,
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
492
m_adj(mb, sizeof (struct ipoib_pseudoheader));
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
495
phead = mtod(mb, void *);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
496
if (mb->m_len < hlen) {
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
499
m_freem(mb);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
502
m_adj(mb, hlen);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
504
if (unlikely(mb->m_pkthdr.len - IPOIB_ENCAP_LEN > priv->mcast_mtu)) {
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
506
mb->m_pkthdr.len, priv->mcast_mtu);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
508
ipoib_cm_mb_too_long(priv, mb, priv->mcast_mtu);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
516
mb->m_pkthdr.len, address, qpn);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
526
tx_req->mb = mb;
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
529
if (tx_req->mb)
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
530
m_freem(tx_req->mb);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
534
if (mb->m_pkthdr.csum_flags & (CSUM_IP|CSUM_TCP|CSUM_UDP))
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
553
m_freem(mb);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
711
if (priv->rx_ring[i].mb)
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
810
m_freem(tx_req->mb);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
819
if (!rx_req->mb)
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
822
m_freem(rx_req->mb);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
823
rx_req->mb = NULL;
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_ib.c
99
for (i = 0, m = rx_req->mb; m != NULL; m = m->m_next, i++)
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_main.c
517
struct mbuf *mb;
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_main.c
548
_IF_DEQUEUE(&path->queue, mb);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_main.c
549
if (mb == NULL)
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_main.c
551
_IF_ENQUEUE(&mbqueue, mb);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_main.c
572
_IF_DEQUEUE(&mbqueue, mb);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_main.c
573
if (mb == NULL)
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_main.c
575
mb->m_pkthdr.rcvif = dev;
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_main.c
576
if (if_transmit(dev, mb))
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_main.c
670
ipoib_unicast_send(struct mbuf *mb, struct ipoib_dev_priv *priv, struct ipoib_header *eh)
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_main.c
684
_IF_ENQUEUE(&path->queue, mb);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_main.c
687
m_freem(mb);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_main.c
698
m_freem(mb);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_main.c
705
ipoib_cm_send(priv, mb, ipoib_cm_get(path));
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_main.c
707
ipoib_send(priv, mb, path->ah, IPOIB_QPN(eh->hwaddr));
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_main.c
710
_IF_ENQUEUE(&path->queue, mb);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_main.c
713
m_freem(mb);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_main.c
718
ipoib_send_one(struct ipoib_dev_priv *priv, struct mbuf *mb)
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_main.c
722
eh = mtod(mb, struct ipoib_header *);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_main.c
728
ipoib_mcast_send(priv, eh->hwaddr + 4, mb);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_main.c
730
ipoib_unicast_send(mb, priv, eh);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_main.c
738
struct mbuf *mb;
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_main.c
744
mb = if_dequeue(dev);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_main.c
745
if (mb == NULL)
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_main.c
747
infiniband_bpf_mtap(dev, mb);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_main.c
748
ipoib_send_one(priv, mb);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_main.c
775
struct mbuf *mb;
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_main.c
781
mb = if_dequeue(dev);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_main.c
782
if (mb == NULL)
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_main.c
784
m_freem(mb);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_multicast.c
233
struct mbuf *mb;
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_multicast.c
234
_IF_DEQUEUE(&mcast->pkt_queue, mb);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_multicast.c
235
mb->m_pkthdr.rcvif = dev;
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_multicast.c
237
if (if_transmit(dev, mb))
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_multicast.c
628
ipoib_mcast_send(struct ipoib_dev_priv *priv, void *mgid, struct mbuf *mb)
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_multicast.c
637
m_freem(mb);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_multicast.c
652
m_freem(mb);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_multicast.c
664
_IF_ENQUEUE(&mcast->pkt_queue, mb);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_multicast.c
667
m_freem(mb);
sys/ofed/drivers/infiniband/ulp/ipoib/ipoib_multicast.c
685
ipoib_send(priv, mb, mcast->ah, IB_MULTICAST_QPN);
sys/ofed/drivers/infiniband/ulp/sdp/sdp.h
136
#define BZCOPY_STATE(mb) (SDP_SKB_CB(mb)->bz)
sys/ofed/drivers/infiniband/ulp/sdp/sdp.h
137
#define RX_SRCAVAIL_STATE(mb) (SDP_SKB_CB(mb)->rx_sa)
sys/ofed/drivers/infiniband/ulp/sdp/sdp.h
138
#define TX_SRCAVAIL_STATE(mb) (SDP_SKB_CB(mb)->tx_sa)
sys/ofed/drivers/infiniband/ulp/sdp/sdp.h
203
struct mbuf *mb;
sys/ofed/drivers/infiniband/ulp/sdp/sdp.h
507
struct mbuf *mb;
sys/ofed/drivers/infiniband/ulp/sdp/sdp.h
509
MGETHDR(mb, wait, MT_DATA);
sys/ofed/drivers/infiniband/ulp/sdp/sdp.h
510
if (mb == NULL)
sys/ofed/drivers/infiniband/ulp/sdp/sdp.h
512
mb->m_pkthdr.len = mb->m_len = sizeof(struct sdp_bsdh);
sys/ofed/drivers/infiniband/ulp/sdp/sdp.h
513
h = mtod(mb, struct sdp_bsdh *);
sys/ofed/drivers/infiniband/ulp/sdp/sdp.h
516
return mb;
sys/ofed/drivers/infiniband/ulp/sdp/sdp.h
531
mb_put(struct mbuf *mb, int len)
sys/ofed/drivers/infiniband/ulp/sdp/sdp.h
535
data = mb->m_data;
sys/ofed/drivers/infiniband/ulp/sdp/sdp.h
536
data += mb->m_len;
sys/ofed/drivers/infiniband/ulp/sdp/sdp.h
537
mb->m_len += len;
sys/ofed/drivers/infiniband/ulp/sdp/sdp.h
544
struct mbuf *mb;
sys/ofed/drivers/infiniband/ulp/sdp/sdp.h
547
mb = sdp_alloc_mb(sk, SDP_MID_CHRCVBUF_ACK, sizeof(*resp_size), wait);
sys/ofed/drivers/infiniband/ulp/sdp/sdp.h
548
if (mb == NULL)
sys/ofed/drivers/infiniband/ulp/sdp/sdp.h
550
resp_size = (struct sdp_chrecvbuf *)mb_put(mb, sizeof *resp_size);
sys/ofed/drivers/infiniband/ulp/sdp/sdp.h
553
return mb;
sys/ofed/drivers/infiniband/ulp/sdp/sdp.h
559
struct mbuf *mb;
sys/ofed/drivers/infiniband/ulp/sdp/sdp.h
562
mb = sdp_alloc_mb(sk, SDP_MID_SRCAVAIL, sizeof(*srcah), wait);
sys/ofed/drivers/infiniband/ulp/sdp/sdp.h
563
if (mb == NULL)
sys/ofed/drivers/infiniband/ulp/sdp/sdp.h
565
srcah = (struct sdp_srcah *)mb_put(mb, sizeof(*srcah));
sys/ofed/drivers/infiniband/ulp/sdp/sdp.h
570
return mb;
sys/ofed/drivers/infiniband/ulp/sdp/sdp.h
582
struct mbuf *mb;
sys/ofed/drivers/infiniband/ulp/sdp/sdp.h
585
mb = sdp_alloc_mb(sk, SDP_MID_RDMARDCOMPL, sizeof(*rrch), wait);
sys/ofed/drivers/infiniband/ulp/sdp/sdp.h
586
if (mb == NULL)
sys/ofed/drivers/infiniband/ulp/sdp/sdp.h
588
rrch = (struct sdp_rrch *)mb_put(mb, sizeof(*rrch));
sys/ofed/drivers/infiniband/ulp/sdp/sdp.h
591
return mb;
sys/ofed/drivers/infiniband/ulp/sdp/sdp.h
628
struct mbuf *mb;
sys/ofed/drivers/infiniband/ulp/sdp/sdp.h
632
for (i = 0, mb = sbuf->mb; mb != NULL; mb = mb->m_next, i++)
sys/ofed/drivers/infiniband/ulp/sdp/sdp.h
633
ib_dma_unmap_single(dev, sbuf->mapping[i], mb->m_len, dir);
sys/ofed/drivers/infiniband/ulp/sdp/sdp.h
639
void sdp_urg(struct sdp_sock *ssk, struct mbuf *mb);
sys/ofed/drivers/infiniband/ulp/sdp/sdp.h
652
void sdp_post_send(struct sdp_sock *ssk, struct mbuf *mb);
sys/ofed/drivers/infiniband/ulp/sdp/sdp.h
673
int sdp_rdma_to_iovec(struct socket *sk, struct iovec *iov, struct mbuf *mb,
sys/ofed/drivers/infiniband/ulp/sdp/sdp_bcopy.c
107
sdp_nagle_off(struct sdp_sock *ssk, struct mbuf *mb)
sys/ofed/drivers/infiniband/ulp/sdp/sdp_bcopy.c
112
h = mtod(mb, struct sdp_bsdh *);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_bcopy.c
115
BZCOPY_STATE(mb) ||
sys/ofed/drivers/infiniband/ulp/sdp/sdp_bcopy.c
120
mb->m_pkthdr.len >= ssk->xmit_size_goal / 4 ||
sys/ofed/drivers/infiniband/ulp/sdp/sdp_bcopy.c
121
(mb->m_flags & M_PUSH);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_bcopy.c
168
struct mbuf *mb;
sys/ofed/drivers/infiniband/ulp/sdp/sdp_bcopy.c
190
mb = sdp_alloc_mb_chrcvbuf_ack(sk,
sys/ofed/drivers/infiniband/ulp/sdp/sdp_bcopy.c
192
if (mb == NULL)
sys/ofed/drivers/infiniband/ulp/sdp/sdp_bcopy.c
195
sdp_post_send(ssk, mb);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_bcopy.c
206
sdp_tx_ring_slots_left(ssk) && (mb = sk->so_snd.sb_sndptr) &&
sys/ofed/drivers/infiniband/ulp/sdp/sdp_bcopy.c
207
sdp_nagle_off(ssk, mb)) {
sys/ofed/drivers/infiniband/ulp/sdp/sdp_bcopy.c
211
sk->so_snd.sb_sndptr = mb->m_nextpkt;
sys/ofed/drivers/infiniband/ulp/sdp/sdp_bcopy.c
212
sk->so_snd.sb_mb = mb->m_nextpkt;
sys/ofed/drivers/infiniband/ulp/sdp/sdp_bcopy.c
213
mb->m_nextpkt = NULL;
sys/ofed/drivers/infiniband/ulp/sdp/sdp_bcopy.c
215
for (n = mb; n != NULL; n = n->m_next)
sys/ofed/drivers/infiniband/ulp/sdp/sdp_bcopy.c
218
sdp_post_send(ssk, mb);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_bcopy.c
224
mb = sdp_alloc_mb_data(ssk->socket, wait);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_bcopy.c
225
if (mb == NULL)
sys/ofed/drivers/infiniband/ulp/sdp/sdp_bcopy.c
227
sdp_post_send(ssk, mb);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_bcopy.c
240
mb = sdp_alloc_mb_disconnect(sk, wait);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_bcopy.c
241
if (mb == NULL)
sys/ofed/drivers/infiniband/ulp/sdp/sdp_bcopy.c
244
sdp_post_send(ssk, mb);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_bcopy.c
42
struct mbuf *mb, const struct sdp_bsdh *h)
sys/ofed/drivers/infiniband/ulp/sdp/sdp_bcopy.c
53
str, mb, h->mid, mid2str(h->mid), h->flags,
sys/ofed/drivers/infiniband/ulp/sdp/sdp_dbg.h
112
#define SDP_DUMP_PACKET(sk, str, mb, h) \
sys/ofed/drivers/infiniband/ulp/sdp/sdp_dbg.h
115
dump_packet(sk, str, mb, h); \
sys/ofed/drivers/infiniband/ulp/sdp/sdp_dbg.h
119
#define SDP_DUMP_PACKET(sk, str, mb, h)
sys/ofed/drivers/infiniband/ulp/sdp/sdp_dbg.h
162
struct mbuf *mb, const struct sdp_bsdh *h);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_dbg.h
163
#define dump_packet(sk, str, mb, h) \
sys/ofed/drivers/infiniband/ulp/sdp/sdp_dbg.h
164
_dump_packet(__func__, __LINE__, sk, str, mb, h)
sys/ofed/drivers/infiniband/ulp/sdp/sdp_dbg.h
39
struct mbuf *mb;
sys/ofed/drivers/infiniband/ulp/sdp/sdp_dbg.h
73
l->mb = s; \
sys/ofed/drivers/infiniband/ulp/sdp/sdp_main.c
1501
sdp_urg(struct sdp_sock *ssk, struct mbuf *mb)
sys/ofed/drivers/infiniband/ulp/sdp/sdp_main.c
1510
so->so_oobmark = sbused(&so->so_rcv) + mb->m_pkthdr.len - 1;
sys/ofed/drivers/infiniband/ulp/sdp/sdp_main.c
1514
for (m = mb; m->m_next != NULL; m = m->m_next);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_main.c
1518
mb->m_pkthdr.len--;
sys/ofed/drivers/infiniband/ulp/sdp/sdp_main.c
839
sdp_append(struct sdp_sock *ssk, struct sockbuf *sb, struct mbuf *mb, int cnt)
sys/ofed/drivers/infiniband/ulp/sdp/sdp_main.c
846
KASSERT(mb->m_flags & M_PKTHDR,
sys/ofed/drivers/infiniband/ulp/sdp/sdp_main.c
847
("sdp_append: %p Missing packet header.\n", mb));
sys/ofed/drivers/infiniband/ulp/sdp/sdp_main.c
853
sb->sb_lastrecord = sb->sb_mb = sb->sb_sndptr = mb;
sys/ofed/drivers/infiniband/ulp/sdp/sdp_main.c
854
for (; mb; mb = mb->m_next) {
sys/ofed/drivers/infiniband/ulp/sdp/sdp_main.c
855
sb->sb_mbtail = mb;
sys/ofed/drivers/infiniband/ulp/sdp/sdp_main.c
856
sballoc(sb, mb);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_main.c
873
n->m_pkthdr.len + mb->m_pkthdr.len - SDP_HEAD_SIZE <
sys/ofed/drivers/infiniband/ulp/sdp/sdp_main.c
875
m_adj(mb, SDP_HEAD_SIZE);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_main.c
876
n->m_pkthdr.len += mb->m_pkthdr.len;
sys/ofed/drivers/infiniband/ulp/sdp/sdp_main.c
877
n->m_flags |= mb->m_flags & (M_PUSH | M_URG);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_main.c
878
m_demote(mb, 1, 0);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_main.c
879
sbcompress(sb, mb, sb->sb_mbtail);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_main.c
886
sb->sb_lastrecord->m_nextpkt = mb;
sys/ofed/drivers/infiniband/ulp/sdp/sdp_main.c
887
sb->sb_lastrecord = mb;
sys/ofed/drivers/infiniband/ulp/sdp/sdp_main.c
889
sb->sb_sndptr = mb;
sys/ofed/drivers/infiniband/ulp/sdp/sdp_main.c
890
for (; mb; mb = mb->m_next) {
sys/ofed/drivers/infiniband/ulp/sdp/sdp_main.c
891
sb->sb_mbtail = mb;
sys/ofed/drivers/infiniband/ulp/sdp/sdp_main.c
892
sballoc(sb, mb);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_proc.c
374
l->mb, l->func, l->line);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
107
struct mbuf *mb, *m;
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
112
sdp_prf(ssk->socket, mb, "Posting mb");
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
113
mb = m_getm2(NULL, ssk->recv_bytes, M_NOWAIT, MT_DATA, M_PKTHDR);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
114
if (mb == NULL) {
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
120
for (m = mb; m != NULL; m = m->m_next) {
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
122
mb->m_pkthdr.len += m->m_len;
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
124
h = mtod(mb, struct sdp_bsdh *);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
126
rx_req->mb = mb;
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
128
for (i = 0; mb != NULL; i++, mb = mb->m_next, sge++) {
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
129
addr = ib_dma_map_single(dev, mb->m_data, mb->m_len,
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
136
sge->length = mb->m_len;
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
149
m_freem(mb);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
202
sdp_sock_queue_rcv_mb(struct socket *sk, struct mbuf *mb)
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
207
h = mtod(mb, struct sdp_bsdh *);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
210
SDP_SKB_CB(mb)->seq = rcv_nxt(ssk);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
217
ssk->rx_sa = rx_sa = RX_SRCAVAIL_STATE(mb) = kzalloc(
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
234
atomic_add(mb->len, &ssk->rcv_nxt);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
240
atomic_add(mb->m_pkthdr.len, &ssk->rcv_nxt);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
243
m_adj(mb, SDP_HEAD_SIZE);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
246
sdp_urg(ssk, mb);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
247
sbappend_locked(&sk->so_rcv, mb, 0);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
249
return mb;
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
306
struct mbuf *mb;
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
316
mb = rx_req->mb;
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
321
return mb;
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
325
sdp_process_rx_ctl_mb(struct sdp_sock *ssk, struct mbuf *mb)
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
333
h = mtod(mb, struct sdp_bsdh *);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
393
m_freem(mb);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
397
sdp_process_rx_mb(struct sdp_sock *ssk, struct mbuf *mb)
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
404
h = mtod(mb, struct sdp_bsdh *);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
413
m_freem(mb);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
426
sdp_prf1(ssk->socket, mb, "RX %s +%d c:%d->%d mseq:%d ack:%d\n",
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
431
mb->m_pkthdr.len == SDP_HEAD_SIZE)) {
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
433
m_freem(mb);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
456
if (mbufq_enqueue(&ssk->rxctlq, mb) != 0)
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
457
m_freem(mb);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
462
mb = sdp_sock_queue_rcv_mb(sk, mb);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
472
struct mbuf *mb;
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
477
mb = sdp_recv_completion(ssk, wc->wr_id);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
478
if (unlikely(!mb))
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
490
m_freem(mb);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
499
m_freem(mb);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
503
m_adj(mb, -(mb->m_pkthdr.len - wc->byte_len));
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
504
h = mtod(mb, struct sdp_bsdh *);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
506
SDP_DUMP_PACKET(ssk->socket, "RX", mb, h);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
509
ssk->rx_bytes += mb->m_pkthdr.len;
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
517
return mb;
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
538
struct mbuf *mb;
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
546
mb = sdp_process_rx_wc(ssk, wc);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
547
if (!mb)
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
550
sdp_process_rx_mb(ssk, mb);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
596
struct mbuf *mb;
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
604
while ((mb = mbufq_dequeue(&ssk->rxctlq)) != NULL)
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
605
sdp_process_rx_ctl_mb(ssk, mb);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
680
struct mbuf *mb;
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
681
mb = sdp_recv_completion(ssk, ring_tail(ssk->rx_ring));
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
682
if (!mb)
sys/ofed/drivers/infiniband/ulp/sdp/sdp_rx.c
684
m_freem(mb);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_tx.c
102
TX_SRCAVAIL_STATE(mb)->mseq = mseq;
sys/ofed/drivers/infiniband/ulp/sdp/sdp_tx.c
106
if (unlikely(mb->m_flags & M_URG))
sys/ofed/drivers/infiniband/ulp/sdp/sdp_tx.c
111
mb->m_flags |= M_RDONLY; /* Don't allow compression once sent. */
sys/ofed/drivers/infiniband/ulp/sdp/sdp_tx.c
113
h->len = htonl(mb->m_pkthdr.len);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_tx.c
117
sdp_prf1(ssk->socket, mb, "TX: %s bufs: %d mseq:%ld ack:%d",
sys/ofed/drivers/infiniband/ulp/sdp/sdp_tx.c
121
SDP_DUMP_PACKET(ssk->socket, "TX", mb, h);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_tx.c
124
tx_req->mb = mb;
sys/ofed/drivers/infiniband/ulp/sdp/sdp_tx.c
127
for (i = 0; mb != NULL; i++, mb = mb->m_next, sge++) {
sys/ofed/drivers/infiniband/ulp/sdp/sdp_tx.c
128
addr = ib_dma_map_single(dev, mb->m_data, mb->m_len,
sys/ofed/drivers/infiniband/ulp/sdp/sdp_tx.c
135
sge->length = mb->m_len;
sys/ofed/drivers/infiniband/ulp/sdp/sdp_tx.c
144
if (unlikely(tx_req->mb->m_flags & M_URG))
sys/ofed/drivers/infiniband/ulp/sdp/sdp_tx.c
155
m_freem(tx_req->mb);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_tx.c
171
struct mbuf *mb = NULL;
sys/ofed/drivers/infiniband/ulp/sdp/sdp_tx.c
182
mb = tx_req->mb;
sys/ofed/drivers/infiniband/ulp/sdp/sdp_tx.c
187
if (BZCOPY_STATE(mb))
sys/ofed/drivers/infiniband/ulp/sdp/sdp_tx.c
188
BZCOPY_STATE(mb)->busy--;
sys/ofed/drivers/infiniband/ulp/sdp/sdp_tx.c
194
return mb;
sys/ofed/drivers/infiniband/ulp/sdp/sdp_tx.c
200
struct mbuf *mb = NULL;
sys/ofed/drivers/infiniband/ulp/sdp/sdp_tx.c
205
sdp_prf(ssk->socket, mb, "Send completion with error. "
sys/ofed/drivers/infiniband/ulp/sdp/sdp_tx.c
213
mb = sdp_send_completion(ssk, wc->wr_id);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_tx.c
214
if (unlikely(!mb))
sys/ofed/drivers/infiniband/ulp/sdp/sdp_tx.c
217
h = mtod(mb, struct sdp_bsdh *);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_tx.c
218
sdp_prf1(ssk->socket, mb, "tx completion. mseq:%d", ntohl(h->mseq));
sys/ofed/drivers/infiniband/ulp/sdp/sdp_tx.c
220
mb, mb->m_pkthdr.len, ntohl(h->mseq));
sys/ofed/drivers/infiniband/ulp/sdp/sdp_tx.c
221
m_freem(mb);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_tx.c
381
struct mbuf *mb;
sys/ofed/drivers/infiniband/ulp/sdp/sdp_tx.c
382
mb = sdp_send_completion(ssk, ring_tail(ssk->tx_ring));
sys/ofed/drivers/infiniband/ulp/sdp/sdp_tx.c
383
if (!mb)
sys/ofed/drivers/infiniband/ulp/sdp/sdp_tx.c
385
m_freem(mb);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_tx.c
66
sdp_post_send(struct sdp_sock *ssk, struct mbuf *mb)
sys/ofed/drivers/infiniband/ulp/sdp/sdp_tx.c
80
SDPSTATS_HIST(send_size, mb->len);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_tx.c
83
m_freem(mb);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_tx.c
88
h = mtod(mb, struct sdp_bsdh *);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_tx.c
90
ssk->tx_bytes += mb->m_pkthdr.len;
sys/ofed/drivers/infiniband/ulp/sdp/sdp_tx.c
94
struct tx_srcavail_state *tx_sa = TX_SRCAVAIL_STATE(mb);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_tx.c
99
m_freem(mb);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_zcopy.c
103
mb_entail(sk, ssk, mb);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_zcopy.c
106
SDP_SKB_CB(mb)->end_seq += payload_len;
sys/ofed/drivers/infiniband/ulp/sdp/sdp_zcopy.c
119
struct mbuf *mb;
sys/ofed/drivers/infiniband/ulp/sdp/sdp_zcopy.c
123
mb = sdp_alloc_mb_srcavail_cancel(sk, 0);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_zcopy.c
124
mb_entail(sk, ssk, mb);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_zcopy.c
286
struct mbuf *mb;
sys/ofed/drivers/infiniband/ulp/sdp/sdp_zcopy.c
292
mb = sdp_alloc_mb_rdmardcompl(ssk->socket, copied, 0);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_zcopy.c
297
sdp_post_send(ssk, mb);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_zcopy.c
304
struct mbuf *mb = sdp_alloc_mb_sendsm(sk, 0);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_zcopy.c
306
sdp_post_send(sdp_sk(sk), mb);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_zcopy.c
54
struct mbuf *mb;
sys/ofed/drivers/infiniband/ulp/sdp/sdp_zcopy.c
549
int sdp_rdma_to_iovec(struct socket *sk, struct iovec *iov, struct mbuf *mb,
sys/ofed/drivers/infiniband/ulp/sdp/sdp_zcopy.c
553
struct rx_srcavail_state *rx_sa = RX_SRCAVAIL_STATE(mb);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_zcopy.c
584
sdp_prf(sk, mb, "Finished posting(rc=%d), now to wait", rc);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_zcopy.c
592
sdp_prf(sk, mb, "Finished waiting(rc=%d)", rc);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_zcopy.c
75
mb = sdp_alloc_mb_srcavail(sk, len, tx_sa->fmr->fmr->lkey, off, 0);
sys/ofed/drivers/infiniband/ulp/sdp/sdp_zcopy.c
76
if (!mb) {
sys/ofed/drivers/infiniband/ulp/sdp/sdp_zcopy.c
81
TX_SRCAVAIL_STATE(mb) = tx_sa; /* tx_sa is hanged on the mb
sys/ofed/drivers/infiniband/ulp/sdp/sdp_zcopy.c
94
mb_fill_page_desc(mb, mb_shinfo(mb)->nr_frags,
sys/ofed/drivers/infiniband/ulp/sdp/sdp_zcopy.c
97
mb->len += payload_len;
sys/ofed/drivers/infiniband/ulp/sdp/sdp_zcopy.c
98
mb->data_len = payload_len;
sys/ofed/drivers/infiniband/ulp/sdp/sdp_zcopy.c
99
mb->truesize += payload_len;
sys/powerpc/aim/mmu_oea.c
1065
vm_page_t *mb, vm_offset_t b_offset, int xfersize)
sys/powerpc/aim/mmu_oea.c
1078
b_cp = (char *)VM_PAGE_TO_PHYS(mb[b_offset >> PAGE_SHIFT]) +
sys/powerpc/aim/mmu_oea.c
281
vm_page_t *mb, vm_offset_t b_offset, int xfersize);
sys/powerpc/aim/mmu_oea64.c
1458
vm_page_t *mb, vm_offset_t b_offset, int xfersize)
sys/powerpc/aim/mmu_oea64.c
1473
VM_PAGE_TO_PHYS(mb[b_offset >> PAGE_SHIFT])) +
sys/powerpc/aim/mmu_oea64.c
1484
vm_page_t *mb, vm_offset_t b_offset, int xfersize)
sys/powerpc/aim/mmu_oea64.c
1500
VM_PAGE_TO_PHYS(mb[b_offset >> PAGE_SHIFT]));
sys/powerpc/aim/mmu_oea64.c
360
vm_page_t *mb, vm_offset_t b_offset, int xfersize);
sys/powerpc/aim/mmu_oea64.c
362
vm_page_t *mb, vm_offset_t b_offset, int xfersize);
sys/powerpc/aim/mmu_radix.c
2685
mmu_radix_copy_pages(vm_page_t ma[], vm_offset_t a_offset, vm_page_t mb[],
sys/powerpc/aim/mmu_radix.c
2693
a_offset, mb, b_offset, xfersize);
sys/powerpc/aim/mmu_radix.c
2704
VM_PAGE_TO_PHYS(mb[b_offset >> PAGE_SHIFT])) +
sys/powerpc/aim/mmu_radix.c
482
vm_page_t *mb, vm_offset_t b_offset, int xfersize);
sys/powerpc/booke/pmap.c
2678
mb();
sys/powerpc/booke/pmap_32.c
846
vm_page_t *mb, vm_offset_t b_offset, int xfersize)
sys/powerpc/booke/pmap_32.c
862
VM_PAGE_TO_PHYS(mb[b_offset >> PAGE_SHIFT]));
sys/powerpc/booke/pmap_64.c
706
vm_page_t *mb, vm_offset_t b_offset, int xfersize)
sys/powerpc/booke/pmap_64.c
718
pb = mb[b_offset >> PAGE_SHIFT];
sys/powerpc/mambo/mambo_disk.c
102
intmax_t mb;
sys/powerpc/mambo/mambo_disk.c
125
mb = d->d_mediasize >> 20; /* 1MiB == 1 << 20 */
sys/powerpc/mambo/mambo_disk.c
127
if (mb >= 10240) { /* 1GiB = 1024 MiB */
sys/powerpc/mambo/mambo_disk.c
129
mb /= 1024;
sys/powerpc/mambo/mambo_disk.c
131
device_printf(dev, "%ju%cB, %d byte sectors\n", mb, unit,
sys/powerpc/powermac/nvbl.c
134
mb();
sys/powerpc/powermac/pmu.c
1132
mb();
sys/powerpc/powernv/opal_pci.c
197
mb();
sys/powerpc/powernv/opal_pci.c
199
mb();
sys/powerpc/powerpc/cpu.c
849
mb();
sys/powerpc/powerpc/db_disasm.c
930
u_int mb, mbl, mbh;
sys/powerpc/powerpc/db_disasm.c
933
mb = mbh << 4 | mbl;
sys/powerpc/powerpc/db_disasm.c
934
APP_PSTR(", %d", mb);
sys/powerpc/powerpc/pmap_dispatch.c
168
(vm_page_t ma[], vm_offset_t a_offset, vm_page_t mb[],
sys/powerpc/ps3/ps3disk.c
159
intmax_t mb;
sys/powerpc/ps3/ps3disk.c
248
mb = d->d_mediasize >> 20;
sys/powerpc/ps3/ps3disk.c
250
if (mb >= 10240) {
sys/powerpc/ps3/ps3disk.c
252
mb /= 1024;
sys/powerpc/ps3/ps3disk.c
258
device_printf(dev, "region %d %ju%cB%s\n", i, mb, unit,
sys/powerpc/pseries/platform_chrp.c
172
mb();
sys/powerpc/pseries/xics.c
444
mb();
sys/riscv/riscv/pmap.c
4074
pmap_copy_pages(vm_page_t ma[], vm_offset_t a_offset, vm_page_t mb[],
sys/riscv/riscv/pmap.c
4088
m_b = mb[b_offset >> PAGE_SHIFT];
sys/riscv/riscv/sbi_ipi.c
110
mb();
sys/riscv/riscv/sbi_ipi.c
116
mb();
sys/riscv/sifive/sifive_ccache.c
105
mb();
sys/riscv/sifive/sifive_ccache.c
92
mb();
sys/riscv/vmm/vmm_fence.c
197
mb();
sys/rpc/rpcm_subs.h
81
{ if ((s) > M_TRAILINGSPACE(mb)) { \
sys/rpc/rpcm_subs.h
85
mb->m_next = mb2; \
sys/rpc/rpcm_subs.h
86
mb = mb2; \
sys/rpc/rpcm_subs.h
87
mb->m_len = 0; \
sys/rpc/rpcm_subs.h
88
bpos = mtod(mb, caddr_t); \
sys/rpc/rpcm_subs.h
91
mb->m_len += (s); \
sys/security/mac_biba/mac_biba.c
156
biba_free(struct mac_biba *mb)
sys/security/mac_biba/mac_biba.c
159
if (mb != NULL)
sys/security/mac_biba/mac_biba.c
160
uma_zfree(zone_biba, mb);
sys/security/mac_biba/mac_biba.c
166
biba_atmostflags(struct mac_biba *mb, int flags)
sys/security/mac_biba/mac_biba.c
169
if ((mb->mb_flags & flags) != mb->mb_flags)
sys/security/mac_biba/mac_biba.c
227
biba_subject_dominate_high(struct mac_biba *mb)
sys/security/mac_biba/mac_biba.c
231
KASSERT((mb->mb_flags & MAC_BIBA_FLAG_EFFECTIVE) != 0,
sys/security/mac_biba/mac_biba.c
233
element = &mb->mb_effective;
sys/security/mac_biba/mac_biba.c
301
biba_contains_equal(struct mac_biba *mb)
sys/security/mac_biba/mac_biba.c
304
if (mb->mb_flags & MAC_BIBA_FLAG_EFFECTIVE) {
sys/security/mac_biba/mac_biba.c
305
if (mb->mb_effective.mbe_type == MAC_BIBA_TYPE_EQUAL)
sys/security/mac_biba/mac_biba.c
309
if (mb->mb_flags & MAC_BIBA_FLAG_RANGE) {
sys/security/mac_biba/mac_biba.c
310
if (mb->mb_rangelow.mbe_type == MAC_BIBA_TYPE_EQUAL)
sys/security/mac_biba/mac_biba.c
312
if (mb->mb_rangehigh.mbe_type == MAC_BIBA_TYPE_EQUAL)
sys/security/mac_biba/mac_biba.c
320
biba_subject_privileged(struct mac_biba *mb)
sys/security/mac_biba/mac_biba.c
323
KASSERT((mb->mb_flags & MAC_BIBA_FLAGS_BOTH) == MAC_BIBA_FLAGS_BOTH,
sys/security/mac_biba/mac_biba.c
327
if (mb->mb_effective.mbe_type == MAC_BIBA_TYPE_EQUAL)
sys/security/mac_biba/mac_biba.c
331
if (mb->mb_rangelow.mbe_type == MAC_BIBA_TYPE_EQUAL ||
sys/security/mac_biba/mac_biba.c
332
mb->mb_rangehigh.mbe_type == MAC_BIBA_TYPE_EQUAL)
sys/security/mac_biba/mac_biba.c
336
if (mb->mb_rangelow.mbe_type == MAC_BIBA_TYPE_LOW &&
sys/security/mac_biba/mac_biba.c
337
mb->mb_rangehigh.mbe_type == MAC_BIBA_TYPE_HIGH)
sys/security/mac_biba/mac_biba.c
345
biba_high_effective(struct mac_biba *mb)
sys/security/mac_biba/mac_biba.c
348
KASSERT((mb->mb_flags & MAC_BIBA_FLAG_EFFECTIVE) != 0,
sys/security/mac_biba/mac_biba.c
351
return (mb->mb_effective.mbe_type == MAC_BIBA_TYPE_HIGH);
sys/security/mac_biba/mac_biba.c
355
biba_valid(struct mac_biba *mb)
sys/security/mac_biba/mac_biba.c
358
if (mb->mb_flags & MAC_BIBA_FLAG_EFFECTIVE) {
sys/security/mac_biba/mac_biba.c
359
switch (mb->mb_effective.mbe_type) {
sys/security/mac_biba/mac_biba.c
366
if (mb->mb_effective.mbe_grade != 0 ||
sys/security/mac_biba/mac_biba.c
368
mb->mb_effective.mbe_compartments))
sys/security/mac_biba/mac_biba.c
376
if (mb->mb_effective.mbe_type != MAC_BIBA_TYPE_UNDEF)
sys/security/mac_biba/mac_biba.c
380
if (mb->mb_flags & MAC_BIBA_FLAG_RANGE) {
sys/security/mac_biba/mac_biba.c
381
switch (mb->mb_rangelow.mbe_type) {
sys/security/mac_biba/mac_biba.c
388
if (mb->mb_rangelow.mbe_grade != 0 ||
sys/security/mac_biba/mac_biba.c
390
mb->mb_rangelow.mbe_compartments))
sys/security/mac_biba/mac_biba.c
398
switch (mb->mb_rangehigh.mbe_type) {
sys/security/mac_biba/mac_biba.c
405
if (mb->mb_rangehigh.mbe_grade != 0 ||
sys/security/mac_biba/mac_biba.c
407
mb->mb_rangehigh.mbe_compartments))
sys/security/mac_biba/mac_biba.c
414
if (!biba_dominate_element(&mb->mb_rangehigh,
sys/security/mac_biba/mac_biba.c
415
&mb->mb_rangelow))
sys/security/mac_biba/mac_biba.c
418
if (mb->mb_rangelow.mbe_type != MAC_BIBA_TYPE_UNDEF ||
sys/security/mac_biba/mac_biba.c
419
mb->mb_rangehigh.mbe_type != MAC_BIBA_TYPE_UNDEF)
sys/security/mac_biba/mac_biba.c
427
biba_set_range(struct mac_biba *mb, u_short typelow, u_short gradelow,
sys/security/mac_biba/mac_biba.c
432
mb->mb_rangelow.mbe_type = typelow;
sys/security/mac_biba/mac_biba.c
433
mb->mb_rangelow.mbe_grade = gradelow;
sys/security/mac_biba/mac_biba.c
435
memcpy(mb->mb_rangelow.mbe_compartments, compartmentslow,
sys/security/mac_biba/mac_biba.c
436
sizeof(mb->mb_rangelow.mbe_compartments));
sys/security/mac_biba/mac_biba.c
437
mb->mb_rangehigh.mbe_type = typehigh;
sys/security/mac_biba/mac_biba.c
438
mb->mb_rangehigh.mbe_grade = gradehigh;
sys/security/mac_biba/mac_biba.c
440
memcpy(mb->mb_rangehigh.mbe_compartments, compartmentshigh,
sys/security/mac_biba/mac_biba.c
441
sizeof(mb->mb_rangehigh.mbe_compartments));
sys/security/mac_biba/mac_biba.c
442
mb->mb_flags |= MAC_BIBA_FLAG_RANGE;
sys/security/mac_biba/mac_biba.c
446
biba_set_effective(struct mac_biba *mb, u_short type, u_short grade,
sys/security/mac_biba/mac_biba.c
450
mb->mb_effective.mbe_type = type;
sys/security/mac_biba/mac_biba.c
451
mb->mb_effective.mbe_grade = grade;
sys/security/mac_biba/mac_biba.c
453
memcpy(mb->mb_effective.mbe_compartments, compartments,
sys/security/mac_biba/mac_biba.c
454
sizeof(mb->mb_effective.mbe_compartments));
sys/security/mac_biba/mac_biba.c
455
mb->mb_flags |= MAC_BIBA_FLAG_EFFECTIVE;
sys/security/mac_biba/mac_biba.c
586
biba_to_string(struct sbuf *sb, struct mac_biba *mb)
sys/security/mac_biba/mac_biba.c
589
if (mb->mb_flags & MAC_BIBA_FLAG_EFFECTIVE) {
sys/security/mac_biba/mac_biba.c
590
if (biba_element_to_string(sb, &mb->mb_effective) == -1)
sys/security/mac_biba/mac_biba.c
594
if (mb->mb_flags & MAC_BIBA_FLAG_RANGE) {
sys/security/mac_biba/mac_biba.c
598
if (biba_element_to_string(sb, &mb->mb_rangelow) == -1)
sys/security/mac_biba/mac_biba.c
604
if (biba_element_to_string(sb, &mb->mb_rangehigh) == -1)
sys/security/mac_biba/mac_biba.c
618
struct mac_biba *mb;
sys/security/mac_biba/mac_biba.c
625
mb = SLOT(label);
sys/security/mac_biba/mac_biba.c
626
return (biba_to_string(sb, mb));
sys/security/mac_biba/mac_biba.c
686
biba_parse(struct mac_biba *mb, char *string)
sys/security/mac_biba/mac_biba.c
713
bzero(mb, sizeof(*mb));
sys/security/mac_biba/mac_biba.c
715
error = biba_parse_element(&mb->mb_effective, effective);
sys/security/mac_biba/mac_biba.c
718
mb->mb_flags |= MAC_BIBA_FLAG_EFFECTIVE;
sys/security/mac_biba/mac_biba.c
722
error = biba_parse_element(&mb->mb_rangelow, rangelow);
sys/security/mac_biba/mac_biba.c
725
error = biba_parse_element(&mb->mb_rangehigh, rangehigh);
sys/security/mac_biba/mac_biba.c
728
mb->mb_flags |= MAC_BIBA_FLAG_RANGE;
sys/security/mac_biba/mac_biba.c
731
error = biba_valid(mb);
sys/security/mac_biba/mac_biba.c
742
struct mac_biba *mb, mb_temp;
sys/security/mac_biba/mac_biba.c
754
mb = SLOT(label);
sys/security/mac_biba/mac_biba.c
755
*mb = mb_temp;
sys/security/mac_biba/mac_biba.c
941
struct mac_biba *mb;
sys/security/mac_biba/mac_biba.c
945
mb = SLOT(delabel);
sys/security/mac_biba/mac_biba.c
959
biba_set_effective(mb, biba_type, 0, NULL);
sys/security/mac_biba/mac_biba.c
966
struct mac_biba *mb;
sys/security/mac_biba/mac_biba.c
968
mb = SLOT(delabel);
sys/security/mac_biba/mac_biba.c
970
biba_set_effective(mb, MAC_BIBA_TYPE_HIGH, 0, NULL);
sys/sys/ktls.h
261
ktls_mbuf_crypto_st_t ktls_mbuf_crypto_state(struct mbuf *mb, int offset, int len);
sys/sys/sockbuf.h
250
void sbsndptr_adv(struct sockbuf *sb, struct mbuf *mb, u_int len);
sys/sys/stats.h
651
struct metablob *mb; /* Template metadata */
sys/vm/pmap.h
129
vm_page_t mb[], vm_offset_t b_offset, int xfersize);
sys/x86/x86/cpu_machdep.c
592
struct monitorbuf *mb;
sys/x86/x86/cpu_machdep.c
613
mb = &pcpu_find(0)->pc_monitorbuf;
sys/x86/x86/cpu_machdep.c
614
atomic_store_int(&mb->stop_state,
sys/x86/x86/cpu_machdep.c
820
struct monitorbuf *mb;
sys/x86/x86/cpu_machdep.c
823
mb = &pcpu_find(cpu)->pc_monitorbuf;
sys/x86/x86/cpu_machdep.c
824
state = &mb->idle_state;
sys/x86/x86/mp_x86.c
1527
struct monitorbuf *mb;
sys/x86/x86/mp_x86.c
1538
mb = PCPU_PTR(monitorbuf);
sys/x86/x86/mp_x86.c
1539
atomic_store_int(&mb->stop_state,
sys/x86/x86/mp_x86.c
1549
cpu_monitor(mb, 0, 0);
sys/x86/x86/mp_x86.c
1550
if (atomic_load_int(&mb->stop_state) ==
sys/xen/xen-os.h
167
#define xen_mb() mb()
usr.bin/localedef/wide.c
141
show_mb(const char *mb)
usr.bin/localedef/wide.c
146
if (isascii(*mb) && isgraph(*mb)) {
usr.bin/localedef/wide.c
147
buf[0] = *mb;
usr.bin/localedef/wide.c
152
while (*mb != 0) {
usr.bin/localedef/wide.c
154
(void) snprintf(scr, sizeof (scr), "\\x%02x", *mb);
usr.bin/localedef/wide.c
156
mb++;
usr.bin/localedef/wide.c
181
towide_none(wchar_t *c, const char *mb, unsigned n __unused)
usr.bin/localedef/wide.c
187
*c = (uint8_t)*mb;
usr.bin/localedef/wide.c
192
tomb_none(char *mb, wchar_t wc)
usr.bin/localedef/wide.c
198
*(uint8_t *)mb = (wc & 0xff);
usr.bin/localedef/wide.c
199
mb[1] = 0;
usr.bin/localedef/wide.c
207
towide_utf8(wchar_t *wc, const char *mb, unsigned n)
usr.bin/localedef/wide.c
213
const uint8_t *s = (const uint8_t *)mb;
usr.bin/localedef/wide.c
238
werr("utf8 encoding too large (%s)", show_mb(mb));
usr.bin/localedef/wide.c
242
werr("incomplete utf8 sequence (%s)", show_mb(mb));
usr.bin/localedef/wide.c
256
werr("illegal redundant utf8 encoding (%s)", show_mb(mb));
usr.bin/localedef/wide.c
264
tomb_utf8(char *mb, wchar_t wc)
usr.bin/localedef/wide.c
266
uint8_t *s = (uint8_t *)mb;
usr.bin/localedef/wide.c
312
towide_dbcs(wchar_t *wc, const char *mb, unsigned n)
usr.bin/localedef/wide.c
316
c = *(const uint8_t *)mb;
usr.bin/localedef/wide.c
324
werr("incomplete character sequence (%s)", show_mb(mb));
usr.bin/localedef/wide.c
330
c |= (uint8_t)(mb[1]);
usr.bin/localedef/wide.c
341
tomb_mbs(char *mb, wchar_t wc)
usr.bin/localedef/wide.c
343
uint8_t *s = (uint8_t *)mb;
usr.bin/localedef/wide.c
371
towide_big5(wchar_t *wc, const char *mb, unsigned n)
usr.bin/localedef/wide.c
373
return (towide_dbcs(wc, mb, n));
usr.bin/localedef/wide.c
381
towide_gbk(wchar_t *wc, const char *mb, unsigned n)
usr.bin/localedef/wide.c
383
return (towide_dbcs(wc, mb, n));
usr.bin/localedef/wide.c
391
towide_gb2312(wchar_t *wc, const char *mb, unsigned n)
usr.bin/localedef/wide.c
393
return (towide_dbcs(wc, mb, n));
usr.bin/localedef/wide.c
402
towide_gb18030(wchar_t *wc, const char *mb, unsigned n)
usr.bin/localedef/wide.c
406
c = *(const uint8_t *)mb;
usr.bin/localedef/wide.c
414
werr("incomplete character sequence (%s)", show_mb(mb));
usr.bin/localedef/wide.c
420
c |= (uint8_t)(mb[1]);
usr.bin/localedef/wide.c
425
show_mb(mb));
usr.bin/localedef/wide.c
429
c |= (uint8_t)(mb[2]);
usr.bin/localedef/wide.c
431
c |= (uint8_t)(mb[3]);
usr.bin/localedef/wide.c
445
towide_mskanji(wchar_t *wc, const char *mb, unsigned n)
usr.bin/localedef/wide.c
449
c = *(const uint8_t *)mb;
usr.bin/localedef/wide.c
458
werr("incomplete character sequence (%s)", show_mb(mb));
usr.bin/localedef/wide.c
464
c |= (uint8_t)(mb[1]);
usr.bin/localedef/wide.c
478
towide_euc_impl(wchar_t *wc, const char *mb, unsigned n,
usr.bin/localedef/wide.c
485
c = *(const uint8_t *)mb;
usr.bin/localedef/wide.c
512
werr("incomplete character sequence (%s)", show_mb(mb));
usr.bin/localedef/wide.c
519
c |= (uint8_t)(mb[i]);
usr.bin/localedef/wide.c
535
towide_euccn(wchar_t *wc, const char *mb, unsigned n)
usr.bin/localedef/wide.c
537
return (towide_euc_impl(wc, mb, n, 0x8e, 4, 0, 0));
usr.bin/localedef/wide.c
549
towide_eucjp(wchar_t *wc, const char *mb, unsigned n)
usr.bin/localedef/wide.c
551
return (towide_euc_impl(wc, mb, n, 0x8e, 2, 0x8f, 3));
usr.bin/localedef/wide.c
563
towide_euckr(wchar_t *wc, const char *mb, unsigned n)
usr.bin/localedef/wide.c
565
return (towide_euc_impl(wc, mb, n, 0, 0, 0, 0));
usr.bin/localedef/wide.c
577
towide_euctw(wchar_t *wc, const char *mb, unsigned n)
usr.bin/localedef/wide.c
579
return (towide_euc_impl(wc, mb, n, 0x8e, 4, 0, 0));
usr.bin/localedef/wide.c
587
to_wide(wchar_t *wc, const char *mb)
usr.bin/localedef/wide.c
590
return (_towide(wc, mb, strlen(mb)));
usr.bin/localedef/wide.c
594
to_mb(char *mb, wchar_t wc)
usr.bin/localedef/wide.c
598
if ((rv = _tomb(mb, wc)) < 0) {
usr.sbin/ppp/mbuf.c
106
struct mbucket **mb;
usr.sbin/ppp/mbuf.c
121
mb = M_BUCKET(m_len);
usr.sbin/ppp/mbuf.c
124
if (*mb) {
usr.sbin/ppp/mbuf.c
126
bp = &(*mb)->u.m;
usr.sbin/ppp/mbuf.c
127
if (--(*mb)->u.f.count == 0)
usr.sbin/ppp/mbuf.c
128
*mb = (*mb)->u.f.next;
usr.sbin/ppp/mbuf.c
130
((struct mbucket *)((char *)*mb + size))->u.f.count = (*mb)->u.f.count;
usr.sbin/ppp/mbuf.c
131
*mb = (struct mbucket *)((char *)*mb + size);
usr.sbin/ppp/mbuf.c
132
(*mb)->u.f.next = NULL;
usr.sbin/ppp/mbuf.c
139
*mb = (struct mbucket *)malloc(BUCKET_CHUNK * size);
usr.sbin/ppp/mbuf.c
140
if (*mb == NULL) {
usr.sbin/ppp/mbuf.c
145
bp = &(*mb)->u.m;
usr.sbin/ppp/mbuf.c
146
*mb = (struct mbucket *)((char *)*mb + size);
usr.sbin/ppp/mbuf.c
147
(*mb)->u.f.count = BUCKET_CHUNK - 1;
usr.sbin/ppp/mbuf.c
148
(*mb)->u.f.next = NULL;
usr.sbin/ppp/mbuf.c
167
struct mbucket **mb, *f;
usr.sbin/ppp/mbuf.c
175
mb = M_BUCKET(bp->m_size);
usr.sbin/ppp/mbuf.c
176
f->u.f.next = *mb;
usr.sbin/ppp/mbuf.c
178
*mb = f;