mlx5_wqe_eth_seg
overhead_ud_lso += sizeof(struct mlx5_wqe_eth_seg);
sizeof(struct mlx5_wqe_eth_seg);
struct mlx5_wqe_eth_seg *eseg = *seg;
memset(eseg, 0, sizeof(struct mlx5_wqe_eth_seg));
stride = ALIGN(sizeof(struct mlx5_wqe_eth_seg) -
*seg += sizeof(struct mlx5_wqe_eth_seg);
*size += sizeof(struct mlx5_wqe_eth_seg) / 16;
struct mlx5_wqe_eth_seg eth;
#define INL_HDR_START_SZ (sizeof(((struct mlx5_wqe_eth_seg *)NULL)->inline_hdr.start))
static inline void mlx5e_eseg_swp_offsets_add_vlan(struct mlx5_wqe_eth_seg *eseg)
mlx5e_set_eseg_swp(struct sk_buff *skb, struct mlx5_wqe_eth_seg *eseg,
struct mlx5_wqe_eth_seg *eseg = priv;
struct mlx5_wqe_eth_seg *eseg;
#define MLX5E_ACCEL_ESEG_LEN offsetof(struct mlx5_wqe_eth_seg, mss)
struct mlx5_wqe_eth_seg *eseg, u16 ihs)
mlx5e_tx_tunnel_accel(struct sk_buff *skb, struct mlx5_wqe_eth_seg *eseg, u16 ihs)
struct mlx5_wqe_eth_seg *eseg)
struct mlx5_wqe_eth_seg *eseg, u8 mode,
struct mlx5_wqe_eth_seg *eseg)
static inline bool mlx5e_ipsec_eseg_meta(struct mlx5_wqe_eth_seg *eseg)
struct mlx5_wqe_eth_seg *eseg)
static inline bool mlx5e_ipsec_eseg_meta(struct mlx5_wqe_eth_seg *eseg)
struct mlx5_wqe_eth_seg *eseg);
struct mlx5_wqe_eth_seg *eseg)
struct mlx5_wqe_eth_seg *eseg);
struct mlx5_wqe_eth_seg *eseg)
struct mlx5_wqe_eth_seg *eseg)
struct mlx5_wqe_eth_seg *eseg)
struct mlx5_wqe_eth_seg *eseg);
struct mlx5_wqe_eth_seg *eseg)
struct mlx5_wqe_eth_seg *eseg)
struct mlx5_wqe_eth_seg *eseg, bool xmit_more)
struct mlx5_wqe_eth_seg *eseg;
static bool mlx5e_tx_mpwqe_same_eseg(struct mlx5e_txqsq *sq, struct mlx5_wqe_eth_seg *eseg)
struct mlx5_wqe_eth_seg *eseg)
struct mlx5_wqe_eth_seg *eseg, bool xmit_more)
struct mlx5_wqe_eth_seg *eseg)
struct mlx5_wqe_eth_seg *eseg, u16 ihs)
struct mlx5_wqe_eth_seg eseg = {};
struct mlx5_wqe_eth_seg *eseg;
struct mlx5_wqe_eth_seg eth;