mlx5e_umr_wqe
(ALIGN_DOWN(MLX5_SEND_WQE_MAX_SIZE - sizeof(struct mlx5e_umr_wqe), \
static_assert(offsetof(struct mlx5e_umr_wqe, inline_mtts) == sizeof(struct mlx5e_umr_wqe_hdr),
max_pages_per_wqe = ALIGN_DOWN(max_wqe_size - sizeof(struct mlx5e_umr_wqe),
umr_wqe_sz = sizeof(struct mlx5e_umr_wqe) +
(sizeof(struct mlx5e_umr_wqe) +\
(((wqe_size) - sizeof(struct mlx5e_umr_wqe)) / sizeof(struct mlx5_ksm))
struct mlx5e_umr_wqe *umr_wqe;
memcpy(umr_wqe, &rq->mpwqe.umr_wqe, sizeof(struct mlx5e_umr_wqe));
struct mlx5e_umr_wqe *wqe)
struct mlx5e_umr_wqe, hdr));
struct mlx5e_umr_wqe *umr_wqe;
memcpy(umr_wqe, &rq->mpwqe.umr_wqe, sizeof(struct mlx5e_umr_wqe));