user_index
u32 *user_index)
*user_index = cmd_uidx;
*user_index = MLX5_IB_DEFAULT_UIDX;
u32 *user_index)
return verify_assign_uidx(cqe_version, ucmd->uidx, user_index);
u32 *user_index)
return verify_assign_uidx(cqe_version, ucmd->uidx, user_index);
u32 user_index;
#define MLX5_USER_ASSIGNED_UIDX_MASK __mlx5_mask(qpc, user_index)
MLX5_SET(sqc, sqc, user_index, MLX5_GET(qpc, qpc, user_index));
MLX5_SET(rqc, rqc, user_index, MLX5_GET(qpc, qpc, user_index));
MLX5_SET(qpc, qpc, user_index, uidx);
MLX5_SET(qpc, qpc, user_index, uidx);
MLX5_SET(qpc, qpc, user_index, uidx);
MLX5_SET(qpc, qpc, user_index, uidx);
MLX5_SET(dctc, dctc, user_index, uidx);
MLX5_SET(rqc, rqc, user_index, rwq->user_index);
rwq->user_index = ucmd.user_index;
in->user_index = MLX5_IB_DEFAULT_UIDX;
in->user_index = uidx;
u32 user_index;
MLX5_SET(xrc_srqc, xrc_srqc, user_index, in->user_index);
MLX5_SET(xrqc, xrqc, user_index, in->user_index);
MLX5_SET(qpc, qpc, user_index, 0xFFFFFF);
MLX5_SET(qpc, qpc, user_index, 0xFFFFFF);
if (!htt_stats_buf->user_index)
if (htt_stats_buf->user_index <
htt_stats_buf->user_index,
htt_stats_buf->user_index,
htt_stats_buf->user_index,
htt_stats_buf->user_index,
htt_stats_buf->user_index,
htt_stats_buf->user_index,
htt_stats_buf->user_index,
if (!htt_stats_buf->user_index)
if (htt_stats_buf->user_index <
htt_stats_buf->user_index,
htt_stats_buf->user_index,
htt_stats_buf->user_index,
htt_stats_buf->user_index,
htt_stats_buf->user_index,
htt_stats_buf->user_index,
htt_stats_buf->user_index,
if (!htt_stats_buf->user_index)
if (htt_stats_buf->user_index < HTT_TX_PDEV_STATS_NUM_OFDMA_USER_STATS) {
htt_stats_buf->user_index,
htt_stats_buf->user_index,
htt_stats_buf->user_index,
htt_stats_buf->user_index,
htt_stats_buf->user_index,
htt_stats_buf->user_index,
htt_stats_buf->user_index,
u32 user_index;
u32 user_index;
user_index = __le32_to_cpu(htt_stats_buf->user_index);
if (!user_index)
if (user_index < ATH12K_HTT_TX_NUM_AC_MUMIMO_USER_STATS) {
user_index,
user_index,
user_index,
user_index,
user_index,
user_index,
user_index,
if (!user_index)
if (user_index < ATH12K_HTT_TX_NUM_AX_MUMIMO_USER_STATS) {
user_index,
user_index,
user_index,
user_index,
user_index,
user_index,
user_index,
if (!user_index)
if (user_index < ATH12K_HTT_TX_NUM_OFDMA_USER_STATS) {
user_index,
user_index,
user_index,
user_index,
user_index,
user_index,
user_index,
u32 user_index;
user_index = __le32_to_cpu(htt_stats_buf->user_index);
if (!user_index)
user_index,
user_index,
user_index,
user_index,
"rx_ulofdma_non_data_nusers_%u = %u\n", user_index,
user_index,
__le32 user_index;
__le32 user_index;
u16 __user *user_index = (u16 __user *)uv_ioctl->argument_addr;
if (buf_len > UVIO_RETR_SECRET_MAX_LEN || buf_len < sizeof(*user_index))
if (get_user(uvcb.secret_idx, user_index))
MLX5_SET(qpc, qpc, user_index, 0xFFFFFF);
#define MLX5_USER_INDEX_LEN (MLX5_FLD_SZ_BYTES(qpc, user_index) * 8)
u8 user_index[0x18];
u8 user_index[0x18];
u8 user_index[0x18];
u8 user_index[0x18];
u8 user_index[0x18];
u8 user_index[0x18];
__u32 user_index;
user_index,
static u32 user_index = 0xFFFFFFFF;
if (user_index != 0xFFFFFFFF) {
func = index_to_func(user_index);
errx(EXIT_FAILURE, "Invalid input leaf (0x%x)", user_index);
user_index, user_sub);
user_index = strtoul(optarg, NULL, 0);