MLX5_CAP_GEN_2
if (MLX5_CAP_GEN_2(mdev, dp_ordering_force) &&
if (!MLX5_CAP_GEN(mdev, ib_virt) || !MLX5_CAP_GEN_2(mdev, multiplane))
!MLX5_CAP_GEN_2(dev->mdev, query_vuid))
!MLX5_CAP_GEN_2(dev->mdev, query_vuid))
!MLX5_CAP_GEN_2(mparent->mdev, multiplane_qp_ud))
MLX5_CAP_GEN_2(dev->mdev, max_mkey_log_entity_size_mtt);
max_log_size = MLX5_CAP_GEN_2(
!MLX5_CAP_GEN_2(dev->mdev, umr_log_entity_size_5)))
MLX5_CAP_GEN_2(dev->mdev,
min_log_size = MLX5_CAP_GEN_2(
MLX5_CAP_GEN_2(dev->mdev, log_min_mkey_entity_size);
if (!MLX5_CAP_GEN_2(mdev, dp_ordering_force))
if (MLX5_CAP_GEN_2(dev->mdev, umr_log_entity_size_5))
u8 min_page_shift = MLX5_CAP_GEN_2(mdev, log_min_mkey_entity_size);
1 << MLX5_CAP_GEN_2(ptpsq->txqsq.mdev,
min(MLX5_CAP_GEN_2(c->mdev, ts_cqe_metadata_size2wqe_counter),
unsigned int max_num_vhca_id = MLX5_CAP_GEN_2(mdev, max_rqt_vhca_id);
MLX5_CAP_GEN_2(mdev, flow_table_type_2_type) & MLX5_FT_NIC_RX_2_NIC_RX_RDMA &&
MLX5_CAP_GEN_2(mdev, flow_table_type_2_type) & MLX5_FT_NIC_TX_RDMA_2_NIC_TX)
!MLX5_CAP_GEN_2(mdev, ts_cqe_metadata_size2wqe_counter))
if (!MLX5_CAP_GEN_2(mdev, min_mkey_log_entity_size_fixed_buffer_valid))
if (MLX5_CAP_GEN_2(mdev, min_mkey_log_entity_size_fixed_buffer) >
int max_eqs_sf = MLX5_CAP_GEN_2(dev, sf_eq_usage) ?
MLX5_CAP_GEN_2(dev, max_num_eqs_24b) :
if (!MLX5_CAP_GEN_2(esw->dev, delegated_vhca_max))
u32 delegated_vhca_max = MLX5_CAP_GEN_2(esw->dev, delegated_vhca_max);
MLX5_CAP_GEN_2(esw->dev, max_reformat_insert_size) < sizeof(vlan_hdr) ||
MLX5_CAP_GEN_2(esw->dev, max_reformat_insert_offset) <
MLX5_CAP_GEN_2(esw->dev, max_reformat_remove_size) >= sizeof(struct vlan_hdr) &&
MLX5_CAP_GEN_2(esw->dev, max_reformat_remove_offset) >=
MLX5_CAP_GEN_2((esw->dev), ec_vf_vport_base), \
MLX5_CAP_GEN_2((esw->dev), ec_vf_vport_base) +\
if (!MLX5_CAP_GEN_2(esw->dev, max_num_eqs_24b)) {
if (!MLX5_CAP_GEN_2(esw->dev, max_num_eqs_24b)) {
use_l4_type = MLX5_CAP_GEN_2(dev, pcc_ifa2) &&
use_l4_type = MLX5_CAP_GEN_2(dev, pcc_ifa2) &&
use_l4_type = MLX5_CAP_GEN_2(dev, pcc_ifa2) &&
use_l4_type = MLX5_CAP_GEN_2(dev, pcc_ifa2) &&
u32 obj_supp = MLX5_CAP_GEN_2(mdev, cross_vhca_object_to_object_supported);
if (host_buses > MLX5_CAP_GEN_2(dev, max_rqt_vhca_id))
u32 obj_supp = MLX5_CAP_GEN_2(dev, cross_vhca_object_to_object_supported);
return MLX5_CAP_GEN_2(dev, ec_vf_vport_base);
if (MLX5_CAP_GEN_2(dev, max_num_eqs_24b))
return MLX5_CAP_GEN_2(dev, max_num_eqs_24b);
if (!vport->max_eqs_set && MLX5_CAP_GEN_2(dev, max_num_eqs_24b)) {
MLX5_CAP_GEN_2(mdev, max_reformat_insert_size) < params->size ||
MLX5_CAP_GEN_2(mdev, max_reformat_insert_offset) < params->param_1) {
MLX5_CAP_GEN_2(dmn->mdev, max_reformat_insert_size) < data_sz ||
MLX5_CAP_GEN_2(dmn->mdev, max_reformat_insert_offset) < reformat_param_1) {
MLX5_CAP_GEN_2(dmn->mdev, max_reformat_remove_size) < data_sz ||
MLX5_CAP_GEN_2(dmn->mdev, max_reformat_remove_offset) < reformat_param_1) {
if (MLX5_CAP_GEN_2(master_mdev, sw_vhca_id_valid)) {
MLX5_CAP_GEN_2(master_mdev, sw_vhca_id));
if (MLX5_CAP_GEN_2(mdev, load_balance_id) &&
MLX5_CAP_GEN_2(mdev, lag_per_mp_group))
buf[(*len)++] = MLX5_CAP_GEN_2(mdev, load_balance_id);
if (!MLX5_CAP_GEN_2(dev, function_id_type_vhca_id))
if (!(MLX5_CAP_GEN_2(mvdev->mdev, migration_multi_load) &&
MLX5_CAP_GEN_2(mvdev->mdev, migration_tracking_state)))
if (MLX5_CAP_GEN_2(mvdev->mdev, migration_in_chunks))
if (((MLX5_CAP_GEN_2(mdev, flow_table_type_2_type) &