Lines Matching refs:MLX5_CAP_GEN
326 if (!MLX5_CAP_GEN(dev->mdev, qpc_extension) || in mlx5_ib_qp_err_syndrome()
327 !MLX5_CAP_GEN(dev->mdev, qp_error_syndrome)) in mlx5_ib_qp_err_syndrome()
440 if (cap->max_recv_wr > (1 << MLX5_CAP_GEN(dev->mdev, log_max_qp_sz))) in set_rq_size()
475 if (wqe_size > MLX5_CAP_GEN(dev->mdev, max_wqe_sz_rq)) { in set_rq_size()
478 MLX5_CAP_GEN(dev->mdev, in set_rq_size()
605 if (wqe_size > MLX5_CAP_GEN(dev->mdev, max_wqe_sz_sq)) { in calc_sq_size()
607 wqe_size, MLX5_CAP_GEN(dev->mdev, max_wqe_sz_sq)); in calc_sq_size()
617 if (qp->sq.wqe_cnt > (1 << MLX5_CAP_GEN(dev->mdev, log_max_qp_sz))) { in calc_sq_size()
621 1 << MLX5_CAP_GEN(dev->mdev, log_max_qp_sz)); in calc_sq_size()
644 if (desc_sz > MLX5_CAP_GEN(dev->mdev, max_wqe_sz_sq)) { in set_user_buf_size()
646 desc_sz, MLX5_CAP_GEN(dev->mdev, max_wqe_sz_sq)); in set_user_buf_size()
658 if (qp->sq.wqe_cnt > (1 << MLX5_CAP_GEN(dev->mdev, log_max_qp_sz))) { in set_user_buf_size()
661 1 << MLX5_CAP_GEN(dev->mdev, log_max_qp_sz)); in set_user_buf_size()
1116 qp->bf.buf_size = (1 << MLX5_CAP_GEN(dev->mdev, log_bf_reg_size)) / 2; in _create_kernel_qp()
1291 u8 ts_cap = MLX5_CAP_GEN(dev->mdev, rq_ts_format); in get_rq_ts_format()
1299 u8 ts_cap = MLX5_CAP_GEN(dev->mdev, sq_ts_format); in get_sq_ts_format()
1388 if (MLX5_CAP_GEN(dev->mdev, eth_net_offloads) && in create_raw_packet_qp_sq()
1939 MLX5_CAP_GEN(dev->mdev, dc_req_scat_data_cqe)) in configure_requester_scat_cqe()
1966 u8 atomic = MLX5_CAP_GEN(dev->mdev, atomic); in get_atomic_mode()
2041 if (MLX5_CAP_GEN(mdev, cqe_version) == MLX5_CQE_VERSION_V1) in create_xrc_tgt_qp()
2059 if (MLX5_CAP_GEN(mdev, ece_support)) in create_xrc_tgt_qp()
2115 if (ucmd->sq_wqe_count > (1 << MLX5_CAP_GEN(mdev, log_max_qp_sz))) in create_dci()
2129 if (MLX5_CAP_GEN(mdev, ece_support)) in create_dci()
2186 if (MLX5_CAP_GEN(mdev, cqe_version) == MLX5_CQE_VERSION_V1) in create_dci()
2204 if (MLX5_CAP_GEN(mdev, ece_support)) in create_dci()
2282 if (ucmd->sq_wqe_count > (1 << MLX5_CAP_GEN(mdev, log_max_qp_sz))) in create_user_qp()
2300 if (MLX5_CAP_GEN(mdev, ece_support)) in create_user_qp()
2381 if (MLX5_CAP_GEN(mdev, cqe_version) == MLX5_CQE_VERSION_V1) in create_user_qp()
2407 if (MLX5_CAP_GEN(mdev, ece_support)) in create_user_qp()
2524 if (MLX5_CAP_GEN(mdev, cqe_version) == MLX5_CQE_VERSION_V1) in create_kernel_qp()
2532 MLX5_CAP_GEN(mdev, go_back_n)) in create_kernel_qp()
2740 if (mlx5_lag_is_active(dev->mdev) && !MLX5_CAP_GEN(dev->mdev, lag_dct)) in create_dct()
2754 if (MLX5_CAP_GEN(dev->mdev, ece_support)) in create_dct()
2771 if (attr->qp_type == IB_QPT_DRIVER && !MLX5_CAP_GEN(dev->mdev, dct)) in check_qp_type()
2777 if (!MLX5_CAP_GEN(dev->mdev, xrc)) in check_qp_type()
2853 return MLX5_CAP_GEN(mdev, dp_ordering_ooo_all_rc); in get_dp_ooo_cap()
2856 return MLX5_CAP_GEN(mdev, dp_ordering_ooo_all_xrc); in get_dp_ooo_cap()
2858 return MLX5_CAP_GEN(mdev, dp_ordering_ooo_all_uc); in get_dp_ooo_cap()
2860 return MLX5_CAP_GEN(mdev, dp_ordering_ooo_all_ud); in get_dp_ooo_cap()
2863 return MLX5_CAP_GEN(mdev, dp_ordering_ooo_all_dc); in get_dp_ooo_cap()
2929 MLX5_CAP_GEN(mdev, log_max_dci_stream_channels), in process_vendor_flags()
2934 MLX5_CAP_GEN(mdev, sctr_data_cqe), qp); in process_vendor_flags()
2936 MLX5_CAP_GEN(mdev, sctr_data_cqe), qp); in process_vendor_flags()
2955 MLX5_CAP_GEN(mdev, qp_packet_based), qp); in process_vendor_flags()
3010 MLX5_CAP_GEN(mdev, sho), qp); in process_create_flags()
3013 MLX5_CAP_GEN(mdev, block_lb_mc), qp); in process_create_flags()
3015 MLX5_CAP_GEN(mdev, cd), qp); in process_create_flags()
3017 MLX5_CAP_GEN(mdev, cd), qp); in process_create_flags()
3019 MLX5_CAP_GEN(mdev, cd), qp); in process_create_flags()
3024 MLX5_CAP_GEN(mdev, ipoib_basic_offloads), in process_create_flags()
3026 cond = MLX5_CAP_GEN(mdev, port_type) == MLX5_CAP_PORT_TYPE_IB; in process_create_flags()
3032 cond = MLX5_CAP_GEN(mdev, eth_net_offloads) && in process_create_flags()
3037 cond = MLX5_CAP_GEN(mdev, eth_net_offloads) && in process_create_flags()
3045 MLX5_CAP_GEN(mdev, end_pad), qp); in process_create_flags()
3460 stat_rate_support = MLX5_CAP_GEN(dev->mdev, stat_rate_support); in mlx5r_ib_rate()
3819 if (MLX5_CAP_GEN(dev->mdev, modify_rq_counter_set_id)) { in modify_raw_packet_qp_rq()
4017 (dev->lag_active ? dev->lag_ports : MLX5_CAP_GEN(dev->mdev, num_lag_ports)) + 1; in get_tx_affinity_rr()
4213 MLX5_CAP_GEN(dev->mdev, init2_lag_tx_port_affinity)) in __mlx5_ib_modify_qp()
4233 MLX5_CAP_GEN(dev->mdev, log_max_msg)); in __mlx5_ib_modify_qp()
4410 MLX5_CAP_GEN(dev->mdev, ece_support) ? in __mlx5_ib_modify_qp()
4537 if (MLX5_CAP_GEN(dev->mdev, ece_support) && ucmd->ece_options) in mlx5_ib_modify_dct()
4622 if (MLX5_CAP_GEN(dev->mdev, ece_support)) in mlx5_ib_modify_dct()
4657 log_max_ra_res = 1 << MLX5_CAP_GEN(dev->mdev, in validate_rd_atomic()
4659 log_max_ra_req = 1 << MLX5_CAP_GEN(dev->mdev, in validate_rd_atomic()
4662 log_max_ra_res = 1 << MLX5_CAP_GEN(dev->mdev, in validate_rd_atomic()
4664 log_max_ra_req = 1 << MLX5_CAP_GEN(dev->mdev, in validate_rd_atomic()
5172 if (!MLX5_CAP_GEN(dev->mdev, xrc)) in mlx5_ib_alloc_xrcd()
5267 if (!MLX5_CAP_GEN(dev->mdev, end_pad)) { in create_rq()
5299 has_net_offloads = MLX5_CAP_GEN(dev->mdev, eth_net_offloads); in create_rq()
5350 if (wq_init_attr->max_wr > (1 << MLX5_CAP_GEN(dev->mdev, log_max_wq_sz))) in set_user_rq_size()
5372 if (!MLX5_CAP_GEN(dev->mdev, ext_stride_num_range) && in log_of_strides_valid()
5412 if (!MLX5_CAP_GEN(dev->mdev, striding_rq)) { in prepare_user_rq()
5432 MLX5_CAP_GEN(dev->mdev, ext_stride_num_range) ? in prepare_user_rq()
5560 MLX5_CAP_GEN(dev->mdev, log_max_rqt_size)) { in mlx5_ib_create_rwq_ind_table()
5563 MLX5_CAP_GEN(dev->mdev, log_max_rqt_size)); in mlx5_ib_create_rwq_ind_table()
5666 if (!(MLX5_CAP_GEN(dev->mdev, eth_net_offloads) && in mlx5_ib_modify_wq()
5689 if (MLX5_CAP_GEN(dev->mdev, modify_rq_counter_set_id)) { in mlx5_ib_modify_wq()
5853 if (!MLX5_CAP_GEN(dev->mdev, rts2rts_qp_counters_set_id)) { in mlx5_ib_qp_set_counter()