|
@@ -66,15 +66,13 @@ static int mlx5_ib_query_device(struct ib_device *ibdev,
|
|
|
struct ib_device_attr *props)
|
|
|
{
|
|
|
struct mlx5_ib_dev *dev = to_mdev(ibdev);
|
|
|
+ struct mlx5_core_dev *mdev = dev->mdev;
|
|
|
struct ib_smp *in_mad = NULL;
|
|
|
struct ib_smp *out_mad = NULL;
|
|
|
- struct mlx5_general_caps *gen;
|
|
|
int err = -ENOMEM;
|
|
|
int max_rq_sg;
|
|
|
int max_sq_sg;
|
|
|
- u64 flags;
|
|
|
|
|
|
- gen = &dev->mdev->caps.gen;
|
|
|
in_mad = kzalloc(sizeof(*in_mad), GFP_KERNEL);
|
|
|
out_mad = kmalloc(sizeof(*out_mad), GFP_KERNEL);
|
|
|
if (!in_mad || !out_mad)
|
|
@@ -96,18 +94,18 @@ static int mlx5_ib_query_device(struct ib_device *ibdev,
|
|
|
IB_DEVICE_PORT_ACTIVE_EVENT |
|
|
|
IB_DEVICE_SYS_IMAGE_GUID |
|
|
|
IB_DEVICE_RC_RNR_NAK_GEN;
|
|
|
- flags = gen->flags;
|
|
|
- if (flags & MLX5_DEV_CAP_FLAG_BAD_PKEY_CNTR)
|
|
|
+
|
|
|
+ if (MLX5_CAP_GEN(mdev, pkv))
|
|
|
props->device_cap_flags |= IB_DEVICE_BAD_PKEY_CNTR;
|
|
|
- if (flags & MLX5_DEV_CAP_FLAG_BAD_QKEY_CNTR)
|
|
|
+ if (MLX5_CAP_GEN(mdev, qkv))
|
|
|
props->device_cap_flags |= IB_DEVICE_BAD_QKEY_CNTR;
|
|
|
- if (flags & MLX5_DEV_CAP_FLAG_APM)
|
|
|
+ if (MLX5_CAP_GEN(mdev, apm))
|
|
|
props->device_cap_flags |= IB_DEVICE_AUTO_PATH_MIG;
|
|
|
props->device_cap_flags |= IB_DEVICE_LOCAL_DMA_LKEY;
|
|
|
- if (flags & MLX5_DEV_CAP_FLAG_XRC)
|
|
|
+ if (MLX5_CAP_GEN(mdev, xrc))
|
|
|
props->device_cap_flags |= IB_DEVICE_XRC;
|
|
|
props->device_cap_flags |= IB_DEVICE_MEM_MGT_EXTENSIONS;
|
|
|
- if (flags & MLX5_DEV_CAP_FLAG_SIG_HAND_OVER) {
|
|
|
+ if (MLX5_CAP_GEN(mdev, sho)) {
|
|
|
props->device_cap_flags |= IB_DEVICE_SIGNATURE_HANDOVER;
|
|
|
/* At this stage no support for signature handover */
|
|
|
props->sig_prot_cap = IB_PROT_T10DIF_TYPE_1 |
|
|
@@ -116,7 +114,7 @@ static int mlx5_ib_query_device(struct ib_device *ibdev,
|
|
|
props->sig_guard_cap = IB_GUARD_T10DIF_CRC |
|
|
|
IB_GUARD_T10DIF_CSUM;
|
|
|
}
|
|
|
- if (flags & MLX5_DEV_CAP_FLAG_BLOCK_MCAST)
|
|
|
+ if (MLX5_CAP_GEN(mdev, block_lb_mc))
|
|
|
props->device_cap_flags |= IB_DEVICE_BLOCK_MULTICAST_LOOPBACK;
|
|
|
|
|
|
props->vendor_id = be32_to_cpup((__be32 *)(out_mad->data + 36)) &
|
|
@@ -126,37 +124,38 @@ static int mlx5_ib_query_device(struct ib_device *ibdev,
|
|
|
memcpy(&props->sys_image_guid, out_mad->data + 4, 8);
|
|
|
|
|
|
props->max_mr_size = ~0ull;
|
|
|
- props->page_size_cap = gen->min_page_sz;
|
|
|
- props->max_qp = 1 << gen->log_max_qp;
|
|
|
- props->max_qp_wr = gen->max_wqes;
|
|
|
- max_rq_sg = gen->max_rq_desc_sz / sizeof(struct mlx5_wqe_data_seg);
|
|
|
- max_sq_sg = (gen->max_sq_desc_sz - sizeof(struct mlx5_wqe_ctrl_seg)) /
|
|
|
- sizeof(struct mlx5_wqe_data_seg);
|
|
|
+ props->page_size_cap = 1ull << MLX5_CAP_GEN(mdev, log_pg_sz);
|
|
|
+ props->max_qp = 1 << MLX5_CAP_GEN(mdev, log_max_qp);
|
|
|
+ props->max_qp_wr = 1 << MLX5_CAP_GEN(mdev, log_max_qp_sz);
|
|
|
+ max_rq_sg = MLX5_CAP_GEN(mdev, max_wqe_sz_rq) /
|
|
|
+ sizeof(struct mlx5_wqe_data_seg);
|
|
|
+ max_sq_sg = (MLX5_CAP_GEN(mdev, max_wqe_sz_sq) -
|
|
|
+ sizeof(struct mlx5_wqe_ctrl_seg)) /
|
|
|
+ sizeof(struct mlx5_wqe_data_seg);
|
|
|
props->max_sge = min(max_rq_sg, max_sq_sg);
|
|
|
- props->max_cq = 1 << gen->log_max_cq;
|
|
|
- props->max_cqe = gen->max_cqes - 1;
|
|
|
- props->max_mr = 1 << gen->log_max_mkey;
|
|
|
- props->max_pd = 1 << gen->log_max_pd;
|
|
|
- props->max_qp_rd_atom = 1 << gen->log_max_ra_req_qp;
|
|
|
- props->max_qp_init_rd_atom = 1 << gen->log_max_ra_res_qp;
|
|
|
- props->max_srq = 1 << gen->log_max_srq;
|
|
|
- props->max_srq_wr = gen->max_srq_wqes - 1;
|
|
|
- props->local_ca_ack_delay = gen->local_ca_ack_delay;
|
|
|
+ props->max_cq = 1 << MLX5_CAP_GEN(mdev, log_max_cq);
|
|
|
+ props->max_cqe = (1 << MLX5_CAP_GEN(mdev, log_max_eq_sz)) - 1;
|
|
|
+ props->max_mr = 1 << MLX5_CAP_GEN(mdev, log_max_mkey);
|
|
|
+ props->max_pd = 1 << MLX5_CAP_GEN(mdev, log_max_pd);
|
|
|
+ props->max_qp_rd_atom = 1 << MLX5_CAP_GEN(mdev, log_max_ra_req_qp);
|
|
|
+ props->max_qp_init_rd_atom = 1 << MLX5_CAP_GEN(mdev, log_max_ra_res_qp);
|
|
|
+ props->max_srq = 1 << MLX5_CAP_GEN(mdev, log_max_srq);
|
|
|
+ props->max_srq_wr = (1 << MLX5_CAP_GEN(mdev, log_max_srq_sz)) - 1;
|
|
|
+ props->local_ca_ack_delay = MLX5_CAP_GEN(mdev, local_ca_ack_delay);
|
|
|
props->max_res_rd_atom = props->max_qp_rd_atom * props->max_qp;
|
|
|
props->max_srq_sge = max_rq_sg - 1;
|
|
|
props->max_fast_reg_page_list_len = (unsigned int)-1;
|
|
|
- props->local_ca_ack_delay = gen->local_ca_ack_delay;
|
|
|
props->atomic_cap = IB_ATOMIC_NONE;
|
|
|
props->masked_atomic_cap = IB_ATOMIC_NONE;
|
|
|
props->max_pkeys = be16_to_cpup((__be16 *)(out_mad->data + 28));
|
|
|
- props->max_mcast_grp = 1 << gen->log_max_mcg;
|
|
|
- props->max_mcast_qp_attach = gen->max_qp_mcg;
|
|
|
+ props->max_mcast_grp = 1 << MLX5_CAP_GEN(mdev, log_max_mcg);
|
|
|
+ props->max_mcast_qp_attach = MLX5_CAP_GEN(mdev, max_qp_mcg);
|
|
|
props->max_total_mcast_qp_attach = props->max_mcast_qp_attach *
|
|
|
props->max_mcast_grp;
|
|
|
props->max_map_per_fmr = INT_MAX; /* no limit in ConnectIB */
|
|
|
|
|
|
#ifdef CONFIG_INFINIBAND_ON_DEMAND_PAGING
|
|
|
- if (dev->mdev->caps.gen.flags & MLX5_DEV_CAP_FLAG_ON_DMND_PG)
|
|
|
+ if (MLX5_CAP_GEN(mdev, pg))
|
|
|
props->device_cap_flags |= IB_DEVICE_ON_DEMAND_PAGING;
|
|
|
props->odp_caps = dev->odp_caps;
|
|
|
#endif
|
|
@@ -172,14 +171,13 @@ int mlx5_ib_query_port(struct ib_device *ibdev, u8 port,
|
|
|
struct ib_port_attr *props)
|
|
|
{
|
|
|
struct mlx5_ib_dev *dev = to_mdev(ibdev);
|
|
|
+ struct mlx5_core_dev *mdev = dev->mdev;
|
|
|
struct ib_smp *in_mad = NULL;
|
|
|
struct ib_smp *out_mad = NULL;
|
|
|
- struct mlx5_general_caps *gen;
|
|
|
int ext_active_speed;
|
|
|
int err = -ENOMEM;
|
|
|
|
|
|
- gen = &dev->mdev->caps.gen;
|
|
|
- if (port < 1 || port > gen->num_ports) {
|
|
|
+ if (port < 1 || port > MLX5_CAP_GEN(mdev, num_ports)) {
|
|
|
mlx5_ib_warn(dev, "invalid port number %d\n", port);
|
|
|
return -EINVAL;
|
|
|
}
|
|
@@ -210,8 +208,8 @@ int mlx5_ib_query_port(struct ib_device *ibdev, u8 port,
|
|
|
props->phys_state = out_mad->data[33] >> 4;
|
|
|
props->port_cap_flags = be32_to_cpup((__be32 *)(out_mad->data + 20));
|
|
|
props->gid_tbl_len = out_mad->data[50];
|
|
|
- props->max_msg_sz = 1 << gen->log_max_msg;
|
|
|
- props->pkey_tbl_len = gen->port[port - 1].pkey_table_len;
|
|
|
+ props->max_msg_sz = 1 << MLX5_CAP_GEN(mdev, log_max_msg);
|
|
|
+ props->pkey_tbl_len = mdev->port_caps[port - 1].pkey_table_len;
|
|
|
props->bad_pkey_cntr = be16_to_cpup((__be16 *)(out_mad->data + 46));
|
|
|
props->qkey_viol_cntr = be16_to_cpup((__be16 *)(out_mad->data + 48));
|
|
|
props->active_width = out_mad->data[31] & 0xf;
|
|
@@ -238,7 +236,7 @@ int mlx5_ib_query_port(struct ib_device *ibdev, u8 port,
|
|
|
|
|
|
/* If reported active speed is QDR, check if is FDR-10 */
|
|
|
if (props->active_speed == 4) {
|
|
|
- if (gen->ext_port_cap[port - 1] &
|
|
|
+ if (mdev->port_caps[port - 1].ext_port_cap &
|
|
|
MLX_EXT_PORT_CAP_FLAG_EXTENDED_PORT_INFO) {
|
|
|
init_query_mad(in_mad);
|
|
|
in_mad->attr_id = MLX5_ATTR_EXTENDED_PORT_INFO;
|
|
@@ -392,7 +390,6 @@ static struct ib_ucontext *mlx5_ib_alloc_ucontext(struct ib_device *ibdev,
|
|
|
struct mlx5_ib_alloc_ucontext_req_v2 req;
|
|
|
struct mlx5_ib_alloc_ucontext_resp resp;
|
|
|
struct mlx5_ib_ucontext *context;
|
|
|
- struct mlx5_general_caps *gen;
|
|
|
struct mlx5_uuar_info *uuari;
|
|
|
struct mlx5_uar *uars;
|
|
|
int gross_uuars;
|
|
@@ -403,7 +400,6 @@ static struct ib_ucontext *mlx5_ib_alloc_ucontext(struct ib_device *ibdev,
|
|
|
int i;
|
|
|
size_t reqlen;
|
|
|
|
|
|
- gen = &dev->mdev->caps.gen;
|
|
|
if (!dev->ib_active)
|
|
|
return ERR_PTR(-EAGAIN);
|
|
|
|
|
@@ -436,14 +432,14 @@ static struct ib_ucontext *mlx5_ib_alloc_ucontext(struct ib_device *ibdev,
|
|
|
|
|
|
num_uars = req.total_num_uuars / MLX5_NON_FP_BF_REGS_PER_PAGE;
|
|
|
gross_uuars = num_uars * MLX5_BF_REGS_PER_PAGE;
|
|
|
- resp.qp_tab_size = 1 << gen->log_max_qp;
|
|
|
- resp.bf_reg_size = gen->bf_reg_size;
|
|
|
- resp.cache_line_size = L1_CACHE_BYTES;
|
|
|
- resp.max_sq_desc_sz = gen->max_sq_desc_sz;
|
|
|
- resp.max_rq_desc_sz = gen->max_rq_desc_sz;
|
|
|
- resp.max_send_wqebb = gen->max_wqes;
|
|
|
- resp.max_recv_wr = gen->max_wqes;
|
|
|
- resp.max_srq_recv_wr = gen->max_srq_wqes;
|
|
|
+ resp.qp_tab_size = 1 << MLX5_CAP_GEN(dev->mdev, log_max_qp);
|
|
|
+ resp.bf_reg_size = 1 << MLX5_CAP_GEN(dev->mdev, log_bf_reg_size);
|
|
|
+ resp.cache_line_size = L1_CACHE_BYTES;
|
|
|
+ resp.max_sq_desc_sz = MLX5_CAP_GEN(dev->mdev, max_wqe_sz_sq);
|
|
|
+ resp.max_rq_desc_sz = MLX5_CAP_GEN(dev->mdev, max_wqe_sz_rq);
|
|
|
+ resp.max_send_wqebb = 1 << MLX5_CAP_GEN(dev->mdev, log_max_qp_sz);
|
|
|
+ resp.max_recv_wr = 1 << MLX5_CAP_GEN(dev->mdev, log_max_qp_sz);
|
|
|
+ resp.max_srq_recv_wr = 1 << MLX5_CAP_GEN(dev->mdev, log_max_srq_sz);
|
|
|
|
|
|
context = kzalloc(sizeof(*context), GFP_KERNEL);
|
|
|
if (!context)
|
|
@@ -493,7 +489,7 @@ static struct ib_ucontext *mlx5_ib_alloc_ucontext(struct ib_device *ibdev,
|
|
|
mutex_init(&context->db_page_mutex);
|
|
|
|
|
|
resp.tot_uuars = req.total_num_uuars;
|
|
|
- resp.num_ports = gen->num_ports;
|
|
|
+ resp.num_ports = MLX5_CAP_GEN(dev->mdev, num_ports);
|
|
|
err = ib_copy_to_udata(udata, &resp,
|
|
|
sizeof(resp) - sizeof(resp.reserved));
|
|
|
if (err)
|
|
@@ -895,11 +891,9 @@ static void mlx5_ib_event(struct mlx5_core_dev *dev, void *context,
|
|
|
|
|
|
static void get_ext_port_caps(struct mlx5_ib_dev *dev)
|
|
|
{
|
|
|
- struct mlx5_general_caps *gen;
|
|
|
int port;
|
|
|
|
|
|
- gen = &dev->mdev->caps.gen;
|
|
|
- for (port = 1; port <= gen->num_ports; port++)
|
|
|
+ for (port = 1; port <= MLX5_CAP_GEN(dev->mdev, num_ports); port++)
|
|
|
mlx5_query_ext_port_caps(dev, port);
|
|
|
}
|
|
|
|
|
@@ -907,11 +901,9 @@ static int get_port_caps(struct mlx5_ib_dev *dev)
|
|
|
{
|
|
|
struct ib_device_attr *dprops = NULL;
|
|
|
struct ib_port_attr *pprops = NULL;
|
|
|
- struct mlx5_general_caps *gen;
|
|
|
int err = -ENOMEM;
|
|
|
int port;
|
|
|
|
|
|
- gen = &dev->mdev->caps.gen;
|
|
|
pprops = kmalloc(sizeof(*pprops), GFP_KERNEL);
|
|
|
if (!pprops)
|
|
|
goto out;
|
|
@@ -926,14 +918,17 @@ static int get_port_caps(struct mlx5_ib_dev *dev)
|
|
|
goto out;
|
|
|
}
|
|
|
|
|
|
- for (port = 1; port <= gen->num_ports; port++) {
|
|
|
+ for (port = 1; port <= MLX5_CAP_GEN(dev->mdev, num_ports); port++) {
|
|
|
err = mlx5_ib_query_port(&dev->ib_dev, port, pprops);
|
|
|
if (err) {
|
|
|
- mlx5_ib_warn(dev, "query_port %d failed %d\n", port, err);
|
|
|
+ mlx5_ib_warn(dev, "query_port %d failed %d\n",
|
|
|
+ port, err);
|
|
|
break;
|
|
|
}
|
|
|
- gen->port[port - 1].pkey_table_len = dprops->max_pkeys;
|
|
|
- gen->port[port - 1].gid_table_len = pprops->gid_tbl_len;
|
|
|
+ dev->mdev->port_caps[port - 1].pkey_table_len =
|
|
|
+ dprops->max_pkeys;
|
|
|
+ dev->mdev->port_caps[port - 1].gid_table_len =
|
|
|
+ pprops->gid_tbl_len;
|
|
|
mlx5_ib_dbg(dev, "pkey_table_len %d, gid_table_len %d\n",
|
|
|
dprops->max_pkeys, pprops->gid_tbl_len);
|
|
|
}
|
|
@@ -1207,8 +1202,8 @@ static void *mlx5_ib_add(struct mlx5_core_dev *mdev)
|
|
|
strlcpy(dev->ib_dev.name, "mlx5_%d", IB_DEVICE_NAME_MAX);
|
|
|
dev->ib_dev.owner = THIS_MODULE;
|
|
|
dev->ib_dev.node_type = RDMA_NODE_IB_CA;
|
|
|
- dev->ib_dev.local_dma_lkey = mdev->caps.gen.reserved_lkey;
|
|
|
- dev->num_ports = mdev->caps.gen.num_ports;
|
|
|
+ dev->ib_dev.local_dma_lkey = 0 /* not supported for now */;
|
|
|
+ dev->num_ports = MLX5_CAP_GEN(mdev, num_ports);
|
|
|
dev->ib_dev.phys_port_cnt = dev->num_ports;
|
|
|
dev->ib_dev.num_comp_vectors =
|
|
|
dev->mdev->priv.eq_table.num_comp_vectors;
|
|
@@ -1286,9 +1281,9 @@ static void *mlx5_ib_add(struct mlx5_core_dev *mdev)
|
|
|
dev->ib_dev.free_fast_reg_page_list = mlx5_ib_free_fast_reg_page_list;
|
|
|
dev->ib_dev.check_mr_status = mlx5_ib_check_mr_status;
|
|
|
|
|
|
- mlx5_ib_internal_query_odp_caps(dev);
|
|
|
+ mlx5_ib_internal_fill_odp_caps(dev);
|
|
|
|
|
|
- if (mdev->caps.gen.flags & MLX5_DEV_CAP_FLAG_XRC) {
|
|
|
+ if (MLX5_CAP_GEN(mdev, xrc)) {
|
|
|
dev->ib_dev.alloc_xrcd = mlx5_ib_alloc_xrcd;
|
|
|
dev->ib_dev.dealloc_xrcd = mlx5_ib_dealloc_xrcd;
|
|
|
dev->ib_dev.uverbs_cmd_mask |=
|