RDMA/mana_ib: query device capabilities
With RDMA device registered, use it to query on hardware capabilities and cache this information for future query requests to the driver. Signed-off-by: Long Li <longli@microsoft.com> Link: https://lore.kernel.org/r/1702692255-23640-3-git-send-email-longli@linuxonhyperv.com Signed-off-by: Leon Romanovsky <leon@kernel.org>
This commit is contained in:
parent
a7f0636d22
commit
2c20e20b22
@ -26,7 +26,7 @@ int mana_ib_create_cq(struct ib_cq *ibcq, const struct ib_cq_init_attr *attr,
|
||||
return err;
|
||||
}
|
||||
|
||||
if (attr->cqe > MAX_SEND_BUFFERS_PER_QUEUE) {
|
||||
if (attr->cqe > mdev->adapter_caps.max_qp_wr) {
|
||||
ibdev_dbg(ibdev, "CQE %d exceeding limit\n", attr->cqe);
|
||||
return -EINVAL;
|
||||
}
|
||||
|
@ -85,6 +85,13 @@ static int mana_ib_probe(struct auxiliary_device *adev,
|
||||
}
|
||||
dev->gdma_dev = &mdev->gdma_context->mana_ib;
|
||||
|
||||
ret = mana_ib_gd_query_adapter_caps(dev);
|
||||
if (ret) {
|
||||
ibdev_err(&dev->ib_dev, "Failed to query device caps, ret %d",
|
||||
ret);
|
||||
goto deregister_device;
|
||||
}
|
||||
|
||||
ret = ib_register_device(&dev->ib_dev, "mana_%d",
|
||||
mdev->gdma_context->dev);
|
||||
if (ret)
|
||||
|
@ -486,20 +486,17 @@ int mana_ib_get_port_immutable(struct ib_device *ibdev, u32 port_num,
|
||||
int mana_ib_query_device(struct ib_device *ibdev, struct ib_device_attr *props,
|
||||
struct ib_udata *uhw)
|
||||
{
|
||||
props->max_qp = MANA_MAX_NUM_QUEUES;
|
||||
props->max_qp_wr = MAX_SEND_BUFFERS_PER_QUEUE;
|
||||
|
||||
/*
|
||||
* max_cqe could be potentially much bigger.
|
||||
* As this version of driver only support RAW QP, set it to the same
|
||||
* value as max_qp_wr
|
||||
*/
|
||||
props->max_cqe = MAX_SEND_BUFFERS_PER_QUEUE;
|
||||
struct mana_ib_dev *dev = container_of(ibdev,
|
||||
struct mana_ib_dev, ib_dev);
|
||||
|
||||
props->max_qp = dev->adapter_caps.max_qp_count;
|
||||
props->max_qp_wr = dev->adapter_caps.max_qp_wr;
|
||||
props->max_cq = dev->adapter_caps.max_cq_count;
|
||||
props->max_cqe = dev->adapter_caps.max_qp_wr;
|
||||
props->max_mr = dev->adapter_caps.max_mr_count;
|
||||
props->max_mr_size = MANA_IB_MAX_MR_SIZE;
|
||||
props->max_mr = MANA_IB_MAX_MR;
|
||||
props->max_send_sge = MAX_TX_WQE_SGL_ENTRIES;
|
||||
props->max_recv_sge = MAX_RX_WQE_SGL_ENTRIES;
|
||||
props->max_send_sge = dev->adapter_caps.max_send_sge_count;
|
||||
props->max_recv_sge = dev->adapter_caps.max_recv_sge_count;
|
||||
|
||||
return 0;
|
||||
}
|
||||
@ -521,3 +518,45 @@ int mana_ib_query_gid(struct ib_device *ibdev, u32 port, int index,
|
||||
void mana_ib_disassociate_ucontext(struct ib_ucontext *ibcontext)
|
||||
{
|
||||
}
|
||||
|
||||
int mana_ib_gd_query_adapter_caps(struct mana_ib_dev *dev)
|
||||
{
|
||||
struct mana_ib_adapter_caps *caps = &dev->adapter_caps;
|
||||
struct mana_ib_query_adapter_caps_resp resp = {};
|
||||
struct mana_ib_query_adapter_caps_req req = {};
|
||||
int err;
|
||||
|
||||
mana_gd_init_req_hdr(&req.hdr, MANA_IB_GET_ADAPTER_CAP, sizeof(req),
|
||||
sizeof(resp));
|
||||
req.hdr.resp.msg_version = GDMA_MESSAGE_V3;
|
||||
req.hdr.dev_id = dev->gdma_dev->dev_id;
|
||||
|
||||
err = mana_gd_send_request(dev->gdma_dev->gdma_context, sizeof(req),
|
||||
&req, sizeof(resp), &resp);
|
||||
|
||||
if (err) {
|
||||
ibdev_err(&dev->ib_dev,
|
||||
"Failed to query adapter caps err %d", err);
|
||||
return err;
|
||||
}
|
||||
|
||||
caps->max_sq_id = resp.max_sq_id;
|
||||
caps->max_rq_id = resp.max_rq_id;
|
||||
caps->max_cq_id = resp.max_cq_id;
|
||||
caps->max_qp_count = resp.max_qp_count;
|
||||
caps->max_cq_count = resp.max_cq_count;
|
||||
caps->max_mr_count = resp.max_mr_count;
|
||||
caps->max_pd_count = resp.max_pd_count;
|
||||
caps->max_inbound_read_limit = resp.max_inbound_read_limit;
|
||||
caps->max_outbound_read_limit = resp.max_outbound_read_limit;
|
||||
caps->mw_count = resp.mw_count;
|
||||
caps->max_srq_count = resp.max_srq_count;
|
||||
caps->max_qp_wr = min_t(u32,
|
||||
resp.max_requester_sq_size / GDMA_MAX_SQE_SIZE,
|
||||
resp.max_requester_rq_size / GDMA_MAX_RQE_SIZE);
|
||||
caps->max_inline_data_size = resp.max_inline_data_size;
|
||||
caps->max_send_sge_count = resp.max_send_sge_count;
|
||||
caps->max_recv_sge_count = resp.max_recv_sge_count;
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
@ -27,9 +27,28 @@
|
||||
*/
|
||||
#define MANA_IB_MAX_MR 0xFFFFFFu
|
||||
|
||||
struct mana_ib_adapter_caps {
|
||||
u32 max_sq_id;
|
||||
u32 max_rq_id;
|
||||
u32 max_cq_id;
|
||||
u32 max_qp_count;
|
||||
u32 max_cq_count;
|
||||
u32 max_mr_count;
|
||||
u32 max_pd_count;
|
||||
u32 max_inbound_read_limit;
|
||||
u32 max_outbound_read_limit;
|
||||
u32 mw_count;
|
||||
u32 max_srq_count;
|
||||
u32 max_qp_wr;
|
||||
u32 max_send_sge_count;
|
||||
u32 max_recv_sge_count;
|
||||
u32 max_inline_data_size;
|
||||
};
|
||||
|
||||
struct mana_ib_dev {
|
||||
struct ib_device ib_dev;
|
||||
struct gdma_dev *gdma_dev;
|
||||
struct mana_ib_adapter_caps adapter_caps;
|
||||
};
|
||||
|
||||
struct mana_ib_wq {
|
||||
@ -92,6 +111,36 @@ struct mana_ib_rwq_ind_table {
|
||||
struct ib_rwq_ind_table ib_ind_table;
|
||||
};
|
||||
|
||||
enum mana_ib_command_code {
|
||||
MANA_IB_GET_ADAPTER_CAP = 0x30001,
|
||||
};
|
||||
|
||||
struct mana_ib_query_adapter_caps_req {
|
||||
struct gdma_req_hdr hdr;
|
||||
}; /*HW Data */
|
||||
|
||||
struct mana_ib_query_adapter_caps_resp {
|
||||
struct gdma_resp_hdr hdr;
|
||||
u32 max_sq_id;
|
||||
u32 max_rq_id;
|
||||
u32 max_cq_id;
|
||||
u32 max_qp_count;
|
||||
u32 max_cq_count;
|
||||
u32 max_mr_count;
|
||||
u32 max_pd_count;
|
||||
u32 max_inbound_read_limit;
|
||||
u32 max_outbound_read_limit;
|
||||
u32 mw_count;
|
||||
u32 max_srq_count;
|
||||
u32 max_requester_sq_size;
|
||||
u32 max_responder_sq_size;
|
||||
u32 max_requester_rq_size;
|
||||
u32 max_responder_rq_size;
|
||||
u32 max_send_sge_count;
|
||||
u32 max_recv_sge_count;
|
||||
u32 max_inline_data_size;
|
||||
}; /* HW Data */
|
||||
|
||||
int mana_ib_gd_create_dma_region(struct mana_ib_dev *dev, struct ib_umem *umem,
|
||||
mana_handle_t *gdma_region);
|
||||
|
||||
@ -159,4 +208,5 @@ int mana_ib_query_gid(struct ib_device *ibdev, u32 port, int index,
|
||||
|
||||
void mana_ib_disassociate_ucontext(struct ib_ucontext *ibcontext);
|
||||
|
||||
int mana_ib_gd_query_adapter_caps(struct mana_ib_dev *mdev);
|
||||
#endif
|
||||
|
@ -130,7 +130,7 @@ static int mana_ib_create_qp_rss(struct ib_qp *ibqp, struct ib_pd *pd,
|
||||
return ret;
|
||||
}
|
||||
|
||||
if (attr->cap.max_recv_wr > MAX_SEND_BUFFERS_PER_QUEUE) {
|
||||
if (attr->cap.max_recv_wr > mdev->adapter_caps.max_qp_wr) {
|
||||
ibdev_dbg(&mdev->ib_dev,
|
||||
"Requested max_recv_wr %d exceeding limit\n",
|
||||
attr->cap.max_recv_wr);
|
||||
@ -296,7 +296,7 @@ static int mana_ib_create_qp_raw(struct ib_qp *ibqp, struct ib_pd *ibpd,
|
||||
if (port < 1 || port > mc->num_ports)
|
||||
return -EINVAL;
|
||||
|
||||
if (attr->cap.max_send_wr > MAX_SEND_BUFFERS_PER_QUEUE) {
|
||||
if (attr->cap.max_send_wr > mdev->adapter_caps.max_qp_wr) {
|
||||
ibdev_dbg(&mdev->ib_dev,
|
||||
"Requested max_send_wr %d exceeding limit\n",
|
||||
attr->cap.max_send_wr);
|
||||
|
@ -150,6 +150,7 @@ struct gdma_general_req {
|
||||
|
||||
#define GDMA_MESSAGE_V1 1
|
||||
#define GDMA_MESSAGE_V2 2
|
||||
#define GDMA_MESSAGE_V3 3
|
||||
|
||||
struct gdma_general_resp {
|
||||
struct gdma_resp_hdr hdr;
|
||||
|
Loading…
x
Reference in New Issue
Block a user