|
@@ -645,7 +645,8 @@ static int nvme_rdma_init_io_queues(struct nvme_rdma_ctrl *ctrl)
|
|
|
int i, ret;
|
|
|
|
|
|
for (i = 1; i < ctrl->queue_count; i++) {
|
|
|
- ret = nvme_rdma_init_queue(ctrl, i, ctrl->ctrl.sqsize);
|
|
|
+ ret = nvme_rdma_init_queue(ctrl, i,
|
|
|
+ ctrl->ctrl.opts->queue_size);
|
|
|
if (ret) {
|
|
|
dev_info(ctrl->ctrl.device,
|
|
|
"failed to initialize i/o queue: %d\n", ret);
|
|
@@ -1286,8 +1287,13 @@ static int nvme_rdma_route_resolved(struct nvme_rdma_queue *queue)
|
|
|
priv.hrqsize = cpu_to_le16(NVMF_AQ_DEPTH);
|
|
|
priv.hsqsize = cpu_to_le16(NVMF_AQ_DEPTH - 1);
|
|
|
} else {
|
|
|
+ /*
|
|
|
+ * current interpretation of the fabrics spec
|
|
|
+ * is at minimum you make hrqsize sqsize+1, or a
|
|
|
+ * 1's based representation of sqsize.
|
|
|
+ */
|
|
|
priv.hrqsize = cpu_to_le16(queue->queue_size);
|
|
|
- priv.hsqsize = cpu_to_le16(queue->queue_size);
|
|
|
+ priv.hsqsize = cpu_to_le16(queue->ctrl->ctrl.sqsize);
|
|
|
}
|
|
|
|
|
|
ret = rdma_connect(queue->cm_id, ¶m);
|
|
@@ -1825,7 +1831,7 @@ static int nvme_rdma_create_io_queues(struct nvme_rdma_ctrl *ctrl)
|
|
|
|
|
|
memset(&ctrl->tag_set, 0, sizeof(ctrl->tag_set));
|
|
|
ctrl->tag_set.ops = &nvme_rdma_mq_ops;
|
|
|
- ctrl->tag_set.queue_depth = ctrl->ctrl.sqsize;
|
|
|
+ ctrl->tag_set.queue_depth = ctrl->ctrl.opts->queue_size;
|
|
|
ctrl->tag_set.reserved_tags = 1; /* fabric connect */
|
|
|
ctrl->tag_set.numa_node = NUMA_NO_NODE;
|
|
|
ctrl->tag_set.flags = BLK_MQ_F_SHOULD_MERGE;
|
|
@@ -1923,7 +1929,7 @@ static struct nvme_ctrl *nvme_rdma_create_ctrl(struct device *dev,
|
|
|
spin_lock_init(&ctrl->lock);
|
|
|
|
|
|
ctrl->queue_count = opts->nr_io_queues + 1; /* +1 for admin queue */
|
|
|
- ctrl->ctrl.sqsize = opts->queue_size;
|
|
|
+ ctrl->ctrl.sqsize = opts->queue_size - 1;
|
|
|
ctrl->ctrl.kato = opts->kato;
|
|
|
|
|
|
ret = -ENOMEM;
|