nvme-rdma: split nvme_rdma_alloc_tagset
[linux-2.6-microblaze.git] / drivers / nvme / host / rdma.c
index 46c2dcf..3100643 100644 (file)
@@ -29,7 +29,7 @@
 #include "fabrics.h"
 
 
-#define NVME_RDMA_CONNECT_TIMEOUT_MS   3000            /* 3 second */
+#define NVME_RDMA_CM_TIMEOUT_MS                3000            /* 3 second */
 
 #define NVME_RDMA_MAX_SEGMENTS         256
 
@@ -248,12 +248,9 @@ static int nvme_rdma_wait_for_cm(struct nvme_rdma_queue *queue)
 {
        int ret;
 
-       ret = wait_for_completion_interruptible_timeout(&queue->cm_done,
-                       msecs_to_jiffies(NVME_RDMA_CONNECT_TIMEOUT_MS) + 1);
-       if (ret < 0)
+       ret = wait_for_completion_interruptible(&queue->cm_done);
+       if (ret)
                return ret;
-       if (ret == 0)
-               return -ETIMEDOUT;
        WARN_ON_ONCE(queue->cm_error > 0);
        return queue->cm_error;
 }
@@ -612,7 +609,7 @@ static int nvme_rdma_alloc_queue(struct nvme_rdma_ctrl *ctrl,
        queue->cm_error = -ETIMEDOUT;
        ret = rdma_resolve_addr(queue->cm_id, src_addr,
                        (struct sockaddr *)&ctrl->addr,
-                       NVME_RDMA_CONNECT_TIMEOUT_MS);
+                       NVME_RDMA_CM_TIMEOUT_MS);
        if (ret) {
                dev_info(ctrl->ctrl.device,
                        "rdma_resolve_addr failed (%d).\n", ret);
@@ -790,58 +787,62 @@ out_free_queues:
        return ret;
 }
 
-static struct blk_mq_tag_set *nvme_rdma_alloc_tagset(struct nvme_ctrl *nctrl,
-               bool admin)
+static int nvme_rdma_alloc_admin_tag_set(struct nvme_ctrl *nctrl)
 {
        struct nvme_rdma_ctrl *ctrl = to_rdma_ctrl(nctrl);
-       struct blk_mq_tag_set *set;
+       struct blk_mq_tag_set *set = &ctrl->admin_tag_set;
        int ret;
 
-       if (admin) {
-               set = &ctrl->admin_tag_set;
-               memset(set, 0, sizeof(*set));
-               set->ops = &nvme_rdma_admin_mq_ops;
-               set->queue_depth = NVME_AQ_MQ_TAG_DEPTH;
-               set->reserved_tags = NVMF_RESERVED_TAGS;
-               set->numa_node = nctrl->numa_node;
-               set->cmd_size = sizeof(struct nvme_rdma_request) +
-                               NVME_RDMA_DATA_SGL_SIZE;
-               set->driver_data = ctrl;
-               set->nr_hw_queues = 1;
-               set->timeout = NVME_ADMIN_TIMEOUT;
-               set->flags = BLK_MQ_F_NO_SCHED;
-       } else {
-               set = &ctrl->tag_set;
-               memset(set, 0, sizeof(*set));
-               set->ops = &nvme_rdma_mq_ops;
-               set->queue_depth = nctrl->sqsize + 1;
-               set->reserved_tags = NVMF_RESERVED_TAGS;
-               set->numa_node = nctrl->numa_node;
-               set->flags = BLK_MQ_F_SHOULD_MERGE;
-               set->cmd_size = sizeof(struct nvme_rdma_request) +
-                               NVME_RDMA_DATA_SGL_SIZE;
-               if (nctrl->max_integrity_segments)
-                       set->cmd_size += sizeof(struct nvme_rdma_sgl) +
-                                        NVME_RDMA_METADATA_SGL_SIZE;
-               set->driver_data = ctrl;
-               set->nr_hw_queues = nctrl->queue_count - 1;
-               set->timeout = NVME_IO_TIMEOUT;
-               set->nr_maps = nctrl->opts->nr_poll_queues ? HCTX_MAX_TYPES : 2;
-       }
-
+       memset(set, 0, sizeof(*set));
+       set->ops = &nvme_rdma_admin_mq_ops;
+       set->queue_depth = NVME_AQ_MQ_TAG_DEPTH;
+       set->reserved_tags = NVMF_RESERVED_TAGS;
+       set->numa_node = nctrl->numa_node;
+       set->cmd_size = sizeof(struct nvme_rdma_request) +
+                       NVME_RDMA_DATA_SGL_SIZE;
+       set->driver_data = ctrl;
+       set->nr_hw_queues = 1;
+       set->timeout = NVME_ADMIN_TIMEOUT;
+       set->flags = BLK_MQ_F_NO_SCHED;
        ret = blk_mq_alloc_tag_set(set);
-       if (ret)
-               return ERR_PTR(ret);
+       if (!ret)
+               ctrl->ctrl.admin_tagset = set;
+       return ret;
+}
+
+static int nvme_rdma_alloc_tag_set(struct nvme_ctrl *nctrl)
+{
+       struct nvme_rdma_ctrl *ctrl = to_rdma_ctrl(nctrl);
+       struct blk_mq_tag_set *set = &ctrl->tag_set;
+       int ret;
 
-       return set;
+       memset(set, 0, sizeof(*set));
+       set->ops = &nvme_rdma_mq_ops;
+       set->queue_depth = nctrl->sqsize + 1;
+       set->reserved_tags = NVMF_RESERVED_TAGS;
+       set->numa_node = nctrl->numa_node;
+       set->flags = BLK_MQ_F_SHOULD_MERGE;
+       set->cmd_size = sizeof(struct nvme_rdma_request) +
+                       NVME_RDMA_DATA_SGL_SIZE;
+       if (nctrl->max_integrity_segments)
+               set->cmd_size += sizeof(struct nvme_rdma_sgl) +
+                                NVME_RDMA_METADATA_SGL_SIZE;
+       set->driver_data = ctrl;
+       set->nr_hw_queues = nctrl->queue_count - 1;
+       set->timeout = NVME_IO_TIMEOUT;
+       set->nr_maps = nctrl->opts->nr_poll_queues ? HCTX_MAX_TYPES : 2;
+       ret = blk_mq_alloc_tag_set(set);
+       if (!ret)
+               ctrl->ctrl.tagset = set;
+       return ret;
 }
 
 static void nvme_rdma_destroy_admin_queue(struct nvme_rdma_ctrl *ctrl,
                bool remove)
 {
        if (remove) {
-               blk_cleanup_queue(ctrl->ctrl.admin_q);
-               blk_cleanup_queue(ctrl->ctrl.fabrics_q);
+               blk_mq_destroy_queue(ctrl->ctrl.admin_q);
+               blk_mq_destroy_queue(ctrl->ctrl.fabrics_q);
                blk_mq_free_tag_set(ctrl->ctrl.admin_tagset);
        }
        if (ctrl->async_event_sqe.data) {
@@ -885,11 +886,9 @@ static int nvme_rdma_configure_admin_queue(struct nvme_rdma_ctrl *ctrl,
                goto out_free_queue;
 
        if (new) {
-               ctrl->ctrl.admin_tagset = nvme_rdma_alloc_tagset(&ctrl->ctrl, true);
-               if (IS_ERR(ctrl->ctrl.admin_tagset)) {
-                       error = PTR_ERR(ctrl->ctrl.admin_tagset);
+               error = nvme_rdma_alloc_admin_tag_set(&ctrl->ctrl);
+               if (error)
                        goto out_free_async_qe;
-               }
 
                ctrl->ctrl.fabrics_q = blk_mq_init_queue(&ctrl->admin_tag_set);
                if (IS_ERR(ctrl->ctrl.fabrics_q)) {
@@ -935,10 +934,10 @@ out_stop_queue:
        nvme_cancel_admin_tagset(&ctrl->ctrl);
 out_cleanup_queue:
        if (new)
-               blk_cleanup_queue(ctrl->ctrl.admin_q);
+               blk_mq_destroy_queue(ctrl->ctrl.admin_q);
 out_cleanup_fabrics_q:
        if (new)
-               blk_cleanup_queue(ctrl->ctrl.fabrics_q);
+               blk_mq_destroy_queue(ctrl->ctrl.fabrics_q);
 out_free_tagset:
        if (new)
                blk_mq_free_tag_set(ctrl->ctrl.admin_tagset);
@@ -957,7 +956,7 @@ static void nvme_rdma_destroy_io_queues(struct nvme_rdma_ctrl *ctrl,
                bool remove)
 {
        if (remove) {
-               blk_cleanup_queue(ctrl->ctrl.connect_q);
+               blk_mq_destroy_queue(ctrl->ctrl.connect_q);
                blk_mq_free_tag_set(ctrl->ctrl.tagset);
        }
        nvme_rdma_free_io_queues(ctrl);
@@ -972,11 +971,9 @@ static int nvme_rdma_configure_io_queues(struct nvme_rdma_ctrl *ctrl, bool new)
                return ret;
 
        if (new) {
-               ctrl->ctrl.tagset = nvme_rdma_alloc_tagset(&ctrl->ctrl, false);
-               if (IS_ERR(ctrl->ctrl.tagset)) {
-                       ret = PTR_ERR(ctrl->ctrl.tagset);
+               ret = nvme_rdma_alloc_tag_set(&ctrl->ctrl);
+               if (ret)
                        goto out_free_io_queues;
-               }
 
                ret = nvme_ctrl_init_connect_q(&(ctrl->ctrl));
                if (ret)
@@ -1012,7 +1009,7 @@ out_wait_freeze_timed_out:
 out_cleanup_connect_q:
        nvme_cancel_tagset(&ctrl->ctrl);
        if (new)
-               blk_cleanup_queue(ctrl->ctrl.connect_q);
+               blk_mq_destroy_queue(ctrl->ctrl.connect_q);
 out_free_tag_set:
        if (new)
                blk_mq_free_tag_set(ctrl->ctrl.tagset);
@@ -1205,6 +1202,7 @@ static void nvme_rdma_error_recovery_work(struct work_struct *work)
        struct nvme_rdma_ctrl *ctrl = container_of(work,
                        struct nvme_rdma_ctrl, err_work);
 
+       nvme_auth_stop(&ctrl->ctrl);
        nvme_stop_keep_alive(&ctrl->ctrl);
        flush_work(&ctrl->ctrl.async_event_work);
        nvme_rdma_teardown_io_queues(ctrl, false);
@@ -1894,7 +1892,7 @@ static int nvme_rdma_addr_resolved(struct nvme_rdma_queue *queue)
 
        if (ctrl->opts->tos >= 0)
                rdma_set_service_type(queue->cm_id, ctrl->opts->tos);
-       ret = rdma_resolve_route(queue->cm_id, NVME_RDMA_CONNECT_TIMEOUT_MS);
+       ret = rdma_resolve_route(queue->cm_id, NVME_RDMA_CM_TIMEOUT_MS);
        if (ret) {
                dev_err(ctrl->device, "rdma_resolve_route failed (%d).\n",
                        queue->cm_error);
@@ -2021,8 +2019,7 @@ static void nvme_rdma_complete_timed_out(struct request *rq)
        nvmf_complete_timed_out_request(rq);
 }
 
-static enum blk_eh_timer_return
-nvme_rdma_timeout(struct request *rq, bool reserved)
+static enum blk_eh_timer_return nvme_rdma_timeout(struct request *rq)
 {
        struct nvme_rdma_request *req = blk_mq_rq_to_pdu(rq);
        struct nvme_rdma_queue *queue = req->queue;