Merge branch 'for-5.9/block' into for-5.9/block-merge
[linux-2.6-microblaze.git] / drivers / nvme / host / rdma.c
index 13506a8..e881f87 100644 (file)
@@ -149,6 +149,7 @@ MODULE_PARM_DESC(register_always,
 static int nvme_rdma_cm_handler(struct rdma_cm_id *cm_id,
                struct rdma_cm_event *event);
 static void nvme_rdma_recv_done(struct ib_cq *cq, struct ib_wc *wc);
+static void nvme_rdma_complete_rq(struct request *rq);
 
 static const struct blk_mq_ops nvme_rdma_mq_ops;
 static const struct blk_mq_ops nvme_rdma_admin_mq_ops;
@@ -1149,6 +1150,16 @@ static void nvme_rdma_error_recovery(struct nvme_rdma_ctrl *ctrl)
        queue_work(nvme_reset_wq, &ctrl->err_work);
 }
 
+static void nvme_rdma_end_request(struct nvme_rdma_request *req)
+{
+       struct request *rq = blk_mq_rq_from_pdu(req);
+
+       if (!refcount_dec_and_test(&req->ref))
+               return;
+       if (!nvme_end_request(rq, req->status, req->result))
+               nvme_rdma_complete_rq(rq);
+}
+
 static void nvme_rdma_wr_error(struct ib_cq *cq, struct ib_wc *wc,
                const char *op)
 {
@@ -1173,16 +1184,11 @@ static void nvme_rdma_inv_rkey_done(struct ib_cq *cq, struct ib_wc *wc)
 {
        struct nvme_rdma_request *req =
                container_of(wc->wr_cqe, struct nvme_rdma_request, reg_cqe);
-       struct request *rq = blk_mq_rq_from_pdu(req);
 
-       if (unlikely(wc->status != IB_WC_SUCCESS)) {
+       if (unlikely(wc->status != IB_WC_SUCCESS))
                nvme_rdma_wr_error(cq, wc, "LOCAL_INV");
-               return;
-       }
-
-       if (refcount_dec_and_test(&req->ref))
-               nvme_end_request(rq, req->status, req->result);
-
+       else
+               nvme_rdma_end_request(req);
 }
 
 static int nvme_rdma_inv_rkey(struct nvme_rdma_queue *queue,
@@ -1547,15 +1553,11 @@ static void nvme_rdma_send_done(struct ib_cq *cq, struct ib_wc *wc)
                container_of(wc->wr_cqe, struct nvme_rdma_qe, cqe);
        struct nvme_rdma_request *req =
                container_of(qe, struct nvme_rdma_request, sqe);
-       struct request *rq = blk_mq_rq_from_pdu(req);
 
-       if (unlikely(wc->status != IB_WC_SUCCESS)) {
+       if (unlikely(wc->status != IB_WC_SUCCESS))
                nvme_rdma_wr_error(cq, wc, "SEND");
-               return;
-       }
-
-       if (refcount_dec_and_test(&req->ref))
-               nvme_end_request(rq, req->status, req->result);
+       else
+               nvme_rdma_end_request(req);
 }
 
 static int nvme_rdma_post_send(struct nvme_rdma_queue *queue,
@@ -1697,8 +1699,7 @@ static void nvme_rdma_process_nvme_rsp(struct nvme_rdma_queue *queue,
                return;
        }
 
-       if (refcount_dec_and_test(&req->ref))
-               nvme_end_request(rq, req->status, req->result);
+       nvme_rdma_end_request(req);
 }
 
 static void nvme_rdma_recv_done(struct ib_cq *cq, struct ib_wc *wc)