RDMA/rdmavt: Decouple QP and SGE lists allocations
authorLeon Romanovsky <leonro@nvidia.com>
Fri, 23 Jul 2021 11:39:49 +0000 (14:39 +0300)
committerJason Gunthorpe <jgg@nvidia.com>
Tue, 3 Aug 2021 16:44:27 +0000 (13:44 -0300)
The rdmavt QP has fields that are both needed for the control and data
path. Such mixed declaration caused to the very specific allocation flow
with kzalloc_node and SGE list embedded into the struct rvt_qp.

This patch separates QP creation to two: regular memory allocation for the
control path and specific code for the SGE list, while the access to the
later is performed through derefenced pointer.

Such pointer and its context are expected to be in the cache, so
performance difference is expected to be negligible, if any exists.

Link: https://lore.kernel.org/r/f66c1e20ccefba0db3c69c58ca9c897f062b4d1c.1627040189.git.leonro@nvidia.com
Signed-off-by: Leon Romanovsky <leonro@nvidia.com>
Signed-off-by: Jason Gunthorpe <jgg@nvidia.com>
drivers/infiniband/sw/rdmavt/qp.c
include/rdma/rdmavt_qp.h

index e9f3d35..1490086 100644 (file)
@@ -1078,7 +1078,7 @@ struct ib_qp *rvt_create_qp(struct ib_pd *ibpd,
        int err;
        struct rvt_swqe *swq = NULL;
        size_t sz;
-       size_t sg_list_sz;
+       size_t sg_list_sz = 0;
        struct ib_qp *ret = ERR_PTR(-ENOMEM);
        struct rvt_dev_info *rdi = ib_to_rvt(ibpd->device);
        void *priv = NULL;
@@ -1126,8 +1126,6 @@ struct ib_qp *rvt_create_qp(struct ib_pd *ibpd,
                if (!swq)
                        return ERR_PTR(-ENOMEM);
 
-               sz = sizeof(*qp);
-               sg_list_sz = 0;
                if (init_attr->srq) {
                        struct rvt_srq *srq = ibsrq_to_rvtsrq(init_attr->srq);
 
@@ -1137,10 +1135,13 @@ struct ib_qp *rvt_create_qp(struct ib_pd *ibpd,
                } else if (init_attr->cap.max_recv_sge > 1)
                        sg_list_sz = sizeof(*qp->r_sg_list) *
                                (init_attr->cap.max_recv_sge - 1);
-               qp = kzalloc_node(sz + sg_list_sz, GFP_KERNEL,
-                                 rdi->dparms.node);
+               qp = kzalloc_node(sizeof(*qp), GFP_KERNEL, rdi->dparms.node);
                if (!qp)
                        goto bail_swq;
+               qp->r_sg_list =
+                       kzalloc_node(sg_list_sz, GFP_KERNEL, rdi->dparms.node);
+               if (!qp->r_sg_list)
+                       goto bail_qp;
                qp->allowed_ops = get_allowed_ops(init_attr->qp_type);
 
                RCU_INIT_POINTER(qp->next, NULL);
@@ -1328,6 +1329,7 @@ bail_driver_priv:
 
 bail_qp:
        kfree(qp->s_ack_queue);
+       kfree(qp->r_sg_list);
        kfree(qp);
 
 bail_swq:
@@ -1762,6 +1764,7 @@ int rvt_destroy_qp(struct ib_qp *ibqp, struct ib_udata *udata)
        kvfree(qp->r_rq.kwq);
        rdi->driver_f.qp_priv_free(rdi, qp);
        kfree(qp->s_ack_queue);
+       kfree(qp->r_sg_list);
        rdma_destroy_ah_attr(&qp->remote_ah_attr);
        rdma_destroy_ah_attr(&qp->alt_ah_attr);
        free_ud_wq_attr(qp);
index 8275954..2e58d5e 100644 (file)
@@ -444,7 +444,7 @@ struct rvt_qp {
        /*
         * This sge list MUST be last. Do not add anything below here.
         */
-       struct rvt_sge r_sg_list[] /* verified SGEs */
+       struct rvt_sge *r_sg_list /* verified SGEs */
                ____cacheline_aligned_in_smp;
 };