blk-mq-sched: Rename blk_mq_sched_free_{requests -> rqs}()
[linux-2.6-microblaze.git] / block / blk-mq-sched.c
index 0f006ca..bdbb6c3 100644 (file)
@@ -515,9 +515,9 @@ void blk_mq_sched_insert_requests(struct blk_mq_hw_ctx *hctx,
        percpu_ref_put(&q->q_usage_counter);
 }
 
-static int blk_mq_sched_alloc_tags(struct request_queue *q,
-                                  struct blk_mq_hw_ctx *hctx,
-                                  unsigned int hctx_idx)
+static int blk_mq_sched_alloc_map_and_rqs(struct request_queue *q,
+                                         struct blk_mq_hw_ctx *hctx,
+                                         unsigned int hctx_idx)
 {
        struct blk_mq_tag_set *set = q->tag_set;
        int ret;
@@ -606,18 +606,18 @@ int blk_mq_init_sched(struct request_queue *q, struct elevator_type *e)
         * Additionally, this is a per-hw queue depth.
         */
        q->nr_requests = 2 * min_t(unsigned int, q->tag_set->queue_depth,
-                                  BLKDEV_MAX_RQ);
+                                  BLKDEV_DEFAULT_RQ);
 
        queue_for_each_hw_ctx(q, hctx, i) {
-               ret = blk_mq_sched_alloc_tags(q, hctx, i);
+               ret = blk_mq_sched_alloc_map_and_rqs(q, hctx, i);
                if (ret)
-                       goto err_free_tags;
+                       goto err_free_map_and_rqs;
        }
 
        if (blk_mq_is_sbitmap_shared(q->tag_set->flags)) {
                ret = blk_mq_init_sched_shared_sbitmap(q);
                if (ret)
-                       goto err_free_tags;
+                       goto err_free_map_and_rqs;
        }
 
        ret = e->ops.init_sched(q, e);
@@ -631,7 +631,7 @@ int blk_mq_init_sched(struct request_queue *q, struct elevator_type *e)
                        ret = e->ops.init_hctx(hctx, i);
                        if (ret) {
                                eq = q->elevator;
-                               blk_mq_sched_free_requests(q);
+                               blk_mq_sched_free_rqs(q);
                                blk_mq_exit_sched(q, eq);
                                kobject_put(&eq->kobj);
                                return ret;
@@ -645,8 +645,8 @@ int blk_mq_init_sched(struct request_queue *q, struct elevator_type *e)
 err_free_sbitmap:
        if (blk_mq_is_sbitmap_shared(q->tag_set->flags))
                blk_mq_exit_sched_shared_sbitmap(q);
-err_free_tags:
-       blk_mq_sched_free_requests(q);
+err_free_map_and_rqs:
+       blk_mq_sched_free_rqs(q);
        blk_mq_sched_tags_teardown(q);
        q->elevator = NULL;
        return ret;
@@ -656,7 +656,7 @@ err_free_tags:
  * called in either blk_queue_cleanup or elevator_switch, tagset
  * is required for freeing requests
  */
-void blk_mq_sched_free_requests(struct request_queue *q)
+void blk_mq_sched_free_rqs(struct request_queue *q)
 {
        struct blk_mq_hw_ctx *hctx;
        int i;