From: Xiubo Li <[email protected]>

There are at least 6 places are using the same combined GFP flags,
switch them to one macro instead to make the code get cleaner.

Signed-off-by: Xiubo Li <[email protected]>
Reviewed-by: Ming Lei <[email protected]>
---
 block/blk-mq.c | 19 ++++++++++---------
 1 file changed, 10 insertions(+), 9 deletions(-)

diff --git a/block/blk-mq.c b/block/blk-mq.c
index 9c52e4dfe132..3d3b3e5787b0 100644
--- a/block/blk-mq.c
+++ b/block/blk-mq.c
@@ -39,6 +39,8 @@
 #include "blk-mq-sched.h"
 #include "blk-rq-qos.h"
 
+#define BLK_MQ_GFP_FLAGS (GFP_KERNEL | __GFP_NOWARN | __GFP_NORETRY)
+
 static void blk_mq_poll_stats_start(struct request_queue *q);
 static void blk_mq_poll_stats_fn(struct blk_stat_callback *cb);
 
@@ -2091,21 +2093,19 @@ struct blk_mq_tags *blk_mq_alloc_rq_map(struct 
blk_mq_tag_set *set,
 
        tags = blk_mq_init_tags(nr_tags, reserved_tags, node,
                                BLK_MQ_FLAG_TO_ALLOC_POLICY(set->flags),
-                               GFP_NOIO | __GFP_NOWARN | __GFP_NORETRY);
+                               BLK_MQ_GFP_FLAGS);
        if (!tags)
                return NULL;
 
        tags->rqs = kcalloc_node(nr_tags, sizeof(struct request *),
-                                GFP_NOIO | __GFP_NOWARN | __GFP_NORETRY,
-                                node);
+                                BLK_MQ_GFP_FLAGS, node);
        if (!tags->rqs) {
                blk_mq_free_tags(tags);
                return NULL;
        }
 
        tags->static_rqs = kcalloc_node(nr_tags, sizeof(struct request *),
-                                       GFP_NOIO | __GFP_NOWARN | __GFP_NORETRY,
-                                       node);
+                                       BLK_MQ_GFP_FLAGS, node);
        if (!tags->static_rqs) {
                kfree(tags->rqs);
                blk_mq_free_tags(tags);
@@ -2167,7 +2167,7 @@ int blk_mq_alloc_rqs(struct blk_mq_tag_set *set, struct 
blk_mq_tags *tags,
 
                do {
                        page = alloc_pages_node(node,
-                               GFP_NOIO | __GFP_NOWARN | __GFP_NORETRY | 
__GFP_ZERO,
+                               BLK_MQ_GFP_FLAGS | __GFP_ZERO,
                                this_order);
                        if (page)
                                break;
@@ -2188,7 +2188,8 @@ int blk_mq_alloc_rqs(struct blk_mq_tag_set *set, struct 
blk_mq_tags *tags,
                 * Allow kmemleak to scan these pages as they contain pointers
                 * to additional allocations like via ops->init_request().
                 */
-               kmemleak_alloc(p, order_to_size(this_order), 1, GFP_NOIO);
+               kmemleak_alloc(p, order_to_size(this_order), 1,
+                              BLK_MQ_GFP_FLAGS);
                entries_per_page = order_to_size(this_order) / rq_size;
                to_do = min(entries_per_page, depth - i);
                left -= to_do * rq_size;
@@ -2333,7 +2334,7 @@ blk_mq_alloc_hctx(struct request_queue *q, struct 
blk_mq_tag_set *set,
                int node)
 {
        struct blk_mq_hw_ctx *hctx;
-       gfp_t gfp = GFP_NOIO | __GFP_NOWARN | __GFP_NORETRY;
+       gfp_t gfp = BLK_MQ_GFP_FLAGS;
 
        hctx = kzalloc_node(blk_mq_hw_ctx_size(set), gfp, node);
        if (!hctx)
@@ -3194,7 +3195,7 @@ static bool blk_mq_elv_switch_none(struct list_head *head,
        if (!q->elevator)
                return true;
 
-       qe = kmalloc(sizeof(*qe), GFP_NOIO | __GFP_NOWARN | __GFP_NORETRY);
+       qe = kmalloc(sizeof(*qe), BLK_MQ_GFP_FLAGS);
        if (!qe)
                return false;
 
-- 
2.21.0

Reply via email to