On Fri, Jul 01, 2016 at 09:01:20PM +0100, Mel Gorman wrote:
> Earlier patches focused on having direct reclaim and kswapd use data that
> is node-centric for reclaiming but shrink_node() itself still uses too
> much zone information.  This patch removes unnecessary zone-based
> information with the most important decision being whether to continue
> reclaim or not.  Some memcg APIs are adjusted as a result even though
> memcg itself still uses some zone information.
> 
> Signed-off-by: Mel Gorman <mgor...@techsingularity.net>
> Acked-by: Michal Hocko <mho...@suse.com>
> Acked-by: Vlastimil Babka <vba...@suse.cz>
> ---
>  include/linux/memcontrol.h | 19 ++++++++--------
>  include/linux/mmzone.h     |  4 ++--
>  include/linux/swap.h       |  2 +-
>  mm/memcontrol.c            |  4 ++--
>  mm/page_alloc.c            |  2 +-
>  mm/vmscan.c                | 57 
> ++++++++++++++++++++++++++--------------------
>  mm/workingset.c            |  6 ++---
>  7 files changed, 51 insertions(+), 43 deletions(-)
> 
> diff --git a/include/linux/memcontrol.h b/include/linux/memcontrol.h
> index 1927dcb6921e..48b43c709ed7 100644
> --- a/include/linux/memcontrol.h
> +++ b/include/linux/memcontrol.h
> @@ -325,22 +325,23 @@ mem_cgroup_zone_zoneinfo(struct mem_cgroup *memcg, 
> struct zone *zone)
>  }
>  
>  /**
> - * mem_cgroup_zone_lruvec - get the lru list vector for a zone and memcg
> + * mem_cgroup_lruvec - get the lru list vector for a node or a memcg zone
> + * @node: node of the wanted lruvec
>   * @zone: zone of the wanted lruvec
>   * @memcg: memcg of the wanted lruvec
>   *
> - * Returns the lru list vector holding pages for the given @zone and
> - * @mem.  This can be the global zone lruvec, if the memory controller
> + * Returns the lru list vector holding pages for a given @node or a given
> + * @memcg and @zone. This can be the node lruvec, if the memory controller
>   * is disabled.
>   */
> -static inline struct lruvec *mem_cgroup_zone_lruvec(struct zone *zone,
> -                                                 struct mem_cgroup *memcg)
> +static inline struct lruvec *mem_cgroup_lruvec(struct pglist_data *pgdat,
> +                             struct zone *zone, struct mem_cgroup *memcg)
>  {
>       struct mem_cgroup_per_zone *mz;
>       struct lruvec *lruvec;
>  
>       if (mem_cgroup_disabled()) {
> -             lruvec = zone_lruvec(zone);
> +             lruvec = node_lruvec(pgdat);
>               goto out;
>       }
>  
> @@ -610,10 +611,10 @@ static inline void mem_cgroup_migrate(struct page *old, 
> struct page *new)
>  {
>  }
>  
> -static inline struct lruvec *mem_cgroup_zone_lruvec(struct zone *zone,
> -                                                 struct mem_cgroup *memcg)
> +static inline struct lruvec *mem_cgroup_lruvec(struct pglist_data *pgdat,
> +                             struct zone *zone, struct mem_cgroup *memcg)
>  {
> -     return zone_lruvec(zone);
> +     return node_lruvec(pgdat);
>  }
>  
>  static inline struct lruvec *mem_cgroup_page_lruvec(struct page *page,
> diff --git a/include/linux/mmzone.h b/include/linux/mmzone.h
> index eb74e63df5cf..f88cbbb476c8 100644
> --- a/include/linux/mmzone.h
> +++ b/include/linux/mmzone.h
> @@ -739,9 +739,9 @@ static inline spinlock_t *zone_lru_lock(struct zone *zone)
>       return &zone->zone_pgdat->lru_lock;
>  }
>  
> -static inline struct lruvec *zone_lruvec(struct zone *zone)
> +static inline struct lruvec *node_lruvec(struct pglist_data *pgdat)
>  {
> -     return &zone->zone_pgdat->lruvec;
> +     return &pgdat->lruvec;
>  }
>  
>  static inline unsigned long pgdat_end_pfn(pg_data_t *pgdat)
> diff --git a/include/linux/swap.h b/include/linux/swap.h
> index 916e2eddecd6..0ad616d7c381 100644
> --- a/include/linux/swap.h
> +++ b/include/linux/swap.h
> @@ -316,7 +316,7 @@ extern unsigned long try_to_free_mem_cgroup_pages(struct 
> mem_cgroup *memcg,
>                                                 unsigned long nr_pages,
>                                                 gfp_t gfp_mask,
>                                                 bool may_swap);
> -extern unsigned long mem_cgroup_shrink_node_zone(struct mem_cgroup *mem,
> +extern unsigned long mem_cgroup_shrink_node(struct mem_cgroup *mem,
>                                               gfp_t gfp_mask, bool noswap,
>                                               struct zone *zone,
>                                               unsigned long *nr_scanned);
> diff --git a/mm/memcontrol.c b/mm/memcontrol.c
> index 50c86ad121bc..c9ebec98e92a 100644
> --- a/mm/memcontrol.c
> +++ b/mm/memcontrol.c
> @@ -1432,8 +1432,8 @@ static int mem_cgroup_soft_reclaim(struct mem_cgroup 
> *root_memcg,
>                       }
>                       continue;
>               }
> -             total += mem_cgroup_shrink_node_zone(victim, gfp_mask, false,
> -                                                  zone, &nr_scanned);
> +             total += mem_cgroup_shrink_node(victim, gfp_mask, false,
> +                                     zone, &nr_scanned);
>               *total_scanned += nr_scanned;
>               if (!soft_limit_excess(root_memcg))
>                       break;
> diff --git a/mm/page_alloc.c b/mm/page_alloc.c
> index f58548139bf2..b76ea2527c09 100644
> --- a/mm/page_alloc.c
> +++ b/mm/page_alloc.c
> @@ -5954,6 +5954,7 @@ static void __paginginit free_area_init_core(struct 
> pglist_data *pgdat)
>  #endif
>       pgdat_page_ext_init(pgdat);
>       spin_lock_init(&pgdat->lru_lock);
> +     lruvec_init(node_lruvec(pgdat));
>  
>       for (j = 0; j < MAX_NR_ZONES; j++) {
>               struct zone *zone = pgdat->node_zones + j;
> @@ -6016,7 +6017,6 @@ static void __paginginit free_area_init_core(struct 
> pglist_data *pgdat)
>               /* For bootup, initialized properly in watermark setup */
>               mod_zone_page_state(zone, NR_ALLOC_BATCH, zone->managed_pages);
>  
> -             lruvec_init(zone_lruvec(zone));
>               if (!size)
>                       continue;
>  
> diff --git a/mm/vmscan.c b/mm/vmscan.c
> index 2f898ba2ee2e..b8e0f76b6e00 100644
> --- a/mm/vmscan.c
> +++ b/mm/vmscan.c
> @@ -2226,10 +2226,11 @@ static inline void init_tlb_ubc(void)
>  /*
>   * This is a basic per-zone page freer.  Used by both kswapd and direct 
> reclaim.
         
                      per-node freer

trivial:

Reply via email to