On Fri, Nov 16, 2012 at 07:38:41PM -0800, Yinghai Lu wrote:
> Now init_memory_mapping is called two times, later will be called for every
> ram ranges.

What is 'later'? Can you say in which of the patches it will
be called for every RAM range?

> 
> Could put all related init_mem calling together and out of setup.c.

That looks like a stray comment.

> 
> Actually, it reverts commit 1bbbbe7
>     x86: Exclude E820_RESERVED regions and memory holes above 4 GB from 
> direct mapping.
> will address that later with complete solution include handling hole under 4g.

.. I presume that the later is one of the patches. Can you say:
'will address that in patch XYZ' ?

> 
> Signed-off-by: Yinghai Lu <ying...@kernel.org>
> Reviewed-by: Pekka Enberg <penb...@kernel.org>
> ---
>  arch/x86/include/asm/init.h    |    1 -
>  arch/x86/include/asm/pgtable.h |    2 +-
>  arch/x86/kernel/setup.c        |   27 +--------------------------
>  arch/x86/mm/init.c             |   19 ++++++++++++++++++-
>  4 files changed, 20 insertions(+), 29 deletions(-)
> 
> diff --git a/arch/x86/include/asm/init.h b/arch/x86/include/asm/init.h
> index adcc0ae..4f13998 100644
> --- a/arch/x86/include/asm/init.h
> +++ b/arch/x86/include/asm/init.h
> @@ -12,7 +12,6 @@ kernel_physical_mapping_init(unsigned long start,
>                            unsigned long end,
>                            unsigned long page_size_mask);
>  
> -

Stray whitespace.

>  extern unsigned long __initdata pgt_buf_start;
>  extern unsigned long __meminitdata pgt_buf_end;
>  extern unsigned long __meminitdata pgt_buf_top;
> diff --git a/arch/x86/include/asm/pgtable.h b/arch/x86/include/asm/pgtable.h
> index 98ac76d..dd1a888 100644
> --- a/arch/x86/include/asm/pgtable.h
> +++ b/arch/x86/include/asm/pgtable.h
> @@ -602,7 +602,7 @@ static inline int pgd_none(pgd_t pgd)
>  #ifndef __ASSEMBLY__
>  
>  extern int direct_gbpages;
> -void probe_page_size_mask(void);
> +void init_mem_mapping(void);
>  
>  /* local pte updates need not use xchg for locking */
>  static inline pte_t native_local_ptep_get_and_clear(pte_t *ptep)
> diff --git a/arch/x86/kernel/setup.c b/arch/x86/kernel/setup.c
> index 01fb5f9..23b079f 100644
> --- a/arch/x86/kernel/setup.c
> +++ b/arch/x86/kernel/setup.c
> @@ -913,34 +913,9 @@ void __init setup_arch(char **cmdline_p)
>       setup_real_mode();
>  
>       init_gbpages();
> -     probe_page_size_mask();
>  
> -     /* max_pfn_mapped is updated here */
> -     max_low_pfn_mapped = init_memory_mapping(0, max_low_pfn<<PAGE_SHIFT);
> -     max_pfn_mapped = max_low_pfn_mapped;
> +     init_mem_mapping();
>  
> -#ifdef CONFIG_X86_64
> -     if (max_pfn > max_low_pfn) {
> -             int i;
> -             unsigned long start, end;
> -             unsigned long start_pfn, end_pfn;
> -
> -             for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn,
> -                                                      NULL) {
> -
> -                     end = PFN_PHYS(end_pfn);
> -                     if (end <= (1UL<<32))
> -                             continue;
> -
> -                     start = PFN_PHYS(start_pfn);
> -                     max_pfn_mapped = init_memory_mapping(
> -                                             max((1UL<<32), start), end);
> -             }
> -
> -             /* can we preseve max_low_pfn ?*/
> -             max_low_pfn = max_pfn;
> -     }
> -#endif
>       memblock.current_limit = get_max_mapped();
>       dma_contiguous_reserve(0);
>  
> diff --git a/arch/x86/mm/init.c b/arch/x86/mm/init.c
> index 701abbc..9e17f9e 100644
> --- a/arch/x86/mm/init.c
> +++ b/arch/x86/mm/init.c
> @@ -37,7 +37,7 @@ struct map_range {
>  
>  static int page_size_mask;
>  
> -void probe_page_size_mask(void)
> +static void __init probe_page_size_mask(void)
>  {
>  #if !defined(CONFIG_DEBUG_PAGEALLOC) && !defined(CONFIG_KMEMCHECK)
>       /*
> @@ -315,6 +315,23 @@ unsigned long __init_refok init_memory_mapping(unsigned 
> long start,
>       return ret >> PAGE_SHIFT;
>  }
>  
> +void __init init_mem_mapping(void)
> +{
> +     probe_page_size_mask();
> +
> +     /* max_pfn_mapped is updated here */
> +     max_low_pfn_mapped = init_memory_mapping(0, max_low_pfn<<PAGE_SHIFT);
> +     max_pfn_mapped = max_low_pfn_mapped;
> +
> +#ifdef CONFIG_X86_64
> +     if (max_pfn > max_low_pfn) {
> +             max_pfn_mapped = init_memory_mapping(1UL<<32,
> +                                                  max_pfn<<PAGE_SHIFT);
> +             /* can we preseve max_low_pfn ?*/
> +             max_low_pfn = max_pfn;
> +     }
> +#endif
> +}
>  
>  /*
>   * devmem_is_allowed() checks to see if /dev/mem access to a certain address
> -- 
> 1.7.7
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Reply via email to