"Aneesh Kumar K.V" <aneesh.ku...@linux.vnet.ibm.com> writes:

> While computing slice mask for the free area we need make sure we only search
> in the addr limit applicable for this mmap. We update the slb_addr_limit
> after we request for a mmap above 128TB. But the following mmap request
> with hint addr below 128TB should still limit its search to below 128TB. ie.
> we should not use slb_addr_limit to compute slice mask in this case. Instead,
> we should derive high addr limit based on the mmap hint addr value.
>
> Signed-off-by: Aneesh Kumar K.V <aneesh.ku...@linux.vnet.ibm.com>
> ---
>  arch/powerpc/mm/slice.c | 34 ++++++++++++++++++++++------------
>  1 file changed, 22 insertions(+), 12 deletions(-)

How does this relate to the fixes Nick has sent?

cheers

> diff --git a/arch/powerpc/mm/slice.c b/arch/powerpc/mm/slice.c
> index 564fff06f5c1..23ec2c5e3b78 100644
> --- a/arch/powerpc/mm/slice.c
> +++ b/arch/powerpc/mm/slice.c
> @@ -122,7 +122,8 @@ static int slice_high_has_vma(struct mm_struct *mm, 
> unsigned long slice)
>       return !slice_area_is_free(mm, start, end - start);
>  }
>  
> -static void slice_mask_for_free(struct mm_struct *mm, struct slice_mask *ret)
> +static void slice_mask_for_free(struct mm_struct *mm, struct slice_mask *ret,
> +                             unsigned long high_limit)
>  {
>       unsigned long i;
>  
> @@ -133,15 +134,16 @@ static void slice_mask_for_free(struct mm_struct *mm, 
> struct slice_mask *ret)
>               if (!slice_low_has_vma(mm, i))
>                       ret->low_slices |= 1u << i;
>  
> -     if (mm->context.slb_addr_limit <= SLICE_LOW_TOP)
> +     if (high_limit <= SLICE_LOW_TOP)
>               return;
>  
> -     for (i = 0; i < GET_HIGH_SLICE_INDEX(mm->context.slb_addr_limit); i++)
> +     for (i = 0; i < GET_HIGH_SLICE_INDEX(high_limit); i++)
>               if (!slice_high_has_vma(mm, i))
>                       __set_bit(i, ret->high_slices);
>  }
>  
> -static void slice_mask_for_size(struct mm_struct *mm, int psize, struct 
> slice_mask *ret)
> +static void slice_mask_for_size(struct mm_struct *mm, int psize, struct 
> slice_mask *ret,
> +                             unsigned long high_limit)
>  {
>       unsigned char *hpsizes;
>       int index, mask_index;
> @@ -156,8 +158,11 @@ static void slice_mask_for_size(struct mm_struct *mm, 
> int psize, struct slice_ma
>               if (((lpsizes >> (i * 4)) & 0xf) == psize)
>                       ret->low_slices |= 1u << i;
>  
> +     if (high_limit <= SLICE_LOW_TOP)
> +             return;
> +
>       hpsizes = mm->context.high_slices_psize;
> -     for (i = 0; i < GET_HIGH_SLICE_INDEX(mm->context.slb_addr_limit); i++) {
> +     for (i = 0; i < GET_HIGH_SLICE_INDEX(high_limit); i++) {
>               mask_index = i & 0x1;
>               index = i >> 1;
>               if (((hpsizes[index] >> (mask_index * 4)) & 0xf) == psize)
> @@ -169,6 +174,10 @@ static int slice_check_fit(struct mm_struct *mm,
>                          struct slice_mask mask, struct slice_mask available)
>  {
>       DECLARE_BITMAP(result, SLICE_NUM_HIGH);
> +     /*
> +      * Make sure we just do bit compare only to the max
> +      * addr limit and not the full bit map size.
> +      */
>       unsigned long slice_count = 
> GET_HIGH_SLICE_INDEX(mm->context.slb_addr_limit);
>  
>       bitmap_and(result, mask.high_slices,
> @@ -472,7 +481,7 @@ unsigned long slice_get_unmapped_area(unsigned long addr, 
> unsigned long len,
>       /* First make up a "good" mask of slices that have the right size
>        * already
>        */
> -     slice_mask_for_size(mm, psize, &good_mask);
> +     slice_mask_for_size(mm, psize, &good_mask, high_limit);
>       slice_print_mask(" good_mask", good_mask);
>  
>       /*
> @@ -497,7 +506,7 @@ unsigned long slice_get_unmapped_area(unsigned long addr, 
> unsigned long len,
>  #ifdef CONFIG_PPC_64K_PAGES
>       /* If we support combo pages, we can allow 64k pages in 4k slices */
>       if (psize == MMU_PAGE_64K) {
> -             slice_mask_for_size(mm, MMU_PAGE_4K, &compat_mask);
> +             slice_mask_for_size(mm, MMU_PAGE_4K, &compat_mask, high_limit);
>               if (fixed)
>                       slice_or_mask(&good_mask, &compat_mask);
>       }
> @@ -530,11 +539,11 @@ unsigned long slice_get_unmapped_area(unsigned long 
> addr, unsigned long len,
>                       return newaddr;
>               }
>       }
> -
> -     /* We don't fit in the good mask, check what other slices are
> +     /*
> +      * We don't fit in the good mask, check what other slices are
>        * empty and thus can be converted
>        */
> -     slice_mask_for_free(mm, &potential_mask);
> +     slice_mask_for_free(mm, &potential_mask, high_limit);
>       slice_or_mask(&potential_mask, &good_mask);
>       slice_print_mask(" potential", potential_mask);
>  
> @@ -744,17 +753,18 @@ int is_hugepage_only_range(struct mm_struct *mm, 
> unsigned long addr,
>  {
>       struct slice_mask mask, available;
>       unsigned int psize = mm->context.user_psize;
> +     unsigned long high_limit = mm->context.slb_addr_limit;
>  
>       if (radix_enabled())
>               return 0;
>  
>       slice_range_to_mask(addr, len, &mask);
> -     slice_mask_for_size(mm, psize, &available);
> +     slice_mask_for_size(mm, psize, &available, high_limit);
>  #ifdef CONFIG_PPC_64K_PAGES
>       /* We need to account for 4k slices too */
>       if (psize == MMU_PAGE_64K) {
>               struct slice_mask compat_mask;
> -             slice_mask_for_size(mm, MMU_PAGE_4K, &compat_mask);
> +             slice_mask_for_size(mm, MMU_PAGE_4K, &compat_mask, high_limit);
>               slice_or_mask(&available, &compat_mask);
>       }
>  #endif
> -- 
> 2.13.6

Reply via email to