On Tue, Mar 05, 2024 at 12:37:49PM +0800, pet...@redhat.com wrote:
> From: Peter Xu <pet...@redhat.com>
> 
> They're not used anymore, drop all of them.
> 
> Reviewed-by: Jason Gunthorpe <j...@nvidia.com>
> Signed-off-by: Peter Xu <pet...@redhat.com>

Reviewed-by: Mike Rapoport (IBM) <r...@kernel.org>

> ---
>  arch/arm/include/asm/pgtable-2level.h        |  1 -
>  arch/arm/include/asm/pgtable-3level.h        |  1 -
>  arch/loongarch/kvm/mmu.c                     |  2 +-
>  arch/powerpc/include/asm/book3s/64/pgtable.h |  4 +---
>  arch/powerpc/include/asm/pgtable.h           |  4 ----
>  arch/s390/include/asm/pgtable.h              |  8 ++++----
>  arch/sparc/include/asm/pgtable_64.h          |  8 ++++----
>  arch/x86/include/asm/pgtable.h               | 19 +++++++------------
>  arch/x86/kvm/mmu/mmu.c                       |  2 +-
>  9 files changed, 18 insertions(+), 31 deletions(-)
> 
> diff --git a/arch/arm/include/asm/pgtable-2level.h 
> b/arch/arm/include/asm/pgtable-2level.h
> index ce543cd9380c..b0a262566eb9 100644
> --- a/arch/arm/include/asm/pgtable-2level.h
> +++ b/arch/arm/include/asm/pgtable-2level.h
> @@ -213,7 +213,6 @@ static inline pmd_t *pmd_offset(pud_t *pud, unsigned long 
> addr)
>  
>  #define pmd_pfn(pmd)         (__phys_to_pfn(pmd_val(pmd) & PHYS_MASK))
>  
> -#define pmd_large(pmd)               (pmd_val(pmd) & 2)
>  #define pmd_leaf(pmd)                (pmd_val(pmd) & 2)
>  #define pmd_bad(pmd)         (pmd_val(pmd) & 2)
>  #define pmd_present(pmd)     (pmd_val(pmd))
> diff --git a/arch/arm/include/asm/pgtable-3level.h 
> b/arch/arm/include/asm/pgtable-3level.h
> index 71c3add6417f..4b1d9eb3908a 100644
> --- a/arch/arm/include/asm/pgtable-3level.h
> +++ b/arch/arm/include/asm/pgtable-3level.h
> @@ -118,7 +118,6 @@
>                                                PMD_TYPE_TABLE)
>  #define pmd_sect(pmd)                ((pmd_val(pmd) & PMD_TYPE_MASK) == \
>                                                PMD_TYPE_SECT)
> -#define pmd_large(pmd)               pmd_sect(pmd)
>  #define pmd_leaf(pmd)                pmd_sect(pmd)
>  
>  #define pud_clear(pudp)                      \
> diff --git a/arch/loongarch/kvm/mmu.c b/arch/loongarch/kvm/mmu.c
> index 50a6acd7ffe4..a556cff35740 100644
> --- a/arch/loongarch/kvm/mmu.c
> +++ b/arch/loongarch/kvm/mmu.c
> @@ -723,7 +723,7 @@ static int host_pfn_mapping_level(struct kvm *kvm, gfn_t 
> gfn,
>       /*
>        * Read each entry once.  As above, a non-leaf entry can be promoted to
>        * a huge page _during_ this walk.  Re-reading the entry could send the
> -      * walk into the weeks, e.g. p*d_large() returns false (sees the old
> +      * walk into the weeks, e.g. p*d_leaf() returns false (sees the old
>        * value) and then p*d_offset() walks into the target huge page instead
>        * of the old page table (sees the new value).
>        */
> diff --git a/arch/powerpc/include/asm/book3s/64/pgtable.h 
> b/arch/powerpc/include/asm/book3s/64/pgtable.h
> index 3e99e409774a..df66dce8306f 100644
> --- a/arch/powerpc/include/asm/book3s/64/pgtable.h
> +++ b/arch/powerpc/include/asm/book3s/64/pgtable.h
> @@ -1437,17 +1437,15 @@ static inline bool is_pte_rw_upgrade(unsigned long 
> old_val, unsigned long new_va
>  }
>  
>  /*
> - * Like pmd_huge() and pmd_large(), but works regardless of config options
> + * Like pmd_huge(), but works regardless of config options
>   */
>  #define pmd_leaf pmd_leaf
> -#define pmd_large pmd_leaf
>  static inline bool pmd_leaf(pmd_t pmd)
>  {
>       return !!(pmd_raw(pmd) & cpu_to_be64(_PAGE_PTE));
>  }
>  
>  #define pud_leaf pud_leaf
> -#define pud_large pud_leaf
>  static inline bool pud_leaf(pud_t pud)
>  {
>       return !!(pud_raw(pud) & cpu_to_be64(_PAGE_PTE));
> diff --git a/arch/powerpc/include/asm/pgtable.h 
> b/arch/powerpc/include/asm/pgtable.h
> index e6edf1cdbc5b..239709a2f68e 100644
> --- a/arch/powerpc/include/asm/pgtable.h
> +++ b/arch/powerpc/include/asm/pgtable.h
> @@ -101,10 +101,6 @@ void poking_init(void);
>  extern unsigned long ioremap_bot;
>  extern const pgprot_t protection_map[16];
>  
> -#ifndef pmd_large
> -#define pmd_large(pmd)               0
> -#endif
> -
>  /* can we use this in kvm */
>  unsigned long vmalloc_to_phys(void *vmalloc_addr);
>  
> diff --git a/arch/s390/include/asm/pgtable.h b/arch/s390/include/asm/pgtable.h
> index a5f16a244a64..9e08af5b9247 100644
> --- a/arch/s390/include/asm/pgtable.h
> +++ b/arch/s390/include/asm/pgtable.h
> @@ -705,16 +705,16 @@ static inline int pud_none(pud_t pud)
>       return pud_val(pud) == _REGION3_ENTRY_EMPTY;
>  }
>  
> -#define pud_leaf     pud_large
> -static inline int pud_large(pud_t pud)
> +#define pud_leaf pud_leaf
> +static inline int pud_leaf(pud_t pud)
>  {
>       if ((pud_val(pud) & _REGION_ENTRY_TYPE_MASK) != _REGION_ENTRY_TYPE_R3)
>               return 0;
>       return !!(pud_val(pud) & _REGION3_ENTRY_LARGE);
>  }
>  
> -#define pmd_leaf     pmd_large
> -static inline int pmd_large(pmd_t pmd)
> +#define pmd_leaf pmd_leaf
> +static inline int pmd_leaf(pmd_t pmd)
>  {
>       return (pmd_val(pmd) & _SEGMENT_ENTRY_LARGE) != 0;
>  }
> diff --git a/arch/sparc/include/asm/pgtable_64.h 
> b/arch/sparc/include/asm/pgtable_64.h
> index 652af9d63fa2..6ff0a28d5fd1 100644
> --- a/arch/sparc/include/asm/pgtable_64.h
> +++ b/arch/sparc/include/asm/pgtable_64.h
> @@ -680,8 +680,8 @@ static inline unsigned long pte_special(pte_t pte)
>       return pte_val(pte) & _PAGE_SPECIAL;
>  }
>  
> -#define pmd_leaf     pmd_large
> -static inline unsigned long pmd_large(pmd_t pmd)
> +#define pmd_leaf pmd_leaf
> +static inline unsigned long pmd_leaf(pmd_t pmd)
>  {
>       pte_t pte = __pte(pmd_val(pmd));
>  
> @@ -867,8 +867,8 @@ static inline pmd_t *pud_pgtable(pud_t pud)
>  /* only used by the stubbed out hugetlb gup code, should never be called */
>  #define p4d_page(p4d)                        NULL
>  
> -#define pud_leaf     pud_large
> -static inline unsigned long pud_large(pud_t pud)
> +#define pud_leaf pud_leaf
> +static inline unsigned long pud_leaf(pud_t pud)
>  {
>       pte_t pte = __pte(pud_val(pud));
>  
> diff --git a/arch/x86/include/asm/pgtable.h b/arch/x86/include/asm/pgtable.h
> index 9db7a38a0e9f..cfc84c55d0e6 100644
> --- a/arch/x86/include/asm/pgtable.h
> +++ b/arch/x86/include/asm/pgtable.h
> @@ -251,8 +251,8 @@ static inline unsigned long pgd_pfn(pgd_t pgd)
>       return (pgd_val(pgd) & PTE_PFN_MASK) >> PAGE_SHIFT;
>  }
>  
> -#define p4d_leaf     p4d_large
> -static inline int p4d_large(p4d_t p4d)
> +#define p4d_leaf p4d_leaf
> +static inline int p4d_leaf(p4d_t p4d)
>  {
>       /* No 512 GiB pages yet */
>       return 0;
> @@ -260,14 +260,14 @@ static inline int p4d_large(p4d_t p4d)
>  
>  #define pte_page(pte)        pfn_to_page(pte_pfn(pte))
>  
> -#define pmd_leaf     pmd_large
> -static inline int pmd_large(pmd_t pte)
> +#define pmd_leaf pmd_leaf
> +static inline int pmd_leaf(pmd_t pte)
>  {
>       return pmd_flags(pte) & _PAGE_PSE;
>  }
>  
>  #ifdef CONFIG_TRANSPARENT_HUGEPAGE
> -/* NOTE: when predicate huge page, consider also pmd_devmap, or use 
> pmd_large */
> +/* NOTE: when predicate huge page, consider also pmd_devmap, or use pmd_leaf 
> */
>  static inline int pmd_trans_huge(pmd_t pmd)
>  {
>       return (pmd_val(pmd) & (_PAGE_PSE|_PAGE_DEVMAP)) == _PAGE_PSE;
> @@ -1085,8 +1085,8 @@ static inline pmd_t *pud_pgtable(pud_t pud)
>   */
>  #define pud_page(pud)        pfn_to_page(pud_pfn(pud))
>  
> -#define pud_leaf     pud_large
> -static inline int pud_large(pud_t pud)
> +#define pud_leaf pud_leaf
> +static inline int pud_leaf(pud_t pud)
>  {
>       return (pud_val(pud) & (_PAGE_PSE | _PAGE_PRESENT)) ==
>               (_PAGE_PSE | _PAGE_PRESENT);
> @@ -1096,11 +1096,6 @@ static inline int pud_bad(pud_t pud)
>  {
>       return (pud_flags(pud) & ~(_KERNPG_TABLE | _PAGE_USER)) != 0;
>  }
> -#else
> -static inline int pud_large(pud_t pud)
> -{
> -     return 0;
> -}
>  #endif       /* CONFIG_PGTABLE_LEVELS > 2 */
>  
>  #if CONFIG_PGTABLE_LEVELS > 3
> diff --git a/arch/x86/kvm/mmu/mmu.c b/arch/x86/kvm/mmu/mmu.c
> index 5cb5bc4a72c4..58f5e6b637b4 100644
> --- a/arch/x86/kvm/mmu/mmu.c
> +++ b/arch/x86/kvm/mmu/mmu.c
> @@ -3110,7 +3110,7 @@ static int host_pfn_mapping_level(struct kvm *kvm, 
> gfn_t gfn,
>       /*
>        * Read each entry once.  As above, a non-leaf entry can be promoted to
>        * a huge page _during_ this walk.  Re-reading the entry could send the
> -      * walk into the weeks, e.g. p*d_large() returns false (sees the old
> +      * walk into the weeks, e.g. p*d_leaf() returns false (sees the old
>        * value) and then p*d_offset() walks into the target huge page instead
>        * of the old page table (sees the new value).
>        */
> -- 
> 2.44.0
> 
> 

-- 
Sincerely yours,
Mike.

Reply via email to