On Fri, Apr 25, 2025 at 10:36:55PM +0200, David Hildenbrand wrote:
> On 25.04.25 22:23, Peter Xu wrote:
> > On Fri, Apr 25, 2025 at 10:17:09AM +0200, David Hildenbrand wrote:
> > > Let's use our new interface. In remap_pfn_range(), we'll now decide
> > > whether we have to track (full VMA covered) or only sanitize the pgprot
> > > (partial VMA covered).
> > > 
> > > Remember what we have to untrack by linking it from the VMA. When
> > > duplicating VMAs (e.g., splitting, mremap, fork), we'll handle it similar
> > > to anon VMA names, and use a kref to share the tracking.
> > > 
> > > Once the last VMA un-refs our tracking data, we'll do the untracking,
> > > which simplifies things a lot and should sort our various issues we saw
> > > recently, for example, when partially unmapping/zapping a tracked VMA.
> > > 
> > > This change implies that we'll keep tracking the original PFN range even
> > > after splitting + partially unmapping it: not too bad, because it was
> > > not working reliably before. The only thing that kind-of worked before
> > > was shrinking such a mapping using mremap(): we managed to adjust the
> > > reservation in a hacky way, now we won't adjust the reservation but
> > > leave it around until all involved VMAs are gone.
> > > 
> > > Signed-off-by: David Hildenbrand <da...@redhat.com>
> > > ---
> > >   include/linux/mm_inline.h |  2 +
> > >   include/linux/mm_types.h  | 11 ++++++
> > >   kernel/fork.c             | 54 ++++++++++++++++++++++++--
> > >   mm/memory.c               | 81 +++++++++++++++++++++++++++++++--------
> > >   mm/mremap.c               |  4 --
> > >   5 files changed, 128 insertions(+), 24 deletions(-)
> > > 
> > > diff --git a/include/linux/mm_inline.h b/include/linux/mm_inline.h
> > > index f9157a0c42a5c..89b518ff097e6 100644
> > > --- a/include/linux/mm_inline.h
> > > +++ b/include/linux/mm_inline.h
> > > @@ -447,6 +447,8 @@ static inline bool anon_vma_name_eq(struct 
> > > anon_vma_name *anon_name1,
> > >   #endif  /* CONFIG_ANON_VMA_NAME */
> > > +void pfnmap_track_ctx_release(struct kref *ref);
> > > +
> > >   static inline void init_tlb_flush_pending(struct mm_struct *mm)
> > >   {
> > >           atomic_set(&mm->tlb_flush_pending, 0);
> > > diff --git a/include/linux/mm_types.h b/include/linux/mm_types.h
> > > index 56d07edd01f91..91124761cfda8 100644
> > > --- a/include/linux/mm_types.h
> > > +++ b/include/linux/mm_types.h
> > > @@ -764,6 +764,14 @@ struct vma_numab_state {
> > >           int prev_scan_seq;
> > >   };
> > > +#ifdef __HAVE_PFNMAP_TRACKING
> > > +struct pfnmap_track_ctx {
> > > + struct kref kref;
> > > + unsigned long pfn;
> > > + unsigned long size;
> > > +};
> > > +#endif
> > > +
> > >   /*
> > >    * This struct describes a virtual memory area. There is one of these
> > >    * per VM-area/task. A VM area is any part of the process virtual memory
> > > @@ -877,6 +885,9 @@ struct vm_area_struct {
> > >           struct anon_vma_name *anon_name;
> > >   #endif
> > >           struct vm_userfaultfd_ctx vm_userfaultfd_ctx;
> > > +#ifdef __HAVE_PFNMAP_TRACKING
> > > + struct pfnmap_track_ctx *pfnmap_track_ctx;
> > > +#endif
> > 
> > So this was originally the small concern (or is it small?) that this will
> > grow every vma on x86, am I right?
> 
> Yeah, and last time I looked into this, it would have grown it such that it 
> would
> require a bigger slab. Right now:

Probably due to what config you have.  E.g., when I'm looking mine it's
much bigger and already consuming 256B, but it's because I enabled more
things (userfaultfd, lockdep, etc.).

> 
> Before this change:
> 
> struct vm_area_struct {
>       union {
>               struct {
>                       long unsigned int vm_start;      /*     0     8 */
>                       long unsigned int vm_end;        /*     8     8 */
>               };                                       /*     0    16 */
>               freeptr_t          vm_freeptr;           /*     0     8 */
>       };                                               /*     0    16 */
>       struct mm_struct *         vm_mm;                /*    16     8 */
>       pgprot_t                   vm_page_prot;         /*    24     8 */
>       union {
>               const vm_flags_t   vm_flags;             /*    32     8 */
>               vm_flags_t         __vm_flags;           /*    32     8 */
>       };                                               /*    32     8 */
>       unsigned int               vm_lock_seq;          /*    40     4 */
> 
>       /* XXX 4 bytes hole, try to pack */
> 
>       struct list_head           anon_vma_chain;       /*    48    16 */
>       /* --- cacheline 1 boundary (64 bytes) --- */
>       struct anon_vma *          anon_vma;             /*    64     8 */
>       const struct vm_operations_struct  * vm_ops;     /*    72     8 */
>       long unsigned int          vm_pgoff;             /*    80     8 */
>       struct file *              vm_file;              /*    88     8 */
>       void *                     vm_private_data;      /*    96     8 */
>       atomic_long_t              swap_readahead_info;  /*   104     8 */
>       struct mempolicy *         vm_policy;            /*   112     8 */
>       struct vma_numab_state *   numab_state;          /*   120     8 */
>       /* --- cacheline 2 boundary (128 bytes) --- */
>       refcount_t                 vm_refcnt __attribute__((__aligned__(64))); 
> /*   128     4 */
> 
>       /* XXX 4 bytes hole, try to pack */
> 
>       struct {
>               struct rb_node     rb __attribute__((__aligned__(8))); /*   136 
>    24 */
>               long unsigned int  rb_subtree_last;      /*   160     8 */
>       } __attribute__((__aligned__(8))) shared 
> __attribute__((__aligned__(8)));        /*   136    32 */
>       struct anon_vma_name *     anon_name;            /*   168     8 */
>       struct vm_userfaultfd_ctx  vm_userfaultfd_ctx;   /*   176     0 */
> 
>       /* size: 192, cachelines: 3, members: 18 */
>       /* sum members: 168, holes: 2, sum holes: 8 */
>       /* padding: 16 */
>       /* forced alignments: 2, forced holes: 1, sum forced holes: 4 */
> } __attribute__((__aligned__(64)));
> 
> After this change:
> 
> struct vm_area_struct {
>       union {
>               struct {
>                       long unsigned int vm_start;      /*     0     8 */
>                       long unsigned int vm_end;        /*     8     8 */
>               };                                       /*     0    16 */
>               freeptr_t          vm_freeptr;           /*     0     8 */
>       };                                               /*     0    16 */
>       struct mm_struct *         vm_mm;                /*    16     8 */
>       pgprot_t                   vm_page_prot;         /*    24     8 */
>       union {
>               const vm_flags_t   vm_flags;             /*    32     8 */
>               vm_flags_t         __vm_flags;           /*    32     8 */
>       };                                               /*    32     8 */
>       unsigned int               vm_lock_seq;          /*    40     4 */
> 
>       /* XXX 4 bytes hole, try to pack */
> 
>       struct list_head           anon_vma_chain;       /*    48    16 */
>       /* --- cacheline 1 boundary (64 bytes) --- */
>       struct anon_vma *          anon_vma;             /*    64     8 */
>       const struct vm_operations_struct  * vm_ops;     /*    72     8 */
>       long unsigned int          vm_pgoff;             /*    80     8 */
>       struct file *              vm_file;              /*    88     8 */
>       void *                     vm_private_data;      /*    96     8 */
>       atomic_long_t              swap_readahead_info;  /*   104     8 */
>       struct mempolicy *         vm_policy;            /*   112     8 */
>       struct vma_numab_state *   numab_state;          /*   120     8 */
>       /* --- cacheline 2 boundary (128 bytes) --- */
>       refcount_t                 vm_refcnt __attribute__((__aligned__(64))); 
> /*   128     4 */
> 
>       /* XXX 4 bytes hole, try to pack */
> 
>       struct {
>               struct rb_node     rb __attribute__((__aligned__(8))); /*   136 
>    24 */
>               long unsigned int  rb_subtree_last;      /*   160     8 */
>       } __attribute__((__aligned__(8))) shared 
> __attribute__((__aligned__(8)));        /*   136    32 */
>       struct anon_vma_name *     anon_name;            /*   168     8 */
>       struct vm_userfaultfd_ctx  vm_userfaultfd_ctx;   /*   176     0 */
>       struct pfnmap_track_ctx *  pfnmap_track_ctx;     /*   176     8 */
> 
>       /* size: 192, cachelines: 3, members: 19 */
>       /* sum members: 176, holes: 2, sum holes: 8 */
>       /* padding: 8 */
>       /* forced alignments: 2, forced holes: 1, sum forced holes: 4 */
> } __attribute__((__aligned__(64)));
> 
> Observe that we allocate 192 bytes with or without pfnmap_track_ctx. (IIRC,
> slab sizes are ... 128, 192, 256, 512, ...)

True. I just double checked, vm_area_cachep has SLAB_HWCACHE_ALIGN set, I
think it means it's working like that on x86_64 at least indeed.  So looks
like the new field at least isn't an immediate concern.

Thanks,

-- 
Peter Xu

Reply via email to