Hi Daniel, Thank you for this info, we will fix this issue.
Almost miss this mail, sorry! -Yonghua > -----Original Message----- > From: Daniel Vetter <dan...@ffwll.ch> > Sent: Wednesday, August 10, 2022 20:20 > To: Huang, Yonghua <yonghua.hu...@intel.com> > Cc: gre...@linuxfoundation.org; linux-ker...@vger.kernel.org; > sta...@vger.kernel.org; Chatre, Reinette <reinette.cha...@intel.com>; Wang, > Zhi A <zhi.a.w...@intel.com>; Wang, Yu1 <yu1.w...@intel.com>; Li, Fei1 > <fei1...@intel.com>; Linux MM <linux...@kvack.org>; DRI Development <dri- > de...@lists.freedesktop.org> > Subject: Re: [PATCH] virt: acrn: obtain pa from VMA with PFNMAP flag > > On Mon, Feb 28, 2022 at 05:22:12AM +0300, Yonghua Huang wrote: > > acrn_vm_ram_map can't pin the user pages with VM_PFNMAP flag by > > calling get_user_pages_fast(), the PA(physical pages) may be mapped > > by kernel driver and set PFNMAP flag. > > > > This patch fixes logic to setup EPT mapping for PFN mapped RAM region > > by checking the memory attribute before adding EPT mapping for them. > > > > Fixes: 88f537d5e8dd ("virt: acrn: Introduce EPT mapping management") > > Signed-off-by: Yonghua Huang <yonghua.hu...@intel.com> > > Signed-off-by: Fei Li <fei1...@intel.com> > > --- > > drivers/virt/acrn/mm.c | 24 ++++++++++++++++++++++++ > > 1 file changed, 24 insertions(+) > > > > diff --git a/drivers/virt/acrn/mm.c b/drivers/virt/acrn/mm.c index > > c4f2e15c8a2b..3b1b1e7a844b 100644 > > --- a/drivers/virt/acrn/mm.c > > +++ b/drivers/virt/acrn/mm.c > > @@ -162,10 +162,34 @@ int acrn_vm_ram_map(struct acrn_vm *vm, struct > acrn_vm_memmap *memmap) > > void *remap_vaddr; > > int ret, pinned; > > u64 user_vm_pa; > > + unsigned long pfn; > > + struct vm_area_struct *vma; > > > > if (!vm || !memmap) > > return -EINVAL; > > > > + mmap_read_lock(current->mm); > > + vma = vma_lookup(current->mm, memmap->vma_base); > > + if (vma && ((vma->vm_flags & VM_PFNMAP) != 0)) { > > + if ((memmap->vma_base + memmap->len) > vma->vm_end) { > > + mmap_read_unlock(current->mm); > > + return -EINVAL; > > + } > > + > > + ret = follow_pfn(vma, memmap->vma_base, &pfn); > > This races, don't use follow_pfn() and most definitely don't add new users. In > some cases follow_pte, but the pte/pfn is still only valid for as long as you > hold > the pte spinlock. > > > + mmap_read_unlock(current->mm); > > Definitely after here there's zero guarantees about this pfn and it could > point at > anything. > > Please fix, I tried pretty hard to get rid of follow_pfn(), but some of them > are > just too hard to fix (e.g. kvm needs a pretty hug rewrite to get it all > sorted). > > Cheers, Daniel > > > + if (ret < 0) { > > + dev_dbg(acrn_dev.this_device, > > + "Failed to lookup PFN at VMA:%pK.\n", (void > *)memmap->vma_base); > > + return ret; > > + } > > + > > + return acrn_mm_region_add(vm, memmap->user_vm_pa, > > + PFN_PHYS(pfn), memmap->len, > > + ACRN_MEM_TYPE_WB, memmap->attr); > > + } > > + mmap_read_unlock(current->mm); > > + > > /* Get the page number of the map region */ > > nr_pages = memmap->len >> PAGE_SHIFT; > > pages = vzalloc(nr_pages * sizeof(struct page *)); > > > > base-commit: 73878e5eb1bd3c9656685ca60bc3a49d17311e0c > > -- > > 2.25.1 > > > > -- > Daniel Vetter > Software Engineer, Intel Corporation > http://blog.ffwll.ch