"Kalyazin, Nikita" <[email protected]> writes:

>
> [...snip...]
>
>  static vm_fault_t kvm_gmem_fault_user_mapping(struct vm_fault *vmf)
>  {
>       struct inode *inode = file_inode(vmf->vma->vm_file);
>       struct folio *folio;
>       vm_fault_t ret = VM_FAULT_LOCKED;
> +     int err;
>
>       if (((loff_t)vmf->pgoff << PAGE_SHIFT) >= i_size_read(inode))
>               return VM_FAULT_SIGBUS;
> @@ -418,6 +454,14 @@ static vm_fault_t kvm_gmem_fault_user_mapping(struct 
> vm_fault *vmf)
>               folio_mark_uptodate(folio);
>       }
>
> +     if (kvm_gmem_no_direct_map(folio_inode(folio))) {
> +             err = kvm_gmem_folio_zap_direct_map(folio);
> +             if (err) {
> +                     ret = vmf_error(err);
> +                     goto out_folio;
> +             }
> +     }
> +
>       vmf->page = folio_file_page(folio, vmf->pgoff);
>

Sashiko pointed out that kvm_gmem_populate() might try and write to
direct-map-removed folios, but I think that's handled because populate
will first try and GUP folios, which is already blocked for
direct-map-removed folios.

>  out_folio:
> @@ -528,6 +572,9 @@ static void kvm_gmem_free_folio(struct folio *folio)
>       kvm_pfn_t pfn = page_to_pfn(page);
>       int order = folio_order(folio);
>
> +     if (kvm_gmem_folio_no_direct_map(folio))
> +             kvm_gmem_folio_restore_direct_map(folio);
> +
>       kvm_arch_gmem_invalidate(pfn, pfn + (1ul << order));
>  }
>

Sashiko says to invalidate then restore direct map, I think in this case
it doesn't matter since if the folio needed invalidation, it must be
private, and the host shouldn't be writing to the private pages anyway.

One benefit of retaining this order (restore, invalidate) is that it
opens the invalidate hook to possibly do something regarding memory
contents?

Or perhaps we should just take the suggestion (invalidate, restore) and
align that invalidate should not touch memory contents.

> @@ -591,6 +638,9 @@ static int __kvm_gmem_create(struct kvm *kvm, loff_t 
> size, u64 flags)
>       /* Unmovable mappings are supposed to be marked unevictable as well. */
>       WARN_ON_ONCE(!mapping_unevictable(inode->i_mapping));
>
> +     if (flags & GUEST_MEMFD_FLAG_NO_DIRECT_MAP)
> +             mapping_set_no_direct_map(inode->i_mapping);
> +
>       GMEM_I(inode)->flags = flags;
>
>       file = alloc_file_pseudo(inode, kvm_gmem_mnt, name, O_RDWR, 
> &kvm_gmem_fops);
> @@ -803,13 +853,22 @@ int kvm_gmem_get_pfn(struct kvm *kvm, struct 
> kvm_memory_slot *slot,
>       }
>
>       r = kvm_gmem_prepare_folio(kvm, slot, gfn, folio);
> +     if (r)
> +             goto out_unlock;
>
> +     if (kvm_gmem_no_direct_map(folio_inode(folio))) {
> +             r = kvm_gmem_folio_zap_direct_map(folio);
> +             if (r)
> +                     goto out_unlock;
> +     }
> +
>
> [...snip...]
>

Preparing a folio used to involve zeroing, but that has since been
refactored out, so I believe zapping can come before preparing.

Similar to the above point on invalidation: perhaps we should take the
suggestion to zap then prepare

+ And align that preparation should not touch memory contents
+ Avoid needing to undo the preparation on zapping failure (.free_folio
  is not called on folio_put(), it is only called folio on removal from
  filemap).

Reply via email to