I am dropping this patch based on our conversation, where we agreed, we
need to rootcause the migration failure.

On Thu, Jul 23, 2020 at 11:43:44AM +0530, Bharata B Rao wrote:
> On Fri, Jul 17, 2020 at 01:00:26AM -0700, Ram Pai wrote:
> > @@ -812,7 +842,7 @@ unsigned long kvmppc_h_svm_page_in(struct kvm *kvm, 
> > unsigned long gpa,
> >     struct vm_area_struct *vma;
> >     int srcu_idx;
> >     unsigned long gfn = gpa >> page_shift;
> > -   int ret;
> > +   int ret, repeat_count = REPEAT_COUNT;
> >  
> >     if (!(kvm->arch.secure_guest & KVMPPC_SECURE_INIT_START))
> >             return H_UNSUPPORTED;
> > @@ -826,34 +856,44 @@ unsigned long kvmppc_h_svm_page_in(struct kvm *kvm, 
> > unsigned long gpa,
> >     if (flags & H_PAGE_IN_SHARED)
> >             return kvmppc_share_page(kvm, gpa, page_shift);
> >  
> > -   ret = H_PARAMETER;
> >     srcu_idx = srcu_read_lock(&kvm->srcu);
> > -   mmap_read_lock(kvm->mm);
> >  
> > -   start = gfn_to_hva(kvm, gfn);
> > -   if (kvm_is_error_hva(start))
> > -           goto out;
> > -
> > -   mutex_lock(&kvm->arch.uvmem_lock);
> >     /* Fail the page-in request of an already paged-in page */
> > -   if (kvmppc_gfn_is_uvmem_pfn(gfn, kvm, NULL))
> > -           goto out_unlock;
> > +   mutex_lock(&kvm->arch.uvmem_lock);
> > +   ret = kvmppc_gfn_is_uvmem_pfn(gfn, kvm, NULL);
> > +   mutex_unlock(&kvm->arch.uvmem_lock);
> 
> Same comment as for the prev patch. I don't think you can release
> the lock here.
> 
> > +   if (ret) {
> > +           srcu_read_unlock(&kvm->srcu, srcu_idx);
> > +           return H_PARAMETER;
> > +   }
> >  
> > -   end = start + (1UL << page_shift);
> > -   vma = find_vma_intersection(kvm->mm, start, end);
> > -   if (!vma || vma->vm_start > start || vma->vm_end < end)
> > -           goto out_unlock;
> > +   do {
> > +           ret = H_PARAMETER;
> > +           mmap_read_lock(kvm->mm);
> >  
> > -   if (kvmppc_svm_migrate_page(vma, start, end, gpa, kvm, page_shift,
> > -                           true))
> > -           goto out_unlock;
> > +           start = gfn_to_hva(kvm, gfn);
> > +           if (kvm_is_error_hva(start)) {
> > +                   mmap_read_unlock(kvm->mm);
> > +                   break;
> > +           }
> >  
> > -   ret = H_SUCCESS;
> > +           end = start + (1UL << page_shift);
> > +           vma = find_vma_intersection(kvm->mm, start, end);
> > +           if (!vma || vma->vm_start > start || vma->vm_end < end) {
> > +                   mmap_read_unlock(kvm->mm);
> > +                   break;
> > +           }
> > +
> > +           mutex_lock(&kvm->arch.uvmem_lock);
> > +           ret = kvmppc_svm_migrate_page(vma, start, end, gpa, kvm, 
> > page_shift, true);
> > +           mutex_unlock(&kvm->arch.uvmem_lock);
> > +
> > +           mmap_read_unlock(kvm->mm);
> > +   } while (ret == -2 && repeat_count--);
> > +
> > +   if (ret == -2)
> > +           ret = H_BUSY;
> >  
> > -out_unlock:
> > -   mutex_unlock(&kvm->arch.uvmem_lock);
> > -out:
> > -   mmap_read_unlock(kvm->mm);
> >     srcu_read_unlock(&kvm->srcu, srcu_idx);
> >     return ret;
> >  }
> > -- 
> > 1.8.3.1

-- 
Ram Pai

Reply via email to