> From: Lu Baolu <baolu...@linux.intel.com>
> Sent: Thursday, July 9, 2020 3:06 PM
> 
> A pasid might be bound to a page table from a VM guest via the iommu
> ops.sva_bind_gpasid. In this case, when a DMA page fault is detected
> on the physical IOMMU, we need to inject the page fault request into
> the guest. After the guest completes handling the page fault, a page
> response need to be sent back via the iommu ops.page_response().
> 
> This adds support to report a page request fault. Any external module
> which is interested in handling this fault should regiester a notifier
> with iommu_register_device_fault_handler().
> 
> Co-developed-by: Jacob Pan <jacob.jun....@linux.intel.com>
> Signed-off-by: Jacob Pan <jacob.jun....@linux.intel.com>
> Co-developed-by: Liu Yi L <yi.l....@intel.com>
> Signed-off-by: Liu Yi L <yi.l....@intel.com>
> Signed-off-by: Lu Baolu <baolu...@linux.intel.com>
> ---
>  drivers/iommu/intel/svm.c | 103 +++++++++++++++++++++++++++++++-------
>  1 file changed, 85 insertions(+), 18 deletions(-)
> 
> diff --git a/drivers/iommu/intel/svm.c b/drivers/iommu/intel/svm.c
> index c23167877b2b..d24e71bac8db 100644
> --- a/drivers/iommu/intel/svm.c
> +++ b/drivers/iommu/intel/svm.c
> @@ -815,8 +815,63 @@ static void intel_svm_drain_prq(struct device *dev,
> int pasid)
>       }
>  }
> 
> +static int prq_to_iommu_prot(struct page_req_dsc *req)
> +{
> +     int prot = 0;
> +
> +     if (req->rd_req)
> +             prot |= IOMMU_FAULT_PERM_READ;
> +     if (req->wr_req)
> +             prot |= IOMMU_FAULT_PERM_WRITE;
> +     if (req->exe_req)
> +             prot |= IOMMU_FAULT_PERM_EXEC;
> +     if (req->pm_req)
> +             prot |= IOMMU_FAULT_PERM_PRIV;
> +
> +     return prot;
> +}
> +
> +static int
> +intel_svm_prq_report(struct device *dev, struct page_req_dsc *desc)
> +{
> +     struct iommu_fault_event event;
> +
> +     /* Fill in event data for device specific processing */
> +     memset(&event, 0, sizeof(struct iommu_fault_event));
> +     event.fault.type = IOMMU_FAULT_PAGE_REQ;
> +     event.fault.prm.addr = desc->addr;
> +     event.fault.prm.pasid = desc->pasid;
> +     event.fault.prm.grpid = desc->prg_index;
> +     event.fault.prm.perm = prq_to_iommu_prot(desc);
> +
> +     if (!dev || !dev_is_pci(dev))
> +             return -ENODEV;

move the check before memset.

> +
> +     if (desc->lpig)
> +             event.fault.prm.flags |=
> IOMMU_FAULT_PAGE_REQUEST_LAST_PAGE;
> +     if (desc->pasid_present) {
> +             event.fault.prm.flags |=
> IOMMU_FAULT_PAGE_REQUEST_PASID_VALID;
> +             event.fault.prm.flags |=
> IOMMU_FAULT_PAGE_RESPONSE_NEEDS_PASID;
> +     }

if pasid is not present, should we return error directly instead of
submitting the req and let iommu core to figure out? I don't have
a strong preference, thus:

Reviewed-by: Kevin Tian <kevin.t...@intel.com>

> +     if (desc->priv_data_present) {
> +             /*
> +              * Set last page in group bit if private data is present,
> +              * page response is required as it does for LPIG.
> +              * iommu_report_device_fault() doesn't understand this
> vendor
> +              * specific requirement thus we set last_page as a
> workaround.
> +              */
> +             event.fault.prm.flags |=
> IOMMU_FAULT_PAGE_REQUEST_LAST_PAGE;
> +             event.fault.prm.flags |=
> IOMMU_FAULT_PAGE_REQUEST_PRIV_DATA;
> +             memcpy(event.fault.prm.private_data, desc->priv_data,
> +                    sizeof(desc->priv_data));
> +     }
> +
> +     return iommu_report_device_fault(dev, &event);
> +}
> +
>  static irqreturn_t prq_event_thread(int irq, void *d)
>  {
> +     struct intel_svm_dev *sdev = NULL;
>       struct intel_iommu *iommu = d;
>       struct intel_svm *svm = NULL;
>       int head, tail, handled = 0;
> @@ -828,7 +883,6 @@ static irqreturn_t prq_event_thread(int irq, void *d)
>       tail = dmar_readq(iommu->reg + DMAR_PQT_REG) &
> PRQ_RING_MASK;
>       head = dmar_readq(iommu->reg + DMAR_PQH_REG) &
> PRQ_RING_MASK;
>       while (head != tail) {
> -             struct intel_svm_dev *sdev;
>               struct vm_area_struct *vma;
>               struct page_req_dsc *req;
>               struct qi_desc resp;
> @@ -864,6 +918,20 @@ static irqreturn_t prq_event_thread(int irq, void *d)
>                       }
>               }
> 
> +             if (!sdev || sdev->sid != req->rid) {
> +                     struct intel_svm_dev *t;
> +
> +                     sdev = NULL;
> +                     rcu_read_lock();
> +                     list_for_each_entry_rcu(t, &svm->devs, list) {
> +                             if (t->sid == req->rid) {
> +                                     sdev = t;
> +                                     break;
> +                             }
> +                     }
> +                     rcu_read_unlock();
> +             }
> +
>               result = QI_RESP_INVALID;
>               /* Since we're using init_mm.pgd directly, we should never
> take
>                * any faults on kernel addresses. */
> @@ -874,6 +942,17 @@ static irqreturn_t prq_event_thread(int irq, void *d)
>               if (!is_canonical_address(address))
>                       goto bad_req;
> 
> +             /*
> +              * If prq is to be handled outside iommu driver via receiver of
> +              * the fault notifiers, we skip the page response here.
> +              */
> +             if (svm->flags & SVM_FLAG_GUEST_MODE) {
> +                     if (sdev && !intel_svm_prq_report(sdev->dev, req))
> +                             goto prq_advance;
> +                     else
> +                             goto bad_req;
> +             }
> +
>               /* If the mm is already defunct, don't handle faults. */
>               if (!mmget_not_zero(svm->mm))
>                       goto bad_req;
> @@ -892,24 +971,11 @@ static irqreturn_t prq_event_thread(int irq, void *d)
>                       goto invalid;
> 
>               result = QI_RESP_SUCCESS;
> -     invalid:
> +invalid:
>               mmap_read_unlock(svm->mm);
>               mmput(svm->mm);
> -     bad_req:
> -             /* Accounting for major/minor faults? */
> -             rcu_read_lock();
> -             list_for_each_entry_rcu(sdev, &svm->devs, list) {
> -                     if (sdev->sid == req->rid)
> -                             break;
> -             }
> -             /* Other devices can go away, but the drivers are not
> permitted
> -              * to unbind while any page faults might be in flight. So it's
> -              * OK to drop the 'lock' here now we have it. */
> -             rcu_read_unlock();
> -
> -             if (WARN_ON(&sdev->list == &svm->devs))
> -                     sdev = NULL;
> -
> +bad_req:
> +             WARN_ON(!sdev);
>               if (sdev && sdev->ops && sdev->ops->fault_cb) {
>                       int rwxp = (req->rd_req << 3) | (req->wr_req << 2) |
>                               (req->exe_req << 1) | (req->pm_req);
> @@ -920,7 +986,7 @@ static irqreturn_t prq_event_thread(int irq, void *d)
>                  and these can be NULL. Do not use them below this point!
> */
>               sdev = NULL;
>               svm = NULL;
> -     no_pasid:
> +no_pasid:
>               if (req->lpig || req->priv_data_present) {
>                       /*
>                        * Per VT-d spec. v3.0 ch7.7, system software must
> @@ -945,6 +1011,7 @@ static irqreturn_t prq_event_thread(int irq, void *d)
>                       resp.qw3 = 0;
>                       qi_submit_sync(iommu, &resp, 1, 0);
>               }
> +prq_advance:
>               head = (head + sizeof(*req)) & PRQ_RING_MASK;
>       }
> 
> --
> 2.17.1

Reply via email to