On Sat, 26 Oct 2019 09:00:51 +0800
Lu Baolu <baolu...@linux.intel.com> wrote:

> Hi,
> 
> On 10/25/19 3:55 AM, Jacob Pan wrote:
> > Use combined macros for_each_svm_dev() to simplify SVM device
> > iteration and error checking.
> > 
> > Suggested-by: Andy Shevchenko <andriy.shevche...@linux.intel.com>
> > Signed-off-by: Jacob Pan <jacob.jun....@linux.intel.com>
> > Reviewed-by: Eric Auger <eric.au...@redhat.com>
> > ---
> >   drivers/iommu/intel-svm.c | 89
> > ++++++++++++++++++++++------------------------- 1 file changed, 42
> > insertions(+), 47 deletions(-)
> > 
> > diff --git a/drivers/iommu/intel-svm.c b/drivers/iommu/intel-svm.c
> > index a9a7f85a09bc..a18b02a9709d 100644
> > --- a/drivers/iommu/intel-svm.c
> > +++ b/drivers/iommu/intel-svm.c
> > @@ -212,6 +212,10 @@ static const struct mmu_notifier_ops
> > intel_mmuops = { static DEFINE_MUTEX(pasid_mutex);
> >   static LIST_HEAD(global_svm_list);
> >   
> > +#define for_each_svm_dev(svm, dev)                 \
> > +   list_for_each_entry(sdev, &svm->devs, list)     \
> > +   if (dev == sdev->dev)                           \
> > +
> >   int intel_svm_bind_mm(struct device *dev, int *pasid, int flags,
> > struct svm_dev_ops *ops) {
> >     struct intel_iommu *iommu =
> > intel_svm_device_to_iommu(dev); @@ -257,15 +261,13 @@ int
> > intel_svm_bind_mm(struct device *dev, int *pasid, int flags, struct
> > svm_dev_ goto out; }
> >   
> > -                   list_for_each_entry(sdev, &svm->devs,
> > list) {
> > -                           if (dev == sdev->dev) {
> > -                                   if (sdev->ops != ops) {
> > -                                           ret = -EBUSY;
> > -                                           goto out;
> > -                                   }
> > -                                   sdev->users++;
> > -                                   goto success;
> > +                   for_each_svm_dev(svm, dev) {
> > +                           if (sdev->ops != ops) {
> > +                                   ret = -EBUSY;
> > +                                   goto out;
> >                             }
> > +                           sdev->users++;
> > +                           goto success;
> >                     }
> >   
> >                     break;
> > @@ -402,50 +404,43 @@ int intel_svm_unbind_mm(struct device *dev,
> > int pasid) goto out;
> >   
> >     svm = ioasid_find(NULL, pasid, NULL);
> > -   if (IS_ERR(svm)) {
> > +   if (IS_ERR_OR_NULL(svm)) {
> >             ret = PTR_ERR(svm);
> >             goto out;
> >     }
> >   
> > -   if (!svm)
> > -           goto out;  
> 
> If svm == NULL here, this function will return success. This isn't
> expected, right?
> 
you are right, should handle separately.

Thanks!
> Others looks good to me.
> 
> Reviewed-by: Lu Baolu <baolu...@linux.intel.com>
> 
> Best regards,
> baolu
> 
> > -
> > -   list_for_each_entry(sdev, &svm->devs, list) {
> > -           if (dev == sdev->dev) {
> > -                   ret = 0;
> > -                   sdev->users--;
> > -                   if (!sdev->users) {
> > -                           list_del_rcu(&sdev->list);
> > -                           /* Flush the PASID cache and IOTLB
> > for this device.
> > -                            * Note that we do depend on the
> > hardware *not* using
> > -                            * the PASID any more. Just as we
> > depend on other
> > -                            * devices never using PASIDs that
> > they have no right
> > -                            * to use. We have a *shared*
> > PASID table, because it's
> > -                            * large and has to be physically
> > contiguous. So it's
> > -                            * hard to be as defensive as we
> > might like. */
> > -                           intel_pasid_tear_down_entry(iommu,
> > dev, svm->pasid);
> > -                           intel_flush_svm_range_dev(svm,
> > sdev, 0, -1, 0);
> > -                           kfree_rcu(sdev, rcu);
> > -
> > -                           if (list_empty(&svm->devs)) {
> > -                                   /* Clear private data so
> > that free pass check */
> > -
> > ioasid_set_data(svm->pasid, NULL);
> > -                                   ioasid_free(svm->pasid);
> > -                                   if (svm->mm)
> > -
> > mmu_notifier_unregister(&svm->notifier, svm->mm); -
> > -                                   list_del(&svm->list);
> > -
> > -                                   /* We mandate that no page
> > faults may be outstanding
> > -                                    * for the PASID when
> > intel_svm_unbind_mm() is called.
> > -                                    * If that is not obeyed,
> > subtle errors will happen.
> > -                                    * Let's make them less
> > subtle... */
> > -                                   memset(svm, 0x6b,
> > sizeof(*svm));
> > -                                   kfree(svm);
> > -                           }
> > +   for_each_svm_dev(svm, dev) {
> > +           ret = 0;
> > +           sdev->users--;
> > +           if (!sdev->users) {
> > +                   list_del_rcu(&sdev->list);
> > +                   /* Flush the PASID cache and IOTLB for
> > this device.
> > +                    * Note that we do depend on the hardware
> > *not* using
> > +                    * the PASID any more. Just as we depend
> > on other
> > +                    * devices never using PASIDs that they
> > have no right
> > +                    * to use. We have a *shared* PASID table,
> > because it's
> > +                    * large and has to be physically
> > contiguous. So it's
> > +                    * hard to be as defensive as we might
> > like. */
> > +                   intel_pasid_tear_down_entry(iommu, dev,
> > svm->pasid);
> > +                   intel_flush_svm_range_dev(svm, sdev, 0,
> > -1, 0);
> > +                   kfree_rcu(sdev, rcu);
> > +
> > +                   if (list_empty(&svm->devs)) {
> > +                           /* Clear private data so that free
> > pass check */
> > +                           ioasid_set_data(svm->pasid, NULL);
> > +                           ioasid_free(svm->pasid);
> > +                           if (svm->mm)
> > +
> > mmu_notifier_unregister(&svm->notifier, svm->mm);
> > +                           list_del(&svm->list);
> > +                           /* We mandate that no page faults
> > may be outstanding
> > +                            * for the PASID when
> > intel_svm_unbind_mm() is called.
> > +                            * If that is not obeyed, subtle
> > errors will happen.
> > +                            * Let's make them less subtle...
> > */
> > +                           memset(svm, 0x6b, sizeof(*svm));
> > +                           kfree(svm);
> >                     }
> > -                   break;
> >             }
> > +           break;
> >     }
> >    out:
> >     mutex_unlock(&pasid_mutex);
> > @@ -581,7 +576,7 @@ static irqreturn_t prq_event_thread(int irq,
> > void *d)
> >                      * to unbind the mm while any page faults
> > are outstanding.
> >                      * So we only need RCU to protect the
> > internal idr code. */ rcu_read_unlock();
> > -                   if (IS_ERR(svm) || !svm) {
> > +                   if (IS_ERR_OR_NULL(svm)) {
> >                             pr_err("%s: Page request for
> > invalid PASID %d: %08llx %08llx\n", iommu->name, req->pasid,
> > ((unsigned long long *)req)[0], ((unsigned long long *)req)[1]);
> >   

[Jacob Pan]
_______________________________________________
iommu mailing list
iommu@lists.linux-foundation.org
https://lists.linuxfoundation.org/mailman/listinfo/iommu

Reply via email to