On Fri, Feb 07, 2025 at 10:46:01AM -0400, Jason Gunthorpe wrote:
> To make way for a domain_alloc_paging conversion add the typical global
> static IDENTITY domain. This supports VMMs that have a
> VIRTIO_IOMMU_F_BYPASS_CONFIG config.
> 
> If the VMM does not have support then the domain_alloc path is still used,
> which creates an IDENTITY domain out of a paging domain.
> 
> Signed-off-by: Jason Gunthorpe <j...@nvidia.com>
> ---
>  drivers/iommu/virtio-iommu.c | 86 ++++++++++++++++++++++++++++--------
>  1 file changed, 67 insertions(+), 19 deletions(-)
> 
> diff --git a/drivers/iommu/virtio-iommu.c b/drivers/iommu/virtio-iommu.c
> index b85ce6310ddbda..c71a996760bddb 100644
> --- a/drivers/iommu/virtio-iommu.c
> +++ b/drivers/iommu/virtio-iommu.c
> @@ -48,6 +48,7 @@ struct viommu_dev {
>       u64                             pgsize_bitmap;
>       u32                             first_domain;
>       u32                             last_domain;
> +     u32                             identity_domain_id;
>       /* Supported MAP flags */
>       u32                             map_flags;
>       u32                             probe_size;
> @@ -70,7 +71,6 @@ struct viommu_domain {
>       struct rb_root_cached           mappings;
>  
>       unsigned long                   nr_endpoints;
> -     bool                            bypass;
>  };
>  
>  struct viommu_endpoint {
> @@ -305,6 +305,22 @@ static int viommu_send_req_sync(struct viommu_dev 
> *viommu, void *buf,
>       return ret;
>  }
>  
> +static int viommu_send_attach_req(struct viommu_dev *viommu, struct device 
> *dev,
> +                               struct virtio_iommu_req_attach *req)
> +{
> +     struct iommu_fwspec *fwspec = dev_iommu_fwspec_get(dev);
> +     int ret;
> +     int i;
> +
> +     for (i = 0; i < fwspec->num_ids; i++) {
> +             req->endpoint = cpu_to_le32(fwspec->ids[i]);
> +             ret = viommu_send_req_sync(viommu, req, sizeof(*req));
> +             if (ret)
> +                     return ret;
> +     }
> +     return 0;
> +}
> +
>  /*
>   * viommu_add_mapping - add a mapping to the internal tree
>   *
> @@ -687,12 +703,6 @@ static int viommu_domain_finalise(struct viommu_endpoint 
> *vdev,
>       vdomain->viommu         = viommu;
>  
>       if (domain->type == IOMMU_DOMAIN_IDENTITY) {
> -             if (virtio_has_feature(viommu->vdev,
> -                                    VIRTIO_IOMMU_F_BYPASS_CONFIG)) {
> -                     vdomain->bypass = true;
> -                     return 0;
> -             }
> -
>               ret = viommu_domain_map_identity(vdev, vdomain);
>               if (ret) {
>                       ida_free(&viommu->domain_ids, vdomain->id);
> @@ -719,10 +729,8 @@ static void viommu_domain_free(struct iommu_domain 
> *domain)
>  
>  static int viommu_attach_dev(struct iommu_domain *domain, struct device *dev)
>  {
> -     int i;
>       int ret = 0;
>       struct virtio_iommu_req_attach req;
> -     struct iommu_fwspec *fwspec = dev_iommu_fwspec_get(dev);
>       struct viommu_endpoint *vdev = dev_iommu_priv_get(dev);
>       struct viommu_domain *vdomain = to_viommu_domain(domain);
>  
> @@ -761,16 +769,9 @@ static int viommu_attach_dev(struct iommu_domain 
> *domain, struct device *dev)
>               .domain         = cpu_to_le32(vdomain->id),
>       };
>  
> -     if (vdomain->bypass)
> -             req.flags |= cpu_to_le32(VIRTIO_IOMMU_ATTACH_F_BYPASS);
> -
> -     for (i = 0; i < fwspec->num_ids; i++) {
> -             req.endpoint = cpu_to_le32(fwspec->ids[i]);
> -
> -             ret = viommu_send_req_sync(vdomain->viommu, &req, sizeof(req));
> -             if (ret)
> -                     return ret;
> -     }
> +     ret = viommu_send_attach_req(vdomain->viommu, dev, &req);
> +     if (ret)
> +             return ret;
>  
>       if (!vdomain->nr_endpoints) {
>               /*
> @@ -788,6 +789,40 @@ static int viommu_attach_dev(struct iommu_domain 
> *domain, struct device *dev)
>       return 0;
>  }
>  
> +static int viommu_attach_identity_domain(struct iommu_domain *domain,
> +                                      struct device *dev)
> +{
> +     int ret = 0;
> +     struct virtio_iommu_req_attach req;
> +     struct viommu_endpoint *vdev = dev_iommu_priv_get(dev);
> +     struct viommu_domain *vdomain = to_viommu_domain(domain);
> +
> +     req = (struct virtio_iommu_req_attach) {
> +             .head.type      = VIRTIO_IOMMU_T_ATTACH,
> +             .domain         = cpu_to_le32(vdev->viommu->identity_domain_id),
> +             .flags          = cpu_to_le32(VIRTIO_IOMMU_ATTACH_F_BYPASS),
> +     };
> +
> +     ret = viommu_send_attach_req(vdev->viommu, dev, &req);
> +     if (ret)
> +             return ret;
> +
> +     if (vdev->vdomain) {
> +             vdev->vdomain->nr_endpoints--;
> +             vdomain->nr_endpoints++;
> +             vdev->vdomain = vdomain;

These two need to be unconditional

> +     }
> +     return 0;
> +}
> +
> +static struct viommu_domain viommu_identity_domain = {
> +     .domain = { .type = IOMMU_DOMAIN_IDENTITY,
> +                 .ops =
> +                         &(const struct iommu_domain_ops){
> +                                 .attach_dev = viommu_attach_identity_domain,
> +                         } }
> +};

nit: how about

        static struct viommu_domain viommu_identity_domain = {
                .domain = {
                        .type = IOMMU_DOMAIN_IDENTITY,
                        .ops = &(const struct iommu_domain_ops) {
                                .attach_dev = viommu_attach_identity_domain,
                        },
                },
        };

> +
>  static void viommu_detach_dev(struct viommu_endpoint *vdev)
>  {
>       int i;
> @@ -1061,6 +1096,7 @@ static bool viommu_capable(struct device *dev, enum 
> iommu_cap cap)
>  }
>  
>  static struct iommu_ops viommu_ops = {
> +     .identity_domain        = &viommu_identity_domain.domain,
>       .capable                = viommu_capable,
>       .domain_alloc           = viommu_domain_alloc,
>       .probe_device           = viommu_probe_device,
> @@ -1184,6 +1220,18 @@ static int viommu_probe(struct virtio_device *vdev)
>       if (virtio_has_feature(vdev, VIRTIO_IOMMU_F_MMIO))
>               viommu->map_flags |= VIRTIO_IOMMU_MAP_F_MMIO;
>  
> +     /* Reserve an ID to use as the bypass domain */
> +     if (virtio_has_feature(viommu->vdev, VIRTIO_IOMMU_F_BYPASS_CONFIG)) {
> +             viommu->identity_domain_id = viommu->first_domain;
> +             viommu->first_domain++;
> +     } else {
> +             /*
> +              * Assume the VMM is sensible and it either supports bypass on
> +              * all instances or no instances.
> +              */

Maybe also a WARN_ON(!viommu_ops.identity_domain) above?

Thanks,
Jean



> +             viommu_ops.identity_domain = NULL;
> +     }
> +
>       viommu_ops.pgsize_bitmap = viommu->pgsize_bitmap;
>  
>       virtio_device_ready(vdev);
> -- 
> 2.43.0
> 
> 

Reply via email to