On Mon, Apr 05, 2021 at 12:11:09PM -0700, Isaac J. Manjarres wrote:
> Implement the unmap_pages() callback for the ARM LPAE io-pgtable
> format.
> 
> Signed-off-by: Isaac J. Manjarres <isa...@codeaurora.org>
> Suggested-by: Will Deacon <w...@kernel.org>
> ---
>  drivers/iommu/io-pgtable-arm.c | 124 +++++++++++++++++++++++++++------
>  1 file changed, 104 insertions(+), 20 deletions(-)

[...]

> +static size_t arm_lpae_unmap_pages(struct io_pgtable_ops *ops, unsigned long 
> iova,
> +                                size_t pgsize, size_t pgcount,
> +                                struct iommu_iotlb_gather *gather)
> +{
> +     struct arm_lpae_io_pgtable *data = io_pgtable_ops_to_data(ops);
> +     struct io_pgtable_cfg *cfg = &data->iop.cfg;
> +     arm_lpae_iopte *ptep = data->pgd;
> +     long iaext = (s64)iova >> cfg->ias;
> +     size_t unmapped = 0, unmapped_page;
> +     int last_lvl;
> +     size_t table_size, pages, tbl_offset, max_entries;
> +
> +     if (WARN_ON(!pgsize || (pgsize & cfg->pgsize_bitmap) != pgsize || 
> !pgcount))
> +             return 0;
> +
> +     if (cfg->quirks & IO_PGTABLE_QUIRK_ARM_TTBR1)
> +             iaext = ~iaext;
> +     if (WARN_ON(iaext))
> +             return 0;
> +
> +     /*
> +      * Calculating the page table size here helps avoid situations where
> +      * a page range that is being unmapped may be mapped at the same level
> +      * but not mapped by the same tables. Allowing such a scenario to
> +      * occur can complicate the logic in arm_lpae_split_blk_unmap().
> +      */
> +     last_lvl = ARM_LPAE_BLOCK_SIZE_LVL(pgsize, data);
> +
> +     if (last_lvl == data->start_level)
> +             table_size = ARM_LPAE_PGD_SIZE(data);
> +     else
> +             table_size = ARM_LPAE_GRANULE(data);
> +
> +     max_entries = table_size / sizeof(*ptep);
> +
> +     while (pgcount) {
> +             tbl_offset = ARM_LPAE_LVL_IDX(iova, last_lvl, data);
> +             pages = min_t(size_t, pgcount, max_entries - tbl_offset);
> +             unmapped_page = __arm_lpae_unmap(data, gather, iova, pgsize,
> +                                              pages, data->start_level,
> +                                              ptep);
> +             if (!unmapped_page)
> +                     break;
> +
> +             unmapped += unmapped_page;
> +             iova += unmapped_page;
> +             pgcount -= pages;
> +     }

Robin has some comments on the first version of this patch, and I don't think 
you
addressed them:

https://lore.kernel.org/r/b93fa0b1-e2a4-1aad-8b88-4d0dfecdf...@arm.com

I'm inclined to agree that iterating here doesn't make a lot of sense --
we've already come back out of the page-table walk, so I think we should
just return to the caller (who is well prepared to handle a partial unmap).
Same for the map side of things.

If we get numbers showing that this is causing a performance issue, then
we should rework the page-table code to handle this at the lower level
(because I doubt the loop that you have is really worse than returning to
the caller anyway).

Will
_______________________________________________
iommu mailing list
iommu@lists.linux-foundation.org
https://lists.linuxfoundation.org/mailman/listinfo/iommu

Reply via email to