On Thu, Aug 01, 2019 at 10:35:02AM +0200, Lucas Stach wrote:
> Hi Christoph,
> 
> Am Donnerstag, den 01.08.2019, 09:29 +0200 schrieb Christoph Hellwig:
> > Hi Lukas,
> > 
> > have you tried the latest 5.3-rc kernel, where we limited the NVMe
> > I/O size based on the swiotlb buffer size?
> 
> Yes, the issue was reproduced on 5.3-rc2. I now see your commit
> limiting the request size, so I guess I need to dig in to see why I'm
> still getting requests larger than the SWIOTLB max segment size. Thanks
> for the pointer!

a similar setup to yours the
dma_addressing_limited doesn't work, but if we changed it to a <=
it does.  The result is counter to what I'd expect, but because I'm on
vacation I didn't have time to look into why it works.  This is his
patch, let me know if this works for you:


diff --git a/include/linux/dma-mapping.h b/include/linux/dma-mapping.h
index f7d1eea32c78..89ac1cf754cc 100644
--- a/include/linux/dma-mapping.h
+++ b/include/linux/dma-mapping.h
@@ -689,7 +689,7 @@ static inline int dma_coerce_mask_and_coherent(struct 
device *dev, u64 mask)
  */
 static inline bool dma_addressing_limited(struct device *dev)
 {
-       return min_not_zero(dma_get_mask(dev), dev->bus_dma_mask) <
+       return min_not_zero(dma_get_mask(dev), dev->bus_dma_mask) <=
                            dma_get_required_mask(dev);
 }
 
_______________________________________________
iommu mailing list
iommu@lists.linux-foundation.org
https://lists.linuxfoundation.org/mailman/listinfo/iommu

Reply via email to