On 3/12/2018 1:29 PM, Tom Lendacky wrote:
> On 3/5/2018 11:46 AM, Christoph Hellwig wrote:
>> Give the basic phys_to_dma and dma_to_phys helpers a __-prefix and add
>> the memory encryption mask to the non-prefixed versions.  Use the
>> __-prefixed versions directly instead of clearing the mask again in
>> various places.
>>
>> With that in place the generic dma-direct routines can be used to
>> allocate non-encrypted bounce buffers, and the x86 SEV case can use
>> the generic swiotlb ops.
>>
>> Signed-off-by: Christoph Hellwig <h...@lst.de>
> 
> So this patch results in my system failing to boot when SME is active.
> I'm investigating further to see why.  I'll follow up with more details
> as I find them.

Ok, I found one issue that allows this to work when the IOMMU isn't
enabled (see below).

But the bigger issue is when the IOMMU is enabled.  The IOMMU code uses
a common mapping routine to create the I/O page tables.  This routine
assumes that all memory being mapped is encrypted and therefore sets the
encryption bit in the I/O page tables.  With this patch, the call to
dma_alloc_direct() now returns un-encrypted memory which results in an
encryption mis-match.  I think keeping dma_alloc_direct() as it was prior
to this patch is the way to go.  It allows SME DMA allocations to remain
encrypted and avoids added complexity in the amd_iommu.c file.  This
would mean that SEV would still have special DMA operations (so that the
alloc/free can change the memory to un-encrypted).

What do you think?

> 
> Additionally, when running with SME (not SEV), this is forcing all DMA
> coherent allocations to be decrypted, when that isn't required with SME
> (as long as the device can perform 48-bit or greater DMA).  So it may
> be worth looking at only doing the decrypted allocations for SEV.
> 
> Thanks,
> Tom
> 
>> ---
>>  arch/arm/include/asm/dma-direct.h                  |  4 +-
>>  arch/mips/cavium-octeon/dma-octeon.c               | 10 +--
>>  .../include/asm/mach-cavium-octeon/dma-coherence.h |  4 +-
>>  .../include/asm/mach-loongson64/dma-coherence.h    | 10 +--
>>  arch/mips/loongson64/common/dma-swiotlb.c          |  4 +-
>>  arch/powerpc/include/asm/dma-direct.h              |  4 +-
>>  arch/x86/Kconfig                                   |  2 +-
>>  arch/x86/include/asm/dma-direct.h                  | 25 +-------
>>  arch/x86/mm/mem_encrypt.c                          | 73 
>> +---------------------
>>  arch/x86/pci/sta2x11-fixup.c                       |  6 +-
>>  include/linux/dma-direct.h                         | 21 ++++++-
>>  lib/dma-direct.c                                   | 21 +++++--
>>  lib/swiotlb.c                                      | 25 +++-----
>>  13 files changed, 70 insertions(+), 139 deletions(-)
>>

< ... SNIP ... >

>> diff --git a/lib/dma-direct.c b/lib/dma-direct.c
>> index c9e8e21cb334..84f50b5982fc 100644
>> --- a/lib/dma-direct.c
>> +++ b/lib/dma-direct.c
>> @@ -9,6 +9,7 @@
>>  #include <linux/scatterlist.h>
>>  #include <linux/dma-contiguous.h>
>>  #include <linux/pfn.h>
>> +#include <linux/set_memory.h>
>>  
>>  #define DIRECT_MAPPING_ERROR                0
>>  
>> @@ -35,9 +36,13 @@ check_addr(struct device *dev, dma_addr_t dma_addr, 
>> size_t size,
>>      return true;
>>  }
>>  
>> +/*
>> + * Since we will be clearing the encryption bit, check the mask with it 
>> already
>> + * cleared.
>> + */
>>  static bool dma_coherent_ok(struct device *dev, phys_addr_t phys, size_t 
>> size)
>>  {
>> -    return phys_to_dma(dev, phys) + size - 1 <= dev->coherent_dma_mask;
>> +    return __phys_to_dma(dev, phys) + size - 1 <= dev->coherent_dma_mask;
>>  }
>>  
>>  void *dma_direct_alloc(struct device *dev, size_t size, dma_addr_t 
>> *dma_handle,
>> @@ -46,6 +51,7 @@ void *dma_direct_alloc(struct device *dev, size_t size, 
>> dma_addr_t *dma_handle,
>>      unsigned int count = PAGE_ALIGN(size) >> PAGE_SHIFT;
>>      int page_order = get_order(size);
>>      struct page *page = NULL;
>> +    void *ret;
>>  
>>      /* GFP_DMA32 and GFP_DMA are no ops without the corresponding zones: */
>>      if (dev->coherent_dma_mask <= DMA_BIT_MASK(ARCH_ZONE_DMA_BITS))
>> @@ -78,10 +84,11 @@ void *dma_direct_alloc(struct device *dev, size_t size, 
>> dma_addr_t *dma_handle,
>>  
>>      if (!page)
>>              return NULL;
>> -
>> -    *dma_handle = phys_to_dma(dev, page_to_phys(page));
>> -    memset(page_address(page), 0, size);
>> -    return page_address(page);
>> +    *dma_handle = __phys_to_dma(dev, page_to_phys(page));
>> +    ret = page_address(page);
>> +    set_memory_decrypted((unsigned long)ret, page_order);

The second parameter should be 1 << page_order to get the number of
pages.

Thanks,
Tom

>> +    memset(ret, 0, size);
>> +    return ret;
>>  }
>>  
>>  /*
>> @@ -92,9 +99,11 @@ void dma_direct_free(struct device *dev, size_t size, 
>> void *cpu_addr,
>>              dma_addr_t dma_addr, unsigned long attrs)
>>  {
>>      unsigned int count = PAGE_ALIGN(size) >> PAGE_SHIFT;
>> +    unsigned int page_order = get_order(size);
>>  
>> +    set_memory_encrypted((unsigned long)cpu_addr, 1 << page_order);
>>      if (!dma_release_from_contiguous(dev, virt_to_page(cpu_addr), count))
>> -            free_pages((unsigned long)cpu_addr, get_order(size));
>> +            free_pages((unsigned long)cpu_addr, page_order);
>>  }
>>  
>>  static dma_addr_t dma_direct_map_page(struct device *dev, struct page *page,
>> diff --git a/lib/swiotlb.c b/lib/swiotlb.c
>> index c43ec2271469..ca8eeaead925 100644
>> --- a/lib/swiotlb.c
>> +++ b/lib/swiotlb.c
>> @@ -158,13 +158,6 @@ unsigned long swiotlb_size_or_default(void)
>>  
>>  void __weak swiotlb_set_mem_attributes(void *vaddr, unsigned long size) { }
>>  
>> -/* For swiotlb, clear memory encryption mask from dma addresses */
>> -static dma_addr_t swiotlb_phys_to_dma(struct device *hwdev,
>> -                                  phys_addr_t address)
>> -{
>> -    return __sme_clr(phys_to_dma(hwdev, address));
>> -}
>> -
>>  /* Note that this doesn't work with highmem page */
>>  static dma_addr_t swiotlb_virt_to_bus(struct device *hwdev,
>>                                    volatile void *address)
>> @@ -622,7 +615,7 @@ map_single(struct device *hwdev, phys_addr_t phys, 
>> size_t size,
>>              return SWIOTLB_MAP_ERROR;
>>      }
>>  
>> -    start_dma_addr = swiotlb_phys_to_dma(hwdev, io_tlb_start);
>> +    start_dma_addr = __phys_to_dma(hwdev, io_tlb_start);
>>      return swiotlb_tbl_map_single(hwdev, start_dma_addr, phys, size,
>>                                    dir, attrs);
>>  }
>> @@ -726,12 +719,12 @@ swiotlb_alloc_buffer(struct device *dev, size_t size, 
>> dma_addr_t *dma_handle,
>>              goto out_warn;
>>  
>>      phys_addr = swiotlb_tbl_map_single(dev,
>> -                    swiotlb_phys_to_dma(dev, io_tlb_start),
>> +                    __phys_to_dma(dev, io_tlb_start),
>>                      0, size, DMA_FROM_DEVICE, 0);
>>      if (phys_addr == SWIOTLB_MAP_ERROR)
>>              goto out_warn;
>>  
>> -    *dma_handle = swiotlb_phys_to_dma(dev, phys_addr);
>> +    *dma_handle = __phys_to_dma(dev, phys_addr);
>>      if (dma_coherent_ok(dev, *dma_handle, size))
>>              goto out_unmap;
>>  
>> @@ -867,10 +860,10 @@ dma_addr_t swiotlb_map_page(struct device *dev, struct 
>> page *page,
>>      map = map_single(dev, phys, size, dir, attrs);
>>      if (map == SWIOTLB_MAP_ERROR) {
>>              swiotlb_full(dev, size, dir, 1);
>> -            return swiotlb_phys_to_dma(dev, io_tlb_overflow_buffer);
>> +            return __phys_to_dma(dev, io_tlb_overflow_buffer);
>>      }
>>  
>> -    dev_addr = swiotlb_phys_to_dma(dev, map);
>> +    dev_addr = __phys_to_dma(dev, map);
>>  
>>      /* Ensure that the address returned is DMA'ble */
>>      if (dma_capable(dev, dev_addr, size))
>> @@ -879,7 +872,7 @@ dma_addr_t swiotlb_map_page(struct device *dev, struct 
>> page *page,
>>      attrs |= DMA_ATTR_SKIP_CPU_SYNC;
>>      swiotlb_tbl_unmap_single(dev, map, size, dir, attrs);
>>  
>> -    return swiotlb_phys_to_dma(dev, io_tlb_overflow_buffer);
>> +    return __phys_to_dma(dev, io_tlb_overflow_buffer);
>>  }
>>  
>>  /*
>> @@ -1009,7 +1002,7 @@ swiotlb_map_sg_attrs(struct device *hwdev, struct 
>> scatterlist *sgl, int nelems,
>>                              sg_dma_len(sgl) = 0;
>>                              return 0;
>>                      }
>> -                    sg->dma_address = swiotlb_phys_to_dma(hwdev, map);
>> +                    sg->dma_address = __phys_to_dma(hwdev, map);
>>              } else
>>                      sg->dma_address = dev_addr;
>>              sg_dma_len(sg) = sg->length;
>> @@ -1073,7 +1066,7 @@ swiotlb_sync_sg_for_device(struct device *hwdev, 
>> struct scatterlist *sg,
>>  int
>>  swiotlb_dma_mapping_error(struct device *hwdev, dma_addr_t dma_addr)
>>  {
>> -    return (dma_addr == swiotlb_phys_to_dma(hwdev, io_tlb_overflow_buffer));
>> +    return (dma_addr == __phys_to_dma(hwdev, io_tlb_overflow_buffer));
>>  }
>>  
>>  /*
>> @@ -1085,7 +1078,7 @@ swiotlb_dma_mapping_error(struct device *hwdev, 
>> dma_addr_t dma_addr)
>>  int
>>  swiotlb_dma_supported(struct device *hwdev, u64 mask)
>>  {
>> -    return swiotlb_phys_to_dma(hwdev, io_tlb_end - 1) <= mask;
>> +    return __phys_to_dma(hwdev, io_tlb_end - 1) <= mask;
>>  }
>>  
>>  #ifdef CONFIG_DMA_DIRECT_OPS
>>
_______________________________________________
iommu mailing list
iommu@lists.linux-foundation.org
https://lists.linuxfoundation.org/mailman/listinfo/iommu

Reply via email to