On Tue, Jun 12, 2018 at 09:05:25PM +0200, Eric Auger wrote: > When an IOMMUMemoryRegion is in front of a virtio device, > address_space_cache_init does not set cache->ptr as the memory > region is not RAM. However when the device performs an access, > we end up in glue() which performs the translation and then uses > MAP_RAM. This latter uses the unset ptr and returns a wrong value > which leads to a SIGSEV in address_space_lduw_internal_cached_slow, > for instance. Let's test whether the cache->ptr is set, and in > the negative use the old macro definition. This fixes the > use cases featuring vIOMMU (Intel and ARM SMMU) which lead to > a SIGSEV. > > Fixes: 48564041a73a (exec: reintroduce MemoryRegion caching) > Signed-off-by: Eric Auger <eric.au...@redhat.com> > > --- > > I am not sure whether it doesn't break any targeted optimization > but at least it removes the SIGSEV. > > Signed-off-by: Eric Auger <eric.au...@redhat.com> > --- > exec.c | 4 +++- > 1 file changed, 3 insertions(+), 1 deletion(-) > > diff --git a/exec.c b/exec.c > index f6645ed..46fbd25 100644 > --- a/exec.c > +++ b/exec.c > @@ -3800,7 +3800,9 @@ address_space_write_cached_slow(MemoryRegionCache > *cache, hwaddr addr, > #define SUFFIX _cached_slow > #define TRANSLATE(...) address_space_translate_cached(cache, > __VA_ARGS__) > #define IS_DIRECT(mr, is_write) memory_access_is_direct(mr, is_write) > -#define MAP_RAM(mr, ofs) (cache->ptr + (ofs - cache->xlat)) > +#define MAP_RAM(mr, ofs) (cache->ptr ? \ > + (cache->ptr + (ofs - cache->xlat)) : \ > + qemu_map_ram_ptr((mr)->ram_block, ofs))
A pure question: if the MR is not a RAM (I think the only case for virtio case should be an IOMMU MR), then why we'll call MAP_RAM() after all? An glue() example: void glue(address_space_stb, SUFFIX)(ARG1_DECL, hwaddr addr, uint32_t val, MemTxAttrs attrs, MemTxResult *result) { uint8_t *ptr; MemoryRegion *mr; hwaddr l = 1; hwaddr addr1; MemTxResult r; bool release_lock = false; RCU_READ_LOCK(); mr = TRANSLATE(addr, &addr1, &l, true, attrs); if (!IS_DIRECT(mr, true)) { <----------------- [1] release_lock |= prepare_mmio_access(mr); r = memory_region_dispatch_write(mr, addr1, val, 1, attrs); } else { /* RAM case */ ptr = MAP_RAM(mr, addr1); stb_p(ptr, val); INVALIDATE(mr, addr1, 1); r = MEMTX_OK; } if (result) { *result = r; } if (release_lock) { qemu_mutex_unlock_iothread(); } RCU_READ_UNLOCK(); } At [1] we should check first against whether it's direct after all. AFAIU IOMMU MR should not be direct then it'll go the slow path rather than calling MAP_RAM()? Since at it, I have another (pure) question about the address space cache. I don't think it's urgent since I think it's never a problem for virtio, but I'm still asking anyways... Still taking the stb example: static inline void address_space_stb_cached(MemoryRegionCache *cache, hwaddr addr, uint32_t val, MemTxAttrs attrs, MemTxResult *result) { assert(addr < cache->len); <----------------------------- [2] if (likely(cache->ptr)) { stb_p(cache->ptr + addr, val); } else { address_space_stb_cached_slow(cache, addr, val, attrs, result); } } Here at [2] what if the region cached is smaller than provided when doing address_space_cache_init()? AFAIU the "len" provided to address_space_cache_init() can actually shrink (though for virtio it should never) when do: l = len; ... cache->mrs = *address_space_translate_internal(d, addr, &cache->xlat, &l, true); ... cache->len = l; And here not sure whether we should not assert, instead we only run the fast path if the address falls into the cache region, say: static inline void address_space_stb_cached(MemoryRegionCache *cache, hwaddr addr, uint32_t val, MemTxAttrs attrs, MemTxResult *result) { if (likely(cache->ptr && addr < cache->len)) { stb_p(cache->ptr + addr, val); } else { address_space_stb_cached_slow(cache, addr, val, attrs, result); } } Or we should add a check in address_space_cache_init() to make sure the region won't shrink. Regards, -- Peter Xu