On Thu, Aug 17, 2017 at 11:40:48AM +0200, Paolo Bonzini wrote: > On 17/08/2017 07:56, Peter Xu wrote: > > In vtd_switch_address_space() we did the memory region switch, however > > it's possible that the caller of it has not taken the BQL at all. Make > > sure we have it. > > > > CC: Paolo Bonzini <pbonz...@redhat.com> > > CC: Jason Wang <jasow...@redhat.com> > > CC: Michael S. Tsirkin <m...@redhat.com> > > Signed-off-by: Peter Xu <pet...@redhat.com> > > --- > > > > Paolo: I noticed this qemu_mutex_iothread_locked() function, which might > > simplify the fix, so I decided to use it. Using bottom half should be ok > > as well, but after a second thought it can be complicated: consider the > > case when guest firstly triggered the pt fast path then quickly > > re-enables the IOMMU region before the bottom half being executed. Then > > looks like we need special care on the sync of bottom half task as well. > > No, we don't, because the bottom half (as you correctly do below) would > only have to cover vtd_switch_address_space. So the worst that can > happen is that on of the two calls to vtd_switch_address_space does nothing.
Ah, yes, the state is shared... :) > > The patch below is okay. However, vtd_switch_address_space is > expensive, which is why I suggested the bottom half. But still, shall we just do it this way? It looks cleaner. For the slowness (as I mentioned below), one thing to mention is that, this fast path should even not be used when PT is enabled. When "iommu=pt" is set, the IOMMU regions are off start from the very beginning. In other words, this patch should only affect a very corner use case, and to make sure that use case is safe, though it brings the first IO of that use case slower. How do you think? Thanks, > > Paolo > > > That's over-complicated I guess (if with that, I'd prefer to remove the > > pt fast path since it's even not really the default path when pt is > > used...). Please let me know if you don't think so. > > --- > > hw/i386/intel_iommu.c | 15 +++++++++++++++ > > 1 file changed, 15 insertions(+) > > > > diff --git a/hw/i386/intel_iommu.c b/hw/i386/intel_iommu.c > > index a7bf87a..3a5bb0b 100644 > > --- a/hw/i386/intel_iommu.c > > +++ b/hw/i386/intel_iommu.c > > @@ -957,6 +957,8 @@ static bool vtd_dev_pt_enabled(VTDAddressSpace *as) > > static bool vtd_switch_address_space(VTDAddressSpace *as) > > { > > bool use_iommu; > > + /* Whether we need to take the BQL on our own */ > > + bool take_bql = !qemu_mutex_iothread_locked(); > > > > assert(as); > > > > @@ -967,6 +969,15 @@ static bool vtd_switch_address_space(VTDAddressSpace > > *as) > > VTD_PCI_FUNC(as->devfn), > > use_iommu); > > > > + /* > > + * It's possible that we reach here without BQL, e.g., when called > > + * from vtd_pt_enable_fast_path(). However the memory APIs need > > + * it. We'd better make sure we have had it already, or, take it. > > + */ > > + if (take_bql) { > > + qemu_mutex_lock_iothread(); > > + } > > + > > /* Turn off first then on the other */ > > if (use_iommu) { > > memory_region_set_enabled(&as->sys_alias, false); > > @@ -976,6 +987,10 @@ static bool vtd_switch_address_space(VTDAddressSpace > > *as) > > memory_region_set_enabled(&as->sys_alias, true); > > } > > > > + if (take_bql) { > > + qemu_mutex_unlock_iothread(); > > + } > > + > > return use_iommu; > > } > > > > > -- Peter Xu