On 06/09/2018 14:14, Auger Eric wrote: > Hi Jean-Philippe, > > On 09/06/2018 02:42 PM, Jean-Philippe Brucker wrote: >> On 06/09/2018 10:25, Auger Eric wrote: >>>> + mutex_lock(&fparam->lock); >>>> + list_add_tail(&evt_pending->list, &fparam->faults); >>> same doubt as Yi Liu. You cannot rely on the userspace willingness to >>> void the queue and deallocate this memory. > > By the way I saw there is a kind of garbage collectors for faults which > wouldn't have received any responses. However I am not sure this removes > the concern of having the fault list on kernel side growing beyond > acceptable limits.
How about per-device quotas? (https://lkml.org/lkml/2018/4/23/706 for reference) With PRI the IOMMU driver already sets per-device credits when initializing the device (pci_enable_pri), so if the device behaves properly it shouldn't send new page requests once the number of outstanding ones is maxed out. The stall mode of SMMU doesn't have per-device limit, and depending on the implementation it might be easy for one guest using stall to prevent other guests from receiving faults. For this reason we'll have to enforce a per-device stall quota in the SMMU driver, and immediately terminate faults that exceed this quota. We could easily do the same for PRI, if we don't trust devices to follow the spec. The difficult part is finding the right number of credits... >> Host device drivers that use this API to be notified on fault can't deal >> with arch-specific event formats (SMMU event, Vt-d fault event, etc), so >> the APIs should be arch-agnostic. Given that requirement, using a single >> iommu_fault_event structure for both PRI and event queues made sense, >> especially since the even queue can have stall events that look a lot >> like PRI page requests. > I understand the data structure needs to be generic. Now I guess PRI > events and other standard translator error events (that can happen > without PRI) may have different characteristics in event fields, Right, an event contains more information than a PRI page request. Stage-2 fields (CLASS, S2, IPA, TTRnW) cannot be represented by iommu_fault_event at the moment. For precise emulation it might be useful to at least add the S2 flag (as a new iommu_fault_reason?), so that when the guest maps stage-1 to an invalid GPA, QEMU could for example inject an external abort. > queue > size, that may deserve to create different APIs and internal data > structs. Also this may help separating the concerns. It might duplicate them. If the consumer of the event report is a host device driver, the SMMU needs to report a "generic" iommu_fault_event, and if the consumer is VFIO it would report a specialized one > My remark also > stems from the fact the SMMU uses 2 different queues, whose size can > also be different. Hm, for PRI requests the kernel-userspace queue size should actually be the number of PRI credits for that device. Hadn't thought about it before, where do we pass that info to userspace? For fault events, the queue could be as big as the SMMU event queue, though using all that space might be wasteful. Non-stalled events should be rare and reporting them isn't urgent. Stalled ones would need the number of stall credits I mentioned above, which realistically will be a lot less than the SMMU event queue size. Given that a device will use either PRI or stall but not both, I still think events and PRI could go through the same queue. Thanks, Jean _______________________________________________ iommu mailing list iommu@lists.linux-foundation.org https://lists.linuxfoundation.org/mailman/listinfo/iommu