On Tue, Jan 07, 2025 at 09:10:09AM -0800, Nicolin Chen wrote:

> +static ssize_t iommufd_veventq_fops_read(struct iommufd_eventq *eventq,
> +                                      char __user *buf, size_t count,
> +                                      loff_t *ppos)
> +{
> +     size_t done = 0;
> +     int rc = 0;
> +
> +     if (*ppos)
> +             return -ESPIPE;
> +
> +     mutex_lock(&eventq->mutex);
> +     while (!list_empty(&eventq->deliver) && count > done) {
> +             struct iommufd_vevent *cur = list_first_entry(
> +                     &eventq->deliver, struct iommufd_vevent, node);
> +
> +             if (cur->data_len > count - done)
> +                     break;
> +
> +             if (copy_to_user(buf + done, cur->event_data, cur->data_len)) {
> +                     rc = -EFAULT;
> +                     break;
> +             }

Now that I look at this more closely, the fault path this is copied
from is not great.

This copy_to_user() can block while waiting on a page fault, possibily
for a long time. While blocked the mutex is held and we can't add more
entries to the list.

That will cause the shared IRQ handler in the iommu driver to back up,
which would cause a global DOS.

This probably wants to be organized to look more like

while (itm = eventq_get_next_item(eventq)) {
   if (..) {
       eventq_restore_failed_item(eventq);
       return -1;
   }
}

Where the next_item would just be a simple spinlock across the linked
list manipulation.

Jason

Reply via email to