On Tue, 16 Jan 2018, Keith Busch wrote:

> This is all way over my head, but the part that obviously shows
> something's gone wrong:
> 
>   kworker/u674:3-1421  [028] d...   335.307051: irq_matrix_reserve_managed: 
> bit=56 cpu=0 online=1 avl=86 alloc=116 managed=3 online_maps=112 
> global_avl=22084, global_rsvd=157, total_alloc=570
>   kworker/u674:3-1421  [028] d...   335.307053: irq_matrix_remove_managed: 
> bit=56 cpu=0 online=1 avl=87 alloc=116 managed=2 online_maps=112 
> global_avl=22085, global_rsvd=157, total_alloc=570
>   kworker/u674:3-1421  [028] ....   335.307054: vector_reserve_managed: 
> irq=45 ret=-28
>   kworker/u674:3-1421  [028] ....   335.307054: vector_setup: irq=45 
> is_legacy=0 ret=-28
>   kworker/u674:3-1421  [028] d...   335.307055: vector_teardown: irq=45 
> is_managed=1 has_reserved=0
> 
> Which leads me to x86_vector_alloc_irqs goto error:
> 
> error:
>       x86_vector_free_irqs(domain, virq, i + 1);
> 
> The last parameter looks weird. It's the nr_irqs, and since we failed and
> bailed, I would think we'd need to subtract 1 rather than add 1. Adding
> 1 would doublely remove the failed one, and remove the next one that
> was never setup, right?

Right. That's fishy. Let me stare at it.

> Or maybe irq_matrix_reserve_managed wasn't expected to fail in the
> first place?

Well, it can faul. I don't know why it fails in that case, but let me look
a bit more.

Thanks,

        tglx


Reply via email to