On Wed, Oct 12, 2022 at 11:20:38AM -0600, Jason A. Donenfeld wrote: > On Wed, Oct 12, 2022 at 09:44:52AM -0700, Guenter Roeck wrote: > > On Wed, Oct 12, 2022 at 09:49:26AM -0600, Jason A. Donenfeld wrote: > > > On Wed, Oct 12, 2022 at 07:18:27AM -0700, Guenter Roeck wrote: > > > > NIP [c000000000031630] .replay_soft_interrupts+0x60/0x300 > > > > LR [c000000000031964] .arch_local_irq_restore+0x94/0x1c0 > > > > Call Trace: > > > > [c000000007df3870] [c000000000031964] > > > > .arch_local_irq_restore+0x94/0x1c0 (unreliable) > > > > [c000000007df38f0] [c000000000f8a444] .__schedule+0x664/0xa50 > > > > [c000000007df39d0] [c000000000f8a8b0] .schedule+0x80/0x140 > > > > [c000000007df3a50] [c00000000092f0dc] > > > > .try_to_generate_entropy+0x118/0x174 > > > > [c000000007df3b40] [c00000000092e2e4] .urandom_read_iter+0x74/0x140 > > > > [c000000007df3bc0] [c0000000003b0044] .vfs_read+0x284/0x2d0 > > > > [c000000007df3cd0] [c0000000003b0d2c] .ksys_read+0xdc/0x130 > > > > [c000000007df3d80] [c00000000002a88c] .system_call_exception+0x19c/0x330 > > > > [c000000007df3e10] [c00000000000c1d4] system_call_common+0xf4/0x258 > > > > > > Obviously the first couple lines of this concern me a bit. But I think > > > actually this might just be a catalyst for another bug. You could view > > > that function as basically just: > > > > > > while (something) > > > schedule(); > > > > > > And I guess in the process of calling the scheduler a lot, which toggles > > > interrupts a lot, something got wedged. > > > > > > Curious, though, I did try to reproduce this, to no avail. My .config is > > > https://xn--4db.cc/rBvHWfDZ . What's yours? > > > > > > > Attached. My qemu command line is > > Okay, thanks, I reproduced it. In this case, I suspect > try_to_generate_entropy() is just the messenger. There's an earlier > problem: >
That problem is not new but has existed for a couple of releases, and has never caused a hang until now. > BUG: using smp_processor_id() in preemptible [00000000] code: swapper/0/1 > caller is .__flush_tlb_pending+0x40/0xf0 > CPU: 0 PID: 1 Comm: swapper/0 Not tainted 6.0.0-28380-gde492c83cae0-dirty #4 > Hardware name: PowerMac3,1 PPC970FX 0x3c0301 PowerMac > Call Trace: > [c0000000044c3540] [c000000000f93ef0] .dump_stack_lvl+0x7c/0xc4 (unreliable) > [c0000000044c35d0] [c000000000fc9550] .check_preemption_disabled+0x140/0x150 > [c0000000044c3660] [c000000000073dd0] .__flush_tlb_pending+0x40/0xf0 > [c0000000044c36f0] [c000000000334434] .__apply_to_page_range+0x764/0xa30 > [c0000000044c3840] [c00000000006cad0] .change_memory_attr+0xf0/0x160 > [c0000000044c38d0] [c0000000002a1d70] .bpf_prog_select_runtime+0x150/0x230 > [c0000000044c3970] [c000000000d405d4] .bpf_prepare_filter+0x504/0x6f0 > [c0000000044c3a30] [c000000000d4085c] .bpf_prog_create+0x9c/0x140 > [c0000000044c3ac0] [c000000002051d9c] .ptp_classifier_init+0x44/0x78 > [c0000000044c3b50] [c000000002050f3c] .sock_init+0xe0/0x100 > [c0000000044c3bd0] [c000000000010bd4] .do_one_initcall+0xa4/0x438 > [c0000000044c3cc0] [c000000002005008] .kernel_init_freeable+0x378/0x428 > [c0000000044c3da0] [c0000000000113d8] .kernel_init+0x28/0x1a0 > [c0000000044c3e10] [c00000000000ca3c] .ret_from_kernel_thread+0x58/0x60 > > This in turn is because __flush_tlb_pending() calls: > > static inline int mm_is_thread_local(struct mm_struct *mm) > { > return cpumask_equal(mm_cpumask(mm), > cpumask_of(smp_processor_id())); > } > > __flush_tlb_pending() has a comment about this: > > * Must be called from within some kind of spinlock/non-preempt region... > */ > void __flush_tlb_pending(struct ppc64_tlb_batch *batch) > > So I guess that didn't happen for some reason? Maybe this is indicative > of some lock imbalance that then gets hit later? > > I've also managed to not hit this bug a few times. When it triggers, > after "kprobes: kprobe jump-optimization is enabled. All kprobes are > optimized if possible.", there's a long hang - tens seconds before it > continues. When it doesn't trigger, there's no hang at that point in the > boot process. > That probably explains why my attempts to bisect the problem were unsuccessful. Thanks, Guenter