On Wed, 11 Sep 2013 14:01:13 -0400 Konrad Rzeszutek Wilk <konrad.w...@oracle.com> wrote:
> <confused> > > I am thins would still work: > > > 47 static __always_inline void arch_spin_unlock(arch_spinlock_t *lock) > > 148 { > > 149 if (TICKET_SLOWPATH_FLAG && > > 150 static_key_false(¶virt_ticketlocks_enabled)) { > > > (from arch/x86/include/asm/spinlock.h) as the static_key_false > would check the key->enabled. Which had been incremented? > > Granted there are no patching done yet, but that should still allow > this code path to be taken? Lets look at static_key_false(): If jump labels is not enabled, you are correct. It simply looks like this: static __always_inline bool static_key_false(struct static_key *key) { if (unlikely(atomic_read(&key->enabled)) > 0) return true; return false; } But that's not the case here. Here we have code modifying jump labels, where static_key_false() looks like this: static __always_inline bool static_key_false(struct static_key *key) { return arch_static_branch(key); } static __always_inline bool arch_static_branch(struct static_key *key) { asm goto("1:" ".byte " __stringify(STATIC_KEY_INIT_NOP) "\n\t" ".pushsection __jump_table, \"aw\" \n\t" _ASM_ALIGN "\n\t" _ASM_PTR "1b, %l[l_yes], %c0 \n\t" ".popsection \n\t" : : "i" (key) : : l_yes); return false; l_yes: return true; } Look in that assembly. That "STATIC_KEY_INIT_NOP" is the byte code for a nop, and until we modify it, arch_static_branch() will always return false, no matter what "key->enable" is. In fact, your call trace you posted earlier proves my point! [ 4.966912] [<ffffffff810542e0>] ? poke_int3_handler+0x40/0x40 [ 4.966916] [<ffffffff816a0cf3>] dump_stack+0x59/0x7b [ 4.966920] [<ffffffff81051e0a>] __jump_label_transform+0x18a/0x230 [ 4.966923] [<ffffffff81162980>] ? fire_user_return_notifiers+0x70/0x70 [ 4.966926] [<ffffffff81051f15>] arch_jump_label_transform_static+0x65/0x90 [ 4.966930] [<ffffffff81cfbbfb>] jump_label_init+0x75/0xa3 [ 4.966932] [<ffffffff81cd3e3c>] start_kernel+0x168/0x3ff [ 4.966934] [<ffffffff81cd3af2>] ? repair_env_string+0x5b/0x5b [ 4.966938] [<ffffffff81cd35f3>] x86_64_start_reservations+0x2a/0x2c [ 4.966941] [<ffffffff81cd833a>] xen_start_kernel+0x594/0x596 This blew up in your patch: if (type == JUMP_LABEL_ENABLE) { /* * We are enabling this jump label. If it is not a nop * then something must have gone wrong. */ - if (unlikely(memcmp((void *)entry->code, ideal_nop, 5) != 0)) - bug_at((void *)entry->code, __LINE__); + if (init) { + if (unlikely(memcmp((void *)entry->code, ideal_nop, 5) != 0)) { + static int log = 0; + + if (log == 0) { + pr_warning("op %pS\n", (void *)entry->code); + dump_stack(); + } + log++; + } + } It was expecting to have the ideal nop, because on boot up it didn't expect to have something already marked for enable. It only thought this to be the case after initialization. This explains your origin error message: Unexpected op at trace_clock_global+0x6b/0x120 [ffffffff8113a21b] (0f 1f 44 00 00) The NOP was still the default nop, but it was expecting the ideal nop because it normally only gets into this path after the init was already done. My point is, it wasn't until jump_label_init() where it did the conversion from nop to calling the label. I'm looking to NAK your patch because it is obvious that the jump label code isn't doing what you expect it to be doing. And it wasn't until my checks were in place for you to notice. -- Steve -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majord...@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/