Instead of suppressing the change of .text to become readonly, make the SMP locks patching code properly adjust/restore the page access rights.
On x86-64 additionally remove all mappings past the kernel image, and remove leftovers from the removal of the more general (but abandoned) SMP alternatives. Note that while I expect this to work properly on it own, it was only tested together with the change_page_attr() patch submitted before, which namely fixed the reference counting on split pages (which affected earlier versions of this patch). Signed-off-by: Jan Beulich <[EMAIL PROTECTED]> arch/i386/kernel/alternative.c | 67 +++++++++++++++++++++++++++++++++------ arch/i386/mm/init.c | 13 +------ arch/x86_64/kernel/vmlinux.lds.S | 9 ----- arch/x86_64/mm/init.c | 9 ++--- 4 files changed, 64 insertions(+), 34 deletions(-) --- linux-2.6.22-rc4/arch/i386/kernel/alternative.c 2007-06-11 18:09:52.000000000 +0200 +++ 2.6.22-rc4-x86-alt-page-attr/arch/i386/kernel/alternative.c 2007-06-11 09:13:59.000000000 +0200 @@ -2,8 +2,11 @@ #include <linux/sched.h> #include <linux/spinlock.h> #include <linux/list.h> +#include <linux/pfn.h> #include <asm/alternative.h> #include <asm/sections.h> +#include <asm/pgtable.h> +#include <asm/cacheflush.h> static int noreplace_smp = 0; static int smp_alt_once = 0; @@ -150,6 +153,46 @@ static void nop_out(void *insns, unsigne } } +#ifdef CONFIG_DEBUG_RODATA + +#ifdef CONFIG_X86_32 +#include <asm/highmem.h> +#define MODULES_VADDR VMALLOC_START +#endif + +static inline void make_writable(const void *instr, unsigned int len) +{ + unsigned long va = (unsigned long)instr; + + if (va < MODULES_VADDR) { + change_page_attr(virt_to_page(instr), + PFN_UP(va + len) - PFN_DOWN(va), + PAGE_KERNEL_EXEC); + global_flush_tlb(); + } +} + +static inline void make_readonly(const void *instr, unsigned int len) +{ + unsigned long va = (unsigned long)instr; + + if (va < MODULES_VADDR) { + change_page_attr(virt_to_page(instr), + PFN_UP(va + len) - PFN_DOWN(va), +#ifdef CONFIG_X86_64 + PAGE_KERNEL_RO); +#else + PAGE_KERNEL_RX); +#endif + global_flush_tlb(); + } +} + +#else /* !CONFIG_DEBUG_RODATA */ +#define make_writable(instr, len) ((void)0) +#define make_readonly(instr, len) ((void)0) +#endif + extern struct alt_instr __alt_instructions[], __alt_instructions_end[]; extern u8 *__smp_locks[], *__smp_locks_end[]; @@ -196,7 +239,9 @@ static void alternatives_smp_lock(u8 **s continue; if (*ptr > text_end) continue; + make_writable(*ptr, 1); **ptr = 0xf0; /* lock prefix */ + make_readonly(*ptr, 1); }; } @@ -212,7 +257,9 @@ static void alternatives_smp_unlock(u8 * continue; if (*ptr > text_end) continue; + make_writable(*ptr, 1); nop_out(*ptr, 1); + make_readonly(*ptr, 1); }; } @@ -239,7 +286,6 @@ void alternatives_smp_module_add(struct void *text, void *text_end) { struct smp_alt_module *smp; - unsigned long flags; if (noreplace_smp) return; @@ -265,39 +311,37 @@ void alternatives_smp_module_add(struct __FUNCTION__, smp->locks, smp->locks_end, smp->text, smp->text_end, smp->name); - spin_lock_irqsave(&smp_alt, flags); + spin_lock(&smp_alt); list_add_tail(&smp->next, &smp_alt_modules); if (boot_cpu_has(X86_FEATURE_UP)) alternatives_smp_unlock(smp->locks, smp->locks_end, smp->text, smp->text_end); - spin_unlock_irqrestore(&smp_alt, flags); + spin_unlock(&smp_alt); } void alternatives_smp_module_del(struct module *mod) { struct smp_alt_module *item; - unsigned long flags; if (smp_alt_once || noreplace_smp) return; - spin_lock_irqsave(&smp_alt, flags); + spin_lock(&smp_alt); list_for_each_entry(item, &smp_alt_modules, next) { if (mod != item->mod) continue; list_del(&item->next); - spin_unlock_irqrestore(&smp_alt, flags); + spin_unlock(&smp_alt); DPRINTK("%s: %s\n", __FUNCTION__, item->name); kfree(item); return; } - spin_unlock_irqrestore(&smp_alt, flags); + spin_unlock(&smp_alt); } void alternatives_smp_switch(int smp) { struct smp_alt_module *mod; - unsigned long flags; #ifdef CONFIG_LOCKDEP /* @@ -313,7 +357,7 @@ void alternatives_smp_switch(int smp) return; BUG_ON(!smp && (num_online_cpus() > 1)); - spin_lock_irqsave(&smp_alt, flags); + spin_lock(&smp_alt); if (smp) { printk(KERN_INFO "SMP alternatives: switching to SMP code\n"); clear_bit(X86_FEATURE_UP, boot_cpu_data.x86_capability); @@ -329,7 +373,7 @@ void alternatives_smp_switch(int smp) alternatives_smp_unlock(mod->locks, mod->locks_end, mod->text, mod->text_end); } - spin_unlock_irqrestore(&smp_alt, flags); + spin_unlock(&smp_alt); } #endif @@ -369,6 +413,7 @@ void __init alternative_instructions(voi local_irq_save(flags); apply_alternatives(__alt_instructions, __alt_instructions_end); + local_irq_restore(flags); /* switch to patch-once-at-boottime-only mode and free the * tables in case we know the number of CPUs will never ever @@ -399,6 +444,8 @@ void __init alternative_instructions(voi alternatives_smp_switch(0); } #endif + + local_irq_save(flags); apply_paravirt(__parainstructions, __parainstructions_end); local_irq_restore(flags); } --- linux-2.6.22-rc4/arch/i386/mm/init.c 2007-06-11 18:09:52.000000000 +0200 +++ 2.6.22-rc4-x86-alt-page-attr/arch/i386/mm/init.c 2007-06-11 09:13:59.000000000 +0200 @@ -799,16 +799,9 @@ void mark_rodata_ro(void) unsigned long start = PFN_ALIGN(_text); unsigned long size = PFN_ALIGN(_etext) - start; -#ifdef CONFIG_HOTPLUG_CPU - /* It must still be possible to apply SMP alternatives. */ - if (num_possible_cpus() <= 1) -#endif - { - change_page_attr(virt_to_page(start), - size >> PAGE_SHIFT, PAGE_KERNEL_RX); - printk("Write protecting the kernel text: %luk\n", size >> 10); - } - + change_page_attr(virt_to_page(start), + size >> PAGE_SHIFT, PAGE_KERNEL_RX); + printk("Write protecting the kernel text: %luk\n", size >> 10); start += size; size = (unsigned long)__end_rodata - start; change_page_attr(virt_to_page(start), --- linux-2.6.22-rc4/arch/x86_64/kernel/vmlinux.lds.S 2007-06-11 18:10:04.000000000 +0200 +++ 2.6.22-rc4-x86-alt-page-attr/arch/x86_64/kernel/vmlinux.lds.S 2007-06-11 09:13:59.000000000 +0200 @@ -131,20 +131,11 @@ SECTIONS /* might get freed after init */ . = ALIGN(4096); __smp_alt_begin = .; - __smp_alt_instructions = .; - .smp_altinstructions : AT(ADDR(.smp_altinstructions) - LOAD_OFFSET) { - *(.smp_altinstructions) - } - __smp_alt_instructions_end = .; - . = ALIGN(8); __smp_locks = .; .smp_locks : AT(ADDR(.smp_locks) - LOAD_OFFSET) { *(.smp_locks) } __smp_locks_end = .; - .smp_altinstr_replacement : AT(ADDR(.smp_altinstr_replacement) - LOAD_OFFSET) { - *(.smp_altinstr_replacement) - } . = ALIGN(4096); __smp_alt_end = .; --- linux-2.6.22-rc4/arch/x86_64/mm/init.c 2007-06-11 18:10:04.000000000 +0200 +++ 2.6.22-rc4-x86-alt-page-attr/arch/x86_64/mm/init.c 2007-06-11 10:29:55.000000000 +0200 @@ -590,6 +590,10 @@ void free_initmem(void) free_init_pages("unused kernel memory", (unsigned long)(&__init_begin), (unsigned long)(&__init_end)); + change_page_attr_addr(PFN_ALIGN(_end), + (KERNEL_TEXT_SIZE - __pa_symbol(_end)) >> PAGE_SHIFT, + __pgprot(0)); + global_flush_tlb(); } #ifdef CONFIG_DEBUG_RODATA @@ -598,11 +602,6 @@ void mark_rodata_ro(void) { unsigned long start = (unsigned long)_stext, end; -#ifdef CONFIG_HOTPLUG_CPU - /* It must still be possible to apply SMP alternatives. */ - if (num_possible_cpus() > 1) - start = (unsigned long)_etext; -#endif end = (unsigned long)__end_rodata; start = (start + PAGE_SIZE - 1) & PAGE_MASK; end &= PAGE_MASK; - To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to [EMAIL PROTECTED] More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/