This switch most of the early feature check to use the non static key variant of the function. In later patches we will be switching cpu_has_feature and mmu_has_feature to use static keys and we can use them only after static key/jump label is initialized. Any check for feature before jump label init should be done using this new helper.
Signed-off-by: Aneesh Kumar K.V <aneesh.ku...@linux.vnet.ibm.com> --- arch/powerpc/include/asm/book3s/64/mmu-hash.h | 4 ++-- arch/powerpc/include/asm/book3s/64/pgtable.h | 2 +- arch/powerpc/kernel/paca.c | 2 +- arch/powerpc/kernel/setup-common.c | 6 +++--- arch/powerpc/kernel/setup_32.c | 14 +++++++------- arch/powerpc/kernel/setup_64.c | 12 ++++++------ arch/powerpc/kernel/smp.c | 2 +- arch/powerpc/kvm/book3s_hv_builtin.c | 2 +- arch/powerpc/mm/44x_mmu.c | 6 +++--- arch/powerpc/mm/hash_native_64.c | 2 +- arch/powerpc/mm/hash_utils_64.c | 12 ++++++------ arch/powerpc/mm/hugetlbpage.c | 2 +- arch/powerpc/mm/mmu_context_nohash.c | 4 ++-- arch/powerpc/mm/pgtable-hash64.c | 2 +- arch/powerpc/mm/ppc_mmu_32.c | 2 +- arch/powerpc/platforms/44x/iss4xx.c | 2 +- arch/powerpc/platforms/44x/ppc476.c | 2 +- arch/powerpc/platforms/85xx/smp.c | 6 +++--- arch/powerpc/platforms/cell/pervasive.c | 2 +- arch/powerpc/platforms/cell/smp.c | 2 +- arch/powerpc/platforms/powermac/setup.c | 2 +- arch/powerpc/platforms/powermac/smp.c | 4 ++-- arch/powerpc/platforms/powernv/setup.c | 2 +- arch/powerpc/platforms/powernv/smp.c | 4 ++-- arch/powerpc/platforms/powernv/subcore.c | 2 +- arch/powerpc/platforms/pseries/lpar.c | 4 ++-- arch/powerpc/platforms/pseries/smp.c | 6 +++--- 27 files changed, 56 insertions(+), 56 deletions(-) diff --git a/arch/powerpc/include/asm/book3s/64/mmu-hash.h b/arch/powerpc/include/asm/book3s/64/mmu-hash.h index 6ec21aad8ccc..e908a8cc1942 100644 --- a/arch/powerpc/include/asm/book3s/64/mmu-hash.h +++ b/arch/powerpc/include/asm/book3s/64/mmu-hash.h @@ -239,7 +239,7 @@ static inline unsigned long hpte_encode_avpn(unsigned long vpn, int psize, */ v = (vpn >> (23 - VPN_SHIFT)) & ~(mmu_psize_defs[psize].avpnm); v <<= HPTE_V_AVPN_SHIFT; - if (!cpu_has_feature(CPU_FTR_ARCH_300)) + if (!__cpu_has_feature(CPU_FTR_ARCH_300)) v |= ((unsigned long) ssize) << HPTE_V_SSIZE_SHIFT; return v; } @@ -267,7 +267,7 @@ static inline unsigned long hpte_encode_r(unsigned long pa, int base_psize, int actual_psize, int ssize) { - if (cpu_has_feature(CPU_FTR_ARCH_300)) + if (__cpu_has_feature(CPU_FTR_ARCH_300)) pa |= ((unsigned long) ssize) << HPTE_R_3_0_SSIZE_SHIFT; /* A 4K page needs no special encoding */ diff --git a/arch/powerpc/include/asm/book3s/64/pgtable.h b/arch/powerpc/include/asm/book3s/64/pgtable.h index d3ab97e3c744..bf3452fbfad6 100644 --- a/arch/powerpc/include/asm/book3s/64/pgtable.h +++ b/arch/powerpc/include/asm/book3s/64/pgtable.h @@ -805,7 +805,7 @@ static inline int __meminit vmemmap_create_mapping(unsigned long start, unsigned long page_size, unsigned long phys) { - if (radix_enabled()) + if (__radix_enabled()) return radix__vmemmap_create_mapping(start, page_size, phys); return hash__vmemmap_create_mapping(start, page_size, phys); } diff --git a/arch/powerpc/kernel/paca.c b/arch/powerpc/kernel/paca.c index 93dae296b6be..1b0b89e80824 100644 --- a/arch/powerpc/kernel/paca.c +++ b/arch/powerpc/kernel/paca.c @@ -184,7 +184,7 @@ void setup_paca(struct paca_struct *new_paca) * if we do a GET_PACA() before the feature fixups have been * applied */ - if (cpu_has_feature(CPU_FTR_HVMODE)) + if (__cpu_has_feature(CPU_FTR_HVMODE)) mtspr(SPRN_SPRG_HPACA, local_paca); #endif mtspr(SPRN_SPRG_PACA, local_paca); diff --git a/arch/powerpc/kernel/setup-common.c b/arch/powerpc/kernel/setup-common.c index 8ca79b7503d8..f43d2d76d81f 100644 --- a/arch/powerpc/kernel/setup-common.c +++ b/arch/powerpc/kernel/setup-common.c @@ -236,7 +236,7 @@ static int show_cpuinfo(struct seq_file *m, void *v) seq_printf(m, "unknown (%08x)", pvr); #ifdef CONFIG_ALTIVEC - if (cpu_has_feature(CPU_FTR_ALTIVEC)) + if (__cpu_has_feature(CPU_FTR_ALTIVEC)) seq_printf(m, ", altivec supported"); #endif /* CONFIG_ALTIVEC */ @@ -484,7 +484,7 @@ void __init smp_setup_cpu_maps(void) } /* If no SMT supported, nthreads is forced to 1 */ - if (!cpu_has_feature(CPU_FTR_SMT)) { + if (!__cpu_has_feature(CPU_FTR_SMT)) { DBG(" SMT disabled ! nthreads forced to 1\n"); nthreads = 1; } @@ -510,7 +510,7 @@ void __init smp_setup_cpu_maps(void) maxcpus = be32_to_cpup(ireg + num_addr_cell + num_size_cell); /* Double maxcpus for processors which have SMT capability */ - if (cpu_has_feature(CPU_FTR_SMT)) + if (__cpu_has_feature(CPU_FTR_SMT)) maxcpus *= nthreads; if (maxcpus > nr_cpu_ids) { diff --git a/arch/powerpc/kernel/setup_32.c b/arch/powerpc/kernel/setup_32.c index d544fa311757..ecdc42d44951 100644 --- a/arch/powerpc/kernel/setup_32.c +++ b/arch/powerpc/kernel/setup_32.c @@ -132,14 +132,14 @@ notrace void __init machine_init(u64 dt_ptr) setup_kdump_trampoline(); #ifdef CONFIG_6xx - if (cpu_has_feature(CPU_FTR_CAN_DOZE) || - cpu_has_feature(CPU_FTR_CAN_NAP)) + if (__cpu_has_feature(CPU_FTR_CAN_DOZE) || + __cpu_has_feature(CPU_FTR_CAN_NAP)) ppc_md.power_save = ppc6xx_idle; #endif #ifdef CONFIG_E500 - if (cpu_has_feature(CPU_FTR_CAN_DOZE) || - cpu_has_feature(CPU_FTR_CAN_NAP)) + if (__cpu_has_feature(CPU_FTR_CAN_DOZE) || + __cpu_has_feature(CPU_FTR_CAN_NAP)) ppc_md.power_save = e500_idle; #endif if (ppc_md.progress) @@ -149,7 +149,7 @@ notrace void __init machine_init(u64 dt_ptr) /* Checks "l2cr=xxxx" command-line option */ int __init ppc_setup_l2cr(char *str) { - if (cpu_has_feature(CPU_FTR_L2CR)) { + if (__cpu_has_feature(CPU_FTR_L2CR)) { unsigned long val = simple_strtoul(str, NULL, 0); printk(KERN_INFO "l2cr set to %lx\n", val); _set_L2CR(0); /* force invalidate by disable cache */ @@ -162,7 +162,7 @@ __setup("l2cr=", ppc_setup_l2cr); /* Checks "l3cr=xxxx" command-line option */ int __init ppc_setup_l3cr(char *str) { - if (cpu_has_feature(CPU_FTR_L3CR)) { + if (__cpu_has_feature(CPU_FTR_L3CR)) { unsigned long val = simple_strtoul(str, NULL, 0); printk(KERN_INFO "l3cr set to %lx\n", val); _set_L3CR(val); /* and enable it */ @@ -294,7 +294,7 @@ void __init setup_arch(char **cmdline_p) dcache_bsize = cur_cpu_spec->dcache_bsize; icache_bsize = cur_cpu_spec->icache_bsize; ucache_bsize = 0; - if (cpu_has_feature(CPU_FTR_UNIFIED_ID_CACHE)) + if (__cpu_has_feature(CPU_FTR_UNIFIED_ID_CACHE)) ucache_bsize = icache_bsize = dcache_bsize; if (ppc_md.panic) diff --git a/arch/powerpc/kernel/setup_64.c b/arch/powerpc/kernel/setup_64.c index 5530bb55a78b..05dde6318b79 100644 --- a/arch/powerpc/kernel/setup_64.c +++ b/arch/powerpc/kernel/setup_64.c @@ -125,7 +125,7 @@ static void setup_tlb_core_data(void) * will be racy and could produce duplicate entries. */ if (smt_enabled_at_boot >= 2 && - !mmu_has_feature(MMU_FTR_USE_TLBRSRV) && + !__mmu_has_feature(MMU_FTR_USE_TLBRSRV) && book3e_htw_mode != PPC_HTW_E6500) { /* Should we panic instead? */ WARN_ONCE("%s: unsupported MMU configuration -- expect problems\n", @@ -216,8 +216,8 @@ static void cpu_ready_for_interrupts(void) * not in hypervisor mode, we enable relocation-on interrupts later * in pSeries_setup_arch() using the H_SET_MODE hcall. */ - if (cpu_has_feature(CPU_FTR_HVMODE) && - cpu_has_feature(CPU_FTR_ARCH_207S)) { + if (__cpu_has_feature(CPU_FTR_HVMODE) && + __cpu_has_feature(CPU_FTR_ARCH_207S)) { unsigned long lpcr = mfspr(SPRN_LPCR); mtspr(SPRN_LPCR, lpcr | LPCR_AIL_3); } @@ -588,13 +588,13 @@ static u64 safe_stack_limit(void) { #ifdef CONFIG_PPC_BOOK3E /* Freescale BookE bolts the entire linear mapping */ - if (mmu_has_feature(MMU_FTR_TYPE_FSL_E)) + if (__mmu_has_feature(MMU_FTR_TYPE_FSL_E)) return linear_map_top; /* Other BookE, we assume the first GB is bolted */ return 1ul << 30; #else /* BookS, the first segment is bolted */ - if (mmu_has_feature(MMU_FTR_1T_SEGMENT)) + if (__mmu_has_feature(MMU_FTR_1T_SEGMENT)) return 1UL << SID_SHIFT_1T; return 1UL << SID_SHIFT; #endif @@ -639,7 +639,7 @@ static void __init exc_lvl_early_init(void) paca[i].mc_kstack = __va(sp + THREAD_SIZE); } - if (cpu_has_feature(CPU_FTR_DEBUG_LVL_EXC)) + if (__cpu_has_feature(CPU_FTR_DEBUG_LVL_EXC)) patch_exception(0x040, exc_debug_debug_book3e); } #else diff --git a/arch/powerpc/kernel/smp.c b/arch/powerpc/kernel/smp.c index 5a1f015ea9f3..d1a7234c1c33 100644 --- a/arch/powerpc/kernel/smp.c +++ b/arch/powerpc/kernel/smp.c @@ -96,7 +96,7 @@ int smp_generic_cpu_bootable(unsigned int nr) /* Special case - we inhibit secondary thread startup * during boot if the user requests it. */ - if (system_state == SYSTEM_BOOTING && cpu_has_feature(CPU_FTR_SMT)) { + if (system_state == SYSTEM_BOOTING && __cpu_has_feature(CPU_FTR_SMT)) { if (!smt_enabled_at_boot && cpu_thread_in_core(nr) != 0) return 0; if (smt_enabled_at_boot diff --git a/arch/powerpc/kvm/book3s_hv_builtin.c b/arch/powerpc/kvm/book3s_hv_builtin.c index 5f0380db3eab..cadb2d0f9892 100644 --- a/arch/powerpc/kvm/book3s_hv_builtin.c +++ b/arch/powerpc/kvm/book3s_hv_builtin.c @@ -80,7 +80,7 @@ void __init kvm_cma_reserve(void) /* * We need CMA reservation only when we are in HV mode */ - if (!cpu_has_feature(CPU_FTR_HVMODE)) + if (!__cpu_has_feature(CPU_FTR_HVMODE)) return; /* * We cannot use memblock_phys_mem_size() here, because diff --git a/arch/powerpc/mm/44x_mmu.c b/arch/powerpc/mm/44x_mmu.c index 82b1ff759e26..0b17851b0f90 100644 --- a/arch/powerpc/mm/44x_mmu.c +++ b/arch/powerpc/mm/44x_mmu.c @@ -187,12 +187,12 @@ unsigned long __init mmu_mapin_ram(unsigned long top) * initial 256M mapping established in head_44x.S */ for (addr = memstart + PPC_PIN_SIZE; addr < lowmem_end_addr; addr += PPC_PIN_SIZE) { - if (mmu_has_feature(MMU_FTR_TYPE_47x)) + if (__mmu_has_feature(MMU_FTR_TYPE_47x)) ppc47x_pin_tlb(addr + PAGE_OFFSET, addr); else ppc44x_pin_tlb(addr + PAGE_OFFSET, addr); } - if (mmu_has_feature(MMU_FTR_TYPE_47x)) { + if (__mmu_has_feature(MMU_FTR_TYPE_47x)) { ppc47x_update_boltmap(); #ifdef DEBUG @@ -245,7 +245,7 @@ void mmu_init_secondary(int cpu) */ for (addr = memstart + PPC_PIN_SIZE; addr < lowmem_end_addr; addr += PPC_PIN_SIZE) { - if (mmu_has_feature(MMU_FTR_TYPE_47x)) + if (__mmu_has_feature(MMU_FTR_TYPE_47x)) ppc47x_pin_tlb(addr + PAGE_OFFSET, addr); else ppc44x_pin_tlb(addr + PAGE_OFFSET, addr); diff --git a/arch/powerpc/mm/hash_native_64.c b/arch/powerpc/mm/hash_native_64.c index 277047528a3a..2208780587a0 100644 --- a/arch/powerpc/mm/hash_native_64.c +++ b/arch/powerpc/mm/hash_native_64.c @@ -746,6 +746,6 @@ void __init hpte_init_native(void) ppc_md.flush_hash_range = native_flush_hash_range; ppc_md.hugepage_invalidate = native_hugepage_invalidate; - if (cpu_has_feature(CPU_FTR_ARCH_300)) + if (__cpu_has_feature(CPU_FTR_ARCH_300)) ppc_md.register_process_table = native_register_proc_table; } diff --git a/arch/powerpc/mm/hash_utils_64.c b/arch/powerpc/mm/hash_utils_64.c index 47d59a1f12f1..3509337502f6 100644 --- a/arch/powerpc/mm/hash_utils_64.c +++ b/arch/powerpc/mm/hash_utils_64.c @@ -529,7 +529,7 @@ static bool might_have_hea(void) * we will never see an HEA ethernet device. */ #ifdef CONFIG_IBMEBUS - return !cpu_has_feature(CPU_FTR_ARCH_207S); + return !__cpu_has_feature(CPU_FTR_ARCH_207S); #else return false; #endif @@ -559,7 +559,7 @@ static void __init htab_init_page_sizes(void) * Not in the device-tree, let's fallback on known size * list for 16M capable GP & GR */ - if (mmu_has_feature(MMU_FTR_16M_PAGE)) + if (__mmu_has_feature(MMU_FTR_16M_PAGE)) memcpy(mmu_psize_defs, mmu_psize_defaults_gp, sizeof(mmu_psize_defaults_gp)); found: @@ -589,7 +589,7 @@ found: mmu_vmalloc_psize = MMU_PAGE_64K; if (mmu_linear_psize == MMU_PAGE_4K) mmu_linear_psize = MMU_PAGE_64K; - if (mmu_has_feature(MMU_FTR_CI_LARGE_PAGE)) { + if (__mmu_has_feature(MMU_FTR_CI_LARGE_PAGE)) { /* * When running on pSeries using 64k pages for ioremap * would stop us accessing the HEA ethernet. So if we @@ -763,7 +763,7 @@ static void __init htab_initialize(void) /* Initialize page sizes */ htab_init_page_sizes(); - if (mmu_has_feature(MMU_FTR_1T_SEGMENT)) { + if (__mmu_has_feature(MMU_FTR_1T_SEGMENT)) { mmu_kernel_ssize = MMU_SEGSIZE_1T; mmu_highuser_ssize = MMU_SEGSIZE_1T; printk(KERN_INFO "Using 1TB segments\n"); @@ -815,7 +815,7 @@ static void __init htab_initialize(void) /* Initialize the HPT with no entries */ memset((void *)table, 0, htab_size_bytes); - if (!cpu_has_feature(CPU_FTR_ARCH_300)) + if (!__cpu_has_feature(CPU_FTR_ARCH_300)) /* Set SDR1 */ mtspr(SPRN_SDR1, _SDR1); else @@ -952,7 +952,7 @@ void hash__early_init_mmu_secondary(void) { /* Initialize hash table for that CPU */ if (!firmware_has_feature(FW_FEATURE_LPAR)) { - if (!cpu_has_feature(CPU_FTR_ARCH_300)) + if (!__cpu_has_feature(CPU_FTR_ARCH_300)) mtspr(SPRN_SDR1, _SDR1); else mtspr(SPRN_PTCR, diff --git a/arch/powerpc/mm/hugetlbpage.c b/arch/powerpc/mm/hugetlbpage.c index 119d18611500..3be9c9e918b6 100644 --- a/arch/powerpc/mm/hugetlbpage.c +++ b/arch/powerpc/mm/hugetlbpage.c @@ -828,7 +828,7 @@ static int __init hugetlbpage_init(void) { int psize; - if (!radix_enabled() && !mmu_has_feature(MMU_FTR_16M_PAGE)) + if (!radix_enabled() && !__mmu_has_feature(MMU_FTR_16M_PAGE)) return -ENODEV; for (psize = 0; psize < MMU_PAGE_COUNT; ++psize) { diff --git a/arch/powerpc/mm/mmu_context_nohash.c b/arch/powerpc/mm/mmu_context_nohash.c index 7d95bc402dba..4ec513e506fb 100644 --- a/arch/powerpc/mm/mmu_context_nohash.c +++ b/arch/powerpc/mm/mmu_context_nohash.c @@ -442,11 +442,11 @@ void __init mmu_context_init(void) * present if needed. * -- BenH */ - if (mmu_has_feature(MMU_FTR_TYPE_8xx)) { + if (__mmu_has_feature(MMU_FTR_TYPE_8xx)) { first_context = 0; last_context = 15; no_selective_tlbil = true; - } else if (mmu_has_feature(MMU_FTR_TYPE_47x)) { + } else if (__mmu_has_feature(MMU_FTR_TYPE_47x)) { first_context = 1; last_context = 65535; no_selective_tlbil = false; diff --git a/arch/powerpc/mm/pgtable-hash64.c b/arch/powerpc/mm/pgtable-hash64.c index c23e286a6b8f..d9b5804bdce9 100644 --- a/arch/powerpc/mm/pgtable-hash64.c +++ b/arch/powerpc/mm/pgtable-hash64.c @@ -313,7 +313,7 @@ pmd_t hash__pmdp_huge_get_and_clear(struct mm_struct *mm, int hash__has_transparent_hugepage(void) { - if (!mmu_has_feature(MMU_FTR_16M_PAGE)) + if (!__mmu_has_feature(MMU_FTR_16M_PAGE)) return 0; /* * We support THP only if PMD_SIZE is 16MB. diff --git a/arch/powerpc/mm/ppc_mmu_32.c b/arch/powerpc/mm/ppc_mmu_32.c index 2a049fb8523d..0915733d8ae4 100644 --- a/arch/powerpc/mm/ppc_mmu_32.c +++ b/arch/powerpc/mm/ppc_mmu_32.c @@ -187,7 +187,7 @@ void __init MMU_init_hw(void) extern unsigned int hash_page[]; extern unsigned int flush_hash_patch_A[], flush_hash_patch_B[]; - if (!mmu_has_feature(MMU_FTR_HPTE_TABLE)) { + if (!__mmu_has_feature(MMU_FTR_HPTE_TABLE)) { /* * Put a blr (procedure return) instruction at the * start of hash_page, since we can still get DSI diff --git a/arch/powerpc/platforms/44x/iss4xx.c b/arch/powerpc/platforms/44x/iss4xx.c index c7c6758b3cfe..506b711828b0 100644 --- a/arch/powerpc/platforms/44x/iss4xx.c +++ b/arch/powerpc/platforms/44x/iss4xx.c @@ -131,7 +131,7 @@ static struct smp_ops_t iss_smp_ops = { static void __init iss4xx_smp_init(void) { - if (mmu_has_feature(MMU_FTR_TYPE_47x)) + if (__mmu_has_feature(MMU_FTR_TYPE_47x)) smp_ops = &iss_smp_ops; } diff --git a/arch/powerpc/platforms/44x/ppc476.c b/arch/powerpc/platforms/44x/ppc476.c index c11ce6516c8f..895dc63d6a49 100644 --- a/arch/powerpc/platforms/44x/ppc476.c +++ b/arch/powerpc/platforms/44x/ppc476.c @@ -201,7 +201,7 @@ static struct smp_ops_t ppc47x_smp_ops = { static void __init ppc47x_smp_init(void) { - if (mmu_has_feature(MMU_FTR_TYPE_47x)) + if (__mmu_has_feature(MMU_FTR_TYPE_47x)) smp_ops = &ppc47x_smp_ops; } diff --git a/arch/powerpc/platforms/85xx/smp.c b/arch/powerpc/platforms/85xx/smp.c index fe9f19e5e935..a4705d964187 100644 --- a/arch/powerpc/platforms/85xx/smp.c +++ b/arch/powerpc/platforms/85xx/smp.c @@ -280,7 +280,7 @@ static int smp_85xx_kick_cpu(int nr) #ifdef CONFIG_PPC64 if (threads_per_core == 2) { - if (WARN_ON_ONCE(!cpu_has_feature(CPU_FTR_SMT))) + if (WARN_ON_ONCE(!__cpu_has_feature(CPU_FTR_SMT))) return -ENOENT; booting_thread_hwid = cpu_thread_in_core(nr); @@ -462,7 +462,7 @@ static void mpc85xx_smp_machine_kexec(struct kimage *image) static void smp_85xx_basic_setup(int cpu_nr) { - if (cpu_has_feature(CPU_FTR_DBELL)) + if (__cpu_has_feature(CPU_FTR_DBELL)) doorbell_setup_this_cpu(); } @@ -485,7 +485,7 @@ void __init mpc85xx_smp_init(void) } else smp_85xx_ops.setup_cpu = smp_85xx_basic_setup; - if (cpu_has_feature(CPU_FTR_DBELL)) { + if (__cpu_has_feature(CPU_FTR_DBELL)) { /* * If left NULL, .message_pass defaults to * smp_muxed_ipi_message_pass diff --git a/arch/powerpc/platforms/cell/pervasive.c b/arch/powerpc/platforms/cell/pervasive.c index d17e98bc0c10..f053602e63fa 100644 --- a/arch/powerpc/platforms/cell/pervasive.c +++ b/arch/powerpc/platforms/cell/pervasive.c @@ -115,7 +115,7 @@ void __init cbe_pervasive_init(void) { int cpu; - if (!cpu_has_feature(CPU_FTR_PAUSE_ZERO)) + if (!__cpu_has_feature(CPU_FTR_PAUSE_ZERO)) return; for_each_possible_cpu(cpu) { diff --git a/arch/powerpc/platforms/cell/smp.c b/arch/powerpc/platforms/cell/smp.c index 895560f4be69..4d373c6375a8 100644 --- a/arch/powerpc/platforms/cell/smp.c +++ b/arch/powerpc/platforms/cell/smp.c @@ -148,7 +148,7 @@ void __init smp_init_cell(void) smp_ops = &bpa_iic_smp_ops; /* Mark threads which are still spinning in hold loops. */ - if (cpu_has_feature(CPU_FTR_SMT)) { + if (__cpu_has_feature(CPU_FTR_SMT)) { for_each_present_cpu(i) { if (cpu_thread_in_core(i) == 0) cpumask_set_cpu(i, &of_spin_map); diff --git a/arch/powerpc/platforms/powermac/setup.c b/arch/powerpc/platforms/powermac/setup.c index 8dd78f4e1af4..615bb39b82d3 100644 --- a/arch/powerpc/platforms/powermac/setup.c +++ b/arch/powerpc/platforms/powermac/setup.c @@ -248,7 +248,7 @@ static void __init ohare_init(void) static void __init l2cr_init(void) { /* Checks "l2cr-value" property in the registry */ - if (cpu_has_feature(CPU_FTR_L2CR)) { + if (__cpu_has_feature(CPU_FTR_L2CR)) { struct device_node *np = of_find_node_by_name(NULL, "cpus"); if (np == 0) np = of_find_node_by_type(NULL, "cpu"); diff --git a/arch/powerpc/platforms/powermac/smp.c b/arch/powerpc/platforms/powermac/smp.c index 28a147ca32ba..d917ebad551e 100644 --- a/arch/powerpc/platforms/powermac/smp.c +++ b/arch/powerpc/platforms/powermac/smp.c @@ -670,7 +670,7 @@ volatile static long int core99_l3_cache; static void core99_init_caches(int cpu) { #ifndef CONFIG_PPC64 - if (!cpu_has_feature(CPU_FTR_L2CR)) + if (!__cpu_has_feature(CPU_FTR_L2CR)) return; if (cpu == 0) { @@ -683,7 +683,7 @@ static void core99_init_caches(int cpu) printk("CPU%d: L2CR set to %lx\n", cpu, core99_l2_cache); } - if (!cpu_has_feature(CPU_FTR_L3CR)) + if (!__cpu_has_feature(CPU_FTR_L3CR)) return; if (cpu == 0){ diff --git a/arch/powerpc/platforms/powernv/setup.c b/arch/powerpc/platforms/powernv/setup.c index 8492bbbcfc08..607a05233119 100644 --- a/arch/powerpc/platforms/powernv/setup.c +++ b/arch/powerpc/platforms/powernv/setup.c @@ -273,7 +273,7 @@ static int __init pnv_probe(void) if (!of_flat_dt_is_compatible(root, "ibm,powernv")) return 0; - if (IS_ENABLED(CONFIG_PPC_RADIX_MMU) && radix_enabled()) + if (IS_ENABLED(CONFIG_PPC_RADIX_MMU) && __radix_enabled()) radix_init_native(); else if (IS_ENABLED(CONFIG_PPC_STD_MMU_64)) hpte_init_native(); diff --git a/arch/powerpc/platforms/powernv/smp.c b/arch/powerpc/platforms/powernv/smp.c index ad7b1a3dbed0..a9f20306d305 100644 --- a/arch/powerpc/platforms/powernv/smp.c +++ b/arch/powerpc/platforms/powernv/smp.c @@ -50,7 +50,7 @@ static void pnv_smp_setup_cpu(int cpu) xics_setup_cpu(); #ifdef CONFIG_PPC_DOORBELL - if (cpu_has_feature(CPU_FTR_DBELL)) + if (__cpu_has_feature(CPU_FTR_DBELL)) doorbell_setup_this_cpu(); #endif } @@ -233,7 +233,7 @@ static int pnv_cpu_bootable(unsigned int nr) * switches. So on those machines we ignore the smt_enabled_at_boot * setting (smt-enabled on the kernel command line). */ - if (cpu_has_feature(CPU_FTR_ARCH_207S)) + if (__cpu_has_feature(CPU_FTR_ARCH_207S)) return 1; return smp_generic_cpu_bootable(nr); diff --git a/arch/powerpc/platforms/powernv/subcore.c b/arch/powerpc/platforms/powernv/subcore.c index 0babef11136f..abf308fbb385 100644 --- a/arch/powerpc/platforms/powernv/subcore.c +++ b/arch/powerpc/platforms/powernv/subcore.c @@ -407,7 +407,7 @@ static DEVICE_ATTR(subcores_per_core, 0644, static int subcore_init(void) { - if (!cpu_has_feature(CPU_FTR_SUBCORE)) + if (!__cpu_has_feature(CPU_FTR_SUBCORE)) return 0; /* diff --git a/arch/powerpc/platforms/pseries/lpar.c b/arch/powerpc/platforms/pseries/lpar.c index 03ff9867a610..a54de1cff935 100644 --- a/arch/powerpc/platforms/pseries/lpar.c +++ b/arch/powerpc/platforms/pseries/lpar.c @@ -76,10 +76,10 @@ void vpa_init(int cpu) */ WARN_ON(cpu != smp_processor_id()); - if (cpu_has_feature(CPU_FTR_ALTIVEC)) + if (__cpu_has_feature(CPU_FTR_ALTIVEC)) lppaca_of(cpu).vmxregs_in_use = 1; - if (cpu_has_feature(CPU_FTR_ARCH_207S)) + if (__cpu_has_feature(CPU_FTR_ARCH_207S)) lppaca_of(cpu).ebb_regs_in_use = 1; addr = __pa(&lppaca_of(cpu)); diff --git a/arch/powerpc/platforms/pseries/smp.c b/arch/powerpc/platforms/pseries/smp.c index f6f83aeccaaa..57111bae6eec 100644 --- a/arch/powerpc/platforms/pseries/smp.c +++ b/arch/powerpc/platforms/pseries/smp.c @@ -143,7 +143,7 @@ static void smp_setup_cpu(int cpu) { if (cpu != boot_cpuid) xics_setup_cpu(); - if (cpu_has_feature(CPU_FTR_DBELL)) + if (__cpu_has_feature(CPU_FTR_DBELL)) doorbell_setup_this_cpu(); if (firmware_has_feature(FW_FEATURE_SPLPAR)) @@ -200,7 +200,7 @@ static __init void pSeries_smp_probe(void) { xics_smp_probe(); - if (cpu_has_feature(CPU_FTR_DBELL)) { + if (__cpu_has_feature(CPU_FTR_DBELL)) { xics_cause_ipi = smp_ops->cause_ipi; smp_ops->cause_ipi = pSeries_cause_ipi_mux; } @@ -232,7 +232,7 @@ void __init smp_init_pseries(void) * query-cpu-stopped-state. */ if (rtas_token("query-cpu-stopped-state") == RTAS_UNKNOWN_SERVICE) { - if (cpu_has_feature(CPU_FTR_SMT)) { + if (__cpu_has_feature(CPU_FTR_SMT)) { for_each_present_cpu(i) { if (cpu_thread_in_core(i) == 0) cpumask_set_cpu(i, of_spin_mask); -- 2.7.4 _______________________________________________ Linuxppc-dev mailing list Linuxppc-dev@lists.ozlabs.org https://lists.ozlabs.org/listinfo/linuxppc-dev