Add num_{possible,present,active}_cpus_{eq,gt,le} and replace num_*_cpus()
with one of new functions where appropriate. This allows num_*_cpus_*()
to return earlier depending on the condition.

Signed-off-by: Yury Norov <yury.no...@gmail.com>
---
 arch/arc/kernel/smp.c                         |  2 +-
 arch/arm/kernel/machine_kexec.c               |  2 +-
 arch/arm/mach-exynos/exynos.c                 |  2 +-
 arch/arm/mm/cache-b15-rac.c                   |  2 +-
 arch/arm64/kernel/smp.c                       |  2 +-
 arch/arm64/mm/context.c                       |  2 +-
 arch/csky/mm/asid.c                           |  2 +-
 arch/csky/mm/context.c                        |  2 +-
 arch/ia64/mm/tlb.c                            |  6 ++---
 arch/mips/kernel/i8253.c                      |  2 +-
 arch/mips/kernel/perf_event_mipsxx.c          |  4 ++--
 arch/mips/kernel/rtlx-cmp.c                   |  2 +-
 arch/mips/kernel/smp.c                        |  4 ++--
 arch/mips/kernel/vpe-cmp.c                    |  2 +-
 .../loongson2ef/common/cs5536/cs5536_mfgpt.c  |  2 +-
 arch/mips/mm/context.c                        |  2 +-
 arch/mips/mm/tlbex.c                          |  2 +-
 arch/nios2/kernel/cpuinfo.c                   |  2 +-
 arch/powerpc/platforms/85xx/smp.c             |  2 +-
 arch/powerpc/platforms/pseries/hotplug-cpu.c  |  4 ++--
 arch/powerpc/sysdev/mpic.c                    |  2 +-
 arch/powerpc/xmon/xmon.c                      |  6 ++---
 arch/riscv/kvm/vmid.c                         |  2 +-
 arch/sparc/kernel/mdesc.c                     |  6 ++---
 arch/x86/events/amd/core.c                    |  2 +-
 arch/x86/kernel/alternative.c                 |  8 +++----
 arch/x86/kernel/apic/apic.c                   |  4 ++--
 arch/x86/kernel/apic/apic_flat_64.c           |  2 +-
 arch/x86/kernel/apic/probe_32.c               |  2 +-
 arch/x86/kernel/cpu/mce/dev-mcelog.c          |  2 +-
 arch/x86/kernel/hpet.c                        |  2 +-
 arch/x86/kernel/i8253.c                       |  2 +-
 arch/x86/kernel/kvm.c                         |  2 +-
 arch/x86/kernel/kvmclock.c                    |  2 +-
 arch/x86/kernel/tsc.c                         |  2 +-
 arch/x86/xen/smp_pv.c                         |  2 +-
 arch/x86/xen/spinlock.c                       |  2 +-
 drivers/clk/samsung/clk-exynos4.c             |  2 +-
 drivers/clocksource/ingenic-timer.c           |  3 +--
 drivers/cpufreq/pcc-cpufreq.c                 |  2 +-
 drivers/dma/mv_xor.c                          |  5 ++--
 drivers/gpu/drm/i810/i810_drv.c               |  2 +-
 drivers/irqchip/irq-gic.c                     |  2 +-
 drivers/net/caif/caif_virtio.c                |  2 +-
 .../cavium/liquidio/cn23xx_vf_device.c        |  2 +-
 drivers/net/ethernet/hisilicon/hns/hns_enet.c |  2 +-
 .../net/ethernet/marvell/mvpp2/mvpp2_main.c   |  2 +-
 drivers/net/wireless/ath/ath9k/hw.c           |  2 +-
 drivers/net/wireless/marvell/mwifiex/main.c   |  4 ++--
 drivers/net/wireless/st/cw1200/queue.c        |  3 +--
 drivers/nvdimm/region.c                       |  2 +-
 drivers/nvme/host/pci.c                       |  2 +-
 drivers/perf/arm_pmu.c                        |  2 +-
 .../intel/speed_select_if/isst_if_common.c    |  6 ++---
 drivers/soc/bcm/brcmstb/biuctrl.c             |  2 +-
 drivers/soc/fsl/dpio/dpio-service.c           |  4 ++--
 drivers/spi/spi-dw-bt1.c                      |  2 +-
 drivers/virt/acrn/hsm.c                       |  2 +-
 fs/xfs/xfs_sysfs.c                            |  2 +-
 include/linux/cpumask.h                       | 23 +++++++++++++++++++
 include/linux/kdb.h                           |  2 +-
 kernel/debug/kdb/kdb_bt.c                     |  2 +-
 kernel/printk/printk.c                        |  2 +-
 kernel/reboot.c                               |  4 ++--
 kernel/time/clockevents.c                     |  2 +-
 mm/percpu.c                                   |  6 ++---
 mm/slab.c                                     |  2 +-
 67 files changed, 110 insertions(+), 90 deletions(-)

diff --git a/arch/arc/kernel/smp.c b/arch/arc/kernel/smp.c
index 78e6d069b1c1..d4f2765755c9 100644
--- a/arch/arc/kernel/smp.c
+++ b/arch/arc/kernel/smp.c
@@ -103,7 +103,7 @@ void __init smp_prepare_cpus(unsigned int max_cpus)
         * if platform didn't set the present map already, do it now
         * boot cpu is set to present already by init/main.c
         */
-       if (num_present_cpus() <= 1)
+       if (num_present_cpus_le(2))
                init_cpu_present(cpu_possible_mask);
 }
 
diff --git a/arch/arm/kernel/machine_kexec.c b/arch/arm/kernel/machine_kexec.c
index f567032a09c0..8875e2ee0083 100644
--- a/arch/arm/kernel/machine_kexec.c
+++ b/arch/arm/kernel/machine_kexec.c
@@ -44,7 +44,7 @@ int machine_kexec_prepare(struct kimage *image)
         * and implements CPU hotplug for the current HW. If not, we won't be
         * able to kexec reliably, so fail the prepare operation.
         */
-       if (num_possible_cpus() > 1 && platform_can_secondary_boot() &&
+       if (num_possible_cpus_gt(1) && platform_can_secondary_boot() &&
            !platform_can_cpu_hotplug())
                return -EINVAL;
 
diff --git a/arch/arm/mach-exynos/exynos.c b/arch/arm/mach-exynos/exynos.c
index 8b48326be9fd..ba658402ac1e 100644
--- a/arch/arm/mach-exynos/exynos.c
+++ b/arch/arm/mach-exynos/exynos.c
@@ -120,7 +120,7 @@ void exynos_set_delayed_reset_assertion(bool enable)
        if (of_machine_is_compatible("samsung,exynos4")) {
                unsigned int tmp, core_id;
 
-               for (core_id = 0; core_id < num_possible_cpus(); core_id++) {
+               for (core_id = 0; num_possible_cpus_gt(core_id); core_id++) {
                        tmp = pmu_raw_readl(EXYNOS_ARM_CORE_OPTION(core_id));
                        if (enable)
                                tmp |= S5P_USE_DELAYED_RESET_ASSERTION;
diff --git a/arch/arm/mm/cache-b15-rac.c b/arch/arm/mm/cache-b15-rac.c
index bdc07030997b..202c3a6cf98b 100644
--- a/arch/arm/mm/cache-b15-rac.c
+++ b/arch/arm/mm/cache-b15-rac.c
@@ -296,7 +296,7 @@ static int __init b15_rac_init(void)
        if (!dn)
                return -ENODEV;
 
-       if (WARN(num_possible_cpus() > 4, "RAC only supports 4 CPUs\n"))
+       if (WARN(num_possible_cpus_gt(4), "RAC only supports 4 CPUs\n"))
                goto out;
 
        b15_rac_base = of_iomap(dn, 0);
diff --git a/arch/arm64/kernel/smp.c b/arch/arm64/kernel/smp.c
index 27df5c1e6baa..bd1280e5081b 100644
--- a/arch/arm64/kernel/smp.c
+++ b/arch/arm64/kernel/smp.c
@@ -1099,7 +1099,7 @@ static bool have_cpu_die(void)
 
 bool cpus_are_stuck_in_kernel(void)
 {
-       bool smp_spin_tables = (num_possible_cpus() > 1 && !have_cpu_die());
+       bool smp_spin_tables = (num_possible_cpus_gt(1) && !have_cpu_die());
 
        return !!cpus_stuck_in_kernel || smp_spin_tables ||
                is_protected_kvm_enabled();
diff --git a/arch/arm64/mm/context.c b/arch/arm64/mm/context.c
index cd72576ae2b7..702248dc105e 100644
--- a/arch/arm64/mm/context.c
+++ b/arch/arm64/mm/context.c
@@ -384,7 +384,7 @@ static int asids_update_limit(void)
         * Expect allocation after rollover to fail if we don't have at least
         * one more ASID than CPUs. ASID #0 is reserved for init_mm.
         */
-       WARN_ON(num_available_asids - 1 <= num_possible_cpus());
+       WARN_ON(num_possible_cpus_gt(num_available_asids - 2));
        pr_info("ASID allocator initialised with %lu entries\n",
                num_available_asids);
 
diff --git a/arch/csky/mm/asid.c b/arch/csky/mm/asid.c
index b2e914745c1d..4dd6eb62a9e0 100644
--- a/arch/csky/mm/asid.c
+++ b/arch/csky/mm/asid.c
@@ -176,7 +176,7 @@ int asid_allocator_init(struct asid_info *info,
         * Expect allocation after rollover to fail if we don't have at least
         * one more ASID than CPUs. ASID #0 is always reserved.
         */
-       WARN_ON(NUM_CTXT_ASIDS(info) - 1 <= num_possible_cpus());
+       WARN_ON(num_possible_cpus_gt(NUM_CTXT_ASIDS(info) - 2));
        atomic64_set(&info->generation, ASID_FIRST_VERSION(info));
        info->map = kcalloc(BITS_TO_LONGS(NUM_CTXT_ASIDS(info)),
                            sizeof(*info->map), GFP_KERNEL);
diff --git a/arch/csky/mm/context.c b/arch/csky/mm/context.c
index 0d95bdd93846..c12312215bde 100644
--- a/arch/csky/mm/context.c
+++ b/arch/csky/mm/context.c
@@ -28,7 +28,7 @@ static void asid_flush_cpu_ctxt(void)
 
 static int asids_init(void)
 {
-       BUG_ON(((1 << CONFIG_CPU_ASID_BITS) - 1) <= num_possible_cpus());
+       BUG_ON(num_possible_cpus_gt((1 << CONFIG_CPU_ASID_BITS) - 2));
 
        if (asid_allocator_init(&asid_info, CONFIG_CPU_ASID_BITS, 1,
                                asid_flush_cpu_ctxt))
diff --git a/arch/ia64/mm/tlb.c b/arch/ia64/mm/tlb.c
index a5bce13ab047..44f623f5dc5e 100644
--- a/arch/ia64/mm/tlb.c
+++ b/arch/ia64/mm/tlb.c
@@ -202,7 +202,7 @@ setup_ptcg_sem(int max_purges, int nptcg_from)
                goto resetsema;
        }
        if (kp_override) {
-               need_ptcg_sem = num_possible_cpus() > nptcg;
+               need_ptcg_sem = num_possible_cpus_gt(nptcg);
                return;
        }
 
@@ -221,7 +221,7 @@ setup_ptcg_sem(int max_purges, int nptcg_from)
        }
        if (palo_override) {
                if (nptcg != PALO_MAX_TLB_PURGES)
-                       need_ptcg_sem = (num_possible_cpus() > nptcg);
+                       need_ptcg_sem = num_possible_cpus_gt(nptcg);
                return;
        }
 
@@ -238,7 +238,7 @@ setup_ptcg_sem(int max_purges, int nptcg_from)
                need_ptcg_sem = 0;
                return;
        } else
-               need_ptcg_sem = (num_possible_cpus() > nptcg);
+               need_ptcg_sem = num_possible_cpus_gt(nptcg);
 
 resetsema:
        spinaphore_init(&ptcg_sem, max_purges);
diff --git a/arch/mips/kernel/i8253.c b/arch/mips/kernel/i8253.c
index ca21210e06b5..89a63538be4a 100644
--- a/arch/mips/kernel/i8253.c
+++ b/arch/mips/kernel/i8253.c
@@ -29,7 +29,7 @@ void __init setup_pit_timer(void)
 
 static int __init init_pit_clocksource(void)
 {
-       if (num_possible_cpus() > 1 || /* PIT does not scale! */
+       if (num_possible_cpus_gt(1) || /* PIT does not scale! */
            !clockevent_state_periodic(&i8253_clockevent))
                return 0;
 
diff --git a/arch/mips/kernel/perf_event_mipsxx.c 
b/arch/mips/kernel/perf_event_mipsxx.c
index 1641d274fe37..4b6458899b05 100644
--- a/arch/mips/kernel/perf_event_mipsxx.c
+++ b/arch/mips/kernel/perf_event_mipsxx.c
@@ -135,7 +135,7 @@ static DEFINE_RWLOCK(pmuint_rwlock);
 /* Copied from op_model_mipsxx.c */
 static unsigned int vpe_shift(void)
 {
-       if (num_possible_cpus() > 1)
+       if (num_possible_cpus_gt(1))
                return 1;
 
        return 0;
@@ -704,7 +704,7 @@ static unsigned int mipspmu_perf_event_encode(const struct 
mips_perf_event *pev)
  * event_id.
  */
 #ifdef CONFIG_MIPS_MT_SMP
-       if (num_possible_cpus() > 1)
+       if (num_possible_cpus_gt(1))
                return ((unsigned int)pev->range << 24) |
                        (pev->cntr_mask & 0xffff00) |
                        (pev->event_id & 0xff);
diff --git a/arch/mips/kernel/rtlx-cmp.c b/arch/mips/kernel/rtlx-cmp.c
index d26dcc4b46e7..e4bb83bc46c6 100644
--- a/arch/mips/kernel/rtlx-cmp.c
+++ b/arch/mips/kernel/rtlx-cmp.c
@@ -54,7 +54,7 @@ int __init rtlx_module_init(void)
                return -ENODEV;
        }
 
-       if (num_possible_cpus() - aprp_cpu_index() < 1) {
+       if (num_possible_cpus_le(aprp_cpu_index() + 1)) {
                pr_warn("No TCs reserved for AP/SP, not initializing RTLX.\n"
                        "Pass maxcpus=<n> argument as kernel argument\n");
 
diff --git a/arch/mips/kernel/smp.c b/arch/mips/kernel/smp.c
index d542fb7af3ba..6a0bbf249528 100644
--- a/arch/mips/kernel/smp.c
+++ b/arch/mips/kernel/smp.c
@@ -248,7 +248,7 @@ int mips_smp_ipi_allocate(const struct cpumask *mask)
         * setup, if we're running with only a single CPU.
         */
        if (!ipidomain) {
-               BUG_ON(num_present_cpus() > 1);
+               BUG_ON(num_present_cpus_gt(1));
                return 0;
        }
 
@@ -314,7 +314,7 @@ int mips_smp_ipi_free(const struct cpumask *mask)
 
 static int __init mips_smp_ipi_init(void)
 {
-       if (num_possible_cpus() == 1)
+       if (num_possible_cpus_eq(1))
                return 0;
 
        mips_smp_ipi_allocate(cpu_possible_mask);
diff --git a/arch/mips/kernel/vpe-cmp.c b/arch/mips/kernel/vpe-cmp.c
index e673603e11e5..c1dc00cda666 100644
--- a/arch/mips/kernel/vpe-cmp.c
+++ b/arch/mips/kernel/vpe-cmp.c
@@ -98,7 +98,7 @@ int __init vpe_module_init(void)
                return -ENODEV;
        }
 
-       if (num_possible_cpus() - aprp_cpu_index() < 1) {
+       if (num_possible_cpus_le(aprp_cpu_index() + 1)) {
                pr_warn("No VPEs reserved for AP/SP, not initialize VPE 
loader\n"
                        "Pass maxcpus=<n> argument as kernel argument\n");
                return -ENODEV;
diff --git a/arch/mips/loongson2ef/common/cs5536/cs5536_mfgpt.c 
b/arch/mips/loongson2ef/common/cs5536/cs5536_mfgpt.c
index f21a540a1dd2..37166fa866c4 100644
--- a/arch/mips/loongson2ef/common/cs5536/cs5536_mfgpt.c
+++ b/arch/mips/loongson2ef/common/cs5536/cs5536_mfgpt.c
@@ -194,7 +194,7 @@ static struct clocksource clocksource_mfgpt = {
 
 int __init init_mfgpt_clocksource(void)
 {
-       if (num_possible_cpus() > 1)    /* MFGPT does not scale! */
+       if (num_possible_cpus_gt(1))    /* MFGPT does not scale! */
                return 0;
 
        return clocksource_register_hz(&clocksource_mfgpt, MFGPT_TICK_RATE);
diff --git a/arch/mips/mm/context.c b/arch/mips/mm/context.c
index b25564090939..bf508e38d30a 100644
--- a/arch/mips/mm/context.c
+++ b/arch/mips/mm/context.c
@@ -274,7 +274,7 @@ static int mmid_init(void)
         * one more MMID than CPUs.
         */
        num_mmids = asid_first_version(0);
-       WARN_ON(num_mmids <= num_possible_cpus());
+       WARN_ON(num_possible_cpus_gt(num_mmids - 1));
 
        atomic64_set(&mmid_version, asid_first_version(0));
        mmid_map = kcalloc(BITS_TO_LONGS(num_mmids), sizeof(*mmid_map),
diff --git a/arch/mips/mm/tlbex.c b/arch/mips/mm/tlbex.c
index bede66b072a7..92dae5cfa0a4 100644
--- a/arch/mips/mm/tlbex.c
+++ b/arch/mips/mm/tlbex.c
@@ -363,7 +363,7 @@ static struct work_registers build_get_work_registers(u32 
**p)
                return r;
        }
 
-       if (num_possible_cpus() > 1) {
+       if (num_possible_cpus_gt(1)) {
                /* Get smp_processor_id */
                UASM_i_CPUID_MFC0(p, K0, SMP_CPUID_REG);
                UASM_i_SRL_SAFE(p, K0, K0, SMP_CPUID_REGSHIFT);
diff --git a/arch/nios2/kernel/cpuinfo.c b/arch/nios2/kernel/cpuinfo.c
index 203870c4b86d..7bdc511eba60 100644
--- a/arch/nios2/kernel/cpuinfo.c
+++ b/arch/nios2/kernel/cpuinfo.c
@@ -172,7 +172,7 @@ static void *cpuinfo_start(struct seq_file *m, loff_t *pos)
 {
        unsigned long i = *pos;
 
-       return i < num_possible_cpus() ? (void *) (i + 1) : NULL;
+       return num_possible_cpus_gt(i) ? (void *) (i + 1) : NULL;
 }
 
 static void *cpuinfo_next(struct seq_file *m, void *v, loff_t *pos)
diff --git a/arch/powerpc/platforms/85xx/smp.c 
b/arch/powerpc/platforms/85xx/smp.c
index 83f4a6389a28..15573310fab4 100644
--- a/arch/powerpc/platforms/85xx/smp.c
+++ b/arch/powerpc/platforms/85xx/smp.c
@@ -280,7 +280,7 @@ static int smp_85xx_kick_cpu(int nr)
        int primary = nr;
 #endif
 
-       WARN_ON(nr < 0 || nr >= num_possible_cpus());
+       WARN_ON(nr < 0 || num_possible_cpus_le(nr + 1));
 
        pr_debug("kick CPU #%d\n", nr);
 
diff --git a/arch/powerpc/platforms/pseries/hotplug-cpu.c 
b/arch/powerpc/platforms/pseries/hotplug-cpu.c
index 5ab44600c8d3..b0d66de92309 100644
--- a/arch/powerpc/platforms/pseries/hotplug-cpu.c
+++ b/arch/powerpc/platforms/pseries/hotplug-cpu.c
@@ -365,7 +365,7 @@ static int dlpar_offline_cpu(struct device_node *dn)
                        cpu_maps_update_begin();
                        break;
                }
-               if (cpu == num_possible_cpus()) {
+               if (num_possible_cpus_eq(cpu)) {
                        pr_warn("Could not find cpu to offline with physical id 
0x%x\n",
                                thread);
                }
@@ -408,7 +408,7 @@ static int dlpar_online_cpu(struct device_node *dn)
 
                        break;
                }
-               if (cpu == num_possible_cpus())
+               if (num_possible_cpus_eq(cpu))
                        printk(KERN_WARNING "Could not find cpu to online "
                               "with physical id 0x%x\n", thread);
        }
diff --git a/arch/powerpc/sysdev/mpic.c b/arch/powerpc/sysdev/mpic.c
index 995fb2ada507..ded5007f2af9 100644
--- a/arch/powerpc/sysdev/mpic.c
+++ b/arch/powerpc/sysdev/mpic.c
@@ -1440,7 +1440,7 @@ struct mpic * __init mpic_alloc(struct device_node *node,
         * The MPIC driver will crash if there are more cores than we
         * can initialize, so we may as well catch that problem here.
         */
-       BUG_ON(num_possible_cpus() > MPIC_MAX_CPUS);
+       BUG_ON(num_possible_cpus_gt(MPIC_MAX_CPUS));
 
        /* Map the per-CPU registers */
        for_each_possible_cpu(i) {
diff --git a/arch/powerpc/xmon/xmon.c b/arch/powerpc/xmon/xmon.c
index 2073be312fe9..938346f9af7d 100644
--- a/arch/powerpc/xmon/xmon.c
+++ b/arch/powerpc/xmon/xmon.c
@@ -2747,7 +2747,7 @@ static void dump_all_pacas(void)
 {
        int cpu;
 
-       if (num_possible_cpus() == 0) {
+       if (num_possible_cpus_eq(0)) {
                printf("No possible cpus, use 'dp #' to dump individual 
cpus\n");
                return;
        }
@@ -2809,7 +2809,7 @@ static void dump_all_xives(void)
 {
        int cpu;
 
-       if (num_possible_cpus() == 0) {
+       if (num_possible_cpus_eq(0)) {
                printf("No possible cpus, use 'dx #' to dump individual 
cpus\n");
                return;
        }
@@ -3692,7 +3692,7 @@ symbol_lookup(void)
                    ptr >= (void __percpu *)__per_cpu_start &&
                    ptr < (void __percpu *)__per_cpu_end)
                {
-                       if (scanhex(&cpu) && cpu < num_possible_cpus()) {
+                       if (scanhex(&cpu) && num_possible_cpus_gt(cpu)) {
                                addr = (unsigned long)per_cpu_ptr(ptr, cpu);
                        } else {
                                cpu = raw_smp_processor_id();
diff --git a/arch/riscv/kvm/vmid.c b/arch/riscv/kvm/vmid.c
index 2c6253b293bc..6e176baedf65 100644
--- a/arch/riscv/kvm/vmid.c
+++ b/arch/riscv/kvm/vmid.c
@@ -36,7 +36,7 @@ void kvm_riscv_stage2_vmid_detect(void)
        __kvm_riscv_hfence_gvma_all();
 
        /* We don't use VMID bits if they are not sufficient */
-       if ((1UL << vmid_bits) < num_possible_cpus())
+       if (num_possible_cpus_gt(1UL << vmid_bits))
                vmid_bits = 0;
 }
 
diff --git a/arch/sparc/kernel/mdesc.c b/arch/sparc/kernel/mdesc.c
index 30f171b7b00c..b779c6607ff3 100644
--- a/arch/sparc/kernel/mdesc.c
+++ b/arch/sparc/kernel/mdesc.c
@@ -885,7 +885,7 @@ static void __mark_core_id(struct mdesc_handle *hp, u64 
node,
 {
        const u64 *id = mdesc_get_property(hp, node, "id", NULL);
 
-       if (*id < num_possible_cpus())
+       if (num_possible_cpus_gt(*id))
                cpu_data(*id).core_id = core_id;
 }
 
@@ -894,7 +894,7 @@ static void __mark_max_cache_id(struct mdesc_handle *hp, 
u64 node,
 {
        const u64 *id = mdesc_get_property(hp, node, "id", NULL);
 
-       if (*id < num_possible_cpus()) {
+       if (num_possible_cpus_gt(*id)) {
                cpu_data(*id).max_cache_id = max_cache_id;
 
                /**
@@ -986,7 +986,7 @@ static void set_sock_ids_by_socket(struct mdesc_handle *hp, 
u64 mp)
                                continue;
 
                        id = mdesc_get_property(hp, t, "id", NULL);
-                       if (*id < num_possible_cpus())
+                       if (num_possible_cpus_gt(*id))
                                cpu_data(*id).sock_id = idx;
                }
                idx++;
diff --git a/arch/x86/events/amd/core.c b/arch/x86/events/amd/core.c
index 9687a8aef01c..d69ed09a85b0 100644
--- a/arch/x86/events/amd/core.c
+++ b/arch/x86/events/amd/core.c
@@ -1007,7 +1007,7 @@ __init int amd_pmu_init(void)
        if (ret)
                return ret;
 
-       if (num_possible_cpus() == 1) {
+       if (num_possible_cpus_eq(1)) {
                /*
                 * No point in allocating data structures to serialize
                 * against other CPUs, when there is only the one CPU.
diff --git a/arch/x86/kernel/alternative.c b/arch/x86/kernel/alternative.c
index 23fb4d51a5da..55fd70fdb213 100644
--- a/arch/x86/kernel/alternative.c
+++ b/arch/x86/kernel/alternative.c
@@ -574,7 +574,7 @@ void __init_or_module alternatives_smp_module_add(struct 
module *mod,
        if (!uniproc_patched)
                goto unlock;
 
-       if (num_possible_cpus() == 1)
+       if (num_possible_cpus_eq(1))
                /* Don't bother remembering, we'll never have to undo it. */
                goto smp_unlock;
 
@@ -620,7 +620,7 @@ void alternatives_enable_smp(void)
        struct smp_alt_module *mod;
 
        /* Why bother if there are no other CPUs? */
-       BUG_ON(num_possible_cpus() == 1);
+       BUG_ON(num_possible_cpus_eq(1));
 
        mutex_lock(&text_mutex);
 
@@ -833,14 +833,14 @@ void __init alternative_instructions(void)
 
 #ifdef CONFIG_SMP
        /* Patch to UP if other cpus not imminent. */
-       if (!noreplace_smp && (num_present_cpus() == 1 || setup_max_cpus <= 1)) 
{
+       if (!noreplace_smp && (num_present_cpus_eq(1) || setup_max_cpus <= 1)) {
                uniproc_patched = true;
                alternatives_smp_module_add(NULL, "core kernel",
                                            __smp_locks, __smp_locks_end,
                                            _text, _etext);
        }
 
-       if (!uniproc_patched || num_possible_cpus() == 1) {
+       if (!uniproc_patched || num_possible_cpus_eq(1)) {
                free_init_pages("SMP alternatives",
                                (unsigned long)__smp_locks,
                                (unsigned long)__smp_locks_end);
diff --git a/arch/x86/kernel/apic/apic.c b/arch/x86/kernel/apic/apic.c
index b70344bf6600..9a3d0748ca86 100644
--- a/arch/x86/kernel/apic/apic.c
+++ b/arch/x86/kernel/apic/apic.c
@@ -1020,7 +1020,7 @@ void __init setup_boot_APIC_clock(void)
        if (disable_apic_timer) {
                pr_info("Disabling APIC timer\n");
                /* No broadcast on UP ! */
-               if (num_possible_cpus() > 1) {
+               if (num_possible_cpus_gt(1)) {
                        lapic_clockevent.mult = 1;
                        setup_APIC_timer();
                }
@@ -1029,7 +1029,7 @@ void __init setup_boot_APIC_clock(void)
 
        if (calibrate_APIC_clock()) {
                /* No broadcast on UP ! */
-               if (num_possible_cpus() > 1)
+               if (num_possible_cpus_gt(1))
                        setup_APIC_timer();
                return;
        }
diff --git a/arch/x86/kernel/apic/apic_flat_64.c 
b/arch/x86/kernel/apic/apic_flat_64.c
index 8f72b4351c9f..3dfd4c5d30dc 100644
--- a/arch/x86/kernel/apic/apic_flat_64.c
+++ b/arch/x86/kernel/apic/apic_flat_64.c
@@ -189,7 +189,7 @@ static void physflat_init_apic_ldr(void)
 
 static int physflat_probe(void)
 {
-       if (apic == &apic_physflat || num_possible_cpus() > 8 ||
+       if (apic == &apic_physflat || num_possible_cpus_gt(8) ||
            jailhouse_paravirt())
                return 1;
 
diff --git a/arch/x86/kernel/apic/probe_32.c b/arch/x86/kernel/apic/probe_32.c
index a61f642b1b90..b65c1572aaf5 100644
--- a/arch/x86/kernel/apic/probe_32.c
+++ b/arch/x86/kernel/apic/probe_32.c
@@ -138,7 +138,7 @@ void __init default_setup_apic_routing(void)
 {
        int version = boot_cpu_apic_version;
 
-       if (num_possible_cpus() > 8) {
+       if (num_possible_cpus_gt(8)) {
                switch (boot_cpu_data.x86_vendor) {
                case X86_VENDOR_INTEL:
                        if (!APIC_XAPIC(version)) {
diff --git a/arch/x86/kernel/cpu/mce/dev-mcelog.c 
b/arch/x86/kernel/cpu/mce/dev-mcelog.c
index 100fbeebdc72..34e44b0d9546 100644
--- a/arch/x86/kernel/cpu/mce/dev-mcelog.c
+++ b/arch/x86/kernel/cpu/mce/dev-mcelog.c
@@ -310,7 +310,7 @@ static ssize_t mce_chrdev_write(struct file *filp, const 
char __user *ubuf,
        if (copy_from_user(&m, ubuf, usize))
                return -EFAULT;
 
-       if (m.extcpu >= num_possible_cpus() || !cpu_online(m.extcpu))
+       if (num_possible_cpus_le(m.extcpu + 1) || !cpu_online(m.extcpu))
                return -EINVAL;
 
        /*
diff --git a/arch/x86/kernel/hpet.c b/arch/x86/kernel/hpet.c
index 882213df3713..e432e6248599 100644
--- a/arch/x86/kernel/hpet.c
+++ b/arch/x86/kernel/hpet.c
@@ -737,7 +737,7 @@ static void __init hpet_select_clockevents(void)
                hc->irq = irq;
                hc->mode = HPET_MODE_CLOCKEVT;
 
-               if (++hpet_base.nr_clockevents == num_possible_cpus())
+               if (num_possible_cpus_eq(++hpet_base.nr_clockevents))
                        break;
        }
 
diff --git a/arch/x86/kernel/i8253.c b/arch/x86/kernel/i8253.c
index 2b7999a1a50a..e6e30a7bc80f 100644
--- a/arch/x86/kernel/i8253.c
+++ b/arch/x86/kernel/i8253.c
@@ -57,7 +57,7 @@ static int __init init_pit_clocksource(void)
          * - when HPET is enabled
          * - when local APIC timer is active (PIT is switched off)
          */
-       if (num_possible_cpus() > 1 || is_hpet_enabled() ||
+       if (num_possible_cpus_gt(1) || is_hpet_enabled() ||
            !clockevent_state_periodic(&i8253_clockevent))
                return 0;
 
diff --git a/arch/x86/kernel/kvm.c b/arch/x86/kernel/kvm.c
index 59abbdad7729..375226dcf29e 100644
--- a/arch/x86/kernel/kvm.c
+++ b/arch/x86/kernel/kvm.c
@@ -1057,7 +1057,7 @@ void __init kvm_spinlock_init(void)
                goto out;
        }
 
-       if (num_possible_cpus() == 1) {
+       if (num_possible_cpus_eq(1)) {
                pr_info("PV spinlocks disabled, single CPU\n");
                goto out;
        }
diff --git a/arch/x86/kernel/kvmclock.c b/arch/x86/kernel/kvmclock.c
index 462dd8e9b03d..12c1fb1dfd07 100644
--- a/arch/x86/kernel/kvmclock.c
+++ b/arch/x86/kernel/kvmclock.c
@@ -205,7 +205,7 @@ static void __init kvmclock_init_mem(void)
        struct page *p;
        int r;
 
-       if (HVC_BOOT_ARRAY_SIZE >= num_possible_cpus())
+       if (num_possible_cpus_le(HVC_BOOT_ARRAY_SIZE + 1))
                return;
 
        ncpus = num_possible_cpus() - HVC_BOOT_ARRAY_SIZE;
diff --git a/arch/x86/kernel/tsc.c b/arch/x86/kernel/tsc.c
index 2e076a459a0c..2245c9721d4a 100644
--- a/arch/x86/kernel/tsc.c
+++ b/arch/x86/kernel/tsc.c
@@ -1223,7 +1223,7 @@ int unsynchronized_tsc(void)
         */
        if (boot_cpu_data.x86_vendor != X86_VENDOR_INTEL) {
                /* assume multi socket systems are not synchronized: */
-               if (num_possible_cpus() > 1)
+               if (num_possible_cpus_gt(1))
                        return 1;
        }
 
diff --git a/arch/x86/xen/smp_pv.c b/arch/x86/xen/smp_pv.c
index 6a8f3b53ab83..b32ca28292ae 100644
--- a/arch/x86/xen/smp_pv.c
+++ b/arch/x86/xen/smp_pv.c
@@ -254,7 +254,7 @@ static void __init xen_pv_smp_prepare_cpus(unsigned int 
max_cpus)
        cpumask_copy(xen_cpu_initialized_map, cpumask_of(0));
 
        /* Restrict the possible_map according to max_cpus. */
-       while ((num_possible_cpus() > 1) && (num_possible_cpus() > max_cpus)) {
+       while (num_possible_cpus_gt(max(1, max_cpus))) {
                for (cpu = nr_cpu_ids - 1; !cpu_possible(cpu); cpu--)
                        continue;
                set_cpu_possible(cpu, false);
diff --git a/arch/x86/xen/spinlock.c b/arch/x86/xen/spinlock.c
index 043c73dfd2c9..58caaa9aec3e 100644
--- a/arch/x86/xen/spinlock.c
+++ b/arch/x86/xen/spinlock.c
@@ -125,7 +125,7 @@ PV_CALLEE_SAVE_REGS_THUNK(xen_vcpu_stolen);
 void __init xen_init_spinlocks(void)
 {
        /*  Don't need to use pvqspinlock code if there is only 1 vCPU. */
-       if (num_possible_cpus() == 1 || nopvspin)
+       if (num_possible_cpus_eq(1) || nopvspin)
                xen_pvspin = false;
 
        if (!xen_pvspin) {
diff --git a/drivers/clk/samsung/clk-exynos4.c 
b/drivers/clk/samsung/clk-exynos4.c
index 22009cb53428..64d7de6b885c 100644
--- a/drivers/clk/samsung/clk-exynos4.c
+++ b/drivers/clk/samsung/clk-exynos4.c
@@ -1178,7 +1178,7 @@ static void __init exynos4x12_core_down_clock(void)
                PWR_CTRL1_USE_CORE1_WFE | PWR_CTRL1_USE_CORE0_WFE |
                PWR_CTRL1_USE_CORE1_WFI | PWR_CTRL1_USE_CORE0_WFI);
        /* On Exynos4412 enable it also on core 2 and 3 */
-       if (num_possible_cpus() == 4)
+       if (num_possible_cpus_eq(4))
                tmp |= PWR_CTRL1_USE_CORE3_WFE | PWR_CTRL1_USE_CORE2_WFE |
                       PWR_CTRL1_USE_CORE3_WFI | PWR_CTRL1_USE_CORE2_WFI;
        writel_relaxed(tmp, reg_base + PWR_CTRL1);
diff --git a/drivers/clocksource/ingenic-timer.c 
b/drivers/clocksource/ingenic-timer.c
index 24ed0f1f089b..c4a34d26357c 100644
--- a/drivers/clocksource/ingenic-timer.c
+++ b/drivers/clocksource/ingenic-timer.c
@@ -302,8 +302,7 @@ static int __init ingenic_tcu_init(struct device_node *np)
                             (u32 *)&tcu->pwm_channels_mask);
 
        /* Verify that we have at least num_possible_cpus() + 1 free channels */
-       if (hweight8(tcu->pwm_channels_mask) >
-                       soc_info->num_channels - num_possible_cpus() + 1) {
+       if (num_possible_cpus_gt(soc_info->num_channels + 1 - 
hweight8(tcu->pwm_channels_mask))) {
                pr_crit("%s: Invalid PWM channel mask: 0x%02lx\n", __func__,
                        tcu->pwm_channels_mask);
                ret = -EINVAL;
diff --git a/drivers/cpufreq/pcc-cpufreq.c b/drivers/cpufreq/pcc-cpufreq.c
index 9f3fc7a073d0..8bf76eaa9e1e 100644
--- a/drivers/cpufreq/pcc-cpufreq.c
+++ b/drivers/cpufreq/pcc-cpufreq.c
@@ -593,7 +593,7 @@ static int __init pcc_cpufreq_init(void)
                return ret;
        }
 
-       if (num_present_cpus() > 4) {
+       if (num_present_cpus_gt(4)) {
                pcc_cpufreq_driver.flags |= CPUFREQ_NO_AUTO_DYNAMIC_SWITCHING;
                pr_err("%s: Too many CPUs, dynamic performance scaling 
disabled\n",
                       __func__);
diff --git a/drivers/dma/mv_xor.c b/drivers/dma/mv_xor.c
index 23b232b57518..f99177e72158 100644
--- a/drivers/dma/mv_xor.c
+++ b/drivers/dma/mv_xor.c
@@ -1293,7 +1293,7 @@ static int mv_xor_probe(struct platform_device *pdev)
        struct mv_xor_device *xordev;
        struct mv_xor_platform_data *pdata = dev_get_platdata(&pdev->dev);
        struct resource *res;
-       unsigned int max_engines, max_channels;
+       unsigned int max_channels;
        int i, ret;
 
        dev_notice(&pdev->dev, "Marvell shared XOR driver\n");
@@ -1362,7 +1362,6 @@ static int mv_xor_probe(struct platform_device *pdev)
         * separate engines when possible.  For dual-CPU Armada 3700
         * SoC with single XOR engine allow using its both channels.
         */
-       max_engines = num_present_cpus();
        if (xordev->xor_type == XOR_ARMADA_37XX)
                max_channels =  num_present_cpus();
        else
@@ -1370,7 +1369,7 @@ static int mv_xor_probe(struct platform_device *pdev)
                                     MV_XOR_MAX_CHANNELS,
                                     DIV_ROUND_UP(num_present_cpus(), 2));
 
-       if (mv_xor_engine_count >= max_engines)
+       if (num_present_cpus_le(mv_xor_engine_count + 1))
                return 0;
 
        if (pdev->dev.of_node) {
diff --git a/drivers/gpu/drm/i810/i810_drv.c b/drivers/gpu/drm/i810/i810_drv.c
index 0e53a066d4db..c70745fa4166 100644
--- a/drivers/gpu/drm/i810/i810_drv.c
+++ b/drivers/gpu/drm/i810/i810_drv.c
@@ -80,7 +80,7 @@ static struct pci_driver i810_pci_driver = {
 
 static int __init i810_init(void)
 {
-       if (num_possible_cpus() > 1) {
+       if (num_possible_cpus_gt(1)) {
                pr_err("drm/i810 does not support SMP\n");
                return -EINVAL;
        }
diff --git a/drivers/irqchip/irq-gic.c b/drivers/irqchip/irq-gic.c
index b8bb46c65a97..4e319e4ba9dc 100644
--- a/drivers/irqchip/irq-gic.c
+++ b/drivers/irqchip/irq-gic.c
@@ -430,7 +430,7 @@ static u8 gic_get_cpumask(struct gic_chip_data *gic)
                        break;
        }
 
-       if (!mask && num_possible_cpus() > 1)
+       if (!mask && num_possible_cpus_gt(1))
                pr_crit("GIC CPU mask not found - kernel will fail to boot.\n");
 
        return mask;
diff --git a/drivers/net/caif/caif_virtio.c b/drivers/net/caif/caif_virtio.c
index 91230894692d..c7aa3f6dc635 100644
--- a/drivers/net/caif/caif_virtio.c
+++ b/drivers/net/caif/caif_virtio.c
@@ -537,7 +537,7 @@ static netdev_tx_t cfv_netdev_tx(struct sk_buff *skb, 
struct net_device *netdev)
         *
         * Flow-on is triggered when sufficient buffers are freed
         */
-       if (unlikely(cfv->vq_tx->num_free <= num_present_cpus())) {
+       if (unlikely(num_present_cpus_gt(cfv->vq_tx->num_free - 1))) {
                flow_off = true;
                cfv->stats.tx_full_ring++;
        }
diff --git a/drivers/net/ethernet/cavium/liquidio/cn23xx_vf_device.c 
b/drivers/net/ethernet/cavium/liquidio/cn23xx_vf_device.c
index fda49404968c..79d5ded30b65 100644
--- a/drivers/net/ethernet/cavium/liquidio/cn23xx_vf_device.c
+++ b/drivers/net/ethernet/cavium/liquidio/cn23xx_vf_device.c
@@ -649,7 +649,7 @@ int cn23xx_setup_octeon_vf_device(struct octeon_device *oct)
                         rings_per_vf);
                oct->sriov_info.rings_per_vf = rings_per_vf;
        } else {
-               if (rings_per_vf > num_present_cpus()) {
+               if (num_present_cpus_le(rings_per_vf)) {
                        dev_warn(&oct->pci_dev->dev,
                                 "PF configured rings_per_vf:%d greater than 
num_cpu:%d. Using rings_per_vf:%d equal to num cpus\n",
                                 rings_per_vf,
diff --git a/drivers/net/ethernet/hisilicon/hns/hns_enet.c 
b/drivers/net/ethernet/hisilicon/hns/hns_enet.c
index 22a463e15678..7d97939413d2 100644
--- a/drivers/net/ethernet/hisilicon/hns/hns_enet.c
+++ b/drivers/net/ethernet/hisilicon/hns/hns_enet.c
@@ -1239,7 +1239,7 @@ static int hns_nic_init_affinity_mask(int q_num, int 
ring_idx,
         * The cpu mask set by ring index according to the ring flag
         * which indicate the ring is tx or rx.
         */
-       if (q_num == num_possible_cpus()) {
+       if (num_possible_cpus_eq(q_num)) {
                if (is_tx_ring(ring))
                        cpu = ring_idx;
                else
diff --git a/drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 
b/drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
index a48e804c46f2..34ad59fd51d6 100644
--- a/drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
+++ b/drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
@@ -3315,7 +3315,7 @@ static int mvpp2_setup_txqs(struct mvpp2_port *port)
                        goto err_cleanup;
 
                /* Assign this queue to a CPU */
-               if (queue < num_possible_cpus())
+               if (num_possible_cpus_gt(queue))
                        netif_set_xps_queue(port->dev, cpumask_of(queue), 
queue);
        }
 
diff --git a/drivers/net/wireless/ath/ath9k/hw.c 
b/drivers/net/wireless/ath/ath9k/hw.c
index 172081ffe477..33d3cddc6c7b 100644
--- a/drivers/net/wireless/ath/ath9k/hw.c
+++ b/drivers/net/wireless/ath/ath9k/hw.c
@@ -429,7 +429,7 @@ static void ath9k_hw_init_config(struct ath_hw *ah)
         * This issue is not present on PCI-Express devices or pre-AR5416
         * devices (legacy, 802.11abg).
         */
-       if (num_possible_cpus() > 1)
+       if (num_possible_cpus_gt(1))
                ah->config.serialize_regmode = SER_REG_MODE_AUTO;
 
        if (NR_CPUS > 1 && ah->config.serialize_regmode == SER_REG_MODE_AUTO) {
diff --git a/drivers/net/wireless/marvell/mwifiex/main.c 
b/drivers/net/wireless/marvell/mwifiex/main.c
index 19b996c6a260..6ce0236a3203 100644
--- a/drivers/net/wireless/marvell/mwifiex/main.c
+++ b/drivers/net/wireless/marvell/mwifiex/main.c
@@ -1536,7 +1536,7 @@ mwifiex_reinit_sw(struct mwifiex_adapter *adapter)
        adapter->cmd_wait_q.status = 0;
        adapter->scan_wait_q_woken = false;
 
-       if ((num_possible_cpus() > 1) || adapter->iface_type == MWIFIEX_USB)
+       if (num_possible_cpus_gt(1) || adapter->iface_type == MWIFIEX_USB)
                adapter->rx_work_enabled = true;
 
        adapter->workqueue =
@@ -1691,7 +1691,7 @@ mwifiex_add_card(void *card, struct completion *fw_done,
        adapter->cmd_wait_q.status = 0;
        adapter->scan_wait_q_woken = false;
 
-       if ((num_possible_cpus() > 1) || adapter->iface_type == MWIFIEX_USB)
+       if (num_possible_cpus_gt(1) || adapter->iface_type == MWIFIEX_USB)
                adapter->rx_work_enabled = true;
 
        adapter->workqueue =
diff --git a/drivers/net/wireless/st/cw1200/queue.c 
b/drivers/net/wireless/st/cw1200/queue.c
index 12952b1c29df..4d47a1e26d55 100644
--- a/drivers/net/wireless/st/cw1200/queue.c
+++ b/drivers/net/wireless/st/cw1200/queue.c
@@ -312,8 +312,7 @@ int cw1200_queue_put(struct cw1200_queue *queue,
                 * Leave extra queue slots so we don't overflow.
                 */
                if (queue->overfull == false &&
-                   queue->num_queued >=
-                   (queue->capacity - (num_present_cpus() - 1))) {
+                   num_present_cpus_gt(queue->capacity - queue->num_queued)) {
                        queue->overfull = true;
                        __cw1200_queue_lock(queue);
                        mod_timer(&queue->gc, jiffies);
diff --git a/drivers/nvdimm/region.c b/drivers/nvdimm/region.c
index e0c34120df37..474f1ed5d9b9 100644
--- a/drivers/nvdimm/region.c
+++ b/drivers/nvdimm/region.c
@@ -17,7 +17,7 @@ static int nd_region_probe(struct device *dev)
        struct nd_region *nd_region = to_nd_region(dev);
 
        if (nd_region->num_lanes > num_online_cpus()
-                       && nd_region->num_lanes < num_possible_cpus()
+                       && num_possible_cpus_gt(nd_region->num_lanes)
                        && !test_and_set_bit(0, &once)) {
                dev_dbg(dev, "online cpus (%d) < concurrent i/o lanes (%d) < 
possible cpus (%d)\n",
                                num_online_cpus(), nd_region->num_lanes,
diff --git a/drivers/nvme/host/pci.c b/drivers/nvme/host/pci.c
index ca2ee806d74b..34958f775ad8 100644
--- a/drivers/nvme/host/pci.c
+++ b/drivers/nvme/host/pci.c
@@ -79,7 +79,7 @@ static int io_queue_count_set(const char *val, const struct 
kernel_param *kp)
        int ret;
 
        ret = kstrtouint(val, 10, &n);
-       if (ret != 0 || n > num_possible_cpus())
+       if (ret != 0 || num_possible_cpus_le(n))
                return -EINVAL;
        return param_set_uint(val, kp);
 }
diff --git a/drivers/perf/arm_pmu.c b/drivers/perf/arm_pmu.c
index a31b302b0ade..5f43a7bde55d 100644
--- a/drivers/perf/arm_pmu.c
+++ b/drivers/perf/arm_pmu.c
@@ -637,7 +637,7 @@ int armpmu_request_irq(int irq, int cpu)
 
                err = irq_force_affinity(irq, cpumask_of(cpu));
 
-               if (err && num_possible_cpus() > 1) {
+               if (err && num_possible_cpus_gt(1)) {
                        pr_warn("unable to set irq affinity (irq=%d, cpu=%u)\n",
                                irq, cpu);
                        goto err_out;
diff --git a/drivers/platform/x86/intel/speed_select_if/isst_if_common.c 
b/drivers/platform/x86/intel/speed_select_if/isst_if_common.c
index c9a85eb2e860..c25902969475 100644
--- a/drivers/platform/x86/intel/speed_select_if/isst_if_common.c
+++ b/drivers/platform/x86/intel/speed_select_if/isst_if_common.c
@@ -297,7 +297,7 @@ static struct pci_dev *_isst_if_get_pci_dev(int cpu, int 
bus_no, int dev, int fn
        int i, bus_number;
 
        if (bus_no < 0 || bus_no > 1 || cpu < 0 || cpu >= nr_cpu_ids ||
-           cpu >= num_possible_cpus())
+           num_possible_cpus_le(cpu + 1))
                return NULL;
 
        bus_number = isst_cpu_info[cpu].bus_info[bus_no];
@@ -362,7 +362,7 @@ struct pci_dev *isst_if_get_pci_dev(int cpu, int bus_no, 
int dev, int fn)
        struct pci_dev *pci_dev;
 
        if (bus_no < 0 || bus_no > 1 || cpu < 0 || cpu >= nr_cpu_ids ||
-           cpu >= num_possible_cpus())
+           num_possible_cpus_le(cpu + 1))
                return NULL;
 
        pci_dev = isst_cpu_info[cpu].pci_dev[bus_no];
@@ -442,7 +442,7 @@ static long isst_if_proc_phyid_req(u8 *cmd_ptr, int 
*write_only, int resume)
 
        cpu_map = (struct isst_if_cpu_map *)cmd_ptr;
        if (cpu_map->logical_cpu >= nr_cpu_ids ||
-           cpu_map->logical_cpu >= num_possible_cpus())
+           num_possible_cpus_le(cpu_map->logical_cpu + 1))
                return -EINVAL;
 
        *write_only = 0;
diff --git a/drivers/soc/bcm/brcmstb/biuctrl.c 
b/drivers/soc/bcm/brcmstb/biuctrl.c
index 2c975d79fe8e..6a75cbe836a4 100644
--- a/drivers/soc/bcm/brcmstb/biuctrl.c
+++ b/drivers/soc/bcm/brcmstb/biuctrl.c
@@ -181,7 +181,7 @@ static void __init a72_b53_rac_enable_all(struct 
device_node *np)
        if (IS_ENABLED(CONFIG_CACHE_B15_RAC))
                return;
 
-       if (WARN(num_possible_cpus() > 4, "RAC only supports 4 CPUs\n"))
+       if (WARN(num_possible_cpus_gt(4), "RAC only supports 4 CPUs\n"))
                return;
 
        pref_dist = cbc_readl(RAC_CONFIG1_REG);
diff --git a/drivers/soc/fsl/dpio/dpio-service.c 
b/drivers/soc/fsl/dpio/dpio-service.c
index 1d2b27e3ea63..b38c519f2294 100644
--- a/drivers/soc/fsl/dpio/dpio-service.c
+++ b/drivers/soc/fsl/dpio/dpio-service.c
@@ -60,7 +60,7 @@ static inline struct dpaa2_io *service_select_by_cpu(struct 
dpaa2_io *d,
        if (d)
                return d;
 
-       if (cpu != DPAA2_IO_ANY_CPU && cpu >= num_possible_cpus())
+       if (cpu != DPAA2_IO_ANY_CPU && num_possible_cpus_le(cpu + 1))
                return NULL;
 
        /*
@@ -140,7 +140,7 @@ struct dpaa2_io *dpaa2_io_create(const struct dpaa2_io_desc 
*desc,
                return NULL;
 
        /* check if CPU is out of range (-1 means any cpu) */
-       if (desc->cpu != DPAA2_IO_ANY_CPU && desc->cpu >= num_possible_cpus()) {
+       if (desc->cpu != DPAA2_IO_ANY_CPU && num_possible_cpus_le(desc->cpu + 
1)) {
                kfree(obj);
                return NULL;
        }
diff --git a/drivers/spi/spi-dw-bt1.c b/drivers/spi/spi-dw-bt1.c
index c06553416123..ab6b6a32a0d6 100644
--- a/drivers/spi/spi-dw-bt1.c
+++ b/drivers/spi/spi-dw-bt1.c
@@ -241,7 +241,7 @@ static int dw_spi_bt1_sys_init(struct platform_device *pdev,
         * though, but still tends to be not fast enough at low CPU
         * frequencies.
         */
-       if (num_possible_cpus() > 1)
+       if (num_possible_cpus_gt(1))
                dws->max_mem_freq = 10000000U;
        else
                dws->max_mem_freq = 20000000U;
diff --git a/drivers/virt/acrn/hsm.c b/drivers/virt/acrn/hsm.c
index 5419794fccf1..50cd69012dcf 100644
--- a/drivers/virt/acrn/hsm.c
+++ b/drivers/virt/acrn/hsm.c
@@ -431,7 +431,7 @@ static ssize_t remove_cpu_store(struct device *dev,
        if (kstrtoull(buf, 0, &cpu) < 0)
                return -EINVAL;
 
-       if (cpu >= num_possible_cpus() || cpu == 0 || !cpu_is_hotpluggable(cpu))
+       if (num_possible_cpus_le(cpu + 1) || cpu == 0 || 
!cpu_is_hotpluggable(cpu))
                return -EINVAL;
 
        if (cpu_online(cpu))
diff --git a/fs/xfs/xfs_sysfs.c b/fs/xfs/xfs_sysfs.c
index 8608f804388f..5580d60ec962 100644
--- a/fs/xfs/xfs_sysfs.c
+++ b/fs/xfs/xfs_sysfs.c
@@ -211,7 +211,7 @@ pwork_threads_store(
        if (ret)
                return ret;
 
-       if (val < -1 || val > num_possible_cpus())
+       if (val < -1 || num_possible_cpus_le(val))
                return -EINVAL;
 
        xfs_globals.pwork_threads = val;
diff --git a/include/linux/cpumask.h b/include/linux/cpumask.h
index b5e50cf74785..ea0699fa4d4c 100644
--- a/include/linux/cpumask.h
+++ b/include/linux/cpumask.h
@@ -945,8 +945,19 @@ static inline unsigned int num_online_cpus(void)
        return atomic_read(&__num_online_cpus);
 }
 #define num_possible_cpus()    cpumask_weight(cpu_possible_mask)
+#define num_possible_cpus_eq(n)        cpumask_weight_eq(cpu_possible_mask, 
(n))
+#define num_possible_cpus_gt(n)        cpumask_weight_gt(cpu_possible_mask, 
(n))
+#define num_possible_cpus_le(n)        cpumask_weight_le(cpu_possible_mask, 
(n))
+
 #define num_present_cpus()     cpumask_weight(cpu_present_mask)
+#define num_present_cpus_eq(n) cpumask_weight_eq(cpu_present_mask, (n))
+#define num_present_cpus_gt(n) cpumask_weight_gt(cpu_present_mask, (n))
+#define num_present_cpus_le(n) cpumask_weight_le(cpu_present_mask, (n))
+
 #define num_active_cpus()      cpumask_weight(cpu_active_mask)
+#define num_active_cpus_eq(n)  cpumask_weight_eq(cpu_active_mask, (n))
+#define num_active_cpus_gt(n)  cpumask_weight_gt(cpu_active_mask, (n))
+#define num_active_cpus_le(n)  cpumask_weight_le(cpu_active_mask, (n))
 
 static inline bool cpu_online(unsigned int cpu)
 {
@@ -976,9 +987,21 @@ static inline bool cpu_dying(unsigned int cpu)
 #else
 
 #define num_online_cpus()      1U
+
 #define num_possible_cpus()    1U
+#define num_possible_cpus_eq(n)        (1U == (n))
+#define num_possible_cpus_gt(n)        (1U > (n))
+#define num_possible_cpus_le(n)        (1U < (n))
+
 #define num_present_cpus()     1U
+#define num_present_cpus_eq(n) (1U == (n))
+#define num_present_cpus_gt(n) (1U > (n))
+#define num_present_cpus_le(n) (1U < (n))
+
 #define num_active_cpus()      1U
+#define num_active_cpus_eq(n)  (1U == (n))
+#define num_active_cpus_gt(n)  (1U > (n))
+#define num_active_cpus_le(n)  (1U < (n))
 
 static inline bool cpu_online(unsigned int cpu)
 {
diff --git a/include/linux/kdb.h b/include/linux/kdb.h
index ea0f5e580fac..48269d32b038 100644
--- a/include/linux/kdb.h
+++ b/include/linux/kdb.h
@@ -191,7 +191,7 @@ static inline
 int kdb_process_cpu(const struct task_struct *p)
 {
        unsigned int cpu = task_cpu(p);
-       if (cpu > num_possible_cpus())
+       if (num_possible_cpus_le(cpu))
                cpu = 0;
        return cpu;
 }
diff --git a/kernel/debug/kdb/kdb_bt.c b/kernel/debug/kdb/kdb_bt.c
index 10b454554ab0..b6435a41a537 100644
--- a/kernel/debug/kdb/kdb_bt.c
+++ b/kernel/debug/kdb/kdb_bt.c
@@ -108,7 +108,7 @@ kdb_bt_cpu(unsigned long cpu)
 {
        struct task_struct *kdb_tsk;
 
-       if (cpu >= num_possible_cpus() || !cpu_online(cpu)) {
+       if (num_possible_cpus_le(cpu + 1) || !cpu_online(cpu)) {
                kdb_printf("WARNING: no process for cpu %ld\n", cpu);
                return;
        }
diff --git a/kernel/printk/printk.c b/kernel/printk/printk.c
index cbc35d586afb..08e6df52eb4d 100644
--- a/kernel/printk/printk.c
+++ b/kernel/printk/printk.c
@@ -1005,7 +1005,7 @@ static void __init log_buf_add_cpu(void)
         * set_cpu_possible() after setup_arch() but just in
         * case lets ensure this is valid.
         */
-       if (num_possible_cpus() == 1)
+       if (num_possible_cpus_eq(1))
                return;
 
        cpu_extra = (num_possible_cpus() - 1) * __LOG_CPU_MAX_BUF_LEN;
diff --git a/kernel/reboot.c b/kernel/reboot.c
index 6bcc5d6a6572..f21c2c20505d 100644
--- a/kernel/reboot.c
+++ b/kernel/reboot.c
@@ -639,7 +639,7 @@ static int __init reboot_setup(char *str)
                        if (isdigit(str[0])) {
                                int cpu = simple_strtoul(str, NULL, 0);
 
-                               if (cpu >= num_possible_cpus()) {
+                               if (num_possible_cpus_le(cpu + 1)) {
                                        pr_err("Ignoring the CPU number in 
reboot= option. "
                                        "CPU %d exceeds possible cpu number 
%d\n",
                                        cpu, num_possible_cpus());
@@ -844,7 +844,7 @@ static ssize_t cpu_store(struct kobject *kobj, struct 
kobj_attribute *attr,
        if (rc)
                return rc;
 
-       if (cpunum >= num_possible_cpus())
+       if (num_possible_cpus_le(cpunum + 1))
                return -ERANGE;
 
        reboot_default = 0;
diff --git a/kernel/time/clockevents.c b/kernel/time/clockevents.c
index 32d6629a55b2..c1fdfa4084c3 100644
--- a/kernel/time/clockevents.c
+++ b/kernel/time/clockevents.c
@@ -448,7 +448,7 @@ void clockevents_register_device(struct clock_event_device 
*dev)
        clockevent_set_state(dev, CLOCK_EVT_STATE_DETACHED);
 
        if (!dev->cpumask) {
-               WARN_ON(num_possible_cpus() > 1);
+               WARN_ON(num_possible_cpus_gt(1));
                dev->cpumask = cpumask_of(smp_processor_id());
        }
 
diff --git a/mm/percpu.c b/mm/percpu.c
index 293009cc03ef..76e846b3d48e 100644
--- a/mm/percpu.c
+++ b/mm/percpu.c
@@ -2936,7 +2936,7 @@ static struct pcpu_alloc_info * __init __flatten 
pcpu_build_alloc_info(
                 * greater-than comparison ensures upa==1 always
                 * passes the following check.
                 */
-               if (wasted > num_possible_cpus() / 3)
+               if (num_possible_cpus_le(wasted * 3))
                        continue;
 
                /* and then don't consume more memory */
@@ -3193,7 +3193,7 @@ int __init pcpu_page_first_chunk(size_t reserved_size,
 
        /* allocate pages */
        j = 0;
-       for (unit = 0; unit < num_possible_cpus(); unit++) {
+       for (unit = 0; num_possible_cpus_gt(unit); unit++) {
                unsigned int cpu = ai->groups[0].cpu_map[unit];
                for (i = 0; i < unit_pages; i++) {
                        void *ptr;
@@ -3215,7 +3215,7 @@ int __init pcpu_page_first_chunk(size_t reserved_size,
        vm.size = num_possible_cpus() * ai->unit_size;
        vm_area_register_early(&vm, PAGE_SIZE);
 
-       for (unit = 0; unit < num_possible_cpus(); unit++) {
+       for (unit = 0; num_possible_cpus_gt(unit); unit++) {
                unsigned long unit_addr =
                        (unsigned long)vm.addr + unit * ai->unit_size;
 
diff --git a/mm/slab.c b/mm/slab.c
index ca4822f6b2b6..3cf2ee629b23 100644
--- a/mm/slab.c
+++ b/mm/slab.c
@@ -3929,7 +3929,7 @@ static int enable_cpucache(struct kmem_cache *cachep, 
gfp_t gfp)
         * to a larger limit. Thus disabled by default.
         */
        shared = 0;
-       if (cachep->size <= PAGE_SIZE && num_possible_cpus() > 1)
+       if (cachep->size <= PAGE_SIZE && num_possible_cpus_gt(1))
                shared = 8;
 
 #if DEBUG
-- 
2.25.1

Reply via email to