Use the byte-order aware big endian accessors, allowing for kernels
running under big-endian.

Signed-off-by: Matthew Leach <matt...@mattleach.net>
---
CC: Thomas Gleixner <t...@linutronix.de>
CC: Jason Cooper <ja...@lakedaemon.net>
CC: Marc Zyngier <marc.zyng...@arm.com>
CC: Kukjin Kim <kg...@kernel.org>
CC: Krzysztof Kozlowski <k.kozlow...@samsung.com>
CC: linux-kernel@vger.kernel.org
CC: linux-arm-ker...@lists.infradead.org
CC: linux-samsung-...@vger.kernel.org
---
 drivers/irqchip/exynos-combiner.c | 14 +++++++-------
 1 file changed, 7 insertions(+), 7 deletions(-)

diff --git a/drivers/irqchip/exynos-combiner.c 
b/drivers/irqchip/exynos-combiner.c
index ead15be..b78a169 100644
--- a/drivers/irqchip/exynos-combiner.c
+++ b/drivers/irqchip/exynos-combiner.c
@@ -55,14 +55,14 @@ static void combiner_mask_irq(struct irq_data *data)
 {
        u32 mask = 1 << (data->hwirq % 32);
 
-       __raw_writel(mask, combiner_base(data) + COMBINER_ENABLE_CLEAR);
+       writel_relaxed(mask, combiner_base(data) + COMBINER_ENABLE_CLEAR);
 }
 
 static void combiner_unmask_irq(struct irq_data *data)
 {
        u32 mask = 1 << (data->hwirq % 32);
 
-       __raw_writel(mask, combiner_base(data) + COMBINER_ENABLE_SET);
+       writel_relaxed(mask, combiner_base(data) + COMBINER_ENABLE_SET);
 }
 
 static void combiner_handle_cascade_irq(struct irq_desc *desc)
@@ -75,7 +75,7 @@ static void combiner_handle_cascade_irq(struct irq_desc *desc)
        chained_irq_enter(chip, desc);
 
        spin_lock(&irq_controller_lock);
-       status = __raw_readl(chip_data->base + COMBINER_INT_STATUS);
+       status = readl_relaxed(chip_data->base + COMBINER_INT_STATUS);
        spin_unlock(&irq_controller_lock);
        status &= chip_data->irq_mask;
 
@@ -135,7 +135,7 @@ static void __init combiner_init_one(struct 
combiner_chip_data *combiner_data,
        combiner_data->parent_irq = irq;
 
        /* Disable all interrupts */
-       __raw_writel(combiner_data->irq_mask, base + COMBINER_ENABLE_CLEAR);
+       writel_relaxed(combiner_data->irq_mask, base + COMBINER_ENABLE_CLEAR);
 }
 
 static int combiner_irq_domain_xlate(struct irq_domain *d,
@@ -218,7 +218,7 @@ static int combiner_suspend(void)
 
        for (i = 0; i < max_nr; i++)
                combiner_data[i].pm_save =
-                       __raw_readl(combiner_data[i].base + 
COMBINER_ENABLE_SET);
+                       readl_relaxed(combiner_data[i].base + 
COMBINER_ENABLE_SET);
 
        return 0;
 }
@@ -235,9 +235,9 @@ static void combiner_resume(void)
        int i;
 
        for (i = 0; i < max_nr; i++) {
-               __raw_writel(combiner_data[i].irq_mask,
+               writel_relaxed(combiner_data[i].irq_mask,
                             combiner_data[i].base + COMBINER_ENABLE_CLEAR);
-               __raw_writel(combiner_data[i].pm_save,
+               writel_relaxed(combiner_data[i].pm_save,
                             combiner_data[i].base + COMBINER_ENABLE_SET);
        }
 }
-- 
2.8.3

Reply via email to