__ioremap_caller() do the same thing. Define a common one.

__ioremap() is not reused because most of the tests included in
it are unnecessary when coming from __ioremap_caller()

Signed-off-by: Christophe Leroy <christophe.le...@c-s.fr>
---
 arch/powerpc/mm/ioremap.c    | 99 ++++++++++++++++++++++++++++++++++++++++++++
 arch/powerpc/mm/pgtable_32.c | 75 ---------------------------------
 arch/powerpc/mm/pgtable_64.c | 61 ---------------------------
 3 files changed, 99 insertions(+), 136 deletions(-)

diff --git a/arch/powerpc/mm/ioremap.c b/arch/powerpc/mm/ioremap.c
index 889ee656cf64..537c9148cea1 100644
--- a/arch/powerpc/mm/ioremap.c
+++ b/arch/powerpc/mm/ioremap.c
@@ -76,6 +76,105 @@ void __iomem *ioremap_prot(phys_addr_t addr, unsigned long 
size, unsigned long f
 }
 EXPORT_SYMBOL(ioremap_prot);
 
+int __weak ioremap_range(unsigned long ea, phys_addr_t pa, unsigned long size,
+                        pgprot_t prot, int nid)
+{
+       unsigned long i;
+
+       for (i = 0; i < size; i += PAGE_SIZE) {
+               int err = map_kernel_page(ea + i, pa + i, prot);
+
+               if (err) {
+                       if (slab_is_available())
+                               unmap_kernel_range(ea, size);
+                       else
+                               WARN_ON_ONCE(1); /* Should clean up */
+                       return err;
+               }
+       }
+
+       return 0;
+}
+
+void __iomem *__ioremap_caller(phys_addr_t addr, unsigned long size,
+                              pgprot_t prot, void *caller)
+{
+       phys_addr_t pa = addr & PAGE_MASK;
+       int ret;
+       unsigned long va;
+
+       size = PAGE_ALIGN(addr + size) - pa;
+
+#ifdef CONFIG_PPC64
+       /* We don't support the 4K PFN hack with ioremap */
+       if (pgprot_val(prot) & H_PAGE_4K_PFN)
+               return NULL;
+#else
+       /*
+        * If the address lies within the first 16 MB, assume it's in ISA
+        * memory space
+        */
+       if (pa < SZ_16M)
+               pa += _ISA_MEM_BASE;
+
+#ifndef CONFIG_CRASH_DUMP
+       /*
+        * Don't allow anybody to remap normal RAM that we're using.
+        * mem_init() sets high_memory so only do the check after that.
+        */
+       if (slab_is_available() && pa <= virt_to_phys(high_memory - 1) &&
+           page_is_ram(__phys_to_pfn(pa))) {
+               pr_err("%s(): phys addr 0x%llx is RAM lr %ps\n", __func__,
+                      (unsigned long long)pa, __builtin_return_address(0));
+               return NULL;
+       }
+#endif
+#endif /* CONFIG_PPC64 */
+
+       if (size == 0 || pa == 0)
+               return NULL;
+
+       /*
+        * Is it already mapped?  Perhaps overlapped by a previous
+        * mapping.
+        */
+       va = p_block_mapped(pa);
+       if (va)
+               return (void __iomem *)va + (addr & ~PAGE_MASK);
+
+       /*
+        * Choose an address to map it to.
+        * Once the vmalloc system is running, we use it.
+        * Before that, we map using addresses going
+        * down from ioremap_bot.  vmalloc will use
+        * the addresses from IOREMAP_BASE through
+        * ioremap_bot
+        *
+        */
+       if (slab_is_available()) {
+               struct vm_struct *area;
+
+               area = __get_vm_area_caller(size, VM_IOREMAP, IOREMAP_BASE,
+                                           ioremap_bot, caller);
+               if (area == NULL)
+                       return NULL;
+
+               area->phys_addr = pa;
+               va = (unsigned long)area->addr;
+       } else {
+               ioremap_bot -= size;
+               va = ioremap_bot;
+       }
+       ret = ioremap_range(va, pa, size, prot, NUMA_NO_NODE);
+       if (!ret)
+               return (void __iomem *)va + (addr & ~PAGE_MASK);
+
+       if (!slab_is_available())
+               ioremap_bot += size;
+
+       return NULL;
+}
+
 /*
  * Unmap an IO region and remove it from vmalloc'd list.
  * Access to IO memory should be serialized by driver.
diff --git a/arch/powerpc/mm/pgtable_32.c b/arch/powerpc/mm/pgtable_32.c
index 4597f45e4dc6..bacf3b85191c 100644
--- a/arch/powerpc/mm/pgtable_32.c
+++ b/arch/powerpc/mm/pgtable_32.c
@@ -35,81 +35,6 @@
 
 extern char etext[], _stext[], _sinittext[], _einittext[];
 
-void __iomem *
-__ioremap_caller(phys_addr_t addr, unsigned long size, pgprot_t prot, void 
*caller)
-{
-       unsigned long v, i;
-       phys_addr_t p;
-       int err;
-
-       /*
-        * Choose an address to map it to.
-        * Once the vmalloc system is running, we use it.
-        * Before then, we use space going down from IOREMAP_TOP
-        * (ioremap_bot records where we're up to).
-        */
-       p = addr & PAGE_MASK;
-       size = PAGE_ALIGN(addr + size) - p;
-
-       /*
-        * If the address lies within the first 16 MB, assume it's in ISA
-        * memory space
-        */
-       if (p < 16*1024*1024)
-               p += _ISA_MEM_BASE;
-
-#ifndef CONFIG_CRASH_DUMP
-       /*
-        * Don't allow anybody to remap normal RAM that we're using.
-        * mem_init() sets high_memory so only do the check after that.
-        */
-       if (slab_is_available() && p <= virt_to_phys(high_memory - 1) &&
-           page_is_ram(__phys_to_pfn(p))) {
-               printk("__ioremap(): phys addr 0x%llx is RAM lr %ps\n",
-                      (unsigned long long)p, __builtin_return_address(0));
-               return NULL;
-       }
-#endif
-
-       if (size == 0)
-               return NULL;
-
-       /*
-        * Is it already mapped?  Perhaps overlapped by a previous
-        * mapping.
-        */
-       v = p_block_mapped(p);
-       if (v)
-               goto out;
-
-       if (slab_is_available()) {
-               struct vm_struct *area;
-               area = get_vm_area_caller(size, VM_IOREMAP, caller);
-               if (area == 0)
-                       return NULL;
-               area->phys_addr = p;
-               v = (unsigned long) area->addr;
-       } else {
-               v = (ioremap_bot -= size);
-       }
-
-       /*
-        * Should check if it is a candidate for a BAT mapping
-        */
-
-       err = 0;
-       for (i = 0; i < size && err == 0; i += PAGE_SIZE)
-               err = map_kernel_page(v + i, p + i, prot);
-       if (err) {
-               if (slab_is_available())
-                       vunmap((void *)v);
-               return NULL;
-       }
-
-out:
-       return (void __iomem *) (v + ((unsigned long)addr & ~PAGE_MASK));
-}
-
 static void __init *early_alloc_pgtable(unsigned long size)
 {
        void *ptr = memblock_alloc(size, size);
diff --git a/arch/powerpc/mm/pgtable_64.c b/arch/powerpc/mm/pgtable_64.c
index 32220f7381d7..781263df9f5e 100644
--- a/arch/powerpc/mm/pgtable_64.c
+++ b/arch/powerpc/mm/pgtable_64.c
@@ -101,67 +101,6 @@ unsigned long __pte_frag_size_shift;
 EXPORT_SYMBOL(__pte_frag_size_shift);
 #endif
 
-int __weak ioremap_range(unsigned long ea, phys_addr_t pa, unsigned long size, 
pgprot_t prot, int nid)
-{
-       unsigned long i;
-
-       for (i = 0; i < size; i += PAGE_SIZE) {
-               int err = map_kernel_page(ea + i, pa + i, prot);
-               if (err) {
-                       if (slab_is_available())
-                               unmap_kernel_range(ea, size);
-                       else
-                               WARN_ON_ONCE(1); /* Should clean up */
-                       return err;
-               }
-       }
-
-       return 0;
-}
-
-void __iomem * __ioremap_caller(phys_addr_t addr, unsigned long size,
-                               pgprot_t prot, void *caller)
-{
-       phys_addr_t paligned;
-       void __iomem *ret;
-
-       /*
-        * Choose an address to map it to.
-        * Once the vmalloc system is running, we use it.
-        * Before that, we map using addresses going
-        * down from ioremap_bot.  vmalloc will use
-        * the addresses from IOREMAP_BASE through
-        * ioremap_bot
-        * 
-        */
-       paligned = addr & PAGE_MASK;
-       size = PAGE_ALIGN(addr + size) - paligned;
-
-       if ((size == 0) || (paligned == 0))
-               return NULL;
-
-       if (slab_is_available()) {
-               struct vm_struct *area;
-
-               area = __get_vm_area_caller(size, VM_IOREMAP,
-                                           IOREMAP_BASE, ioremap_bot,
-                                           caller);
-               if (area == NULL)
-                       return NULL;
-
-               area->phys_addr = paligned;
-               ret = __ioremap_at(paligned, area->addr, size, prot);
-       } else {
-               ret = __ioremap_at(paligned, (void *)ioremap_bot - size, size, 
prot);
-               if (ret)
-                       ioremap_bot -= size;
-       }
-
-       if (ret)
-               ret += addr & ~PAGE_MASK;
-       return ret;
-}
-
 #ifndef __PAGETABLE_PUD_FOLDED
 /* 4 level page table */
 struct page *pgd_page(pgd_t pgd)
-- 
2.13.3

Reply via email to