x86-64: fall back to regular page vmemmap on allocation failure
[platform/adaptation/renesas_rcar/renesas_kernel.git] / arch / x86 / mm / init_64.c
index 528c143..71ff55a 100644 (file)
@@ -1281,7 +1281,8 @@ static long __meminitdata addr_start, addr_end;
 static void __meminitdata *p_start, *p_end;
 static int __meminitdata node_start;
 
-int __meminit vmemmap_populate(unsigned long start, unsigned long end, int node)
+static int __meminit vmemmap_populate_hugepages(unsigned long start,
+                                               unsigned long end, int node)
 {
        unsigned long addr;
        unsigned long next;
@@ -1290,7 +1291,7 @@ int __meminit vmemmap_populate(unsigned long start, unsigned long end, int node)
        pmd_t *pmd;
 
        for (addr = start; addr < end; addr = next) {
-               void *p = NULL;
+               next = pmd_addr_end(addr, end);
 
                pgd = vmemmap_pgd_populate(addr, node);
                if (!pgd)
@@ -1300,31 +1301,14 @@ int __meminit vmemmap_populate(unsigned long start, unsigned long end, int node)
                if (!pud)
                        return -ENOMEM;
 
-               if (!cpu_has_pse) {
-                       next = (addr + PAGE_SIZE) & PAGE_MASK;
-                       pmd = vmemmap_pmd_populate(pud, addr, node);
-
-                       if (!pmd)
-                               return -ENOMEM;
-
-                       p = vmemmap_pte_populate(pmd, addr, node);
+               pmd = pmd_offset(pud, addr);
+               if (pmd_none(*pmd)) {
+                       void *p;
 
-                       if (!p)
-                               return -ENOMEM;
-
-                       addr_end = addr + PAGE_SIZE;
-                       p_end = p + PAGE_SIZE;
-               } else {
-                       next = pmd_addr_end(addr, end);
-
-                       pmd = pmd_offset(pud, addr);
-                       if (pmd_none(*pmd)) {
+                       p = vmemmap_alloc_block_buf(PMD_SIZE, node);
+                       if (p) {
                                pte_t entry;
 
-                               p = vmemmap_alloc_block_buf(PMD_SIZE, node);
-                               if (!p)
-                                       return -ENOMEM;
-
                                entry = pfn_pte(__pa(p) >> PAGE_SHIFT,
                                                PAGE_KERNEL_LARGE);
                                set_pmd(pmd, __pmd(pte_val(entry)));
@@ -1341,15 +1325,32 @@ int __meminit vmemmap_populate(unsigned long start, unsigned long end, int node)
 
                                addr_end = addr + PMD_SIZE;
                                p_end = p + PMD_SIZE;
-                       } else
-                               vmemmap_verify((pte_t *)pmd, node, addr, next);
+                               continue;
+                       }
+               } else if (pmd_large(*pmd)) {
+                       vmemmap_verify((pte_t *)pmd, node, addr, next);
+                       continue;
                }
-
+               pr_warn_once("vmemmap: falling back to regular page backing\n");
+               if (vmemmap_populate_basepages(addr, next, node))
+                       return -ENOMEM;
        }
-       sync_global_pgds(start, end - 1);
        return 0;
 }
 
+int __meminit vmemmap_populate(unsigned long start, unsigned long end, int node)
+{
+       int err;
+
+       if (cpu_has_pse)
+               err = vmemmap_populate_hugepages(start, end, node);
+       else
+               err = vmemmap_populate_basepages(start, end, node);
+       if (!err)
+               sync_global_pgds(start, end - 1);
+       return err;
+}
+
 #if defined(CONFIG_MEMORY_HOTPLUG_SPARSE) && defined(CONFIG_HAVE_BOOTMEM_INFO_NODE)
 void register_page_bootmem_memmap(unsigned long section_nr,
                                  struct page *start_page, unsigned long size)