hugetlb: be sure to free demoted CMA pages to CMA
authorMike Kravetz <mike.kravetz@oracle.com>
Fri, 5 Nov 2021 20:41:27 +0000 (13:41 -0700)
committerLinus Torvalds <torvalds@linux-foundation.org>
Sat, 6 Nov 2021 20:30:39 +0000 (13:30 -0700)
When huge page demotion is fully implemented, gigantic pages can be
demoted to a smaller huge page size.  For example, on x86 a 1G page can
be demoted to 512 2M pages.  However, gigantic pages can potentially be
allocated from CMA.  If a gigantic page which was allocated from CMA is
demoted, the corresponding demoted pages needs to be returned to CMA.

Use the new interface cma_pages_valid() to determine if a non-gigantic
hugetlb page should be freed to CMA.  Also, clear mapping field of these
pages as expected by cma_release.

This also requires a change to CMA region creation for gigantic pages.
CMA uses a per-region bit map to track allocations.  When setting up the
region, you specify how many pages each bit represents.  Currently, only
gigantic pages are allocated/freed from CMA so the region is set up such
that one bit represents a gigantic page size allocation.

With demote, a gigantic page (allocation) could be split into smaller
size pages.  And, these smaller size pages will be freed to CMA.  So,
since the per-region bit map needs to be set up to represent the
smallest allocation/free size, it now needs to be set to the smallest
huge page size which can be freed to CMA.

Unfortunately, we set up the CMA region for huge pages before we set up
huge pages sizes (hstates).  So, technically we do not know the smallest
huge page size as this can change via command line options and
architecture specific code.  Therefore, at region setup time we use
HUGETLB_PAGE_ORDER as the smallest possible huge page size that can be
given back to CMA.  It is possible that this value is sub-optimal for
some architectures/config options.  If needed, this can be addressed in
follow on work.

Link: https://lkml.kernel.org/r/20211007181918.136982-4-mike.kravetz@oracle.com
Signed-off-by: Mike Kravetz <mike.kravetz@oracle.com>
Cc: "Aneesh Kumar K . V" <aneesh.kumar@linux.ibm.com>
Cc: David Hildenbrand <david@redhat.com>
Cc: David Rientjes <rientjes@google.com>
Cc: Michal Hocko <mhocko@suse.com>
Cc: Muchun Song <songmuchun@bytedance.com>
Cc: Naoya Horiguchi <naoya.horiguchi@linux.dev>
Cc: Nghia Le <nghialm78@gmail.com>
Cc: Oscar Salvador <osalvador@suse.de>
Cc: Zi Yan <ziy@nvidia.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
mm/hugetlb.c

index 1a18ff2..6662a99 100644 (file)
@@ -50,6 +50,16 @@ struct hstate hstates[HUGE_MAX_HSTATE];
 
 #ifdef CONFIG_CMA
 static struct cma *hugetlb_cma[MAX_NUMNODES];
+static bool hugetlb_cma_page(struct page *page, unsigned int order)
+{
+       return cma_pages_valid(hugetlb_cma[page_to_nid(page)], page,
+                               1 << order);
+}
+#else
+static bool hugetlb_cma_page(struct page *page, unsigned int order)
+{
+       return false;
+}
 #endif
 static unsigned long hugetlb_cma_size __initdata;
 
@@ -1272,6 +1282,7 @@ static void destroy_compound_gigantic_page(struct page *page,
        atomic_set(compound_pincount_ptr(page), 0);
 
        for (i = 1; i < nr_pages; i++, p = mem_map_next(p, page, i)) {
+               p->mapping = NULL;
                clear_compound_head(p);
                set_page_refcounted(p);
        }
@@ -1476,7 +1487,13 @@ static void __update_and_free_page(struct hstate *h, struct page *page)
                                1 << PG_active | 1 << PG_private |
                                1 << PG_writeback);
        }
-       if (hstate_is_gigantic(h)) {
+
+       /*
+        * Non-gigantic pages demoted from CMA allocated gigantic pages
+        * need to be given back to CMA in free_gigantic_page.
+        */
+       if (hstate_is_gigantic(h) ||
+           hugetlb_cma_page(page, huge_page_order(h))) {
                destroy_compound_gigantic_page(page, huge_page_order(h));
                free_gigantic_page(page, huge_page_order(h));
        } else {
@@ -3001,9 +3018,13 @@ static void __init hugetlb_init_hstates(void)
                 * h->demote_order is initially 0.
                 * - We can not demote gigantic pages if runtime freeing
                 *   is not supported, so skip this.
+                * - If CMA allocation is possible, we can not demote
+                *   HUGETLB_PAGE_ORDER or smaller size pages.
                 */
                if (hstate_is_gigantic(h) && !gigantic_page_runtime_supported())
                        continue;
+               if (hugetlb_cma_size && h->order <= HUGETLB_PAGE_ORDER)
+                       continue;
                for_each_hstate(h2) {
                        if (h2 == h)
                                continue;
@@ -3555,6 +3576,8 @@ static ssize_t demote_size_store(struct kobject *kobj,
        if (!demote_hstate)
                return -EINVAL;
        demote_order = demote_hstate->order;
+       if (demote_order < HUGETLB_PAGE_ORDER)
+               return -EINVAL;
 
        /* demote order must be smaller than hstate order */
        h = kobj_to_hstate(kobj, &nid);
@@ -6543,6 +6566,7 @@ void __init hugetlb_cma_reserve(int order)
        if (hugetlb_cma_size < (PAGE_SIZE << order)) {
                pr_warn("hugetlb_cma: cma area should be at least %lu MiB\n",
                        (PAGE_SIZE << order) / SZ_1M);
+               hugetlb_cma_size = 0;
                return;
        }
 
@@ -6563,7 +6587,13 @@ void __init hugetlb_cma_reserve(int order)
                size = round_up(size, PAGE_SIZE << order);
 
                snprintf(name, sizeof(name), "hugetlb%d", nid);
-               res = cma_declare_contiguous_nid(0, size, 0, PAGE_SIZE << order,
+               /*
+                * Note that 'order per bit' is based on smallest size that
+                * may be returned to CMA allocator in the case of
+                * huge page demotion.
+                */
+               res = cma_declare_contiguous_nid(0, size, 0,
+                                               PAGE_SIZE << HUGETLB_PAGE_ORDER,
                                                 0, false, name,
                                                 &hugetlb_cma[nid], nid);
                if (res) {
@@ -6579,6 +6609,13 @@ void __init hugetlb_cma_reserve(int order)
                if (reserved >= hugetlb_cma_size)
                        break;
        }
+
+       if (!reserved)
+               /*
+                * hugetlb_cma_size is used to determine if allocations from
+                * cma are possible.  Set to zero if no cma regions are set up.
+                */
+               hugetlb_cma_size = 0;
 }
 
 void __init hugetlb_cma_check(void)