mm: make __free_pages_bootmem() only available at boot time
authorJiang Liu <[email protected]>
Wed, 3 Jul 2013 22:03:17 +0000 (15:03 -0700)
committerLinus Torvalds <[email protected]>
Wed, 3 Jul 2013 23:07:33 +0000 (16:07 -0700)
In order to simpilify management of totalram_pages and
zone->managed_pages, make __free_pages_bootmem() only available at boot
time.  With this change applied, __free_pages_bootmem() will only be
used by bootmem.c and nobootmem.c at boot time, so mark it as __init.
Other callers of __free_pages_bootmem() have been converted to use
free_reserved_page(), which handles totalram_pages and
zone->managed_pages in a safer way.

This patch also fix a bug in free_pagetable() for x86_64, which should
increase zone->managed_pages instead of zone->present_pages when freeing
reserved pages.

And now we have managed_pages_count_lock to protect totalram_pages and
zone->managed_pages, so remove the redundant ppb_lock lock in
put_page_bootmem().  This greatly simplifies the locking rules.

Signed-off-by: Jiang Liu <[email protected]>
Cc: Thomas Gleixner <[email protected]>
Cc: Ingo Molnar <[email protected]>
Cc: "H. Peter Anvin" <[email protected]>
Cc: Yinghai Lu <[email protected]>
Cc: Wen Congyang <[email protected]>
Cc: Tang Chen <[email protected]>
Cc: Yasuaki Ishimatsu <[email protected]>
Cc: Mel Gorman <[email protected]>
Cc: Minchan Kim <[email protected]>
Cc: "Michael S. Tsirkin" <[email protected]>
Cc: <[email protected]>
Cc: Arnd Bergmann <[email protected]>
Cc: Catalin Marinas <[email protected]>
Cc: Chris Metcalf <[email protected]>
Cc: David Howells <[email protected]>
Cc: Geert Uytterhoeven <[email protected]>
Cc: Jeremy Fitzhardinge <[email protected]>
Cc: Jianguo Wu <[email protected]>
Cc: Joonsoo Kim <[email protected]>
Cc: Kamezawa Hiroyuki <[email protected]>
Cc: Konrad Rzeszutek Wilk <[email protected]>
Cc: Marek Szyprowski <[email protected]>
Cc: Michel Lespinasse <[email protected]>
Cc: Rik van Riel <[email protected]>
Cc: Rusty Russell <[email protected]>
Cc: Tejun Heo <[email protected]>
Cc: Will Deacon <[email protected]>
Cc: Russell King <[email protected]>
Signed-off-by: Andrew Morton <[email protected]>
Signed-off-by: Linus Torvalds <[email protected]>
arch/x86/mm/init_64.c
mm/memory_hotplug.c
mm/page_alloc.c

index b7bdf7bebf3b7192895af20a6f8340e78b53ae73..ec312a92b13798109af09c68b78fb5a376bf7217 100644 (file)
@@ -712,36 +712,22 @@ EXPORT_SYMBOL_GPL(arch_add_memory);
 
 static void __meminit free_pagetable(struct page *page, int order)
 {
-       struct zone *zone;
-       bool bootmem = false;
        unsigned long magic;
        unsigned int nr_pages = 1 << order;
 
        /* bootmem page has reserved flag */
        if (PageReserved(page)) {
                __ClearPageReserved(page);
-               bootmem = true;
 
                magic = (unsigned long)page->lru.next;
                if (magic == SECTION_INFO || magic == MIX_SECTION_INFO) {
                        while (nr_pages--)
                                put_page_bootmem(page++);
                } else
-                       __free_pages_bootmem(page, order);
+                       while (nr_pages--)
+                               free_reserved_page(page++);
        } else
                free_pages((unsigned long)page_address(page), order);
-
-       /*
-        * SECTION_INFO pages and MIX_SECTION_INFO pages
-        * are all allocated by bootmem.
-        */
-       if (bootmem) {
-               zone = page_zone(page);
-               zone_span_writelock(zone);
-               zone->present_pages += nr_pages;
-               zone_span_writeunlock(zone);
-               totalram_pages += nr_pages;
-       }
 }
 
 static void __meminit free_pte_table(pte_t *pte_start, pmd_t *pmd)
index 6096cb91873505c29a0917f1ce9cbe4ff6d80c74..814ecb2d262f76b4d3492b38722f4b097a9b197b 100644 (file)
@@ -101,12 +101,9 @@ void get_page_bootmem(unsigned long info,  struct page *page,
        atomic_inc(&page->_count);
 }
 
-/* reference to __meminit __free_pages_bootmem is valid
- * so use __ref to tell modpost not to generate a warning */
-void __ref put_page_bootmem(struct page *page)
+void put_page_bootmem(struct page *page)
 {
        unsigned long type;
-       static DEFINE_MUTEX(ppb_lock);
 
        type = (unsigned long) page->lru.next;
        BUG_ON(type < MEMORY_HOTPLUG_MIN_BOOTMEM_TYPE ||
@@ -116,17 +113,8 @@ void __ref put_page_bootmem(struct page *page)
                ClearPagePrivate(page);
                set_page_private(page, 0);
                INIT_LIST_HEAD(&page->lru);
-
-               /*
-                * Please refer to comment for __free_pages_bootmem()
-                * for why we serialize here.
-                */
-               mutex_lock(&ppb_lock);
-               __free_pages_bootmem(page, 0);
-               mutex_unlock(&ppb_lock);
-               totalram_pages++;
+               free_reserved_page(page);
        }
-
 }
 
 #ifdef CONFIG_HAVE_BOOTMEM_INFO_NODE
index 93f292a60cb03d0f799aed15ce94b761e4d3715f..2437a7e17aba2f8bf90011625d45a65e99d82af1 100644 (file)
@@ -745,14 +745,7 @@ static void __free_pages_ok(struct page *page, unsigned int order)
        local_irq_restore(flags);
 }
 
-/*
- * Read access to zone->managed_pages is safe because it's unsigned long,
- * but we still need to serialize writers. Currently all callers of
- * __free_pages_bootmem() except put_page_bootmem() should only be used
- * at boot time. So for shorter boot time, we shift the burden to
- * put_page_bootmem() to serialize writers.
- */
-void __meminit __free_pages_bootmem(struct page *page, unsigned int order)
+void __init __free_pages_bootmem(struct page *page, unsigned int order)
 {
        unsigned int nr_pages = 1 << order;
        unsigned int loop;