|
@@ -1393,15 +1393,18 @@ static void __init deferred_free_range(struct page *page,
|
|
|
return;
|
|
|
|
|
|
/* Free a large naturally-aligned chunk if possible */
|
|
|
- if (nr_pages == MAX_ORDER_NR_PAGES &&
|
|
|
- (pfn & (MAX_ORDER_NR_PAGES-1)) == 0) {
|
|
|
+ if (nr_pages == pageblock_nr_pages &&
|
|
|
+ (pfn & (pageblock_nr_pages - 1)) == 0) {
|
|
|
set_pageblock_migratetype(page, MIGRATE_MOVABLE);
|
|
|
- __free_pages_boot_core(page, MAX_ORDER-1);
|
|
|
+ __free_pages_boot_core(page, pageblock_order);
|
|
|
return;
|
|
|
}
|
|
|
|
|
|
- for (i = 0; i < nr_pages; i++, page++)
|
|
|
+ for (i = 0; i < nr_pages; i++, page++, pfn++) {
|
|
|
+ if ((pfn & (pageblock_nr_pages - 1)) == 0)
|
|
|
+ set_pageblock_migratetype(page, MIGRATE_MOVABLE);
|
|
|
__free_pages_boot_core(page, 0);
|
|
|
+ }
|
|
|
}
|
|
|
|
|
|
/* Completion tracking for deferred_init_memmap() threads */
|
|
@@ -1469,9 +1472,9 @@ static int __init deferred_init_memmap(void *data)
|
|
|
|
|
|
/*
|
|
|
* Ensure pfn_valid is checked every
|
|
|
- * MAX_ORDER_NR_PAGES for memory holes
|
|
|
+ * pageblock_nr_pages for memory holes
|
|
|
*/
|
|
|
- if ((pfn & (MAX_ORDER_NR_PAGES - 1)) == 0) {
|
|
|
+ if ((pfn & (pageblock_nr_pages - 1)) == 0) {
|
|
|
if (!pfn_valid(pfn)) {
|
|
|
page = NULL;
|
|
|
goto free_range;
|
|
@@ -1484,7 +1487,7 @@ static int __init deferred_init_memmap(void *data)
|
|
|
}
|
|
|
|
|
|
/* Minimise pfn page lookups and scheduler checks */
|
|
|
- if (page && (pfn & (MAX_ORDER_NR_PAGES - 1)) != 0) {
|
|
|
+ if (page && (pfn & (pageblock_nr_pages - 1)) != 0) {
|
|
|
page++;
|
|
|
} else {
|
|
|
nr_pages += nr_to_free;
|
|
@@ -1520,6 +1523,9 @@ free_range:
|
|
|
free_base_page = NULL;
|
|
|
free_base_pfn = nr_to_free = 0;
|
|
|
}
|
|
|
+ /* Free the last block of pages to allocator */
|
|
|
+ nr_pages += nr_to_free;
|
|
|
+ deferred_free_range(free_base_page, free_base_pfn, nr_to_free);
|
|
|
|
|
|
first_init_pfn = max(end_pfn, first_init_pfn);
|
|
|
}
|