|
@@ -69,7 +69,7 @@ static struct mem_section noinline __init_refok *sparse_index_alloc(int nid)
|
|
else
|
|
else
|
|
section = kzalloc(array_size, GFP_KERNEL);
|
|
section = kzalloc(array_size, GFP_KERNEL);
|
|
} else {
|
|
} else {
|
|
- section = alloc_bootmem_node(NODE_DATA(nid), array_size);
|
|
|
|
|
|
+ section = memblock_virt_alloc_node(array_size, nid);
|
|
}
|
|
}
|
|
|
|
|
|
return section;
|
|
return section;
|
|
@@ -279,8 +279,9 @@ sparse_early_usemaps_alloc_pgdat_section(struct pglist_data *pgdat,
|
|
limit = goal + (1UL << PA_SECTION_SHIFT);
|
|
limit = goal + (1UL << PA_SECTION_SHIFT);
|
|
nid = early_pfn_to_nid(goal >> PAGE_SHIFT);
|
|
nid = early_pfn_to_nid(goal >> PAGE_SHIFT);
|
|
again:
|
|
again:
|
|
- p = ___alloc_bootmem_node_nopanic(NODE_DATA(nid), size,
|
|
|
|
- SMP_CACHE_BYTES, goal, limit);
|
|
|
|
|
|
+ p = memblock_virt_alloc_try_nid_nopanic(size,
|
|
|
|
+ SMP_CACHE_BYTES, goal, limit,
|
|
|
|
+ nid);
|
|
if (!p && limit) {
|
|
if (!p && limit) {
|
|
limit = 0;
|
|
limit = 0;
|
|
goto again;
|
|
goto again;
|
|
@@ -331,7 +332,7 @@ static unsigned long * __init
|
|
sparse_early_usemaps_alloc_pgdat_section(struct pglist_data *pgdat,
|
|
sparse_early_usemaps_alloc_pgdat_section(struct pglist_data *pgdat,
|
|
unsigned long size)
|
|
unsigned long size)
|
|
{
|
|
{
|
|
- return alloc_bootmem_node_nopanic(pgdat, size);
|
|
|
|
|
|
+ return memblock_virt_alloc_node_nopanic(size, pgdat->node_id);
|
|
}
|
|
}
|
|
|
|
|
|
static void __init check_usemap_section_nr(int nid, unsigned long *usemap)
|
|
static void __init check_usemap_section_nr(int nid, unsigned long *usemap)
|
|
@@ -376,8 +377,9 @@ struct page __init *sparse_mem_map_populate(unsigned long pnum, int nid)
|
|
return map;
|
|
return map;
|
|
|
|
|
|
size = PAGE_ALIGN(sizeof(struct page) * PAGES_PER_SECTION);
|
|
size = PAGE_ALIGN(sizeof(struct page) * PAGES_PER_SECTION);
|
|
- map = __alloc_bootmem_node_high(NODE_DATA(nid), size,
|
|
|
|
- PAGE_SIZE, __pa(MAX_DMA_ADDRESS));
|
|
|
|
|
|
+ map = memblock_virt_alloc_try_nid(size,
|
|
|
|
+ PAGE_SIZE, __pa(MAX_DMA_ADDRESS),
|
|
|
|
+ BOOTMEM_ALLOC_ACCESSIBLE, nid);
|
|
return map;
|
|
return map;
|
|
}
|
|
}
|
|
void __init sparse_mem_maps_populate_node(struct page **map_map,
|
|
void __init sparse_mem_maps_populate_node(struct page **map_map,
|
|
@@ -401,8 +403,9 @@ void __init sparse_mem_maps_populate_node(struct page **map_map,
|
|
}
|
|
}
|
|
|
|
|
|
size = PAGE_ALIGN(size);
|
|
size = PAGE_ALIGN(size);
|
|
- map = __alloc_bootmem_node_high(NODE_DATA(nodeid), size * map_count,
|
|
|
|
- PAGE_SIZE, __pa(MAX_DMA_ADDRESS));
|
|
|
|
|
|
+ map = memblock_virt_alloc_try_nid(size * map_count,
|
|
|
|
+ PAGE_SIZE, __pa(MAX_DMA_ADDRESS),
|
|
|
|
+ BOOTMEM_ALLOC_ACCESSIBLE, nodeid);
|
|
if (map) {
|
|
if (map) {
|
|
for (pnum = pnum_begin; pnum < pnum_end; pnum++) {
|
|
for (pnum = pnum_begin; pnum < pnum_end; pnum++) {
|
|
if (!present_section_nr(pnum))
|
|
if (!present_section_nr(pnum))
|
|
@@ -545,7 +548,7 @@ void __init sparse_init(void)
|
|
* sparse_early_mem_map_alloc, so allocate usemap_map at first.
|
|
* sparse_early_mem_map_alloc, so allocate usemap_map at first.
|
|
*/
|
|
*/
|
|
size = sizeof(unsigned long *) * NR_MEM_SECTIONS;
|
|
size = sizeof(unsigned long *) * NR_MEM_SECTIONS;
|
|
- usemap_map = alloc_bootmem(size);
|
|
|
|
|
|
+ usemap_map = memblock_virt_alloc(size, 0);
|
|
if (!usemap_map)
|
|
if (!usemap_map)
|
|
panic("can not allocate usemap_map\n");
|
|
panic("can not allocate usemap_map\n");
|
|
alloc_usemap_and_memmap(sparse_early_usemaps_alloc_node,
|
|
alloc_usemap_and_memmap(sparse_early_usemaps_alloc_node,
|
|
@@ -553,7 +556,7 @@ void __init sparse_init(void)
|
|
|
|
|
|
#ifdef CONFIG_SPARSEMEM_ALLOC_MEM_MAP_TOGETHER
|
|
#ifdef CONFIG_SPARSEMEM_ALLOC_MEM_MAP_TOGETHER
|
|
size2 = sizeof(struct page *) * NR_MEM_SECTIONS;
|
|
size2 = sizeof(struct page *) * NR_MEM_SECTIONS;
|
|
- map_map = alloc_bootmem(size2);
|
|
|
|
|
|
+ map_map = memblock_virt_alloc(size2, 0);
|
|
if (!map_map)
|
|
if (!map_map)
|
|
panic("can not allocate map_map\n");
|
|
panic("can not allocate map_map\n");
|
|
alloc_usemap_and_memmap(sparse_early_mem_maps_alloc_node,
|
|
alloc_usemap_and_memmap(sparse_early_mem_maps_alloc_node,
|
|
@@ -583,9 +586,9 @@ void __init sparse_init(void)
|
|
vmemmap_populate_print_last();
|
|
vmemmap_populate_print_last();
|
|
|
|
|
|
#ifdef CONFIG_SPARSEMEM_ALLOC_MEM_MAP_TOGETHER
|
|
#ifdef CONFIG_SPARSEMEM_ALLOC_MEM_MAP_TOGETHER
|
|
- free_bootmem(__pa(map_map), size2);
|
|
|
|
|
|
+ memblock_free_early(__pa(map_map), size2);
|
|
#endif
|
|
#endif
|
|
- free_bootmem(__pa(usemap_map), size);
|
|
|
|
|
|
+ memblock_free_early(__pa(usemap_map), size);
|
|
}
|
|
}
|
|
|
|
|
|
#ifdef CONFIG_MEMORY_HOTPLUG
|
|
#ifdef CONFIG_MEMORY_HOTPLUG
|