Commit bb016b84 authored by Santosh Shilimkar's avatar Santosh Shilimkar Committed by Linus Torvalds

mm/sparse: use memblock apis for early memory allocations

Switch to memblock interfaces for early memory allocator instead of
bootmem allocator.  No functional change in beahvior than what it is in
current code from bootmem users points of view.

Archs already converted to NO_BOOTMEM now directly use memblock
interfaces instead of bootmem wrappers build on top of memblock.  And
the archs which still uses bootmem, these new apis just fallback to
exiting bootmem APIs.
Signed-off-by: default avatarSantosh Shilimkar <santosh.shilimkar@ti.com>
Cc: "Rafael J. Wysocki" <rjw@sisk.pl>
Cc: Arnd Bergmann <arnd@arndb.de>
Cc: Christoph Lameter <cl@linux-foundation.org>
Cc: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
Cc: Grygorii Strashko <grygorii.strashko@ti.com>
Cc: H. Peter Anvin <hpa@zytor.com>
Cc: Johannes Weiner <hannes@cmpxchg.org>
Cc: KAMEZAWA Hiroyuki <kamezawa.hiroyu@jp.fujitsu.com>
Cc: Konrad Rzeszutek Wilk <konrad.wilk@oracle.com>
Cc: Michal Hocko <mhocko@suse.cz>
Cc: Paul Walmsley <paul@pwsan.com>
Cc: Pavel Machek <pavel@ucw.cz>
Cc: Russell King <linux@arm.linux.org.uk>
Cc: Tejun Heo <tj@kernel.org>
Cc: Tony Lindgren <tony@atomide.com>
Cc: Yinghai Lu <yinghai@kernel.org>
Signed-off-by: default avatarAndrew Morton <akpm@linux-foundation.org>
Signed-off-by: default avatarLinus Torvalds <torvalds@linux-foundation.org>
parent c1529500
...@@ -40,7 +40,8 @@ static void * __init_refok __earlyonly_bootmem_alloc(int node, ...@@ -40,7 +40,8 @@ static void * __init_refok __earlyonly_bootmem_alloc(int node,
unsigned long align, unsigned long align,
unsigned long goal) unsigned long goal)
{ {
return __alloc_bootmem_node_high(NODE_DATA(node), size, align, goal); return memblock_virt_alloc_try_nid(size, align, goal,
BOOTMEM_ALLOC_ACCESSIBLE, node);
} }
static void *vmemmap_buf; static void *vmemmap_buf;
...@@ -226,7 +227,8 @@ void __init sparse_mem_maps_populate_node(struct page **map_map, ...@@ -226,7 +227,8 @@ void __init sparse_mem_maps_populate_node(struct page **map_map,
if (vmemmap_buf_start) { if (vmemmap_buf_start) {
/* need to free left buf */ /* need to free left buf */
free_bootmem(__pa(vmemmap_buf), vmemmap_buf_end - vmemmap_buf); memblock_free_early(__pa(vmemmap_buf),
vmemmap_buf_end - vmemmap_buf);
vmemmap_buf = NULL; vmemmap_buf = NULL;
vmemmap_buf_end = NULL; vmemmap_buf_end = NULL;
} }
......
...@@ -69,7 +69,7 @@ static struct mem_section noinline __init_refok *sparse_index_alloc(int nid) ...@@ -69,7 +69,7 @@ static struct mem_section noinline __init_refok *sparse_index_alloc(int nid)
else else
section = kzalloc(array_size, GFP_KERNEL); section = kzalloc(array_size, GFP_KERNEL);
} else { } else {
section = alloc_bootmem_node(NODE_DATA(nid), array_size); section = memblock_virt_alloc_node(array_size, nid);
} }
return section; return section;
...@@ -279,8 +279,9 @@ sparse_early_usemaps_alloc_pgdat_section(struct pglist_data *pgdat, ...@@ -279,8 +279,9 @@ sparse_early_usemaps_alloc_pgdat_section(struct pglist_data *pgdat,
limit = goal + (1UL << PA_SECTION_SHIFT); limit = goal + (1UL << PA_SECTION_SHIFT);
nid = early_pfn_to_nid(goal >> PAGE_SHIFT); nid = early_pfn_to_nid(goal >> PAGE_SHIFT);
again: again:
p = ___alloc_bootmem_node_nopanic(NODE_DATA(nid), size, p = memblock_virt_alloc_try_nid_nopanic(size,
SMP_CACHE_BYTES, goal, limit); SMP_CACHE_BYTES, goal, limit,
nid);
if (!p && limit) { if (!p && limit) {
limit = 0; limit = 0;
goto again; goto again;
...@@ -331,7 +332,7 @@ static unsigned long * __init ...@@ -331,7 +332,7 @@ static unsigned long * __init
sparse_early_usemaps_alloc_pgdat_section(struct pglist_data *pgdat, sparse_early_usemaps_alloc_pgdat_section(struct pglist_data *pgdat,
unsigned long size) unsigned long size)
{ {
return alloc_bootmem_node_nopanic(pgdat, size); return memblock_virt_alloc_node_nopanic(size, pgdat->node_id);
} }
static void __init check_usemap_section_nr(int nid, unsigned long *usemap) static void __init check_usemap_section_nr(int nid, unsigned long *usemap)
...@@ -376,8 +377,9 @@ struct page __init *sparse_mem_map_populate(unsigned long pnum, int nid) ...@@ -376,8 +377,9 @@ struct page __init *sparse_mem_map_populate(unsigned long pnum, int nid)
return map; return map;
size = PAGE_ALIGN(sizeof(struct page) * PAGES_PER_SECTION); size = PAGE_ALIGN(sizeof(struct page) * PAGES_PER_SECTION);
map = __alloc_bootmem_node_high(NODE_DATA(nid), size, map = memblock_virt_alloc_try_nid(size,
PAGE_SIZE, __pa(MAX_DMA_ADDRESS)); PAGE_SIZE, __pa(MAX_DMA_ADDRESS),
BOOTMEM_ALLOC_ACCESSIBLE, nid);
return map; return map;
} }
void __init sparse_mem_maps_populate_node(struct page **map_map, void __init sparse_mem_maps_populate_node(struct page **map_map,
...@@ -401,8 +403,9 @@ void __init sparse_mem_maps_populate_node(struct page **map_map, ...@@ -401,8 +403,9 @@ void __init sparse_mem_maps_populate_node(struct page **map_map,
} }
size = PAGE_ALIGN(size); size = PAGE_ALIGN(size);
map = __alloc_bootmem_node_high(NODE_DATA(nodeid), size * map_count, map = memblock_virt_alloc_try_nid(size * map_count,
PAGE_SIZE, __pa(MAX_DMA_ADDRESS)); PAGE_SIZE, __pa(MAX_DMA_ADDRESS),
BOOTMEM_ALLOC_ACCESSIBLE, nodeid);
if (map) { if (map) {
for (pnum = pnum_begin; pnum < pnum_end; pnum++) { for (pnum = pnum_begin; pnum < pnum_end; pnum++) {
if (!present_section_nr(pnum)) if (!present_section_nr(pnum))
...@@ -545,7 +548,7 @@ void __init sparse_init(void) ...@@ -545,7 +548,7 @@ void __init sparse_init(void)
* sparse_early_mem_map_alloc, so allocate usemap_map at first. * sparse_early_mem_map_alloc, so allocate usemap_map at first.
*/ */
size = sizeof(unsigned long *) * NR_MEM_SECTIONS; size = sizeof(unsigned long *) * NR_MEM_SECTIONS;
usemap_map = alloc_bootmem(size); usemap_map = memblock_virt_alloc(size, 0);
if (!usemap_map) if (!usemap_map)
panic("can not allocate usemap_map\n"); panic("can not allocate usemap_map\n");
alloc_usemap_and_memmap(sparse_early_usemaps_alloc_node, alloc_usemap_and_memmap(sparse_early_usemaps_alloc_node,
...@@ -553,7 +556,7 @@ void __init sparse_init(void) ...@@ -553,7 +556,7 @@ void __init sparse_init(void)
#ifdef CONFIG_SPARSEMEM_ALLOC_MEM_MAP_TOGETHER #ifdef CONFIG_SPARSEMEM_ALLOC_MEM_MAP_TOGETHER
size2 = sizeof(struct page *) * NR_MEM_SECTIONS; size2 = sizeof(struct page *) * NR_MEM_SECTIONS;
map_map = alloc_bootmem(size2); map_map = memblock_virt_alloc(size2, 0);
if (!map_map) if (!map_map)
panic("can not allocate map_map\n"); panic("can not allocate map_map\n");
alloc_usemap_and_memmap(sparse_early_mem_maps_alloc_node, alloc_usemap_and_memmap(sparse_early_mem_maps_alloc_node,
...@@ -583,9 +586,9 @@ void __init sparse_init(void) ...@@ -583,9 +586,9 @@ void __init sparse_init(void)
vmemmap_populate_print_last(); vmemmap_populate_print_last();
#ifdef CONFIG_SPARSEMEM_ALLOC_MEM_MAP_TOGETHER #ifdef CONFIG_SPARSEMEM_ALLOC_MEM_MAP_TOGETHER
free_bootmem(__pa(map_map), size2); memblock_free_early(__pa(map_map), size2);
#endif #endif
free_bootmem(__pa(usemap_map), size); memblock_free_early(__pa(usemap_map), size);
} }
#ifdef CONFIG_MEMORY_HOTPLUG #ifdef CONFIG_MEMORY_HOTPLUG
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment