Commit 8676af1f authored by Aslan Bakirov's avatar Aslan Bakirov Committed by Linus Torvalds

mm: cma: NUMA node interface

I've noticed that there is no interface exposed by CMA which would let
me to declare contigous memory on particular NUMA node.

This patchset adds the ability to try to allocate contiguous memory on a
specific node.  It will fallback to other nodes if the specified one
doesn't work.

Implement a new method for declaring contigous memory on particular node
and keep cma_declare_contiguous() as a wrapper.

[akpm@linux-foundation.org: build fix]
Signed-off-by: default avatarAslan Bakirov <aslan@fb.com>
Signed-off-by: default avatarRoman Gushchin <guro@fb.com>
Signed-off-by: default avatarAndrew Morton <akpm@linux-foundation.org>
Acked-by: default avatarMichal Hocko <mhocko@kernel.org>
Cc: Andreas Schaufler <andreas.schaufler@gmx.de>
Cc: Mike Kravetz <mike.kravetz@oracle.com>
Cc: Rik van Riel <riel@surriel.com>
Cc: Joonsoo Kim <js1304@gmail.com>
Link: http://lkml.kernel.org/r/20200407163840.92263-2-guro@fb.comSigned-off-by: default avatarLinus Torvalds <torvalds@linux-foundation.org>
parent 783fda85
...@@ -4,6 +4,7 @@ ...@@ -4,6 +4,7 @@
#include <linux/init.h> #include <linux/init.h>
#include <linux/types.h> #include <linux/types.h>
#include <linux/numa.h>
/* /*
* There is always at least global CMA area and a few optional * There is always at least global CMA area and a few optional
...@@ -24,10 +25,19 @@ extern phys_addr_t cma_get_base(const struct cma *cma); ...@@ -24,10 +25,19 @@ extern phys_addr_t cma_get_base(const struct cma *cma);
extern unsigned long cma_get_size(const struct cma *cma); extern unsigned long cma_get_size(const struct cma *cma);
extern const char *cma_get_name(const struct cma *cma); extern const char *cma_get_name(const struct cma *cma);
extern int __init cma_declare_contiguous(phys_addr_t base, extern int __init cma_declare_contiguous_nid(phys_addr_t base,
phys_addr_t size, phys_addr_t limit, phys_addr_t size, phys_addr_t limit,
phys_addr_t alignment, unsigned int order_per_bit, phys_addr_t alignment, unsigned int order_per_bit,
bool fixed, const char *name, struct cma **res_cma); bool fixed, const char *name, struct cma **res_cma,
int nid);
static inline int __init cma_declare_contiguous(phys_addr_t base,
phys_addr_t size, phys_addr_t limit,
phys_addr_t alignment, unsigned int order_per_bit,
bool fixed, const char *name, struct cma **res_cma)
{
return cma_declare_contiguous_nid(base, size, limit, alignment,
order_per_bit, fixed, name, res_cma, NUMA_NO_NODE);
}
extern int cma_init_reserved_mem(phys_addr_t base, phys_addr_t size, extern int cma_init_reserved_mem(phys_addr_t base, phys_addr_t size,
unsigned int order_per_bit, unsigned int order_per_bit,
const char *name, const char *name,
......
...@@ -348,6 +348,9 @@ static inline int memblock_get_region_node(const struct memblock_region *r) ...@@ -348,6 +348,9 @@ static inline int memblock_get_region_node(const struct memblock_region *r)
phys_addr_t memblock_phys_alloc_range(phys_addr_t size, phys_addr_t align, phys_addr_t memblock_phys_alloc_range(phys_addr_t size, phys_addr_t align,
phys_addr_t start, phys_addr_t end); phys_addr_t start, phys_addr_t end);
phys_addr_t memblock_alloc_range_nid(phys_addr_t size,
phys_addr_t align, phys_addr_t start,
phys_addr_t end, int nid, bool exact_nid);
phys_addr_t memblock_phys_alloc_try_nid(phys_addr_t size, phys_addr_t align, int nid); phys_addr_t memblock_phys_alloc_try_nid(phys_addr_t size, phys_addr_t align, int nid);
static inline phys_addr_t memblock_phys_alloc(phys_addr_t size, static inline phys_addr_t memblock_phys_alloc(phys_addr_t size,
......
...@@ -220,7 +220,7 @@ int __init cma_init_reserved_mem(phys_addr_t base, phys_addr_t size, ...@@ -220,7 +220,7 @@ int __init cma_init_reserved_mem(phys_addr_t base, phys_addr_t size,
} }
/** /**
* cma_declare_contiguous() - reserve custom contiguous area * cma_declare_contiguous_nid() - reserve custom contiguous area
* @base: Base address of the reserved area optional, use 0 for any * @base: Base address of the reserved area optional, use 0 for any
* @size: Size of the reserved area (in bytes), * @size: Size of the reserved area (in bytes),
* @limit: End address of the reserved memory (optional, 0 for any). * @limit: End address of the reserved memory (optional, 0 for any).
...@@ -229,6 +229,7 @@ int __init cma_init_reserved_mem(phys_addr_t base, phys_addr_t size, ...@@ -229,6 +229,7 @@ int __init cma_init_reserved_mem(phys_addr_t base, phys_addr_t size,
* @fixed: hint about where to place the reserved area * @fixed: hint about where to place the reserved area
* @name: The name of the area. See function cma_init_reserved_mem() * @name: The name of the area. See function cma_init_reserved_mem()
* @res_cma: Pointer to store the created cma region. * @res_cma: Pointer to store the created cma region.
* @nid: nid of the free area to find, %NUMA_NO_NODE for any node
* *
* This function reserves memory from early allocator. It should be * This function reserves memory from early allocator. It should be
* called by arch specific code once the early allocator (memblock or bootmem) * called by arch specific code once the early allocator (memblock or bootmem)
...@@ -238,10 +239,11 @@ int __init cma_init_reserved_mem(phys_addr_t base, phys_addr_t size, ...@@ -238,10 +239,11 @@ int __init cma_init_reserved_mem(phys_addr_t base, phys_addr_t size,
* If @fixed is true, reserve contiguous area at exactly @base. If false, * If @fixed is true, reserve contiguous area at exactly @base. If false,
* reserve in range from @base to @limit. * reserve in range from @base to @limit.
*/ */
int __init cma_declare_contiguous(phys_addr_t base, int __init cma_declare_contiguous_nid(phys_addr_t base,
phys_addr_t size, phys_addr_t limit, phys_addr_t size, phys_addr_t limit,
phys_addr_t alignment, unsigned int order_per_bit, phys_addr_t alignment, unsigned int order_per_bit,
bool fixed, const char *name, struct cma **res_cma) bool fixed, const char *name, struct cma **res_cma,
int nid)
{ {
phys_addr_t memblock_end = memblock_end_of_DRAM(); phys_addr_t memblock_end = memblock_end_of_DRAM();
phys_addr_t highmem_start; phys_addr_t highmem_start;
...@@ -336,14 +338,14 @@ int __init cma_declare_contiguous(phys_addr_t base, ...@@ -336,14 +338,14 @@ int __init cma_declare_contiguous(phys_addr_t base,
* memory in case of failure. * memory in case of failure.
*/ */
if (base < highmem_start && limit > highmem_start) { if (base < highmem_start && limit > highmem_start) {
addr = memblock_phys_alloc_range(size, alignment, addr = memblock_alloc_range_nid(size, alignment,
highmem_start, limit); highmem_start, limit, nid, false);
limit = highmem_start; limit = highmem_start;
} }
if (!addr) { if (!addr) {
addr = memblock_phys_alloc_range(size, alignment, base, addr = memblock_alloc_range_nid(size, alignment, base,
limit); limit, nid, false);
if (!addr) { if (!addr) {
ret = -ENOMEM; ret = -ENOMEM;
goto err; goto err;
......
...@@ -1349,7 +1349,7 @@ __next_mem_pfn_range_in_zone(u64 *idx, struct zone *zone, ...@@ -1349,7 +1349,7 @@ __next_mem_pfn_range_in_zone(u64 *idx, struct zone *zone,
* Return: * Return:
* Physical address of allocated memory block on success, %0 on failure. * Physical address of allocated memory block on success, %0 on failure.
*/ */
static phys_addr_t __init memblock_alloc_range_nid(phys_addr_t size, phys_addr_t __init memblock_alloc_range_nid(phys_addr_t size,
phys_addr_t align, phys_addr_t start, phys_addr_t align, phys_addr_t start,
phys_addr_t end, int nid, phys_addr_t end, int nid,
bool exact_nid) bool exact_nid)
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment