Commit 524e00b3 authored by Liam R. Howlett's avatar Liam R. Howlett Committed by Andrew Morton

mm: remove rb tree.

Remove the RB tree and start using the maple tree for vm_area_struct
tracking.

Drop validate_mm() calls in expand_upwards() and expand_downwards() as the
lock is not held.

Link: https://lkml.kernel.org/r/20220906194824.2110408-18-Liam.Howlett@oracle.comSigned-off-by: default avatarLiam R. Howlett <Liam.Howlett@Oracle.com>
Tested-by: default avatarYu Zhao <yuzhao@google.com>
Cc: Catalin Marinas <catalin.marinas@arm.com>
Cc: David Hildenbrand <david@redhat.com>
Cc: David Howells <dhowells@redhat.com>
Cc: Davidlohr Bueso <dave@stgolabs.net>
Cc: "Matthew Wilcox (Oracle)" <willy@infradead.org>
Cc: SeongJae Park <sj@kernel.org>
Cc: Sven Schnelle <svens@linux.ibm.com>
Cc: Vlastimil Babka <vbabka@suse.cz>
Cc: Will Deacon <will@kernel.org>
Signed-off-by: default avatarAndrew Morton <akpm@linux-foundation.org>
parent 0c563f14
...@@ -95,7 +95,6 @@ void __init tboot_probe(void) ...@@ -95,7 +95,6 @@ void __init tboot_probe(void)
static pgd_t *tboot_pg_dir; static pgd_t *tboot_pg_dir;
static struct mm_struct tboot_mm = { static struct mm_struct tboot_mm = {
.mm_rb = RB_ROOT,
.mm_mt = MTREE_INIT_EXT(mm_mt, MM_MT_FLAGS, tboot_mm.mmap_lock), .mm_mt = MTREE_INIT_EXT(mm_mt, MM_MT_FLAGS, tboot_mm.mmap_lock),
.pgd = swapper_pg_dir, .pgd = swapper_pg_dir,
.mm_users = ATOMIC_INIT(2), .mm_users = ATOMIC_INIT(2),
......
...@@ -57,7 +57,6 @@ static unsigned long __initdata mem_reserve = EFI_INVALID_TABLE_ADDR; ...@@ -57,7 +57,6 @@ static unsigned long __initdata mem_reserve = EFI_INVALID_TABLE_ADDR;
static unsigned long __initdata rt_prop = EFI_INVALID_TABLE_ADDR; static unsigned long __initdata rt_prop = EFI_INVALID_TABLE_ADDR;
struct mm_struct efi_mm = { struct mm_struct efi_mm = {
.mm_rb = RB_ROOT,
.mm_mt = MTREE_INIT_EXT(mm_mt, MM_MT_FLAGS, efi_mm.mmap_lock), .mm_mt = MTREE_INIT_EXT(mm_mt, MM_MT_FLAGS, efi_mm.mmap_lock),
.mm_users = ATOMIC_INIT(2), .mm_users = ATOMIC_INIT(2),
.mm_count = ATOMIC_INIT(1), .mm_count = ATOMIC_INIT(1),
......
...@@ -2654,8 +2654,6 @@ extern int __split_vma(struct mm_struct *, struct vm_area_struct *, ...@@ -2654,8 +2654,6 @@ extern int __split_vma(struct mm_struct *, struct vm_area_struct *,
extern int split_vma(struct mm_struct *, struct vm_area_struct *, extern int split_vma(struct mm_struct *, struct vm_area_struct *,
unsigned long addr, int new_below); unsigned long addr, int new_below);
extern int insert_vm_struct(struct mm_struct *, struct vm_area_struct *); extern int insert_vm_struct(struct mm_struct *, struct vm_area_struct *);
extern void __vma_link_rb(struct mm_struct *, struct vm_area_struct *,
struct rb_node **, struct rb_node *);
extern void unlink_file_vma(struct vm_area_struct *); extern void unlink_file_vma(struct vm_area_struct *);
extern struct vm_area_struct *copy_vma(struct vm_area_struct **, extern struct vm_area_struct *copy_vma(struct vm_area_struct **,
unsigned long addr, unsigned long len, pgoff_t pgoff, unsigned long addr, unsigned long len, pgoff_t pgoff,
......
...@@ -410,19 +410,6 @@ struct vm_area_struct { ...@@ -410,19 +410,6 @@ struct vm_area_struct {
/* linked list of VM areas per task, sorted by address */ /* linked list of VM areas per task, sorted by address */
struct vm_area_struct *vm_next, *vm_prev; struct vm_area_struct *vm_next, *vm_prev;
struct rb_node vm_rb;
/*
* Largest free memory gap in bytes to the left of this VMA.
* Either between this VMA and vma->vm_prev, or between one of the
* VMAs below us in the VMA rbtree and its ->vm_prev. This helps
* get_unmapped_area find a free area of the right size.
*/
unsigned long rb_subtree_gap;
/* Second cache line starts here. */
struct mm_struct *vm_mm; /* The address space we belong to. */ struct mm_struct *vm_mm; /* The address space we belong to. */
/* /*
...@@ -488,7 +475,6 @@ struct mm_struct { ...@@ -488,7 +475,6 @@ struct mm_struct {
struct { struct {
struct vm_area_struct *mmap; /* list of VMAs */ struct vm_area_struct *mmap; /* list of VMAs */
struct maple_tree mm_mt; struct maple_tree mm_mt;
struct rb_root mm_rb;
u64 vmacache_seqnum; /* per-thread vmacache */ u64 vmacache_seqnum; /* per-thread vmacache */
#ifdef CONFIG_MMU #ifdef CONFIG_MMU
unsigned long (*get_unmapped_area) (struct file *filp, unsigned long (*get_unmapped_area) (struct file *filp,
......
...@@ -581,7 +581,6 @@ static __latent_entropy int dup_mmap(struct mm_struct *mm, ...@@ -581,7 +581,6 @@ static __latent_entropy int dup_mmap(struct mm_struct *mm,
struct mm_struct *oldmm) struct mm_struct *oldmm)
{ {
struct vm_area_struct *mpnt, *tmp, *prev, **pprev; struct vm_area_struct *mpnt, *tmp, *prev, **pprev;
struct rb_node **rb_link, *rb_parent;
int retval; int retval;
unsigned long charge = 0; unsigned long charge = 0;
LIST_HEAD(uf); LIST_HEAD(uf);
...@@ -608,8 +607,6 @@ static __latent_entropy int dup_mmap(struct mm_struct *mm, ...@@ -608,8 +607,6 @@ static __latent_entropy int dup_mmap(struct mm_struct *mm,
mm->exec_vm = oldmm->exec_vm; mm->exec_vm = oldmm->exec_vm;
mm->stack_vm = oldmm->stack_vm; mm->stack_vm = oldmm->stack_vm;
rb_link = &mm->mm_rb.rb_node;
rb_parent = NULL;
pprev = &mm->mmap; pprev = &mm->mmap;
retval = ksm_fork(mm, oldmm); retval = ksm_fork(mm, oldmm);
if (retval) if (retval)
...@@ -701,10 +698,6 @@ static __latent_entropy int dup_mmap(struct mm_struct *mm, ...@@ -701,10 +698,6 @@ static __latent_entropy int dup_mmap(struct mm_struct *mm,
tmp->vm_prev = prev; tmp->vm_prev = prev;
prev = tmp; prev = tmp;
__vma_link_rb(mm, tmp, rb_link, rb_parent);
rb_link = &tmp->vm_rb.rb_right;
rb_parent = &tmp->vm_rb;
/* Link the vma into the MT */ /* Link the vma into the MT */
mas.index = tmp->vm_start; mas.index = tmp->vm_start;
mas.last = tmp->vm_end - 1; mas.last = tmp->vm_end - 1;
...@@ -1133,7 +1126,6 @@ static struct mm_struct *mm_init(struct mm_struct *mm, struct task_struct *p, ...@@ -1133,7 +1126,6 @@ static struct mm_struct *mm_init(struct mm_struct *mm, struct task_struct *p,
struct user_namespace *user_ns) struct user_namespace *user_ns)
{ {
mm->mmap = NULL; mm->mmap = NULL;
mm->mm_rb = RB_ROOT;
mt_init_flags(&mm->mm_mt, MM_MT_FLAGS); mt_init_flags(&mm->mm_mt, MM_MT_FLAGS);
mt_set_external_lock(&mm->mm_mt, &mm->mmap_lock); mt_set_external_lock(&mm->mm_mt, &mm->mmap_lock);
mm->vmacache_seqnum = 0; mm->vmacache_seqnum = 0;
......
// SPDX-License-Identifier: GPL-2.0 // SPDX-License-Identifier: GPL-2.0
#include <linux/mm_types.h> #include <linux/mm_types.h>
#include <linux/rbtree.h>
#include <linux/maple_tree.h> #include <linux/maple_tree.h>
#include <linux/rwsem.h> #include <linux/rwsem.h>
#include <linux/spinlock.h> #include <linux/spinlock.h>
...@@ -29,7 +28,6 @@ ...@@ -29,7 +28,6 @@
* and size this cpu_bitmask to NR_CPUS. * and size this cpu_bitmask to NR_CPUS.
*/ */
struct mm_struct init_mm = { struct mm_struct init_mm = {
.mm_rb = RB_ROOT,
.mm_mt = MTREE_INIT_EXT(mm_mt, MM_MT_FLAGS, init_mm.mmap_lock), .mm_mt = MTREE_INIT_EXT(mm_mt, MM_MT_FLAGS, init_mm.mmap_lock),
.pgd = swapper_pg_dir, .pgd = swapper_pg_dir,
.mm_users = ATOMIC_INIT(2), .mm_users = ATOMIC_INIT(2),
......
This diff is collapsed.
...@@ -566,9 +566,9 @@ void vma_mas_remove(struct vm_area_struct *vma, struct ma_state *mas) ...@@ -566,9 +566,9 @@ void vma_mas_remove(struct vm_area_struct *vma, struct ma_state *mas)
*/ */
static void add_vma_to_mm(struct mm_struct *mm, struct vm_area_struct *vma) static void add_vma_to_mm(struct mm_struct *mm, struct vm_area_struct *vma)
{ {
struct vm_area_struct *pvma, *prev;
struct address_space *mapping; struct address_space *mapping;
struct rb_node **p, *parent, *rb_prev; struct vm_area_struct *prev;
MA_STATE(mas, &mm->mm_mt, vma->vm_start, vma->vm_end);
BUG_ON(!vma->vm_region); BUG_ON(!vma->vm_region);
...@@ -586,42 +586,10 @@ static void add_vma_to_mm(struct mm_struct *mm, struct vm_area_struct *vma) ...@@ -586,42 +586,10 @@ static void add_vma_to_mm(struct mm_struct *mm, struct vm_area_struct *vma)
i_mmap_unlock_write(mapping); i_mmap_unlock_write(mapping);
} }
prev = mas_prev(&mas, 0);
mas_reset(&mas);
/* add the VMA to the tree */ /* add the VMA to the tree */
parent = rb_prev = NULL; vma_mas_store(vma, &mas);
p = &mm->mm_rb.rb_node;
while (*p) {
parent = *p;
pvma = rb_entry(parent, struct vm_area_struct, vm_rb);
/* sort by: start addr, end addr, VMA struct addr in that order
* (the latter is necessary as we may get identical VMAs) */
if (vma->vm_start < pvma->vm_start)
p = &(*p)->rb_left;
else if (vma->vm_start > pvma->vm_start) {
rb_prev = parent;
p = &(*p)->rb_right;
} else if (vma->vm_end < pvma->vm_end)
p = &(*p)->rb_left;
else if (vma->vm_end > pvma->vm_end) {
rb_prev = parent;
p = &(*p)->rb_right;
} else if (vma < pvma)
p = &(*p)->rb_left;
else if (vma > pvma) {
rb_prev = parent;
p = &(*p)->rb_right;
} else
BUG();
}
rb_link_node(&vma->vm_rb, parent, p);
rb_insert_color(&vma->vm_rb, &mm->mm_rb);
/* add VMA to the VMA list also */
prev = NULL;
if (rb_prev)
prev = rb_entry(rb_prev, struct vm_area_struct, vm_rb);
__vma_link_list(mm, vma, prev); __vma_link_list(mm, vma, prev);
} }
...@@ -634,6 +602,7 @@ static void delete_vma_from_mm(struct vm_area_struct *vma) ...@@ -634,6 +602,7 @@ static void delete_vma_from_mm(struct vm_area_struct *vma)
struct address_space *mapping; struct address_space *mapping;
struct mm_struct *mm = vma->vm_mm; struct mm_struct *mm = vma->vm_mm;
struct task_struct *curr = current; struct task_struct *curr = current;
MA_STATE(mas, &vma->vm_mm->mm_mt, 0, 0);
mm->map_count--; mm->map_count--;
for (i = 0; i < VMACACHE_SIZE; i++) { for (i = 0; i < VMACACHE_SIZE; i++) {
...@@ -656,8 +625,7 @@ static void delete_vma_from_mm(struct vm_area_struct *vma) ...@@ -656,8 +625,7 @@ static void delete_vma_from_mm(struct vm_area_struct *vma)
} }
/* remove from the MM's tree and list */ /* remove from the MM's tree and list */
rb_erase(&vma->vm_rb, &mm->mm_rb); vma_mas_remove(vma, &mas);
__vma_unlink_list(mm, vma); __vma_unlink_list(mm, vma);
} }
...@@ -681,24 +649,19 @@ static void delete_vma(struct mm_struct *mm, struct vm_area_struct *vma) ...@@ -681,24 +649,19 @@ static void delete_vma(struct mm_struct *mm, struct vm_area_struct *vma)
struct vm_area_struct *find_vma(struct mm_struct *mm, unsigned long addr) struct vm_area_struct *find_vma(struct mm_struct *mm, unsigned long addr)
{ {
struct vm_area_struct *vma; struct vm_area_struct *vma;
MA_STATE(mas, &mm->mm_mt, addr, addr);
/* check the cache first */ /* check the cache first */
vma = vmacache_find(mm, addr); vma = vmacache_find(mm, addr);
if (likely(vma)) if (likely(vma))
return vma; return vma;
/* trawl the list (there may be multiple mappings in which addr vma = mas_walk(&mas);
* resides) */
for (vma = mm->mmap; vma; vma = vma->vm_next) {
if (vma->vm_start > addr)
return NULL;
if (vma->vm_end > addr) {
vmacache_update(addr, vma);
return vma;
}
}
return NULL; if (vma)
vmacache_update(addr, vma);
return vma;
} }
EXPORT_SYMBOL(find_vma); EXPORT_SYMBOL(find_vma);
...@@ -730,26 +693,23 @@ static struct vm_area_struct *find_vma_exact(struct mm_struct *mm, ...@@ -730,26 +693,23 @@ static struct vm_area_struct *find_vma_exact(struct mm_struct *mm,
{ {
struct vm_area_struct *vma; struct vm_area_struct *vma;
unsigned long end = addr + len; unsigned long end = addr + len;
MA_STATE(mas, &mm->mm_mt, addr, addr);
/* check the cache first */ /* check the cache first */
vma = vmacache_find_exact(mm, addr, end); vma = vmacache_find_exact(mm, addr, end);
if (vma) if (vma)
return vma; return vma;
/* trawl the list (there may be multiple mappings in which addr vma = mas_walk(&mas);
* resides) */ if (!vma)
for (vma = mm->mmap; vma; vma = vma->vm_next) { return NULL;
if (vma->vm_start < addr) if (vma->vm_start != addr)
continue; return NULL;
if (vma->vm_start > addr) if (vma->vm_end != end)
return NULL; return NULL;
if (vma->vm_end == end) {
vmacache_update(addr, vma);
return vma;
}
}
return NULL; vmacache_update(addr, vma);
return vma;
} }
/* /*
...@@ -1546,6 +1506,7 @@ void exit_mmap(struct mm_struct *mm) ...@@ -1546,6 +1506,7 @@ void exit_mmap(struct mm_struct *mm)
delete_vma(mm, vma); delete_vma(mm, vma);
cond_resched(); cond_resched();
} }
__mt_destroy(&mm->mm_mt);
} }
int vm_brk(unsigned long addr, unsigned long len) int vm_brk(unsigned long addr, unsigned long len)
......
...@@ -288,6 +288,8 @@ void __vma_link_list(struct mm_struct *mm, struct vm_area_struct *vma, ...@@ -288,6 +288,8 @@ void __vma_link_list(struct mm_struct *mm, struct vm_area_struct *vma,
vma->vm_next = next; vma->vm_next = next;
if (next) if (next)
next->vm_prev = vma; next->vm_prev = vma;
else
mm->highest_vm_end = vm_end_gap(vma);
} }
void __vma_unlink_list(struct mm_struct *mm, struct vm_area_struct *vma) void __vma_unlink_list(struct mm_struct *mm, struct vm_area_struct *vma)
...@@ -300,8 +302,14 @@ void __vma_unlink_list(struct mm_struct *mm, struct vm_area_struct *vma) ...@@ -300,8 +302,14 @@ void __vma_unlink_list(struct mm_struct *mm, struct vm_area_struct *vma)
prev->vm_next = next; prev->vm_next = next;
else else
mm->mmap = next; mm->mmap = next;
if (next) if (next) {
next->vm_prev = prev; next->vm_prev = prev;
} else {
if (prev)
mm->highest_vm_end = vm_end_gap(prev);
else
mm->highest_vm_end = 0;
}
} }
/* Check if the vma is being used as a stack by this task */ /* Check if the vma is being used as a stack by this task */
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment