Commit d4f9d02b authored by Rik van Riel's avatar Rik van Riel Committed by Linus Torvalds

[PATCH] token based thrashing control

The following experimental patch implements token based thrashing
protection, using the algorithm described in:

	http://www.cs.wm.edu/~sjiang/token.htm

When there are pageins going on, a task can grab a token, that protects the
task from pageout (except by itself) until it is no longer doing heavy
pageins, or until the maximum hold time of the token is over.

If the maximum hold time is exceeded, the task isn't eligable to hold the
token for a while more, since it wasn't doing it much good anyway.

I have run a very unscientific benchmark on my system to test the
effectiveness of the patch, timing how a 230MB two-process qsbench run
takes, with and without the token thrashing protection present.

normal 2.6.8-rc6:	6m45s
2.6.8-rc6 + token:	4m24s

This is a quick hack, implemented without having talked to the inventor of
the algorithm.  He's copied on the mail and I suspect we'll be able to do
better than my quick implementation ...
Signed-off-by: default avatarAndrew Morton <akpm@osdl.org>
Signed-off-by: default avatarLinus Torvalds <torvalds@osdl.org>
parent f09de30c
...@@ -220,6 +220,10 @@ struct mm_struct { ...@@ -220,6 +220,10 @@ struct mm_struct {
/* Architecture-specific MM context */ /* Architecture-specific MM context */
mm_context_t context; mm_context_t context;
/* Token based thrashing protection. */
unsigned long swap_token_time;
char recent_pagein;
/* coredumping support */ /* coredumping support */
int core_waiters; int core_waiters;
struct completion *core_startup_done, core_done; struct completion *core_startup_done, core_done;
......
...@@ -204,7 +204,6 @@ extern void free_pages_and_swap_cache(struct page **, int); ...@@ -204,7 +204,6 @@ extern void free_pages_and_swap_cache(struct page **, int);
extern struct page * lookup_swap_cache(swp_entry_t); extern struct page * lookup_swap_cache(swp_entry_t);
extern struct page * read_swap_cache_async(swp_entry_t, struct vm_area_struct *vma, extern struct page * read_swap_cache_async(swp_entry_t, struct vm_area_struct *vma,
unsigned long addr); unsigned long addr);
/* linux/mm/swapfile.c */ /* linux/mm/swapfile.c */
extern long total_swap_pages; extern long total_swap_pages;
extern unsigned int nr_swapfiles; extern unsigned int nr_swapfiles;
...@@ -229,6 +228,22 @@ extern spinlock_t swaplock; ...@@ -229,6 +228,22 @@ extern spinlock_t swaplock;
#define swap_device_lock(p) spin_lock(&p->sdev_lock) #define swap_device_lock(p) spin_lock(&p->sdev_lock)
#define swap_device_unlock(p) spin_unlock(&p->sdev_lock) #define swap_device_unlock(p) spin_unlock(&p->sdev_lock)
/* linux/mm/thrash.c */
extern struct mm_struct * swap_token_mm;
extern void grab_swap_token(void);
extern void __put_swap_token(struct mm_struct *);
static inline int has_swap_token(struct mm_struct *mm)
{
return (mm == swap_token_mm);
}
static inline void put_swap_token(struct mm_struct *mm)
{
if (has_swap_token(mm))
__put_swap_token(mm);
}
#else /* CONFIG_SWAP */ #else /* CONFIG_SWAP */
#define total_swap_pages 0 #define total_swap_pages 0
...@@ -266,6 +281,11 @@ static inline swp_entry_t get_swap_page(void) ...@@ -266,6 +281,11 @@ static inline swp_entry_t get_swap_page(void)
return entry; return entry;
} }
/* linux/mm/thrash.c */
#define put_swap_token(x) do { } while(0)
#define grab_swap_token() do { } while(0)
#define has_swap_token(x) 0
#endif /* CONFIG_SWAP */ #endif /* CONFIG_SWAP */
#endif /* __KERNEL__*/ #endif /* __KERNEL__*/
#endif /* _LINUX_SWAP_H */ #endif /* _LINUX_SWAP_H */
...@@ -29,6 +29,7 @@ ...@@ -29,6 +29,7 @@
#include <linux/fs.h> #include <linux/fs.h>
#include <linux/cpu.h> #include <linux/cpu.h>
#include <linux/security.h> #include <linux/security.h>
#include <linux/swap.h>
#include <linux/syscalls.h> #include <linux/syscalls.h>
#include <linux/jiffies.h> #include <linux/jiffies.h>
#include <linux/futex.h> #include <linux/futex.h>
...@@ -461,6 +462,7 @@ void mmput(struct mm_struct *mm) ...@@ -461,6 +462,7 @@ void mmput(struct mm_struct *mm)
spin_unlock(&mmlist_lock); spin_unlock(&mmlist_lock);
exit_aio(mm); exit_aio(mm);
exit_mmap(mm); exit_mmap(mm);
put_swap_token(mm);
mmdrop(mm); mmdrop(mm);
} }
} }
......
...@@ -12,6 +12,6 @@ obj-y := bootmem.o filemap.o mempool.o oom_kill.o fadvise.o \ ...@@ -12,6 +12,6 @@ obj-y := bootmem.o filemap.o mempool.o oom_kill.o fadvise.o \
readahead.o slab.o swap.o truncate.o vmscan.o \ readahead.o slab.o swap.o truncate.o vmscan.o \
$(mmu-y) $(mmu-y)
obj-$(CONFIG_SWAP) += page_io.o swap_state.o swapfile.o obj-$(CONFIG_SWAP) += page_io.o swap_state.o swapfile.o thrash.o
obj-$(CONFIG_HUGETLBFS) += hugetlb.o obj-$(CONFIG_HUGETLBFS) += hugetlb.o
obj-$(CONFIG_NUMA) += mempolicy.o obj-$(CONFIG_NUMA) += mempolicy.o
...@@ -1195,6 +1195,7 @@ struct page * filemap_nopage(struct vm_area_struct * area, unsigned long address ...@@ -1195,6 +1195,7 @@ struct page * filemap_nopage(struct vm_area_struct * area, unsigned long address
* effect. * effect.
*/ */
error = page_cache_read(file, pgoff); error = page_cache_read(file, pgoff);
grab_swap_token();
/* /*
* The page we want has now been added to the page cache. * The page we want has now been added to the page cache.
......
...@@ -1351,6 +1351,7 @@ static int do_swap_page(struct mm_struct * mm, ...@@ -1351,6 +1351,7 @@ static int do_swap_page(struct mm_struct * mm,
/* Had to read the page from swap area: Major fault */ /* Had to read the page from swap area: Major fault */
ret = VM_FAULT_MAJOR; ret = VM_FAULT_MAJOR;
inc_page_state(pgmajfault); inc_page_state(pgmajfault);
grab_swap_token();
} }
mark_page_accessed(page); mark_page_accessed(page);
......
...@@ -230,6 +230,9 @@ static int page_referenced_one(struct page *page, ...@@ -230,6 +230,9 @@ static int page_referenced_one(struct page *page,
if (ptep_clear_flush_young(vma, address, pte)) if (ptep_clear_flush_young(vma, address, pte))
referenced++; referenced++;
if (mm != current->mm && has_swap_token(mm))
referenced++;
(*mapcount)--; (*mapcount)--;
out_unmap: out_unmap:
......
/*
* mm/thrash.c
*
* Copyright (C) 2004, Red Hat, Inc.
* Copyright (C) 2004, Rik van Riel <riel@redhat.com>
* Released under the GPL, see the file COPYING for details.
*
* Simple token based thrashing protection, using the algorithm
* described in: http://www.cs.wm.edu/~sjiang/token.pdf
*/
#include <linux/jiffies.h>
#include <linux/mm.h>
#include <linux/sched.h>
#include <linux/swap.h>
static spinlock_t swap_token_lock = SPIN_LOCK_UNLOCKED;
static unsigned long swap_token_timeout;
unsigned long swap_token_check;
struct mm_struct * swap_token_mm = &init_mm;
#define SWAP_TOKEN_CHECK_INTERVAL (HZ * 2)
#define SWAP_TOKEN_TIMEOUT (HZ * 300)
/*
* Take the token away if the process had no page faults
* in the last interval, or if it has held the token for
* too long.
*/
#define SWAP_TOKEN_ENOUGH_RSS 1
#define SWAP_TOKEN_TIMED_OUT 2
static int should_release_swap_token(struct mm_struct *mm)
{
int ret = 0;
if (!mm->recent_pagein)
ret = SWAP_TOKEN_ENOUGH_RSS;
else if (time_after(jiffies, swap_token_timeout))
ret = SWAP_TOKEN_TIMED_OUT;
mm->recent_pagein = 0;
return ret;
}
/*
* Try to grab the swapout protection token. We only try to
* grab it once every TOKEN_CHECK_INTERVAL, both to prevent
* SMP lock contention and to check that the process that held
* the token before is no longer thrashing.
*/
void grab_swap_token(void)
{
struct mm_struct *mm;
int reason;
/* We have the token. Let others know we still need it. */
if (has_swap_token(current->mm)) {
current->mm->recent_pagein = 1;
return;
}
if (time_after(jiffies, swap_token_check)) {
/* Can't get swapout protection if we exceed our RSS limit. */
// if (current->mm->rss > current->mm->rlimit_rss)
// return;
/* ... or if we recently held the token. */
if (time_before(jiffies, current->mm->swap_token_time))
return;
if (!spin_trylock(&swap_token_lock))
return;
swap_token_check = jiffies + SWAP_TOKEN_CHECK_INTERVAL;
mm = swap_token_mm;
if ((reason = should_release_swap_token(mm))) {
unsigned long eligible = jiffies;
if (reason == SWAP_TOKEN_TIMED_OUT) {
eligible += SWAP_TOKEN_TIMEOUT;
}
mm->swap_token_time = eligible;
swap_token_timeout = jiffies + SWAP_TOKEN_TIMEOUT;
swap_token_mm = current->mm;
}
spin_unlock(&swap_token_lock);
}
return;
}
/* Called on process exit. */
void __put_swap_token(struct mm_struct *mm)
{
spin_lock(&swap_token_lock);
if (likely(mm == swap_token_mm)) {
swap_token_mm = &init_mm;
swap_token_check = jiffies;
}
spin_unlock(&swap_token_lock);
}
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment