Commit 643b1138 authored by Christoph Lameter's avatar Christoph Lameter Committed by Linus Torvalds

slub: enable tracking of full slabs

If slab tracking is on then build a list of full slabs so that we can verify
the integrity of all slabs and are also able to built list of alloc/free
callers.
Signed-off-by: default avatarChristoph Lameter <clameter@sgi.com>
Signed-off-by: default avatarAndrew Morton <akpm@linux-foundation.org>
Signed-off-by: default avatarLinus Torvalds <torvalds@linux-foundation.org>
parent 77c5e2d0
...@@ -16,6 +16,7 @@ struct kmem_cache_node { ...@@ -16,6 +16,7 @@ struct kmem_cache_node {
unsigned long nr_partial; unsigned long nr_partial;
atomic_long_t nr_slabs; atomic_long_t nr_slabs;
struct list_head partial; struct list_head partial;
struct list_head full;
}; };
/* /*
......
...@@ -661,6 +661,40 @@ static int on_freelist(struct kmem_cache *s, struct page *page, void *search) ...@@ -661,6 +661,40 @@ static int on_freelist(struct kmem_cache *s, struct page *page, void *search)
return search == NULL; return search == NULL;
} }
/*
* Tracking of fully allocated slabs for debugging
*/
static void add_full(struct kmem_cache *s, struct page *page)
{
struct kmem_cache_node *n;
VM_BUG_ON(!irqs_disabled());
VM_BUG_ON(!irqs_disabled());
if (!(s->flags & SLAB_STORE_USER))
return;
n = get_node(s, page_to_nid(page));
spin_lock(&n->list_lock);
list_add(&page->lru, &n->full);
spin_unlock(&n->list_lock);
}
static void remove_full(struct kmem_cache *s, struct page *page)
{
struct kmem_cache_node *n;
if (!(s->flags & SLAB_STORE_USER))
return;
n = get_node(s, page_to_nid(page));
spin_lock(&n->list_lock);
list_del(&page->lru);
spin_unlock(&n->list_lock);
}
static int alloc_object_checks(struct kmem_cache *s, struct page *page, static int alloc_object_checks(struct kmem_cache *s, struct page *page,
void *object) void *object)
{ {
...@@ -1090,6 +1124,8 @@ static void putback_slab(struct kmem_cache *s, struct page *page) ...@@ -1090,6 +1124,8 @@ static void putback_slab(struct kmem_cache *s, struct page *page)
if (page->inuse) { if (page->inuse) {
if (page->freelist) if (page->freelist)
add_partial(s, page); add_partial(s, page);
else if (PageError(page))
add_full(s, page);
slab_unlock(page); slab_unlock(page);
} else { } else {
slab_unlock(page); slab_unlock(page);
...@@ -1302,7 +1338,7 @@ static void slab_free(struct kmem_cache *s, struct page *page, ...@@ -1302,7 +1338,7 @@ static void slab_free(struct kmem_cache *s, struct page *page,
slab_empty: slab_empty:
if (prior) if (prior)
/* /*
* Partially used slab that is on the partial list. * Slab on the partial list.
*/ */
remove_partial(s, page); remove_partial(s, page);
...@@ -1314,6 +1350,8 @@ static void slab_free(struct kmem_cache *s, struct page *page, ...@@ -1314,6 +1350,8 @@ static void slab_free(struct kmem_cache *s, struct page *page,
debug: debug:
if (!free_object_checks(s, page, x)) if (!free_object_checks(s, page, x))
goto out_unlock; goto out_unlock;
if (!PageActive(page) && !page->freelist)
remove_full(s, page);
if (s->flags & SLAB_STORE_USER) if (s->flags & SLAB_STORE_USER)
set_track(s, x, TRACK_FREE, addr); set_track(s, x, TRACK_FREE, addr);
goto checks_ok; goto checks_ok;
...@@ -1466,6 +1504,7 @@ static void init_kmem_cache_node(struct kmem_cache_node *n) ...@@ -1466,6 +1504,7 @@ static void init_kmem_cache_node(struct kmem_cache_node *n)
atomic_long_set(&n->nr_slabs, 0); atomic_long_set(&n->nr_slabs, 0);
spin_lock_init(&n->list_lock); spin_lock_init(&n->list_lock);
INIT_LIST_HEAD(&n->partial); INIT_LIST_HEAD(&n->partial);
INIT_LIST_HEAD(&n->full);
} }
#ifdef CONFIG_NUMA #ifdef CONFIG_NUMA
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment