Commit 2a7684a2 authored by Wu Fengguang's avatar Wu Fengguang Committed by Andi Kleen

HWPOISON: check and isolate corrupted free pages v2

If memory corruption hits the free buddy pages, we can safely ignore them.
No one will access them until page allocation time, then prep_new_page()
will automatically check and isolate PG_hwpoison page for us (for 0-order
allocation).

This patch expands prep_new_page() to check every component page in a high
order page allocation, in order to completely stop PG_hwpoison pages from
being recirculated.

Note that the common case -- only allocating a single page, doesn't
do any more work than before. Allocating > order 0 does a bit more work,
but that's relatively uncommon.

This simple implementation may drop some innocent neighbor pages, hopefully
it is not a big problem because the event should be rare enough.

This patch adds some runtime costs to high order page users.

[AK: Improved description]

v2: Andi Kleen:
Port to -mm code
Move check into separate function.
Don't dump stack in bad_pages for hwpoisoned pages.
Signed-off-by: default avatarWu Fengguang <fengguang.wu@intel.com>
Signed-off-by: default avatarAndi Kleen <ak@linux.intel.com>
parent 888b9f7c
...@@ -234,6 +234,12 @@ static void bad_page(struct page *page) ...@@ -234,6 +234,12 @@ static void bad_page(struct page *page)
static unsigned long nr_shown; static unsigned long nr_shown;
static unsigned long nr_unshown; static unsigned long nr_unshown;
/* Don't complain about poisoned pages */
if (PageHWPoison(page)) {
__ClearPageBuddy(page);
return;
}
/* /*
* Allow a burst of 60 reports, then keep quiet for that minute; * Allow a burst of 60 reports, then keep quiet for that minute;
* or allow a steady drip of one report per second. * or allow a steady drip of one report per second.
...@@ -646,7 +652,7 @@ static inline void expand(struct zone *zone, struct page *page, ...@@ -646,7 +652,7 @@ static inline void expand(struct zone *zone, struct page *page,
/* /*
* This page is about to be returned from the page allocator * This page is about to be returned from the page allocator
*/ */
static int prep_new_page(struct page *page, int order, gfp_t gfp_flags) static inline int check_new_page(struct page *page)
{ {
if (unlikely(page_mapcount(page) | if (unlikely(page_mapcount(page) |
(page->mapping != NULL) | (page->mapping != NULL) |
...@@ -655,6 +661,18 @@ static int prep_new_page(struct page *page, int order, gfp_t gfp_flags) ...@@ -655,6 +661,18 @@ static int prep_new_page(struct page *page, int order, gfp_t gfp_flags)
bad_page(page); bad_page(page);
return 1; return 1;
} }
return 0;
}
static int prep_new_page(struct page *page, int order, gfp_t gfp_flags)
{
int i;
for (i = 0; i < (1 << order); i++) {
struct page *p = page + i;
if (unlikely(check_new_page(p)))
return 1;
}
set_page_private(page, 0); set_page_private(page, 0);
set_page_refcounted(page); set_page_refcounted(page);
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment