Commit 12bb4c68 authored by Alexander Gordeev's avatar Alexander Gordeev Committed by Heiko Carstens

s390/vmem: make variable and function names consistent

Rename some variable and functions to better clarify
what they are and what they do.
Signed-off-by: default avatarAlexander Gordeev <agordeev@linux.ibm.com>
Signed-off-by: default avatarHeiko Carstens <hca@linux.ibm.com>
parent af71657c
...@@ -76,20 +76,20 @@ static void vmem_pte_free(unsigned long *table) ...@@ -76,20 +76,20 @@ static void vmem_pte_free(unsigned long *table)
/* /*
* The unused vmemmap range, which was not yet memset(PAGE_UNUSED) ranges * The unused vmemmap range, which was not yet memset(PAGE_UNUSED) ranges
* from unused_pmd_start to next PMD_SIZE boundary. * from unused_sub_pmd_start to next PMD_SIZE boundary.
*/ */
static unsigned long unused_pmd_start; static unsigned long unused_sub_pmd_start;
static void vmemmap_flush_unused_pmd(void) static void vmemmap_flush_unused_sub_pmd(void)
{ {
if (!unused_pmd_start) if (!unused_sub_pmd_start)
return; return;
memset(__va(unused_pmd_start), PAGE_UNUSED, memset(__va(unused_sub_pmd_start), PAGE_UNUSED,
ALIGN(unused_pmd_start, PMD_SIZE) - unused_pmd_start); ALIGN(unused_sub_pmd_start, PMD_SIZE) - unused_sub_pmd_start);
unused_pmd_start = 0; unused_sub_pmd_start = 0;
} }
static void __vmemmap_use_sub_pmd(unsigned long start, unsigned long end) static void vmemmap_mark_sub_pmd_used(unsigned long start, unsigned long end)
{ {
/* /*
* As we expect to add in the same granularity as we remove, it's * As we expect to add in the same granularity as we remove, it's
...@@ -106,24 +106,24 @@ static void vmemmap_use_sub_pmd(unsigned long start, unsigned long end) ...@@ -106,24 +106,24 @@ static void vmemmap_use_sub_pmd(unsigned long start, unsigned long end)
* We only optimize if the new used range directly follows the * We only optimize if the new used range directly follows the
* previously unused range (esp., when populating consecutive sections). * previously unused range (esp., when populating consecutive sections).
*/ */
if (unused_pmd_start == start) { if (unused_sub_pmd_start == start) {
unused_pmd_start = end; unused_sub_pmd_start = end;
if (likely(IS_ALIGNED(unused_pmd_start, PMD_SIZE))) if (likely(IS_ALIGNED(unused_sub_pmd_start, PMD_SIZE)))
unused_pmd_start = 0; unused_sub_pmd_start = 0;
return; return;
} }
vmemmap_flush_unused_pmd(); vmemmap_flush_unused_sub_pmd();
__vmemmap_use_sub_pmd(start, end); vmemmap_mark_sub_pmd_used(start, end);
} }
static void vmemmap_use_new_sub_pmd(unsigned long start, unsigned long end) static void vmemmap_use_new_sub_pmd(unsigned long start, unsigned long end)
{ {
void *page = __va(ALIGN_DOWN(start, PMD_SIZE)); void *page = __va(ALIGN_DOWN(start, PMD_SIZE));
vmemmap_flush_unused_pmd(); vmemmap_flush_unused_sub_pmd();
/* Could be our memmap page is filled with PAGE_UNUSED already ... */ /* Could be our memmap page is filled with PAGE_UNUSED already ... */
__vmemmap_use_sub_pmd(start, end); vmemmap_mark_sub_pmd_used(start, end);
/* Mark the unused parts of the new memmap page PAGE_UNUSED. */ /* Mark the unused parts of the new memmap page PAGE_UNUSED. */
if (!IS_ALIGNED(start, PMD_SIZE)) if (!IS_ALIGNED(start, PMD_SIZE))
...@@ -134,7 +134,7 @@ static void vmemmap_use_new_sub_pmd(unsigned long start, unsigned long end) ...@@ -134,7 +134,7 @@ static void vmemmap_use_new_sub_pmd(unsigned long start, unsigned long end)
* unused range in the populated PMD. * unused range in the populated PMD.
*/ */
if (!IS_ALIGNED(end, PMD_SIZE)) if (!IS_ALIGNED(end, PMD_SIZE))
unused_pmd_start = end; unused_sub_pmd_start = end;
} }
/* Returns true if the PMD is completely unused and can be freed. */ /* Returns true if the PMD is completely unused and can be freed. */
...@@ -142,7 +142,7 @@ static bool vmemmap_unuse_sub_pmd(unsigned long start, unsigned long end) ...@@ -142,7 +142,7 @@ static bool vmemmap_unuse_sub_pmd(unsigned long start, unsigned long end)
{ {
void *page = __va(ALIGN_DOWN(start, PMD_SIZE)); void *page = __va(ALIGN_DOWN(start, PMD_SIZE));
vmemmap_flush_unused_pmd(); vmemmap_flush_unused_sub_pmd();
memset(__va(start), PAGE_UNUSED, end - start); memset(__va(start), PAGE_UNUSED, end - start);
return !memchr_inv(page, PAGE_UNUSED, PMD_SIZE); return !memchr_inv(page, PAGE_UNUSED, PMD_SIZE);
} }
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment