Commit 91030ca1 authored by Hugh Dickins's avatar Hugh Dickins Committed by Ingo Molnar

x86: unsigned long pte_pfn

pte_pfn() has always been of type unsigned long, even on 32-bit PAE;
but in the current tip/next/mm tree it works out to be unsigned long
long on 64-bit, which gives an irritating warning if you try to printk
a pfn with the usual %lx.

Now use the same pte_pfn() function, moved from pgtable-3level.h
to pgtable.h, for all models: as suggested by Jeremy Fitzhardinge.
And pte_page() can well move along with it (remaining a macro to
avoid dependence on mm_types.h).
Signed-off-by: default avatarHugh Dickins <hugh@veritas.com>
Acked-by: default avatarJeremy Fitzhardinge <jeremy@goop.org>
Signed-off-by: default avatarIngo Molnar <mingo@elte.hu>
parent adee14b2
...@@ -53,9 +53,7 @@ static inline pte_t native_ptep_get_and_clear(pte_t *xp) ...@@ -53,9 +53,7 @@ static inline pte_t native_ptep_get_and_clear(pte_t *xp)
#define native_ptep_get_and_clear(xp) native_local_ptep_get_and_clear(xp) #define native_ptep_get_and_clear(xp) native_local_ptep_get_and_clear(xp)
#endif #endif
#define pte_page(x) pfn_to_page(pte_pfn(x))
#define pte_none(x) (!(x).pte_low) #define pte_none(x) (!(x).pte_low)
#define pte_pfn(x) (pte_val(x) >> PAGE_SHIFT)
/* /*
* Bits 0, 6 and 7 are taken, split up the 29 bits of offset * Bits 0, 6 and 7 are taken, split up the 29 bits of offset
......
...@@ -151,18 +151,11 @@ static inline int pte_same(pte_t a, pte_t b) ...@@ -151,18 +151,11 @@ static inline int pte_same(pte_t a, pte_t b)
return a.pte_low == b.pte_low && a.pte_high == b.pte_high; return a.pte_low == b.pte_low && a.pte_high == b.pte_high;
} }
#define pte_page(x) pfn_to_page(pte_pfn(x))
static inline int pte_none(pte_t pte) static inline int pte_none(pte_t pte)
{ {
return !pte.pte_low && !pte.pte_high; return !pte.pte_low && !pte.pte_high;
} }
static inline unsigned long pte_pfn(pte_t pte)
{
return (pte_val(pte) & PTE_PFN_MASK) >> PAGE_SHIFT;
}
/* /*
* Bits 0, 6 and 7 are taken in the low part of the pte, * Bits 0, 6 and 7 are taken in the low part of the pte,
* put the 32 bits of offset into the high part. * put the 32 bits of offset into the high part.
......
...@@ -186,6 +186,13 @@ static inline int pte_special(pte_t pte) ...@@ -186,6 +186,13 @@ static inline int pte_special(pte_t pte)
return pte_val(pte) & _PAGE_SPECIAL; return pte_val(pte) & _PAGE_SPECIAL;
} }
static inline unsigned long pte_pfn(pte_t pte)
{
return (pte_val(pte) & PTE_PFN_MASK) >> PAGE_SHIFT;
}
#define pte_page(pte) pfn_to_page(pte_pfn(pte))
static inline int pmd_large(pmd_t pte) static inline int pmd_large(pmd_t pte)
{ {
return (pmd_val(pte) & (_PAGE_PSE | _PAGE_PRESENT)) == return (pmd_val(pte) & (_PAGE_PSE | _PAGE_PRESENT)) ==
......
...@@ -175,8 +175,6 @@ static inline int pmd_bad(pmd_t pmd) ...@@ -175,8 +175,6 @@ static inline int pmd_bad(pmd_t pmd)
#define pte_present(x) (pte_val((x)) & (_PAGE_PRESENT | _PAGE_PROTNONE)) #define pte_present(x) (pte_val((x)) & (_PAGE_PRESENT | _PAGE_PROTNONE))
#define pages_to_mb(x) ((x) >> (20 - PAGE_SHIFT)) /* FIXME: is this right? */ #define pages_to_mb(x) ((x) >> (20 - PAGE_SHIFT)) /* FIXME: is this right? */
#define pte_page(x) pfn_to_page(pte_pfn((x)))
#define pte_pfn(x) ((pte_val((x)) & __PHYSICAL_MASK) >> PAGE_SHIFT)
/* /*
* Macro to mark a page protection value as "uncacheable". * Macro to mark a page protection value as "uncacheable".
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment