Commit 0cbe3e26 authored by Aneesh Kumar K.V's avatar Aneesh Kumar K.V Committed by Linus Torvalds

mm: update ptep_modify_prot_start/commit to take vm_area_struct as arg

Patch series "NestMMU pte upgrade workaround for mprotect", v5.

We can upgrade pte access (R -> RW transition) via mprotect.  We need to
make sure we follow the recommended pte update sequence as outlined in
commit bd5050e3 ("powerpc/mm/radix: Change pte relax sequence to
handle nest MMU hang") for such updates.  This patch series does that.

This patch (of 5):

Some architectures may want to call flush_tlb_range from these helpers.

Link: http://lkml.kernel.org/r/20190116085035.29729-2-aneesh.kumar@linux.ibm.comSigned-off-by: default avatarAneesh Kumar K.V <aneesh.kumar@linux.ibm.com>
Cc: Nicholas Piggin <npiggin@gmail.com>
Cc: Benjamin Herrenschmidt <benh@kernel.crashing.org>
Cc: Paul Mackerras <paulus@samba.org>
Cc: Michael Ellerman <mpe@ellerman.id.au>
Cc: Heiko Carstens <heiko.carstens@de.ibm.com>
Cc: Martin Schwidefsky <schwidefsky@de.ibm.com>
Cc: Ingo Molnar <mingo@elte.hu>
Cc: "H. Peter Anvin" <hpa@zytor.com>
Cc: Thomas Gleixner <tglx@linutronix.de>
Signed-off-by: default avatarAndrew Morton <akpm@linux-foundation.org>
Signed-off-by: default avatarLinus Torvalds <torvalds@linux-foundation.org>
parent 8bb4e7a2
...@@ -1069,8 +1069,8 @@ static inline pte_t ptep_get_and_clear(struct mm_struct *mm, ...@@ -1069,8 +1069,8 @@ static inline pte_t ptep_get_and_clear(struct mm_struct *mm,
} }
#define __HAVE_ARCH_PTEP_MODIFY_PROT_TRANSACTION #define __HAVE_ARCH_PTEP_MODIFY_PROT_TRANSACTION
pte_t ptep_modify_prot_start(struct mm_struct *, unsigned long, pte_t *); pte_t ptep_modify_prot_start(struct vm_area_struct *, unsigned long, pte_t *);
void ptep_modify_prot_commit(struct mm_struct *, unsigned long, pte_t *, pte_t); void ptep_modify_prot_commit(struct vm_area_struct *, unsigned long, pte_t *, pte_t);
#define __HAVE_ARCH_PTEP_CLEAR_FLUSH #define __HAVE_ARCH_PTEP_CLEAR_FLUSH
static inline pte_t ptep_clear_flush(struct vm_area_struct *vma, static inline pte_t ptep_clear_flush(struct vm_area_struct *vma,
......
...@@ -301,12 +301,13 @@ pte_t ptep_xchg_lazy(struct mm_struct *mm, unsigned long addr, ...@@ -301,12 +301,13 @@ pte_t ptep_xchg_lazy(struct mm_struct *mm, unsigned long addr,
} }
EXPORT_SYMBOL(ptep_xchg_lazy); EXPORT_SYMBOL(ptep_xchg_lazy);
pte_t ptep_modify_prot_start(struct mm_struct *mm, unsigned long addr, pte_t ptep_modify_prot_start(struct vm_area_struct *vma, unsigned long addr,
pte_t *ptep) pte_t *ptep)
{ {
pgste_t pgste; pgste_t pgste;
pte_t old; pte_t old;
int nodat; int nodat;
struct mm_struct *mm = vma->vm_mm;
preempt_disable(); preempt_disable();
pgste = ptep_xchg_start(mm, addr, ptep); pgste = ptep_xchg_start(mm, addr, ptep);
...@@ -319,10 +320,11 @@ pte_t ptep_modify_prot_start(struct mm_struct *mm, unsigned long addr, ...@@ -319,10 +320,11 @@ pte_t ptep_modify_prot_start(struct mm_struct *mm, unsigned long addr,
return old; return old;
} }
void ptep_modify_prot_commit(struct mm_struct *mm, unsigned long addr, void ptep_modify_prot_commit(struct vm_area_struct *vma, unsigned long addr,
pte_t *ptep, pte_t pte) pte_t *ptep, pte_t pte)
{ {
pgste_t pgste; pgste_t pgste;
struct mm_struct *mm = vma->vm_mm;
if (!MACHINE_HAS_NX) if (!MACHINE_HAS_NX)
pte_val(pte) &= ~_PAGE_NOEXEC; pte_val(pte) &= ~_PAGE_NOEXEC;
......
...@@ -422,25 +422,26 @@ static inline pgdval_t pgd_val(pgd_t pgd) ...@@ -422,25 +422,26 @@ static inline pgdval_t pgd_val(pgd_t pgd)
} }
#define __HAVE_ARCH_PTEP_MODIFY_PROT_TRANSACTION #define __HAVE_ARCH_PTEP_MODIFY_PROT_TRANSACTION
static inline pte_t ptep_modify_prot_start(struct mm_struct *mm, unsigned long addr, static inline pte_t ptep_modify_prot_start(struct vm_area_struct *vma, unsigned long addr,
pte_t *ptep) pte_t *ptep)
{ {
pteval_t ret; pteval_t ret;
ret = PVOP_CALL3(pteval_t, mmu.ptep_modify_prot_start, mm, addr, ptep); ret = PVOP_CALL3(pteval_t, mmu.ptep_modify_prot_start, vma, addr, ptep);
return (pte_t) { .pte = ret }; return (pte_t) { .pte = ret };
} }
static inline void ptep_modify_prot_commit(struct mm_struct *mm, unsigned long addr, static inline void ptep_modify_prot_commit(struct vm_area_struct *vma, unsigned long addr,
pte_t *ptep, pte_t pte) pte_t *ptep, pte_t pte)
{ {
if (sizeof(pteval_t) > sizeof(long)) if (sizeof(pteval_t) > sizeof(long))
/* 5 arg words */ /* 5 arg words */
pv_ops.mmu.ptep_modify_prot_commit(mm, addr, ptep, pte); pv_ops.mmu.ptep_modify_prot_commit(vma, addr, ptep, pte);
else else
PVOP_VCALL4(mmu.ptep_modify_prot_commit, PVOP_VCALL4(mmu.ptep_modify_prot_commit,
mm, addr, ptep, pte.pte); vma, addr, ptep, pte.pte);
} }
static inline void set_pte(pte_t *ptep, pte_t pte) static inline void set_pte(pte_t *ptep, pte_t pte)
......
...@@ -55,6 +55,7 @@ struct task_struct; ...@@ -55,6 +55,7 @@ struct task_struct;
struct cpumask; struct cpumask;
struct flush_tlb_info; struct flush_tlb_info;
struct mmu_gather; struct mmu_gather;
struct vm_area_struct;
/* /*
* Wrapper type for pointers to code which uses the non-standard * Wrapper type for pointers to code which uses the non-standard
...@@ -254,9 +255,9 @@ struct pv_mmu_ops { ...@@ -254,9 +255,9 @@ struct pv_mmu_ops {
pte_t *ptep, pte_t pteval); pte_t *ptep, pte_t pteval);
void (*set_pmd)(pmd_t *pmdp, pmd_t pmdval); void (*set_pmd)(pmd_t *pmdp, pmd_t pmdval);
pte_t (*ptep_modify_prot_start)(struct mm_struct *mm, unsigned long addr, pte_t (*ptep_modify_prot_start)(struct vm_area_struct *vma, unsigned long addr,
pte_t *ptep); pte_t *ptep);
void (*ptep_modify_prot_commit)(struct mm_struct *mm, unsigned long addr, void (*ptep_modify_prot_commit)(struct vm_area_struct *vma, unsigned long addr,
pte_t *ptep, pte_t pte); pte_t *ptep, pte_t pte);
struct paravirt_callee_save pte_val; struct paravirt_callee_save pte_val;
......
...@@ -17,8 +17,8 @@ bool __set_phys_to_machine(unsigned long pfn, unsigned long mfn); ...@@ -17,8 +17,8 @@ bool __set_phys_to_machine(unsigned long pfn, unsigned long mfn);
void set_pte_mfn(unsigned long vaddr, unsigned long pfn, pgprot_t flags); void set_pte_mfn(unsigned long vaddr, unsigned long pfn, pgprot_t flags);
pte_t xen_ptep_modify_prot_start(struct mm_struct *mm, unsigned long addr, pte_t *ptep); pte_t xen_ptep_modify_prot_start(struct vm_area_struct *vma, unsigned long addr, pte_t *ptep);
void xen_ptep_modify_prot_commit(struct mm_struct *mm, unsigned long addr, void xen_ptep_modify_prot_commit(struct vm_area_struct *vma, unsigned long addr,
pte_t *ptep, pte_t pte); pte_t *ptep, pte_t pte);
unsigned long xen_read_cr2_direct(void); unsigned long xen_read_cr2_direct(void);
......
...@@ -306,20 +306,20 @@ static void xen_set_pte_at(struct mm_struct *mm, unsigned long addr, ...@@ -306,20 +306,20 @@ static void xen_set_pte_at(struct mm_struct *mm, unsigned long addr,
__xen_set_pte(ptep, pteval); __xen_set_pte(ptep, pteval);
} }
pte_t xen_ptep_modify_prot_start(struct mm_struct *mm, pte_t xen_ptep_modify_prot_start(struct vm_area_struct *vma,
unsigned long addr, pte_t *ptep) unsigned long addr, pte_t *ptep)
{ {
/* Just return the pte as-is. We preserve the bits on commit */ /* Just return the pte as-is. We preserve the bits on commit */
trace_xen_mmu_ptep_modify_prot_start(mm, addr, ptep, *ptep); trace_xen_mmu_ptep_modify_prot_start(vma->vm_mm, addr, ptep, *ptep);
return *ptep; return *ptep;
} }
void xen_ptep_modify_prot_commit(struct mm_struct *mm, unsigned long addr, void xen_ptep_modify_prot_commit(struct vm_area_struct *vma, unsigned long addr,
pte_t *ptep, pte_t pte) pte_t *ptep, pte_t pte)
{ {
struct mmu_update u; struct mmu_update u;
trace_xen_mmu_ptep_modify_prot_commit(mm, addr, ptep, pte); trace_xen_mmu_ptep_modify_prot_commit(vma->vm_mm, addr, ptep, pte);
xen_mc_batch(); xen_mc_batch();
u.ptr = virt_to_machine(ptep).maddr | MMU_PT_UPDATE_PRESERVE_AD; u.ptr = virt_to_machine(ptep).maddr | MMU_PT_UPDATE_PRESERVE_AD;
......
...@@ -948,10 +948,10 @@ static inline void clear_soft_dirty(struct vm_area_struct *vma, ...@@ -948,10 +948,10 @@ static inline void clear_soft_dirty(struct vm_area_struct *vma,
pte_t ptent = *pte; pte_t ptent = *pte;
if (pte_present(ptent)) { if (pte_present(ptent)) {
ptent = ptep_modify_prot_start(vma->vm_mm, addr, pte); ptent = ptep_modify_prot_start(vma, addr, pte);
ptent = pte_wrprotect(ptent); ptent = pte_wrprotect(ptent);
ptent = pte_clear_soft_dirty(ptent); ptent = pte_clear_soft_dirty(ptent);
ptep_modify_prot_commit(vma->vm_mm, addr, pte, ptent); ptep_modify_prot_commit(vma, addr, pte, ptent);
} else if (is_swap_pte(ptent)) { } else if (is_swap_pte(ptent)) {
ptent = pte_swp_clear_soft_dirty(ptent); ptent = pte_swp_clear_soft_dirty(ptent);
set_pte_at(vma->vm_mm, addr, pte, ptent); set_pte_at(vma->vm_mm, addr, pte, ptent);
......
...@@ -606,7 +606,7 @@ static inline int pmd_none_or_clear_bad(pmd_t *pmd) ...@@ -606,7 +606,7 @@ static inline int pmd_none_or_clear_bad(pmd_t *pmd)
return 0; return 0;
} }
static inline pte_t __ptep_modify_prot_start(struct mm_struct *mm, static inline pte_t __ptep_modify_prot_start(struct vm_area_struct *vma,
unsigned long addr, unsigned long addr,
pte_t *ptep) pte_t *ptep)
{ {
...@@ -615,10 +615,10 @@ static inline pte_t __ptep_modify_prot_start(struct mm_struct *mm, ...@@ -615,10 +615,10 @@ static inline pte_t __ptep_modify_prot_start(struct mm_struct *mm,
* non-present, preventing the hardware from asynchronously * non-present, preventing the hardware from asynchronously
* updating it. * updating it.
*/ */
return ptep_get_and_clear(mm, addr, ptep); return ptep_get_and_clear(vma->vm_mm, addr, ptep);
} }
static inline void __ptep_modify_prot_commit(struct mm_struct *mm, static inline void __ptep_modify_prot_commit(struct vm_area_struct *vma,
unsigned long addr, unsigned long addr,
pte_t *ptep, pte_t pte) pte_t *ptep, pte_t pte)
{ {
...@@ -626,7 +626,7 @@ static inline void __ptep_modify_prot_commit(struct mm_struct *mm, ...@@ -626,7 +626,7 @@ static inline void __ptep_modify_prot_commit(struct mm_struct *mm,
* The pte is non-present, so there's no hardware state to * The pte is non-present, so there's no hardware state to
* preserve. * preserve.
*/ */
set_pte_at(mm, addr, ptep, pte); set_pte_at(vma->vm_mm, addr, ptep, pte);
} }
#ifndef __HAVE_ARCH_PTEP_MODIFY_PROT_TRANSACTION #ifndef __HAVE_ARCH_PTEP_MODIFY_PROT_TRANSACTION
...@@ -644,22 +644,22 @@ static inline void __ptep_modify_prot_commit(struct mm_struct *mm, ...@@ -644,22 +644,22 @@ static inline void __ptep_modify_prot_commit(struct mm_struct *mm,
* queue the update to be done at some later time. The update must be * queue the update to be done at some later time. The update must be
* actually committed before the pte lock is released, however. * actually committed before the pte lock is released, however.
*/ */
static inline pte_t ptep_modify_prot_start(struct mm_struct *mm, static inline pte_t ptep_modify_prot_start(struct vm_area_struct *vma,
unsigned long addr, unsigned long addr,
pte_t *ptep) pte_t *ptep)
{ {
return __ptep_modify_prot_start(mm, addr, ptep); return __ptep_modify_prot_start(vma, addr, ptep);
} }
/* /*
* Commit an update to a pte, leaving any hardware-controlled bits in * Commit an update to a pte, leaving any hardware-controlled bits in
* the PTE unmodified. * the PTE unmodified.
*/ */
static inline void ptep_modify_prot_commit(struct mm_struct *mm, static inline void ptep_modify_prot_commit(struct vm_area_struct *vma,
unsigned long addr, unsigned long addr,
pte_t *ptep, pte_t pte) pte_t *ptep, pte_t pte)
{ {
__ptep_modify_prot_commit(mm, addr, ptep, pte); __ptep_modify_prot_commit(vma, addr, ptep, pte);
} }
#endif /* __HAVE_ARCH_PTEP_MODIFY_PROT_TRANSACTION */ #endif /* __HAVE_ARCH_PTEP_MODIFY_PROT_TRANSACTION */
#endif /* CONFIG_MMU */ #endif /* CONFIG_MMU */
......
...@@ -3619,12 +3619,12 @@ static vm_fault_t do_numa_page(struct vm_fault *vmf) ...@@ -3619,12 +3619,12 @@ static vm_fault_t do_numa_page(struct vm_fault *vmf)
* Make it present again, Depending on how arch implementes non * Make it present again, Depending on how arch implementes non
* accessible ptes, some can allow access by kernel mode. * accessible ptes, some can allow access by kernel mode.
*/ */
pte = ptep_modify_prot_start(vma->vm_mm, vmf->address, vmf->pte); pte = ptep_modify_prot_start(vma, vmf->address, vmf->pte);
pte = pte_modify(pte, vma->vm_page_prot); pte = pte_modify(pte, vma->vm_page_prot);
pte = pte_mkyoung(pte); pte = pte_mkyoung(pte);
if (was_writable) if (was_writable)
pte = pte_mkwrite(pte); pte = pte_mkwrite(pte);
ptep_modify_prot_commit(vma->vm_mm, vmf->address, vmf->pte, pte); ptep_modify_prot_commit(vma, vmf->address, vmf->pte, pte);
update_mmu_cache(vma, vmf->address, vmf->pte); update_mmu_cache(vma, vmf->address, vmf->pte);
page = vm_normal_page(vma, vmf->address, pte); page = vm_normal_page(vma, vmf->address, pte);
......
...@@ -110,7 +110,7 @@ static unsigned long change_pte_range(struct vm_area_struct *vma, pmd_t *pmd, ...@@ -110,7 +110,7 @@ static unsigned long change_pte_range(struct vm_area_struct *vma, pmd_t *pmd,
continue; continue;
} }
ptent = ptep_modify_prot_start(mm, addr, pte); ptent = ptep_modify_prot_start(vma, addr, pte);
ptent = pte_modify(ptent, newprot); ptent = pte_modify(ptent, newprot);
if (preserve_write) if (preserve_write)
ptent = pte_mk_savedwrite(ptent); ptent = pte_mk_savedwrite(ptent);
...@@ -121,7 +121,7 @@ static unsigned long change_pte_range(struct vm_area_struct *vma, pmd_t *pmd, ...@@ -121,7 +121,7 @@ static unsigned long change_pte_range(struct vm_area_struct *vma, pmd_t *pmd,
!(vma->vm_flags & VM_SOFTDIRTY))) { !(vma->vm_flags & VM_SOFTDIRTY))) {
ptent = pte_mkwrite(ptent); ptent = pte_mkwrite(ptent);
} }
ptep_modify_prot_commit(mm, addr, pte, ptent); ptep_modify_prot_commit(vma, addr, pte, ptent);
pages++; pages++;
} else if (IS_ENABLED(CONFIG_MIGRATION)) { } else if (IS_ENABLED(CONFIG_MIGRATION)) {
swp_entry_t entry = pte_to_swp_entry(oldpte); swp_entry_t entry = pte_to_swp_entry(oldpte);
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment