Commit e059853d authored by Catalin Marinas's avatar Catalin Marinas Committed by Marc Zyngier

arm64: mte: Fix/clarify the PG_mte_tagged semantics

Currently the PG_mte_tagged page flag mostly means the page contains
valid tags and it should be set after the tags have been cleared or
restored. However, in mte_sync_tags() it is set before setting the tags
to avoid, in theory, a race with concurrent mprotect(PROT_MTE) for
shared pages. However, a concurrent mprotect(PROT_MTE) with a copy on
write in another thread can cause the new page to have stale tags.
Similarly, tag reading via ptrace() can read stale tags if the
PG_mte_tagged flag is set before actually clearing/restoring the tags.

Fix the PG_mte_tagged semantics so that it is only set after the tags
have been cleared or restored. This is safe for swap restoring into a
MAP_SHARED or CoW page since the core code takes the page lock. Add two
functions to test and set the PG_mte_tagged flag with acquire and
release semantics. The downside is that concurrent mprotect(PROT_MTE) on
a MAP_SHARED page may cause tag loss. This is already the case for KVM
guests if a VMM changes the page protection while the guest triggers a
user_mem_abort().
Signed-off-by: default avatarCatalin Marinas <catalin.marinas@arm.com>
[pcc@google.com: fix build with CONFIG_ARM64_MTE disabled]
Signed-off-by: default avatarPeter Collingbourne <pcc@google.com>
Reviewed-by: default avatarCornelia Huck <cohuck@redhat.com>
Reviewed-by: default avatarSteven Price <steven.price@arm.com>
Cc: Will Deacon <will@kernel.org>
Cc: Marc Zyngier <maz@kernel.org>
Cc: Peter Collingbourne <pcc@google.com>
Signed-off-by: default avatarMarc Zyngier <maz@kernel.org>
Link: https://lore.kernel.org/r/20221104011041.290951-3-pcc@google.com
parent b0284cd2
...@@ -37,6 +37,29 @@ void mte_free_tag_storage(char *storage); ...@@ -37,6 +37,29 @@ void mte_free_tag_storage(char *storage);
/* track which pages have valid allocation tags */ /* track which pages have valid allocation tags */
#define PG_mte_tagged PG_arch_2 #define PG_mte_tagged PG_arch_2
static inline void set_page_mte_tagged(struct page *page)
{
/*
* Ensure that the tags written prior to this function are visible
* before the page flags update.
*/
smp_wmb();
set_bit(PG_mte_tagged, &page->flags);
}
static inline bool page_mte_tagged(struct page *page)
{
bool ret = test_bit(PG_mte_tagged, &page->flags);
/*
* If the page is tagged, ensure ordering with a likely subsequent
* read of the tags.
*/
if (ret)
smp_rmb();
return ret;
}
void mte_zero_clear_page_tags(void *addr); void mte_zero_clear_page_tags(void *addr);
void mte_sync_tags(pte_t old_pte, pte_t pte); void mte_sync_tags(pte_t old_pte, pte_t pte);
void mte_copy_page_tags(void *kto, const void *kfrom); void mte_copy_page_tags(void *kto, const void *kfrom);
...@@ -56,6 +79,13 @@ size_t mte_probe_user_range(const char __user *uaddr, size_t size); ...@@ -56,6 +79,13 @@ size_t mte_probe_user_range(const char __user *uaddr, size_t size);
/* unused if !CONFIG_ARM64_MTE, silence the compiler */ /* unused if !CONFIG_ARM64_MTE, silence the compiler */
#define PG_mte_tagged 0 #define PG_mte_tagged 0
static inline void set_page_mte_tagged(struct page *page)
{
}
static inline bool page_mte_tagged(struct page *page)
{
return false;
}
static inline void mte_zero_clear_page_tags(void *addr) static inline void mte_zero_clear_page_tags(void *addr)
{ {
} }
......
...@@ -1050,7 +1050,7 @@ static inline void arch_swap_invalidate_area(int type) ...@@ -1050,7 +1050,7 @@ static inline void arch_swap_invalidate_area(int type)
static inline void arch_swap_restore(swp_entry_t entry, struct folio *folio) static inline void arch_swap_restore(swp_entry_t entry, struct folio *folio)
{ {
if (system_supports_mte() && mte_restore_tags(entry, &folio->page)) if (system_supports_mte() && mte_restore_tags(entry, &folio->page))
set_bit(PG_mte_tagged, &folio->flags); set_page_mte_tagged(&folio->page);
} }
#endif /* CONFIG_ARM64_MTE */ #endif /* CONFIG_ARM64_MTE */
......
...@@ -2050,8 +2050,10 @@ static void cpu_enable_mte(struct arm64_cpu_capabilities const *cap) ...@@ -2050,8 +2050,10 @@ static void cpu_enable_mte(struct arm64_cpu_capabilities const *cap)
* Clear the tags in the zero page. This needs to be done via the * Clear the tags in the zero page. This needs to be done via the
* linear map which has the Tagged attribute. * linear map which has the Tagged attribute.
*/ */
if (!test_and_set_bit(PG_mte_tagged, &ZERO_PAGE(0)->flags)) if (!page_mte_tagged(ZERO_PAGE(0))) {
mte_clear_page_tags(lm_alias(empty_zero_page)); mte_clear_page_tags(lm_alias(empty_zero_page));
set_page_mte_tagged(ZERO_PAGE(0));
}
kasan_init_hw_tags_cpu(); kasan_init_hw_tags_cpu();
} }
......
...@@ -47,7 +47,7 @@ static int mte_dump_tag_range(struct coredump_params *cprm, ...@@ -47,7 +47,7 @@ static int mte_dump_tag_range(struct coredump_params *cprm,
* Pages mapped in user space as !pte_access_permitted() (e.g. * Pages mapped in user space as !pte_access_permitted() (e.g.
* PROT_EXEC only) may not have the PG_mte_tagged flag set. * PROT_EXEC only) may not have the PG_mte_tagged flag set.
*/ */
if (!test_bit(PG_mte_tagged, &page->flags)) { if (!page_mte_tagged(page)) {
put_page(page); put_page(page);
dump_skip(cprm, MTE_PAGE_TAG_STORAGE); dump_skip(cprm, MTE_PAGE_TAG_STORAGE);
continue; continue;
......
...@@ -271,7 +271,7 @@ static int swsusp_mte_save_tags(void) ...@@ -271,7 +271,7 @@ static int swsusp_mte_save_tags(void)
if (!page) if (!page)
continue; continue;
if (!test_bit(PG_mte_tagged, &page->flags)) if (!page_mte_tagged(page))
continue; continue;
ret = save_tags(page, pfn); ret = save_tags(page, pfn);
......
...@@ -41,8 +41,10 @@ static void mte_sync_page_tags(struct page *page, pte_t old_pte, ...@@ -41,8 +41,10 @@ static void mte_sync_page_tags(struct page *page, pte_t old_pte,
if (check_swap && is_swap_pte(old_pte)) { if (check_swap && is_swap_pte(old_pte)) {
swp_entry_t entry = pte_to_swp_entry(old_pte); swp_entry_t entry = pte_to_swp_entry(old_pte);
if (!non_swap_entry(entry) && mte_restore_tags(entry, page)) if (!non_swap_entry(entry) && mte_restore_tags(entry, page)) {
set_page_mte_tagged(page);
return; return;
}
} }
if (!pte_is_tagged) if (!pte_is_tagged)
...@@ -52,8 +54,10 @@ static void mte_sync_page_tags(struct page *page, pte_t old_pte, ...@@ -52,8 +54,10 @@ static void mte_sync_page_tags(struct page *page, pte_t old_pte,
* Test PG_mte_tagged again in case it was racing with another * Test PG_mte_tagged again in case it was racing with another
* set_pte_at(). * set_pte_at().
*/ */
if (!test_and_set_bit(PG_mte_tagged, &page->flags)) if (!page_mte_tagged(page)) {
mte_clear_page_tags(page_address(page)); mte_clear_page_tags(page_address(page));
set_page_mte_tagged(page);
}
} }
void mte_sync_tags(pte_t old_pte, pte_t pte) void mte_sync_tags(pte_t old_pte, pte_t pte)
...@@ -69,9 +73,11 @@ void mte_sync_tags(pte_t old_pte, pte_t pte) ...@@ -69,9 +73,11 @@ void mte_sync_tags(pte_t old_pte, pte_t pte)
/* if PG_mte_tagged is set, tags have already been initialised */ /* if PG_mte_tagged is set, tags have already been initialised */
for (i = 0; i < nr_pages; i++, page++) { for (i = 0; i < nr_pages; i++, page++) {
if (!test_bit(PG_mte_tagged, &page->flags)) if (!page_mte_tagged(page)) {
mte_sync_page_tags(page, old_pte, check_swap, mte_sync_page_tags(page, old_pte, check_swap,
pte_is_tagged); pte_is_tagged);
set_page_mte_tagged(page);
}
} }
/* ensure the tags are visible before the PTE is set */ /* ensure the tags are visible before the PTE is set */
...@@ -96,8 +102,7 @@ int memcmp_pages(struct page *page1, struct page *page2) ...@@ -96,8 +102,7 @@ int memcmp_pages(struct page *page1, struct page *page2)
* pages is tagged, set_pte_at() may zero or change the tags of the * pages is tagged, set_pte_at() may zero or change the tags of the
* other page via mte_sync_tags(). * other page via mte_sync_tags().
*/ */
if (test_bit(PG_mte_tagged, &page1->flags) || if (page_mte_tagged(page1) || page_mte_tagged(page2))
test_bit(PG_mte_tagged, &page2->flags))
return addr1 != addr2; return addr1 != addr2;
return ret; return ret;
...@@ -454,7 +459,7 @@ static int __access_remote_tags(struct mm_struct *mm, unsigned long addr, ...@@ -454,7 +459,7 @@ static int __access_remote_tags(struct mm_struct *mm, unsigned long addr,
put_page(page); put_page(page);
break; break;
} }
WARN_ON_ONCE(!test_bit(PG_mte_tagged, &page->flags)); WARN_ON_ONCE(!page_mte_tagged(page));
/* limit access to the end of the page */ /* limit access to the end of the page */
offset = offset_in_page(addr); offset = offset_in_page(addr);
......
...@@ -1059,7 +1059,7 @@ long kvm_vm_ioctl_mte_copy_tags(struct kvm *kvm, ...@@ -1059,7 +1059,7 @@ long kvm_vm_ioctl_mte_copy_tags(struct kvm *kvm,
maddr = page_address(page); maddr = page_address(page);
if (!write) { if (!write) {
if (test_bit(PG_mte_tagged, &page->flags)) if (page_mte_tagged(page))
num_tags = mte_copy_tags_to_user(tags, maddr, num_tags = mte_copy_tags_to_user(tags, maddr,
MTE_GRANULES_PER_PAGE); MTE_GRANULES_PER_PAGE);
else else
...@@ -1076,7 +1076,7 @@ long kvm_vm_ioctl_mte_copy_tags(struct kvm *kvm, ...@@ -1076,7 +1076,7 @@ long kvm_vm_ioctl_mte_copy_tags(struct kvm *kvm,
* completed fully * completed fully
*/ */
if (num_tags == MTE_GRANULES_PER_PAGE) if (num_tags == MTE_GRANULES_PER_PAGE)
set_bit(PG_mte_tagged, &page->flags); set_page_mte_tagged(page);
kvm_release_pfn_dirty(pfn); kvm_release_pfn_dirty(pfn);
} }
......
...@@ -1110,9 +1110,9 @@ static int sanitise_mte_tags(struct kvm *kvm, kvm_pfn_t pfn, ...@@ -1110,9 +1110,9 @@ static int sanitise_mte_tags(struct kvm *kvm, kvm_pfn_t pfn,
return -EFAULT; return -EFAULT;
for (i = 0; i < nr_pages; i++, page++) { for (i = 0; i < nr_pages; i++, page++) {
if (!test_bit(PG_mte_tagged, &page->flags)) { if (!page_mte_tagged(page)) {
mte_clear_page_tags(page_address(page)); mte_clear_page_tags(page_address(page));
set_bit(PG_mte_tagged, &page->flags); set_page_mte_tagged(page);
} }
} }
......
...@@ -21,9 +21,10 @@ void copy_highpage(struct page *to, struct page *from) ...@@ -21,9 +21,10 @@ void copy_highpage(struct page *to, struct page *from)
copy_page(kto, kfrom); copy_page(kto, kfrom);
if (system_supports_mte() && test_bit(PG_mte_tagged, &from->flags)) { if (system_supports_mte() && page_mte_tagged(from)) {
set_bit(PG_mte_tagged, &to->flags); page_kasan_tag_reset(to);
mte_copy_page_tags(kto, kfrom); mte_copy_page_tags(kto, kfrom);
set_page_mte_tagged(to);
} }
} }
EXPORT_SYMBOL(copy_highpage); EXPORT_SYMBOL(copy_highpage);
......
...@@ -934,5 +934,5 @@ struct page *alloc_zeroed_user_highpage_movable(struct vm_area_struct *vma, ...@@ -934,5 +934,5 @@ struct page *alloc_zeroed_user_highpage_movable(struct vm_area_struct *vma,
void tag_clear_highpage(struct page *page) void tag_clear_highpage(struct page *page)
{ {
mte_zero_clear_page_tags(page_address(page)); mte_zero_clear_page_tags(page_address(page));
set_bit(PG_mte_tagged, &page->flags); set_page_mte_tagged(page);
} }
...@@ -24,7 +24,7 @@ int mte_save_tags(struct page *page) ...@@ -24,7 +24,7 @@ int mte_save_tags(struct page *page)
{ {
void *tag_storage, *ret; void *tag_storage, *ret;
if (!test_bit(PG_mte_tagged, &page->flags)) if (!page_mte_tagged(page))
return 0; return 0;
tag_storage = mte_allocate_tag_storage(); tag_storage = mte_allocate_tag_storage();
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment