Commit 3b8c9f1c authored by Will Deacon's avatar Will Deacon

arm64: IPI each CPU after invalidating the I-cache for kernel mappings

When invalidating the instruction cache for a kernel mapping via
flush_icache_range(), it is also necessary to flush the pipeline for
other CPUs so that instructions fetched into the pipeline before the
I-cache invalidation are discarded. For example, if module 'foo' is
unloaded and then module 'bar' is loaded into the same area of memory,
a CPU could end up executing instructions from 'foo' when branching into
'bar' if these instructions were fetched into the pipeline before 'foo'
was unloaded.

Whilst this is highly unlikely to occur in practice, particularly as
any exception acts as a context-synchronizing operation, following the
letter of the architecture requires us to execute an ISB on each CPU
in order for the new instruction stream to be visible.
Acked-by: default avatarCatalin Marinas <catalin.marinas@arm.com>
Signed-off-by: default avatarWill Deacon <will.deacon@arm.com>
parent 7373fed2
...@@ -19,6 +19,7 @@ ...@@ -19,6 +19,7 @@
#ifndef __ASM_CACHEFLUSH_H #ifndef __ASM_CACHEFLUSH_H
#define __ASM_CACHEFLUSH_H #define __ASM_CACHEFLUSH_H
#include <linux/kgdb.h>
#include <linux/mm.h> #include <linux/mm.h>
/* /*
...@@ -71,7 +72,7 @@ ...@@ -71,7 +72,7 @@
* - kaddr - page address * - kaddr - page address
* - size - region size * - size - region size
*/ */
extern void flush_icache_range(unsigned long start, unsigned long end); extern void __flush_icache_range(unsigned long start, unsigned long end);
extern int invalidate_icache_range(unsigned long start, unsigned long end); extern int invalidate_icache_range(unsigned long start, unsigned long end);
extern void __flush_dcache_area(void *addr, size_t len); extern void __flush_dcache_area(void *addr, size_t len);
extern void __inval_dcache_area(void *addr, size_t len); extern void __inval_dcache_area(void *addr, size_t len);
...@@ -81,6 +82,30 @@ extern void __clean_dcache_area_pou(void *addr, size_t len); ...@@ -81,6 +82,30 @@ extern void __clean_dcache_area_pou(void *addr, size_t len);
extern long __flush_cache_user_range(unsigned long start, unsigned long end); extern long __flush_cache_user_range(unsigned long start, unsigned long end);
extern void sync_icache_aliases(void *kaddr, unsigned long len); extern void sync_icache_aliases(void *kaddr, unsigned long len);
static inline void flush_icache_range(unsigned long start, unsigned long end)
{
__flush_icache_range(start, end);
/*
* IPI all online CPUs so that they undergo a context synchronization
* event and are forced to refetch the new instructions.
*/
#ifdef CONFIG_KGDB
/*
* KGDB performs cache maintenance with interrupts disabled, so we
* will deadlock trying to IPI the secondary CPUs. In theory, we can
* set CACHE_FLUSH_IS_SAFE to 0 to avoid this known issue, but that
* just means that KGDB will elide the maintenance altogether! As it
* turns out, KGDB uses IPIs to round-up the secondary CPUs during
* the patching operation, so we don't need extra IPIs here anyway.
* In which case, add a KGDB-specific bodge and return early.
*/
if (kgdb_connected && irqs_disabled())
return;
#endif
kick_all_cpus_sync();
}
static inline void flush_cache_mm(struct mm_struct *mm) static inline void flush_cache_mm(struct mm_struct *mm)
{ {
} }
......
...@@ -107,7 +107,7 @@ static void __copy_hyp_vect_bpi(int slot, const char *hyp_vecs_start, ...@@ -107,7 +107,7 @@ static void __copy_hyp_vect_bpi(int slot, const char *hyp_vecs_start,
for (i = 0; i < SZ_2K; i += 0x80) for (i = 0; i < SZ_2K; i += 0x80)
memcpy(dst + i, hyp_vecs_start, hyp_vecs_end - hyp_vecs_start); memcpy(dst + i, hyp_vecs_start, hyp_vecs_end - hyp_vecs_start);
flush_icache_range((uintptr_t)dst, (uintptr_t)dst + SZ_2K); __flush_icache_range((uintptr_t)dst, (uintptr_t)dst + SZ_2K);
} }
static void __install_bp_hardening_cb(bp_hardening_cb_t fn, static void __install_bp_hardening_cb(bp_hardening_cb_t fn,
......
...@@ -216,7 +216,7 @@ int __kprobes aarch64_insn_patch_text_nosync(void *addr, u32 insn) ...@@ -216,7 +216,7 @@ int __kprobes aarch64_insn_patch_text_nosync(void *addr, u32 insn)
ret = aarch64_insn_write(tp, insn); ret = aarch64_insn_write(tp, insn);
if (ret == 0) if (ret == 0)
flush_icache_range((uintptr_t)tp, __flush_icache_range((uintptr_t)tp,
(uintptr_t)tp + AARCH64_INSN_SIZE); (uintptr_t)tp + AARCH64_INSN_SIZE);
return ret; return ret;
...@@ -283,18 +283,8 @@ int __kprobes aarch64_insn_patch_text(void *addrs[], u32 insns[], int cnt) ...@@ -283,18 +283,8 @@ int __kprobes aarch64_insn_patch_text(void *addrs[], u32 insns[], int cnt)
if (ret) if (ret)
return ret; return ret;
if (aarch64_insn_hotpatch_safe(insn, insns[0])) { if (aarch64_insn_hotpatch_safe(insn, insns[0]))
/* return aarch64_insn_patch_text_nosync(addrs[0], insns[0]);
* ARMv8 architecture doesn't guarantee all CPUs see
* the new instruction after returning from function
* aarch64_insn_patch_text_nosync(). So send IPIs to
* all other CPUs to achieve instruction
* synchronization.
*/
ret = aarch64_insn_patch_text_nosync(addrs[0], insns[0]);
kick_all_cpus_sync();
return ret;
}
} }
return aarch64_insn_patch_text_sync(addrs, insns, cnt); return aarch64_insn_patch_text_sync(addrs, insns, cnt);
......
...@@ -35,7 +35,7 @@ ...@@ -35,7 +35,7 @@
* - start - virtual start address of region * - start - virtual start address of region
* - end - virtual end address of region * - end - virtual end address of region
*/ */
ENTRY(flush_icache_range) ENTRY(__flush_icache_range)
/* FALLTHROUGH */ /* FALLTHROUGH */
/* /*
...@@ -77,7 +77,7 @@ alternative_else_nop_endif ...@@ -77,7 +77,7 @@ alternative_else_nop_endif
9: 9:
mov x0, #-EFAULT mov x0, #-EFAULT
b 1b b 1b
ENDPROC(flush_icache_range) ENDPROC(__flush_icache_range)
ENDPROC(__flush_cache_user_range) ENDPROC(__flush_cache_user_range)
/* /*
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment