Commit a76ab5c1 authored by Ralf Baechle's avatar Ralf Baechle

[MIPS] MT: Fix bug in multithreaded kernels.

When GDB writes a breakpoint into address area of inferior process the
kernel needs to invalidate the modified memory in the inferior which
is done by calling flush_cache_page which in turns calls
r4k_flush_cache_page and local_r4k_flush_cache_page for VSMP or SMTC
kernel via r4k_on_each_cpu().

As the VSMP and SMTC SMP kernels for 34K are running on a single shared
caches it is possible to get away without interprocessor function calls.
This optimization is implemented in r4k_on_each_cpu, so
local_r4k_flush_cache_page is only ever called on the local CPU.

This is where the following code in local_r4k_flush_cache_page() strikes:

        /*
         * If ownes no valid ASID yet, cannot possibly have gotten
         * this page into the cache.
         */
        if (cpu_context(smp_processor_id(), mm) == 0)
                return;

On VSMP and SMTC had a function of cpu_context() for each CPU(TC).

So in case another CPU than the CPU executing local_r4k_cache_flush_page
has not accessed the mm but one of the other CPUs has there may be data
to be flushed in the cache yet local_r4k_cache_flush_page will falsely
return leaving the I-cache inconsistent for the breakpoint.

While the issue was discovered with GDB it also exists in
local_r4k_flush_cache_range() and local_r4k_flush_cache().

Fixed by introducing a new function has_valid_asid which on MT kernels
returns true if a mm is active on any processor in the system.

This is relativly expensive since for memory acccesses in that loop
cache misses have to be assumed but it seems the most viable solution
for 2.6.23 and older -stable kernels.
Signed-off-by: default avatarRalf Baechle <ralf@linux-mips.org>
parent a3706055
...@@ -345,11 +345,26 @@ static void r4k___flush_cache_all(void) ...@@ -345,11 +345,26 @@ static void r4k___flush_cache_all(void)
r4k_on_each_cpu(local_r4k___flush_cache_all, NULL, 1, 1); r4k_on_each_cpu(local_r4k___flush_cache_all, NULL, 1, 1);
} }
static inline int has_valid_asid(const struct mm_struct *mm)
{
#if defined(CONFIG_MIPS_MT_SMP) || defined(CONFIG_MIPS_MT_SMTC)
int i;
for_each_online_cpu(i)
if (cpu_context(i, mm))
return 1;
return 0;
#else
return cpu_context(smp_processor_id(), mm);
#endif
}
static inline void local_r4k_flush_cache_range(void * args) static inline void local_r4k_flush_cache_range(void * args)
{ {
struct vm_area_struct *vma = args; struct vm_area_struct *vma = args;
if (!(cpu_context(smp_processor_id(), vma->vm_mm))) if (!(has_valid_asid(vma->vm_mm)))
return; return;
r4k_blast_dcache(); r4k_blast_dcache();
...@@ -368,7 +383,7 @@ static inline void local_r4k_flush_cache_mm(void * args) ...@@ -368,7 +383,7 @@ static inline void local_r4k_flush_cache_mm(void * args)
{ {
struct mm_struct *mm = args; struct mm_struct *mm = args;
if (!cpu_context(smp_processor_id(), mm)) if (!has_valid_asid(mm))
return; return;
/* /*
...@@ -420,7 +435,7 @@ static inline void local_r4k_flush_cache_page(void *args) ...@@ -420,7 +435,7 @@ static inline void local_r4k_flush_cache_page(void *args)
* If ownes no valid ASID yet, cannot possibly have gotten * If ownes no valid ASID yet, cannot possibly have gotten
* this page into the cache. * this page into the cache.
*/ */
if (cpu_context(smp_processor_id(), mm) == 0) if (!has_valid_asid(mm))
return; return;
addr &= PAGE_MASK; addr &= PAGE_MASK;
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment