Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Support
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
L
linux
Project overview
Project overview
Details
Activity
Releases
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Issues
0
Issues
0
List
Boards
Labels
Milestones
Merge Requests
0
Merge Requests
0
Analytics
Analytics
Repository
Value Stream
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Create a new issue
Commits
Issue Boards
Open sidebar
nexedi
linux
Commits
f5cfa980
Commit
f5cfa980
authored
Oct 01, 2005
by
Ralf Baechle
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
Use R4000 TLB routines for SB1 also.
Signed-off-by:
Ralf Baechle
<
ralf@linux-mips.org
>
parent
c5c96e13
Changes
3
Show whitespace changes
Inline
Side-by-side
Showing
3 changed files
with
5 additions
and
388 deletions
+5
-388
arch/mips/mm/Makefile
arch/mips/mm/Makefile
+1
-1
arch/mips/mm/tlb-sb1.c
arch/mips/mm/tlb-sb1.c
+0
-385
include/asm-mips/hazards.h
include/asm-mips/hazards.h
+4
-2
No files found.
arch/mips/mm/Makefile
View file @
f5cfa980
...
...
@@ -22,7 +22,7 @@ obj-$(CONFIG_CPU_R8000) += c-r4k.o cex-gen.o pg-r4k.o tlb-r8k.o
obj-$(CONFIG_CPU_RM7000)
+=
c-r4k.o cex-gen.o pg-r4k.o tlb-r4k.o
obj-$(CONFIG_CPU_RM9000)
+=
c-r4k.o cex-gen.o pg-r4k.o tlb-r4k.o
obj-$(CONFIG_CPU_SB1)
+=
c-sb1.o cerr-sb1.o cex-sb1.o pg-sb1.o
\
tlb-
sb1
.o
tlb-
r4k
.o
obj-$(CONFIG_CPU_TX39XX)
+=
c-tx39.o pg-r4k.o tlb-r3k.o
obj-$(CONFIG_CPU_TX49XX)
+=
c-r4k.o cex-gen.o pg-r4k.o tlb-r4k.o
obj-$(CONFIG_CPU_VR41XX)
+=
c-r4k.o cex-gen.o pg-r4k.o tlb-r4k.o
...
...
arch/mips/mm/tlb-sb1.c
deleted
100644 → 0
View file @
c5c96e13
/*
* Copyright (C) 1996 David S. Miller (dm@engr.sgi.com)
* Copyright (C) 1997, 2001 Ralf Baechle (ralf@gnu.org)
* Copyright (C) 2000, 2001, 2002, 2003 Broadcom Corporation
*
* This program is free software; you can redistribute it and/or
* modify it under the terms of the GNU General Public License
* as published by the Free Software Foundation; either version 2
* of the License, or (at your option) any later version.
*
* This program is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
* GNU General Public License for more details.
*
* You should have received a copy of the GNU General Public License
* along with this program; if not, write to the Free Software
* Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA.
*/
#include <linux/init.h>
#include <asm/mmu_context.h>
#include <asm/bootinfo.h>
#include <asm/cpu.h>
extern
void
build_tlb_refill_handler
(
void
);
#define UNIQUE_ENTRYHI(idx) (CKSEG0 + ((idx) << (PAGE_SHIFT + 1)))
/* Dump the current entry* and pagemask registers */
static
inline
void
dump_cur_tlb_regs
(
void
)
{
unsigned
int
entryhihi
,
entryhilo
,
entrylo0hi
,
entrylo0lo
,
entrylo1hi
;
unsigned
int
entrylo1lo
,
pagemask
;
__asm__
__volatile__
(
".set push
\n
"
".set noreorder
\n
"
".set mips64
\n
"
".set noat
\n
"
" tlbr
\n
"
" dmfc0 $1, $10
\n
"
" dsrl32 %0, $1, 0
\n
"
" sll %1, $1, 0
\n
"
" dmfc0 $1, $2
\n
"
" dsrl32 %2, $1, 0
\n
"
" sll %3, $1, 0
\n
"
" dmfc0 $1, $3
\n
"
" dsrl32 %4, $1, 0
\n
"
" sll %5, $1, 0
\n
"
" mfc0 %6, $5
\n
"
".set pop
\n
"
:
"=r"
(
entryhihi
),
"=r"
(
entryhilo
),
"=r"
(
entrylo0hi
),
"=r"
(
entrylo0lo
),
"=r"
(
entrylo1hi
),
"=r"
(
entrylo1lo
),
"=r"
(
pagemask
));
printk
(
"%08X%08X %08X%08X %08X%08X %08X"
,
entryhihi
,
entryhilo
,
entrylo0hi
,
entrylo0lo
,
entrylo1hi
,
entrylo1lo
,
pagemask
);
}
void
sb1_dump_tlb
(
void
)
{
unsigned
long
old_ctx
;
unsigned
long
flags
;
int
entry
;
local_irq_save
(
flags
);
old_ctx
=
read_c0_entryhi
();
printk
(
"Current TLB registers state:
\n
"
" EntryHi EntryLo0 EntryLo1 PageMask Index
\n
"
"--------------------------------------------------------------------
\n
"
);
dump_cur_tlb_regs
();
printk
(
" %08X
\n
"
,
read_c0_index
());
printk
(
"
\n\n
Full TLB Dump:
\n
"
"Idx EntryHi EntryLo0 EntryLo1 PageMask
\n
"
"--------------------------------------------------------------
\n
"
);
for
(
entry
=
0
;
entry
<
current_cpu_data
.
tlbsize
;
entry
++
)
{
write_c0_index
(
entry
);
printk
(
"
\n
%02i "
,
entry
);
dump_cur_tlb_regs
();
}
printk
(
"
\n
"
);
write_c0_entryhi
(
old_ctx
);
local_irq_restore
(
flags
);
}
void
local_flush_tlb_all
(
void
)
{
unsigned
long
flags
;
unsigned
long
old_ctx
;
int
entry
;
local_irq_save
(
flags
);
/* Save old context and create impossible VPN2 value */
old_ctx
=
read_c0_entryhi
();
write_c0_entrylo0
(
0
);
write_c0_entrylo1
(
0
);
entry
=
read_c0_wired
();
while
(
entry
<
current_cpu_data
.
tlbsize
)
{
write_c0_entryhi
(
UNIQUE_ENTRYHI
(
entry
));
write_c0_index
(
entry
);
tlb_write_indexed
();
entry
++
;
}
write_c0_entryhi
(
old_ctx
);
local_irq_restore
(
flags
);
}
/*
* Use a bogus region of memory (starting at 0) to sanitize the TLB's.
* Use increments of the maximum page size (16MB), and check for duplicate
* entries before doing a given write. Then, when we're safe from collisions
* with the firmware, go back and give all the entries invalid addresses with
* the normal flush routine. Wired entries will be killed as well!
*/
static
void
__init
sb1_sanitize_tlb
(
void
)
{
int
entry
;
long
addr
=
0
;
long
inc
=
1
<<
24
;
/* 16MB */
/* Save old context and create impossible VPN2 value */
write_c0_entrylo0
(
0
);
write_c0_entrylo1
(
0
);
for
(
entry
=
0
;
entry
<
current_cpu_data
.
tlbsize
;
entry
++
)
{
do
{
addr
+=
inc
;
write_c0_entryhi
(
addr
);
tlb_probe
();
}
while
((
int
)(
read_c0_index
())
>=
0
);
write_c0_index
(
entry
);
tlb_write_indexed
();
}
/* Now that we know we're safe from collisions, we can safely flush
the TLB with the "normal" routine. */
local_flush_tlb_all
();
}
void
local_flush_tlb_range
(
struct
vm_area_struct
*
vma
,
unsigned
long
start
,
unsigned
long
end
)
{
struct
mm_struct
*
mm
=
vma
->
vm_mm
;
int
cpu
=
smp_processor_id
();
if
(
cpu_context
(
cpu
,
mm
)
!=
0
)
{
unsigned
long
flags
;
int
size
;
size
=
(
end
-
start
+
(
PAGE_SIZE
-
1
))
>>
PAGE_SHIFT
;
size
=
(
size
+
1
)
>>
1
;
local_irq_save
(
flags
);
if
(
size
<=
(
current_cpu_data
.
tlbsize
/
2
))
{
int
oldpid
=
read_c0_entryhi
();
int
newpid
=
cpu_asid
(
cpu
,
mm
);
start
&=
(
PAGE_MASK
<<
1
);
end
+=
((
PAGE_SIZE
<<
1
)
-
1
);
end
&=
(
PAGE_MASK
<<
1
);
while
(
start
<
end
)
{
int
idx
;
write_c0_entryhi
(
start
|
newpid
);
start
+=
(
PAGE_SIZE
<<
1
);
tlb_probe
();
idx
=
read_c0_index
();
write_c0_entrylo0
(
0
);
write_c0_entrylo1
(
0
);
if
(
idx
<
0
)
continue
;
write_c0_entryhi
(
UNIQUE_ENTRYHI
(
idx
));
tlb_write_indexed
();
}
write_c0_entryhi
(
oldpid
);
}
else
{
drop_mmu_context
(
mm
,
cpu
);
}
local_irq_restore
(
flags
);
}
}
void
local_flush_tlb_kernel_range
(
unsigned
long
start
,
unsigned
long
end
)
{
unsigned
long
flags
;
int
size
;
size
=
(
end
-
start
+
(
PAGE_SIZE
-
1
))
>>
PAGE_SHIFT
;
size
=
(
size
+
1
)
>>
1
;
local_irq_save
(
flags
);
if
(
size
<=
(
current_cpu_data
.
tlbsize
/
2
))
{
int
pid
=
read_c0_entryhi
();
start
&=
(
PAGE_MASK
<<
1
);
end
+=
((
PAGE_SIZE
<<
1
)
-
1
);
end
&=
(
PAGE_MASK
<<
1
);
while
(
start
<
end
)
{
int
idx
;
write_c0_entryhi
(
start
);
start
+=
(
PAGE_SIZE
<<
1
);
tlb_probe
();
idx
=
read_c0_index
();
write_c0_entrylo0
(
0
);
write_c0_entrylo1
(
0
);
if
(
idx
<
0
)
continue
;
write_c0_entryhi
(
UNIQUE_ENTRYHI
(
idx
));
tlb_write_indexed
();
}
write_c0_entryhi
(
pid
);
}
else
{
local_flush_tlb_all
();
}
local_irq_restore
(
flags
);
}
void
local_flush_tlb_page
(
struct
vm_area_struct
*
vma
,
unsigned
long
page
)
{
int
cpu
=
smp_processor_id
();
if
(
cpu_context
(
cpu
,
vma
->
vm_mm
)
!=
0
)
{
unsigned
long
flags
;
int
oldpid
,
newpid
,
idx
;
newpid
=
cpu_asid
(
cpu
,
vma
->
vm_mm
);
page
&=
(
PAGE_MASK
<<
1
);
local_irq_save
(
flags
);
oldpid
=
read_c0_entryhi
();
write_c0_entryhi
(
page
|
newpid
);
tlb_probe
();
idx
=
read_c0_index
();
write_c0_entrylo0
(
0
);
write_c0_entrylo1
(
0
);
if
(
idx
<
0
)
goto
finish
;
/* Make sure all entries differ. */
write_c0_entryhi
(
UNIQUE_ENTRYHI
(
idx
));
tlb_write_indexed
();
finish:
write_c0_entryhi
(
oldpid
);
local_irq_restore
(
flags
);
}
}
/*
* Remove one kernel space TLB entry. This entry is assumed to be marked
* global so we don't do the ASID thing.
*/
void
local_flush_tlb_one
(
unsigned
long
page
)
{
unsigned
long
flags
;
int
oldpid
,
idx
;
local_irq_save
(
flags
);
oldpid
=
read_c0_entryhi
();
page
&=
(
PAGE_MASK
<<
1
);
write_c0_entryhi
(
page
);
tlb_probe
();
idx
=
read_c0_index
();
write_c0_entrylo0
(
0
);
write_c0_entrylo1
(
0
);
if
(
idx
>=
0
)
{
/* Make sure all entries differ. */
write_c0_entryhi
(
UNIQUE_ENTRYHI
(
idx
));
tlb_write_indexed
();
}
write_c0_entryhi
(
oldpid
);
local_irq_restore
(
flags
);
}
/* All entries common to a mm share an asid. To effectively flush
these entries, we just bump the asid. */
void
local_flush_tlb_mm
(
struct
mm_struct
*
mm
)
{
int
cpu
;
preempt_disable
();
cpu
=
smp_processor_id
();
if
(
cpu_context
(
cpu
,
mm
)
!=
0
)
{
drop_mmu_context
(
mm
,
cpu
);
}
preempt_enable
();
}
/* Stolen from mips32 routines */
void
__update_tlb
(
struct
vm_area_struct
*
vma
,
unsigned
long
address
,
pte_t
pte
)
{
unsigned
long
flags
;
pgd_t
*
pgdp
;
pud_t
*
pudp
;
pmd_t
*
pmdp
;
pte_t
*
ptep
;
int
idx
,
pid
;
/*
* Handle debugger faulting in for debugee.
*/
if
(
current
->
active_mm
!=
vma
->
vm_mm
)
return
;
local_irq_save
(
flags
);
pid
=
read_c0_entryhi
()
&
ASID_MASK
;
address
&=
(
PAGE_MASK
<<
1
);
write_c0_entryhi
(
address
|
pid
);
pgdp
=
pgd_offset
(
vma
->
vm_mm
,
address
);
tlb_probe
();
pudp
=
pud_offset
(
pgdp
,
address
);
pmdp
=
pmd_offset
(
pudp
,
address
);
idx
=
read_c0_index
();
ptep
=
pte_offset_map
(
pmdp
,
address
);
#if defined(CONFIG_64BIT_PHYS_ADDR) && defined(CONFIG_CPU_MIPS32_R1)
write_c0_entrylo0
(
ptep
->
pte_high
);
ptep
++
;
write_c0_entrylo1
(
ptep
->
pte_high
);
#else
write_c0_entrylo0
(
pte_val
(
*
ptep
++
)
>>
6
);
write_c0_entrylo1
(
pte_val
(
*
ptep
)
>>
6
);
#endif
if
(
idx
<
0
)
tlb_write_random
();
else
tlb_write_indexed
();
local_irq_restore
(
flags
);
}
void
__init
add_wired_entry
(
unsigned
long
entrylo0
,
unsigned
long
entrylo1
,
unsigned
long
entryhi
,
unsigned
long
pagemask
)
{
unsigned
long
flags
;
unsigned
long
wired
;
unsigned
long
old_pagemask
;
unsigned
long
old_ctx
;
local_irq_save
(
flags
);
/* Save old context and create impossible VPN2 value */
old_ctx
=
read_c0_entryhi
();
old_pagemask
=
read_c0_pagemask
();
wired
=
read_c0_wired
();
write_c0_wired
(
wired
+
1
);
write_c0_index
(
wired
);
write_c0_pagemask
(
pagemask
);
write_c0_entryhi
(
entryhi
);
write_c0_entrylo0
(
entrylo0
);
write_c0_entrylo1
(
entrylo1
);
tlb_write_indexed
();
write_c0_entryhi
(
old_ctx
);
write_c0_pagemask
(
old_pagemask
);
local_flush_tlb_all
();
local_irq_restore
(
flags
);
}
/*
* This is called from loadmmu.c. We have to set up all the
* memory management function pointers, as well as initialize
* the caches and tlbs
*/
void
tlb_init
(
void
)
{
write_c0_pagemask
(
PM_DEFAULT_MASK
);
write_c0_wired
(
0
);
/*
* We don't know what state the firmware left the TLB's in, so this is
* the ultra-conservative way to flush the TLB's and avoid machine
* check exceptions due to duplicate TLB entries
*/
sb1_sanitize_tlb
();
build_tlb_refill_handler
();
}
include/asm-mips/hazards.h
View file @
f5cfa980
...
...
@@ -74,7 +74,8 @@
#define irq_disable_hazard
_ehb
#elif defined(CONFIG_CPU_R10000) || defined(CONFIG_CPU_RM9000)
#elif defined(CONFIG_CPU_R10000) || defined(CONFIG_CPU_RM9000) || \
defined(CONFIG_CPU_SB1)
/*
* R10000 rocks - all hazards handled in hardware, so this becomes a nobrainer.
...
...
@@ -180,7 +181,8 @@ __asm__(
__asm__ __volatile__( \
"back_to_back_c0_hazard")
#elif defined(CONFIG_CPU_R10000) || defined(CONFIG_CPU_RM9000)
#elif defined(CONFIG_CPU_R10000) || defined(CONFIG_CPU_RM9000) || \
defined(CONFIG_CPU_SB1)
/*
* R10000 rocks - all hazards handled in hardware, so this becomes a nobrainer.
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment