Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Support
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
L
linux
Project overview
Project overview
Details
Activity
Releases
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Issues
0
Issues
0
List
Boards
Labels
Milestones
Merge Requests
0
Merge Requests
0
Analytics
Analytics
Repository
Value Stream
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Create a new issue
Commits
Issue Boards
Open sidebar
Kirill Smelkov
linux
Commits
d1df9cdc
Commit
d1df9cdc
authored
Dec 19, 2003
by
Tony Luck
Committed by
David Mosberger
Dec 19, 2003
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
[PATCH] ia64: enable recovery from TLB errors
Here's the updated version of the MCA TLB recovery patch.
parent
d273c362
Changes
7
Show whitespace changes
Inline
Side-by-side
Showing
7 changed files
with
307 additions
and
126 deletions
+307
-126
arch/ia64/kernel/asm-offsets.c
arch/ia64/kernel/asm-offsets.c
+4
-0
arch/ia64/kernel/efi.c
arch/ia64/kernel/efi.c
+12
-0
arch/ia64/kernel/mca.c
arch/ia64/kernel/mca.c
+18
-6
arch/ia64/kernel/mca_asm.S
arch/ia64/kernel/mca_asm.S
+236
-120
arch/ia64/mm/init.c
arch/ia64/mm/init.c
+21
-0
include/asm-ia64/mca.h
include/asm-ia64/mca.h
+12
-0
include/asm-ia64/pgtable.h
include/asm-ia64/pgtable.h
+4
-0
No files found.
arch/ia64/kernel/asm-offsets.c
View file @
d1df9cdc
...
@@ -12,6 +12,7 @@
...
@@ -12,6 +12,7 @@
#include <asm-ia64/ptrace.h>
#include <asm-ia64/ptrace.h>
#include <asm-ia64/siginfo.h>
#include <asm-ia64/siginfo.h>
#include <asm-ia64/sigcontext.h>
#include <asm-ia64/sigcontext.h>
#include <asm-ia64/mca.h>
#include "../kernel/sigframe.h"
#include "../kernel/sigframe.h"
...
@@ -204,4 +205,7 @@ void foo(void)
...
@@ -204,4 +205,7 @@ void foo(void)
# error "CLONE_SETTLS_BIT incorrect, please fix"
# error "CLONE_SETTLS_BIT incorrect, please fix"
#endif
#endif
BLANK
();
DEFINE
(
IA64_MCA_TLB_INFO_SIZE
,
sizeof
(
struct
ia64_mca_tlb_info
));
}
}
arch/ia64/kernel/efi.c
View file @
d1df9cdc
...
@@ -30,6 +30,7 @@
...
@@ -30,6 +30,7 @@
#include <asm/kregs.h>
#include <asm/kregs.h>
#include <asm/pgtable.h>
#include <asm/pgtable.h>
#include <asm/processor.h>
#include <asm/processor.h>
#include <asm/mca.h>
#define EFI_DEBUG 0
#define EFI_DEBUG 0
...
@@ -395,6 +396,9 @@ efi_map_pal_code (void)
...
@@ -395,6 +396,9 @@ efi_map_pal_code (void)
int
pal_code_count
=
0
;
int
pal_code_count
=
0
;
u64
mask
,
psr
;
u64
mask
,
psr
;
u64
vaddr
;
u64
vaddr
;
#ifdef CONFIG_IA64_MCA
int
cpu
;
#endif
efi_map_start
=
__va
(
ia64_boot_param
->
efi_memmap
);
efi_map_start
=
__va
(
ia64_boot_param
->
efi_memmap
);
efi_map_end
=
efi_map_start
+
ia64_boot_param
->
efi_memmap_size
;
efi_map_end
=
efi_map_start
+
ia64_boot_param
->
efi_memmap_size
;
...
@@ -455,6 +459,14 @@ efi_map_pal_code (void)
...
@@ -455,6 +459,14 @@ efi_map_pal_code (void)
IA64_GRANULE_SHIFT
);
IA64_GRANULE_SHIFT
);
ia64_set_psr
(
psr
);
/* restore psr */
ia64_set_psr
(
psr
);
/* restore psr */
ia64_srlz_i
();
ia64_srlz_i
();
#ifdef CONFIG_IA64_MCA
cpu
=
smp_processor_id
();
/* insert this TR into our list for MCA recovery purposes */
ia64_mca_tlb_list
[
cpu
].
pal_base
=
vaddr
&
mask
;
ia64_mca_tlb_list
[
cpu
].
pal_paddr
=
pte_val
(
mk_pte_phys
(
md
->
phys_addr
,
PAGE_KERNEL
));
#endif
}
}
}
}
...
...
arch/ia64/kernel/mca.c
View file @
d1df9cdc
...
@@ -83,9 +83,8 @@ u64 ia64_mca_stack[1024] __attribute__((aligned(16)));
...
@@ -83,9 +83,8 @@ u64 ia64_mca_stack[1024] __attribute__((aligned(16)));
u64
ia64_mca_stackframe
[
32
];
u64
ia64_mca_stackframe
[
32
];
u64
ia64_mca_bspstore
[
1024
];
u64
ia64_mca_bspstore
[
1024
];
u64
ia64_init_stack
[
KERNEL_STACK_SIZE
/
8
]
__attribute__
((
aligned
(
16
)));
u64
ia64_init_stack
[
KERNEL_STACK_SIZE
/
8
]
__attribute__
((
aligned
(
16
)));
u64
ia64_mca_sal_data_area
[
1356
];
u64
ia64_tlb_functional
;
u64
ia64_os_mca_recovery_successful
;
u64
ia64_os_mca_recovery_successful
;
u64
ia64_mca_serialize
;
static
void
ia64_mca_wakeup_ipi_wait
(
void
);
static
void
ia64_mca_wakeup_ipi_wait
(
void
);
static
void
ia64_mca_wakeup
(
int
cpu
);
static
void
ia64_mca_wakeup
(
int
cpu
);
static
void
ia64_mca_wakeup_all
(
void
);
static
void
ia64_mca_wakeup_all
(
void
);
...
@@ -95,6 +94,8 @@ extern void ia64_slave_init_handler (void);
...
@@ -95,6 +94,8 @@ extern void ia64_slave_init_handler (void);
static
u64
ia64_log_get
(
int
sal_info_type
,
u8
**
buffer
);
static
u64
ia64_log_get
(
int
sal_info_type
,
u8
**
buffer
);
extern
struct
hw_interrupt_type
irq_type_iosapic_level
;
extern
struct
hw_interrupt_type
irq_type_iosapic_level
;
struct
ia64_mca_tlb_info
ia64_mca_tlb_list
[
NR_CPUS
];
static
struct
irqaction
cmci_irqaction
=
{
static
struct
irqaction
cmci_irqaction
=
{
.
handler
=
ia64_mca_cmc_int_handler
,
.
handler
=
ia64_mca_cmc_int_handler
,
.
flags
=
SA_INTERRUPT
,
.
flags
=
SA_INTERRUPT
,
...
@@ -984,6 +985,9 @@ ia64_mca_wakeup_int_handler(int wakeup_irq, void *arg, struct pt_regs *ptregs)
...
@@ -984,6 +985,9 @@ ia64_mca_wakeup_int_handler(int wakeup_irq, void *arg, struct pt_regs *ptregs)
void
void
ia64_return_to_sal_check
(
void
)
ia64_return_to_sal_check
(
void
)
{
{
pal_processor_state_info_t
*
psp
=
(
pal_processor_state_info_t
*
)
&
ia64_sal_to_os_handoff_state
.
proc_state_param
;
/* Copy over some relevant stuff from the sal_to_os_mca_handoff
/* Copy over some relevant stuff from the sal_to_os_mca_handoff
* so that it can be used at the time of os_mca_to_sal_handoff
* so that it can be used at the time of os_mca_to_sal_handoff
*/
*/
...
@@ -993,14 +997,22 @@ ia64_return_to_sal_check(void)
...
@@ -993,14 +997,22 @@ ia64_return_to_sal_check(void)
ia64_os_to_sal_handoff_state
.
imots_sal_check_ra
=
ia64_os_to_sal_handoff_state
.
imots_sal_check_ra
=
ia64_sal_to_os_handoff_state
.
imsto_sal_check_ra
;
ia64_sal_to_os_handoff_state
.
imsto_sal_check_ra
;
/* Cold Boot for uncorrectable MCA */
/*
* Did we correct the error? At the moment the only error that
* we fix is a TLB error, if any other kind of error occurred
* we must reboot.
*/
if
(
psp
->
cc
==
1
&&
psp
->
bc
==
1
&&
psp
->
rc
==
1
&&
psp
->
uc
==
1
)
ia64_os_to_sal_handoff_state
.
imots_os_status
=
IA64_MCA_COLD_BOOT
;
ia64_os_to_sal_handoff_state
.
imots_os_status
=
IA64_MCA_COLD_BOOT
;
else
ia64_os_to_sal_handoff_state
.
imots_os_status
=
IA64_MCA_CORRECTED
;
/* Default = tell SAL to return to same context */
/* Default = tell SAL to return to same context */
ia64_os_to_sal_handoff_state
.
imots_context
=
IA64_MCA_SAME_CONTEXT
;
ia64_os_to_sal_handoff_state
.
imots_context
=
IA64_MCA_SAME_CONTEXT
;
ia64_os_to_sal_handoff_state
.
imots_new_min_state
=
ia64_os_to_sal_handoff_state
.
imots_new_min_state
=
(
u64
*
)
ia64_sal_to_os_handoff_state
.
pal_min_state
;
(
u64
*
)
ia64_sal_to_os_handoff_state
.
pal_min_state
;
}
}
/*
/*
...
@@ -1359,8 +1371,8 @@ ia64_init_handler (struct pt_regs *pt, struct switch_stack *sw)
...
@@ -1359,8 +1371,8 @@ ia64_init_handler (struct pt_regs *pt, struct switch_stack *sw)
void
void
ia64_log_prt_guid
(
efi_guid_t
*
p_guid
,
prfunc_t
prfunc
)
ia64_log_prt_guid
(
efi_guid_t
*
p_guid
,
prfunc_t
prfunc
)
{
{
char
out
[
40
];
//
char out[40];
printk
(
KERN_DEBUG
"GUID = %s
\n
"
,
efi_guid_unparse
(
p_guid
,
out
));
//
printk(KERN_DEBUG "GUID = %s\n", efi_guid_unparse(p_guid, out));
}
}
static
void
static
void
...
...
arch/ia64/kernel/mca_asm.S
View file @
d1df9cdc
...
@@ -14,6 +14,7 @@
...
@@ -14,6 +14,7 @@
//
3.
Move
stack
ptr
16
bytes
to
conform
to
C
calling
convention
//
3.
Move
stack
ptr
16
bytes
to
conform
to
C
calling
convention
//
//
#include <linux/config.h>
#include <linux/config.h>
#include <linux/threads.h>
#include <asm/asmmacro.h>
#include <asm/asmmacro.h>
#include <asm/pgtable.h>
#include <asm/pgtable.h>
...
@@ -22,20 +23,15 @@
...
@@ -22,20 +23,15 @@
#include <asm/mca.h>
#include <asm/mca.h>
/*
/*
*
When
we
get
a
n
machine
check
,
the
kernel
stack
pointer
is
no
longer
*
When
we
get
a
machine
check
,
the
kernel
stack
pointer
is
no
longer
*
valid
,
so
we
need
to
set
a
new
stack
pointer
.
*
valid
,
so
we
need
to
set
a
new
stack
pointer
.
*/
*/
#define MINSTATE_PHYS /* Make sure stack access is physical for MINSTATE */
#define MINSTATE_PHYS /* Make sure stack access is physical for MINSTATE */
/*
*
Needed
for
ia64_sal
call
*/
#define SAL_GET_STATE_INFO 0x01000001
/*
/*
*
Needed
for
return
context
to
SAL
*
Needed
for
return
context
to
SAL
*/
*/
#define IA64_MCA_SAME_CONTEXT 0
x0
#define IA64_MCA_SAME_CONTEXT 0
#define IA64_MCA_COLD_BOOT -2
#define IA64_MCA_COLD_BOOT -2
#include "minstate.h"
#include "minstate.h"
...
@@ -71,19 +67,36 @@
...
@@ -71,19 +67,36 @@
*
returns
ptr
to
SAL
rtn
save
loc
in
_tmp
*
returns
ptr
to
SAL
rtn
save
loc
in
_tmp
*/
*/
#define OS_MCA_TO_SAL_HANDOFF_STATE_RESTORE(_tmp) \
#define OS_MCA_TO_SAL_HANDOFF_STATE_RESTORE(_tmp) \
LOAD_PHYSICAL
(
p6
,
_tmp
,
ia64_sal_to_os_handoff_state
)
;; \
movl
_tmp
=
ia64_os_to_sal_handoff_state
;; \
LOAD_PHYSICAL
(
p7
,
_tmp
,
ia64_os_to_sal_handoff_state
)
;; \
DATA_VA_TO_PA
(
_tmp
)
;; \
(
p6
)
movl
r8
=
IA64_MCA_COLD_BOOT
; \
ld8
r8
=[
_tmp
],
0x08
;; \
(
p6
)
movl
r10
=
IA64_MCA_SAME_CONTEXT
; \
ld8
r9
=[
_tmp
],
0x08
;; \
(
p6
)
add
_tmp
=
0x18
,
_tmp
;; \
ld8
r10
=[
_tmp
],
0x08
;; \
(
p6
)
ld8
r9
=[
_tmp
],
0x10
; \
ld8
r22
=[
_tmp
],
0x08
;;
(
p6
)
mov
r22
=
r0
;; \
(
p7
)
ld8
r8
=[
_tmp
],
0x08
;; \
(
p7
)
ld8
r9
=[
_tmp
],
0x08
;; \
(
p7
)
ld8
r10
=[
_tmp
],
0x08
;; \
(
p7
)
ld8
r22
=[
_tmp
],
0x08
;;
//
now
_tmp
is
pointing
to
SAL
rtn
save
location
//
now
_tmp
is
pointing
to
SAL
rtn
save
location
/*
*
COLD_BOOT_HANDOFF_STATE
()
sets
ia64_mca_os_to_sal_state
*
imots_os_status
=
IA64_MCA_COLD_BOOT
*
imots_sal_gp
=
SAL
GP
*
imots_context
=
IA64_MCA_SAME_CONTEXT
*
imots_new_min_state
=
Min
state
save
area
pointer
*
imots_sal_check_ra
=
Return
address
to
location
within
SAL_CHECK
*
*/
#define COLD_BOOT_HANDOFF_STATE(sal_to_os_handoff,os_to_sal_handoff,tmp)\
movl
tmp
=
IA64_MCA_COLD_BOOT
; \
movl
sal_to_os_handoff
=
__pa
(
ia64_sal_to_os_handoff_state
)
; \
movl
os_to_sal_handoff
=
__pa
(
ia64_os_to_sal_handoff_state
)
;; \
st8
[
os_to_sal_handoff
]=
tmp
,
8
;; \
ld8
tmp
=[
sal_to_os_handoff
],
48
;; \
st8
[
os_to_sal_handoff
]=
tmp
,
8
;; \
movl
tmp
=
IA64_MCA_SAME_CONTEXT
;; \
st8
[
os_to_sal_handoff
]=
tmp
,
8
;; \
ld8
tmp
=[
sal_to_os_handoff
],-
8
;; \
st8
[
os_to_sal_handoff
]=
tmp
,
8
;; \
ld8
tmp
=[
sal_to_os_handoff
]
;; \
st8
[
os_to_sal_handoff
]=
tmp
;;
.
global
ia64_os_mca_dispatch
.
global
ia64_os_mca_dispatch
.
global
ia64_os_mca_dispatch_end
.
global
ia64_os_mca_dispatch_end
...
@@ -94,20 +107,21 @@
...
@@ -94,20 +107,21 @@
.
global
ia64_mca_stackframe
.
global
ia64_mca_stackframe
.
global
ia64_mca_bspstore
.
global
ia64_mca_bspstore
.
global
ia64_init_stack
.
global
ia64_init_stack
.
global
ia64_mca_sal_data_area
.
global
ia64_tlb_functional
.
text
.
text
.
align
16
.
align
16
ia64_os_mca_dispatch
:
ia64_os_mca_dispatch
:
#if defined(MCA_TEST)
//
Serialize
all
MCA
processing
//
Pretend
that
we
are
in
interrupt
context
//
movl
r2
=
ia64_mca_serialize
mov
r2
=
psr
mov
r3
=
1
;;
dep
r2
=
0
,
r2
,
PSR_IC
,
2
;
//
DATA_VA_TO_PA
(
r2
)
;;
mov
psr
.
l
=
r2
LOAD_PHYSICAL
(
p0
,
r2
,
ia64_mca_serialize
)
;;
#endif /* #if defined(MCA_TEST) */
ia64_os_mca_spin
:
xchg8
r4
=[
r2
],
r3
;;
cmp.ne
p6
,
p0
=
r4
,
r0
(
p6
)
br
ia64_os_mca_spin
//
Save
the
SAL
to
OS
MCA
handoff
state
as
defined
//
Save
the
SAL
to
OS
MCA
handoff
state
as
defined
//
by
SAL
SPEC
3
.0
//
by
SAL
SPEC
3
.0
...
@@ -124,6 +138,191 @@ begin_os_mca_dump:
...
@@ -124,6 +138,191 @@ begin_os_mca_dump:
ia64_os_mca_done_dump
:
ia64_os_mca_done_dump
:
//
movl
r16
=
__pa
(
ia64_sal_to_os_handoff_state
)+
56
LOAD_PHYSICAL
(
p0
,
r16
,
ia64_sal_to_os_handoff_state
+56)
;;
ld8
r18
=[
r16
]
//
Get
processor
state
parameter
on
existing
PALE_CHECK
.
;;
tbit.nz
p6
,
p7
=
r18
,
60
(
p7
)
br.spnt
done_tlb_purge_and_reload
//
The
following
code
purges
TC
and
TR
entries
.
Then
reload
all
TC
entries
.
//
Purge
percpu
data
TC
entries
.
begin_tlb_purge_and_reload
:
mov
r16
=
cr
.
lid
//
movl
r17
=
__pa
(
ia64_mca_tlb_list
)
//
Physical
address
of
ia64_mca_tlb_list
LOAD_PHYSICAL
(
p0
,
r17
,
ia64_mca_tlb_list
)
//
Physical
address
of
ia64_mca_tlb_list
mov
r19
=
0
mov
r20
=
NR_CPUS
;;
1
:
cmp.eq
p6
,
p7
=
r19
,
r20
(
p6
)
br.spnt.few
err
ld8
r18
=[
r17
],
IA64_MCA_TLB_INFO_SIZE
;;
add
r19
=
1
,
r19
cmp.eq
p6
,
p7
=
r18
,
r16
(
p7
)
br.sptk.few
1
b
;;
adds
r17
=-
IA64_MCA_TLB_INFO_SIZE
,
r17
;;
mov
r23
=
r17
//
save
current
ia64_mca_percpu_info
addr
pointer
.
adds
r17
=
16
,
r17
;;
.
global
aegl
aegl
:
ld8
r18
=[
r17
],
8
//
r18
=
ptce_base
;;
ld4
r19
=[
r17
],
4
//
r19
=
ptce_count
[
0
]
;;
ld4
r20
=[
r17
],
4
//
r20
=
ptce_count
[
1
]
;;
ld4
r21
=[
r17
],
4
//
r21
=
ptce_stride
[
0
]
mov
r24
=
0
;;
ld4
r22
=[
r17
],
4
//
r22
=
ptce_stride
[
1
]
adds
r20
=-
1
,
r20
;;
2
:
cmp.ltu
p6
,
p7
=
r24
,
r19
(
p7
)
br.cond.dpnt.few
4
f
mov
ar
.
lc
=
r20
3
:
ptc.e
r18
;;
add
r18
=
r22
,
r18
br.cloop.sptk.few
3
b
;;
add
r18
=
r21
,
r18
add
r24
=
1
,
r24
;;
br.sptk.few
2
b
4
:
srlz.i
//
srlz
.
i
implies
srlz
.
d
;;
//
Now
purge
addresses
formerly
mapped
by
TR
registers
//
1
.
Purge
ITR
&D
TR
for
kernel
.
movl
r16
=
KERNEL_START
mov
r18
=
KERNEL_TR_PAGE_SHIFT
<<
2
;;
ptr.i
r16
,
r18
ptr.d
r16
,
r18
;;
srlz.i
;;
srlz.d
;;
//
2
.
Purge
DTR
for
PERCPU
data
.
movl
r16
=
PERCPU_ADDR
mov
r18
=
PERCPU_PAGE_SHIFT
<<
2
;;
ptr.d
r16
,
r18
;;
srlz.d
;;
//
3
.
Purge
ITR
for
PAL
code
.
adds
r17
=
48
,
r23
;;
ld8
r16
=[
r17
]
mov
r18
=
IA64_GRANULE_SHIFT
<<
2
;;
ptr.i
r16
,
r18
;;
srlz.i
;;
//
4
.
Purge
DTR
for
stack
.
mov
r16
=
IA64_KR
(
CURRENT_STACK
)
;;
shl
r16
=
r16
,
IA64_GRANULE_SHIFT
movl
r19
=
PAGE_OFFSET
;;
add
r16
=
r19
,
r16
mov
r18
=
IA64_GRANULE_SHIFT
<<
2
;;
ptr.d
r16
,
r18
;;
srlz.i
;;
//
Finally
reload
the
TR
registers
.
//
1
.
Reload
DTR
/
ITR
registers
for
kernel
.
mov
r18
=
KERNEL_TR_PAGE_SHIFT
<<
2
movl
r17
=
KERNEL_START
;;
mov
cr
.
itir
=
r18
mov
cr
.
ifa
=
r17
mov
r16
=
IA64_TR_KERNEL
mov
r19
=
ip
movl
r18
=
PAGE_KERNEL
;;
dep
r17
=
0
,
r19
,
0
,
KERNEL_TR_PAGE_SHIFT
;;
or
r18
=
r17
,
r18
;;
itr.i
itr
[
r16
]=
r18
;;
itr.d
dtr
[
r16
]=
r18
;;
srlz.i
srlz.d
;;
//
2
.
Reload
DTR
register
for
PERCPU
data
.
adds
r17
=
8
,
r23
movl
r16
=
PERCPU_ADDR
//
vaddr
movl
r18
=
PERCPU_PAGE_SHIFT
<<
2
;;
mov
cr
.
itir
=
r18
mov
cr
.
ifa
=
r16
;;
ld8
r18
=[
r17
]
//
pte
mov
r16
=
IA64_TR_PERCPU_DATA
;
;;
itr.d
dtr
[
r16
]=
r18
;;
srlz.d
;;
//
3
.
Reload
ITR
for
PAL
code
.
adds
r17
=
40
,
r23
;;
ld8
r18
=[
r17
],
8
//
pte
;;
ld8
r16
=[
r17
]
//
vaddr
mov
r19
=
IA64_GRANULE_SHIFT
<<
2
;;
mov
cr
.
itir
=
r19
mov
cr
.
ifa
=
r16
mov
r20
=
IA64_TR_PALCODE
;;
itr.i
itr
[
r20
]=
r18
;;
srlz.i
;;
//
4
.
Reload
DTR
for
stack
.
mov
r16
=
IA64_KR
(
CURRENT_STACK
)
;;
shl
r16
=
r16
,
IA64_GRANULE_SHIFT
movl
r19
=
PAGE_OFFSET
;;
add
r18
=
r19
,
r16
movl
r20
=
PAGE_KERNEL
;;
add
r16
=
r20
,
r16
mov
r19
=
IA64_GRANULE_SHIFT
<<
2
;;
mov
cr
.
itir
=
r19
mov
cr
.
ifa
=
r18
mov
r20
=
IA64_TR_CURRENT_STACK
;;
itr.d
dtr
[
r20
]=
r16
;;
srlz.d
;;
br.sptk.many
done_tlb_purge_and_reload
err
:
COLD_BOOT_HANDOFF_STATE
(
r20
,
r21
,
r22
)
br.sptk.many
ia64_os_mca_done_restore
done_tlb_purge_and_reload
:
//
Setup
new
stack
frame
for
OS_MCA
handling
//
Setup
new
stack
frame
for
OS_MCA
handling
movl
r2
=
ia64_mca_bspstore
;; // local bspstore area location in r2
movl
r2
=
ia64_mca_bspstore
;; // local bspstore area location in r2
DATA_VA_TO_PA
(
r2
)
;;
DATA_VA_TO_PA
(
r2
)
;;
...
@@ -137,17 +336,11 @@ ia64_os_mca_done_dump:
...
@@ -137,17 +336,11 @@ ia64_os_mca_done_dump:
//
(
C
calling
convention
)
//
(
C
calling
convention
)
DATA_VA_TO_PA
(
r12
)
;;
DATA_VA_TO_PA
(
r12
)
;;
//
Check
to
see
if
the
MCA
resulted
from
a
TLB
error
//
Enter
virtual
mode
from
physical
mode
begin_tlb_error_check
:
br
ia64_os_mca_tlb_error_check
;;
done_tlb_error_check
:
//
If
TLB
is
functional
,
enter
virtual
mode
from
physical
mode
VIRTUAL_MODE_ENTER
(
r2
,
r3
,
ia64_os_mca_virtual_begin
,
r4
)
VIRTUAL_MODE_ENTER
(
r2
,
r3
,
ia64_os_mca_virtual_begin
,
r4
)
ia64_os_mca_virtual_begin
:
ia64_os_mca_virtual_begin
:
//
call
our
handler
//
Call
virtual
mode
handler
movl
r2
=
ia64_mca_ucmc_handler
;;
movl
r2
=
ia64_mca_ucmc_handler
;;
mov
b6
=
r2
;;
mov
b6
=
r2
;;
br.call.sptk.many
b0
=
b6
;;
br.call.sptk.many
b0
=
b6
;;
...
@@ -156,13 +349,6 @@ ia64_os_mca_virtual_begin:
...
@@ -156,13 +349,6 @@ ia64_os_mca_virtual_begin:
PHYSICAL_MODE_ENTER
(
r2
,
r3
,
ia64_os_mca_virtual_end
,
r4
)
PHYSICAL_MODE_ENTER
(
r2
,
r3
,
ia64_os_mca_virtual_end
,
r4
)
ia64_os_mca_virtual_end
:
ia64_os_mca_virtual_end
:
#if defined(MCA_TEST)
//
Pretend
that
we
are
in
interrupt
context
mov
r2
=
psr
;;
dep
r2
=
0
,
r2
,
PSR_IC
,
2
;;
mov
psr
.
l
=
r2
;;
#endif /* #if defined(MCA_TEST) */
//
restore
the
original
stack
frame
here
//
restore
the
original
stack
frame
here
movl
r2
=
ia64_mca_stackframe
//
restore
stack
frame
from
memory
at
r2
movl
r2
=
ia64_mca_stackframe
//
restore
stack
frame
from
memory
at
r2
;;
;;
...
@@ -178,14 +364,16 @@ begin_os_mca_restore:
...
@@ -178,14 +364,16 @@ begin_os_mca_restore:
br
ia64_os_mca_proc_state_restore
;;
br
ia64_os_mca_proc_state_restore
;;
ia64_os_mca_done_restore
:
ia64_os_mca_done_restore
:
movl
r3
=
ia64_tlb_functional
;;
DATA_VA_TO_PA
(
r3
)
;;
ld8
r3
=[
r3
]
;;
cmp.eq
p6
,
p7
=
r0
,
r3
;;
OS_MCA_TO_SAL_HANDOFF_STATE_RESTORE
(
r2
)
;;
OS_MCA_TO_SAL_HANDOFF_STATE_RESTORE
(
r2
)
;;
//
branch
back
to
SALE_CHECK
//
branch
back
to
SALE_CHECK
ld8
r3
=[
r2
]
;;
ld8
r3
=[
r2
]
;;
mov
b0
=
r3
;; // SAL_CHECK return address
mov
b0
=
r3
;; // SAL_CHECK return address
//
release
lock
movl
r3
=
ia64_mca_serialize
;;
DATA_VA_TO_PA
(
r3
)
;;
st8.rel
[
r3
]=
r0
br
b0
br
b0
;;
;;
ia64_os_mca_dispatch_end
:
ia64_os_mca_dispatch_end
:
...
@@ -205,8 +393,9 @@ ia64_os_mca_dispatch_end:
...
@@ -205,8 +393,9 @@ ia64_os_mca_dispatch_end:
ia64_os_mca_proc_state_dump
:
ia64_os_mca_proc_state_dump
:
//
Save
bank
1
GRs
16
-
31
which
will
be
used
by
c
-
language
code
when
we
switch
//
Save
bank
1
GRs
16
-
31
which
will
be
used
by
c
-
language
code
when
we
switch
//
to
virtual
addressing
mode
.
//
to
virtual
addressing
mode
.
movl
r2
=
ia64_mca_proc_state_dump
;; // Os state dump area
//
movl
r2
=
ia64_mca_proc_state_dump
;; // Os state dump area
DATA_VA_TO_PA
(
r2
)
//
convert
to
to
physical
address
//
DATA_VA_TO_PA
(
r2
)
//
convert
to
to
physical
address
LOAD_PHYSICAL
(
p0
,
r2
,
ia64_mca_proc_state_dump
)//
convert
OS
state
dump
area
to
physical
address
//
save
ar
.
NaT
//
save
ar
.
NaT
mov
r5
=
ar
.
unat
//
ar
.
unat
mov
r5
=
ar
.
unat
//
ar
.
unat
...
@@ -658,79 +847,6 @@ end_os_mca_restore:
...
@@ -658,79 +847,6 @@ end_os_mca_restore:
//
EndStub
//////////////////////////////////////////////////////////////////////
//
EndStub
//////////////////////////////////////////////////////////////////////
//++
//
Name
:
//
ia64_os_mca_tlb_error_check
()
//
//
Stub
Description
:
//
//
This
stub
checks
to
see
if
the
MCA
resulted
from
a
TLB
error
//
//--
ia64_os_mca_tlb_error_check
:
//
Retrieve
sal
data
structure
for
uncorrected
MCA
//
Make
the
ia64_sal_get_state_info
()
call
movl
r4
=
ia64_mca_sal_data_area
;;
movl
r7
=
ia64_sal
;;
mov
r6
=
r1
//
save
gp
DATA_VA_TO_PA
(
r4
)
//
convert
to
physical
address
DATA_VA_TO_PA
(
r7
)
;; // convert to physical address
ld8
r7
=[
r7
]
//
get
addr
of
pdesc
from
ia64_sal
movl
r3
=
SAL_GET_STATE_INFO
;;
DATA_VA_TO_PA
(
r7
)
;; // convert to physical address
ld8
r8
=[
r7
],
8
;; // get pdesc function pointer
dep
r8
=
0
,
r8
,
61
,
3
;; // convert SAL VA to PA
ld8
r1
=[
r7
]
;; // set new (ia64_sal) gp
dep
r1
=
0
,
r1
,
61
,
3
;; // convert SAL VA to PA
mov
b6
=
r8
alloc
r5
=
ar
.
pfs
,
8
,
0
,
8
,
0
;; // allocate stack frame for SAL call
mov
out0
=
r3
//
which
SAL
proc
to
call
mov
out1
=
r0
//
error
type
==
MCA
mov
out2
=
r0
//
null
arg
mov
out3
=
r4
//
data
copy
area
mov
out4
=
r0
//
null
arg
mov
out5
=
r0
//
null
arg
mov
out6
=
r0
//
null
arg
mov
out7
=
r0
;; // null arg
br.call.sptk.few
b0
=
b6
;;
mov
r1
=
r6
//
restore
gp
mov
ar
.
pfs
=
r5
;; // restore ar.pfs
movl
r6
=
ia64_tlb_functional
;;
DATA_VA_TO_PA
(
r6
)
//
needed
later
cmp.eq
p6
,
p7
=
r0
,
r8
;; // check SAL call return address
(
p7
)
st8
[
r6
]=
r0
//
clear
tlb_functional
flag
(
p7
)
br
tlb_failure
//
error
; return to SAL
//
examine
processor
error
log
for
type
of
error
add
r4
=
40
+
24
,
r4
;; // parse past record header (length=40)
//
and
section
header
(
length
=
24
)
ld4
r4
=[
r4
]
//
get
valid
field
of
processor
log
mov
r5
=
0xf00
;;
and
r5
=
r4
,
r5
;; // read bits 8-11 of valid field
//
to
determine
if
we
have
a
TLB
error
movl
r3
=
0x1
cmp.eq
p6
,
p7
=
r0
,
r5
;;
//
if
no
TLB
failure
,
set
tlb_functional
flag
(
p6
)
st8
[
r6
]=
r3
//
else
clear
flag
(
p7
)
st8
[
r6
]=
r0
//
if
no
TLB
failure
,
continue
with
normal
virtual
mode
logging
(
p6
)
br
done_tlb_error_check
//
else
no
point
in
entering
virtual
mode
for
logging
tlb_failure
:
br
ia64_os_mca_virtual_end
//
EndStub
//////////////////////////////////////////////////////////////////////
//
ok
,
the
issue
here
is
that
we
need
to
save
state
information
so
//
ok
,
the
issue
here
is
that
we
need
to
save
state
information
so
//
it
can
be
useable
by
the
kernel
debugger
and
show
regs
routines
.
//
it
can
be
useable
by
the
kernel
debugger
and
show
regs
routines
.
...
...
arch/ia64/mm/init.c
View file @
d1df9cdc
...
@@ -34,6 +34,7 @@
...
@@ -34,6 +34,7 @@
#include <asm/tlb.h>
#include <asm/tlb.h>
#include <asm/uaccess.h>
#include <asm/uaccess.h>
#include <asm/unistd.h>
#include <asm/unistd.h>
#include <asm/mca.h>
DEFINE_PER_CPU
(
struct
mmu_gather
,
mmu_gathers
);
DEFINE_PER_CPU
(
struct
mmu_gather
,
mmu_gathers
);
...
@@ -277,6 +278,10 @@ ia64_mmu_init (void *my_cpu_data)
...
@@ -277,6 +278,10 @@ ia64_mmu_init (void *my_cpu_data)
{
{
unsigned
long
psr
,
pta
,
impl_va_bits
;
unsigned
long
psr
,
pta
,
impl_va_bits
;
extern
void
__init
tlb_init
(
void
);
extern
void
__init
tlb_init
(
void
);
#ifdef CONFIG_IA64_MCA
int
cpu
;
#endif
#ifdef CONFIG_DISABLE_VHPT
#ifdef CONFIG_DISABLE_VHPT
# define VHPT_ENABLE_BIT 0
# define VHPT_ENABLE_BIT 0
#else
#else
...
@@ -335,6 +340,22 @@ ia64_mmu_init (void *my_cpu_data)
...
@@ -335,6 +340,22 @@ ia64_mmu_init (void *my_cpu_data)
ia64_set_pta
(
pta
|
(
0
<<
8
)
|
(
vmlpt_bits
<<
2
)
|
VHPT_ENABLE_BIT
);
ia64_set_pta
(
pta
|
(
0
<<
8
)
|
(
vmlpt_bits
<<
2
)
|
VHPT_ENABLE_BIT
);
ia64_tlb_init
();
ia64_tlb_init
();
#ifdef CONFIG_IA64_MCA
cpu
=
smp_processor_id
();
/* mca handler uses cr.lid as key to pick the right entry */
ia64_mca_tlb_list
[
cpu
].
cr_lid
=
ia64_getreg
(
_IA64_REG_CR_LID
);
/* insert this percpu data information into our list for MCA recovery purposes */
ia64_mca_tlb_list
[
cpu
].
percpu_paddr
=
pte_val
(
mk_pte_phys
(
__pa
(
my_cpu_data
),
PAGE_KERNEL
));
/* Also save per-cpu tlb flush recipe for use in physical mode mca handler */
ia64_mca_tlb_list
[
cpu
].
ptce_base
=
local_cpu_data
->
ptce_base
;
ia64_mca_tlb_list
[
cpu
].
ptce_count
[
0
]
=
local_cpu_data
->
ptce_count
[
0
];
ia64_mca_tlb_list
[
cpu
].
ptce_count
[
1
]
=
local_cpu_data
->
ptce_count
[
1
];
ia64_mca_tlb_list
[
cpu
].
ptce_stride
[
0
]
=
local_cpu_data
->
ptce_stride
[
0
];
ia64_mca_tlb_list
[
cpu
].
ptce_stride
[
1
]
=
local_cpu_data
->
ptce_stride
[
1
];
#endif
}
}
#ifdef CONFIG_VIRTUAL_MEM_MAP
#ifdef CONFIG_VIRTUAL_MEM_MAP
...
...
include/asm-ia64/mca.h
View file @
d1df9cdc
...
@@ -18,6 +18,7 @@
...
@@ -18,6 +18,7 @@
#include <asm/param.h>
#include <asm/param.h>
#include <asm/sal.h>
#include <asm/sal.h>
#include <asm/processor.h>
#include <asm/processor.h>
#include <asm/mca_asm.h>
/* These are the return codes from all the IA64_MCA specific interfaces */
/* These are the return codes from all the IA64_MCA specific interfaces */
typedef
int
ia64_mca_return_code_t
;
typedef
int
ia64_mca_return_code_t
;
...
@@ -61,6 +62,17 @@ enum {
...
@@ -61,6 +62,17 @@ enum {
IA64_MCA_RENDEZ_CHECKIN_DONE
=
0x1
IA64_MCA_RENDEZ_CHECKIN_DONE
=
0x1
};
};
/* the following data structure is used for TLB error recovery purposes */
extern
struct
ia64_mca_tlb_info
{
u64
cr_lid
;
u64
percpu_paddr
;
u64
ptce_base
;
u32
ptce_count
[
2
];
u32
ptce_stride
[
2
];
u64
pal_paddr
;
u64
pal_base
;
}
ia64_mca_tlb_list
[
NR_CPUS
];
/* Information maintained by the MC infrastructure */
/* Information maintained by the MC infrastructure */
typedef
struct
ia64_mc_info_s
{
typedef
struct
ia64_mc_info_s
{
u64
imi_mca_handler
;
u64
imi_mca_handler
;
...
...
include/asm-ia64/pgtable.h
View file @
d1df9cdc
...
@@ -230,6 +230,10 @@ ia64_phys_addr_valid (unsigned long addr)
...
@@ -230,6 +230,10 @@ ia64_phys_addr_valid (unsigned long addr)
#define mk_pte(page, pgprot) pfn_pte(page_to_pfn(page), (pgprot))
#define mk_pte(page, pgprot) pfn_pte(page_to_pfn(page), (pgprot))
/* This takes a physical page address that is used by the remapping functions */
#define mk_pte_phys(physpage, pgprot) \
({ pte_t __pte; pte_val(__pte) = physpage + pgprot_val(pgprot); __pte; })
#define pte_modify(_pte, newprot) \
#define pte_modify(_pte, newprot) \
(__pte((pte_val(_pte) & ~_PAGE_CHG_MASK) | (pgprot_val(newprot) & _PAGE_CHG_MASK)))
(__pte((pte_val(_pte) & ~_PAGE_CHG_MASK) | (pgprot_val(newprot) & _PAGE_CHG_MASK)))
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment