Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Support
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
L
linux
Project overview
Project overview
Details
Activity
Releases
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Issues
0
Issues
0
List
Boards
Labels
Milestones
Merge Requests
0
Merge Requests
0
Analytics
Analytics
Repository
Value Stream
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Create a new issue
Commits
Issue Boards
Open sidebar
Kirill Smelkov
linux
Commits
c66b31f3
Commit
c66b31f3
authored
Mar 31, 2009
by
Tony Luck
Committed by
Tony Luck
Mar 31, 2009
Browse files
Options
Browse Files
Download
Plain Diff
Pull pvops into release branch
parents
50fbe56c
7120569c
Changes
45
Show whitespace changes
Inline
Side-by-side
Showing
45 changed files
with
3134 additions
and
212 deletions
+3134
-212
arch/ia64/include/asm/intrinsics.h
arch/ia64/include/asm/intrinsics.h
+5
-1
arch/ia64/include/asm/module.h
arch/ia64/include/asm/module.h
+6
-0
arch/ia64/include/asm/native/inst.h
arch/ia64/include/asm/native/inst.h
+13
-0
arch/ia64/include/asm/native/patchlist.h
arch/ia64/include/asm/native/patchlist.h
+38
-0
arch/ia64/include/asm/native/pvchk_inst.h
arch/ia64/include/asm/native/pvchk_inst.h
+8
-0
arch/ia64/include/asm/paravirt.h
arch/ia64/include/asm/paravirt.h
+65
-0
arch/ia64/include/asm/paravirt_patch.h
arch/ia64/include/asm/paravirt_patch.h
+143
-0
arch/ia64/include/asm/paravirt_privop.h
arch/ia64/include/asm/paravirt_privop.h
+361
-4
arch/ia64/include/asm/timex.h
arch/ia64/include/asm/timex.h
+1
-0
arch/ia64/include/asm/xen/hypervisor.h
arch/ia64/include/asm/xen/hypervisor.h
+18
-21
arch/ia64/include/asm/xen/inst.h
arch/ia64/include/asm/xen/inst.h
+28
-0
arch/ia64/include/asm/xen/interface.h
arch/ia64/include/asm/xen/interface.h
+9
-0
arch/ia64/include/asm/xen/minstate.h
arch/ia64/include/asm/xen/minstate.h
+10
-1
arch/ia64/include/asm/xen/patchlist.h
arch/ia64/include/asm/xen/patchlist.h
+38
-0
arch/ia64/include/asm/xen/privop.h
arch/ia64/include/asm/xen/privop.h
+7
-1
arch/ia64/kernel/Makefile
arch/ia64/kernel/Makefile
+9
-30
arch/ia64/kernel/Makefile.gate
arch/ia64/kernel/Makefile.gate
+27
-0
arch/ia64/kernel/asm-offsets.c
arch/ia64/kernel/asm-offsets.c
+2
-0
arch/ia64/kernel/efi.c
arch/ia64/kernel/efi.c
+1
-0
arch/ia64/kernel/entry.S
arch/ia64/kernel/entry.S
+2
-2
arch/ia64/kernel/fsys.S
arch/ia64/kernel/fsys.S
+18
-17
arch/ia64/kernel/gate.S
arch/ia64/kernel/gate.S
+90
-81
arch/ia64/kernel/gate.lds.S
arch/ia64/kernel/gate.lds.S
+9
-8
arch/ia64/kernel/head.S
arch/ia64/kernel/head.S
+8
-2
arch/ia64/kernel/ivt.S
arch/ia64/kernel/ivt.S
+1
-1
arch/ia64/kernel/module.c
arch/ia64/kernel/module.c
+32
-0
arch/ia64/kernel/paravirt.c
arch/ia64/kernel/paravirt.c
+535
-4
arch/ia64/kernel/paravirt_patch.c
arch/ia64/kernel/paravirt_patch.c
+514
-0
arch/ia64/kernel/paravirt_patchlist.c
arch/ia64/kernel/paravirt_patchlist.c
+79
-0
arch/ia64/kernel/paravirt_patchlist.h
arch/ia64/kernel/paravirt_patchlist.h
+28
-0
arch/ia64/kernel/paravirtentry.S
arch/ia64/kernel/paravirtentry.S
+80
-19
arch/ia64/kernel/patch.c
arch/ia64/kernel/patch.c
+30
-10
arch/ia64/kernel/setup.c
arch/ia64/kernel/setup.c
+2
-0
arch/ia64/kernel/time.c
arch/ia64/kernel/time.c
+9
-0
arch/ia64/kernel/vmlinux.lds.S
arch/ia64/kernel/vmlinux.lds.S
+30
-0
arch/ia64/kvm/kvm-ia64.c
arch/ia64/kvm/kvm-ia64.c
+1
-1
arch/ia64/kvm/vcpu.c
arch/ia64/kvm/vcpu.c
+1
-1
arch/ia64/kvm/vtlb.c
arch/ia64/kvm/vtlb.c
+2
-0
arch/ia64/mm/init.c
arch/ia64/mm/init.c
+6
-3
arch/ia64/scripts/pvcheck.sed
arch/ia64/scripts/pvcheck.sed
+1
-0
arch/ia64/xen/Makefile
arch/ia64/xen/Makefile
+17
-2
arch/ia64/xen/gate-data.S
arch/ia64/xen/gate-data.S
+3
-0
arch/ia64/xen/hypercall.S
arch/ia64/xen/hypercall.S
+2
-0
arch/ia64/xen/time.c
arch/ia64/xen/time.c
+48
-0
arch/ia64/xen/xen_pv_ops.c
arch/ia64/xen/xen_pv_ops.c
+797
-3
No files found.
arch/ia64/include/asm/intrinsics.h
View file @
c66b31f3
...
...
@@ -202,7 +202,11 @@ extern long ia64_cmpxchg_called_with_bad_pointer (void);
#ifndef __ASSEMBLY__
#if defined(CONFIG_PARAVIRT) && defined(__KERNEL__)
#define IA64_INTRINSIC_API(name) pv_cpu_ops.name
#ifdef ASM_SUPPORTED
# define IA64_INTRINSIC_API(name) paravirt_ ## name
#else
# define IA64_INTRINSIC_API(name) pv_cpu_ops.name
#endif
#define IA64_INTRINSIC_MACRO(name) paravirt_ ## name
#else
#define IA64_INTRINSIC_API(name) ia64_native_ ## name
...
...
arch/ia64/include/asm/module.h
View file @
c66b31f3
...
...
@@ -16,6 +16,12 @@ struct mod_arch_specific {
struct
elf64_shdr
*
got
;
/* global offset table */
struct
elf64_shdr
*
opd
;
/* official procedure descriptors */
struct
elf64_shdr
*
unwind
;
/* unwind-table section */
#ifdef CONFIG_PARAVIRT
struct
elf64_shdr
*
paravirt_bundles
;
/* paravirt_alt_bundle_patch table */
struct
elf64_shdr
*
paravirt_insts
;
/* paravirt_alt_inst_patch table */
#endif
unsigned
long
gp
;
/* global-pointer for module */
void
*
core_unw_table
;
/* core unwind-table cookie returned by unwinder */
...
...
arch/ia64/include/asm/native/inst.h
View file @
c66b31f3
...
...
@@ -30,6 +30,9 @@
#define __paravirt_work_processed_syscall_target \
ia64_work_processed_syscall
#define paravirt_fsyscall_table ia64_native_fsyscall_table
#define paravirt_fsys_bubble_down ia64_native_fsys_bubble_down
#ifdef CONFIG_PARAVIRT_GUEST_ASM_CLOBBER_CHECK
# define PARAVIRT_POISON 0xdeadbeefbaadf00d
# define CLOBBER(clob) \
...
...
@@ -74,6 +77,11 @@
(pred) mov reg = psr \
CLOBBER(clob)
#define MOV_FROM_ITC(pred, pred_clob, reg, clob) \
(pred) mov reg = ar.itc \
CLOBBER(clob) \
CLOBBER_PRED(pred_clob)
#define MOV_TO_IFA(reg, clob) \
mov cr.ifa = reg \
CLOBBER(clob)
...
...
@@ -158,6 +166,11 @@
#define RSM_PSR_DT \
rsm psr.dt
#define RSM_PSR_BE_I(clob0, clob1) \
rsm psr.be | psr.i \
CLOBBER(clob0) \
CLOBBER(clob1)
#define SSM_PSR_DT_AND_SRLZ_I \
ssm psr.dt \
;; \
...
...
arch/ia64/include/asm/native/patchlist.h
0 → 100644
View file @
c66b31f3
/******************************************************************************
* arch/ia64/include/asm/native/inst.h
*
* Copyright (c) 2008 Isaku Yamahata <yamahata at valinux co jp>
* VA Linux Systems Japan K.K.
*
* This program is free software; you can redistribute it and/or modify
* it under the terms of the GNU General Public License as published by
* the Free Software Foundation; either version 2 of the License, or
* (at your option) any later version.
*
* This program is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
* GNU General Public License for more details.
*
* You should have received a copy of the GNU General Public License
* along with this program; if not, write to the Free Software
* Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
*
*/
#define __paravirt_start_gate_fsyscall_patchlist \
__ia64_native_start_gate_fsyscall_patchlist
#define __paravirt_end_gate_fsyscall_patchlist \
__ia64_native_end_gate_fsyscall_patchlist
#define __paravirt_start_gate_brl_fsys_bubble_down_patchlist \
__ia64_native_start_gate_brl_fsys_bubble_down_patchlist
#define __paravirt_end_gate_brl_fsys_bubble_down_patchlist \
__ia64_native_end_gate_brl_fsys_bubble_down_patchlist
#define __paravirt_start_gate_vtop_patchlist \
__ia64_native_start_gate_vtop_patchlist
#define __paravirt_end_gate_vtop_patchlist \
__ia64_native_end_gate_vtop_patchlist
#define __paravirt_start_gate_mckinley_e9_patchlist \
__ia64_native_start_gate_mckinley_e9_patchlist
#define __paravirt_end_gate_mckinley_e9_patchlist \
__ia64_native_end_gate_mckinley_e9_patchlist
arch/ia64/include/asm/native/pvchk_inst.h
View file @
c66b31f3
...
...
@@ -180,6 +180,11 @@
IS_PRED_IN(pred) \
IS_RREG_OUT(reg) \
IS_RREG_CLOB(clob)
#define MOV_FROM_ITC(pred, pred_clob, reg, clob) \
IS_PRED_IN(pred) \
IS_PRED_CLOB(pred_clob) \
IS_RREG_OUT(reg) \
IS_RREG_CLOB(clob)
#define MOV_TO_IFA(reg, clob) \
IS_RREG_IN(reg) \
IS_RREG_CLOB(clob)
...
...
@@ -246,6 +251,9 @@
IS_RREG_CLOB(clob2)
#define RSM_PSR_DT \
nop 0
#define RSM_PSR_BE_I(clob0, clob1) \
IS_RREG_CLOB(clob0) \
IS_RREG_CLOB(clob1)
#define SSM_PSR_DT_AND_SRLZ_I \
nop 0
#define BSW_0(clob0, clob1, clob2) \
...
...
arch/ia64/include/asm/paravirt.h
View file @
c66b31f3
...
...
@@ -22,6 +22,56 @@
#ifndef __ASM_PARAVIRT_H
#define __ASM_PARAVIRT_H
#ifndef __ASSEMBLY__
/******************************************************************************
* fsys related addresses
*/
struct
pv_fsys_data
{
unsigned
long
*
fsyscall_table
;
void
*
fsys_bubble_down
;
};
extern
struct
pv_fsys_data
pv_fsys_data
;
unsigned
long
*
paravirt_get_fsyscall_table
(
void
);
char
*
paravirt_get_fsys_bubble_down
(
void
);
/******************************************************************************
* patchlist addresses for gate page
*/
enum
pv_gate_patchlist
{
PV_GATE_START_FSYSCALL
,
PV_GATE_END_FSYSCALL
,
PV_GATE_START_BRL_FSYS_BUBBLE_DOWN
,
PV_GATE_END_BRL_FSYS_BUBBLE_DOWN
,
PV_GATE_START_VTOP
,
PV_GATE_END_VTOP
,
PV_GATE_START_MCKINLEY_E9
,
PV_GATE_END_MCKINLEY_E9
,
};
struct
pv_patchdata
{
unsigned
long
start_fsyscall_patchlist
;
unsigned
long
end_fsyscall_patchlist
;
unsigned
long
start_brl_fsys_bubble_down_patchlist
;
unsigned
long
end_brl_fsys_bubble_down_patchlist
;
unsigned
long
start_vtop_patchlist
;
unsigned
long
end_vtop_patchlist
;
unsigned
long
start_mckinley_e9_patchlist
;
unsigned
long
end_mckinley_e9_patchlist
;
void
*
gate_section
;
};
extern
struct
pv_patchdata
pv_patchdata
;
unsigned
long
paravirt_get_gate_patchlist
(
enum
pv_gate_patchlist
type
);
void
*
paravirt_get_gate_section
(
void
);
#endif
#ifdef CONFIG_PARAVIRT_GUEST
#define PARAVIRT_HYPERVISOR_TYPE_DEFAULT 0
...
...
@@ -68,6 +118,14 @@ struct pv_init_ops {
int
(
*
arch_setup_nomca
)(
void
);
void
(
*
post_smp_prepare_boot_cpu
)(
void
);
#ifdef ASM_SUPPORTED
unsigned
long
(
*
patch_bundle
)(
void
*
sbundle
,
void
*
ebundle
,
unsigned
long
type
);
unsigned
long
(
*
patch_inst
)(
unsigned
long
stag
,
unsigned
long
etag
,
unsigned
long
type
);
#endif
void
(
*
patch_branch
)(
unsigned
long
tag
,
unsigned
long
type
);
};
extern
struct
pv_init_ops
pv_init_ops
;
...
...
@@ -210,6 +268,8 @@ struct pv_time_ops {
int
(
*
do_steal_accounting
)(
unsigned
long
*
new_itm
);
void
(
*
clocksource_resume
)(
void
);
unsigned
long
long
(
*
sched_clock
)(
void
);
};
extern
struct
pv_time_ops
pv_time_ops
;
...
...
@@ -227,6 +287,11 @@ paravirt_do_steal_accounting(unsigned long *new_itm)
return
pv_time_ops
.
do_steal_accounting
(
new_itm
);
}
static
inline
unsigned
long
long
paravirt_sched_clock
(
void
)
{
return
pv_time_ops
.
sched_clock
();
}
#endif
/* !__ASSEMBLY__ */
#else
...
...
arch/ia64/include/asm/paravirt_patch.h
0 → 100644
View file @
c66b31f3
/******************************************************************************
* Copyright (c) 2008 Isaku Yamahata <yamahata at valinux co jp>
* VA Linux Systems Japan K.K.
*
* This program is free software; you can redistribute it and/or modify
* it under the terms of the GNU General Public License as published by
* the Free Software Foundation; either version 2 of the License, or
* (at your option) any later version.
*
* This program is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
* GNU General Public License for more details.
*
* You should have received a copy of the GNU General Public License
* along with this program; if not, write to the Free Software
* Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
*
*/
#ifndef __ASM_PARAVIRT_PATCH_H
#define __ASM_PARAVIRT_PATCH_H
#ifdef __ASSEMBLY__
.
section
.
paravirt_branches
,
"a"
.
previous
#define PARAVIRT_PATCH_SITE_BR(type) \
{ \
[1:] ; \
br.cond.sptk.many 2f ; \
nop.b 0 ; \
nop.b 0;; ; \
} ; \
2: \
.xdata8 ".paravirt_branches", 1b, type
#else
#include <linux/stringify.h>
#include <asm/intrinsics.h>
/* for binary patch */
struct
paravirt_patch_site_bundle
{
void
*
sbundle
;
void
*
ebundle
;
unsigned
long
type
;
};
/* label means the beginning of new bundle */
#define paravirt_alt_bundle(instr, privop) \
"\t998:\n" \
"\t" instr "\n" \
"\t999:\n" \
"\t.pushsection .paravirt_bundles, \"a\"\n" \
"\t.popsection\n" \
"\t.xdata8 \".paravirt_bundles\", 998b, 999b, " \
__stringify(privop) "\n"
struct
paravirt_patch_bundle_elem
{
const
void
*
sbundle
;
const
void
*
ebundle
;
unsigned
long
type
;
};
struct
paravirt_patch_site_inst
{
unsigned
long
stag
;
unsigned
long
etag
;
unsigned
long
type
;
};
#define paravirt_alt_inst(instr, privop) \
"\t[998:]\n" \
"\t" instr "\n" \
"\t[999:]\n" \
"\t.pushsection .paravirt_insts, \"a\"\n" \
"\t.popsection\n" \
"\t.xdata8 \".paravirt_insts\", 998b, 999b, " \
__stringify(privop) "\n"
struct
paravirt_patch_site_branch
{
unsigned
long
tag
;
unsigned
long
type
;
};
struct
paravirt_patch_branch_target
{
const
void
*
entry
;
unsigned
long
type
;
};
void
__paravirt_patch_apply_branch
(
unsigned
long
tag
,
unsigned
long
type
,
const
struct
paravirt_patch_branch_target
*
entries
,
unsigned
int
nr_entries
);
void
paravirt_patch_reloc_br
(
unsigned
long
tag
,
const
void
*
target
);
void
paravirt_patch_reloc_brl
(
unsigned
long
tag
,
const
void
*
target
);
#if defined(ASM_SUPPORTED) && defined(CONFIG_PARAVIRT)
unsigned
long
ia64_native_patch_bundle
(
void
*
sbundle
,
void
*
ebundle
,
unsigned
long
type
);
unsigned
long
__paravirt_patch_apply_bundle
(
void
*
sbundle
,
void
*
ebundle
,
unsigned
long
type
,
const
struct
paravirt_patch_bundle_elem
*
elems
,
unsigned
long
nelems
,
const
struct
paravirt_patch_bundle_elem
**
found
);
void
paravirt_patch_apply_bundle
(
const
struct
paravirt_patch_site_bundle
*
start
,
const
struct
paravirt_patch_site_bundle
*
end
);
void
paravirt_patch_apply_inst
(
const
struct
paravirt_patch_site_inst
*
start
,
const
struct
paravirt_patch_site_inst
*
end
);
void
paravirt_patch_apply
(
void
);
#else
#define paravirt_patch_apply_bundle(start, end) do { } while (0)
#define paravirt_patch_apply_inst(start, end) do { } while (0)
#define paravirt_patch_apply() do { } while (0)
#endif
#endif
/* !__ASSEMBLEY__ */
#endif
/* __ASM_PARAVIRT_PATCH_H */
/*
* Local variables:
* mode: C
* c-set-style: "linux"
* c-basic-offset: 8
* tab-width: 8
* indent-tabs-mode: t
* End:
*/
arch/ia64/include/asm/paravirt_privop.h
View file @
c66b31f3
...
...
@@ -33,7 +33,7 @@
*/
struct
pv_cpu_ops
{
void
(
*
fc
)(
unsigned
long
addr
);
void
(
*
fc
)(
void
*
addr
);
unsigned
long
(
*
thash
)(
unsigned
long
addr
);
unsigned
long
(
*
get_cpuid
)(
int
index
);
unsigned
long
(
*
get_pmd
)(
int
index
);
...
...
@@ -60,12 +60,18 @@ extern unsigned long ia64_native_getreg_func(int regnum);
/* Instructions paravirtualized for performance */
/************************************************/
#ifndef ASM_SUPPORTED
#define paravirt_ssm_i() pv_cpu_ops.ssm_i()
#define paravirt_rsm_i() pv_cpu_ops.rsm_i()
#define __paravirt_getreg() pv_cpu_ops.getreg()
#endif
/* mask for ia64_native_ssm/rsm() must be constant.("i" constraing).
* static inline function doesn't satisfy it. */
#define paravirt_ssm(mask) \
do { \
if ((mask) == IA64_PSR_I) \
p
v_cpu_ops.
ssm_i(); \
p
aravirt_
ssm_i(); \
else \
ia64_native_ssm(mask); \
} while (0)
...
...
@@ -73,7 +79,7 @@ extern unsigned long ia64_native_getreg_func(int regnum);
#define paravirt_rsm(mask) \
do { \
if ((mask) == IA64_PSR_I) \
p
v_cpu_ops.
rsm_i(); \
p
aravirt_
rsm_i(); \
else \
ia64_native_rsm(mask); \
} while (0)
...
...
@@ -86,7 +92,7 @@ extern unsigned long ia64_native_getreg_func(int regnum);
if ((reg) == _IA64_REG_IP) \
res = ia64_native_getreg(_IA64_REG_IP); \
else \
res =
pv_cpu_ops.
getreg(reg); \
res =
__paravirt_
getreg(reg); \
res; \
})
...
...
@@ -112,6 +118,12 @@ void paravirt_cpu_asm_init(const struct pv_cpu_asm_switch *cpu_asm_switch);
#endif
/* CONFIG_PARAVIRT */
#if defined(CONFIG_PARAVIRT) && defined(ASM_SUPPORTED)
#define paravirt_dv_serialize_data() ia64_dv_serialize_data()
#else
#define paravirt_dv_serialize_data()
/* nothing */
#endif
/* these routines utilize privilege-sensitive or performance-sensitive
* privileged instructions so the code must be replaced with
* paravirtualized versions */
...
...
@@ -121,4 +133,349 @@ void paravirt_cpu_asm_init(const struct pv_cpu_asm_switch *cpu_asm_switch);
IA64_PARAVIRT_ASM_FUNC(work_processed_syscall)
#define ia64_leave_kernel IA64_PARAVIRT_ASM_FUNC(leave_kernel)
#if defined(CONFIG_PARAVIRT)
/******************************************************************************
* binary patching infrastructure
*/
#define PARAVIRT_PATCH_TYPE_FC 1
#define PARAVIRT_PATCH_TYPE_THASH 2
#define PARAVIRT_PATCH_TYPE_GET_CPUID 3
#define PARAVIRT_PATCH_TYPE_GET_PMD 4
#define PARAVIRT_PATCH_TYPE_PTCGA 5
#define PARAVIRT_PATCH_TYPE_GET_RR 6
#define PARAVIRT_PATCH_TYPE_SET_RR 7
#define PARAVIRT_PATCH_TYPE_SET_RR0_TO_RR4 8
#define PARAVIRT_PATCH_TYPE_SSM_I 9
#define PARAVIRT_PATCH_TYPE_RSM_I 10
#define PARAVIRT_PATCH_TYPE_GET_PSR_I 11
#define PARAVIRT_PATCH_TYPE_INTRIN_LOCAL_IRQ_RESTORE 12
/* PARAVIRT_PATY_TYPE_[GS]ETREG + _IA64_REG_xxx */
#define PARAVIRT_PATCH_TYPE_GETREG 0x10000000
#define PARAVIRT_PATCH_TYPE_SETREG 0x20000000
/*
* struct task_struct* (*ia64_switch_to)(void* next_task);
* void *ia64_leave_syscall;
* void *ia64_work_processed_syscall
* void *ia64_leave_kernel;
*/
#define PARAVIRT_PATCH_TYPE_BR_START 0x30000000
#define PARAVIRT_PATCH_TYPE_BR_SWITCH_TO \
(PARAVIRT_PATCH_TYPE_BR_START + 0)
#define PARAVIRT_PATCH_TYPE_BR_LEAVE_SYSCALL \
(PARAVIRT_PATCH_TYPE_BR_START + 1)
#define PARAVIRT_PATCH_TYPE_BR_WORK_PROCESSED_SYSCALL \
(PARAVIRT_PATCH_TYPE_BR_START + 2)
#define PARAVIRT_PATCH_TYPE_BR_LEAVE_KERNEL \
(PARAVIRT_PATCH_TYPE_BR_START + 3)
#ifdef ASM_SUPPORTED
#include <asm/paravirt_patch.h>
/*
* pv_cpu_ops calling stub.
* normal function call convension can't be written by gcc
* inline assembly.
*
* from the caller's point of view,
* the following registers will be clobbered.
* r2, r3
* r8-r15
* r16, r17
* b6, b7
* p6-p15
* ar.ccv
*
* from the callee's point of view ,
* the following registers can be used.
* r2, r3: scratch
* r8: scratch, input argument0 and return value
* r0-r15: scratch, input argument1-5
* b6: return pointer
* b7: scratch
* p6-p15: scratch
* ar.ccv: scratch
*
* other registers must not be changed. especially
* b0: rp: preserved. gcc ignores b0 in clobbered register.
* r16: saved gp
*/
/* 5 bundles */
#define __PARAVIRT_BR \
";;\n" \
"{ .mlx\n" \
"nop 0\n" \
"movl r2 = %[op_addr]\n"
/* get function pointer address */
\
";;\n" \
"}\n" \
"1:\n" \
"{ .mii\n" \
"ld8 r2 = [r2]\n"
/* load function descriptor address */
\
"mov r17 = ip\n"
/* get ip to calc return address */
\
"mov r16 = gp\n"
/* save gp */
\
";;\n" \
"}\n" \
"{ .mii\n" \
"ld8 r3 = [r2], 8\n"
/* load entry address */
\
"adds r17 = 1f - 1b, r17\n"
/* calculate return address */
\
";;\n" \
"mov b7 = r3\n"
/* set entry address */
\
"}\n" \
"{ .mib\n" \
"ld8 gp = [r2]\n"
/* load gp value */
\
"mov b6 = r17\n"
/* set return address */
\
"br.cond.sptk.few b7\n"
/* intrinsics are very short isns */
\
"}\n" \
"1:\n" \
"{ .mii\n" \
"mov gp = r16\n"
/* restore gp value */
\
"nop 0\n" \
"nop 0\n" \
";;\n" \
"}\n"
#define PARAVIRT_OP(op) \
[op_addr] "i"(&pv_cpu_ops.op)
#define PARAVIRT_TYPE(type) \
PARAVIRT_PATCH_TYPE_ ## type
#define PARAVIRT_REG_CLOBBERS0 \
"r2", "r3",
/*"r8",*/
"r9", "r10", "r11", "r14", \
"r15", "r16", "r17"
#define PARAVIRT_REG_CLOBBERS1 \
"r2","r3",
/*"r8",*/
"r9", "r10", "r11", "r14", \
"r15", "r16", "r17"
#define PARAVIRT_REG_CLOBBERS2 \
"r2", "r3",
/*"r8", "r9",*/
"r10", "r11", "r14", \
"r15", "r16", "r17"
#define PARAVIRT_REG_CLOBBERS5 \
"r2", "r3",
/*"r8", "r9", "r10", "r11", "r14",*/
\
"r15", "r16", "r17"
#define PARAVIRT_BR_CLOBBERS \
"b6", "b7"
#define PARAVIRT_PR_CLOBBERS \
"p6", "p7", "p8", "p9", "p10", "p11", "p12", "p13", "p14", "p15"
#define PARAVIRT_AR_CLOBBERS \
"ar.ccv"
#define PARAVIRT_CLOBBERS0 \
PARAVIRT_REG_CLOBBERS0, \
PARAVIRT_BR_CLOBBERS, \
PARAVIRT_PR_CLOBBERS, \
PARAVIRT_AR_CLOBBERS, \
"memory"
#define PARAVIRT_CLOBBERS1 \
PARAVIRT_REG_CLOBBERS1, \
PARAVIRT_BR_CLOBBERS, \
PARAVIRT_PR_CLOBBERS, \
PARAVIRT_AR_CLOBBERS, \
"memory"
#define PARAVIRT_CLOBBERS2 \
PARAVIRT_REG_CLOBBERS2, \
PARAVIRT_BR_CLOBBERS, \
PARAVIRT_PR_CLOBBERS, \
PARAVIRT_AR_CLOBBERS, \
"memory"
#define PARAVIRT_CLOBBERS5 \
PARAVIRT_REG_CLOBBERS5, \
PARAVIRT_BR_CLOBBERS, \
PARAVIRT_PR_CLOBBERS, \
PARAVIRT_AR_CLOBBERS, \
"memory"
#define PARAVIRT_BR0(op, type) \
register unsigned long ia64_clobber asm ("r8"); \
asm volatile (paravirt_alt_bundle(__PARAVIRT_BR, \
PARAVIRT_TYPE(type)) \
: "=r"(ia64_clobber) \
: PARAVIRT_OP(op) \
: PARAVIRT_CLOBBERS0)
#define PARAVIRT_BR0_RET(op, type) \
register unsigned long ia64_intri_res asm ("r8"); \
asm volatile (paravirt_alt_bundle(__PARAVIRT_BR, \
PARAVIRT_TYPE(type)) \
: "=r"(ia64_intri_res) \
: PARAVIRT_OP(op) \
: PARAVIRT_CLOBBERS0)
#define PARAVIRT_BR1(op, type, arg1) \
register unsigned long __##arg1 asm ("r8") = arg1; \
register unsigned long ia64_clobber asm ("r8"); \
asm volatile (paravirt_alt_bundle(__PARAVIRT_BR, \
PARAVIRT_TYPE(type)) \
: "=r"(ia64_clobber) \
: PARAVIRT_OP(op), "0"(__##arg1) \
: PARAVIRT_CLOBBERS1)
#define PARAVIRT_BR1_RET(op, type, arg1) \
register unsigned long ia64_intri_res asm ("r8"); \
register unsigned long __##arg1 asm ("r8") = arg1; \
asm volatile (paravirt_alt_bundle(__PARAVIRT_BR, \
PARAVIRT_TYPE(type)) \
: "=r"(ia64_intri_res) \
: PARAVIRT_OP(op), "0"(__##arg1) \
: PARAVIRT_CLOBBERS1)
#define PARAVIRT_BR1_VOID(op, type, arg1) \
register void *__##arg1 asm ("r8") = arg1; \
register unsigned long ia64_clobber asm ("r8"); \
asm volatile (paravirt_alt_bundle(__PARAVIRT_BR, \
PARAVIRT_TYPE(type)) \
: "=r"(ia64_clobber) \
: PARAVIRT_OP(op), "0"(__##arg1) \
: PARAVIRT_CLOBBERS1)
#define PARAVIRT_BR2(op, type, arg1, arg2) \
register unsigned long __##arg1 asm ("r8") = arg1; \
register unsigned long __##arg2 asm ("r9") = arg2; \
register unsigned long ia64_clobber1 asm ("r8"); \
register unsigned long ia64_clobber2 asm ("r9"); \
asm volatile (paravirt_alt_bundle(__PARAVIRT_BR, \
PARAVIRT_TYPE(type)) \
: "=r"(ia64_clobber1), "=r"(ia64_clobber2) \
: PARAVIRT_OP(op), "0"(__##arg1), "1"(__##arg2) \
: PARAVIRT_CLOBBERS2)
#define PARAVIRT_DEFINE_CPU_OP0(op, type) \
static inline void \
paravirt_ ## op (void) \
{ \
PARAVIRT_BR0(op, type); \
}
#define PARAVIRT_DEFINE_CPU_OP0_RET(op, type) \
static inline unsigned long \
paravirt_ ## op (void) \
{ \
PARAVIRT_BR0_RET(op, type); \
return ia64_intri_res; \
}
#define PARAVIRT_DEFINE_CPU_OP1_VOID(op, type) \
static inline void \
paravirt_ ## op (void *arg1) \
{ \
PARAVIRT_BR1_VOID(op, type, arg1); \
}
#define PARAVIRT_DEFINE_CPU_OP1(op, type) \
static inline void \
paravirt_ ## op (unsigned long arg1) \
{ \
PARAVIRT_BR1(op, type, arg1); \
}
#define PARAVIRT_DEFINE_CPU_OP1_RET(op, type) \
static inline unsigned long \
paravirt_ ## op (unsigned long arg1) \
{ \
PARAVIRT_BR1_RET(op, type, arg1); \
return ia64_intri_res; \
}
#define PARAVIRT_DEFINE_CPU_OP2(op, type) \
static inline void \
paravirt_ ## op (unsigned long arg1, \
unsigned long arg2) \
{ \
PARAVIRT_BR2(op, type, arg1, arg2); \
}
PARAVIRT_DEFINE_CPU_OP1_VOID
(
fc
,
FC
);
PARAVIRT_DEFINE_CPU_OP1_RET
(
thash
,
THASH
)
PARAVIRT_DEFINE_CPU_OP1_RET
(
get_cpuid
,
GET_CPUID
)
PARAVIRT_DEFINE_CPU_OP1_RET
(
get_pmd
,
GET_PMD
)
PARAVIRT_DEFINE_CPU_OP2
(
ptcga
,
PTCGA
)
PARAVIRT_DEFINE_CPU_OP1_RET
(
get_rr
,
GET_RR
)
PARAVIRT_DEFINE_CPU_OP2
(
set_rr
,
SET_RR
)
PARAVIRT_DEFINE_CPU_OP0
(
ssm_i
,
SSM_I
)
PARAVIRT_DEFINE_CPU_OP0
(
rsm_i
,
RSM_I
)
PARAVIRT_DEFINE_CPU_OP0_RET
(
get_psr_i
,
GET_PSR_I
)
PARAVIRT_DEFINE_CPU_OP1
(
intrin_local_irq_restore
,
INTRIN_LOCAL_IRQ_RESTORE
)
static
inline
void
paravirt_set_rr0_to_rr4
(
unsigned
long
val0
,
unsigned
long
val1
,
unsigned
long
val2
,
unsigned
long
val3
,
unsigned
long
val4
)
{
register
unsigned
long
__val0
asm
(
"r8"
)
=
val0
;
register
unsigned
long
__val1
asm
(
"r9"
)
=
val1
;
register
unsigned
long
__val2
asm
(
"r10"
)
=
val2
;
register
unsigned
long
__val3
asm
(
"r11"
)
=
val3
;
register
unsigned
long
__val4
asm
(
"r14"
)
=
val4
;
register
unsigned
long
ia64_clobber0
asm
(
"r8"
);
register
unsigned
long
ia64_clobber1
asm
(
"r9"
);
register
unsigned
long
ia64_clobber2
asm
(
"r10"
);
register
unsigned
long
ia64_clobber3
asm
(
"r11"
);
register
unsigned
long
ia64_clobber4
asm
(
"r14"
);
asm
volatile
(
paravirt_alt_bundle
(
__PARAVIRT_BR
,
PARAVIRT_TYPE
(
SET_RR0_TO_RR4
))
:
"=r"
(
ia64_clobber0
),
"=r"
(
ia64_clobber1
),
"=r"
(
ia64_clobber2
),
"=r"
(
ia64_clobber3
),
"=r"
(
ia64_clobber4
)
:
PARAVIRT_OP
(
set_rr0_to_rr4
),
"0"
(
__val0
),
"1"
(
__val1
),
"2"
(
__val2
),
"3"
(
__val3
),
"4"
(
__val4
)
:
PARAVIRT_CLOBBERS5
);
}
/* unsigned long paravirt_getreg(int reg) */
#define __paravirt_getreg(reg) \
({ \
register unsigned long ia64_intri_res asm ("r8"); \
register unsigned long __reg asm ("r8") = (reg); \
\
BUILD_BUG_ON(!__builtin_constant_p(reg)); \
asm volatile (paravirt_alt_bundle(__PARAVIRT_BR, \
PARAVIRT_TYPE(GETREG) \
+ (reg)) \
: "=r"(ia64_intri_res) \
: PARAVIRT_OP(getreg), "0"(__reg) \
: PARAVIRT_CLOBBERS1); \
\
ia64_intri_res; \
})
/* void paravirt_setreg(int reg, unsigned long val) */
#define paravirt_setreg(reg, val) \
do { \
register unsigned long __val asm ("r8") = val; \
register unsigned long __reg asm ("r9") = reg; \
register unsigned long ia64_clobber1 asm ("r8"); \
register unsigned long ia64_clobber2 asm ("r9"); \
\
BUILD_BUG_ON(!__builtin_constant_p(reg)); \
asm volatile (paravirt_alt_bundle(__PARAVIRT_BR, \
PARAVIRT_TYPE(SETREG) \
+ (reg)) \
: "=r"(ia64_clobber1), \
"=r"(ia64_clobber2) \
: PARAVIRT_OP(setreg), \
"1"(__reg), "0"(__val) \
: PARAVIRT_CLOBBERS2); \
} while (0)
#endif
/* ASM_SUPPORTED */
#endif
/* CONFIG_PARAVIRT && ASM_SUPPOTED */
#endif
/* _ASM_IA64_PARAVIRT_PRIVOP_H */
arch/ia64/include/asm/timex.h
View file @
c66b31f3
...
...
@@ -40,5 +40,6 @@ get_cycles (void)
}
extern
void
ia64_cpu_local_tick
(
void
);
extern
unsigned
long
long
ia64_native_sched_clock
(
void
);
#endif
/* _ASM_IA64_TIMEX_H */
arch/ia64/include/asm/xen/hypervisor.h
View file @
c66b31f3
...
...
@@ -33,9 +33,6 @@
#ifndef _ASM_IA64_XEN_HYPERVISOR_H
#define _ASM_IA64_XEN_HYPERVISOR_H
#ifdef CONFIG_XEN
#include <linux/init.h>
#include <xen/interface/xen.h>
#include <xen/interface/version.h>
/* to compile feature.c */
#include <xen/features.h>
/* to comiple xen-netfront.c */
...
...
@@ -43,22 +40,32 @@
/* xen_domain_type is set before executing any C code by early_xen_setup */
enum
xen_domain_type
{
XEN_NATIVE
,
XEN_PV_DOMAIN
,
XEN_HVM_DOMAIN
,
XEN_NATIVE
,
/* running on bare hardware */
XEN_PV_DOMAIN
,
/* running in a PV domain */
XEN_HVM_DOMAIN
,
/* running in a Xen hvm domain*/
};
#ifdef CONFIG_XEN
extern
enum
xen_domain_type
xen_domain_type
;
#else
#define xen_domain_type XEN_NATIVE
#endif
#define xen_domain() (xen_domain_type != XEN_NATIVE)
#define xen_pv_domain() (xen_domain_type == XEN_PV_DOMAIN)
#define xen_pv_domain() (xen_domain() && \
xen_domain_type == XEN_PV_DOMAIN)
#define xen_hvm_domain() (xen_domain() && \
xen_domain_type == XEN_HVM_DOMAIN)
#ifdef CONFIG_XEN_DOM0
#define xen_initial_domain() (xen_pv_domain() && \
(xen_start_info->flags & SIF_INITDOMAIN))
#define xen_hvm_domain() (xen_domain_type == XEN_HVM_DOMAIN)
#else
#define xen_initial_domain() (0)
#endif
/* deprecated. remove this */
#define is_running_on_xen() (xen_domain_type == XEN_PV_DOMAIN)
#ifdef CONFIG_XEN
extern
struct
shared_info
*
HYPERVISOR_shared_info
;
extern
struct
start_info
*
xen_start_info
;
...
...
@@ -74,16 +81,6 @@ void force_evtchn_callback(void);
/* For setup_arch() in arch/ia64/kernel/setup.c */
void
xen_ia64_enable_opt_feature
(
void
);
#else
/* CONFIG_XEN */
#define xen_domain() (0)
#define xen_pv_domain() (0)
#define xen_initial_domain() (0)
#define xen_hvm_domain() (0)
#define is_running_on_xen() (0)
/* deprecated. remove this */
#endif
#define is_initial_xendomain() (0)
/* deprecated. remove this */
#endif
/* _ASM_IA64_XEN_HYPERVISOR_H */
arch/ia64/include/asm/xen/inst.h
View file @
c66b31f3
...
...
@@ -33,6 +33,9 @@
#define __paravirt_work_processed_syscall_target \
xen_work_processed_syscall
#define paravirt_fsyscall_table xen_fsyscall_table
#define paravirt_fsys_bubble_down xen_fsys_bubble_down
#define MOV_FROM_IFA(reg) \
movl reg = XSI_IFA; \
;; \
...
...
@@ -110,6 +113,27 @@
.
endm
#define MOV_FROM_PSR(pred, reg, clob) __MOV_FROM_PSR pred, reg, clob
/* assuming ar.itc is read with interrupt disabled. */
#define MOV_FROM_ITC(pred, pred_clob, reg, clob) \
(pred) movl clob = XSI_ITC_OFFSET; \
;; \
(pred) ld8 clob = [clob]; \
(pred) mov reg = ar.itc; \
;; \
(pred) add reg = reg, clob; \
;; \
(pred) movl clob = XSI_ITC_LAST; \
;; \
(pred) ld8 clob = [clob]; \
;; \
(pred) cmp.geu.unc pred_clob, p0 = clob, reg; \
;; \
(pred_clob) add reg = 1, clob; \
;; \
(pred) movl clob = XSI_ITC_LAST; \
;; \
(pred) st8 [clob] = reg
#define MOV_TO_IFA(reg, clob) \
movl clob = XSI_IFA; \
...
...
@@ -362,6 +386,10 @@
#define RSM_PSR_DT \
XEN_HYPER_RSM_PSR_DT
#define RSM_PSR_BE_I(clob0, clob1) \
RSM_PSR_I(p0, clob0, clob1); \
rum psr.be
#define SSM_PSR_DT_AND_SRLZ_I \
XEN_HYPER_SSM_PSR_DT
...
...
arch/ia64/include/asm/xen/interface.h
View file @
c66b31f3
...
...
@@ -209,6 +209,15 @@ struct mapped_regs {
unsigned
long
krs
[
8
];
/* kernel registers */
unsigned
long
tmp
[
16
];
/* temp registers
(e.g. for hyperprivops) */
/* itc paravirtualization
* vAR.ITC = mAR.ITC + itc_offset
* itc_last is one which was lastly passed to
* the guest OS in order to prevent it from
* going backwords.
*/
unsigned
long
itc_offset
;
unsigned
long
itc_last
;
};
};
};
...
...
arch/ia64/include/asm/xen/minstate.h
View file @
c66b31f3
#ifdef CONFIG_VIRT_CPU_ACCOUNTING
/* read ar.itc in advance, and use it before leaving bank 0 */
#define XEN_ACCOUNT_GET_STAMP \
MOV_FROM_ITC(pUStk, p6, r20, r2);
#else
#define XEN_ACCOUNT_GET_STAMP
#endif
/*
* DO_SAVE_MIN switches to the kernel stacks (if necessary) and saves
* the minimum state necessary that allows us to turn psr.ic back
...
...
@@ -123,7 +132,7 @@
;; \
.mem.offset 0,0; st8.spill [r16]=r2,16; \
.mem.offset 8,0; st8.spill [r17]=r3,16; \
ACCOUNT_GET_STAMP \
XEN_
ACCOUNT_GET_STAMP \
adds r2=IA64_PT_REGS_R16_OFFSET,r1; \
;; \
EXTRA; \
...
...
arch/ia64/include/asm/xen/patchlist.h
0 → 100644
View file @
c66b31f3
/******************************************************************************
* arch/ia64/include/asm/xen/patchlist.h
*
* Copyright (c) 2008 Isaku Yamahata <yamahata at valinux co jp>
* VA Linux Systems Japan K.K.
*
* This program is free software; you can redistribute it and/or modify
* it under the terms of the GNU General Public License as published by
* the Free Software Foundation; either version 2 of the License, or
* (at your option) any later version.
*
* This program is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
* GNU General Public License for more details.
*
* You should have received a copy of the GNU General Public License
* along with this program; if not, write to the Free Software
* Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
*
*/
#define __paravirt_start_gate_fsyscall_patchlist \
__xen_start_gate_fsyscall_patchlist
#define __paravirt_end_gate_fsyscall_patchlist \
__xen_end_gate_fsyscall_patchlist
#define __paravirt_start_gate_brl_fsys_bubble_down_patchlist \
__xen_start_gate_brl_fsys_bubble_down_patchlist
#define __paravirt_end_gate_brl_fsys_bubble_down_patchlist \
__xen_end_gate_brl_fsys_bubble_down_patchlist
#define __paravirt_start_gate_vtop_patchlist \
__xen_start_gate_vtop_patchlist
#define __paravirt_end_gate_vtop_patchlist \
__xen_end_gate_vtop_patchlist
#define __paravirt_start_gate_mckinley_e9_patchlist \
__xen_start_gate_mckinley_e9_patchlist
#define __paravirt_end_gate_mckinley_e9_patchlist \
__xen_end_gate_mckinley_e9_patchlist
arch/ia64/include/asm/xen/privop.h
View file @
c66b31f3
...
...
@@ -55,6 +55,8 @@
#define XSI_BANK1_R16 (XSI_BASE + XSI_BANK1_R16_OFS)
#define XSI_BANKNUM (XSI_BASE + XSI_BANKNUM_OFS)
#define XSI_IHA (XSI_BASE + XSI_IHA_OFS)
#define XSI_ITC_OFFSET (XSI_BASE + XSI_ITC_OFFSET_OFS)
#define XSI_ITC_LAST (XSI_BASE + XSI_ITC_LAST_OFS)
#endif
#ifndef __ASSEMBLY__
...
...
@@ -67,7 +69,7 @@
* may have different semantics depending on whether they are executed
* at PL0 vs PL!=0. When paravirtualized, these instructions mustn't
* be allowed to execute directly, lest incorrect semantics result. */
extern
void
xen_fc
(
unsigned
long
addr
);
extern
void
xen_fc
(
void
*
addr
);
extern
unsigned
long
xen_thash
(
unsigned
long
addr
);
/* Note that "ttag" and "cover" are also privilege-sensitive; "ttag"
...
...
@@ -80,8 +82,10 @@ extern unsigned long xen_thash(unsigned long addr);
extern
unsigned
long
xen_get_cpuid
(
int
index
);
extern
unsigned
long
xen_get_pmd
(
int
index
);
#ifndef ASM_SUPPORTED
extern
unsigned
long
xen_get_eflag
(
void
);
/* see xen_ia64_getreg */
extern
void
xen_set_eflag
(
unsigned
long
);
/* see xen_ia64_setreg */
#endif
/************************************************/
/* Instructions paravirtualized for performance */
...
...
@@ -106,6 +110,7 @@ extern void xen_set_eflag(unsigned long); /* see xen_ia64_setreg */
#define xen_get_virtual_pend() \
(*(((uint8_t *)XEN_MAPPEDREGS->interrupt_mask_addr) - 1))
#ifndef ASM_SUPPORTED
/* Although all privileged operations can be left to trap and will
* be properly handled by Xen, some are frequent enough that we use
* hyperprivops for performance. */
...
...
@@ -123,6 +128,7 @@ extern void xen_set_rr0_to_rr4(unsigned long val0, unsigned long val1,
unsigned
long
val4
);
extern
void
xen_set_kr
(
unsigned
long
index
,
unsigned
long
val
);
extern
void
xen_ptcga
(
unsigned
long
addr
,
unsigned
long
size
);
#endif
/* !ASM_SUPPORTED */
#endif
/* !__ASSEMBLY__ */
...
...
arch/ia64/kernel/Makefile
View file @
c66b31f3
...
...
@@ -5,7 +5,7 @@
extra-y
:=
head.o init_task.o vmlinux.lds
obj-y
:=
acpi.o entry.o efi.o efi_stub.o gate-data.o fsys.o ia64_ksyms.o irq.o irq_ia64.o
\
irq_lsapic.o ivt.o machvec.o pal.o patch.o process.o perfmon.o ptrace.o sal.o
\
irq_lsapic.o ivt.o machvec.o pal.o pa
ravirt_patchlist.o pa
tch.o process.o perfmon.o ptrace.o sal.o
\
salinfo.o setup.o signal.o sys_ia64.o time.o traps.o unaligned.o
\
unwind.o mca.o mca_asm.o topology.o dma-mapping.o
...
...
@@ -36,7 +36,8 @@ obj-$(CONFIG_PCI_MSI) += msi_ia64.o
mca_recovery-y
+=
mca_drv.o mca_drv_asm.o
obj-$(CONFIG_IA64_MC_ERR_INJECT)
+=
err_inject.o
obj-$(CONFIG_PARAVIRT)
+=
paravirt.o paravirtentry.o
obj-$(CONFIG_PARAVIRT)
+=
paravirt.o paravirtentry.o
\
paravirt_patch.o
obj-$(CONFIG_IA64_ESI)
+=
esi.o
ifneq
($(CONFIG_IA64_ESI),)
...
...
@@ -45,35 +46,13 @@ endif
obj-$(CONFIG_DMAR)
+=
pci-dma.o
obj-$(CONFIG_SWIOTLB)
+=
pci-swiotlb.o
# The gate DSO image is built using a special linker script.
targets
+=
gate.so gate-syms.o
extra-y
+=
gate.so gate-syms.o gate.lds gate.o
# fp_emulate() expects f2-f5,f16-f31 to contain the user-level state.
CFLAGS_traps.o
+=
-mfixed-range
=
f2-f5,f16-f31
CPPFLAGS_gate.lds
:=
-P
-C
-U
$(ARCH)
quiet_cmd_gate
=
GATE
$@
cmd_gate
=
$(CC)
-nostdlib
$
(
GATECFLAGS_
$
(
@F
))
-Wl
,-T,
$(
filter-out
FORCE,
$^
)
-o
$@
GATECFLAGS_gate.so
=
-shared
-s
-Wl
,-soname
=
linux-gate.so.1
\
$(
call
ld-option,
-Wl
$(comma)
--hash-style
=
sysv
)
$(obj)/gate.so
:
$(obj)/gate.lds $(obj)/gate.o FORCE
$(
call
if_changed,gate
)
$(obj)/built-in.o
:
$(obj)/gate-syms.o
$(obj)/built-in.o
:
ld_flags += -R $(obj)/gate-syms.o
GATECFLAGS_gate-syms.o
=
-r
$(obj)/gate-syms.o
:
$(obj)/gate.lds $(obj)/gate.o FORCE
$(
call
if_changed,gate
)
# gate-data.o contains the gate DSO image as data in section .data.gate.
# We must build gate.so before we can assemble it.
# Note: kbuild does not track this dependency due to usage of .incbin
$(obj)/gate-data.o
:
$(obj)/gate.so
# The gate DSO image is built using a special linker script.
include
$(srctree)/arch/ia64/kernel/Makefile.gate
# tell compiled for native
CPPFLAGS_gate.lds
+=
-D__IA64_GATE_PARAVIRTUALIZED_NATIVE
# Calculate NR_IRQ = max(IA64_NATIVE_NR_IRQS, XEN_NR_IRQS, ...) based on config
define
sed-y
...
...
@@ -109,9 +88,9 @@ include/asm-ia64/nr-irqs.h: arch/$(SRCARCH)/kernel/nr-irqs.s
clean-files
+=
$(objtree)
/include/asm-ia64/nr-irqs.h
#
# native ivt.S
and entry
.S
# native ivt.S
, entry.S and fsys
.S
#
ASM_PARAVIRT_OBJS
=
ivt.o entry.o
ASM_PARAVIRT_OBJS
=
ivt.o entry.o
fsys.o
define
paravirtualized_native
AFLAGS_$(1)
+=
-D__IA64_ASM_PARAVIRTUALIZED_NATIVE
AFLAGS_pvchk-sed-$(1)
+=
-D__IA64_ASM_PARAVIRTUALIZED_PVCHECK
...
...
arch/ia64/kernel/Makefile.gate
0 → 100644
View file @
c66b31f3
# The gate DSO image is built using a special linker script.
targets
+=
gate.so gate-syms.o
extra-y
+=
gate.so gate-syms.o gate.lds gate.o
CPPFLAGS_gate.lds
:=
-P
-C
-U
$(ARCH)
quiet_cmd_gate
=
GATE
$@
cmd_gate
=
$(CC)
-nostdlib
$
(
GATECFLAGS_
$
(
@F
))
-Wl
,-T,
$(
filter-out
FORCE,
$^
)
-o
$@
GATECFLAGS_gate.so
=
-shared
-s
-Wl
,-soname
=
linux-gate.so.1
\
$(
call
ld-option,
-Wl
$(comma)
--hash-style
=
sysv
)
$(obj)/gate.so
:
$(obj)/gate.lds $(obj)/gate.o FORCE
$(
call
if_changed,gate
)
$(obj)/built-in.o
:
$(obj)/gate-syms.o
$(obj)/built-in.o
:
ld_flags += -R $(obj)/gate-syms.o
GATECFLAGS_gate-syms.o
=
-r
$(obj)/gate-syms.o
:
$(obj)/gate.lds $(obj)/gate.o FORCE
$(
call
if_changed,gate
)
# gate-data.o contains the gate DSO image as data in section .data.gate.
# We must build gate.so before we can assemble it.
# Note: kbuild does not track this dependency due to usage of .incbin
$(obj)/gate-data.o
:
$(obj)/gate.so
arch/ia64/kernel/asm-offsets.c
View file @
c66b31f3
...
...
@@ -316,5 +316,7 @@ void foo(void)
DEFINE_MAPPED_REG_OFS
(
XSI_BANK1_R16_OFS
,
bank1_regs
[
0
]);
DEFINE_MAPPED_REG_OFS
(
XSI_B0NATS_OFS
,
vbnat
);
DEFINE_MAPPED_REG_OFS
(
XSI_B1NATS_OFS
,
vnat
);
DEFINE_MAPPED_REG_OFS
(
XSI_ITC_OFFSET_OFS
,
itc_offset
);
DEFINE_MAPPED_REG_OFS
(
XSI_ITC_LAST_OFS
,
itc_last
);
#endif
/* CONFIG_XEN */
}
arch/ia64/kernel/efi.c
View file @
c66b31f3
...
...
@@ -456,6 +456,7 @@ efi_map_pal_code (void)
GRANULEROUNDDOWN
((
unsigned
long
)
pal_vaddr
),
pte_val
(
pfn_pte
(
__pa
(
pal_vaddr
)
>>
PAGE_SHIFT
,
PAGE_KERNEL
)),
IA64_GRANULE_SHIFT
);
paravirt_dv_serialize_data
();
ia64_set_psr
(
psr
);
/* restore psr */
}
...
...
arch/ia64/kernel/entry.S
View file @
c66b31f3
...
...
@@ -735,7 +735,7 @@ GLOBAL_ENTRY(__paravirt_leave_syscall)
__paravirt_work_processed_syscall
:
#ifdef CONFIG_VIRT_CPU_ACCOUNTING
adds
r2
=
PT
(
LOADRS
)+
16
,
r12
(
pUStk
)
mov.m
r22
=
ar
.
itc
//
fetch
time
at
leave
MOV_FROM_ITC
(
pUStk
,
p9
,
r22
,
r19
)
//
fetch
time
at
leave
adds
r18
=
TI_FLAGS
+
IA64_TASK_SIZE
,
r13
;;
(
p6
)
ld4
r31
=[
r18
]
//
load
current_thread_info
()->
flags
...
...
@@ -984,7 +984,7 @@ GLOBAL_ENTRY(__paravirt_leave_kernel)
#ifdef CONFIG_VIRT_CPU_ACCOUNTING
.
pred.rel.mutex
pUStk
,
pKStk
MOV_FROM_PSR
(
pKStk
,
r22
,
r29
)
//
M2
read
PSR
now
that
interrupts
are
disabled
(
pUStk
)
mov.m
r22
=
ar
.
itc
//
M
fetch
time
at
leave
MOV_FROM_ITC
(
pUStk
,
p9
,
r22
,
r29
)
//
M
fetch
time
at
leave
nop.i
0
;;
#else
...
...
arch/ia64/kernel/fsys.S
View file @
c66b31f3
...
...
@@ -25,6 +25,7 @@
#include <asm/unistd.h>
#include "entry.h"
#include "paravirt_inst.h"
/*
*
See
Documentation
/
ia64
/
fsys
.
txt
for
details
on
fsyscalls
.
...
...
@@ -279,7 +280,7 @@ ENTRY(fsys_gettimeofday)
(
p9
)
cmp.eq
p13
,
p0
=
0
,
r30
//
if
mmio_ptr
,
clear
p13
jitter
control
;;
.
pred.rel.mutex
p8
,
p9
(
p8
)
mov
r2
=
ar
.
itc
//
CPU_TIMER
.
36
clocks
latency
!!!
MOV_FROM_ITC
(
p8
,
p6
,
r2
,
r10
)
//
CPU_TIMER
.
36
clocks
latency
!!!
(
p9
)
ld8
r2
=
[
r30
]
//
MMIO_TIMER
.
Could
also
have
latency
issues
..
(
p13
)
ld8
r25
=
[
r19
]
//
get
itc_lastcycle
value
ld8
r9
=
[
r22
],
IA64_TIMESPEC_TV_NSEC_OFFSET
//
tv_sec
...
...
@@ -418,7 +419,7 @@ EX(.fail_efault, ld8 r14=[r33]) // r14 <- *set
mov
r17
=(
1
<<
(
SIGKILL
-
1
))
|
(
1
<<
(
SIGSTOP
-
1
))
;;
rsm
psr
.
i
//
mask
interrupt
delivery
RSM_PSR_I
(
p0
,
r18
,
r19
)
//
mask
interrupt
delivery
mov
ar
.
ccv
=
0
andcm
r14
=
r14
,
r17
//
filter
out
SIGKILL
&
SIGSTOP
...
...
@@ -491,7 +492,7 @@ EX(.fail_efault, ld8 r14=[r33]) // r14 <- *set
#ifdef CONFIG_SMP
st4.rel
[
r31
]=
r0
//
release
the
lock
#endif
ssm
psr
.
i
SSM_PSR_I
(
p0
,
p9
,
r31
)
;;
srlz.d
//
ensure
psr
.
i
is
set
again
...
...
@@ -513,7 +514,7 @@ EX(.fail_efault, (p15) st8 [r34]=r3)
#ifdef CONFIG_SMP
st4.rel
[
r31
]=
r0
//
release
the
lock
#endif
ssm
psr
.
i
SSM_PSR_I
(
p0
,
p9
,
r17
)
;;
srlz.d
br.sptk.many
fsys_fallback_syscall
//
with
signal
pending
,
do
the
heavy
-
weight
syscall
...
...
@@ -521,7 +522,7 @@ EX(.fail_efault, (p15) st8 [r34]=r3)
#ifdef CONFIG_SMP
.
lock_contention
:
/
*
Rather
than
spinning
here
,
fall
back
on
doing
a
heavy
-
weight
syscall
.
*/
ssm
psr
.
i
SSM_PSR_I
(
p0
,
p9
,
r17
)
;;
srlz.d
br.sptk.many
fsys_fallback_syscall
...
...
@@ -592,17 +593,17 @@ ENTRY(fsys_fallback_syscall)
adds
r17
=-
1024
,
r15
movl
r14
=
sys_call_table
;;
rsm
psr
.
i
RSM_PSR_I
(
p0
,
r26
,
r27
)
shladd
r18
=
r17
,
3
,
r14
;;
ld8
r18
=[
r18
]
//
load
normal
(
heavy
-
weight
)
syscall
entry
-
point
mov
r29
=
psr
//
read
psr
(
12
cyc
load
latency
)
MOV_FROM_PSR
(
p0
,
r29
,
r26
)
//
read
psr
(
12
cyc
load
latency
)
mov
r27
=
ar
.
rsc
mov
r21
=
ar
.
fpsr
mov
r26
=
ar
.
pfs
END
(
fsys_fallback_syscall
)
/
*
FALL
THROUGH
*/
GLOBAL_ENTRY
(
fsys_bubble_down
)
GLOBAL_ENTRY
(
paravirt_
fsys_bubble_down
)
.
prologue
.
altrp
b6
.
body
...
...
@@ -640,7 +641,7 @@ GLOBAL_ENTRY(fsys_bubble_down)
*
*
PSR
.
BE
:
already
is
turned
off
in
__kernel_syscall_via_epc
()
*
PSR
.
AC
:
don
't care (kernel normally turns PSR.AC on)
*
PSR
.
I
:
already
turned
off
by
the
time
fsys_bubble_down
gets
*
PSR
.
I
:
already
turned
off
by
the
time
paravirt_
fsys_bubble_down
gets
*
invoked
*
PSR
.
DFL
:
always
0
(
kernel
never
turns
it
on
)
*
PSR
.
DFH
:
don
't care --- kernel never touches f32-f127 on its own
...
...
@@ -650,7 +651,7 @@ GLOBAL_ENTRY(fsys_bubble_down)
*
PSR
.
DB
:
don
't care --- kernel never enables kernel-level
*
breakpoints
*
PSR
.
TB
:
must
be
0
already
; if it wasn't zero on entry to
*
__kernel_syscall_via_epc
,
the
branch
to
fsys_bubble_down
*
__kernel_syscall_via_epc
,
the
branch
to
paravirt_
fsys_bubble_down
*
will
trigger
a
taken
branch
; the taken-trap-handler then
*
converts
the
syscall
into
a
break
-
based
system
-
call
.
*/
...
...
@@ -683,7 +684,7 @@ GLOBAL_ENTRY(fsys_bubble_down)
;;
mov
ar
.
rsc
=
0
//
M2
set
enforced
lazy
mode
,
pl
0
,
LE
,
loadrs
=
0
#ifdef CONFIG_VIRT_CPU_ACCOUNTING
mov.m
r30
=
ar
.
itc
//
M
get
cycle
for
accounting
MOV_FROM_ITC
(
p0
,
p6
,
r30
,
r23
)
//
M
get
cycle
for
accounting
#else
nop.m
0
#endif
...
...
@@ -734,21 +735,21 @@ GLOBAL_ENTRY(fsys_bubble_down)
mov
rp
=
r14
//
I0
set
the
real
return
addr
and
r3
=
_TIF_SYSCALL_TRACEAUDIT
,
r3
//
A
;;
ssm
psr
.
i
//
M2
we
're on kernel stacks now, reenable irqs
SSM_PSR_I
(
p0
,
p6
,
r22
)
//
M2
we
're on kernel stacks now, reenable irqs
cmp.eq
p8
,
p0
=
r3
,
r0
//
A
(
p10
)
br.cond.spnt.many
ia64_ret_from_syscall
//
B
return
if
bad
call
-
frame
or
r15
is
a
NaT
nop.m
0
(
p8
)
br.call.sptk.many
b6
=
b6
//
B
(
ignore
return
address
)
br.cond.spnt
ia64_trace_syscall
//
B
END
(
fsys_bubble_down
)
END
(
paravirt_
fsys_bubble_down
)
.
rodata
.
align
8
.
globl
fsyscall_table
.
globl
paravirt_
fsyscall_table
data8
fsys_bubble_down
fsyscall_table
:
data8
paravirt_
fsys_bubble_down
paravirt_
fsyscall_table
:
data8
fsys_ni_syscall
data8
0
//
exit
//
1025
data8
0
//
read
...
...
@@ -1033,4 +1034,4 @@ fsyscall_table:
//
fill
in
zeros
for
the
remaining
entries
.
zero
:
.
space
fsyscall_table
+
8
*
NR_syscalls
-
.
zero
,
0
.
space
paravirt_
fsyscall_table
+
8
*
NR_syscalls
-
.
zero
,
0
arch/ia64/kernel/gate.S
View file @
c66b31f3
...
...
@@ -13,6 +13,7 @@
#include <asm/sigcontext.h>
#include <asm/system.h>
#include <asm/unistd.h>
#include "paravirt_inst.h"
/*
*
We
can
't easily refer to symbols inside the kernel. To avoid full runtime relocation,
...
...
@@ -48,87 +49,6 @@ GLOBAL_ENTRY(__kernel_syscall_via_break)
}
END
(
__kernel_syscall_via_break
)
/*
*
On
entry
:
*
r11
=
saved
ar
.
pfs
*
r15
=
system
call
#
*
b0
=
saved
return
address
*
b6
=
return
address
*
On
exit
:
*
r11
=
saved
ar
.
pfs
*
r15
=
system
call
#
*
b0
=
saved
return
address
*
all
other
"scratch"
registers
:
undefined
*
all
"preserved"
registers
:
same
as
on
entry
*/
GLOBAL_ENTRY
(
__kernel_syscall_via_epc
)
.
prologue
.
altrp
b6
.
body
{
/
*
*
Note
:
the
kernel
cannot
assume
that
the
first
two
instructions
in
this
*
bundle
get
executed
.
The
remaining
code
must
be
safe
even
if
*
they
do
not
get
executed
.
*/
adds
r17
=-
1024
,
r15
//
A
mov
r10
=
0
//
A
default
to
successful
syscall
execution
epc
//
B
causes
split
-
issue
}
;;
rsm
psr
.
be
|
psr
.
i
//
M2
(
5
cyc
to
srlz
.
d
)
LOAD_FSYSCALL_TABLE
(
r14
)
//
X
;;
mov
r16
=
IA64_KR
(
CURRENT
)
//
M2
(
12
cyc
)
shladd
r18
=
r17
,
3
,
r14
//
A
mov
r19
=
NR_syscalls
-
1
//
A
;;
lfetch
[
r18
]
//
M0
|
1
mov
r29
=
psr
//
M2
(
12
cyc
)
//
If
r17
is
a
NaT
,
p6
will
be
zero
cmp.geu
p6
,
p7
=
r19
,
r17
//
A
(
sysnr
>
0
&&
sysnr
<
1024
+
NR_syscalls
)?
;;
mov
r21
=
ar
.
fpsr
//
M2
(
12
cyc
)
tnat.nz
p10
,
p9
=
r15
//
I0
mov.i
r26
=
ar
.
pfs
//
I0
(
would
stall
anyhow
due
to
srlz
.
d
...
)
;;
srlz.d
//
M0
(
forces
split
-
issue
)
ensure
PSR
.
BE
==
0
(
p6
)
ld8
r18
=[
r18
]
//
M0
|
1
nop.i
0
;;
nop.m
0
(
p6
)
tbit.z.unc
p8
,
p0
=
r18
,
0
//
I0
(
dual
-
issues
with
"mov b7=r18"
!)
nop.i
0
;;
(
p8
)
ssm
psr
.
i
(
p6
)
mov
b7
=
r18
//
I0
(
p8
)
br.dptk.many
b7
//
B
mov
r27
=
ar
.
rsc
//
M2
(
12
cyc
)
/*
*
brl
.
cond
doesn
't work as intended because the linker would convert this branch
*
into
a
branch
to
a
PLT
.
Perhaps
there
will
be
a
way
to
avoid
this
with
some
*
future
version
of
the
linker
.
In
the
meantime
,
we
just
use
an
indirect
branch
*
instead
.
*/
#ifdef CONFIG_ITANIUM
(
p6
)
add
r14
=-
8
,
r14
//
r14
<-
addr
of
fsys_bubble_down
entry
;;
(
p6
)
ld8
r14
=[
r14
]
//
r14
<-
fsys_bubble_down
;;
(
p6
)
mov
b7
=
r14
(
p6
)
br.sptk.many
b7
#else
BRL_COND_FSYS_BUBBLE_DOWN
(
p6
)
#endif
ssm
psr
.
i
mov
r10
=-
1
(
p10
)
mov
r8
=
EINVAL
(
p9
)
mov
r8
=
ENOSYS
FSYS_RETURN
END
(
__kernel_syscall_via_epc
)
# define ARG0_OFF (16 + IA64_SIGFRAME_ARG0_OFFSET)
# define ARG1_OFF (16 + IA64_SIGFRAME_ARG1_OFFSET)
# define ARG2_OFF (16 + IA64_SIGFRAME_ARG2_OFFSET)
...
...
@@ -374,3 +294,92 @@ restore_rbs:
//
invala
not
necessary
as
that
will
happen
when
returning
to
user
-
mode
br.cond.sptk
back_from_restore_rbs
END
(
__kernel_sigtramp
)
/*
*
On
entry
:
*
r11
=
saved
ar
.
pfs
*
r15
=
system
call
#
*
b0
=
saved
return
address
*
b6
=
return
address
*
On
exit
:
*
r11
=
saved
ar
.
pfs
*
r15
=
system
call
#
*
b0
=
saved
return
address
*
all
other
"scratch"
registers
:
undefined
*
all
"preserved"
registers
:
same
as
on
entry
*/
GLOBAL_ENTRY
(
__kernel_syscall_via_epc
)
.
prologue
.
altrp
b6
.
body
{
/
*
*
Note
:
the
kernel
cannot
assume
that
the
first
two
instructions
in
this
*
bundle
get
executed
.
The
remaining
code
must
be
safe
even
if
*
they
do
not
get
executed
.
*/
adds
r17
=-
1024
,
r15
//
A
mov
r10
=
0
//
A
default
to
successful
syscall
execution
epc
//
B
causes
split
-
issue
}
;;
RSM_PSR_BE_I
(
r20
,
r22
)
//
M2
(
5
cyc
to
srlz
.
d
)
LOAD_FSYSCALL_TABLE
(
r14
)
//
X
;;
mov
r16
=
IA64_KR
(
CURRENT
)
//
M2
(
12
cyc
)
shladd
r18
=
r17
,
3
,
r14
//
A
mov
r19
=
NR_syscalls
-
1
//
A
;;
lfetch
[
r18
]
//
M0
|
1
MOV_FROM_PSR
(
p0
,
r29
,
r8
)
//
M2
(
12
cyc
)
//
If
r17
is
a
NaT
,
p6
will
be
zero
cmp.geu
p6
,
p7
=
r19
,
r17
//
A
(
sysnr
>
0
&&
sysnr
<
1024
+
NR_syscalls
)?
;;
mov
r21
=
ar
.
fpsr
//
M2
(
12
cyc
)
tnat.nz
p10
,
p9
=
r15
//
I0
mov.i
r26
=
ar
.
pfs
//
I0
(
would
stall
anyhow
due
to
srlz
.
d
...
)
;;
srlz.d
//
M0
(
forces
split
-
issue
)
ensure
PSR
.
BE
==
0
(
p6
)
ld8
r18
=[
r18
]
//
M0
|
1
nop.i
0
;;
nop.m
0
(
p6
)
tbit.z.unc
p8
,
p0
=
r18
,
0
//
I0
(
dual
-
issues
with
"mov b7=r18"
!)
nop.i
0
;;
SSM_PSR_I
(
p8
,
p14
,
r25
)
(
p6
)
mov
b7
=
r18
//
I0
(
p8
)
br.dptk.many
b7
//
B
mov
r27
=
ar
.
rsc
//
M2
(
12
cyc
)
/*
*
brl
.
cond
doesn
't work as intended because the linker would convert this branch
*
into
a
branch
to
a
PLT
.
Perhaps
there
will
be
a
way
to
avoid
this
with
some
*
future
version
of
the
linker
.
In
the
meantime
,
we
just
use
an
indirect
branch
*
instead
.
*/
#ifdef CONFIG_ITANIUM
(
p6
)
add
r14
=-
8
,
r14
//
r14
<-
addr
of
fsys_bubble_down
entry
;;
(
p6
)
ld8
r14
=[
r14
]
//
r14
<-
fsys_bubble_down
;;
(
p6
)
mov
b7
=
r14
(
p6
)
br.sptk.many
b7
#else
BRL_COND_FSYS_BUBBLE_DOWN
(
p6
)
#endif
SSM_PSR_I
(
p0
,
p14
,
r10
)
mov
r10
=-
1
(
p10
)
mov
r8
=
EINVAL
(
p9
)
mov
r8
=
ENOSYS
FSYS_RETURN
#ifdef CONFIG_PARAVIRT
/
*
*
padd
to
make
the
size
of
this
symbol
constant
*
independent
of
paravirtualization
.
*/
.
align
PAGE_SIZE
/
8
#endif
END
(
__kernel_syscall_via_epc
)
arch/ia64/kernel/gate.lds.S
View file @
c66b31f3
...
...
@@ -7,6 +7,7 @@
#include <asm/system.h>
#include "paravirt_patchlist.h"
SECTIONS
{
...
...
@@ -33,21 +34,21 @@ SECTIONS
.
=
GATE_ADDR
+
0x600
;
.
data.patch
:
{
__start_gate_mckinley_e9_patchlist
=
.
;
__
paravirt_
start_gate_mckinley_e9_patchlist
=
.
;
*(.
data.patch.
mckinley_e9
)
__end_gate_mckinley_e9_patchlist
=
.
;
__
paravirt_
end_gate_mckinley_e9_patchlist
=
.
;
__start_gate_vtop_patchlist
=
.
;
__
paravirt_
start_gate_vtop_patchlist
=
.
;
*(.
data.patch.vtop
)
__end_gate_vtop_patchlist
=
.
;
__
paravirt_
end_gate_vtop_patchlist
=
.
;
__start_gate_fsyscall_patchlist
=
.
;
__
paravirt_
start_gate_fsyscall_patchlist
=
.
;
*(.
data.patch.
fsyscall_table
)
__end_gate_fsyscall_patchlist
=
.
;
__
paravirt_
end_gate_fsyscall_patchlist
=
.
;
__start_gate_brl_fsys_bubble_down_patchlist
=
.
;
__
paravirt_
start_gate_brl_fsys_bubble_down_patchlist
=
.
;
*(.
data.patch.
brl_fsys_bubble_down
)
__end_gate_brl_fsys_bubble_down_patchlist
=
.
;
__
paravirt_
end_gate_brl_fsys_bubble_down_patchlist
=
.
;
}
:
readable
.
IA_64
.
unwind_info
:
{
*(
.
IA_64
.
unwind_info
*)
}
...
...
arch/ia64/kernel/head.S
View file @
c66b31f3
...
...
@@ -1050,7 +1050,7 @@ END(ia64_delay_loop)
*
except
that
the
multiplication
and
the
shift
are
done
with
128
-
bit
*
intermediate
precision
so
that
we
can
produce
a
full
64
-
bit
result
.
*/
GLOBAL_ENTRY
(
sched_clock
)
GLOBAL_ENTRY
(
ia64_native_
sched_clock
)
addl
r8
=
THIS_CPU
(
cpu_info
)
+
IA64_CPUINFO_NSEC_PER_CYC_OFFSET
,
r0
mov.m
r9
=
ar
.
itc
//
fetch
cycle
-
counter
(
35
cyc
)
;;
...
...
@@ -1066,7 +1066,13 @@ GLOBAL_ENTRY(sched_clock)
;;
shrp
r8
=
r9
,
r8
,
IA64_NSEC_PER_CYC_SHIFT
br.ret.sptk.many
rp
END
(
sched_clock
)
END
(
ia64_native_sched_clock
)
#ifndef CONFIG_PARAVIRT
//
unsigned
long
long
//
sched_clock
(
void
)
__attribute__
((
alias
(
"ia64_native_sched_clock"
)))
;
.
global
sched_clock
sched_clock
=
ia64_native_sched_clock
#endif
#ifdef CONFIG_VIRT_CPU_ACCOUNTING
GLOBAL_ENTRY
(
cycle_to_cputime
)
...
...
arch/ia64/kernel/ivt.S
View file @
c66b31f3
...
...
@@ -804,7 +804,7 @@ ENTRY(break_fault)
///////////////////////////////////////////////////////////////////////
st1
[
r16
]=
r0
//
M2
|
3
clear
current
->
thread
.
on_ustack
flag
#ifdef CONFIG_VIRT_CPU_ACCOUNTING
mov.m
r30
=
ar
.
itc
//
M
get
cycle
for
accounting
MOV_FROM_ITC
(
p0
,
p14
,
r30
,
r18
)
//
M
get
cycle
for
accounting
#else
mov
b6
=
r30
//
I0
setup
syscall
handler
branch
reg
early
#endif
...
...
arch/ia64/kernel/module.c
View file @
c66b31f3
...
...
@@ -446,6 +446,14 @@ module_frob_arch_sections (Elf_Ehdr *ehdr, Elf_Shdr *sechdrs, char *secstrings,
mod
->
arch
.
opd
=
s
;
else
if
(
strcmp
(
".IA_64.unwind"
,
secstrings
+
s
->
sh_name
)
==
0
)
mod
->
arch
.
unwind
=
s
;
#ifdef CONFIG_PARAVIRT
else
if
(
strcmp
(
".paravirt_bundles"
,
secstrings
+
s
->
sh_name
)
==
0
)
mod
->
arch
.
paravirt_bundles
=
s
;
else
if
(
strcmp
(
".paravirt_insts"
,
secstrings
+
s
->
sh_name
)
==
0
)
mod
->
arch
.
paravirt_insts
=
s
;
#endif
if
(
!
mod
->
arch
.
core_plt
||
!
mod
->
arch
.
init_plt
||
!
mod
->
arch
.
got
||
!
mod
->
arch
.
opd
)
{
printk
(
KERN_ERR
"%s: sections missing
\n
"
,
mod
->
name
);
...
...
@@ -921,6 +929,30 @@ module_finalize (const Elf_Ehdr *hdr, const Elf_Shdr *sechdrs, struct module *mo
DEBUGP
(
"%s: init: entry=%p
\n
"
,
__func__
,
mod
->
init
);
if
(
mod
->
arch
.
unwind
)
register_unwind_table
(
mod
);
#ifdef CONFIG_PARAVIRT
if
(
mod
->
arch
.
paravirt_bundles
)
{
struct
paravirt_patch_site_bundle
*
start
=
(
struct
paravirt_patch_site_bundle
*
)
mod
->
arch
.
paravirt_bundles
->
sh_addr
;
struct
paravirt_patch_site_bundle
*
end
=
(
struct
paravirt_patch_site_bundle
*
)
(
mod
->
arch
.
paravirt_bundles
->
sh_addr
+
mod
->
arch
.
paravirt_bundles
->
sh_size
);
paravirt_patch_apply_bundle
(
start
,
end
);
}
if
(
mod
->
arch
.
paravirt_insts
)
{
struct
paravirt_patch_site_inst
*
start
=
(
struct
paravirt_patch_site_inst
*
)
mod
->
arch
.
paravirt_insts
->
sh_addr
;
struct
paravirt_patch_site_inst
*
end
=
(
struct
paravirt_patch_site_inst
*
)
(
mod
->
arch
.
paravirt_insts
->
sh_addr
+
mod
->
arch
.
paravirt_insts
->
sh_size
);
paravirt_patch_apply_inst
(
start
,
end
);
}
#endif
return
0
;
}
...
...
arch/ia64/kernel/paravirt.c
View file @
c66b31f3
...
...
@@ -46,13 +46,23 @@ struct pv_info pv_info = {
* initialization hooks.
*/
struct
pv_init_ops
pv_init_ops
;
static
void
__init
ia64_native_patch_branch
(
unsigned
long
tag
,
unsigned
long
type
);
struct
pv_init_ops
pv_init_ops
=
{
#ifdef ASM_SUPPORTED
.
patch_bundle
=
ia64_native_patch_bundle
,
#endif
.
patch_branch
=
ia64_native_patch_branch
,
};
/***************************************************************************
* pv_cpu_ops
* intrinsics hooks.
*/
#ifndef ASM_SUPPORTED
/* ia64_native_xxx are macros so that we have to make them real functions */
#define DEFINE_VOID_FUNC1(name) \
...
...
@@ -60,7 +70,14 @@ struct pv_init_ops pv_init_ops;
ia64_native_ ## name ## _func(unsigned long arg) \
{ \
ia64_native_ ## name(arg); \
} \
}
#define DEFINE_VOID_FUNC1_VOID(name) \
static void \
ia64_native_ ## name ## _func(void *arg) \
{ \
ia64_native_ ## name(arg); \
}
#define DEFINE_VOID_FUNC2(name) \
static void \
...
...
@@ -68,7 +85,7 @@ struct pv_init_ops pv_init_ops;
unsigned long arg1) \
{ \
ia64_native_ ## name(arg0, arg1); \
}
\
}
#define DEFINE_FUNC0(name) \
static unsigned long \
...
...
@@ -84,7 +101,7 @@ struct pv_init_ops pv_init_ops;
return ia64_native_ ## name(arg); \
} \
DEFINE_VOID_FUNC1
(
fc
);
DEFINE_VOID_FUNC1
_VOID
(
fc
);
DEFINE_VOID_FUNC1
(
intrin_local_irq_restore
);
DEFINE_VOID_FUNC2
(
ptcga
);
...
...
@@ -274,6 +291,266 @@ ia64_native_setreg_func(int regnum, unsigned long val)
break
;
}
}
#else
#define __DEFINE_FUNC(name, code) \
extern const char ia64_native_ ## name ## _direct_start[]; \
extern const char ia64_native_ ## name ## _direct_end[]; \
asm (".align 32\n" \
".proc ia64_native_" #name "_func\n" \
"ia64_native_" #name "_func:\n" \
"ia64_native_" #name "_direct_start:\n" \
code \
"ia64_native_" #name "_direct_end:\n" \
"br.cond.sptk.many b6\n" \
".endp ia64_native_" #name "_func\n")
#define DEFINE_VOID_FUNC0(name, code) \
extern void \
ia64_native_ ## name ## _func(void); \
__DEFINE_FUNC(name, code)
#define DEFINE_VOID_FUNC1(name, code) \
extern void \
ia64_native_ ## name ## _func(unsigned long arg); \
__DEFINE_FUNC(name, code)
#define DEFINE_VOID_FUNC1_VOID(name, code) \
extern void \
ia64_native_ ## name ## _func(void *arg); \
__DEFINE_FUNC(name, code)
#define DEFINE_VOID_FUNC2(name, code) \
extern void \
ia64_native_ ## name ## _func(unsigned long arg0, \
unsigned long arg1); \
__DEFINE_FUNC(name, code)
#define DEFINE_FUNC0(name, code) \
extern unsigned long \
ia64_native_ ## name ## _func(void); \
__DEFINE_FUNC(name, code)
#define DEFINE_FUNC1(name, type, code) \
extern unsigned long \
ia64_native_ ## name ## _func(type arg); \
__DEFINE_FUNC(name, code)
DEFINE_VOID_FUNC1_VOID
(
fc
,
"fc r8
\n
"
);
DEFINE_VOID_FUNC1
(
intrin_local_irq_restore
,
";;
\n
"
" cmp.ne p6, p7 = r8, r0
\n
"
";;
\n
"
"(p6) ssm psr.i
\n
"
"(p7) rsm psr.i
\n
"
";;
\n
"
"(p6) srlz.d
\n
"
);
DEFINE_VOID_FUNC2
(
ptcga
,
"ptc.ga r8, r9
\n
"
);
DEFINE_VOID_FUNC2
(
set_rr
,
"mov rr[r8] = r9
\n
"
);
/* ia64_native_getreg(_IA64_REG_PSR) & IA64_PSR_I */
DEFINE_FUNC0
(
get_psr_i
,
"mov r2 = "
__stringify
(
1
<<
IA64_PSR_I_BIT
)
"
\n
"
"mov r8 = psr
\n
"
";;
\n
"
"and r8 = r2, r8
\n
"
);
DEFINE_FUNC1
(
thash
,
unsigned
long
,
"thash r8 = r8
\n
"
);
DEFINE_FUNC1
(
get_cpuid
,
int
,
"mov r8 = cpuid[r8]
\n
"
);
DEFINE_FUNC1
(
get_pmd
,
int
,
"mov r8 = pmd[r8]
\n
"
);
DEFINE_FUNC1
(
get_rr
,
unsigned
long
,
"mov r8 = rr[r8]
\n
"
);
DEFINE_VOID_FUNC0
(
ssm_i
,
"ssm psr.i
\n
"
);
DEFINE_VOID_FUNC0
(
rsm_i
,
"rsm psr.i
\n
"
);
extern
void
ia64_native_set_rr0_to_rr4_func
(
unsigned
long
val0
,
unsigned
long
val1
,
unsigned
long
val2
,
unsigned
long
val3
,
unsigned
long
val4
);
__DEFINE_FUNC
(
set_rr0_to_rr4
,
"mov rr[r0] = r8
\n
"
"movl r2 = 0x2000000000000000
\n
"
";;
\n
"
"mov rr[r2] = r9
\n
"
"shl r3 = r2, 1
\n
"
/* movl r3 = 0x4000000000000000 */
";;
\n
"
"add r2 = r2, r3
\n
"
/* movl r2 = 0x6000000000000000 */
"mov rr[r3] = r10
\n
"
";;
\n
"
"mov rr[r2] = r11
\n
"
"shl r3 = r3, 1
\n
"
/* movl r3 = 0x8000000000000000 */
";;
\n
"
"mov rr[r3] = r14
\n
"
);
extern
unsigned
long
ia64_native_getreg_func
(
int
regnum
);
asm
(
".global ia64_native_getreg_func
\n
"
);
#define __DEFINE_GET_REG(id, reg) \
"mov r2 = " __stringify(_IA64_REG_ ## id) "\n" \
";;\n" \
"cmp.eq p6, p0 = r2, r8\n" \
";;\n" \
"(p6) mov r8 = " #reg "\n" \
"(p6) br.cond.sptk.many b6\n" \
";;\n"
#define __DEFINE_GET_AR(id, reg) __DEFINE_GET_REG(AR_ ## id, ar.reg)
#define __DEFINE_GET_CR(id, reg) __DEFINE_GET_REG(CR_ ## id, cr.reg)
__DEFINE_FUNC
(
getreg
,
__DEFINE_GET_REG
(
GP
,
gp
)
/*__DEFINE_GET_REG(IP, ip)*/
/* returned ip value shouldn't be constant */
__DEFINE_GET_REG
(
PSR
,
psr
)
__DEFINE_GET_REG
(
TP
,
tp
)
__DEFINE_GET_REG
(
SP
,
sp
)
__DEFINE_GET_REG
(
AR_KR0
,
ar0
)
__DEFINE_GET_REG
(
AR_KR1
,
ar1
)
__DEFINE_GET_REG
(
AR_KR2
,
ar2
)
__DEFINE_GET_REG
(
AR_KR3
,
ar3
)
__DEFINE_GET_REG
(
AR_KR4
,
ar4
)
__DEFINE_GET_REG
(
AR_KR5
,
ar5
)
__DEFINE_GET_REG
(
AR_KR6
,
ar6
)
__DEFINE_GET_REG
(
AR_KR7
,
ar7
)
__DEFINE_GET_AR
(
RSC
,
rsc
)
__DEFINE_GET_AR
(
BSP
,
bsp
)
__DEFINE_GET_AR
(
BSPSTORE
,
bspstore
)
__DEFINE_GET_AR
(
RNAT
,
rnat
)
__DEFINE_GET_AR
(
FCR
,
fcr
)
__DEFINE_GET_AR
(
EFLAG
,
eflag
)
__DEFINE_GET_AR
(
CSD
,
csd
)
__DEFINE_GET_AR
(
SSD
,
ssd
)
__DEFINE_GET_REG
(
AR_CFLAG
,
ar27
)
__DEFINE_GET_AR
(
FSR
,
fsr
)
__DEFINE_GET_AR
(
FIR
,
fir
)
__DEFINE_GET_AR
(
FDR
,
fdr
)
__DEFINE_GET_AR
(
CCV
,
ccv
)
__DEFINE_GET_AR
(
UNAT
,
unat
)
__DEFINE_GET_AR
(
FPSR
,
fpsr
)
__DEFINE_GET_AR
(
ITC
,
itc
)
__DEFINE_GET_AR
(
PFS
,
pfs
)
__DEFINE_GET_AR
(
LC
,
lc
)
__DEFINE_GET_AR
(
EC
,
ec
)
__DEFINE_GET_CR
(
DCR
,
dcr
)
__DEFINE_GET_CR
(
ITM
,
itm
)
__DEFINE_GET_CR
(
IVA
,
iva
)
__DEFINE_GET_CR
(
PTA
,
pta
)
__DEFINE_GET_CR
(
IPSR
,
ipsr
)
__DEFINE_GET_CR
(
ISR
,
isr
)
__DEFINE_GET_CR
(
IIP
,
iip
)
__DEFINE_GET_CR
(
IFA
,
ifa
)
__DEFINE_GET_CR
(
ITIR
,
itir
)
__DEFINE_GET_CR
(
IIPA
,
iipa
)
__DEFINE_GET_CR
(
IFS
,
ifs
)
__DEFINE_GET_CR
(
IIM
,
iim
)
__DEFINE_GET_CR
(
IHA
,
iha
)
__DEFINE_GET_CR
(
LID
,
lid
)
__DEFINE_GET_CR
(
IVR
,
ivr
)
__DEFINE_GET_CR
(
TPR
,
tpr
)
__DEFINE_GET_CR
(
EOI
,
eoi
)
__DEFINE_GET_CR
(
IRR0
,
irr0
)
__DEFINE_GET_CR
(
IRR1
,
irr1
)
__DEFINE_GET_CR
(
IRR2
,
irr2
)
__DEFINE_GET_CR
(
IRR3
,
irr3
)
__DEFINE_GET_CR
(
ITV
,
itv
)
__DEFINE_GET_CR
(
PMV
,
pmv
)
__DEFINE_GET_CR
(
CMCV
,
cmcv
)
__DEFINE_GET_CR
(
LRR0
,
lrr0
)
__DEFINE_GET_CR
(
LRR1
,
lrr1
)
"mov r8 = -1
\n
"
/* unsupported case */
);
extern
void
ia64_native_setreg_func
(
int
regnum
,
unsigned
long
val
);
asm
(
".global ia64_native_setreg_func
\n
"
);
#define __DEFINE_SET_REG(id, reg) \
"mov r2 = " __stringify(_IA64_REG_ ## id) "\n" \
";;\n" \
"cmp.eq p6, p0 = r2, r9\n" \
";;\n" \
"(p6) mov " #reg " = r8\n" \
"(p6) br.cond.sptk.many b6\n" \
";;\n"
#define __DEFINE_SET_AR(id, reg) __DEFINE_SET_REG(AR_ ## id, ar.reg)
#define __DEFINE_SET_CR(id, reg) __DEFINE_SET_REG(CR_ ## id, cr.reg)
__DEFINE_FUNC
(
setreg
,
"mov r2 = "
__stringify
(
_IA64_REG_PSR_L
)
"
\n
"
";;
\n
"
"cmp.eq p6, p0 = r2, r9
\n
"
";;
\n
"
"(p6) mov psr.l = r8
\n
"
#ifdef HAVE_SERIALIZE_DIRECTIVE
".serialize.data
\n
"
#endif
"(p6) br.cond.sptk.many b6
\n
"
__DEFINE_SET_REG
(
GP
,
gp
)
__DEFINE_SET_REG
(
SP
,
sp
)
__DEFINE_SET_REG
(
AR_KR0
,
ar0
)
__DEFINE_SET_REG
(
AR_KR1
,
ar1
)
__DEFINE_SET_REG
(
AR_KR2
,
ar2
)
__DEFINE_SET_REG
(
AR_KR3
,
ar3
)
__DEFINE_SET_REG
(
AR_KR4
,
ar4
)
__DEFINE_SET_REG
(
AR_KR5
,
ar5
)
__DEFINE_SET_REG
(
AR_KR6
,
ar6
)
__DEFINE_SET_REG
(
AR_KR7
,
ar7
)
__DEFINE_SET_AR
(
RSC
,
rsc
)
__DEFINE_SET_AR
(
BSP
,
bsp
)
__DEFINE_SET_AR
(
BSPSTORE
,
bspstore
)
__DEFINE_SET_AR
(
RNAT
,
rnat
)
__DEFINE_SET_AR
(
FCR
,
fcr
)
__DEFINE_SET_AR
(
EFLAG
,
eflag
)
__DEFINE_SET_AR
(
CSD
,
csd
)
__DEFINE_SET_AR
(
SSD
,
ssd
)
__DEFINE_SET_REG
(
AR_CFLAG
,
ar27
)
__DEFINE_SET_AR
(
FSR
,
fsr
)
__DEFINE_SET_AR
(
FIR
,
fir
)
__DEFINE_SET_AR
(
FDR
,
fdr
)
__DEFINE_SET_AR
(
CCV
,
ccv
)
__DEFINE_SET_AR
(
UNAT
,
unat
)
__DEFINE_SET_AR
(
FPSR
,
fpsr
)
__DEFINE_SET_AR
(
ITC
,
itc
)
__DEFINE_SET_AR
(
PFS
,
pfs
)
__DEFINE_SET_AR
(
LC
,
lc
)
__DEFINE_SET_AR
(
EC
,
ec
)
__DEFINE_SET_CR
(
DCR
,
dcr
)
__DEFINE_SET_CR
(
ITM
,
itm
)
__DEFINE_SET_CR
(
IVA
,
iva
)
__DEFINE_SET_CR
(
PTA
,
pta
)
__DEFINE_SET_CR
(
IPSR
,
ipsr
)
__DEFINE_SET_CR
(
ISR
,
isr
)
__DEFINE_SET_CR
(
IIP
,
iip
)
__DEFINE_SET_CR
(
IFA
,
ifa
)
__DEFINE_SET_CR
(
ITIR
,
itir
)
__DEFINE_SET_CR
(
IIPA
,
iipa
)
__DEFINE_SET_CR
(
IFS
,
ifs
)
__DEFINE_SET_CR
(
IIM
,
iim
)
__DEFINE_SET_CR
(
IHA
,
iha
)
__DEFINE_SET_CR
(
LID
,
lid
)
__DEFINE_SET_CR
(
IVR
,
ivr
)
__DEFINE_SET_CR
(
TPR
,
tpr
)
__DEFINE_SET_CR
(
EOI
,
eoi
)
__DEFINE_SET_CR
(
IRR0
,
irr0
)
__DEFINE_SET_CR
(
IRR1
,
irr1
)
__DEFINE_SET_CR
(
IRR2
,
irr2
)
__DEFINE_SET_CR
(
IRR3
,
irr3
)
__DEFINE_SET_CR
(
ITV
,
itv
)
__DEFINE_SET_CR
(
PMV
,
pmv
)
__DEFINE_SET_CR
(
CMCV
,
cmcv
)
__DEFINE_SET_CR
(
LRR0
,
lrr0
)
__DEFINE_SET_CR
(
LRR1
,
lrr1
)
);
#endif
struct
pv_cpu_ops
pv_cpu_ops
=
{
.
fc
=
ia64_native_fc_func
,
...
...
@@ -366,4 +643,258 @@ ia64_native_do_steal_accounting(unsigned long *new_itm)
struct
pv_time_ops
pv_time_ops
=
{
.
do_steal_accounting
=
ia64_native_do_steal_accounting
,
.
sched_clock
=
ia64_native_sched_clock
,
};
/***************************************************************************
* binary pacthing
* pv_init_ops.patch_bundle
*/
#ifdef ASM_SUPPORTED
#define IA64_NATIVE_PATCH_DEFINE_GET_REG(name, reg) \
__DEFINE_FUNC(get_ ## name, \
";;\n" \
"mov r8 = " #reg "\n" \
";;\n")
#define IA64_NATIVE_PATCH_DEFINE_SET_REG(name, reg) \
__DEFINE_FUNC(set_ ## name, \
";;\n" \
"mov " #reg " = r8\n" \
";;\n")
#define IA64_NATIVE_PATCH_DEFINE_REG(name, reg) \
IA64_NATIVE_PATCH_DEFINE_GET_REG(name, reg); \
IA64_NATIVE_PATCH_DEFINE_SET_REG(name, reg) \
#define IA64_NATIVE_PATCH_DEFINE_AR(name, reg) \
IA64_NATIVE_PATCH_DEFINE_REG(ar_ ## name, ar.reg)
#define IA64_NATIVE_PATCH_DEFINE_CR(name, reg) \
IA64_NATIVE_PATCH_DEFINE_REG(cr_ ## name, cr.reg)
IA64_NATIVE_PATCH_DEFINE_GET_REG
(
psr
,
psr
);
IA64_NATIVE_PATCH_DEFINE_GET_REG
(
tp
,
tp
);
/* IA64_NATIVE_PATCH_DEFINE_SET_REG(psr_l, psr.l); */
__DEFINE_FUNC
(
set_psr_l
,
";;
\n
"
"mov psr.l = r8
\n
"
#ifdef HAVE_SERIALIZE_DIRECTIVE
".serialize.data
\n
"
#endif
";;
\n
"
);
IA64_NATIVE_PATCH_DEFINE_REG
(
gp
,
gp
);
IA64_NATIVE_PATCH_DEFINE_REG
(
sp
,
sp
);
IA64_NATIVE_PATCH_DEFINE_REG
(
kr0
,
ar0
);
IA64_NATIVE_PATCH_DEFINE_REG
(
kr1
,
ar1
);
IA64_NATIVE_PATCH_DEFINE_REG
(
kr2
,
ar2
);
IA64_NATIVE_PATCH_DEFINE_REG
(
kr3
,
ar3
);
IA64_NATIVE_PATCH_DEFINE_REG
(
kr4
,
ar4
);
IA64_NATIVE_PATCH_DEFINE_REG
(
kr5
,
ar5
);
IA64_NATIVE_PATCH_DEFINE_REG
(
kr6
,
ar6
);
IA64_NATIVE_PATCH_DEFINE_REG
(
kr7
,
ar7
);
IA64_NATIVE_PATCH_DEFINE_AR
(
rsc
,
rsc
);
IA64_NATIVE_PATCH_DEFINE_AR
(
bsp
,
bsp
);
IA64_NATIVE_PATCH_DEFINE_AR
(
bspstore
,
bspstore
);
IA64_NATIVE_PATCH_DEFINE_AR
(
rnat
,
rnat
);
IA64_NATIVE_PATCH_DEFINE_AR
(
fcr
,
fcr
);
IA64_NATIVE_PATCH_DEFINE_AR
(
eflag
,
eflag
);
IA64_NATIVE_PATCH_DEFINE_AR
(
csd
,
csd
);
IA64_NATIVE_PATCH_DEFINE_AR
(
ssd
,
ssd
);
IA64_NATIVE_PATCH_DEFINE_REG
(
ar27
,
ar27
);
IA64_NATIVE_PATCH_DEFINE_AR
(
fsr
,
fsr
);
IA64_NATIVE_PATCH_DEFINE_AR
(
fir
,
fir
);
IA64_NATIVE_PATCH_DEFINE_AR
(
fdr
,
fdr
);
IA64_NATIVE_PATCH_DEFINE_AR
(
ccv
,
ccv
);
IA64_NATIVE_PATCH_DEFINE_AR
(
unat
,
unat
);
IA64_NATIVE_PATCH_DEFINE_AR
(
fpsr
,
fpsr
);
IA64_NATIVE_PATCH_DEFINE_AR
(
itc
,
itc
);
IA64_NATIVE_PATCH_DEFINE_AR
(
pfs
,
pfs
);
IA64_NATIVE_PATCH_DEFINE_AR
(
lc
,
lc
);
IA64_NATIVE_PATCH_DEFINE_AR
(
ec
,
ec
);
IA64_NATIVE_PATCH_DEFINE_CR
(
dcr
,
dcr
);
IA64_NATIVE_PATCH_DEFINE_CR
(
itm
,
itm
);
IA64_NATIVE_PATCH_DEFINE_CR
(
iva
,
iva
);
IA64_NATIVE_PATCH_DEFINE_CR
(
pta
,
pta
);
IA64_NATIVE_PATCH_DEFINE_CR
(
ipsr
,
ipsr
);
IA64_NATIVE_PATCH_DEFINE_CR
(
isr
,
isr
);
IA64_NATIVE_PATCH_DEFINE_CR
(
iip
,
iip
);
IA64_NATIVE_PATCH_DEFINE_CR
(
ifa
,
ifa
);
IA64_NATIVE_PATCH_DEFINE_CR
(
itir
,
itir
);
IA64_NATIVE_PATCH_DEFINE_CR
(
iipa
,
iipa
);
IA64_NATIVE_PATCH_DEFINE_CR
(
ifs
,
ifs
);
IA64_NATIVE_PATCH_DEFINE_CR
(
iim
,
iim
);
IA64_NATIVE_PATCH_DEFINE_CR
(
iha
,
iha
);
IA64_NATIVE_PATCH_DEFINE_CR
(
lid
,
lid
);
IA64_NATIVE_PATCH_DEFINE_CR
(
ivr
,
ivr
);
IA64_NATIVE_PATCH_DEFINE_CR
(
tpr
,
tpr
);
IA64_NATIVE_PATCH_DEFINE_CR
(
eoi
,
eoi
);
IA64_NATIVE_PATCH_DEFINE_CR
(
irr0
,
irr0
);
IA64_NATIVE_PATCH_DEFINE_CR
(
irr1
,
irr1
);
IA64_NATIVE_PATCH_DEFINE_CR
(
irr2
,
irr2
);
IA64_NATIVE_PATCH_DEFINE_CR
(
irr3
,
irr3
);
IA64_NATIVE_PATCH_DEFINE_CR
(
itv
,
itv
);
IA64_NATIVE_PATCH_DEFINE_CR
(
pmv
,
pmv
);
IA64_NATIVE_PATCH_DEFINE_CR
(
cmcv
,
cmcv
);
IA64_NATIVE_PATCH_DEFINE_CR
(
lrr0
,
lrr0
);
IA64_NATIVE_PATCH_DEFINE_CR
(
lrr1
,
lrr1
);
static
const
struct
paravirt_patch_bundle_elem
ia64_native_patch_bundle_elems
[]
__initdata_or_module
=
{
#define IA64_NATIVE_PATCH_BUNDLE_ELEM(name, type) \
{ \
(void*)ia64_native_ ## name ## _direct_start, \
(void*)ia64_native_ ## name ## _direct_end, \
PARAVIRT_PATCH_TYPE_ ## type, \
}
IA64_NATIVE_PATCH_BUNDLE_ELEM
(
fc
,
FC
),
IA64_NATIVE_PATCH_BUNDLE_ELEM
(
thash
,
THASH
),
IA64_NATIVE_PATCH_BUNDLE_ELEM
(
get_cpuid
,
GET_CPUID
),
IA64_NATIVE_PATCH_BUNDLE_ELEM
(
get_pmd
,
GET_PMD
),
IA64_NATIVE_PATCH_BUNDLE_ELEM
(
ptcga
,
PTCGA
),
IA64_NATIVE_PATCH_BUNDLE_ELEM
(
get_rr
,
GET_RR
),
IA64_NATIVE_PATCH_BUNDLE_ELEM
(
set_rr
,
SET_RR
),
IA64_NATIVE_PATCH_BUNDLE_ELEM
(
set_rr0_to_rr4
,
SET_RR0_TO_RR4
),
IA64_NATIVE_PATCH_BUNDLE_ELEM
(
ssm_i
,
SSM_I
),
IA64_NATIVE_PATCH_BUNDLE_ELEM
(
rsm_i
,
RSM_I
),
IA64_NATIVE_PATCH_BUNDLE_ELEM
(
get_psr_i
,
GET_PSR_I
),
IA64_NATIVE_PATCH_BUNDLE_ELEM
(
intrin_local_irq_restore
,
INTRIN_LOCAL_IRQ_RESTORE
),
#define IA64_NATIVE_PATCH_BUNDLE_ELEM_GETREG(name, reg) \
{ \
(void*)ia64_native_get_ ## name ## _direct_start, \
(void*)ia64_native_get_ ## name ## _direct_end, \
PARAVIRT_PATCH_TYPE_GETREG + _IA64_REG_ ## reg, \
}
#define IA64_NATIVE_PATCH_BUNDLE_ELEM_SETREG(name, reg) \
{ \
(void*)ia64_native_set_ ## name ## _direct_start, \
(void*)ia64_native_set_ ## name ## _direct_end, \
PARAVIRT_PATCH_TYPE_SETREG + _IA64_REG_ ## reg, \
}
#define IA64_NATIVE_PATCH_BUNDLE_ELEM_REG(name, reg) \
IA64_NATIVE_PATCH_BUNDLE_ELEM_GETREG(name, reg), \
IA64_NATIVE_PATCH_BUNDLE_ELEM_SETREG(name, reg) \
#define IA64_NATIVE_PATCH_BUNDLE_ELEM_AR(name, reg) \
IA64_NATIVE_PATCH_BUNDLE_ELEM_REG(ar_ ## name, AR_ ## reg)
#define IA64_NATIVE_PATCH_BUNDLE_ELEM_CR(name, reg) \
IA64_NATIVE_PATCH_BUNDLE_ELEM_REG(cr_ ## name, CR_ ## reg)
IA64_NATIVE_PATCH_BUNDLE_ELEM_GETREG
(
psr
,
PSR
),
IA64_NATIVE_PATCH_BUNDLE_ELEM_GETREG
(
tp
,
TP
),
IA64_NATIVE_PATCH_BUNDLE_ELEM_SETREG
(
psr_l
,
PSR_L
),
IA64_NATIVE_PATCH_BUNDLE_ELEM_REG
(
gp
,
GP
),
IA64_NATIVE_PATCH_BUNDLE_ELEM_REG
(
sp
,
SP
),
IA64_NATIVE_PATCH_BUNDLE_ELEM_REG
(
kr0
,
AR_KR0
),
IA64_NATIVE_PATCH_BUNDLE_ELEM_REG
(
kr1
,
AR_KR1
),
IA64_NATIVE_PATCH_BUNDLE_ELEM_REG
(
kr2
,
AR_KR2
),
IA64_NATIVE_PATCH_BUNDLE_ELEM_REG
(
kr3
,
AR_KR3
),
IA64_NATIVE_PATCH_BUNDLE_ELEM_REG
(
kr4
,
AR_KR4
),
IA64_NATIVE_PATCH_BUNDLE_ELEM_REG
(
kr5
,
AR_KR5
),
IA64_NATIVE_PATCH_BUNDLE_ELEM_REG
(
kr6
,
AR_KR6
),
IA64_NATIVE_PATCH_BUNDLE_ELEM_REG
(
kr7
,
AR_KR7
),
IA64_NATIVE_PATCH_BUNDLE_ELEM_AR
(
rsc
,
RSC
),
IA64_NATIVE_PATCH_BUNDLE_ELEM_AR
(
bsp
,
BSP
),
IA64_NATIVE_PATCH_BUNDLE_ELEM_AR
(
bspstore
,
BSPSTORE
),
IA64_NATIVE_PATCH_BUNDLE_ELEM_AR
(
rnat
,
RNAT
),
IA64_NATIVE_PATCH_BUNDLE_ELEM_AR
(
fcr
,
FCR
),
IA64_NATIVE_PATCH_BUNDLE_ELEM_AR
(
eflag
,
EFLAG
),
IA64_NATIVE_PATCH_BUNDLE_ELEM_AR
(
csd
,
CSD
),
IA64_NATIVE_PATCH_BUNDLE_ELEM_AR
(
ssd
,
SSD
),
IA64_NATIVE_PATCH_BUNDLE_ELEM_REG
(
ar27
,
AR_CFLAG
),
IA64_NATIVE_PATCH_BUNDLE_ELEM_AR
(
fsr
,
FSR
),
IA64_NATIVE_PATCH_BUNDLE_ELEM_AR
(
fir
,
FIR
),
IA64_NATIVE_PATCH_BUNDLE_ELEM_AR
(
fdr
,
FDR
),
IA64_NATIVE_PATCH_BUNDLE_ELEM_AR
(
ccv
,
CCV
),
IA64_NATIVE_PATCH_BUNDLE_ELEM_AR
(
unat
,
UNAT
),
IA64_NATIVE_PATCH_BUNDLE_ELEM_AR
(
fpsr
,
FPSR
),
IA64_NATIVE_PATCH_BUNDLE_ELEM_AR
(
itc
,
ITC
),
IA64_NATIVE_PATCH_BUNDLE_ELEM_AR
(
pfs
,
PFS
),
IA64_NATIVE_PATCH_BUNDLE_ELEM_AR
(
lc
,
LC
),
IA64_NATIVE_PATCH_BUNDLE_ELEM_AR
(
ec
,
EC
),
IA64_NATIVE_PATCH_BUNDLE_ELEM_CR
(
dcr
,
DCR
),
IA64_NATIVE_PATCH_BUNDLE_ELEM_CR
(
itm
,
ITM
),
IA64_NATIVE_PATCH_BUNDLE_ELEM_CR
(
iva
,
IVA
),
IA64_NATIVE_PATCH_BUNDLE_ELEM_CR
(
pta
,
PTA
),
IA64_NATIVE_PATCH_BUNDLE_ELEM_CR
(
ipsr
,
IPSR
),
IA64_NATIVE_PATCH_BUNDLE_ELEM_CR
(
isr
,
ISR
),
IA64_NATIVE_PATCH_BUNDLE_ELEM_CR
(
iip
,
IIP
),
IA64_NATIVE_PATCH_BUNDLE_ELEM_CR
(
ifa
,
IFA
),
IA64_NATIVE_PATCH_BUNDLE_ELEM_CR
(
itir
,
ITIR
),
IA64_NATIVE_PATCH_BUNDLE_ELEM_CR
(
iipa
,
IIPA
),
IA64_NATIVE_PATCH_BUNDLE_ELEM_CR
(
ifs
,
IFS
),
IA64_NATIVE_PATCH_BUNDLE_ELEM_CR
(
iim
,
IIM
),
IA64_NATIVE_PATCH_BUNDLE_ELEM_CR
(
iha
,
IHA
),
IA64_NATIVE_PATCH_BUNDLE_ELEM_CR
(
lid
,
LID
),
IA64_NATIVE_PATCH_BUNDLE_ELEM_CR
(
ivr
,
IVR
),
IA64_NATIVE_PATCH_BUNDLE_ELEM_CR
(
tpr
,
TPR
),
IA64_NATIVE_PATCH_BUNDLE_ELEM_CR
(
eoi
,
EOI
),
IA64_NATIVE_PATCH_BUNDLE_ELEM_CR
(
irr0
,
IRR0
),
IA64_NATIVE_PATCH_BUNDLE_ELEM_CR
(
irr1
,
IRR1
),
IA64_NATIVE_PATCH_BUNDLE_ELEM_CR
(
irr2
,
IRR2
),
IA64_NATIVE_PATCH_BUNDLE_ELEM_CR
(
irr3
,
IRR3
),
IA64_NATIVE_PATCH_BUNDLE_ELEM_CR
(
itv
,
ITV
),
IA64_NATIVE_PATCH_BUNDLE_ELEM_CR
(
pmv
,
PMV
),
IA64_NATIVE_PATCH_BUNDLE_ELEM_CR
(
cmcv
,
CMCV
),
IA64_NATIVE_PATCH_BUNDLE_ELEM_CR
(
lrr0
,
LRR0
),
IA64_NATIVE_PATCH_BUNDLE_ELEM_CR
(
lrr1
,
LRR1
),
};
unsigned
long
__init_or_module
ia64_native_patch_bundle
(
void
*
sbundle
,
void
*
ebundle
,
unsigned
long
type
)
{
const
unsigned
long
nelems
=
sizeof
(
ia64_native_patch_bundle_elems
)
/
sizeof
(
ia64_native_patch_bundle_elems
[
0
]);
return
__paravirt_patch_apply_bundle
(
sbundle
,
ebundle
,
type
,
ia64_native_patch_bundle_elems
,
nelems
,
NULL
);
}
#endif
/* ASM_SUPPOTED */
extern
const
char
ia64_native_switch_to
[];
extern
const
char
ia64_native_leave_syscall
[];
extern
const
char
ia64_native_work_processed_syscall
[];
extern
const
char
ia64_native_leave_kernel
[];
const
struct
paravirt_patch_branch_target
ia64_native_branch_target
[]
__initconst
=
{
#define PARAVIRT_BR_TARGET(name, type) \
{ \
ia64_native_ ## name, \
PARAVIRT_PATCH_TYPE_BR_ ## type, \
}
PARAVIRT_BR_TARGET
(
switch_to
,
SWITCH_TO
),
PARAVIRT_BR_TARGET
(
leave_syscall
,
LEAVE_SYSCALL
),
PARAVIRT_BR_TARGET
(
work_processed_syscall
,
WORK_PROCESSED_SYSCALL
),
PARAVIRT_BR_TARGET
(
leave_kernel
,
LEAVE_KERNEL
),
};
static
void
__init
ia64_native_patch_branch
(
unsigned
long
tag
,
unsigned
long
type
)
{
const
unsigned
long
nelem
=
sizeof
(
ia64_native_branch_target
)
/
sizeof
(
ia64_native_branch_target
[
0
]);
__paravirt_patch_apply_branch
(
tag
,
type
,
ia64_native_branch_target
,
nelem
);
}
arch/ia64/kernel/paravirt_patch.c
0 → 100644
View file @
c66b31f3
/******************************************************************************
* linux/arch/ia64/xen/paravirt_patch.c
*
* Copyright (c) 2008 Isaku Yamahata <yamahata at valinux co jp>
* VA Linux Systems Japan K.K.
*
* This program is free software; you can redistribute it and/or modify
* it under the terms of the GNU General Public License as published by
* the Free Software Foundation; either version 2 of the License, or
* (at your option) any later version.
*
* This program is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
* GNU General Public License for more details.
*
* You should have received a copy of the GNU General Public License
* along with this program; if not, write to the Free Software
* Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
*
*/
#include <linux/init.h>
#include <asm/intrinsics.h>
#include <asm/kprobes.h>
#include <asm/paravirt.h>
#include <asm/paravirt_patch.h>
typedef
union
ia64_inst
{
struct
{
unsigned
long
long
qp
:
6
;
unsigned
long
long
:
31
;
unsigned
long
long
opcode
:
4
;
unsigned
long
long
reserved
:
23
;
}
generic
;
unsigned
long
long
l
;
}
ia64_inst_t
;
/*
* flush_icache_range() can't be used here.
* we are here before cpu_init() which initializes
* ia64_i_cache_stride_shift. flush_icache_range() uses it.
*/
void
__init_or_module
paravirt_flush_i_cache_range
(
const
void
*
instr
,
unsigned
long
size
)
{
extern
void
paravirt_fc_i
(
const
void
*
addr
);
unsigned
long
i
;
for
(
i
=
0
;
i
<
size
;
i
+=
sizeof
(
bundle_t
))
paravirt_fc_i
(
instr
+
i
);
}
bundle_t
*
__init_or_module
paravirt_get_bundle
(
unsigned
long
tag
)
{
return
(
bundle_t
*
)(
tag
&
~
3UL
);
}
unsigned
long
__init_or_module
paravirt_get_slot
(
unsigned
long
tag
)
{
return
tag
&
3UL
;
}
unsigned
long
__init_or_module
paravirt_get_num_inst
(
unsigned
long
stag
,
unsigned
long
etag
)
{
bundle_t
*
sbundle
=
paravirt_get_bundle
(
stag
);
unsigned
long
sslot
=
paravirt_get_slot
(
stag
);
bundle_t
*
ebundle
=
paravirt_get_bundle
(
etag
);
unsigned
long
eslot
=
paravirt_get_slot
(
etag
);
return
(
ebundle
-
sbundle
)
*
3
+
eslot
-
sslot
+
1
;
}
unsigned
long
__init_or_module
paravirt_get_next_tag
(
unsigned
long
tag
)
{
unsigned
long
slot
=
paravirt_get_slot
(
tag
);
switch
(
slot
)
{
case
0
:
case
1
:
return
tag
+
1
;
case
2
:
{
bundle_t
*
bundle
=
paravirt_get_bundle
(
tag
);
return
(
unsigned
long
)(
bundle
+
1
);
}
default:
BUG
();
}
/* NOTREACHED */
}
ia64_inst_t
__init_or_module
paravirt_read_slot0
(
const
bundle_t
*
bundle
)
{
ia64_inst_t
inst
;
inst
.
l
=
bundle
->
quad0
.
slot0
;
return
inst
;
}
ia64_inst_t
__init_or_module
paravirt_read_slot1
(
const
bundle_t
*
bundle
)
{
ia64_inst_t
inst
;
inst
.
l
=
bundle
->
quad0
.
slot1_p0
|
((
unsigned
long
long
)
bundle
->
quad1
.
slot1_p1
<<
18UL
);
return
inst
;
}
ia64_inst_t
__init_or_module
paravirt_read_slot2
(
const
bundle_t
*
bundle
)
{
ia64_inst_t
inst
;
inst
.
l
=
bundle
->
quad1
.
slot2
;
return
inst
;
}
ia64_inst_t
__init_or_module
paravirt_read_inst
(
unsigned
long
tag
)
{
bundle_t
*
bundle
=
paravirt_get_bundle
(
tag
);
unsigned
long
slot
=
paravirt_get_slot
(
tag
);
switch
(
slot
)
{
case
0
:
return
paravirt_read_slot0
(
bundle
);
case
1
:
return
paravirt_read_slot1
(
bundle
);
case
2
:
return
paravirt_read_slot2
(
bundle
);
default:
BUG
();
}
/* NOTREACHED */
}
void
__init_or_module
paravirt_write_slot0
(
bundle_t
*
bundle
,
ia64_inst_t
inst
)
{
bundle
->
quad0
.
slot0
=
inst
.
l
;
}
void
__init_or_module
paravirt_write_slot1
(
bundle_t
*
bundle
,
ia64_inst_t
inst
)
{
bundle
->
quad0
.
slot1_p0
=
inst
.
l
;
bundle
->
quad1
.
slot1_p1
=
inst
.
l
>>
18UL
;
}
void
__init_or_module
paravirt_write_slot2
(
bundle_t
*
bundle
,
ia64_inst_t
inst
)
{
bundle
->
quad1
.
slot2
=
inst
.
l
;
}
void
__init_or_module
paravirt_write_inst
(
unsigned
long
tag
,
ia64_inst_t
inst
)
{
bundle_t
*
bundle
=
paravirt_get_bundle
(
tag
);
unsigned
long
slot
=
paravirt_get_slot
(
tag
);
switch
(
slot
)
{
case
0
:
paravirt_write_slot0
(
bundle
,
inst
);
break
;
case
1
:
paravirt_write_slot1
(
bundle
,
inst
);
break
;
case
2
:
paravirt_write_slot2
(
bundle
,
inst
);
break
;
default:
BUG
();
break
;
}
paravirt_flush_i_cache_range
(
bundle
,
sizeof
(
*
bundle
));
}
/* for debug */
void
paravirt_print_bundle
(
const
bundle_t
*
bundle
)
{
const
unsigned
long
*
quad
=
(
const
unsigned
long
*
)
bundle
;
ia64_inst_t
slot0
=
paravirt_read_slot0
(
bundle
);
ia64_inst_t
slot1
=
paravirt_read_slot1
(
bundle
);
ia64_inst_t
slot2
=
paravirt_read_slot2
(
bundle
);
printk
(
KERN_DEBUG
"bundle 0x%p 0x%016lx 0x%016lx
\n
"
,
bundle
,
quad
[
0
],
quad
[
1
]);
printk
(
KERN_DEBUG
"bundle template 0x%x
\n
"
,
bundle
->
quad0
.
template
);
printk
(
KERN_DEBUG
"slot0 0x%lx slot1_p0 0x%lx slot1_p1 0x%lx slot2 0x%lx
\n
"
,
(
unsigned
long
)
bundle
->
quad0
.
slot0
,
(
unsigned
long
)
bundle
->
quad0
.
slot1_p0
,
(
unsigned
long
)
bundle
->
quad1
.
slot1_p1
,
(
unsigned
long
)
bundle
->
quad1
.
slot2
);
printk
(
KERN_DEBUG
"slot0 0x%016llx slot1 0x%016llx slot2 0x%016llx
\n
"
,
slot0
.
l
,
slot1
.
l
,
slot2
.
l
);
}
static
int
noreplace_paravirt
__init_or_module
=
0
;
static
int
__init
setup_noreplace_paravirt
(
char
*
str
)
{
noreplace_paravirt
=
1
;
return
1
;
}
__setup
(
"noreplace-paravirt"
,
setup_noreplace_paravirt
);
#ifdef ASM_SUPPORTED
static
void
__init_or_module
fill_nop_bundle
(
void
*
sbundle
,
void
*
ebundle
)
{
extern
const
char
paravirt_nop_bundle
[];
extern
const
unsigned
long
paravirt_nop_bundle_size
;
void
*
bundle
=
sbundle
;
BUG_ON
((((
unsigned
long
)
sbundle
)
%
sizeof
(
bundle_t
))
!=
0
);
BUG_ON
((((
unsigned
long
)
ebundle
)
%
sizeof
(
bundle_t
))
!=
0
);
while
(
bundle
<
ebundle
)
{
memcpy
(
bundle
,
paravirt_nop_bundle
,
paravirt_nop_bundle_size
);
bundle
+=
paravirt_nop_bundle_size
;
}
}
/* helper function */
unsigned
long
__init_or_module
__paravirt_patch_apply_bundle
(
void
*
sbundle
,
void
*
ebundle
,
unsigned
long
type
,
const
struct
paravirt_patch_bundle_elem
*
elems
,
unsigned
long
nelems
,
const
struct
paravirt_patch_bundle_elem
**
found
)
{
unsigned
long
used
=
0
;
unsigned
long
i
;
BUG_ON
((((
unsigned
long
)
sbundle
)
%
sizeof
(
bundle_t
))
!=
0
);
BUG_ON
((((
unsigned
long
)
ebundle
)
%
sizeof
(
bundle_t
))
!=
0
);
found
=
NULL
;
for
(
i
=
0
;
i
<
nelems
;
i
++
)
{
const
struct
paravirt_patch_bundle_elem
*
p
=
&
elems
[
i
];
if
(
p
->
type
==
type
)
{
unsigned
long
need
=
p
->
ebundle
-
p
->
sbundle
;
unsigned
long
room
=
ebundle
-
sbundle
;
if
(
found
!=
NULL
)
*
found
=
p
;
if
(
room
<
need
)
{
/* no room to replace. skip it */
printk
(
KERN_DEBUG
"the space is too small to put "
"bundles. type %ld need %ld room %ld
\n
"
,
type
,
need
,
room
);
break
;
}
used
=
need
;
memcpy
(
sbundle
,
p
->
sbundle
,
used
);
break
;
}
}
return
used
;
}
void
__init_or_module
paravirt_patch_apply_bundle
(
const
struct
paravirt_patch_site_bundle
*
start
,
const
struct
paravirt_patch_site_bundle
*
end
)
{
const
struct
paravirt_patch_site_bundle
*
p
;
if
(
noreplace_paravirt
)
return
;
if
(
pv_init_ops
.
patch_bundle
==
NULL
)
return
;
for
(
p
=
start
;
p
<
end
;
p
++
)
{
unsigned
long
used
;
used
=
(
*
pv_init_ops
.
patch_bundle
)(
p
->
sbundle
,
p
->
ebundle
,
p
->
type
);
if
(
used
==
0
)
continue
;
fill_nop_bundle
(
p
->
sbundle
+
used
,
p
->
ebundle
);
paravirt_flush_i_cache_range
(
p
->
sbundle
,
p
->
ebundle
-
p
->
sbundle
);
}
ia64_sync_i
();
ia64_srlz_i
();
}
/*
* nop.i, nop.m, nop.f instruction are same format.
* but nop.b has differennt format.
* This doesn't support nop.b for now.
*/
static
void
__init_or_module
fill_nop_inst
(
unsigned
long
stag
,
unsigned
long
etag
)
{
extern
const
bundle_t
paravirt_nop_mfi_inst_bundle
[];
unsigned
long
tag
;
const
ia64_inst_t
nop_inst
=
paravirt_read_slot0
(
paravirt_nop_mfi_inst_bundle
);
for
(
tag
=
stag
;
tag
<
etag
;
tag
=
paravirt_get_next_tag
(
tag
))
paravirt_write_inst
(
tag
,
nop_inst
);
}
void
__init_or_module
paravirt_patch_apply_inst
(
const
struct
paravirt_patch_site_inst
*
start
,
const
struct
paravirt_patch_site_inst
*
end
)
{
const
struct
paravirt_patch_site_inst
*
p
;
if
(
noreplace_paravirt
)
return
;
if
(
pv_init_ops
.
patch_inst
==
NULL
)
return
;
for
(
p
=
start
;
p
<
end
;
p
++
)
{
unsigned
long
tag
;
bundle_t
*
sbundle
;
bundle_t
*
ebundle
;
tag
=
(
*
pv_init_ops
.
patch_inst
)(
p
->
stag
,
p
->
etag
,
p
->
type
);
if
(
tag
==
p
->
stag
)
continue
;
fill_nop_inst
(
tag
,
p
->
etag
);
sbundle
=
paravirt_get_bundle
(
p
->
stag
);
ebundle
=
paravirt_get_bundle
(
p
->
etag
)
+
1
;
paravirt_flush_i_cache_range
(
sbundle
,
(
ebundle
-
sbundle
)
*
sizeof
(
bundle_t
));
}
ia64_sync_i
();
ia64_srlz_i
();
}
#endif
/* ASM_SUPPOTED */
/* brl.cond.sptk.many <target64> X3 */
typedef
union
inst_x3_op
{
ia64_inst_t
inst
;
struct
{
unsigned
long
qp
:
6
;
unsigned
long
btyp
:
3
;
unsigned
long
unused
:
3
;
unsigned
long
p
:
1
;
unsigned
long
imm20b
:
20
;
unsigned
long
wh
:
2
;
unsigned
long
d
:
1
;
unsigned
long
i
:
1
;
unsigned
long
opcode
:
4
;
};
unsigned
long
l
;
}
inst_x3_op_t
;
typedef
union
inst_x3_imm
{
ia64_inst_t
inst
;
struct
{
unsigned
long
unused
:
2
;
unsigned
long
imm39
:
39
;
};
unsigned
long
l
;
}
inst_x3_imm_t
;
void
__init_or_module
paravirt_patch_reloc_brl
(
unsigned
long
tag
,
const
void
*
target
)
{
unsigned
long
tag_op
=
paravirt_get_next_tag
(
tag
);
unsigned
long
tag_imm
=
tag
;
bundle_t
*
bundle
=
paravirt_get_bundle
(
tag
);
ia64_inst_t
inst_op
=
paravirt_read_inst
(
tag_op
);
ia64_inst_t
inst_imm
=
paravirt_read_inst
(
tag_imm
);
inst_x3_op_t
inst_x3_op
=
{
.
l
=
inst_op
.
l
};
inst_x3_imm_t
inst_x3_imm
=
{
.
l
=
inst_imm
.
l
};
unsigned
long
imm60
=
((
unsigned
long
)
target
-
(
unsigned
long
)
bundle
)
>>
4
;
BUG_ON
(
paravirt_get_slot
(
tag
)
!=
1
);
/* MLX */
BUG_ON
(((
unsigned
long
)
target
&
(
sizeof
(
bundle_t
)
-
1
))
!=
0
);
/* imm60[59] 1bit */
inst_x3_op
.
i
=
(
imm60
>>
59
)
&
1
;
/* imm60[19:0] 20bit */
inst_x3_op
.
imm20b
=
imm60
&
((
1UL
<<
20
)
-
1
);
/* imm60[58:20] 39bit */
inst_x3_imm
.
imm39
=
(
imm60
>>
20
)
&
((
1UL
<<
39
)
-
1
);
inst_op
.
l
=
inst_x3_op
.
l
;
inst_imm
.
l
=
inst_x3_imm
.
l
;
paravirt_write_inst
(
tag_op
,
inst_op
);
paravirt_write_inst
(
tag_imm
,
inst_imm
);
}
/* br.cond.sptk.many <target25> B1 */
typedef
union
inst_b1
{
ia64_inst_t
inst
;
struct
{
unsigned
long
qp
:
6
;
unsigned
long
btype
:
3
;
unsigned
long
unused
:
3
;
unsigned
long
p
:
1
;
unsigned
long
imm20b
:
20
;
unsigned
long
wh
:
2
;
unsigned
long
d
:
1
;
unsigned
long
s
:
1
;
unsigned
long
opcode
:
4
;
};
unsigned
long
l
;
}
inst_b1_t
;
void
__init
paravirt_patch_reloc_br
(
unsigned
long
tag
,
const
void
*
target
)
{
bundle_t
*
bundle
=
paravirt_get_bundle
(
tag
);
ia64_inst_t
inst
=
paravirt_read_inst
(
tag
);
unsigned
long
target25
=
(
unsigned
long
)
target
-
(
unsigned
long
)
bundle
;
inst_b1_t
inst_b1
;
BUG_ON
(((
unsigned
long
)
target
&
(
sizeof
(
bundle_t
)
-
1
))
!=
0
);
inst_b1
.
l
=
inst
.
l
;
if
(
target25
&
(
1UL
<<
63
))
inst_b1
.
s
=
1
;
else
inst_b1
.
s
=
0
;
inst_b1
.
imm20b
=
target25
>>
4
;
inst
.
l
=
inst_b1
.
l
;
paravirt_write_inst
(
tag
,
inst
);
}
void
__init
__paravirt_patch_apply_branch
(
unsigned
long
tag
,
unsigned
long
type
,
const
struct
paravirt_patch_branch_target
*
entries
,
unsigned
int
nr_entries
)
{
unsigned
int
i
;
for
(
i
=
0
;
i
<
nr_entries
;
i
++
)
{
if
(
entries
[
i
].
type
==
type
)
{
paravirt_patch_reloc_br
(
tag
,
entries
[
i
].
entry
);
break
;
}
}
}
static
void
__init
paravirt_patch_apply_branch
(
const
struct
paravirt_patch_site_branch
*
start
,
const
struct
paravirt_patch_site_branch
*
end
)
{
const
struct
paravirt_patch_site_branch
*
p
;
if
(
noreplace_paravirt
)
return
;
if
(
pv_init_ops
.
patch_branch
==
NULL
)
return
;
for
(
p
=
start
;
p
<
end
;
p
++
)
(
*
pv_init_ops
.
patch_branch
)(
p
->
tag
,
p
->
type
);
ia64_sync_i
();
ia64_srlz_i
();
}
void
__init
paravirt_patch_apply
(
void
)
{
extern
const
char
__start_paravirt_bundles
[];
extern
const
char
__stop_paravirt_bundles
[];
extern
const
char
__start_paravirt_insts
[];
extern
const
char
__stop_paravirt_insts
[];
extern
const
char
__start_paravirt_branches
[];
extern
const
char
__stop_paravirt_branches
[];
paravirt_patch_apply_bundle
((
const
struct
paravirt_patch_site_bundle
*
)
__start_paravirt_bundles
,
(
const
struct
paravirt_patch_site_bundle
*
)
__stop_paravirt_bundles
);
paravirt_patch_apply_inst
((
const
struct
paravirt_patch_site_inst
*
)
__start_paravirt_insts
,
(
const
struct
paravirt_patch_site_inst
*
)
__stop_paravirt_insts
);
paravirt_patch_apply_branch
((
const
struct
paravirt_patch_site_branch
*
)
__start_paravirt_branches
,
(
const
struct
paravirt_patch_site_branch
*
)
__stop_paravirt_branches
);
}
/*
* Local variables:
* mode: C
* c-set-style: "linux"
* c-basic-offset: 8
* tab-width: 8
* indent-tabs-mode: t
* End:
*/
arch/ia64/kernel/paravirt_patchlist.c
0 → 100644
View file @
c66b31f3
/******************************************************************************
* Copyright (c) 2008 Isaku Yamahata <yamahata at valinux co jp>
* VA Linux Systems Japan K.K.
*
* This program is free software; you can redistribute it and/or modify
* it under the terms of the GNU General Public License as published by
* the Free Software Foundation; either version 2 of the License, or
* (at your option) any later version.
*
* This program is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
* GNU General Public License for more details.
*
* You should have received a copy of the GNU General Public License
* along with this program; if not, write to the Free Software
* Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
*
*/
#include <linux/bug.h>
#include <asm/paravirt.h>
#define DECLARE(name) \
extern unsigned long \
__ia64_native_start_gate_##name##_patchlist[]; \
extern unsigned long \
__ia64_native_end_gate_##name##_patchlist[]
DECLARE
(
fsyscall
);
DECLARE
(
brl_fsys_bubble_down
);
DECLARE
(
vtop
);
DECLARE
(
mckinley_e9
);
extern
unsigned
long
__start_gate_section
[];
#define ASSIGN(name) \
.start_##name##_patchlist = \
(unsigned long)__ia64_native_start_gate_##name##_patchlist, \
.end_##name##_patchlist = \
(unsigned long)__ia64_native_end_gate_##name##_patchlist
struct
pv_patchdata
pv_patchdata
__initdata
=
{
ASSIGN
(
fsyscall
),
ASSIGN
(
brl_fsys_bubble_down
),
ASSIGN
(
vtop
),
ASSIGN
(
mckinley_e9
),
.
gate_section
=
(
void
*
)
__start_gate_section
,
};
unsigned
long
__init
paravirt_get_gate_patchlist
(
enum
pv_gate_patchlist
type
)
{
#define CASE(NAME, name) \
case PV_GATE_START_##NAME: \
return pv_patchdata.start_##name##_patchlist; \
case PV_GATE_END_##NAME: \
return pv_patchdata.end_##name##_patchlist; \
switch
(
type
)
{
CASE
(
FSYSCALL
,
fsyscall
);
CASE
(
BRL_FSYS_BUBBLE_DOWN
,
brl_fsys_bubble_down
);
CASE
(
VTOP
,
vtop
);
CASE
(
MCKINLEY_E9
,
mckinley_e9
);
default:
BUG
();
break
;
}
return
0
;
}
void
*
__init
paravirt_get_gate_section
(
void
)
{
return
pv_patchdata
.
gate_section
;
}
arch/ia64/kernel/paravirt_patchlist.h
0 → 100644
View file @
c66b31f3
/******************************************************************************
* linux/arch/ia64/xen/paravirt_patchlist.h
*
* Copyright (c) 2008 Isaku Yamahata <yamahata at valinux co jp>
* VA Linux Systems Japan K.K.
*
* This program is free software; you can redistribute it and/or modify
* it under the terms of the GNU General Public License as published by
* the Free Software Foundation; either version 2 of the License, or
* (at your option) any later version.
*
* This program is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
* GNU General Public License for more details.
*
* You should have received a copy of the GNU General Public License
* along with this program; if not, write to the Free Software
* Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
*
*/
#if defined(__IA64_GATE_PARAVIRTUALIZED_XEN)
#include <asm/xen/patchlist.h>
#else
#include <asm/native/patchlist.h>
#endif
arch/ia64/kernel/paravirtentry.S
View file @
c66b31f3
...
...
@@ -20,8 +20,11 @@
*
*/
#include <linux/init.h>
#include <asm/asmmacro.h>
#include <asm/asm-offsets.h>
#include <asm/paravirt_privop.h>
#include <asm/paravirt_patch.h>
#include "entry.h"
#define DATA8(sym, init_value) \
...
...
@@ -32,7 +35,9 @@
data8
init_value
; \
.
popsection
#define BRANCH(targ, reg, breg) \
#define BRANCH(targ, reg, breg, type) \
PARAVIRT_PATCH_SITE_BR
(
PARAVIRT_PATCH_TYPE_BR_
##
type
)
; \
;; \
movl
reg
=
targ
; \
;; \
ld8
reg
=[
reg
]
; \
...
...
@@ -40,21 +45,77 @@
mov
breg
=
reg
; \
br.cond.sptk.many
breg
#define BRANCH_PROC(sym, reg, breg) \
#define BRANCH_PROC(sym, reg, breg
, type
) \
DATA8
(
paravirt_
##
sym
##
_targ
,
ia64_native_
##
sym
)
; \
GLOBAL_ENTRY
(
paravirt_
##
sym
)
; \
BRANCH
(
paravirt_
##
sym
##
_targ
,
reg
,
breg
)
; \
BRANCH
(
paravirt_
##
sym
##
_targ
,
reg
,
breg
,
type
)
; \
END
(
paravirt_
##
sym
)
#define BRANCH_PROC_UNWINFO(sym, reg, breg) \
#define BRANCH_PROC_UNWINFO(sym, reg, breg
, type
) \
DATA8
(
paravirt_
##
sym
##
_targ
,
ia64_native_
##
sym
)
; \
GLOBAL_ENTRY
(
paravirt_
##
sym
)
; \
PT_REGS_UNWIND_INFO
(0)
; \
BRANCH
(
paravirt_
##
sym
##
_targ
,
reg
,
breg
)
; \
BRANCH
(
paravirt_
##
sym
##
_targ
,
reg
,
breg
,
type
)
; \
END
(
paravirt_
##
sym
)
BRANCH_PROC
(
switch_to
,
r22
,
b7
)
BRANCH_PROC_UNWINFO
(
leave_syscall
,
r22
,
b7
)
BRANCH_PROC
(
work_processed_syscall
,
r2
,
b7
)
BRANCH_PROC_UNWINFO
(
leave_kernel
,
r22
,
b7
)
BRANCH_PROC
(
switch_to
,
r22
,
b7
,
SWITCH_TO
)
BRANCH_PROC_UNWINFO
(
leave_syscall
,
r22
,
b7
,
LEAVE_SYSCALL
)
BRANCH_PROC
(
work_processed_syscall
,
r2
,
b7
,
WORK_PROCESSED_SYSCALL
)
BRANCH_PROC_UNWINFO
(
leave_kernel
,
r22
,
b7
,
LEAVE_KERNEL
)
#ifdef CONFIG_MODULES
#define __INIT_OR_MODULE .text
#define __INITDATA_OR_MODULE .data
#else
#define __INIT_OR_MODULE __INIT
#define __INITDATA_OR_MODULE __INITDATA
#endif /* CONFIG_MODULES */
__INIT_OR_MODULE
GLOBAL_ENTRY
(
paravirt_fc_i
)
fc.i
r32
br.ret.sptk.many
rp
END
(
paravirt_fc_i
)
__FINIT
__INIT_OR_MODULE
.
align
32
GLOBAL_ENTRY
(
paravirt_nop_b_inst_bundle
)
{
nop.b
0
nop.b
0
nop.b
0
}
END
(
paravirt_nop_b_inst_bundle
)
__FINIT
/
*
NOTE
:
nop
.
[
mfi
]
has
same
format
*/
__INIT_OR_MODULE
GLOBAL_ENTRY
(
paravirt_nop_mfi_inst_bundle
)
{
nop.m
0
nop.f
0
nop.i
0
}
END
(
paravirt_nop_mfi_inst_bundle
)
__FINIT
__INIT_OR_MODULE
GLOBAL_ENTRY
(
paravirt_nop_bundle
)
paravirt_nop_bundle_start
:
{
nop
0
nop
0
nop
0
}
paravirt_nop_bundle_end
:
END
(
paravirt_nop_bundle
)
__FINIT
__INITDATA_OR_MODULE
.
align
8
.
global
paravirt_nop_bundle_size
paravirt_nop_bundle_size
:
data8
paravirt_nop_bundle_end
-
paravirt_nop_bundle_start
arch/ia64/kernel/patch.c
View file @
c66b31f3
...
...
@@ -7,6 +7,7 @@
#include <linux/init.h>
#include <linux/string.h>
#include <asm/paravirt.h>
#include <asm/patch.h>
#include <asm/processor.h>
#include <asm/sections.h>
...
...
@@ -169,16 +170,35 @@ ia64_patch_mckinley_e9 (unsigned long start, unsigned long end)
ia64_srlz_i
();
}
extern
unsigned
long
ia64_native_fsyscall_table
[
NR_syscalls
];
extern
char
ia64_native_fsys_bubble_down
[];
struct
pv_fsys_data
pv_fsys_data
__initdata
=
{
.
fsyscall_table
=
(
unsigned
long
*
)
ia64_native_fsyscall_table
,
.
fsys_bubble_down
=
(
void
*
)
ia64_native_fsys_bubble_down
,
};
unsigned
long
*
__init
paravirt_get_fsyscall_table
(
void
)
{
return
pv_fsys_data
.
fsyscall_table
;
}
char
*
__init
paravirt_get_fsys_bubble_down
(
void
)
{
return
pv_fsys_data
.
fsys_bubble_down
;
}
static
void
__init
patch_fsyscall_table
(
unsigned
long
start
,
unsigned
long
end
)
{
extern
unsigned
long
fsyscall_table
[
NR_syscalls
]
;
u64
fsyscall_table
=
(
u64
)
paravirt_get_fsyscall_table
()
;
s32
*
offp
=
(
s32
*
)
start
;
u64
ip
;
while
(
offp
<
(
s32
*
)
end
)
{
ip
=
(
u64
)
ia64_imva
((
char
*
)
offp
+
*
offp
);
ia64_patch_imm64
(
ip
,
(
u64
)
fsyscall_table
);
ia64_patch_imm64
(
ip
,
fsyscall_table
);
ia64_fc
((
void
*
)
ip
);
++
offp
;
}
...
...
@@ -189,7 +209,7 @@ patch_fsyscall_table (unsigned long start, unsigned long end)
static
void
__init
patch_brl_fsys_bubble_down
(
unsigned
long
start
,
unsigned
long
end
)
{
extern
char
fsys_bubble_down
[]
;
u64
fsys_bubble_down
=
(
u64
)
paravirt_get_fsys_bubble_down
()
;
s32
*
offp
=
(
s32
*
)
start
;
u64
ip
;
...
...
@@ -207,13 +227,13 @@ patch_brl_fsys_bubble_down (unsigned long start, unsigned long end)
void
__init
ia64_patch_gate
(
void
)
{
# define START(name)
((unsigned long) __start_gate_##name##_patchlist
)
# define END(name)
((unsigned long)__end_gate_##name##_patchlist
)
# define START(name)
paravirt_get_gate_patchlist(PV_GATE_START_##name
)
# define END(name)
paravirt_get_gate_patchlist(PV_GATE_END_##name
)
patch_fsyscall_table
(
START
(
fsyscall
),
END
(
fsyscall
));
patch_brl_fsys_bubble_down
(
START
(
brl_fsys_bubble_down
),
END
(
brl_fsys_bubble_down
));
ia64_patch_vtop
(
START
(
vtop
),
END
(
vtop
));
ia64_patch_mckinley_e9
(
START
(
mckinley_e9
),
END
(
mckinley_e
9
));
patch_fsyscall_table
(
START
(
FSYSCALL
),
END
(
FSYSCALL
));
patch_brl_fsys_bubble_down
(
START
(
BRL_FSYS_BUBBLE_DOWN
),
END
(
BRL_FSYS_BUBBLE_DOWN
));
ia64_patch_vtop
(
START
(
VTOP
),
END
(
VTOP
));
ia64_patch_mckinley_e9
(
START
(
MCKINLEY_E9
),
END
(
MCKINLEY_E
9
));
}
void
ia64_patch_phys_stack_reg
(
unsigned
long
val
)
...
...
@@ -229,7 +249,7 @@ void ia64_patch_phys_stack_reg(unsigned long val)
while
(
offp
<
end
)
{
ip
=
(
u64
)
offp
+
*
offp
;
ia64_patch
(
ip
,
mask
,
imm
);
ia64_fc
(
ip
);
ia64_fc
(
(
void
*
)
ip
);
++
offp
;
}
ia64_sync_i
();
...
...
arch/ia64/kernel/setup.c
View file @
c66b31f3
...
...
@@ -52,6 +52,7 @@
#include <asm/meminit.h>
#include <asm/page.h>
#include <asm/paravirt.h>
#include <asm/paravirt_patch.h>
#include <asm/patch.h>
#include <asm/pgtable.h>
#include <asm/processor.h>
...
...
@@ -537,6 +538,7 @@ setup_arch (char **cmdline_p)
paravirt_arch_setup_early
();
ia64_patch_vtop
((
u64
)
__start___vtop_patchlist
,
(
u64
)
__end___vtop_patchlist
);
paravirt_patch_apply
();
*
cmdline_p
=
__va
(
ia64_boot_param
->
command_line
);
strlcpy
(
boot_command_line
,
*
cmdline_p
,
COMMAND_LINE_SIZE
);
...
...
arch/ia64/kernel/time.c
View file @
c66b31f3
...
...
@@ -49,6 +49,15 @@ EXPORT_SYMBOL(last_cli_ip);
#endif
#ifdef CONFIG_PARAVIRT
/* We need to define a real function for sched_clock, to override the
weak default version */
unsigned
long
long
sched_clock
(
void
)
{
return
paravirt_sched_clock
();
}
#endif
#ifdef CONFIG_PARAVIRT
static
void
paravirt_clocksource_resume
(
void
)
...
...
arch/ia64/kernel/vmlinux.lds.S
View file @
c66b31f3
...
...
@@ -169,6 +169,30 @@ SECTIONS
__end___mckinley_e9_bundles
=
.
;
}
#if defined(CONFIG_PARAVIRT)
.
=
ALIGN
(
16
)
;
.
paravirt_bundles
:
AT
(
ADDR
(
.
paravirt_bundles
)
-
LOAD_OFFSET
)
{
__start_paravirt_bundles
=
.
;
*(.
paravirt_bundles
)
__stop_paravirt_bundles
=
.
;
}
.
=
ALIGN
(
16
)
;
.
paravirt_insts
:
AT
(
ADDR
(
.
paravirt_insts
)
-
LOAD_OFFSET
)
{
__start_paravirt_insts
=
.
;
*(.
paravirt_insts
)
__stop_paravirt_insts
=
.
;
}
.
=
ALIGN
(
16
)
;
.
paravirt_branches
:
AT
(
ADDR
(
.
paravirt_branches
)
-
LOAD_OFFSET
)
{
__start_paravirt_branches
=
.
;
*(.
paravirt_branches
)
__stop_paravirt_branches
=
.
;
}
#endif
#if defined(CONFIG_IA64_GENERIC)
/
*
Machine
Vector
*/
.
=
ALIGN
(
16
)
;
...
...
@@ -201,6 +225,12 @@ SECTIONS
__start_gate_section
=
.
;
*(.
data.gate
)
__stop_gate_section
=
.
;
#ifdef CONFIG_XEN
.
=
ALIGN
(
PAGE_SIZE
)
;
__xen_start_gate_section
=
.
;
*(.
data.gate.xen
)
__xen_stop_gate_section
=
.
;
#endif
}
.
=
ALIGN
(
PAGE_SIZE
)
; /* make sure the gate page doesn't expose
*
kernel
data
...
...
arch/ia64/kvm/kvm-ia64.c
View file @
c66b31f3
...
...
@@ -70,7 +70,7 @@ static void kvm_flush_icache(unsigned long start, unsigned long len)
int
l
;
for
(
l
=
0
;
l
<
(
len
+
32
);
l
+=
32
)
ia64_fc
(
start
+
l
);
ia64_fc
(
(
void
*
)(
start
+
l
)
);
ia64_sync_i
();
ia64_srlz_i
();
...
...
arch/ia64/kvm/vcpu.c
View file @
c66b31f3
...
...
@@ -386,7 +386,7 @@ void set_rse_reg(struct kvm_pt_regs *regs, unsigned long r1,
else
*
rnat_addr
=
(
*
rnat_addr
)
&
(
~
nat_mask
);
ia64_setreg
(
_IA64_REG_AR_BSPSTORE
,
bspstore
);
ia64_setreg
(
_IA64_REG_AR_BSPSTORE
,
(
unsigned
long
)
bspstore
);
ia64_setreg
(
_IA64_REG_AR_RNAT
,
rnat
);
}
local_irq_restore
(
psr
);
...
...
arch/ia64/kvm/vtlb.c
View file @
c66b31f3
...
...
@@ -210,6 +210,7 @@ void thash_vhpt_insert(struct kvm_vcpu *v, u64 pte, u64 itir, u64 va, int type)
phy_pte
&=
~
PAGE_FLAGS_RV_MASK
;
psr
=
ia64_clear_ic
();
ia64_itc
(
type
,
va
,
phy_pte
,
itir_ps
(
itir
));
paravirt_dv_serialize_data
();
ia64_set_psr
(
psr
);
}
...
...
@@ -456,6 +457,7 @@ void thash_purge_and_insert(struct kvm_vcpu *v, u64 pte, u64 itir,
phy_pte
&=
~
PAGE_FLAGS_RV_MASK
;
psr
=
ia64_clear_ic
();
ia64_itc
(
type
,
ifa
,
phy_pte
,
ps
);
paravirt_dv_serialize_data
();
ia64_set_psr
(
psr
);
}
if
(
!
(
pte
&
VTLB_PTE_IO
))
...
...
arch/ia64/mm/init.c
View file @
c66b31f3
...
...
@@ -35,6 +35,7 @@
#include <asm/uaccess.h>
#include <asm/unistd.h>
#include <asm/mca.h>
#include <asm/paravirt.h>
DEFINE_PER_CPU
(
struct
mmu_gather
,
mmu_gathers
);
...
...
@@ -259,6 +260,7 @@ put_kernel_page (struct page *page, unsigned long address, pgprot_t pgprot)
static
void
__init
setup_gate
(
void
)
{
void
*
gate_section
;
struct
page
*
page
;
/*
...
...
@@ -266,10 +268,11 @@ setup_gate (void)
* headers etc. and once execute-only page to enable
* privilege-promotion via "epc":
*/
page
=
virt_to_page
(
ia64_imva
(
__start_gate_section
));
gate_section
=
paravirt_get_gate_section
();
page
=
virt_to_page
(
ia64_imva
(
gate_section
));
put_kernel_page
(
page
,
GATE_ADDR
,
PAGE_READONLY
);
#ifdef HAVE_BUGGY_SEGREL
page
=
virt_to_page
(
ia64_imva
(
__start_
gate_section
+
PAGE_SIZE
));
page
=
virt_to_page
(
ia64_imva
(
gate_section
+
PAGE_SIZE
));
put_kernel_page
(
page
,
GATE_ADDR
+
PAGE_SIZE
,
PAGE_GATE
);
#else
put_kernel_page
(
page
,
GATE_ADDR
+
PERCPU_PAGE_SIZE
,
PAGE_GATE
);
...
...
@@ -667,8 +670,8 @@ mem_init (void)
* code can tell them apart.
*/
for
(
i
=
0
;
i
<
NR_syscalls
;
++
i
)
{
extern
unsigned
long
fsyscall_table
[
NR_syscalls
];
extern
unsigned
long
sys_call_table
[
NR_syscalls
];
unsigned
long
*
fsyscall_table
=
paravirt_get_fsyscall_table
();
if
(
!
fsyscall_table
[
i
]
||
nolwsys
)
fsyscall_table
[
i
]
=
sys_call_table
[
i
]
|
1
;
...
...
arch/ia64/scripts/pvcheck.sed
View file @
c66b31f3
...
...
@@ -17,6 +17,7 @@ s/mov.*=.*cr\.iip/.warning \"cr.iip should not used directly\"/g
s
/
mov
.*
=
.*
cr
\.
ivr
/
.warning
\"
cr.ivr should not used directly
\"
/
g
s
/
mov
.*
=
[^
\.
]
*
psr
/
.warning
\"
psr should not used directly
\"
/
g
# avoid ar.fpsr
s
/
mov
.*
=
.*
ar
\.
eflags
/
.warning
\"
ar.eflags should not used directly
\"
/
g
s
/
mov
.*
=
.*
ar
\.
itc
.*
/
.warning
\"
ar.itc should not used directly
\"
/
g
s
/
mov
.*
cr
\.
ifa
.*
=
.*
/
.warning
\"
cr.ifa should not used directly
\"
/
g
s
/
mov
.*
cr
\.
itir
.*
=
.*
/
.warning
\"
cr.itir should not used directly
\"
/
g
s
/
mov
.*
cr
\.
iha
.*
=
.*
/
.warning
\"
cr.iha should not used directly
\"
/
g
...
...
arch/ia64/xen/Makefile
View file @
c66b31f3
...
...
@@ -3,14 +3,29 @@
#
obj-y
:=
hypercall.o xenivt.o xensetup.o xen_pv_ops.o irq_xen.o
\
hypervisor.o xencomm.o xcom_hcall.o grant-table.o time.o suspend.o
hypervisor.o xencomm.o xcom_hcall.o grant-table.o time.o suspend.o
\
gate-data.o
obj-$(CONFIG_IA64_GENERIC)
+=
machvec.o
# The gate DSO image is built using a special linker script.
include
$(srctree)/arch/ia64/kernel/Makefile.gate
# tell compiled for xen
CPPFLAGS_gate.lds
+=
-D__IA64_GATE_PARAVIRTUALIZED_XEN
AFLAGS_gate.o
+=
-D__IA64_ASM_PARAVIRTUALIZED_XEN
-D__IA64_GATE_PARAVIRTUALIZED_XEN
# use same file of native.
$(obj)/gate.o
:
$(src)/../kernel/gate.S FORCE
$(
call
if_changed_dep,as_o_S
)
$(obj)/gate.lds
:
$(src)/../kernel/gate.lds.S FORCE
$(
call
if_changed_dep,cpp_lds_S
)
AFLAGS_xenivt.o
+=
-D__IA64_ASM_PARAVIRTUALIZED_XEN
# xen multi compile
ASM_PARAVIRT_MULTI_COMPILE_SRCS
=
ivt.S entry.S
ASM_PARAVIRT_MULTI_COMPILE_SRCS
=
ivt.S entry.S
fsys.S
ASM_PARAVIRT_OBJS
=
$(
addprefix
xen-,
$(ASM_PARAVIRT_MULTI_COMPILE_SRCS:.S=.o)
)
obj-y
+=
$(ASM_PARAVIRT_OBJS)
define
paravirtualized_xen
...
...
arch/ia64/xen/gate-data.S
0 → 100644
View file @
c66b31f3
.
section
.
data
.
gate
.
xen
,
"aw"
.
incbin
"arch/ia64/xen/gate.so"
arch/ia64/xen/hypercall.S
View file @
c66b31f3
...
...
@@ -9,6 +9,7 @@
#include <asm/intrinsics.h>
#include <asm/xen/privop.h>
#ifdef __INTEL_COMPILER
/*
*
Hypercalls
without
parameter
.
*/
...
...
@@ -72,6 +73,7 @@ GLOBAL_ENTRY(xen_set_rr0_to_rr4)
br.ret.sptk.many
rp
;;
END
(
xen_set_rr0_to_rr4
)
#endif
GLOBAL_ENTRY
(
xen_send_ipi
)
mov
r14
=
r32
...
...
arch/ia64/xen/time.c
View file @
c66b31f3
...
...
@@ -175,10 +175,58 @@ static void xen_itc_jitter_data_reset(void)
}
while
(
unlikely
(
ret
!=
lcycle
));
}
/* based on xen_sched_clock() in arch/x86/xen/time.c. */
/*
* This relies on HAVE_UNSTABLE_SCHED_CLOCK. If it can't be defined,
* something similar logic should be implemented here.
*/
/*
* Xen sched_clock implementation. Returns the number of unstolen
* nanoseconds, which is nanoseconds the VCPU spent in RUNNING+BLOCKED
* states.
*/
static
unsigned
long
long
xen_sched_clock
(
void
)
{
struct
vcpu_runstate_info
runstate
;
unsigned
long
long
now
;
unsigned
long
long
offset
;
unsigned
long
long
ret
;
/*
* Ideally sched_clock should be called on a per-cpu basis
* anyway, so preempt should already be disabled, but that's
* not current practice at the moment.
*/
preempt_disable
();
/*
* both ia64_native_sched_clock() and xen's runstate are
* based on mAR.ITC. So difference of them makes sense.
*/
now
=
ia64_native_sched_clock
();
get_runstate_snapshot
(
&
runstate
);
WARN_ON
(
runstate
.
state
!=
RUNSTATE_running
);
offset
=
0
;
if
(
now
>
runstate
.
state_entry_time
)
offset
=
now
-
runstate
.
state_entry_time
;
ret
=
runstate
.
time
[
RUNSTATE_blocked
]
+
runstate
.
time
[
RUNSTATE_running
]
+
offset
;
preempt_enable
();
return
ret
;
}
struct
pv_time_ops
xen_time_ops
__initdata
=
{
.
init_missing_ticks_accounting
=
xen_init_missing_ticks_accounting
,
.
do_steal_accounting
=
xen_do_steal_accounting
,
.
clocksource_resume
=
xen_itc_jitter_data_reset
,
.
sched_clock
=
xen_sched_clock
,
};
/* Called after suspend, to resume time. */
...
...
arch/ia64/xen/xen_pv_ops.c
View file @
c66b31f3
...
...
@@ -24,6 +24,7 @@
#include <linux/irq.h>
#include <linux/kernel.h>
#include <linux/pm.h>
#include <linux/unistd.h>
#include <asm/xen/hypervisor.h>
#include <asm/xen/xencomm.h>
...
...
@@ -153,6 +154,13 @@ xen_post_smp_prepare_boot_cpu(void)
xen_setup_vcpu_info_placement
();
}
#ifdef ASM_SUPPORTED
static
unsigned
long
__init_or_module
xen_patch_bundle
(
void
*
sbundle
,
void
*
ebundle
,
unsigned
long
type
);
#endif
static
void
__init
xen_patch_branch
(
unsigned
long
tag
,
unsigned
long
type
);
static
const
struct
pv_init_ops
xen_init_ops
__initconst
=
{
.
banner
=
xen_banner
,
...
...
@@ -163,6 +171,53 @@ static const struct pv_init_ops xen_init_ops __initconst = {
.
arch_setup_nomca
=
xen_arch_setup_nomca
,
.
post_smp_prepare_boot_cpu
=
xen_post_smp_prepare_boot_cpu
,
#ifdef ASM_SUPPORTED
.
patch_bundle
=
xen_patch_bundle
,
#endif
.
patch_branch
=
xen_patch_branch
,
};
/***************************************************************************
* pv_fsys_data
* addresses for fsys
*/
extern
unsigned
long
xen_fsyscall_table
[
NR_syscalls
];
extern
char
xen_fsys_bubble_down
[];
struct
pv_fsys_data
xen_fsys_data
__initdata
=
{
.
fsyscall_table
=
(
unsigned
long
*
)
xen_fsyscall_table
,
.
fsys_bubble_down
=
(
void
*
)
xen_fsys_bubble_down
,
};
/***************************************************************************
* pv_patchdata
* patchdata addresses
*/
#define DECLARE(name) \
extern unsigned long __xen_start_gate_##name##_patchlist[]; \
extern unsigned long __xen_end_gate_##name##_patchlist[]
DECLARE
(
fsyscall
);
DECLARE
(
brl_fsys_bubble_down
);
DECLARE
(
vtop
);
DECLARE
(
mckinley_e9
);
extern
unsigned
long
__xen_start_gate_section
[];
#define ASSIGN(name) \
.start_##name##_patchlist = \
(unsigned long)__xen_start_gate_##name##_patchlist, \
.end_##name##_patchlist = \
(unsigned long)__xen_end_gate_##name##_patchlist
static
struct
pv_patchdata
xen_patchdata
__initdata
=
{
ASSIGN
(
fsyscall
),
ASSIGN
(
brl_fsys_bubble_down
),
ASSIGN
(
vtop
),
ASSIGN
(
mckinley_e9
),
.
gate_section
=
(
void
*
)
__xen_start_gate_section
,
};
/***************************************************************************
...
...
@@ -170,6 +225,76 @@ static const struct pv_init_ops xen_init_ops __initconst = {
* intrinsics hooks.
*/
#ifndef ASM_SUPPORTED
static
void
xen_set_itm_with_offset
(
unsigned
long
val
)
{
/* ia64_cpu_local_tick() calls this with interrupt enabled. */
/* WARN_ON(!irqs_disabled()); */
xen_set_itm
(
val
-
XEN_MAPPEDREGS
->
itc_offset
);
}
static
unsigned
long
xen_get_itm_with_offset
(
void
)
{
/* unused at this moment */
printk
(
KERN_DEBUG
"%s is called.
\n
"
,
__func__
);
WARN_ON
(
!
irqs_disabled
());
return
ia64_native_getreg
(
_IA64_REG_CR_ITM
)
+
XEN_MAPPEDREGS
->
itc_offset
;
}
/* ia64_set_itc() is only called by
* cpu_init() with ia64_set_itc(0) and ia64_sync_itc().
* So XEN_MAPPEDRESG->itc_offset cal be considered as almost constant.
*/
static
void
xen_set_itc
(
unsigned
long
val
)
{
unsigned
long
mitc
;
WARN_ON
(
!
irqs_disabled
());
mitc
=
ia64_native_getreg
(
_IA64_REG_AR_ITC
);
XEN_MAPPEDREGS
->
itc_offset
=
val
-
mitc
;
XEN_MAPPEDREGS
->
itc_last
=
val
;
}
static
unsigned
long
xen_get_itc
(
void
)
{
unsigned
long
res
;
unsigned
long
itc_offset
;
unsigned
long
itc_last
;
unsigned
long
ret_itc_last
;
itc_offset
=
XEN_MAPPEDREGS
->
itc_offset
;
do
{
itc_last
=
XEN_MAPPEDREGS
->
itc_last
;
res
=
ia64_native_getreg
(
_IA64_REG_AR_ITC
);
res
+=
itc_offset
;
if
(
itc_last
>=
res
)
res
=
itc_last
+
1
;
ret_itc_last
=
cmpxchg
(
&
XEN_MAPPEDREGS
->
itc_last
,
itc_last
,
res
);
}
while
(
unlikely
(
ret_itc_last
!=
itc_last
));
return
res
;
#if 0
/* ia64_itc_udelay() calls ia64_get_itc() with interrupt enabled.
Should it be paravirtualized instead? */
WARN_ON(!irqs_disabled());
itc_offset = XEN_MAPPEDREGS->itc_offset;
itc_last = XEN_MAPPEDREGS->itc_last;
res = ia64_native_getreg(_IA64_REG_AR_ITC);
res += itc_offset;
if (itc_last >= res)
res = itc_last + 1;
XEN_MAPPEDREGS->itc_last = res;
return res;
#endif
}
static
void
xen_setreg
(
int
regnum
,
unsigned
long
val
)
{
switch
(
regnum
)
{
...
...
@@ -181,11 +306,14 @@ static void xen_setreg(int regnum, unsigned long val)
xen_set_eflag
(
val
);
break
;
#endif
case
_IA64_REG_AR_ITC
:
xen_set_itc
(
val
);
break
;
case
_IA64_REG_CR_TPR
:
xen_set_tpr
(
val
);
break
;
case
_IA64_REG_CR_ITM
:
xen_set_itm
(
val
);
xen_set_itm
_with_offset
(
val
);
break
;
case
_IA64_REG_CR_EOI
:
xen_eoi
(
val
);
...
...
@@ -209,6 +337,12 @@ static unsigned long xen_getreg(int regnum)
res
=
xen_get_eflag
();
break
;
#endif
case
_IA64_REG_AR_ITC
:
res
=
xen_get_itc
();
break
;
case
_IA64_REG_CR_ITM
:
res
=
xen_get_itm_with_offset
();
break
;
case
_IA64_REG_CR_IVR
:
res
=
xen_get_ivr
();
break
;
...
...
@@ -259,8 +393,417 @@ xen_intrin_local_irq_restore(unsigned long mask)
else
xen_rsm_i
();
}
#else
#define __DEFINE_FUNC(name, code) \
extern const char xen_ ## name ## _direct_start[]; \
extern const char xen_ ## name ## _direct_end[]; \
asm (".align 32\n" \
".proc xen_" #name "\n" \
"xen_" #name ":\n" \
"xen_" #name "_direct_start:\n" \
code \
"xen_" #name "_direct_end:\n" \
"br.cond.sptk.many b6\n" \
".endp xen_" #name "\n")
#define DEFINE_VOID_FUNC0(name, code) \
extern void \
xen_ ## name (void); \
__DEFINE_FUNC(name, code)
#define DEFINE_VOID_FUNC1(name, code) \
extern void \
xen_ ## name (unsigned long arg); \
__DEFINE_FUNC(name, code)
#define DEFINE_VOID_FUNC1_VOID(name, code) \
extern void \
xen_ ## name (void *arg); \
__DEFINE_FUNC(name, code)
#define DEFINE_VOID_FUNC2(name, code) \
extern void \
xen_ ## name (unsigned long arg0, \
unsigned long arg1); \
__DEFINE_FUNC(name, code)
#define DEFINE_FUNC0(name, code) \
extern unsigned long \
xen_ ## name (void); \
__DEFINE_FUNC(name, code)
#define DEFINE_FUNC1(name, type, code) \
extern unsigned long \
xen_ ## name (type arg); \
__DEFINE_FUNC(name, code)
#define XEN_PSR_I_ADDR_ADDR (XSI_BASE + XSI_PSR_I_ADDR_OFS)
/*
* static void xen_set_itm_with_offset(unsigned long val)
* xen_set_itm(val - XEN_MAPPEDREGS->itc_offset);
*/
/* 2 bundles */
DEFINE_VOID_FUNC1
(
set_itm_with_offset
,
"mov r2 = "
__stringify
(
XSI_BASE
)
" + "
__stringify
(
XSI_ITC_OFFSET_OFS
)
"
\n
"
";;
\n
"
"ld8 r3 = [r2]
\n
"
";;
\n
"
"sub r8 = r8, r3
\n
"
"break "
__stringify
(
HYPERPRIVOP_SET_ITM
)
"
\n
"
);
/*
* static unsigned long xen_get_itm_with_offset(void)
* return ia64_native_getreg(_IA64_REG_CR_ITM) + XEN_MAPPEDREGS->itc_offset;
*/
/* 2 bundles */
DEFINE_FUNC0
(
get_itm_with_offset
,
"mov r2 = "
__stringify
(
XSI_BASE
)
" + "
__stringify
(
XSI_ITC_OFFSET_OFS
)
"
\n
"
";;
\n
"
"ld8 r3 = [r2]
\n
"
"mov r8 = cr.itm
\n
"
";;
\n
"
"add r8 = r8, r2
\n
"
);
/*
* static void xen_set_itc(unsigned long val)
* unsigned long mitc;
*
* WARN_ON(!irqs_disabled());
* mitc = ia64_native_getreg(_IA64_REG_AR_ITC);
* XEN_MAPPEDREGS->itc_offset = val - mitc;
* XEN_MAPPEDREGS->itc_last = val;
*/
/* 2 bundles */
DEFINE_VOID_FUNC1
(
set_itc
,
"mov r2 = "
__stringify
(
XSI_BASE
)
" + "
__stringify
(
XSI_ITC_LAST_OFS
)
"
\n
"
"mov r3 = ar.itc
\n
"
";;
\n
"
"sub r3 = r8, r3
\n
"
"st8 [r2] = r8, "
__stringify
(
XSI_ITC_LAST_OFS
)
" - "
__stringify
(
XSI_ITC_OFFSET_OFS
)
"
\n
"
";;
\n
"
"st8 [r2] = r3
\n
"
);
/*
* static unsigned long xen_get_itc(void)
* unsigned long res;
* unsigned long itc_offset;
* unsigned long itc_last;
* unsigned long ret_itc_last;
*
* itc_offset = XEN_MAPPEDREGS->itc_offset;
* do {
* itc_last = XEN_MAPPEDREGS->itc_last;
* res = ia64_native_getreg(_IA64_REG_AR_ITC);
* res += itc_offset;
* if (itc_last >= res)
* res = itc_last + 1;
* ret_itc_last = cmpxchg(&XEN_MAPPEDREGS->itc_last,
* itc_last, res);
* } while (unlikely(ret_itc_last != itc_last));
* return res;
*/
/* 5 bundles */
DEFINE_FUNC0
(
get_itc
,
"mov r2 = "
__stringify
(
XSI_BASE
)
" + "
__stringify
(
XSI_ITC_OFFSET_OFS
)
"
\n
"
";;
\n
"
"ld8 r9 = [r2], "
__stringify
(
XSI_ITC_LAST_OFS
)
" - "
__stringify
(
XSI_ITC_OFFSET_OFS
)
"
\n
"
/* r9 = itc_offset */
/* r2 = XSI_ITC_OFFSET */
"888:
\n
"
"mov r8 = ar.itc
\n
"
/* res = ar.itc */
";;
\n
"
"ld8 r3 = [r2]
\n
"
/* r3 = itc_last */
"add r8 = r8, r9
\n
"
/* res = ar.itc + itc_offset */
";;
\n
"
"cmp.gtu p6, p0 = r3, r8
\n
"
";;
\n
"
"(p6) add r8 = 1, r3
\n
"
/* if (itc_last > res) itc_last + 1 */
";;
\n
"
"mov ar.ccv = r8
\n
"
";;
\n
"
"cmpxchg8.acq r10 = [r2], r8, ar.ccv
\n
"
";;
\n
"
"cmp.ne p6, p0 = r10, r3
\n
"
"(p6) hint @pause
\n
"
"(p6) br.cond.spnt 888b
\n
"
);
DEFINE_VOID_FUNC1_VOID
(
fc
,
"break "
__stringify
(
HYPERPRIVOP_FC
)
"
\n
"
);
/*
* psr_i_addr_addr = XEN_PSR_I_ADDR_ADDR
* masked_addr = *psr_i_addr_addr
* pending_intr_addr = masked_addr - 1
* if (val & IA64_PSR_I) {
* masked = *masked_addr
* *masked_addr = 0:xen_set_virtual_psr_i(1)
* compiler barrier
* if (masked) {
* uint8_t pending = *pending_intr_addr;
* if (pending)
* XEN_HYPER_SSM_I
* }
* } else {
* *masked_addr = 1:xen_set_virtual_psr_i(0)
* }
*/
/* 6 bundles */
DEFINE_VOID_FUNC1
(
intrin_local_irq_restore
,
/* r8 = input value: 0 or IA64_PSR_I
* p6 = (flags & IA64_PSR_I)
* = if clause
* p7 = !(flags & IA64_PSR_I)
* = else clause
*/
"cmp.ne p6, p7 = r8, r0
\n
"
"mov r9 = "
__stringify
(
XEN_PSR_I_ADDR_ADDR
)
"
\n
"
";;
\n
"
/* r9 = XEN_PSR_I_ADDR */
"ld8 r9 = [r9]
\n
"
";;
\n
"
/* r10 = masked previous value */
"(p6) ld1.acq r10 = [r9]
\n
"
";;
\n
"
/* p8 = !masked interrupt masked previously? */
"(p6) cmp.ne.unc p8, p0 = r10, r0
\n
"
/* p7 = else clause */
"(p7) mov r11 = 1
\n
"
";;
\n
"
/* masked = 1 */
"(p7) st1.rel [r9] = r11
\n
"
/* p6 = if clause */
/* masked = 0
* r9 = masked_addr - 1
* = pending_intr_addr
*/
"(p8) st1.rel [r9] = r0, -1
\n
"
";;
\n
"
/* r8 = pending_intr */
"(p8) ld1.acq r11 = [r9]
\n
"
";;
\n
"
/* p9 = interrupt pending? */
"(p8) cmp.ne.unc p9, p10 = r11, r0
\n
"
";;
\n
"
"(p10) mf
\n
"
/* issue hypercall to trigger interrupt */
"(p9) break "
__stringify
(
HYPERPRIVOP_SSM_I
)
"
\n
"
);
DEFINE_VOID_FUNC2
(
ptcga
,
"break "
__stringify
(
HYPERPRIVOP_PTC_GA
)
"
\n
"
);
DEFINE_VOID_FUNC2
(
set_rr
,
"break "
__stringify
(
HYPERPRIVOP_SET_RR
)
"
\n
"
);
/*
* tmp = XEN_MAPPEDREGS->interrupt_mask_addr = XEN_PSR_I_ADDR_ADDR;
* tmp = *tmp
* tmp = *tmp;
* psr_i = tmp? 0: IA64_PSR_I;
*/
/* 4 bundles */
DEFINE_FUNC0
(
get_psr_i
,
"mov r9 = "
__stringify
(
XEN_PSR_I_ADDR_ADDR
)
"
\n
"
";;
\n
"
"ld8 r9 = [r9]
\n
"
/* r9 = XEN_PSR_I_ADDR */
"mov r8 = 0
\n
"
/* psr_i = 0 */
";;
\n
"
"ld1.acq r9 = [r9]
\n
"
/* r9 = XEN_PSR_I */
";;
\n
"
"cmp.eq.unc p6, p0 = r9, r0
\n
"
/* p6 = (XEN_PSR_I != 0) */
";;
\n
"
"(p6) mov r8 = "
__stringify
(
1
<<
IA64_PSR_I_BIT
)
"
\n
"
);
static
const
struct
pv_cpu_ops
xen_cpu_ops
__initdata
=
{
DEFINE_FUNC1
(
thash
,
unsigned
long
,
"break "
__stringify
(
HYPERPRIVOP_THASH
)
"
\n
"
);
DEFINE_FUNC1
(
get_cpuid
,
int
,
"break "
__stringify
(
HYPERPRIVOP_GET_CPUID
)
"
\n
"
);
DEFINE_FUNC1
(
get_pmd
,
int
,
"break "
__stringify
(
HYPERPRIVOP_GET_PMD
)
"
\n
"
);
DEFINE_FUNC1
(
get_rr
,
unsigned
long
,
"break "
__stringify
(
HYPERPRIVOP_GET_RR
)
"
\n
"
);
/*
* void xen_privop_ssm_i(void)
*
* int masked = !xen_get_virtual_psr_i();
* // masked = *(*XEN_MAPPEDREGS->interrupt_mask_addr)
* xen_set_virtual_psr_i(1)
* // *(*XEN_MAPPEDREGS->interrupt_mask_addr) = 0
* // compiler barrier
* if (masked) {
* uint8_t* pend_int_addr =
* (uint8_t*)(*XEN_MAPPEDREGS->interrupt_mask_addr) - 1;
* uint8_t pending = *pend_int_addr;
* if (pending)
* XEN_HYPER_SSM_I
* }
*/
/* 4 bundles */
DEFINE_VOID_FUNC0
(
ssm_i
,
"mov r8 = "
__stringify
(
XEN_PSR_I_ADDR_ADDR
)
"
\n
"
";;
\n
"
"ld8 r8 = [r8]
\n
"
/* r8 = XEN_PSR_I_ADDR */
";;
\n
"
"ld1.acq r9 = [r8]
\n
"
/* r9 = XEN_PSR_I */
";;
\n
"
"st1.rel [r8] = r0, -1
\n
"
/* psr_i = 0. enable interrupt
* r8 = XEN_PSR_I_ADDR - 1
* = pend_int_addr
*/
"cmp.eq.unc p0, p6 = r9, r0
\n
"
/* p6 = !XEN_PSR_I
* previously interrupt
* masked?
*/
";;
\n
"
"(p6) ld1.acq r8 = [r8]
\n
"
/* r8 = xen_pend_int */
";;
\n
"
"(p6) cmp.eq.unc p6, p7 = r8, r0
\n
"
/*interrupt pending?*/
";;
\n
"
/* issue hypercall to get interrupt */
"(p7) break "
__stringify
(
HYPERPRIVOP_SSM_I
)
"
\n
"
";;
\n
"
);
/*
* psr_i_addr_addr = XEN_MAPPEDREGS->interrupt_mask_addr
* = XEN_PSR_I_ADDR_ADDR;
* psr_i_addr = *psr_i_addr_addr;
* *psr_i_addr = 1;
*/
/* 2 bundles */
DEFINE_VOID_FUNC0
(
rsm_i
,
"mov r8 = "
__stringify
(
XEN_PSR_I_ADDR_ADDR
)
"
\n
"
/* r8 = XEN_PSR_I_ADDR */
"mov r9 = 1
\n
"
";;
\n
"
"ld8 r8 = [r8]
\n
"
/* r8 = XEN_PSR_I */
";;
\n
"
"st1.rel [r8] = r9
\n
"
);
/* XEN_PSR_I = 1 */
extern
void
xen_set_rr0_to_rr4
(
unsigned
long
val0
,
unsigned
long
val1
,
unsigned
long
val2
,
unsigned
long
val3
,
unsigned
long
val4
);
__DEFINE_FUNC
(
set_rr0_to_rr4
,
"break "
__stringify
(
HYPERPRIVOP_SET_RR0_TO_RR4
)
"
\n
"
);
extern
unsigned
long
xen_getreg
(
int
regnum
);
#define __DEFINE_GET_REG(id, privop) \
"mov r2 = " __stringify(_IA64_REG_ ## id) "\n" \
";;\n" \
"cmp.eq p6, p0 = r2, r8\n" \
";;\n" \
"(p6) break " __stringify(HYPERPRIVOP_GET_ ## privop) "\n" \
"(p6) br.cond.sptk.many b6\n" \
";;\n"
__DEFINE_FUNC
(
getreg
,
__DEFINE_GET_REG
(
PSR
,
PSR
)
#ifdef CONFIG_IA32_SUPPORT
__DEFINE_GET_REG
(
AR_EFLAG
,
EFLAG
)
#endif
/* get_itc */
"mov r2 = "
__stringify
(
_IA64_REG_AR_ITC
)
"
\n
"
";;
\n
"
"cmp.eq p6, p0 = r2, r8
\n
"
";;
\n
"
"(p6) br.cond.spnt xen_get_itc
\n
"
";;
\n
"
/* get itm */
"mov r2 = "
__stringify
(
_IA64_REG_CR_ITM
)
"
\n
"
";;
\n
"
"cmp.eq p6, p0 = r2, r8
\n
"
";;
\n
"
"(p6) br.cond.spnt xen_get_itm_with_offset
\n
"
";;
\n
"
__DEFINE_GET_REG
(
CR_IVR
,
IVR
)
__DEFINE_GET_REG
(
CR_TPR
,
TPR
)
/* fall back */
"movl r2 = ia64_native_getreg_func
\n
"
";;
\n
"
"mov b7 = r2
\n
"
";;
\n
"
"br.cond.sptk.many b7
\n
"
);
extern
void
xen_setreg
(
int
regnum
,
unsigned
long
val
);
#define __DEFINE_SET_REG(id, privop) \
"mov r2 = " __stringify(_IA64_REG_ ## id) "\n" \
";;\n" \
"cmp.eq p6, p0 = r2, r9\n" \
";;\n" \
"(p6) break " __stringify(HYPERPRIVOP_ ## privop) "\n" \
"(p6) br.cond.sptk.many b6\n" \
";;\n"
__DEFINE_FUNC
(
setreg
,
/* kr0 .. kr 7*/
/*
* if (_IA64_REG_AR_KR0 <= regnum &&
* regnum <= _IA64_REG_AR_KR7) {
* register __index asm ("r8") = regnum - _IA64_REG_AR_KR0
* register __val asm ("r9") = val
* "break HYPERPRIVOP_SET_KR"
* }
*/
"mov r17 = r9
\n
"
"mov r2 = "
__stringify
(
_IA64_REG_AR_KR0
)
"
\n
"
";;
\n
"
"cmp.ge p6, p0 = r9, r2
\n
"
"sub r17 = r17, r2
\n
"
";;
\n
"
"(p6) cmp.ge.unc p7, p0 = "
__stringify
(
_IA64_REG_AR_KR7
)
" - "
__stringify
(
_IA64_REG_AR_KR0
)
", r17
\n
"
";;
\n
"
"(p7) mov r9 = r8
\n
"
";;
\n
"
"(p7) mov r8 = r17
\n
"
"(p7) break "
__stringify
(
HYPERPRIVOP_SET_KR
)
"
\n
"
/* set itm */
"mov r2 = "
__stringify
(
_IA64_REG_CR_ITM
)
"
\n
"
";;
\n
"
"cmp.eq p6, p0 = r2, r8
\n
"
";;
\n
"
"(p6) br.cond.spnt xen_set_itm_with_offset
\n
"
/* set itc */
"mov r2 = "
__stringify
(
_IA64_REG_AR_ITC
)
"
\n
"
";;
\n
"
"cmp.eq p6, p0 = r2, r8
\n
"
";;
\n
"
"(p6) br.cond.spnt xen_set_itc
\n
"
#ifdef CONFIG_IA32_SUPPORT
__DEFINE_SET_REG
(
AR_EFLAG
,
SET_EFLAG
)
#endif
__DEFINE_SET_REG
(
CR_TPR
,
SET_TPR
)
__DEFINE_SET_REG
(
CR_EOI
,
EOI
)
/* fall back */
"movl r2 = ia64_native_setreg_func
\n
"
";;
\n
"
"mov b7 = r2
\n
"
";;
\n
"
"br.cond.sptk.many b7
\n
"
);
#endif
static
const
struct
pv_cpu_ops
xen_cpu_ops
__initconst
=
{
.
fc
=
xen_fc
,
.
thash
=
xen_thash
,
.
get_cpuid
=
xen_get_cpuid
,
...
...
@@ -337,7 +880,7 @@ xen_iosapic_write(char __iomem *iosapic, unsigned int reg, u32 val)
HYPERVISOR_physdev_op
(
PHYSDEVOP_apic_write
,
&
apic_op
);
}
static
const
struct
pv_iosapic_ops
xen_iosapic_ops
__initconst
=
{
static
struct
pv_iosapic_ops
xen_iosapic_ops
__initdata
=
{
.
pcat_compat_init
=
xen_pcat_compat_init
,
.
__get_irq_chip
=
xen_iosapic_get_irq_chip
,
...
...
@@ -355,6 +898,8 @@ xen_setup_pv_ops(void)
xen_info_init
();
pv_info
=
xen_info
;
pv_init_ops
=
xen_init_ops
;
pv_fsys_data
=
xen_fsys_data
;
pv_patchdata
=
xen_patchdata
;
pv_cpu_ops
=
xen_cpu_ops
;
pv_iosapic_ops
=
xen_iosapic_ops
;
pv_irq_ops
=
xen_irq_ops
;
...
...
@@ -362,3 +907,252 @@ xen_setup_pv_ops(void)
paravirt_cpu_asm_init
(
&
xen_cpu_asm_switch
);
}
#ifdef ASM_SUPPORTED
/***************************************************************************
* binary pacthing
* pv_init_ops.patch_bundle
*/
#define DEFINE_FUNC_GETREG(name, privop) \
DEFINE_FUNC0(get_ ## name, \
"break "__stringify(HYPERPRIVOP_GET_ ## privop) "\n")
DEFINE_FUNC_GETREG
(
psr
,
PSR
);
DEFINE_FUNC_GETREG
(
eflag
,
EFLAG
);
DEFINE_FUNC_GETREG
(
ivr
,
IVR
);
DEFINE_FUNC_GETREG
(
tpr
,
TPR
);
#define DEFINE_FUNC_SET_KR(n) \
DEFINE_VOID_FUNC0(set_kr ## n, \
";;\n" \
"mov r9 = r8\n" \
"mov r8 = " #n "\n" \
"break " __stringify(HYPERPRIVOP_SET_KR) "\n")
DEFINE_FUNC_SET_KR
(
0
);
DEFINE_FUNC_SET_KR
(
1
);
DEFINE_FUNC_SET_KR
(
2
);
DEFINE_FUNC_SET_KR
(
3
);
DEFINE_FUNC_SET_KR
(
4
);
DEFINE_FUNC_SET_KR
(
5
);
DEFINE_FUNC_SET_KR
(
6
);
DEFINE_FUNC_SET_KR
(
7
);
#define __DEFINE_FUNC_SETREG(name, privop) \
DEFINE_VOID_FUNC0(name, \
"break "__stringify(HYPERPRIVOP_ ## privop) "\n")
#define DEFINE_FUNC_SETREG(name, privop) \
__DEFINE_FUNC_SETREG(set_ ## name, SET_ ## privop)
DEFINE_FUNC_SETREG
(
eflag
,
EFLAG
);
DEFINE_FUNC_SETREG
(
tpr
,
TPR
);
__DEFINE_FUNC_SETREG
(
eoi
,
EOI
);
extern
const
char
xen_check_events
[];
extern
const
char
__xen_intrin_local_irq_restore_direct_start
[];
extern
const
char
__xen_intrin_local_irq_restore_direct_end
[];
extern
const
unsigned
long
__xen_intrin_local_irq_restore_direct_reloc
;
asm
(
".align 32
\n
"
".proc xen_check_events
\n
"
"xen_check_events:
\n
"
/* masked = 0
* r9 = masked_addr - 1
* = pending_intr_addr
*/
"st1.rel [r9] = r0, -1
\n
"
";;
\n
"
/* r8 = pending_intr */
"ld1.acq r11 = [r9]
\n
"
";;
\n
"
/* p9 = interrupt pending? */
"cmp.ne p9, p10 = r11, r0
\n
"
";;
\n
"
"(p10) mf
\n
"
/* issue hypercall to trigger interrupt */
"(p9) break "
__stringify
(
HYPERPRIVOP_SSM_I
)
"
\n
"
"br.cond.sptk.many b6
\n
"
".endp xen_check_events
\n
"
"
\n
"
".align 32
\n
"
".proc __xen_intrin_local_irq_restore_direct
\n
"
"__xen_intrin_local_irq_restore_direct:
\n
"
"__xen_intrin_local_irq_restore_direct_start:
\n
"
"1:
\n
"
"{
\n
"
"cmp.ne p6, p7 = r8, r0
\n
"
"mov r17 = ip
\n
"
/* get ip to calc return address */
"mov r9 = "
__stringify
(
XEN_PSR_I_ADDR_ADDR
)
"
\n
"
";;
\n
"
"}
\n
"
"{
\n
"
/* r9 = XEN_PSR_I_ADDR */
"ld8 r9 = [r9]
\n
"
";;
\n
"
/* r10 = masked previous value */
"(p6) ld1.acq r10 = [r9]
\n
"
"adds r17 = 1f - 1b, r17
\n
"
/* calculate return address */
";;
\n
"
"}
\n
"
"{
\n
"
/* p8 = !masked interrupt masked previously? */
"(p6) cmp.ne.unc p8, p0 = r10, r0
\n
"
"
\n
"
/* p7 = else clause */
"(p7) mov r11 = 1
\n
"
";;
\n
"
"(p8) mov b6 = r17
\n
"
/* set return address */
"}
\n
"
"{
\n
"
/* masked = 1 */
"(p7) st1.rel [r9] = r11
\n
"
"
\n
"
"[99:]
\n
"
"(p8) brl.cond.dptk.few xen_check_events
\n
"
"}
\n
"
/* pv calling stub is 5 bundles. fill nop to adjust return address */
"{
\n
"
"nop 0
\n
"
"nop 0
\n
"
"nop 0
\n
"
"}
\n
"
"1:
\n
"
"__xen_intrin_local_irq_restore_direct_end:
\n
"
".endp __xen_intrin_local_irq_restore_direct
\n
"
"
\n
"
".align 8
\n
"
"__xen_intrin_local_irq_restore_direct_reloc:
\n
"
"data8 99b
\n
"
);
static
struct
paravirt_patch_bundle_elem
xen_patch_bundle_elems
[]
__initdata_or_module
=
{
#define XEN_PATCH_BUNDLE_ELEM(name, type) \
{ \
(void*)xen_ ## name ## _direct_start, \
(void*)xen_ ## name ## _direct_end, \
PARAVIRT_PATCH_TYPE_ ## type, \
}
XEN_PATCH_BUNDLE_ELEM
(
fc
,
FC
),
XEN_PATCH_BUNDLE_ELEM
(
thash
,
THASH
),
XEN_PATCH_BUNDLE_ELEM
(
get_cpuid
,
GET_CPUID
),
XEN_PATCH_BUNDLE_ELEM
(
get_pmd
,
GET_PMD
),
XEN_PATCH_BUNDLE_ELEM
(
ptcga
,
PTCGA
),
XEN_PATCH_BUNDLE_ELEM
(
get_rr
,
GET_RR
),
XEN_PATCH_BUNDLE_ELEM
(
set_rr
,
SET_RR
),
XEN_PATCH_BUNDLE_ELEM
(
set_rr0_to_rr4
,
SET_RR0_TO_RR4
),
XEN_PATCH_BUNDLE_ELEM
(
ssm_i
,
SSM_I
),
XEN_PATCH_BUNDLE_ELEM
(
rsm_i
,
RSM_I
),
XEN_PATCH_BUNDLE_ELEM
(
get_psr_i
,
GET_PSR_I
),
{
(
void
*
)
__xen_intrin_local_irq_restore_direct_start
,
(
void
*
)
__xen_intrin_local_irq_restore_direct_end
,
PARAVIRT_PATCH_TYPE_INTRIN_LOCAL_IRQ_RESTORE
,
},
#define XEN_PATCH_BUNDLE_ELEM_GETREG(name, reg) \
{ \
xen_get_ ## name ## _direct_start, \
xen_get_ ## name ## _direct_end, \
PARAVIRT_PATCH_TYPE_GETREG + _IA64_REG_ ## reg, \
}
XEN_PATCH_BUNDLE_ELEM_GETREG
(
psr
,
PSR
),
XEN_PATCH_BUNDLE_ELEM_GETREG
(
eflag
,
AR_EFLAG
),
XEN_PATCH_BUNDLE_ELEM_GETREG
(
ivr
,
CR_IVR
),
XEN_PATCH_BUNDLE_ELEM_GETREG
(
tpr
,
CR_TPR
),
XEN_PATCH_BUNDLE_ELEM_GETREG
(
itc
,
AR_ITC
),
XEN_PATCH_BUNDLE_ELEM_GETREG
(
itm_with_offset
,
CR_ITM
),
#define __XEN_PATCH_BUNDLE_ELEM_SETREG(name, reg) \
{ \
xen_ ## name ## _direct_start, \
xen_ ## name ## _direct_end, \
PARAVIRT_PATCH_TYPE_SETREG + _IA64_REG_ ## reg, \
}
#define XEN_PATCH_BUNDLE_ELEM_SETREG(name, reg) \
__XEN_PATCH_BUNDLE_ELEM_SETREG(set_ ## name, reg)
XEN_PATCH_BUNDLE_ELEM_SETREG
(
kr0
,
AR_KR0
),
XEN_PATCH_BUNDLE_ELEM_SETREG
(
kr1
,
AR_KR1
),
XEN_PATCH_BUNDLE_ELEM_SETREG
(
kr2
,
AR_KR2
),
XEN_PATCH_BUNDLE_ELEM_SETREG
(
kr3
,
AR_KR3
),
XEN_PATCH_BUNDLE_ELEM_SETREG
(
kr4
,
AR_KR4
),
XEN_PATCH_BUNDLE_ELEM_SETREG
(
kr5
,
AR_KR5
),
XEN_PATCH_BUNDLE_ELEM_SETREG
(
kr6
,
AR_KR6
),
XEN_PATCH_BUNDLE_ELEM_SETREG
(
kr7
,
AR_KR7
),
XEN_PATCH_BUNDLE_ELEM_SETREG
(
eflag
,
AR_EFLAG
),
XEN_PATCH_BUNDLE_ELEM_SETREG
(
tpr
,
CR_TPR
),
__XEN_PATCH_BUNDLE_ELEM_SETREG
(
eoi
,
CR_EOI
),
XEN_PATCH_BUNDLE_ELEM_SETREG
(
itc
,
AR_ITC
),
XEN_PATCH_BUNDLE_ELEM_SETREG
(
itm_with_offset
,
CR_ITM
),
};
static
unsigned
long
__init_or_module
xen_patch_bundle
(
void
*
sbundle
,
void
*
ebundle
,
unsigned
long
type
)
{
const
unsigned
long
nelems
=
sizeof
(
xen_patch_bundle_elems
)
/
sizeof
(
xen_patch_bundle_elems
[
0
]);
unsigned
long
used
;
const
struct
paravirt_patch_bundle_elem
*
found
;
used
=
__paravirt_patch_apply_bundle
(
sbundle
,
ebundle
,
type
,
xen_patch_bundle_elems
,
nelems
,
&
found
);
if
(
found
==
NULL
)
/* fallback */
return
ia64_native_patch_bundle
(
sbundle
,
ebundle
,
type
);
if
(
used
==
0
)
return
used
;
/* relocation */
switch
(
type
)
{
case
PARAVIRT_PATCH_TYPE_INTRIN_LOCAL_IRQ_RESTORE
:
{
unsigned
long
reloc
=
__xen_intrin_local_irq_restore_direct_reloc
;
unsigned
long
reloc_offset
=
reloc
-
(
unsigned
long
)
__xen_intrin_local_irq_restore_direct_start
;
unsigned
long
tag
=
(
unsigned
long
)
sbundle
+
reloc_offset
;
paravirt_patch_reloc_brl
(
tag
,
xen_check_events
);
break
;
}
default:
/* nothing */
break
;
}
return
used
;
}
#endif
/* ASM_SUPPOTED */
const
struct
paravirt_patch_branch_target
xen_branch_target
[]
__initconst
=
{
#define PARAVIRT_BR_TARGET(name, type) \
{ \
&xen_ ## name, \
PARAVIRT_PATCH_TYPE_BR_ ## type, \
}
PARAVIRT_BR_TARGET
(
switch_to
,
SWITCH_TO
),
PARAVIRT_BR_TARGET
(
leave_syscall
,
LEAVE_SYSCALL
),
PARAVIRT_BR_TARGET
(
work_processed_syscall
,
WORK_PROCESSED_SYSCALL
),
PARAVIRT_BR_TARGET
(
leave_kernel
,
LEAVE_KERNEL
),
};
static
void
__init
xen_patch_branch
(
unsigned
long
tag
,
unsigned
long
type
)
{
const
unsigned
long
nelem
=
sizeof
(
xen_branch_target
)
/
sizeof
(
xen_branch_target
[
0
]);
__paravirt_patch_apply_branch
(
tag
,
type
,
xen_branch_target
,
nelem
);
}
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment