Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Support
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
L
linux
Project overview
Project overview
Details
Activity
Releases
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Issues
0
Issues
0
List
Boards
Labels
Milestones
Merge Requests
0
Merge Requests
0
Analytics
Analytics
Repository
Value Stream
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Create a new issue
Commits
Issue Boards
Open sidebar
nexedi
linux
Commits
63b1ad24
Commit
63b1ad24
authored
May 31, 2007
by
Avi Kivity
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
KVM: MMU: Fold fix_read_pf() into set_pte_common()
Signed-off-by:
Avi Kivity
<
avi@qumranet.com
>
parent
6598c8b2
Changes
2
Hide whitespace changes
Inline
Side-by-side
Showing
2 changed files
with
23 additions
and
28 deletions
+23
-28
drivers/kvm/mmu.c
drivers/kvm/mmu.c
+0
-17
drivers/kvm/paging_tmpl.h
drivers/kvm/paging_tmpl.h
+23
-11
No files found.
drivers/kvm/mmu.c
View file @
63b1ad24
...
@@ -972,23 +972,6 @@ static void inject_page_fault(struct kvm_vcpu *vcpu,
...
@@ -972,23 +972,6 @@ static void inject_page_fault(struct kvm_vcpu *vcpu,
kvm_arch_ops
->
inject_page_fault
(
vcpu
,
addr
,
err_code
);
kvm_arch_ops
->
inject_page_fault
(
vcpu
,
addr
,
err_code
);
}
}
static
inline
int
fix_read_pf
(
u64
*
shadow_ent
)
{
if
((
*
shadow_ent
&
PT_SHADOW_USER_MASK
)
&&
!
(
*
shadow_ent
&
PT_USER_MASK
))
{
/*
* If supervisor write protect is disabled, we shadow kernel
* pages as user pages so we can trap the write access.
*/
*
shadow_ent
|=
PT_USER_MASK
;
*
shadow_ent
&=
~
PT_WRITABLE_MASK
;
return
1
;
}
return
0
;
}
static
void
paging_free
(
struct
kvm_vcpu
*
vcpu
)
static
void
paging_free
(
struct
kvm_vcpu
*
vcpu
)
{
{
nonpaging_free
(
vcpu
);
nonpaging_free
(
vcpu
);
...
...
drivers/kvm/paging_tmpl.h
View file @
63b1ad24
...
@@ -197,6 +197,7 @@ static void FNAME(set_pte_common)(struct kvm_vcpu *vcpu,
...
@@ -197,6 +197,7 @@ static void FNAME(set_pte_common)(struct kvm_vcpu *vcpu,
gpa_t
gaddr
,
gpa_t
gaddr
,
pt_element_t
*
gpte
,
pt_element_t
*
gpte
,
u64
access_bits
,
u64
access_bits
,
int
write_fault
,
gfn_t
gfn
)
gfn_t
gfn
)
{
{
hpa_t
paddr
;
hpa_t
paddr
;
...
@@ -219,6 +220,17 @@ static void FNAME(set_pte_common)(struct kvm_vcpu *vcpu,
...
@@ -219,6 +220,17 @@ static void FNAME(set_pte_common)(struct kvm_vcpu *vcpu,
*
shadow_pte
|=
paddr
;
*
shadow_pte
|=
paddr
;
if
(
!
write_fault
&&
(
*
shadow_pte
&
PT_SHADOW_USER_MASK
)
&&
!
(
*
shadow_pte
&
PT_USER_MASK
))
{
/*
* If supervisor write protect is disabled, we shadow kernel
* pages as user pages so we can trap the write access.
*/
*
shadow_pte
|=
PT_USER_MASK
;
*
shadow_pte
&=
~
PT_WRITABLE_MASK
;
access_bits
&=
~
PT_WRITABLE_MASK
;
}
if
(
access_bits
&
PT_WRITABLE_MASK
)
{
if
(
access_bits
&
PT_WRITABLE_MASK
)
{
struct
kvm_mmu_page
*
shadow
;
struct
kvm_mmu_page
*
shadow
;
...
@@ -242,13 +254,14 @@ static void FNAME(set_pte_common)(struct kvm_vcpu *vcpu,
...
@@ -242,13 +254,14 @@ static void FNAME(set_pte_common)(struct kvm_vcpu *vcpu,
}
}
static
void
FNAME
(
set_pte
)(
struct
kvm_vcpu
*
vcpu
,
pt_element_t
*
gpte
,
static
void
FNAME
(
set_pte
)(
struct
kvm_vcpu
*
vcpu
,
pt_element_t
*
gpte
,
u64
*
shadow_pte
,
u64
access_bits
,
gfn_t
gfn
)
u64
*
shadow_pte
,
u64
access_bits
,
int
write_fault
,
gfn_t
gfn
)
{
{
ASSERT
(
*
shadow_pte
==
0
);
ASSERT
(
*
shadow_pte
==
0
);
access_bits
&=
*
gpte
;
access_bits
&=
*
gpte
;
*
shadow_pte
=
(
*
gpte
&
PT_PTE_COPY_MASK
);
*
shadow_pte
=
(
*
gpte
&
PT_PTE_COPY_MASK
);
FNAME
(
set_pte_common
)(
vcpu
,
shadow_pte
,
*
gpte
&
PT_BASE_ADDR_MASK
,
FNAME
(
set_pte_common
)(
vcpu
,
shadow_pte
,
*
gpte
&
PT_BASE_ADDR_MASK
,
gpte
,
access_bits
,
gfn
);
gpte
,
access_bits
,
write_fault
,
gfn
);
}
}
static
void
FNAME
(
update_pte
)(
struct
kvm_vcpu
*
vcpu
,
struct
kvm_mmu_page
*
page
,
static
void
FNAME
(
update_pte
)(
struct
kvm_vcpu
*
vcpu
,
struct
kvm_mmu_page
*
page
,
...
@@ -262,12 +275,13 @@ static void FNAME(update_pte)(struct kvm_vcpu *vcpu, struct kvm_mmu_page *page,
...
@@ -262,12 +275,13 @@ static void FNAME(update_pte)(struct kvm_vcpu *vcpu, struct kvm_mmu_page *page,
if
(
~
gpte
&
(
PT_PRESENT_MASK
|
PT_ACCESSED_MASK
))
if
(
~
gpte
&
(
PT_PRESENT_MASK
|
PT_ACCESSED_MASK
))
return
;
return
;
pgprintk
(
"%s: gpte %llx spte %p
\n
"
,
__FUNCTION__
,
(
u64
)
gpte
,
spte
);
pgprintk
(
"%s: gpte %llx spte %p
\n
"
,
__FUNCTION__
,
(
u64
)
gpte
,
spte
);
FNAME
(
set_pte
)(
vcpu
,
&
gpte
,
spte
,
PT_USER_MASK
|
PT_WRITABLE_MASK
,
FNAME
(
set_pte
)(
vcpu
,
&
gpte
,
spte
,
PT_USER_MASK
|
PT_WRITABLE_MASK
,
0
,
(
gpte
&
PT_BASE_ADDR_MASK
)
>>
PAGE_SHIFT
);
(
gpte
&
PT_BASE_ADDR_MASK
)
>>
PAGE_SHIFT
);
}
}
static
void
FNAME
(
set_pde
)(
struct
kvm_vcpu
*
vcpu
,
pt_element_t
*
gpde
,
static
void
FNAME
(
set_pde
)(
struct
kvm_vcpu
*
vcpu
,
pt_element_t
*
gpde
,
u64
*
shadow_pte
,
u64
access_bits
,
gfn_t
gfn
)
u64
*
shadow_pte
,
u64
access_bits
,
int
write_fault
,
gfn_t
gfn
)
{
{
gpa_t
gaddr
;
gpa_t
gaddr
;
...
@@ -279,14 +293,14 @@ static void FNAME(set_pde)(struct kvm_vcpu *vcpu, pt_element_t *gpde,
...
@@ -279,14 +293,14 @@ static void FNAME(set_pde)(struct kvm_vcpu *vcpu, pt_element_t *gpde,
(
32
-
PT32_DIR_PSE36_SHIFT
);
(
32
-
PT32_DIR_PSE36_SHIFT
);
*
shadow_pte
=
*
gpde
&
PT_PTE_COPY_MASK
;
*
shadow_pte
=
*
gpde
&
PT_PTE_COPY_MASK
;
FNAME
(
set_pte_common
)(
vcpu
,
shadow_pte
,
gaddr
,
FNAME
(
set_pte_common
)(
vcpu
,
shadow_pte
,
gaddr
,
gpde
,
access_bits
,
gfn
);
gpde
,
access_bits
,
write_fault
,
gfn
);
}
}
/*
/*
* Fetch a shadow pte for a specific level in the paging hierarchy.
* Fetch a shadow pte for a specific level in the paging hierarchy.
*/
*/
static
u64
*
FNAME
(
fetch
)(
struct
kvm_vcpu
*
vcpu
,
gva_t
addr
,
static
u64
*
FNAME
(
fetch
)(
struct
kvm_vcpu
*
vcpu
,
gva_t
addr
,
struct
guest_walker
*
walker
)
struct
guest_walker
*
walker
,
int
write_fault
)
{
{
hpa_t
shadow_addr
;
hpa_t
shadow_addr
;
int
level
;
int
level
;
...
@@ -351,12 +365,12 @@ static u64 *FNAME(fetch)(struct kvm_vcpu *vcpu, gva_t addr,
...
@@ -351,12 +365,12 @@ static u64 *FNAME(fetch)(struct kvm_vcpu *vcpu, gva_t addr,
if
(
prev_shadow_ent
)
if
(
prev_shadow_ent
)
*
prev_shadow_ent
|=
PT_SHADOW_PS_MARK
;
*
prev_shadow_ent
|=
PT_SHADOW_PS_MARK
;
FNAME
(
set_pde
)(
vcpu
,
guest_ent
,
shadow_ent
,
FNAME
(
set_pde
)(
vcpu
,
guest_ent
,
shadow_ent
,
walker
->
inherited_ar
,
walker
->
gfn
);
walker
->
inherited_ar
,
w
rite_fault
,
w
alker
->
gfn
);
}
else
{
}
else
{
ASSERT
(
walker
->
level
==
PT_PAGE_TABLE_LEVEL
);
ASSERT
(
walker
->
level
==
PT_PAGE_TABLE_LEVEL
);
FNAME
(
set_pte
)(
vcpu
,
guest_ent
,
shadow_ent
,
FNAME
(
set_pte
)(
vcpu
,
guest_ent
,
shadow_ent
,
walker
->
inherited_ar
,
walker
->
inherited_ar
,
walker
->
gfn
);
w
rite_fault
,
w
alker
->
gfn
);
}
}
return
shadow_ent
;
return
shadow_ent
;
}
}
...
@@ -489,7 +503,7 @@ static int FNAME(page_fault)(struct kvm_vcpu *vcpu, gva_t addr,
...
@@ -489,7 +503,7 @@ static int FNAME(page_fault)(struct kvm_vcpu *vcpu, gva_t addr,
return
0
;
return
0
;
}
}
shadow_pte
=
FNAME
(
fetch
)(
vcpu
,
addr
,
&
walker
);
shadow_pte
=
FNAME
(
fetch
)(
vcpu
,
addr
,
&
walker
,
write_fault
);
pgprintk
(
"%s: shadow pte %p %llx
\n
"
,
__FUNCTION__
,
pgprintk
(
"%s: shadow pte %p %llx
\n
"
,
__FUNCTION__
,
shadow_pte
,
*
shadow_pte
);
shadow_pte
,
*
shadow_pte
);
...
@@ -499,8 +513,6 @@ static int FNAME(page_fault)(struct kvm_vcpu *vcpu, gva_t addr,
...
@@ -499,8 +513,6 @@ static int FNAME(page_fault)(struct kvm_vcpu *vcpu, gva_t addr,
if
(
write_fault
)
if
(
write_fault
)
fixed
=
FNAME
(
fix_write_pf
)(
vcpu
,
shadow_pte
,
&
walker
,
addr
,
fixed
=
FNAME
(
fix_write_pf
)(
vcpu
,
shadow_pte
,
&
walker
,
addr
,
user_fault
,
&
write_pt
);
user_fault
,
&
write_pt
);
else
fixed
=
fix_read_pf
(
shadow_pte
);
pgprintk
(
"%s: updated shadow pte %p %llx
\n
"
,
__FUNCTION__
,
pgprintk
(
"%s: updated shadow pte %p %llx
\n
"
,
__FUNCTION__
,
shadow_pte
,
*
shadow_pte
);
shadow_pte
,
*
shadow_pte
);
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment