Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Support
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in / Register
Toggle navigation
L
linux
Project overview
Project overview
Details
Activity
Releases
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Issues
0
Issues
0
List
Boards
Labels
Milestones
Merge Requests
0
Merge Requests
0
Analytics
Analytics
Repository
Value Stream
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Create a new issue
Commits
Issue Boards
Open sidebar
Kirill Smelkov
linux
Commits
b749567a
Commit
b749567a
authored
Apr 03, 2002
by
Linus Torvalds
Browse files
Options
Browse Files
Download
Plain Diff
Merge
parents
023f147f
7ee912ba
Changes
43
Show whitespace changes
Inline
Side-by-side
Showing
43 changed files
with
523 additions
and
472 deletions
+523
-472
arch/alpha/kernel/core_irongate.c
arch/alpha/kernel/core_irongate.c
+2
-0
arch/alpha/kernel/sys_alcor.c
arch/alpha/kernel/sys_alcor.c
+1
-0
arch/alpha/kernel/sys_cabriolet.c
arch/alpha/kernel/sys_cabriolet.c
+1
-0
arch/alpha/kernel/sys_dp264.c
arch/alpha/kernel/sys_dp264.c
+1
-0
arch/alpha/kernel/sys_eb64p.c
arch/alpha/kernel/sys_eb64p.c
+1
-0
arch/alpha/kernel/sys_eiger.c
arch/alpha/kernel/sys_eiger.c
+1
-0
arch/alpha/kernel/sys_jensen.c
arch/alpha/kernel/sys_jensen.c
+1
-0
arch/alpha/kernel/sys_miata.c
arch/alpha/kernel/sys_miata.c
+1
-0
arch/alpha/kernel/sys_mikasa.c
arch/alpha/kernel/sys_mikasa.c
+1
-0
arch/alpha/kernel/sys_nautilus.c
arch/alpha/kernel/sys_nautilus.c
+1
-0
arch/alpha/kernel/sys_noritake.c
arch/alpha/kernel/sys_noritake.c
+1
-0
arch/alpha/kernel/sys_rawhide.c
arch/alpha/kernel/sys_rawhide.c
+1
-0
arch/alpha/kernel/sys_ruffian.c
arch/alpha/kernel/sys_ruffian.c
+1
-0
arch/alpha/kernel/sys_rx164.c
arch/alpha/kernel/sys_rx164.c
+1
-0
arch/alpha/kernel/sys_sable.c
arch/alpha/kernel/sys_sable.c
+1
-0
arch/alpha/kernel/sys_sio.c
arch/alpha/kernel/sys_sio.c
+1
-0
arch/alpha/kernel/sys_sx164.c
arch/alpha/kernel/sys_sx164.c
+1
-0
arch/alpha/kernel/sys_takara.c
arch/alpha/kernel/sys_takara.c
+1
-0
arch/alpha/kernel/sys_titan.c
arch/alpha/kernel/sys_titan.c
+1
-0
arch/alpha/kernel/sys_wildfire.c
arch/alpha/kernel/sys_wildfire.c
+1
-0
arch/alpha/mm/fault.c
arch/alpha/mm/fault.c
+1
-1
arch/i386/kernel/acpi.c
arch/i386/kernel/acpi.c
+1
-0
arch/i386/kernel/vm86.c
arch/i386/kernel/vm86.c
+2
-0
arch/i386/mm/init.c
arch/i386/mm/init.c
+95
-1
arch/i386/mm/ioremap.c
arch/i386/mm/ioremap.c
+3
-0
fs/binfmt_aout.c
fs/binfmt_aout.c
+1
-0
include/asm-alpha/cacheflush.h
include/asm-alpha/cacheflush.h
+64
-0
include/asm-alpha/pgalloc.h
include/asm-alpha/pgalloc.h
+2
-222
include/asm-alpha/tlbflush.h
include/asm-alpha/tlbflush.h
+155
-0
include/asm-i386/cacheflush.h
include/asm-i386/cacheflush.h
+18
-0
include/asm-i386/pgalloc.h
include/asm-i386/pgalloc.h
+4
-182
include/asm-i386/pgtable.h
include/asm-i386/pgtable.h
+0
-65
include/asm-i386/tlbflush.h
include/asm-i386/tlbflush.h
+141
-0
include/linux/highmem.h
include/linux/highmem.h
+1
-0
kernel/fork.c
kernel/fork.c
+2
-0
kernel/module.c
kernel/module.c
+1
-0
mm/memory.c
mm/memory.c
+1
-0
mm/mmap.c
mm/mmap.c
+1
-0
mm/mprotect.c
mm/mprotect.c
+3
-1
mm/mremap.c
mm/mremap.c
+3
-0
mm/msync.c
mm/msync.c
+1
-0
mm/vmalloc.c
mm/vmalloc.c
+1
-0
mm/vmscan.c
mm/vmscan.c
+1
-0
No files found.
arch/alpha/kernel/core_irongate.c
View file @
b749567a
...
...
@@ -19,6 +19,8 @@
#include <asm/system.h>
#include <asm/pci.h>
#include <asm/hwrpb.h>
#include <asm/cacheflush.h>
#include <asm/tlbflush.h>
#define __EXTERN_INLINE inline
#include <asm/io.h>
...
...
arch/alpha/kernel/sys_alcor.c
View file @
b749567a
...
...
@@ -26,6 +26,7 @@
#include <asm/irq.h>
#include <asm/pgtable.h>
#include <asm/core_cia.h>
#include <asm/tlbflush.h>
#include "proto.h"
#include "irq_impl.h"
...
...
arch/alpha/kernel/sys_cabriolet.c
View file @
b749567a
...
...
@@ -28,6 +28,7 @@
#include <asm/core_apecs.h>
#include <asm/core_cia.h>
#include <asm/core_lca.h>
#include <asm/tlbflush.h>
#include "proto.h"
#include "irq_impl.h"
...
...
arch/alpha/kernel/sys_dp264.c
View file @
b749567a
...
...
@@ -30,6 +30,7 @@
#include <asm/pgtable.h>
#include <asm/core_tsunami.h>
#include <asm/hwrpb.h>
#include <asm/tlbflush.h>
#include "proto.h"
#include "irq_impl.h"
...
...
arch/alpha/kernel/sys_eb64p.c
View file @
b749567a
...
...
@@ -27,6 +27,7 @@
#include <asm/core_apecs.h>
#include <asm/core_lca.h>
#include <asm/hwrpb.h>
#include <asm/tlbflush.h>
#include "proto.h"
#include "irq_impl.h"
...
...
arch/alpha/kernel/sys_eiger.c
View file @
b749567a
...
...
@@ -27,6 +27,7 @@
#include <asm/pgtable.h>
#include <asm/core_tsunami.h>
#include <asm/hwrpb.h>
#include <asm/tlbflush.h>
#include "proto.h"
#include "irq_impl.h"
...
...
arch/alpha/kernel/sys_jensen.c
View file @
b749567a
...
...
@@ -26,6 +26,7 @@
#include <asm/irq.h>
#include <asm/mmu_context.h>
#include <asm/pgtable.h>
#include <asm/tlbflush.h>
#include "proto.h"
#include "irq_impl.h"
...
...
arch/alpha/kernel/sys_miata.c
View file @
b749567a
...
...
@@ -24,6 +24,7 @@
#include <asm/io.h>
#include <asm/pgtable.h>
#include <asm/core_cia.h>
#include <asm/tlbflush.h>
#include "proto.h"
#include "irq_impl.h"
...
...
arch/alpha/kernel/sys_mikasa.c
View file @
b749567a
...
...
@@ -26,6 +26,7 @@
#include <asm/pgtable.h>
#include <asm/core_apecs.h>
#include <asm/core_cia.h>
#include <asm/tlbflush.h>
#include "proto.h"
#include "irq_impl.h"
...
...
arch/alpha/kernel/sys_nautilus.c
View file @
b749567a
...
...
@@ -43,6 +43,7 @@
#include <asm/pgtable.h>
#include <asm/core_irongate.h>
#include <asm/hwrpb.h>
#include <asm/tlbflush.h>
#include "proto.h"
#include "irq_impl.h"
...
...
arch/alpha/kernel/sys_noritake.c
View file @
b749567a
...
...
@@ -27,6 +27,7 @@
#include <asm/pgtable.h>
#include <asm/core_apecs.h>
#include <asm/core_cia.h>
#include <asm/tlbflush.h>
#include "proto.h"
#include "irq_impl.h"
...
...
arch/alpha/kernel/sys_rawhide.c
View file @
b749567a
...
...
@@ -23,6 +23,7 @@
#include <asm/io.h>
#include <asm/pgtable.h>
#include <asm/core_mcpcia.h>
#include <asm/tlbflush.h>
#include "proto.h"
#include "irq_impl.h"
...
...
arch/alpha/kernel/sys_ruffian.c
View file @
b749567a
...
...
@@ -24,6 +24,7 @@
#include <asm/io.h>
#include <asm/pgtable.h>
#include <asm/core_cia.h>
#include <asm/tlbflush.h>
#include "proto.h"
#include "irq_impl.h"
...
...
arch/alpha/kernel/sys_rx164.c
View file @
b749567a
...
...
@@ -24,6 +24,7 @@
#include <asm/io.h>
#include <asm/pgtable.h>
#include <asm/core_polaris.h>
#include <asm/tlbflush.h>
#include "proto.h"
#include "irq_impl.h"
...
...
arch/alpha/kernel/sys_sable.c
View file @
b749567a
...
...
@@ -24,6 +24,7 @@
#include <asm/io.h>
#include <asm/pgtable.h>
#include <asm/core_t2.h>
#include <asm/tlbflush.h>
#include "proto.h"
#include "irq_impl.h"
...
...
arch/alpha/kernel/sys_sio.c
View file @
b749567a
...
...
@@ -29,6 +29,7 @@
#include <asm/pgtable.h>
#include <asm/core_apecs.h>
#include <asm/core_lca.h>
#include <asm/tlbflush.h>
#include "proto.h"
#include "irq_impl.h"
...
...
arch/alpha/kernel/sys_sx164.c
View file @
b749567a
...
...
@@ -25,6 +25,7 @@
#include <asm/pgtable.h>
#include <asm/core_cia.h>
#include <asm/hwrpb.h>
#include <asm/tlbflush.h>
#include "proto.h"
#include "irq_impl.h"
...
...
arch/alpha/kernel/sys_takara.c
View file @
b749567a
...
...
@@ -23,6 +23,7 @@
#include <asm/io.h>
#include <asm/pgtable.h>
#include <asm/core_cia.h>
#include <asm/tlbflush.h>
#include "proto.h"
#include "irq_impl.h"
...
...
arch/alpha/kernel/sys_titan.c
View file @
b749567a
...
...
@@ -28,6 +28,7 @@
#include <asm/pgtable.h>
#include <asm/core_titan.h>
#include <asm/hwrpb.h>
#include <asm/tlbflush.h>
#include "proto.h"
#include "irq_impl.h"
...
...
arch/alpha/kernel/sys_wildfire.c
View file @
b749567a
...
...
@@ -23,6 +23,7 @@
#include <asm/pgtable.h>
#include <asm/core_wildfire.h>
#include <asm/hwrpb.h>
#include <asm/tlbflush.h>
#include "proto.h"
#include "irq_impl.h"
...
...
arch/alpha/mm/fault.c
View file @
b749567a
...
...
@@ -12,7 +12,7 @@
#define __EXTERN_INLINE inline
#include <asm/mmu_context.h>
#include <asm/
pgalloc
.h>
#include <asm/
tlbflush
.h>
#undef __EXTERN_INLINE
#include <linux/signal.h>
...
...
arch/i386/kernel/acpi.c
View file @
b749567a
...
...
@@ -43,6 +43,7 @@
#include <asm/pgtable.h>
#include <asm/pgalloc.h>
#include <asm/io_apic.h>
#include <asm/tlbflush.h>
#define PREFIX "ACPI: "
...
...
arch/i386/kernel/vm86.c
View file @
b749567a
...
...
@@ -12,10 +12,12 @@
#include <linux/mm.h>
#include <linux/smp.h>
#include <linux/smp_lock.h>
#include <linux/highmem.h>
#include <asm/uaccess.h>
#include <asm/pgalloc.h>
#include <asm/io.h>
#include <asm/tlbflush.h>
/*
* Known problems:
...
...
arch/i386/mm/init.c
View file @
b749567a
...
...
@@ -37,6 +37,7 @@
#include <asm/e820.h>
#include <asm/apic.h>
#include <asm/tlb.h>
#include <asm/tlbflush.h>
mmu_gather_t
mmu_gathers
[
NR_CPUS
];
unsigned
long
highstart_pfn
,
highend_pfn
;
...
...
@@ -573,7 +574,8 @@ void si_meminfo(struct sysinfo *val)
}
#if defined(CONFIG_X86_PAE)
struct
kmem_cache_s
*
pae_pgd_cachep
;
static
struct
kmem_cache_s
*
pae_pgd_cachep
;
void
__init
pgtable_cache_init
(
void
)
{
/*
...
...
@@ -584,4 +586,96 @@ void __init pgtable_cache_init(void)
if
(
!
pae_pgd_cachep
)
panic
(
"init_pae(): Cannot alloc pae_pgd SLAB cache"
);
}
pgd_t
*
pgd_alloc
(
struct
mm_struct
*
mm
)
{
int
i
;
pgd_t
*
pgd
=
kmem_cache_alloc
(
pae_pgd_cachep
,
GFP_KERNEL
);
if
(
pgd
)
{
for
(
i
=
0
;
i
<
USER_PTRS_PER_PGD
;
i
++
)
{
unsigned
long
pmd
=
__get_free_page
(
GFP_KERNEL
);
if
(
!
pmd
)
goto
out_oom
;
clear_page
(
pmd
);
set_pgd
(
pgd
+
i
,
__pgd
(
1
+
__pa
(
pmd
)));
}
memcpy
(
pgd
+
USER_PTRS_PER_PGD
,
swapper_pg_dir
+
USER_PTRS_PER_PGD
,
(
PTRS_PER_PGD
-
USER_PTRS_PER_PGD
)
*
sizeof
(
pgd_t
));
}
return
pgd
;
out_oom:
for
(
i
--
;
i
>=
0
;
i
--
)
free_page
((
unsigned
long
)
__va
(
pgd_val
(
pgd
[
i
])
-
1
));
kmem_cache_free
(
pae_pgd_cachep
,
pgd
);
return
NULL
;
}
void
pgd_free
(
pgd_t
*
pgd
)
{
int
i
;
for
(
i
=
0
;
i
<
USER_PTRS_PER_PGD
;
i
++
)
free_page
((
unsigned
long
)
__va
(
pgd_val
(
pgd
[
i
])
-
1
));
kmem_cache_free
(
pae_pgd_cachep
,
pgd
);
}
#else
pgd_t
*
pgd_alloc
(
struct
mm_struct
*
mm
)
{
pgd_t
*
pgd
=
(
pgd_t
*
)
__get_free_page
(
GFP_KERNEL
);
if
(
pgd
)
{
memset
(
pgd
,
0
,
USER_PTRS_PER_PGD
*
sizeof
(
pgd_t
));
memcpy
(
pgd
+
USER_PTRS_PER_PGD
,
swapper_pg_dir
+
USER_PTRS_PER_PGD
,
(
PTRS_PER_PGD
-
USER_PTRS_PER_PGD
)
*
sizeof
(
pgd_t
));
}
return
pgd
;
}
void
pgd_free
(
pgd_t
*
pgd
)
{
free_page
((
unsigned
long
)
pgd
);
}
#endif
/* CONFIG_X86_PAE */
pte_t
*
pte_alloc_one_kernel
(
struct
mm_struct
*
mm
,
unsigned
long
address
)
{
int
count
=
0
;
pte_t
*
pte
;
do
{
pte
=
(
pte_t
*
)
__get_free_page
(
GFP_KERNEL
);
if
(
pte
)
clear_page
(
pte
);
else
{
current
->
state
=
TASK_UNINTERRUPTIBLE
;
schedule_timeout
(
HZ
);
}
}
while
(
!
pte
&&
(
count
++
<
10
));
return
pte
;
}
struct
page
*
pte_alloc_one
(
struct
mm_struct
*
mm
,
unsigned
long
address
)
{
int
count
=
0
;
struct
page
*
pte
;
do
{
#if CONFIG_HIGHPTE
pte
=
alloc_pages
(
GFP_KERNEL
|
__GFP_HIGHMEM
,
0
);
#else
pte
=
alloc_pages
(
GFP_KERNEL
,
0
);
#endif
if
(
pte
)
clear_highpage
(
pte
);
else
{
current
->
state
=
TASK_UNINTERRUPTIBLE
;
schedule_timeout
(
HZ
);
}
}
while
(
!
pte
&&
(
count
++
<
10
));
return
pte
;
}
arch/i386/mm/ioremap.c
View file @
b749567a
...
...
@@ -12,6 +12,9 @@
#include <asm/io.h>
#include <asm/pgalloc.h>
#include <asm/fixmap.h>
#include <asm/cacheflush.h>
#include <asm/tlbflush.h>
static
inline
void
remap_area_pte
(
pte_t
*
pte
,
unsigned
long
address
,
unsigned
long
size
,
unsigned
long
phys_addr
,
unsigned
long
flags
)
...
...
fs/binfmt_aout.c
View file @
b749567a
...
...
@@ -28,6 +28,7 @@
#include <asm/system.h>
#include <asm/uaccess.h>
#include <asm/pgalloc.h>
#include <asm/cacheflush.h>
static
int
load_aout_binary
(
struct
linux_binprm
*
,
struct
pt_regs
*
regs
);
static
int
load_aout_library
(
struct
file
*
);
...
...
include/asm-alpha/cacheflush.h
0 → 100644
View file @
b749567a
#ifndef _ALPHA_CACHEFLUSH_H
#define _ALPHA_CACHEFLUSH_H
#include <linux/config.h>
#include <linux/mm.h>
/* Caches aren't brain-dead on the Alpha. */
#define flush_cache_all() do { } while (0)
#define flush_cache_mm(mm) do { } while (0)
#define flush_cache_range(vma, start, end) do { } while (0)
#define flush_cache_page(vma, vmaddr) do { } while (0)
#define flush_page_to_ram(page) do { } while (0)
#define flush_dcache_page(page) do { } while (0)
/* Note that the following two definitions are _highly_ dependent
on the contexts in which they are used in the kernel. I personally
think it is criminal how loosely defined these macros are. */
/* We need to flush the kernel's icache after loading modules. The
only other use of this macro is in load_aout_interp which is not
used on Alpha.
Note that this definition should *not* be used for userspace
icache flushing. While functional, it is _way_ overkill. The
icache is tagged with ASNs and it suffices to allocate a new ASN
for the process. */
#ifndef CONFIG_SMP
#define flush_icache_range(start, end) imb()
#else
#define flush_icache_range(start, end) smp_imb()
extern
void
smp_imb
(
void
);
#endif
/* We need to flush the userspace icache after setting breakpoints in
ptrace.
Instead of indiscriminately using imb, take advantage of the fact
that icache entries are tagged with the ASN and load a new mm context. */
/* ??? Ought to use this in arch/alpha/kernel/signal.c too. */
#ifndef CONFIG_SMP
extern
void
__load_new_mm_context
(
struct
mm_struct
*
);
static
inline
void
flush_icache_user_range
(
struct
vm_area_struct
*
vma
,
struct
page
*
page
,
unsigned
long
addr
,
int
len
)
{
if
(
vma
->
vm_flags
&
VM_EXEC
)
{
struct
mm_struct
*
mm
=
vma
->
vm_mm
;
if
(
current
->
active_mm
==
mm
)
__load_new_mm_context
(
mm
);
else
mm
->
context
[
smp_processor_id
()]
=
0
;
}
}
#else
extern
void
flush_icache_user_range
(
struct
vm_area_struct
*
vma
,
struct
page
*
page
,
unsigned
long
addr
,
int
len
);
#endif
/* This is used only in do_no_page and do_swap_page. */
#define flush_icache_page(vma, page) \
flush_icache_user_range((vma), (page), 0, 0)
#endif
/* _ALPHA_CACHEFLUSH_H */
include/asm-alpha/pgalloc.h
View file @
b749567a
...
...
@@ -3,228 +3,6 @@
#include <linux/config.h>
#ifndef __EXTERN_INLINE
#define __EXTERN_INLINE extern inline
#define __MMU_EXTERN_INLINE
#endif
extern
void
__load_new_mm_context
(
struct
mm_struct
*
);
/* Caches aren't brain-dead on the Alpha. */
#define flush_cache_all() do { } while (0)
#define flush_cache_mm(mm) do { } while (0)
#define flush_cache_range(vma, start, end) do { } while (0)
#define flush_cache_page(vma, vmaddr) do { } while (0)
#define flush_page_to_ram(page) do { } while (0)
#define flush_dcache_page(page) do { } while (0)
/* Note that the following two definitions are _highly_ dependent
on the contexts in which they are used in the kernel. I personally
think it is criminal how loosely defined these macros are. */
/* We need to flush the kernel's icache after loading modules. The
only other use of this macro is in load_aout_interp which is not
used on Alpha.
Note that this definition should *not* be used for userspace
icache flushing. While functional, it is _way_ overkill. The
icache is tagged with ASNs and it suffices to allocate a new ASN
for the process. */
#ifndef CONFIG_SMP
#define flush_icache_range(start, end) imb()
#else
#define flush_icache_range(start, end) smp_imb()
extern
void
smp_imb
(
void
);
#endif
/*
* Use a few helper functions to hide the ugly broken ASN
* numbers on early Alphas (ev4 and ev45)
*/
__EXTERN_INLINE
void
ev4_flush_tlb_current
(
struct
mm_struct
*
mm
)
{
__load_new_mm_context
(
mm
);
tbiap
();
}
__EXTERN_INLINE
void
ev5_flush_tlb_current
(
struct
mm_struct
*
mm
)
{
__load_new_mm_context
(
mm
);
}
static
inline
void
flush_tlb_other
(
struct
mm_struct
*
mm
)
{
long
*
mmc
=
&
mm
->
context
[
smp_processor_id
()];
/*
* Check it's not zero first to avoid cacheline ping pong when
* possible.
*/
if
(
*
mmc
)
*
mmc
=
0
;
}
/* We need to flush the userspace icache after setting breakpoints in
ptrace.
Instead of indiscriminately using imb, take advantage of the fact
that icache entries are tagged with the ASN and load a new mm context. */
/* ??? Ought to use this in arch/alpha/kernel/signal.c too. */
#ifndef CONFIG_SMP
static
inline
void
flush_icache_user_range
(
struct
vm_area_struct
*
vma
,
struct
page
*
page
,
unsigned
long
addr
,
int
len
)
{
if
(
vma
->
vm_flags
&
VM_EXEC
)
{
struct
mm_struct
*
mm
=
vma
->
vm_mm
;
if
(
current
->
active_mm
==
mm
)
__load_new_mm_context
(
mm
);
else
mm
->
context
[
smp_processor_id
()]
=
0
;
}
}
#else
extern
void
flush_icache_user_range
(
struct
vm_area_struct
*
vma
,
struct
page
*
page
,
unsigned
long
addr
,
int
len
);
#endif
/* this is used only in do_no_page and do_swap_page */
#define flush_icache_page(vma, page) flush_icache_user_range((vma), (page), 0, 0)
/*
* Flush just one page in the current TLB set.
* We need to be very careful about the icache here, there
* is no way to invalidate a specific icache page..
*/
__EXTERN_INLINE
void
ev4_flush_tlb_current_page
(
struct
mm_struct
*
mm
,
struct
vm_area_struct
*
vma
,
unsigned
long
addr
)
{
int
tbi_flag
=
2
;
if
(
vma
->
vm_flags
&
VM_EXEC
)
{
__load_new_mm_context
(
mm
);
tbi_flag
=
3
;
}
tbi
(
tbi_flag
,
addr
);
}
__EXTERN_INLINE
void
ev5_flush_tlb_current_page
(
struct
mm_struct
*
mm
,
struct
vm_area_struct
*
vma
,
unsigned
long
addr
)
{
if
(
vma
->
vm_flags
&
VM_EXEC
)
__load_new_mm_context
(
mm
);
else
tbi
(
2
,
addr
);
}
#ifdef CONFIG_ALPHA_GENERIC
# define flush_tlb_current alpha_mv.mv_flush_tlb_current
# define flush_tlb_current_page alpha_mv.mv_flush_tlb_current_page
#else
# ifdef CONFIG_ALPHA_EV4
# define flush_tlb_current ev4_flush_tlb_current
# define flush_tlb_current_page ev4_flush_tlb_current_page
# else
# define flush_tlb_current ev5_flush_tlb_current
# define flush_tlb_current_page ev5_flush_tlb_current_page
# endif
#endif
#ifdef __MMU_EXTERN_INLINE
#undef __EXTERN_INLINE
#undef __MMU_EXTERN_INLINE
#endif
/*
* Flush current user mapping.
*/
static
inline
void
flush_tlb
(
void
)
{
flush_tlb_current
(
current
->
active_mm
);
}
/*
* Flush a specified range of user mapping page tables
* from TLB.
* Although Alpha uses VPTE caches, this can be a nop, as Alpha does
* not have finegrained tlb flushing, so it will flush VPTE stuff
* during next flush_tlb_range.
*/
static
inline
void
flush_tlb_pgtables
(
struct
mm_struct
*
mm
,
unsigned
long
start
,
unsigned
long
end
)
{
}
#ifndef CONFIG_SMP
/*
* Flush everything (kernel mapping may also have
* changed due to vmalloc/vfree)
*/
static
inline
void
flush_tlb_all
(
void
)
{
tbia
();
}
/*
* Flush a specified user mapping
*/
static
inline
void
flush_tlb_mm
(
struct
mm_struct
*
mm
)
{
if
(
mm
==
current
->
active_mm
)
flush_tlb_current
(
mm
);
else
flush_tlb_other
(
mm
);
}
/*
* Page-granular tlb flush.
*
* do a tbisd (type = 2) normally, and a tbis (type = 3)
* if it is an executable mapping. We want to avoid the
* itlb flush, because that potentially also does a
* icache flush.
*/
static
inline
void
flush_tlb_page
(
struct
vm_area_struct
*
vma
,
unsigned
long
addr
)
{
struct
mm_struct
*
mm
=
vma
->
vm_mm
;
if
(
mm
==
current
->
active_mm
)
flush_tlb_current_page
(
mm
,
vma
,
addr
);
else
flush_tlb_other
(
mm
);
}
/*
* Flush a specified range of user mapping: on the
* Alpha we flush the whole user tlb.
*/
static
inline
void
flush_tlb_range
(
struct
vm_area_struct
*
vma
,
unsigned
long
start
,
unsigned
long
end
)
{
flush_tlb_mm
(
vma
->
vm_mm
);
}
#else
/* CONFIG_SMP */
extern
void
flush_tlb_all
(
void
);
extern
void
flush_tlb_mm
(
struct
mm_struct
*
);
extern
void
flush_tlb_page
(
struct
vm_area_struct
*
,
unsigned
long
);
extern
void
flush_tlb_range
(
struct
vm_area_struct
*
,
unsigned
long
,
unsigned
long
);
#endif
/* CONFIG_SMP */
/*
* Allocate and free page tables. The xxx_kernel() versions are
* used to allocate a kernel page table - this turns on ASN bits
...
...
@@ -292,4 +70,6 @@ pte_free(struct page *page)
__free_page
(
page
);
}
#define check_pgt_cache() do { } while (0)
#endif
/* _ALPHA_PGALLOC_H */
include/asm-alpha/tlbflush.h
0 → 100644
View file @
b749567a
#ifndef _ALPHA_TLBFLUSH_H
#define _ALPHA_TLBFLUSH_H
#include <linux/config.h>
#include <linux/mm.h>
#ifndef __EXTERN_INLINE
#define __EXTERN_INLINE extern inline
#define __MMU_EXTERN_INLINE
#endif
extern
void
__load_new_mm_context
(
struct
mm_struct
*
);
/* Use a few helper functions to hide the ugly broken ASN
numbers on early Alphas (ev4 and ev45). */
__EXTERN_INLINE
void
ev4_flush_tlb_current
(
struct
mm_struct
*
mm
)
{
__load_new_mm_context
(
mm
);
tbiap
();
}
__EXTERN_INLINE
void
ev5_flush_tlb_current
(
struct
mm_struct
*
mm
)
{
__load_new_mm_context
(
mm
);
}
/* Flush just one page in the current TLB set. We need to be very
careful about the icache here, there is no way to invalidate a
specific icache page. */
__EXTERN_INLINE
void
ev4_flush_tlb_current_page
(
struct
mm_struct
*
mm
,
struct
vm_area_struct
*
vma
,
unsigned
long
addr
)
{
int
tbi_flag
=
2
;
if
(
vma
->
vm_flags
&
VM_EXEC
)
{
__load_new_mm_context
(
mm
);
tbi_flag
=
3
;
}
tbi
(
tbi_flag
,
addr
);
}
__EXTERN_INLINE
void
ev5_flush_tlb_current_page
(
struct
mm_struct
*
mm
,
struct
vm_area_struct
*
vma
,
unsigned
long
addr
)
{
if
(
vma
->
vm_flags
&
VM_EXEC
)
__load_new_mm_context
(
mm
);
else
tbi
(
2
,
addr
);
}
#ifdef CONFIG_ALPHA_GENERIC
# define flush_tlb_current alpha_mv.mv_flush_tlb_current
# define flush_tlb_current_page alpha_mv.mv_flush_tlb_current_page
#else
# ifdef CONFIG_ALPHA_EV4
# define flush_tlb_current ev4_flush_tlb_current
# define flush_tlb_current_page ev4_flush_tlb_current_page
# else
# define flush_tlb_current ev5_flush_tlb_current
# define flush_tlb_current_page ev5_flush_tlb_current_page
# endif
#endif
#ifdef __MMU_EXTERN_INLINE
#undef __EXTERN_INLINE
#undef __MMU_EXTERN_INLINE
#endif
/* Flush current user mapping. */
static
inline
void
flush_tlb
(
void
)
{
flush_tlb_current
(
current
->
active_mm
);
}
/* Flush someone else's user mapping. */
static
inline
void
flush_tlb_other
(
struct
mm_struct
*
mm
)
{
long
*
mmc
=
&
mm
->
context
[
smp_processor_id
()];
/* Check it's not zero first to avoid cacheline ping pong
when possible. */
if
(
*
mmc
)
*
mmc
=
0
;
}
/* Flush a specified range of user mapping page tables from TLB.
Although Alpha uses VPTE caches, this can be a nop, as Alpha does
not have finegrained tlb flushing, so it will flush VPTE stuff
during next flush_tlb_range. */
static
inline
void
flush_tlb_pgtables
(
struct
mm_struct
*
mm
,
unsigned
long
start
,
unsigned
long
end
)
{
}
#ifndef CONFIG_SMP
/* Flush everything (kernel mapping may also have changed
due to vmalloc/vfree). */
static
inline
void
flush_tlb_all
(
void
)
{
tbia
();
}
/* Flush a specified user mapping. */
static
inline
void
flush_tlb_mm
(
struct
mm_struct
*
mm
)
{
if
(
mm
==
current
->
active_mm
)
flush_tlb_current
(
mm
);
else
flush_tlb_other
(
mm
);
}
/* Page-granular tlb flush. */
static
inline
void
flush_tlb_page
(
struct
vm_area_struct
*
vma
,
unsigned
long
addr
)
{
struct
mm_struct
*
mm
=
vma
->
vm_mm
;
if
(
mm
==
current
->
active_mm
)
flush_tlb_current_page
(
mm
,
vma
,
addr
);
else
flush_tlb_other
(
mm
);
}
/* Flush a specified range of user mapping. On the Alpha we flush
the whole user tlb. */
static
inline
void
flush_tlb_range
(
struct
vm_area_struct
*
vma
,
unsigned
long
start
,
unsigned
long
end
)
{
flush_tlb_mm
(
vma
->
vm_mm
);
}
#else
/* CONFIG_SMP */
extern
void
flush_tlb_all
(
void
);
extern
void
flush_tlb_mm
(
struct
mm_struct
*
);
extern
void
flush_tlb_page
(
struct
vm_area_struct
*
,
unsigned
long
);
extern
void
flush_tlb_range
(
struct
vm_area_struct
*
,
unsigned
long
,
unsigned
long
);
#endif
/* CONFIG_SMP */
#endif
/* _ALPHA_TLBFLUSH_H */
include/asm-i386/cacheflush.h
0 → 100644
View file @
b749567a
#ifndef _I386_CACHEFLUSH_H
#define _I386_CACHEFLUSH_H
/* Keep includes the same across arches. */
#include <linux/mm.h>
/* Caches aren't brain-dead on the intel. */
#define flush_cache_all() do { } while (0)
#define flush_cache_mm(mm) do { } while (0)
#define flush_cache_range(vma, start, end) do { } while (0)
#define flush_cache_page(vma, vmaddr) do { } while (0)
#define flush_page_to_ram(page) do { } while (0)
#define flush_dcache_page(page) do { } while (0)
#define flush_icache_range(start, end) do { } while (0)
#define flush_icache_page(vma,pg) do { } while (0)
#define flush_icache_user_range(vma,pg,adr,len) do { } while (0)
#endif
/* _I386_CACHEFLUSH_H */
include/asm-i386/pgalloc.h
View file @
b749567a
...
...
@@ -5,7 +5,6 @@
#include <asm/processor.h>
#include <asm/fixmap.h>
#include <linux/threads.h>
#include <linux/highmem.h>
#define pmd_populate_kernel(mm, pmd, pte) \
set_pmd(pmd, __pmd(_PAGE_TABLE + __pa(pte)))
...
...
@@ -20,109 +19,11 @@ static inline void pmd_populate(struct mm_struct *mm, pmd_t *pmd, struct page *p
* Allocate and free page tables.
*/
#if defined (CONFIG_X86_PAE)
/*
* We can't include <linux/slab.h> here, thus these uglinesses.
*/
struct
kmem_cache_s
;
extern
struct
kmem_cache_s
*
pae_pgd_cachep
;
extern
void
*
kmem_cache_alloc
(
struct
kmem_cache_s
*
,
int
);
extern
void
kmem_cache_free
(
struct
kmem_cache_s
*
,
void
*
);
static
inline
pgd_t
*
pgd_alloc
(
struct
mm_struct
*
mm
)
{
int
i
;
pgd_t
*
pgd
=
kmem_cache_alloc
(
pae_pgd_cachep
,
GFP_KERNEL
);
if
(
pgd
)
{
for
(
i
=
0
;
i
<
USER_PTRS_PER_PGD
;
i
++
)
{
unsigned
long
pmd
=
__get_free_page
(
GFP_KERNEL
);
if
(
!
pmd
)
goto
out_oom
;
clear_page
(
pmd
);
set_pgd
(
pgd
+
i
,
__pgd
(
1
+
__pa
(
pmd
)));
}
memcpy
(
pgd
+
USER_PTRS_PER_PGD
,
swapper_pg_dir
+
USER_PTRS_PER_PGD
,
(
PTRS_PER_PGD
-
USER_PTRS_PER_PGD
)
*
sizeof
(
pgd_t
));
}
return
pgd
;
out_oom:
for
(
i
--
;
i
>=
0
;
i
--
)
free_page
((
unsigned
long
)
__va
(
pgd_val
(
pgd
[
i
])
-
1
));
kmem_cache_free
(
pae_pgd_cachep
,
pgd
);
return
NULL
;
}
#else
static
inline
pgd_t
*
pgd_alloc
(
struct
mm_struct
*
mm
)
{
pgd_t
*
pgd
=
(
pgd_t
*
)
__get_free_page
(
GFP_KERNEL
);
if
(
pgd
)
{
memset
(
pgd
,
0
,
USER_PTRS_PER_PGD
*
sizeof
(
pgd_t
));
memcpy
(
pgd
+
USER_PTRS_PER_PGD
,
swapper_pg_dir
+
USER_PTRS_PER_PGD
,
(
PTRS_PER_PGD
-
USER_PTRS_PER_PGD
)
*
sizeof
(
pgd_t
));
}
return
pgd
;
}
#endif
/* CONFIG_X86_PAE */
extern
pgd_t
*
pgd_alloc
(
struct
mm_struct
*
);
extern
void
pgd_free
(
pgd_t
*
pgd
);
static
inline
void
pgd_free
(
pgd_t
*
pgd
)
{
#if defined(CONFIG_X86_PAE)
int
i
;
for
(
i
=
0
;
i
<
USER_PTRS_PER_PGD
;
i
++
)
free_page
((
unsigned
long
)
__va
(
pgd_val
(
pgd
[
i
])
-
1
));
kmem_cache_free
(
pae_pgd_cachep
,
pgd
);
#else
free_page
((
unsigned
long
)
pgd
);
#endif
}
static
inline
pte_t
*
pte_alloc_one_kernel
(
struct
mm_struct
*
mm
,
unsigned
long
address
)
{
int
count
=
0
;
pte_t
*
pte
;
do
{
pte
=
(
pte_t
*
)
__get_free_page
(
GFP_KERNEL
);
if
(
pte
)
clear_page
(
pte
);
else
{
current
->
state
=
TASK_UNINTERRUPTIBLE
;
schedule_timeout
(
HZ
);
}
}
while
(
!
pte
&&
(
count
++
<
10
));
return
pte
;
}
static
inline
struct
page
*
pte_alloc_one
(
struct
mm_struct
*
mm
,
unsigned
long
address
)
{
int
count
=
0
;
struct
page
*
pte
;
do
{
#if CONFIG_HIGHPTE
pte
=
alloc_pages
(
GFP_KERNEL
|
__GFP_HIGHMEM
,
0
);
#else
pte
=
alloc_pages
(
GFP_KERNEL
,
0
);
#endif
if
(
pte
)
clear_highpage
(
pte
);
else
{
current
->
state
=
TASK_UNINTERRUPTIBLE
;
schedule_timeout
(
HZ
);
}
}
while
(
!
pte
&&
(
count
++
<
10
));
return
pte
;
}
extern
pte_t
*
pte_alloc_one_kernel
(
struct
mm_struct
*
,
unsigned
long
);
extern
struct
page
*
pte_alloc_one
(
struct
mm_struct
*
,
unsigned
long
);
static
inline
void
pte_free_kernel
(
pte_t
*
pte
)
{
...
...
@@ -144,85 +45,6 @@ static inline void pte_free(struct page *pte)
#define pmd_free(x) do { } while (0)
#define pgd_populate(mm, pmd, pte) BUG()
/*
* TLB flushing:
*
* - flush_tlb() flushes the current mm struct TLBs
* - flush_tlb_all() flushes all processes TLBs
* - flush_tlb_mm(mm) flushes the specified mm context TLB's
* - flush_tlb_page(vma, vmaddr) flushes one page
* - flush_tlb_range(vma, start, end) flushes a range of pages
* - flush_tlb_pgtables(mm, start, end) flushes a range of page tables
*
* ..but the i386 has somewhat limited tlb flushing capabilities,
* and page-granular flushes are available only on i486 and up.
*/
#ifndef CONFIG_SMP
#define flush_tlb() __flush_tlb()
#define flush_tlb_all() __flush_tlb_all()
#define local_flush_tlb() __flush_tlb()
static
inline
void
flush_tlb_mm
(
struct
mm_struct
*
mm
)
{
if
(
mm
==
current
->
active_mm
)
__flush_tlb
();
}
static
inline
void
flush_tlb_page
(
struct
vm_area_struct
*
vma
,
unsigned
long
addr
)
{
if
(
vma
->
vm_mm
==
current
->
active_mm
)
__flush_tlb_one
(
addr
);
}
static
inline
void
flush_tlb_range
(
struct
vm_area_struct
*
vma
,
unsigned
long
start
,
unsigned
long
end
)
{
if
(
vma
->
vm_mm
==
current
->
active_mm
)
__flush_tlb
();
}
#else
#include <asm/smp.h>
#define local_flush_tlb() \
__flush_tlb()
extern
void
flush_tlb_all
(
void
);
extern
void
flush_tlb_current_task
(
void
);
extern
void
flush_tlb_mm
(
struct
mm_struct
*
);
extern
void
flush_tlb_page
(
struct
vm_area_struct
*
,
unsigned
long
);
#define flush_tlb() flush_tlb_current_task()
static
inline
void
flush_tlb_range
(
struct
vm_area_struct
*
vma
,
unsigned
long
start
,
unsigned
long
end
)
{
flush_tlb_mm
(
vma
->
vm_mm
);
}
#define TLBSTATE_OK 1
#define TLBSTATE_LAZY 2
struct
tlb_state
{
struct
mm_struct
*
active_mm
;
int
state
;
char
__cacheline_padding
[
24
];
};
extern
struct
tlb_state
cpu_tlbstate
[
NR_CPUS
];
#endif
static
inline
void
flush_tlb_pgtables
(
struct
mm_struct
*
mm
,
unsigned
long
start
,
unsigned
long
end
)
{
/* i386 does not keep any page table caches in TLB */
}
#define check_pgt_cache() do { } while (0)
#endif
/* _I386_PGALLOC_H */
include/asm-i386/pgtable.h
View file @
b749567a
...
...
@@ -24,71 +24,6 @@
extern
pgd_t
swapper_pg_dir
[
1024
];
extern
void
paging_init
(
void
);
/* Caches aren't brain-dead on the intel. */
#define flush_cache_all() do { } while (0)
#define flush_cache_mm(mm) do { } while (0)
#define flush_cache_range(vma, start, end) do { } while (0)
#define flush_cache_page(vma, vmaddr) do { } while (0)
#define flush_page_to_ram(page) do { } while (0)
#define flush_dcache_page(page) do { } while (0)
#define flush_icache_range(start, end) do { } while (0)
#define flush_icache_page(vma,pg) do { } while (0)
#define flush_icache_user_range(vma,pg,adr,len) do { } while (0)
#define __flush_tlb() \
do { \
unsigned int tmpreg; \
\
__asm__ __volatile__( \
"movl %%cr3, %0; # flush TLB \n" \
"movl %0, %%cr3; \n" \
: "=r" (tmpreg) \
:: "memory"); \
} while (0)
/*
* Global pages have to be flushed a bit differently. Not a real
* performance problem because this does not happen often.
*/
#define __flush_tlb_global() \
do { \
unsigned int tmpreg; \
\
__asm__ __volatile__( \
"movl %1, %%cr4; # turn off PGE \n" \
"movl %%cr3, %0; # flush TLB \n" \
"movl %0, %%cr3; \n" \
"movl %2, %%cr4; # turn PGE back on \n" \
: "=&r" (tmpreg) \
: "r" (mmu_cr4_features & ~X86_CR4_PGE), \
"r" (mmu_cr4_features) \
: "memory"); \
} while (0)
extern
unsigned
long
pgkern_mask
;
/*
* Do not check the PGE bit unnecesserily if this is a PPro+ kernel.
*/
#ifdef CONFIG_X86_PGE
# define __flush_tlb_all() __flush_tlb_global()
#else
# define __flush_tlb_all() \
do { \
if (cpu_has_pge) \
__flush_tlb_global(); \
else \
__flush_tlb(); \
} while (0)
#endif
#ifndef CONFIG_X86_INVLPG
#define __flush_tlb_one(addr) __flush_tlb()
#else
#define __flush_tlb_one(addr) \
__asm__ __volatile__("invlpg %0": :"m" (*(char *) addr))
#endif
/*
* ZERO_PAGE is a global shared page that is always zero: used
* for zero-mapped memory areas etc..
...
...
include/asm-i386/tlbflush.h
0 → 100644
View file @
b749567a
#ifndef _I386_TLBFLUSH_H
#define _I386_TLBFLUSH_H
#include <linux/config.h>
#include <linux/mm.h>
#include <asm/processor.h>
#define __flush_tlb() \
do { \
unsigned int tmpreg; \
\
__asm__ __volatile__( \
"movl %%cr3, %0; # flush TLB \n" \
"movl %0, %%cr3; \n" \
: "=r" (tmpreg) \
:: "memory"); \
} while (0)
/*
* Global pages have to be flushed a bit differently. Not a real
* performance problem because this does not happen often.
*/
#define __flush_tlb_global() \
do { \
unsigned int tmpreg; \
\
__asm__ __volatile__( \
"movl %1, %%cr4; # turn off PGE \n" \
"movl %%cr3, %0; # flush TLB \n" \
"movl %0, %%cr3; \n" \
"movl %2, %%cr4; # turn PGE back on \n" \
: "=&r" (tmpreg) \
: "r" (mmu_cr4_features & ~X86_CR4_PGE), \
"r" (mmu_cr4_features) \
: "memory"); \
} while (0)
extern
unsigned
long
pgkern_mask
;
/*
* Do not check the PGE bit unnecesserily if this is a PPro+ kernel.
*/
#ifdef CONFIG_X86_PGE
# define __flush_tlb_all() __flush_tlb_global()
#else
# define __flush_tlb_all() \
do { \
if (cpu_has_pge) \
__flush_tlb_global(); \
else \
__flush_tlb(); \
} while (0)
#endif
#ifndef CONFIG_X86_INVLPG
#define __flush_tlb_one(addr) __flush_tlb()
#else
#define __flush_tlb_one(addr) \
__asm__ __volatile__("invlpg %0": :"m" (*(char *) addr))
#endif
/*
* TLB flushing:
*
* - flush_tlb() flushes the current mm struct TLBs
* - flush_tlb_all() flushes all processes TLBs
* - flush_tlb_mm(mm) flushes the specified mm context TLB's
* - flush_tlb_page(vma, vmaddr) flushes one page
* - flush_tlb_range(vma, start, end) flushes a range of pages
* - flush_tlb_pgtables(mm, start, end) flushes a range of page tables
*
* ..but the i386 has somewhat limited tlb flushing capabilities,
* and page-granular flushes are available only on i486 and up.
*/
#ifndef CONFIG_SMP
#define flush_tlb() __flush_tlb()
#define flush_tlb_all() __flush_tlb_all()
#define local_flush_tlb() __flush_tlb()
static
inline
void
flush_tlb_mm
(
struct
mm_struct
*
mm
)
{
if
(
mm
==
current
->
active_mm
)
__flush_tlb
();
}
static
inline
void
flush_tlb_page
(
struct
vm_area_struct
*
vma
,
unsigned
long
addr
)
{
if
(
vma
->
vm_mm
==
current
->
active_mm
)
__flush_tlb_one
(
addr
);
}
static
inline
void
flush_tlb_range
(
struct
vm_area_struct
*
vma
,
unsigned
long
start
,
unsigned
long
end
)
{
if
(
vma
->
vm_mm
==
current
->
active_mm
)
__flush_tlb
();
}
#else
#include <asm/smp.h>
#define local_flush_tlb() \
__flush_tlb()
extern
void
flush_tlb_all
(
void
);
extern
void
flush_tlb_current_task
(
void
);
extern
void
flush_tlb_mm
(
struct
mm_struct
*
);
extern
void
flush_tlb_page
(
struct
vm_area_struct
*
,
unsigned
long
);
#define flush_tlb() flush_tlb_current_task()
static
inline
void
flush_tlb_range
(
struct
vm_area_struct
*
vma
,
unsigned
long
start
,
unsigned
long
end
)
{
flush_tlb_mm
(
vma
->
vm_mm
);
}
#define TLBSTATE_OK 1
#define TLBSTATE_LAZY 2
struct
tlb_state
{
struct
mm_struct
*
active_mm
;
int
state
;
char
__cacheline_padding
[
24
];
};
extern
struct
tlb_state
cpu_tlbstate
[
NR_CPUS
];
#endif
static
inline
void
flush_tlb_pgtables
(
struct
mm_struct
*
mm
,
unsigned
long
start
,
unsigned
long
end
)
{
/* i386 does not keep any page table caches in TLB */
}
#endif
/* _I386_TLBFLUSH_H */
include/linux/highmem.h
View file @
b749567a
...
...
@@ -4,6 +4,7 @@
#include <linux/config.h>
#include <linux/bio.h>
#include <linux/fs.h>
#include <asm/cacheflush.h>
#ifdef CONFIG_HIGHMEM
...
...
kernel/fork.c
View file @
b749567a
...
...
@@ -29,6 +29,8 @@
#include <asm/pgalloc.h>
#include <asm/uaccess.h>
#include <asm/mmu_context.h>
#include <asm/cacheflush.h>
#include <asm/tlbflush.h>
static
kmem_cache_t
*
task_struct_cachep
;
...
...
kernel/module.c
View file @
b749567a
...
...
@@ -11,6 +11,7 @@
#include <linux/kmod.h>
#include <linux/seq_file.h>
#include <linux/fs.h>
#include <asm/cacheflush.h>
/*
* Originally by Anonymous (as far as I know...)
...
...
mm/memory.c
View file @
b749567a
...
...
@@ -48,6 +48,7 @@
#include <asm/pgalloc.h>
#include <asm/uaccess.h>
#include <asm/tlb.h>
#include <asm/tlbflush.h>
unsigned
long
max_mapnr
;
unsigned
long
num_physpages
;
...
...
mm/mmap.c
View file @
b749567a
...
...
@@ -17,6 +17,7 @@
#include <asm/uaccess.h>
#include <asm/pgalloc.h>
#include <asm/tlbflush.h>
/*
* WARNING: the debugging will use recursive algorithms so never enable this
...
...
mm/mprotect.c
View file @
b749567a
...
...
@@ -9,11 +9,13 @@
#include <linux/shm.h>
#include <linux/mman.h>
#include <linux/fs.h>
#include <linux/highmem.h>
#include <asm/uaccess.h>
#include <asm/pgalloc.h>
#include <asm/pgtable.h>
#include <linux/highmem.h>
#include <asm/cacheflush.h>
#include <asm/tlbflush.h>
static
inline
void
change_pte_range
(
pmd_t
*
pmd
,
unsigned
long
address
,
unsigned
long
size
,
pgprot_t
newprot
)
...
...
mm/mremap.c
View file @
b749567a
...
...
@@ -11,9 +11,12 @@
#include <linux/mman.h>
#include <linux/swap.h>
#include <linux/fs.h>
#include <linux/highmem.h>
#include <asm/uaccess.h>
#include <asm/pgalloc.h>
#include <asm/cacheflush.h>
#include <asm/tlbflush.h>
extern
int
vm_enough_memory
(
long
pages
);
...
...
mm/msync.c
View file @
b749567a
...
...
@@ -14,6 +14,7 @@
#include <asm/pgtable.h>
#include <asm/pgalloc.h>
#include <asm/tlbflush.h>
/*
* Called with mm->page_table_lock held to protect against other
...
...
mm/vmalloc.c
View file @
b749567a
...
...
@@ -15,6 +15,7 @@
#include <asm/uaccess.h>
#include <asm/pgalloc.h>
#include <asm/tlbflush.h>
rwlock_t
vmlist_lock
=
RW_LOCK_UNLOCKED
;
struct
vm_struct
*
vmlist
;
...
...
mm/vmscan.c
View file @
b749567a
...
...
@@ -24,6 +24,7 @@
#include <linux/compiler.h>
#include <asm/pgalloc.h>
#include <asm/tlbflush.h>
/*
* The "priority" of VM scanning is how much of the queues we
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment