Commit 3c5a8af5 authored by Andrew Morton's avatar Andrew Morton Committed by David S. Miller

[PATCH] allow CONFIG_SWAP=n for i386

Patch from Christoph Hellwig <hch@sgi.com>

There's a bunch of minor fixes needed to disable the swap code for systems
with mmu.
parent d3a3fe93
...@@ -19,8 +19,13 @@ config MMU ...@@ -19,8 +19,13 @@ config MMU
default y default y
config SWAP config SWAP
bool bool "Support for paging of anonymous memory"
default y default y
help
This option allows you to choose whether you want to have support
for socalled swap devices or swap files in your kernel that are
used to provide more virtual memory than the actual RAM present
in your computer. If unusre say Y.
config SBUS config SBUS
bool bool
......
...@@ -11,6 +11,7 @@ ...@@ -11,6 +11,7 @@
#include <linux/smp.h> #include <linux/smp.h>
#include <linux/highmem.h> #include <linux/highmem.h>
#include <linux/slab.h> #include <linux/slab.h>
#include <linux/pagemap.h>
#include <asm/system.h> #include <asm/system.h>
#include <asm/pgtable.h> #include <asm/pgtable.h>
......
...@@ -68,10 +68,11 @@ typedef struct { ...@@ -68,10 +68,11 @@ typedef struct {
#ifdef __KERNEL__ #ifdef __KERNEL__
struct sysinfo;
struct address_space; struct address_space;
struct zone; struct pte_chain;
struct sysinfo;
struct writeback_control; struct writeback_control;
struct zone;
/* /*
* A swap extent maps a range of a swapfile's PAGE_SIZE pages onto a range of * A swap extent maps a range of a swapfile's PAGE_SIZE pages onto a range of
...@@ -140,6 +141,9 @@ struct swap_list_t { ...@@ -140,6 +141,9 @@ struct swap_list_t {
/* Swap 50% full? Release swapcache more aggressively.. */ /* Swap 50% full? Release swapcache more aggressively.. */
#define vm_swap_full() (nr_swap_pages*2 < total_swap_pages) #define vm_swap_full() (nr_swap_pages*2 < total_swap_pages)
/* linux/mm/oom_kill.c */
extern void out_of_memory(void);
/* linux/mm/page_alloc.c */ /* linux/mm/page_alloc.c */
extern unsigned long totalram_pages; extern unsigned long totalram_pages;
extern unsigned long totalhigh_pages; extern unsigned long totalhigh_pages;
...@@ -149,13 +153,11 @@ extern unsigned int nr_free_pages_pgdat(pg_data_t *pgdat); ...@@ -149,13 +153,11 @@ extern unsigned int nr_free_pages_pgdat(pg_data_t *pgdat);
extern unsigned int nr_free_buffer_pages(void); extern unsigned int nr_free_buffer_pages(void);
extern unsigned int nr_free_pagecache_pages(void); extern unsigned int nr_free_pagecache_pages(void);
/* linux/mm/filemap.c */
extern void FASTCALL(mark_page_accessed(struct page *));
/* linux/mm/swap.c */ /* linux/mm/swap.c */
extern void FASTCALL(lru_cache_add(struct page *)); extern void FASTCALL(lru_cache_add(struct page *));
extern void FASTCALL(lru_cache_add_active(struct page *)); extern void FASTCALL(lru_cache_add_active(struct page *));
extern void FASTCALL(activate_page(struct page *)); extern void FASTCALL(activate_page(struct page *));
extern void FASTCALL(mark_page_accessed(struct page *));
extern void lru_add_drain(void); extern void lru_add_drain(void);
extern int rotate_reclaimable_page(struct page *page); extern int rotate_reclaimable_page(struct page *page);
extern void swap_setup(void); extern void swap_setup(void);
...@@ -165,11 +167,8 @@ extern int try_to_free_pages(struct zone *, unsigned int, unsigned int); ...@@ -165,11 +167,8 @@ extern int try_to_free_pages(struct zone *, unsigned int, unsigned int);
extern int shrink_all_memory(int); extern int shrink_all_memory(int);
extern int vm_swappiness; extern int vm_swappiness;
/* linux/mm/oom_kill.c */
extern void out_of_memory(void);
/* linux/mm/rmap.c */ /* linux/mm/rmap.c */
struct pte_chain; #ifdef CONFIG_MMU
int FASTCALL(page_referenced(struct page *)); int FASTCALL(page_referenced(struct page *));
struct pte_chain *FASTCALL(page_add_rmap(struct page *, pte_t *, struct pte_chain *FASTCALL(page_add_rmap(struct page *, pte_t *,
struct pte_chain *)); struct pte_chain *));
...@@ -186,6 +185,11 @@ int FASTCALL(page_over_rsslimit(struct page *)); ...@@ -186,6 +185,11 @@ int FASTCALL(page_over_rsslimit(struct page *));
/* linux/mm/shmem.c */ /* linux/mm/shmem.c */
extern int shmem_unuse(swp_entry_t entry, struct page *page); extern int shmem_unuse(swp_entry_t entry, struct page *page);
#else
#define page_referenced(page) \
TestClearPageReferenced(page)
#endif /* CONFIG_MMU */
#ifdef CONFIG_SWAP #ifdef CONFIG_SWAP
/* linux/mm/page_io.c */ /* linux/mm/page_io.c */
extern int swap_readpage(struct file *, struct page *); extern int swap_readpage(struct file *, struct page *);
...@@ -242,8 +246,6 @@ extern spinlock_t swaplock; ...@@ -242,8 +246,6 @@ extern spinlock_t swaplock;
page_cache_release(page) page_cache_release(page)
#define free_pages_and_swap_cache(pages, nr) \ #define free_pages_and_swap_cache(pages, nr) \
release_pages((pages), (nr), 0); release_pages((pages), (nr), 0);
#define page_referenced(page) \
TestClearPageReferenced(page)
#define show_swap_cache_info() /*NOTHING*/ #define show_swap_cache_info() /*NOTHING*/
#define free_swap_and_cache(swp) /*NOTHING*/ #define free_swap_and_cache(swp) /*NOTHING*/
......
...@@ -363,5 +363,4 @@ void __init swap_setup(void) ...@@ -363,5 +363,4 @@ void __init swap_setup(void)
* Right now other parts of the system means that we * Right now other parts of the system means that we
* _really_ don't want to cluster much more * _really_ don't want to cluster much more
*/ */
init_MUTEX(&swapper_space.i_shared_sem);
} }
...@@ -44,6 +44,7 @@ struct address_space swapper_space = { ...@@ -44,6 +44,7 @@ struct address_space swapper_space = {
.backing_dev_info = &swap_backing_dev_info, .backing_dev_info = &swap_backing_dev_info,
.i_mmap = LIST_HEAD_INIT(swapper_space.i_mmap), .i_mmap = LIST_HEAD_INIT(swapper_space.i_mmap),
.i_mmap_shared = LIST_HEAD_INIT(swapper_space.i_mmap_shared), .i_mmap_shared = LIST_HEAD_INIT(swapper_space.i_mmap_shared),
.i_shared_sem = __MUTEX_INITIALIZER(swapper_space.i_shared_sem),
.private_lock = SPIN_LOCK_UNLOCKED, .private_lock = SPIN_LOCK_UNLOCKED,
.private_list = LIST_HEAD_INIT(swapper_space.private_list), .private_list = LIST_HEAD_INIT(swapper_space.private_list),
}; };
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment