Commit e72e2bd6 authored by KAMEZAWA Hiroyuki's avatar KAMEZAWA Hiroyuki Committed by Linus Torvalds

memcg: rename scan global lru

Rename scan_global_lru() to scanning_global_lru().
Signed-off-by: default avatarKAMEZAWA Hiroyuki <kamezawa.hiroyu@jp.fujitsu.com>
Cc: Balbir Singh <balbir@in.ibm.com>
Cc: Daisuke Nishimura <nishimura@mxp.nes.nec.co.jp>
Cc: Hugh Dickins <hugh@veritas.com>
Cc: KOSAKI Motohiro <kosaki.motohiro@jp.fujitsu.com>
Signed-off-by: default avatarAndrew Morton <akpm@linux-foundation.org>
Signed-off-by: default avatarLinus Torvalds <torvalds@linux-foundation.org>
parent 7f016ee8
...@@ -125,15 +125,15 @@ static LIST_HEAD(shrinker_list); ...@@ -125,15 +125,15 @@ static LIST_HEAD(shrinker_list);
static DECLARE_RWSEM(shrinker_rwsem); static DECLARE_RWSEM(shrinker_rwsem);
#ifdef CONFIG_CGROUP_MEM_RES_CTLR #ifdef CONFIG_CGROUP_MEM_RES_CTLR
#define scan_global_lru(sc) (!(sc)->mem_cgroup) #define scanning_global_lru(sc) (!(sc)->mem_cgroup)
#else #else
#define scan_global_lru(sc) (1) #define scanning_global_lru(sc) (1)
#endif #endif
static struct zone_reclaim_stat *get_reclaim_stat(struct zone *zone, static struct zone_reclaim_stat *get_reclaim_stat(struct zone *zone,
struct scan_control *sc) struct scan_control *sc)
{ {
if (!scan_global_lru(sc)) if (!scanning_global_lru(sc))
return mem_cgroup_get_reclaim_stat(sc->mem_cgroup, zone); return mem_cgroup_get_reclaim_stat(sc->mem_cgroup, zone);
return &zone->reclaim_stat; return &zone->reclaim_stat;
...@@ -142,7 +142,7 @@ static struct zone_reclaim_stat *get_reclaim_stat(struct zone *zone, ...@@ -142,7 +142,7 @@ static struct zone_reclaim_stat *get_reclaim_stat(struct zone *zone,
static unsigned long zone_nr_pages(struct zone *zone, struct scan_control *sc, static unsigned long zone_nr_pages(struct zone *zone, struct scan_control *sc,
enum lru_list lru) enum lru_list lru)
{ {
if (!scan_global_lru(sc)) if (!scanning_global_lru(sc))
return mem_cgroup_zone_nr_pages(sc->mem_cgroup, zone, lru); return mem_cgroup_zone_nr_pages(sc->mem_cgroup, zone, lru);
return zone_page_state(zone, NR_LRU_BASE + lru); return zone_page_state(zone, NR_LRU_BASE + lru);
...@@ -1090,7 +1090,7 @@ static unsigned long shrink_inactive_list(unsigned long max_scan, ...@@ -1090,7 +1090,7 @@ static unsigned long shrink_inactive_list(unsigned long max_scan,
__mod_zone_page_state(zone, NR_INACTIVE_ANON, __mod_zone_page_state(zone, NR_INACTIVE_ANON,
-count[LRU_INACTIVE_ANON]); -count[LRU_INACTIVE_ANON]);
if (scan_global_lru(sc)) if (scanning_global_lru(sc))
zone->pages_scanned += nr_scan; zone->pages_scanned += nr_scan;
reclaim_stat->recent_scanned[0] += count[LRU_INACTIVE_ANON]; reclaim_stat->recent_scanned[0] += count[LRU_INACTIVE_ANON];
...@@ -1129,7 +1129,7 @@ static unsigned long shrink_inactive_list(unsigned long max_scan, ...@@ -1129,7 +1129,7 @@ static unsigned long shrink_inactive_list(unsigned long max_scan,
if (current_is_kswapd()) { if (current_is_kswapd()) {
__count_zone_vm_events(PGSCAN_KSWAPD, zone, nr_scan); __count_zone_vm_events(PGSCAN_KSWAPD, zone, nr_scan);
__count_vm_events(KSWAPD_STEAL, nr_freed); __count_vm_events(KSWAPD_STEAL, nr_freed);
} else if (scan_global_lru(sc)) } else if (scanning_global_lru(sc))
__count_zone_vm_events(PGSCAN_DIRECT, zone, nr_scan); __count_zone_vm_events(PGSCAN_DIRECT, zone, nr_scan);
__count_zone_vm_events(PGSTEAL, zone, nr_freed); __count_zone_vm_events(PGSTEAL, zone, nr_freed);
...@@ -1228,7 +1228,7 @@ static void shrink_active_list(unsigned long nr_pages, struct zone *zone, ...@@ -1228,7 +1228,7 @@ static void shrink_active_list(unsigned long nr_pages, struct zone *zone,
* zone->pages_scanned is used for detect zone's oom * zone->pages_scanned is used for detect zone's oom
* mem_cgroup remembers nr_scan by itself. * mem_cgroup remembers nr_scan by itself.
*/ */
if (scan_global_lru(sc)) { if (scanning_global_lru(sc)) {
zone->pages_scanned += pgscanned; zone->pages_scanned += pgscanned;
} }
reclaim_stat->recent_scanned[!!file] += pgmoved; reclaim_stat->recent_scanned[!!file] += pgmoved;
...@@ -1337,7 +1337,7 @@ static int inactive_anon_is_low(struct zone *zone, struct scan_control *sc) ...@@ -1337,7 +1337,7 @@ static int inactive_anon_is_low(struct zone *zone, struct scan_control *sc)
{ {
int low; int low;
if (scan_global_lru(sc)) if (scanning_global_lru(sc))
low = inactive_anon_is_low_global(zone); low = inactive_anon_is_low_global(zone);
else else
low = mem_cgroup_inactive_anon_is_low(sc->mem_cgroup, zone); low = mem_cgroup_inactive_anon_is_low(sc->mem_cgroup, zone);
...@@ -1390,7 +1390,7 @@ static void get_scan_ratio(struct zone *zone, struct scan_control *sc, ...@@ -1390,7 +1390,7 @@ static void get_scan_ratio(struct zone *zone, struct scan_control *sc,
file = zone_nr_pages(zone, sc, LRU_ACTIVE_FILE) + file = zone_nr_pages(zone, sc, LRU_ACTIVE_FILE) +
zone_nr_pages(zone, sc, LRU_INACTIVE_FILE); zone_nr_pages(zone, sc, LRU_INACTIVE_FILE);
if (scan_global_lru(sc)) { if (scanning_global_lru(sc)) {
free = zone_page_state(zone, NR_FREE_PAGES); free = zone_page_state(zone, NR_FREE_PAGES);
/* If we have very few page cache pages, /* If we have very few page cache pages,
force-scan anon pages. */ force-scan anon pages. */
...@@ -1474,7 +1474,7 @@ static void shrink_zone(int priority, struct zone *zone, ...@@ -1474,7 +1474,7 @@ static void shrink_zone(int priority, struct zone *zone,
scan >>= priority; scan >>= priority;
scan = (scan * percent[file]) / 100; scan = (scan * percent[file]) / 100;
} }
if (scan_global_lru(sc)) { if (scanning_global_lru(sc)) {
zone->lru[l].nr_scan += scan; zone->lru[l].nr_scan += scan;
nr[l] = zone->lru[l].nr_scan; nr[l] = zone->lru[l].nr_scan;
if (nr[l] >= swap_cluster_max) if (nr[l] >= swap_cluster_max)
...@@ -1550,7 +1550,7 @@ static void shrink_zones(int priority, struct zonelist *zonelist, ...@@ -1550,7 +1550,7 @@ static void shrink_zones(int priority, struct zonelist *zonelist,
* Take care memory controller reclaiming has small influence * Take care memory controller reclaiming has small influence
* to global LRU. * to global LRU.
*/ */
if (scan_global_lru(sc)) { if (scanning_global_lru(sc)) {
if (!cpuset_zone_allowed_hardwall(zone, GFP_KERNEL)) if (!cpuset_zone_allowed_hardwall(zone, GFP_KERNEL))
continue; continue;
note_zone_scanning_priority(zone, priority); note_zone_scanning_priority(zone, priority);
...@@ -1603,12 +1603,12 @@ static unsigned long do_try_to_free_pages(struct zonelist *zonelist, ...@@ -1603,12 +1603,12 @@ static unsigned long do_try_to_free_pages(struct zonelist *zonelist,
delayacct_freepages_start(); delayacct_freepages_start();
if (scan_global_lru(sc)) if (scanning_global_lru(sc))
count_vm_event(ALLOCSTALL); count_vm_event(ALLOCSTALL);
/* /*
* mem_cgroup will not do shrink_slab. * mem_cgroup will not do shrink_slab.
*/ */
if (scan_global_lru(sc)) { if (scanning_global_lru(sc)) {
for_each_zone_zonelist(zone, z, zonelist, high_zoneidx) { for_each_zone_zonelist(zone, z, zonelist, high_zoneidx) {
if (!cpuset_zone_allowed_hardwall(zone, GFP_KERNEL)) if (!cpuset_zone_allowed_hardwall(zone, GFP_KERNEL))
...@@ -1627,7 +1627,7 @@ static unsigned long do_try_to_free_pages(struct zonelist *zonelist, ...@@ -1627,7 +1627,7 @@ static unsigned long do_try_to_free_pages(struct zonelist *zonelist,
* Don't shrink slabs when reclaiming memory from * Don't shrink slabs when reclaiming memory from
* over limit cgroups * over limit cgroups
*/ */
if (scan_global_lru(sc)) { if (scanning_global_lru(sc)) {
shrink_slab(sc->nr_scanned, sc->gfp_mask, lru_pages); shrink_slab(sc->nr_scanned, sc->gfp_mask, lru_pages);
if (reclaim_state) { if (reclaim_state) {
sc->nr_reclaimed += reclaim_state->reclaimed_slab; sc->nr_reclaimed += reclaim_state->reclaimed_slab;
...@@ -1658,7 +1658,7 @@ static unsigned long do_try_to_free_pages(struct zonelist *zonelist, ...@@ -1658,7 +1658,7 @@ static unsigned long do_try_to_free_pages(struct zonelist *zonelist,
congestion_wait(WRITE, HZ/10); congestion_wait(WRITE, HZ/10);
} }
/* top priority shrink_zones still had more to do? don't OOM, then */ /* top priority shrink_zones still had more to do? don't OOM, then */
if (!sc->all_unreclaimable && scan_global_lru(sc)) if (!sc->all_unreclaimable && scanning_global_lru(sc))
ret = sc->nr_reclaimed; ret = sc->nr_reclaimed;
out: out:
/* /*
...@@ -1671,7 +1671,7 @@ static unsigned long do_try_to_free_pages(struct zonelist *zonelist, ...@@ -1671,7 +1671,7 @@ static unsigned long do_try_to_free_pages(struct zonelist *zonelist,
if (priority < 0) if (priority < 0)
priority = 0; priority = 0;
if (scan_global_lru(sc)) { if (scanning_global_lru(sc)) {
for_each_zone_zonelist(zone, z, zonelist, high_zoneidx) { for_each_zone_zonelist(zone, z, zonelist, high_zoneidx) {
if (!cpuset_zone_allowed_hardwall(zone, GFP_KERNEL)) if (!cpuset_zone_allowed_hardwall(zone, GFP_KERNEL))
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment