Donate to e Foundation | Murena handsets with /e/OS | Own a part of Murena! Learn more

Commit 44c241f1 authored by KOSAKI Motohiro's avatar KOSAKI Motohiro Committed by Linus Torvalds
Browse files

mm: rename pgmoved variable in shrink_active_list()



Currently the pgmoved variable has two meanings.  It causes harder
reviewing.  This patch separates it.

Signed-off-by: default avatarKOSAKI Motohiro <kosaki.motohiro@jp.fujitsu.com>
Reviewed-by: default avatarRik van Riel <riel@redhat.com>
Reviewed-by: default avatarMinchan Kim <minchan.kim@gmail.com>
Signed-off-by: default avatarAndrew Morton <akpm@linux-foundation.org>
Signed-off-by: default avatarLinus Torvalds <torvalds@linux-foundation.org>
parent b259fbde
Loading
Loading
Loading
Loading
+8 −8
Original line number Original line Diff line number Diff line
@@ -1244,7 +1244,7 @@ static void move_active_pages_to_lru(struct zone *zone,
static void shrink_active_list(unsigned long nr_pages, struct zone *zone,
static void shrink_active_list(unsigned long nr_pages, struct zone *zone,
			struct scan_control *sc, int priority, int file)
			struct scan_control *sc, int priority, int file)
{
{
	unsigned long pgmoved;
	unsigned long nr_taken;
	unsigned long pgscanned;
	unsigned long pgscanned;
	unsigned long vm_flags;
	unsigned long vm_flags;
	LIST_HEAD(l_hold);	/* The pages which were snipped off */
	LIST_HEAD(l_hold);	/* The pages which were snipped off */
@@ -1252,10 +1252,11 @@ static void shrink_active_list(unsigned long nr_pages, struct zone *zone,
	LIST_HEAD(l_inactive);
	LIST_HEAD(l_inactive);
	struct page *page;
	struct page *page;
	struct zone_reclaim_stat *reclaim_stat = get_reclaim_stat(zone, sc);
	struct zone_reclaim_stat *reclaim_stat = get_reclaim_stat(zone, sc);
	unsigned long nr_rotated = 0;


	lru_add_drain();
	lru_add_drain();
	spin_lock_irq(&zone->lru_lock);
	spin_lock_irq(&zone->lru_lock);
	pgmoved = sc->isolate_pages(nr_pages, &l_hold, &pgscanned, sc->order,
	nr_taken = sc->isolate_pages(nr_pages, &l_hold, &pgscanned, sc->order,
					ISOLATE_ACTIVE, zone,
					ISOLATE_ACTIVE, zone,
					sc->mem_cgroup, 1, file);
					sc->mem_cgroup, 1, file);
	/*
	/*
@@ -1265,16 +1266,15 @@ static void shrink_active_list(unsigned long nr_pages, struct zone *zone,
	if (scanning_global_lru(sc)) {
	if (scanning_global_lru(sc)) {
		zone->pages_scanned += pgscanned;
		zone->pages_scanned += pgscanned;
	}
	}
	reclaim_stat->recent_scanned[!!file] += pgmoved;
	reclaim_stat->recent_scanned[!!file] += nr_taken;


	__count_zone_vm_events(PGREFILL, zone, pgscanned);
	__count_zone_vm_events(PGREFILL, zone, pgscanned);
	if (file)
	if (file)
		__mod_zone_page_state(zone, NR_ACTIVE_FILE, -pgmoved);
		__mod_zone_page_state(zone, NR_ACTIVE_FILE, -nr_taken);
	else
	else
		__mod_zone_page_state(zone, NR_ACTIVE_ANON, -pgmoved);
		__mod_zone_page_state(zone, NR_ACTIVE_ANON, -nr_taken);
	spin_unlock_irq(&zone->lru_lock);
	spin_unlock_irq(&zone->lru_lock);


	pgmoved = 0;  /* count referenced (mapping) mapped pages */
	while (!list_empty(&l_hold)) {
	while (!list_empty(&l_hold)) {
		cond_resched();
		cond_resched();
		page = lru_to_page(&l_hold);
		page = lru_to_page(&l_hold);
@@ -1288,7 +1288,7 @@ static void shrink_active_list(unsigned long nr_pages, struct zone *zone,
		/* page_referenced clears PageReferenced */
		/* page_referenced clears PageReferenced */
		if (page_mapping_inuse(page) &&
		if (page_mapping_inuse(page) &&
		    page_referenced(page, 0, sc->mem_cgroup, &vm_flags)) {
		    page_referenced(page, 0, sc->mem_cgroup, &vm_flags)) {
			pgmoved++;
			nr_rotated++;
			/*
			/*
			 * Identify referenced, file-backed active pages and
			 * Identify referenced, file-backed active pages and
			 * give them one more trip around the active list. So
			 * give them one more trip around the active list. So
@@ -1317,7 +1317,7 @@ static void shrink_active_list(unsigned long nr_pages, struct zone *zone,
	 * helps balance scan pressure between file and anonymous pages in
	 * helps balance scan pressure between file and anonymous pages in
	 * get_scan_ratio.
	 * get_scan_ratio.
	 */
	 */
	reclaim_stat->recent_rotated[!!file] += pgmoved;
	reclaim_stat->recent_rotated[!!file] += nr_rotated;


	move_active_pages_to_lru(zone, &l_active,
	move_active_pages_to_lru(zone, &l_active,
						LRU_ACTIVE + file * LRU_FILE);
						LRU_ACTIVE + file * LRU_FILE);