|
@@ -1862,7 +1862,7 @@ static void get_scan_count(struct lruvec *lruvec, struct scan_control *sc,
|
|
struct zone *zone = lruvec_zone(lruvec);
|
|
struct zone *zone = lruvec_zone(lruvec);
|
|
unsigned long anon_prio, file_prio;
|
|
unsigned long anon_prio, file_prio;
|
|
enum scan_balance scan_balance;
|
|
enum scan_balance scan_balance;
|
|
- unsigned long anon, file, free;
|
|
|
|
|
|
+ unsigned long anon, file;
|
|
bool force_scan = false;
|
|
bool force_scan = false;
|
|
unsigned long ap, fp;
|
|
unsigned long ap, fp;
|
|
enum lru_list lru;
|
|
enum lru_list lru;
|
|
@@ -1915,20 +1915,6 @@ static void get_scan_count(struct lruvec *lruvec, struct scan_control *sc,
|
|
file = get_lru_size(lruvec, LRU_ACTIVE_FILE) +
|
|
file = get_lru_size(lruvec, LRU_ACTIVE_FILE) +
|
|
get_lru_size(lruvec, LRU_INACTIVE_FILE);
|
|
get_lru_size(lruvec, LRU_INACTIVE_FILE);
|
|
|
|
|
|
- /*
|
|
|
|
- * If it's foreseeable that reclaiming the file cache won't be
|
|
|
|
- * enough to get the zone back into a desirable shape, we have
|
|
|
|
- * to swap. Better start now and leave the - probably heavily
|
|
|
|
- * thrashing - remaining file pages alone.
|
|
|
|
- */
|
|
|
|
- if (global_reclaim(sc)) {
|
|
|
|
- free = zone_page_state(zone, NR_FREE_PAGES);
|
|
|
|
- if (unlikely(file + free <= high_wmark_pages(zone))) {
|
|
|
|
- scan_balance = SCAN_ANON;
|
|
|
|
- goto out;
|
|
|
|
- }
|
|
|
|
- }
|
|
|
|
-
|
|
|
|
/*
|
|
/*
|
|
* There is enough inactive page cache, do not reclaim
|
|
* There is enough inactive page cache, do not reclaim
|
|
* anything from the anonymous working set right now.
|
|
* anything from the anonymous working set right now.
|