|
@@ -147,7 +147,7 @@ static bool global_reclaim(struct scan_control *sc)
|
|
|
}
|
|
|
#endif
|
|
|
|
|
|
-unsigned long zone_reclaimable_pages(struct zone *zone)
|
|
|
+static unsigned long zone_reclaimable_pages(struct zone *zone)
|
|
|
{
|
|
|
int nr;
|
|
|
|
|
@@ -3315,27 +3315,6 @@ void wakeup_kswapd(struct zone *zone, int order, enum zone_type classzone_idx)
|
|
|
wake_up_interruptible(&pgdat->kswapd_wait);
|
|
|
}
|
|
|
|
|
|
-/*
|
|
|
- * The reclaimable count would be mostly accurate.
|
|
|
- * The less reclaimable pages may be
|
|
|
- * - mlocked pages, which will be moved to unevictable list when encountered
|
|
|
- * - mapped pages, which may require several travels to be reclaimed
|
|
|
- * - dirty pages, which is not "instantly" reclaimable
|
|
|
- */
|
|
|
-unsigned long global_reclaimable_pages(void)
|
|
|
-{
|
|
|
- int nr;
|
|
|
-
|
|
|
- nr = global_page_state(NR_ACTIVE_FILE) +
|
|
|
- global_page_state(NR_INACTIVE_FILE);
|
|
|
-
|
|
|
- if (get_nr_swap_pages() > 0)
|
|
|
- nr += global_page_state(NR_ACTIVE_ANON) +
|
|
|
- global_page_state(NR_INACTIVE_ANON);
|
|
|
-
|
|
|
- return nr;
|
|
|
-}
|
|
|
-
|
|
|
#ifdef CONFIG_HIBERNATION
|
|
|
/*
|
|
|
* Try to free `nr_to_reclaim' of memory, system-wide, and return the number of
|