From: Andrew Morton Date: Fri, 7 Sep 2012 00:23:11 +0000 (+1000) Subject: mm-vmscan-scale-number-of-pages-reclaimed-by-reclaim-compaction-based-on-failures-fix X-Git-Tag: next-20120907~1^2~173 X-Git-Url: https://git.karo-electronics.de/?a=commitdiff_plain;h=e90f5690156711f021ed769e43a07c3f71b6f4ed;p=karo-tx-linux.git mm-vmscan-scale-number-of-pages-reclaimed-by-reclaim-compaction-based-on-failures-fix fix build Cc: Mel Gorman Cc: Minchan Kim Cc: Rik van Riel Signed-off-by: Andrew Morton --- diff --git a/mm/vmscan.c b/mm/vmscan.c index 5228fc7492c2..f6fdae7a8d4c 100644 --- a/mm/vmscan.c +++ b/mm/vmscan.c @@ -1729,6 +1729,28 @@ static bool in_reclaim_compaction(struct scan_control *sc) return false; } +#ifdef CONFIG_COMPACTION +/* + * If compaction is deferred for sc->order then scale the number of pages + * reclaimed based on the number of consecutive allocation failures + */ +static unsigned long scale_for_compaction(unsigned long pages_for_compaction, + struct lruvec *lruvec, struct scan_control *sc) +{ + struct zone *zone = lruvec_zone(lruvec); + + if (zone->compact_order_failed <= sc->order) + pages_for_compaction <<= zone->compact_defer_shift; + return pages_for_compaction; +} +#else +static unsigned long scale_for_compaction(unsigned long pages_for_compaction, + struct lruvec *lruvec, struct scan_control *sc) +{ + return pages_for_compaction; +} +#endif + /* * Reclaim/compaction is used for high-order allocation requests. It reclaims * order-0 pages before compacting the zone. should_continue_reclaim() returns @@ -1743,7 +1765,6 @@ static inline bool should_continue_reclaim(struct lruvec *lruvec, { unsigned long pages_for_compaction; unsigned long inactive_lru_pages; - struct zone *zone; /* If not in reclaim/compaction mode, stop */ if (!in_reclaim_compaction(sc)) @@ -1778,14 +1799,8 @@ static inline bool should_continue_reclaim(struct lruvec *lruvec, */ pages_for_compaction = (2UL << sc->order); - /* - * If compaction is deferred for sc->order then scale the number of - * pages reclaimed based on the number of consecutive allocation - * failures - */ - zone = lruvec_zone(lruvec); - if (zone->compact_order_failed <= sc->order) - pages_for_compaction <<= zone->compact_defer_shift; + pages_for_compaction = scale_for_compaction(pages_for_compaction, + lruvec, sc); inactive_lru_pages = get_lru_size(lruvec, LRU_INACTIVE_FILE); if (nr_swap_pages > 0) inactive_lru_pages += get_lru_size(lruvec, LRU_INACTIVE_ANON);