]> git.karo-electronics.de Git - karo-tx-linux.git/commitdiff
cma: count free CMA pages
authorBartlomiej Zolnierkiewicz <b.zolnierkie@samsung.com>
Fri, 28 Sep 2012 00:19:38 +0000 (10:19 +1000)
committerStephen Rothwell <sfr@canb.auug.org.au>
Fri, 5 Oct 2012 04:01:01 +0000 (14:01 +1000)
Add NR_FREE_CMA_PAGES counter to be later used for checking watermark in
__zone_watermark_ok().  For simplicity and to avoid #ifdef hell make this
counter always available (not only when CONFIG_CMA=y).

Signed-off-by: Bartlomiej Zolnierkiewicz <b.zolnierkie@samsung.com>
Signed-off-by: Kyungmin Park <kyungmin.park@samsung.com>
Cc: Marek Szyprowski <m.szyprowski@samsung.com>
Cc: Michal Nazarewicz <mina86@mina86.com>
Cc: Minchan Kim <minchan@kernel.org>
Cc: Mel Gorman <mgorman@suse.de>
Cc: Hugh Dickins <hughd@google.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
include/linux/mmzone.h
include/linux/vmstat.h
mm/page_alloc.c
mm/page_isolation.c
mm/vmstat.c

index 36405f1819544e5a2a2cdd9f5bf12c2ba1ccf190..9705f2d9dff096a2db4a74c4271830fc94ad39aa 100644 (file)
@@ -141,6 +141,7 @@ enum zone_stat_item {
        NUMA_OTHER,             /* allocation from other node */
 #endif
        NR_ANON_TRANSPARENT_HUGEPAGES,
+       NR_FREE_CMA_PAGES,
        NR_VM_ZONE_STAT_ITEMS };
 
 /*
index ad2cfd53dadce54b28ff52adb73f923c6dde2f55..a5bb15018b5c4f70d8cfcdc1dad04c6517df393a 100644 (file)
@@ -253,6 +253,14 @@ static inline void refresh_zone_stat_thresholds(void) { }
 
 #endif         /* CONFIG_SMP */
 
+static inline void __mod_zone_freepage_state(struct zone *zone, int nr_pages,
+                                            int migratetype)
+{
+       __mod_zone_page_state(zone, NR_FREE_PAGES, nr_pages);
+       if (is_migrate_cma(migratetype))
+               __mod_zone_page_state(zone, NR_FREE_CMA_PAGES, nr_pages);
+}
+
 extern const char * const vmstat_text[];
 
 #endif /* _LINUX_VMSTAT_H */
index 2e403419a30c7b73a14ab90ba0022af906b9d475..36c3460f7c5b33e708fa17fd231dea8ed5106f6d 100644 (file)
@@ -557,7 +557,8 @@ static inline void __free_one_page(struct page *page,
                if (page_is_guard(buddy)) {
                        clear_page_guard_flag(buddy);
                        set_page_private(page, 0);
-                       __mod_zone_page_state(zone, NR_FREE_PAGES, 1 << order);
+                       __mod_zone_freepage_state(zone, 1 << order,
+                                                 migratetype);
                } else {
                        list_del(&buddy->lru);
                        zone->free_area[order].nr_free--;
@@ -676,6 +677,8 @@ static void free_pcppages_bulk(struct zone *zone, int count,
                        /* MIGRATE_MOVABLE list may include MIGRATE_RESERVEs */
                        __free_one_page(page, zone, 0, mt);
                        trace_mm_page_pcpu_drain(page, 0, mt);
+                       if (is_migrate_cma(mt))
+                               __mod_zone_page_state(zone, NR_FREE_CMA_PAGES, 1);
                } while (--to_free && --batch_free && !list_empty(list));
        }
        __mod_zone_page_state(zone, NR_FREE_PAGES, count);
@@ -691,7 +694,7 @@ static void free_one_page(struct zone *zone, struct page *page, int order,
 
        __free_one_page(page, zone, order, migratetype);
        if (unlikely(migratetype != MIGRATE_ISOLATE))
-               __mod_zone_page_state(zone, NR_FREE_PAGES, 1 << order);
+               __mod_zone_freepage_state(zone, 1 << order, migratetype);
        spin_unlock(&zone->lock);
 }
 
@@ -814,7 +817,8 @@ static inline void expand(struct zone *zone, struct page *page,
                        set_page_guard_flag(&page[size]);
                        set_page_private(&page[size], high);
                        /* Guard pages are not available for any usage */
-                       __mod_zone_page_state(zone, NR_FREE_PAGES, -(1 << high));
+                       __mod_zone_freepage_state(zone, -(1 << high),
+                                                 migratetype);
                        continue;
                }
 #endif
@@ -1140,6 +1144,9 @@ static int rmqueue_bulk(struct zone *zone, unsigned int order,
                }
                set_page_private(page, mt);
                list = &page->lru;
+               if (is_migrate_cma(mt))
+                       __mod_zone_page_state(zone, NR_FREE_CMA_PAGES,
+                                             -(1 << order));
        }
        __mod_zone_page_state(zone, NR_FREE_PAGES, -(i << order));
        spin_unlock(&zone->lock);
@@ -1411,7 +1418,7 @@ int capture_free_page(struct page *page, int alloc_order, int migratetype)
 
        mt = get_pageblock_migratetype(page);
        if (unlikely(mt != MIGRATE_ISOLATE))
-               __mod_zone_page_state(zone, NR_FREE_PAGES, -(1UL << order));
+               __mod_zone_freepage_state(zone, -(1UL << order), mt);
 
        if (alloc_order != order)
                expand(zone, page, alloc_order, order,
@@ -1515,7 +1522,8 @@ again:
                spin_unlock(&zone->lock);
                if (!page)
                        goto failed;
-               __mod_zone_page_state(zone, NR_FREE_PAGES, -(1 << order));
+               __mod_zone_freepage_state(zone, -(1 << order),
+                                         get_pageblock_migratetype(page));
        }
 
        __count_zone_vm_events(PGALLOC, zone, 1 << order);
@@ -2889,7 +2897,8 @@ void show_free_areas(unsigned int filter)
                " unevictable:%lu"
                " dirty:%lu writeback:%lu unstable:%lu\n"
                " free:%lu slab_reclaimable:%lu slab_unreclaimable:%lu\n"
-               " mapped:%lu shmem:%lu pagetables:%lu bounce:%lu\n",
+               " mapped:%lu shmem:%lu pagetables:%lu bounce:%lu\n"
+               " free_cma:%lu\n",
                global_page_state(NR_ACTIVE_ANON),
                global_page_state(NR_INACTIVE_ANON),
                global_page_state(NR_ISOLATED_ANON),
@@ -2906,7 +2915,8 @@ void show_free_areas(unsigned int filter)
                global_page_state(NR_FILE_MAPPED),
                global_page_state(NR_SHMEM),
                global_page_state(NR_PAGETABLE),
-               global_page_state(NR_BOUNCE));
+               global_page_state(NR_BOUNCE),
+               global_page_state(NR_FREE_CMA_PAGES));
 
        for_each_populated_zone(zone) {
                int i;
@@ -2938,6 +2948,7 @@ void show_free_areas(unsigned int filter)
                        " pagetables:%lukB"
                        " unstable:%lukB"
                        " bounce:%lukB"
+                       " free_cma:%lukB"
                        " writeback_tmp:%lukB"
                        " pages_scanned:%lu"
                        " all_unreclaimable? %s"
@@ -2967,6 +2978,7 @@ void show_free_areas(unsigned int filter)
                        K(zone_page_state(zone, NR_PAGETABLE)),
                        K(zone_page_state(zone, NR_UNSTABLE_NFS)),
                        K(zone_page_state(zone, NR_BOUNCE)),
+                       K(zone_page_state(zone, NR_FREE_CMA_PAGES)),
                        K(zone_page_state(zone, NR_WRITEBACK_TEMP)),
                        zone->pages_scanned,
                        (zone->all_unreclaimable ? "yes" : "no")
index 3ca1716471bc5af48fef3560863b2c9e6bd15d60..bce97c93451f363aa55fb6023519911a15f6f57c 100644 (file)
@@ -77,11 +77,12 @@ int set_migratetype_isolate(struct page *page)
 out:
        if (!ret) {
                unsigned long nr_pages;
+               int mt = get_pageblock_migratetype(page);
 
                set_pageblock_isolate(page);
                nr_pages = move_freepages_block(zone, page, MIGRATE_ISOLATE);
 
-               __mod_zone_page_state(zone, NR_FREE_PAGES, -nr_pages);
+               __mod_zone_freepage_state(zone, -nr_pages, mt);
        }
 
        spin_unlock_irqrestore(&zone->lock, flags);
@@ -100,7 +101,7 @@ void unset_migratetype_isolate(struct page *page, unsigned migratetype)
        if (get_pageblock_migratetype(page) != MIGRATE_ISOLATE)
                goto out;
        nr_pages = move_freepages_block(zone, page, migratetype);
-       __mod_zone_page_state(zone, NR_FREE_PAGES, nr_pages);
+       __mod_zone_freepage_state(zone, nr_pages, migratetype);
        restore_pageblock_isolate(page, migratetype);
 out:
        spin_unlock_irqrestore(&zone->lock, flags);
index 3b0924892c14c3d489d5d3d0132db62dc69879ce..660d427c05b38ac5165ce95e61ee386db4f612bb 100644 (file)
@@ -721,6 +721,7 @@ const char * const vmstat_text[] = {
        "numa_other",
 #endif
        "nr_anon_transparent_hugepages",
+       "nr_free_cma",
        "nr_dirty_threshold",
        "nr_dirty_background_threshold",