remove PAGE_GROUP_BY_MOBILITY
[sfrench/cifs-2.6.git] / mm / page_alloc.c
index e1d87ee1d9c6fac46de70ba00fe55980f3685349..8aec4d4601e715bfaa9508c3f8175a61d28a4d2d 100644 (file)
@@ -158,7 +158,6 @@ int nr_node_ids __read_mostly = MAX_NUMNODES;
 EXPORT_SYMBOL(nr_node_ids);
 #endif
 
-#ifdef CONFIG_PAGE_GROUP_BY_MOBILITY
 int page_group_by_mobility_disabled __read_mostly;
 
 static inline int get_pageblock_migratetype(struct page *page)
@@ -192,22 +191,6 @@ static inline int allocflags_to_migratetype(gfp_t gfp_flags, int order)
                ((gfp_flags & __GFP_RECLAIMABLE) != 0);
 }
 
-#else
-static inline int get_pageblock_migratetype(struct page *page)
-{
-       return MIGRATE_UNMOVABLE;
-}
-
-static void set_pageblock_migratetype(struct page *page, int migratetype)
-{
-}
-
-static inline int allocflags_to_migratetype(gfp_t gfp_flags, int order)
-{
-       return MIGRATE_UNMOVABLE;
-}
-#endif /* CONFIG_PAGE_GROUP_BY_MOBILITY */
-
 #ifdef CONFIG_DEBUG_VM
 static int page_outside_zone_boundaries(struct zone *zone, struct page *page)
 {
@@ -687,16 +670,47 @@ static int prep_new_page(struct page *page, int order, gfp_t gfp_flags)
        return 0;
 }
 
-#ifdef CONFIG_PAGE_GROUP_BY_MOBILITY
+/*
+ * Go through the free lists for the given migratetype and remove
+ * the smallest available page from the freelists
+ */
+static struct page *__rmqueue_smallest(struct zone *zone, unsigned int order,
+                                               int migratetype)
+{
+       unsigned int current_order;
+       struct free_area * area;
+       struct page *page;
+
+       /* Find a page of the appropriate size in the preferred list */
+       for (current_order = order; current_order < MAX_ORDER; ++current_order) {
+               area = &(zone->free_area[current_order]);
+               if (list_empty(&area->free_list[migratetype]))
+                       continue;
+
+               page = list_entry(area->free_list[migratetype].next,
+                                                       struct page, lru);
+               list_del(&page->lru);
+               rmv_page_order(page);
+               area->nr_free--;
+               __mod_zone_page_state(zone, NR_FREE_PAGES, - (1UL << order));
+               expand(zone, page, order, current_order, area, migratetype);
+               return page;
+       }
+
+       return NULL;
+}
+
+
 /*
  * This array describes the order lists are fallen back to when
  * the free lists for the desirable migrate type are depleted
  */
 static int fallbacks[MIGRATE_TYPES][MIGRATE_TYPES-1] = {
-       [MIGRATE_UNMOVABLE]   = { MIGRATE_RECLAIMABLE, MIGRATE_MOVABLE,  MIGRATE_HIGHATOMIC },
-       [MIGRATE_RECLAIMABLE] = { MIGRATE_UNMOVABLE,   MIGRATE_MOVABLE,  MIGRATE_HIGHATOMIC },
-       [MIGRATE_MOVABLE]     = { MIGRATE_RECLAIMABLE, MIGRATE_UNMOVABLE,MIGRATE_HIGHATOMIC },
-       [MIGRATE_HIGHATOMIC]  = { MIGRATE_RECLAIMABLE, MIGRATE_UNMOVABLE,MIGRATE_MOVABLE},
+       [MIGRATE_UNMOVABLE]   = { MIGRATE_RECLAIMABLE, MIGRATE_MOVABLE,   MIGRATE_HIGHATOMIC, MIGRATE_RESERVE },
+       [MIGRATE_RECLAIMABLE] = { MIGRATE_UNMOVABLE,   MIGRATE_MOVABLE,   MIGRATE_HIGHATOMIC, MIGRATE_RESERVE },
+       [MIGRATE_MOVABLE]     = { MIGRATE_RECLAIMABLE, MIGRATE_UNMOVABLE, MIGRATE_HIGHATOMIC, MIGRATE_RESERVE },
+       [MIGRATE_HIGHATOMIC]  = { MIGRATE_RECLAIMABLE, MIGRATE_UNMOVABLE, MIGRATE_MOVABLE,    MIGRATE_RESERVE },
+       [MIGRATE_RESERVE]     = { MIGRATE_RESERVE,     MIGRATE_RESERVE,   MIGRATE_RESERVE,    MIGRATE_RESERVE }, /* Never used */
 };
 
 /*
@@ -718,7 +732,7 @@ int move_freepages(struct zone *zone,
         * CONFIG_HOLES_IN_ZONE is set. This bug check is probably redundant
         * anyway as we check zone boundaries in move_freepages_block().
         * Remove at a later date when no bug reports exist related to
-        * CONFIG_PAGE_GROUP_BY_MOBILITY
+        * grouping pages by mobility
         */
        BUG_ON(page_zone(start_page) != page_zone(end_page));
 #endif
@@ -799,6 +813,9 @@ retry:
                for (i = 0; i < MIGRATE_TYPES - 1; i++) {
                        migratetype = fallbacks[start_migratetype][i];
 
+                       /* MIGRATE_RESERVE handled later if necessary */
+                       if (migratetype == MIGRATE_RESERVE)
+                               continue;
                        /*
                         * Make it hard to fallback to blocks used for
                         * high-order atomic allocations
@@ -821,11 +838,23 @@ retry:
 
                        /*
                         * If breaking a large block of pages, move all free
-                        * pages to the preferred allocation list
+                        * pages to the preferred allocation list. If falling
+                        * back for a reclaimable kernel allocation, be more
+                        * agressive about taking ownership of free pages
                         */
-                       if (unlikely(current_order >= MAX_ORDER / 2)) {
+                       if (unlikely(current_order >= MAX_ORDER / 2) ||
+                                       start_migratetype == MIGRATE_RECLAIMABLE) {
+                               unsigned long pages;
+                               pages = move_freepages_block(zone, page,
+                                                               start_migratetype);
+
+                               /* Claim the whole block if over half of it is free */
+                               if ((pages << current_order) >= (1 << (MAX_ORDER-2)) &&
+                                               migratetype != MIGRATE_HIGHATOMIC)
+                                       set_pageblock_migratetype(page,
+                                                               start_migratetype);
+
                                migratetype = start_migratetype;
-                               move_freepages_block(zone, page, migratetype);
                        }
 
                        /* Remove the page from the freelists */
@@ -849,46 +878,23 @@ retry:
                goto retry;
        }
 
-       return NULL;
-}
-#else
-static struct page *__rmqueue_fallback(struct zone *zone, int order,
-                                               int start_migratetype)
-{
-       return NULL;
+       /* Use MIGRATE_RESERVE rather than fail an allocation */
+       return __rmqueue_smallest(zone, order, MIGRATE_RESERVE);
 }
-#endif /* CONFIG_PAGE_GROUP_BY_MOBILITY */
 
-/* 
+/*
  * Do the hard work of removing an element from the buddy allocator.
  * Call me with the zone->lock already held.
  */
 static struct page *__rmqueue(struct zone *zone, unsigned int order,
                                                int migratetype)
 {
-       struct free_area * area;
-       unsigned int current_order;
        struct page *page;
 
-       /* Find a page of the appropriate size in the preferred list */
-       for (current_order = order; current_order < MAX_ORDER; ++current_order) {
-               area = &(zone->free_area[current_order]);
-               if (list_empty(&area->free_list[migratetype]))
-                       continue;
-
-               page = list_entry(area->free_list[migratetype].next,
-                                                       struct page, lru);
-               list_del(&page->lru);
-               rmv_page_order(page);
-               area->nr_free--;
-               __mod_zone_page_state(zone, NR_FREE_PAGES, - (1UL << order));
-               expand(zone, page, order, current_order, area, migratetype);
-               goto got_page;
-       }
-
-       page = __rmqueue_fallback(zone, order, migratetype);
+       page = __rmqueue_smallest(zone, order, migratetype);
 
-got_page:
+       if (unlikely(!page))
+               page = __rmqueue_fallback(zone, order, migratetype);
 
        return page;
 }
@@ -1002,7 +1008,6 @@ void mark_free_pages(struct zone *zone)
 }
 #endif /* CONFIG_PM */
 
-#if defined(CONFIG_HIBERNATION) || defined(CONFIG_PAGE_GROUP_BY_MOBILITY)
 /*
  * Spill all of this CPU's per-cpu pages back into the buddy allocator.
  */
@@ -1033,9 +1038,6 @@ void drain_all_local_pages(void)
 
        smp_call_function(smp_drain_local_pages, NULL, 0, 1);
 }
-#else
-void drain_all_local_pages(void) {}
-#endif /* CONFIG_HIBERNATION || CONFIG_PAGE_GROUP_BY_MOBILITY */
 
 /*
  * Free a 0-order page
@@ -1126,7 +1128,6 @@ again:
                                goto failed;
                }
 
-#ifdef CONFIG_PAGE_GROUP_BY_MOBILITY
                /* Find a page of the appropriate migrate type */
                list_for_each_entry(page, &pcp->list, lru)
                        if (page_private(page) == migratetype)
@@ -1138,9 +1139,6 @@ again:
                                        pcp->batch, &pcp->list, migratetype);
                        page = list_entry(pcp->list.next, struct page, lru);
                }
-#else
-               page = list_entry(pcp->list.next, struct page, lru);
-#endif /* CONFIG_PAGE_GROUP_BY_MOBILITY */
 
                list_del(&page->lru);
                pcp->count--;
@@ -2494,6 +2492,60 @@ static inline unsigned long wait_table_bits(unsigned long size)
 
 #define LONG_ALIGN(x) (((x)+(sizeof(long))-1)&~((sizeof(long))-1))
 
+/*
+ * Mark a number of MAX_ORDER_NR_PAGES blocks as MIGRATE_RESERVE. The number
+ * of blocks reserved is based on zone->pages_min. The memory within the
+ * reserve will tend to store contiguous free pages. Setting min_free_kbytes
+ * higher will lead to a bigger reserve which will get freed as contiguous
+ * blocks as reclaim kicks in
+ */
+static void setup_zone_migrate_reserve(struct zone *zone)
+{
+       unsigned long start_pfn, pfn, end_pfn;
+       struct page *page;
+       unsigned long reserve, block_migratetype;
+
+       /* Get the start pfn, end pfn and the number of blocks to reserve */
+       start_pfn = zone->zone_start_pfn;
+       end_pfn = start_pfn + zone->spanned_pages;
+       reserve = roundup(zone->pages_min, MAX_ORDER_NR_PAGES) >> (MAX_ORDER-1);
+
+       for (pfn = start_pfn; pfn < end_pfn; pfn += MAX_ORDER_NR_PAGES) {
+               if (!pfn_valid(pfn))
+                       continue;
+               page = pfn_to_page(pfn);
+
+               /* Blocks with reserved pages will never free, skip them. */
+               if (PageReserved(page))
+                       continue;
+
+               block_migratetype = get_pageblock_migratetype(page);
+
+               /* If this block is reserved, account for it */
+               if (reserve > 0 && block_migratetype == MIGRATE_RESERVE) {
+                       reserve--;
+                       continue;
+               }
+
+               /* Suitable for reserving if this block is movable */
+               if (reserve > 0 && block_migratetype == MIGRATE_MOVABLE) {
+                       set_pageblock_migratetype(page, MIGRATE_RESERVE);
+                       move_freepages_block(zone, page, MIGRATE_RESERVE);
+                       reserve--;
+                       continue;
+               }
+
+               /*
+                * If the reserve is met and this is a previous reserved block,
+                * take it back
+                */
+               if (block_migratetype == MIGRATE_RESERVE) {
+                       set_pageblock_migratetype(page, MIGRATE_MOVABLE);
+                       move_freepages_block(zone, page, MIGRATE_MOVABLE);
+               }
+       }
+}
+
 /*
  * Initially all pages are reserved - free ones are freed
  * up by free_all_bootmem() once the early boot process is
@@ -2529,9 +2581,12 @@ void __meminit memmap_init_zone(unsigned long size, int nid, unsigned long zone,
                 * movable at startup. This will force kernel allocations
                 * to reserve their blocks rather than leaking throughout
                 * the address space during boot when many long-lived
-                * kernel allocations are made
+                * kernel allocations are made. Later some blocks near
+                * the start are marked MIGRATE_RESERVE by
+                * setup_zone_migrate_reserve()
                 */
-               set_pageblock_migratetype(page, MIGRATE_MOVABLE);
+               if ((pfn & (MAX_ORDER_NR_PAGES-1)))
+                       set_pageblock_migratetype(page, MIGRATE_MOVABLE);
 
                INIT_LIST_HEAD(&page->lru);
 #ifdef WANT_PAGE_VIRTUAL
@@ -4066,6 +4121,7 @@ void setup_per_zone_pages_min(void)
 
                zone->pages_low   = zone->pages_min + (tmp >> 2);
                zone->pages_high  = zone->pages_min + (tmp >> 1);
+               setup_zone_migrate_reserve(zone);
                spin_unlock_irqrestore(&zone->lru_lock, flags);
        }