mm: fix boundary checking in free_bootmem_core
[sfrench/cifs-2.6.git] / mm / bootmem.c
index 50353e0dac12dc41441b64c78b59c852140f76c4..2ccea700968f7ecaaa265023e79e52d540a82871 100644 (file)
@@ -8,17 +8,15 @@
  *  free memory collector. It's used to deal with reserved
  *  system memory and memory holes as well.
  */
-
-#include <linux/mm.h>
-#include <linux/kernel_stat.h>
-#include <linux/swap.h>
-#include <linux/interrupt.h>
 #include <linux/init.h>
+#include <linux/pfn.h>
 #include <linux/bootmem.h>
-#include <linux/mmzone.h>
 #include <linux/module.h>
-#include <asm/dma.h>
+
+#include <asm/bug.h>
 #include <asm/io.h>
+#include <asm/processor.h>
+
 #include "internal.h"
 
 /*
@@ -29,8 +27,6 @@ unsigned long max_low_pfn;
 unsigned long min_low_pfn;
 unsigned long max_pfn;
 
-EXPORT_UNUSED_SYMBOL(max_pfn);  /*  June 2006  */
-
 static LIST_HEAD(bdata_list);
 #ifdef CONFIG_CRASH_DUMP
 /*
@@ -41,7 +37,7 @@ unsigned long saved_max_pfn;
 #endif
 
 /* return the number of _pages_ that will be allocated for the boot bitmap */
-unsigned long __init bootmem_bootmap_pages (unsigned long pages)
+unsigned long __init bootmem_bootmap_pages(unsigned long pages)
 {
        unsigned long mapsize;
 
@@ -51,12 +47,14 @@ unsigned long __init bootmem_bootmap_pages (unsigned long pages)
 
        return mapsize;
 }
+
 /*
  * link bdata in order
  */
-static void link_bootmem(bootmem_data_t *bdata)
+static void __init link_bootmem(bootmem_data_t *bdata)
 {
        bootmem_data_t *ent;
+
        if (list_empty(&bdata_list)) {
                list_add(&bdata->list, &bdata_list);
                return;
@@ -69,22 +67,32 @@ static void link_bootmem(bootmem_data_t *bdata)
                }
        }
        list_add_tail(&bdata->list, &bdata_list);
-       return;
 }
 
+/*
+ * Given an initialised bdata, it returns the size of the boot bitmap
+ */
+static unsigned long __init get_mapsize(bootmem_data_t *bdata)
+{
+       unsigned long mapsize;
+       unsigned long start = PFN_DOWN(bdata->node_boot_start);
+       unsigned long end = bdata->node_low_pfn;
+
+       mapsize = ((end - start) + 7) / 8;
+       return ALIGN(mapsize, sizeof(long));
+}
 
 /*
  * Called once to set up the allocator itself.
  */
-static unsigned long __init init_bootmem_core (pg_data_t *pgdat,
+static unsigned long __init init_bootmem_core(pg_data_t *pgdat,
        unsigned long mapstart, unsigned long start, unsigned long end)
 {
        bootmem_data_t *bdata = pgdat->bdata;
-       unsigned long mapsize = ((end - start)+7)/8;
+       unsigned long mapsize;
 
-       mapsize = ALIGN(mapsize, sizeof(long));
-       bdata->node_bootmem_map = phys_to_virt(mapstart << PAGE_SHIFT);
-       bdata->node_boot_start = (start << PAGE_SHIFT);
+       bdata->node_bootmem_map = phys_to_virt(PFN_PHYS(mapstart));
+       bdata->node_boot_start = PFN_PHYS(start);
        bdata->node_low_pfn = end;
        link_bootmem(bdata);
 
@@ -92,6 +100,7 @@ static unsigned long __init init_bootmem_core (pg_data_t *pgdat,
         * Initially all pages are reserved - setup_arch() has to
         * register free RAM areas explicitly.
         */
+       mapsize = get_mapsize(bdata);
        memset(bdata->node_bootmem_map, 0xff, mapsize);
 
        return mapsize;
@@ -102,54 +111,77 @@ static unsigned long __init init_bootmem_core (pg_data_t *pgdat,
  * might be used for boot-time allocations - or it might get added
  * to the free page pool later on.
  */
-static void __init reserve_bootmem_core(bootmem_data_t *bdata, unsigned long addr, unsigned long size)
+static int __init reserve_bootmem_core(bootmem_data_t *bdata,
+                       unsigned long addr, unsigned long size, int flags)
 {
+       unsigned long sidx, eidx;
        unsigned long i;
+       int ret;
+
        /*
         * round up, partially reserved pages are considered
         * fully reserved.
         */
-       unsigned long sidx = (addr - bdata->node_boot_start)/PAGE_SIZE;
-       unsigned long eidx = (addr + size - bdata->node_boot_start + 
-                                                       PAGE_SIZE-1)/PAGE_SIZE;
-       unsigned long end = (addr + size + PAGE_SIZE-1)/PAGE_SIZE;
-
        BUG_ON(!size);
-       BUG_ON(sidx >= eidx);
-       BUG_ON((addr >> PAGE_SHIFT) >= bdata->node_low_pfn);
-       BUG_ON(end > bdata->node_low_pfn);
+       BUG_ON(PFN_DOWN(addr) >= bdata->node_low_pfn);
+       BUG_ON(PFN_UP(addr + size) > bdata->node_low_pfn);
+       BUG_ON(addr < bdata->node_boot_start);
+
+       sidx = PFN_DOWN(addr - bdata->node_boot_start);
+       eidx = PFN_UP(addr + size - bdata->node_boot_start);
 
        for (i = sidx; i < eidx; i++)
                if (test_and_set_bit(i, bdata->node_bootmem_map)) {
 #ifdef CONFIG_DEBUG_BOOTMEM
                        printk("hm, page %08lx reserved twice.\n", i*PAGE_SIZE);
 #endif
+                       if (flags & BOOTMEM_EXCLUSIVE) {
+                               ret = -EBUSY;
+                               goto err;
+                       }
                }
+
+       return 0;
+
+err:
+       /* unreserve memory we accidentally reserved */
+       for (i--; i >= sidx; i--)
+               clear_bit(i, bdata->node_bootmem_map);
+
+       return ret;
 }
 
-static void __init free_bootmem_core(bootmem_data_t *bdata, unsigned long addr, unsigned long size)
+static void __init free_bootmem_core(bootmem_data_t *bdata, unsigned long addr,
+                                    unsigned long size)
 {
+       unsigned long sidx, eidx;
        unsigned long i;
-       unsigned long start;
+
+       BUG_ON(!size);
+
+       /* out range */
+       if (addr + size < bdata->node_boot_start ||
+               PFN_DOWN(addr) > bdata->node_low_pfn)
+               return;
        /*
         * round down end of usable mem, partially free pages are
         * considered reserved.
         */
-       unsigned long sidx;
-       unsigned long eidx = (addr + size - bdata->node_boot_start)/PAGE_SIZE;
-       unsigned long end = (addr + size)/PAGE_SIZE;
-
-       BUG_ON(!size);
-       BUG_ON(end > bdata->node_low_pfn);
 
-       if (addr < bdata->last_success)
+       if (addr >= bdata->node_boot_start && addr < bdata->last_success)
                bdata->last_success = addr;
 
        /*
-        * Round up the beginning of the address.
+        * Round up to index to the range.
         */
-       start = (addr + PAGE_SIZE-1) / PAGE_SIZE;
-       sidx = start - (bdata->node_boot_start/PAGE_SIZE);
+       if (PFN_UP(addr) > PFN_DOWN(bdata->node_boot_start))
+               sidx = PFN_UP(addr) - PFN_DOWN(bdata->node_boot_start);
+       else
+               sidx = 0;
+
+       eidx = PFN_DOWN(addr + size - bdata->node_boot_start);
+       if (eidx > bdata->node_low_pfn - PFN_DOWN(bdata->node_boot_start))
+               eidx = bdata->node_low_pfn - PFN_DOWN(bdata->node_boot_start);
 
        for (i = sidx; i < eidx; i++) {
                if (unlikely(!test_and_clear_bit(i, bdata->node_bootmem_map)))
@@ -175,10 +207,10 @@ __alloc_bootmem_core(struct bootmem_data *bdata, unsigned long size,
              unsigned long align, unsigned long goal, unsigned long limit)
 {
        unsigned long offset, remaining_size, areasize, preferred;
-       unsigned long i, start = 0, incr, eidx, end_pfn = bdata->node_low_pfn;
+       unsigned long i, start = 0, incr, eidx, end_pfn;
        void *ret;
 
-       if(!size) {
+       if (!size) {
                printk("__alloc_bootmem_core(): zero-sized request\n");
                BUG();
        }
@@ -187,23 +219,26 @@ __alloc_bootmem_core(struct bootmem_data *bdata, unsigned long size,
        if (limit && bdata->node_boot_start >= limit)
                return NULL;
 
-        limit >>=PAGE_SHIFT;
+       /* on nodes without memory - bootmem_map is NULL */
+       if (!bdata->node_bootmem_map)
+               return NULL;
+
+       end_pfn = bdata->node_low_pfn;
+       limit = PFN_DOWN(limit);
        if (limit && end_pfn > limit)
                end_pfn = limit;
 
-       eidx = end_pfn - (bdata->node_boot_start >> PAGE_SHIFT);
+       eidx = end_pfn - PFN_DOWN(bdata->node_boot_start);
        offset = 0;
-       if (align &&
-           (bdata->node_boot_start & (align - 1UL)) != 0)
-               offset = (align - (bdata->node_boot_start & (align - 1UL)));
-       offset >>= PAGE_SHIFT;
+       if (align && (bdata->node_boot_start & (align - 1UL)) != 0)
+               offset = align - (bdata->node_boot_start & (align - 1UL));
+       offset = PFN_DOWN(offset);
 
        /*
         * We try to allocate bootmem pages above 'goal'
         * first, then we try to allocate lower pages.
         */
-       if (goal && (goal >= bdata->node_boot_start) && 
-           ((goal >> PAGE_SHIFT) < end_pfn)) {
+       if (goal && goal >= bdata->node_boot_start && PFN_DOWN(goal) < end_pfn) {
                preferred = goal - bdata->node_boot_start;
 
                if (bdata->last_success >= preferred)
@@ -212,9 +247,8 @@ __alloc_bootmem_core(struct bootmem_data *bdata, unsigned long size,
        } else
                preferred = 0;
 
-       preferred = ALIGN(preferred, align) >> PAGE_SHIFT;
-       preferred += offset;
-       areasize = (size+PAGE_SIZE-1)/PAGE_SIZE;
+       preferred = PFN_DOWN(ALIGN(preferred, align)) + offset;
+       areasize = (size + PAGE_SIZE-1) / PAGE_SIZE;
        incr = align >> PAGE_SHIFT ? : 1;
 
 restart_scan:
@@ -229,7 +263,7 @@ restart_scan:
                for (j = i + 1; j < i + areasize; ++j) {
                        if (j >= eidx)
                                goto fail_block;
-                       if (test_bit (j, bdata->node_bootmem_map))
+                       if (test_bit(j, bdata->node_bootmem_map))
                                goto fail_block;
                }
                start = i;
@@ -245,7 +279,7 @@ restart_scan:
        return NULL;
 
 found:
-       bdata->last_success = start << PAGE_SHIFT;
+       bdata->last_success = PFN_PHYS(start);
        BUG_ON(start >= eidx);
 
        /*
@@ -257,19 +291,21 @@ found:
            bdata->last_offset && bdata->last_pos+1 == start) {
                offset = ALIGN(bdata->last_offset, align);
                BUG_ON(offset > PAGE_SIZE);
-               remaining_size = PAGE_SIZE-offset;
+               remaining_size = PAGE_SIZE - offset;
                if (size < remaining_size) {
                        areasize = 0;
                        /* last_pos unchanged */
-                       bdata->last_offset = offset+size;
-                       ret = phys_to_virt(bdata->last_pos*PAGE_SIZE + offset +
-                                               bdata->node_boot_start);
+                       bdata->last_offset = offset + size;
+                       ret = phys_to_virt(bdata->last_pos * PAGE_SIZE +
+                                          offset +
+                                          bdata->node_boot_start);
                } else {
                        remaining_size = size - remaining_size;
-                       areasize = (remaining_size+PAGE_SIZE-1)/PAGE_SIZE;
-                       ret = phys_to_virt(bdata->last_pos*PAGE_SIZE + offset +
-                                               bdata->node_boot_start);
-                       bdata->last_pos = start+areasize-1;
+                       areasize = (remaining_size + PAGE_SIZE-1) / PAGE_SIZE;
+                       ret = phys_to_virt(bdata->last_pos * PAGE_SIZE +
+                                          offset +
+                                          bdata->node_boot_start);
+                       bdata->last_pos = start + areasize - 1;
                        bdata->last_offset = remaining_size;
                }
                bdata->last_offset &= ~PAGE_MASK;
@@ -282,7 +318,7 @@ found:
        /*
         * Reserve the area now:
         */
-       for (i = start; i < start+areasize; i++)
+       for (i = start; i < start + areasize; i++)
                if (unlikely(test_and_set_bit(i, bdata->node_bootmem_map)))
                        BUG();
        memset(ret, 0, size);
@@ -303,8 +339,8 @@ static unsigned long __init free_all_bootmem_core(pg_data_t *pgdat)
 
        count = 0;
        /* first extant page of the node */
-       pfn = bdata->node_boot_start >> PAGE_SHIFT;
-       idx = bdata->node_low_pfn - (bdata->node_boot_start >> PAGE_SHIFT);
+       pfn = PFN_DOWN(bdata->node_boot_start);
+       idx = bdata->node_low_pfn - pfn;
        map = bdata->node_bootmem_map;
        /* Check physaddr is O(LOG2(BITS_PER_LONG)) page aligned */
        if (bdata->node_boot_start == 0 ||
@@ -333,7 +369,7 @@ static unsigned long __init free_all_bootmem_core(pg_data_t *pgdat)
                                }
                        }
                } else {
-                       i+=BITS_PER_LONG;
+                       i += BITS_PER_LONG;
                }
                pfn += BITS_PER_LONG;
        }
@@ -345,9 +381,10 @@ static unsigned long __init free_all_bootmem_core(pg_data_t *pgdat)
         */
        page = virt_to_page(bdata->node_bootmem_map);
        count = 0;
-       for (i = 0; i < ((bdata->node_low_pfn-(bdata->node_boot_start >> PAGE_SHIFT))/8 + PAGE_SIZE-1)/PAGE_SIZE; i++,page++) {
-               count++;
+       idx = (get_mapsize(bdata) + PAGE_SIZE-1) >> PAGE_SHIFT;
+       for (i = 0; i < idx; i++, page++) {
                __free_pages_bootmem(page, 0);
+               count++;
        }
        total += count;
        bdata->node_bootmem_map = NULL;
@@ -355,64 +392,75 @@ static unsigned long __init free_all_bootmem_core(pg_data_t *pgdat)
        return total;
 }
 
-unsigned long __init init_bootmem_node (pg_data_t *pgdat, unsigned long freepfn, unsigned long startpfn, unsigned long endpfn)
+unsigned long __init init_bootmem_node(pg_data_t *pgdat, unsigned long freepfn,
+                               unsigned long startpfn, unsigned long endpfn)
 {
-       return(init_bootmem_core(pgdat, freepfn, startpfn, endpfn));
+       return init_bootmem_core(pgdat, freepfn, startpfn, endpfn);
 }
 
-void __init reserve_bootmem_node (pg_data_t *pgdat, unsigned long physaddr, unsigned long size)
+void __init reserve_bootmem_node(pg_data_t *pgdat, unsigned long physaddr,
+                                unsigned long size, int flags)
 {
-       reserve_bootmem_core(pgdat->bdata, physaddr, size);
+       reserve_bootmem_core(pgdat->bdata, physaddr, size, flags);
 }
 
-void __init free_bootmem_node (pg_data_t *pgdat, unsigned long physaddr, unsigned long size)
+void __init free_bootmem_node(pg_data_t *pgdat, unsigned long physaddr,
+                             unsigned long size)
 {
        free_bootmem_core(pgdat->bdata, physaddr, size);
 }
 
-unsigned long __init free_all_bootmem_node (pg_data_t *pgdat)
+unsigned long __init free_all_bootmem_node(pg_data_t *pgdat)
 {
-       return(free_all_bootmem_core(pgdat));
+       return free_all_bootmem_core(pgdat);
 }
 
-unsigned long __init init_bootmem (unsigned long start, unsigned long pages)
+unsigned long __init init_bootmem(unsigned long start, unsigned long pages)
 {
        max_low_pfn = pages;
        min_low_pfn = start;
-       return(init_bootmem_core(NODE_DATA(0), start, 0, pages));
+       return init_bootmem_core(NODE_DATA(0), start, 0, pages);
 }
 
 #ifndef CONFIG_HAVE_ARCH_BOOTMEM_NODE
-void __init reserve_bootmem (unsigned long addr, unsigned long size)
+int __init reserve_bootmem(unsigned long addr, unsigned long size,
+                           int flags)
 {
-       reserve_bootmem_core(NODE_DATA(0)->bdata, addr, size);
+       return reserve_bootmem_core(NODE_DATA(0)->bdata, addr, size, flags);
 }
 #endif /* !CONFIG_HAVE_ARCH_BOOTMEM_NODE */
 
-void __init free_bootmem (unsigned long addr, unsigned long size)
+void __init free_bootmem(unsigned long addr, unsigned long size)
 {
-       free_bootmem_core(NODE_DATA(0)->bdata, addr, size);
+       bootmem_data_t *bdata;
+       list_for_each_entry(bdata, &bdata_list, list)
+               free_bootmem_core(bdata, addr, size);
 }
 
-unsigned long __init free_all_bootmem (void)
+unsigned long __init free_all_bootmem(void)
 {
-       return(free_all_bootmem_core(NODE_DATA(0)));
+       return free_all_bootmem_core(NODE_DATA(0));
 }
 
-void * __init __alloc_bootmem_nopanic(unsigned long size, unsigned long align, unsigned long goal)
+void * __init __alloc_bootmem_nopanic(unsigned long size, unsigned long align,
+                                     unsigned long goal)
 {
        bootmem_data_t *bdata;
        void *ptr;
 
-       list_for_each_entry(bdata, &bdata_list, list)
-               if ((ptr = __alloc_bootmem_core(bdata, size, align, goal, 0)))
-                       return(ptr);
+       list_for_each_entry(bdata, &bdata_list, list) {
+               ptr = __alloc_bootmem_core(bdata, size, align, goal, 0);
+               if (ptr)
+                       return ptr;
+       }
        return NULL;
 }
 
-void * __init __alloc_bootmem(unsigned long size, unsigned long align, unsigned long goal)
+void * __init __alloc_bootmem(unsigned long size, unsigned long align,
+                             unsigned long goal)
 {
        void *mem = __alloc_bootmem_nopanic(size,align,goal);
+
        if (mem)
                return mem;
        /*
@@ -424,29 +472,34 @@ void * __init __alloc_bootmem(unsigned long size, unsigned long align, unsigned
 }
 
 
-void * __init __alloc_bootmem_node(pg_data_t *pgdat, unsigned long size, unsigned long align,
-                                  unsigned long goal)
+void * __init __alloc_bootmem_node(pg_data_t *pgdat, unsigned long size,
+                                  unsigned long align, unsigned long goal)
 {
        void *ptr;
 
        ptr = __alloc_bootmem_core(pgdat->bdata, size, align, goal, 0);
        if (ptr)
-               return (ptr);
+               return ptr;
 
        return __alloc_bootmem(size, align, goal);
 }
 
-#define LOW32LIMIT 0xffffffff
+#ifndef ARCH_LOW_ADDRESS_LIMIT
+#define ARCH_LOW_ADDRESS_LIMIT 0xffffffffUL
+#endif
 
-void * __init __alloc_bootmem_low(unsigned long size, unsigned long align, unsigned long goal)
+void * __init __alloc_bootmem_low(unsigned long size, unsigned long align,
+                                 unsigned long goal)
 {
        bootmem_data_t *bdata;
        void *ptr;
 
-       list_for_each_entry(bdata, &bdata_list, list)
-               if ((ptr = __alloc_bootmem_core(bdata, size,
-                                                align, goal, LOW32LIMIT)))
-                       return(ptr);
+       list_for_each_entry(bdata, &bdata_list, list) {
+               ptr = __alloc_bootmem_core(bdata, size, align, goal,
+                                               ARCH_LOW_ADDRESS_LIMIT);
+               if (ptr)
+                       return ptr;
+       }
 
        /*
         * Whoops, we cannot satisfy the allocation request.
@@ -459,5 +512,6 @@ void * __init __alloc_bootmem_low(unsigned long size, unsigned long align, unsig
 void * __init __alloc_bootmem_low_node(pg_data_t *pgdat, unsigned long size,
                                       unsigned long align, unsigned long goal)
 {
-       return __alloc_bootmem_core(pgdat->bdata, size, align, goal, LOW32LIMIT);
+       return __alloc_bootmem_core(pgdat->bdata, size, align, goal,
+                                   ARCH_LOW_ADDRESS_LIMIT);
 }