|
@@ -1601,7 +1601,6 @@ static void *__vmalloc_node(unsigned long size, unsigned long align,
|
|
|
static void *__vmalloc_area_node(struct vm_struct *area, gfp_t gfp_mask,
|
|
|
pgprot_t prot, int node)
|
|
|
{
|
|
|
- const int order = 0;
|
|
|
struct page **pages;
|
|
|
unsigned int nr_pages, array_size, i;
|
|
|
const gfp_t nested_gfp = (gfp_mask & GFP_RECLAIM_MASK) | __GFP_ZERO;
|
|
@@ -1629,9 +1628,9 @@ static void *__vmalloc_area_node(struct vm_struct *area, gfp_t gfp_mask,
|
|
|
struct page *page;
|
|
|
|
|
|
if (node == NUMA_NO_NODE)
|
|
|
- page = alloc_pages(alloc_mask, order);
|
|
|
+ page = alloc_page(alloc_mask);
|
|
|
else
|
|
|
- page = alloc_pages_node(node, alloc_mask, order);
|
|
|
+ page = alloc_pages_node(node, alloc_mask, 0);
|
|
|
|
|
|
if (unlikely(!page)) {
|
|
|
/* Successfully allocated i pages, free them in __vunmap() */
|
|
@@ -1648,8 +1647,8 @@ static void *__vmalloc_area_node(struct vm_struct *area, gfp_t gfp_mask,
|
|
|
return area->addr;
|
|
|
|
|
|
fail:
|
|
|
- warn_alloc_failed(gfp_mask, order,
|
|
|
- "vmalloc: allocation failure, allocated %ld of %ld bytes\n",
|
|
|
+ warn_alloc(gfp_mask,
|
|
|
+ "vmalloc: allocation failure, allocated %ld of %ld bytes",
|
|
|
(area->nr_pages*PAGE_SIZE), area->size);
|
|
|
vfree(area->addr);
|
|
|
return NULL;
|
|
@@ -1710,9 +1709,8 @@ void *__vmalloc_node_range(unsigned long size, unsigned long align,
|
|
|
return addr;
|
|
|
|
|
|
fail:
|
|
|
- warn_alloc_failed(gfp_mask, 0,
|
|
|
- "vmalloc: allocation failure: %lu bytes\n",
|
|
|
- real_size);
|
|
|
+ warn_alloc(gfp_mask,
|
|
|
+ "vmalloc: allocation failure: %lu bytes", real_size);
|
|
|
return NULL;
|
|
|
}
|
|
|
|