|
@@ -1044,16 +1044,17 @@ bad:
|
|
|
}
|
|
|
|
|
|
/* Supports checking bulk free of a constructed freelist */
|
|
|
-static noinline struct kmem_cache_node *free_debug_processing(
|
|
|
+static noinline int free_debug_processing(
|
|
|
struct kmem_cache *s, struct page *page,
|
|
|
void *head, void *tail, int bulk_cnt,
|
|
|
- unsigned long addr, unsigned long *flags)
|
|
|
+ unsigned long addr)
|
|
|
{
|
|
|
struct kmem_cache_node *n = get_node(s, page_to_nid(page));
|
|
|
void *object = head;
|
|
|
int cnt = 0;
|
|
|
+ unsigned long uninitialized_var(flags);
|
|
|
|
|
|
- spin_lock_irqsave(&n->list_lock, *flags);
|
|
|
+ spin_lock_irqsave(&n->list_lock, flags);
|
|
|
slab_lock(page);
|
|
|
|
|
|
if (!check_slab(s, page))
|
|
@@ -1106,17 +1107,14 @@ out:
|
|
|
bulk_cnt, cnt);
|
|
|
|
|
|
slab_unlock(page);
|
|
|
- /*
|
|
|
- * Keep node_lock to preserve integrity
|
|
|
- * until the object is actually freed
|
|
|
- */
|
|
|
- return n;
|
|
|
+ spin_unlock_irqrestore(&n->list_lock, flags);
|
|
|
+ return 1;
|
|
|
|
|
|
fail:
|
|
|
slab_unlock(page);
|
|
|
- spin_unlock_irqrestore(&n->list_lock, *flags);
|
|
|
+ spin_unlock_irqrestore(&n->list_lock, flags);
|
|
|
slab_fix(s, "Object at 0x%p not freed", object);
|
|
|
- return NULL;
|
|
|
+ return 0;
|
|
|
}
|
|
|
|
|
|
static int __init setup_slub_debug(char *str)
|
|
@@ -1207,10 +1205,10 @@ static inline void setup_object_debug(struct kmem_cache *s,
|
|
|
static inline int alloc_debug_processing(struct kmem_cache *s,
|
|
|
struct page *page, void *object, unsigned long addr) { return 0; }
|
|
|
|
|
|
-static inline struct kmem_cache_node *free_debug_processing(
|
|
|
+static inline int free_debug_processing(
|
|
|
struct kmem_cache *s, struct page *page,
|
|
|
void *head, void *tail, int bulk_cnt,
|
|
|
- unsigned long addr, unsigned long *flags) { return NULL; }
|
|
|
+ unsigned long addr) { return 0; }
|
|
|
|
|
|
static inline int slab_pad_check(struct kmem_cache *s, struct page *page)
|
|
|
{ return 1; }
|
|
@@ -2588,8 +2586,7 @@ static void __slab_free(struct kmem_cache *s, struct page *page,
|
|
|
stat(s, FREE_SLOWPATH);
|
|
|
|
|
|
if (kmem_cache_debug(s) &&
|
|
|
- !(n = free_debug_processing(s, page, head, tail, cnt,
|
|
|
- addr, &flags)))
|
|
|
+ !free_debug_processing(s, page, head, tail, cnt, addr))
|
|
|
return;
|
|
|
|
|
|
do {
|