|
@@ -374,7 +374,7 @@ static inline bool __cmpxchg_double_slab(struct kmem_cache *s, struct page *page
|
|
|
if (cmpxchg_double(&page->freelist, &page->counters,
|
|
|
freelist_old, counters_old,
|
|
|
freelist_new, counters_new))
|
|
|
- return 1;
|
|
|
+ return true;
|
|
|
} else
|
|
|
#endif
|
|
|
{
|
|
@@ -384,7 +384,7 @@ static inline bool __cmpxchg_double_slab(struct kmem_cache *s, struct page *page
|
|
|
page->freelist = freelist_new;
|
|
|
set_page_slub_counters(page, counters_new);
|
|
|
slab_unlock(page);
|
|
|
- return 1;
|
|
|
+ return true;
|
|
|
}
|
|
|
slab_unlock(page);
|
|
|
}
|
|
@@ -396,7 +396,7 @@ static inline bool __cmpxchg_double_slab(struct kmem_cache *s, struct page *page
|
|
|
pr_info("%s %s: cmpxchg double redo ", n, s->name);
|
|
|
#endif
|
|
|
|
|
|
- return 0;
|
|
|
+ return false;
|
|
|
}
|
|
|
|
|
|
static inline bool cmpxchg_double_slab(struct kmem_cache *s, struct page *page,
|
|
@@ -410,7 +410,7 @@ static inline bool cmpxchg_double_slab(struct kmem_cache *s, struct page *page,
|
|
|
if (cmpxchg_double(&page->freelist, &page->counters,
|
|
|
freelist_old, counters_old,
|
|
|
freelist_new, counters_new))
|
|
|
- return 1;
|
|
|
+ return true;
|
|
|
} else
|
|
|
#endif
|
|
|
{
|
|
@@ -424,7 +424,7 @@ static inline bool cmpxchg_double_slab(struct kmem_cache *s, struct page *page,
|
|
|
set_page_slub_counters(page, counters_new);
|
|
|
slab_unlock(page);
|
|
|
local_irq_restore(flags);
|
|
|
- return 1;
|
|
|
+ return true;
|
|
|
}
|
|
|
slab_unlock(page);
|
|
|
local_irq_restore(flags);
|
|
@@ -437,7 +437,7 @@ static inline bool cmpxchg_double_slab(struct kmem_cache *s, struct page *page,
|
|
|
pr_info("%s %s: cmpxchg double redo ", n, s->name);
|
|
|
#endif
|
|
|
|
|
|
- return 0;
|
|
|
+ return false;
|
|
|
}
|
|
|
|
|
|
#ifdef CONFIG_SLUB_DEBUG
|