|
@@ -681,7 +681,7 @@ static void print_trailer(struct kmem_cache *s, struct page *page, u8 *p)
|
|
|
print_section(KERN_ERR, "Bytes b4 ", p - 16, 16);
|
|
|
|
|
|
print_section(KERN_ERR, "Object ", p,
|
|
|
- min_t(unsigned long, s->object_size, PAGE_SIZE));
|
|
|
+ min_t(unsigned int, s->object_size, PAGE_SIZE));
|
|
|
if (s->flags & SLAB_RED_ZONE)
|
|
|
print_section(KERN_ERR, "Redzone ", p + s->object_size,
|
|
|
s->inuse - s->object_size);
|
|
@@ -2399,7 +2399,7 @@ slab_out_of_memory(struct kmem_cache *s, gfp_t gfpflags, int nid)
|
|
|
|
|
|
pr_warn("SLUB: Unable to allocate memory on node %d, gfp=%#x(%pGg)\n",
|
|
|
nid, gfpflags, &gfpflags);
|
|
|
- pr_warn(" cache: %s, object size: %d, buffer size: %d, default order: %d, min order: %d\n",
|
|
|
+ pr_warn(" cache: %s, object size: %u, buffer size: %d, default order: %d, min order: %d\n",
|
|
|
s->name, s->object_size, s->size, oo_order(s->oo),
|
|
|
oo_order(s->min));
|
|
|
|
|
@@ -4255,7 +4255,7 @@ __kmem_cache_alias(const char *name, unsigned int size, unsigned int align,
|
|
|
* Adjust the object sizes so that we clear
|
|
|
* the complete object on kzalloc.
|
|
|
*/
|
|
|
- s->object_size = max(s->object_size, (int)size);
|
|
|
+ s->object_size = max(s->object_size, size);
|
|
|
s->inuse = max(s->inuse, ALIGN(size, sizeof(void *)));
|
|
|
|
|
|
for_each_memcg_cache(c, s) {
|
|
@@ -4901,7 +4901,7 @@ SLAB_ATTR_RO(align);
|
|
|
|
|
|
static ssize_t object_size_show(struct kmem_cache *s, char *buf)
|
|
|
{
|
|
|
- return sprintf(buf, "%d\n", s->object_size);
|
|
|
+ return sprintf(buf, "%u\n", s->object_size);
|
|
|
}
|
|
|
SLAB_ATTR_RO(object_size);
|
|
|
|