|
@@ -227,7 +227,7 @@ static void kmem_cache_node_init(struct kmem_cache_node *parent)
|
|
INIT_LIST_HEAD(&parent->slabs_full);
|
|
INIT_LIST_HEAD(&parent->slabs_full);
|
|
INIT_LIST_HEAD(&parent->slabs_partial);
|
|
INIT_LIST_HEAD(&parent->slabs_partial);
|
|
INIT_LIST_HEAD(&parent->slabs_free);
|
|
INIT_LIST_HEAD(&parent->slabs_free);
|
|
- parent->active_slabs = 0;
|
|
|
|
|
|
+ parent->total_slabs = 0;
|
|
parent->free_slabs = 0;
|
|
parent->free_slabs = 0;
|
|
parent->shared = NULL;
|
|
parent->shared = NULL;
|
|
parent->alien = NULL;
|
|
parent->alien = NULL;
|
|
@@ -1381,20 +1381,18 @@ slab_out_of_memory(struct kmem_cache *cachep, gfp_t gfpflags, int nodeid)
|
|
cachep->name, cachep->size, cachep->gfporder);
|
|
cachep->name, cachep->size, cachep->gfporder);
|
|
|
|
|
|
for_each_kmem_cache_node(cachep, node, n) {
|
|
for_each_kmem_cache_node(cachep, node, n) {
|
|
- unsigned long active_objs = 0, free_objs = 0;
|
|
|
|
- unsigned long active_slabs, num_slabs;
|
|
|
|
|
|
+ unsigned long total_slabs, free_slabs, free_objs;
|
|
|
|
|
|
spin_lock_irqsave(&n->list_lock, flags);
|
|
spin_lock_irqsave(&n->list_lock, flags);
|
|
- active_slabs = n->active_slabs;
|
|
|
|
- num_slabs = active_slabs + n->free_slabs;
|
|
|
|
-
|
|
|
|
- active_objs += (num_slabs * cachep->num) - n->free_objects;
|
|
|
|
- free_objs += n->free_objects;
|
|
|
|
|
|
+ total_slabs = n->total_slabs;
|
|
|
|
+ free_slabs = n->free_slabs;
|
|
|
|
+ free_objs = n->free_objects;
|
|
spin_unlock_irqrestore(&n->list_lock, flags);
|
|
spin_unlock_irqrestore(&n->list_lock, flags);
|
|
|
|
|
|
- pr_warn(" node %d: slabs: %ld/%ld, objs: %ld/%ld, free: %ld\n",
|
|
|
|
- node, active_slabs, num_slabs, active_objs,
|
|
|
|
- num_slabs * cachep->num, free_objs);
|
|
|
|
|
|
+ pr_warn(" node %d: slabs: %ld/%ld, objs: %ld/%ld\n",
|
|
|
|
+ node, total_slabs - free_slabs, total_slabs,
|
|
|
|
+ (total_slabs * cachep->num) - free_objs,
|
|
|
|
+ total_slabs * cachep->num);
|
|
}
|
|
}
|
|
#endif
|
|
#endif
|
|
}
|
|
}
|
|
@@ -2307,6 +2305,7 @@ static int drain_freelist(struct kmem_cache *cache,
|
|
page = list_entry(p, struct page, lru);
|
|
page = list_entry(p, struct page, lru);
|
|
list_del(&page->lru);
|
|
list_del(&page->lru);
|
|
n->free_slabs--;
|
|
n->free_slabs--;
|
|
|
|
+ n->total_slabs--;
|
|
/*
|
|
/*
|
|
* Safe to drop the lock. The slab is no longer linked
|
|
* Safe to drop the lock. The slab is no longer linked
|
|
* to the cache.
|
|
* to the cache.
|
|
@@ -2741,13 +2740,12 @@ static void cache_grow_end(struct kmem_cache *cachep, struct page *page)
|
|
n = get_node(cachep, page_to_nid(page));
|
|
n = get_node(cachep, page_to_nid(page));
|
|
|
|
|
|
spin_lock(&n->list_lock);
|
|
spin_lock(&n->list_lock);
|
|
|
|
+ n->total_slabs++;
|
|
if (!page->active) {
|
|
if (!page->active) {
|
|
list_add_tail(&page->lru, &(n->slabs_free));
|
|
list_add_tail(&page->lru, &(n->slabs_free));
|
|
n->free_slabs++;
|
|
n->free_slabs++;
|
|
- } else {
|
|
|
|
|
|
+ } else
|
|
fixup_slab_list(cachep, n, page, &list);
|
|
fixup_slab_list(cachep, n, page, &list);
|
|
- n->active_slabs++;
|
|
|
|
- }
|
|
|
|
|
|
|
|
STATS_INC_GROWN(cachep);
|
|
STATS_INC_GROWN(cachep);
|
|
n->free_objects += cachep->num - page->active;
|
|
n->free_objects += cachep->num - page->active;
|
|
@@ -2874,7 +2872,7 @@ static inline void fixup_slab_list(struct kmem_cache *cachep,
|
|
|
|
|
|
/* Try to find non-pfmemalloc slab if needed */
|
|
/* Try to find non-pfmemalloc slab if needed */
|
|
static noinline struct page *get_valid_first_slab(struct kmem_cache_node *n,
|
|
static noinline struct page *get_valid_first_slab(struct kmem_cache_node *n,
|
|
- struct page *page, bool *page_is_free, bool pfmemalloc)
|
|
|
|
|
|
+ struct page *page, bool pfmemalloc)
|
|
{
|
|
{
|
|
if (!page)
|
|
if (!page)
|
|
return NULL;
|
|
return NULL;
|
|
@@ -2893,10 +2891,9 @@ static noinline struct page *get_valid_first_slab(struct kmem_cache_node *n,
|
|
|
|
|
|
/* Move pfmemalloc slab to the end of list to speed up next search */
|
|
/* Move pfmemalloc slab to the end of list to speed up next search */
|
|
list_del(&page->lru);
|
|
list_del(&page->lru);
|
|
- if (*page_is_free) {
|
|
|
|
- WARN_ON(page->active);
|
|
|
|
|
|
+ if (!page->active) {
|
|
list_add_tail(&page->lru, &n->slabs_free);
|
|
list_add_tail(&page->lru, &n->slabs_free);
|
|
- *page_is_free = false;
|
|
|
|
|
|
+ n->free_slabs++;
|
|
} else
|
|
} else
|
|
list_add_tail(&page->lru, &n->slabs_partial);
|
|
list_add_tail(&page->lru, &n->slabs_partial);
|
|
|
|
|
|
@@ -2908,7 +2905,7 @@ static noinline struct page *get_valid_first_slab(struct kmem_cache_node *n,
|
|
n->free_touched = 1;
|
|
n->free_touched = 1;
|
|
list_for_each_entry(page, &n->slabs_free, lru) {
|
|
list_for_each_entry(page, &n->slabs_free, lru) {
|
|
if (!PageSlabPfmemalloc(page)) {
|
|
if (!PageSlabPfmemalloc(page)) {
|
|
- *page_is_free = true;
|
|
|
|
|
|
+ n->free_slabs--;
|
|
return page;
|
|
return page;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
@@ -2919,26 +2916,19 @@ static noinline struct page *get_valid_first_slab(struct kmem_cache_node *n,
|
|
static struct page *get_first_slab(struct kmem_cache_node *n, bool pfmemalloc)
|
|
static struct page *get_first_slab(struct kmem_cache_node *n, bool pfmemalloc)
|
|
{
|
|
{
|
|
struct page *page;
|
|
struct page *page;
|
|
- bool page_is_free = false;
|
|
|
|
|
|
|
|
assert_spin_locked(&n->list_lock);
|
|
assert_spin_locked(&n->list_lock);
|
|
- page = list_first_entry_or_null(&n->slabs_partial,
|
|
|
|
- struct page, lru);
|
|
|
|
|
|
+ page = list_first_entry_or_null(&n->slabs_partial, struct page, lru);
|
|
if (!page) {
|
|
if (!page) {
|
|
n->free_touched = 1;
|
|
n->free_touched = 1;
|
|
- page = list_first_entry_or_null(&n->slabs_free,
|
|
|
|
- struct page, lru);
|
|
|
|
|
|
+ page = list_first_entry_or_null(&n->slabs_free, struct page,
|
|
|
|
+ lru);
|
|
if (page)
|
|
if (page)
|
|
- page_is_free = true;
|
|
|
|
|
|
+ n->free_slabs--;
|
|
}
|
|
}
|
|
|
|
|
|
if (sk_memalloc_socks())
|
|
if (sk_memalloc_socks())
|
|
- page = get_valid_first_slab(n, page, &page_is_free, pfmemalloc);
|
|
|
|
-
|
|
|
|
- if (page && page_is_free) {
|
|
|
|
- n->active_slabs++;
|
|
|
|
- n->free_slabs--;
|
|
|
|
- }
|
|
|
|
|
|
+ page = get_valid_first_slab(n, page, pfmemalloc);
|
|
|
|
|
|
return page;
|
|
return page;
|
|
}
|
|
}
|
|
@@ -3441,7 +3431,6 @@ static void free_block(struct kmem_cache *cachep, void **objpp,
|
|
if (page->active == 0) {
|
|
if (page->active == 0) {
|
|
list_add(&page->lru, &n->slabs_free);
|
|
list_add(&page->lru, &n->slabs_free);
|
|
n->free_slabs++;
|
|
n->free_slabs++;
|
|
- n->active_slabs--;
|
|
|
|
} else {
|
|
} else {
|
|
/* Unconditionally move a slab to the end of the
|
|
/* Unconditionally move a slab to the end of the
|
|
* partial list on free - maximum time for the
|
|
* partial list on free - maximum time for the
|
|
@@ -3457,6 +3446,7 @@ static void free_block(struct kmem_cache *cachep, void **objpp,
|
|
page = list_last_entry(&n->slabs_free, struct page, lru);
|
|
page = list_last_entry(&n->slabs_free, struct page, lru);
|
|
list_move(&page->lru, list);
|
|
list_move(&page->lru, list);
|
|
n->free_slabs--;
|
|
n->free_slabs--;
|
|
|
|
+ n->total_slabs--;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
@@ -4109,8 +4099,8 @@ out:
|
|
void get_slabinfo(struct kmem_cache *cachep, struct slabinfo *sinfo)
|
|
void get_slabinfo(struct kmem_cache *cachep, struct slabinfo *sinfo)
|
|
{
|
|
{
|
|
unsigned long active_objs, num_objs, active_slabs;
|
|
unsigned long active_objs, num_objs, active_slabs;
|
|
- unsigned long num_slabs = 0, free_objs = 0, shared_avail = 0;
|
|
|
|
- unsigned long num_slabs_free = 0;
|
|
|
|
|
|
+ unsigned long total_slabs = 0, free_objs = 0, shared_avail = 0;
|
|
|
|
+ unsigned long free_slabs = 0;
|
|
int node;
|
|
int node;
|
|
struct kmem_cache_node *n;
|
|
struct kmem_cache_node *n;
|
|
|
|
|
|
@@ -4118,9 +4108,8 @@ void get_slabinfo(struct kmem_cache *cachep, struct slabinfo *sinfo)
|
|
check_irq_on();
|
|
check_irq_on();
|
|
spin_lock_irq(&n->list_lock);
|
|
spin_lock_irq(&n->list_lock);
|
|
|
|
|
|
- num_slabs += n->active_slabs + n->free_slabs;
|
|
|
|
- num_slabs_free += n->free_slabs;
|
|
|
|
-
|
|
|
|
|
|
+ total_slabs += n->total_slabs;
|
|
|
|
+ free_slabs += n->free_slabs;
|
|
free_objs += n->free_objects;
|
|
free_objs += n->free_objects;
|
|
|
|
|
|
if (n->shared)
|
|
if (n->shared)
|
|
@@ -4128,15 +4117,14 @@ void get_slabinfo(struct kmem_cache *cachep, struct slabinfo *sinfo)
|
|
|
|
|
|
spin_unlock_irq(&n->list_lock);
|
|
spin_unlock_irq(&n->list_lock);
|
|
}
|
|
}
|
|
- num_objs = num_slabs * cachep->num;
|
|
|
|
- active_slabs = num_slabs - num_slabs_free;
|
|
|
|
-
|
|
|
|
|
|
+ num_objs = total_slabs * cachep->num;
|
|
|
|
+ active_slabs = total_slabs - free_slabs;
|
|
active_objs = num_objs - free_objs;
|
|
active_objs = num_objs - free_objs;
|
|
|
|
|
|
sinfo->active_objs = active_objs;
|
|
sinfo->active_objs = active_objs;
|
|
sinfo->num_objs = num_objs;
|
|
sinfo->num_objs = num_objs;
|
|
sinfo->active_slabs = active_slabs;
|
|
sinfo->active_slabs = active_slabs;
|
|
- sinfo->num_slabs = num_slabs;
|
|
|
|
|
|
+ sinfo->num_slabs = total_slabs;
|
|
sinfo->shared_avail = shared_avail;
|
|
sinfo->shared_avail = shared_avail;
|
|
sinfo->limit = cachep->limit;
|
|
sinfo->limit = cachep->limit;
|
|
sinfo->batchcount = cachep->batchcount;
|
|
sinfo->batchcount = cachep->batchcount;
|