diff options
author | Christoph Lameter <clameter@sgi.com> | 2007-07-30 13:06:46 -0700 |
---|---|---|
committer | Christoph Lameter <clameter@sgi.com> | 2007-08-09 21:57:15 -0700 |
commit | fcda3d89bf1366f6801447eab2d8a75ac5b9c4ce (patch) | |
tree | 47382ad57b48a1a0e46a6190991f88a9e6973e45 /mm | |
parent | 6adb31c90c47262c8a25bf5097de9b3426caf3ae (diff) | |
download | linux-fcda3d89bf1366f6801447eab2d8a75ac5b9c4ce.tar.gz linux-fcda3d89bf1366f6801447eab2d8a75ac5b9c4ce.tar.bz2 linux-fcda3d89bf1366f6801447eab2d8a75ac5b9c4ce.zip |
SLUB: Remove checks for MAX_PARTIAL from kmem_cache_shrink
The MAX_PARTIAL checks were supposed to be an optimization. However, slab
shrinking is a manually triggered process either through running slabinfo
or by the kernel calling kmem_cache_shrink.
If one really wants to shrink a slab then all operations should be done
regardless of the size of the partial list. This also fixes an issue that
could surface if the number of partial slabs was initially above MAX_PARTIAL
in kmem_cache_shrink and later drops below MAX_PARTIAL through the
elimination of empty slabs on the partial list (rare). In that case a few
slabs may be left off the partial list (and only be put back when they
are empty).
Signed-off-by: Christoph Lameter <clameter@sgi.com>
Diffstat (limited to 'mm')
-rw-r--r-- | mm/slub.c | 9 |
1 files changed, 2 insertions, 7 deletions
diff --git a/mm/slub.c b/mm/slub.c index 6c6d74ff0694..64fd80bdae30 100644 --- a/mm/slub.c +++ b/mm/slub.c @@ -2500,15 +2500,11 @@ int kmem_cache_shrink(struct kmem_cache *s) slab_unlock(page); discard_slab(s, page); } else { - if (n->nr_partial > MAX_PARTIAL) - list_move(&page->lru, - slabs_by_inuse + page->inuse); + list_move(&page->lru, + slabs_by_inuse + page->inuse); } } - if (n->nr_partial <= MAX_PARTIAL) - goto out; - /* * Rebuild the partial list with the slabs filled up most * first and the least used slabs at the end. @@ -2516,7 +2512,6 @@ int kmem_cache_shrink(struct kmem_cache *s) for (i = s->objects - 1; i >= 0; i--) list_splice(slabs_by_inuse + i, n->partial.prev); - out: spin_unlock_irqrestore(&n->list_lock, flags); } |