diff options
author | Jann Horn <jannh@google.com> | 2020-03-17 01:28:45 +0100 |
---|---|---|
committer | Greg Kroah-Hartman <gregkh@linuxfoundation.org> | 2020-03-21 08:11:58 +0100 |
commit | ae119b7e12472517bc35c1c003d5abf26653674a (patch) | |
tree | a979f7540a8cadb7e3994b1de9b8203c5544e0b1 /samples | |
parent | 47b6d269dbbdbee277f8089ebceae9c9666bd918 (diff) | |
download | linux-stable-ae119b7e12472517bc35c1c003d5abf26653674a.tar.gz linux-stable-ae119b7e12472517bc35c1c003d5abf26653674a.tar.bz2 linux-stable-ae119b7e12472517bc35c1c003d5abf26653674a.zip |
mm: slub: add missing TID bump in kmem_cache_alloc_bulk()
commit fd4d9c7d0c71866ec0c2825189ebd2ce35bd95b8 upstream.
When kmem_cache_alloc_bulk() attempts to allocate N objects from a percpu
freelist of length M, and N > M > 0, it will first remove the M elements
from the percpu freelist, then call ___slab_alloc() to allocate the next
element and repopulate the percpu freelist. ___slab_alloc() can re-enable
IRQs via allocate_slab(), so the TID must be bumped before ___slab_alloc()
to properly commit the freelist head change.
Fix it by unconditionally bumping c->tid when entering the slowpath.
Cc: stable@vger.kernel.org
Fixes: ebe909e0fdb3 ("slub: improve bulk alloc strategy")
Signed-off-by: Jann Horn <jannh@google.com>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
Diffstat (limited to 'samples')
0 files changed, 0 insertions, 0 deletions