mm: slub: add missing TID bump in kmem_cache_alloc_bulk()

When kmem_cache_alloc_bulk() attempts to allocate N objects from a percpu
freelist of length M, and N > M > 0, it will first remove the M elements
from the percpu freelist, then call ___slab_alloc() to allocate the next
element and repopulate the percpu freelist. ___slab_alloc() can re-enable
IRQs via allocate_slab(), so the TID must be bumped before ___slab_alloc()
to properly commit the freelist head change.

Fix it by unconditionally bumping c->tid when entering the slowpath.

Cc: stable@vger.kernel.org
Fixes: ebe909e0fdb3 ("slub: improve bulk alloc strategy")
Signed-off-by: Jann Horn <jannh@google.com>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>

authored by Jann Horn and committed by Linus Torvalds fd4d9c7d ac309e77

+9
+9
mm/slub.c
··· 3175 3175 3176 3176 if (unlikely(!object)) { 3177 3177 /* 3178 + * We may have removed an object from c->freelist using 3179 + * the fastpath in the previous iteration; in that case, 3180 + * c->tid has not been bumped yet. 3181 + * Since ___slab_alloc() may reenable interrupts while 3182 + * allocating memory, we should bump c->tid now. 3183 + */ 3184 + c->tid = next_tid(c->tid); 3185 + 3186 + /* 3178 3187 * Invoking slow path likely have side-effect 3179 3188 * of re-populating per CPU c->freelist 3180 3189 */