Linux kernel mirror (for testing) git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel os linux

md/raid5: dynamically allocate the md-raid5 shrinker

In preparation for implementing lockless slab shrink, use new APIs to
dynamically allocate the md-raid5 shrinker, so that it can be freed
asynchronously via RCU. Then it doesn't need to wait for RCU read-side
critical section when releasing the struct r5conf.

Link: https://lkml.kernel.org/r/20230911094444.68966-26-zhengqi.arch@bytedance.com
Signed-off-by: Qi Zheng <zhengqi.arch@bytedance.com>
Reviewed-by: Muchun Song <songmuchun@bytedance.com>
Reviewed-by: Song Liu <song@kernel.org>
Cc: Abhinav Kumar <quic_abhinavk@quicinc.com>
Cc: Alasdair Kergon <agk@redhat.com>
Cc: Alexander Viro <viro@zeniv.linux.org.uk>
Cc: Alyssa Rosenzweig <alyssa.rosenzweig@collabora.com>
Cc: Andreas Dilger <adilger.kernel@dilger.ca>
Cc: Andreas Gruenbacher <agruenba@redhat.com>
Cc: Anna Schumaker <anna@kernel.org>
Cc: Arnd Bergmann <arnd@arndb.de>
Cc: Bob Peterson <rpeterso@redhat.com>
Cc: Borislav Petkov <bp@alien8.de>
Cc: Carlos Llamas <cmllamas@google.com>
Cc: Chandan Babu R <chandan.babu@oracle.com>
Cc: Chao Yu <chao@kernel.org>
Cc: Chris Mason <clm@fb.com>
Cc: Christian Brauner <brauner@kernel.org>
Cc: Christian Koenig <christian.koenig@amd.com>
Cc: Chuck Lever <cel@kernel.org>
Cc: Coly Li <colyli@suse.de>
Cc: Dai Ngo <Dai.Ngo@oracle.com>
Cc: Daniel Vetter <daniel@ffwll.ch>
Cc: Daniel Vetter <daniel.vetter@ffwll.ch>
Cc: "Darrick J. Wong" <djwong@kernel.org>
Cc: Dave Chinner <david@fromorbit.com>
Cc: Dave Hansen <dave.hansen@linux.intel.com>
Cc: David Airlie <airlied@gmail.com>
Cc: David Hildenbrand <david@redhat.com>
Cc: David Sterba <dsterba@suse.com>
Cc: Dmitry Baryshkov <dmitry.baryshkov@linaro.org>
Cc: Gao Xiang <hsiangkao@linux.alibaba.com>
Cc: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
Cc: Huang Rui <ray.huang@amd.com>
Cc: Ingo Molnar <mingo@redhat.com>
Cc: Jaegeuk Kim <jaegeuk@kernel.org>
Cc: Jani Nikula <jani.nikula@linux.intel.com>
Cc: Jan Kara <jack@suse.cz>
Cc: Jason Wang <jasowang@redhat.com>
Cc: Jeff Layton <jlayton@kernel.org>
Cc: Jeffle Xu <jefflexu@linux.alibaba.com>
Cc: Joel Fernandes (Google) <joel@joelfernandes.org>
Cc: Joonas Lahtinen <joonas.lahtinen@linux.intel.com>
Cc: Josef Bacik <josef@toxicpanda.com>
Cc: Juergen Gross <jgross@suse.com>
Cc: Kent Overstreet <kent.overstreet@gmail.com>
Cc: Kirill Tkhai <tkhai@ya.ru>
Cc: Marijn Suijten <marijn.suijten@somainline.org>
Cc: "Michael S. Tsirkin" <mst@redhat.com>
Cc: Mike Snitzer <snitzer@kernel.org>
Cc: Minchan Kim <minchan@kernel.org>
Cc: Muchun Song <muchun.song@linux.dev>
Cc: Nadav Amit <namit@vmware.com>
Cc: Neil Brown <neilb@suse.de>
Cc: Oleksandr Tyshchenko <oleksandr_tyshchenko@epam.com>
Cc: Olga Kornievskaia <kolga@netapp.com>
Cc: Paul E. McKenney <paulmck@kernel.org>
Cc: Richard Weinberger <richard@nod.at>
Cc: Rob Clark <robdclark@gmail.com>
Cc: Rob Herring <robh@kernel.org>
Cc: Rodrigo Vivi <rodrigo.vivi@intel.com>
Cc: Roman Gushchin <roman.gushchin@linux.dev>
Cc: Sean Paul <sean@poorly.run>
Cc: Sergey Senozhatsky <senozhatsky@chromium.org>
Cc: Stefano Stabellini <sstabellini@kernel.org>
Cc: Steven Price <steven.price@arm.com>
Cc: "Theodore Ts'o" <tytso@mit.edu>
Cc: Thomas Gleixner <tglx@linutronix.de>
Cc: Tomeu Vizoso <tomeu.vizoso@collabora.com>
Cc: Tom Talpey <tom@talpey.com>
Cc: Trond Myklebust <trond.myklebust@hammerspace.com>
Cc: Tvrtko Ursulin <tvrtko.ursulin@linux.intel.com>
Cc: Vlastimil Babka <vbabka@suse.cz>
Cc: Xuan Zhuo <xuanzhuo@linux.alibaba.com>
Cc: Yue Hu <huyue2@coolpad.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>

authored by

Qi Zheng and committed by
Andrew Morton
86298d8b ba3d6aca

+16 -12
+15 -11
drivers/md/raid5.c
··· 7401 7401 7402 7402 log_exit(conf); 7403 7403 7404 - unregister_shrinker(&conf->shrinker); 7404 + shrinker_free(conf->shrinker); 7405 7405 free_thread_groups(conf); 7406 7406 shrink_stripes(conf); 7407 7407 raid5_free_percpu(conf); ··· 7449 7449 static unsigned long raid5_cache_scan(struct shrinker *shrink, 7450 7450 struct shrink_control *sc) 7451 7451 { 7452 - struct r5conf *conf = container_of(shrink, struct r5conf, shrinker); 7452 + struct r5conf *conf = shrink->private_data; 7453 7453 unsigned long ret = SHRINK_STOP; 7454 7454 7455 7455 if (mutex_trylock(&conf->cache_size_mutex)) { ··· 7470 7470 static unsigned long raid5_cache_count(struct shrinker *shrink, 7471 7471 struct shrink_control *sc) 7472 7472 { 7473 - struct r5conf *conf = container_of(shrink, struct r5conf, shrinker); 7473 + struct r5conf *conf = shrink->private_data; 7474 7474 7475 7475 if (conf->max_nr_stripes < conf->min_nr_stripes) 7476 7476 /* unlikely, but not impossible */ ··· 7705 7705 * it reduces the queue depth and so can hurt throughput. 7706 7706 * So set it rather large, scaled by number of devices. 7707 7707 */ 7708 - conf->shrinker.seeks = DEFAULT_SEEKS * conf->raid_disks * 4; 7709 - conf->shrinker.scan_objects = raid5_cache_scan; 7710 - conf->shrinker.count_objects = raid5_cache_count; 7711 - conf->shrinker.batch = 128; 7712 - conf->shrinker.flags = 0; 7713 - ret = register_shrinker(&conf->shrinker, "md-raid5:%s", mdname(mddev)); 7714 - if (ret) { 7715 - pr_warn("md/raid:%s: couldn't register shrinker.\n", 7708 + conf->shrinker = shrinker_alloc(0, "md-raid5:%s", mdname(mddev)); 7709 + if (!conf->shrinker) { 7710 + ret = -ENOMEM; 7711 + pr_warn("md/raid:%s: couldn't allocate shrinker.\n", 7716 7712 mdname(mddev)); 7717 7713 goto abort; 7718 7714 } 7715 + 7716 + conf->shrinker->seeks = DEFAULT_SEEKS * conf->raid_disks * 4; 7717 + conf->shrinker->scan_objects = raid5_cache_scan; 7718 + conf->shrinker->count_objects = raid5_cache_count; 7719 + conf->shrinker->batch = 128; 7720 + conf->shrinker->private_data = conf; 7721 + 7722 + shrinker_register(conf->shrinker); 7719 7723 7720 7724 sprintf(pers_name, "raid%d", mddev->new_level); 7721 7725 rcu_assign_pointer(conf->thread,
+1 -1
drivers/md/raid5.h
··· 670 670 wait_queue_head_t wait_for_stripe; 671 671 wait_queue_head_t wait_for_overlap; 672 672 unsigned long cache_state; 673 - struct shrinker shrinker; 673 + struct shrinker *shrinker; 674 674 int pool_size; /* number of disks in stripeheads in pool */ 675 675 spinlock_t device_lock; 676 676 struct disk_info *disks;