Linux kernel mirror (for testing) git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel os linux

mm: skip CMA pages when they are not available

This patch fixes unproductive reclaiming of CMA pages by skipping them
when they are not available for current context. It arises from the below
OOM issue, which was caused by a large proportion of MIGRATE_CMA pages
among free pages.

[ 36.172486] [03-19 10:05:52.172] ActivityManager: page allocation failure: order:0, mode:0xc00(GFP_NOIO), nodemask=(null),cpuset=foreground,mems_allowed=0
[ 36.189447] [03-19 10:05:52.189] DMA32: 0*4kB 447*8kB (C) 217*16kB (C) 124*32kB (C) 136*64kB (C) 70*128kB (C) 22*256kB (C) 3*512kB (C) 0*1024kB 0*2048kB 0*4096kB = 35848kB
[ 36.193125] [03-19 10:05:52.193] Normal: 231*4kB (UMEH) 49*8kB (MEH) 14*16kB (H) 13*32kB (H) 8*64kB (H) 2*128kB (H) 0*256kB 1*512kB (H) 0*1024kB 0*2048kB 0*4096kB = 3236kB
...
[ 36.234447] [03-19 10:05:52.234] SLUB: Unable to allocate memory on node -1, gfp=0xa20(GFP_ATOMIC)
[ 36.234455] [03-19 10:05:52.234] cache: ext4_io_end, object size: 64, buffer size: 64, default order: 0, min order: 0
[ 36.234459] [03-19 10:05:52.234] node 0: slabs: 53,objs: 3392, free: 0

This change further decreases the chance for wrong OOMs in the presence
of a lot of CMA memory.

[david@redhat.com: changelog addition]
Link: https://lkml.kernel.org/r/1685501461-19290-1-git-send-email-zhaoyang.huang@unisoc.com
Signed-off-by: Zhaoyang Huang <zhaoyang.huang@unisoc.com>
Acked-by: David Hildenbrand <david@redhat.com>
Cc: ke.wang <ke.wang@unisoc.com>
Cc: Matthew Wilcox <willy@infradead.org>
Cc: Minchan Kim <minchan@kernel.org>
Cc: Suren Baghdasaryan <surenb@google.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>

authored by

Zhaoyang Huang and committed by
Andrew Morton
5da226db ce5df776

+21 -1
+21 -1
mm/vmscan.c
··· 2271 2271 2272 2272 } 2273 2273 2274 + #ifdef CONFIG_CMA 2275 + /* 2276 + * It is waste of effort to scan and reclaim CMA pages if it is not available 2277 + * for current allocation context. Kswapd can not be enrolled as it can not 2278 + * distinguish this scenario by using sc->gfp_mask = GFP_KERNEL 2279 + */ 2280 + static bool skip_cma(struct folio *folio, struct scan_control *sc) 2281 + { 2282 + return !current_is_kswapd() && 2283 + gfp_migratetype(sc->gfp_mask) != MIGRATE_MOVABLE && 2284 + get_pageblock_migratetype(&folio->page) == MIGRATE_CMA; 2285 + } 2286 + #else 2287 + static bool skip_cma(struct folio *folio, struct scan_control *sc) 2288 + { 2289 + return false; 2290 + } 2291 + #endif 2292 + 2274 2293 /* 2275 2294 * Isolating page from the lruvec to fill in @dst list by nr_to_scan times. 2276 2295 * ··· 2336 2317 nr_pages = folio_nr_pages(folio); 2337 2318 total_scan += nr_pages; 2338 2319 2339 - if (folio_zonenum(folio) > sc->reclaim_idx) { 2320 + if (folio_zonenum(folio) > sc->reclaim_idx || 2321 + skip_cma(folio, sc)) { 2340 2322 nr_skipped[folio_zonenum(folio)] += nr_pages; 2341 2323 move_to = &folios_skipped; 2342 2324 goto move;