at for-next 3.2 kB view raw
1/* SPDX-License-Identifier: GPL-2.0 */ 2#ifndef _LINUX_PAGE_COUNTER_H 3#define _LINUX_PAGE_COUNTER_H 4 5#include <linux/atomic.h> 6#include <linux/cache.h> 7#include <linux/limits.h> 8#include <asm/page.h> 9 10struct page_counter { 11 /* 12 * Make sure 'usage' does not share cacheline with any other field. The 13 * memcg->memory.usage is a hot member of struct mem_cgroup. 14 */ 15 atomic_long_t usage; 16 CACHELINE_PADDING(_pad1_); 17 18 /* effective memory.min and memory.min usage tracking */ 19 unsigned long emin; 20 atomic_long_t min_usage; 21 atomic_long_t children_min_usage; 22 23 /* effective memory.low and memory.low usage tracking */ 24 unsigned long elow; 25 atomic_long_t low_usage; 26 atomic_long_t children_low_usage; 27 28 unsigned long watermark; 29 /* Latest cg2 reset watermark */ 30 unsigned long local_watermark; 31 unsigned long failcnt; 32 33 /* Keep all the read most fields in a separete cacheline. */ 34 CACHELINE_PADDING(_pad2_); 35 36 bool protection_support; 37 unsigned long min; 38 unsigned long low; 39 unsigned long high; 40 unsigned long max; 41 struct page_counter *parent; 42} ____cacheline_internodealigned_in_smp; 43 44#if BITS_PER_LONG == 32 45#define PAGE_COUNTER_MAX LONG_MAX 46#else 47#define PAGE_COUNTER_MAX (LONG_MAX / PAGE_SIZE) 48#endif 49 50/* 51 * Protection is supported only for the first counter (with id 0). 52 */ 53static inline void page_counter_init(struct page_counter *counter, 54 struct page_counter *parent, 55 bool protection_support) 56{ 57 counter->usage = (atomic_long_t)ATOMIC_LONG_INIT(0); 58 counter->max = PAGE_COUNTER_MAX; 59 counter->parent = parent; 60 counter->protection_support = protection_support; 61} 62 63static inline unsigned long page_counter_read(struct page_counter *counter) 64{ 65 return atomic_long_read(&counter->usage); 66} 67 68void page_counter_cancel(struct page_counter *counter, unsigned long nr_pages); 69void page_counter_charge(struct page_counter *counter, unsigned long nr_pages); 70bool page_counter_try_charge(struct page_counter *counter, 71 unsigned long nr_pages, 72 struct page_counter **fail); 73void page_counter_uncharge(struct page_counter *counter, unsigned long nr_pages); 74void page_counter_set_min(struct page_counter *counter, unsigned long nr_pages); 75void page_counter_set_low(struct page_counter *counter, unsigned long nr_pages); 76 77static inline void page_counter_set_high(struct page_counter *counter, 78 unsigned long nr_pages) 79{ 80 WRITE_ONCE(counter->high, nr_pages); 81} 82 83int page_counter_set_max(struct page_counter *counter, unsigned long nr_pages); 84int page_counter_memparse(const char *buf, const char *max, 85 unsigned long *nr_pages); 86 87static inline void page_counter_reset_watermark(struct page_counter *counter) 88{ 89 unsigned long usage = page_counter_read(counter); 90 91 /* 92 * Update local_watermark first, so it's always <= watermark 93 * (modulo CPU/compiler re-ordering) 94 */ 95 counter->local_watermark = usage; 96 counter->watermark = usage; 97} 98 99#ifdef CONFIG_MEMCG 100void page_counter_calculate_protection(struct page_counter *root, 101 struct page_counter *counter, 102 bool recursive_protection); 103#else 104static inline void page_counter_calculate_protection(struct page_counter *root, 105 struct page_counter *counter, 106 bool recursive_protection) {} 107#endif 108 109#endif /* _LINUX_PAGE_COUNTER_H */