Linux kernel mirror (for testing)
git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel
os
linux
1/* SPDX-License-Identifier: GPL-2.0 */
2#ifndef _BCACHEFS_ALLOC_FOREGROUND_H
3#define _BCACHEFS_ALLOC_FOREGROUND_H
4
5#include "bcachefs.h"
6#include "buckets.h"
7#include "alloc_types.h"
8#include "extents.h"
9#include "io_write_types.h"
10#include "sb-members.h"
11
12#include <linux/hash.h>
13
14struct bkey;
15struct bch_dev;
16struct bch_fs;
17struct bch_devs_List;
18
19extern const char * const bch2_watermarks[];
20
21void bch2_reset_alloc_cursors(struct bch_fs *);
22
23struct dev_alloc_list {
24 unsigned nr;
25 u8 data[BCH_SB_MEMBERS_MAX];
26};
27
28struct alloc_request {
29 unsigned nr_replicas;
30 unsigned target;
31 bool ec;
32 enum bch_watermark watermark;
33 enum bch_write_flags flags;
34 enum bch_data_type data_type;
35 struct bch_devs_list *devs_have;
36 struct write_point *wp;
37
38 /* These fields are used primarily by open_bucket_add_buckets */
39 struct open_buckets ptrs;
40 unsigned nr_effective; /* sum of @ptrs durability */
41 bool have_cache; /* have we allocated from a 0 durability dev */
42 struct bch_devs_mask devs_may_alloc;
43
44 /* bch2_bucket_alloc_set_trans(): */
45 struct dev_alloc_list devs_sorted;
46 struct bch_dev_usage usage;
47
48 /* bch2_bucket_alloc_trans(): */
49 struct bch_dev *ca;
50
51 enum {
52 BTREE_BITMAP_NO,
53 BTREE_BITMAP_YES,
54 BTREE_BITMAP_ANY,
55 } btree_bitmap;
56
57 struct {
58 u64 buckets_seen;
59 u64 skipped_open;
60 u64 skipped_need_journal_commit;
61 u64 need_journal_commit;
62 u64 skipped_nocow;
63 u64 skipped_nouse;
64 u64 skipped_mi_btree_bitmap;
65 } counters;
66
67 unsigned scratch_nr_replicas;
68 unsigned scratch_nr_effective;
69 bool scratch_have_cache;
70 enum bch_data_type scratch_data_type;
71 struct open_buckets scratch_ptrs;
72 struct bch_devs_mask scratch_devs_may_alloc;
73};
74
75void bch2_dev_alloc_list(struct bch_fs *,
76 struct dev_stripe_state *,
77 struct bch_devs_mask *,
78 struct dev_alloc_list *);
79void bch2_dev_stripe_increment(struct bch_dev *, struct dev_stripe_state *);
80
81static inline struct bch_dev *ob_dev(struct bch_fs *c, struct open_bucket *ob)
82{
83 return bch2_dev_have_ref(c, ob->dev);
84}
85
86static inline unsigned bch2_open_buckets_reserved(enum bch_watermark watermark)
87{
88 switch (watermark) {
89 case BCH_WATERMARK_interior_updates:
90 return 0;
91 case BCH_WATERMARK_reclaim:
92 return OPEN_BUCKETS_COUNT / 6;
93 case BCH_WATERMARK_btree:
94 case BCH_WATERMARK_btree_copygc:
95 return OPEN_BUCKETS_COUNT / 4;
96 case BCH_WATERMARK_copygc:
97 return OPEN_BUCKETS_COUNT / 3;
98 default:
99 return OPEN_BUCKETS_COUNT / 2;
100 }
101}
102
103struct open_bucket *bch2_bucket_alloc(struct bch_fs *, struct bch_dev *,
104 enum bch_watermark, enum bch_data_type,
105 struct closure *);
106
107static inline void ob_push(struct bch_fs *c, struct open_buckets *obs,
108 struct open_bucket *ob)
109{
110 BUG_ON(obs->nr >= ARRAY_SIZE(obs->v));
111
112 obs->v[obs->nr++] = ob - c->open_buckets;
113}
114
115#define open_bucket_for_each(_c, _obs, _ob, _i) \
116 for ((_i) = 0; \
117 (_i) < (_obs)->nr && \
118 ((_ob) = (_c)->open_buckets + (_obs)->v[_i], true); \
119 (_i)++)
120
121static inline struct open_bucket *ec_open_bucket(struct bch_fs *c,
122 struct open_buckets *obs)
123{
124 struct open_bucket *ob;
125 unsigned i;
126
127 open_bucket_for_each(c, obs, ob, i)
128 if (ob->ec)
129 return ob;
130
131 return NULL;
132}
133
134void bch2_open_bucket_write_error(struct bch_fs *,
135 struct open_buckets *, unsigned, int);
136
137void __bch2_open_bucket_put(struct bch_fs *, struct open_bucket *);
138
139static inline void bch2_open_bucket_put(struct bch_fs *c, struct open_bucket *ob)
140{
141 if (atomic_dec_and_test(&ob->pin))
142 __bch2_open_bucket_put(c, ob);
143}
144
145static inline void bch2_open_buckets_put(struct bch_fs *c,
146 struct open_buckets *ptrs)
147{
148 struct open_bucket *ob;
149 unsigned i;
150
151 open_bucket_for_each(c, ptrs, ob, i)
152 bch2_open_bucket_put(c, ob);
153 ptrs->nr = 0;
154}
155
156static inline void bch2_alloc_sectors_done_inlined(struct bch_fs *c, struct write_point *wp)
157{
158 struct open_buckets ptrs = { .nr = 0 }, keep = { .nr = 0 };
159 struct open_bucket *ob;
160 unsigned i;
161
162 open_bucket_for_each(c, &wp->ptrs, ob, i)
163 ob_push(c, ob->sectors_free < block_sectors(c)
164 ? &ptrs
165 : &keep, ob);
166 wp->ptrs = keep;
167
168 mutex_unlock(&wp->lock);
169
170 bch2_open_buckets_put(c, &ptrs);
171}
172
173static inline void bch2_open_bucket_get(struct bch_fs *c,
174 struct write_point *wp,
175 struct open_buckets *ptrs)
176{
177 struct open_bucket *ob;
178 unsigned i;
179
180 open_bucket_for_each(c, &wp->ptrs, ob, i) {
181 ob->data_type = wp->data_type;
182 atomic_inc(&ob->pin);
183 ob_push(c, ptrs, ob);
184 }
185}
186
187static inline open_bucket_idx_t *open_bucket_hashslot(struct bch_fs *c,
188 unsigned dev, u64 bucket)
189{
190 return c->open_buckets_hash +
191 (jhash_3words(dev, bucket, bucket >> 32, 0) &
192 (OPEN_BUCKETS_COUNT - 1));
193}
194
195static inline bool bch2_bucket_is_open(struct bch_fs *c, unsigned dev, u64 bucket)
196{
197 open_bucket_idx_t slot = *open_bucket_hashslot(c, dev, bucket);
198
199 while (slot) {
200 struct open_bucket *ob = &c->open_buckets[slot];
201
202 if (ob->dev == dev && ob->bucket == bucket)
203 return true;
204
205 slot = ob->hash;
206 }
207
208 return false;
209}
210
211static inline bool bch2_bucket_is_open_safe(struct bch_fs *c, unsigned dev, u64 bucket)
212{
213 bool ret;
214
215 if (bch2_bucket_is_open(c, dev, bucket))
216 return true;
217
218 spin_lock(&c->freelist_lock);
219 ret = bch2_bucket_is_open(c, dev, bucket);
220 spin_unlock(&c->freelist_lock);
221
222 return ret;
223}
224
225enum bch_write_flags;
226int bch2_bucket_alloc_set_trans(struct btree_trans *, struct alloc_request *,
227 struct dev_stripe_state *, struct closure *);
228
229int bch2_alloc_sectors_start_trans(struct btree_trans *,
230 unsigned, unsigned,
231 struct write_point_specifier,
232 struct bch_devs_list *,
233 unsigned, unsigned,
234 enum bch_watermark,
235 enum bch_write_flags,
236 struct closure *,
237 struct write_point **);
238
239static inline struct bch_extent_ptr bch2_ob_ptr(struct bch_fs *c, struct open_bucket *ob)
240{
241 struct bch_dev *ca = ob_dev(c, ob);
242
243 return (struct bch_extent_ptr) {
244 .type = 1 << BCH_EXTENT_ENTRY_ptr,
245 .gen = ob->gen,
246 .dev = ob->dev,
247 .offset = bucket_to_sector(ca, ob->bucket) +
248 ca->mi.bucket_size -
249 ob->sectors_free,
250 };
251}
252
253/*
254 * Append pointers to the space we just allocated to @k, and mark @sectors space
255 * as allocated out of @ob
256 */
257static inline void
258bch2_alloc_sectors_append_ptrs_inlined(struct bch_fs *c, struct write_point *wp,
259 struct bkey_i *k, unsigned sectors,
260 bool cached)
261{
262 struct open_bucket *ob;
263 unsigned i;
264
265 BUG_ON(sectors > wp->sectors_free);
266 wp->sectors_free -= sectors;
267 wp->sectors_allocated += sectors;
268
269 open_bucket_for_each(c, &wp->ptrs, ob, i) {
270 struct bch_dev *ca = ob_dev(c, ob);
271 struct bch_extent_ptr ptr = bch2_ob_ptr(c, ob);
272
273 ptr.cached = cached ||
274 (!ca->mi.durability &&
275 wp->data_type == BCH_DATA_user);
276
277 bch2_bkey_append_ptr(k, ptr);
278
279 BUG_ON(sectors > ob->sectors_free);
280 ob->sectors_free -= sectors;
281 }
282}
283
284void bch2_alloc_sectors_append_ptrs(struct bch_fs *, struct write_point *,
285 struct bkey_i *, unsigned, bool);
286void bch2_alloc_sectors_done(struct bch_fs *, struct write_point *);
287
288void bch2_open_buckets_stop(struct bch_fs *c, struct bch_dev *, bool);
289
290static inline struct write_point_specifier writepoint_hashed(unsigned long v)
291{
292 return (struct write_point_specifier) { .v = v | 1 };
293}
294
295static inline struct write_point_specifier writepoint_ptr(struct write_point *wp)
296{
297 return (struct write_point_specifier) { .v = (unsigned long) wp };
298}
299
300void bch2_fs_allocator_foreground_init(struct bch_fs *);
301
302void bch2_open_bucket_to_text(struct printbuf *, struct bch_fs *, struct open_bucket *);
303void bch2_open_buckets_to_text(struct printbuf *, struct bch_fs *, struct bch_dev *);
304void bch2_open_buckets_partial_to_text(struct printbuf *, struct bch_fs *);
305
306void bch2_write_points_to_text(struct printbuf *, struct bch_fs *);
307
308void bch2_fs_alloc_debug_to_text(struct printbuf *, struct bch_fs *);
309void bch2_dev_alloc_debug_to_text(struct printbuf *, struct bch_dev *);
310
311void __bch2_wait_on_allocator(struct bch_fs *, struct closure *);
312static inline void bch2_wait_on_allocator(struct bch_fs *c, struct closure *cl)
313{
314 if (cl->closure_get_happened)
315 __bch2_wait_on_allocator(c, cl);
316}
317
318#endif /* _BCACHEFS_ALLOC_FOREGROUND_H */