at for-next 7.1 kB view raw
1/* SPDX-License-Identifier: GPL-2.0 */ 2#ifndef _BCACHEFS_EC_H 3#define _BCACHEFS_EC_H 4 5#include "ec_types.h" 6#include "buckets_types.h" 7#include "extents_types.h" 8 9enum bch_validate_flags; 10 11int bch2_stripe_validate(struct bch_fs *, struct bkey_s_c, enum bch_validate_flags); 12void bch2_stripe_to_text(struct printbuf *, struct bch_fs *, 13 struct bkey_s_c); 14int bch2_trigger_stripe(struct btree_trans *, enum btree_id, unsigned, 15 struct bkey_s_c, struct bkey_s, 16 enum btree_iter_update_trigger_flags); 17 18#define bch2_bkey_ops_stripe ((struct bkey_ops) { \ 19 .key_validate = bch2_stripe_validate, \ 20 .val_to_text = bch2_stripe_to_text, \ 21 .swab = bch2_ptr_swab, \ 22 .trigger = bch2_trigger_stripe, \ 23 .min_val_size = 8, \ 24}) 25 26static inline unsigned stripe_csums_per_device(const struct bch_stripe *s) 27{ 28 return DIV_ROUND_UP(le16_to_cpu(s->sectors), 29 1 << s->csum_granularity_bits); 30} 31 32static inline unsigned stripe_csum_offset(const struct bch_stripe *s, 33 unsigned dev, unsigned csum_idx) 34{ 35 EBUG_ON(s->csum_type >= BCH_CSUM_NR); 36 37 unsigned csum_bytes = bch_crc_bytes[s->csum_type]; 38 39 return sizeof(struct bch_stripe) + 40 sizeof(struct bch_extent_ptr) * s->nr_blocks + 41 (dev * stripe_csums_per_device(s) + csum_idx) * csum_bytes; 42} 43 44static inline unsigned stripe_blockcount_offset(const struct bch_stripe *s, 45 unsigned idx) 46{ 47 return stripe_csum_offset(s, s->nr_blocks, 0) + 48 sizeof(u16) * idx; 49} 50 51static inline unsigned stripe_blockcount_get(const struct bch_stripe *s, 52 unsigned idx) 53{ 54 return le16_to_cpup((void *) s + stripe_blockcount_offset(s, idx)); 55} 56 57static inline void stripe_blockcount_set(struct bch_stripe *s, 58 unsigned idx, unsigned v) 59{ 60 __le16 *p = (void *) s + stripe_blockcount_offset(s, idx); 61 62 *p = cpu_to_le16(v); 63} 64 65static inline unsigned stripe_val_u64s(const struct bch_stripe *s) 66{ 67 return DIV_ROUND_UP(stripe_blockcount_offset(s, s->nr_blocks), 68 sizeof(u64)); 69} 70 71static inline void *stripe_csum(struct bch_stripe *s, 72 unsigned block, unsigned csum_idx) 73{ 74 EBUG_ON(block >= s->nr_blocks); 75 EBUG_ON(csum_idx >= stripe_csums_per_device(s)); 76 77 return (void *) s + stripe_csum_offset(s, block, csum_idx); 78} 79 80static inline struct bch_csum stripe_csum_get(struct bch_stripe *s, 81 unsigned block, unsigned csum_idx) 82{ 83 struct bch_csum csum = { 0 }; 84 85 memcpy(&csum, stripe_csum(s, block, csum_idx), bch_crc_bytes[s->csum_type]); 86 return csum; 87} 88 89static inline void stripe_csum_set(struct bch_stripe *s, 90 unsigned block, unsigned csum_idx, 91 struct bch_csum csum) 92{ 93 memcpy(stripe_csum(s, block, csum_idx), &csum, bch_crc_bytes[s->csum_type]); 94} 95 96static inline bool __bch2_ptr_matches_stripe(const struct bch_extent_ptr *stripe_ptr, 97 const struct bch_extent_ptr *data_ptr, 98 unsigned sectors) 99{ 100 return (data_ptr->dev == stripe_ptr->dev || 101 data_ptr->dev == BCH_SB_MEMBER_INVALID || 102 stripe_ptr->dev == BCH_SB_MEMBER_INVALID) && 103 data_ptr->gen == stripe_ptr->gen && 104 data_ptr->offset >= stripe_ptr->offset && 105 data_ptr->offset < stripe_ptr->offset + sectors; 106} 107 108static inline bool bch2_ptr_matches_stripe(const struct bch_stripe *s, 109 struct extent_ptr_decoded p) 110{ 111 unsigned nr_data = s->nr_blocks - s->nr_redundant; 112 113 BUG_ON(!p.has_ec); 114 115 if (p.ec.block >= nr_data) 116 return false; 117 118 return __bch2_ptr_matches_stripe(&s->ptrs[p.ec.block], &p.ptr, 119 le16_to_cpu(s->sectors)); 120} 121 122static inline bool bch2_ptr_matches_stripe_m(const struct gc_stripe *m, 123 struct extent_ptr_decoded p) 124{ 125 unsigned nr_data = m->nr_blocks - m->nr_redundant; 126 127 BUG_ON(!p.has_ec); 128 129 if (p.ec.block >= nr_data) 130 return false; 131 132 return __bch2_ptr_matches_stripe(&m->ptrs[p.ec.block], &p.ptr, 133 m->sectors); 134} 135 136struct bch_read_bio; 137 138struct ec_stripe_buf { 139 /* might not be buffering the entire stripe: */ 140 unsigned offset; 141 unsigned size; 142 unsigned long valid[BITS_TO_LONGS(BCH_BKEY_PTRS_MAX)]; 143 144 void *data[BCH_BKEY_PTRS_MAX]; 145 146 __BKEY_PADDED(key, 255); 147}; 148 149struct ec_stripe_head; 150 151enum ec_stripe_ref { 152 STRIPE_REF_io, 153 STRIPE_REF_stripe, 154 STRIPE_REF_NR 155}; 156 157struct ec_stripe_new { 158 struct bch_fs *c; 159 struct ec_stripe_head *h; 160 struct mutex lock; 161 struct list_head list; 162 163 struct hlist_node hash; 164 u64 idx; 165 166 struct closure iodone; 167 168 atomic_t ref[STRIPE_REF_NR]; 169 170 int err; 171 172 u8 nr_data; 173 u8 nr_parity; 174 bool allocated; 175 bool pending; 176 bool have_existing_stripe; 177 178 unsigned long blocks_gotten[BITS_TO_LONGS(BCH_BKEY_PTRS_MAX)]; 179 unsigned long blocks_allocated[BITS_TO_LONGS(BCH_BKEY_PTRS_MAX)]; 180 open_bucket_idx_t blocks[BCH_BKEY_PTRS_MAX]; 181 struct disk_reservation res; 182 183 struct ec_stripe_buf new_stripe; 184 struct ec_stripe_buf existing_stripe; 185}; 186 187struct ec_stripe_head { 188 struct list_head list; 189 struct mutex lock; 190 191 unsigned disk_label; 192 unsigned algo; 193 unsigned redundancy; 194 enum bch_watermark watermark; 195 bool insufficient_devs; 196 197 unsigned long rw_devs_change_count; 198 199 u64 nr_created; 200 201 struct bch_devs_mask devs; 202 unsigned nr_active_devs; 203 204 unsigned blocksize; 205 206 struct dev_stripe_state block_stripe; 207 struct dev_stripe_state parity_stripe; 208 209 struct ec_stripe_new *s; 210}; 211 212int bch2_ec_read_extent(struct btree_trans *, struct bch_read_bio *, struct bkey_s_c); 213 214void *bch2_writepoint_ec_buf(struct bch_fs *, struct write_point *); 215 216void bch2_ec_bucket_cancel(struct bch_fs *, struct open_bucket *); 217 218int bch2_ec_stripe_new_alloc(struct bch_fs *, struct ec_stripe_head *); 219 220void bch2_ec_stripe_head_put(struct bch_fs *, struct ec_stripe_head *); 221struct ec_stripe_head *bch2_ec_stripe_head_get(struct btree_trans *, 222 unsigned, unsigned, unsigned, 223 enum bch_watermark, struct closure *); 224 225void bch2_stripes_heap_update(struct bch_fs *, struct stripe *, size_t); 226void bch2_stripes_heap_del(struct bch_fs *, struct stripe *, size_t); 227void bch2_stripes_heap_insert(struct bch_fs *, struct stripe *, size_t); 228 229void bch2_do_stripe_deletes(struct bch_fs *); 230void bch2_ec_do_stripe_creates(struct bch_fs *); 231void bch2_ec_stripe_new_free(struct bch_fs *, struct ec_stripe_new *); 232 233static inline void ec_stripe_new_get(struct ec_stripe_new *s, 234 enum ec_stripe_ref ref) 235{ 236 atomic_inc(&s->ref[ref]); 237} 238 239static inline void ec_stripe_new_put(struct bch_fs *c, struct ec_stripe_new *s, 240 enum ec_stripe_ref ref) 241{ 242 BUG_ON(atomic_read(&s->ref[ref]) <= 0); 243 244 if (atomic_dec_and_test(&s->ref[ref])) 245 switch (ref) { 246 case STRIPE_REF_stripe: 247 bch2_ec_stripe_new_free(c, s); 248 break; 249 case STRIPE_REF_io: 250 bch2_ec_do_stripe_creates(c); 251 break; 252 default: 253 BUG(); 254 } 255} 256 257int bch2_dev_remove_stripes(struct bch_fs *, unsigned); 258 259void bch2_ec_stop_dev(struct bch_fs *, struct bch_dev *); 260void bch2_fs_ec_stop(struct bch_fs *); 261void bch2_fs_ec_flush(struct bch_fs *); 262 263int bch2_stripes_read(struct bch_fs *); 264 265void bch2_stripes_heap_to_text(struct printbuf *, struct bch_fs *); 266void bch2_new_stripes_to_text(struct printbuf *, struct bch_fs *); 267 268void bch2_fs_ec_exit(struct bch_fs *); 269void bch2_fs_ec_init_early(struct bch_fs *); 270int bch2_fs_ec_init(struct bch_fs *); 271 272#endif /* _BCACHEFS_EC_H */