Linux kernel mirror (for testing) git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel os linux
1
fork

Configure Feed

Select the types of activity you want to include in your feed.

at v6.7-rc3 228 lines 6.6 kB view raw
1/* SPDX-License-Identifier: GPL-2.0 */ 2#ifndef _BCACHEFS_BTREE_IO_H 3#define _BCACHEFS_BTREE_IO_H 4 5#include "bkey_methods.h" 6#include "bset.h" 7#include "btree_locking.h" 8#include "checksum.h" 9#include "extents.h" 10#include "io_write_types.h" 11 12struct bch_fs; 13struct btree_write; 14struct btree; 15struct btree_iter; 16struct btree_node_read_all; 17 18static inline void set_btree_node_dirty_acct(struct bch_fs *c, struct btree *b) 19{ 20 if (!test_and_set_bit(BTREE_NODE_dirty, &b->flags)) 21 atomic_inc(&c->btree_cache.dirty); 22} 23 24static inline void clear_btree_node_dirty_acct(struct bch_fs *c, struct btree *b) 25{ 26 if (test_and_clear_bit(BTREE_NODE_dirty, &b->flags)) 27 atomic_dec(&c->btree_cache.dirty); 28} 29 30static inline unsigned btree_ptr_sectors_written(struct bkey_i *k) 31{ 32 return k->k.type == KEY_TYPE_btree_ptr_v2 33 ? le16_to_cpu(bkey_i_to_btree_ptr_v2(k)->v.sectors_written) 34 : 0; 35} 36 37struct btree_read_bio { 38 struct bch_fs *c; 39 struct btree *b; 40 struct btree_node_read_all *ra; 41 u64 start_time; 42 unsigned have_ioref:1; 43 unsigned idx:7; 44 struct extent_ptr_decoded pick; 45 struct work_struct work; 46 struct bio bio; 47}; 48 49struct btree_write_bio { 50 struct work_struct work; 51 __BKEY_PADDED(key, BKEY_BTREE_PTR_VAL_U64s_MAX); 52 void *data; 53 unsigned data_bytes; 54 unsigned sector_offset; 55 struct bch_write_bio wbio; 56}; 57 58void bch2_btree_node_io_unlock(struct btree *); 59void bch2_btree_node_io_lock(struct btree *); 60void __bch2_btree_node_wait_on_read(struct btree *); 61void __bch2_btree_node_wait_on_write(struct btree *); 62void bch2_btree_node_wait_on_read(struct btree *); 63void bch2_btree_node_wait_on_write(struct btree *); 64 65enum compact_mode { 66 COMPACT_LAZY, 67 COMPACT_ALL, 68}; 69 70bool bch2_compact_whiteouts(struct bch_fs *, struct btree *, 71 enum compact_mode); 72 73static inline bool should_compact_bset_lazy(struct btree *b, 74 struct bset_tree *t) 75{ 76 unsigned total_u64s = bset_u64s(t); 77 unsigned dead_u64s = bset_dead_u64s(b, t); 78 79 return dead_u64s > 64 && dead_u64s * 3 > total_u64s; 80} 81 82static inline bool bch2_maybe_compact_whiteouts(struct bch_fs *c, struct btree *b) 83{ 84 struct bset_tree *t; 85 86 for_each_bset(b, t) 87 if (should_compact_bset_lazy(b, t)) 88 return bch2_compact_whiteouts(c, b, COMPACT_LAZY); 89 90 return false; 91} 92 93static inline struct nonce btree_nonce(struct bset *i, unsigned offset) 94{ 95 return (struct nonce) {{ 96 [0] = cpu_to_le32(offset), 97 [1] = ((__le32 *) &i->seq)[0], 98 [2] = ((__le32 *) &i->seq)[1], 99 [3] = ((__le32 *) &i->journal_seq)[0]^BCH_NONCE_BTREE, 100 }}; 101} 102 103static inline int bset_encrypt(struct bch_fs *c, struct bset *i, unsigned offset) 104{ 105 struct nonce nonce = btree_nonce(i, offset); 106 int ret; 107 108 if (!offset) { 109 struct btree_node *bn = container_of(i, struct btree_node, keys); 110 unsigned bytes = (void *) &bn->keys - (void *) &bn->flags; 111 112 ret = bch2_encrypt(c, BSET_CSUM_TYPE(i), nonce, 113 &bn->flags, bytes); 114 if (ret) 115 return ret; 116 117 nonce = nonce_add(nonce, round_up(bytes, CHACHA_BLOCK_SIZE)); 118 } 119 120 return bch2_encrypt(c, BSET_CSUM_TYPE(i), nonce, i->_data, 121 vstruct_end(i) - (void *) i->_data); 122} 123 124void bch2_btree_sort_into(struct bch_fs *, struct btree *, struct btree *); 125 126void bch2_btree_node_drop_keys_outside_node(struct btree *); 127 128void bch2_btree_build_aux_trees(struct btree *); 129void bch2_btree_init_next(struct btree_trans *, struct btree *); 130 131int bch2_btree_node_read_done(struct bch_fs *, struct bch_dev *, 132 struct btree *, bool, bool *); 133void bch2_btree_node_read(struct bch_fs *, struct btree *, bool); 134int bch2_btree_root_read(struct bch_fs *, enum btree_id, 135 const struct bkey_i *, unsigned); 136 137void bch2_btree_complete_write(struct bch_fs *, struct btree *, 138 struct btree_write *); 139 140bool bch2_btree_post_write_cleanup(struct bch_fs *, struct btree *); 141 142enum btree_write_flags { 143 __BTREE_WRITE_ONLY_IF_NEED = BTREE_WRITE_TYPE_BITS, 144 __BTREE_WRITE_ALREADY_STARTED, 145}; 146#define BTREE_WRITE_ONLY_IF_NEED BIT(__BTREE_WRITE_ONLY_IF_NEED) 147#define BTREE_WRITE_ALREADY_STARTED BIT(__BTREE_WRITE_ALREADY_STARTED) 148 149void __bch2_btree_node_write(struct bch_fs *, struct btree *, unsigned); 150void bch2_btree_node_write(struct bch_fs *, struct btree *, 151 enum six_lock_type, unsigned); 152 153static inline void btree_node_write_if_need(struct bch_fs *c, struct btree *b, 154 enum six_lock_type lock_held) 155{ 156 bch2_btree_node_write(c, b, lock_held, BTREE_WRITE_ONLY_IF_NEED); 157} 158 159bool bch2_btree_flush_all_reads(struct bch_fs *); 160bool bch2_btree_flush_all_writes(struct bch_fs *); 161 162static inline void compat_bformat(unsigned level, enum btree_id btree_id, 163 unsigned version, unsigned big_endian, 164 int write, struct bkey_format *f) 165{ 166 if (version < bcachefs_metadata_version_inode_btree_change && 167 btree_id == BTREE_ID_inodes) { 168 swap(f->bits_per_field[BKEY_FIELD_INODE], 169 f->bits_per_field[BKEY_FIELD_OFFSET]); 170 swap(f->field_offset[BKEY_FIELD_INODE], 171 f->field_offset[BKEY_FIELD_OFFSET]); 172 } 173 174 if (version < bcachefs_metadata_version_snapshot && 175 (level || btree_type_has_snapshots(btree_id))) { 176 u64 max_packed = 177 ~(~0ULL << f->bits_per_field[BKEY_FIELD_SNAPSHOT]); 178 179 f->field_offset[BKEY_FIELD_SNAPSHOT] = write 180 ? 0 181 : cpu_to_le64(U32_MAX - max_packed); 182 } 183} 184 185static inline void compat_bpos(unsigned level, enum btree_id btree_id, 186 unsigned version, unsigned big_endian, 187 int write, struct bpos *p) 188{ 189 if (big_endian != CPU_BIG_ENDIAN) 190 bch2_bpos_swab(p); 191 192 if (version < bcachefs_metadata_version_inode_btree_change && 193 btree_id == BTREE_ID_inodes) 194 swap(p->inode, p->offset); 195} 196 197static inline void compat_btree_node(unsigned level, enum btree_id btree_id, 198 unsigned version, unsigned big_endian, 199 int write, 200 struct btree_node *bn) 201{ 202 if (version < bcachefs_metadata_version_inode_btree_change && 203 btree_id_is_extents(btree_id) && 204 !bpos_eq(bn->min_key, POS_MIN) && 205 write) 206 bn->min_key = bpos_nosnap_predecessor(bn->min_key); 207 208 if (version < bcachefs_metadata_version_snapshot && 209 write) 210 bn->max_key.snapshot = 0; 211 212 compat_bpos(level, btree_id, version, big_endian, write, &bn->min_key); 213 compat_bpos(level, btree_id, version, big_endian, write, &bn->max_key); 214 215 if (version < bcachefs_metadata_version_snapshot && 216 !write) 217 bn->max_key.snapshot = U32_MAX; 218 219 if (version < bcachefs_metadata_version_inode_btree_change && 220 btree_id_is_extents(btree_id) && 221 !bpos_eq(bn->min_key, POS_MIN) && 222 !write) 223 bn->min_key = bpos_nosnap_successor(bn->min_key); 224} 225 226void bch2_btree_write_stats_to_text(struct printbuf *, struct bch_fs *); 227 228#endif /* _BCACHEFS_BTREE_IO_H */