Book a Demo!
CoCalc Logo Icon
StoreFeaturesDocsShareSupportNewsAboutPoliciesSign UpSign In
torvalds
GitHub Repository: torvalds/linux
Path: blob/master/fs/bcachefs/backpointers.h
26282 views
1
/* SPDX-License-Identifier: GPL-2.0 */
2
#ifndef _BCACHEFS_BACKPOINTERS_H
3
#define _BCACHEFS_BACKPOINTERS_H
4
5
#include "btree_cache.h"
6
#include "btree_iter.h"
7
#include "btree_update.h"
8
#include "buckets.h"
9
#include "error.h"
10
#include "super.h"
11
12
static inline u64 swab40(u64 x)
13
{
14
return (((x & 0x00000000ffULL) << 32)|
15
((x & 0x000000ff00ULL) << 16)|
16
((x & 0x0000ff0000ULL) >> 0)|
17
((x & 0x00ff000000ULL) >> 16)|
18
((x & 0xff00000000ULL) >> 32));
19
}
20
21
int bch2_backpointer_validate(struct bch_fs *, struct bkey_s_c k,
22
struct bkey_validate_context);
23
void bch2_backpointer_to_text(struct printbuf *, struct bch_fs *, struct bkey_s_c);
24
void bch2_backpointer_swab(struct bkey_s);
25
26
#define bch2_bkey_ops_backpointer ((struct bkey_ops) { \
27
.key_validate = bch2_backpointer_validate, \
28
.val_to_text = bch2_backpointer_to_text, \
29
.swab = bch2_backpointer_swab, \
30
.min_val_size = 32, \
31
})
32
33
#define MAX_EXTENT_COMPRESS_RATIO_SHIFT 10
34
35
/*
36
* Convert from pos in backpointer btree to pos of corresponding bucket in alloc
37
* btree:
38
*/
39
static inline struct bpos bp_pos_to_bucket(const struct bch_dev *ca, struct bpos bp_pos)
40
{
41
u64 bucket_sector = bp_pos.offset >> MAX_EXTENT_COMPRESS_RATIO_SHIFT;
42
43
return POS(bp_pos.inode, sector_to_bucket(ca, bucket_sector));
44
}
45
46
static inline struct bpos bp_pos_to_bucket_and_offset(const struct bch_dev *ca, struct bpos bp_pos,
47
u32 *bucket_offset)
48
{
49
u64 bucket_sector = bp_pos.offset >> MAX_EXTENT_COMPRESS_RATIO_SHIFT;
50
51
return POS(bp_pos.inode, sector_to_bucket_and_offset(ca, bucket_sector, bucket_offset));
52
}
53
54
static inline bool bp_pos_to_bucket_nodev_noerror(struct bch_fs *c, struct bpos bp_pos, struct bpos *bucket)
55
{
56
guard(rcu)();
57
struct bch_dev *ca = bch2_dev_rcu_noerror(c, bp_pos.inode);
58
if (ca)
59
*bucket = bp_pos_to_bucket(ca, bp_pos);
60
return ca != NULL;
61
}
62
63
static inline struct bpos bucket_pos_to_bp_noerror(const struct bch_dev *ca,
64
struct bpos bucket,
65
u64 bucket_offset)
66
{
67
return POS(bucket.inode,
68
(bucket_to_sector(ca, bucket.offset) <<
69
MAX_EXTENT_COMPRESS_RATIO_SHIFT) + bucket_offset);
70
}
71
72
/*
73
* Convert from pos in alloc btree + bucket offset to pos in backpointer btree:
74
*/
75
static inline struct bpos bucket_pos_to_bp(const struct bch_dev *ca,
76
struct bpos bucket,
77
u64 bucket_offset)
78
{
79
struct bpos ret = bucket_pos_to_bp_noerror(ca, bucket, bucket_offset);
80
EBUG_ON(!bkey_eq(bucket, bp_pos_to_bucket(ca, ret)));
81
return ret;
82
}
83
84
static inline struct bpos bucket_pos_to_bp_start(const struct bch_dev *ca, struct bpos bucket)
85
{
86
return bucket_pos_to_bp(ca, bucket, 0);
87
}
88
89
static inline struct bpos bucket_pos_to_bp_end(const struct bch_dev *ca, struct bpos bucket)
90
{
91
return bpos_nosnap_predecessor(bucket_pos_to_bp(ca, bpos_nosnap_successor(bucket), 0));
92
}
93
94
int bch2_bucket_backpointer_mod_nowritebuffer(struct btree_trans *,
95
struct bkey_s_c,
96
struct bkey_i_backpointer *,
97
bool);
98
99
static inline int bch2_bucket_backpointer_mod(struct btree_trans *trans,
100
struct bkey_s_c orig_k,
101
struct bkey_i_backpointer *bp,
102
bool insert)
103
{
104
if (static_branch_unlikely(&bch2_backpointers_no_use_write_buffer))
105
return bch2_bucket_backpointer_mod_nowritebuffer(trans, orig_k, bp, insert);
106
107
if (!insert) {
108
bp->k.type = KEY_TYPE_deleted;
109
set_bkey_val_u64s(&bp->k, 0);
110
}
111
112
return bch2_trans_update_buffered(trans, BTREE_ID_backpointers, &bp->k_i);
113
}
114
115
static inline enum bch_data_type bch2_bkey_ptr_data_type(struct bkey_s_c k,
116
struct extent_ptr_decoded p,
117
const union bch_extent_entry *entry)
118
{
119
switch (k.k->type) {
120
case KEY_TYPE_btree_ptr:
121
case KEY_TYPE_btree_ptr_v2:
122
return BCH_DATA_btree;
123
case KEY_TYPE_extent:
124
case KEY_TYPE_reflink_v:
125
if (p.has_ec)
126
return BCH_DATA_stripe;
127
if (p.ptr.cached)
128
return BCH_DATA_cached;
129
else
130
return BCH_DATA_user;
131
case KEY_TYPE_stripe: {
132
const struct bch_extent_ptr *ptr = &entry->ptr;
133
struct bkey_s_c_stripe s = bkey_s_c_to_stripe(k);
134
135
BUG_ON(ptr < s.v->ptrs ||
136
ptr >= s.v->ptrs + s.v->nr_blocks);
137
138
return ptr >= s.v->ptrs + s.v->nr_blocks - s.v->nr_redundant
139
? BCH_DATA_parity
140
: BCH_DATA_user;
141
}
142
default:
143
BUG();
144
}
145
}
146
147
static inline void bch2_extent_ptr_to_bp(struct bch_fs *c,
148
enum btree_id btree_id, unsigned level,
149
struct bkey_s_c k, struct extent_ptr_decoded p,
150
const union bch_extent_entry *entry,
151
struct bkey_i_backpointer *bp)
152
{
153
bkey_backpointer_init(&bp->k_i);
154
bp->k.p.inode = p.ptr.dev;
155
156
if (k.k->type != KEY_TYPE_stripe)
157
bp->k.p.offset = ((u64) p.ptr.offset << MAX_EXTENT_COMPRESS_RATIO_SHIFT) + p.crc.offset;
158
else {
159
/*
160
* Put stripe backpointers where they won't collide with the
161
* extent backpointers within the stripe:
162
*/
163
struct bkey_s_c_stripe s = bkey_s_c_to_stripe(k);
164
bp->k.p.offset = ((u64) (p.ptr.offset + le16_to_cpu(s.v->sectors)) <<
165
MAX_EXTENT_COMPRESS_RATIO_SHIFT) - 1;
166
}
167
168
bp->v = (struct bch_backpointer) {
169
.btree_id = btree_id,
170
.level = level,
171
.data_type = bch2_bkey_ptr_data_type(k, p, entry),
172
.bucket_gen = p.ptr.gen,
173
.bucket_len = ptr_disk_sectors(level ? btree_sectors(c) : k.k->size, p),
174
.pos = k.k->p,
175
};
176
}
177
178
struct bkey_buf;
179
struct bkey_s_c bch2_backpointer_get_key(struct btree_trans *, struct bkey_s_c_backpointer,
180
struct btree_iter *, unsigned, struct bkey_buf *);
181
struct btree *bch2_backpointer_get_node(struct btree_trans *, struct bkey_s_c_backpointer,
182
struct btree_iter *, struct bkey_buf *);
183
184
int bch2_check_bucket_backpointer_mismatch(struct btree_trans *, struct bch_dev *, u64,
185
bool, struct bkey_buf *);
186
187
int bch2_check_btree_backpointers(struct bch_fs *);
188
int bch2_check_extents_to_backpointers(struct bch_fs *);
189
int bch2_check_backpointers_to_extents(struct bch_fs *);
190
191
static inline bool bch2_bucket_bitmap_test(struct bucket_bitmap *b, u64 i)
192
{
193
unsigned long *bitmap = READ_ONCE(b->buckets);
194
return bitmap && test_bit(i, bitmap);
195
}
196
197
int bch2_bucket_bitmap_resize(struct bch_dev *, struct bucket_bitmap *, u64, u64);
198
void bch2_bucket_bitmap_free(struct bucket_bitmap *);
199
200
#endif /* _BCACHEFS_BACKPOINTERS_BACKGROUND_H */
201
202