Commit cf49f8a8 authored by Kent Overstreet's avatar Kent Overstreet

bcachefs: rename version -> bversion

give bversions a more distinct name, to aid in grepping
Signed-off-by: default avatarKent Overstreet <kent.overstreet@linux.dev>
parent fd65378d
...@@ -501,7 +501,7 @@ static int check_extent_checksum(struct btree_trans *trans, ...@@ -501,7 +501,7 @@ static int check_extent_checksum(struct btree_trans *trans,
prt_printf(&buf, "\n %s ", bch2_btree_id_str(o_btree)); prt_printf(&buf, "\n %s ", bch2_btree_id_str(o_btree));
bch2_bkey_val_to_text(&buf, c, extent2); bch2_bkey_val_to_text(&buf, c, extent2);
struct nonce nonce = extent_nonce(extent.k->version, p.crc); struct nonce nonce = extent_nonce(extent.k->bversion, p.crc);
struct bch_csum csum = bch2_checksum(c, p.crc.csum_type, nonce, data_buf, bytes); struct bch_csum csum = bch2_checksum(c, p.crc.csum_type, nonce, data_buf, bytes);
if (fsck_err_on(bch2_crc_cmp(csum, p.crc.csum), if (fsck_err_on(bch2_crc_cmp(csum, p.crc.csum),
trans, dup_backpointer_to_bad_csum_extent, trans, dup_backpointer_to_bad_csum_extent,
......
...@@ -217,7 +217,7 @@ struct bkey { ...@@ -217,7 +217,7 @@ struct bkey {
#if __BYTE_ORDER__ == __ORDER_LITTLE_ENDIAN__ #if __BYTE_ORDER__ == __ORDER_LITTLE_ENDIAN__
__u8 pad[1]; __u8 pad[1];
struct bversion version; struct bversion bversion;
__u32 size; /* extent size, in sectors */ __u32 size; /* extent size, in sectors */
struct bpos p; struct bpos p;
#elif __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__ #elif __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
...@@ -328,8 +328,8 @@ enum bch_bkey_fields { ...@@ -328,8 +328,8 @@ enum bch_bkey_fields {
bkey_format_field(OFFSET, p.offset), \ bkey_format_field(OFFSET, p.offset), \
bkey_format_field(SNAPSHOT, p.snapshot), \ bkey_format_field(SNAPSHOT, p.snapshot), \
bkey_format_field(SIZE, size), \ bkey_format_field(SIZE, size), \
bkey_format_field(VERSION_HI, version.hi), \ bkey_format_field(VERSION_HI, bversion.hi), \
bkey_format_field(VERSION_LO, version.lo), \ bkey_format_field(VERSION_LO, bversion.lo), \
}, \ }, \
}) })
......
...@@ -554,8 +554,8 @@ static inline void bch2_bkey_pack_test(void) {} ...@@ -554,8 +554,8 @@ static inline void bch2_bkey_pack_test(void) {}
x(BKEY_FIELD_OFFSET, p.offset) \ x(BKEY_FIELD_OFFSET, p.offset) \
x(BKEY_FIELD_SNAPSHOT, p.snapshot) \ x(BKEY_FIELD_SNAPSHOT, p.snapshot) \
x(BKEY_FIELD_SIZE, size) \ x(BKEY_FIELD_SIZE, size) \
x(BKEY_FIELD_VERSION_HI, version.hi) \ x(BKEY_FIELD_VERSION_HI, bversion.hi) \
x(BKEY_FIELD_VERSION_LO, version.lo) x(BKEY_FIELD_VERSION_LO, bversion.lo)
struct bkey_format_state { struct bkey_format_state {
u64 field_min[BKEY_NR_FIELDS]; u64 field_min[BKEY_NR_FIELDS];
......
...@@ -289,7 +289,7 @@ void bch2_bkey_to_text(struct printbuf *out, const struct bkey *k) ...@@ -289,7 +289,7 @@ void bch2_bkey_to_text(struct printbuf *out, const struct bkey *k)
bch2_bpos_to_text(out, k->p); bch2_bpos_to_text(out, k->p);
prt_printf(out, " len %u ver %llu", k->size, k->version.lo); prt_printf(out, " len %u ver %llu", k->size, k->bversion.lo);
} else { } else {
prt_printf(out, "(null)"); prt_printf(out, "(null)");
} }
......
...@@ -70,7 +70,7 @@ bool bch2_bkey_normalize(struct bch_fs *, struct bkey_s); ...@@ -70,7 +70,7 @@ bool bch2_bkey_normalize(struct bch_fs *, struct bkey_s);
static inline bool bch2_bkey_maybe_mergable(const struct bkey *l, const struct bkey *r) static inline bool bch2_bkey_maybe_mergable(const struct bkey *l, const struct bkey *r)
{ {
return l->type == r->type && return l->type == r->type &&
!bversion_cmp(l->version, r->version) && !bversion_cmp(l->bversion, r->bversion) &&
bpos_eq(l->p, bkey_start_pos(r)); bpos_eq(l->p, bkey_start_pos(r));
} }
......
...@@ -601,15 +601,15 @@ static int bch2_gc_mark_key(struct btree_trans *trans, enum btree_id btree_id, ...@@ -601,15 +601,15 @@ static int bch2_gc_mark_key(struct btree_trans *trans, enum btree_id btree_id,
if (initial) { if (initial) {
BUG_ON(bch2_journal_seq_verify && BUG_ON(bch2_journal_seq_verify &&
k.k->version.lo > atomic64_read(&c->journal.seq)); k.k->bversion.lo > atomic64_read(&c->journal.seq));
if (fsck_err_on(btree_id != BTREE_ID_accounting && if (fsck_err_on(btree_id != BTREE_ID_accounting &&
k.k->version.lo > atomic64_read(&c->key_version), k.k->bversion.lo > atomic64_read(&c->key_version),
trans, bkey_version_in_future, trans, bkey_version_in_future,
"key version number higher than recorded %llu\n %s", "key version number higher than recorded %llu\n %s",
atomic64_read(&c->key_version), atomic64_read(&c->key_version),
(bch2_bkey_val_to_text(&buf, c, k), buf.buf))) (bch2_bkey_val_to_text(&buf, c, k), buf.buf)))
atomic64_set(&c->key_version, k.k->version.lo); atomic64_set(&c->key_version, k.k->bversion.lo);
} }
if (mustfix_fsck_err_on(level && !bch2_dev_btree_bitmap_marked(c, k), if (mustfix_fsck_err_on(level && !bch2_dev_btree_bitmap_marked(c, k),
......
...@@ -1223,7 +1223,7 @@ int bch2_btree_node_read_done(struct bch_fs *c, struct bch_dev *ca, ...@@ -1223,7 +1223,7 @@ int bch2_btree_node_read_done(struct bch_fs *c, struct bch_dev *ca,
ret = bch2_bkey_val_validate(c, u.s_c, READ); ret = bch2_bkey_val_validate(c, u.s_c, READ);
if (ret == -BCH_ERR_fsck_delete_bkey || if (ret == -BCH_ERR_fsck_delete_bkey ||
(bch2_inject_invalid_keys && (bch2_inject_invalid_keys &&
!bversion_cmp(u.k->version, MAX_VERSION))) { !bversion_cmp(u.k->bversion, MAX_VERSION))) {
btree_keys_account_key_drop(&b->nr, 0, k); btree_keys_account_key_drop(&b->nr, 0, k);
i->u64s = cpu_to_le16(le16_to_cpu(i->u64s) - k->u64s); i->u64s = cpu_to_le16(le16_to_cpu(i->u64s) - k->u64s);
......
...@@ -684,10 +684,10 @@ bch2_trans_commit_write_locked(struct btree_trans *trans, unsigned flags, ...@@ -684,10 +684,10 @@ bch2_trans_commit_write_locked(struct btree_trans *trans, unsigned flags,
!(flags & BCH_TRANS_COMMIT_no_journal_res)) { !(flags & BCH_TRANS_COMMIT_no_journal_res)) {
if (bch2_journal_seq_verify) if (bch2_journal_seq_verify)
trans_for_each_update(trans, i) trans_for_each_update(trans, i)
i->k->k.version.lo = trans->journal_res.seq; i->k->k.bversion.lo = trans->journal_res.seq;
else if (bch2_inject_invalid_keys) else if (bch2_inject_invalid_keys)
trans_for_each_update(trans, i) trans_for_each_update(trans, i)
i->k->k.version = MAX_VERSION; i->k->k.bversion = MAX_VERSION;
} }
h = trans->hooks; h = trans->hooks;
...@@ -709,9 +709,9 @@ bch2_trans_commit_write_locked(struct btree_trans *trans, unsigned flags, ...@@ -709,9 +709,9 @@ bch2_trans_commit_write_locked(struct btree_trans *trans, unsigned flags,
if (jset_entry_is_key(entry) && entry->start->k.type == KEY_TYPE_accounting) { if (jset_entry_is_key(entry) && entry->start->k.type == KEY_TYPE_accounting) {
struct bkey_i_accounting *a = bkey_i_to_accounting(entry->start); struct bkey_i_accounting *a = bkey_i_to_accounting(entry->start);
a->k.version = journal_pos_to_bversion(&trans->journal_res, a->k.bversion = journal_pos_to_bversion(&trans->journal_res,
(u64 *) entry - (u64 *) trans->journal_entries); (u64 *) entry - (u64 *) trans->journal_entries);
BUG_ON(bversion_zero(a->k.version)); BUG_ON(bversion_zero(a->k.bversion));
ret = bch2_accounting_mem_mod_locked(trans, accounting_i_to_s_c(a), BCH_ACCOUNTING_normal); ret = bch2_accounting_mem_mod_locked(trans, accounting_i_to_s_c(a), BCH_ACCOUNTING_normal);
if (ret) if (ret)
goto revert_fs_usage; goto revert_fs_usage;
......
...@@ -639,7 +639,7 @@ int bch2_data_update_init(struct btree_trans *trans, ...@@ -639,7 +639,7 @@ int bch2_data_update_init(struct btree_trans *trans,
bch2_write_op_init(&m->op, c, io_opts); bch2_write_op_init(&m->op, c, io_opts);
m->op.pos = bkey_start_pos(k.k); m->op.pos = bkey_start_pos(k.k);
m->op.version = k.k->version; m->op.version = k.k->bversion;
m->op.target = data_opts.target; m->op.target = data_opts.target;
m->op.write_point = wp; m->op.write_point = wp;
m->op.nr_replicas = 0; m->op.nr_replicas = 0;
......
...@@ -291,7 +291,7 @@ static int __bch2_accounting_mem_insert(struct bch_fs *c, struct bkey_s_c_accoun ...@@ -291,7 +291,7 @@ static int __bch2_accounting_mem_insert(struct bch_fs *c, struct bkey_s_c_accoun
struct accounting_mem_entry n = { struct accounting_mem_entry n = {
.pos = a.k->p, .pos = a.k->p,
.version = a.k->version, .bversion = a.k->bversion,
.nr_counters = bch2_accounting_counters(a.k), .nr_counters = bch2_accounting_counters(a.k),
.v[0] = __alloc_percpu_gfp(n.nr_counters * sizeof(u64), .v[0] = __alloc_percpu_gfp(n.nr_counters * sizeof(u64),
sizeof(u64), GFP_KERNEL), sizeof(u64), GFP_KERNEL),
...@@ -636,7 +636,7 @@ int bch2_accounting_read(struct bch_fs *c) ...@@ -636,7 +636,7 @@ int bch2_accounting_read(struct bch_fs *c)
accounting_pos_cmp, &k.k->p); accounting_pos_cmp, &k.k->p);
bool applied = idx < acc->k.nr && bool applied = idx < acc->k.nr &&
bversion_cmp(acc->k.data[idx].version, k.k->version) >= 0; bversion_cmp(acc->k.data[idx].bversion, k.k->bversion) >= 0;
if (applied) if (applied)
continue; continue;
...@@ -644,7 +644,7 @@ int bch2_accounting_read(struct bch_fs *c) ...@@ -644,7 +644,7 @@ int bch2_accounting_read(struct bch_fs *c)
if (i + 1 < &darray_top(*keys) && if (i + 1 < &darray_top(*keys) &&
i[1].k->k.type == KEY_TYPE_accounting && i[1].k->k.type == KEY_TYPE_accounting &&
!journal_key_cmp(i, i + 1)) { !journal_key_cmp(i, i + 1)) {
BUG_ON(bversion_cmp(i[0].k->k.version, i[1].k->k.version) >= 0); BUG_ON(bversion_cmp(i[0].k->k.bversion, i[1].k->k.bversion) >= 0);
i[1].journal_seq = i[0].journal_seq; i[1].journal_seq = i[0].journal_seq;
......
...@@ -36,8 +36,8 @@ static inline void bch2_accounting_accumulate(struct bkey_i_accounting *dst, ...@@ -36,8 +36,8 @@ static inline void bch2_accounting_accumulate(struct bkey_i_accounting *dst,
for (unsigned i = 0; i < bch2_accounting_counters(&dst->k); i++) for (unsigned i = 0; i < bch2_accounting_counters(&dst->k); i++)
dst->v.d[i] += src.v->d[i]; dst->v.d[i] += src.v->d[i];
if (bversion_cmp(dst->k.version, src.k->version) < 0) if (bversion_cmp(dst->k.bversion, src.k->bversion) < 0)
dst->k.version = src.k->version; dst->k.bversion = src.k->bversion;
} }
static inline void fs_usage_data_type_to_base(struct bch_fs_usage_base *fs_usage, static inline void fs_usage_data_type_to_base(struct bch_fs_usage_base *fs_usage,
......
...@@ -6,7 +6,7 @@ ...@@ -6,7 +6,7 @@
struct accounting_mem_entry { struct accounting_mem_entry {
struct bpos pos; struct bpos pos;
struct bversion version; struct bversion bversion;
unsigned nr_counters; unsigned nr_counters;
u64 __percpu *v[2]; u64 __percpu *v[2];
}; };
......
...@@ -517,7 +517,7 @@ static int __bch2_rbio_narrow_crcs(struct btree_trans *trans, ...@@ -517,7 +517,7 @@ static int __bch2_rbio_narrow_crcs(struct btree_trans *trans,
if ((ret = bkey_err(k))) if ((ret = bkey_err(k)))
goto out; goto out;
if (bversion_cmp(k.k->version, rbio->version) || if (bversion_cmp(k.k->bversion, rbio->version) ||
!bch2_bkey_matches_ptr(c, k, rbio->pick.ptr, data_offset)) !bch2_bkey_matches_ptr(c, k, rbio->pick.ptr, data_offset))
goto out; goto out;
...@@ -1031,7 +1031,7 @@ int __bch2_read_extent(struct btree_trans *trans, struct bch_read_bio *orig, ...@@ -1031,7 +1031,7 @@ int __bch2_read_extent(struct btree_trans *trans, struct bch_read_bio *orig,
rbio->read_pos = read_pos; rbio->read_pos = read_pos;
rbio->data_btree = data_btree; rbio->data_btree = data_btree;
rbio->data_pos = data_pos; rbio->data_pos = data_pos;
rbio->version = k.k->version; rbio->version = k.k->bversion;
rbio->promote = promote; rbio->promote = promote;
INIT_WORK(&rbio->work, NULL); INIT_WORK(&rbio->work, NULL);
......
...@@ -697,7 +697,7 @@ static void init_append_extent(struct bch_write_op *op, ...@@ -697,7 +697,7 @@ static void init_append_extent(struct bch_write_op *op,
e = bkey_extent_init(op->insert_keys.top); e = bkey_extent_init(op->insert_keys.top);
e->k.p = op->pos; e->k.p = op->pos;
e->k.size = crc.uncompressed_size; e->k.size = crc.uncompressed_size;
e->k.version = version; e->k.bversion = version;
if (crc.csum_type || if (crc.csum_type ||
crc.compression_type || crc.compression_type ||
...@@ -1544,7 +1544,7 @@ static void bch2_write_data_inline(struct bch_write_op *op, unsigned data_len) ...@@ -1544,7 +1544,7 @@ static void bch2_write_data_inline(struct bch_write_op *op, unsigned data_len)
id = bkey_inline_data_init(op->insert_keys.top); id = bkey_inline_data_init(op->insert_keys.top);
id->k.p = op->pos; id->k.p = op->pos;
id->k.version = op->version; id->k.bversion = op->version;
id->k.size = sectors; id->k.size = sectors;
iter = bio->bi_iter; iter = bio->bi_iter;
......
...@@ -151,7 +151,7 @@ static int bch2_journal_replay_accounting_key(struct btree_trans *trans, ...@@ -151,7 +151,7 @@ static int bch2_journal_replay_accounting_key(struct btree_trans *trans,
struct bkey_s_c old = bch2_btree_path_peek_slot(btree_iter_path(trans, &iter), &u); struct bkey_s_c old = bch2_btree_path_peek_slot(btree_iter_path(trans, &iter), &u);
/* Has this delta already been applied to the btree? */ /* Has this delta already been applied to the btree? */
if (bversion_cmp(old.k->version, k->k->k.version) >= 0) { if (bversion_cmp(old.k->bversion, k->k->k.bversion) >= 0) {
ret = 0; ret = 0;
goto out; goto out;
} }
......
...@@ -367,7 +367,7 @@ static int bch2_make_extent_indirect(struct btree_trans *trans, ...@@ -367,7 +367,7 @@ static int bch2_make_extent_indirect(struct btree_trans *trans,
r_v->k.type = bkey_type_to_indirect(&orig->k); r_v->k.type = bkey_type_to_indirect(&orig->k);
r_v->k.p = reflink_iter.pos; r_v->k.p = reflink_iter.pos;
bch2_key_resize(&r_v->k, orig->k.size); bch2_key_resize(&r_v->k, orig->k.size);
r_v->k.version = orig->k.version; r_v->k.bversion = orig->k.bversion;
set_bkey_val_bytes(&r_v->k, sizeof(__le64) + bkey_val_bytes(&orig->k)); set_bkey_val_bytes(&r_v->k, sizeof(__le64) + bkey_val_bytes(&orig->k));
......
...@@ -394,7 +394,7 @@ static int insert_test_extent(struct bch_fs *c, ...@@ -394,7 +394,7 @@ static int insert_test_extent(struct bch_fs *c,
k.k_i.k.p.offset = end; k.k_i.k.p.offset = end;
k.k_i.k.p.snapshot = U32_MAX; k.k_i.k.p.snapshot = U32_MAX;
k.k_i.k.size = end - start; k.k_i.k.size = end - start;
k.k_i.k.version.lo = test_version++; k.k_i.k.bversion.lo = test_version++;
ret = bch2_btree_insert(c, BTREE_ID_extents, &k.k_i, NULL, 0, 0); ret = bch2_btree_insert(c, BTREE_ID_extents, &k.k_i, NULL, 0, 0);
bch_err_fn(c, ret); bch_err_fn(c, ret);
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment