Commit 531a0095 authored by Kent Overstreet's avatar Kent Overstreet Committed by Kent Overstreet

bcachefs: Improve btree iterator tracepoints

This patch adds some new tracepoints to the btree iterator code, and
adds new fields to the existing tracepoints - primarily for the iterator
position.
Signed-off-by: default avatarKent Overstreet <kent.overstreet@gmail.com>
parent f7beb4ca
......@@ -816,7 +816,10 @@ struct btree *bch2_btree_node_get(struct bch_fs *c, struct btree_iter *iter,
if (bch2_btree_node_relock(iter, level + 1))
goto retry;
trace_trans_restart_btree_node_reused(iter->trans->ip);
trace_trans_restart_btree_node_reused(iter->trans->ip,
trace_ip,
iter->btree_id,
&iter->real_pos);
return ERR_PTR(-EINTR);
}
}
......
......@@ -178,8 +178,8 @@ static bool bch2_btree_node_upgrade(struct btree_iter *iter, unsigned level)
return true;
}
static inline bool btree_iter_get_locks(struct btree_iter *iter,
bool upgrade, bool trace)
static inline bool btree_iter_get_locks(struct btree_iter *iter, bool upgrade,
unsigned long trace_ip)
{
unsigned l = iter->level;
int fail_idx = -1;
......@@ -191,16 +191,17 @@ static inline bool btree_iter_get_locks(struct btree_iter *iter,
if (!(upgrade
? bch2_btree_node_upgrade(iter, l)
: bch2_btree_node_relock(iter, l))) {
if (trace)
(upgrade
? trace_node_upgrade_fail
: trace_node_relock_fail)(l, iter->l[l].lock_seq,
is_btree_node(iter, l)
? 0
: (unsigned long) iter->l[l].b,
is_btree_node(iter, l)
? iter->l[l].b->c.lock.state.seq
: 0);
(upgrade
? trace_node_upgrade_fail
: trace_node_relock_fail)(iter->trans->ip, trace_ip,
iter->btree_id, &iter->real_pos,
l, iter->l[l].lock_seq,
is_btree_node(iter, l)
? 0
: (unsigned long) iter->l[l].b,
is_btree_node(iter, l)
? iter->l[l].b->c.lock.state.seq
: 0);
fail_idx = l;
btree_iter_set_dirty(iter, BTREE_ITER_NEED_TRAVERSE);
......@@ -379,9 +380,9 @@ static inline void bch2_btree_iter_verify_locks(struct btree_iter *iter) {}
#endif
__flatten
bool bch2_btree_iter_relock(struct btree_iter *iter, bool trace)
static bool bch2_btree_iter_relock(struct btree_iter *iter, unsigned long trace_ip)
{
return btree_iter_get_locks(iter, false, trace);
return btree_iter_get_locks(iter, false, trace_ip);
}
bool __bch2_btree_iter_upgrade(struct btree_iter *iter,
......@@ -393,7 +394,7 @@ bool __bch2_btree_iter_upgrade(struct btree_iter *iter,
iter->locks_want = new_locks_want;
if (btree_iter_get_locks(iter, true, true))
if (btree_iter_get_locks(iter, true, _THIS_IP_))
return true;
/*
......@@ -421,7 +422,7 @@ bool __bch2_btree_iter_upgrade(struct btree_iter *iter,
linked->btree_id == iter->btree_id &&
linked->locks_want < new_locks_want) {
linked->locks_want = new_locks_want;
btree_iter_get_locks(linked, true, false);
btree_iter_get_locks(linked, true, _THIS_IP_);
}
return false;
......@@ -467,8 +468,9 @@ bool bch2_trans_relock(struct btree_trans *trans)
struct btree_iter *iter;
trans_for_each_iter(trans, iter)
if (!bch2_btree_iter_relock(iter, true)) {
trace_trans_restart_relock(trans->ip);
if (!bch2_btree_iter_relock(iter, _RET_IP_)) {
trace_trans_restart_relock(trans->ip, _RET_IP_,
iter->btree_id, &iter->real_pos);
return false;
}
return true;
......@@ -1182,7 +1184,8 @@ static __always_inline int btree_iter_down(struct btree_iter *iter,
static int btree_iter_traverse_one(struct btree_iter *, unsigned long);
static int __btree_iter_traverse_all(struct btree_trans *trans, int ret)
static int __btree_iter_traverse_all(struct btree_trans *trans, int ret,
unsigned long trace_ip)
{
struct bch_fs *c = trans->c;
struct btree_iter *iter;
......@@ -1199,7 +1202,7 @@ static int __btree_iter_traverse_all(struct btree_trans *trans, int ret)
relock_fail = false;
trans_for_each_iter(trans, iter) {
if (!bch2_btree_iter_relock(iter, true))
if (!bch2_btree_iter_relock(iter, _THIS_IP_))
relock_fail = true;
sorted[nr_sorted++] = iter->idx;
}
......@@ -1276,13 +1279,13 @@ static int __btree_iter_traverse_all(struct btree_trans *trans, int ret)
trans->in_traverse_all = false;
trace_trans_traverse_all(trans->ip);
trace_trans_traverse_all(trans->ip, trace_ip);
return ret;
}
int bch2_btree_iter_traverse_all(struct btree_trans *trans)
{
return __btree_iter_traverse_all(trans, 0);
return __btree_iter_traverse_all(trans, 0, _RET_IP_);
}
static inline bool btree_iter_good_node(struct btree_iter *iter,
......@@ -1327,6 +1330,7 @@ static int btree_iter_traverse_one(struct btree_iter *iter,
unsigned long trace_ip)
{
unsigned depth_want = iter->level;
int ret = 0;
/*
* if we need interior nodes locked, call btree_iter_relock() to make
......@@ -1334,16 +1338,18 @@ static int btree_iter_traverse_one(struct btree_iter *iter,
*/
if (iter->uptodate == BTREE_ITER_NEED_RELOCK ||
iter->locks_want > 1)
bch2_btree_iter_relock(iter, false);
bch2_btree_iter_relock(iter, _THIS_IP_);
if (btree_iter_type(iter) == BTREE_ITER_CACHED)
return bch2_btree_iter_traverse_cached(iter);
if (btree_iter_type(iter) == BTREE_ITER_CACHED) {
ret = bch2_btree_iter_traverse_cached(iter);
goto out;
}
if (iter->uptodate < BTREE_ITER_NEED_RELOCK)
return 0;
goto out;
if (unlikely(iter->level >= BTREE_MAX_DEPTH))
return 0;
goto out;
iter->level = btree_iter_up_until_good_node(iter, 0);
......@@ -1354,12 +1360,18 @@ static int btree_iter_traverse_one(struct btree_iter *iter,
* btree_iter_lock_root() comes next and that it can't fail
*/
while (iter->level > depth_want) {
int ret = btree_iter_node(iter, iter->level)
ret = btree_iter_node(iter, iter->level)
? btree_iter_down(iter, trace_ip)
: btree_iter_lock_root(iter, depth_want, trace_ip);
if (unlikely(ret)) {
if (ret == 1)
return 0;
if (ret == 1) {
/*
* Got to the end of the btree (in
* BTREE_ITER_NODES mode)
*/
ret = 0;
goto out;
}
iter->level = depth_want;
......@@ -1371,14 +1383,16 @@ static int btree_iter_traverse_one(struct btree_iter *iter,
iter->l[iter->level].b =
BTREE_ITER_NO_NODE_DOWN;
}
return ret;
goto out;
}
}
iter->uptodate = BTREE_ITER_NEED_PEEK;
out:
trace_iter_traverse(iter->trans->ip, trace_ip,
iter->btree_id, &iter->real_pos, ret);
bch2_btree_iter_verify(iter);
return 0;
return ret;
}
static int __must_check __bch2_btree_iter_traverse(struct btree_iter *iter)
......@@ -1389,7 +1403,7 @@ static int __must_check __bch2_btree_iter_traverse(struct btree_iter *iter)
ret = bch2_trans_cond_resched(trans) ?:
btree_iter_traverse_one(iter, _RET_IP_);
if (unlikely(ret))
ret = __btree_iter_traverse_all(trans, ret);
ret = __btree_iter_traverse_all(trans, ret, _RET_IP_);
return ret;
}
......@@ -1505,6 +1519,7 @@ struct btree *bch2_btree_iter_next_node(struct btree_iter *iter)
static void btree_iter_set_search_pos(struct btree_iter *iter, struct bpos new_pos)
{
struct bpos old_pos = iter->real_pos;
int cmp = bpos_cmp(new_pos, iter->real_pos);
unsigned l = iter->level;
......@@ -1515,7 +1530,7 @@ static void btree_iter_set_search_pos(struct btree_iter *iter, struct bpos new_p
if (unlikely(btree_iter_type(iter) == BTREE_ITER_CACHED)) {
btree_node_unlock(iter, 0);
iter->l[0].b = BTREE_ITER_NO_NODE_UP;
iter->l[0].b = BTREE_ITER_NO_NODE_CACHED;
btree_iter_set_dirty(iter, BTREE_ITER_NEED_TRAVERSE);
return;
}
......@@ -1544,6 +1559,11 @@ static void btree_iter_set_search_pos(struct btree_iter *iter, struct bpos new_p
btree_iter_set_dirty(iter, BTREE_ITER_NEED_PEEK);
bch2_btree_iter_verify(iter);
#ifdef CONFIG_BCACHEFS_DEBUG
trace_iter_set_search_pos(iter->trans->ip, _RET_IP_,
iter->btree_id,
&old_pos, &new_pos, l);
#endif
}
inline bool bch2_btree_iter_advance(struct btree_iter *iter)
......@@ -2062,13 +2082,6 @@ struct btree_iter *__bch2_trans_get_iter(struct btree_trans *trans,
best = iter;
}
trace_trans_get_iter(_RET_IP_, trans->ip,
btree_id,
&real_pos, locks_want,
best ? &best->real_pos : &pos_min,
best ? best->locks_want : 0,
best ? best->uptodate : BTREE_ITER_NEED_TRAVERSE);
if (!best) {
iter = btree_trans_iter_alloc(trans);
bch2_btree_iter_init(trans, iter, btree_id);
......@@ -2097,7 +2110,7 @@ struct btree_iter *__bch2_trans_get_iter(struct btree_trans *trans,
locks_want = min(locks_want, BTREE_MAX_DEPTH);
if (locks_want > iter->locks_want) {
iter->locks_want = locks_want;
btree_iter_get_locks(iter, true, false);
btree_iter_get_locks(iter, true, _THIS_IP_);
}
while (iter->level != depth) {
......@@ -2115,6 +2128,13 @@ struct btree_iter *__bch2_trans_get_iter(struct btree_trans *trans,
bch2_btree_iter_set_pos(iter, pos);
btree_iter_set_search_pos(iter, real_pos);
trace_trans_get_iter(_RET_IP_, trans->ip,
btree_id,
&real_pos, locks_want, iter->uptodate,
best ? &best->real_pos : &pos_min,
best ? best->locks_want : U8_MAX,
best ? best->uptodate : U8_MAX);
return iter;
}
......
......@@ -111,7 +111,6 @@ void bch2_btree_node_iter_fix(struct btree_iter *, struct btree *,
struct btree_node_iter *, struct bkey_packed *,
unsigned, unsigned);
bool bch2_btree_iter_relock(struct btree_iter *, bool);
bool bch2_trans_relock(struct btree_trans *);
void bch2_trans_unlock(struct btree_trans *);
......
......@@ -230,6 +230,7 @@ enum btree_iter_uptodate {
#define BTREE_ITER_NO_NODE_DOWN ((struct btree *) 5)
#define BTREE_ITER_NO_NODE_INIT ((struct btree *) 6)
#define BTREE_ITER_NO_NODE_ERROR ((struct btree *) 7)
#define BTREE_ITER_NO_NODE_CACHED ((struct btree *) 8)
/*
* @pos - iterator's current position
......
......@@ -955,7 +955,9 @@ bch2_btree_update_start(struct btree_iter *iter, unsigned level,
* instead of locking/reserving all the way to the root:
*/
if (!bch2_btree_iter_upgrade(iter, U8_MAX)) {
trace_trans_restart_iter_upgrade(trans->ip);
trace_trans_restart_iter_upgrade(trans->ip, _RET_IP_,
iter->btree_id,
&iter->real_pos);
return ERR_PTR(-EINTR);
}
......@@ -996,7 +998,7 @@ bch2_btree_update_start(struct btree_iter *iter, unsigned level,
* closure argument
*/
if (flags & BTREE_INSERT_NOUNLOCK) {
trace_trans_restart_journal_preres_get(trans->ip);
trace_trans_restart_journal_preres_get(trans->ip, _RET_IP_);
ret = -EINTR;
goto err;
}
......@@ -1012,7 +1014,7 @@ bch2_btree_update_start(struct btree_iter *iter, unsigned level,
BTREE_UPDATE_JOURNAL_RES,
journal_flags);
if (ret) {
trace_trans_restart_journal_preres_get(trans->ip);
trace_trans_restart_journal_preres_get(trans->ip, _RET_IP_);
goto err;
}
......
......@@ -228,7 +228,8 @@ static inline void btree_insert_entry_checks(struct btree_trans *trans,
}
static noinline int
bch2_trans_journal_preres_get_cold(struct btree_trans *trans, unsigned u64s)
bch2_trans_journal_preres_get_cold(struct btree_trans *trans, unsigned u64s,
unsigned long trace_ip)
{
struct bch_fs *c = trans->c;
int ret;
......@@ -241,7 +242,7 @@ bch2_trans_journal_preres_get_cold(struct btree_trans *trans, unsigned u64s)
return ret;
if (!bch2_trans_relock(trans)) {
trace_trans_restart_journal_preres_get(trans->ip);
trace_trans_restart_journal_preres_get(trans->ip, trace_ip);
return -EINTR;
}
......@@ -368,7 +369,8 @@ static noinline void bch2_trans_mark_gc(struct btree_trans *trans)
static inline int
bch2_trans_commit_write_locked(struct btree_trans *trans,
struct btree_insert_entry **stopped_at)
struct btree_insert_entry **stopped_at,
unsigned long trace_ip)
{
struct bch_fs *c = trans->c;
struct btree_insert_entry *i;
......@@ -378,7 +380,7 @@ bch2_trans_commit_write_locked(struct btree_trans *trans,
int ret;
if (race_fault()) {
trace_trans_restart_fault_inject(trans->ip);
trace_trans_restart_fault_inject(trans->ip, trace_ip);
return -EINTR;
}
......@@ -525,7 +527,8 @@ static noinline int maybe_do_btree_merge(struct btree_trans *trans, struct btree
* Get journal reservation, take write locks, and attempt to do btree update(s):
*/
static inline int do_bch2_trans_commit(struct btree_trans *trans,
struct btree_insert_entry **stopped_at)
struct btree_insert_entry **stopped_at,
unsigned long trace_ip)
{
struct bch_fs *c = trans->c;
struct btree_insert_entry *i;
......@@ -559,7 +562,7 @@ static inline int do_bch2_trans_commit(struct btree_trans *trans,
? JOURNAL_RES_GET_RESERVED : 0));
if (unlikely(ret == -EAGAIN))
ret = bch2_trans_journal_preres_get_cold(trans,
trans->journal_preres_u64s);
trans->journal_preres_u64s, trace_ip);
if (unlikely(ret))
return ret;
......@@ -578,7 +581,9 @@ static inline int do_bch2_trans_commit(struct btree_trans *trans,
if (iter->nodes_locked != iter->nodes_intent_locked) {
if (btree_iter_keep(trans, iter)) {
if (!bch2_btree_iter_upgrade(iter, 1)) {
trace_trans_restart_upgrade(trans->ip);
trace_trans_restart_upgrade(trans->ip, trace_ip,
iter->btree_id,
&iter->real_pos);
return -EINTR;
}
} else {
......@@ -606,7 +611,7 @@ static inline int do_bch2_trans_commit(struct btree_trans *trans,
bch2_btree_node_lock_for_insert(c,
iter_l(i->iter)->b, i->iter);
ret = bch2_trans_commit_write_locked(trans, stopped_at);
ret = bch2_trans_commit_write_locked(trans, stopped_at, trace_ip);
trans_for_each_update2(trans, i)
if (!same_leaf_as_prev(trans, i))
......@@ -644,7 +649,7 @@ static int journal_reclaim_wait_done(struct bch_fs *c)
static noinline
int bch2_trans_commit_error(struct btree_trans *trans,
struct btree_insert_entry *i,
int ret)
int ret, unsigned long trace_ip)
{
struct bch_fs *c = trans->c;
unsigned flags = trans->flags;
......@@ -685,7 +690,9 @@ int bch2_trans_commit_error(struct btree_trans *trans,
if (!ret ||
ret == -EINTR ||
(flags & BTREE_INSERT_NOUNLOCK)) {
trace_trans_restart_btree_node_split(trans->ip);
trace_trans_restart_btree_node_split(trans->ip, trace_ip,
i->iter->btree_id,
&i->iter->real_pos);
ret = -EINTR;
}
break;
......@@ -703,7 +710,7 @@ int bch2_trans_commit_error(struct btree_trans *trans,
if (bch2_trans_relock(trans))
return 0;
trace_trans_restart_mark_replicas(trans->ip);
trace_trans_restart_mark_replicas(trans->ip, trace_ip);
ret = -EINTR;
break;
case BTREE_INSERT_NEED_JOURNAL_RES:
......@@ -720,13 +727,13 @@ int bch2_trans_commit_error(struct btree_trans *trans,
if (bch2_trans_relock(trans))
return 0;
trace_trans_restart_journal_res_get(trans->ip);
trace_trans_restart_journal_res_get(trans->ip, trace_ip);
ret = -EINTR;
break;
case BTREE_INSERT_NEED_JOURNAL_RECLAIM:
bch2_trans_unlock(trans);
trace_trans_blocked_journal_reclaim(trans->ip);
trace_trans_blocked_journal_reclaim(trans->ip, trace_ip);
wait_event_freezable(c->journal.reclaim_wait,
(ret = journal_reclaim_wait_done(c)));
......@@ -736,7 +743,7 @@ int bch2_trans_commit_error(struct btree_trans *trans,
if (bch2_trans_relock(trans))
return 0;
trace_trans_restart_journal_reclaim(trans->ip);
trace_trans_restart_journal_reclaim(trans->ip, trace_ip);
ret = -EINTR;
break;
default:
......@@ -950,7 +957,9 @@ int __bch2_trans_commit(struct btree_trans *trans)
i->trigger_flags);
if (unlikely(ret)) {
if (ret == -EINTR)
trace_trans_restart_mark(trans->ip);
trace_trans_restart_mark(trans->ip, _RET_IP_,
i->iter->btree_id,
&i->iter->pos);
goto out;
}
}
......@@ -976,12 +985,16 @@ int __bch2_trans_commit(struct btree_trans *trans)
trans_for_each_update2(trans, i) {
ret = bch2_btree_iter_traverse(i->iter);
if (unlikely(ret)) {
trace_trans_restart_traverse(trans->ip);
trace_trans_restart_traverse(trans->ip, _RET_IP_,
i->iter->btree_id,
&i->iter->pos);
goto out;
}
if (unlikely(!bch2_btree_iter_upgrade(i->iter, i->level + 1))) {
trace_trans_restart_upgrade(trans->ip);
trace_trans_restart_upgrade(trans->ip, _RET_IP_,
i->iter->btree_id,
&i->iter->pos);
ret = -EINTR;
goto out;
}
......@@ -997,7 +1010,7 @@ int __bch2_trans_commit(struct btree_trans *trans)
retry:
memset(&trans->journal_res, 0, sizeof(trans->journal_res));
ret = do_bch2_trans_commit(trans, &i);
ret = do_bch2_trans_commit(trans, &i, _RET_IP_);
/* make sure we didn't drop or screw up locks: */
bch2_btree_trans_verify_locks(trans);
......@@ -1023,7 +1036,7 @@ int __bch2_trans_commit(struct btree_trans *trans)
return ret;
err:
ret = bch2_trans_commit_error(trans, i, ret);
ret = bch2_trans_commit_error(trans, i, ret, _RET_IP_);
if (ret)
goto out;
......
This diff is collapsed.
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment