diff options
author | Kent Overstreet <kent.overstreet@gmail.com> | 2022-01-11 23:24:43 -0500 |
---|---|---|
committer | Kent Overstreet <kent.overstreet@linux.dev> | 2023-10-22 17:09:29 -0400 |
commit | 5f417394033a0d8bfb31d02b3becf7381dc13867 (patch) | |
tree | d5fae9e13063e62c8c90dbf661e992e9e5dbbedc /fs | |
parent | 31f63fd1244d9609265eb5cfc522c142b35cdacc (diff) |
bcachefs: bch2_btree_update_start() refactoring
This simplifies the logic in bch2_btree_update_start() a bit, handling
the unlock/block logic more locally.
Signed-off-by: Kent Overstreet <kent.overstreet@linux.dev>
Diffstat (limited to 'fs')
-rw-r--r-- | fs/bcachefs/btree_update_interior.c | 80 | ||||
-rw-r--r-- | fs/bcachefs/trace.h | 23 |
2 files changed, 54 insertions, 49 deletions
diff --git a/fs/bcachefs/btree_update_interior.c b/fs/bcachefs/btree_update_interior.c index cd4332f891dc..ff56c374ff2e 100644 --- a/fs/bcachefs/btree_update_interior.c +++ b/fs/bcachefs/btree_update_interior.c @@ -413,19 +413,24 @@ static void bch2_btree_reserve_put(struct btree_update *as) mutex_unlock(&c->btree_reserve_cache_lock); } -static int bch2_btree_reserve_get(struct btree_update *as, unsigned nr_nodes[2], - unsigned flags, struct closure *cl) +static int bch2_btree_reserve_get(struct btree_update *as, + unsigned nr_nodes[2], + unsigned flags, + struct closure *cl) { struct bch_fs *c = as->c; struct btree *b; unsigned interior; - int ret; + int ret = 0; BUG_ON(nr_nodes[0] + nr_nodes[1] > BTREE_RESERVE_MAX); /* * Protects reaping from the btree node cache and using the btree node * open bucket reserve: + * + * BTREE_INSERT_NOWAIT only applies to btree node allocation, not + * blocking on this lock: */ ret = bch2_btree_cache_cannibalize_lock(c, cl); if (ret) @@ -436,9 +441,8 @@ static int bch2_btree_reserve_get(struct btree_update *as, unsigned nr_nodes[2], while (p->nr < nr_nodes[interior]) { b = __bch2_btree_node_alloc(c, &as->disk_res, - flags & BTREE_INSERT_NOWAIT - ? NULL : cl, - interior, flags); + flags & BTREE_INSERT_NOWAIT ? NULL : cl, + interior, flags); if (IS_ERR(b)) { ret = PTR_ERR(b); goto err; @@ -447,12 +451,8 @@ static int bch2_btree_reserve_get(struct btree_update *as, unsigned nr_nodes[2], p->b[p->nr++] = b; } } - - bch2_btree_cache_cannibalize_unlock(c); - return 0; err: bch2_btree_cache_cannibalize_unlock(c); - trace_btree_reserve_get_fail(c, nr_nodes[0] + nr_nodes[1], cl); return ret; } @@ -958,21 +958,18 @@ bch2_btree_update_start(struct btree_trans *trans, struct btree_path *path, { struct bch_fs *c = trans->c; struct btree_update *as; - struct closure cl; u64 start_time = local_clock(); int disk_res_flags = (flags & BTREE_INSERT_NOFAIL) ? BCH_DISK_RESERVATION_NOFAIL : 0; - unsigned nr_nodes[2]; + unsigned nr_nodes[2] = { 0, 0 }; unsigned update_level = level; int journal_flags = flags & JOURNAL_WATERMARK_MASK; int ret = 0; BUG_ON(!path->should_be_locked); - closure_init_stack(&cl); -retry: - nr_nodes[0] = nr_nodes[1] = 0; - update_level = level; + if (flags & BTREE_INSERT_JOURNAL_RECLAIM) + journal_flags |= JOURNAL_RES_GET_NONBLOCK; while (1) { nr_nodes[!!update_level] += 1 + split; @@ -1044,27 +1041,21 @@ retry: ret = bch2_journal_preres_get(&c->journal, &as->journal_preres, BTREE_UPDATE_JOURNAL_RES, journal_flags|JOURNAL_RES_GET_NONBLOCK); - if (ret == -EAGAIN) { + if (ret) { bch2_trans_unlock(trans); - if (flags & BTREE_INSERT_JOURNAL_RECLAIM) { - bch2_btree_update_free(as); - btree_trans_restart(trans); - return ERR_PTR(ret); - } - ret = bch2_journal_preres_get(&c->journal, &as->journal_preres, - BTREE_UPDATE_JOURNAL_RES, - journal_flags); + BTREE_UPDATE_JOURNAL_RES, + journal_flags); if (ret) { trace_trans_restart_journal_preres_get(trans->fn, _RET_IP_); + btree_trans_restart(trans); goto err; } - if (!bch2_trans_relock(trans)) { - ret = -EINTR; + ret = bch2_trans_relock(trans); + if (ret) goto err; - } } ret = bch2_disk_reservation_get(c, &as->disk_res, @@ -1074,23 +1065,34 @@ retry: if (ret) goto err; - ret = bch2_btree_reserve_get(as, nr_nodes, flags, &cl); - if (ret) - goto err; + ret = bch2_btree_reserve_get(as, nr_nodes, flags, NULL); + if (ret) { + struct closure cl; - return as; -err: - bch2_btree_update_free(as); + closure_init_stack(&cl); - if (ret == -EAGAIN) { bch2_trans_unlock(trans); - closure_sync(&cl); - ret = -EINTR; + + do { + ret = bch2_btree_reserve_get(as, nr_nodes, flags, &cl); + closure_sync(&cl); + } while (ret == -EAGAIN); } - if (ret == -EINTR && bch2_trans_relock(trans)) - goto retry; + if (ret) { + trace_btree_reserve_get_fail(trans->fn, _RET_IP_, + nr_nodes[0] + nr_nodes[1]); + goto err; + } + if (!bch2_trans_relock(trans)) { + ret = -EINTR; + goto err; + } + + return as; +err: + bch2_btree_update_free(as); return ERR_PTR(ret); } diff --git a/fs/bcachefs/trace.h b/fs/bcachefs/trace.h index 54260349c07e..89207fd7b617 100644 --- a/fs/bcachefs/trace.h +++ b/fs/bcachefs/trace.h @@ -278,24 +278,27 @@ DEFINE_EVENT(bch_fs, btree_node_cannibalize_unlock, ); TRACE_EVENT(btree_reserve_get_fail, - TP_PROTO(struct bch_fs *c, size_t required, struct closure *cl), - TP_ARGS(c, required, cl), + TP_PROTO(const char *trans_fn, + unsigned long caller_ip, + size_t required), + TP_ARGS(trans_fn, caller_ip, required), TP_STRUCT__entry( - __field(dev_t, dev ) + __array(char, trans_fn, 24 ) + __field(unsigned long, caller_ip ) __field(size_t, required ) - __field(struct closure *, cl ) ), TP_fast_assign( - __entry->dev = c->dev; - __entry->required = required; - __entry->cl = cl; + strlcpy(__entry->trans_fn, trans_fn, sizeof(__entry->trans_fn)); + __entry->caller_ip = caller_ip; + __entry->required = required; ), - TP_printk("%d,%d required %zu by %p", - MAJOR(__entry->dev), MINOR(__entry->dev), - __entry->required, __entry->cl) + TP_printk("%s %pS required %zu", + __entry->trans_fn, + (void *) __entry->caller_ip, + __entry->required) ); DEFINE_EVENT(btree_node, btree_split, |