summaryrefslogtreecommitdiff
path: root/fs
diff options
context:
space:
mode:
authorKent Overstreet <kent.overstreet@gmail.com>2022-01-11 23:24:43 -0500
committerKent Overstreet <kent.overstreet@linux.dev>2023-10-22 17:09:29 -0400
commit5f417394033a0d8bfb31d02b3becf7381dc13867 (patch)
treed5fae9e13063e62c8c90dbf661e992e9e5dbbedc /fs
parent31f63fd1244d9609265eb5cfc522c142b35cdacc (diff)
bcachefs: bch2_btree_update_start() refactoring
This simplifies the logic in bch2_btree_update_start() a bit, handling the unlock/block logic more locally. Signed-off-by: Kent Overstreet <kent.overstreet@linux.dev>
Diffstat (limited to 'fs')
-rw-r--r--fs/bcachefs/btree_update_interior.c80
-rw-r--r--fs/bcachefs/trace.h23
2 files changed, 54 insertions, 49 deletions
diff --git a/fs/bcachefs/btree_update_interior.c b/fs/bcachefs/btree_update_interior.c
index cd4332f891dc..ff56c374ff2e 100644
--- a/fs/bcachefs/btree_update_interior.c
+++ b/fs/bcachefs/btree_update_interior.c
@@ -413,19 +413,24 @@ static void bch2_btree_reserve_put(struct btree_update *as)
mutex_unlock(&c->btree_reserve_cache_lock);
}
-static int bch2_btree_reserve_get(struct btree_update *as, unsigned nr_nodes[2],
- unsigned flags, struct closure *cl)
+static int bch2_btree_reserve_get(struct btree_update *as,
+ unsigned nr_nodes[2],
+ unsigned flags,
+ struct closure *cl)
{
struct bch_fs *c = as->c;
struct btree *b;
unsigned interior;
- int ret;
+ int ret = 0;
BUG_ON(nr_nodes[0] + nr_nodes[1] > BTREE_RESERVE_MAX);
/*
* Protects reaping from the btree node cache and using the btree node
* open bucket reserve:
+ *
+ * BTREE_INSERT_NOWAIT only applies to btree node allocation, not
+ * blocking on this lock:
*/
ret = bch2_btree_cache_cannibalize_lock(c, cl);
if (ret)
@@ -436,9 +441,8 @@ static int bch2_btree_reserve_get(struct btree_update *as, unsigned nr_nodes[2],
while (p->nr < nr_nodes[interior]) {
b = __bch2_btree_node_alloc(c, &as->disk_res,
- flags & BTREE_INSERT_NOWAIT
- ? NULL : cl,
- interior, flags);
+ flags & BTREE_INSERT_NOWAIT ? NULL : cl,
+ interior, flags);
if (IS_ERR(b)) {
ret = PTR_ERR(b);
goto err;
@@ -447,12 +451,8 @@ static int bch2_btree_reserve_get(struct btree_update *as, unsigned nr_nodes[2],
p->b[p->nr++] = b;
}
}
-
- bch2_btree_cache_cannibalize_unlock(c);
- return 0;
err:
bch2_btree_cache_cannibalize_unlock(c);
- trace_btree_reserve_get_fail(c, nr_nodes[0] + nr_nodes[1], cl);
return ret;
}
@@ -958,21 +958,18 @@ bch2_btree_update_start(struct btree_trans *trans, struct btree_path *path,
{
struct bch_fs *c = trans->c;
struct btree_update *as;
- struct closure cl;
u64 start_time = local_clock();
int disk_res_flags = (flags & BTREE_INSERT_NOFAIL)
? BCH_DISK_RESERVATION_NOFAIL : 0;
- unsigned nr_nodes[2];
+ unsigned nr_nodes[2] = { 0, 0 };
unsigned update_level = level;
int journal_flags = flags & JOURNAL_WATERMARK_MASK;
int ret = 0;
BUG_ON(!path->should_be_locked);
- closure_init_stack(&cl);
-retry:
- nr_nodes[0] = nr_nodes[1] = 0;
- update_level = level;
+ if (flags & BTREE_INSERT_JOURNAL_RECLAIM)
+ journal_flags |= JOURNAL_RES_GET_NONBLOCK;
while (1) {
nr_nodes[!!update_level] += 1 + split;
@@ -1044,27 +1041,21 @@ retry:
ret = bch2_journal_preres_get(&c->journal, &as->journal_preres,
BTREE_UPDATE_JOURNAL_RES,
journal_flags|JOURNAL_RES_GET_NONBLOCK);
- if (ret == -EAGAIN) {
+ if (ret) {
bch2_trans_unlock(trans);
- if (flags & BTREE_INSERT_JOURNAL_RECLAIM) {
- bch2_btree_update_free(as);
- btree_trans_restart(trans);
- return ERR_PTR(ret);
- }
-
ret = bch2_journal_preres_get(&c->journal, &as->journal_preres,
- BTREE_UPDATE_JOURNAL_RES,
- journal_flags);
+ BTREE_UPDATE_JOURNAL_RES,
+ journal_flags);
if (ret) {
trace_trans_restart_journal_preres_get(trans->fn, _RET_IP_);
+ btree_trans_restart(trans);
goto err;
}
- if (!bch2_trans_relock(trans)) {
- ret = -EINTR;
+ ret = bch2_trans_relock(trans);
+ if (ret)
goto err;
- }
}
ret = bch2_disk_reservation_get(c, &as->disk_res,
@@ -1074,23 +1065,34 @@ retry:
if (ret)
goto err;
- ret = bch2_btree_reserve_get(as, nr_nodes, flags, &cl);
- if (ret)
- goto err;
+ ret = bch2_btree_reserve_get(as, nr_nodes, flags, NULL);
+ if (ret) {
+ struct closure cl;
- return as;
-err:
- bch2_btree_update_free(as);
+ closure_init_stack(&cl);
- if (ret == -EAGAIN) {
bch2_trans_unlock(trans);
- closure_sync(&cl);
- ret = -EINTR;
+
+ do {
+ ret = bch2_btree_reserve_get(as, nr_nodes, flags, &cl);
+ closure_sync(&cl);
+ } while (ret == -EAGAIN);
}
- if (ret == -EINTR && bch2_trans_relock(trans))
- goto retry;
+ if (ret) {
+ trace_btree_reserve_get_fail(trans->fn, _RET_IP_,
+ nr_nodes[0] + nr_nodes[1]);
+ goto err;
+ }
+ if (!bch2_trans_relock(trans)) {
+ ret = -EINTR;
+ goto err;
+ }
+
+ return as;
+err:
+ bch2_btree_update_free(as);
return ERR_PTR(ret);
}
diff --git a/fs/bcachefs/trace.h b/fs/bcachefs/trace.h
index 54260349c07e..89207fd7b617 100644
--- a/fs/bcachefs/trace.h
+++ b/fs/bcachefs/trace.h
@@ -278,24 +278,27 @@ DEFINE_EVENT(bch_fs, btree_node_cannibalize_unlock,
);
TRACE_EVENT(btree_reserve_get_fail,
- TP_PROTO(struct bch_fs *c, size_t required, struct closure *cl),
- TP_ARGS(c, required, cl),
+ TP_PROTO(const char *trans_fn,
+ unsigned long caller_ip,
+ size_t required),
+ TP_ARGS(trans_fn, caller_ip, required),
TP_STRUCT__entry(
- __field(dev_t, dev )
+ __array(char, trans_fn, 24 )
+ __field(unsigned long, caller_ip )
__field(size_t, required )
- __field(struct closure *, cl )
),
TP_fast_assign(
- __entry->dev = c->dev;
- __entry->required = required;
- __entry->cl = cl;
+ strlcpy(__entry->trans_fn, trans_fn, sizeof(__entry->trans_fn));
+ __entry->caller_ip = caller_ip;
+ __entry->required = required;
),
- TP_printk("%d,%d required %zu by %p",
- MAJOR(__entry->dev), MINOR(__entry->dev),
- __entry->required, __entry->cl)
+ TP_printk("%s %pS required %zu",
+ __entry->trans_fn,
+ (void *) __entry->caller_ip,
+ __entry->required)
);
DEFINE_EVENT(btree_node, btree_split,