public inbox for linux-bcachefs@vger.kernel.org
 help / color / mirror / Atom feed
From: Jul Lang <jullanggit@proton.me>
To: linux-bcachefs@vger.kernel.org
Cc: kent.overstreet@linux.dev, Jul Lang <jullanggit@proton.me>
Subject: [PATCH 2/2] fix typos
Date: Sat, 24 Jan 2026 11:19:41 +0000	[thread overview]
Message-ID: <20260124111913.6263-3-jullanggit@proton.me> (raw)
In-Reply-To: <20260124111913.6263-1-jullanggit@proton.me>

Signed-off-by: Jul Lang <jullanggit@proton.me>
---
 fs/bcachefs/alloc/accounting.c            |  2 +-
 fs/bcachefs/alloc/buckets.c               |  2 +-
 fs/bcachefs/alloc/check.c                 |  2 +-
 fs/bcachefs/alloc/foreground.c            |  6 +++---
 fs/bcachefs/bcachefs.h                    |  2 +-
 fs/bcachefs/bcachefs_format.h             | 10 +++++-----
 fs/bcachefs/bcachefs_ioctl.h              |  4 ++--
 fs/bcachefs/btree/bkey_methods.c          |  4 ++--
 fs/bcachefs/btree/bkey_methods.h          |  2 +-
 fs/bcachefs/btree/bset.h                  |  4 ++--
 fs/bcachefs/btree/cache.c                 |  2 +-
 fs/bcachefs/btree/interior.h              |  2 +-
 fs/bcachefs/btree/iter.c                  |  4 ++--
 fs/bcachefs/btree/update.c                |  4 ++--
 fs/bcachefs/data/checksum.c               |  2 +-
 fs/bcachefs/data/reconcile/work.c         |  2 +-
 fs/bcachefs/debug/async_objs.c            |  2 +-
 fs/bcachefs/fs/check.c                    |  2 +-
 fs/bcachefs/fs/inode.c                    |  2 +-
 fs/bcachefs/init/passes.c                 |  2 +-
 fs/bcachefs/init/progress.h               |  2 +-
 fs/bcachefs/journal/journal.c             |  4 ++--
 fs/bcachefs/journal/types.h               |  4 ++--
 fs/bcachefs/journal/write.c               |  2 +-
 fs/bcachefs/sb/io.c                       |  2 +-
 fs/bcachefs/sb/members_format.h           |  2 +-
 fs/bcachefs/snapshots/format.h            |  4 ++--
 fs/bcachefs/util/eytzinger.h              |  2 +-
 fs/bcachefs/util/mean_and_variance.h      |  4 ++--
 fs/bcachefs/util/mean_and_variance_test.c |  2 +-
 fs/bcachefs/util/printbuf.c               |  2 +-
 fs/bcachefs/util/printbuf.h               |  4 ++--
 fs/bcachefs/util/rcu_pending.c            |  2 +-
 fs/bcachefs/util/siphash.h                |  2 +-
 fs/bcachefs/util/six.c                    |  6 +++---
 fs/bcachefs/util/six.h                    |  4 ++--
 fs/bcachefs/util/time_stats.h             |  2 +-
 fs/bcachefs/util/util.c                   |  2 +-
 fs/bcachefs/vfs/io.c                      |  2 +-
 39 files changed, 58 insertions(+), 58 deletions(-)

diff --git a/fs/bcachefs/alloc/accounting.c b/fs/bcachefs/alloc/accounting.c
index c86a01410..0e3de80b5 100644
--- a/fs/bcachefs/alloc/accounting.c
+++ b/fs/bcachefs/alloc/accounting.c
@@ -525,7 +525,7 @@ void bch2_accounting_mem_gc(struct bch_fs *c)
  * Read out accounting keys for replicas entries, as an array of
  * bch_replicas_usage entries.
  *
- * Note: this may be deprecated/removed at smoe point in the future and replaced
+ * Note: this may be deprecated/removed at some point in the future and replaced
  * with something more general, it exists to support the ioctl used by the
  * 'bcachefs fs usage' command.
  */
diff --git a/fs/bcachefs/alloc/buckets.c b/fs/bcachefs/alloc/buckets.c
index e6261a6cc..4a3373572 100644
--- a/fs/bcachefs/alloc/buckets.c
+++ b/fs/bcachefs/alloc/buckets.c
@@ -353,7 +353,7 @@ int bch2_check_fix_ptrs(struct btree_trans *trans,
 
 			/*
 			 * no locking, we're single threaded and not rw yet, see
-			 * the big assertino above that we repeat here:
+			 * the big assertion above that we repeat here:
 			 */
 			BUG_ON(test_bit(BCH_FS_rw, &c->flags));
 
diff --git a/fs/bcachefs/alloc/check.c b/fs/bcachefs/alloc/check.c
index eef2d8df3..db91af741 100644
--- a/fs/bcachefs/alloc/check.c
+++ b/fs/bcachefs/alloc/check.c
@@ -397,7 +397,7 @@ int __bch2_check_discard_freespace_key(struct btree_trans *trans, struct btree_i
 	if (!bch2_dev_bucket_exists(c, bucket)) {
 		if (__fsck_err(trans, fsck_flags,
 			       need_discard_freespace_key_to_invalid_dev_bucket,
-			       "entry in %s btree for nonexistant dev:bucket %llu:%llu",
+			       "entry in %s btree for nonexistent dev:bucket %llu:%llu",
 			       bch2_btree_id_str(iter->btree_id), bucket.inode, bucket.offset))
 			ret = delete_discard_freespace_key(trans, iter, async_repair);
 		else
diff --git a/fs/bcachefs/alloc/foreground.c b/fs/bcachefs/alloc/foreground.c
index 7f4a0674d..c7bf0c3f9 100644
--- a/fs/bcachefs/alloc/foreground.c
+++ b/fs/bcachefs/alloc/foreground.c
@@ -388,7 +388,7 @@ static struct open_bucket *bch2_bucket_alloc_freelist(struct btree_trans *trans,
 					 POS(ca->dev_idx, U64_MAX),
 					 0, k, ret) {
 		/*
-		 * peek normally dosen't trim extents - they can span iter.pos,
+		 * peek normally doesn't trim extents - they can span iter.pos,
 		 * which is not what we want here:
 		 */
 		iter.k.size = iter.k.p.offset - iter.pos.offset;
@@ -633,7 +633,7 @@ static noinline void bch2_stripe_state_rescale(struct dev_stripe_state *stripe)
 	 * to the whole filesystem.
 	 */
 	u64 scale_max = U64_MAX;	/* maximum we can subtract without underflow */
-	u64 scale_min = 0;		/* minumum we must subtract to avoid overflow */
+	u64 scale_min = 0;		/* minimum we must subtract to avoid overflow */
 
 	for (u64 *v = stripe->next_alloc;
 	     v < stripe->next_alloc + ARRAY_SIZE(stripe->next_alloc); v++) {
@@ -1348,7 +1348,7 @@ void bch2_fs_allocator_foreground_init(struct bch_fs *c)
 	mutex_init(&a->write_points_hash_lock);
 	a->write_points_nr = ARRAY_SIZE(a->write_points);
 
-	/* open bucket 0 is a sentinal NULL: */
+	/* open bucket 0 is a sentinel NULL: */
 	spin_lock_init(&a->open_buckets[0].lock);
 
 	for (ob = a->open_buckets + 1;
diff --git a/fs/bcachefs/bcachefs.h b/fs/bcachefs/bcachefs.h
index 8b877099a..d3cfce3a4 100644
--- a/fs/bcachefs/bcachefs.h
+++ b/fs/bcachefs/bcachefs.h
@@ -778,7 +778,7 @@ struct bch_fs {
 	struct enumerated_ref	writes;
 
 	/*
-	 * Analagous to c->writes, for asynchronous ops that don't necessarily
+	 * Analogous to c->writes, for asynchronous ops that don't necessarily
 	 * need fs to be read-write
 	 */
 	refcount_t		ro_ref;
diff --git a/fs/bcachefs/bcachefs_format.h b/fs/bcachefs/bcachefs_format.h
index aa3ffe695..10368565d 100644
--- a/fs/bcachefs/bcachefs_format.h
+++ b/fs/bcachefs/bcachefs_format.h
@@ -238,7 +238,7 @@ struct bkey {
  *
  * Specifically, when i was designing bkey, I wanted the header to be no
  * bigger than necessary so that bkey_packed could use the rest. That means that
- * decently offten extent keys will fit into only 8 bytes, instead of spilling over
+ * decently often extent keys will fit into only 8 bytes, instead of spilling over
  * to 16.
  *
  * But packed_bkey treats the part after the header - the packed section -
@@ -250,7 +250,7 @@ struct bkey {
  * So that constrains the key part of a bkig endian bkey to start right
  * after the header.
  *
- * If we ever do a bkey_v2 and need to expand the hedaer by another byte for
+ * If we ever do a bkey_v2 and need to expand the header by another byte for
  * some reason - that will clean up this wart.
  */
 __aligned(8)
@@ -844,7 +844,7 @@ struct bch_sb_field_ext {
 	x(btree_node_accounting,	BCH_VERSION(1, 31))		\
 	x(sb_field_extent_type_u64s,	BCH_VERSION(1, 32))		\
 	x(reconcile,			BCH_VERSION(1, 33))		\
-	x(extented_key_type_error,	BCH_VERSION(1, 34))		\
+	x(extended_key_type_error,	BCH_VERSION(1, 34))		\
 	x(bucket_stripe_index,		BCH_VERSION(1, 35))
 
 enum bcachefs_metadata_version {
@@ -925,7 +925,7 @@ struct bch_sb {
 
 /*
  * Flags:
- * BCH_SB_INITALIZED	- set on first mount
+ * BCH_SB_INITIALIZED	- set on first mount
  * BCH_SB_CLEAN		- did we shut down cleanly? Just a hint, doesn't affect
  *			  behaviour of mount/recovery path:
  * BCH_SB_INODE_32BIT	- limit inode numbers to 32 bits
@@ -1523,7 +1523,7 @@ static inline bool btree_id_can_reconstruct(enum btree_id btree)
 }
 
 /*
- * We can reconstruct BTREE_ID_alloc, but reconstucting it from scratch is not
+ * We can reconstruct BTREE_ID_alloc, but reconstructing it from scratch is not
  * so cheap and OOMs on huge filesystems (until we have online
  * check_allocations)
  */
diff --git a/fs/bcachefs/bcachefs_ioctl.h b/fs/bcachefs/bcachefs_ioctl.h
index dd926db42..05e170f06 100644
--- a/fs/bcachefs/bcachefs_ioctl.h
+++ b/fs/bcachefs/bcachefs_ioctl.h
@@ -113,7 +113,7 @@ struct bch_ioctl_query_uuid {
  * may be either offline or offline.
  *
  * Will fail removing @dev would leave us with insufficient read write devices
- * or degraded/unavailable data, unless the approprate BCH_FORCE_IF_* flags are
+ * or degraded/unavailable data, unless the appropriate BCH_FORCE_IF_* flags are
  * set.
  */
 
@@ -136,7 +136,7 @@ struct bch_ioctl_query_uuid {
  *
  * Will fail (similarly to BCH_IOCTL_DISK_SET_STATE) if offlining @dev would
  * leave us with insufficient read write devices or degraded/unavailable data,
- * unless the approprate BCH_FORCE_IF_* flags are set.
+ * unless the appropriate BCH_FORCE_IF_* flags are set.
  */
 
 struct bch_ioctl_disk {
diff --git a/fs/bcachefs/btree/bkey_methods.c b/fs/bcachefs/btree/bkey_methods.c
index a694adcca..95a92696a 100644
--- a/fs/bcachefs/btree/bkey_methods.c
+++ b/fs/bcachefs/btree/bkey_methods.c
@@ -89,7 +89,7 @@ void bch2_set_bkey_error(struct bch_fs *c, struct bkey_i *k, enum bch_key_type_e
 {
 	k->k.type = KEY_TYPE_error;
 
-	if (!bch2_request_incompat_feature(c, bcachefs_metadata_version_extented_key_type_error)) {
+	if (!bch2_request_incompat_feature(c, bcachefs_metadata_version_extended_key_type_error)) {
 		set_bkey_val_bytes(&k->k, sizeof(struct bch_error));
 
 		struct bkey_i_error *e = bkey_i_to_error(k);
@@ -389,7 +389,7 @@ bool bch2_bkey_merge(struct bch_fs *c, struct bkey_s l, struct bkey_s_c r)
 	const struct bkey_ops *ops = bch2_bkey_type_ops(l.k->type);
 
 	return ops->key_merge &&
-		bch2_bkey_maybe_mergable(l.k, r.k) &&
+		bch2_bkey_maybe_mergeable(l.k, r.k) &&
 		(u64) l.k->size + r.k->size <= KEY_SIZE_MAX &&
 		!static_branch_unlikely(&bch2_key_merging_disabled) &&
 		ops->key_merge(c, l, r);
diff --git a/fs/bcachefs/btree/bkey_methods.h b/fs/bcachefs/btree/bkey_methods.h
index 3af1f35b4..08e74d8be 100644
--- a/fs/bcachefs/btree/bkey_methods.h
+++ b/fs/bcachefs/btree/bkey_methods.h
@@ -65,7 +65,7 @@ void bch2_bkey_val_to_text(struct printbuf *, struct bch_fs *,
 
 void bch2_bkey_swab_val(const struct bch_fs *c, struct bkey_s);
 
-static inline bool bch2_bkey_maybe_mergable(const struct bkey *l, const struct bkey *r)
+static inline bool bch2_bkey_maybe_mergeable(const struct bkey *l, const struct bkey *r)
 {
 	return l->type == r->type &&
 		!bversion_cmp(l->bversion, r->bversion) &&
diff --git a/fs/bcachefs/btree/bset.h b/fs/bcachefs/btree/bset.h
index f626d2ff8..b51144c7f 100644
--- a/fs/bcachefs/btree/bset.h
+++ b/fs/bcachefs/btree/bset.h
@@ -45,7 +45,7 @@
  * 4 in memory - we lazily resort as needed.
  *
  * We implement code here for creating and maintaining auxiliary search trees
- * (described below) for searching an individial bset, and on top of that we
+ * (described below) for searching an individual bset, and on top of that we
  * implement a btree iterator.
  *
  * BTREE ITERATOR:
@@ -179,7 +179,7 @@ static inline enum bset_aux_tree_type bset_aux_tree_type(const struct bset_tree
  * memory if it was 128.
  *
  * It definites the number of bytes (in struct bset) per struct bkey_float in
- * the auxiliar search tree - when we're done searching the bset_float tree we
+ * the auxiliary search tree - when we're done searching the bset_float tree we
  * have this many bytes left that we do a linear search over.
  *
  * Since (after level 5) every level of the bset_tree is on a new cacheline,
diff --git a/fs/bcachefs/btree/cache.c b/fs/bcachefs/btree/cache.c
index ce2799114..2a7196782 100644
--- a/fs/bcachefs/btree/cache.c
+++ b/fs/bcachefs/btree/cache.c
@@ -1212,7 +1212,7 @@ struct btree *bch2_btree_node_get(struct btree_trans *trans, struct btree_path *
 	/*
 	 * Check b->hash_val _before_ calling btree_node_lock() - this might not
 	 * be the node we want anymore, and trying to lock the wrong node could
-	 * cause an unneccessary transaction restart:
+	 * cause an unnecessary transaction restart:
 	 */
 	if (unlikely(!c->opts.btree_node_mem_ptr_optimization ||
 		     !b ||
diff --git a/fs/bcachefs/btree/interior.h b/fs/bcachefs/btree/interior.h
index 952a8cd45..1996395f8 100644
--- a/fs/bcachefs/btree/interior.h
+++ b/fs/bcachefs/btree/interior.h
@@ -117,7 +117,7 @@ struct btree_update {
 	struct keylist			parent_keys;
 	/*
 	 * Enough room for btree_split's keys without realloc - btree node
-	 * pointers never have crc/compression info, so we only need to acount
+	 * pointers never have crc/compression info, so we only need to account
 	 * for the pointers for three keys
 	 */
 	u64				inline_keys[BKEY_BTREE_PTR_U64s_MAX * 3];
diff --git a/fs/bcachefs/btree/iter.c b/fs/bcachefs/btree/iter.c
index f763449db..2a6b8908a 100644
--- a/fs/bcachefs/btree/iter.c
+++ b/fs/bcachefs/btree/iter.c
@@ -3479,7 +3479,7 @@ u32 bch2_trans_begin(struct btree_trans *trans)
 
 		/*
 		 * If the transaction wasn't restarted, we're presuming to be
-		 * doing something new: dont keep iterators excpt the ones that
+		 * doing something new: dont keep iterators except the ones that
 		 * are in use - except for the subvolumes btree:
 		 */
 		if (!trans->restarted && path->btree_id != BTREE_ID_subvolumes)
@@ -3623,7 +3623,7 @@ struct btree_trans *__bch2_trans_get(struct bch_fs *c, unsigned fn_idx)
 
 	*trans_paths_nr(trans->paths) = BTREE_ITER_INITIAL;
 
-	/* Reserve path 0 for our sentinal value */
+	/* Reserve path 0 for our sentinel value */
 	trans->paths_allocated[0] = 1;
 
 	static struct lock_class_key lockdep_key;
diff --git a/fs/bcachefs/btree/update.c b/fs/bcachefs/btree/update.c
index 5f2de66b4..f03c24b2e 100644
--- a/fs/bcachefs/btree/update.c
+++ b/fs/bcachefs/btree/update.c
@@ -250,7 +250,7 @@ static int bch2_trans_update_extent(struct btree_trans *trans,
 		goto out;
 
 	if (bkey_eq(k.k->p, bkey_start_pos(&insert->k))) {
-		if (bch2_bkey_maybe_mergable(k.k, &insert->k))
+		if (bch2_bkey_maybe_mergeable(k.k, &insert->k))
 			try(extent_front_merge(trans, &iter, k, &insert, flags));
 
 		goto next;
@@ -293,7 +293,7 @@ static int bch2_trans_update_extent(struct btree_trans *trans,
 			goto out;
 	}
 
-	if (bch2_bkey_maybe_mergable(&insert->k, k.k))
+	if (bch2_bkey_maybe_mergeable(&insert->k, k.k))
 		try(extent_back_merge(trans, &iter, insert, k));
 out:
 	return !bkey_deleted(&insert->k)
diff --git a/fs/bcachefs/data/checksum.c b/fs/bcachefs/data/checksum.c
index 999024202..9b587887f 100644
--- a/fs/bcachefs/data/checksum.c
+++ b/fs/bcachefs/data/checksum.c
@@ -20,7 +20,7 @@
 /*
  * bch2_checksum state is an abstraction of the checksum state calculated over different pages.
  * it features page merging without having the checksum algorithm lose its state.
- * for native checksum aglorithms (like crc), a default seed value will do.
+ * for native checksum algorithms (like crc), a default seed value will do.
  * for hash-like algorithms, a state needs to be stored
  */
 
diff --git a/fs/bcachefs/data/reconcile/work.c b/fs/bcachefs/data/reconcile/work.c
index 1abe4354b..a761eb56d 100644
--- a/fs/bcachefs/data/reconcile/work.c
+++ b/fs/bcachefs/data/reconcile/work.c
@@ -1322,7 +1322,7 @@ void bch2_reconcile_stop(struct bch_fs *c)
 	c->reconcile.thread = NULL;
 
 	if (p) {
-		/* for sychronizing with bch2_reconcile_wakeup() */
+		/* for synchronizing with bch2_reconcile_wakeup() */
 		synchronize_rcu();
 
 		kthread_stop(p);
diff --git a/fs/bcachefs/debug/async_objs.c b/fs/bcachefs/debug/async_objs.c
index 6717deadd..9bec6c9e1 100644
--- a/fs/bcachefs/debug/async_objs.c
+++ b/fs/bcachefs/debug/async_objs.c
@@ -1,7 +1,7 @@
 // SPDX-License-Identifier: GPL-2.0
 /*
  * Async obj debugging: keep asynchronous objects on (very fast) lists, make
- * them visibile in debugfs:
+ * them visible in debugfs:
  */
 
 #include "bcachefs.h"
diff --git a/fs/bcachefs/fs/check.c b/fs/bcachefs/fs/check.c
index 6b6da8406..c32fef694 100644
--- a/fs/bcachefs/fs/check.c
+++ b/fs/bcachefs/fs/check.c
@@ -1886,7 +1886,7 @@ int bch2_fix_reflink_p(struct bch_fs *c)
 			fix_reflink_p_key(trans, &iter, k));
 }
 
-/* translate to return code of fsck commad - man(8) fsck */
+/* translate to return code of fsck command - man(8) fsck */
 int bch2_fs_fsck_errcode(struct bch_fs *c, struct printbuf *msg)
 {
 	int ret = 0;
diff --git a/fs/bcachefs/fs/inode.c b/fs/bcachefs/fs/inode.c
index 5a0925edd..ef1105811 100644
--- a/fs/bcachefs/fs/inode.c
+++ b/fs/bcachefs/fs/inode.c
@@ -1367,7 +1367,7 @@ int bch2_delete_dead_inodes(struct bch_fs *c)
 	/*
 	 * if we ran check_inodes() unlinked inodes will have already been
 	 * cleaned up but the write buffer will be out of sync; therefore we
-	 * alway need a write buffer flush
+	 * always need a write buffer flush
 	 *
 	 * Weird transaction restart handling here because on successful delete,
 	 * bch2_inode_rm_snapshot() will return a nested transaction restart,
diff --git a/fs/bcachefs/init/passes.c b/fs/bcachefs/init/passes.c
index 5e49ca218..1a88cecf9 100644
--- a/fs/bcachefs/init/passes.c
+++ b/fs/bcachefs/init/passes.c
@@ -554,7 +554,7 @@ int bch2_run_recovery_passes(struct bch_fs *c, u64 orig_passes_to_run, bool fail
 
 		if (r->rewound_to) {
 			r->rewound_from	= max(r->rewound_from, pass);
-			/* Restore r->current_passses up to and including r->rewound_to */
+			/* Restore r->current_passes up to and including r->rewound_to */
 			r->current_passes |= orig_passes_to_run & (~0ULL << r->rewound_to);
 			r->rewound_to = 0;
 		} else if (!ret2) {
diff --git a/fs/bcachefs/init/progress.h b/fs/bcachefs/init/progress.h
index 877b0e311..74dbee395 100644
--- a/fs/bcachefs/init/progress.h
+++ b/fs/bcachefs/init/progress.h
@@ -8,7 +8,7 @@
  * Lame progress indicators
  *
  * We don't like to use these because they print to the dmesg console, which is
- * spammy - we much prefer to be wired up to a userspace programm (e.g. via
+ * spammy - we much prefer to be wired up to a userspace program (e.g. via
  * thread_with_file) and have it print the progress indicator.
  *
  * But some code is old and doesn't support that, or runs in a context where
diff --git a/fs/bcachefs/journal/journal.c b/fs/bcachefs/journal/journal.c
index 1a680c9a7..2242bdb89 100644
--- a/fs/bcachefs/journal/journal.c
+++ b/fs/bcachefs/journal/journal.c
@@ -123,7 +123,7 @@ journal_error_check_stuck(struct journal *j, int error, unsigned flags)
 	CLASS(bch_log_msg, msg)(c);
 	msg.m.suppress = true; /* only print once, when we go ERO */
 
-	prt_printf(&msg.m, "Journal stuck! Hava a pre-reservation but journal full (error %s)",
+	prt_printf(&msg.m, "Journal stuck! Have a pre-reservation but journal full (error %s)",
 		   bch2_err_str(error));
 	bch2_journal_debug_to_text(&msg.m, j);
 
@@ -879,7 +879,7 @@ bool bch2_journal_noflush_seq(struct journal *j, u64 start, u64 end)
 	     unwritten_seq++) {
 		struct journal_buf *buf = journal_seq_to_buf(j, unwritten_seq);
 
-		/* journal flush already in flight, or flush requseted */
+		/* journal flush already in flight, or flush requested */
 		if (buf->must_flush)
 			return false;
 
diff --git a/fs/bcachefs/journal/types.h b/fs/bcachefs/journal/types.h
index 0d2ec3e04..5a43f619a 100644
--- a/fs/bcachefs/journal/types.h
+++ b/fs/bcachefs/journal/types.h
@@ -130,7 +130,7 @@ union journal_res_state {
 #define JOURNAL_ENTRY_SIZE_MAX		(4U  << 20) /* 4M */
 
 /*
- * We stash some journal state as sentinal values in cur_entry_offset:
+ * We stash some journal state as sentinel values in cur_entry_offset:
  * note - cur_entry_offset is in units of u64s
  */
 #define JOURNAL_ENTRY_OFFSET_MAX	((1U << 22) - 1)
@@ -212,7 +212,7 @@ struct journal {
 	darray_u64		early_journal_entries;
 
 	/*
-	 * Protects journal_buf->data, when accessing without a jorunal
+	 * Protects journal_buf->data, when accessing without a journal
 	 * reservation: for synchronization between the btree write buffer code
 	 * and the journal write path:
 	 */
diff --git a/fs/bcachefs/journal/write.c b/fs/bcachefs/journal/write.c
index 4d1449a3a..8e9ee9349 100644
--- a/fs/bcachefs/journal/write.c
+++ b/fs/bcachefs/journal/write.c
@@ -289,7 +289,7 @@ static CLOSURE_CALLBACK(journal_write_done)
 				 * can return an error if appending to
 				 * replicas_refs failed, but we don't
 				 * care - it's a preallocated darray so
-				 * it'll allways be able to do some
+				 * it'll always be able to do some
 				 * work, and we have to retry anyways,
 				 * because we have to drop j->lock to
 				 * put the replicas refs before updating
diff --git a/fs/bcachefs/sb/io.c b/fs/bcachefs/sb/io.c
index 849d71bc5..80743a474 100644
--- a/fs/bcachefs/sb/io.c
+++ b/fs/bcachefs/sb/io.c
@@ -479,7 +479,7 @@ int bch2_sb_validate(struct bch_sb *sb, struct bch_opts *opts, u64 read_offset,
 		/*
 		 * Been seeing a bug where these are getting inexplicably
 		 * zeroed, so we're now validating them, but we have to be
-		 * careful not to preven people's filesystems from mounting:
+		 * careful not to prevent people's filesystems from mounting:
 		 */
 		if (!BCH_SB_JOURNAL_FLUSH_DELAY(sb))
 			SET_BCH_SB_JOURNAL_FLUSH_DELAY(sb, 1000);
diff --git a/fs/bcachefs/sb/members_format.h b/fs/bcachefs/sb/members_format.h
index 2c9f72653..941507d63 100644
--- a/fs/bcachefs/sb/members_format.h
+++ b/fs/bcachefs/sb/members_format.h
@@ -9,7 +9,7 @@
 #define BCH_SB_MEMBERS_MAX		64
 
 /*
- * Sentinal value - indicates a device that does not exist
+ * Sentinel value - indicates a device that does not exist
  */
 #define BCH_SB_MEMBER_INVALID		255
 
diff --git a/fs/bcachefs/snapshots/format.h b/fs/bcachefs/snapshots/format.h
index 72c28bd9c..a6e4fd33f 100644
--- a/fs/bcachefs/snapshots/format.h
+++ b/fs/bcachefs/snapshots/format.h
@@ -58,7 +58,7 @@ struct bch_snapshot {
  * something, key should be deleted)
  *
  * NO_KEYS: we don't remove interior snapshot nodes from snapshot trees at
- * runtime, since we can't do the adjustements for the depth/skiplist field
+ * runtime, since we can't do the adjustments for the depth/skiplist field
  * atomically - and that breaks e.g. is_ancestor(). Instead, we mark it to be
  * deleted at the next remount; this tells us that we don't need to run the full
  * delete_dead_snapshots().
@@ -78,7 +78,7 @@ LE32_BITMASK(BCH_SNAPSHOT_NO_KEYS,	struct bch_snapshot, flags,  3,  4)
 /*
  * Snapshot trees:
  *
- * The snapshot_trees btree gives us persistent indentifier for each tree of
+ * The snapshot_trees btree gives us persistent identifier for each tree of
  * bch_snapshot nodes, and allow us to record and easily find the root/master
  * subvolume that other snapshots were created from:
  */
diff --git a/fs/bcachefs/util/eytzinger.h b/fs/bcachefs/util/eytzinger.h
index b14ae1ff7..43a66cab5 100644
--- a/fs/bcachefs/util/eytzinger.h
+++ b/fs/bcachefs/util/eytzinger.h
@@ -13,7 +13,7 @@
 #endif
 
 /*
- * Traversal for trees in eytzinger layout - a full binary tree layed out in an
+ * Traversal for trees in eytzinger layout - a full binary tree laid out in an
  * array.
  *
  * Consider using an eytzinger tree any time you would otherwise be doing binary
diff --git a/fs/bcachefs/util/mean_and_variance.h b/fs/bcachefs/util/mean_and_variance.h
index 47e4a3c3d..281d6f9d1 100644
--- a/fs/bcachefs/util/mean_and_variance.h
+++ b/fs/bcachefs/util/mean_and_variance.h
@@ -152,7 +152,7 @@ struct mean_and_variance {
 	u128_u	sum_squares;
 };
 
-/* expontentially weighted variant */
+/* exponentially weighted variant */
 struct mean_and_variance_weighted {
 	s64	mean;
 	u64	variance;
@@ -200,4 +200,4 @@ u64 mean_and_variance_weighted_get_variance(struct mean_and_variance_weighted s,
 u32 mean_and_variance_weighted_get_stddev(struct mean_and_variance_weighted s,
 		u8 weight);
 
-#endif // MEAN_AND_VAIRANCE_H_
+#endif // MEAN_AND_VARIANCE_H_
diff --git a/fs/bcachefs/util/mean_and_variance_test.c b/fs/bcachefs/util/mean_and_variance_test.c
index e9d9c0212..86f38db11 100644
--- a/fs/bcachefs/util/mean_and_variance_test.c
+++ b/fs/bcachefs/util/mean_and_variance_test.c
@@ -25,7 +25,7 @@ static void mean_and_variance_basic_test(struct kunit *test)
 }
 
 /*
- * Test values computed using a spreadsheet from the psuedocode at the bottom:
+ * Test values computed using a spreadsheet from the pseudocode at the bottom:
  * https://fanf2.user.srcf.net/hermes/doc/antiforgery/stats.pdf
  */
 
diff --git a/fs/bcachefs/util/printbuf.c b/fs/bcachefs/util/printbuf.c
index 0a6f44fd5..80e101915 100644
--- a/fs/bcachefs/util/printbuf.c
+++ b/fs/bcachefs/util/printbuf.c
@@ -417,7 +417,7 @@ void bch2_prt_tab_rjust(struct printbuf *buf)
  * @str:	string to print
  * @count:	number of bytes to print
  *
- * The following contol characters are handled as so:
+ * The following control characters are handled as so:
  *   \n: prt_newline	newline that obeys current indent level
  *   \t: prt_tab	advance to next tabstop
  *   \r: prt_tab_rjust	advance to next tabstop, with right justification
diff --git a/fs/bcachefs/util/printbuf.h b/fs/bcachefs/util/printbuf.h
index 41e8d71d9..fe41736bd 100644
--- a/fs/bcachefs/util/printbuf.h
+++ b/fs/bcachefs/util/printbuf.h
@@ -25,7 +25,7 @@
  * everything to the kernel log buffer, and then those pretty-printers can be
  * used by other code that outputs to kernel log, sysfs, debugfs, etc.
  *
- * Memory allocation: Outputing to a printbuf may allocate memory. This
+ * Memory allocation: Outputting to a printbuf may allocate memory. This
  * allocation is done with GFP_KERNEL, by default: use the newer
  * memalloc_*_(save|restore) functions as needed.
  *
@@ -56,7 +56,7 @@
  * next tabstop - right justifying it.
  *
  * Make sure you use prt_newline() instead of \n in the format string for indent
- * level and tabstops to work corretly.
+ * level and tabstops to work correctly.
  *
  * Output units: printbuf->units exists to tell pretty-printers how to output
  * numbers: a raw value (e.g. directly from a superblock field), as bytes, or as
diff --git a/fs/bcachefs/util/rcu_pending.c b/fs/bcachefs/util/rcu_pending.c
index 964ed493c..794f2d985 100644
--- a/fs/bcachefs/util/rcu_pending.c
+++ b/fs/bcachefs/util/rcu_pending.c
@@ -349,7 +349,7 @@ rcu_pending_enqueue_list(struct rcu_pending_pcpu *p, rcu_gp_poll_state_t seq,
 			/*
 			 * kvfree_rcu_mightsleep(): we weren't passed an
 			 * rcu_head, but we need one: use the low bit of the
-			 * ponter to free to flag that the head needs to be
+			 * pointer to free to flag that the head needs to be
 			 * freed as well:
 			 */
 			ptr = (void *)(((unsigned long) ptr)|1UL);
diff --git a/fs/bcachefs/util/siphash.h b/fs/bcachefs/util/siphash.h
index 3dfaf34a4..7744b4fe0 100644
--- a/fs/bcachefs/util/siphash.h
+++ b/fs/bcachefs/util/siphash.h
@@ -36,7 +36,7 @@
  * optimized for speed on short messages returning a 64bit hash/digest value.
  *
  * The number of rounds is defined during the initialization:
- *  SipHash24_Init() for the fast and resonable strong version
+ *  SipHash24_Init() for the fast and reasonably strong version
  *  SipHash48_Init() for the strong version (half as fast)
  *
  * struct SIPHASH_CTX ctx;
diff --git a/fs/bcachefs/util/six.c b/fs/bcachefs/util/six.c
index 08083d6ca..cdf7efb0c 100644
--- a/fs/bcachefs/util/six.c
+++ b/fs/bcachefs/util/six.c
@@ -111,7 +111,7 @@ static inline unsigned pcpu_read_count(struct six_lock *lock)
  * Returns 1 on success, 0 on failure
  *
  * In percpu reader mode, a failed trylock may cause a spurious trylock failure
- * for anoter thread taking the competing lock type, and we may havve to do a
+ * for another thread taking the competing lock type, and we may have to do a
  * wakeup: when a wakeup is required, we return -1 - wakeup_type.
  */
 static int __do_six_trylock(struct six_lock *lock, enum six_lock_type type,
@@ -591,7 +591,7 @@ static void do_six_unlock_type(struct six_lock *lock, enum six_lock_type type)
  * @type:	SIX_LOCK_read, SIX_LOCK_intent, or SIX_LOCK_write
  * @ip:		ip parameter for lockdep/lockstat, i.e. _THIS_IP_
  *
- * When a lock is held multiple times (because six_lock_incement()) was used),
+ * When a lock is held multiple times (because six_lock_increment()) was used),
  * this decrements the 'lock held' counter by one.
  *
  * For example:
@@ -632,7 +632,7 @@ EXPORT_SYMBOL_GPL(six_unlock_ip);
 
 /**
  * six_lock_downgrade - convert an intent lock to a read lock
- * @lock:	lock to dowgrade
+ * @lock:	lock to downgrade
  *
  * @lock will have read count incremented and intent count decremented
  */
diff --git a/fs/bcachefs/util/six.h b/fs/bcachefs/util/six.h
index 59b851cf8..2f55de935 100644
--- a/fs/bcachefs/util/six.h
+++ b/fs/bcachefs/util/six.h
@@ -79,7 +79,7 @@
  *     six_unlock_read(&foo->lock);
  *   foo->lock is now fully unlocked.
  *
- *   Since the intent state supercedes read, it's legal to increment the read
+ *   Since the intent state supersedes read, it's legal to increment the read
  *   counter when holding an intent lock, but not the reverse.
  *
  *   A lock may only be held once for write: six_lock_increment(.., SIX_LOCK_write)
@@ -298,7 +298,7 @@ void six_unlock_ip(struct six_lock *lock, enum six_lock_type type, unsigned long
  * @lock:	lock to unlock
  * @type:	SIX_LOCK_read, SIX_LOCK_intent, or SIX_LOCK_write
  *
- * When a lock is held multiple times (because six_lock_incement()) was used),
+ * When a lock is held multiple times (because six_lock_increment()) was used),
  * this decrements the 'lock held' counter by one.
  *
  * For example:
diff --git a/fs/bcachefs/util/time_stats.h b/fs/bcachefs/util/time_stats.h
index 7c7069673..37cf1a286 100644
--- a/fs/bcachefs/util/time_stats.h
+++ b/fs/bcachefs/util/time_stats.h
@@ -12,7 +12,7 @@
  *  - sum of all event durations
  *  - average event duration, standard and weighted
  *  - standard deviation of event durations, standard and weighted
- * and analagous statistics for the frequency of events
+ * and analogous statistics for the frequency of events
  *
  * We provide both mean and weighted mean (exponentially weighted), and standard
  * deviation and weighted standard deviation, to give an efficient-to-compute
diff --git a/fs/bcachefs/util/util.c b/fs/bcachefs/util/util.c
index d538758b0..a444ec7cd 100644
--- a/fs/bcachefs/util/util.c
+++ b/fs/bcachefs/util/util.c
@@ -501,7 +501,7 @@ void bch2_ratelimit_increment(struct bch_ratelimit *d, u64 done)
 /* pd controller: */
 
 /*
- * Updates pd_controller. Attempts to scale inputed values to units per second.
+ * Updates pd_controller. Attempts to scale inputted values to units per second.
  * @target: desired value
  * @actual: current value
  *
diff --git a/fs/bcachefs/vfs/io.c b/fs/bcachefs/vfs/io.c
index 8ea29763d..91d0922a1 100644
--- a/fs/bcachefs/vfs/io.c
+++ b/fs/bcachefs/vfs/io.c
@@ -945,7 +945,7 @@ loff_t bch2_remap_file_range(struct file *file_src, loff_t pos_src,
 		goto err;
 
 	/*
-	 * due to alignment, we might have remapped slightly more than requsted
+	 * due to alignment, we might have remapped slightly more than requested
 	 */
 	ret = min((u64) ret << 9, (u64) len);
 
-- 
2.52.0



      parent reply	other threads:[~2026-01-24 11:19 UTC|newest]

Thread overview: 3+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2026-01-24 11:19 [PATCH 0/2] Various typo fixes Jul Lang
2026-01-24 11:19 ` [PATCH 1/2] docs: fix typo in bch2_trans_begin inline comment Jul Lang
2026-01-24 11:19 ` Jul Lang [this message]

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20260124111913.6263-3-jullanggit@proton.me \
    --to=jullanggit@proton.me \
    --cc=kent.overstreet@linux.dev \
    --cc=linux-bcachefs@vger.kernel.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox