bcachefs: Fix assorted checkpatch nits
authorKent Overstreet <kent.overstreet@linux.dev>
Mon, 7 Aug 2023 16:04:05 +0000 (12:04 -0400)
committerKent Overstreet <kent.overstreet@linux.dev>
Sun, 22 Oct 2023 21:10:10 +0000 (17:10 -0400)
Signed-off-by: Kent Overstreet <kent.overstreet@linux.dev>
22 files changed:
fs/bcachefs/alloc_types.h
fs/bcachefs/btree_gc.c
fs/bcachefs/btree_io.h
fs/bcachefs/btree_iter.c
fs/bcachefs/btree_update.h
fs/bcachefs/buckets.c
fs/bcachefs/chardev.h
fs/bcachefs/checksum.c
fs/bcachefs/compress.c
fs/bcachefs/extents.c
fs/bcachefs/fs-io.c
fs/bcachefs/io.c
fs/bcachefs/io.h
fs/bcachefs/journal.c
fs/bcachefs/journal_io.c
fs/bcachefs/journal_reclaim.c
fs/bcachefs/recovery.c
fs/bcachefs/super-io.c
fs/bcachefs/super-io.h
fs/bcachefs/util.c
fs/bcachefs/util.h
fs/bcachefs/varint.c

index 804a843..b91b7a4 100644 (file)
@@ -105,7 +105,7 @@ struct write_point {
                struct dev_stripe_state stripe;
 
                u64                     sectors_allocated;
-       } __attribute__((__aligned__(SMP_CACHE_BYTES)));
+       } __aligned(SMP_CACHE_BYTES);
 
        struct {
                struct work_struct      index_update_work;
@@ -116,7 +116,7 @@ struct write_point {
                enum write_point_state  state;
                u64                     last_state_change;
                u64                     time[WRITE_POINT_STATE_NR];
-       } __attribute__((__aligned__(SMP_CACHE_BYTES)));
+       } __aligned(SMP_CACHE_BYTES);
 };
 
 struct write_point_specifier {
index 3f5b4d8..1a749d4 100644 (file)
@@ -535,7 +535,7 @@ int bch2_check_topology(struct bch_fs *c)
 
        bch2_trans_init(&trans, c, 0, 0);
 
-       for (i = 0; i < btree_id_nr_alive(c)&& !ret; i++) {
+       for (i = 0; i < btree_id_nr_alive(c) && !ret; i++) {
                struct btree_root *r = bch2_btree_id_root(c, i);
 
                if (!r->alive)
index 0cadf65..cd99bbb 100644 (file)
@@ -143,8 +143,8 @@ enum btree_write_flags {
        __BTREE_WRITE_ONLY_IF_NEED = BTREE_WRITE_TYPE_BITS,
        __BTREE_WRITE_ALREADY_STARTED,
 };
-#define BTREE_WRITE_ONLY_IF_NEED       (1U << __BTREE_WRITE_ONLY_IF_NEED )
-#define BTREE_WRITE_ALREADY_STARTED    (1U << __BTREE_WRITE_ALREADY_STARTED)
+#define BTREE_WRITE_ONLY_IF_NEED       BIT(__BTREE_WRITE_ONLY_IF_NEED)
+#define BTREE_WRITE_ALREADY_STARTED    BIT(__BTREE_WRITE_ALREADY_STARTED)
 
 void __bch2_btree_node_write(struct bch_fs *, struct btree *, unsigned);
 void bch2_btree_node_write(struct bch_fs *, struct btree *,
index ecc123b..feb23e9 100644 (file)
@@ -1008,7 +1008,7 @@ retry_all:
        /*
         * We used to assert that all paths had been traversed here
         * (path->uptodate < BTREE_ITER_NEED_TRAVERSE); however, since
-        * path->Should_be_locked is not set yet, we we might have unlocked and
+        * path->should_be_locked is not set yet, we might have unlocked and
         * then failed to relock a path - that's fine.
         */
 err:
@@ -2738,9 +2738,9 @@ void bch2_trans_node_iter_init(struct btree_trans *trans,
                               unsigned depth,
                               unsigned flags)
 {
-       flags |= BTREE_ITER_NOT_EXTENTS;
-       flags |= __BTREE_ITER_ALL_SNAPSHOTS;
-       flags |= BTREE_ITER_ALL_SNAPSHOTS;
+       flags |= BTREE_ITER_NOT_EXTENTS;
+       flags |= __BTREE_ITER_ALL_SNAPSHOTS;
+       flags |= BTREE_ITER_ALL_SNAPSHOTS;
 
        bch2_trans_iter_init_common(trans, iter, btree_id, pos, locks_want, depth,
                               __bch2_btree_iter_flags(trans, btree_id, flags),
index 2281140..901c42b 100644 (file)
@@ -268,10 +268,10 @@ static inline struct bkey_i *__bch2_bkey_get_mut_noupdate(struct btree_trans *tr
 {
        struct bkey_s_c k = __bch2_bkey_get_iter(trans, iter,
                                btree_id, pos, flags|BTREE_ITER_INTENT, type);
-       struct bkey_i *ret = unlikely(IS_ERR(k.k))
+       struct bkey_i *ret = IS_ERR(k.k)
                ? ERR_CAST(k.k)
                : __bch2_bkey_make_mut_noupdate(trans, k, 0, min_bytes);
-       if (unlikely(IS_ERR(ret)))
+       if (IS_ERR(ret))
                bch2_trans_iter_exit(trans, iter);
        return ret;
 }
index 7bb7f0c..c02c8c9 100644 (file)
@@ -1924,6 +1924,7 @@ static int __bch2_trans_mark_dev_sb(struct btree_trans *trans,
 int bch2_trans_mark_dev_sb(struct bch_fs *c, struct bch_dev *ca)
 {
        int ret = bch2_trans_run(c, __bch2_trans_mark_dev_sb(&trans, ca));
+
        if (ret)
                bch_err_fn(c, ret);
        return ret;
index 3a4890d..0f563ca 100644 (file)
@@ -17,7 +17,7 @@ int __init bch2_chardev_init(void);
 static inline long bch2_fs_ioctl(struct bch_fs *c,
                                unsigned cmd, void __user * arg)
 {
-       return -ENOSYS;
+       return -ENOTTY;
 }
 
 static inline void bch2_fs_chardev_exit(struct bch_fs *c) {}
index bf03d42..76cf2e7 100644 (file)
@@ -265,9 +265,10 @@ static struct bch_csum __bch2_checksum_bio(struct bch_fs *c, unsigned type,
 
 #ifdef CONFIG_HIGHMEM
                __bio_for_each_segment(bv, bio, *iter, *iter) {
-                       void *p = kmap_atomic(bv.bv_page) + bv.bv_offset;
+                       void *p = kmap_local_page(bv.bv_page) + bv.bv_offset;
+
                        bch2_checksum_update(&state, p, bv.bv_len);
-                       kunmap_atomic(p);
+                       kunmap_local(p);
                }
 #else
                __bio_for_each_bvec(bv, bio, *iter, *iter)
@@ -287,10 +288,10 @@ static struct bch_csum __bch2_checksum_bio(struct bch_fs *c, unsigned type,
 
 #ifdef CONFIG_HIGHMEM
                __bio_for_each_segment(bv, bio, *iter, *iter) {
-                       void *p = kmap_atomic(bv.bv_page) + bv.bv_offset;
+                       void *p = kmap_local_page(bv.bv_page) + bv.bv_offset;
 
                        crypto_shash_update(desc, p, bv.bv_len);
-                       kunmap_atomic(p);
+                       kunmap_local(p);
                }
 #else
                __bio_for_each_bvec(bv, bio, *iter, *iter)
@@ -427,8 +428,9 @@ int bch2_rechecksum_bio(struct bch_fs *c, struct bio *bio,
                                extent_nonce(version, crc_old), bio);
 
        if (bch2_crc_cmp(merged, crc_old.csum) && !c->opts.no_data_io) {
-               bch_err(c, "checksum error in bch2_rechecksum_bio() (memory corruption or bug?)\n"
+               bch_err(c, "checksum error in %s() (memory corruption or bug?)\n"
                        "expected %0llx:%0llx got %0llx:%0llx (old type %s new type %s)",
+                       __func__,
                        crc_old.csum.hi,
                        crc_old.csum.lo,
                        merged.hi,
index c9ca7cc..6b17f7c 100644 (file)
@@ -643,7 +643,8 @@ static int __bch2_fs_compress_init(struct bch_fs *c, u64 features)
 static u64 compression_opt_to_feature(unsigned v)
 {
        unsigned type = bch2_compression_decode(v).type;
-       return 1ULL << bch2_compression_opt_to_feature[type];
+
+       return BIT_ULL(bch2_compression_opt_to_feature[type]);
 }
 
 int bch2_fs_compress_init(struct bch_fs *c)
index 7a3f42f..d7f74db 100644 (file)
@@ -517,7 +517,7 @@ static void bch2_extent_crc_pack(union bch_extent_crc *dst,
        switch (type) {
        case BCH_EXTENT_ENTRY_crc32:
                set_common_fields(dst->crc32, src);
-               dst->crc32.csum         = (u32 __force) *((__le32 *) &src.csum.lo);
+               dst->crc32.csum         = (u32 __force) *((__le32 *) &src.csum.lo);
                break;
        case BCH_EXTENT_ENTRY_crc64:
                set_common_fields(dst->crc64, src);
@@ -915,11 +915,11 @@ bool bch2_extents_match(struct bkey_s_c k1, struct bkey_s_c k2)
 
                bkey_for_each_ptr_decode(k1.k, ptrs1, p1, entry1)
                        bkey_for_each_ptr_decode(k2.k, ptrs2, p2, entry2)
-                       if (p1.ptr.dev          == p2.ptr.dev &&
-                           p1.ptr.gen          == p2.ptr.gen &&
-                           (s64) p1.ptr.offset + p1.crc.offset - bkey_start_offset(k1.k) ==
-                           (s64) p2.ptr.offset + p2.crc.offset - bkey_start_offset(k2.k))
-                               return true;
+                               if (p1.ptr.dev          == p2.ptr.dev &&
+                                   p1.ptr.gen          == p2.ptr.gen &&
+                                   (s64) p1.ptr.offset + p1.crc.offset - bkey_start_offset(k1.k) ==
+                                   (s64) p2.ptr.offset + p2.crc.offset - bkey_start_offset(k2.k))
+                                       return true;
 
                return false;
        } else {
index 917ad1c..40bfd0b 100644 (file)
@@ -2867,7 +2867,7 @@ static int __bch2_truncate_folio(struct bch_inode_info *inode,
 
                folio = __filemap_get_folio(mapping, index,
                                            FGP_LOCK|FGP_CREAT, GFP_KERNEL);
-               if (unlikely(IS_ERR_OR_NULL(folio))) {
+               if (IS_ERR_OR_NULL(folio)) {
                        ret = -ENOMEM;
                        goto out;
                }
index 5bacc6a..f42d9da 100644 (file)
@@ -2435,6 +2435,7 @@ static void __bch2_read_endio(struct work_struct *work)
 
                if (rbio->bounce) {
                        struct bvec_iter src_iter = src->bi_iter;
+
                        bio_copy_data_iter(dst, &dst_iter, src, &src_iter);
                }
        }
index 1476380..831e3f1 100644 (file)
@@ -52,7 +52,7 @@ enum __bch_write_flags {
 };
 
 enum bch_write_flags {
-#define x(f)   BCH_WRITE_##f = 1U << __BCH_WRITE_##f,
+#define x(f)   BCH_WRITE_##f = BIT(__BCH_WRITE_##f),
        BCH_WRITE_FLAGS()
 #undef x
 };
index 80a612c..055920c 100644 (file)
@@ -63,6 +63,7 @@ journal_seq_to_buf(struct journal *j, u64 seq)
 static void journal_pin_list_init(struct journal_entry_pin_list *p, int count)
 {
        unsigned i;
+
        for (i = 0; i < ARRAY_SIZE(p->list); i++)
                INIT_LIST_HEAD(&p->list[i]);
        INIT_LIST_HEAD(&p->flushed);
@@ -514,8 +515,7 @@ int bch2_journal_res_get_slowpath(struct journal *j, struct journal_res *res,
        int ret;
 
        closure_wait_event(&j->async_wait,
-                  (ret = __journal_res_get(j, res, flags)) !=
-                  -BCH_ERR_journal_res_get_blocked||
+                  (ret = __journal_res_get(j, res, flags)) != -BCH_ERR_journal_res_get_blocked ||
                   (flags & JOURNAL_RES_GET_NONBLOCK));
        return ret;
 }
index 798d4c8..42c9700 100644 (file)
@@ -1053,6 +1053,7 @@ found:
                bch_err(c, "cur_idx %u/%u", ja->cur_idx, ja->nr);
                for (i = 0; i < 3; i++) {
                        unsigned idx = (ja->cur_idx + ja->nr - 1 + i) % ja->nr;
+
                        bch_err(c, "bucket_seq[%u] = %llu", idx, ja->bucket_seq[idx]);
                }
                ja->sectors_free = 0;
@@ -1629,7 +1630,6 @@ static void do_journal_write(struct closure *cl)
        }
 
        continue_at(cl, journal_write_done, c->io_complete_wq);
-       return;
 }
 
 static void bch2_journal_entries_postprocess(struct bch_fs *c, struct jset *jset)
index 8de83e1..9a2a534 100644 (file)
@@ -345,7 +345,7 @@ static inline bool __journal_pin_drop(struct journal *j,
        list_del_init(&pin->list);
 
        /*
-        * Unpinning a journal entry make make journal_next_bucket() succeed, if
+        * Unpinning a journal entry may make journal_next_bucket() succeed, if
         * writing a new last_seq will now make another bucket available:
         */
        return atomic_dec_and_test(&pin_list->count) &&
index 35b67c5..5dbe1b2 100644 (file)
@@ -648,7 +648,7 @@ static int bch2_journal_replay(struct bch_fs *c)
        move_gap(keys->d, keys->nr, keys->size, keys->gap, keys->nr);
        keys->gap = keys->nr;
 
-       keys_sorted = kvmalloc_array(sizeof(*keys_sorted), keys->nr, GFP_KERNEL);
+       keys_sorted = kvmalloc_array(keys->nr, sizeof(*keys_sorted), GFP_KERNEL);
        if (!keys_sorted)
                return -BCH_ERR_ENOMEM_journal_replay;
 
@@ -1403,7 +1403,7 @@ use_clean:
        }
 
        c->journal_replay_seq_start     = last_seq;
-       c->journal_replay_seq_end       = blacklist_seq - 1;;
+       c->journal_replay_seq_end       = blacklist_seq - 1;
 
        if (c->opts.reconstruct_alloc) {
                c->sb.compat &= ~(1ULL << BCH_COMPAT_alloc_info);
index a58b975..405ea74 100644 (file)
@@ -553,7 +553,9 @@ static int __copy_super(struct bch_sb_handle *dst_handle, struct bch_sb *src)
                d = (src_f ? le32_to_cpu(src_f->u64s) : 0) -
                    (dst_f ? le32_to_cpu(dst_f->u64s) : 0);
                if (d > 0) {
-                       int ret = bch2_sb_realloc(dst_handle, le32_to_cpu(dst_handle->sb->u64s) + d);
+                       int ret = bch2_sb_realloc(dst_handle,
+                                       le32_to_cpu(dst_handle->sb->u64s) + d);
+
                        if (ret)
                                return ret;
 
index 904adea..6e59b01 100644 (file)
@@ -58,6 +58,7 @@ struct bch_sb_field_ops {
 static inline __le64 bch2_sb_magic(struct bch_fs *c)
 {
        __le64 ret;
+
        memcpy(&ret, &c->sb.uuid, sizeof(ret));
        return ret;
 }
index 6374d8a..e4f21fc 100644 (file)
@@ -216,6 +216,7 @@ u64 bch2_read_flag_list(char *opt, const char * const list[])
 
        while ((p = strsep(&s, ","))) {
                int flag = match_string(list, -1, p);
+
                if (flag < 0) {
                        ret = -1;
                        break;
@@ -797,9 +798,10 @@ void memcpy_to_bio(struct bio *dst, struct bvec_iter dst_iter, const void *src)
        struct bvec_iter iter;
 
        __bio_for_each_segment(bv, dst, iter, dst_iter) {
-               void *dstp = kmap_atomic(bv.bv_page);
+               void *dstp = kmap_local_page(bv.bv_page);
+
                memcpy(dstp + bv.bv_offset, src, bv.bv_len);
-               kunmap_atomic(dstp);
+               kunmap_local(dstp);
 
                src += bv.bv_len;
        }
@@ -811,9 +813,10 @@ void memcpy_from_bio(void *dst, struct bio *src, struct bvec_iter src_iter)
        struct bvec_iter iter;
 
        __bio_for_each_segment(bv, src, iter, src_iter) {
-               void *srcp = kmap_atomic(bv.bv_page);
+               void *srcp = kmap_local_page(bv.bv_page);
+
                memcpy(dst, srcp + bv.bv_offset, bv.bv_len);
-               kunmap_atomic(srcp);
+               kunmap_local(srcp);
 
                dst += bv.bv_len;
        }
index 8e37ce0..3cec617 100644 (file)
@@ -467,8 +467,10 @@ struct bch_pd_controller {
        s64                     last_change;
        s64                     last_target;
 
-       /* If true, the rate will not increase if bch2_ratelimit_delay()
-        * is not being called often enough. */
+       /*
+        * If true, the rate will not increase if bch2_ratelimit_delay()
+        * is not being called often enough.
+        */
        bool                    backpressure;
 };
 
@@ -604,6 +606,7 @@ static inline void __memcpy_u64s(void *dst, const void *src,
 {
 #ifdef CONFIG_X86_64
        long d0, d1, d2;
+
        asm volatile("rep ; movsq"
                     : "=&c" (d0), "=&D" (d1), "=&S" (d2)
                     : "0" (u64s), "1" (dst), "2" (src)
@@ -680,6 +683,7 @@ static inline void __memmove_u64s_up(void *_dst, const void *_src,
 
 #ifdef CONFIG_X86_64
        long d0, d1, d2;
+
        asm volatile("std ;\n"
                     "rep ; movsq\n"
                     "cld ;\n"
index ef030fc..2a2ab86 100644 (file)
@@ -59,6 +59,7 @@ int bch2_varint_decode(const u8 *in, const u8 *end, u64 *out)
 
        if (likely(bytes < 9)) {
                __le64 v_le = 0;
+
                memcpy(&v_le, in, bytes);
                v = le64_to_cpu(v_le);
                v >>= bytes;