struct btrfs_block_group_cache *cache)
 {
        struct rb_node *node;
+
        spin_lock(&root->fs_info->block_group_cache_lock);
+
+       /* If our block group was removed, we need a full search. */
+       if (RB_EMPTY_NODE(&cache->cache_node)) {
+               const u64 next_bytenr = cache->key.objectid + cache->key.offset;
+
+               spin_unlock(&root->fs_info->block_group_cache_lock);
+               btrfs_put_block_group(cache);
+               cache = btrfs_lookup_first_block_group(root->fs_info,
+                                                      next_bytenr);
+               return cache;
+       }
        node = rb_next(&cache->cache_node);
        btrfs_put_block_group(cache);
        if (node) {
        spin_lock(&root->fs_info->block_group_cache_lock);
        rb_erase(&block_group->cache_node,
                 &root->fs_info->block_group_cache_tree);
+       RB_CLEAR_NODE(&block_group->cache_node);
 
        if (root->fs_info->first_logical_byte == block_group->key.objectid)
                root->fs_info->first_logical_byte = (u64)-1;