/* * Copyright (C) STRATO AG 2011. All rights reserved. * * This program is free software; you can redistribute it and/or * modify it under the terms of the GNU General Public * License v2 as published by the Free Software Foundation. * * This program is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * General Public License for more details. * * You should have received a copy of the GNU General Public * License along with this program; if not, write to the * Free Software Foundation, Inc., 59 Temple Place - Suite 330, * Boston, MA 021110-1307, USA. */ /* * This module can be used to catch cases when the btrfs kernel * code executes write requests to the disk that bring the file * system in an inconsistent state. In such a state, a power-loss * or kernel panic event would cause that the data on disk is * lost or at least damaged. * * Code is added that examines all block write requests during * runtime (including writes of the super block). Three rules * are verified and an error is printed on violation of the * rules: * 1. It is not allowed to write a disk block which is * currently referenced by the super block (either directly * or indirectly). * 2. When a super block is written, it is verified that all * referenced (directly or indirectly) blocks fulfill the * following requirements: * 2a. All referenced blocks have either been present when * the file system was mounted, (i.e., they have been * referenced by the super block) or they have been * written since then and the write completion callback * was called and no write error was indicated and a * FLUSH request to the device where these blocks are * located was received and completed. * 2b. All referenced blocks need to have a generation * number which is equal to the parent's number. * * One issue that was found using this module was that the log * tree on disk became temporarily corrupted because disk blocks * that had been in use for the log tree had been freed and * reused too early, while being referenced by the written super * block. * * The search term in the kernel log that can be used to filter * on the existence of detected integrity issues is * "btrfs: attempt". * * The integrity check is enabled via mount options. These * mount options are only supported if the integrity check * tool is compiled by defining BTRFS_FS_CHECK_INTEGRITY. * * Example #1, apply integrity checks to all metadata: * mount /dev/sdb1 /mnt -o check_int * * Example #2, apply integrity checks to all metadata and * to data extents: * mount /dev/sdb1 /mnt -o check_int_data * * Example #3, apply integrity checks to all metadata and dump * the tree that the super block references to kernel messages * each time after a super block was written: * mount /dev/sdb1 /mnt -o check_int,check_int_print_mask=263 * * If the integrity check tool is included and activated in * the mount options, plenty of kernel memory is used, and * plenty of additional CPU cycles are spent. Enabling this * functionality is not intended for normal use. In most * cases, unless you are a btrfs developer who needs to verify * the integrity of (super)-block write requests, do not * enable the config option BTRFS_FS_CHECK_INTEGRITY to * include and compile the integrity check tool. * * Expect millions of lines of information in the kernel log with an * enabled check_int_print_mask. Therefore set LOG_BUF_SHIFT in the * kernel config to at least 26 (which is 64MB). Usually the value is * limited to 21 (which is 2MB) in init/Kconfig. The file needs to be * changed like this before LOG_BUF_SHIFT can be set to a high value: * config LOG_BUF_SHIFT * int "Kernel log buffer size (16 => 64KB, 17 => 128KB)" * range 12 30 */ #include <linux/sched.h> #include <linux/slab.h> #include <linux/buffer_head.h> #include <linux/mutex.h> #include <linux/genhd.h> #include <linux/blkdev.h> #include <linux/vmalloc.h> #include "ctree.h" #include "disk-io.h" #include "hash.h" #include "transaction.h" #include "extent_io.h" #include "volumes.h" #include "print-tree.h" #include "locking.h" #include "check-integrity.h" #include "rcu-string.h" #define BTRFSIC_BLOCK_HASHTABLE_SIZE 0x10000 #define BTRFSIC_BLOCK_LINK_HASHTABLE_SIZE 0x10000 #define BTRFSIC_DEV2STATE_HASHTABLE_SIZE 0x100 #define BTRFSIC_BLOCK_MAGIC_NUMBER 0x14491051 #define BTRFSIC_BLOCK_LINK_MAGIC_NUMBER 0x11070807 #define BTRFSIC_DEV2STATE_MAGIC_NUMBER 0x20111530 #define BTRFSIC_BLOCK_STACK_FRAME_MAGIC_NUMBER 20111300 #define BTRFSIC_TREE_DUMP_MAX_INDENT_LEVEL (200 - 6) /* in characters, * excluding " [...]" */ #define BTRFSIC_GENERATION_UNKNOWN ((u64)-1) /* * The definition of the bitmask fields for the print_mask. * They are specified with the mount option check_integrity_print_mask. */ #define BTRFSIC_PRINT_MASK_SUPERBLOCK_WRITE 0x00000001 #define BTRFSIC_PRINT_MASK_ROOT_CHUNK_LOG_TREE_LOCATION 0x00000002 #define BTRFSIC_PRINT_MASK_TREE_AFTER_SB_WRITE 0x00000004 #define BTRFSIC_PRINT_MASK_TREE_BEFORE_SB_WRITE 0x00000008 #define BTRFSIC_PRINT_MASK_SUBMIT_BIO_BH 0x00000010 #define BTRFSIC_PRINT_MASK_END_IO_BIO_BH 0x00000020 #define BTRFSIC_PRINT_MASK_VERBOSE 0x00000040 #define BTRFSIC_PRINT_MASK_VERY_VERBOSE 0x00000080 #define BTRFSIC_PRINT_MASK_INITIAL_TREE 0x00000100 #define BTRFSIC_PRINT_MASK_INITIAL_ALL_TREES 0x00000200 #define BTRFSIC_PRINT_MASK_INITIAL_DATABASE 0x00000400 #define BTRFSIC_PRINT_MASK_NUM_COPIES 0x00000800 #define BTRFSIC_PRINT_MASK_TREE_WITH_ALL_MIRRORS 0x00001000 #define BTRFSIC_PRINT_MASK_SUBMIT_BIO_BH_VERBOSE 0x00002000 struct btrfsic_dev_state; struct btrfsic_state; struct btrfsic_block { u32 magic_num; /* only used for debug purposes */ unsigned int is_metadata:1; /* if it is meta-data, not data-data */ unsigned int is_superblock:1; /* if it is one of the superblocks */ unsigned int is_iodone:1; /* if is done by lower subsystem */ unsigned int iodone_w_error:1; /* error was indicated to endio */ unsigned int never_written:1; /* block was added because it was * referenced, not because it was * written */ unsigned int mirror_num; /* large enough to hold * BTRFS_SUPER_MIRROR_MAX */ struct btrfsic_dev_state *dev_state; u64 dev_bytenr; /* key, physical byte num on disk */ u64 logical_bytenr; /* logical byte num on disk */ u64 generation; struct btrfs_disk_key disk_key; /* extra info to print in case of * issues, will not always be correct */ struct list_head collision_resolving_node; /* list node */ struct list_head all_blocks_node; /* list node */ /* the following two lists contain block_link items */ struct list_head ref_to_list; /* list */ struct list_head ref_from_list; /* list */ struct btrfsic_block *next_in_same_bio; void *orig_bio_bh_private; union { bio_end_io_t *bio; bh_end_io_t *bh; } orig_bio_bh_end_io; int submit_bio_bh_rw; u64 flush_gen; /* only valid if !never_written */ }; /* * Elements of this type are allocated dynamically and required because * each block object can refer to and can be ref from multiple blocks. * The key to lookup them in the hashtable is the dev_bytenr of * the block ref to plus the one from the block refered from. * The fact that they are searchable via a hashtable and that a * ref_cnt is maintained is not required for the btrfs integrity * check algorithm itself, it is only used to make the output more * beautiful in case that an error is detected (an error is defined * as a write operation to a block while that block is still referenced). */ struct btrfsic_block_link { u32 magic_num; /* only used for debug purposes */ u32 ref_cnt; struct list_head node_ref_to; /* list node */ struct list_head node_ref_from; /* list node */ struct list_head collision_resolving_node; /* list node */ struct btrfsic_block *block_ref_to; struct btrfsic_block *block_ref_from; u64 parent_generation; }; struct btrfsic_dev_state { u32 magic_num; /* only used for debug purposes */ struct block_device *bdev; struct btrfsic_state *state; struct list_head collision_resolving_node; /* list node */ struct btrfsic_block dummy_block_for_bio_bh_flush; u64 last_flush_gen; char name[BDEVNAME_SIZE]; }; struct btrfsic_block_hashtable { struct list_head table[BTRFSIC_BLOCK_HASHTABLE_SIZE]; }; struct btrfsic_block_link_hashtable { struct list_head table[BTRFSIC_BLOCK_LINK_HASHTABLE_SIZE]; }; struct btrfsic_dev_state_hashtable { struct list_head table[BTRFSIC_DEV2STATE_HASHTABLE_SIZE]; }; struct btrfsic_block_data_ctx { u64 start; /* virtual bytenr */ u64 dev_bytenr; /* physical bytenr on device */ u32 len; struct btrfsic_dev_state *dev; char **datav; struct page **pagev; void *mem_to_free; }; /* This structure is used to implement recursion without occupying * any stack space, refer to btrfsic_process_metablock() */ struct btrfsic_stack_frame { u32 magic; u32 nr; int error; int i; int limit_nesting; int num_copies; int mirror_num; struct btrfsic_block *block; struct btrfsic_block_data_ctx *block_ctx; struct btrfsic_block *next_block; struct btrfsic_block_data_ctx next_block_ctx; struct btrfs_header *hdr; struct btrfsic_stack_frame *prev; }; /* Some state per mounted filesystem */ struct btrfsic_state { u32 print_mask; int include_extent_data; int csum_size; struct list_head all_blocks_list; struct btrfsic_block_hashtable block_hashtable; struct btrfsic_block_link_hashtable block_link_hashtable; struct btrfs_root *root; u64 max_superblock_generation; struct btrfsic_block *latest_superblock; u32 metablock_size; u32 datablock_size; }; static void btrfsic_block_init(struct btrfsic_block *b); static struct btrfsic_block *btrfsic_block_alloc(void); static void btrfsic_block_free(struct btrfsic_block *b); static void btrfsic_block_link_init(struct btrfsic_block_link *n); static struct btrfsic_block_link *btrfsic_block_link_alloc(void); static void btrfsic_block_link_free(struct btrfsic_block_link *n); static void btrfsic_dev_state_init(struct btrfsic_dev_state *ds); static struct btrfsic_dev_state *btrfsic_dev_state_alloc(void); static void btrfsic_dev_state_free(struct btrfsic_dev_state *ds); static void btrfsic_block_hashtable_init(struct btrfsic_block_hashtable *h); static void btrfsic_block_hashtable_add(struct btrfsic_block *b, struct btrfsic_block_hashtable *h); static void btrfsic_block_hashtable_remove(struct btrfsic_block *b); static struct btrfsic_block *btrfsic_block_hashtable_lookup( struct block_device *bdev, u64 dev_bytenr, struct btrfsic_block_hashtable *h); static void btrfsic_block_link_hashtable_init( struct btrfsic_block_link_hashtable *h); static void btrfsic_block_link_hashtable_add( struct btrfsic_block_link *l, struct btrfsic_block_link_hashtable *h); static void btrfsic_block_link_hashtable_remove(struct btrfsic_block_link *l); static struct btrfsic_block_link *btrfsic_block_link_hashtable_lookup( struct block_device *bdev_ref_to, u64 dev_bytenr_ref_to, struct block_device *bdev_ref_from, u64 dev_bytenr_ref_from, struct btrfsic_block_link_hashtable *h); static void btrfsic_dev_state_hashtable_init( struct btrfsic_dev_state_hashtable *h); static void btrfsic_dev_state_hashtable_add( struct btrfsic_dev_state *ds, struct btrfsic_dev_state_hashtable *h); static void btrfsic_dev_state_hashtable_remove(struct btrfsic_dev_state *ds); static struct btrfsic_dev_state *btrfsic_dev_state_hashtable_lookup( struct block_device *bdev, struct btrfsic_dev_state_hashtable *h); static struct btrfsic_stack_frame *btrfsic_stack_frame_alloc(void); static void btrfsic_stack_frame_free(struct btrfsic_stack_frame *sf); static int btrfsic_process_superblock(struct btrfsic_state *state, struct btrfs_fs_devices *fs_devices); static int btrfsic_process_metablock(struct btrfsic_state *state, struct btrfsic_block *block, struct btrfsic_block_data_ctx *block_ctx, int limit_nesting, int force_iodone_flag); static void btrfsic_read_from_block_data( struct btrfsic_block_data_ctx *block_ctx, void *dst, u32 offset, size_t len); static int btrfsic_create_link_to_next_block( struct btrfsic_state *state, struct btrfsic_block *block, struct btrfsic_block_data_ctx *block_ctx, u64 next_bytenr, int limit_nesting, struct btrfsic_block_data_ctx *next_block_ctx, struct btrfsic_block **next_blockp, int force_iodone_flag, int *num_copiesp, int *mirror_nump, struct btrfs_disk_key *disk_key, u64 parent_generation); static int btrfsic_handle_extent_data(struct btrfsic_state *state, struct btrfsic_block *block, struct btrfsic_block_data_ctx *block_ctx, u32 item_offset, int force_iodone_flag); static int btrfsic_map_block(struct btrfsic_state *state, u64 bytenr, u32 len, struct btrfsic_block_data_ctx *block_ctx_out, int mirror_num); static void btrfsic_release_block_ctx(struct btrfsic_block_data_ctx *block_ctx); static int btrfsic_read_block(struct btrfsic_state *state, struct btrfsic_block_data_ctx *block_ctx); static void btrfsic_dump_database(struct btrfsic_state *state); static int btrfsic_test_for_metadata(struct btrfsic_state *state, char **datav, unsigned int num_pages); static void btrfsic_process_written_block(struct btrfsic_dev_state *dev_state, u64 dev_bytenr, char **mapped_datav, unsigned int num_pages, struct bio *bio, int *bio_is_patched, struct buffer_head *bh, int submit_bio_bh_rw); static int btrfsic_process_written_superblock( struct btrfsic_state *state, struct btrfsic_block *const block, struct btrfs_super_block *const super_hdr); static void btrfsic_bio_end_io(struct bio *bp, int bio_error_status); static void btrfsic_bh_end_io(struct buffer_head *bh, int uptodate); static int btrfsic_is_block_ref_by_superblock(const struct btrfsic_state *state, const struct btrfsic_block *block, int recursion_level); static int btrfsic_check_all_ref_blocks(struct btrfsic_state *state, struct btrfsic_block *const block, int recursion_level); static void btrfsic_print_add_link(const struct btrfsic_state *state, const struct btrfsic_block_link *l); static void btrfsic_print_rem_link(const struct btrfsic_state *state, const struct btrfsic_block_link *l); static char btrfsic_get_block_type(const struct btrfsic_state *state, const struct btrfsic_block *block); static void btrfsic_dump_tree(const struct btrfsic_state *state); static void btrfsic_dump_tree_sub(const struct btrfsic_state *state, const struct btrfsic_block *block, int indent_level); static struct btrfsic_block_link *btrfsic_block_link_lookup_or_add( struct btrfsic_state *state, struct btrfsic_block_data_ctx *next_block_ctx, struct btrfsic_block *next_block, struct btrfsic_block *from_block, u64 parent_generation); static struct btrfsic_block *btrfsic_block_lookup_or_add( struct btrfsic_state *state, struct btrfsic_block_data_ctx *block_ctx, const char *additional_string, int is_metadata, int is_iodone, int never_written, int mirror_num, int *was_created); static int btrfsic_process_superblock_dev_mirror( struct btrfsic_state *state, struct btrfsic_dev_state *dev_state, struct btrfs_device *device, int superblock_mirror_num, struct btrfsic_dev_state **selected_dev_state, struct btrfs_super_block *selected_super); static struct btrfsic_dev_state *btrfsic_dev_state_lookup( struct block_device *bdev); static void btrfsic_cmp_log_and_dev_bytenr(struct btrfsic_state *state, u64 bytenr, struct btrfsic_dev_state *dev_state, u64 dev_bytenr); static struct mutex btrfsic_mutex; static int btrfsic_is_initialized; static struct btrfsic_dev_state_hashtable btrfsic_dev_state_hashtable; static void btrfsic_block_init(struct btrfsic_block *b) { b->magic_num = BTRFSIC_BLOCK_MAGIC_NUMBER; b->dev_state = NULL; b->dev_bytenr = 0; b->logical_bytenr = 0; b->generation = BTRFSIC_GENERATION_UNKNOWN; b->disk_key.objectid = 0; b->disk_key.type = 0; b->disk_key.offset = 0; b->is_metadata = 0; b->is_superblock = 0; b->is_iodone = 0; b->iodone_w_error = 0; b->never_written = 0; b->mirror_num = 0; b->next_in_same_bio = NULL; b->orig_bio_bh_private = NULL; b->orig_bio_bh_end_io.bio = NULL; INIT_LIST_HEAD(&b->collision_resolving_node); INIT_LIST_HEAD(&b->all_blocks_node); INIT_LIST_HEAD(&b->ref_to_list); INIT_LIST_HEAD(&b->ref_from_list); b->submit_bio_bh_rw = 0; b->flush_gen = 0; } static struct btrfsic_block *btrfsic_block_alloc(void) { struct btrfsic_block *b; b = kzalloc(sizeof(*b), GFP_NOFS); if (NULL != b) btrfsic_block_init(b); return b; } static void btrfsic_block_free(struct btrfsic_block *b) { BUG_ON(!(NULL == b || BTRFSIC_BLOCK_MAGIC_NUMBER == b->magic_num)); kfree(b); } static void btrfsic_block_link_init(struct btrfsic_block_link *l) { l->magic_num = BTRFSIC_BLOCK_LINK_MAGIC_NUMBER; l->ref_cnt = 1; INIT_LIST_HEAD(&l->node_ref_to); INIT_LIST_HEAD(&l->node_ref_from); INIT_LIST_HEAD(&l->collision_resolving_node); l->block_ref_to = NULL; l->block_ref_from = NULL; } static struct btrfsic_block_link *btrfsic_block_link_alloc(void) { struct btrfsic_block_link *l; l = kzalloc(sizeof(*l), GFP_NOFS); if (NULL != l) btrfsic_block_link_init(l); return l; } static void btrfsic_block_link_free(struct btrfsic_block_link *l) { BUG_ON(!(NULL == l || BTRFSIC_BLOCK_LINK_MAGIC_NUMBER == l->magic_num)); kfree(l); } static void btrfsic_dev_state_init(struct btrfsic_dev_state *ds) { ds->magic_num = BTRFSIC_DEV2STATE_MAGIC_NUMBER; ds->bdev = NULL; ds->state = NULL; ds->name[0] = '\0'; INIT_LIST_HEAD(&ds->collision_resolving_node); ds->last_flush_gen = 0; btrfsic_block_init(&ds->dummy_block_for_bio_bh_flush); ds->dummy_block_for_bio_bh_flush.is_iodone = 1; ds->dummy_block_for_bio_bh_flush.dev_state = ds; } static struct btrfsic_dev_state *btrfsic_dev_state_alloc(void) { struct btrfsic_dev_state *ds; ds = kzalloc(sizeof(*ds), GFP_NOFS); if (NULL != ds) btrfsic_dev_state_init(ds); return ds; } static void btrfsic_dev_state_free(struct btrfsic_dev_state *ds) { BUG_ON(!(NULL == ds || BTRFSIC_DEV2STATE_MAGIC_NUMBER == ds->magic_num)); kfree(ds); } static void btrfsic_block_hashtable_init(struct btrfsic_block_hashtable *h) { int i; for (i = 0; i < BTRFSIC_BLOCK_HASHTABLE_SIZE; i++) INIT_LIST_HEAD(h->table + i); } static void btrfsic_block_hashtable_add(struct btrfsic_block *b, struct btrfsic_block_hashtable *h) { const unsigned int hashval = (((unsigned int)(b->dev_bytenr >> 16)) ^ ((unsigned int)((uintptr_t)b->dev_state->bdev))) & (BTRFSIC_BLOCK_HASHTABLE_SIZE - 1); list_add(&b->collision_resolving_node, h->table + hashval); } static void btrfsic_block_hashtable_remove(struct btrfsic_block *b) { list_del(&b->collision_resolving_node); } static struct btrfsic_block *btrfsic_block_hashtable_lookup( struct block_device *bdev, u64 dev_bytenr, struct btrfsic_block_hashtable *h) { const unsigned int hashval = (((unsigned int)(dev_bytenr >> 16)) ^ ((unsigned int)((uintptr_t)bdev))) & (BTRFSIC_BLOCK_HASHTABLE_SIZE - 1); struct list_head *elem; list_for_each(elem, h->table + hashval) { struct btrfsic_block *const b = list_entry(elem, struct btrfsic_block, collision_resolving_node); if (b->dev_state->bdev == bdev && b->dev_bytenr == dev_bytenr) return b; } return NULL; } static void btrfsic_block_link_hashtable_init( struct btrfsic_block_link_hashtable *h) { int i; for (i = 0; i < BTRFSIC_BLOCK_LINK_HASHTABLE_SIZE; i++) INIT_LIST_HEAD(h->table + i); } static void btrfsic_block_link_hashtable_add( struct btrfsic_block_link *l, struct btrfsic_block_link_hashtable *h) { const unsigned int hashval = (((unsigned int)(l->block_ref_to->dev_bytenr >> 16)) ^ ((unsigned int)(l->block_ref_from->dev_bytenr >> 16)) ^ ((unsigned int)((uintptr_t)l->block_ref_to->dev_state->bdev)) ^ ((unsigned int)((uintptr_t)l->block_ref_from->dev_state->bdev))) & (BTRFSIC_BLOCK_LINK_HASHTABLE_SIZE - 1); BUG_ON(NULL == l->block_ref_to); BUG_ON(NULL == l->block_ref_from); list_add(&l->collision_resolving_node, h->table + hashval); } static void btrfsic_block_link_hashtable_remove(struct btrfsic_block_link *l) { list_del(&l->collision_resolving_node); } static struct btrfsic_block_link *btrfsic_block_link_hashtable_lookup( struct block_device *bdev_ref_to, u64 dev_bytenr_ref_to, struct block_device *bdev_ref_from, u64 dev_bytenr_ref_from, struct btrfsic_block_link_hashtable *h) { const unsigned int hashval = (((unsigned int)(dev_bytenr_ref_to >> 16)) ^ ((unsigned int)(dev_bytenr_ref_from >> 16)) ^ ((unsigned int)((uintptr_t)bdev_ref_to)) ^ ((unsigned int)((uintptr_t)bdev_ref_from))) & (BTRFSIC_BLOCK_LINK_HASHTABLE_SIZE - 1); struct list_head *elem; list_for_each(elem, h->table + hashval) { struct btrfsic_block_link *const l = list_entry(elem, struct btrfsic_block_link, collision_resolving_node); BUG_ON(NULL == l->block_ref_to); BUG_ON(NULL == l->block_ref_from); if (l->block_ref_to->dev_state->bdev == bdev_ref_to && l->block_ref_to->dev_bytenr == dev_bytenr_ref_to && l->block_ref_from->dev_state->bdev == bdev_ref_from && l->block_ref_from->dev_bytenr == dev_bytenr_ref_from) return l; } return NULL; } static void btrfsic_dev_state_hashtable_init( struct btrfsic_dev_state_hashtable *h) { int i; for (i = 0; i < BTRFSIC_DEV2STATE_HASHTABLE_SIZE; i++) INIT_LIST_HEAD(h->table + i); } static void btrfsic_dev_state_hashtable_add( struct btrfsic_dev_state *ds, struct btrfsic_dev_state_hashtable *h) { const unsigned int hashval = (((unsigned int)((uintptr_t)ds->bdev)) & (BTRFSIC_DEV2STATE_HASHTABLE_SIZE - 1)); list_add(&ds->collision_resolving_node, h->table + hashval); } static void btrfsic_dev_state_hashtable_remove(struct btrfsic_dev_state *ds) { list_del(&ds->collision_resolving_node); } static struct btrfsic_dev_state *btrfsic_dev_state_hashtable_lookup( struct block_device *bdev, struct btrfsic_dev_state_hashtable *h) { const unsigned int hashval = (((unsigned int)((uintptr_t)bdev)) & (BTRFSIC_DEV2STATE_HASHTABLE_SIZE - 1)); struct list_head *elem; list_for_each(elem, h->table + hashval) { struct btrfsic_dev_state *const ds = list_entry(elem, struct btrfsic_dev_state, collision_resolving_node); if (ds->bdev == bdev) return ds; } return NULL; } static int btrfsic_process_superblock(struct btrfsic_state *state, struct btrfs_fs_devices *fs_devices) { int ret = 0; struct btrfs_super_block *selected_super; struct list_head *dev_head = &fs_devices->devices; struct btrfs_device *device; struct btrfsic_dev_state *selected_dev_state = NULL; int pass; BUG_ON(NULL == state); selected_super = kzalloc(sizeof(*selected_super), GFP_NOFS); if (NULL == selected_super) { printk(KERN_INFO "btrfsic: error, kmalloc failed!\n"); return -1; } list_for_each_entry(device, dev_head, dev_list) { int i; struct btrfsic_dev_state *dev_state; if (!device->bdev || !device->name) continue; dev_state = btrfsic_dev_state_lookup(device->bdev); BUG_ON(NULL == dev_state); for (i = 0; i < BTRFS_SUPER_MIRROR_MAX; i++) { ret = btrfsic_process_superblock_dev_mirror( state, dev_state, device, i, &selected_dev_state, selected_super); if (0 != ret && 0 == i) { kfree(selected_super); return ret; } } } if (NULL == state->latest_superblock) { printk(KERN_INFO "btrfsic: no superblock found!\n"); kfree(selected_super); return -1; } state->csum_size = btrfs_super_csum_size(selected_super); for (pass = 0; pass < 3; pass++) { int num_copies; int mirror_num; u64 next_bytenr; switch (pass) { case 0: next_bytenr = btrfs_super_root(selected_super); if (state->print_mask & BTRFSIC_PRINT_MASK_ROOT_CHUNK_LOG_TREE_LOCATION) printk(KERN_INFO "root@%llu\n", next_bytenr); break; case 1: next_bytenr = btrfs_super_chunk_root(selected_super); if (state->print_mask & BTRFSIC_PRINT_MASK_ROOT_CHUNK_LOG_TREE_LOCATION) printk(KERN_INFO "chunk@%llu\n", next_bytenr); break; case 2: next_bytenr = btrfs_super_log_root(selected_super); if (0 == next_bytenr) continue; if (state->print_mask & BTRFSIC_PRINT_MASK_ROOT_CHUNK_LOG_TREE_LOCATION) printk(KERN_INFO "log@%llu\n", next_bytenr); break; } num_copies = btrfs_num_copies(state->root->fs_info, next_bytenr, state->metablock_size); if (state->print_mask & BTRFSIC_PRINT_MASK_NUM_COPIES) printk(KERN_INFO "num_copies(log_bytenr=%llu) = %d\n", next_bytenr, num_copies); for (mirror_num = 1; mirror_num <= num_copies; mirror_num++) { struct btrfsic_block *next_block; struct btrfsic_block_data_ctx tmp_next_block_ctx; struct btrfsic_block_link *l; ret = btrfsic_map_block(state, next_bytenr, state->metablock_size, &tmp_next_block_ctx, mirror_num); if (ret) { printk(KERN_INFO "btrfsic:" " btrfsic_map_block(root @%llu," " mirror %d) failed!\n", next_bytenr, mirror_num); kfree(selected_super); return -1; } next_block = btrfsic_block_hashtable_lookup( tmp_next_block_ctx.dev->bdev, tmp_next_block_ctx.dev_bytenr, &state->block_hashtable); BUG_ON(NULL == next_block); l = btrfsic_block_link_hashtable_lookup( tmp_next_block_ctx.dev->bdev, tmp_next_block_ctx.dev_bytenr, state->latest_superblock->dev_state-> bdev, state->latest_superblock->dev_bytenr, &state->block_link_hashtable); BUG_ON(NULL == l); ret = btrfsic_read_block(state, &tmp_next_block_ctx); if (ret < (int)PAGE_CACHE_SIZE) { printk(KERN_INFO "btrfsic: read @logical %llu failed!\n", tmp_next_block_ctx.start); btrfsic_release_block_ctx(&tmp_next_block_ctx); kfree(selected_super); return -1; } ret = btrfsic_process_metablock(state, next_block, &tmp_next_block_ctx, BTRFS_MAX_LEVEL + 3, 1); btrfsic_release_block_ctx(&tmp_next_block_ctx); } } kfree(selected_super); return ret; } static int btrfsic_process_superblock_dev_mirror( struct btrfsic_state *state, struct btrfsic_dev_state *dev_state, struct btrfs_device *device, int superblock_mirror_num, struct btrfsic_dev_state **selected_dev_state, struct btrfs_super_block *selected_super) { struct btrfs_super_block *super_tmp; u64 dev_bytenr; struct buffer_head *bh; struct btrfsic_block *superblock_tmp; int pass; struct block_device *const superblock_bdev = device->bdev; /* super block bytenr is always the unmapped device bytenr */ dev_bytenr = btrfs_sb_offset(superblock_mirror_num); if (dev_bytenr + BTRFS_SUPER_INFO_SIZE > device->commit_total_bytes) return -1; bh = __bread(superblock_bdev, dev_bytenr / 4096, BTRFS_SUPER_INFO_SIZE); if (NULL == bh) return -1; super_tmp = (struct btrfs_super_block *) (bh->b_data + (dev_bytenr & 4095)); if (btrfs_super_bytenr(super_tmp) != dev_bytenr || btrfs_super_magic(super_tmp) != BTRFS_MAGIC || memcmp(device->uuid, super_tmp->dev_item.uuid, BTRFS_UUID_SIZE) || btrfs_super_nodesize(super_tmp) != state->metablock_size || btrfs_super_sectorsize(super_tmp) != state->datablock_size) { brelse(bh); return 0; } superblock_tmp = btrfsic_block_hashtable_lookup(superblock_bdev, dev_bytenr, &state->block_hashtable); if (NULL == superblock_tmp) { superblock_tmp = btrfsic_block_alloc(); if (NULL == superblock_tmp) { printk(KERN_INFO "btrfsic: error, kmalloc failed!\n"); brelse(bh); return -1; } /* for superblock, only the dev_bytenr makes sense */ superblock_tmp->dev_bytenr = dev_bytenr; superblock_tmp->dev_state = dev_state; superblock_tmp->logical_bytenr = dev_bytenr; superblock_tmp->generation = btrfs_super_generation(super_tmp); superblock_tmp->is_metadata = 1; superblock_tmp->is_superblock = 1; superblock_tmp->is_iodone = 1; superblock_tmp->never_written = 0; superblock_tmp->mirror_num = 1 + superblock_mirror_num; if (state->print_mask & BTRFSIC_PRINT_MASK_SUPERBLOCK_WRITE) printk_in_rcu(KERN_INFO "New initial S-block (bdev %p, %s)" " @%llu (%s/%llu/%d)\n", superblock_bdev, rcu_str_deref(device->name), dev_bytenr, dev_state->name, dev_bytenr, superblock_mirror_num); list_add(&superblock_tmp->all_blocks_node, &state->all_blocks_list); btrfsic_block_hashtable_add(superblock_tmp, &state->block_hashtable); } /* select the one with the highest generation field */ if (btrfs_super_generation(super_tmp) > state->max_superblock_generation || 0 == state->max_superblock_generation) { memcpy(selected_super, super_tmp, sizeof(*selected_super)); *selected_dev_state = dev_state; state->max_superblock_generation = btrfs_super_generation(super_tmp); state->latest_superblock = superblock_tmp; } for (pass = 0; pass < 3; pass++) { u64 next_bytenr; int num_copies; int mirror_num; const char *additional_string = NULL; struct btrfs_disk_key tmp_disk_key; tmp_disk_key.type = BTRFS_ROOT_ITEM_KEY; tmp_disk_key.offset = 0; switch (pass) { case 0: btrfs_set_disk_key_objectid(&tmp_disk_key, BTRFS_ROOT_TREE_OBJECTID); additional_string = "initial root "; next_bytenr = btrfs_super_root(super_tmp); break; case 1: btrfs_set_disk_key_objectid(&tmp_disk_key, BTRFS_CHUNK_TREE_OBJECTID); additional_string = "initial chunk "; next_bytenr = btrfs_super_chunk_root(super_tmp); break; case 2: btrfs_set_disk_key_objectid(&tmp_disk_key, BTRFS_TREE_LOG_OBJECTID); additional_string = "initial log "; next_bytenr = btrfs_super_log_root(super_tmp); if (0 == next_bytenr) continue; break; } num_copies = btrfs_num_copies(state->root->fs_info, next_bytenr, state->metablock_size); if (state->print_mask & BTRFSIC_PRINT_MASK_NUM_COPIES) printk(KERN_INFO "num_copies(log_bytenr=%llu) = %d\n", next_bytenr, num_copies); for (mirror_num = 1; mirror_num <= num_copies; mirror_num++) { struct btrfsic_block *next_block; struct btrfsic_block_data_ctx tmp_next_block_ctx; struct btrfsic_block_link *l; if (btrfsic_map_block(state, next_bytenr, state->metablock_size, &tmp_next_block_ctx, mirror_num)) { printk(KERN_INFO "btrfsic: btrfsic_map_block(" "bytenr @%llu, mirror %d) failed!\n", next_bytenr, mirror_num); brelse(bh); return -1; } next_block = btrfsic_block_lookup_or_add( state, &tmp_next_block_ctx, additional_string, 1, 1, 0, mirror_num, NULL); if (NULL == next_block) { btrfsic_release_block_ctx(&tmp_next_block_ctx); brelse(bh); return -1; } next_block->disk_key = tmp_disk_key; next_block->generation = BTRFSIC_GENERATION_UNKNOWN; l = btrfsic_block_link_lookup_or_add( state, &tmp_next_block_ctx, next_block, superblock_tmp, BTRFSIC_GENERATION_UNKNOWN); btrfsic_release_block_ctx(&tmp_next_block_ctx); if (NULL == l) { brelse(bh); return -1; } } } if (state->print_mask & BTRFSIC_PRINT_MASK_INITIAL_ALL_TREES) btrfsic_dump_tree_sub(state, superblock_tmp, 0); brelse(bh); return 0; } static struct btrfsic_stack_frame *btrfsic_stack_frame_alloc(void) { struct btrfsic_stack_frame *sf; sf = kzalloc(sizeof(*sf), GFP_NOFS); if (NULL == sf) printk(KERN_INFO "btrfsic: alloc memory failed!\n"); else sf->magic = BTRFSIC_BLOCK_STACK_FRAME_MAGIC_NUMBER; return sf; } static void btrfsic_stack_frame_free(struct btrfsic_stack_frame *sf) { BUG_ON(!(NULL == sf || BTRFSIC_BLOCK_STACK_FRAME_MAGIC_NUMBER == sf->magic)); kfree(sf); } static int btrfsic_process_metablock( struct btrfsic_state *state, struct btrfsic_block *const first_block, struct btrfsic_block_data_ctx *const first_block_ctx, int first_limit_nesting, int force_iodone_flag) { struct btrfsic_stack_frame initial_stack_frame = { 0 }; struct btrfsic_stack_frame *sf; struct btrfsic_stack_frame *next_stack; struct btrfs_header *const first_hdr = (struct btrfs_header *)first_block_ctx->datav[0]; BUG_ON(!first_hdr); sf = &initial_stack_frame; sf->error = 0; sf->i = -1; sf->limit_nesting = first_limit_nesting; sf->block = first_block; sf->block_ctx = first_block_ctx; sf->next_block = NULL; sf->hdr = first_hdr; sf->prev = NULL; continue_with_new_stack_frame: sf->block->generation = le64_to_cpu(sf->hdr->generation); if (0 == sf->hdr->level) { struct btrfs_leaf *const leafhdr = (struct btrfs_leaf *)sf->hdr; if (-1 == sf->i) { sf->nr = btrfs_stack_header_nritems(&leafhdr->header); if (state->print_mask & BTRFSIC_PRINT_MASK_VERBOSE) printk(KERN_INFO "leaf %llu items %d generation %llu" " owner %llu\n", sf->block_ctx->start, sf->nr, btrfs_stack_header_generation( &leafhdr->header), btrfs_stack_header_owner( &leafhdr->header)); } continue_with_current_leaf_stack_frame: if (0 == sf->num_copies || sf->mirror_num > sf->num_copies) { sf->i++; sf->num_copies = 0; } if (sf->i < sf->nr) { struct btrfs_item disk_item; u32 disk_item_offset = (uintptr_t)(leafhdr->items + sf->i) - (uintptr_t)leafhdr; struct btrfs_disk_key *disk_key; u8 type; u32 item_offset; u32 item_size; if (disk_item_offset + sizeof(struct btrfs_item) > sf->block_ctx->len) { leaf_item_out_of_bounce_error: printk(KERN_INFO "btrfsic: leaf item out of bounce at logical %llu, dev %s\n", sf->block_ctx->start, sf->block_ctx->dev->name); goto one_stack_frame_backwards; } btrfsic_read_from_block_data(sf->block_ctx, &disk_item, disk_item_offset, sizeof(struct btrfs_item)); item_offset = btrfs_stack_item_offset(&disk_item); item_size = btrfs_stack_item_size(&disk_item); disk_key = &disk_item.key; type = btrfs_disk_key_type(disk_key); if (BTRFS_ROOT_ITEM_KEY == type) { struct btrfs_root_item root_item; u32 root_item_offset; u64 next_bytenr; root_item_offset = item_offset + offsetof(struct btrfs_leaf, items); if (root_item_offset + item_size > sf->block_ctx->len) goto leaf_item_out_of_bounce_error; btrfsic_read_from_block_data( sf->block_ctx, &root_item, root_item_offset, item_size); next_bytenr = btrfs_root_bytenr(&root_item); sf->error = btrfsic_create_link_to_next_block( state, sf->block, sf->block_ctx, next_bytenr, sf->limit_nesting, &sf->next_block_ctx, &sf->next_block, force_iodone_flag, &sf->num_copies, &sf->mirror_num, disk_key, btrfs_root_generation( &root_item)); if (sf->error) goto one_stack_frame_backwards; if (NULL != sf->next_block) { struct btrfs_header *const next_hdr = (struct btrfs_header *) sf->next_block_ctx.datav[0]; next_stack = btrfsic_stack_frame_alloc(); if (NULL == next_stack) { sf->error = -1; btrfsic_release_block_ctx( &sf-> next_block_ctx); goto one_stack_frame_backwards; } next_stack->i = -1; next_stack->block = sf->next_block; next_stack->block_ctx = &sf->next_block_ctx; next_stack->next_block = NULL; next_stack->hdr = next_hdr; next_stack->limit_nesting = sf->limit_nesting - 1; next_stack->prev = sf; sf = next_stack; goto continue_with_new_stack_frame; } } else if (BTRFS_EXTENT_DATA_KEY == type && state->include_extent_data) { sf->error = btrfsic_handle_extent_data( state, sf->block, sf->block_ctx, item_offset, force_iodone_flag); if (sf->error) goto one_stack_frame_backwards; } goto continue_with_current_leaf_stack_frame; } } else { struct btrfs_node *const nodehdr = (struct btrfs_node *)sf->hdr; if (-1 == sf->i) { sf->nr = btrfs_stack_header_nritems(&nodehdr->header); if (state->print_mask & BTRFSIC_PRINT_MASK_VERBOSE) printk(KERN_INFO "node %llu level %d items %d" " generation %llu owner %llu\n", sf->block_ctx->start, nodehdr->header.level, sf->nr, btrfs_stack_header_generation( &nodehdr->header), btrfs_stack_header_owner( &nodehdr->header)); } continue_with_current_node_stack_frame: if (0 == sf->num_copies || sf->mirror_num > sf->num_copies) { sf->i++; sf->num_copies = 0; } if (sf->i < sf->nr) { struct btrfs_key_ptr key_ptr; u32 key_ptr_offset; u64 next_bytenr; key_ptr_offset = (uintptr_t)(nodehdr->ptrs + sf->i) - (uintptr_t)nodehdr; if (key_ptr_offset + sizeof(struct btrfs_key_ptr) > sf->block_ctx->len) { printk(KERN_INFO "btrfsic: node item out of bounce at logical %llu, dev %s\n", sf->block_ctx->start, sf->block_ctx->dev->name); goto one_stack_frame_backwards; } btrfsic_read_from_block_data( sf->block_ctx, &key_ptr, key_ptr_offset, sizeof(struct btrfs_key_ptr)); next_bytenr = btrfs_stack_key_blockptr(&key_ptr); sf->error = btrfsic_create_link_to_next_block( state, sf->block, sf->block_ctx, next_bytenr, sf->limit_nesting, &sf->next_block_ctx, &sf->next_block, force_iodone_flag, &sf->num_copies, &sf->mirror_num, &key_ptr.key, btrfs_stack_key_generation(&key_ptr)); if (sf->error) goto one_stack_frame_backwards; if (NULL != sf->next_block) { struct btrfs_header *const next_hdr = (struct btrfs_header *) sf->next_block_ctx.datav[0]; next_stack = btrfsic_stack_frame_alloc(); if (NULL == next_stack) { sf->error = -1; goto one_stack_frame_backwards; } next_stack->i = -1; next_stack->block = sf->next_block; next_stack->block_ctx = &sf->next_block_ctx; next_stack->next_block = NULL; next_stack->hdr = next_hdr; next_stack->limit_nesting = sf->limit_nesting - 1; next_stack->prev = sf; sf = next_stack; goto continue_with_new_stack_frame; } goto continue_with_current_node_stack_frame; } } one_stack_frame_backwards: if (NULL != sf->prev) { struct btrfsic_stack_frame *const prev = sf->prev; /* the one for the initial block is freed in the caller */ btrfsic_release_block_ctx(sf->block_ctx); if (sf->error) { prev->error = sf->error; btrfsic_stack_frame_free(sf); sf = prev; goto one_stack_frame_backwards; } btrfsic_stack_frame_free(sf); sf = prev; goto continue_with_new_stack_frame; } else { BUG_ON(&initial_stack_frame != sf); } return sf->error; } static void btrfsic_read_from_block_data( struct btrfsic_block_data_ctx *block_ctx, void *dstv, u32 offset, size_t len) { size_t cur; size_t offset_in_page; char *kaddr; char *dst = (char *)dstv; size_t start_offset = block_ctx->start & ((u64)PAGE_CACHE_SIZE - 1); unsigned long i = (start_offset + offset) >> PAGE_CACHE_SHIFT; WARN_ON(offset + len > block_ctx->len); offset_in_page = (start_offset + offset) & (PAGE_CACHE_SIZE - 1); while (len > 0) { cur = min(len, ((size_t)PAGE_CACHE_SIZE - offset_in_page)); BUG_ON(i >= DIV_ROUND_UP(block_ctx->len, PAGE_CACHE_SIZE)); kaddr = block_ctx->datav[i]; memcpy(dst, kaddr + offset_in_page, cur); dst += cur; len -= cur; offset_in_page = 0; i++; } } static int btrfsic_create_link_to_next_block( struct btrfsic_state *state, struct btrfsic_block *block, struct btrfsic_block_data_ctx *block_ctx, u64 next_bytenr, int limit_nesting, struct btrfsic_block_data_ctx *next_block_ctx, struct btrfsic_block **next_blockp, int force_iodone_flag, int *num_copiesp, int *mirror_nump, struct btrfs_disk_key *disk_key, u64 parent_generation) { struct btrfsic_block *next_block = NULL; int ret; struct btrfsic_block_link *l; int did_alloc_block_link; int block_was_created; *next_blockp = NULL; if (0 == *num_copiesp) { *num_copiesp = btrfs_num_copies(state->root->fs_info, next_bytenr, state->metablock_size); if (state->print_mask & BTRFSIC_PRINT_MASK_NUM_COPIES) printk(KERN_INFO "num_copies(log_bytenr=%llu) = %d\n", next_bytenr, *num_copiesp); *mirror_nump = 1; } if (*mirror_nump > *num_copiesp) return 0; if (state->print_mask & BTRFSIC_PRINT_MASK_VERBOSE) printk(KERN_INFO "btrfsic_create_link_to_next_block(mirror_num=%d)\n", *mirror_nump); ret = btrfsic_map_block(state, next_bytenr, state->metablock_size, next_block_ctx, *mirror_nump); if (ret) { printk(KERN_INFO "btrfsic: btrfsic_map_block(@%llu, mirror=%d) failed!\n", next_bytenr, *mirror_nump); btrfsic_release_block_ctx(next_block_ctx); *next_blockp = NULL; return -1; } next_block = btrfsic_block_lookup_or_add(state, next_block_ctx, "referenced ", 1, force_iodone_flag, !force_iodone_flag, *mirror_nump, &block_was_created); if (NULL == next_block) { btrfsic_release_block_ctx(next_block_ctx); *next_blockp = NULL; return -1; } if (block_was_created) { l = NULL; next_block->generation = BTRFSIC_GENERATION_UNKNOWN; } else { if (state->print_mask & BTRFSIC_PRINT_MASK_VERBOSE) { if (next_block->logical_bytenr != next_bytenr && !(!next_block->is_metadata && 0 == next_block->logical_bytenr)) printk(KERN_INFO "Referenced block @%llu (%s/%llu/%d) found in hash table, %c, bytenr mismatch (!= stored %llu).\n", next_bytenr, next_block_ctx->dev->name, next_block_ctx->dev_bytenr, *mirror_nump, btrfsic_get_block_type(state, next_block), next_block->logical_bytenr); else printk(KERN_INFO "Referenced block @%llu (%s/%llu/%d) found in hash table, %c.\n", next_bytenr, next_block_ctx->dev->name, next_block_ctx->dev_bytenr, *mirror_nump, btrfsic_get_block_type(state, next_block)); } next_block->logical_bytenr = next_bytenr; next_block->mirror_num = *mirror_nump; l = btrfsic_block_link_hashtable_lookup( next_block_ctx->dev->bdev, next_block_ctx->dev_bytenr, block_ctx->dev->bdev, block_ctx->dev_bytenr, &state->block_link_hashtable); } next_block->disk_key = *disk_key; if (NULL == l) { l = btrfsic_block_link_alloc(); if (NULL == l) { printk(KERN_INFO "btrfsic: error, kmalloc failed!\n"); btrfsic_release_block_ctx(next_block_ctx); *next_blockp = NULL; return -1; } did_alloc_block_link = 1; l->block_ref_to = next_block; l->block_ref_from = block; l->ref_cnt = 1; l->parent_generation = parent_generation; if (state->print_mask & BTRFSIC_PRINT_MASK_VERBOSE) btrfsic_print_add_link(state, l); list_add(&l->node_ref_to, &block->ref_to_list); list_add(&l->node_ref_from, &next_block->ref_from_list); btrfsic_block_link_hashtable_add(l, &state->block_link_hashtable); } else { did_alloc_block_link = 0; if (0 == limit_nesting) { l->ref_cnt++; l->parent_generation = parent_generation; if (state->print_mask & BTRFSIC_PRINT_MASK_VERBOSE) btrfsic_print_add_link(state, l); } } if (limit_nesting > 0 && did_alloc_block_link) { ret = btrfsic_read_block(state, next_block_ctx); if (ret < (int)next_block_ctx->len) { printk(KERN_INFO "btrfsic: read block @logical %llu failed!\n", next_bytenr); btrfsic_release_block_ctx(next_block_ctx); *next_blockp = NULL; return -1; } *next_blockp = next_block; } else { *next_blockp = NULL; } (*mirror_nump)++; return 0; } static int btrfsic_handle_extent_data( struct btrfsic_state *state, struct btrfsic_block *block, struct btrfsic_block_data_ctx *block_ctx, u32 item_offset, int force_iodone_flag) { int ret; struct btrfs_file_extent_item file_extent_item; u64 file_extent_item_offset; u64 next_bytenr; u64 num_bytes; u64 generation; struct btrfsic_block_link *l; file_extent_item_offset = offsetof(struct btrfs_leaf, items) + item_offset; if (file_extent_item_offset + offsetof(struct btrfs_file_extent_item, disk_num_bytes) > block_ctx->len) { printk(KERN_INFO "btrfsic: file item out of bounce at logical %llu, dev %s\n", block_ctx->start, block_ctx->dev->name); return -1; } btrfsic_read_from_block_data(block_ctx, &file_extent_item, file_extent_item_offset, offsetof(struct btrfs_file_extent_item, disk_num_bytes)); if (BTRFS_FILE_EXTENT_REG != file_extent_item.type || btrfs_stack_file_extent_disk_bytenr(&file_extent_item) == 0) { if (state->print_mask & BTRFSIC_PRINT_MASK_VERY_VERBOSE) printk(KERN_INFO "extent_data: type %u, disk_bytenr = %llu\n", file_extent_item.type, btrfs_stack_file_extent_disk_bytenr( &file_extent_item)); return 0; } if (file_extent_item_offset + sizeof(struct btrfs_file_extent_item) > block_ctx->len) { printk(KERN_INFO "btrfsic: file item out of bounce at logical %llu, dev %s\n", block_ctx->start, block_ctx->dev->name); return -1; } btrfsic_read_from_block_data(block_ctx, &file_extent_item, file_extent_item_offset, sizeof(struct btrfs_file_extent_item)); next_bytenr = btrfs_stack_file_extent_disk_bytenr(&file_extent_item); if (btrfs_stack_file_extent_compression(&file_extent_item) == BTRFS_COMPRESS_NONE) { next_bytenr += btrfs_stack_file_extent_offset(&file_extent_item); num_bytes = btrfs_stack_file_extent_num_bytes(&file_extent_item); } else { num_bytes = btrfs_stack_file_extent_disk_num_bytes(&file_extent_item); } generation = btrfs_stack_file_extent_generation(&file_extent_item); if (state->print_mask & BTRFSIC_PRINT_MASK_VERY_VERBOSE) printk(KERN_INFO "extent_data: type %u, disk_bytenr = %llu," " offset = %llu, num_bytes = %llu\n", file_extent_item.type, btrfs_stack_file_extent_disk_bytenr(&file_extent_item), btrfs_stack_file_extent_offset(&file_extent_item), num_bytes); while (num_bytes > 0) { u32 chunk_len; int num_copies; int mirror_num; if (num_bytes > state->datablock_size) chunk_len = state->datablock_size; else chunk_len = num_bytes; num_copies = btrfs_num_copies(state->root->fs_info, next_bytenr, state->datablock_size); if (state->print_mask & BTRFSIC_PRINT_MASK_NUM_COPIES) printk(KERN_INFO "num_copies(log_bytenr=%llu) = %d\n", next_bytenr, num_copies); for (mirror_num = 1; mirror_num <= num_copies; mirror_num++) { struct btrfsic_block_data_ctx next_block_ctx; struct btrfsic_block *next_block; int block_was_created; if (state->print_mask & BTRFSIC_PRINT_MASK_VERBOSE) printk(KERN_INFO "btrfsic_handle_extent_data(" "mirror_num=%d)\n", mirror_num); if (state->print_mask & BTRFSIC_PRINT_MASK_VERY_VERBOSE) printk(KERN_INFO "\tdisk_bytenr = %llu, num_bytes %u\n", next_bytenr, chunk_len); ret = btrfsic_map_block(state, next_bytenr, chunk_len, &next_block_ctx, mirror_num); if (ret) { printk(KERN_INFO "btrfsic: btrfsic_map_block(@%llu," " mirror=%d) failed!\n", next_bytenr, mirror_num); return -1; } next_block = btrfsic_block_lookup_or_add( state, &next_block_ctx, "referenced ", 0, force_iodone_flag, !force_iodone_flag, mirror_num, &block_was_created); if (NULL == next_block) { printk(KERN_INFO "btrfsic: error, kmalloc failed!\n"); btrfsic_release_block_ctx(&next_block_ctx); return -1; } if (!block_was_created) { if ((state->print_mask & BTRFSIC_PRINT_MASK_VERBOSE) && next_block->logical_bytenr != next_bytenr && !(!next_block->is_metadata && 0 == next_block->logical_bytenr)) { printk(KERN_INFO "Referenced block" " @%llu (%s/%llu/%d)" " found in hash table, D," " bytenr mismatch" " (!= stored %llu).\n", next_bytenr, next_block_ctx.dev->name, next_block_ctx.dev_bytenr, mirror_num, next_block->logical_bytenr); } next_block->logical_bytenr = next_bytenr; next_block->mirror_num = mirror_num; } l = btrfsic_block_link_lookup_or_add(state, &next_block_ctx, next_block, block, generation); btrfsic_release_block_ctx(&next_block_ctx); if (NULL == l) return -1; } next_bytenr += chunk_len; num_bytes -= chunk_len; } return 0; } static int btrfsic_map_block(struct btrfsic_state *state, u64 bytenr, u32 len, struct btrfsic_block_data_ctx *block_ctx_out, int mirror_num) { int ret; u64 length; struct btrfs_bio *multi = NULL; struct btrfs_device *device; length = len; ret = btrfs_map_block(state->root->fs_info, READ, bytenr, &length, &multi, mirror_num); if (ret) { block_ctx_out->start = 0; block_ctx_out->dev_bytenr = 0; block_ctx_out->len = 0; block_ctx_out->dev = NULL; block_ctx_out->datav = NULL; block_ctx_out->pagev = NULL; block_ctx_out->mem_to_free = NULL; return ret; } device = multi->stripes[0].dev; block_ctx_out->dev = btrfsic_dev_state_lookup(device->bdev); block_ctx_out->dev_bytenr = multi->stripes[0].physical; block_ctx_out->start = bytenr; block_ctx_out->len = len; block_ctx_out->datav = NULL; block_ctx_out->pagev = NULL; block_ctx_out->mem_to_free = NULL; kfree(multi); if (NULL == block_ctx_out->dev) { ret = -ENXIO; printk(KERN_INFO "btrfsic: error, cannot lookup dev (#1)!\n"); } return ret; } static void btrfsic_release_block_ctx(struct btrfsic_block_data_ctx *block_ctx) { if (block_ctx->mem_to_free) { unsigned int num_pages; BUG_ON(!block_ctx->datav); BUG_ON(!block_ctx->pagev); num_pages = (block_ctx->len + (u64)PAGE_CACHE_SIZE - 1) >> PAGE_CACHE_SHIFT; while (num_pages > 0) { num_pages--; if (block_ctx->datav[num_pages]) { kunmap(block_ctx->pagev[num_pages]); block_ctx->datav[num_pages] = NULL; } if (block_ctx->pagev[num_pages]) { __free_page(block_ctx->pagev[num_pages]); block_ctx->pagev[num_pages] = NULL; } } kfree(block_ctx->mem_to_free); block_ctx->mem_to_free = NULL; block_ctx->pagev = NULL; block_ctx->datav = NULL; } } static int btrfsic_read_block(struct btrfsic_state *state, struct btrfsic_block_data_ctx *block_ctx) { unsigned int num_pages; unsigned int i; u64 dev_bytenr; int ret; BUG_ON(block_ctx->datav); BUG_ON(block_ctx->pagev); BUG_ON(block_ctx->mem_to_free); if (block_ctx->dev_bytenr & ((u64)PAGE_CACHE_SIZE - 1)) { printk(KERN_INFO "btrfsic: read_block() with unaligned bytenr %llu\n", block_ctx->dev_bytenr); return -1; } num_pages = (block_ctx->len + (u64)PAGE_CACHE_SIZE - 1) >> PAGE_CACHE_SHIFT; block_ctx->mem_to_free = kzalloc((sizeof(*block_ctx->datav) + sizeof(*block_ctx->pagev)) * num_pages, GFP_NOFS); if (!block_ctx->mem_to_free) return -1; block_ctx->datav = block_ctx->mem_to_free; block_ctx->pagev = (struct page **)(block_ctx->datav + num_pages); for (i = 0; i < num_pages; i++) { block_ctx->pagev[i] = alloc_page(GFP_NOFS); if (!block_ctx->pagev[i]) return -1; } dev_bytenr = block_ctx->dev_bytenr; for (i = 0; i < num_pages;) { struct bio *bio; unsigned int j; bio = btrfs_io_bio_alloc(GFP_NOFS, num_pages - i); if (!bio) { printk(KERN_INFO "btrfsic: bio_alloc() for %u pages failed!\n", num_pages - i); return -1; } bio->bi_bdev = block_ctx->dev->bdev; bio->bi_iter.bi_sector = dev_bytenr >> 9; for (j = i; j < num_pages; j++) { ret = bio_add_page(bio, block_ctx->pagev[j], PAGE_CACHE_SIZE, 0); if (PAGE_CACHE_SIZE != ret) break; } if (j == i) { printk(KERN_INFO "btrfsic: error, failed to add a single page!\n"); return -1; } if (submit_bio_wait(READ, bio)) { printk(KERN_INFO "btrfsic: read error at logical %llu dev %s!\n", block_ctx->start, block_ctx->dev->name); bio_put(bio); return -1; } bio_put(bio); dev_bytenr += (j - i) * PAGE_CACHE_SIZE; i = j; } for (i = 0; i < num_pages; i++) { block_ctx->datav[i] = kmap(block_ctx->pagev[i]); if (!block_ctx->datav[i]) { printk(KERN_INFO "btrfsic: kmap() failed (dev %s)!\n", block_ctx->dev->name); return -1; } } return block_ctx->len; } static void btrfsic_dump_database(struct btrfsic_state *state) { struct list_head *elem_all; BUG_ON(NULL == state); printk(KERN_INFO "all_blocks_list:\n"); list_for_each(elem_all, &state->all_blocks_list) { const struct btrfsic_block *const b_all = list_entry(elem_all, struct btrfsic_block, all_blocks_node); struct list_head *elem_ref_to; struct list_head *elem_ref_from; printk(KERN_INFO "%c-block @%llu (%s/%llu/%d)\n", btrfsic_get_block_type(state, b_all), b_all->logical_bytenr, b_all->dev_state->name, b_all->dev_bytenr, b_all->mirror_num); list_for_each(elem_ref_to, &b_all->ref_to_list) { const struct btrfsic_block_link *const l = list_entry(elem_ref_to, struct btrfsic_block_link, node_ref_to); printk(KERN_INFO " %c @%llu (%s/%llu/%d)" " refers %u* to" " %c @%llu (%s/%llu/%d)\n", btrfsic_get_block_type(state, b_all), b_all->logical_bytenr, b_all->dev_state->name, b_all->dev_bytenr, b_all->mirror_num, l->ref_cnt, btrfsic_get_block_type(state, l->block_ref_to), l->block_ref_to->logical_bytenr, l->block_ref_to->dev_state->name, l->block_ref_to->dev_bytenr, l->block_ref_to->mirror_num); } list_for_each(elem_ref_from, &b_all->ref_from_list) { const struct btrfsic_block_link *const l = list_entry(elem_ref_from, struct btrfsic_block_link, node_ref_from); printk(KERN_INFO " %c @%llu (%s/%llu/%d)" " is ref %u* from" " %c @%llu (%s/%llu/%d)\n", btrfsic_get_block_type(state, b_all), b_all->logical_bytenr, b_all->dev_state->name, b_all->dev_bytenr, b_all->mirror_num, l->ref_cnt, btrfsic_get_block_type(state, l->block_ref_from), l->block_ref_from->logical_bytenr, l->block_ref_from->dev_state->name, l->block_ref_from->dev_bytenr, l->block_ref_from->mirror_num); } printk(KERN_INFO "\n"); } } /* * Test whether the disk block contains a tree block (leaf or node) * (note that this test fails for the super block) */ static int btrfsic_test_for_metadata(struct btrfsic_state *state, char **datav, unsigned int num_pages) { struct btrfs_header *h; u8 csum[BTRFS_CSUM_SIZE]; u32 crc = ~(u32)0; unsigned int i; if (num_pages * PAGE_CACHE_SIZE < state->metablock_size) return 1; /* not metadata */ num_pages = state->metablock_size >> PAGE_CACHE_SHIFT; h = (struct btrfs_header *)datav[0]; if (memcmp(h->fsid, state->root->fs_info->fsid, BTRFS_UUID_SIZE)) return 1; for (i = 0; i < num_pages; i++) { u8 *data = i ? datav[i] : (datav[i] + BTRFS_CSUM_SIZE); size_t sublen = i ? PAGE_CACHE_SIZE : (PAGE_CACHE_SIZE - BTRFS_CSUM_SIZE); crc = btrfs_crc32c(crc, data, sublen); } btrfs_csum_final(crc, csum); if (memcmp(csum, h->csum, state->csum_size)) return 1; return 0; /* is metadata */ } static void btrfsic_process_written_block(struct btrfsic_dev_state *dev_state, u64 dev_bytenr, char **mapped_datav, unsigned int num_pages, struct bio *bio, int *bio_is_patched, struct buffer_head *bh, int submit_bio_bh_rw) { int is_metadata; struct btrfsic_block *block; struct btrfsic_block_data_ctx block_ctx; int ret; struct btrfsic_state *state = dev_state->state; struct block_device *bdev = dev_state->bdev; unsigned int processed_len; if (NULL != bio_is_patched) *bio_is_patched = 0; again: if (num_pages == 0) return; processed_len = 0; is_metadata = (0 == btrfsic_test_for_metadata(state, mapped_datav, num_pages)); block = btrfsic_block_hashtable_lookup(bdev, dev_bytenr, &state->block_hashtable); if (NULL != block) { u64 bytenr = 0; struct list_head *elem_ref_to; struct list_head *tmp_ref_to; if (block->is_superblock) { bytenr = btrfs_super_bytenr((struct btrfs_super_block *) mapped_datav[0]); if (num_pages * PAGE_CACHE_SIZE < BTRFS_SUPER_INFO_SIZE) { printk(KERN_INFO "btrfsic: cannot work with too short bios!\n"); return; } is_metadata = 1; BUG_ON(BTRFS_SUPER_INFO_SIZE & (PAGE_CACHE_SIZE - 1)); processed_len = BTRFS_SUPER_INFO_SIZE; if (state->print_mask & BTRFSIC_PRINT_MASK_TREE_BEFORE_SB_WRITE) { printk(KERN_INFO "[before new superblock is written]:\n"); btrfsic_dump_tree_sub(state, block, 0); } } if (is_metadata) { if (!block->is_superblock) { if (num_pages * PAGE_CACHE_SIZE < state->metablock_size) { printk(KERN_INFO "btrfsic: cannot work with too short bios!\n"); return; } processed_len = state->metablock_size; bytenr = btrfs_stack_header_bytenr( (struct btrfs_header *) mapped_datav[0]); btrfsic_cmp_log_and_dev_bytenr(state, bytenr, dev_state, dev_bytenr); } if (state->print_mask & BTRFSIC_PRINT_MASK_VERBOSE) { if (block->logical_bytenr != bytenr && !(!block->is_metadata && block->logical_bytenr == 0)) printk(KERN_INFO "Written block @%llu (%s/%llu/%d) found in hash table, %c, bytenr mismatch (!= stored %llu).\n", bytenr, dev_state->name, dev_bytenr, block->mirror_num, btrfsic_get_block_type(state, block), block->logical_bytenr); else printk(KERN_INFO "Written block @%llu (%s/%llu/%d) found in hash table, %c.\n", bytenr, dev_state->name, dev_bytenr, block->mirror_num, btrfsic_get_block_type(state, block)); } block->logical_bytenr = bytenr; } else { if (num_pages * PAGE_CACHE_SIZE < state->datablock_size) { printk(KERN_INFO "btrfsic: cannot work with too short bios!\n"); return; } processed_len = state->datablock_size; bytenr = block->logical_bytenr; if (state->print_mask & BTRFSIC_PRINT_MASK_VERBOSE) printk(KERN_INFO "Written block @%llu (%s/%llu/%d)" " found in hash table, %c.\n", bytenr, dev_state->name, dev_bytenr, block->mirror_num, btrfsic_get_block_type(state, block)); } if (state->print_mask & BTRFSIC_PRINT_MASK_VERBOSE) printk(KERN_INFO "ref_to_list: %cE, ref_from_list: %cE\n", list_empty(&block->ref_to_list) ? ' ' : '!', list_empty(&block->ref_from_list) ? ' ' : '!'); if (btrfsic_is_block_ref_by_superblock(state, block, 0)) { printk(KERN_INFO "btrfs: attempt to overwrite %c-block" " @%llu (%s/%llu/%d), old(gen=%llu," " objectid=%llu, type=%d, offset=%llu)," " new(gen=%llu)," " which is referenced by most recent superblock" " (superblockgen=%llu)!\n", btrfsic_get_block_type(state, block), bytenr, dev_state->name, dev_bytenr, block->mirror_num, block->generation, btrfs_disk_key_objectid(&block->disk_key), block->disk_key.type, btrfs_disk_key_offset(&block->disk_key), btrfs_stack_header_generation( (struct btrfs_header *) mapped_datav[0]), state->max_superblock_generation); btrfsic_dump_tree(state); } if (!block->is_iodone && !block->never_written) { printk(KERN_INFO "btrfs: attempt to overwrite %c-block" " @%llu (%s/%llu/%d), oldgen=%llu, newgen=%llu," " which is not yet iodone!\n", btrfsic_get_block_type(state, block), bytenr, dev_state->name, dev_bytenr, block->mirror_num, block->generation, btrfs_stack_header_generation( (struct btrfs_header *) mapped_datav[0])); /* it would not be safe to go on */ btrfsic_dump_tree(state); goto continue_loop; } /* * Clear all references of this block. Do not free * the block itself even if is not referenced anymore * because it still carries valueable information * like whether it was ever written and IO completed. */ list_for_each_safe(elem_ref_to, tmp_ref_to, &block->ref_to_list) { struct btrfsic_block_link *const l = list_entry(elem_ref_to, struct btrfsic_block_link, node_ref_to); if (state->print_mask & BTRFSIC_PRINT_MASK_VERBOSE) btrfsic_print_rem_link(state, l); l->ref_cnt--; if (0 == l->ref_cnt) { list_del(&l->node_ref_to); list_del(&l->node_ref_from); btrfsic_block_link_hashtable_remove(l); btrfsic_block_link_free(l); } } block_ctx.dev = dev_state; block_ctx.dev_bytenr = dev_bytenr; block_ctx.start = bytenr; block_ctx.len = processed_len; block_ctx.pagev = NULL; block_ctx.mem_to_free = NULL; block_ctx.datav = mapped_datav; if (is_metadata || state->include_extent_data) { block->never_written = 0; block->iodone_w_error = 0; if (NULL != bio) { block->is_iodone = 0; BUG_ON(NULL == bio_is_patched); if (!*bio_is_patched) { block->orig_bio_bh_private = bio->bi_private; block->orig_bio_bh_end_io.bio = bio->bi_end_io; block->next_in_same_bio = NULL; bio->bi_private = block; bio->bi_end_io = btrfsic_bio_end_io; *bio_is_patched = 1; } else { struct btrfsic_block *chained_block = (struct btrfsic_block *) bio->bi_private; BUG_ON(NULL == chained_block); block->orig_bio_bh_private = chained_block->orig_bio_bh_private; block->orig_bio_bh_end_io.bio = chained_block->orig_bio_bh_end_io. bio; block->next_in_same_bio = chained_block; bio->bi_private = block; } } else if (NULL != bh) { block->is_iodone = 0; block->orig_bio_bh_private = bh->b_private; block->orig_bio_bh_end_io.bh = bh->b_end_io; block->next_in_same_bio = NULL; bh->b_private = block; bh->b_end_io = btrfsic_bh_end_io; } else { block->is_iodone = 1; block->orig_bio_bh_private = NULL; block->orig_bio_bh_end_io.bio = NULL; block->next_in_same_bio = NULL; } } block->flush_gen = dev_state->last_flush_gen + 1; block->submit_bio_bh_rw = submit_bio_bh_rw; if (is_metadata) { block->logical_bytenr = bytenr; block->is_metadata = 1; if (block->is_superblock) { BUG_ON(PAGE_CACHE_SIZE != BTRFS_SUPER_INFO_SIZE); ret = btrfsic_process_written_superblock( state, block, (struct btrfs_super_block *) mapped_datav[0]); if (state->print_mask & BTRFSIC_PRINT_MASK_TREE_AFTER_SB_WRITE) { printk(KERN_INFO "[after new superblock is written]:\n"); btrfsic_dump_tree_sub(state, block, 0); } } else { block->mirror_num = 0; /* unknown */ ret = btrfsic_process_metablock( state, block, &block_ctx, 0, 0); } if (ret) printk(KERN_INFO "btrfsic: btrfsic_process_metablock" "(root @%llu) failed!\n", dev_bytenr); } else { block->is_metadata = 0; block->mirror_num = 0; /* unknown */ block->generation = BTRFSIC_GENERATION_UNKNOWN; if (!state->include_extent_data && list_empty(&block->ref_from_list)) { /* * disk block is overwritten with extent * data (not meta data) and we are configured * to not include extent data: take the * chance and free the block's memory */ btrfsic_block_hashtable_remove(block); list_del(&block->all_blocks_node); btrfsic_block_free(block); } } btrfsic_release_block_ctx(&block_ctx); } else { /* block has not been found in hash table */ u64 bytenr; if (!is_metadata) { processed_len = state->datablock_size; if (state->print_mask & BTRFSIC_PRINT_MASK_VERBOSE) printk(KERN_INFO "Written block (%s/%llu/?)" " !found in hash table, D.\n", dev_state->name, dev_bytenr); if (!state->include_extent_data) { /* ignore that written D block */ goto continue_loop; } /* this is getting ugly for the * include_extent_data case... */ bytenr = 0; /* unknown */ } else { processed_len = state->metablock_size; bytenr = btrfs_stack_header_bytenr( (struct btrfs_header *) mapped_datav[0]); btrfsic_cmp_log_and_dev_bytenr(state, bytenr, dev_state, dev_bytenr); if (state->print_mask & BTRFSIC_PRINT_MASK_VERBOSE) printk(KERN_INFO "Written block @%llu (%s/%llu/?)" " !found in hash table, M.\n", bytenr, dev_state->name, dev_bytenr); } block_ctx.dev = dev_state; block_ctx.dev_bytenr = dev_bytenr; block_ctx.start = bytenr; block_ctx.len = processed_len; block_ctx.pagev = NULL; block_ctx.mem_to_free = NULL; block_ctx.datav = mapped_datav; block = btrfsic_block_alloc(); if (NULL == block) { printk(KERN_INFO "btrfsic: error, kmalloc failed!\n"); btrfsic_release_block_ctx(&block_ctx); goto continue_loop; } block->dev_state = dev_state; block->dev_bytenr = dev_bytenr; block->logical_bytenr = bytenr; block->is_metadata = is_metadata; block->never_written = 0; block->iodone_w_error = 0; block->mirror_num = 0; /* unknown */ block->flush_gen = dev_state->last_flush_gen + 1; block->submit_bio_bh_rw = submit_bio_bh_rw; if (NULL != bio) { block->is_iodone = 0; BUG_ON(NULL == bio_is_patched); if (!*bio_is_patched) { block->orig_bio_bh_private = bio->bi_private; block->orig_bio_bh_end_io.bio = bio->bi_end_io; block->next_in_same_bio = NULL; bio->bi_private = block; bio->bi_end_io = btrfsic_bio_end_io; *bio_is_patched = 1; } else { struct btrfsic_block *chained_block = (struct btrfsic_block *) bio->bi_private; BUG_ON(NULL == chained_block); block->orig_bio_bh_private = chained_block->orig_bio_bh_private; block->orig_bio_bh_end_io.bio = chained_block->orig_bio_bh_end_io.bio; block->next_in_same_bio = chained_block; bio->bi_private = block; } } else if (NULL != bh) { block->is_iodone = 0; block->orig_bio_bh_private = bh->b_private; block->orig_bio_bh_end_io.bh = bh->b_end_io; block->next_in_same_bio = NULL; bh->b_private = block; bh->b_end_io = btrfsic_bh_end_io; } else { block->is_iodone = 1; block->orig_bio_bh_private = NULL; block->orig_bio_bh_end_io.bio = NULL; block->next_in_same_bio = NULL; } if (state->print_mask & BTRFSIC_PRINT_MASK_VERBOSE) printk(KERN_INFO "New written %c-block @%llu (%s/%llu/%d)\n", is_metadata ? 'M' : 'D', block->logical_bytenr, block->dev_state->name, block->dev_bytenr, block->mirror_num); list_add(&block->all_blocks_node, &state->all_blocks_list); btrfsic_block_hashtable_add(block, &state->block_hashtable); if (is_metadata) { ret = btrfsic_process_metablock(state, block, &block_ctx, 0, 0); if (ret) printk(KERN_INFO "btrfsic: process_metablock(root @%llu)" " failed!\n", dev_bytenr); } btrfsic_release_block_ctx(&block_ctx); } continue_loop: BUG_ON(!processed_len); dev_bytenr += processed_len; mapped_datav += processed_len >> PAGE_CACHE_SHIFT; num_pages -= processed_len >> PAGE_CACHE_SHIFT; goto again; } static void btrfsic_bio_end_io(struct bio *bp, int bio_error_status) { struct btrfsic_block *block = (struct btrfsic_block *)bp->bi_private; int iodone_w_error; /* mutex is not held! This is not save if IO is not yet completed * on umount */ iodone_w_error = 0; if (bio_error_status) iodone_w_error = 1; BUG_ON(NULL == block); bp->bi_private = block->orig_bio_bh_private; bp->bi_end_io = block->orig_bio_bh_end_io.bio; do { struct btrfsic_block *next_block; struct btrfsic_dev_state *const dev_state = block->dev_state; if ((dev_state->state->print_mask & BTRFSIC_PRINT_MASK_END_IO_BIO_BH)) printk(KERN_INFO "bio_end_io(err=%d) for %c @%llu (%s/%llu/%d)\n", bio_error_status, btrfsic_get_block_type(dev_state->state, block), block->logical_bytenr, dev_state->name, block->dev_bytenr, block->mirror_num); next_block = block->next_in_same_bio; block->iodone_w_error = iodone_w_error; if (block->submit_bio_bh_rw & REQ_FLUSH) { dev_state->last_flush_gen++; if ((dev_state->state->print_mask & BTRFSIC_PRINT_MASK_END_IO_BIO_BH)) printk(KERN_INFO "bio_end_io() new %s flush_gen=%llu\n", dev_state->name, dev_state->last_flush_gen); } if (block->submit_bio_bh_rw & REQ_FUA) block->flush_gen = 0; /* FUA completed means block is * on disk */ block->is_iodone = 1; /* for FLUSH, this releases the block */ block = next_block; } while (NULL != block); bp->bi_end_io(bp, bio_error_status); } static void btrfsic_bh_end_io(struct buffer_head *bh, int uptodate) { struct btrfsic_block *block = (struct btrfsic_block *)bh->b_private; int iodone_w_error = !uptodate; struct btrfsic_dev_state *dev_state; BUG_ON(NULL == block); dev_state = block->dev_state; if ((dev_state->state->print_mask & BTRFSIC_PRINT_MASK_END_IO_BIO_BH)) printk(KERN_INFO "bh_end_io(error=%d) for %c @%llu (%s/%llu/%d)\n", iodone_w_error, btrfsic_get_block_type(dev_state->state, block), block->logical_bytenr, block->dev_state->name, block->dev_bytenr, block->mirror_num); block->iodone_w_error = iodone_w_error; if (block->submit_bio_bh_rw & REQ_FLUSH) { dev_state->last_flush_gen++; if ((dev_state->state->print_mask & BTRFSIC_PRINT_MASK_END_IO_BIO_BH)) printk(KERN_INFO "bh_end_io() new %s flush_gen=%llu\n", dev_state->name, dev_state->last_flush_gen); } if (block->submit_bio_bh_rw & REQ_FUA) block->flush_gen = 0; /* FUA completed means block is on disk */ bh->b_private = block->orig_bio_bh_private; bh->b_end_io = block->orig_bio_bh_end_io.bh; block->is_iodone = 1; /* for FLUSH, this releases the block */ bh->b_end_io(bh, uptodate); } static int btrfsic_process_written_superblock( struct btrfsic_state *state, struct btrfsic_block *const superblock, struct btrfs_super_block *const super_hdr) { int pass; superblock->generation = btrfs_super_generation(super_hdr); if (!(superblock->generation > state->max_superblock_generation || 0 == state->max_superblock_generation)) { if (state->print_mask & BTRFSIC_PRINT_MASK_SUPERBLOCK_WRITE) printk(KERN_INFO "btrfsic: superblock @%llu (%s/%llu/%d)" " with old gen %llu <= %llu\n", superblock->logical_bytenr, superblock->dev_state->name, superblock->dev_bytenr, superblock->mirror_num, btrfs_super_generation(super_hdr), state->max_superblock_generation); } else { if (state->print_mask & BTRFSIC_PRINT_MASK_SUPERBLOCK_WRITE) printk(KERN_INFO "btrfsic: got new superblock @%llu (%s/%llu/%d)" " with new gen %llu > %llu\n", superblock->logical_bytenr, superblock->dev_state->name, superblock->dev_bytenr, superblock->mirror_num, btrfs_super_generation(super_hdr), state->max_superblock_generation); state->max_superblock_generation = btrfs_super_generation(super_hdr); state->latest_superblock = superblock; } for (pass = 0; pass < 3; pass++) { int ret; u64 next_bytenr; struct btrfsic_block *next_block; struct btrfsic_block_data_ctx tmp_next_block_ctx; struct btrfsic_block_link *l; int num_copies; int mirror_num; const char *additional_string = NULL; struct btrfs_disk_key tmp_disk_key = {0}; btrfs_set_disk_key_objectid(&tmp_disk_key, BTRFS_ROOT_ITEM_KEY); btrfs_set_disk_key_objectid(&tmp_disk_key, 0); switch (pass) { case 0: btrfs_set_disk_key_objectid(&tmp_disk_key, BTRFS_ROOT_TREE_OBJECTID); additional_string = "root "; next_bytenr = btrfs_super_root(super_hdr); if (state->print_mask & BTRFSIC_PRINT_MASK_ROOT_CHUNK_LOG_TREE_LOCATION) printk(KERN_INFO "root@%llu\n", next_bytenr); break; case 1: btrfs_set_disk_key_objectid(&tmp_disk_key, BTRFS_CHUNK_TREE_OBJECTID); additional_string = "chunk "; next_bytenr = btrfs_super_chunk_root(super_hdr); if (state->print_mask & BTRFSIC_PRINT_MASK_ROOT_CHUNK_LOG_TREE_LOCATION) printk(KERN_INFO "chunk@%llu\n", next_bytenr); break; case 2: btrfs_set_disk_key_objectid(&tmp_disk_key, BTRFS_TREE_LOG_OBJECTID); additional_string = "log "; next_bytenr = btrfs_super_log_root(super_hdr); if (0 == next_bytenr) continue; if (state->print_mask & BTRFSIC_PRINT_MASK_ROOT_CHUNK_LOG_TREE_LOCATION) printk(KERN_INFO "log@%llu\n", next_bytenr); break; } num_copies = btrfs_num_copies(state->root->fs_info, next_bytenr, BTRFS_SUPER_INFO_SIZE); if (state->print_mask & BTRFSIC_PRINT_MASK_NUM_COPIES) printk(KERN_INFO "num_copies(log_bytenr=%llu) = %d\n", next_bytenr, num_copies); for (mirror_num = 1; mirror_num <= num_copies; mirror_num++) { int was_created; if (state->print_mask & BTRFSIC_PRINT_MASK_VERBOSE) printk(KERN_INFO "btrfsic_process_written_superblock(" "mirror_num=%d)\n", mirror_num); ret = btrfsic_map_block(state, next_bytenr, BTRFS_SUPER_INFO_SIZE, &tmp_next_block_ctx, mirror_num); if (ret) { printk(KERN_INFO "btrfsic: btrfsic_map_block(@%llu," " mirror=%d) failed!\n", next_bytenr, mirror_num); return -1; } next_block = btrfsic_block_lookup_or_add( state, &tmp_next_block_ctx, additional_string, 1, 0, 1, mirror_num, &was_created); if (NULL == next_block) { printk(KERN_INFO "btrfsic: error, kmalloc failed!\n"); btrfsic_release_block_ctx(&tmp_next_block_ctx); return -1; } next_block->disk_key = tmp_disk_key; if (was_created) next_block->generation = BTRFSIC_GENERATION_UNKNOWN; l = btrfsic_block_link_lookup_or_add( state, &tmp_next_block_ctx, next_block, superblock, BTRFSIC_GENERATION_UNKNOWN); btrfsic_release_block_ctx(&tmp_next_block_ctx); if (NULL == l) return -1; } } if (WARN_ON(-1 == btrfsic_check_all_ref_blocks(state, superblock, 0))) btrfsic_dump_tree(state); return 0; } static int btrfsic_check_all_ref_blocks(struct btrfsic_state *state, struct btrfsic_block *const block, int recursion_level) { struct list_head *elem_ref_to; int ret = 0; if (recursion_level >= 3 + BTRFS_MAX_LEVEL) { /* * Note that this situation can happen and does not * indicate an error in regular cases. It happens * when disk blocks are freed and later reused. * The check-integrity module is not aware of any * block free operations, it just recognizes block * write operations. Therefore it keeps the linkage * information for a block until a block is * rewritten. This can temporarily cause incorrect * and even circular linkage informations. This * causes no harm unless such blocks are referenced * by the most recent super block. */ if (state->print_mask & BTRFSIC_PRINT_MASK_VERBOSE) printk(KERN_INFO "btrfsic: abort cyclic linkage (case 1).\n"); return ret; } /* * This algorithm is recursive because the amount of used stack * space is very small and the max recursion depth is limited. */ list_for_each(elem_ref_to, &block->ref_to_list) { const struct btrfsic_block_link *const l = list_entry(elem_ref_to, struct btrfsic_block_link, node_ref_to); if (state->print_mask & BTRFSIC_PRINT_MASK_VERBOSE) printk(KERN_INFO "rl=%d, %c @%llu (%s/%llu/%d)" " %u* refers to %c @%llu (%s/%llu/%d)\n", recursion_level, btrfsic_get_block_type(state, block), block->logical_bytenr, block->dev_state->name, block->dev_bytenr, block->mirror_num, l->ref_cnt, btrfsic_get_block_type(state, l->block_ref_to), l->block_ref_to->logical_bytenr, l->block_ref_to->dev_state->name, l->block_ref_to->dev_bytenr, l->block_ref_to->mirror_num); if (l->block_ref_to->never_written) { printk(KERN_INFO "btrfs: attempt to write superblock" " which references block %c @%llu (%s/%llu/%d)" " which is never written!\n", btrfsic_get_block_type(state, l->block_ref_to), l->block_ref_to->logical_bytenr, l->block_ref_to->dev_state->name, l->block_ref_to->dev_bytenr, l->block_ref_to->mirror_num); ret = -1; } else if (!l->block_ref_to->is_iodone) { printk(KERN_INFO "btrfs: attempt to write superblock" " which references block %c @%llu (%s/%llu/%d)" " which is not yet iodone!\n", btrfsic_get_block_type(state, l->block_ref_to), l->block_ref_to->logical_bytenr, l->block_ref_to->dev_state->name, l->block_ref_to->dev_bytenr, l->block_ref_to->mirror_num); ret = -1; } else if (l->block_ref_to->iodone_w_error) { printk(KERN_INFO "btrfs: attempt to write superblock" " which references block %c @%llu (%s/%llu/%d)" " which has write error!\n", btrfsic_get_block_type(state, l->block_ref_to), l->block_ref_to->logical_bytenr, l->block_ref_to->dev_state->name, l->block_ref_to->dev_bytenr, l->block_ref_to->mirror_num); ret = -1; } else if (l->parent_generation != l->block_ref_to->generation && BTRFSIC_GENERATION_UNKNOWN != l->parent_generation && BTRFSIC_GENERATION_UNKNOWN != l->block_ref_to->generation) { printk(KERN_INFO "btrfs: attempt to write superblock" " which references block %c @%llu (%s/%llu/%d)" " with generation %llu !=" " parent generation %llu!\n", btrfsic_get_block_type(state, l->block_ref_to), l->block_ref_to->logical_bytenr, l->block_ref_to->dev_state->name, l->block_ref_to->dev_bytenr, l->block_ref_to->mirror_num, l->block_ref_to->generation, l->parent_generation); ret = -1; } else if (l->block_ref_to->flush_gen > l->block_ref_to->dev_state->last_flush_gen) { printk(KERN_INFO "btrfs: attempt to write superblock" " which references block %c @%llu (%s/%llu/%d)" " which is not flushed out of disk's write cache" " (block flush_gen=%llu," " dev->flush_gen=%llu)!\n", btrfsic_get_block_type(state, l->block_ref_to), l->block_ref_to->logical_bytenr, l->block_ref_to->dev_state->name, l->block_ref_to->dev_bytenr, l->block_ref_to->mirror_num, block->flush_gen, l->block_ref_to->dev_state->last_flush_gen); ret = -1; } else if (-1 == btrfsic_check_all_ref_blocks(state, l->block_ref_to, recursion_level + 1)) { ret = -1; } } return ret; } static int btrfsic_is_block_ref_by_superblock( const struct btrfsic_state *state, const struct btrfsic_block *block, int recursion_level) { struct list_head *elem_ref_from; if (recursion_level >= 3 + BTRFS_MAX_LEVEL) { /* refer to comment at "abort cyclic linkage (case 1)" */ if (state->print_mask & BTRFSIC_PRINT_MASK_VERBOSE) printk(KERN_INFO "btrfsic: abort cyclic linkage (case 2).\n"); return 0; } /* * This algorithm is recursive because the amount of used stack space * is very small and the max recursion depth is limited. */ list_for_each(elem_ref_from, &block->ref_from_list) { const struct btrfsic_block_link *const l = list_entry(elem_ref_from, struct btrfsic_block_link, node_ref_from); if (state->print_mask & BTRFSIC_PRINT_MASK_VERBOSE) printk(KERN_INFO "rl=%d, %c @%llu (%s/%llu/%d)" " is ref %u* from %c @%llu (%s/%llu/%d)\n", recursion_level, btrfsic_get_block_type(state, block), block->logical_bytenr, block->dev_state->name, block->dev_bytenr, block->mirror_num, l->ref_cnt, btrfsic_get_block_type(state, l->block_ref_from), l->block_ref_from->logical_bytenr, l->block_ref_from->dev_state->name, l->block_ref_from->dev_bytenr, l->block_ref_from->mirror_num); if (l->block_ref_from->is_superblock && state->latest_superblock->dev_bytenr == l->block_ref_from->dev_bytenr && state->latest_superblock->dev_state->bdev == l->block_ref_from->dev_state->bdev) return 1; else if (btrfsic_is_block_ref_by_superblock(state, l->block_ref_from, recursion_level + 1)) return 1; } return 0; } static void btrfsic_print_add_link(const struct btrfsic_state *state, const struct btrfsic_block_link *l) { printk(KERN_INFO "Add %u* link from %c @%llu (%s/%llu/%d)" " to %c @%llu (%s/%llu/%d).\n", l->ref_cnt, btrfsic_get_block_type(state, l->block_ref_from), l->block_ref_from->logical_bytenr, l->block_ref_from->dev_state->name, l->block_ref_from->dev_bytenr, l->block_ref_from->mirror_num, btrfsic_get_block_type(state, l->block_ref_to), l->block_ref_to->logical_bytenr, l->block_ref_to->dev_state->name, l->block_ref_to->dev_bytenr, l->block_ref_to->mirror_num); } static void btrfsic_print_rem_link(const struct btrfsic_state *state, const struct btrfsic_block_link *l) { printk(KERN_INFO "Rem %u* link from %c @%llu (%s/%llu/%d)" " to %c @%llu (%s/%llu/%d).\n", l->ref_cnt, btrfsic_get_block_type(state, l->block_ref_from), l->block_ref_from->logical_bytenr, l->block_ref_from->dev_state->name, l->block_ref_from->dev_bytenr, l->block_ref_from->mirror_num, btrfsic_get_block_type(state, l->block_ref_to), l->block_ref_to->logical_bytenr, l->block_ref_to->dev_state->name, l->block_ref_to->dev_bytenr, l->block_ref_to->mirror_num); } static char btrfsic_get_block_type(const struct btrfsic_state *state, const struct btrfsic_block *block) { if (block->is_superblock && state->latest_superblock->dev_bytenr == block->dev_bytenr && state->latest_superblock->dev_state->bdev == block->dev_state->bdev) return 'S'; else if (block->is_superblock) return 's'; else if (block->is_metadata) return 'M'; else return 'D'; } static void btrfsic_dump_tree(const struct btrfsic_state *state) { btrfsic_dump_tree_sub(state, state->latest_superblock, 0); } static void btrfsic_dump_tree_sub(const struct btrfsic_state *state, const struct btrfsic_block *block, int indent_level) { struct list_head *elem_ref_to; int indent_add; static char buf[80]; int cursor_position; /* * Should better fill an on-stack buffer with a complete line and * dump it at once when it is time to print a newline character. */ /* * This algorithm is recursive because the amount of used stack space * is very small and the max recursion depth is limited. */ indent_add = sprintf(buf, "%c-%llu(%s/%llu/%d)", btrfsic_get_block_type(state, block), block->logical_bytenr, block->dev_state->name, block->dev_bytenr, block->mirror_num); if (indent_level + indent_add > BTRFSIC_TREE_DUMP_MAX_INDENT_LEVEL) { printk("[...]\n"); return; } printk(buf); indent_level += indent_add; if (list_empty(&block->ref_to_list)) { printk("\n"); return; } if (block->mirror_num > 1 && !(state->print_mask & BTRFSIC_PRINT_MASK_TREE_WITH_ALL_MIRRORS)) { printk(" [...]\n"); return; } cursor_position = indent_level; list_for_each(elem_ref_to, &block->ref_to_list) { const struct btrfsic_block_link *const l = list_entry(elem_ref_to, struct btrfsic_block_link, node_ref_to); while (cursor_position < indent_level) { printk(" "); cursor_position++; } if (l->ref_cnt > 1) indent_add = sprintf(buf, " %d*--> ", l->ref_cnt); else indent_add = sprintf(buf, " --> "); if (indent_level + indent_add > BTRFSIC_TREE_DUMP_MAX_INDENT_LEVEL) { printk("[...]\n"); cursor_position = 0; continue; } printk(buf); btrfsic_dump_tree_sub(state, l->block_ref_to, indent_level + indent_add); cursor_position = 0; } } static struct btrfsic_block_link *btrfsic_block_link_lookup_or_add( struct btrfsic_state *state, struct btrfsic_block_data_ctx *next_block_ctx, struct btrfsic_block *next_block, struct btrfsic_block *from_block, u64 parent_generation) { struct btrfsic_block_link *l; l = btrfsic_block_link_hashtable_lookup(next_block_ctx->dev->bdev, next_block_ctx->dev_bytenr, from_block->dev_state->bdev, from_block->dev_bytenr, &state->block_link_hashtable); if (NULL == l) { l = btrfsic_block_link_alloc(); if (NULL == l) { printk(KERN_INFO "btrfsic: error, kmalloc" " failed!\n"); return NULL; } l->block_ref_to = next_block; l->block_ref_from = from_block; l->ref_cnt = 1; l->parent_generation = parent_generation; if (state->print_mask & BTRFSIC_PRINT_MASK_VERBOSE) btrfsic_print_add_link(state, l); list_add(&l->node_ref_to, &from_block->ref_to_list); list_add(&l->node_ref_from, &next_block->ref_from_list); btrfsic_block_link_hashtable_add(l, &state->block_link_hashtable); } else { l->ref_cnt++; l->parent_generation = parent_generation; if (state->print_mask & BTRFSIC_PRINT_MASK_VERBOSE) btrfsic_print_add_link(state, l); } return l; } static struct btrfsic_block *btrfsic_block_lookup_or_add( struct btrfsic_state *state, struct btrfsic_block_data_ctx *block_ctx, const char *additional_string, int is_metadata, int is_iodone, int never_written, int mirror_num, int *was_created) { struct btrfsic_block *block; block = btrfsic_block_hashtable_lookup(block_ctx->dev->bdev, block_ctx->dev_bytenr, &state->block_hashtable); if (NULL == block) { struct btrfsic_dev_state *dev_state; block = btrfsic_block_alloc(); if (NULL == block) { printk(KERN_INFO "btrfsic: error, kmalloc failed!\n"); return NULL; } dev_state = btrfsic_dev_state_lookup(block_ctx->dev->bdev); if (NULL == dev_state) { printk(KERN_INFO "btrfsic: error, lookup dev_state failed!\n"); btrfsic_block_free(block); return NULL; } block->dev_state = dev_state; block->dev_bytenr = block_ctx->dev_bytenr; block->logical_bytenr = block_ctx->start; block->is_metadata = is_metadata; block->is_iodone = is_iodone; block->never_written = never_written; block->mirror_num = mirror_num; if (state->print_mask & BTRFSIC_PRINT_MASK_VERBOSE) printk(KERN_INFO "New %s%c-block @%llu (%s/%llu/%d)\n", additional_string, btrfsic_get_block_type(state, block), block->logical_bytenr, dev_state->name, block->dev_bytenr, mirror_num); list_add(&block->all_blocks_node, &state->all_blocks_list); btrfsic_block_hashtable_add(block, &state->block_hashtable); if (NULL != was_created) *was_created = 1; } else { if (NULL != was_created) *was_created = 0; } return block; } static void btrfsic_cmp_log_and_dev_bytenr(struct btrfsic_state *state, u64 bytenr, struct btrfsic_dev_state *dev_state, u64 dev_bytenr) { int num_copies; int mirror_num; int ret; struct btrfsic_block_data_ctx block_ctx; int match = 0; num_copies = btrfs_num_copies(state->root->fs_info, bytenr, state->metablock_size); for (mirror_num = 1; mirror_num <= num_copies; mirror_num++) { ret = btrfsic_map_block(state, bytenr, state->metablock_size, &block_ctx, mirror_num); if (ret) { printk(KERN_INFO "btrfsic:" " btrfsic_map_block(logical @%llu," " mirror %d) failed!\n", bytenr, mirror_num); continue; } if (dev_state->bdev == block_ctx.dev->bdev && dev_bytenr == block_ctx.dev_bytenr) { match++; btrfsic_release_block_ctx(&block_ctx); break; } btrfsic_release_block_ctx(&block_ctx); } if (WARN_ON(!match)) { printk(KERN_INFO "btrfs: attempt to write M-block which contains logical bytenr that doesn't map to dev+physical bytenr of submit_bio," " buffer->log_bytenr=%llu, submit_bio(bdev=%s," " phys_bytenr=%llu)!\n", bytenr, dev_state->name, dev_bytenr); for (mirror_num = 1; mirror_num <= num_copies; mirror_num++) { ret = btrfsic_map_block(state, bytenr, state->metablock_size, &block_ctx, mirror_num); if (ret) continue; printk(KERN_INFO "Read logical bytenr @%llu maps to" " (%s/%llu/%d)\n", bytenr, block_ctx.dev->name, block_ctx.dev_bytenr, mirror_num); } } } static struct btrfsic_dev_state *btrfsic_dev_state_lookup( struct block_device *bdev) { struct btrfsic_dev_state *ds; ds = btrfsic_dev_state_hashtable_lookup(bdev, &btrfsic_dev_state_hashtable); return ds; } int btrfsic_submit_bh(int rw, struct buffer_head *bh) { struct btrfsic_dev_state *dev_state; if (!btrfsic_is_initialized) return submit_bh(rw, bh); mutex_lock(&btrfsic_mutex); /* since btrfsic_submit_bh() might also be called before * btrfsic_mount(), this might return NULL */ dev_state = btrfsic_dev_state_lookup(bh->b_bdev); /* Only called to write the superblock (incl. FLUSH/FUA) */ if (NULL != dev_state && (rw & WRITE) && bh->b_size > 0) { u64 dev_bytenr; dev_bytenr = 4096 * bh->b_blocknr; if (dev_state->state->print_mask & BTRFSIC_PRINT_MASK_SUBMIT_BIO_BH) printk(KERN_INFO "submit_bh(rw=0x%x, blocknr=%llu (bytenr %llu)," " size=%zu, data=%p, bdev=%p)\n", rw, (unsigned long long)bh->b_blocknr, dev_bytenr, bh->b_size, bh->b_data, bh->b_bdev); btrfsic_process_written_block(dev_state, dev_bytenr, &bh->b_data, 1, NULL, NULL, bh, rw); } else if (NULL != dev_state && (rw & REQ_FLUSH)) { if (dev_state->state->print_mask & BTRFSIC_PRINT_MASK_SUBMIT_BIO_BH) printk(KERN_INFO "submit_bh(rw=0x%x FLUSH, bdev=%p)\n", rw, bh->b_bdev); if (!dev_state->dummy_block_for_bio_bh_flush.is_iodone) { if ((dev_state->state->print_mask & (BTRFSIC_PRINT_MASK_SUBMIT_BIO_BH | BTRFSIC_PRINT_MASK_VERBOSE))) printk(KERN_INFO "btrfsic_submit_bh(%s) with FLUSH" " but dummy block already in use" " (ignored)!\n", dev_state->name); } else { struct btrfsic_block *const block = &dev_state->dummy_block_for_bio_bh_flush; block->is_iodone = 0; block->never_written = 0; block->iodone_w_error = 0; block->flush_gen = dev_state->last_flush_gen + 1; block->submit_bio_bh_rw = rw; block->orig_bio_bh_private = bh->b_private; block->orig_bio_bh_end_io.bh = bh->b_end_io; block->next_in_same_bio = NULL; bh->b_private = block; bh->b_end_io = btrfsic_bh_end_io; } } mutex_unlock(&btrfsic_mutex); return submit_bh(rw, bh); } static void __btrfsic_submit_bio(int rw, struct bio *bio) { struct btrfsic_dev_state *dev_state; if (!btrfsic_is_initialized) return; mutex_lock(&btrfsic_mutex); /* since btrfsic_submit_bio() is also called before * btrfsic_mount(), this might return NULL */ dev_state = btrfsic_dev_state_lookup(bio->bi_bdev); if (NULL != dev_state && (rw & WRITE) && NULL != bio->bi_io_vec) { unsigned int i; u64 dev_bytenr; u64 cur_bytenr; int bio_is_patched; char **mapped_datav; dev_bytenr = 512 * bio->bi_iter.bi_sector; bio_is_patched = 0; if (dev_state->state->print_mask & BTRFSIC_PRINT_MASK_SUBMIT_BIO_BH) printk(KERN_INFO "submit_bio(rw=0x%x, bi_vcnt=%u," " bi_sector=%llu (bytenr %llu), bi_bdev=%p)\n", rw, bio->bi_vcnt, (unsigned long long)bio->bi_iter.bi_sector, dev_bytenr, bio->bi_bdev); mapped_datav = kmalloc_array(bio->bi_vcnt, sizeof(*mapped_datav), GFP_NOFS); if (!mapped_datav) goto leave; cur_bytenr = dev_bytenr; for (i = 0; i < bio->bi_vcnt; i++) { BUG_ON(bio->bi_io_vec[i].bv_len != PAGE_CACHE_SIZE); mapped_datav[i] = kmap(bio->bi_io_vec[i].bv_page); if (!mapped_datav[i]) { while (i > 0) { i--; kunmap(bio->bi_io_vec[i].bv_page); } kfree(mapped_datav); goto leave; } if (dev_state->state->print_mask & BTRFSIC_PRINT_MASK_SUBMIT_BIO_BH_VERBOSE) printk(KERN_INFO "#%u: bytenr=%llu, len=%u, offset=%u\n", i, cur_bytenr, bio->bi_io_vec[i].bv_len, bio->bi_io_vec[i].bv_offset); cur_bytenr += bio->bi_io_vec[i].bv_len; } btrfsic_process_written_block(dev_state, dev_bytenr, mapped_datav, bio->bi_vcnt, bio, &bio_is_patched, NULL, rw); while (i > 0) { i--; kunmap(bio->bi_io_vec[i].bv_page); } kfree(mapped_datav); } else if (NULL != dev_state && (rw & REQ_FLUSH)) { if (dev_state->state->print_mask & BTRFSIC_PRINT_MASK_SUBMIT_BIO_BH) printk(KERN_INFO "submit_bio(rw=0x%x FLUSH, bdev=%p)\n", rw, bio->bi_bdev); if (!dev_state->dummy_block_for_bio_bh_flush.is_iodone) { if ((dev_state->state->print_mask & (BTRFSIC_PRINT_MASK_SUBMIT_BIO_BH | BTRFSIC_PRINT_MASK_VERBOSE))) printk(KERN_INFO "btrfsic_submit_bio(%s) with FLUSH" " but dummy block already in use" " (ignored)!\n", dev_state->name); } else { struct btrfsic_block *const block = &dev_state->dummy_block_for_bio_bh_flush; block->is_iodone = 0; block->never_written = 0; block->iodone_w_error = 0; block->flush_gen = dev_state->last_flush_gen + 1; block->submit_bio_bh_rw = rw; block->orig_bio_bh_private = bio->bi_private; block->orig_bio_bh_end_io.bio = bio->bi_end_io; block->next_in_same_bio = NULL; bio->bi_private = block; bio->bi_end_io = btrfsic_bio_end_io; } } leave: mutex_unlock(&btrfsic_mutex); } void btrfsic_submit_bio(int rw, struct bio *bio) { __btrfsic_submit_bio(rw, bio); submit_bio(rw, bio); } int btrfsic_submit_bio_wait(int rw, struct bio *bio) { __btrfsic_submit_bio(rw, bio); return submit_bio_wait(rw, bio); } int btrfsic_mount(struct btrfs_root *root, struct btrfs_fs_devices *fs_devices, int including_extent_data, u32 print_mask) { int ret; struct btrfsic_state *state; struct list_head *dev_head = &fs_devices->devices; struct btrfs_device *device; if (root->nodesize & ((u64)PAGE_CACHE_SIZE - 1)) { printk(KERN_INFO "btrfsic: cannot handle nodesize %d not being a multiple of PAGE_CACHE_SIZE %ld!\n", root->nodesize, PAGE_CACHE_SIZE); return -1; } if (root->sectorsize & ((u64)PAGE_CACHE_SIZE - 1)) { printk(KERN_INFO "btrfsic: cannot handle sectorsize %d not being a multiple of PAGE_CACHE_SIZE %ld!\n", root->sectorsize, PAGE_CACHE_SIZE); return -1; } state = kzalloc(sizeof(*state), GFP_KERNEL | __GFP_NOWARN | __GFP_REPEAT); if (!state) { state = vzalloc(sizeof(*state)); if (!state) { printk(KERN_INFO "btrfs check-integrity: vzalloc() failed!\n"); return -1; } } if (!btrfsic_is_initialized) { mutex_init(&btrfsic_mutex); btrfsic_dev_state_hashtable_init(&btrfsic_dev_state_hashtable); btrfsic_is_initialized = 1; } mutex_lock(&btrfsic_mutex); state->root = root; state->print_mask = print_mask; state->include_extent_data = including_extent_data; state->csum_size = 0; state->metablock_size = root->nodesize; state->datablock_size = root->sectorsize; INIT_LIST_HEAD(&state->all_blocks_list); btrfsic_block_hashtable_init(&state->block_hashtable); btrfsic_block_link_hashtable_init(&state->block_link_hashtable); state->max_superblock_generation = 0; state->latest_superblock = NULL; list_for_each_entry(device, dev_head, dev_list) { struct btrfsic_dev_state *ds; char *p; if (!device->bdev || !device->name) continue; ds = btrfsic_dev_state_alloc(); if (NULL == ds) { printk(KERN_INFO "btrfs check-integrity: kmalloc() failed!\n"); mutex_unlock(&btrfsic_mutex); return -1; } ds->bdev = device->bdev; ds->state = state; bdevname(ds->bdev, ds->name); ds->name[BDEVNAME_SIZE - 1] = '\0'; for (p = ds->name; *p != '\0'; p++); while (p > ds->name && *p != '/') p--; if (*p == '/') p++; strlcpy(ds->name, p, sizeof(ds->name)); btrfsic_dev_state_hashtable_add(ds, &btrfsic_dev_state_hashtable); } ret = btrfsic_process_superblock(state, fs_devices); if (0 != ret) { mutex_unlock(&btrfsic_mutex); btrfsic_unmount(root, fs_devices); return ret; } if (state->print_mask & BTRFSIC_PRINT_MASK_INITIAL_DATABASE) btrfsic_dump_database(state); if (state->print_mask & BTRFSIC_PRINT_MASK_INITIAL_TREE) btrfsic_dump_tree(state); mutex_unlock(&btrfsic_mutex); return 0; } void btrfsic_unmount(struct btrfs_root *root, struct btrfs_fs_devices *fs_devices) { struct list_head *elem_all; struct list_head *tmp_all; struct btrfsic_state *state; struct list_head *dev_head = &fs_devices->devices; struct btrfs_device *device; if (!btrfsic_is_initialized) return; mutex_lock(&btrfsic_mutex); state = NULL; list_for_each_entry(device, dev_head, dev_list) { struct btrfsic_dev_state *ds; if (!device->bdev || !device->name) continue; ds = btrfsic_dev_state_hashtable_lookup( device->bdev, &btrfsic_dev_state_hashtable); if (NULL != ds) { state = ds->state; btrfsic_dev_state_hashtable_remove(ds); btrfsic_dev_state_free(ds); } } if (NULL == state) { printk(KERN_INFO "btrfsic: error, cannot find state information" " on umount!\n"); mutex_unlock(&btrfsic_mutex); return; } /* * Don't care about keeping the lists' state up to date, * just free all memory that was allocated dynamically. * Free the blocks and the block_links. */ list_for_each_safe(elem_all, tmp_all, &state->all_blocks_list) { struct btrfsic_block *const b_all = list_entry(elem_all, struct btrfsic_block, all_blocks_node); struct list_head *elem_ref_to; struct list_head *tmp_ref_to; list_for_each_safe(elem_ref_to, tmp_ref_to, &b_all->ref_to_list) { struct btrfsic_block_link *const l = list_entry(elem_ref_to, struct btrfsic_block_link, node_ref_to); if (state->print_mask & BTRFSIC_PRINT_MASK_VERBOSE) btrfsic_print_rem_link(state, l); l->ref_cnt--; if (0 == l->ref_cnt) btrfsic_block_link_free(l); } if (b_all->is_iodone || b_all->never_written) btrfsic_block_free(b_all); else printk(KERN_INFO "btrfs: attempt to free %c-block" " @%llu (%s/%llu/%d) on umount which is" " not yet iodone!\n", btrfsic_get_block_type(state, b_all), b_all->logical_bytenr, b_all->dev_state->name, b_all->dev_bytenr, b_all->mirror_num); } mutex_unlock(&btrfsic_mutex); kvfree(state); }