diff --git a/0200-btrfs-progs-print-qgroup-excl-as-unsigned.patch b/0200-btrfs-progs-print-qgroup-excl-as-unsigned.patch new file mode 100644 index 0000000..3d2c559 --- /dev/null +++ b/0200-btrfs-progs-print-qgroup-excl-as-unsigned.patch @@ -0,0 +1,65 @@ +From 8b40b00f23806115c9f03344227b6590cb187a96 Mon Sep 17 00:00:00 2001 +From: Mark Fasheh +Date: Thu, 1 May 2014 22:35:15 -0700 +Subject: [PATCH 1/3] btrfs-progs: print qgroup excl as unsigned +References: bnc#865621 +Upstream: submitted + +It's unsigned in the structure definition. + +Signed-off-by: Mark Fasheh +--- + print-tree.c | 12 ++++++------ + qgroup.c | 4 ++-- + 2 files changed, 8 insertions(+), 8 deletions(-) + +diff --git a/print-tree.c b/print-tree.c +index 7263b09..adef94a 100644 +--- a/print-tree.c ++++ b/print-tree.c +@@ -884,18 +884,18 @@ void btrfs_print_leaf(struct btrfs_root *root, struct extent_buffer *l) + qg_info = btrfs_item_ptr(l, i, + struct btrfs_qgroup_info_item); + printf("\t\tgeneration %llu\n" +- "\t\treferenced %lld referenced compressed %lld\n" +- "\t\texclusive %lld exclusive compressed %lld\n", ++ "\t\treferenced %llu referenced compressed %llu\n" ++ "\t\texclusive %llu exclusive compressed %llu\n", + (unsigned long long) + btrfs_qgroup_info_generation(l, qg_info), +- (long long) ++ (unsigned long long) + btrfs_qgroup_info_referenced(l, qg_info), +- (long long) ++ (unsigned long long) + btrfs_qgroup_info_referenced_compressed(l, + qg_info), +- (long long) ++ (unsigned long long) + btrfs_qgroup_info_exclusive(l, qg_info), +- (long long) ++ (unsigned long long) + btrfs_qgroup_info_exclusive_compressed(l, + qg_info)); + break; +diff --git a/qgroup.c b/qgroup.c +index 94d1feb..368b262 100644 +--- a/qgroup.c ++++ b/qgroup.c +@@ -203,11 +203,11 @@ static void print_qgroup_column(struct btrfs_qgroup *qgroup, + print_qgroup_column_add_blank(BTRFS_QGROUP_QGROUPID, len); + break; + case BTRFS_QGROUP_RFER: +- len = printf("%lld", qgroup->rfer); ++ len = printf("%llu", qgroup->rfer); + print_qgroup_column_add_blank(BTRFS_QGROUP_RFER, len); + break; + case BTRFS_QGROUP_EXCL: +- len = printf("%lld", qgroup->excl); ++ len = printf("%llu", qgroup->excl); + print_qgroup_column_add_blank(BTRFS_QGROUP_EXCL, len); + break; + case BTRFS_QGROUP_PARENT: +-- +1.8.4 + diff --git a/0201-btrfs-progs-import-ulist.patch b/0201-btrfs-progs-import-ulist.patch new file mode 100644 index 0000000..b52f9a2 --- /dev/null +++ b/0201-btrfs-progs-import-ulist.patch @@ -0,0 +1,374 @@ +From e1857c491c61040cd845b5e08f1d996d3e3557a8 Mon Sep 17 00:00:00 2001 +From: Mark Fasheh +Date: Thu, 1 May 2014 14:34:30 -0700 +Subject: [PATCH 2/3] btrfs-progs: import ulist +References: bnc#865621 +Upstream: submitted + +qgroup-verify.c wants this for walking root refs. + +Signed-off-by: Mark Fasheh +--- + Makefile | 3 +- + kerncompat.h | 2 +- + ulist.c | 253 +++++++++++++++++++++++++++++++++++++++++++++++++++++++++++ + ulist.h | 66 ++++++++++++++++ + 4 files changed, 322 insertions(+), 2 deletions(-) + create mode 100644 ulist.c + create mode 100644 ulist.h + +Index: btrfs-progs-v3.14.1/Makefile +=================================================================== +--- btrfs-progs-v3.14.1.orig/Makefile ++++ btrfs-progs-v3.14.1/Makefile +@@ -10,7 +10,7 @@ objects = ctree.o disk-io.o radix-tree.o + root-tree.o dir-item.o file-item.o inode-item.o inode-map.o \ + extent-cache.o extent_io.o volumes.o utils.o repair.o \ + qgroup.o raid6.o free-space-cache.o list_sort.o props.o \ +- utils-lib.o string_table.o ++ utils-lib.o string_table.o ulist.o + cmds_objects = cmds-subvolume.o cmds-filesystem.o cmds-device.o cmds-scrub.o \ + cmds-inspect.o cmds-balance.o cmds-send.o cmds-receive.o \ + cmds-quota.o cmds-qgroup.o cmds-replace.o cmds-check.o \ +Index: btrfs-progs-v3.14.1/kerncompat.h +=================================================================== +--- btrfs-progs-v3.14.1.orig/kerncompat.h ++++ btrfs-progs-v3.14.1/kerncompat.h +@@ -235,7 +235,7 @@ static inline long IS_ERR(const void *pt + + #define BUG_ON(c) assert(!(c)) + #define WARN_ON(c) assert(!(c)) +- ++#define ASSERT(c) assert(c) + + #define container_of(ptr, type, member) ({ \ + const typeof( ((type *)0)->member ) *__mptr = (ptr); \ +Index: btrfs-progs-v3.14.1/ulist.c +=================================================================== +--- /dev/null ++++ btrfs-progs-v3.14.1/ulist.c +@@ -0,0 +1,253 @@ ++/* ++ * Copyright (C) 2011 STRATO AG ++ * written by Arne Jansen ++ * Distributed under the GNU GPL license version 2. ++ */ ++ ++//#include ++#include ++#include "kerncompat.h" ++#include "ulist.h" ++#include "ctree.h" ++ ++/* ++ * ulist is a generic data structure to hold a collection of unique u64 ++ * values. The only operations it supports is adding to the list and ++ * enumerating it. ++ * It is possible to store an auxiliary value along with the key. ++ * ++ * A sample usage for ulists is the enumeration of directed graphs without ++ * visiting a node twice. The pseudo-code could look like this: ++ * ++ * ulist = ulist_alloc(); ++ * ulist_add(ulist, root); ++ * ULIST_ITER_INIT(&uiter); ++ * ++ * while ((elem = ulist_next(ulist, &uiter)) { ++ * for (all child nodes n in elem) ++ * ulist_add(ulist, n); ++ * do something useful with the node; ++ * } ++ * ulist_free(ulist); ++ * ++ * This assumes the graph nodes are adressable by u64. This stems from the ++ * usage for tree enumeration in btrfs, where the logical addresses are ++ * 64 bit. ++ * ++ * It is also useful for tree enumeration which could be done elegantly ++ * recursively, but is not possible due to kernel stack limitations. The ++ * loop would be similar to the above. ++ */ ++ ++/** ++ * ulist_init - freshly initialize a ulist ++ * @ulist: the ulist to initialize ++ * ++ * Note: don't use this function to init an already used ulist, use ++ * ulist_reinit instead. ++ */ ++void ulist_init(struct ulist *ulist) ++{ ++ INIT_LIST_HEAD(&ulist->nodes); ++ ulist->root = RB_ROOT; ++ ulist->nnodes = 0; ++} ++ ++/** ++ * ulist_fini - free up additionally allocated memory for the ulist ++ * @ulist: the ulist from which to free the additional memory ++ * ++ * This is useful in cases where the base 'struct ulist' has been statically ++ * allocated. ++ */ ++static void ulist_fini(struct ulist *ulist) ++{ ++ struct ulist_node *node; ++ struct ulist_node *next; ++ ++ list_for_each_entry_safe(node, next, &ulist->nodes, list) { ++ kfree(node); ++ } ++ ulist->root = RB_ROOT; ++ INIT_LIST_HEAD(&ulist->nodes); ++} ++ ++/** ++ * ulist_reinit - prepare a ulist for reuse ++ * @ulist: ulist to be reused ++ * ++ * Free up all additional memory allocated for the list elements and reinit ++ * the ulist. ++ */ ++void ulist_reinit(struct ulist *ulist) ++{ ++ ulist_fini(ulist); ++ ulist_init(ulist); ++} ++ ++/** ++ * ulist_alloc - dynamically allocate a ulist ++ * @gfp_mask: allocation flags to for base allocation ++ * ++ * The allocated ulist will be returned in an initialized state. ++ */ ++struct ulist *ulist_alloc(gfp_t gfp_mask) ++{ ++ struct ulist *ulist = kmalloc(sizeof(*ulist), gfp_mask); ++ ++ if (!ulist) ++ return NULL; ++ ++ ulist_init(ulist); ++ ++ return ulist; ++} ++ ++/** ++ * ulist_free - free dynamically allocated ulist ++ * @ulist: ulist to free ++ * ++ * It is not necessary to call ulist_fini before. ++ */ ++void ulist_free(struct ulist *ulist) ++{ ++ if (!ulist) ++ return; ++ ulist_fini(ulist); ++ kfree(ulist); ++} ++ ++static struct ulist_node *ulist_rbtree_search(struct ulist *ulist, u64 val) ++{ ++ struct rb_node *n = ulist->root.rb_node; ++ struct ulist_node *u = NULL; ++ ++ while (n) { ++ u = rb_entry(n, struct ulist_node, rb_node); ++ if (u->val < val) ++ n = n->rb_right; ++ else if (u->val > val) ++ n = n->rb_left; ++ else ++ return u; ++ } ++ return NULL; ++} ++ ++static int ulist_rbtree_insert(struct ulist *ulist, struct ulist_node *ins) ++{ ++ struct rb_node **p = &ulist->root.rb_node; ++ struct rb_node *parent = NULL; ++ struct ulist_node *cur = NULL; ++ ++ while (*p) { ++ parent = *p; ++ cur = rb_entry(parent, struct ulist_node, rb_node); ++ ++ if (cur->val < ins->val) ++ p = &(*p)->rb_right; ++ else if (cur->val > ins->val) ++ p = &(*p)->rb_left; ++ else ++ return -EEXIST; ++ } ++ rb_link_node(&ins->rb_node, parent, p); ++ rb_insert_color(&ins->rb_node, &ulist->root); ++ return 0; ++} ++ ++/** ++ * ulist_add - add an element to the ulist ++ * @ulist: ulist to add the element to ++ * @val: value to add to ulist ++ * @aux: auxiliary value to store along with val ++ * @gfp_mask: flags to use for allocation ++ * ++ * Note: locking must be provided by the caller. In case of rwlocks write ++ * locking is needed ++ * ++ * Add an element to a ulist. The @val will only be added if it doesn't ++ * already exist. If it is added, the auxiliary value @aux is stored along with ++ * it. In case @val already exists in the ulist, @aux is ignored, even if ++ * it differs from the already stored value. ++ * ++ * ulist_add returns 0 if @val already exists in ulist and 1 if @val has been ++ * inserted. ++ * In case of allocation failure -ENOMEM is returned and the ulist stays ++ * unaltered. ++ */ ++int ulist_add(struct ulist *ulist, u64 val, u64 aux, gfp_t gfp_mask) ++{ ++ return ulist_add_merge(ulist, val, aux, NULL, gfp_mask); ++} ++ ++int ulist_add_merge(struct ulist *ulist, u64 val, u64 aux, ++ u64 *old_aux, gfp_t gfp_mask) ++{ ++ int ret; ++ struct ulist_node *node; ++ ++ node = ulist_rbtree_search(ulist, val); ++ if (node) { ++ if (old_aux) ++ *old_aux = node->aux; ++ return 0; ++ } ++ node = kmalloc(sizeof(*node), gfp_mask); ++ if (!node) ++ return -ENOMEM; ++ ++ node->val = val; ++ node->aux = aux; ++#ifdef CONFIG_BTRFS_DEBUG ++ node->seqnum = ulist->nnodes; ++#endif ++ ++ ret = ulist_rbtree_insert(ulist, node); ++ ASSERT(!ret); ++ list_add_tail(&node->list, &ulist->nodes); ++ ulist->nnodes++; ++ ++ return 1; ++} ++ ++/** ++ * ulist_next - iterate ulist ++ * @ulist: ulist to iterate ++ * @uiter: iterator variable, initialized with ULIST_ITER_INIT(&iterator) ++ * ++ * Note: locking must be provided by the caller. In case of rwlocks only read ++ * locking is needed ++ * ++ * This function is used to iterate an ulist. ++ * It returns the next element from the ulist or %NULL when the ++ * end is reached. No guarantee is made with respect to the order in which ++ * the elements are returned. They might neither be returned in order of ++ * addition nor in ascending order. ++ * It is allowed to call ulist_add during an enumeration. Newly added items ++ * are guaranteed to show up in the running enumeration. ++ */ ++struct ulist_node *ulist_next(struct ulist *ulist, struct ulist_iterator *uiter) ++{ ++ struct ulist_node *node; ++ ++ if (list_empty(&ulist->nodes)) ++ return NULL; ++ if (uiter->cur_list && uiter->cur_list->next == &ulist->nodes) ++ return NULL; ++ if (uiter->cur_list) { ++ uiter->cur_list = uiter->cur_list->next; ++ } else { ++ uiter->cur_list = ulist->nodes.next; ++#ifdef CONFIG_BTRFS_DEBUG ++ uiter->i = 0; ++#endif ++ } ++ node = list_entry(uiter->cur_list, struct ulist_node, list); ++#ifdef CONFIG_BTRFS_DEBUG ++ ASSERT(node->seqnum == uiter->i); ++ ASSERT(uiter->i >= 0 && uiter->i < ulist->nnodes); ++ uiter->i++; ++#endif ++ return node; ++} +Index: btrfs-progs-v3.14.1/ulist.h +=================================================================== +--- /dev/null ++++ btrfs-progs-v3.14.1/ulist.h +@@ -0,0 +1,66 @@ ++/* ++ * Copyright (C) 2011 STRATO AG ++ * written by Arne Jansen ++ * Distributed under the GNU GPL license version 2. ++ * ++ */ ++ ++#ifndef __ULIST__ ++#define __ULIST__ ++ ++#include "kerncompat.h" ++#include "list.h" ++#include "rbtree.h" ++ ++/* ++ * ulist is a generic data structure to hold a collection of unique u64 ++ * values. The only operations it supports is adding to the list and ++ * enumerating it. ++ * It is possible to store an auxiliary value along with the key. ++ * ++ */ ++struct ulist_iterator { ++#ifdef CONFIG_BTRFS_DEBUG ++ int i; ++#endif ++ struct list_head *cur_list; /* hint to start search */ ++}; ++ ++/* ++ * element of the list ++ */ ++struct ulist_node { ++ u64 val; /* value to store */ ++ u64 aux; /* auxiliary value saved along with the val */ ++ ++#ifdef CONFIG_BTRFS_DEBUG ++ int seqnum; /* sequence number this node is added */ ++#endif ++ ++ struct list_head list; /* used to link node */ ++ struct rb_node rb_node; /* used to speed up search */ ++}; ++ ++struct ulist { ++ /* ++ * number of elements stored in list ++ */ ++ unsigned long nnodes; ++ ++ struct list_head nodes; ++ struct rb_root root; ++}; ++ ++void ulist_init(struct ulist *ulist); ++void ulist_reinit(struct ulist *ulist); ++struct ulist *ulist_alloc(gfp_t gfp_mask); ++void ulist_free(struct ulist *ulist); ++int ulist_add(struct ulist *ulist, u64 val, u64 aux, gfp_t gfp_mask); ++int ulist_add_merge(struct ulist *ulist, u64 val, u64 aux, ++ u64 *old_aux, gfp_t gfp_mask); ++struct ulist_node *ulist_next(struct ulist *ulist, ++ struct ulist_iterator *uiter); ++ ++#define ULIST_ITER_INIT(uiter) ((uiter)->cur_list = NULL) ++ ++#endif diff --git a/0202-btrfs-progs-add-quota-group-verify-code.patch b/0202-btrfs-progs-add-quota-group-verify-code.patch new file mode 100644 index 0000000..08cc35c --- /dev/null +++ b/0202-btrfs-progs-add-quota-group-verify-code.patch @@ -0,0 +1,1385 @@ +From c8c1814a8b10fab0fae7a32ef239ec8847a0ca81 Mon Sep 17 00:00:00 2001 +From: Mark Fasheh +Date: Thu, 1 May 2014 22:44:24 -0700 +Subject: [PATCH 3/3] btrfs-progs: add quota group verify code +References: bnc#865621 +Upstream: submitted + +This patch adds functionality (in qgroup-verify.c) to compute bytecounts in +subvolume quota groups. The original groups are read in and stored in memory +so that after we compute our own bytecounts, we can compare them with those +on disk. A print function is provided to do this comparison and show the +results on the console. + +A 'qgroup check' pass is added to btrfsck. If any subvolume quota groups +differ from what we compute, the differences for them are printed. We also +provide an option '--qgroup-report' which will run only the quota check code +and print a report on all quota groups. Other than making it possible to +verify that our qgroup changes work correctly, this mode can also be used in +xfstests for automated checking after qgroup tests. + +This patch does not address the following: +- compressed counts are identical to non compressed, because kernel doesn't + make the distinction yet. Adding the code to verify compressed counts + shouldn't be hard at all though once kernel can do this. +- It is only concerned with subvolume quota groups (like most of + btrfs-progs). + +Signed-off-by: Mark Fasheh +--- + Makefile | 2 +- + cmds-check.c | 24 ++ + ctree.h | 10 + + disk-io.c | 16 +- + print-tree.c | 2 +- + print-tree.h | 1 + + qgroup-verify.c | 1085 +++++++++++++++++++++++++++++++++++++++++++++++++++++++ + qgroup-verify.h | 25 ++ + 8 files changed, 1161 insertions(+), 4 deletions(-) + create mode 100644 qgroup-verify.c + create mode 100644 qgroup-verify.h + +Index: btrfs-progs-v3.14.1/Makefile +=================================================================== +--- btrfs-progs-v3.14.1.orig/Makefile ++++ btrfs-progs-v3.14.1/Makefile +@@ -10,7 +10,7 @@ objects = ctree.o disk-io.o radix-tree.o + root-tree.o dir-item.o file-item.o inode-item.o inode-map.o \ + extent-cache.o extent_io.o volumes.o utils.o repair.o \ + qgroup.o raid6.o free-space-cache.o list_sort.o props.o \ +- utils-lib.o string_table.o ulist.o ++ utils-lib.o string_table.o ulist.o qgroup-verify.o + cmds_objects = cmds-subvolume.o cmds-filesystem.o cmds-device.o cmds-scrub.o \ + cmds-inspect.o cmds-balance.o cmds-send.o cmds-receive.o \ + cmds-quota.o cmds-qgroup.o cmds-replace.o cmds-check.o \ +Index: btrfs-progs-v3.14.1/cmds-check.c +=================================================================== +--- btrfs-progs-v3.14.1.orig/cmds-check.c ++++ btrfs-progs-v3.14.1/cmds-check.c +@@ -38,6 +38,7 @@ + #include "commands.h" + #include "free-space-cache.h" + #include "btrfsck.h" ++#include "qgroup-verify.h" + + static u64 bytes_used = 0; + static u64 total_csum_bytes = 0; +@@ -6472,6 +6473,7 @@ static struct option long_options[] = { + { "init-csum-tree", 0, NULL, 0 }, + { "init-extent-tree", 0, NULL, 0 }, + { "backup", 0, NULL, 0 }, ++ { "qgroup-report", 0, NULL, 'Q' }, + { NULL, 0, NULL, 0} + }; + +@@ -6484,6 +6486,7 @@ const char * const cmd_check_usage[] = { + "--repair try to repair the filesystem", + "--init-csum-tree create a new CRC tree", + "--init-extent-tree create a new extent tree", ++ "--qgroup-report print a report on qgroup consistency", + NULL + }; + +@@ -6498,6 +6501,7 @@ int cmd_check(int argc, char **argv) + u64 num; + int option_index = 0; + int init_csum_tree = 0; ++ int qgroup_report = 0; + enum btrfs_open_ctree_flags ctree_flags = + OPEN_CTREE_PARTIAL | OPEN_CTREE_EXCLUSIVE; + +@@ -6524,6 +6528,9 @@ int cmd_check(int argc, char **argv) + printf("using SB copy %llu, bytenr %llu\n", num, + (unsigned long long)bytenr); + break; ++ case 'Q': ++ qgroup_report = 1; ++ break; + case '?': + case 'h': + usage(cmd_check_usage); +@@ -6588,6 +6595,14 @@ int cmd_check(int argc, char **argv) + } + + uuid_unparse(info->super_copy->fsid, uuidbuf); ++ if (qgroup_report) { ++ printf("Print quota groups for %s\nUUID: %s\n", argv[optind], ++ uuidbuf); ++ ret = qgroup_verify_all(info); ++ if (ret == 0) ++ print_qgroup_report(1); ++ goto close_out; ++ } + printf("Checking filesystem on %s\nUUID: %s\n", argv[optind], uuidbuf); + + if (!extent_buffer_uptodate(info->tree_root->node) || +@@ -6691,11 +6706,20 @@ int cmd_check(int argc, char **argv) + free(bad); + } + ++ if (info->quota_enabled) { ++ int err; ++ fprintf(stderr, "checking quota groups\n"); ++ err = qgroup_verify_all(info); ++ if (err) ++ goto out; ++ } ++ + if (!list_empty(&root->fs_info->recow_ebs)) { + fprintf(stderr, "Transid errors in file system\n"); + ret = 1; + } + out: ++ print_qgroup_report(0); + if (found_old_backref) { /* + * there was a disk format change when mixed + * backref was in testing tree. The old format +Index: btrfs-progs-v3.14.1/ctree.h +=================================================================== +--- btrfs-progs-v3.14.1.orig/ctree.h ++++ btrfs-progs-v3.14.1/ctree.h +@@ -950,6 +950,7 @@ struct btrfs_fs_info { + struct btrfs_root *chunk_root; + struct btrfs_root *dev_root; + struct btrfs_root *csum_root; ++ struct btrfs_root *quota_root; + + struct rb_root fs_root_tree; + +@@ -995,6 +996,7 @@ struct btrfs_fs_info { + unsigned int readonly:1; + unsigned int on_restoring:1; + unsigned int is_chunk_recover:1; ++ unsigned int quota_enabled:1; + + int (*free_extent_hook)(struct btrfs_trans_handle *trans, + struct btrfs_root *root, +@@ -2389,4 +2391,12 @@ int btrfs_csum_truncate(struct btrfs_tra + int btrfs_lookup_uuid_subvol_item(int fd, const u8 *uuid, u64 *subvol_id); + int btrfs_lookup_uuid_received_subvol_item(int fd, const u8 *uuid, + u64 *subvol_id); ++ ++static inline int is_fstree(u64 rootid) ++{ ++ if (rootid == BTRFS_FS_TREE_OBJECTID || ++ (signed long long)rootid >= (signed long long)BTRFS_FIRST_FREE_OBJECTID) ++ return 1; ++ return 0; ++} + #endif +Index: btrfs-progs-v3.14.1/disk-io.c +=================================================================== +--- btrfs-progs-v3.14.1.orig/disk-io.c ++++ btrfs-progs-v3.14.1/disk-io.c +@@ -571,7 +571,6 @@ static int find_and_setup_log_root(struc + return 0; + } + +- + int btrfs_free_fs_root(struct btrfs_root *root) + { + if (root->node) +@@ -697,6 +696,8 @@ struct btrfs_root *btrfs_read_fs_root(st + return fs_info->dev_root; + if (location->objectid == BTRFS_CSUM_TREE_OBJECTID) + return fs_info->csum_root; ++ if (location->objectid == BTRFS_QUOTA_TREE_OBJECTID) ++ return fs_info->csum_root; + + BUG_ON(location->objectid == BTRFS_TREE_RELOC_OBJECTID || + location->offset != (u64)-1); +@@ -723,6 +724,7 @@ void btrfs_free_fs_info(struct btrfs_fs_ + free(fs_info->chunk_root); + free(fs_info->dev_root); + free(fs_info->csum_root); ++ free(fs_info->quota_root); + free(fs_info->super_copy); + free(fs_info->log_root_tree); + free(fs_info); +@@ -743,11 +745,13 @@ struct btrfs_fs_info *btrfs_new_fs_info( + fs_info->chunk_root = malloc(sizeof(struct btrfs_root)); + fs_info->dev_root = malloc(sizeof(struct btrfs_root)); + fs_info->csum_root = malloc(sizeof(struct btrfs_root)); ++ fs_info->quota_root = malloc(sizeof(struct btrfs_root)); + fs_info->super_copy = malloc(BTRFS_SUPER_INFO_SIZE); + + if (!fs_info->tree_root || !fs_info->extent_root || + !fs_info->chunk_root || !fs_info->dev_root || +- !fs_info->csum_root || !fs_info->super_copy) ++ !fs_info->csum_root || !fs_info->quota_root || ++ !fs_info->super_copy) + goto free_all; + + memset(fs_info->super_copy, 0, BTRFS_SUPER_INFO_SIZE); +@@ -756,6 +760,7 @@ struct btrfs_fs_info *btrfs_new_fs_info( + memset(fs_info->chunk_root, 0, sizeof(struct btrfs_root)); + memset(fs_info->dev_root, 0, sizeof(struct btrfs_root)); + memset(fs_info->csum_root, 0, sizeof(struct btrfs_root)); ++ memset(fs_info->quota_root, 0, sizeof(struct btrfs_root)); + + extent_io_tree_init(&fs_info->extent_cache); + extent_io_tree_init(&fs_info->free_space_cache); +@@ -914,6 +919,11 @@ int btrfs_setup_all_roots(struct btrfs_f + } + fs_info->csum_root->track_dirty = 1; + ++ ret = find_and_setup_root(root, fs_info, BTRFS_QUOTA_TREE_OBJECTID, ++ fs_info->quota_root); ++ if (ret == 0) ++ fs_info->quota_enabled = 1; ++ + ret = find_and_setup_log_root(root, fs_info, sb); + if (ret) { + printk("Couldn't setup log root tree\n"); +@@ -938,6 +948,8 @@ int btrfs_setup_all_roots(struct btrfs_f + + void btrfs_release_all_roots(struct btrfs_fs_info *fs_info) + { ++ if (fs_info->quota_root) ++ free_extent_buffer(fs_info->quota_root->node); + if (fs_info->csum_root) + free_extent_buffer(fs_info->csum_root->node); + if (fs_info->dev_root) +Index: btrfs-progs-v3.14.1/print-tree.c +=================================================================== +--- btrfs-progs-v3.14.1.orig/print-tree.c ++++ btrfs-progs-v3.14.1/print-tree.c +@@ -242,7 +242,7 @@ static void print_file_extent_item(struc + btrfs_file_extent_compression(eb, fi)); + } + +-static void print_extent_item(struct extent_buffer *eb, int slot, int metadata) ++void print_extent_item(struct extent_buffer *eb, int slot, int metadata) + { + struct btrfs_extent_item *ei; + struct btrfs_extent_inline_ref *iref; +Index: btrfs-progs-v3.14.1/print-tree.h +=================================================================== +--- btrfs-progs-v3.14.1.orig/print-tree.h ++++ btrfs-progs-v3.14.1/print-tree.h +@@ -21,4 +21,5 @@ + void btrfs_print_leaf(struct btrfs_root *root, struct extent_buffer *l); + void btrfs_print_tree(struct btrfs_root *root, struct extent_buffer *t, int follow); + void btrfs_print_key(struct btrfs_disk_key *disk_key); ++void print_extent_item(struct extent_buffer *eb, int slot, int metadata); + #endif +Index: btrfs-progs-v3.14.1/qgroup-verify.c +=================================================================== +--- /dev/null ++++ btrfs-progs-v3.14.1/qgroup-verify.c +@@ -0,0 +1,1085 @@ ++/* ++ * Copyright (C) 2014 SUSE. All rights reserved. ++ * ++ * This program is free software; you can redistribute it and/or ++ * modify it under the terms of the GNU General Public ++ * License v2 as published by the Free Software Foundation. ++ * ++ * This program is distributed in the hope that it will be useful, ++ * but WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU ++ * General Public License for more details. ++ * ++ * You should have received a copy of the GNU General Public ++ * License along with this program; if not, write to the ++ * Free Software Foundation, Inc., 59 Temple Place - Suite 330, ++ * Boston, MA 021110-1307, USA. ++ * ++ * Authors: Mark Fasheh ++ */ ++ ++#include ++#include ++#include ++#include "kerncompat.h" ++#include "radix-tree.h" ++#include "ctree.h" ++#include "disk-io.h" ++#include "print-tree.h" ++#include "utils.h" ++#include "ulist.h" ++ ++#include "qgroup-verify.h" ++ ++/*#define QGROUP_VERIFY_DEBUG*/ ++static unsigned long tot_extents_scanned = 0; ++ ++static void add_bytes(u64 root_objectid, u64 num_bytes, int exclusive); ++ ++struct qgroup_count { ++ u64 qgroupid; ++ ++ struct btrfs_disk_key key; ++ struct btrfs_qgroup_info_item diskinfo; ++ ++ struct btrfs_qgroup_info_item info; ++ ++ struct rb_node rb_node; ++}; ++ ++struct counts_tree { ++ struct rb_root root; ++ unsigned int num_groups; ++} counts = { .root = RB_ROOT }; ++ ++struct rb_root by_bytenr = RB_ROOT; ++ ++/* ++ * List of interior tree blocks. We walk this list after loading the ++ * extent tree to resolve implied refs. For each interior node we'll ++ * place a shared ref in the ref tree against each child object. This ++ * allows the shared ref resolving code to do the actual work later of ++ * finding roots to account against. ++ * ++ * An implied ref is when a tree block has refs on it that may not ++ * exist in any of it's child nodes. Even though the refs might not ++ * exist further down the tree, the fact that our interior node has a ++ * ref means we need to account anything below it to all it's roots. ++ */ ++struct ulist *tree_blocks = NULL; /* unode->val = bytenr, ->aux ++ * = tree_block pointer */ ++struct tree_block { ++ int level; ++ u64 num_bytes; ++}; ++ ++struct ref { ++ u64 bytenr; ++ u64 num_bytes; ++ u64 parent; ++ u64 root; ++ ++ struct rb_node bytenr_node; ++}; ++ ++#ifdef QGROUP_VERIFY_DEBUG ++static void print_ref(struct ref *ref) ++{ ++ printf("bytenr: %llu\t\tnum_bytes: %llu\t\t parent: %llu\t\t" ++ "root: %llu\n", ref->bytenr, ref->num_bytes, ++ ref->parent, ref->root); ++} ++ ++static void print_all_refs(void) ++{ ++ unsigned long count = 0; ++ struct ref *ref; ++ struct rb_node *node; ++ ++ node = rb_first(&by_bytenr); ++ while (node) { ++ ref = rb_entry(node, struct ref, bytenr_node); ++ ++ print_ref(ref); ++ ++ count++; ++ node = rb_next(node); ++ } ++ ++ printf("%lu extents scanned with %lu refs in total.\n", ++ tot_extents_scanned, count); ++} ++#endif ++ ++/* ++ * Store by bytenr in rbtree ++ * ++ * The tree is sorted in ascending order by bytenr, then parent, then ++ * root. Since full refs have a parent == 0, those will come before ++ * shared refs. ++ */ ++static int compare_ref(struct ref *orig, u64 bytenr, u64 root, u64 parent) ++{ ++ if (bytenr < orig->bytenr) ++ return -1; ++ if (bytenr > orig->bytenr) ++ return 1; ++ ++ if (parent < orig->parent) ++ return -1; ++ if (parent > orig->parent) ++ return 1; ++ ++ if (root < orig->root) ++ return -1; ++ if (root > orig->root) ++ return 1; ++ ++ return 0; ++} ++ ++/* ++ * insert a new ref into the tree. returns the existing ref entry ++ * if one is already there. ++ */ ++static struct ref *insert_ref(struct ref *ref) ++{ ++ int ret; ++ struct rb_node **p = &by_bytenr.rb_node; ++ struct rb_node *parent = NULL; ++ struct ref *curr; ++ ++ while (*p) { ++ parent = *p; ++ curr = rb_entry(parent, struct ref, bytenr_node); ++ ++ ret = compare_ref(curr, ref->bytenr, ref->root, ref->parent); ++ if (ret < 0) ++ p = &(*p)->rb_left; ++ else if (ret > 0) ++ p = &(*p)->rb_right; ++ else ++ return curr; ++ } ++ ++ rb_link_node(&ref->bytenr_node, parent, p); ++ rb_insert_color(&ref->bytenr_node, &by_bytenr); ++ return ref; ++} ++ ++/* ++ * Partial search, returns the first ref with matching bytenr. Caller ++ * can walk forward from there. ++ * ++ * Leftmost refs will be full refs - this is used to our advantage ++ * when resolving roots. ++ */ ++static struct ref *find_ref_bytenr(u64 bytenr) ++{ ++ struct rb_node *n = by_bytenr.rb_node; ++ struct ref *ref; ++ ++ while (n) { ++ ref = rb_entry(n, struct ref, bytenr_node); ++ ++ if (bytenr < ref->bytenr) ++ n = n->rb_left; ++ else if (bytenr > ref->bytenr) ++ n = n->rb_right; ++ else { ++ /* Walk to the left to find the first item */ ++ struct rb_node *node_left = rb_prev(&ref->bytenr_node); ++ struct ref *ref_left; ++ ++ while (node_left) { ++ ref_left = rb_entry(node_left, struct ref, ++ bytenr_node); ++ if (ref_left->bytenr != ref->bytenr) ++ break; ++ ref = ref_left; ++ node_left = rb_prev(node_left); ++ } ++ return ref; ++ } ++ } ++ return NULL; ++} ++ ++static struct ref *find_ref(u64 bytenr, u64 root, u64 parent) ++{ ++ struct rb_node *n = by_bytenr.rb_node; ++ struct ref *ref; ++ int ret; ++ ++ while (n) { ++ ref = rb_entry(n, struct ref, bytenr_node); ++ ++ ret = compare_ref(ref, bytenr, root, parent); ++ if (ret < 0) ++ n = n->rb_left; ++ else if (ret > 0) ++ n = n->rb_right; ++ else ++ return ref; ++ } ++ return NULL; ++} ++ ++static struct ref *alloc_ref(u64 bytenr, u64 root, u64 parent, u64 num_bytes) ++{ ++ struct ref *ref = find_ref(bytenr, root, parent); ++ ++ BUG_ON(parent && root); ++ ++ if (ref == NULL) { ++ ref = calloc(1, sizeof(*ref)); ++ if (ref) { ++ ref->bytenr = bytenr; ++ ref->root = root; ++ ref->parent = parent; ++ ref->num_bytes = num_bytes; ++ ++ insert_ref(ref); ++ } ++ } ++ return ref; ++} ++ ++static void free_ref_node(struct rb_node *node) ++{ ++ struct ref *ref = rb_entry(node, struct ref, bytenr_node); ++ free(ref); ++} ++ ++FREE_RB_BASED_TREE(ref, free_ref_node); ++ ++/* ++ * Resolves all the possible roots for the ref at parent. ++ */ ++static void find_parent_roots(struct ulist *roots, u64 parent) ++{ ++ struct ref *ref; ++ struct rb_node *node; ++ ++ /* ++ * Search the rbtree for the first ref with bytenr == parent. ++ * Walk forward so long as bytenr == parent, adding resolved root ids. ++ * For each unresolved root, we recurse ++ */ ++ ref = find_ref_bytenr(parent); ++ node = &ref->bytenr_node; ++ BUG_ON(ref == NULL); ++ BUG_ON(ref->bytenr != parent); ++ ++ { ++ /* ++ * Random sanity check, are we actually getting the ++ * leftmost node? ++ */ ++ struct rb_node *prev_node = rb_prev(&ref->bytenr_node); ++ struct ref *prev; ++ if (prev_node) { ++ prev = rb_entry(prev_node, struct ref, bytenr_node); ++ BUG_ON(prev->bytenr == parent); ++ } ++ } ++ ++ do { ++ if (ref->root) ++ ulist_add(roots, ref->root, 0, 0); ++ else ++ find_parent_roots(roots, ref->parent); ++ ++ node = rb_next(node); ++ if (node) ++ ref = rb_entry(node, struct ref, bytenr_node); ++ } while (node && ref->bytenr == parent); ++} ++ ++/* ++ * Account each ref. Walk the refs, for each set of refs in a ++ * given bytenr: ++ * ++ * - add the roots for direct refs to the ref roots ulist ++ * ++ * - resolve all possible roots for shared refs, insert each ++ * of those into ref_roots ulist (this is a recursive process) ++ * ++ * - Walk ref_roots ulist, adding extent bytes to each qgroup count that ++ * cooresponds to a found root. ++ */ ++static void account_all_refs(void) ++{ ++ int exclusive; ++ struct ref *ref; ++ struct rb_node *node; ++ u64 bytenr, num_bytes; ++ struct ulist *roots = ulist_alloc(0); ++ struct ulist_iterator uiter; ++ struct ulist_node *unode; ++ ++ node = rb_first(&by_bytenr); ++ while (node) { ++ ulist_reinit(roots); ++ ++ ref = rb_entry(node, struct ref, bytenr_node); ++ /* ++ * Walk forward through the list of refs for this ++ * bytenr, adding roots to our ulist. If it's a full ++ * ref, then we have the easy case. Otherwise we need ++ * to search for roots. ++ */ ++ bytenr = ref->bytenr; ++ num_bytes = ref->num_bytes; ++ do { ++ BUG_ON(ref->bytenr != bytenr); ++ BUG_ON(ref->num_bytes != num_bytes); ++ if (ref->root) ++ ulist_add(roots, ref->root, 0, 0); ++ else ++ find_parent_roots(roots, ref->parent); ++ ++ /* ++ * When we leave this inner loop, node is set ++ * to next in our tree and will be turned into ++ * a ref object up top ++ */ ++ node = rb_next(node); ++ if (node) ++ ref = rb_entry(node, struct ref, bytenr_node); ++ } while (node && ref->bytenr == bytenr); ++ ++ /* ++ * Now that we have all roots, we can properly account ++ * this extent against the corresponding qgroups. ++ */ ++ if (roots->nnodes == 1) ++ exclusive = 1; ++ else ++ exclusive = 0; ++ ++ ULIST_ITER_INIT(&uiter); ++ while ((unode = ulist_next(roots, &uiter))) { ++ BUG_ON(unode->val == 0ULL); ++ /* We only want to account fs trees */ ++ if (is_fstree(unode->val)) ++ add_bytes(unode->val, num_bytes, exclusive); ++ } ++ } ++ ++ ulist_free(roots); ++} ++ ++static u64 resolve_one_root(u64 bytenr) ++{ ++ struct ref *ref = find_ref_bytenr(bytenr); ++ ++ BUG_ON(ref == NULL); ++ ++ if (ref->root) ++ return ref->root; ++ return resolve_one_root(ref->parent); ++} ++ ++static inline struct tree_block *unode_tree_block(struct ulist_node *unode) ++{ ++ return (struct tree_block *)unode->aux; ++} ++static inline u64 unode_bytenr(struct ulist_node *unode) ++{ ++ return unode->val; ++} ++ ++static int alloc_tree_block(u64 bytenr, u64 num_bytes, int level) ++{ ++ struct tree_block *block = calloc(1, sizeof(*block)); ++ ++ if (block) { ++ block->num_bytes = num_bytes; ++ block->level = level; ++ if (ulist_add(tree_blocks, bytenr, (unsigned long long)block, 0) >= 0) ++ return 0; ++ free(block); ++ } ++ return -ENOMEM; ++} ++ ++static void free_tree_blocks(void) ++{ ++ struct ulist_iterator uiter; ++ struct ulist_node *unode; ++ ++ if (!tree_blocks) ++ return; ++ ++ ULIST_ITER_INIT(&uiter); ++ while ((unode = ulist_next(tree_blocks, &uiter))) ++ free(unode_tree_block(unode)); ++ ulist_free(tree_blocks); ++ tree_blocks = NULL; ++} ++ ++#ifdef QGROUP_VERIFY_DEBUG ++static void print_tree_block(u64 bytenr, struct tree_block *block) ++{ ++ struct ref *ref; ++ struct rb_node *node; ++ ++ printf("tree block: %llu\t\tlevel: %d\n", (unsigned long long)bytenr, ++ block->level); ++ ++ ref = find_ref_bytenr(bytenr); ++ node = &ref->bytenr_node; ++ do { ++ print_ref(ref); ++ node = rb_next(node); ++ if (node) ++ ref = rb_entry(node, struct ref, bytenr_node); ++ } while (node && ref->bytenr == bytenr); ++ ++ printf("\n"); ++} ++ ++static void print_all_tree_blocks(void) ++{ ++ struct ulist_iterator uiter; ++ struct ulist_node *unode; ++ ++ if (!tree_blocks) ++ return; ++ ++ printf("Listing all found interior tree nodes:\n"); ++ ++ ULIST_ITER_INIT(&uiter); ++ while ((unode = ulist_next(tree_blocks, &uiter))) ++ print_tree_block(unode_bytenr(unode), unode_tree_block(unode)); ++} ++#endif ++ ++static int add_refs_for_leaf_items(struct extent_buffer *eb, u64 ref_parent) ++{ ++ int nr, i; ++ int extent_type; ++ u64 bytenr, num_bytes; ++ struct btrfs_key key; ++ struct btrfs_disk_key disk_key; ++ struct btrfs_file_extent_item *fi; ++ ++ nr = btrfs_header_nritems(eb); ++ for (i = 0; i < nr; i++) { ++ btrfs_item_key(eb, &disk_key, i); ++ btrfs_disk_key_to_cpu(&key, &disk_key); ++ ++ if (key.type != BTRFS_EXTENT_DATA_KEY) ++ continue; ++ ++ fi = btrfs_item_ptr(eb, i, struct btrfs_file_extent_item); ++ /* filter out: inline, disk_bytenr == 0, compressed? ++ * not if we can avoid it */ ++ extent_type = btrfs_file_extent_type(eb, fi); ++ ++ if (extent_type == BTRFS_FILE_EXTENT_INLINE) ++ continue; ++ ++ bytenr = btrfs_file_extent_disk_bytenr(eb, fi); ++ if (!bytenr) ++ continue; ++ ++ num_bytes = btrfs_file_extent_disk_num_bytes(eb, fi); ++ if (alloc_ref(bytenr, 0, ref_parent, num_bytes) == NULL) ++ return ENOMEM; ++ } ++ ++ return 0; ++} ++ ++static int travel_tree(struct btrfs_fs_info *info, struct btrfs_root *root, ++ u64 bytenr, u64 num_bytes, u64 ref_parent) ++{ ++ int ret, nr, i; ++ struct extent_buffer *eb; ++ u64 new_bytenr; ++ u64 new_num_bytes; ++ ++// printf("travel_tree: bytenr: %llu\tnum_bytes: %llu\tref_parent: %llu\n", ++// bytenr, num_bytes, ref_parent); ++ ++ eb = read_tree_block(root, bytenr, num_bytes, 0); ++ if (!eb) ++ return -EIO; ++ ++ ret = 0; ++ /* Don't add a ref for our starting tree block to itself */ ++ if (bytenr != ref_parent) { ++ if (alloc_ref(bytenr, 0, ref_parent, num_bytes) == NULL) ++ return ENOMEM; ++ } ++ ++ if (btrfs_is_leaf(eb)) { ++ ret = add_refs_for_leaf_items(eb, ref_parent); ++ goto out; ++ } ++ ++ /* ++ * Interior nodes are tuples of (key, bytenr) where key is the ++ * leftmost key in the tree block pointed to by bytenr. We ++ * don't have to care about key here, just follow the bytenr ++ * pointer. ++ */ ++ nr = btrfs_header_nritems(eb); ++ for (i = 0; i < nr; i++) { ++ new_bytenr = btrfs_node_blockptr(eb, i); ++ new_num_bytes = btrfs_level_size(root, ++ btrfs_header_level(eb) - 1); ++ ++ ret = travel_tree(info, root, new_bytenr, new_num_bytes, ++ ref_parent); ++ } ++ ++out: ++ free_extent_buffer(eb); ++ return ret; ++} ++ ++static int add_refs_for_implied(struct btrfs_fs_info *info, u64 bytenr, ++ struct tree_block *block) ++{ ++ int ret; ++ u64 root_bytenr = resolve_one_root(bytenr); ++ struct btrfs_root *root; ++ struct btrfs_key key; ++ ++ key.objectid = root_bytenr; ++ key.type = BTRFS_ROOT_ITEM_KEY; ++ key.offset = (u64)-1; ++ ++ /* ++ * XXX: Don't free the root object as we don't know whether it ++ * came off our fs_info struct or not. ++ */ ++ root = btrfs_read_fs_root(info, &key); ++ if (!root || IS_ERR(root)) ++ return ENOENT; ++ ++ ret = travel_tree(info, root, bytenr, block->num_bytes, bytenr); ++ if (ret) ++ return ret; ++ ++ return 0; ++} ++ ++/* ++ * Place shared refs in the ref tree for each child of an interior tree node. ++ */ ++static int map_implied_refs(struct btrfs_fs_info *info) ++{ ++ int ret = 0; ++ struct ulist_iterator uiter; ++ struct ulist_node *unode; ++ ++ ULIST_ITER_INIT(&uiter); ++ while ((unode = ulist_next(tree_blocks, &uiter))) { ++ ret = add_refs_for_implied(info, unode_bytenr(unode), ++ unode_tree_block(unode)); ++ if (ret) ++ goto out; ++ } ++out: ++ return ret; ++} ++ ++/* ++ * insert a new root into the tree. returns the existing root entry ++ * if one is already there. qgroupid is used ++ * as the key ++ */ ++static int insert_count(struct qgroup_count *qc) ++{ ++ struct rb_node **p = &counts.root.rb_node; ++ struct rb_node *parent = NULL; ++ struct qgroup_count *curr; ++ ++ while (*p) { ++ parent = *p; ++ curr = rb_entry(parent, struct qgroup_count, rb_node); ++ ++ if (qc->qgroupid < curr->qgroupid) ++ p = &(*p)->rb_left; ++ else if (qc->qgroupid > curr->qgroupid) ++ p = &(*p)->rb_right; ++ else ++ return EEXIST; ++ } ++ counts.num_groups++; ++ rb_link_node(&qc->rb_node, parent, p); ++ rb_insert_color(&qc->rb_node, &counts.root); ++ return 0; ++} ++ ++static struct qgroup_count *find_count(u64 qgroupid) ++{ ++ struct rb_node *n = counts.root.rb_node; ++ struct qgroup_count *count; ++ ++ while (n) { ++ count = rb_entry(n, struct qgroup_count, rb_node); ++ ++ if (qgroupid < count->qgroupid) ++ n = n->rb_left; ++ else if (qgroupid > count->qgroupid) ++ n = n->rb_right; ++ else ++ return count; ++ } ++ return NULL; ++} ++ ++static struct qgroup_count *alloc_count(struct btrfs_disk_key *key, ++ struct extent_buffer *leaf, ++ struct btrfs_qgroup_info_item *disk) ++{ ++ struct qgroup_count *c = calloc(1, sizeof(*c)); ++ struct btrfs_qgroup_info_item *item; ++ ++ if (c) { ++ c->qgroupid = btrfs_disk_key_offset(key); ++ c->key = *key; ++ ++ item = &c->diskinfo; ++ item->generation = btrfs_qgroup_info_generation(leaf, disk); ++ item->referenced = btrfs_qgroup_info_referenced(leaf, disk); ++ item->referenced_compressed = ++ btrfs_qgroup_info_referenced_compressed(leaf, disk); ++ item->exclusive = btrfs_qgroup_info_exclusive(leaf, disk); ++ item->exclusive_compressed = ++ btrfs_qgroup_info_exclusive_compressed(leaf, disk); ++ ++ if (insert_count(c)) { ++ free(c); ++ c = NULL; ++ } ++ } ++ return c; ++} ++ ++static void add_bytes(u64 root_objectid, u64 num_bytes, int exclusive) ++{ ++ struct qgroup_count *count = find_count(root_objectid); ++ struct btrfs_qgroup_info_item *qg; ++ ++ BUG_ON(num_bytes < 4096); /* Random sanity check. */ ++ ++ if (!count) ++ return; ++ ++ qg = &count->info; ++ ++ qg->referenced += num_bytes; ++ /* ++ * count of compressed bytes is unimplemented, so we do the ++ * same as kernel. ++ */ ++ qg->referenced_compressed += num_bytes; ++ ++ if (exclusive) { ++ qg->exclusive += num_bytes; ++ qg->exclusive_compressed += num_bytes; ++ } ++} ++ ++static int load_quota_info(struct btrfs_fs_info *info) ++{ ++ int ret; ++ struct btrfs_root *root = info->quota_root; ++ struct btrfs_path path; ++ struct btrfs_key key; ++ struct btrfs_disk_key disk_key; ++ struct extent_buffer *leaf; ++ struct btrfs_qgroup_info_item *item; ++ struct qgroup_count *count; ++ int i, nr; ++ ++ btrfs_init_path(&path); ++ ++ key.offset = 0; ++ key.objectid = 0; ++ key.type = 0; ++ ++ ret = btrfs_search_slot(NULL, root, &key, &path, 0, 0); ++ if (ret < 0) { ++ fprintf(stderr, "ERROR: Couldn't search slot: %d\n", ret); ++ goto out; ++ } ++ ++ while (1) { ++ leaf = path.nodes[0]; ++ ++ nr = btrfs_header_nritems(leaf); ++ for(i = 0; i < nr; i++) { ++ btrfs_item_key(leaf, &disk_key, i); ++ btrfs_disk_key_to_cpu(&key, &disk_key); ++ ++ if (key.type == BTRFS_QGROUP_RELATION_KEY) ++ printf("Ignoring qgroup relation key %llu\n", ++ key.objectid); ++ ++ /* ++ * Ignore: BTRFS_QGROUP_STATUS_KEY, ++ * BTRFS_QGROUP_LIMIT_KEY, BTRFS_QGROUP_RELATION_KEY ++ */ ++ if (key.type != BTRFS_QGROUP_INFO_KEY) ++ continue; ++ ++ item = btrfs_item_ptr(leaf, i, ++ struct btrfs_qgroup_info_item); ++ ++ count = alloc_count(&disk_key, leaf, item); ++ if (!count) { ++ ret = ENOMEM; ++ fprintf(stderr, "ERROR: out of memory\n"); ++ goto out; ++ } ++ } ++ ++ ret = btrfs_next_leaf(root, &path); ++ if (ret != 0) ++ break; ++ } ++ ++ ret = 0; ++ btrfs_release_path(&path); ++out: ++ return ret; ++} ++ ++static int add_inline_refs(struct btrfs_fs_info *info, ++ struct extent_buffer *ei_leaf, int slot, ++ u64 bytenr, u64 num_bytes, int meta_item) ++{ ++ struct btrfs_extent_item *ei; ++ struct btrfs_extent_inline_ref *iref; ++ struct btrfs_extent_data_ref *dref; ++ u64 flags, root_obj, offset, parent; ++ u32 item_size = btrfs_item_size_nr(ei_leaf, slot); ++ int type; ++ unsigned long end; ++ unsigned long ptr; ++ ++ ei = btrfs_item_ptr(ei_leaf, slot, struct btrfs_extent_item); ++ flags = btrfs_extent_flags(ei_leaf, ei); ++ ++ if (flags & BTRFS_EXTENT_FLAG_TREE_BLOCK && !meta_item) { ++ struct btrfs_tree_block_info *tbinfo; ++ tbinfo = (struct btrfs_tree_block_info *)(ei + 1); ++ iref = (struct btrfs_extent_inline_ref *)(tbinfo + 1); ++ } else { ++ iref = (struct btrfs_extent_inline_ref *)(ei + 1); ++ } ++ ++ ptr = (unsigned long)iref; ++ end = (unsigned long)ei + item_size; ++ while (ptr < end) { ++ iref = (struct btrfs_extent_inline_ref *)ptr; ++ ++ parent = root_obj = 0; ++ offset = btrfs_extent_inline_ref_offset(ei_leaf, iref); ++ type = btrfs_extent_inline_ref_type(ei_leaf, iref); ++ switch (type) { ++ case BTRFS_TREE_BLOCK_REF_KEY: ++ root_obj = offset; ++ break; ++ case BTRFS_EXTENT_DATA_REF_KEY: ++ dref = (struct btrfs_extent_data_ref *)(&iref->offset); ++ root_obj = btrfs_extent_data_ref_root(ei_leaf, dref); ++ break; ++ case BTRFS_SHARED_DATA_REF_KEY: ++ case BTRFS_SHARED_BLOCK_REF_KEY: ++ parent = offset; ++ break; ++ default: ++ return 1; ++ } ++ ++ if (alloc_ref(bytenr, root_obj, parent, num_bytes) == NULL) ++ return ENOMEM; ++ ++ ptr += btrfs_extent_inline_ref_size(type); ++ } ++ ++ return 0; ++} ++ ++static int add_keyed_ref(struct btrfs_fs_info *info, ++ struct btrfs_key *key, ++ struct extent_buffer *leaf, int slot, ++ u64 bytenr, u64 num_bytes) ++{ ++ u64 root_obj = 0, parent = 0; ++ struct btrfs_extent_data_ref *dref; ++ ++ switch(key->type) { ++ case BTRFS_TREE_BLOCK_REF_KEY: ++ root_obj = key->offset; ++ break; ++ case BTRFS_EXTENT_DATA_REF_KEY: ++ dref = btrfs_item_ptr(leaf, slot, struct btrfs_extent_data_ref); ++ root_obj = btrfs_extent_data_ref_root(leaf, dref); ++ break; ++ case BTRFS_SHARED_DATA_REF_KEY: ++ case BTRFS_SHARED_BLOCK_REF_KEY: ++ parent = key->offset; ++ break; ++ default: ++ return 1; ++ } ++ ++ if (alloc_ref(bytenr, root_obj, parent, num_bytes) == NULL) ++ return ENOMEM; ++ ++ return 0; ++} ++ ++/* ++ * return value of 0 indicates leaf or not meta data. The code that ++ * calls this does not need to make a distinction between the two as ++ * it is only concerned with intermediate blocks which will always ++ * have level > 0. ++ */ ++static int get_tree_block_level(struct btrfs_key *key, ++ struct extent_buffer *ei_leaf, ++ int slot) ++{ ++ int level = 0; ++ int meta_key = key->type == BTRFS_METADATA_ITEM_KEY; ++ u64 flags; ++ struct btrfs_extent_item *ei; ++ ++ ei = btrfs_item_ptr(ei_leaf, slot, struct btrfs_extent_item); ++ flags = btrfs_extent_flags(ei_leaf, ei); ++ ++ if (flags & BTRFS_EXTENT_FLAG_TREE_BLOCK && !meta_key) { ++ struct btrfs_tree_block_info *tbinfo; ++ tbinfo = (struct btrfs_tree_block_info *)(ei + 1); ++ level = btrfs_tree_block_level(ei_leaf, tbinfo); ++ } else if (meta_key) { ++ /* skinny metadata */ ++ level = (int)key->offset; ++ } ++ return level; ++} ++ ++/* ++ * Walk the extent tree, allocating a ref item for every ref and ++ * storing it in the bytenr tree. ++ */ ++static int scan_extents(struct btrfs_fs_info *info, ++ u64 start, u64 end) ++{ ++ int ret, i, nr, level; ++ struct btrfs_root *root = info->extent_root; ++ struct btrfs_key key; ++ struct btrfs_path path; ++ struct btrfs_disk_key disk_key; ++ struct extent_buffer *leaf; ++ u64 bytenr = 0, num_bytes = 0; ++ ++ btrfs_init_path(&path); ++ ++ key.objectid = start; ++ key.type = 0; ++ key.offset = 0; ++ ++ ret = btrfs_search_slot(NULL, root, &key, &path, 0, 0); ++ if (ret < 0) { ++ fprintf(stderr, "ERROR: Couldn't search slot: %d\n", ret); ++ goto out; ++ } ++ path.reada = 1; ++ ++ while (1) { ++ leaf = path.nodes[0]; ++ ++ nr = btrfs_header_nritems(leaf); ++ for(i = 0; i < nr; i++) { ++ btrfs_item_key(leaf, &disk_key, i); ++ btrfs_disk_key_to_cpu(&key, &disk_key); ++ ++ if (key.objectid < start) ++ continue; ++ ++ if (key.objectid > end) ++ goto done; ++ ++ if (key.type == BTRFS_EXTENT_ITEM_KEY || ++ key.type == BTRFS_METADATA_ITEM_KEY) { ++ int meta = 0; ++ ++ tot_extents_scanned++; ++ ++ bytenr = key.objectid; ++ num_bytes = key.offset; ++ if (key.type == BTRFS_METADATA_ITEM_KEY) { ++ num_bytes = info->extent_root->leafsize; ++ meta = 1; ++ } ++ ++ ret = add_inline_refs(info, leaf, i, bytenr, ++ num_bytes, meta); ++ if (ret) ++ goto out; ++ ++ level = get_tree_block_level(&key, leaf, i); ++ if (level) { ++ if (alloc_tree_block(bytenr, num_bytes, ++ level)) ++ return ENOMEM; ++ } ++ ++ continue; ++ } ++ ++ if (key.type > BTRFS_SHARED_DATA_REF_KEY) ++ continue; ++ if (key.type < BTRFS_TREE_BLOCK_REF_KEY) ++ continue; ++ ++ /* ++ * Keyed refs should come after their extent ++ * item in the tree. As a result, the value of ++ * bytenr and num_bytes should be unchanged ++ * from the above block that catches the ++ * original extent item. ++ */ ++ BUG_ON(key.objectid != bytenr); ++ ++ ret = add_keyed_ref(info, &key, leaf, i, bytenr, ++ num_bytes); ++ if (ret) ++ goto out; ++ } ++ ++ ret = btrfs_next_leaf(root, &path); ++ if (ret != 0) { ++ if (ret < 0) { ++ fprintf(stderr, ++ "ERROR: Next leaf failed: %d\n", ret); ++ goto out; ++ } ++ break; ++ } ++ } ++done: ++ ret = 0; ++out: ++ btrfs_release_path(&path); ++ ++ return ret; ++} ++ ++static void print_fields(u64 bytes, u64 bytes_compressed, char *prefix, ++ char *type) ++{ ++ printf("%s\t\t%s %llu %s compressed %llu\n", ++ prefix, type, (unsigned long long)bytes, type, ++ (unsigned long long)bytes_compressed); ++} ++ ++static void print_fields_signed(long long bytes, ++ long long bytes_compressed, ++ char *prefix, char *type) ++{ ++ printf("%s\t\t%s %lld %s compressed %lld\n", ++ prefix, type, bytes, type, bytes_compressed); ++} ++ ++static void print_qgroup_difference(struct qgroup_count *count, int verbose) ++{ ++ int is_different; ++ struct btrfs_qgroup_info_item *info = &count->info; ++ struct btrfs_qgroup_info_item *disk = &count->diskinfo; ++ long long excl_diff = info->exclusive - disk->exclusive; ++ long long ref_diff = info->referenced - disk->referenced; ++ ++ is_different = excl_diff || ref_diff; ++ ++ if (verbose || is_different) { ++ printf("Counts for qgroup id: %llu %s\n", ++ (unsigned long long)count->qgroupid, ++ is_different ? "are different" : ""); ++ ++ print_fields(info->referenced, info->referenced_compressed, ++ "our:", "referenced"); ++ print_fields(disk->referenced, disk->referenced_compressed, ++ "disk:", "referenced"); ++ if (ref_diff) ++ print_fields_signed(ref_diff, ref_diff, ++ "diff:", "referenced"); ++ print_fields(info->exclusive, info->exclusive_compressed, ++ "our:", "exclusive"); ++ print_fields(disk->exclusive, disk->exclusive_compressed, ++ "disk:", "exclusive"); ++ if (excl_diff) ++ print_fields_signed(excl_diff, excl_diff, ++ "diff:", "exclusive"); ++ } ++} ++ ++void print_qgroup_report(int all) ++{ ++ struct rb_node *node; ++ struct qgroup_count *c; ++ ++ node = rb_first(&counts.root); ++ while (node) { ++ c = rb_entry(node, struct qgroup_count, rb_node); ++ print_qgroup_difference(c, all); ++ node = rb_next(node); ++ } ++} ++ ++int qgroup_verify_all(struct btrfs_fs_info *info) ++{ ++ int ret; ++ ++ if (!info->quota_enabled) ++ return 0; ++ ++ tree_blocks = ulist_alloc(0); ++ if (!tree_blocks) { ++ fprintf(stderr, ++ "ERROR: Out of memory while allocating ulist.\n"); ++ return ENOMEM; ++ } ++ ++ ret = load_quota_info(info); ++ if (ret) { ++ fprintf(stderr, "ERROR: Loading qgroups from disk: %d\n", ret); ++ goto out; ++ } ++ ++ /* ++ * Put all extent refs into our rbtree ++ */ ++ ret = scan_extents(info, 0, ~0ULL); ++ if (ret) { ++ fprintf(stderr, "ERROR: while scanning extent tree: %d\n", ret); ++ goto out; ++ } ++ ++ ret = map_implied_refs(info); ++ if (ret) { ++ fprintf(stderr, "ERROR: while mapping refs: %d\n", ret); ++ goto out; ++ } ++ ++ account_all_refs(); ++ ++out: ++ /* ++ * Don't free the qgroup count records as they will be walked ++ * later via the print function. ++ */ ++ free_tree_blocks(); ++ free_ref_tree(&by_bytenr); ++ return ret; ++} +Index: btrfs-progs-v3.14.1/qgroup-verify.h +=================================================================== +--- /dev/null ++++ btrfs-progs-v3.14.1/qgroup-verify.h +@@ -0,0 +1,25 @@ ++/* ++ * Copyright (C) 2014 SUSE. All rights reserved. ++ * ++ * This program is free software; you can redistribute it and/or ++ * modify it under the terms of the GNU General Public ++ * License v2 as published by the Free Software Foundation. ++ * ++ * This program is distributed in the hope that it will be useful, ++ * but WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU ++ * General Public License for more details. ++ * ++ * You should have received a copy of the GNU General Public ++ * License along with this program; if not, write to the ++ * Free Software Foundation, Inc., 59 Temple Place - Suite 330, ++ * Boston, MA 021110-1307, USA. ++ */ ++ ++#ifndef _BTRFS_QGROUP_VERIFY_H ++#define _BTRFS_QGROUP_VERIFY_H ++ ++int qgroup_verify_all(struct btrfs_fs_info *info); ++void print_qgroup_report(int all); ++ ++#endif /* _BTRFS_QGROUP_VERIFY_H */ diff --git a/btrfsprogs.changes b/btrfsprogs.changes index 4fface7..99087f6 100644 --- a/btrfsprogs.changes +++ b/btrfsprogs.changes @@ -1,3 +1,23 @@ +------------------------------------------------------------------- +Wed May 21 19:32:13 UTC 2014 - mfasheh@suse.com + +- add quota group verify patches, sent to list near end of the week starting + 5/4/2014 + - updates btrfsck with the ability to verify quota groups +- Update the changelog with some sad version of diffstat output to meet + requirements. +- Added patches: + * 0200-btrfs-progs-print-qgroup-excl-as-unsigned.patch + * 0201-btrfs-progs-import-ulist.patch + * 0202-btrfs-progs-add-quota-group-verify-code.patch + +------------------------------------------------------------------- +Tue May 13 20:28:23 UTC 2014 - mfasheh@suse.com + +- add quota group verify patches, sent to list near end of the week starting + 5/4/2014 + - updates btrfsck with the ability to verify quota groups + ------------------------------------------------------------------- Fri May 2 13:37:04 UTC 2014 - dsterba@suse.cz diff --git a/btrfsprogs.spec b/btrfsprogs.spec index 968afaf..ab9dff4 100644 --- a/btrfsprogs.spec +++ b/btrfsprogs.spec @@ -69,6 +69,11 @@ Patch167: 0167-Btrfs-progs-make-find_and_setup_root-return-an-error.patch Patch168: 0168-Btrfs-progs-don-t-bug-out-if-we-can-t-find-the-last-.patch Patch169: 0169-btrfs-progs-Check-metadata-mirrors-in-find-root.patch Patch170: 0170-btrfs-progs-In-find-root-dump-bytenr-for-every-slot.patch + +Patch200: 0200-btrfs-progs-print-qgroup-excl-as-unsigned.patch +Patch201: 0201-btrfs-progs-import-ulist.patch +Patch202: 0202-btrfs-progs-add-quota-group-verify-code.patch + Patch1000: local-version-override.patch BuildRequires: libacl-devel BuildRequires: libattr-devel @@ -143,6 +148,9 @@ build applications to interface with btrfs. %patch168 -p1 %patch169 -p1 %patch170 -p1 +%patch200 -p1 +%patch201 -p1 +%patch202 -p1 %patch1000 -p1 %build