mirror of
				https://github.com/torvalds/linux.git
				synced 2025-10-31 00:28:52 +02:00 
			
		
		
		
	 47bf225a8d
			
		
	
	
		47bf225a8d
		
	
	
	
	
		
			
			At btrfs_del_root_ref(), if btrfs_search_slot() returns an error, we end
up returning from the function with a value of 0 (success). This happens
because the function returns the value stored in the variable 'err',
which is 0, while the error value we got from btrfs_search_slot() is
stored in the 'ret' variable.
So fix it by setting 'err' with the error value.
Fixes: 8289ed9f93 ("btrfs: replace the BUG_ON in btrfs_del_root_ref with proper error handling")
CC: stable@vger.kernel.org # 5.16+
Reviewed-by: Qu Wenruo <wqu@suse.com>
Signed-off-by: Filipe Manana <fdmanana@suse.com>
Reviewed-by: David Sterba <dsterba@suse.com>
Signed-off-by: David Sterba <dsterba@suse.com>
		
	
			
		
			
				
	
	
		
			547 lines
		
	
	
	
		
			14 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			547 lines
		
	
	
	
		
			14 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| // SPDX-License-Identifier: GPL-2.0
 | |
| /*
 | |
|  * Copyright (C) 2007 Oracle.  All rights reserved.
 | |
|  */
 | |
| 
 | |
| #include <linux/err.h>
 | |
| #include <linux/uuid.h>
 | |
| #include "ctree.h"
 | |
| #include "transaction.h"
 | |
| #include "disk-io.h"
 | |
| #include "print-tree.h"
 | |
| #include "qgroup.h"
 | |
| #include "space-info.h"
 | |
| 
 | |
| /*
 | |
|  * Read a root item from the tree. In case we detect a root item smaller then
 | |
|  * sizeof(root_item), we know it's an old version of the root structure and
 | |
|  * initialize all new fields to zero. The same happens if we detect mismatching
 | |
|  * generation numbers as then we know the root was once mounted with an older
 | |
|  * kernel that was not aware of the root item structure change.
 | |
|  */
 | |
| static void btrfs_read_root_item(struct extent_buffer *eb, int slot,
 | |
| 				struct btrfs_root_item *item)
 | |
| {
 | |
| 	u32 len;
 | |
| 	int need_reset = 0;
 | |
| 
 | |
| 	len = btrfs_item_size(eb, slot);
 | |
| 	read_extent_buffer(eb, item, btrfs_item_ptr_offset(eb, slot),
 | |
| 			   min_t(u32, len, sizeof(*item)));
 | |
| 	if (len < sizeof(*item))
 | |
| 		need_reset = 1;
 | |
| 	if (!need_reset && btrfs_root_generation(item)
 | |
| 		!= btrfs_root_generation_v2(item)) {
 | |
| 		if (btrfs_root_generation_v2(item) != 0) {
 | |
| 			btrfs_warn(eb->fs_info,
 | |
| 					"mismatching generation and generation_v2 found in root item. This root was probably mounted with an older kernel. Resetting all new fields.");
 | |
| 		}
 | |
| 		need_reset = 1;
 | |
| 	}
 | |
| 	if (need_reset) {
 | |
| 		/* Clear all members from generation_v2 onwards. */
 | |
| 		memset_startat(item, 0, generation_v2);
 | |
| 		generate_random_guid(item->uuid);
 | |
| 	}
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * btrfs_find_root - lookup the root by the key.
 | |
|  * root: the root of the root tree
 | |
|  * search_key: the key to search
 | |
|  * path: the path we search
 | |
|  * root_item: the root item of the tree we look for
 | |
|  * root_key: the root key of the tree we look for
 | |
|  *
 | |
|  * If ->offset of 'search_key' is -1ULL, it means we are not sure the offset
 | |
|  * of the search key, just lookup the root with the highest offset for a
 | |
|  * given objectid.
 | |
|  *
 | |
|  * If we find something return 0, otherwise > 0, < 0 on error.
 | |
|  */
 | |
| int btrfs_find_root(struct btrfs_root *root, const struct btrfs_key *search_key,
 | |
| 		    struct btrfs_path *path, struct btrfs_root_item *root_item,
 | |
| 		    struct btrfs_key *root_key)
 | |
| {
 | |
| 	struct btrfs_key found_key;
 | |
| 	struct extent_buffer *l;
 | |
| 	int ret;
 | |
| 	int slot;
 | |
| 
 | |
| 	ret = btrfs_search_slot(NULL, root, search_key, path, 0, 0);
 | |
| 	if (ret < 0)
 | |
| 		return ret;
 | |
| 
 | |
| 	if (search_key->offset != -1ULL) {	/* the search key is exact */
 | |
| 		if (ret > 0)
 | |
| 			goto out;
 | |
| 	} else {
 | |
| 		BUG_ON(ret == 0);		/* Logical error */
 | |
| 		if (path->slots[0] == 0)
 | |
| 			goto out;
 | |
| 		path->slots[0]--;
 | |
| 		ret = 0;
 | |
| 	}
 | |
| 
 | |
| 	l = path->nodes[0];
 | |
| 	slot = path->slots[0];
 | |
| 
 | |
| 	btrfs_item_key_to_cpu(l, &found_key, slot);
 | |
| 	if (found_key.objectid != search_key->objectid ||
 | |
| 	    found_key.type != BTRFS_ROOT_ITEM_KEY) {
 | |
| 		ret = 1;
 | |
| 		goto out;
 | |
| 	}
 | |
| 
 | |
| 	if (root_item)
 | |
| 		btrfs_read_root_item(l, slot, root_item);
 | |
| 	if (root_key)
 | |
| 		memcpy(root_key, &found_key, sizeof(found_key));
 | |
| out:
 | |
| 	btrfs_release_path(path);
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| void btrfs_set_root_node(struct btrfs_root_item *item,
 | |
| 			 struct extent_buffer *node)
 | |
| {
 | |
| 	btrfs_set_root_bytenr(item, node->start);
 | |
| 	btrfs_set_root_level(item, btrfs_header_level(node));
 | |
| 	btrfs_set_root_generation(item, btrfs_header_generation(node));
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * copy the data in 'item' into the btree
 | |
|  */
 | |
| int btrfs_update_root(struct btrfs_trans_handle *trans, struct btrfs_root
 | |
| 		      *root, struct btrfs_key *key, struct btrfs_root_item
 | |
| 		      *item)
 | |
| {
 | |
| 	struct btrfs_fs_info *fs_info = root->fs_info;
 | |
| 	struct btrfs_path *path;
 | |
| 	struct extent_buffer *l;
 | |
| 	int ret;
 | |
| 	int slot;
 | |
| 	unsigned long ptr;
 | |
| 	u32 old_len;
 | |
| 
 | |
| 	path = btrfs_alloc_path();
 | |
| 	if (!path)
 | |
| 		return -ENOMEM;
 | |
| 
 | |
| 	ret = btrfs_search_slot(trans, root, key, path, 0, 1);
 | |
| 	if (ret < 0)
 | |
| 		goto out;
 | |
| 
 | |
| 	if (ret > 0) {
 | |
| 		btrfs_crit(fs_info,
 | |
| 			"unable to find root key (%llu %u %llu) in tree %llu",
 | |
| 			key->objectid, key->type, key->offset,
 | |
| 			root->root_key.objectid);
 | |
| 		ret = -EUCLEAN;
 | |
| 		btrfs_abort_transaction(trans, ret);
 | |
| 		goto out;
 | |
| 	}
 | |
| 
 | |
| 	l = path->nodes[0];
 | |
| 	slot = path->slots[0];
 | |
| 	ptr = btrfs_item_ptr_offset(l, slot);
 | |
| 	old_len = btrfs_item_size(l, slot);
 | |
| 
 | |
| 	/*
 | |
| 	 * If this is the first time we update the root item which originated
 | |
| 	 * from an older kernel, we need to enlarge the item size to make room
 | |
| 	 * for the added fields.
 | |
| 	 */
 | |
| 	if (old_len < sizeof(*item)) {
 | |
| 		btrfs_release_path(path);
 | |
| 		ret = btrfs_search_slot(trans, root, key, path,
 | |
| 				-1, 1);
 | |
| 		if (ret < 0) {
 | |
| 			btrfs_abort_transaction(trans, ret);
 | |
| 			goto out;
 | |
| 		}
 | |
| 
 | |
| 		ret = btrfs_del_item(trans, root, path);
 | |
| 		if (ret < 0) {
 | |
| 			btrfs_abort_transaction(trans, ret);
 | |
| 			goto out;
 | |
| 		}
 | |
| 		btrfs_release_path(path);
 | |
| 		ret = btrfs_insert_empty_item(trans, root, path,
 | |
| 				key, sizeof(*item));
 | |
| 		if (ret < 0) {
 | |
| 			btrfs_abort_transaction(trans, ret);
 | |
| 			goto out;
 | |
| 		}
 | |
| 		l = path->nodes[0];
 | |
| 		slot = path->slots[0];
 | |
| 		ptr = btrfs_item_ptr_offset(l, slot);
 | |
| 	}
 | |
| 
 | |
| 	/*
 | |
| 	 * Update generation_v2 so at the next mount we know the new root
 | |
| 	 * fields are valid.
 | |
| 	 */
 | |
| 	btrfs_set_root_generation_v2(item, btrfs_root_generation(item));
 | |
| 
 | |
| 	write_extent_buffer(l, item, ptr, sizeof(*item));
 | |
| 	btrfs_mark_buffer_dirty(path->nodes[0]);
 | |
| out:
 | |
| 	btrfs_free_path(path);
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| int btrfs_insert_root(struct btrfs_trans_handle *trans, struct btrfs_root *root,
 | |
| 		      const struct btrfs_key *key, struct btrfs_root_item *item)
 | |
| {
 | |
| 	/*
 | |
| 	 * Make sure generation v1 and v2 match. See update_root for details.
 | |
| 	 */
 | |
| 	btrfs_set_root_generation_v2(item, btrfs_root_generation(item));
 | |
| 	return btrfs_insert_item(trans, root, key, item, sizeof(*item));
 | |
| }
 | |
| 
 | |
| int btrfs_find_orphan_roots(struct btrfs_fs_info *fs_info)
 | |
| {
 | |
| 	struct btrfs_root *tree_root = fs_info->tree_root;
 | |
| 	struct extent_buffer *leaf;
 | |
| 	struct btrfs_path *path;
 | |
| 	struct btrfs_key key;
 | |
| 	struct btrfs_root *root;
 | |
| 	int err = 0;
 | |
| 	int ret;
 | |
| 
 | |
| 	path = btrfs_alloc_path();
 | |
| 	if (!path)
 | |
| 		return -ENOMEM;
 | |
| 
 | |
| 	key.objectid = BTRFS_ORPHAN_OBJECTID;
 | |
| 	key.type = BTRFS_ORPHAN_ITEM_KEY;
 | |
| 	key.offset = 0;
 | |
| 
 | |
| 	while (1) {
 | |
| 		u64 root_objectid;
 | |
| 
 | |
| 		ret = btrfs_search_slot(NULL, tree_root, &key, path, 0, 0);
 | |
| 		if (ret < 0) {
 | |
| 			err = ret;
 | |
| 			break;
 | |
| 		}
 | |
| 
 | |
| 		leaf = path->nodes[0];
 | |
| 		if (path->slots[0] >= btrfs_header_nritems(leaf)) {
 | |
| 			ret = btrfs_next_leaf(tree_root, path);
 | |
| 			if (ret < 0)
 | |
| 				err = ret;
 | |
| 			if (ret != 0)
 | |
| 				break;
 | |
| 			leaf = path->nodes[0];
 | |
| 		}
 | |
| 
 | |
| 		btrfs_item_key_to_cpu(leaf, &key, path->slots[0]);
 | |
| 		btrfs_release_path(path);
 | |
| 
 | |
| 		if (key.objectid != BTRFS_ORPHAN_OBJECTID ||
 | |
| 		    key.type != BTRFS_ORPHAN_ITEM_KEY)
 | |
| 			break;
 | |
| 
 | |
| 		root_objectid = key.offset;
 | |
| 		key.offset++;
 | |
| 
 | |
| 		root = btrfs_get_fs_root(fs_info, root_objectid, false);
 | |
| 		err = PTR_ERR_OR_ZERO(root);
 | |
| 		if (err && err != -ENOENT) {
 | |
| 			break;
 | |
| 		} else if (err == -ENOENT) {
 | |
| 			struct btrfs_trans_handle *trans;
 | |
| 
 | |
| 			btrfs_release_path(path);
 | |
| 
 | |
| 			trans = btrfs_join_transaction(tree_root);
 | |
| 			if (IS_ERR(trans)) {
 | |
| 				err = PTR_ERR(trans);
 | |
| 				btrfs_handle_fs_error(fs_info, err,
 | |
| 					    "Failed to start trans to delete orphan item");
 | |
| 				break;
 | |
| 			}
 | |
| 			err = btrfs_del_orphan_item(trans, tree_root,
 | |
| 						    root_objectid);
 | |
| 			btrfs_end_transaction(trans);
 | |
| 			if (err) {
 | |
| 				btrfs_handle_fs_error(fs_info, err,
 | |
| 					    "Failed to delete root orphan item");
 | |
| 				break;
 | |
| 			}
 | |
| 			continue;
 | |
| 		}
 | |
| 
 | |
| 		WARN_ON(!test_bit(BTRFS_ROOT_ORPHAN_ITEM_INSERTED, &root->state));
 | |
| 		if (btrfs_root_refs(&root->root_item) == 0) {
 | |
| 			struct btrfs_key drop_key;
 | |
| 
 | |
| 			btrfs_disk_key_to_cpu(&drop_key, &root->root_item.drop_progress);
 | |
| 			/*
 | |
| 			 * If we have a non-zero drop_progress then we know we
 | |
| 			 * made it partly through deleting this snapshot, and
 | |
| 			 * thus we need to make sure we block any balance from
 | |
| 			 * happening until this snapshot is completely dropped.
 | |
| 			 */
 | |
| 			if (drop_key.objectid != 0 || drop_key.type != 0 ||
 | |
| 			    drop_key.offset != 0) {
 | |
| 				set_bit(BTRFS_FS_UNFINISHED_DROPS, &fs_info->flags);
 | |
| 				set_bit(BTRFS_ROOT_UNFINISHED_DROP, &root->state);
 | |
| 			}
 | |
| 
 | |
| 			set_bit(BTRFS_ROOT_DEAD_TREE, &root->state);
 | |
| 			btrfs_add_dead_root(root);
 | |
| 		}
 | |
| 		btrfs_put_root(root);
 | |
| 	}
 | |
| 
 | |
| 	btrfs_free_path(path);
 | |
| 	return err;
 | |
| }
 | |
| 
 | |
| /* drop the root item for 'key' from the tree root */
 | |
| int btrfs_del_root(struct btrfs_trans_handle *trans,
 | |
| 		   const struct btrfs_key *key)
 | |
| {
 | |
| 	struct btrfs_root *root = trans->fs_info->tree_root;
 | |
| 	struct btrfs_path *path;
 | |
| 	int ret;
 | |
| 
 | |
| 	path = btrfs_alloc_path();
 | |
| 	if (!path)
 | |
| 		return -ENOMEM;
 | |
| 	ret = btrfs_search_slot(trans, root, key, path, -1, 1);
 | |
| 	if (ret < 0)
 | |
| 		goto out;
 | |
| 
 | |
| 	BUG_ON(ret != 0);
 | |
| 
 | |
| 	ret = btrfs_del_item(trans, root, path);
 | |
| out:
 | |
| 	btrfs_free_path(path);
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| int btrfs_del_root_ref(struct btrfs_trans_handle *trans, u64 root_id,
 | |
| 		       u64 ref_id, u64 dirid, u64 *sequence, const char *name,
 | |
| 		       int name_len)
 | |
| 
 | |
| {
 | |
| 	struct btrfs_root *tree_root = trans->fs_info->tree_root;
 | |
| 	struct btrfs_path *path;
 | |
| 	struct btrfs_root_ref *ref;
 | |
| 	struct extent_buffer *leaf;
 | |
| 	struct btrfs_key key;
 | |
| 	unsigned long ptr;
 | |
| 	int err = 0;
 | |
| 	int ret;
 | |
| 
 | |
| 	path = btrfs_alloc_path();
 | |
| 	if (!path)
 | |
| 		return -ENOMEM;
 | |
| 
 | |
| 	key.objectid = root_id;
 | |
| 	key.type = BTRFS_ROOT_BACKREF_KEY;
 | |
| 	key.offset = ref_id;
 | |
| again:
 | |
| 	ret = btrfs_search_slot(trans, tree_root, &key, path, -1, 1);
 | |
| 	if (ret < 0) {
 | |
| 		err = ret;
 | |
| 		goto out;
 | |
| 	} else if (ret == 0) {
 | |
| 		leaf = path->nodes[0];
 | |
| 		ref = btrfs_item_ptr(leaf, path->slots[0],
 | |
| 				     struct btrfs_root_ref);
 | |
| 		ptr = (unsigned long)(ref + 1);
 | |
| 		if ((btrfs_root_ref_dirid(leaf, ref) != dirid) ||
 | |
| 		    (btrfs_root_ref_name_len(leaf, ref) != name_len) ||
 | |
| 		    memcmp_extent_buffer(leaf, name, ptr, name_len)) {
 | |
| 			err = -ENOENT;
 | |
| 			goto out;
 | |
| 		}
 | |
| 		*sequence = btrfs_root_ref_sequence(leaf, ref);
 | |
| 
 | |
| 		ret = btrfs_del_item(trans, tree_root, path);
 | |
| 		if (ret) {
 | |
| 			err = ret;
 | |
| 			goto out;
 | |
| 		}
 | |
| 	} else
 | |
| 		err = -ENOENT;
 | |
| 
 | |
| 	if (key.type == BTRFS_ROOT_BACKREF_KEY) {
 | |
| 		btrfs_release_path(path);
 | |
| 		key.objectid = ref_id;
 | |
| 		key.type = BTRFS_ROOT_REF_KEY;
 | |
| 		key.offset = root_id;
 | |
| 		goto again;
 | |
| 	}
 | |
| 
 | |
| out:
 | |
| 	btrfs_free_path(path);
 | |
| 	return err;
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * add a btrfs_root_ref item.  type is either BTRFS_ROOT_REF_KEY
 | |
|  * or BTRFS_ROOT_BACKREF_KEY.
 | |
|  *
 | |
|  * The dirid, sequence, name and name_len refer to the directory entry
 | |
|  * that is referencing the root.
 | |
|  *
 | |
|  * For a forward ref, the root_id is the id of the tree referencing
 | |
|  * the root and ref_id is the id of the subvol  or snapshot.
 | |
|  *
 | |
|  * For a back ref the root_id is the id of the subvol or snapshot and
 | |
|  * ref_id is the id of the tree referencing it.
 | |
|  *
 | |
|  * Will return 0, -ENOMEM, or anything from the CoW path
 | |
|  */
 | |
| int btrfs_add_root_ref(struct btrfs_trans_handle *trans, u64 root_id,
 | |
| 		       u64 ref_id, u64 dirid, u64 sequence, const char *name,
 | |
| 		       int name_len)
 | |
| {
 | |
| 	struct btrfs_root *tree_root = trans->fs_info->tree_root;
 | |
| 	struct btrfs_key key;
 | |
| 	int ret;
 | |
| 	struct btrfs_path *path;
 | |
| 	struct btrfs_root_ref *ref;
 | |
| 	struct extent_buffer *leaf;
 | |
| 	unsigned long ptr;
 | |
| 
 | |
| 	path = btrfs_alloc_path();
 | |
| 	if (!path)
 | |
| 		return -ENOMEM;
 | |
| 
 | |
| 	key.objectid = root_id;
 | |
| 	key.type = BTRFS_ROOT_BACKREF_KEY;
 | |
| 	key.offset = ref_id;
 | |
| again:
 | |
| 	ret = btrfs_insert_empty_item(trans, tree_root, path, &key,
 | |
| 				      sizeof(*ref) + name_len);
 | |
| 	if (ret) {
 | |
| 		btrfs_abort_transaction(trans, ret);
 | |
| 		btrfs_free_path(path);
 | |
| 		return ret;
 | |
| 	}
 | |
| 
 | |
| 	leaf = path->nodes[0];
 | |
| 	ref = btrfs_item_ptr(leaf, path->slots[0], struct btrfs_root_ref);
 | |
| 	btrfs_set_root_ref_dirid(leaf, ref, dirid);
 | |
| 	btrfs_set_root_ref_sequence(leaf, ref, sequence);
 | |
| 	btrfs_set_root_ref_name_len(leaf, ref, name_len);
 | |
| 	ptr = (unsigned long)(ref + 1);
 | |
| 	write_extent_buffer(leaf, name, ptr, name_len);
 | |
| 	btrfs_mark_buffer_dirty(leaf);
 | |
| 
 | |
| 	if (key.type == BTRFS_ROOT_BACKREF_KEY) {
 | |
| 		btrfs_release_path(path);
 | |
| 		key.objectid = ref_id;
 | |
| 		key.type = BTRFS_ROOT_REF_KEY;
 | |
| 		key.offset = root_id;
 | |
| 		goto again;
 | |
| 	}
 | |
| 
 | |
| 	btrfs_free_path(path);
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Old btrfs forgets to init root_item->flags and root_item->byte_limit
 | |
|  * for subvolumes. To work around this problem, we steal a bit from
 | |
|  * root_item->inode_item->flags, and use it to indicate if those fields
 | |
|  * have been properly initialized.
 | |
|  */
 | |
| void btrfs_check_and_init_root_item(struct btrfs_root_item *root_item)
 | |
| {
 | |
| 	u64 inode_flags = btrfs_stack_inode_flags(&root_item->inode);
 | |
| 
 | |
| 	if (!(inode_flags & BTRFS_INODE_ROOT_ITEM_INIT)) {
 | |
| 		inode_flags |= BTRFS_INODE_ROOT_ITEM_INIT;
 | |
| 		btrfs_set_stack_inode_flags(&root_item->inode, inode_flags);
 | |
| 		btrfs_set_root_flags(root_item, 0);
 | |
| 		btrfs_set_root_limit(root_item, 0);
 | |
| 	}
 | |
| }
 | |
| 
 | |
| void btrfs_update_root_times(struct btrfs_trans_handle *trans,
 | |
| 			     struct btrfs_root *root)
 | |
| {
 | |
| 	struct btrfs_root_item *item = &root->root_item;
 | |
| 	struct timespec64 ct;
 | |
| 
 | |
| 	ktime_get_real_ts64(&ct);
 | |
| 	spin_lock(&root->root_item_lock);
 | |
| 	btrfs_set_root_ctransid(item, trans->transid);
 | |
| 	btrfs_set_stack_timespec_sec(&item->ctime, ct.tv_sec);
 | |
| 	btrfs_set_stack_timespec_nsec(&item->ctime, ct.tv_nsec);
 | |
| 	spin_unlock(&root->root_item_lock);
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * btrfs_subvolume_reserve_metadata() - reserve space for subvolume operation
 | |
|  * root: the root of the parent directory
 | |
|  * rsv: block reservation
 | |
|  * items: the number of items that we need do reservation
 | |
|  * use_global_rsv: allow fallback to the global block reservation
 | |
|  *
 | |
|  * This function is used to reserve the space for snapshot/subvolume
 | |
|  * creation and deletion. Those operations are different with the
 | |
|  * common file/directory operations, they change two fs/file trees
 | |
|  * and root tree, the number of items that the qgroup reserves is
 | |
|  * different with the free space reservation. So we can not use
 | |
|  * the space reservation mechanism in start_transaction().
 | |
|  */
 | |
| int btrfs_subvolume_reserve_metadata(struct btrfs_root *root,
 | |
| 				     struct btrfs_block_rsv *rsv, int items,
 | |
| 				     bool use_global_rsv)
 | |
| {
 | |
| 	u64 qgroup_num_bytes = 0;
 | |
| 	u64 num_bytes;
 | |
| 	int ret;
 | |
| 	struct btrfs_fs_info *fs_info = root->fs_info;
 | |
| 	struct btrfs_block_rsv *global_rsv = &fs_info->global_block_rsv;
 | |
| 
 | |
| 	if (test_bit(BTRFS_FS_QUOTA_ENABLED, &fs_info->flags)) {
 | |
| 		/* One for parent inode, two for dir entries */
 | |
| 		qgroup_num_bytes = 3 * fs_info->nodesize;
 | |
| 		ret = btrfs_qgroup_reserve_meta_prealloc(root,
 | |
| 							 qgroup_num_bytes, true,
 | |
| 							 false);
 | |
| 		if (ret)
 | |
| 			return ret;
 | |
| 	}
 | |
| 
 | |
| 	num_bytes = btrfs_calc_insert_metadata_size(fs_info, items);
 | |
| 	rsv->space_info = btrfs_find_space_info(fs_info,
 | |
| 					    BTRFS_BLOCK_GROUP_METADATA);
 | |
| 	ret = btrfs_block_rsv_add(fs_info, rsv, num_bytes,
 | |
| 				  BTRFS_RESERVE_FLUSH_ALL);
 | |
| 
 | |
| 	if (ret == -ENOSPC && use_global_rsv)
 | |
| 		ret = btrfs_block_rsv_migrate(global_rsv, rsv, num_bytes, true);
 | |
| 
 | |
| 	if (ret && qgroup_num_bytes)
 | |
| 		btrfs_qgroup_free_meta_prealloc(root, qgroup_num_bytes);
 | |
| 
 | |
| 	if (!ret) {
 | |
| 		spin_lock(&rsv->lock);
 | |
| 		rsv->qgroup_rsv_reserved += qgroup_num_bytes;
 | |
| 		spin_unlock(&rsv->lock);
 | |
| 	}
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| void btrfs_subvolume_release_metadata(struct btrfs_root *root,
 | |
| 				      struct btrfs_block_rsv *rsv)
 | |
| {
 | |
| 	struct btrfs_fs_info *fs_info = root->fs_info;
 | |
| 	u64 qgroup_to_release;
 | |
| 
 | |
| 	btrfs_block_rsv_release(fs_info, rsv, (u64)-1, &qgroup_to_release);
 | |
| 	btrfs_qgroup_convert_reserved_meta(root, qgroup_to_release);
 | |
| }
 |