Commit 11833d66 authored by Yan Zheng's avatar Yan Zheng Committed by Chris Mason

Btrfs: improve async block group caching

This patch gets rid of two limitations of async block group caching.
The old code delays handling pinned extents when block group is in
caching. To allocate logged file extents, the old code need wait
until block group is fully cached. To get rid of the limitations,
This patch introduces a data structure to track the progress of
caching. Base on the caching progress, we know which extents should
be added to the free space cache when handling the pinned extents.
The logged file extents are also handled in a similar way.

This patch also changes how pinned extents are tracked. The old
code uses one tree to track pinned extents, and copy the pinned
extents tree at transaction commit time. This patch makes it use
two trees to track pinned extents. One tree for extents that are
pinned in the running transaction, one tree for extents that can
be unpinned. At transaction commit time, we swap the two trees.
Signed-off-by: default avatarYan Zheng <zheng.yan@oracle.com>
Signed-off-by: default avatarChris Mason <chris.mason@oracle.com>
parent 6e74057c
...@@ -726,6 +726,15 @@ enum btrfs_caching_type { ...@@ -726,6 +726,15 @@ enum btrfs_caching_type {
BTRFS_CACHE_FINISHED = 2, BTRFS_CACHE_FINISHED = 2,
}; };
struct btrfs_caching_control {
struct list_head list;
struct mutex mutex;
wait_queue_head_t wait;
struct btrfs_block_group_cache *block_group;
u64 progress;
atomic_t count;
};
struct btrfs_block_group_cache { struct btrfs_block_group_cache {
struct btrfs_key key; struct btrfs_key key;
struct btrfs_block_group_item item; struct btrfs_block_group_item item;
...@@ -742,8 +751,9 @@ struct btrfs_block_group_cache { ...@@ -742,8 +751,9 @@ struct btrfs_block_group_cache {
int dirty; int dirty;
/* cache tracking stuff */ /* cache tracking stuff */
wait_queue_head_t caching_q;
int cached; int cached;
struct btrfs_caching_control *caching_ctl;
u64 last_byte_to_unpin;
struct btrfs_space_info *space_info; struct btrfs_space_info *space_info;
...@@ -788,7 +798,8 @@ struct btrfs_fs_info { ...@@ -788,7 +798,8 @@ struct btrfs_fs_info {
spinlock_t block_group_cache_lock; spinlock_t block_group_cache_lock;
struct rb_root block_group_cache_tree; struct rb_root block_group_cache_tree;
struct extent_io_tree pinned_extents; struct extent_io_tree freed_extents[2];
struct extent_io_tree *pinned_extents;
/* logical->physical extent mapping */ /* logical->physical extent mapping */
struct btrfs_mapping_tree mapping_tree; struct btrfs_mapping_tree mapping_tree;
...@@ -825,8 +836,6 @@ struct btrfs_fs_info { ...@@ -825,8 +836,6 @@ struct btrfs_fs_info {
struct mutex drop_mutex; struct mutex drop_mutex;
struct mutex volume_mutex; struct mutex volume_mutex;
struct mutex tree_reloc_mutex; struct mutex tree_reloc_mutex;
struct rw_semaphore extent_commit_sem;
/* /*
* this protects the ordered operations list only while we are * this protects the ordered operations list only while we are
* processing all of the entries on it. This way we make * processing all of the entries on it. This way we make
...@@ -835,10 +844,12 @@ struct btrfs_fs_info { ...@@ -835,10 +844,12 @@ struct btrfs_fs_info {
* before jumping into the main commit. * before jumping into the main commit.
*/ */
struct mutex ordered_operations_mutex; struct mutex ordered_operations_mutex;
struct rw_semaphore extent_commit_sem;
struct list_head trans_list; struct list_head trans_list;
struct list_head hashers; struct list_head hashers;
struct list_head dead_roots; struct list_head dead_roots;
struct list_head caching_block_groups;
atomic_t nr_async_submits; atomic_t nr_async_submits;
atomic_t async_submit_draining; atomic_t async_submit_draining;
...@@ -1920,8 +1931,8 @@ void btrfs_put_block_group(struct btrfs_block_group_cache *cache); ...@@ -1920,8 +1931,8 @@ void btrfs_put_block_group(struct btrfs_block_group_cache *cache);
int btrfs_run_delayed_refs(struct btrfs_trans_handle *trans, int btrfs_run_delayed_refs(struct btrfs_trans_handle *trans,
struct btrfs_root *root, unsigned long count); struct btrfs_root *root, unsigned long count);
int btrfs_lookup_extent(struct btrfs_root *root, u64 start, u64 len); int btrfs_lookup_extent(struct btrfs_root *root, u64 start, u64 len);
int btrfs_update_pinned_extents(struct btrfs_root *root, int btrfs_pin_extent(struct btrfs_root *root,
u64 bytenr, u64 num, int pin); u64 bytenr, u64 num, int reserved);
int btrfs_drop_leaf_ref(struct btrfs_trans_handle *trans, int btrfs_drop_leaf_ref(struct btrfs_trans_handle *trans,
struct btrfs_root *root, struct extent_buffer *leaf); struct btrfs_root *root, struct extent_buffer *leaf);
int btrfs_cross_ref_exist(struct btrfs_trans_handle *trans, int btrfs_cross_ref_exist(struct btrfs_trans_handle *trans,
...@@ -1971,9 +1982,10 @@ int btrfs_free_extent(struct btrfs_trans_handle *trans, ...@@ -1971,9 +1982,10 @@ int btrfs_free_extent(struct btrfs_trans_handle *trans,
u64 root_objectid, u64 owner, u64 offset); u64 root_objectid, u64 owner, u64 offset);
int btrfs_free_reserved_extent(struct btrfs_root *root, u64 start, u64 len); int btrfs_free_reserved_extent(struct btrfs_root *root, u64 start, u64 len);
int btrfs_prepare_extent_commit(struct btrfs_trans_handle *trans,
struct btrfs_root *root);
int btrfs_finish_extent_commit(struct btrfs_trans_handle *trans, int btrfs_finish_extent_commit(struct btrfs_trans_handle *trans,
struct btrfs_root *root, struct btrfs_root *root);
struct extent_io_tree *unpin);
int btrfs_inc_extent_ref(struct btrfs_trans_handle *trans, int btrfs_inc_extent_ref(struct btrfs_trans_handle *trans,
struct btrfs_root *root, struct btrfs_root *root,
u64 bytenr, u64 num_bytes, u64 parent, u64 bytenr, u64 num_bytes, u64 parent,
...@@ -2006,7 +2018,6 @@ void btrfs_delalloc_reserve_space(struct btrfs_root *root, struct inode *inode, ...@@ -2006,7 +2018,6 @@ void btrfs_delalloc_reserve_space(struct btrfs_root *root, struct inode *inode,
u64 bytes); u64 bytes);
void btrfs_delalloc_free_space(struct btrfs_root *root, struct inode *inode, void btrfs_delalloc_free_space(struct btrfs_root *root, struct inode *inode,
u64 bytes); u64 bytes);
void btrfs_free_pinned_extents(struct btrfs_fs_info *info);
/* ctree.c */ /* ctree.c */
int btrfs_bin_search(struct extent_buffer *eb, struct btrfs_key *key, int btrfs_bin_search(struct extent_buffer *eb, struct btrfs_key *key,
int level, int *slot); int level, int *slot);
......
...@@ -1563,6 +1563,7 @@ struct btrfs_root *open_ctree(struct super_block *sb, ...@@ -1563,6 +1563,7 @@ struct btrfs_root *open_ctree(struct super_block *sb,
INIT_LIST_HEAD(&fs_info->hashers); INIT_LIST_HEAD(&fs_info->hashers);
INIT_LIST_HEAD(&fs_info->delalloc_inodes); INIT_LIST_HEAD(&fs_info->delalloc_inodes);
INIT_LIST_HEAD(&fs_info->ordered_operations); INIT_LIST_HEAD(&fs_info->ordered_operations);
INIT_LIST_HEAD(&fs_info->caching_block_groups);
spin_lock_init(&fs_info->delalloc_lock); spin_lock_init(&fs_info->delalloc_lock);
spin_lock_init(&fs_info->new_trans_lock); spin_lock_init(&fs_info->new_trans_lock);
spin_lock_init(&fs_info->ref_cache_lock); spin_lock_init(&fs_info->ref_cache_lock);
...@@ -1621,8 +1622,11 @@ struct btrfs_root *open_ctree(struct super_block *sb, ...@@ -1621,8 +1622,11 @@ struct btrfs_root *open_ctree(struct super_block *sb,
spin_lock_init(&fs_info->block_group_cache_lock); spin_lock_init(&fs_info->block_group_cache_lock);
fs_info->block_group_cache_tree.rb_node = NULL; fs_info->block_group_cache_tree.rb_node = NULL;
extent_io_tree_init(&fs_info->pinned_extents, extent_io_tree_init(&fs_info->freed_extents[0],
fs_info->btree_inode->i_mapping, GFP_NOFS); fs_info->btree_inode->i_mapping, GFP_NOFS);
extent_io_tree_init(&fs_info->freed_extents[1],
fs_info->btree_inode->i_mapping, GFP_NOFS);
fs_info->pinned_extents = &fs_info->freed_extents[0];
fs_info->do_barriers = 1; fs_info->do_barriers = 1;
BTRFS_I(fs_info->btree_inode)->root = tree_root; BTRFS_I(fs_info->btree_inode)->root = tree_root;
...@@ -2359,7 +2363,6 @@ int close_ctree(struct btrfs_root *root) ...@@ -2359,7 +2363,6 @@ int close_ctree(struct btrfs_root *root)
free_extent_buffer(root->fs_info->csum_root->commit_root); free_extent_buffer(root->fs_info->csum_root->commit_root);
btrfs_free_block_groups(root->fs_info); btrfs_free_block_groups(root->fs_info);
btrfs_free_pinned_extents(root->fs_info);
del_fs_roots(fs_info); del_fs_roots(fs_info);
......
...@@ -32,12 +32,12 @@ ...@@ -32,12 +32,12 @@
#include "locking.h" #include "locking.h"
#include "free-space-cache.h" #include "free-space-cache.h"
static int update_reserved_extents(struct btrfs_root *root,
u64 bytenr, u64 num, int reserve);
static int update_block_group(struct btrfs_trans_handle *trans, static int update_block_group(struct btrfs_trans_handle *trans,
struct btrfs_root *root, struct btrfs_root *root,
u64 bytenr, u64 num_bytes, int alloc, u64 bytenr, u64 num_bytes, int alloc,
int mark_free); int mark_free);
static int update_reserved_extents(struct btrfs_block_group_cache *cache,
u64 num_bytes, int reserve);
static int __btrfs_free_extent(struct btrfs_trans_handle *trans, static int __btrfs_free_extent(struct btrfs_trans_handle *trans,
struct btrfs_root *root, struct btrfs_root *root,
u64 bytenr, u64 num_bytes, u64 parent, u64 bytenr, u64 num_bytes, u64 parent,
...@@ -57,10 +57,17 @@ static int alloc_reserved_tree_block(struct btrfs_trans_handle *trans, ...@@ -57,10 +57,17 @@ static int alloc_reserved_tree_block(struct btrfs_trans_handle *trans,
u64 parent, u64 root_objectid, u64 parent, u64 root_objectid,
u64 flags, struct btrfs_disk_key *key, u64 flags, struct btrfs_disk_key *key,
int level, struct btrfs_key *ins); int level, struct btrfs_key *ins);
static int do_chunk_alloc(struct btrfs_trans_handle *trans, static int do_chunk_alloc(struct btrfs_trans_handle *trans,
struct btrfs_root *extent_root, u64 alloc_bytes, struct btrfs_root *extent_root, u64 alloc_bytes,
u64 flags, int force); u64 flags, int force);
static int pin_down_bytes(struct btrfs_trans_handle *trans,
struct btrfs_root *root,
struct btrfs_path *path,
u64 bytenr, u64 num_bytes,
int is_data, int reserved,
struct extent_buffer **must_clean);
static int find_next_key(struct btrfs_path *path, int level,
struct btrfs_key *key);
static noinline int static noinline int
block_group_cache_done(struct btrfs_block_group_cache *cache) block_group_cache_done(struct btrfs_block_group_cache *cache)
...@@ -153,34 +160,34 @@ block_group_cache_tree_search(struct btrfs_fs_info *info, u64 bytenr, ...@@ -153,34 +160,34 @@ block_group_cache_tree_search(struct btrfs_fs_info *info, u64 bytenr,
return ret; return ret;
} }
/* static int add_excluded_extent(struct btrfs_root *root,
* We always set EXTENT_LOCKED for the super mirror extents so we don't u64 start, u64 num_bytes)
* overwrite them, so those bits need to be unset. Also, if we are unmounting
* with pinned extents still sitting there because we had a block group caching,
* we need to clear those now, since we are done.
*/
void btrfs_free_pinned_extents(struct btrfs_fs_info *info)
{ {
u64 start, end, last = 0; u64 end = start + num_bytes - 1;
int ret; set_extent_bits(&root->fs_info->freed_extents[0],
start, end, EXTENT_UPTODATE, GFP_NOFS);
set_extent_bits(&root->fs_info->freed_extents[1],
start, end, EXTENT_UPTODATE, GFP_NOFS);
return 0;
}
while (1) { static void free_excluded_extents(struct btrfs_root *root,
ret = find_first_extent_bit(&info->pinned_extents, last, struct btrfs_block_group_cache *cache)
&start, &end, {
EXTENT_LOCKED|EXTENT_DIRTY); u64 start, end;
if (ret)
break;
clear_extent_bits(&info->pinned_extents, start, end, start = cache->key.objectid;
EXTENT_LOCKED|EXTENT_DIRTY, GFP_NOFS); end = start + cache->key.offset - 1;
last = end+1;
} clear_extent_bits(&root->fs_info->freed_extents[0],
start, end, EXTENT_UPTODATE, GFP_NOFS);
clear_extent_bits(&root->fs_info->freed_extents[1],
start, end, EXTENT_UPTODATE, GFP_NOFS);
} }
static int remove_sb_from_cache(struct btrfs_root *root, static int exclude_super_stripes(struct btrfs_root *root,
struct btrfs_block_group_cache *cache) struct btrfs_block_group_cache *cache)
{ {
struct btrfs_fs_info *fs_info = root->fs_info;
u64 bytenr; u64 bytenr;
u64 *logical; u64 *logical;
int stripe_len; int stripe_len;
...@@ -192,17 +199,41 @@ static int remove_sb_from_cache(struct btrfs_root *root, ...@@ -192,17 +199,41 @@ static int remove_sb_from_cache(struct btrfs_root *root,
cache->key.objectid, bytenr, cache->key.objectid, bytenr,
0, &logical, &nr, &stripe_len); 0, &logical, &nr, &stripe_len);
BUG_ON(ret); BUG_ON(ret);
while (nr--) { while (nr--) {
try_lock_extent(&fs_info->pinned_extents, ret = add_excluded_extent(root, logical[nr],
logical[nr], stripe_len);
logical[nr] + stripe_len - 1, GFP_NOFS); BUG_ON(ret);
} }
kfree(logical); kfree(logical);
} }
return 0; return 0;
} }
static struct btrfs_caching_control *
get_caching_control(struct btrfs_block_group_cache *cache)
{
struct btrfs_caching_control *ctl;
spin_lock(&cache->lock);
if (cache->cached != BTRFS_CACHE_STARTED) {
spin_unlock(&cache->lock);
return NULL;
}
ctl = cache->caching_ctl;
atomic_inc(&ctl->count);
spin_unlock(&cache->lock);
return ctl;
}
static void put_caching_control(struct btrfs_caching_control *ctl)
{
if (atomic_dec_and_test(&ctl->count))
kfree(ctl);
}
/* /*
* this is only called by cache_block_group, since we could have freed extents * this is only called by cache_block_group, since we could have freed extents
* we need to check the pinned_extents for any extents that can't be used yet * we need to check the pinned_extents for any extents that can't be used yet
...@@ -215,9 +246,9 @@ static u64 add_new_free_space(struct btrfs_block_group_cache *block_group, ...@@ -215,9 +246,9 @@ static u64 add_new_free_space(struct btrfs_block_group_cache *block_group,
int ret; int ret;
while (start < end) { while (start < end) {
ret = find_first_extent_bit(&info->pinned_extents, start, ret = find_first_extent_bit(info->pinned_extents, start,
&extent_start, &extent_end, &extent_start, &extent_end,
EXTENT_DIRTY|EXTENT_LOCKED); EXTENT_DIRTY | EXTENT_UPTODATE);
if (ret) if (ret)
break; break;
...@@ -249,22 +280,24 @@ static int caching_kthread(void *data) ...@@ -249,22 +280,24 @@ static int caching_kthread(void *data)
{ {
struct btrfs_block_group_cache *block_group = data; struct btrfs_block_group_cache *block_group = data;
struct btrfs_fs_info *fs_info = block_group->fs_info; struct btrfs_fs_info *fs_info = block_group->fs_info;
u64 last = 0; struct btrfs_caching_control *caching_ctl = block_group->caching_ctl;
struct btrfs_root *extent_root = fs_info->extent_root;
struct btrfs_path *path; struct btrfs_path *path;
int ret = 0;
struct btrfs_key key;
struct extent_buffer *leaf; struct extent_buffer *leaf;
int slot; struct btrfs_key key;
u64 total_found = 0; u64 total_found = 0;
u64 last = 0;
BUG_ON(!fs_info); u32 nritems;
int ret = 0;
path = btrfs_alloc_path(); path = btrfs_alloc_path();
if (!path) if (!path)
return -ENOMEM; return -ENOMEM;
atomic_inc(&block_group->space_info->caching_threads); exclude_super_stripes(extent_root, block_group);
last = max_t(u64, block_group->key.objectid, BTRFS_SUPER_INFO_OFFSET); last = max_t(u64, block_group->key.objectid, BTRFS_SUPER_INFO_OFFSET);
/* /*
* We don't want to deadlock with somebody trying to allocate a new * We don't want to deadlock with somebody trying to allocate a new
* extent for the extent root while also trying to search the extent * extent for the extent root while also trying to search the extent
...@@ -277,74 +310,64 @@ static int caching_kthread(void *data) ...@@ -277,74 +310,64 @@ static int caching_kthread(void *data)
key.objectid = last; key.objectid = last;
key.offset = 0; key.offset = 0;
btrfs_set_key_type(&key, BTRFS_EXTENT_ITEM_KEY); key.type = BTRFS_EXTENT_ITEM_KEY;
again: again:
mutex_lock(&caching_ctl->mutex);
/* need to make sure the commit_root doesn't disappear */ /* need to make sure the commit_root doesn't disappear */
down_read(&fs_info->extent_commit_sem); down_read(&fs_info->extent_commit_sem);
ret = btrfs_search_slot(NULL, fs_info->extent_root, &key, path, 0, 0); ret = btrfs_search_slot(NULL, extent_root, &key, path, 0, 0);
if (ret < 0) if (ret < 0)
goto err; goto err;
leaf = path->nodes[0];
nritems = btrfs_header_nritems(leaf);
while (1) { while (1) {
smp_mb(); smp_mb();
if (block_group->fs_info->closing > 1) { if (fs_info->closing > 1) {
last = (u64)-1; last = (u64)-1;
break; break;
} }
leaf = path->nodes[0]; if (path->slots[0] < nritems) {
slot = path->slots[0]; btrfs_item_key_to_cpu(leaf, &key, path->slots[0]);
if (slot >= btrfs_header_nritems(leaf)) { } else {
ret = btrfs_next_leaf(fs_info->extent_root, path); ret = find_next_key(path, 0, &key);
if (ret < 0) if (ret)
goto err;
else if (ret)
break; break;
if (need_resched() || caching_ctl->progress = last;
btrfs_transaction_in_commit(fs_info)) { btrfs_release_path(extent_root, path);
leaf = path->nodes[0]; up_read(&fs_info->extent_commit_sem);
mutex_unlock(&caching_ctl->mutex);
/* this shouldn't happen, but if the if (btrfs_transaction_in_commit(fs_info))
* leaf is empty just move on.
*/
if (btrfs_header_nritems(leaf) == 0)
break;
/*
* we need to copy the key out so that
* we are sure the next search advances
* us forward in the btree.
*/
btrfs_item_key_to_cpu(leaf, &key, 0);
btrfs_release_path(fs_info->extent_root, path);
up_read(&fs_info->extent_commit_sem);
schedule_timeout(1); schedule_timeout(1);
goto again; else
} cond_resched();
goto again;
}
if (key.objectid < block_group->key.objectid) {
path->slots[0]++;
continue; continue;
} }
btrfs_item_key_to_cpu(leaf, &key, slot);
if (key.objectid < block_group->key.objectid)
goto next;
if (key.objectid >= block_group->key.objectid + if (key.objectid >= block_group->key.objectid +
block_group->key.offset) block_group->key.offset)
break; break;
if (btrfs_key_type(&key) == BTRFS_EXTENT_ITEM_KEY) { if (key.type == BTRFS_EXTENT_ITEM_KEY) {
total_found += add_new_free_space(block_group, total_found += add_new_free_space(block_group,
fs_info, last, fs_info, last,
key.objectid); key.objectid);
last = key.objectid + key.offset; last = key.objectid + key.offset;
}
if (total_found > (1024 * 1024 * 2)) { if (total_found > (1024 * 1024 * 2)) {
total_found = 0; total_found = 0;
wake_up(&block_group->caching_q); wake_up(&caching_ctl->wait);
}
} }
next:
path->slots[0]++; path->slots[0]++;
} }
ret = 0; ret = 0;
...@@ -352,33 +375,65 @@ static int caching_kthread(void *data) ...@@ -352,33 +375,65 @@ static int caching_kthread(void *data)
total_found += add_new_free_space(block_group, fs_info, last, total_found += add_new_free_space(block_group, fs_info, last,
block_group->key.objectid + block_group->key.objectid +
block_group->key.offset); block_group->key.offset);
caching_ctl->progress = (u64)-1;
spin_lock(&block_group->lock); spin_lock(&block_group->lock);
block_group->caching_ctl = NULL;
block_group->cached = BTRFS_CACHE_FINISHED; block_group->cached = BTRFS_CACHE_FINISHED;
spin_unlock(&block_group->lock); spin_unlock(&block_group->lock);
err: err:
btrfs_free_path(path); btrfs_free_path(path);
up_read(&fs_info->extent_commit_sem); up_read(&fs_info->extent_commit_sem);
atomic_dec(&block_group->space_info->caching_threads);
wake_up(&block_group->caching_q);
free_excluded_extents(extent_root, block_group);
mutex_unlock(&caching_ctl->mutex);
wake_up(&caching_ctl->wait);
put_caching_control(caching_ctl);
atomic_dec(&block_group->space_info->caching_threads);
return 0; return 0;
} }
static int cache_block_group(struct btrfs_block_group_cache *cache) static int cache_block_group(struct btrfs_block_group_cache *cache)
{ {
struct btrfs_fs_info *fs_info = cache->fs_info;
struct btrfs_caching_control *caching_ctl;
struct task_struct *tsk; struct task_struct *tsk;
int ret = 0; int ret = 0;
smp_mb();
if (cache->cached != BTRFS_CACHE_NO)
return 0;
caching_ctl = kzalloc(sizeof(*caching_ctl), GFP_KERNEL);
BUG_ON(!caching_ctl);
INIT_LIST_HEAD(&caching_ctl->list);
mutex_init(&caching_ctl->mutex);
init_waitqueue_head(&caching_ctl->wait);
caching_ctl->block_group = cache;
caching_ctl->progress = cache->key.objectid;
/* one for caching kthread, one for caching block group list */
atomic_set(&caching_ctl->count, 2);
spin_lock(&cache->lock); spin_lock(&cache->lock);
if (cache->cached != BTRFS_CACHE_NO) { if (cache->cached != BTRFS_CACHE_NO) {
spin_unlock(&cache->lock); spin_unlock(&cache->lock);
return ret; kfree(caching_ctl);
return 0;
} }
cache->caching_ctl = caching_ctl;
cache->cached = BTRFS_CACHE_STARTED; cache->cached = BTRFS_CACHE_STARTED;
spin_unlock(&cache->lock); spin_unlock(&cache->lock);
down_write(&fs_info->extent_commit_sem);
list_add_tail(&caching_ctl->list, &fs_info->caching_block_groups);
up_write(&fs_info->extent_commit_sem);
atomic_inc(&cache->space_info->caching_threads);
tsk = kthread_run(caching_kthread, cache, "btrfs-cache-%llu\n", tsk = kthread_run(caching_kthread, cache, "btrfs-cache-%llu\n",
cache->key.objectid); cache->key.objectid);
if (IS_ERR(tsk)) { if (IS_ERR(tsk)) {
...@@ -1656,7 +1711,6 @@ static int run_delayed_data_ref(struct btrfs_trans_handle *trans, ...@@ -1656,7 +1711,6 @@ static int run_delayed_data_ref(struct btrfs_trans_handle *trans,
parent, ref_root, flags, parent, ref_root, flags,
ref->objectid, ref->offset, ref->objectid, ref->offset,
&ins, node->ref_mod); &ins, node->ref_mod);
update_reserved_extents(root, ins.objectid, ins.offset, 0);
} else if (node->action == BTRFS_ADD_DELAYED_REF) { } else if (node->action == BTRFS_ADD_DELAYED_REF) {
ret = __btrfs_inc_extent_ref(trans, root, node->bytenr, ret = __btrfs_inc_extent_ref(trans, root, node->bytenr,
node->num_bytes, parent, node->num_bytes, parent,
...@@ -1782,7 +1836,6 @@ static int run_delayed_tree_ref(struct btrfs_trans_handle *trans, ...@@ -1782,7 +1836,6 @@ static int run_delayed_tree_ref(struct btrfs_trans_handle *trans,
extent_op->flags_to_set, extent_op->flags_to_set,
&extent_op->key, &extent_op->key,
ref->level, &ins); ref->level, &ins);
update_reserved_extents(root, ins.objectid, ins.offset, 0);
} else if (node->action == BTRFS_ADD_DELAYED_REF) { } else if (node->action == BTRFS_ADD_DELAYED_REF) {
ret = __btrfs_inc_extent_ref(trans, root, node->bytenr, ret = __btrfs_inc_extent_ref(trans, root, node->bytenr,
node->num_bytes, parent, ref_root, node->num_bytes, parent, ref_root,
...@@ -1817,16 +1870,32 @@ static int run_one_delayed_ref(struct btrfs_trans_handle *trans, ...@@ -1817,16 +1870,32 @@ static int run_one_delayed_ref(struct btrfs_trans_handle *trans,
BUG_ON(extent_op); BUG_ON(extent_op);
head = btrfs_delayed_node_to_head(node); head = btrfs_delayed_node_to_head(node);
if (insert_reserved) { if (insert_reserved) {
int mark_free = 0;
struct extent_buffer *must_clean = NULL;
ret = pin_down_bytes(trans, root, NULL,
node->bytenr, node->num_bytes,
head->is_data, 1, &must_clean);
if (ret > 0)
mark_free = 1;
if (must_clean) {
clean_tree_block(NULL, root, must_clean);
btrfs_tree_unlock(must_clean);
free_extent_buffer(must_clean);
}
if (head->is_data) { if (head->is_data) {
ret = btrfs_del_csums(trans, root, ret = btrfs_del_csums(trans, root,
node->bytenr, node->bytenr,
node->num_bytes); node->num_bytes);
BUG_ON(ret); BUG_ON(ret);
} }
btrfs_update_pinned_extents(root, node->bytenr, if (mark_free) {
node->num_bytes, 1); ret = btrfs_free_reserved_extent(root,
update_reserved_extents(root, node->bytenr, node->bytenr,
node->num_bytes, 0); node->num_bytes);
BUG_ON(ret);
}
} }
mutex_unlock(&head->mutex); mutex_unlock(&head->mutex);
return 0; return 0;
...@@ -3008,10 +3077,12 @@ static int update_block_group(struct btrfs_trans_handle *trans, ...@@ -3008,10 +3077,12 @@ static int update_block_group(struct btrfs_trans_handle *trans,
num_bytes = min(total, cache->key.offset - byte_in_group); num_bytes = min(total, cache->key.offset - byte_in_group);
if (alloc) { if (alloc) {
old_val += num_bytes; old_val += num_bytes;
btrfs_set_block_group_used(&cache->item, old_val);
cache->reserved -= num_bytes;
cache->space_info->bytes_used += num_bytes; cache->space_info->bytes_used += num_bytes;
cache->space_info->bytes_reserved -= num_bytes;
if (cache->ro) if (cache->ro)
cache->space_info->bytes_readonly -= num_bytes; cache->space_info->bytes_readonly -= num_bytes;
btrfs_set_block_group_used(&cache->item, old_val);
spin_unlock(&cache->lock); spin_unlock(&cache->lock);
spin_unlock(&cache->space_info->lock); spin_unlock(&cache->space_info->lock);
} else { } else {
...@@ -3056,127 +3127,136 @@ static u64 first_logical_byte(struct btrfs_root *root, u64 search_start) ...@@ -3056,127 +3127,136 @@ static u64 first_logical_byte(struct btrfs_root *root, u64 search_start)
return bytenr; return bytenr;
} }
int btrfs_update_pinned_extents(struct btrfs_root *root, /*
u64 bytenr, u64 num, int pin) * this function must be called within transaction
*/
int btrfs_pin_extent(struct btrfs_root *root,
u64 bytenr, u64 num_bytes, int reserved)
{ {
u64 len;
struct btrfs_block_group_cache *cache;
struct btrfs_fs_info *fs_info = root->fs_info; struct btrfs_fs_info *fs_info = root->fs_info;
struct btrfs_block_group_cache *cache;
if (pin) cache = btrfs_lookup_block_group(fs_info, bytenr);
set_extent_dirty(&fs_info->pinned_extents, BUG_ON(!cache);
bytenr, bytenr + num - 1, GFP_NOFS);
while (num > 0) {
cache = btrfs_lookup_block_group(fs_info, bytenr);
BUG_ON(!cache);
len = min(num, cache->key.offset -
(bytenr - cache->key.objectid));
if (pin) {
spin_lock(&cache->space_info->lock);
spin_lock(&cache->lock);
cache->pinned += len;
cache->space_info->bytes_pinned += len;
spin_unlock(&cache->lock);
spin_unlock(&cache->space_info->lock);
fs_info->total_pinned += len;
} else {
int unpin = 0;
/* spin_lock(&cache->space_info->lock);
* in order to not race with the block group caching, we spin_lock(&cache->lock);
* only want to unpin the extent if we are cached. If cache->pinned += num_bytes;
* we aren't cached, we want to start async caching this cache->space_info->bytes_pinned += num_bytes;
* block group so we can free the extent the next time if (reserved) {
* around. cache->reserved -= num_bytes;
*/ cache->space_info->bytes_reserved -= num_bytes;
spin_lock(&cache->space_info->lock); }
spin_lock(&cache->lock); spin_unlock(&cache->lock);
unpin = (cache->cached == BTRFS_CACHE_FINISHED); spin_unlock(&cache->space_info->lock);
if (likely(unpin)) {
cache->pinned -= len;
cache->space_info->bytes_pinned -= len;
fs_info->total_pinned -= len;
}
spin_unlock(&cache->lock);
spin_unlock(&cache->space_info->lock);
if (likely(unpin)) btrfs_put_block_group(cache);
clear_extent_dirty(&fs_info->pinned_extents,
bytenr, bytenr + len -1,
GFP_NOFS);
else
cache_block_group(cache);
if (unpin) set_extent_dirty(fs_info->pinned_extents,
btrfs_add_free_space(cache, bytenr, len); bytenr, bytenr + num_bytes - 1, GFP_NOFS);
} return 0;
btrfs_put_block_group(cache); }
bytenr += len;
num -= len; static int update_reserved_extents(struct btrfs_block_group_cache *cache,
u64 num_bytes, int reserve)
{
spin_lock(&cache->space_info->lock);
spin_lock(&cache->lock);
if (reserve) {
cache->reserved += num_bytes;
cache->space_info->bytes_reserved += num_bytes;
} else {
cache->reserved -= num_bytes;
cache->space_info->bytes_reserved -= num_bytes;
} }
spin_unlock(&cache->lock);
spin_unlock(&cache->space_info->lock);
return 0; return 0;
} }
static int update_reserved_extents(struct btrfs_root *root, int btrfs_prepare_extent_commit(struct btrfs_trans_handle *trans,
u64 bytenr, u64 num, int reserve) struct btrfs_root *root)
{ {
u64 len;
struct btrfs_block_group_cache *cache;
struct btrfs_fs_info *fs_info = root->fs_info; struct btrfs_fs_info *fs_info = root->fs_info;
struct btrfs_caching_control *next;
struct btrfs_caching_control *caching_ctl;
struct btrfs_block_group_cache *cache;
while (num > 0) { down_write(&fs_info->extent_commit_sem);
cache = btrfs_lookup_block_group(fs_info, bytenr);
BUG_ON(!cache);
len = min(num, cache->key.offset -
(bytenr - cache->key.objectid));
spin_lock(&cache->space_info->lock); list_for_each_entry_safe(caching_ctl, next,
spin_lock(&cache->lock); &fs_info->caching_block_groups, list) {
if (reserve) { cache = caching_ctl->block_group;
cache->reserved += len; if (block_group_cache_done(cache)) {
cache->space_info->bytes_reserved += len; cache->last_byte_to_unpin = (u64)-1;
list_del_init(&caching_ctl->list);
put_caching_control(caching_ctl);
} else { } else {
cache->reserved -= len; cache->last_byte_to_unpin = caching_ctl->progress;
cache->space_info->bytes_reserved -= len;
} }
spin_unlock(&cache->lock);
spin_unlock(&cache->space_info->lock);
btrfs_put_block_group(cache);
bytenr += len;
num -= len;
} }
if (fs_info->pinned_extents == &fs_info->freed_extents[0])
fs_info->pinned_extents = &fs_info->freed_extents[1];
else
fs_info->pinned_extents = &fs_info->freed_extents[0];
up_write(&fs_info->extent_commit_sem);
return 0; return 0;
} }
int btrfs_copy_pinned(struct btrfs_root *root, struct extent_io_tree *copy) static int unpin_extent_range(struct btrfs_root *root, u64 start, u64 end)
{ {
u64 last = 0; struct btrfs_fs_info *fs_info = root->fs_info;
u64 start; struct btrfs_block_group_cache *cache = NULL;
u64 end; u64 len;
struct extent_io_tree *pinned_extents = &root->fs_info->pinned_extents;
int ret;
while (1) { while (start <= end) {
ret = find_first_extent_bit(pinned_extents, last, if (!cache ||
&start, &end, EXTENT_DIRTY); start >= cache->key.objectid + cache->key.offset) {
if (ret) if (cache)
break; btrfs_put_block_group(cache);
cache = btrfs_lookup_block_group(fs_info, start);
BUG_ON(!cache);
}
len = cache->key.objectid + cache->key.offset - start;
len = min(len, end + 1 - start);
if (start < cache->last_byte_to_unpin) {
len = min(len, cache->last_byte_to_unpin - start);
btrfs_add_free_space(cache, start, len);
}
spin_lock(&cache->space_info->lock);
spin_lock(&cache->lock);
cache->pinned -= len;
cache->space_info->bytes_pinned -= len;
spin_unlock(&cache->lock);
spin_unlock(&cache->space_info->lock);
set_extent_dirty(copy, start, end, GFP_NOFS); start += len;
last = end + 1;
} }
if (cache)
btrfs_put_block_group(cache);
return 0; return 0;
} }
int btrfs_finish_extent_commit(struct btrfs_trans_handle *trans, int btrfs_finish_extent_commit(struct btrfs_trans_handle *trans,
struct btrfs_root *root, struct btrfs_root *root)
struct extent_io_tree *unpin)
{ {
struct btrfs_fs_info *fs_info = root->fs_info;
struct extent_io_tree *unpin;
u64 start; u64 start;
u64 end; u64 end;
int ret; int ret;
if (fs_info->pinned_extents == &fs_info->freed_extents[0])
unpin = &fs_info->freed_extents[1];
else
unpin = &fs_info->freed_extents[0];
while (1) { while (1) {
ret = find_first_extent_bit(unpin, 0, &start, &end, ret = find_first_extent_bit(unpin, 0, &start, &end,
EXTENT_DIRTY); EXTENT_DIRTY);
...@@ -3185,10 +3265,8 @@ int btrfs_finish_extent_commit(struct btrfs_trans_handle *trans, ...@@ -3185,10 +3265,8 @@ int btrfs_finish_extent_commit(struct btrfs_trans_handle *trans,
ret = btrfs_discard_extent(root, start, end + 1 - start); ret = btrfs_discard_extent(root, start, end + 1 - start);
/* unlocks the pinned mutex */
btrfs_update_pinned_extents(root, start, end + 1 - start, 0);
clear_extent_dirty(unpin, start, end, GFP_NOFS); clear_extent_dirty(unpin, start, end, GFP_NOFS);
unpin_extent_range(root, start, end);
cond_resched(); cond_resched();
} }
...@@ -3198,7 +3276,8 @@ int btrfs_finish_extent_commit(struct btrfs_trans_handle *trans, ...@@ -3198,7 +3276,8 @@ int btrfs_finish_extent_commit(struct btrfs_trans_handle *trans,
static int pin_down_bytes(struct btrfs_trans_handle *trans, static int pin_down_bytes(struct btrfs_trans_handle *trans,
struct btrfs_root *root, struct btrfs_root *root,
struct btrfs_path *path, struct btrfs_path *path,
u64 bytenr, u64 num_bytes, int is_data, u64 bytenr, u64 num_bytes,
int is_data, int reserved,
struct extent_buffer **must_clean) struct extent_buffer **must_clean)
{ {
int err = 0; int err = 0;
...@@ -3230,15 +3309,15 @@ static int pin_down_bytes(struct btrfs_trans_handle *trans, ...@@ -3230,15 +3309,15 @@ static int pin_down_bytes(struct btrfs_trans_handle *trans,
} }
free_extent_buffer(buf); free_extent_buffer(buf);
pinit: pinit:
btrfs_set_path_blocking(path); if (path)
btrfs_set_path_blocking(path);
/* unlocks the pinned mutex */ /* unlocks the pinned mutex */
btrfs_update_pinned_extents(root, bytenr, num_bytes, 1); btrfs_pin_extent(root, bytenr, num_bytes, reserved);
BUG_ON(err < 0); BUG_ON(err < 0);
return 0; return 0;
} }
static int __btrfs_free_extent(struct btrfs_trans_handle *trans, static int __btrfs_free_extent(struct btrfs_trans_handle *trans,
struct btrfs_root *root, struct btrfs_root *root,
u64 bytenr, u64 num_bytes, u64 parent, u64 bytenr, u64 num_bytes, u64 parent,
...@@ -3412,7 +3491,7 @@ static int __btrfs_free_extent(struct btrfs_trans_handle *trans, ...@@ -3412,7 +3491,7 @@ static int __btrfs_free_extent(struct btrfs_trans_handle *trans,
} }
ret = pin_down_bytes(trans, root, path, bytenr, ret = pin_down_bytes(trans, root, path, bytenr,
num_bytes, is_data, &must_clean); num_bytes, is_data, 0, &must_clean);
if (ret > 0) if (ret > 0)
mark_free = 1; mark_free = 1;
BUG_ON(ret < 0); BUG_ON(ret < 0);
...@@ -3543,8 +3622,7 @@ int btrfs_free_extent(struct btrfs_trans_handle *trans, ...@@ -3543,8 +3622,7 @@ int btrfs_free_extent(struct btrfs_trans_handle *trans,
if (root_objectid == BTRFS_TREE_LOG_OBJECTID) { if (root_objectid == BTRFS_TREE_LOG_OBJECTID) {
WARN_ON(owner >= BTRFS_FIRST_FREE_OBJECTID); WARN_ON(owner >= BTRFS_FIRST_FREE_OBJECTID);
/* unlocks the pinned mutex */ /* unlocks the pinned mutex */
btrfs_update_pinned_extents(root, bytenr, num_bytes, 1); btrfs_pin_extent(root, bytenr, num_bytes, 1);
update_reserved_extents(root, bytenr, num_bytes, 0);
ret = 0; ret = 0;
} else if (owner < BTRFS_FIRST_FREE_OBJECTID) { } else if (owner < BTRFS_FIRST_FREE_OBJECTID) {
ret = btrfs_add_delayed_tree_ref(trans, bytenr, num_bytes, ret = btrfs_add_delayed_tree_ref(trans, bytenr, num_bytes,
...@@ -3584,19 +3662,33 @@ static noinline int ...@@ -3584,19 +3662,33 @@ static noinline int
wait_block_group_cache_progress(struct btrfs_block_group_cache *cache, wait_block_group_cache_progress(struct btrfs_block_group_cache *cache,
u64 num_bytes) u64 num_bytes)
{ {
struct btrfs_caching_control *caching_ctl;
DEFINE_WAIT(wait); DEFINE_WAIT(wait);
prepare_to_wait(&cache->caching_q, &wait, TASK_UNINTERRUPTIBLE); caching_ctl = get_caching_control(cache);
if (!caching_ctl)
if (block_group_cache_done(cache)) {
finish_wait(&cache->caching_q, &wait);
return 0; return 0;
}
schedule();
finish_wait(&cache->caching_q, &wait);
wait_event(cache->caching_q, block_group_cache_done(cache) || wait_event(caching_ctl->wait, block_group_cache_done(cache) ||
(cache->free_space >= num_bytes)); (cache->free_space >= num_bytes));
put_caching_control(caching_ctl);
return 0;
}
static noinline int
wait_block_group_cache_done(struct btrfs_block_group_cache *cache)
{
struct btrfs_caching_control *caching_ctl;
DEFINE_WAIT(wait);
caching_ctl = get_caching_control(cache);
if (!caching_ctl)
return 0;
wait_event(caching_ctl->wait, block_group_cache_done(cache));
put_caching_control(caching_ctl);
return 0; return 0;
} }
...@@ -3880,6 +3972,8 @@ static noinline int find_free_extent(struct btrfs_trans_handle *trans, ...@@ -3880,6 +3972,8 @@ static noinline int find_free_extent(struct btrfs_trans_handle *trans,
search_start - offset); search_start - offset);
BUG_ON(offset > search_start); BUG_ON(offset > search_start);
update_reserved_extents(block_group, num_bytes, 1);
/* we are all good, lets return */ /* we are all good, lets return */
break; break;
loop: loop:
...@@ -3972,12 +4066,12 @@ static void dump_space_info(struct btrfs_space_info *info, u64 bytes) ...@@ -3972,12 +4066,12 @@ static void dump_space_info(struct btrfs_space_info *info, u64 bytes)
up_read(&info->groups_sem); up_read(&info->groups_sem);
} }
static int __btrfs_reserve_extent(struct btrfs_trans_handle *trans, int btrfs_reserve_extent(struct btrfs_trans_handle *trans,
struct btrfs_root *root, struct btrfs_root *root,
u64 num_bytes, u64 min_alloc_size, u64 num_bytes, u64 min_alloc_size,
u64 empty_size, u64 hint_byte, u64 empty_size, u64 hint_byte,
u64 search_end, struct btrfs_key *ins, u64 search_end, struct btrfs_key *ins,
u64 data) u64 data)
{ {
int ret; int ret;
u64 search_start = 0; u64 search_start = 0;
...@@ -4043,25 +4137,8 @@ int btrfs_free_reserved_extent(struct btrfs_root *root, u64 start, u64 len) ...@@ -4043,25 +4137,8 @@ int btrfs_free_reserved_extent(struct btrfs_root *root, u64 start, u64 len)
ret = btrfs_discard_extent(root, start, len); ret = btrfs_discard_extent(root, start, len);
btrfs_add_free_space(cache, start, len); btrfs_add_free_space(cache, start, len);
update_reserved_extents(cache, len, 0);
btrfs_put_block_group(cache); btrfs_put_block_group(cache);
update_reserved_extents(root, start, len, 0);
return ret;
}
int btrfs_reserve_extent(struct btrfs_trans_handle *trans,
struct btrfs_root *root,
u64 num_bytes, u64 min_alloc_size,
u64 empty_size, u64 hint_byte,
u64 search_end, struct btrfs_key *ins,
u64 data)
{
int ret;
ret = __btrfs_reserve_extent(trans, root, num_bytes, min_alloc_size,
empty_size, hint_byte, search_end, ins,
data);
if (!ret)
update_reserved_extents(root, ins->objectid, ins->offset, 1);
return ret; return ret;
} }
...@@ -4222,15 +4299,46 @@ int btrfs_alloc_logged_file_extent(struct btrfs_trans_handle *trans, ...@@ -4222,15 +4299,46 @@ int btrfs_alloc_logged_file_extent(struct btrfs_trans_handle *trans,
{ {
int ret; int ret;
struct btrfs_block_group_cache *block_group; struct btrfs_block_group_cache *block_group;
struct btrfs_caching_control *caching_ctl;
u64 start = ins->objectid;
u64 num_bytes = ins->offset;
block_group = btrfs_lookup_block_group(root->fs_info, ins->objectid); block_group = btrfs_lookup_block_group(root->fs_info, ins->objectid);
cache_block_group(block_group); cache_block_group(block_group);
wait_event(block_group->caching_q, caching_ctl = get_caching_control(block_group);
block_group_cache_done(block_group));
ret = btrfs_remove_free_space(block_group, ins->objectid, if (!caching_ctl) {
ins->offset); BUG_ON(!block_group_cache_done(block_group));
BUG_ON(ret); ret = btrfs_remove_free_space(block_group, start, num_bytes);
BUG_ON(ret);
} else {
mutex_lock(&caching_ctl->mutex);
if (start >= caching_ctl->progress) {
ret = add_excluded_extent(root, start, num_bytes);
BUG_ON(ret);
} else if (start + num_bytes <= caching_ctl->progress) {
ret = btrfs_remove_free_space(block_group,
start, num_bytes);
BUG_ON(ret);
} else {
num_bytes = caching_ctl->progress - start;
ret = btrfs_remove_free_space(block_group,
start, num_bytes);
BUG_ON(ret);
start = caching_ctl->progress;
num_bytes = ins->objectid + ins->offset -
caching_ctl->progress;
ret = add_excluded_extent(root, start, num_bytes);
BUG_ON(ret);
}
mutex_unlock(&caching_ctl->mutex);
put_caching_control(caching_ctl);
}
update_reserved_extents(block_group, ins->offset, 1);
btrfs_put_block_group(block_group); btrfs_put_block_group(block_group);
ret = alloc_reserved_file_extent(trans, root, 0, root_objectid, ret = alloc_reserved_file_extent(trans, root, 0, root_objectid,
0, owner, offset, ins, 1); 0, owner, offset, ins, 1);
...@@ -4254,9 +4362,9 @@ static int alloc_tree_block(struct btrfs_trans_handle *trans, ...@@ -4254,9 +4362,9 @@ static int alloc_tree_block(struct btrfs_trans_handle *trans,
int ret; int ret;
u64 flags = 0; u64 flags = 0;
ret = __btrfs_reserve_extent(trans, root, num_bytes, num_bytes, ret = btrfs_reserve_extent(trans, root, num_bytes, num_bytes,
empty_size, hint_byte, search_end, empty_size, hint_byte, search_end,
ins, 0); ins, 0);
if (ret) if (ret)
return ret; return ret;
...@@ -4267,7 +4375,6 @@ static int alloc_tree_block(struct btrfs_trans_handle *trans, ...@@ -4267,7 +4375,6 @@ static int alloc_tree_block(struct btrfs_trans_handle *trans,
} else } else
BUG_ON(parent > 0); BUG_ON(parent > 0);
update_reserved_extents(root, ins->objectid, ins->offset, 1);
if (root_objectid != BTRFS_TREE_LOG_OBJECTID) { if (root_objectid != BTRFS_TREE_LOG_OBJECTID) {
struct btrfs_delayed_extent_op *extent_op; struct btrfs_delayed_extent_op *extent_op;
extent_op = kmalloc(sizeof(*extent_op), GFP_NOFS); extent_op = kmalloc(sizeof(*extent_op), GFP_NOFS);
...@@ -7164,8 +7271,18 @@ int btrfs_free_block_groups(struct btrfs_fs_info *info) ...@@ -7164,8 +7271,18 @@ int btrfs_free_block_groups(struct btrfs_fs_info *info)
{ {
struct btrfs_block_group_cache *block_group; struct btrfs_block_group_cache *block_group;
struct btrfs_space_info *space_info; struct btrfs_space_info *space_info;
struct btrfs_caching_control *caching_ctl;
struct rb_node *n; struct rb_node *n;
down_write(&info->extent_commit_sem);
while (!list_empty(&info->caching_block_groups)) {
caching_ctl = list_entry(info->caching_block_groups.next,
struct btrfs_caching_control, list);
list_del(&caching_ctl->list);
put_caching_control(caching_ctl);
}
up_write(&info->extent_commit_sem);
spin_lock(&info->block_group_cache_lock); spin_lock(&info->block_group_cache_lock);
while ((n = rb_last(&info->block_group_cache_tree)) != NULL) { while ((n = rb_last(&info->block_group_cache_tree)) != NULL) {
block_group = rb_entry(n, struct btrfs_block_group_cache, block_group = rb_entry(n, struct btrfs_block_group_cache,
...@@ -7179,8 +7296,7 @@ int btrfs_free_block_groups(struct btrfs_fs_info *info) ...@@ -7179,8 +7296,7 @@ int btrfs_free_block_groups(struct btrfs_fs_info *info)
up_write(&block_group->space_info->groups_sem); up_write(&block_group->space_info->groups_sem);
if (block_group->cached == BTRFS_CACHE_STARTED) if (block_group->cached == BTRFS_CACHE_STARTED)
wait_event(block_group->caching_q, wait_block_group_cache_done(block_group);
block_group_cache_done(block_group));
btrfs_remove_free_space_cache(block_group); btrfs_remove_free_space_cache(block_group);
...@@ -7250,7 +7366,6 @@ int btrfs_read_block_groups(struct btrfs_root *root) ...@@ -7250,7 +7366,6 @@ int btrfs_read_block_groups(struct btrfs_root *root)
spin_lock_init(&cache->lock); spin_lock_init(&cache->lock);
spin_lock_init(&cache->tree_lock); spin_lock_init(&cache->tree_lock);
cache->fs_info = info; cache->fs_info = info;
init_waitqueue_head(&cache->caching_q);
INIT_LIST_HEAD(&cache->list); INIT_LIST_HEAD(&cache->list);
INIT_LIST_HEAD(&cache->cluster_list); INIT_LIST_HEAD(&cache->cluster_list);
...@@ -7272,8 +7387,6 @@ int btrfs_read_block_groups(struct btrfs_root *root) ...@@ -7272,8 +7387,6 @@ int btrfs_read_block_groups(struct btrfs_root *root)
cache->flags = btrfs_block_group_flags(&cache->item); cache->flags = btrfs_block_group_flags(&cache->item);
cache->sectorsize = root->sectorsize; cache->sectorsize = root->sectorsize;
remove_sb_from_cache(root, cache);
/* /*
* check for two cases, either we are full, and therefore * check for two cases, either we are full, and therefore
* don't need to bother with the caching work since we won't * don't need to bother with the caching work since we won't
...@@ -7282,13 +7395,17 @@ int btrfs_read_block_groups(struct btrfs_root *root) ...@@ -7282,13 +7395,17 @@ int btrfs_read_block_groups(struct btrfs_root *root)
* time, particularly in the full case. * time, particularly in the full case.
*/ */
if (found_key.offset == btrfs_block_group_used(&cache->item)) { if (found_key.offset == btrfs_block_group_used(&cache->item)) {
cache->last_byte_to_unpin = (u64)-1;
cache->cached = BTRFS_CACHE_FINISHED; cache->cached = BTRFS_CACHE_FINISHED;
} else if (btrfs_block_group_used(&cache->item) == 0) { } else if (btrfs_block_group_used(&cache->item) == 0) {
exclude_super_stripes(root, cache);
cache->last_byte_to_unpin = (u64)-1;
cache->cached = BTRFS_CACHE_FINISHED; cache->cached = BTRFS_CACHE_FINISHED;
add_new_free_space(cache, root->fs_info, add_new_free_space(cache, root->fs_info,
found_key.objectid, found_key.objectid,
found_key.objectid + found_key.objectid +
found_key.offset); found_key.offset);
free_excluded_extents(root, cache);
} }
ret = update_space_info(info, cache->flags, found_key.offset, ret = update_space_info(info, cache->flags, found_key.offset,
...@@ -7345,7 +7462,6 @@ int btrfs_make_block_group(struct btrfs_trans_handle *trans, ...@@ -7345,7 +7462,6 @@ int btrfs_make_block_group(struct btrfs_trans_handle *trans,
atomic_set(&cache->count, 1); atomic_set(&cache->count, 1);
spin_lock_init(&cache->lock); spin_lock_init(&cache->lock);
spin_lock_init(&cache->tree_lock); spin_lock_init(&cache->tree_lock);
init_waitqueue_head(&cache->caching_q);
INIT_LIST_HEAD(&cache->list); INIT_LIST_HEAD(&cache->list);
INIT_LIST_HEAD(&cache->cluster_list); INIT_LIST_HEAD(&cache->cluster_list);
...@@ -7354,12 +7470,15 @@ int btrfs_make_block_group(struct btrfs_trans_handle *trans, ...@@ -7354,12 +7470,15 @@ int btrfs_make_block_group(struct btrfs_trans_handle *trans,
cache->flags = type; cache->flags = type;
btrfs_set_block_group_flags(&cache->item, type); btrfs_set_block_group_flags(&cache->item, type);
cache->last_byte_to_unpin = (u64)-1;
cache->cached = BTRFS_CACHE_FINISHED; cache->cached = BTRFS_CACHE_FINISHED;
remove_sb_from_cache(root, cache); exclude_super_stripes(root, cache);
add_new_free_space(cache, root->fs_info, chunk_offset, add_new_free_space(cache, root->fs_info, chunk_offset,
chunk_offset + size); chunk_offset + size);
free_excluded_extents(root, cache);
ret = update_space_info(root->fs_info, cache->flags, size, bytes_used, ret = update_space_info(root->fs_info, cache->flags, size, bytes_used,
&cache->space_info); &cache->space_info);
BUG_ON(ret); BUG_ON(ret);
...@@ -7428,8 +7547,7 @@ int btrfs_remove_block_group(struct btrfs_trans_handle *trans, ...@@ -7428,8 +7547,7 @@ int btrfs_remove_block_group(struct btrfs_trans_handle *trans,
up_write(&block_group->space_info->groups_sem); up_write(&block_group->space_info->groups_sem);
if (block_group->cached == BTRFS_CACHE_STARTED) if (block_group->cached == BTRFS_CACHE_STARTED)
wait_event(block_group->caching_q, wait_block_group_cache_done(block_group);
block_group_cache_done(block_group));
btrfs_remove_free_space_cache(block_group); btrfs_remove_free_space_cache(block_group);
......
...@@ -874,7 +874,6 @@ int btrfs_commit_transaction(struct btrfs_trans_handle *trans, ...@@ -874,7 +874,6 @@ int btrfs_commit_transaction(struct btrfs_trans_handle *trans,
unsigned long timeout = 1; unsigned long timeout = 1;
struct btrfs_transaction *cur_trans; struct btrfs_transaction *cur_trans;
struct btrfs_transaction *prev_trans = NULL; struct btrfs_transaction *prev_trans = NULL;
struct extent_io_tree *pinned_copy;
DEFINE_WAIT(wait); DEFINE_WAIT(wait);
int ret; int ret;
int should_grow = 0; int should_grow = 0;
...@@ -915,13 +914,6 @@ int btrfs_commit_transaction(struct btrfs_trans_handle *trans, ...@@ -915,13 +914,6 @@ int btrfs_commit_transaction(struct btrfs_trans_handle *trans,
return 0; return 0;
} }
pinned_copy = kmalloc(sizeof(*pinned_copy), GFP_NOFS);
if (!pinned_copy)
return -ENOMEM;
extent_io_tree_init(pinned_copy,
root->fs_info->btree_inode->i_mapping, GFP_NOFS);
trans->transaction->in_commit = 1; trans->transaction->in_commit = 1;
trans->transaction->blocked = 1; trans->transaction->blocked = 1;
if (cur_trans->list.prev != &root->fs_info->trans_list) { if (cur_trans->list.prev != &root->fs_info->trans_list) {
...@@ -1019,6 +1011,8 @@ int btrfs_commit_transaction(struct btrfs_trans_handle *trans, ...@@ -1019,6 +1011,8 @@ int btrfs_commit_transaction(struct btrfs_trans_handle *trans,
ret = commit_cowonly_roots(trans, root); ret = commit_cowonly_roots(trans, root);
BUG_ON(ret); BUG_ON(ret);
btrfs_prepare_extent_commit(trans, root);
cur_trans = root->fs_info->running_transaction; cur_trans = root->fs_info->running_transaction;
spin_lock(&root->fs_info->new_trans_lock); spin_lock(&root->fs_info->new_trans_lock);
root->fs_info->running_transaction = NULL; root->fs_info->running_transaction = NULL;
...@@ -1042,8 +1036,6 @@ int btrfs_commit_transaction(struct btrfs_trans_handle *trans, ...@@ -1042,8 +1036,6 @@ int btrfs_commit_transaction(struct btrfs_trans_handle *trans,
memcpy(&root->fs_info->super_for_commit, &root->fs_info->super_copy, memcpy(&root->fs_info->super_for_commit, &root->fs_info->super_copy,
sizeof(root->fs_info->super_copy)); sizeof(root->fs_info->super_copy));
btrfs_copy_pinned(root, pinned_copy);
trans->transaction->blocked = 0; trans->transaction->blocked = 0;
wake_up(&root->fs_info->transaction_wait); wake_up(&root->fs_info->transaction_wait);
...@@ -1059,8 +1051,7 @@ int btrfs_commit_transaction(struct btrfs_trans_handle *trans, ...@@ -1059,8 +1051,7 @@ int btrfs_commit_transaction(struct btrfs_trans_handle *trans,
*/ */
mutex_unlock(&root->fs_info->tree_log_mutex); mutex_unlock(&root->fs_info->tree_log_mutex);
btrfs_finish_extent_commit(trans, root, pinned_copy); btrfs_finish_extent_commit(trans, root);
kfree(pinned_copy);
/* do the directory inserts of any pending snapshot creations */ /* do the directory inserts of any pending snapshot creations */
finish_pending_snapshots(trans, root->fs_info); finish_pending_snapshots(trans, root->fs_info);
......
...@@ -263,8 +263,8 @@ static int process_one_buffer(struct btrfs_root *log, ...@@ -263,8 +263,8 @@ static int process_one_buffer(struct btrfs_root *log,
struct walk_control *wc, u64 gen) struct walk_control *wc, u64 gen)
{ {
if (wc->pin) if (wc->pin)
btrfs_update_pinned_extents(log->fs_info->extent_root, btrfs_pin_extent(log->fs_info->extent_root,
eb->start, eb->len, 1); eb->start, eb->len, 0);
if (btrfs_buffer_uptodate(eb, gen)) { if (btrfs_buffer_uptodate(eb, gen)) {
if (wc->write) if (wc->write)
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment