#include <linux/fs.h>
#include <linux/sched.h>
+#include <linux/writeback.h>
+#include <linux/pagemap.h>
#include "ctree.h"
#include "disk-io.h"
#include "transaction.h"
BUG_ON(!cur_trans);
root->fs_info->generation++;
root->fs_info->running_transaction = cur_trans;
- cur_trans->num_writers = 0;
+ cur_trans->num_writers = 1;
+ cur_trans->num_joined = 0;
cur_trans->transid = root->fs_info->generation;
init_waitqueue_head(&cur_trans->writer_wait);
init_waitqueue_head(&cur_trans->commit_wait);
cur_trans->commit_done = 0;
cur_trans->start_time = get_seconds();
list_add_tail(&cur_trans->list, &root->fs_info->trans_list);
- init_bit_radix(&cur_trans->dirty_pages);
+ extent_map_tree_init(&cur_trans->dirty_pages,
+ root->fs_info->btree_inode->i_mapping,
+ GFP_NOFS);
+ } else {
+ cur_trans->num_writers++;
+ cur_trans->num_joined++;
}
- cur_trans->num_writers++;
+
return 0;
}
(unsigned long)root->root_key.objectid,
BTRFS_ROOT_DEFRAG_TAG);
root->commit_root = root->node;
- get_bh(root->node);
+ extent_buffer_get(root->node);
} else {
WARN_ON(1);
}
int btrfs_write_and_wait_transaction(struct btrfs_trans_handle *trans,
struct btrfs_root *root)
{
- unsigned long gang[16];
int ret;
- int i;
int err;
int werr = 0;
+ struct extent_map_tree *dirty_pages;
struct page *page;
- struct radix_tree_root *dirty_pages;
struct inode *btree_inode = root->fs_info->btree_inode;
+ u64 start;
+ u64 end;
+ unsigned long index;
if (!trans || !trans->transaction) {
return filemap_write_and_wait(btree_inode->i_mapping);
}
dirty_pages = &trans->transaction->dirty_pages;
while(1) {
- ret = find_first_radix_bit(dirty_pages, gang,
- 0, ARRAY_SIZE(gang));
- if (!ret)
+ ret = find_first_extent_bit(dirty_pages, 0, &start, &end,
+ EXTENT_DIRTY);
+ if (ret)
break;
- for (i = 0; i < ret; i++) {
- /* FIXME EIO */
- clear_radix_bit(dirty_pages, gang[i]);
- page = find_lock_page(btree_inode->i_mapping,
- gang[i]);
+ clear_extent_dirty(dirty_pages, start, end, GFP_NOFS);
+ while(start <= end) {
+ index = start >> PAGE_CACHE_SHIFT;
+ start = (u64)(index + 1) << PAGE_CACHE_SHIFT;
+ page = find_lock_page(btree_inode->i_mapping, index);
if (!page)
continue;
if (PageWriteback(page)) {
btrfs_write_dirty_block_groups(trans, extent_root);
while(1) {
- old_extent_block = btrfs_root_blocknr(&extent_root->root_item);
- if (old_extent_block == bh_blocknr(extent_root->node))
+ old_extent_block = btrfs_root_bytenr(&extent_root->root_item);
+ if (old_extent_block == extent_root->node->start)
break;
- btrfs_set_root_blocknr(&extent_root->root_item,
- bh_blocknr(extent_root->node));
+ btrfs_set_root_bytenr(&extent_root->root_item,
+ extent_root->node->start);
+ btrfs_set_root_level(&extent_root->root_item,
+ btrfs_header_level(extent_root->node));
ret = btrfs_update_root(trans, tree_root,
&extent_root->root_key,
&extent_root->root_item);
struct dirty_root {
struct list_head list;
struct btrfs_root *root;
+ struct btrfs_root *latest_root;
};
-int btrfs_add_dead_root(struct btrfs_root *root, struct list_head *dead_list)
+int btrfs_add_dead_root(struct btrfs_root *root,
+ struct btrfs_root *latest,
+ struct list_head *dead_list)
{
struct dirty_root *dirty;
if (!dirty)
return -ENOMEM;
dirty->root = root;
+ dirty->latest_root = latest;
list_add(&dirty->list, dead_list);
return 0;
}
(unsigned long)root->root_key.objectid,
BTRFS_ROOT_TRANS_TAG);
if (root->commit_root == root->node) {
- WARN_ON(bh_blocknr(root->node) !=
- btrfs_root_blocknr(&root->root_item));
- brelse(root->commit_root);
+ WARN_ON(root->node->start !=
+ btrfs_root_bytenr(&root->root_item));
+ free_extent_buffer(root->commit_root);
root->commit_root = NULL;
+
+ /* make sure to update the root on disk
+ * so we get any updates to the block used
+ * counts
+ */
+ err = btrfs_update_root(trans,
+ root->fs_info->tree_root,
+ &root->root_key,
+ &root->root_item);
continue;
}
dirty = kmalloc(sizeof(*dirty), GFP_NOFS);
memcpy(dirty->root, root, sizeof(*root));
dirty->root->node = root->commit_root;
+ dirty->latest_root = root;
root->commit_root = NULL;
root->root_key.offset = root->fs_info->generation;
- btrfs_set_root_blocknr(&root->root_item,
- bh_blocknr(root->node));
+ btrfs_set_root_bytenr(&root->root_item,
+ root->node->start);
+ btrfs_set_root_level(&root->root_item,
+ btrfs_header_level(root->node));
err = btrfs_insert_root(trans, root->fs_info->tree_root,
&root->root_key,
&root->root_item);
return err;
}
+int btrfs_defrag_root(struct btrfs_root *root, int cacheonly)
+{
+ struct btrfs_fs_info *info = root->fs_info;
+ int ret;
+ struct btrfs_trans_handle *trans;
+ unsigned long nr;
+
+ if (root->defrag_running)
+ return 0;
+ trans = btrfs_start_transaction(root, 1);
+ while (1) {
+ root->defrag_running = 1;
+ ret = btrfs_defrag_leaves(trans, root, cacheonly);
+ nr = trans->blocks_used;
+ btrfs_end_transaction(trans, root);
+ mutex_unlock(&info->fs_mutex);
+ btrfs_btree_balance_dirty(info->tree_root, nr);
+ cond_resched();
+
+ mutex_lock(&info->fs_mutex);
+ trans = btrfs_start_transaction(root, 1);
+ if (ret != -EAGAIN)
+ break;
+ }
+ root->defrag_running = 0;
+ radix_tree_tag_clear(&info->fs_roots_radix,
+ (unsigned long)root->root_key.objectid,
+ BTRFS_ROOT_DEFRAG_TAG);
+ btrfs_end_transaction(trans, root);
+ return 0;
+}
+
int btrfs_defrag_dirty_roots(struct btrfs_fs_info *info)
{
struct btrfs_root *gang[1];
struct btrfs_root *root;
- struct btrfs_root *tree_root = info->tree_root;
- struct btrfs_trans_handle *trans;
int i;
int ret;
int err = 0;
u64 last = 0;
- trans = btrfs_start_transaction(tree_root, 1);
while(1) {
ret = radix_tree_gang_lookup_tag(&info->fs_roots_radix,
(void **)gang, last,
for (i = 0; i < ret; i++) {
root = gang[i];
last = root->root_key.objectid + 1;
- radix_tree_tag_clear(&info->fs_roots_radix,
- (unsigned long)root->root_key.objectid,
- BTRFS_ROOT_DEFRAG_TAG);
- if (root->defrag_running)
- continue;
-
- while (1) {
- mutex_lock(&root->fs_info->trans_mutex);
- record_root_in_trans(root);
- mutex_unlock(&root->fs_info->trans_mutex);
-
- root->defrag_running = 1;
- err = btrfs_defrag_leaves(trans, root, 1);
- btrfs_end_transaction(trans, tree_root);
- mutex_unlock(&info->fs_mutex);
-
- btrfs_btree_balance_dirty(root);
- cond_resched();
-
- mutex_lock(&info->fs_mutex);
- trans = btrfs_start_transaction(tree_root, 1);
- if (err != -EAGAIN)
- break;
- }
- root->defrag_running = 0;
- radix_tree_tag_clear(&info->fs_roots_radix,
- (unsigned long)root->root_key.objectid,
- BTRFS_ROOT_DEFRAG_TAG);
+ btrfs_defrag_root(root, 1);
}
}
- btrfs_end_transaction(trans, tree_root);
+ btrfs_defrag_root(info->extent_root, 1);
return err;
}
{
struct dirty_root *dirty;
struct btrfs_trans_handle *trans;
+ unsigned long nr;
+ u64 num_bytes;
+ u64 bytes_used;
int ret = 0;
int err;
while(!list_empty(list)) {
+ struct btrfs_root *root;
+
mutex_lock(&tree_root->fs_info->fs_mutex);
dirty = list_entry(list->next, struct dirty_root, list);
list_del_init(&dirty->list);
+ num_bytes = btrfs_root_used(&dirty->root->root_item);
+ root = dirty->latest_root;
+
while(1) {
trans = btrfs_start_transaction(tree_root, 1);
ret = btrfs_drop_snapshot(trans, dirty->root);
if (ret != -EAGAIN) {
break;
}
+
err = btrfs_update_root(trans,
tree_root,
&dirty->root->root_key,
&dirty->root->root_item);
if (err)
ret = err;
+ nr = trans->blocks_used;
ret = btrfs_end_transaction(trans, tree_root);
BUG_ON(ret);
mutex_unlock(&tree_root->fs_info->fs_mutex);
-
- btrfs_btree_balance_dirty(tree_root);
- schedule();
-
+ btrfs_btree_balance_dirty(tree_root, nr);
+ cond_resched();
mutex_lock(&tree_root->fs_info->fs_mutex);
}
BUG_ON(ret);
+
+ num_bytes -= btrfs_root_used(&dirty->root->root_item);
+ bytes_used = btrfs_root_used(&root->root_item);
+ if (num_bytes) {
+ record_root_in_trans(root);
+ btrfs_set_root_used(&root->root_item,
+ bytes_used - num_bytes);
+ }
ret = btrfs_del_root(trans, tree_root, &dirty->root->root_key);
- if (ret)
+ if (ret) {
+ BUG();
break;
+ }
+ nr = trans->blocks_used;
ret = btrfs_end_transaction(trans, tree_root);
BUG_ON(ret);
+ free_extent_buffer(dirty->root->node);
kfree(dirty->root);
kfree(dirty);
mutex_unlock(&tree_root->fs_info->fs_mutex);
- btrfs_btree_balance_dirty(tree_root);
- schedule();
+
+ btrfs_btree_balance_dirty(tree_root, nr);
+ cond_resched();
}
return ret;
}
int btrfs_commit_transaction(struct btrfs_trans_handle *trans,
struct btrfs_root *root)
{
- int ret = 0;
+ unsigned long joined = 0;
+ unsigned long timeout = 1;
struct btrfs_transaction *cur_trans;
struct btrfs_transaction *prev_trans = NULL;
struct list_head dirty_fs_roots;
- struct radix_tree_root pinned_copy;
+ struct extent_map_tree pinned_copy;
DEFINE_WAIT(wait);
+ int ret;
- init_bit_radix(&pinned_copy);
+ extent_map_tree_init(&pinned_copy,
+ root->fs_info->btree_inode->i_mapping, GFP_NOFS);
INIT_LIST_HEAD(&dirty_fs_roots);
mutex_lock(&root->fs_info->trans_mutex);
mutex_unlock(&root->fs_info->fs_mutex);
ret = wait_for_commit(root, cur_trans);
BUG_ON(ret);
+
+ mutex_lock(&root->fs_info->trans_mutex);
put_transaction(cur_trans);
+ mutex_unlock(&root->fs_info->trans_mutex);
+
mutex_lock(&root->fs_info->fs_mutex);
return 0;
}
mutex_unlock(&root->fs_info->trans_mutex);
wait_for_commit(root, prev_trans);
- put_transaction(prev_trans);
mutex_lock(&root->fs_info->fs_mutex);
mutex_lock(&root->fs_info->trans_mutex);
+ put_transaction(prev_trans);
}
}
- while (trans->transaction->num_writers > 1) {
+
+ do {
+ joined = cur_trans->num_joined;
WARN_ON(cur_trans != trans->transaction);
- prepare_to_wait(&trans->transaction->writer_wait, &wait,
+ prepare_to_wait(&cur_trans->writer_wait, &wait,
TASK_UNINTERRUPTIBLE);
- if (trans->transaction->num_writers <= 1)
- break;
+
+ if (cur_trans->num_writers > 1)
+ timeout = MAX_SCHEDULE_TIMEOUT;
+ else
+ timeout = 1;
+
mutex_unlock(&root->fs_info->fs_mutex);
mutex_unlock(&root->fs_info->trans_mutex);
- schedule();
+
+ schedule_timeout(timeout);
+
mutex_lock(&root->fs_info->fs_mutex);
mutex_lock(&root->fs_info->trans_mutex);
- finish_wait(&trans->transaction->writer_wait, &wait);
- }
- finish_wait(&trans->transaction->writer_wait, &wait);
+ finish_wait(&cur_trans->writer_wait, &wait);
+ } while (cur_trans->num_writers > 1 ||
+ (cur_trans->num_joined != joined));
+
WARN_ON(cur_trans != trans->transaction);
ret = add_dirty_roots(trans, &root->fs_info->fs_roots_radix,
&dirty_fs_roots);
btrfs_set_super_generation(&root->fs_info->super_copy,
cur_trans->transid);
btrfs_set_super_root(&root->fs_info->super_copy,
- bh_blocknr(root->fs_info->tree_root->node));
- memcpy(root->fs_info->disk_super, &root->fs_info->super_copy,
- sizeof(root->fs_info->super_copy));
+ root->fs_info->tree_root->node->start);
+ btrfs_set_super_root_level(&root->fs_info->super_copy,
+ btrfs_header_level(root->fs_info->tree_root->node));
+
+ write_extent_buffer(root->fs_info->sb_buffer,
+ &root->fs_info->super_copy, 0,
+ sizeof(root->fs_info->super_copy));
btrfs_copy_pinned(root, &pinned_copy);
btrfs_finish_extent_commit(trans, root, &pinned_copy);
mutex_lock(&root->fs_info->trans_mutex);
cur_trans->commit_done = 1;
+ root->fs_info->last_trans_committed = cur_trans->transid;
wake_up(&cur_trans->commit_wait);
put_transaction(cur_trans);
put_transaction(cur_trans);
+
if (root->fs_info->closing)
list_splice_init(&root->fs_info->dead_roots, &dirty_fs_roots);
else
list_splice_init(&dirty_fs_roots, &root->fs_info->dead_roots);
+
mutex_unlock(&root->fs_info->trans_mutex);
kmem_cache_free(btrfs_trans_handle_cachep, trans);
return ret;
}
+int btrfs_clean_old_snapshots(struct btrfs_root *root)
+{
+ struct list_head dirty_roots;
+ INIT_LIST_HEAD(&dirty_roots);
+
+ mutex_lock(&root->fs_info->trans_mutex);
+ list_splice_init(&root->fs_info->dead_roots, &dirty_roots);
+ mutex_unlock(&root->fs_info->trans_mutex);
+
+ if (!list_empty(&dirty_roots)) {
+ drop_dirty_roots(root, &dirty_roots);
+ }
+ return 0;
+}
+#if LINUX_VERSION_CODE <= KERNEL_VERSION(2,6,18)
+void btrfs_transaction_cleaner(void *p)
+#else
void btrfs_transaction_cleaner(struct work_struct *work)
+#endif
{
+#if LINUX_VERSION_CODE <= KERNEL_VERSION(2,6,18)
+ struct btrfs_fs_info *fs_info = p;
+#else
struct btrfs_fs_info *fs_info = container_of(work,
struct btrfs_fs_info,
trans_work.work);
+#endif
struct btrfs_root *root = fs_info->tree_root;
struct btrfs_transaction *cur;
struct btrfs_trans_handle *trans;
- struct list_head dirty_roots;
unsigned long now;
unsigned long delay = HZ * 30;
int ret;
- INIT_LIST_HEAD(&dirty_roots);
mutex_lock(&root->fs_info->fs_mutex);
mutex_lock(&root->fs_info->trans_mutex);
cur = root->fs_info->running_transaction;
ret = btrfs_commit_transaction(trans, root);
out:
mutex_unlock(&root->fs_info->fs_mutex);
-
- mutex_lock(&root->fs_info->trans_mutex);
- list_splice_init(&root->fs_info->dead_roots, &dirty_roots);
- mutex_unlock(&root->fs_info->trans_mutex);
-
- if (!list_empty(&dirty_roots)) {
- drop_dirty_roots(root, &dirty_roots);
- }
+ btrfs_clean_old_snapshots(root);
btrfs_transaction_queue_work(root, delay);
}
trans_wq = create_workqueue("btrfs");
}
-void __exit btrfs_exit_transaction_sys(void)
+void btrfs_exit_transaction_sys(void)
{
destroy_workqueue(trans_wq);
}