clear_extent_dirty(dirty_pages, start, end, GFP_NOFS);
while(start <= end) {
index = start >> PAGE_CACHE_SHIFT;
- start = (index + 1) << PAGE_CACHE_SHIFT;
+ start = (u64)(index + 1) << PAGE_CACHE_SHIFT;
page = find_lock_page(btree_inode->i_mapping, index);
if (!page)
continue;
if (root->defrag_running)
return 0;
-
trans = btrfs_start_transaction(root, 1);
- while (0) {
+ while (1) {
root->defrag_running = 1;
ret = btrfs_defrag_leaves(trans, root, cacheonly);
nr = trans->blocks_used;
btrfs_end_transaction(trans, root);
mutex_unlock(&info->fs_mutex);
-
btrfs_btree_balance_dirty(info->tree_root, nr);
cond_resched();
btrfs_defrag_root(root, 1);
}
}
- // btrfs_defrag_root(info->extent_root, 1);
+ btrfs_defrag_root(info->extent_root, 1);
return err;
}
BUG_ON(ret);
mutex_unlock(&tree_root->fs_info->fs_mutex);
btrfs_btree_balance_dirty(tree_root, nr);
- schedule();
-
+ cond_resched();
mutex_lock(&tree_root->fs_info->fs_mutex);
}
BUG_ON(ret);
mutex_unlock(&tree_root->fs_info->fs_mutex);
btrfs_btree_balance_dirty(tree_root, nr);
- schedule();
+ cond_resched();
}
return ret;
}
struct btrfs_transaction *cur_trans;
struct btrfs_transaction *prev_trans = NULL;
struct list_head dirty_fs_roots;
- struct extent_map_tree pinned_copy;
+ struct extent_map_tree *pinned_copy;
DEFINE_WAIT(wait);
int ret;
- extent_map_tree_init(&pinned_copy,
- root->fs_info->btree_inode->i_mapping, GFP_NOFS);
INIT_LIST_HEAD(&dirty_fs_roots);
mutex_lock(&root->fs_info->trans_mutex);
mutex_lock(&root->fs_info->fs_mutex);
return 0;
}
+
+ pinned_copy = kmalloc(sizeof(*pinned_copy), GFP_NOFS);
+ if (!pinned_copy)
+ return -ENOMEM;
+
+ extent_map_tree_init(pinned_copy,
+ root->fs_info->btree_inode->i_mapping, GFP_NOFS);
+
trans->transaction->in_commit = 1;
cur_trans = trans->transaction;
if (cur_trans->list.prev != &root->fs_info->trans_list) {
&root->fs_info->super_copy, 0,
sizeof(root->fs_info->super_copy));
- btrfs_copy_pinned(root, &pinned_copy);
+ btrfs_copy_pinned(root, pinned_copy);
mutex_unlock(&root->fs_info->trans_mutex);
mutex_unlock(&root->fs_info->fs_mutex);
ret = btrfs_write_and_wait_transaction(trans, root);
BUG_ON(ret);
write_ctree_super(trans, root);
+
mutex_lock(&root->fs_info->fs_mutex);
- btrfs_finish_extent_commit(trans, root, &pinned_copy);
+ btrfs_finish_extent_commit(trans, root, pinned_copy);
mutex_lock(&root->fs_info->trans_mutex);
+
+ kfree(pinned_copy);
+
cur_trans->commit_done = 1;
root->fs_info->last_trans_committed = cur_trans->transid;
wake_up(&cur_trans->commit_wait);
}
return 0;
}
+#if LINUX_VERSION_CODE <= KERNEL_VERSION(2,6,18)
+void btrfs_transaction_cleaner(void *p)
+#else
void btrfs_transaction_cleaner(struct work_struct *work)
+#endif
{
+#if LINUX_VERSION_CODE <= KERNEL_VERSION(2,6,18)
+ struct btrfs_fs_info *fs_info = p;
+#else
struct btrfs_fs_info *fs_info = container_of(work,
struct btrfs_fs_info,
trans_work.work);
+#endif
struct btrfs_root *root = fs_info->tree_root;
struct btrfs_transaction *cur;
struct btrfs_trans_handle *trans;
trans_wq = create_workqueue("btrfs");
}
-void __exit btrfs_exit_transaction_sys(void)
+void btrfs_exit_transaction_sys(void)
{
destroy_workqueue(trans_wq);
}