aboutsummaryrefslogtreecommitdiff
path: root/fs/btrfs/disk-io.c
diff options
context:
space:
mode:
authorChris Mason <chris.mason@oracle.com>2008-04-18 16:11:30 -0400
committerChris Mason <chris.mason@oracle.com>2008-09-25 11:04:02 -0400
commit9ad6b7bc2e00ba02f915cffd5b6bcd6564bb2c75 (patch)
tree8a07ab45a145ac723bdfe011a187d09eb0cc2803 /fs/btrfs/disk-io.c
parent0afbaf8c8216a0de64b55c2d455de2b6c9dde5eb (diff)
Force page->private removal in btrfs_invalidatepage
btrfs_invalidatepage is not allowed to leave pages around on the lru. Any such pages will trigger an oops later on because the VM will see page->private and assume it is a buffer head. This also forces extra flushes of the async work queues before dropping all the pages on the btree inode during unmount. Left over items on the work queues are one possible cause of busy state ranges during truncate_inode_pages. Signed-off-by: Chris Mason <chris.mason@oracle.com>
Diffstat (limited to 'fs/btrfs/disk-io.c')
-rw-r--r--fs/btrfs/disk-io.c12
1 files changed, 12 insertions, 0 deletions
diff --git a/fs/btrfs/disk-io.c b/fs/btrfs/disk-io.c
index 9d648f2f41a..b479cdfe3ee 100644
--- a/fs/btrfs/disk-io.c
+++ b/fs/btrfs/disk-io.c
@@ -533,6 +533,12 @@ static void btree_invalidatepage(struct page *page, unsigned long offset)
tree = &BTRFS_I(page->mapping->host)->io_tree;
extent_invalidatepage(tree, page, offset);
btree_releasepage(page, GFP_NOFS);
+ if (PagePrivate(page)) {
+ printk("2invalidate page cleaning up after releasepage\n");
+ ClearPagePrivate(page);
+ set_page_private(page, 0);
+ page_cache_release(page);
+ }
}
#if 0
@@ -1484,6 +1490,8 @@ int close_ctree(struct btrfs_root *root)
write_ctree_super(NULL, root);
mutex_unlock(&fs_info->fs_mutex);
+ btrfs_transaction_flush_work(root);
+
if (fs_info->delalloc_bytes) {
printk("btrfs: at unmount delalloc count %Lu\n",
fs_info->delalloc_bytes);
@@ -1514,7 +1522,11 @@ int close_ctree(struct btrfs_root *root)
extent_io_tree_empty_lru(&fs_info->extent_ins);
extent_io_tree_empty_lru(&BTRFS_I(fs_info->btree_inode)->io_tree);
+ flush_workqueue(end_io_workqueue);
+ flush_workqueue(async_submit_workqueue);
+
truncate_inode_pages(fs_info->btree_inode->i_mapping, 0);
+
flush_workqueue(end_io_workqueue);
destroy_workqueue(end_io_workqueue);