mirror of
https://github.com/Fishwaldo/Star64_linux.git
synced 2025-07-23 23:32:14 +00:00
Merge git://git.kernel.org/pub/scm/linux/kernel/git/mason/btrfs-unstable
* git://git.kernel.org/pub/scm/linux/kernel/git/mason/btrfs-unstable: Btrfs: try to free metadata pages when we free btree blocks Btrfs: add extra flushing for renames and truncates Btrfs: make sure btrfs_update_delayed_ref doesn't increase ref_mod Btrfs: optimize fsyncs on old files Btrfs: tree logging unlink/rename fixes Btrfs: Make sure i_nlink doesn't hit zero too soon during log replay Btrfs: limit balancing work while flushing delayed refs Btrfs: readahead checksums during btrfs_finish_ordered_io Btrfs: leave btree locks spinning more often Btrfs: Only let very young transactions grow during commit Btrfs: Check for a blocking lock before taking the spin Btrfs: reduce stack in cow_file_range Btrfs: reduce stalls during transaction commit Btrfs: process the delayed reference queue in clusters Btrfs: try to cleanup delayed refs while freeing extents Btrfs: reduce stack usage in some crucial tree balancing functions Btrfs: do extent allocation and reference count updates in the background Btrfs: don't preallocate metadata blocks during btrfs_search_slot
This commit is contained in:
commit
c226fd659f
24 changed files with 2828 additions and 1688 deletions
194
fs/btrfs/inode.c
194
fs/btrfs/inode.c
|
@ -134,6 +134,7 @@ static noinline int insert_inline_extent(struct btrfs_trans_handle *trans,
|
|||
if (!path)
|
||||
return -ENOMEM;
|
||||
|
||||
path->leave_spinning = 1;
|
||||
btrfs_set_trans_block_group(trans, inode);
|
||||
|
||||
key.objectid = inode->i_ino;
|
||||
|
@ -167,9 +168,9 @@ static noinline int insert_inline_extent(struct btrfs_trans_handle *trans,
|
|||
cur_size = min_t(unsigned long, compressed_size,
|
||||
PAGE_CACHE_SIZE);
|
||||
|
||||
kaddr = kmap(cpage);
|
||||
kaddr = kmap_atomic(cpage, KM_USER0);
|
||||
write_extent_buffer(leaf, kaddr, ptr, cur_size);
|
||||
kunmap(cpage);
|
||||
kunmap_atomic(kaddr, KM_USER0);
|
||||
|
||||
i++;
|
||||
ptr += cur_size;
|
||||
|
@ -204,7 +205,7 @@ fail:
|
|||
* does the checks required to make sure the data is small enough
|
||||
* to fit as an inline extent.
|
||||
*/
|
||||
static int cow_file_range_inline(struct btrfs_trans_handle *trans,
|
||||
static noinline int cow_file_range_inline(struct btrfs_trans_handle *trans,
|
||||
struct btrfs_root *root,
|
||||
struct inode *inode, u64 start, u64 end,
|
||||
size_t compressed_size,
|
||||
|
@ -854,11 +855,6 @@ static int cow_file_range_async(struct inode *inode, struct page *locked_page,
|
|||
u64 cur_end;
|
||||
int limit = 10 * 1024 * 1042;
|
||||
|
||||
if (!btrfs_test_opt(root, COMPRESS)) {
|
||||
return cow_file_range(inode, locked_page, start, end,
|
||||
page_started, nr_written, 1);
|
||||
}
|
||||
|
||||
clear_extent_bit(&BTRFS_I(inode)->io_tree, start, end, EXTENT_LOCKED |
|
||||
EXTENT_DELALLOC, 1, 0, GFP_NOFS);
|
||||
while (start < end) {
|
||||
|
@ -935,7 +931,8 @@ static noinline int csum_exist_in_range(struct btrfs_root *root,
|
|||
* If no cow copies or snapshots exist, we write directly to the existing
|
||||
* blocks on disk
|
||||
*/
|
||||
static int run_delalloc_nocow(struct inode *inode, struct page *locked_page,
|
||||
static noinline int run_delalloc_nocow(struct inode *inode,
|
||||
struct page *locked_page,
|
||||
u64 start, u64 end, int *page_started, int force,
|
||||
unsigned long *nr_written)
|
||||
{
|
||||
|
@ -1133,6 +1130,7 @@ static int run_delalloc_range(struct inode *inode, struct page *locked_page,
|
|||
unsigned long *nr_written)
|
||||
{
|
||||
int ret;
|
||||
struct btrfs_root *root = BTRFS_I(inode)->root;
|
||||
|
||||
if (btrfs_test_flag(inode, NODATACOW))
|
||||
ret = run_delalloc_nocow(inode, locked_page, start, end,
|
||||
|
@ -1140,10 +1138,12 @@ static int run_delalloc_range(struct inode *inode, struct page *locked_page,
|
|||
else if (btrfs_test_flag(inode, PREALLOC))
|
||||
ret = run_delalloc_nocow(inode, locked_page, start, end,
|
||||
page_started, 0, nr_written);
|
||||
else if (!btrfs_test_opt(root, COMPRESS))
|
||||
ret = cow_file_range(inode, locked_page, start, end,
|
||||
page_started, nr_written, 1);
|
||||
else
|
||||
ret = cow_file_range_async(inode, locked_page, start, end,
|
||||
page_started, nr_written);
|
||||
|
||||
return ret;
|
||||
}
|
||||
|
||||
|
@ -1453,6 +1453,7 @@ static int insert_reserved_file_extent(struct btrfs_trans_handle *trans,
|
|||
path = btrfs_alloc_path();
|
||||
BUG_ON(!path);
|
||||
|
||||
path->leave_spinning = 1;
|
||||
ret = btrfs_drop_extents(trans, root, inode, file_pos,
|
||||
file_pos + num_bytes, file_pos, &hint);
|
||||
BUG_ON(ret);
|
||||
|
@ -1475,6 +1476,10 @@ static int insert_reserved_file_extent(struct btrfs_trans_handle *trans,
|
|||
btrfs_set_file_extent_compression(leaf, fi, compression);
|
||||
btrfs_set_file_extent_encryption(leaf, fi, encryption);
|
||||
btrfs_set_file_extent_other_encoding(leaf, fi, other_encoding);
|
||||
|
||||
btrfs_unlock_up_safe(path, 1);
|
||||
btrfs_set_lock_blocking(leaf);
|
||||
|
||||
btrfs_mark_buffer_dirty(leaf);
|
||||
|
||||
inode_add_bytes(inode, num_bytes);
|
||||
|
@ -1487,11 +1492,35 @@ static int insert_reserved_file_extent(struct btrfs_trans_handle *trans,
|
|||
root->root_key.objectid,
|
||||
trans->transid, inode->i_ino, &ins);
|
||||
BUG_ON(ret);
|
||||
|
||||
btrfs_free_path(path);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
/*
|
||||
* helper function for btrfs_finish_ordered_io, this
|
||||
* just reads in some of the csum leaves to prime them into ram
|
||||
* before we start the transaction. It limits the amount of btree
|
||||
* reads required while inside the transaction.
|
||||
*/
|
||||
static noinline void reada_csum(struct btrfs_root *root,
|
||||
struct btrfs_path *path,
|
||||
struct btrfs_ordered_extent *ordered_extent)
|
||||
{
|
||||
struct btrfs_ordered_sum *sum;
|
||||
u64 bytenr;
|
||||
|
||||
sum = list_entry(ordered_extent->list.next, struct btrfs_ordered_sum,
|
||||
list);
|
||||
bytenr = sum->sums[0].bytenr;
|
||||
|
||||
/*
|
||||
* we don't care about the results, the point of this search is
|
||||
* just to get the btree leaves into ram
|
||||
*/
|
||||
btrfs_lookup_csum(NULL, root->fs_info->csum_root, path, bytenr, 0);
|
||||
}
|
||||
|
||||
/* as ordered data IO finishes, this gets called so we can finish
|
||||
* an ordered extent if the range of bytes in the file it covers are
|
||||
* fully written.
|
||||
|
@ -1500,8 +1529,9 @@ static int btrfs_finish_ordered_io(struct inode *inode, u64 start, u64 end)
|
|||
{
|
||||
struct btrfs_root *root = BTRFS_I(inode)->root;
|
||||
struct btrfs_trans_handle *trans;
|
||||
struct btrfs_ordered_extent *ordered_extent;
|
||||
struct btrfs_ordered_extent *ordered_extent = NULL;
|
||||
struct extent_io_tree *io_tree = &BTRFS_I(inode)->io_tree;
|
||||
struct btrfs_path *path;
|
||||
int compressed = 0;
|
||||
int ret;
|
||||
|
||||
|
@ -1509,9 +1539,33 @@ static int btrfs_finish_ordered_io(struct inode *inode, u64 start, u64 end)
|
|||
if (!ret)
|
||||
return 0;
|
||||
|
||||
/*
|
||||
* before we join the transaction, try to do some of our IO.
|
||||
* This will limit the amount of IO that we have to do with
|
||||
* the transaction running. We're unlikely to need to do any
|
||||
* IO if the file extents are new, the disk_i_size checks
|
||||
* covers the most common case.
|
||||
*/
|
||||
if (start < BTRFS_I(inode)->disk_i_size) {
|
||||
path = btrfs_alloc_path();
|
||||
if (path) {
|
||||
ret = btrfs_lookup_file_extent(NULL, root, path,
|
||||
inode->i_ino,
|
||||
start, 0);
|
||||
ordered_extent = btrfs_lookup_ordered_extent(inode,
|
||||
start);
|
||||
if (!list_empty(&ordered_extent->list)) {
|
||||
btrfs_release_path(root, path);
|
||||
reada_csum(root, path, ordered_extent);
|
||||
}
|
||||
btrfs_free_path(path);
|
||||
}
|
||||
}
|
||||
|
||||
trans = btrfs_join_transaction(root, 1);
|
||||
|
||||
ordered_extent = btrfs_lookup_ordered_extent(inode, start);
|
||||
if (!ordered_extent)
|
||||
ordered_extent = btrfs_lookup_ordered_extent(inode, start);
|
||||
BUG_ON(!ordered_extent);
|
||||
if (test_bit(BTRFS_ORDERED_NOCOW, &ordered_extent->flags))
|
||||
goto nocow;
|
||||
|
@ -2101,6 +2155,7 @@ noinline int btrfs_update_inode(struct btrfs_trans_handle *trans,
|
|||
|
||||
path = btrfs_alloc_path();
|
||||
BUG_ON(!path);
|
||||
path->leave_spinning = 1;
|
||||
ret = btrfs_lookup_inode(trans, root, path,
|
||||
&BTRFS_I(inode)->location, 1);
|
||||
if (ret) {
|
||||
|
@ -2147,6 +2202,7 @@ int btrfs_unlink_inode(struct btrfs_trans_handle *trans,
|
|||
goto err;
|
||||
}
|
||||
|
||||
path->leave_spinning = 1;
|
||||
di = btrfs_lookup_dir_item(trans, root, path, dir->i_ino,
|
||||
name, name_len, -1);
|
||||
if (IS_ERR(di)) {
|
||||
|
@ -2190,8 +2246,6 @@ int btrfs_unlink_inode(struct btrfs_trans_handle *trans,
|
|||
ret = btrfs_del_inode_ref_in_log(trans, root, name, name_len,
|
||||
inode, dir->i_ino);
|
||||
BUG_ON(ret != 0 && ret != -ENOENT);
|
||||
if (ret != -ENOENT)
|
||||
BTRFS_I(dir)->log_dirty_trans = trans->transid;
|
||||
|
||||
ret = btrfs_del_dir_entries_in_log(trans, root, name, name_len,
|
||||
dir, index);
|
||||
|
@ -2224,6 +2278,9 @@ static int btrfs_unlink(struct inode *dir, struct dentry *dentry)
|
|||
trans = btrfs_start_transaction(root, 1);
|
||||
|
||||
btrfs_set_trans_block_group(trans, dir);
|
||||
|
||||
btrfs_record_unlink_dir(trans, dir, dentry->d_inode, 0);
|
||||
|
||||
ret = btrfs_unlink_inode(trans, root, dir, dentry->d_inode,
|
||||
dentry->d_name.name, dentry->d_name.len);
|
||||
|
||||
|
@ -2498,6 +2555,7 @@ noinline int btrfs_truncate_inode_items(struct btrfs_trans_handle *trans,
|
|||
key.type = (u8)-1;
|
||||
|
||||
search_again:
|
||||
path->leave_spinning = 1;
|
||||
ret = btrfs_search_slot(trans, root, &key, path, -1, 1);
|
||||
if (ret < 0)
|
||||
goto error;
|
||||
|
@ -2644,6 +2702,7 @@ delete:
|
|||
break;
|
||||
}
|
||||
if (found_extent) {
|
||||
btrfs_set_path_blocking(path);
|
||||
ret = btrfs_free_extent(trans, root, extent_start,
|
||||
extent_num_bytes,
|
||||
leaf->start, root_owner,
|
||||
|
@ -2848,11 +2907,21 @@ static int btrfs_setattr(struct dentry *dentry, struct iattr *attr)
|
|||
if (err)
|
||||
return err;
|
||||
|
||||
if (S_ISREG(inode->i_mode) &&
|
||||
attr->ia_valid & ATTR_SIZE && attr->ia_size > inode->i_size) {
|
||||
err = btrfs_cont_expand(inode, attr->ia_size);
|
||||
if (err)
|
||||
return err;
|
||||
if (S_ISREG(inode->i_mode) && (attr->ia_valid & ATTR_SIZE)) {
|
||||
if (attr->ia_size > inode->i_size) {
|
||||
err = btrfs_cont_expand(inode, attr->ia_size);
|
||||
if (err)
|
||||
return err;
|
||||
} else if (inode->i_size > 0 &&
|
||||
attr->ia_size == 0) {
|
||||
|
||||
/* we're truncating a file that used to have good
|
||||
* data down to zero. Make sure it gets into
|
||||
* the ordered flush list so that any new writes
|
||||
* get down to disk quickly.
|
||||
*/
|
||||
BTRFS_I(inode)->ordered_data_close = 1;
|
||||
}
|
||||
}
|
||||
|
||||
err = inode_setattr(inode, attr);
|
||||
|
@ -2984,13 +3053,14 @@ static noinline void init_btrfs_i(struct inode *inode)
|
|||
bi->disk_i_size = 0;
|
||||
bi->flags = 0;
|
||||
bi->index_cnt = (u64)-1;
|
||||
bi->log_dirty_trans = 0;
|
||||
bi->last_unlink_trans = 0;
|
||||
extent_map_tree_init(&BTRFS_I(inode)->extent_tree, GFP_NOFS);
|
||||
extent_io_tree_init(&BTRFS_I(inode)->io_tree,
|
||||
inode->i_mapping, GFP_NOFS);
|
||||
extent_io_tree_init(&BTRFS_I(inode)->io_failure_tree,
|
||||
inode->i_mapping, GFP_NOFS);
|
||||
INIT_LIST_HEAD(&BTRFS_I(inode)->delalloc_inodes);
|
||||
INIT_LIST_HEAD(&BTRFS_I(inode)->ordered_operations);
|
||||
btrfs_ordered_inode_tree_init(&BTRFS_I(inode)->ordered_tree);
|
||||
mutex_init(&BTRFS_I(inode)->extent_mutex);
|
||||
mutex_init(&BTRFS_I(inode)->log_mutex);
|
||||
|
@ -3449,6 +3519,7 @@ static struct inode *btrfs_new_inode(struct btrfs_trans_handle *trans,
|
|||
sizes[0] = sizeof(struct btrfs_inode_item);
|
||||
sizes[1] = name_len + sizeof(*ref);
|
||||
|
||||
path->leave_spinning = 1;
|
||||
ret = btrfs_insert_empty_items(trans, root, path, key, sizes, 2);
|
||||
if (ret != 0)
|
||||
goto fail;
|
||||
|
@ -3727,6 +3798,8 @@ static int btrfs_link(struct dentry *old_dentry, struct inode *dir,
|
|||
drop_inode = 1;
|
||||
|
||||
nr = trans->blocks_used;
|
||||
|
||||
btrfs_log_new_name(trans, inode, NULL, dentry->d_parent);
|
||||
btrfs_end_transaction_throttle(trans, root);
|
||||
fail:
|
||||
if (drop_inode) {
|
||||
|
@ -4363,6 +4436,8 @@ again:
|
|||
}
|
||||
ClearPageChecked(page);
|
||||
set_page_dirty(page);
|
||||
|
||||
BTRFS_I(inode)->last_trans = root->fs_info->generation + 1;
|
||||
unlock_extent(io_tree, page_start, page_end, GFP_NOFS);
|
||||
|
||||
out_unlock:
|
||||
|
@ -4388,6 +4463,27 @@ static void btrfs_truncate(struct inode *inode)
|
|||
btrfs_wait_ordered_range(inode, inode->i_size & (~mask), (u64)-1);
|
||||
|
||||
trans = btrfs_start_transaction(root, 1);
|
||||
|
||||
/*
|
||||
* setattr is responsible for setting the ordered_data_close flag,
|
||||
* but that is only tested during the last file release. That
|
||||
* could happen well after the next commit, leaving a great big
|
||||
* window where new writes may get lost if someone chooses to write
|
||||
* to this file after truncating to zero
|
||||
*
|
||||
* The inode doesn't have any dirty data here, and so if we commit
|
||||
* this is a noop. If someone immediately starts writing to the inode
|
||||
* it is very likely we'll catch some of their writes in this
|
||||
* transaction, and the commit will find this file on the ordered
|
||||
* data list with good things to send down.
|
||||
*
|
||||
* This is a best effort solution, there is still a window where
|
||||
* using truncate to replace the contents of the file will
|
||||
* end up with a zero length file after a crash.
|
||||
*/
|
||||
if (inode->i_size == 0 && BTRFS_I(inode)->ordered_data_close)
|
||||
btrfs_add_ordered_operation(trans, root, inode);
|
||||
|
||||
btrfs_set_trans_block_group(trans, inode);
|
||||
btrfs_i_size_write(inode, inode->i_size);
|
||||
|
||||
|
@ -4464,12 +4560,15 @@ struct inode *btrfs_alloc_inode(struct super_block *sb)
|
|||
ei->i_acl = BTRFS_ACL_NOT_CACHED;
|
||||
ei->i_default_acl = BTRFS_ACL_NOT_CACHED;
|
||||
INIT_LIST_HEAD(&ei->i_orphan);
|
||||
INIT_LIST_HEAD(&ei->ordered_operations);
|
||||
return &ei->vfs_inode;
|
||||
}
|
||||
|
||||
void btrfs_destroy_inode(struct inode *inode)
|
||||
{
|
||||
struct btrfs_ordered_extent *ordered;
|
||||
struct btrfs_root *root = BTRFS_I(inode)->root;
|
||||
|
||||
WARN_ON(!list_empty(&inode->i_dentry));
|
||||
WARN_ON(inode->i_data.nrpages);
|
||||
|
||||
|
@ -4480,13 +4579,24 @@ void btrfs_destroy_inode(struct inode *inode)
|
|||
BTRFS_I(inode)->i_default_acl != BTRFS_ACL_NOT_CACHED)
|
||||
posix_acl_release(BTRFS_I(inode)->i_default_acl);
|
||||
|
||||
spin_lock(&BTRFS_I(inode)->root->list_lock);
|
||||
/*
|
||||
* Make sure we're properly removed from the ordered operation
|
||||
* lists.
|
||||
*/
|
||||
smp_mb();
|
||||
if (!list_empty(&BTRFS_I(inode)->ordered_operations)) {
|
||||
spin_lock(&root->fs_info->ordered_extent_lock);
|
||||
list_del_init(&BTRFS_I(inode)->ordered_operations);
|
||||
spin_unlock(&root->fs_info->ordered_extent_lock);
|
||||
}
|
||||
|
||||
spin_lock(&root->list_lock);
|
||||
if (!list_empty(&BTRFS_I(inode)->i_orphan)) {
|
||||
printk(KERN_ERR "BTRFS: inode %lu: inode still on the orphan"
|
||||
" list\n", inode->i_ino);
|
||||
dump_stack();
|
||||
}
|
||||
spin_unlock(&BTRFS_I(inode)->root->list_lock);
|
||||
spin_unlock(&root->list_lock);
|
||||
|
||||
while (1) {
|
||||
ordered = btrfs_lookup_first_ordered_extent(inode, (u64)-1);
|
||||
|
@ -4611,8 +4721,36 @@ static int btrfs_rename(struct inode *old_dir, struct dentry *old_dentry,
|
|||
if (ret)
|
||||
goto out_unlock;
|
||||
|
||||
/*
|
||||
* we're using rename to replace one file with another.
|
||||
* and the replacement file is large. Start IO on it now so
|
||||
* we don't add too much work to the end of the transaction
|
||||
*/
|
||||
if (new_inode && old_inode && S_ISREG(old_inode->i_mode) &&
|
||||
new_inode->i_size &&
|
||||
old_inode->i_size > BTRFS_ORDERED_OPERATIONS_FLUSH_LIMIT)
|
||||
filemap_flush(old_inode->i_mapping);
|
||||
|
||||
trans = btrfs_start_transaction(root, 1);
|
||||
|
||||
/*
|
||||
* make sure the inode gets flushed if it is replacing
|
||||
* something.
|
||||
*/
|
||||
if (new_inode && new_inode->i_size &&
|
||||
old_inode && S_ISREG(old_inode->i_mode)) {
|
||||
btrfs_add_ordered_operation(trans, root, old_inode);
|
||||
}
|
||||
|
||||
/*
|
||||
* this is an ugly little race, but the rename is required to make
|
||||
* sure that if we crash, the inode is either at the old name
|
||||
* or the new one. pinning the log transaction lets us make sure
|
||||
* we don't allow a log commit to come in after we unlink the
|
||||
* name but before we add the new name back in.
|
||||
*/
|
||||
btrfs_pin_log_trans(root);
|
||||
|
||||
btrfs_set_trans_block_group(trans, new_dir);
|
||||
|
||||
btrfs_inc_nlink(old_dentry->d_inode);
|
||||
|
@ -4620,6 +4758,9 @@ static int btrfs_rename(struct inode *old_dir, struct dentry *old_dentry,
|
|||
new_dir->i_ctime = new_dir->i_mtime = ctime;
|
||||
old_inode->i_ctime = ctime;
|
||||
|
||||
if (old_dentry->d_parent != new_dentry->d_parent)
|
||||
btrfs_record_unlink_dir(trans, old_dir, old_inode, 1);
|
||||
|
||||
ret = btrfs_unlink_inode(trans, root, old_dir, old_dentry->d_inode,
|
||||
old_dentry->d_name.name,
|
||||
old_dentry->d_name.len);
|
||||
|
@ -4651,7 +4792,14 @@ static int btrfs_rename(struct inode *old_dir, struct dentry *old_dentry,
|
|||
if (ret)
|
||||
goto out_fail;
|
||||
|
||||
btrfs_log_new_name(trans, old_inode, old_dir,
|
||||
new_dentry->d_parent);
|
||||
out_fail:
|
||||
|
||||
/* this btrfs_end_log_trans just allows the current
|
||||
* log-sub transaction to complete
|
||||
*/
|
||||
btrfs_end_log_trans(root);
|
||||
btrfs_end_transaction_throttle(trans, root);
|
||||
out_unlock:
|
||||
return ret;
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue