mirror of
https://github.com/Fishwaldo/Star64_linux.git
synced 2025-03-16 12:14:06 +00:00
xfs: rename xfs_iflush_int()
with xfs_iflush() gone, we can rename xfs_iflush_int() back to xfs_iflush(). Also move it up above xfs_iflush_cluster() so we don't need the forward definition any more. Signed-off-by: Dave Chinner <dchinner@redhat.com> Reviewed-by: Amir Goldstein <amir73il@gmail.com> Reviewed-by: Darrick J. Wong <darrick.wong@oracle.com> Reviewed-by: Brian Foster <bfoster@redhat.com> Signed-off-by: Darrick J. Wong <darrick.wong@oracle.com>
This commit is contained in:
parent
90c60e1640
commit
e6187b3444
1 changed files with 146 additions and 147 deletions
|
@ -44,7 +44,6 @@ kmem_zone_t *xfs_inode_zone;
|
|||
*/
|
||||
#define XFS_ITRUNC_MAX_EXTENTS 2
|
||||
|
||||
STATIC int xfs_iflush_int(struct xfs_inode *, struct xfs_buf *);
|
||||
STATIC int xfs_iunlink(struct xfs_trans *, struct xfs_inode *);
|
||||
STATIC int xfs_iunlink_remove(struct xfs_trans *, struct xfs_inode *);
|
||||
|
||||
|
@ -3450,152 +3449,8 @@ out_release_wip:
|
|||
return error;
|
||||
}
|
||||
|
||||
/*
|
||||
* Non-blocking flush of dirty inode metadata into the backing buffer.
|
||||
*
|
||||
* The caller must have a reference to the inode and hold the cluster buffer
|
||||
* locked. The function will walk across all the inodes on the cluster buffer it
|
||||
* can find and lock without blocking, and flush them to the cluster buffer.
|
||||
*
|
||||
* On success, the caller must write out the buffer returned in *bp and
|
||||
* release it. On failure, the filesystem will be shut down, the buffer will
|
||||
* have been unlocked and released, and EFSCORRUPTED will be returned.
|
||||
*/
|
||||
int
|
||||
xfs_iflush_cluster(
|
||||
struct xfs_inode *ip,
|
||||
struct xfs_buf *bp)
|
||||
{
|
||||
struct xfs_mount *mp = ip->i_mount;
|
||||
struct xfs_perag *pag;
|
||||
unsigned long first_index, mask;
|
||||
int cilist_size;
|
||||
struct xfs_inode **cilist;
|
||||
struct xfs_inode *cip;
|
||||
struct xfs_ino_geometry *igeo = M_IGEO(mp);
|
||||
int error = 0;
|
||||
int nr_found;
|
||||
int clcount = 0;
|
||||
int i;
|
||||
|
||||
pag = xfs_perag_get(mp, XFS_INO_TO_AGNO(mp, ip->i_ino));
|
||||
|
||||
cilist_size = igeo->inodes_per_cluster * sizeof(struct xfs_inode *);
|
||||
cilist = kmem_alloc(cilist_size, KM_MAYFAIL|KM_NOFS);
|
||||
if (!cilist)
|
||||
goto out_put;
|
||||
|
||||
mask = ~(igeo->inodes_per_cluster - 1);
|
||||
first_index = XFS_INO_TO_AGINO(mp, ip->i_ino) & mask;
|
||||
rcu_read_lock();
|
||||
/* really need a gang lookup range call here */
|
||||
nr_found = radix_tree_gang_lookup(&pag->pag_ici_root, (void**)cilist,
|
||||
first_index, igeo->inodes_per_cluster);
|
||||
if (nr_found == 0)
|
||||
goto out_free;
|
||||
|
||||
for (i = 0; i < nr_found; i++) {
|
||||
cip = cilist[i];
|
||||
|
||||
/*
|
||||
* because this is an RCU protected lookup, we could find a
|
||||
* recently freed or even reallocated inode during the lookup.
|
||||
* We need to check under the i_flags_lock for a valid inode
|
||||
* here. Skip it if it is not valid or the wrong inode.
|
||||
*/
|
||||
spin_lock(&cip->i_flags_lock);
|
||||
if (!cip->i_ino ||
|
||||
__xfs_iflags_test(cip, XFS_ISTALE)) {
|
||||
spin_unlock(&cip->i_flags_lock);
|
||||
continue;
|
||||
}
|
||||
|
||||
/*
|
||||
* Once we fall off the end of the cluster, no point checking
|
||||
* any more inodes in the list because they will also all be
|
||||
* outside the cluster.
|
||||
*/
|
||||
if ((XFS_INO_TO_AGINO(mp, cip->i_ino) & mask) != first_index) {
|
||||
spin_unlock(&cip->i_flags_lock);
|
||||
break;
|
||||
}
|
||||
spin_unlock(&cip->i_flags_lock);
|
||||
|
||||
/*
|
||||
* Do an un-protected check to see if the inode is dirty and
|
||||
* is a candidate for flushing. These checks will be repeated
|
||||
* later after the appropriate locks are acquired.
|
||||
*/
|
||||
if (xfs_inode_clean(cip) && xfs_ipincount(cip) == 0)
|
||||
continue;
|
||||
|
||||
/*
|
||||
* Try to get locks. If any are unavailable or it is pinned,
|
||||
* then this inode cannot be flushed and is skipped.
|
||||
*/
|
||||
|
||||
if (!xfs_ilock_nowait(cip, XFS_ILOCK_SHARED))
|
||||
continue;
|
||||
if (!xfs_iflock_nowait(cip)) {
|
||||
xfs_iunlock(cip, XFS_ILOCK_SHARED);
|
||||
continue;
|
||||
}
|
||||
if (xfs_ipincount(cip)) {
|
||||
xfs_ifunlock(cip);
|
||||
xfs_iunlock(cip, XFS_ILOCK_SHARED);
|
||||
continue;
|
||||
}
|
||||
|
||||
|
||||
/*
|
||||
* Check the inode number again, just to be certain we are not
|
||||
* racing with freeing in xfs_reclaim_inode(). See the comments
|
||||
* in that function for more information as to why the initial
|
||||
* check is not sufficient.
|
||||
*/
|
||||
if (!cip->i_ino) {
|
||||
xfs_ifunlock(cip);
|
||||
xfs_iunlock(cip, XFS_ILOCK_SHARED);
|
||||
continue;
|
||||
}
|
||||
|
||||
/*
|
||||
* arriving here means that this inode can be flushed. First
|
||||
* re-check that it's dirty before flushing.
|
||||
*/
|
||||
if (!xfs_inode_clean(cip)) {
|
||||
error = xfs_iflush_int(cip, bp);
|
||||
if (error) {
|
||||
xfs_iunlock(cip, XFS_ILOCK_SHARED);
|
||||
goto out_free;
|
||||
}
|
||||
clcount++;
|
||||
} else {
|
||||
xfs_ifunlock(cip);
|
||||
}
|
||||
xfs_iunlock(cip, XFS_ILOCK_SHARED);
|
||||
}
|
||||
|
||||
if (clcount) {
|
||||
XFS_STATS_INC(mp, xs_icluster_flushcnt);
|
||||
XFS_STATS_ADD(mp, xs_icluster_flushinode, clcount);
|
||||
}
|
||||
|
||||
out_free:
|
||||
rcu_read_unlock();
|
||||
kmem_free(cilist);
|
||||
out_put:
|
||||
xfs_perag_put(pag);
|
||||
if (error) {
|
||||
bp->b_flags |= XBF_ASYNC;
|
||||
xfs_buf_ioend_fail(bp);
|
||||
xfs_force_shutdown(mp, SHUTDOWN_CORRUPT_INCORE);
|
||||
}
|
||||
return error;
|
||||
}
|
||||
|
||||
STATIC int
|
||||
xfs_iflush_int(
|
||||
static int
|
||||
xfs_iflush(
|
||||
struct xfs_inode *ip,
|
||||
struct xfs_buf *bp)
|
||||
{
|
||||
|
@ -3741,6 +3596,150 @@ flush_out:
|
|||
return error;
|
||||
}
|
||||
|
||||
/*
|
||||
* Non-blocking flush of dirty inode metadata into the backing buffer.
|
||||
*
|
||||
* The caller must have a reference to the inode and hold the cluster buffer
|
||||
* locked. The function will walk across all the inodes on the cluster buffer it
|
||||
* can find and lock without blocking, and flush them to the cluster buffer.
|
||||
*
|
||||
* On success, the caller must write out the buffer returned in *bp and
|
||||
* release it. On failure, the filesystem will be shut down, the buffer will
|
||||
* have been unlocked and released, and EFSCORRUPTED will be returned.
|
||||
*/
|
||||
int
|
||||
xfs_iflush_cluster(
|
||||
struct xfs_inode *ip,
|
||||
struct xfs_buf *bp)
|
||||
{
|
||||
struct xfs_mount *mp = ip->i_mount;
|
||||
struct xfs_perag *pag;
|
||||
unsigned long first_index, mask;
|
||||
int cilist_size;
|
||||
struct xfs_inode **cilist;
|
||||
struct xfs_inode *cip;
|
||||
struct xfs_ino_geometry *igeo = M_IGEO(mp);
|
||||
int error = 0;
|
||||
int nr_found;
|
||||
int clcount = 0;
|
||||
int i;
|
||||
|
||||
pag = xfs_perag_get(mp, XFS_INO_TO_AGNO(mp, ip->i_ino));
|
||||
|
||||
cilist_size = igeo->inodes_per_cluster * sizeof(struct xfs_inode *);
|
||||
cilist = kmem_alloc(cilist_size, KM_MAYFAIL|KM_NOFS);
|
||||
if (!cilist)
|
||||
goto out_put;
|
||||
|
||||
mask = ~(igeo->inodes_per_cluster - 1);
|
||||
first_index = XFS_INO_TO_AGINO(mp, ip->i_ino) & mask;
|
||||
rcu_read_lock();
|
||||
/* really need a gang lookup range call here */
|
||||
nr_found = radix_tree_gang_lookup(&pag->pag_ici_root, (void**)cilist,
|
||||
first_index, igeo->inodes_per_cluster);
|
||||
if (nr_found == 0)
|
||||
goto out_free;
|
||||
|
||||
for (i = 0; i < nr_found; i++) {
|
||||
cip = cilist[i];
|
||||
|
||||
/*
|
||||
* because this is an RCU protected lookup, we could find a
|
||||
* recently freed or even reallocated inode during the lookup.
|
||||
* We need to check under the i_flags_lock for a valid inode
|
||||
* here. Skip it if it is not valid or the wrong inode.
|
||||
*/
|
||||
spin_lock(&cip->i_flags_lock);
|
||||
if (!cip->i_ino ||
|
||||
__xfs_iflags_test(cip, XFS_ISTALE)) {
|
||||
spin_unlock(&cip->i_flags_lock);
|
||||
continue;
|
||||
}
|
||||
|
||||
/*
|
||||
* Once we fall off the end of the cluster, no point checking
|
||||
* any more inodes in the list because they will also all be
|
||||
* outside the cluster.
|
||||
*/
|
||||
if ((XFS_INO_TO_AGINO(mp, cip->i_ino) & mask) != first_index) {
|
||||
spin_unlock(&cip->i_flags_lock);
|
||||
break;
|
||||
}
|
||||
spin_unlock(&cip->i_flags_lock);
|
||||
|
||||
/*
|
||||
* Do an un-protected check to see if the inode is dirty and
|
||||
* is a candidate for flushing. These checks will be repeated
|
||||
* later after the appropriate locks are acquired.
|
||||
*/
|
||||
if (xfs_inode_clean(cip) && xfs_ipincount(cip) == 0)
|
||||
continue;
|
||||
|
||||
/*
|
||||
* Try to get locks. If any are unavailable or it is pinned,
|
||||
* then this inode cannot be flushed and is skipped.
|
||||
*/
|
||||
|
||||
if (!xfs_ilock_nowait(cip, XFS_ILOCK_SHARED))
|
||||
continue;
|
||||
if (!xfs_iflock_nowait(cip)) {
|
||||
xfs_iunlock(cip, XFS_ILOCK_SHARED);
|
||||
continue;
|
||||
}
|
||||
if (xfs_ipincount(cip)) {
|
||||
xfs_ifunlock(cip);
|
||||
xfs_iunlock(cip, XFS_ILOCK_SHARED);
|
||||
continue;
|
||||
}
|
||||
|
||||
|
||||
/*
|
||||
* Check the inode number again, just to be certain we are not
|
||||
* racing with freeing in xfs_reclaim_inode(). See the comments
|
||||
* in that function for more information as to why the initial
|
||||
* check is not sufficient.
|
||||
*/
|
||||
if (!cip->i_ino) {
|
||||
xfs_ifunlock(cip);
|
||||
xfs_iunlock(cip, XFS_ILOCK_SHARED);
|
||||
continue;
|
||||
}
|
||||
|
||||
/*
|
||||
* arriving here means that this inode can be flushed. First
|
||||
* re-check that it's dirty before flushing.
|
||||
*/
|
||||
if (!xfs_inode_clean(cip)) {
|
||||
error = xfs_iflush(cip, bp);
|
||||
if (error) {
|
||||
xfs_iunlock(cip, XFS_ILOCK_SHARED);
|
||||
goto out_free;
|
||||
}
|
||||
clcount++;
|
||||
} else {
|
||||
xfs_ifunlock(cip);
|
||||
}
|
||||
xfs_iunlock(cip, XFS_ILOCK_SHARED);
|
||||
}
|
||||
|
||||
if (clcount) {
|
||||
XFS_STATS_INC(mp, xs_icluster_flushcnt);
|
||||
XFS_STATS_ADD(mp, xs_icluster_flushinode, clcount);
|
||||
}
|
||||
|
||||
out_free:
|
||||
rcu_read_unlock();
|
||||
kmem_free(cilist);
|
||||
out_put:
|
||||
xfs_perag_put(pag);
|
||||
if (error) {
|
||||
bp->b_flags |= XBF_ASYNC;
|
||||
xfs_buf_ioend_fail(bp);
|
||||
xfs_force_shutdown(mp, SHUTDOWN_CORRUPT_INCORE);
|
||||
}
|
||||
return error;
|
||||
}
|
||||
|
||||
/* Release an inode. */
|
||||
void
|
||||
xfs_irele(
|
||||
|
|
Loading…
Add table
Reference in a new issue