forked from Minki/linux
[XFS] add infrastructure for waiting on I/O completion at inode reclaim
time SGI-PV: 934766 SGI-Modid: xfs-linux:xfs-kern:196854a Signed-off-by: Christoph Hellwig <hch@sgi.com> Signed-off-by: Nathan Scott <nathans@sgi.com>
This commit is contained in:
parent
592cb26bda
commit
51c91ed52b
@ -139,7 +139,7 @@ linvfs_unwritten_convert(
|
||||
XFS_BUF_SET_FSPRIVATE(bp, NULL);
|
||||
XFS_BUF_CLR_IODONE_FUNC(bp);
|
||||
XFS_BUF_UNDATAIO(bp);
|
||||
iput(LINVFS_GET_IP(vp));
|
||||
vn_iowake(vp);
|
||||
pagebuf_iodone(bp, 0, 0);
|
||||
}
|
||||
|
||||
@ -448,14 +448,7 @@ xfs_map_unwritten(
|
||||
if (!pb)
|
||||
return -EAGAIN;
|
||||
|
||||
/* Take a reference to the inode to prevent it from
|
||||
* being reclaimed while we have outstanding unwritten
|
||||
* extent IO on it.
|
||||
*/
|
||||
if ((igrab(inode)) != inode) {
|
||||
pagebuf_free(pb);
|
||||
return -EAGAIN;
|
||||
}
|
||||
atomic_inc(&LINVFS_GET_VP(inode)->v_iocount);
|
||||
|
||||
/* Set the count to 1 initially, this will stop an I/O
|
||||
* completion callout which happens before we have started
|
||||
|
@ -42,17 +42,33 @@ DEFINE_SPINLOCK(vnumber_lock);
|
||||
*/
|
||||
#define NVSYNC 37
|
||||
#define vptosync(v) (&vsync[((unsigned long)v) % NVSYNC])
|
||||
sv_t vsync[NVSYNC];
|
||||
STATIC wait_queue_head_t vsync[NVSYNC];
|
||||
|
||||
|
||||
void
|
||||
vn_init(void)
|
||||
{
|
||||
register sv_t *svp;
|
||||
register int i;
|
||||
int i;
|
||||
|
||||
for (svp = vsync, i = 0; i < NVSYNC; i++, svp++)
|
||||
init_sv(svp, SV_DEFAULT, "vsy", i);
|
||||
for (i = 0; i < NVSYNC; i++)
|
||||
init_waitqueue_head(&vsync[i]);
|
||||
}
|
||||
|
||||
void
|
||||
vn_iowait(
|
||||
struct vnode *vp)
|
||||
{
|
||||
wait_queue_head_t *wq = vptosync(vp);
|
||||
|
||||
wait_event(*wq, (atomic_read(&vp->v_iocount) == 0));
|
||||
}
|
||||
|
||||
void
|
||||
vn_iowake(
|
||||
struct vnode *vp)
|
||||
{
|
||||
if (atomic_dec_and_test(&vp->v_iocount))
|
||||
wake_up(vptosync(vp));
|
||||
}
|
||||
|
||||
/*
|
||||
@ -111,6 +127,8 @@ vn_initialize(
|
||||
/* Initialize the first behavior and the behavior chain head. */
|
||||
vn_bhv_head_init(VN_BHV_HEAD(vp), "vnode");
|
||||
|
||||
atomic_set(&vp->v_iocount, 0);
|
||||
|
||||
#ifdef XFS_VNODE_TRACE
|
||||
vp->v_trace = ktrace_alloc(VNODE_TRACE_SIZE, KM_SLEEP);
|
||||
#endif /* XFS_VNODE_TRACE */
|
||||
|
@ -80,6 +80,7 @@ typedef struct vnode {
|
||||
vnumber_t v_number; /* in-core vnode number */
|
||||
vn_bhv_head_t v_bh; /* behavior head */
|
||||
spinlock_t v_lock; /* VN_LOCK/VN_UNLOCK */
|
||||
atomic_t v_iocount; /* outstanding I/O count */
|
||||
#ifdef XFS_VNODE_TRACE
|
||||
struct ktrace *v_trace; /* trace header structure */
|
||||
#endif
|
||||
@ -506,6 +507,9 @@ extern int vn_revalidate(struct vnode *);
|
||||
extern void vn_revalidate_core(struct vnode *, vattr_t *);
|
||||
extern void vn_remove(struct vnode *);
|
||||
|
||||
extern void vn_iowait(struct vnode *vp);
|
||||
extern void vn_iowake(struct vnode *vp);
|
||||
|
||||
static inline int vn_count(struct vnode *vp)
|
||||
{
|
||||
return atomic_read(&LINVFS_GET_IP(vp)->i_count);
|
||||
|
@ -3846,51 +3846,10 @@ xfs_reclaim(
|
||||
return 0;
|
||||
}
|
||||
|
||||
if ((ip->i_d.di_mode & S_IFMT) == S_IFREG) {
|
||||
if (ip->i_d.di_size > 0) {
|
||||
/*
|
||||
* Flush and invalidate any data left around that is
|
||||
* a part of this file.
|
||||
*
|
||||
* Get the inode's i/o lock so that buffers are pushed
|
||||
* out while holding the proper lock. We can't hold
|
||||
* the inode lock here since flushing out buffers may
|
||||
* cause us to try to get the lock in xfs_strategy().
|
||||
*
|
||||
* We don't have to call remapf() here, because there
|
||||
* cannot be any mapped file references to this vnode
|
||||
* since it is being reclaimed.
|
||||
*/
|
||||
xfs_ilock(ip, XFS_IOLOCK_EXCL);
|
||||
vn_iowait(vp);
|
||||
|
||||
/*
|
||||
* If we hit an IO error, we need to make sure that the
|
||||
* buffer and page caches of file data for
|
||||
* the file are tossed away. We don't want to use
|
||||
* VOP_FLUSHINVAL_PAGES here because we don't want dirty
|
||||
* pages to stay attached to the vnode, but be
|
||||
* marked P_BAD. pdflush/vnode_pagebad
|
||||
* hates that.
|
||||
*/
|
||||
if (!XFS_FORCED_SHUTDOWN(ip->i_mount)) {
|
||||
VOP_FLUSHINVAL_PAGES(vp, 0, -1, FI_NONE);
|
||||
} else {
|
||||
VOP_TOSS_PAGES(vp, 0, -1, FI_NONE);
|
||||
}
|
||||
|
||||
ASSERT(VN_CACHED(vp) == 0);
|
||||
ASSERT(XFS_FORCED_SHUTDOWN(ip->i_mount) ||
|
||||
ip->i_delayed_blks == 0);
|
||||
xfs_iunlock(ip, XFS_IOLOCK_EXCL);
|
||||
} else if (XFS_FORCED_SHUTDOWN(ip->i_mount)) {
|
||||
/*
|
||||
* di_size field may not be quite accurate if we're
|
||||
* shutting down.
|
||||
*/
|
||||
VOP_TOSS_PAGES(vp, 0, -1, FI_NONE);
|
||||
ASSERT(VN_CACHED(vp) == 0);
|
||||
}
|
||||
}
|
||||
ASSERT(XFS_FORCED_SHUTDOWN(ip->i_mount) || ip->i_delayed_blks == 0);
|
||||
ASSERT(VN_CACHED(vp) == 0);
|
||||
|
||||
/* If we have nothing to flush with this inode then complete the
|
||||
* teardown now, otherwise break the link between the xfs inode
|
||||
|
Loading…
Reference in New Issue
Block a user