ceph: fix pending vmtruncate race
The locking order for pending vmtruncate is wrong, it can lead to following race: write wmtruncate work ------------------------ ---------------------- lock i_mutex check i_truncate_pending check i_truncate_pending truncate_inode_pages() lock i_mutex (blocked) copy data to page cache unlock i_mutex truncate_inode_pages() The fix is take i_mutex before calling __ceph_do_pending_vmtruncate() Fixes: http://tracker.ceph.com/issues/5453 Signed-off-by: Yan, Zheng <zheng.z.yan@intel.com> Reviewed-by: Sage Weil <sage@inktank.com>
This commit is contained in:
parent
5446429630
commit
b415bf4f9f
@ -2057,7 +2057,11 @@ static int try_get_cap_refs(struct ceph_inode_info *ci, int need, int want,
|
|||||||
/* finish pending truncate */
|
/* finish pending truncate */
|
||||||
while (ci->i_truncate_pending) {
|
while (ci->i_truncate_pending) {
|
||||||
spin_unlock(&ci->i_ceph_lock);
|
spin_unlock(&ci->i_ceph_lock);
|
||||||
__ceph_do_pending_vmtruncate(inode, !(need & CEPH_CAP_FILE_WR));
|
if (!(need & CEPH_CAP_FILE_WR))
|
||||||
|
mutex_lock(&inode->i_mutex);
|
||||||
|
__ceph_do_pending_vmtruncate(inode);
|
||||||
|
if (!(need & CEPH_CAP_FILE_WR))
|
||||||
|
mutex_unlock(&inode->i_mutex);
|
||||||
spin_lock(&ci->i_ceph_lock);
|
spin_lock(&ci->i_ceph_lock);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -822,7 +822,7 @@ static loff_t ceph_llseek(struct file *file, loff_t offset, int whence)
|
|||||||
int ret;
|
int ret;
|
||||||
|
|
||||||
mutex_lock(&inode->i_mutex);
|
mutex_lock(&inode->i_mutex);
|
||||||
__ceph_do_pending_vmtruncate(inode, false);
|
__ceph_do_pending_vmtruncate(inode);
|
||||||
|
|
||||||
if (whence == SEEK_END || whence == SEEK_DATA || whence == SEEK_HOLE) {
|
if (whence == SEEK_END || whence == SEEK_DATA || whence == SEEK_HOLE) {
|
||||||
ret = ceph_do_getattr(inode, CEPH_STAT_CAP_SIZE);
|
ret = ceph_do_getattr(inode, CEPH_STAT_CAP_SIZE);
|
||||||
|
@ -1465,7 +1465,9 @@ static void ceph_vmtruncate_work(struct work_struct *work)
|
|||||||
struct inode *inode = &ci->vfs_inode;
|
struct inode *inode = &ci->vfs_inode;
|
||||||
|
|
||||||
dout("vmtruncate_work %p\n", inode);
|
dout("vmtruncate_work %p\n", inode);
|
||||||
__ceph_do_pending_vmtruncate(inode, true);
|
mutex_lock(&inode->i_mutex);
|
||||||
|
__ceph_do_pending_vmtruncate(inode);
|
||||||
|
mutex_unlock(&inode->i_mutex);
|
||||||
iput(inode);
|
iput(inode);
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -1492,7 +1494,7 @@ void ceph_queue_vmtruncate(struct inode *inode)
|
|||||||
* Make sure any pending truncation is applied before doing anything
|
* Make sure any pending truncation is applied before doing anything
|
||||||
* that may depend on it.
|
* that may depend on it.
|
||||||
*/
|
*/
|
||||||
void __ceph_do_pending_vmtruncate(struct inode *inode, bool needlock)
|
void __ceph_do_pending_vmtruncate(struct inode *inode)
|
||||||
{
|
{
|
||||||
struct ceph_inode_info *ci = ceph_inode(inode);
|
struct ceph_inode_info *ci = ceph_inode(inode);
|
||||||
u64 to;
|
u64 to;
|
||||||
@ -1525,11 +1527,7 @@ retry:
|
|||||||
ci->i_truncate_pending, to);
|
ci->i_truncate_pending, to);
|
||||||
spin_unlock(&ci->i_ceph_lock);
|
spin_unlock(&ci->i_ceph_lock);
|
||||||
|
|
||||||
if (needlock)
|
|
||||||
mutex_lock(&inode->i_mutex);
|
|
||||||
truncate_inode_pages(inode->i_mapping, to);
|
truncate_inode_pages(inode->i_mapping, to);
|
||||||
if (needlock)
|
|
||||||
mutex_unlock(&inode->i_mutex);
|
|
||||||
|
|
||||||
spin_lock(&ci->i_ceph_lock);
|
spin_lock(&ci->i_ceph_lock);
|
||||||
if (to == ci->i_truncate_size) {
|
if (to == ci->i_truncate_size) {
|
||||||
@ -1588,7 +1586,7 @@ int ceph_setattr(struct dentry *dentry, struct iattr *attr)
|
|||||||
if (ceph_snap(inode) != CEPH_NOSNAP)
|
if (ceph_snap(inode) != CEPH_NOSNAP)
|
||||||
return -EROFS;
|
return -EROFS;
|
||||||
|
|
||||||
__ceph_do_pending_vmtruncate(inode, false);
|
__ceph_do_pending_vmtruncate(inode);
|
||||||
|
|
||||||
err = inode_change_ok(inode, attr);
|
err = inode_change_ok(inode, attr);
|
||||||
if (err != 0)
|
if (err != 0)
|
||||||
@ -1770,7 +1768,7 @@ int ceph_setattr(struct dentry *dentry, struct iattr *attr)
|
|||||||
ceph_cap_string(dirtied), mask);
|
ceph_cap_string(dirtied), mask);
|
||||||
|
|
||||||
ceph_mdsc_put_request(req);
|
ceph_mdsc_put_request(req);
|
||||||
__ceph_do_pending_vmtruncate(inode, false);
|
__ceph_do_pending_vmtruncate(inode);
|
||||||
return err;
|
return err;
|
||||||
out:
|
out:
|
||||||
spin_unlock(&ci->i_ceph_lock);
|
spin_unlock(&ci->i_ceph_lock);
|
||||||
|
@ -692,7 +692,7 @@ extern int ceph_readdir_prepopulate(struct ceph_mds_request *req,
|
|||||||
extern int ceph_inode_holds_cap(struct inode *inode, int mask);
|
extern int ceph_inode_holds_cap(struct inode *inode, int mask);
|
||||||
|
|
||||||
extern int ceph_inode_set_size(struct inode *inode, loff_t size);
|
extern int ceph_inode_set_size(struct inode *inode, loff_t size);
|
||||||
extern void __ceph_do_pending_vmtruncate(struct inode *inode, bool needlock);
|
extern void __ceph_do_pending_vmtruncate(struct inode *inode);
|
||||||
extern void ceph_queue_vmtruncate(struct inode *inode);
|
extern void ceph_queue_vmtruncate(struct inode *inode);
|
||||||
|
|
||||||
extern void ceph_queue_invalidate(struct inode *inode);
|
extern void ceph_queue_invalidate(struct inode *inode);
|
||||||
|
Loading…
Reference in New Issue
Block a user