forked from Minki/linux
pipe: remove pipe_wait() and fix wakeup race with splice
The pipe splice code still used the old model of waiting for pipe IO by using a non-specific "pipe_wait()" that waited for any pipe event to happen, which depended on all pipe IO being entirely serialized by the pipe lock. So by checking the state you were waiting for, and then adding yourself to the wait queue before dropping the lock, you were guaranteed to see all the wakeups. Strictly speaking, the actual wakeups were not done under the lock, but the pipe_wait() model still worked, because since the waiter held the lock when checking whether it should sleep, it would always see the current state, and the wakeup was always done after updating the state. However, commit0ddad21d3e
("pipe: use exclusive waits when reading or writing") split the single wait-queue into two, and in the process also made the "wait for event" code wait for _two_ wait queues, and that then showed a race with the wakers that were not serialized by the pipe lock. It's only splice that used that "pipe_wait()" model, so the problem wasn't obvious, but Josef Bacik reports: "I hit a hang with fstest btrfs/187, which does a btrfs send into /dev/null. This works by creating a pipe, the write side is given to the kernel to write into, and the read side is handed to a thread that splices into a file, in this case /dev/null. The box that was hung had the write side stuck here [pipe_write] and the read side stuck here [splice_from_pipe_next -> pipe_wait]. [ more details about pipe_wait() scenario ] The problem is we're doing the prepare_to_wait, which sets our state each time, however we can be woken up either with reads or writes. In the case above we race with the WRITER waking us up, and re-set our state to INTERRUPTIBLE, and thus never break out of schedule" Josef had a patch that avoided the issue in pipe_wait() by just making it set the state only once, but the deeper problem is that pipe_wait() depends on a level of synchonization by the pipe mutex that it really shouldn't. And the whole "wait for any pipe state change" model really isn't very good to begin with. So rather than trying to work around things in pipe_wait(), remove that legacy model of "wait for arbitrary pipe event" entirely, and actually create functions that wait for the pipe actually being readable or writable, and can do so without depending on the pipe lock serializing everything. Fixes:0ddad21d3e
("pipe: use exclusive waits when reading or writing") Link: https://lore.kernel.org/linux-fsdevel/bfa88b5ad6f069b2b679316b9e495a970130416c.1601567868.git.josef@toxicpanda.com/ Reported-by: Josef Bacik <josef@toxicpanda.com> Reviewed-and-tested-by: Josef Bacik <josef@toxicpanda.com> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
This commit is contained in:
parent
44b6e23be3
commit
472e5b056f
62
fs/pipe.c
62
fs/pipe.c
@ -106,25 +106,6 @@ void pipe_double_lock(struct pipe_inode_info *pipe1,
|
||||
}
|
||||
}
|
||||
|
||||
/* Drop the inode semaphore and wait for a pipe event, atomically */
|
||||
void pipe_wait(struct pipe_inode_info *pipe)
|
||||
{
|
||||
DEFINE_WAIT(rdwait);
|
||||
DEFINE_WAIT(wrwait);
|
||||
|
||||
/*
|
||||
* Pipes are system-local resources, so sleeping on them
|
||||
* is considered a noninteractive wait:
|
||||
*/
|
||||
prepare_to_wait(&pipe->rd_wait, &rdwait, TASK_INTERRUPTIBLE);
|
||||
prepare_to_wait(&pipe->wr_wait, &wrwait, TASK_INTERRUPTIBLE);
|
||||
pipe_unlock(pipe);
|
||||
schedule();
|
||||
finish_wait(&pipe->rd_wait, &rdwait);
|
||||
finish_wait(&pipe->wr_wait, &wrwait);
|
||||
pipe_lock(pipe);
|
||||
}
|
||||
|
||||
static void anon_pipe_buf_release(struct pipe_inode_info *pipe,
|
||||
struct pipe_buffer *buf)
|
||||
{
|
||||
@ -1035,12 +1016,52 @@ SYSCALL_DEFINE1(pipe, int __user *, fildes)
|
||||
return do_pipe2(fildes, 0);
|
||||
}
|
||||
|
||||
/*
|
||||
* This is the stupid "wait for pipe to be readable or writable"
|
||||
* model.
|
||||
*
|
||||
* See pipe_read/write() for the proper kind of exclusive wait,
|
||||
* but that requires that we wake up any other readers/writers
|
||||
* if we then do not end up reading everything (ie the whole
|
||||
* "wake_next_reader/writer" logic in pipe_read/write()).
|
||||
*/
|
||||
void pipe_wait_readable(struct pipe_inode_info *pipe)
|
||||
{
|
||||
pipe_unlock(pipe);
|
||||
wait_event_interruptible(pipe->rd_wait, pipe_readable(pipe));
|
||||
pipe_lock(pipe);
|
||||
}
|
||||
|
||||
void pipe_wait_writable(struct pipe_inode_info *pipe)
|
||||
{
|
||||
pipe_unlock(pipe);
|
||||
wait_event_interruptible(pipe->wr_wait, pipe_writable(pipe));
|
||||
pipe_lock(pipe);
|
||||
}
|
||||
|
||||
/*
|
||||
* This depends on both the wait (here) and the wakeup (wake_up_partner)
|
||||
* holding the pipe lock, so "*cnt" is stable and we know a wakeup cannot
|
||||
* race with the count check and waitqueue prep.
|
||||
*
|
||||
* Normally in order to avoid races, you'd do the prepare_to_wait() first,
|
||||
* then check the condition you're waiting for, and only then sleep. But
|
||||
* because of the pipe lock, we can check the condition before being on
|
||||
* the wait queue.
|
||||
*
|
||||
* We use the 'rd_wait' waitqueue for pipe partner waiting.
|
||||
*/
|
||||
static int wait_for_partner(struct pipe_inode_info *pipe, unsigned int *cnt)
|
||||
{
|
||||
DEFINE_WAIT(rdwait);
|
||||
int cur = *cnt;
|
||||
|
||||
while (cur == *cnt) {
|
||||
pipe_wait(pipe);
|
||||
prepare_to_wait(&pipe->rd_wait, &rdwait, TASK_INTERRUPTIBLE);
|
||||
pipe_unlock(pipe);
|
||||
schedule();
|
||||
finish_wait(&pipe->rd_wait, &rdwait);
|
||||
pipe_lock(pipe);
|
||||
if (signal_pending(current))
|
||||
break;
|
||||
}
|
||||
@ -1050,7 +1071,6 @@ static int wait_for_partner(struct pipe_inode_info *pipe, unsigned int *cnt)
|
||||
static void wake_up_partner(struct pipe_inode_info *pipe)
|
||||
{
|
||||
wake_up_interruptible_all(&pipe->rd_wait);
|
||||
wake_up_interruptible_all(&pipe->wr_wait);
|
||||
}
|
||||
|
||||
static int fifo_open(struct inode *inode, struct file *filp)
|
||||
|
@ -563,7 +563,7 @@ static int splice_from_pipe_next(struct pipe_inode_info *pipe, struct splice_des
|
||||
sd->need_wakeup = false;
|
||||
}
|
||||
|
||||
pipe_wait(pipe);
|
||||
pipe_wait_readable(pipe);
|
||||
}
|
||||
|
||||
return 1;
|
||||
@ -1077,7 +1077,7 @@ static int wait_for_space(struct pipe_inode_info *pipe, unsigned flags)
|
||||
return -EAGAIN;
|
||||
if (signal_pending(current))
|
||||
return -ERESTARTSYS;
|
||||
pipe_wait(pipe);
|
||||
pipe_wait_writable(pipe);
|
||||
}
|
||||
}
|
||||
|
||||
@ -1454,7 +1454,7 @@ static int ipipe_prep(struct pipe_inode_info *pipe, unsigned int flags)
|
||||
ret = -EAGAIN;
|
||||
break;
|
||||
}
|
||||
pipe_wait(pipe);
|
||||
pipe_wait_readable(pipe);
|
||||
}
|
||||
|
||||
pipe_unlock(pipe);
|
||||
@ -1493,7 +1493,7 @@ static int opipe_prep(struct pipe_inode_info *pipe, unsigned int flags)
|
||||
ret = -ERESTARTSYS;
|
||||
break;
|
||||
}
|
||||
pipe_wait(pipe);
|
||||
pipe_wait_writable(pipe);
|
||||
}
|
||||
|
||||
pipe_unlock(pipe);
|
||||
|
@ -240,8 +240,9 @@ extern unsigned int pipe_max_size;
|
||||
extern unsigned long pipe_user_pages_hard;
|
||||
extern unsigned long pipe_user_pages_soft;
|
||||
|
||||
/* Drop the inode semaphore and wait for a pipe event, atomically */
|
||||
void pipe_wait(struct pipe_inode_info *pipe);
|
||||
/* Wait for a pipe to be readable/writable while dropping the pipe lock */
|
||||
void pipe_wait_readable(struct pipe_inode_info *);
|
||||
void pipe_wait_writable(struct pipe_inode_info *);
|
||||
|
||||
struct pipe_inode_info *alloc_pipe_info(void);
|
||||
void free_pipe_info(struct pipe_inode_info *);
|
||||
|
Loading…
Reference in New Issue
Block a user