Re: [PATCH] tty races

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Jason Baron <[email protected]> wrote:
>
> 
> On Mon, 25 Apr 2005, Andrew Morton wrote:
> 
> > Jason Baron <[email protected]> wrote:
> > >
> > > There are a couple of tty race conditions, which lead to inconsistent tty 
> > >  reference counting and tty layer oopses.
> > > 
> > >  The first is a tty_open vs. tty_close race in drivers/char/tty.io.c. 
> > >  Basically, from the time that the tty->count is deemed to be 1 and that we 
> > >  are going to free it to the time that TTY_CLOSING bit is set, needs to be 
> > >  atomic with respect to the manipulation of tty->count in init_dev(). This 
> > >  atomicity was previously guarded by the BKL. However, this is no longer 
> > >  true with the addition of a down() call in the middle of the 
> > >  release_dev()'s atomic path. So either the down() needs to be moved 
> > >  outside the atomic patch or dropped. I would vote for simply dropping it 
> > >  as i don't see why it is necessary.
> > 
> > The release_dev() changes looks very fishy to me.  It _removes_ locking. 
> > If that fixes the testcase then one of two things is happening:
> > 
> > a) we have lock_kernel() coverage and the down()'s sleeping breaks the
> >    lock_kenrel() coverage or
> > 
> > b) we don't have lock_kernel() coverage, but removing the down() just
> >    alters the timing and makes the race less probable.
> > 
> > I think it's b).  lock_kernel() coverage in there is very incomplete on the
> > open() side.
> > 
> 
> The patch was written for case a. Indeed lock_kernel() may appear 
> incomplete on the open side, but it protects paths where we don't sleep. 
> So, the 'fast_track' path in 'init_dev', is protected against the 
> release_dev path from setting the 'tty_closing' local variable to the 
> setting of the TTY_CLOSING flag. Thus, i believe the dropping of the 
> down() is correct. 

I don't see anywhere which takes lock_kernel() on the tty_open() path.

The normal release_dev() path takes lock_kernel(), but two error-path
callers of lock_kernel() also appear to not take lock_kernel().

> This was the previous locking model for open vs. close afaict, before the 
> down() was introduced in the release_dev path that was supposed to be 
> atomic with respect to init_dev().

We want to move away from lock_kernel()-based locking.

> 
> > I think it would be better to _increase_ the tty_sem coverage in
> > release_dev() and to make sure that all callers of init_dev() are using
> > tty_sem (they are).
> > 
> > One approach would be to require that all callers of release_dev() hold
> > tty_sem, and make release_dev() drop and reacquire tty_sem in those cases
> > where release_dev() needs to go to sleep when waiting for other threads of
> > control to reelase the tty's resources.
> > 
> 
> Indeed, the situation would be improved if it was held around the 
> driver->close() routine. This routine does sometimes look at tty->count 
> value, see con_close(), where in fact the tty_sem is added to avoid just 
> this problem. However, it is incorrect as one can see in release_dev() the 
> schedule(), can cause the tty->count to change via tty_open(). However, i 
> think this is an extremely rare corner case, b/c con_close() keys off 
> tty->count of 1, which implies that this is the last close() and thus the 
> schedule for 'write_wait' would seem impossible, although AL Viro has 
> said that it is possible in this case. Thus, dropping the tty_sem and 
> reacquiring it, probably isn't good, b/c the driver->close() routines can 
> free resources based upon tty->count==1. 

Maybe we can just hold tty_sem across that schedule() in release_dev().

If not, then maybe retest ->count and take avoiding action if it looks like
some other thread is trying to resurrect the tty.  Obviously this is a much
poorer approach.

> The patch was written as the least invasive and low risk way to fix a 
> nasty race condition, which has the potential to corrupt data. The oops in 
> vt_ioctl has also been seen on system boots with some frequency. The patch 
> imo, returns the the tty_open vs. tty_close paths to their original 
> locking assumptions which have been well tested.
> 

I don't think it does, and the original lock_kernel-based locking is
obsolete.

Please, let's do this properly, with real locks.

-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [email protected]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

[Index of Archives]     [Kernel Newbies]     [Netfilter]     [Bugtraq]     [Photo]     [Stuff]     [Gimp]     [Yosemite News]     [MIPS Linux]     [ARM Linux]     [Linux Security]     [Linux RAID]     [Video 4 Linux]     [Linux for the blind]     [Linux Resources]
  Powered by Linux