On Thu, Apr 26, 2007 at 08:44:36PM +0400, Oleg Nesterov wrote:
> On 04/26, Jarek Poplawski wrote:
> >
> > On Wed, Apr 25, 2007 at 04:47:14PM +0400, Oleg Nesterov wrote:
> > ...
> > > > > > + spin_lock_irq(&cwq->lock);
> > > > > > + /* CPU_DEAD in progress may change cwq */
> > > > > > + if (likely(cwq == get_wq_data(work))) {
> > > > > > + list_del_init(&work->entry);
> > > > > > + __set_bit(WORK_STRUCT_PENDING, work_data_bits(work));
> > > > > > + retry = try_to_del_timer_sync(&dwork->timer) < 0;
> > > > > > + }
> > > > > > + spin_unlock_irq(&cwq->lock);
> > > > > > + } while (unlikely(retry));
> > >
> > > > 1. If delayed_work_timer_fn of this work is fired and is waiting
> > > > on the above spin_lock then, after above spin_unlock, the work
> > > > will be queued.
> > >
> > > No, in that case try_to_del_timer_sync() returns -1.
> >
> > Yes. But I think it's safe only after moving work_clear_pending
> > in run_workqueue under a lock; probably otherwise there is a
> > possibility this flag could be cleared, after above unlock.
>
> It doesn't matter in this particular case because we are going to retry
> anyway. But yes, this patch moves work_clear_pending() under lock, because
> otherwise it could be cleared by run_workqueue() if this work is about
> to be executed, but was already deleted from list.
...and it seems to be the same what I meant...
I wanted only to make agree (now it's only for historical reasons)
the lock on _PENDING could matter in run_workqueue.
Jarek P.
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [email protected]
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
[Index of Archives]
[Kernel Newbies]
[Netfilter]
[Bugtraq]
[Photo]
[Stuff]
[Gimp]
[Yosemite News]
[MIPS Linux]
[ARM Linux]
[Linux Security]
[Linux RAID]
[Video 4 Linux]
[Linux for the blind]
[Linux Resources]