Re: [ckrm-tech] Circular Locking Dependency Chain detected in containers code

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



I'm away from work at the moment and can't investigate fully, but it
looks as though this may be the same one that I mentioned in the
introductory email to the patchset. If so, it's a false positive -
there's a point in the container mount code where we need to lock a
newly-created (and hence guaranteed unlocked) directory inode while
holding container mutex. This makes the lockdep code think that
inode->i_mutex nests inside container_mutex, when in fact
container_mutex nests outside inode->i_mutex in all cases except this
one case where i_mutex can't possibly be locked.

I've not learned enough about lockdep yet to figure out how to shut it
up in this case.

Thanks,

Paul

On 8/6/07, Dhaval Giani <[email protected]> wrote:
> Hi Paul,
>
> I have hit upon a circular locking dependency while doing an rmdir on a
> directory inside the containers code. I believe that it is safe as no one
> should be able to rmdir when a container is getting mounted.
>
> To reproduce it, just do a rmdir inside the container.
>
> =======================================================
> [ INFO: possible circular locking dependency detected ]
> 2.6.23-rc1-mm2-container #1
> -------------------------------------------------------
> rmdir/4321 is trying to acquire lock:
>  (container_mutex){--..}, at: [<c03f3fe0>] mutex_lock+0x21/0x24
>
> but task is already holding lock:
>  (&inode->i_mutex){--..}, at: [<c03f3fe0>] mutex_lock+0x21/0x24
>
> which lock already depends on the new lock.
>
>
> the existing dependency chain (in reverse order) is:
>
> -> #1 (&inode->i_mutex){--..}:
>        [<c013bae1>] check_prev_add+0xae/0x18f
>        [<c013bc1c>] check_prevs_add+0x5a/0xc5
>        [<c013bee5>] validate_chain+0x25e/0x2cd
>        [<c013d892>] __lock_acquire+0x629/0x691
>        [<c013ddef>] lock_acquire+0x61/0x7e
>        [<c03f40b5>] __mutex_lock_slowpath+0xc8/0x230
>        [<c03f3fe0>] mutex_lock+0x21/0x24
>        [<c014749f>] container_get_sb+0x22c/0x283
>        [<c0174f02>] vfs_kern_mount+0x3a/0x73
>        [<c0186743>] do_new_mount+0x7e/0xdc
>        [<c0186d8b>] do_mount+0x178/0x191
>        [<c0186ff3>] sys_mount+0x66/0x9d
>        [<c0104cc2>] sysenter_past_esp+0x5f/0x99
>        [<ffffffff>] 0xffffffff
>
> -> #0 (container_mutex){--..}:
>        [<c013ba5e>] check_prev_add+0x2b/0x18f
>        [<c013bc1c>] check_prevs_add+0x5a/0xc5
>        [<c013bee5>] validate_chain+0x25e/0x2cd
>        [<c013d892>] __lock_acquire+0x629/0x691
>        [<c013ddef>] lock_acquire+0x61/0x7e
>        [<c03f40b5>] __mutex_lock_slowpath+0xc8/0x230
>        [<c03f3fe0>] mutex_lock+0x21/0x24
>        [<c01484f3>] container_rmdir+0x15/0x163
>        [<c017b711>] vfs_rmdir+0x59/0x8f
>        [<c017b7d3>] do_rmdir+0x8c/0xbe
>        [<c017b815>] sys_rmdir+0x10/0x12
>        [<c0104cc2>] sysenter_past_esp+0x5f/0x99
>        [<ffffffff>] 0xffffffff
>
> other info that might help us debug this:
>
> 2 locks held by rmdir/4321:
>  #0:  (&inode->i_mutex/1){--..}, at: [<c017b7b3>] do_rmdir+0x6c/0xbe
>  #1:  (&inode->i_mutex){--..}, at: [<c03f3fe0>] mutex_lock+0x21/0x24
>
> stack backtrace:
>  [<c0105ad0>] show_trace_log_lvl+0x12/0x22
>  [<c0105aed>] show_trace+0xd/0xf
>  [<c0105bc3>] dump_stack+0x14/0x16
>  [<c013b401>] print_circular_bug_tail+0x5b/0x64
>  [<c013ba5e>] check_prev_add+0x2b/0x18f
>  [<c013bc1c>] check_prevs_add+0x5a/0xc5
>  [<c013bee5>] validate_chain+0x25e/0x2cd
>  [<c013d892>] __lock_acquire+0x629/0x691
>  [<c013ddef>] lock_acquire+0x61/0x7e
>  [<c03f40b5>] __mutex_lock_slowpath+0xc8/0x230
>  [<c03f3fe0>] mutex_lock+0x21/0x24
>  [<c01484f3>] container_rmdir+0x15/0x163
>  [<c017b711>] vfs_rmdir+0x59/0x8f
>  [<c017b7d3>] do_rmdir+0x8c/0xbe
>  [<c017b815>] sys_rmdir+0x10/0x12
>  [<c0104cc2>] sysenter_past_esp+0x5f/0x99
>  =======================
> --
> regards,
> Dhaval
>
> I would like to change the world but they don't give me the source code!
>
> -------------------------------------------------------------------------
> This SF.net email is sponsored by: Splunk Inc.
> Still grepping through log files to find problems?  Stop.
> Now Search log events and configuration files using AJAX and a browser.
> Download your FREE copy of Splunk now >>  http://get.splunk.com/
> _______________________________________________
> ckrm-tech mailing list
> https://lists.sourceforge.net/lists/listinfo/ckrm-tech
>
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [email protected]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

[Index of Archives]     [Kernel Newbies]     [Netfilter]     [Bugtraq]     [Photo]     [Stuff]     [Gimp]     [Yosemite News]     [MIPS Linux]     [ARM Linux]     [Linux Security]     [Linux RAID]     [Video 4 Linux]     [Linux for the blind]     [Linux Resources]
  Powered by Linux