nfs3: possible recursive locking (Re: BUG: soft lockup detected on CPU#0! (2.6.18.2))

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Wed, Nov 22, 2006 at 01:22:25PM +0100, Rasmus B??g Hansen wrote:
[]
> >> (gitweb down, i can't check history of smbfs, and i have amd64 arch, anyway)
> >>> Nov 12 03:54:57 gere kernel: BUG: soft lockup detected on CPU#0!
> >>> Nov 12 03:54:57 gere kernel:  [softlockup_tick+170/195] softlockup_tick+0xaa/0xc3
> >>> Nov 12 03:54:57 gere kernel:  [update_process_times+56/137] update_process_times+0x38/0x89
> >>> Nov 12 03:54:57 gere kernel:  [smp_apic_timer_interrupt+105/117] smp_apic_timer_interrupt+0x69/0x75
> >>> Nov 12 03:54:57 gere kernel:  [smbiod+238/348] smbiod+0xee/0x15c
> >> this
> >>
> >>> Nov 12 03:54:57 gere kernel:  [apic_timer_interrupt+31/36] apic_timer_interrupt+0x1f/0x24
> >>> Nov 12 03:54:57 gere kernel:  [journal_init_revoke+49/678] journal_init_revoke+0x31/0x2a6
> >>> Nov 12 03:54:57 gere kernel:  [smbiod+238/348] smbiod+0xee/0x15c
> >> and this *may be* double (un)lock.
> >
> > Hopefully lock debugging will tell.
> 
> I got this - I think it was this morning (somehow kernel logging was
> disabled so I can't tell the exact time):

Did this like "my server froze. It was entirely dead and had to be power
cycled."?

> ----
> 
> =============================================
> [ INFO: possible recursive locking detected ]
> ---------------------------------------------
> nfsd/1788 is trying to acquire lock:
>  (&inode->i_mutex){--..}, at: [<c02cc35c>] mutex_lock+0x8/0xa
> 
> but task is already holding lock:
>  (&inode->i_mutex){--..}, at: [<c02cc35c>] mutex_lock+0x8/0xa
> 
> other info that might help us debug this:
> 2 locks held by nfsd/1788:
>  #0:  (hash_sem){----}, at: [<e0930d99>] exp_readlock+0x12/0x16 [nfsd]
>  #1:  (&inode->i_mutex){--..}, at: [<c02cc35c>] mutex_lock+0x8/0xa
> 
> stack backtrace:
>  [<c0103c10>] show_trace+0x27/0x2b
>  [<c0103d2d>] dump_stack+0x26/0x2a
>  [<c01353ba>] print_deadlock_bug+0xb5/0xba
>  [<c0135420>] check_deadlock+0x61/0x71
>  [<c0136da3>] __lock_acquire+0x334/0x9b6
>  [<c0137b3d>] lock_acquire+0x75/0x90
>  [<c02cb9a1>] __mutex_lock_slowpath+0x85/0x270
>  [<c02cc35c>] mutex_lock+0x8/0xa
>  [<e092cb75>] nfsd_setattr+0x3ca/0x574 [nfsd]
>  [<e092e3a8>] nfsd_create_v3+0x3a6/0x540 [nfsd]
>  [<e0934d4b>] nfsd3_proc_create+0x118/0x161 [nfsd]
>  [<e0929751>] nfsd_dispatch+0xd8/0x1ff [nfsd]
>  [<e08e8503>] svc_process+0x4e5/0x6da [sunrpc]
>  [<e09294e7>] nfsd+0x1cc/0x35e [nfsd]
>  [<c01010b9>] kernel_thread_helper+0x5/0xb
> 
> ----
> 
> Also it doesn't seem to be related to smbfs AFAICS - I am not enough

Yes, it doen't.

> into lock debugging to say if this is relevant, useful or just
> noise...

Do you think, i am? In prev. message there was smbiod() function in the
stach dump. And that, what i've pointed, was close enough to forest of
spinlocks in sources of that function. That's it.

Try to search archives for any nfsd bugs. I'm not interested in it, there
are many people on this.

____
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [email protected]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

[Index of Archives]     [Kernel Newbies]     [Netfilter]     [Bugtraq]     [Photo]     [Stuff]     [Gimp]     [Yosemite News]     [MIPS Linux]     [ARM Linux]     [Linux Security]     [Linux RAID]     [Video 4 Linux]     [Linux for the blind]     [Linux Resources]
  Powered by Linux