Re: 2.6.18-rc3-git3 - XFS - BUG: unable to handle kernel NULL pointer dereference at virtual address 00000078

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On 10/08/06, Jesper Juhl <[email protected]> wrote:
On 08/08/06, Nathan Scott <[email protected]> wrote:
> On Tue, Aug 08, 2006 at 10:37:49AM +0200, Jesper Juhl wrote:
> > On 08/08/06, Nathan Scott <[email protected]> wrote:
> > > On Mon, Aug 07, 2006 at 08:39:49PM -0700, Avuton Olrich wrote:
> > > > On 8/6/06, Nathan Scott <[email protected]> wrote:
> > > > > On Sun, Aug 06, 2006 at 12:05:43PM +0800, Tony.Ho wrote:
> > > > > > I'm sorry about prev mail. I test on a wrong kernel.
> > > > > > The panic is not appear again,
> > > >
> > > > Using 2.6.18-rc4, is this the bug that this thread refers to?
> > >
> > > Yes, try http://oss.sgi.com/archives/xfs/2006-08/msg00054.html
> > > and lemme know what happens - thanks.
> > >
> > Come wednesday would you like me to try rc4 + that patch instead of
> > rc3-git3 with your previous patch?
>
> Yep, ignore that first patch.  Thanks!
>

Ok, I booted the server with 2.6.18-rc4 + your patch. Things went well
for ~3 hours and then blew up - not in the same way though.

The machine was under pretty heavy load recieving data via rsync when
the following happened :

Filesystem "dm-51": XFS internal error xfs_trans_cancel at line 1138
of file fs/xfs/xfs_trans.c.  Caller 0xc0210e3f
 [<c0103a3c>] show_trace_log_lvl+0x152/0x165
 [<c0103a5e>] show_trace+0xf/0x13
 [<c0103b59>] dump_stack+0x15/0x19
 [<c0213474>] xfs_trans_cancel+0xcf/0xf8
 [<c0210e3f>] xfs_rename+0x64d/0x936
 [<c0226286>] xfs_vn_rename+0x48/0x9f
 [<c016584e>] vfs_rename_other+0x99/0xcb
 [<c0165a36>] vfs_rename+0x1b6/0x1eb
 [<c0165bda>] do_rename+0x16f/0x193
 [<c0165c45>] sys_renameat+0x47/0x73
 [<c0165c98>] sys_rename+0x27/0x2b
 [<c0102ae3>] syscall_call+0x7/0xb
 [<b7e0a681>] 0xb7e0a681
 [<c0213474>] xfs_trans_cancel+0xcf/0xf8
 [<c0210e3f>] xfs_rename+0x64d/0x936
 [<c0210e3f>] xfs_rename+0x64d/0x936
 [<c0226286>] xfs_vn_rename+0x48/0x9f
 [<c01626bb>] exec_permission_lite+0x46/0xcd
 [<c0162acb>] __link_path_walk+0x4d/0xd0a
 [<c016f8df>] mntput_no_expire+0x1b/0x78
 [<c01637f3>] link_path_walk+0x6b/0xc4
 [<c016584e>] vfs_rename_other+0x99/0xcb
 [<c022623e>] xfs_vn_rename+0x0/0x9f
 [<c0165a36>] vfs_rename+0x1b6/0x1eb
 [<c0165bda>] do_rename+0x16f/0x193
 [<c0154e6b>] sys_fchmodat+0xc2/0xef
 [<c015508f>] sys_lchown+0x50/0x52
 [<c016231b>] do_getname+0x4b/0x73
 [<c0165c45>] sys_renameat+0x47/0x73
 [<c0165c98>] sys_rename+0x27/0x2b
 [<c0102ae3>] syscall_call+0x7/0xb
xfs_force_shutdown(dm-51,0x8) called from line 1139 of file
fs/xfs/xfs_trans.c.  Return address = 0xc0229395
Filesystem "dm-51": Corruption of in-memory data detected.  Shutting
down filesystem: dm-51
Please umount the filesystem, and rectify the problem(s)
xfs_force_shutdown(dm-51,0x1) called from line 424 of file
fs/xfs/xfs_rw.c.  Return address = 0xc0229395

I was doing an lvmextend +xfs_resize of a different (XFS) filesystem
on the same server at roughly the same time. But I'm not sure if
that's related.

I'm currently running xfs_repair on the fs that blew up.


I think lvextend and xfs_growfs are not the cause of the problem since
I've extended 6 more filesystems since one of them blew up and the
problem reported above has not manifested itself again (yet).

--
Jesper Juhl <[email protected]>
Don't top-post  http://www.catb.org/~esr/jargon/html/T/top-post.html
Plain text mails only, please      http://www.expita.com/nomime.html
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [email protected]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

[Index of Archives]     [Kernel Newbies]     [Netfilter]     [Bugtraq]     [Photo]     [Stuff]     [Gimp]     [Yosemite News]     [MIPS Linux]     [ARM Linux]     [Linux Security]     [Linux RAID]     [Video 4 Linux]     [Linux for the blind]     [Linux Resources]
  Powered by Linux