xfs_da_do_buf: bno 16777216
^^^^^^^
That is the XFS bug from 2.6.17 -> 2.6.17.6, if you have never fixed your
FS to begin with, it will stay corrupted. Check the XFS FAQ on how to
fix.
Backup your data first :)
Justin.
On Tue, 19 Sep 2006, Martin Orda wrote:
Morning,
Please do Cc me as I'm not subscribed. Thank you.
I've been experiencing issues with the XFS filesystem. I use SATA
Seagate ST3120026AS disk on which I have two partitions. First one is
used as XFS root partition and the second one is a LVM physical volume.
On my logical volumes I run XFS. I've started getting the lockups /
oopses after having been running the system for a couple of days. It
either results in the complete lock-up or IO operations become very
unstable and dmesg showing errors shown below.
Details:
distro: debian/sid
kernel: 2.6.17.13
xfsprogs: 2.8.11-1
lvm2: 2.02.06-2
I've tried rebooting into single mode to minimize the number of daemons
running and then trying to create files on the partition affected (/usr)
reading them, listing contents, etc. Nothing happened. But it started to
show errors once I issued apt-get install something.
After that when I tried to ls /usr I got the IO Error message. Please
find the kernel output below.
I've gotten this error several times:
0x0: 00 00 00 00 01 06 6f 47 1d 46 63 41 74 6f 6d 69
Filesystem "dm-0": XFS internal error xfs_da_do_buf(2) at line 2212 of file fs/xfs/xfs_da_btree.c. Caller 0xc030193d
<c0312b8a> xfs_corruption_error+0xe2/0x10a <c030193d> xfs_da_read_buf+0x30/0x35
<c033e474> kmem_zone_alloc+0x54/0xcb <c0300ed4> xfs_da_buf_make+0xe2/0x12f
<c0301887> xfs_da_do_buf+0x966/0x98d <c030193d> xfs_da_read_buf+0x30/0x35
<c030193d> xfs_da_read_buf+0x30/0x35 <c030c4f6> xfs_dir2_node_addname+0x480/0xb0e
<c030c4f6> xfs_dir2_node_addname+0x480/0xb0e <c0325743> xlog_write+0x70d/0x7f5
<c0305bf5> xfs_dir2_isleaf+0x1b/0x50 <c0306553> xfs_dir2_createname+0xfc/0x131
<c033e506> kmem_zone_zalloc+0x1b/0x43 <c033c268> xfs_link+0x31b/0x4fb
<c03450f8> xfs_vn_link+0x43/0x90 <c01768cb> mntput_no_expire+0x13/0x72
<c016b2ef> link_path_walk+0x65/0xcb <c03380db> xfs_access+0x34/0x3a
<c0345381> xfs_vn_permission+0x0/0x13 <c0345390> xfs_vn_permission+0xf/0x13
<c0169105> permission+0x9f/0xa4 <c0169227> vfs_link+0x11d/0x168
<c016bfa3> sys_linkat+0xdd/0xf9 <c015b8bd> sys_utime+0x48/0x12c
<c041311d> _atomic_dec_and_lock+0x2d/0x50 <c01729f3> dput+0xfa/0x130
<c015d122> __fput+0x131/0x194 <c016bfee> sys_link+0x2f/0x33
<c0102af3> sysenter_past_esp+0x54/0x75
And this one just once:
xfs_da_do_buf: bno 16777216
dir: inode 17256410
Filesystem "dm-0": XFS internal error xfs_da_do_buf(1) at line 2119 of file fs/xfs/xfs_da_btree.c. Caller 0xc030193d
<c030123e> xfs_da_do_buf+0x31d/0x98d <c030193d> xfs_da_read_buf+0x30/0x35
<c030193d> xfs_da_read_buf+0x30/0x35 <c030d4e7> xfs_dir2_leafn_lookup_int+0x2c0/0x4f6
<c030d4e7> xfs_dir2_leafn_lookup_int+0x2c0/0x4f6 <c030193d> xfs_da_read_buf+0x30/0x35
<c030b88e> xfs_dir2_node_removename+0x30a/0x4ef <c030b88e> xfs_dir2_node_removename+0x30a/0x4ef
<c0306777> xfs_dir2_removename+0xee/0xf0 <c033be2e> xfs_remove+0x389/0x4a8
<c0345381> xfs_vn_permission+0x0/0x13 <c034515c> xfs_vn_unlink+0x17/0x3b
<c01768cb> mntput_no_expire+0x13/0x72 <c016b2ef> link_path_walk+0x65/0xcb
<c03329b1> xfs_trans_unlocked_item+0x28/0x3c <c03380db> xfs_access+0x34/0x3a
<c0345381> xfs_vn_permission+0x0/0x13 <c0345390> xfs_vn_permission+0xf/0x13
<c0169105> permission+0x9f/0xa4 <c01696d2> may_delete+0x2b/0x10d
<c0169c8f> vfs_unlink+0x9c/0xf5 <c016ba7b> do_unlinkat+0xaf/0x12c
<c0102af3> sysenter_past_esp+0x54/0x75
Filesystem "dm-0": XFS internal error xfs_trans_cancel at line 1150 of file fs/xfs/xfs_trans.c. Caller 0xc033be8b
<c0331cea> xfs_trans_cancel+0xe1/0x102 <c033be8b> xfs_remove+0x3e6/0x4a8
<c033be8b> xfs_remove+0x3e6/0x4a8 <c0345381> xfs_vn_permission+0x0/0x13
<c034515c> xfs_vn_unlink+0x17/0x3b <c01768cb> mntput_no_expire+0x13/0x72
<c016b2ef> link_path_walk+0x65/0xcb <c03329b1> xfs_trans_unlocked_item+0x28/0x3c
<c03380db> xfs_access+0x34/0x3a <c0345381> xfs_vn_permission+0x0/0x13
<c0345390> xfs_vn_permission+0xf/0x13 <c0169105> permission+0x9f/0xa4
<c01696d2> may_delete+0x2b/0x10d <c0169c8f> vfs_unlink+0x9c/0xf5
<c016ba7b> do_unlinkat+0xaf/0x12c <c0102af3> sysenter_past_esp+0x54/0x75
xfs_force_shutdown(dm-0,0x8) called from line 1151 of file fs/xfs/xfs_trans.c. Return address = 0xc0348379
Filesystem "dm-0": Corruption of in-memory data detected. Shutting down filesystem: dm-0
Please umount the filesystem, and rectify the problem(s)
This problem affects 2 of my systems which don't run the same hardware
but have roughtly the same software and config. Can you please advice
if this is related to a bug in kernel or userland and suggest some fix
for that if possible?
BTW: please let me know if I've forgotten to mention some details that
you could find helpful. Thanks for your help.
--
Kind regards,
Martin Orda
http://securityshells.com
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [email protected]
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [email protected]
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
[Index of Archives]
[Kernel Newbies]
[Netfilter]
[Bugtraq]
[Photo]
[Stuff]
[Gimp]
[Yosemite News]
[MIPS Linux]
[ARM Linux]
[Linux Security]
[Linux RAID]
[Video 4 Linux]
[Linux for the blind]
[Linux Resources]