In article <[email protected]>,
fitzboy <[email protected]> wrote:
>I've got a very large (2TB) proprietary database that is kept on an XFS
>partition
Ehh.. partition ? Let me tell you about a very common mistake:
>under a debian 2.6.8 kernel. It seemed to work well, until I
>recently did some of my own tests and found that the performance should
>be better then it is...
>
>basically, treat the database as just a bunch of random seeks. The XFS
>partition is sitting on top of a SCSI array (Dell PowerVault) which has
>13+1 disks in a RAID5, stripe size=64k. I have done a number of tests
>that mimic my app's accesses and realized that I want the inode to be
>as large as possible (which in an intel box is only 2k), played with su
>and sw and got those to 64k and 13... and performance got better.
If your RAID has 64K stripes, and the XFS partition is also tuned
for 64K, you expect those two to match, right ? But I bet the start
of the partition isn't aligned to a multiple of 64K..
In those cases I just use the whole disk instead of a partition,
or I dump the partition table with sfdisk, move the start of the
partition I'm using to a multiple of X, and re-read it. You need
to re-mkfs the fs though.
Not sure if it has any impact in your case, just thought I'd mention it.
Mike.
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [email protected]
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
[Index of Archives]
[Kernel Newbies]
[Netfilter]
[Bugtraq]
[Photo]
[Stuff]
[Gimp]
[Yosemite News]
[MIPS Linux]
[ARM Linux]
[Linux Security]
[Linux RAID]
[Video 4 Linux]
[Linux for the blind]
[Linux Resources]