On Thursday 19 April 2007 19:35:04 Christoph Lameter wrote:
> Variable Order Page Cache Patchset
>
> This patchset modifies the core VM so that higher order page cache pages
> become possible. The higher order page cache pages are compound pages
> and can be handled in the same way as regular pages.
>
> The order of the pages is determined by the order set up in the mapping
> (struct address_space). By default the order is set to zero.
> This means that higher order pages are optional. There is no attempt here
> to generally change the page order of the page cache. 4K pages are effective
> for small files.
>
> However, it would be good if the VM would support I/O to higher order pages
> to enable efficient support for large scale I/O. If one wants to write a
> long file of a few gigabytes then the filesystem should have a choice of
> selecting a larger page size for that file and handle larger chunks of
> memory at once.
>
> The support here is only for buffered I/O and only for one filesystem
(ramfs).
> Modification of other filesystems to support higher order pages may require
> extensive work of other components of the kernel. But I hope this shows that
> there is a relatively easy way to that goal that could be taken in steps..
>
> Note that the higher order pages are subject to reclaim. This works in
general
> since we are always operating on a single page struct. Reclaim is fooled to
> think that it is touching page sized objects (there are likely issues to be
> fixed there if we want to go down this road).
>
> What is currently not supported:
> - Buffer heads for higher order pages (possible with the compound pages in
mm
> that do not use page->private requires upgrade of the buffer cache
layers).
> - Higher order pages in the block layer etc.
> - Mmapping higher order pages
>
> Note that this is proof-of-concept. Lots of functionality is missing and
> various issues have not been dealt with. Use of higher order pages may cause
> memory fragmentation. Mel Gorman's anti-fragmentation work is probably
> essential if we want to do this. We likely need actual defragmentation
> support.
>
> The main point of this patchset is to demonstrates that it is basically
> possible to have higher order support with straightforward changes to the
> VM.
>
> The ramfs driver can be used to test higher order page cache functionality
> (and may help troubleshoot the VM support until we get some real filesystem
> and real devices supporting higher order pages).
>
> If you apply this patch and then you can f.e. try this:
>
> mount -tramfs -o10 none /media
>
> Mounts a ramfs filesystem with order 10 pages (4 MB)
>
> cp linux-2.6.21-rc7.tar.gz /media
>
> Populate the ramfs. Note that we allocate 14 pages of 4M each
> instead of 13508..
>
> umount /media
>
> Gets rid of the large pages again
>
> Comments appreciated.
>
> -
> To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
> the body of a message to [email protected]
> More majordomo info at http://vger.kernel.org/majordomo-info.html
> Please read the FAQ at http://www.tux.org/lkml/
>
Hello,
This is exactly what I wanted some time ago,
Thank you very much, I was almost thinking of doing this myself
(but decided that it is too difficult now for me and maybe doesn't worth the
effort)
I want to point out on number of problems that this will solve (and reasons I
wanted to do that)
First of all, today, packet writing on cd/dvd doesn't work well, it is very
slow because
now all file-systems are limited to 4k-barrier and cd/dvd can write only
32k/64k packets.
This is why a pktcdvd was written and it emulates those 4k sectors by doing
read/modify/write cycle
This cause a lot of seeks and read/writing switches and thus it is very slow.
By introducing a bigger that 4k page cache a dvd/cd can be divided is 64k/32k
blocks that will be read an written freely
(Although dvd can read 2k I don't think that reading a 64k block will hurt
since most of time drive is busy seeking and locating a specific sector)
Now I thinking to implement this in an other way, I mean I want to teach udf
filesystem to to packet writing on its own, bypassing disk cache (but not page
cache)
Secondary 32/64k limitation is present of flash devices too, so they can
benefit too, and I almost sure that future hard disks will use bigger block
size too.
To summarize I want to tell that bigger pagesize will allow devices that have
big hardware sectors to work fine in linux.
Best regards,
Maxim Levitsky
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [email protected]
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
[Index of Archives]
[Kernel Newbies]
[Netfilter]
[Bugtraq]
[Photo]
[Stuff]
[Gimp]
[Yosemite News]
[MIPS Linux]
[ARM Linux]
[Linux Security]
[Linux RAID]
[Video 4 Linux]
[Linux for the blind]
[Linux Resources]