Re: [00/17] Large Blocksize Support V3

From: William Lee Irwin III
Date: Thu Apr 26 2007 - 09:51:11 EST

On Thu, 26 Apr 2007, Nick Piggin wrote:
>> OK, I would like to see them. And also discussions of things like why
>> we shouldn't increase PAGE_SIZE instead.

On Thu, Apr 26, 2007 at 12:34:50AM -0700, Christoph Lameter wrote:
> Because 4k is a good page size that is bound to the binary format? Frankly
> there is no point in having my text files in large page sizes. However,
> when I read a dvd then I may want to transfer 64k chunks or when use my
> flash drive I may want to transfer 128k chunks. And yes if a scientific
> application needs to do data dump then it should be able to use very high
> page sizes (megabytes, gigabytes) to be able to continue its work while
> the huge dumps runs at full I/O speed ...

It's possible to divorce PAGE_SIZE from the binary formats, though I
found it difficult to keep up with the update treadmill. Maybe it's
like hch says and I just needed to find more and better API cleanups.
I've only not tried to resurrect it because it's too much for me to do
on my own. I essentially collapsed under the weight of it and my 2.5.x
codebase ended up worse than Katrina as a disaster, which I don't want
to repeat and think collaborators or a different project lead from
myself are needed to avoid that happening again.

It's unclear how much the situation has changed since 32-bit workload
feasibility issues have been relegated to ignorable or deliberate
"f**k 32-bit" status. The effect is doubtless to make it easier, though
to what degree I'm not sure.

Anyway, if that's being kicked around as an alternative, it could be
said that I have some insight into the issues surrounding it.

-- wli
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at
Please read the FAQ at