In article <Pine.LNX.4.44.0208112109110.16360-100000@skynet> you wrote:
> It works by using kernel modules to get a definite view of what the kernel
> is at and to provide reliable, reproducible tests. Modules are divided
> up into 4 catagories. Core modules provide infrastructure for the tool.
> Sense modules tell what is going on in the VM. Test tests particular
> features and bench modules (none yet) will benchmark different sections
> of the VM.
This sounds more like a micro benchmark tool, which is a good start, but the
real problem with VM optimizations is, that they have to take into account
real world load and especially user experience.
A simple example is the fact, that an idle desktop box will feel very sluggy
if a user comes back after a few hours break, because all visible programs
are paged out. To improve this, one could think about adding a flag to
applications like "connected to gui". This feature would need a test then,
which is no usual micro benchmark.
So I think it is a good idea to avoid to introduce slow operations in
hot code path, but it does not help much the developers in the problem of
simulating workload and measuring the interactive and real throughput.
But perhaps you can take this into account?
Greetings
Bernd
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
This archive was generated by hypermail 2b29 : Thu Aug 15 2002 - 22:00:26 EST