On Sun, Aug 26, 2001 at 12:06:42PM -0300, Rik van Riel <riel@conectiva.com.br> wrote:
> > It must be enough, unfortunately ;)
>
> Then you'll need to change the physical structure of
> your disks to eliminate seek time. ;)
well, you could send me 200GB flashdisks, that would certainly help, but
what else could I do that is hardware-cost-neutral?
(there are currently three disks on two ide channels, so there is one
obvious optimization left, but this is completely independent of the
problem ;)
I really think linux should be able to achieve what the hardwrae can,
rather than fixing linux vm shortcomings with faster disks.
(playing around with readahead did indeed give me a very noticable
performance improvement, and ~40 mbits is ok).
> Automatic scaling of readahead window and possibly more agressive
> drop-behind could help your system load.
well, the system "load" is very low (50% idle time ;) here is the top ouput
of the current (typical) load:
PID USER PRI NI SIZE RSS SHARE STAT %CPU %MEM TIME COMMAND
8390 root 20 0 95312 93M 1076 S 43.9 18.5 16:03 myhttpd
8661 root 17 -4 32464 31M 944 R < 28.4 6.2 1:07 get
6279 root 18 14 4856 4856 524 R N 27.3 0.9 122:19 dec
8396 root 10 0 95312 93M 1076 D 6.5 18.5 1:43 myhttpd
8395 root 11 0 95312 93M 1076 D 5.3 18.5 1:46 myhttpd
8394 root 9 0 95312 93M 1076 D 4.4 18.5 1:42 myhttpd
8682 root 19 0 1012 1012 800 R 4.2 0.1 0:01 top
myhttpd is the http serverr, doing about 4MB/s now @ 743 connections.
"get" is a process that reads usenet news from many different servers and
dec is a decoder that decoded news. The news spool is on a 20Gb, 5 disk
SCSI array, together with the system itself. The machine is a dual P-II
300.
procs memory swap io system cpu
r b w swpd free buff cache si so bi bo in cs us sy id
2 3 1 0 3004 11436 171676 0 0 225 170 303 75 30 45 25
# the above line was the total system uptime average, now vmstat 5 output:
1 4 2 0 3056 11060 165760 0 0 7094 103 6905 1046 31 51 18
1 4 2 0 3056 11264 165532 0 0 6173 183 6146 1051 30 41 29
2 4 2 0 3056 10988 167656 0 0 7402 150 6706 1204 31 48 21
0 6 0 0 3056 11196 167344 0 0 7249 265 6760 1318 30 47 23
0 4 0 0 3056 11336 166876 0 0 1718 190 4995 582 25 19 55
2 0 0 0 3056 11536 166988 0 0 1057 264 3264 313 22 12 65
2 5 1 0 2880 11332 152916 0 0 1776 121 2789 280 32 22 46
1 5 0 0 16108 11472 153984 0 0 1040 215 3255 248 29 15 56
1 4 3 0 3056 11624 166800 0 0 4406 179 3329 653 32 23 45
1 4 0 0 3056 10852 167636 0 0 6970 138 5521 1247 34 39 26
2 4 0 0 3056 11016 167440 0 0 7238 162 5997 1118 36 39 25
2 4 1 0 3056 11284 177332 0 0 6247 84 5206 1293 34 36 30
1 4 2 0 3052 11296 181564 0 0 7800 85 5493 1399 35 41 24
There are 4 reader threads ATM, and this coincides nicely with the 4
blocked tasks.
> so we have an idea of what kind of system load we're facing, and
> the active/inactive memory lines from /proc/meminfo ?
I then did: while sleep 5; do grep "\(^In\|^Act\)" </proc/meminfo;done
Active: 144368 kB
Inact_dirty: 29048 kB
Inact_clean: 192 kB
Inact_target: 19348 kB
Active: 154012 kB
Inact_dirty: 14092 kB
Inact_clean: 5556 kB
Inact_target: 19360 kB
Active: 164908 kB
Inact_dirty: 21212 kB
Inact_clean: 5428 kB
Inact_target: 19104 kB
Active: 169788 kB
Inact_dirty: 20652 kB
Inact_clean: 1224 kB
Inact_target: 18912 kB
Active: 147280 kB
Inact_dirty: 37444 kB
Inact_clean: 5080 kB
Inact_target: 19132 kB
Active: 151400 kB
Inact_dirty: 26604 kB
Inact_clean: 10280 kB
Inact_target: 19328 kB
Active: 157288 kB
Inact_dirty: 9312 kB
Inact_clean: 20988 kB
Inact_target: 19500 kB
Active: 160456 kB
Inact_dirty: 11908 kB
Inact_clean: 12112 kB
Inact_target: 19672 kB
> Indeed, something is going wrong ;)
>
> Lets find out exactly what so we can iron out this bug
> properly.
When that happened I can test wether massively increasing the number of
reader threads changes performance ;)
-- -----==- | ----==-- _ | ---==---(_)__ __ ____ __ Marc Lehmann +-- --==---/ / _ \/ // /\ \/ / pcg@goof.com |e| -=====/_/_//_/\_,_/ /_/\_\ XX11-RIPE --+ The choice of a GNU generation | | - To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/
This archive was generated by hypermail 2b29 : Fri Aug 31 2001 - 21:00:20 EST