Date: Wed, 06 Sep 2000 22:00:34 -0400 From: Mike Tancsa <mike@sentex.net> To: Greg Lehey <grog@lemis.com> Cc: stable@FreeBSD.ORG Subject: Re: Vinum (was: RAID) Message-ID: <4.2.2.20000906212120.03316380@mail.sentex.net> In-Reply-To: <20000907103313.C7718@wantadilla.lemis.com> References: <4.2.2.20000906000615.03226880@mail.sentex.net> <00090310375700.05988@www.runapplications.com> <4.2.2.20000904120154.07455bd0@mail.sentex.net> <20000905101653.A49732@wantadilla.lemis.com> <4.2.2.20000904204407.033e2920@mail.sentex.net> <20000905115126.A14470@myhakas.matti.ee> <4.3.2.7.0.20000905090105.053a13f0@marble.sentex.ca> <20000905155032.A27690@futuresouth.com> <4.3.2.7.0.20000905165313.037652c0@marble.sentex.ca> <20000906084835.B21113@wantadilla.lemis.com> <4.2.2.20000906000615.03226880@mail.sentex.net>
next in thread | previous in thread | raw e-mail | index | archive | help
At 10:33 AM 9/7/2000 +0930, Greg Lehey wrote: >I'd really like to see the output of rawio, simply because it's more >repeatable from one platform to another. I will give it a try. > > As for testing, so far so good. Actually, I was quite taken aback > > by some of the results. It really does seem a lot faster, certainly > > from the limited testing I have done. In one, test where I blast > > email at the box as fast as I can from two outside hosts, it took > > half the time deliver mail to 13,000 user mail boxes as compared to > > the time it took on the 428 MegaRAID controller in a 3 disk striped > > config with the same physical drives involved. > >You're saying Vinum is faster? Yes vinum was faster in the mail test. I think the big hit has been the issue of RAID5 on 3 disks. For testing, I put my scratch disks (5 1 gig seagates I got at surplus) and got the following results 1gig segate on an ahc -------Sequential Output-------- ---Sequential Input-- --Random-- -Per Char- --Block--- -Rewrite-- -Per Char- --Block--- --Seeks--- Machine MB K/sec %CPU K/sec %CPU K/sec %CPU K/sec %CPU K/sec %CPU /sec %CPU 400 4684 20.0 4342 5.8 2225 4.0 4466 28.0 4635 3.4 140.6 1.1 5 stripe segate (vinum) -------Sequential Output-------- ---Sequential Input-- --Random-- -Per Char- --Block--- -Rewrite-- -Per Char- --Block--- --Seeks--- Machine MB K/sec %CPU K/sec %CPU K/sec %CPU K/sec %CPU K/sec %CPU /sec %CPU 400 21745 91.5 21862 31.0 3749 8.5 9325 59.3 11582 13.1 348.5 3.8 1000 21484 91.2 20910 29.0 3796 8.3 9429 59.7 11417 12.6 186.7 2.4 raid5 400 1446 6.2 1427 1.9 1191 2.1 6597 41.6 6942 5.4 372.3 3.2 raid10 400 9012 39.1 8692 12.1 3434 6.4 7971 50.3 8209 6.4 377.2 3.0 1gig on amr428 -Per Char- --Block--- -Rewrite-- -Per Char- --Block--- --Seeks--- Machine MB K/sec %CPU K/sec %CPU K/sec %CPU K/sec %CPU K/sec %CPU /sec %CPU raid5 400 10204 43.4 10015 13.8 3106 5.3 7594 47.6 8142 6.1 218.1 1.6 1000 10447 44.5 10593 14.6 3370 5.8 7628 48.0 8200 6.1 105.3 0.9 What I find interesting is the raid5 on 5 disks (SEAGATE ST31055W 40.000MB/s (old 5400 rpm slow pokes, good for RAID5 'wreckin/testin')) I see better read write performance, than on one disk. But, when only using 3 drives in a RAID 5, I see really crappy performance (which I realize now is expected). So, when I was looking at da0 (ATLAS IV 80.000MB/s SCSI-3 device) -------Sequential Output-------- ---Sequential Input-- --Random-- -Per Char- --Block--- -Rewrite-- -Per Char- --Block--- --Seeks--- Machine MB K/sec %CPU K/sec %CPU K/sec %CPU K/sec %CPU K/sec %CPU /sec %CPU 400 20072 84.3 19866 26.7 7477 13.6 14591 91.6 19343 14.9 288.2 2.3 1000 19247 81.6 18716 25.3 7913 17.8 14787 93.4 19310 21.4 146.4 1.8 and looked at the drives in the amr in a 3 drive RAID5 -------Sequential Output-------- ---Sequential Input-- --Random-- -Per Char- --Block--- -Rewrite-- -Per Char- --Block--- --Seeks--- Machine MB K/sec %CPU K/sec %CPU K/sec %CPU K/sec %CPU K/sec %CPU /sec %CPU 400 8984 38.5 8881 12.6 4189 7.2 9420 59.1 9797 7.4 409.5 3.1 1000 9275 39.6 9186 12.6 4341 7.5 9840 61.9 10229 7.7 203.7 1.8 the figures looked disappointing to say the least. I imagine, I would see a fairly big improvement in the amr if I had 2 more of these drives to test with.... Which in the end, is what I will probably go with. RAID5+ 1 hot spare. > > The other neat thing I have found so far, was that on a 256MB > > machine, comparing the amr, mlx, da, ad to the vinum drive, the > > vinum tests gave me the best even distribution of multiple processes > > blasting on the disk. Running 15 bonnie -s 100 at once, all the > >Hmm. I wonder if this is a coincidence. I don't think it's a plus >for Vinum, anyway. I suspect that somewhere in the system we have a >problem with process balancing. It's very obvious initializing Vinum >subdisks, where the elapsed time for performing identical operations >in parallel can vary by a factor of 2:1. The list is rather large, but these particular results can be found at http://www.simianscience.com/multi.txt Some of them are way out of whack. If I have time at the office tomorrow morning, I will get the tests re-run with raw-io. BTW, One quick test I just did from the equipment as is, ad7: 19595MB <QUANTUM FIREBALLP LM20.5> [39813/16/63] at ata3-master using UDMA66 on a promise ata66 vs da0: <QUANTUM ATLAS IV 9 WLS 0B0B> Fixed Direct Access SCSI-3 device da0: 80.000MB/s transfers (40.000MHz, offset 31, 16bit), Tagged Queueing Enabled da0: 8761MB (17942584 512 byte sectors: 255H 63S/T 1116C) Random read Sequential read Random write Sequential write ID K/sec /sec K/sec /sec K/sec /sec K/sec /sec ad7e 1573.7 98 1348.4 82 1804.0 112 2845.7 174 da0e 1889.2 117 1935.6 118 1611.4 99 1736.5 106 ---Mike -------------------------------------------------------------------- Mike Tancsa, tel +1 519 651 3400 Network Administration, mike@sentex.net Sentex Communications www.sentex.net Cambridge, Ontario Canada www.sentex.net/mike To Unsubscribe: send mail to majordomo@FreeBSD.org with "unsubscribe freebsd-stable" in the body of the message
Want to link to this message? Use this URL: <https://mail-archive.FreeBSD.org/cgi/mid.cgi?4.2.2.20000906212120.03316380>