Date: Wed, 02 May 2001 20:29:55 -0500 From: Douglas Egan <degan@calcon.net> To: Greg Lehey <grog@lemis.com> Cc: Brian Skrab <bgs@pinky.us.net>, freebsd-questions@FreeBSD.ORG Subject: Re: Vinum & FreeBDS 4.3-stable??? Message-ID: <3AF0B492.8AB9DCCB@calcon.net> References: <Pine.BSF.4.32.0105021459520.326-100000@pinky.us.net> <20010503084922.B72846@wantadilla.lemis.com>
next in thread | previous in thread | raw e-mail | index | archive | help
Does anyone have any comparisons or comments on RAID5 implemented with vinum and FreeBSD vs RAID5 implemented with RAIDFrame and NetBSD? My nfs/smb server is currently running NetBSD 1.5 with RAID5 using RAIDFrame. When I configured that system, I was unaware of vinum's support of RAID5. Can you comment on vinums stability with RAID5? I have currently configured a 4 disk array; 3 15gig ATA ultra66 and 1 18gig scsi. This is split up into 2 22gig arrays with no hot-spares. So far the RAIDFrame has performed admirably. I would just as soon have all my machines running FreeBSD. Thanks, Doug Egan Greg Lehey wrote: > On Wednesday, 2 May 2001 at 15:13:26 -0400, Brian Skrab wrote: > > Hello, > > > > I recently moved into a nice new Athlon 800Mhz motherboard and > > decided to start with a fresh install of FreeBSD 4.2->4.3-stable. > > With all the spring cleaning going on, I decided to re-create my > > Vinum RAID5 volume for completeness. I have 3 20GB IDE drives > > plugged into the two IDE controllers on the motherboard and have > > been attempting to use the following configuration to build a stable > > RAID5 volume: > > > > ----- vinum.config > > drive d1 device /dev/ad0s1e > > drive d3 device /dev/ad2s1e > > drive d2 device /dev/ad3s1e > > volume raid > > plex name raid.p0 org raid5 512s vol raid > > You don't want to use a stripe size which is a power of 2. That's > described in the man pages, of course. > > > sd name raid.p0.s0 drive d1 plex raid.p0 len 39880704s driveoffset 265s plexoffset 0s > > sd name raid.p0.s1 drive d2 plex raid.p0 len 39880704s driveoffset 265s plexoffset 512s > > sd name raid.p0.s2 drive d3 plex raid.p0 len 39880704s driveoffset 265s plexoffset 1024s > > ----- > > > > I run 'vinum create -f vinum.config', then 'vinum init raid.p0'. > > After the subdisks have completed initialization, I run > > 'newfs -v /dev/vinum/raid' which succeeds. Immediately afterwards, > > running 'fsck -y' results in hundreds of errors reading: > > Strange. > > > Finally, I stop and start vinum, then mount the volume. Everything > > appears to work. I can put files on the volume, and navigate thru > > any directory structures I create, but when I read files from the > > volume, I receive corrupted data. This is especially evident in > > .tar.gz archives which spit archive errors when reading the archive > > contents. > > Well, there's obviously something very wrong. > > > Is there a problem with my configuration? > > I don't know. > > > Any optimization recommendations are welcome. This configuration > > did work with a 4.2-stable system before I moved to the new > > motherboard. Is there some difference between vinum on 4.2 and 4.3 > > that might cause a situation like this? > > None that I know of. > > > Any advice will be greatly appreciated. > > How about reading http://www.vinumvm.org/vinum/how-to-debug.html and > giving me the information I ask for there. > > Greg > -- > When replying to this message, please copy the original recipients. > If you don't, I may ignore the reply. > For more information, see http://www.lemis.com/questions.html > Finger grog@lemis.com for PGP public key > See complete headers for address and phone numbers > > To Unsubscribe: send mail to majordomo@FreeBSD.org > with "unsubscribe freebsd-questions" in the body of the message To Unsubscribe: send mail to majordomo@FreeBSD.org with "unsubscribe freebsd-questions" in the body of the message
Want to link to this message? Use this URL: <https://mail-archive.FreeBSD.org/cgi/mid.cgi?3AF0B492.8AB9DCCB>