Skip site navigation (1)Skip section navigation (2)
Date:      Sat, 25 Mar 2006 14:43:16 +0100
From:      "Olivier Cochard" <olivier@freenas.org>
To:        freebsd-geom@freebsd.org
Subject:   Gvinum configuration changed after a simple reboot
Message-ID:  <3131aa530603250543g2b875564s@mail.gmail.com>

next in thread | raw e-mail | index | archive | help
Hi all!

it's me again torturing gvinum :-)

I've found a strange problem with creating gvinum RAID 5 volume.

I can reproduce this problem at each time.

I'm working with a VMware workstation PC with a 6.1-PRERELEASE and 3 hard
drive:
- ad0: 102Mb
- ad1: 204mb
- ad3: 306mb

And I want to test a RAID 5 volume with these 3 disks.

Then I create this config file:

$ cat /var/etc/raid.conf
drive disk_ad0 device /dev/ad0s1a
drive disk_ad1 device /dev/ad1s1a
drive disk_ad3 device /dev/ad3s1a
volume raid
plex org raid5 256k
sd length 0 drive disk_ad0
sd length 0 drive disk_ad1
sd length 0 drive disk_ad3

With this configuration I should obtain a 204Mb RAID5 volume (2 * 102Mb).
And gvinum confirm that's right with this output:

Software RAID information and status:
3 drives:
D disk_ad3              State: up    /dev/ad3s1a    A: 0/306 MB (0%)
D disk_ad1              State: up    /dev/ad1s1a    A: 0/204 MB (0%)
D disk_ad0              State: up    /dev/ad0s1a    A: 0/102 MB (0%)

1 volume:
V raid                  State: up    Plexes:       1    Size:        204 MB

1 plex:
P raid.p0            R5 State: up    Subdisks:     3    Size:        204 MB

3 subdisks:
S raid.p0.s2            State: up    D: disk_ad3     Size:        306 MB
S raid.p0.s1            State: up    D: disk_ad1     Size:        204 MB
S raid.p0.s0            State: up    D: disk_ad0     Size:        102 MB

I can launch a newfs on it without problem, and put file on it
... All seem ok.

But... when I reboot the VMWare workstation.. I obtain this:

Software RAID status:
3 drives:
D disk_ad0              State: up    /dev/ad0s1a    A: 0/102 MB (0%)
D disk_ad1              State: up    /dev/ad1s1a    A: 0/204 MB (0%)
D disk_ad3              State: up    /dev/ad3s1a    A: 0/306 MB (0%)

1 volume:
V raid                  State: up    Plexes:       1    Size:        613 MB

1 plex:
P raid.p0            R5 State: up    Subdisks:     3    Size:        613 MB

3 subdisks:
S raid.p0.s0            State: up    D: disk_ad0     Size:        102 MB
S raid.p0.s1            State: up    D: disk_ad1     Size:        204 MB
S raid.p0.s2            State: up    D: disk_ad3     Size:        306 MB

The total RAID volume space change... and I can't use it anymore!! (must be
delete).

$ fsck -t ufs /dev/gvinum/raid
Cannot find file system superblock
fsck_ufs: /dev/gvinum/raid: can't read disk label
** /dev/gvinum/raid
ioctl (GCINFO): Inappropriate ioctl for device

Any idea ?

Thanks,

Olivier

--
Olivier Cochard
FreeNAS main developer
http://www.freenas.org
Skype: callto://ocochard



Want to link to this message? Use this URL: <https://mail-archive.FreeBSD.org/cgi/mid.cgi?3131aa530603250543g2b875564s>