Date: Wed, 3 Nov 2004 21:00:09 GMT From: jim feldman <secmgr@jim-liesl.org> To: freebsd-gnats-submit@FreeBSD.org Subject: i386/73499: gvinum can't init raid5 set Message-ID: <200411032100.iA3L09sc026536@www.freebsd.org> Resent-Message-ID: <200411032110.iA3LAW19052971@freefall.freebsd.org>
next in thread | raw e-mail | index | archive | help
>Number: 73499
>Category: i386
>Synopsis: gvinum can't init raid5 set
>Confidential: no
>Severity: non-critical
>Priority: low
>Responsible: freebsd-i386
>State: open
>Quarter:
>Keywords:
>Date-Required:
>Class: sw-bug
>Submitter-Id: current-users
>Arrival-Date: Wed Nov 03 21:10:32 GMT 2004
>Closed-Date:
>Last-Modified:
>Originator: jim feldman
>Release: 5.3 RC2
>Organization:
>Environment:
FreeBSD greybrd.xxx.xxx.net 5.3-STABLE FreeBSD 5.3-STABLE #0: Tue Nov 2 03:52:27 MST 2004 root@greybrd.xxx.xxx.net:/usr/obj/usr/src/sys/GREYBRD i386
>Description:
had working raid 5 set under 5.3rc1 composed of 4 drives (scsi). I updated using cvsup and the 5.3-RELENG tag. after make buildworld && make buildkernel && make installkernel reboot to single, make installworld, mergemaster, I rebooted
all the sub disks in the raid 5 plex showed stale. If I use
"gvinum rm -r volname", it deletes what it should. If I re-create the set, it re-creates the set and the sub disks are still stale.
>How-To-Repeat:
create raid 5 plex based volume with gvinum under rc1. update to rc2 and watch plex become corrupt and unfixable
>Fix:
>Release-Note:
>Audit-Trail:
>Unformatted:
Want to link to this message? Use this URL: <https://mail-archive.FreeBSD.org/cgi/mid.cgi?200411032100.iA3L09sc026536>
