Date: Wed, 10 Nov 2004 14:31:10 -0700 From: secmgr <security@jim-liesl.org> To: msch@snafu.de Cc: freebsd-stable@freebsd.org Subject: Re: freebsd 5.3 have any problem with vinum ? Message-ID: <4192889E.8010506@jim-liesl.org> In-Reply-To: <200411071042.03382.msch@snafu.de> References: <02f201c4ba91$f9f95db0$33017f80@psique> <200411061217.06061.msch@snafu.de> <1099805401.4420.4.camel@emperor> <200411071042.03382.msch@snafu.de>
next in thread | previous in thread | raw e-mail | index | archive | help
ok, your instructions worked like a charm. So i'm running my nice 4
member SCSI gvinum raid5 array (with softupdates turned on), and it's
zipping along. Now I need to test just how robust this is. camcontrol
is too nice. I want to test a more real world failure. I'm running
dbench and just pull one of the drives. My expectation is that I
should see a minor pause, and then the array continue in some slower,
degraded mode. What I get is a kernel trap 12 (boom!). I reboot, and
it will not mount the degraded set till I replace the drive.
I turned off softupdates, and had the same thing happen. Is this a
bogus test? Is it reasonable to expect that a scsi drive failure should
of been tolerated w/o crashing?
(bunch of scsi msgs to console)
sub-disk down
plex degraded
g_access failed:6
trap 12
page fault while in kernel mode
cpuid=1 apic id=01
fault virtual address =0x18c
fault code supervisor write, page not present
instruction pointer =0x8:0xc043d72c
stack pointer =0x10:cbb17bf0
code segment =base 0x0, limit 0xfff, type 0x1b
=DPL0, pres1,def32,gran1
Processor flags interupt enabled, resume,IOPL=0
current process 22(irq11:ahc1)
Matthias Schuendehuette wrote:
>gvinum> start <plexname>
>
>This (as far as I investigated :-)
>
>a) initializes a newly created RAID5-plex or
>
>b) recalculates parity informations on a degraded RAID5-plex with
> a new replaced subdisk.
>
>So, a 'gvinum start raid5.p0' initializes my RAID5-plex if newly
>created. You can monitor the initialization process with subsequent
>'gvinum list' commands.
>
>If you degrade a RAID5-plex with 'camcontrol stop <diskname>' (in case
>of SCSI-Disks) and 'repair' it afterwards with 'camcontrol start
><diskname>', the 'gvinum start raid5.p0' (my volume here is called
>'raid5') command recalculates the parity and revives the subdisk which
>was on disk <diskname>.
>
>
Want to link to this message? Use this URL: <https://mail-archive.FreeBSD.org/cgi/mid.cgi?4192889E.8010506>
