Skip site navigation (1)Skip section navigation (2)
Date:      Sun, 5 Dec 2004 19:16:07 -0800 (PST)
From:      orville weyrich <weyrich_comp@yahoo.com>
To:        freebsd-questions <freebsd-questions@freebsd.org>
Subject:   VINUM Disaster Recovery 
Message-ID:  <20041206031607.52973.qmail@web50710.mail.yahoo.com>

next in thread | raw e-mail | index | archive | help

[-- Attachment #1 --]
I have a 10 disk VINUM configuration and two of the
disks are trashed.  In theory there is still enough
redundant information to get things working again
without data loss.

Vinum has detected a configuration error (duh -- two
disks are toast, plus in recovery I accidently created
two more plexes) and taken upon itself to stop
configuration updates to prevent any further
corruption (thanks! :-).

At this point I have looked at
http://www.vinumvm.org/vinum/how-to-debug.html and
have run a command like the following:

( dd if=/dev/da9s1e skip=8 count=50 | tr -d
'\000-\011\200-\377' ; echo ) > da9s1e.log 

on all 10 disks to obtain a file containing each
disk's on-disk configuration.  As hoped, eight of the
disks show an output similar to the attached file
da1s1e.log (differing only as expected in the first
line).

See attached flog file for a sample output.

PLEASE HELP CONFIRM MY PLAN (FOLLOWING) FOR PROCEEDING
-- I DO NOT WANT TO DO ANYTHING DISASTEROUS.

My thought is that I need to turn on updates, then
delete the two unwanted plexes raid.p2 and
raid.p3(which were accidentally created), detach the
corrupt sdisks, and then hopefully VINUM will forget
about the two disks that are toast (or do I somehow
have to tell VINUM to forget the disks?).

My plan is as follows:

First, selectively start vinum:

vinum -> read /dev/da1s1e /dev/da2s1e /dev/da3s1e
/dev/da4s1e /dev/da5s1e /dev/da6s1e /dev/da7s1e
/dev/da8s1e

Second, enable configuration updates:

vinum->setdaemon 0

Third, save the configuration:

vinum->saveconfig

Fourth, stop and remove the two unwanted plexes and
all attached subdisks:

vinum->stop -f raid.p3
vinum->stop -f raid.p2
vinum->rm -r raid.p3
vinum->rm -r raid.p2

Fifth, stop and detach the corruped subdisks:

vinum->stop -f raid.p0.s0
vinum->stop -f raid.p0.s9
vinum->stop -f raid.p1.s4
vinum->stop -f raid.p1.s5

vinum->detach raid.p0.s0
vinum->detach raid.p0.s9
vinum->detach raid.p1.s4
vinum->detach raid.p1.s5


At this point I expect to have a functional volume
that can be mounted and backed up, prior to the next
step of reinstalling the crashed disks, creating new
subdisks, attaching them to the plexes, and
resynching. 

PLEASE CONFIRM MY APPROACH OR TELL ME WHERE I AM
WRONG!

Thanks

orville








__________________________________________________
Do You Yahoo!?
Tired of spam?  Yahoo! Mail has the best spam protection around 
http://mail.yahoo.com 
[-- Attachment #2 --]
IN VINObashful.weyrich.comahc0t03A'BA,volume raid state up
plex name raid.p0 state up org striped 1024s vol raid 
plex name raid.p1 state corrupt org striped 1024s vol raid 
plex name raid.p2 state faulty org striped 1024s vol raid 
plex name raid.p3 state faulty org striped 1024s vol raid 
sd name raid.p0.s0 drive ahc0t02 plex raid.p0 state up len 4405248s driveoffset 265s plexoffset 0s
sd name raid.p0.s1 drive ahc0t03 plex raid.p0 state up len 4405248s driveoffset 265s plexoffset 1024s
sd name raid.p0.s2 drive ahc0t04 plex raid.p0 state up len 4405248s driveoffset 265s plexoffset 2048s
sd name raid.p0.s3 drive ahc0t09 plex raid.p0 state up len 4405248s driveoffset 265s plexoffset 3072s
sd name raid.p0.s4 drive ahc0t10 plex raid.p0 state up len 4405248s driveoffset 265s plexoffset 4096s
sd name raid.p0.s5 drive ahc0t11 plex raid.p0 state up len 4405248s driveoffset 265s plexoffset 5120s
sd name raid.p0.s6 drive ahc0t12 plex raid.p0 state up len 4405248s driveoffset 265s plexoffset 6144s
sd name raid.p0.s7 drive ahc0t13 plex raid.p0 state up len 4405248s driveoffset 265s plexoffset 7168s
sd name raid.p0.s8 drive ahc0t14 plex raid.p0 state up len 4405248s driveoffset 265s plexoffset 8192s
sd name raid.p0.s9 drive *invalid* plex raid.p0 state up len 4405248s driveoffset 265s plexoffset 9216s
sd name raid.p1.s0 drive ahc0t11 plex raid.p1 state up len 4405248s driveoffset 4405513s plexoffset 0s
sd name raid.p1.s1 drive ahc0t12 plex raid.p1 state up len 4405248s driveoffset 4405513s plexoffset 1024s
sd name raid.p1.s2 drive ahc0t13 plex raid.p1 state up len 4405248s driveoffset 4405513s plexoffset 2048s
sd name raid.p1.s3 drive ahc0t14 plex raid.p1 state up len 4405248s driveoffset 4405513s plexoffset 3072s
sd name raid.p1.s4 drive *invalid* plex raid.p1 state obsolete len 4405248s driveoffset 4405513s plexoffset 4096s
sd name raid.p1.s5 drive ahc0t02 plex raid.p1 state up len 4405248s driveoffset 4405513s plexoffset 5120s
sd name raid.p1.s6 drive ahc0t03 plex raid.p1 state up len 4405248s driveoffset 4405513s plexoffset 6144s
sd name raid.p1.s7 drive ahc0t04 plex raid.p1 state up len 4405248s driveoffset 4405513s plexoffset 7168s
sd name raid.p1.s8 drive ahc0t09 plex raid.p1 state up len 4405248s driveoffset 4405513s plexoffset 8192s
sd name raid.p1.s9 drive ahc0t10 plex raid.p1 state up len 4405248s driveoffset 4405513s plexoffset 9216s
sd name raid.p2.s0 drive ahc0t02 plex raid.p2 state stale len 4405248s driveoffset 265s plexoffset 0s
sd name raid.p2.s1 drive ahc0t03 plex raid.p2 state stale len 4405248s driveoffset 265s plexoffset 1024s
sd name raid.p2.s2 drive ahc0t04 plex raid.p2 state stale len 4405248s driveoffset 265s plexoffset 2048s
sd name raid.p2.s3 drive ahc0t09 plex raid.p2 state stale len 4405248s driveoffset 265s plexoffset 3072s
sd name raid.p2.s4 drive ahc0t10 plex raid.p2 state stale len 4405248s driveoffset 265s plexoffset 4096s
sd name raid.p2.s5 drive ahc0t11 plex raid.p2 state stale len 4405248s driveoffset 265s plexoffset 5120s
sd name raid.p2.s6 drive ahc0t12 plex raid.p2 state stale len 4405248s driveoffset 265s plexoffset 6144s
sd name raid.p2.s7 drive ahc0t13 plex raid.p2 state stale len 4405248s driveoffset 265s plexoffset 7168s
sd name raid.p2.s8 drive ahc0t14 plex raid.p2 state stale len 4405248s driveoffset 265s plexoffset 8192s
sd name raid.p2.s9 drive ahc0t15 plex raid.p2 state stale len 4405248s driveoffset 18446744073709551615s plexoffset 9216s
sd name raid.p3.s0 drive ahc0t11 plex raid.p3 state stale len 4405248s driveoffset 4405513s plexoffset 0s
sd name raid.p3.s1 drive ahc0t12 plex raid.p3 state stale len 4405248s driveoffset 4405513s plexoffset 1024s
sd name raid.p3.s2 drive ahc0t13 plex raid.p3 state stale len 4405248s driveoffset 4405513s plexoffset 2048s
sd name raid.p3.s3 drive ahc0t14 plex raid.p3 state stale len 4405248s driveoffset 4405513s plexoffset 3072s
sd name raid.p3.s4 drive ahc0t15 plex raid.p3 state stale len 4405248s driveoffset 18446744073709551615s plexoffset 4096s
sd name raid.p3.s5 drive ahc0t02 plex raid.p3 state stale len 4405248s driveoffset 4405513s plexoffset 5120s
sd name raid.p3.s6 drive ahc0t03 plex raid.p3 state stale len 4405248s driveoffset 4405513s plexoffset 6144s
sd name raid.p3.s7 drive ahc0t04 plex raid.p3 state stale len 4405248s driveoffset 4405513s plexoffset 7168s
sd name raid.p3.s8 drive ahc0t09 plex raid.p3 state stale len 4405248s driveoffset 4405513s plexoffset 8192s
sd name raid.p3.s9 drive ahc0t10 plex raid.p3 state stale len 4405248s driveoffset 4405513s plexoffset 9216s


Want to link to this message? Use this URL: <https://mail-archive.FreeBSD.org/cgi/mid.cgi?20041206031607.52973.qmail>