Date: Sun, 5 Dec 2004 19:16:07 -0800 (PST) From: orville weyrich <weyrich_comp@yahoo.com> To: freebsd-questions <freebsd-questions@freebsd.org> Subject: VINUM Disaster Recovery Message-ID: <20041206031607.52973.qmail@web50710.mail.yahoo.com>
next in thread | raw e-mail | index | archive | help
[-- Attachment #1 --] I have a 10 disk VINUM configuration and two of the disks are trashed. In theory there is still enough redundant information to get things working again without data loss. Vinum has detected a configuration error (duh -- two disks are toast, plus in recovery I accidently created two more plexes) and taken upon itself to stop configuration updates to prevent any further corruption (thanks! :-). At this point I have looked at http://www.vinumvm.org/vinum/how-to-debug.html and have run a command like the following: ( dd if=/dev/da9s1e skip=8 count=50 | tr -d '\000-\011\200-\377' ; echo ) > da9s1e.log on all 10 disks to obtain a file containing each disk's on-disk configuration. As hoped, eight of the disks show an output similar to the attached file da1s1e.log (differing only as expected in the first line). See attached flog file for a sample output. PLEASE HELP CONFIRM MY PLAN (FOLLOWING) FOR PROCEEDING -- I DO NOT WANT TO DO ANYTHING DISASTEROUS. My thought is that I need to turn on updates, then delete the two unwanted plexes raid.p2 and raid.p3(which were accidentally created), detach the corrupt sdisks, and then hopefully VINUM will forget about the two disks that are toast (or do I somehow have to tell VINUM to forget the disks?). My plan is as follows: First, selectively start vinum: vinum -> read /dev/da1s1e /dev/da2s1e /dev/da3s1e /dev/da4s1e /dev/da5s1e /dev/da6s1e /dev/da7s1e /dev/da8s1e Second, enable configuration updates: vinum->setdaemon 0 Third, save the configuration: vinum->saveconfig Fourth, stop and remove the two unwanted plexes and all attached subdisks: vinum->stop -f raid.p3 vinum->stop -f raid.p2 vinum->rm -r raid.p3 vinum->rm -r raid.p2 Fifth, stop and detach the corruped subdisks: vinum->stop -f raid.p0.s0 vinum->stop -f raid.p0.s9 vinum->stop -f raid.p1.s4 vinum->stop -f raid.p1.s5 vinum->detach raid.p0.s0 vinum->detach raid.p0.s9 vinum->detach raid.p1.s4 vinum->detach raid.p1.s5 At this point I expect to have a functional volume that can be mounted and backed up, prior to the next step of reinstalling the crashed disks, creating new subdisks, attaching them to the plexes, and resynching. PLEASE CONFIRM MY APPROACH OR TELL ME WHERE I AM WRONG! Thanks orville __________________________________________________ Do You Yahoo!? Tired of spam? Yahoo! Mail has the best spam protection around http://mail.yahoo.com [-- Attachment #2 --] IN VINObashful.weyrich.comahc0t03A'BA,volume raid state up plex name raid.p0 state up org striped 1024s vol raid plex name raid.p1 state corrupt org striped 1024s vol raid plex name raid.p2 state faulty org striped 1024s vol raid plex name raid.p3 state faulty org striped 1024s vol raid sd name raid.p0.s0 drive ahc0t02 plex raid.p0 state up len 4405248s driveoffset 265s plexoffset 0s sd name raid.p0.s1 drive ahc0t03 plex raid.p0 state up len 4405248s driveoffset 265s plexoffset 1024s sd name raid.p0.s2 drive ahc0t04 plex raid.p0 state up len 4405248s driveoffset 265s plexoffset 2048s sd name raid.p0.s3 drive ahc0t09 plex raid.p0 state up len 4405248s driveoffset 265s plexoffset 3072s sd name raid.p0.s4 drive ahc0t10 plex raid.p0 state up len 4405248s driveoffset 265s plexoffset 4096s sd name raid.p0.s5 drive ahc0t11 plex raid.p0 state up len 4405248s driveoffset 265s plexoffset 5120s sd name raid.p0.s6 drive ahc0t12 plex raid.p0 state up len 4405248s driveoffset 265s plexoffset 6144s sd name raid.p0.s7 drive ahc0t13 plex raid.p0 state up len 4405248s driveoffset 265s plexoffset 7168s sd name raid.p0.s8 drive ahc0t14 plex raid.p0 state up len 4405248s driveoffset 265s plexoffset 8192s sd name raid.p0.s9 drive *invalid* plex raid.p0 state up len 4405248s driveoffset 265s plexoffset 9216s sd name raid.p1.s0 drive ahc0t11 plex raid.p1 state up len 4405248s driveoffset 4405513s plexoffset 0s sd name raid.p1.s1 drive ahc0t12 plex raid.p1 state up len 4405248s driveoffset 4405513s plexoffset 1024s sd name raid.p1.s2 drive ahc0t13 plex raid.p1 state up len 4405248s driveoffset 4405513s plexoffset 2048s sd name raid.p1.s3 drive ahc0t14 plex raid.p1 state up len 4405248s driveoffset 4405513s plexoffset 3072s sd name raid.p1.s4 drive *invalid* plex raid.p1 state obsolete len 4405248s driveoffset 4405513s plexoffset 4096s sd name raid.p1.s5 drive ahc0t02 plex raid.p1 state up len 4405248s driveoffset 4405513s plexoffset 5120s sd name raid.p1.s6 drive ahc0t03 plex raid.p1 state up len 4405248s driveoffset 4405513s plexoffset 6144s sd name raid.p1.s7 drive ahc0t04 plex raid.p1 state up len 4405248s driveoffset 4405513s plexoffset 7168s sd name raid.p1.s8 drive ahc0t09 plex raid.p1 state up len 4405248s driveoffset 4405513s plexoffset 8192s sd name raid.p1.s9 drive ahc0t10 plex raid.p1 state up len 4405248s driveoffset 4405513s plexoffset 9216s sd name raid.p2.s0 drive ahc0t02 plex raid.p2 state stale len 4405248s driveoffset 265s plexoffset 0s sd name raid.p2.s1 drive ahc0t03 plex raid.p2 state stale len 4405248s driveoffset 265s plexoffset 1024s sd name raid.p2.s2 drive ahc0t04 plex raid.p2 state stale len 4405248s driveoffset 265s plexoffset 2048s sd name raid.p2.s3 drive ahc0t09 plex raid.p2 state stale len 4405248s driveoffset 265s plexoffset 3072s sd name raid.p2.s4 drive ahc0t10 plex raid.p2 state stale len 4405248s driveoffset 265s plexoffset 4096s sd name raid.p2.s5 drive ahc0t11 plex raid.p2 state stale len 4405248s driveoffset 265s plexoffset 5120s sd name raid.p2.s6 drive ahc0t12 plex raid.p2 state stale len 4405248s driveoffset 265s plexoffset 6144s sd name raid.p2.s7 drive ahc0t13 plex raid.p2 state stale len 4405248s driveoffset 265s plexoffset 7168s sd name raid.p2.s8 drive ahc0t14 plex raid.p2 state stale len 4405248s driveoffset 265s plexoffset 8192s sd name raid.p2.s9 drive ahc0t15 plex raid.p2 state stale len 4405248s driveoffset 18446744073709551615s plexoffset 9216s sd name raid.p3.s0 drive ahc0t11 plex raid.p3 state stale len 4405248s driveoffset 4405513s plexoffset 0s sd name raid.p3.s1 drive ahc0t12 plex raid.p3 state stale len 4405248s driveoffset 4405513s plexoffset 1024s sd name raid.p3.s2 drive ahc0t13 plex raid.p3 state stale len 4405248s driveoffset 4405513s plexoffset 2048s sd name raid.p3.s3 drive ahc0t14 plex raid.p3 state stale len 4405248s driveoffset 4405513s plexoffset 3072s sd name raid.p3.s4 drive ahc0t15 plex raid.p3 state stale len 4405248s driveoffset 18446744073709551615s plexoffset 4096s sd name raid.p3.s5 drive ahc0t02 plex raid.p3 state stale len 4405248s driveoffset 4405513s plexoffset 5120s sd name raid.p3.s6 drive ahc0t03 plex raid.p3 state stale len 4405248s driveoffset 4405513s plexoffset 6144s sd name raid.p3.s7 drive ahc0t04 plex raid.p3 state stale len 4405248s driveoffset 4405513s plexoffset 7168s sd name raid.p3.s8 drive ahc0t09 plex raid.p3 state stale len 4405248s driveoffset 4405513s plexoffset 8192s sd name raid.p3.s9 drive ahc0t10 plex raid.p3 state stale len 4405248s driveoffset 4405513s plexoffset 9216s
Want to link to this message? Use this URL: <https://mail-archive.FreeBSD.org/cgi/mid.cgi?20041206031607.52973.qmail>
