From owner-freebsd-questions@FreeBSD.ORG Thu Jun 19 02:20:41 2008 Return-Path: Delivered-To: freebsd-questions@freebsd.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2001:4f8:fff6::34]) by hub.freebsd.org (Postfix) with ESMTP id BAA17106568D for ; Thu, 19 Jun 2008 02:20:41 +0000 (UTC) (envelope-from ryan.coleman@cwis.biz) Received: from foxx.skiltech.com (foxx.skiltech.com [209.41.180.8]) by mx1.freebsd.org (Postfix) with ESMTP id 8B0BB8FC15 for ; Thu, 19 Jun 2008 02:20:41 +0000 (UTC) (envelope-from ryan.coleman@cwis.biz) Received: from localhost (localhost.69599 [127.0.0.1]) by foxx.skiltech.com (Postfix) with ESMTP id D794B8FCFC for ; Wed, 18 Jun 2008 22:20:40 -0400 (EDT) X-Virus-Scanned: amavisd-new at foxx.skiltech.com Received: from foxx.skiltech.com ([127.0.0.1]) by localhost (foxx.skiltech.com [127.0.0.1]) (amavisd-new, port 10024) with ESMTP id v2TVyrUk7B5H for ; Wed, 18 Jun 2008 22:20:28 -0400 (EDT) Received: by foxx.skiltech.com (Postfix, from userid 80) id 570568FCF9; Wed, 18 Jun 2008 22:20:28 -0400 (EDT) Received: from 71.63.150.244 (SquirrelMail authenticated user cwis0001) by www.pictureprints.net with HTTP; Wed, 18 Jun 2008 21:20:28 -0500 (CDT) Message-ID: <2812.71.63.150.244.1213842028.squirrel@www.pictureprints.net> Date: Wed, 18 Jun 2008 21:20:28 -0500 (CDT) From: "Ryan Coleman" To: freebsd-questions@freebsd.org User-Agent: SquirrelMail/1.4.11 MIME-Version: 1.0 Content-Type: text/plain;charset=iso-8859-1 Content-Transfer-Encoding: 8bit X-Priority: 3 (Normal) Importance: Normal Subject: Re: "Fixing" a RAID X-BeenThere: freebsd-questions@freebsd.org X-Mailman-Version: 2.1.5 Precedence: list Reply-To: ryan.coleman@cwis.biz List-Id: User questions List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Thu, 19 Jun 2008 02:20:41 -0000 > Ryan Coleman wrote: >>> Ryan Coleman wrote: >>>> Is there a way to figure out what order drives were supposed to go in >>>> for >>>> a RAID 5? Using a hex tool? >>> Do you mean that you physically unplugged them, and they were not >>> labeled? >>> >>> What kind of disk controller is it? > >> It's a HighPoint pATA controller, one drive went kaput so I replaced it >> with another 250G drive and went to rebuild and it wouldn't go. The >> drive >> itself wasn't actually dead, I did some running tests on it and it spun >> up >> OK in an enclosure and then in another machine. So I tried to put the >> drive back on the array and it doesn't believe in having data anymore. > > Ok. The errors you were witnessing after attempting to re-insert it into > the controller, were they generated at BIOS level within the controller > bootup, or in FreeBSD. I'm completely assuming that your running OS was > ON these disks, so the former is true. > >> This is a 4x250G R5 (so ~750G logical) that does have data on it that I >> would very much like to recover somehow. I know this is very likely a >> fruitless endeavor, > > ah, ah ah, never say never, ever. > >> I just need to try. OnTrack and other recovery places >> are just too expensive for this. > > Recover from backup ;) > > I'm kidding. It's too late for that, isn't it. read on... > >> I can dig up the old logs (I think) from >> when she was firing errors two weeks ago. > > Yes. Post the logs. If they are extensive, perhaps you could email them > off-list, with a notice to the list that you have them in the event > others would like to review them as well. > >> The drive was formatted UFS2 as >> one large logical drive in sysinstall. > > ..so if I understand correctly, you had a RAID-5 with three operational > physical disks, and one hot spare? > > Steve > Actually, this is the data storage temporary before I got my massive 7TB RAID purchased and built. But it crashed out 2 days before it arrived. You'll see below the errors. I couldn't even run a find(1) on it. It was 4 disks that made a 714G functional drive, no hotspare, I didn't have the disks for it at the time -- but I do now. The g_vfs_done() errors threw me a bad thought and my tech said "that's a bad sign, you're toast" and left me hanging. I know more than enough about BSD to get around and tech, but RAIDs are not something I have a lot of experience in. [root@testserver /var/log]# more messages.0 | grep 'ar0' May 31 17:25:18 testserver kernel: ar0: 715425MB status: READY May 31 17:25:18 testserver kernel: ar0: disk0 READY using ad13 at ata6-slave May 31 17:25:18 testserver kernel: ar0: disk1 READY using ad16 at ata8-master May 31 17:25:18 testserver kernel: ar0: disk2 READY using ad15 at ata7-slave May 31 17:25:18 testserver kernel: ar0: disk3 READY using ad17 at ata8-slave Jun 4 22:35:45 testserver kernel: ar0: 715425MB status: READY Jun 4 22:35:45 testserver kernel: ar0: disk0 READY using ad13 at ata6-slave Jun 4 22:35:45 testserver kernel: ar0: disk1 READY using ad16 at ata8-master Jun 4 22:35:45 testserver kernel: ar0: disk2 READY using ad15 at ata7-slave Jun 4 22:35:45 testserver kernel: ar0: disk3 READY using ad17 at ata8-slave Jun 4 22:58:09 testserver kernel: ar0: 715425MB status: READY Jun 4 22:58:09 testserver kernel: ar0: disk0 READY using ad13 at ata6-slave Jun 4 22:58:09 testserver kernel: ar0: disk1 READY using ad16 at ata8-master Jun 4 22:58:09 testserver kernel: ar0: disk2 READY using ad15 at ata7-slave Jun 4 22:58:09 testserver kernel: ar0: disk3 READY using ad17 at ata8-slave Jun 4 23:02:28 testserver kernel: ar0: 715425MB status: READY Jun 4 23:02:28 testserver kernel: ar0: disk0 READY using ad13 at ata6-slave Jun 4 23:02:28 testserver kernel: ar0: disk1 READY using ad16 at ata8-master Jun 4 23:02:28 testserver kernel: ar0: disk2 READY using ad15 at ata7-slave Jun 4 23:02:28 testserver kernel: ar0: disk3 READY using ad17 at ata8-slave Jun 4 23:05:35 testserver kernel: g_vfs_done():ar0s1c[READ(offset=501963358208, length=16384)]error = 5 Jun 4 23:05:35 testserver kernel: g_vfs_done():ar0s1c[READ(offset=397138788352, length=16384)]error = 5 Jun 4 23:05:35 testserver kernel: g_vfs_done():ar0s1c[READ(offset=585206398976, length=16384)]error = 5 Jun 4 23:05:35 testserver kernel: g_vfs_done():ar0s1c[READ(offset=360527265792, length=16384)]error = 5 Jun 4 23:05:35 testserver kernel: g_vfs_done():ar0s1c[READ(offset=279018455040, length=16384)]error = 5 Jun 4 23:05:35 testserver kernel: g_vfs_done():ar0s1c[READ(offset=674808283136, length=16384)]error = 5 Jun 4 23:10:06 testserver kernel: g_vfs_done():ar0s1c[READ(offset=501963358208, length=16384)]error = 5 Jun 4 23:10:06 testserver kernel: g_vfs_done():ar0s1c[READ(offset=397138788352, length=16384)]error = 5 Jun 4 23:10:06 testserver kernel: g_vfs_done():ar0s1c[READ(offset=585206398976, length=16384)]error = 5 Jun 4 23:10:06 testserver kernel: g_vfs_done():ar0s1c[READ(offset=360527265792, length=16384)]error = 5 Jun 4 23:10:06 testserver kernel: g_vfs_done():ar0s1c[READ(offset=279018455040, length=16384)]error = 5 Jun 4 23:10:06 testserver kernel: g_vfs_done():ar0s1c[READ(offset=674808283136, length=16384)]error = 5 Jun 5 21:10:37 testserver kernel: ar0: 715425MB status: READY Jun 5 21:10:37 testserver kernel: ar0: disk0 READY using ad13 at ata6-slave Jun 5 21:10:37 testserver kernel: ar0: disk1 READY using ad16 at ata8-master Jun 5 21:10:37 testserver kernel: ar0: disk2 READY using ad15 at ata7-slave Jun 5 21:10:37 testserver kernel: ar0: disk3 READY using ad17 at ata8-slave [root@testserver /var/log]#