Date: Thu, 28 Oct 2010 19:15:18 -0700 From: Rumen Telbizov <telbizov@gmail.com> To: Artem Belevich <fbsdlist@src.cx> Cc: freebsd-stable@freebsd.org Subject: Re: Degraded zpool cannot detach old/bad drive Message-ID: <AANLkTimMM82=rqMQQfZZYTcaM_CU%2B01xPeZUGAik8H3v@mail.gmail.com> In-Reply-To: <AANLkTikPqgoxuYp7D88Dp0t5LvjXQeO3mCXdFw6onEZN@mail.gmail.com> References: <AANLkTi=EWfVyZjKEYe=c0x6QvsdUcHGo2-iqGr4OaVG7@mail.gmail.com> <AANLkTinjfpnHGMvzJ5Ly8_WFXGvQmQ4D0-_TgbVBi=cf@mail.gmail.com> <AANLkTi=h6ZJtbRHeUOpKX17uOD5_XyYmu01ZTTCCKw=_@mail.gmail.com> <AANLkTikPqgoxuYp7D88Dp0t5LvjXQeO3mCXdFw6onEZN@mail.gmail.com>
next in thread | previous in thread | raw e-mail | index | archive | help
Hello Artem, everyone, Here's an update on my case. After following Artem's advice I downloaded OpenSolaris 2009 06 LiveCD and booted (over IPMI share) from it. Aliasing the proper disk driver I got access to all JBOD disks that I had before. They had different names (OpenSolaris style) but order and configuration seemed fine. I was in fact able to remove those old/nonexisting devices from OpenSolaris without a problem with the same commands that I was using under FreeBSD. The pool started resilvering which wasn't that important at that stage. So I exported the pool and rebooted back into FreeBSD. FreeBSD saw the pool and I managed to mount it fine. All the data was there and resilvering was initiated. There is a problem though. Initially I used gpt labeled partitions to construct the pool. They had names like /dev/gpt/disk-e1:s15 for example and represented a gpt partition on top of a mfidXX device underneat. Now before I import the pool I do see them in /dev/gpt just fine like that: # ls /dev/gpt disk-e1:s10 disk-e1:s11 disk-e1:s12 disk-e1:s13 disk-e1:s14 disk-e1:s15 disk-e1:s16 disk-e1:s18 disk-e1:s19 disk-e1:s20 disk-e1:s21 disk-e1:s22 disk-e1:s23 disk-e1:s3 disk-e1:s4 disk-e1:s5 disk-e1:s6 disk-e1:s7 disk-e1:s8 disk-e1:s9 disk-e2:s0 disk-e2:s1 disk-e2:s10 disk-e2:s11 disk-e2:s2 disk-e2:s3 disk-e2:s4 disk-e2:s5 disk-e2:s6 disk-e2:s7 disk-e2:s8 disk-e2:s9 newdisk-e1:s17 newdisk-e1:s2 But *after* zpool import tank I get a bunch of messages like those (after enabling kern.geom.label.debug) kernel: GEOM_LABEL[1]: Label gptid/a7fb11e8-dfc9-11df-8732-002590087f3a removed. kernel: GEOM_LABEL[1]: Label gpt/newdisk-e1:s2 removed. kernel: GEOM_LABEL[1]: Label gptid/7a36f6f3-b9fd-11df-8105-002590087f3a removed. kernel: GEOM_LABEL[1]: Label gpt/disk-e1:s3 removed. kernel: GEOM_LABEL[1]: Label gptid/7a92d827-b9fd-11df-8105-002590087f3a removed. kernel: GEOM_LABEL[1]: Label gpt/disk-e1:s4 removed. ... and pretty much all of the gpt labels are gone from /dev/gpt. What I have left there is: disk-e1:s10 disk-e1:s20 disk-e2:s0 So my zpool actually falls back somehow to using the mfidXXp1 partition directly instead of the gpt label. Here's how the pool looks like while imported: # zpool status pool: tank state: ONLINE status: One or more devices is currently being resilvered. The pool will continue to function, possibly in a degraded state. action: Wait for the resilver to complete. scrub: resilver in progress for 0h3m, 0.38% done, 16h35m to go config: NAME STATE READ WRITE CKSUM tank ONLINE 0 0 0 raidz1 ONLINE 0 0 0 mfid33p1 ONLINE 0 0 0 4.18G resilvered mfid1p1 ONLINE 0 0 0 28.8M resilvered mfid2p1 ONLINE 0 0 0 28.0M resilvered mfid3p1 ONLINE 0 0 0 18.4M resilvered raidz1 ONLINE 0 0 0 mfid4p1 ONLINE 0 0 0 mfid5p1 ONLINE 0 0 0 mfid6p1 ONLINE 0 0 0 mfid7p1 ONLINE 0 0 0 raidz1 ONLINE 0 0 0 gpt/disk-e1:s10 ONLINE 0 0 0 mfid9p1 ONLINE 0 0 0 mfid10p1 ONLINE 0 0 0 mfid11p1 ONLINE 0 0 0 raidz1 ONLINE 0 0 0 mfid12p1 ONLINE 0 0 0 27.8M resilvered mfid13p1 ONLINE 0 0 0 27.8M resilvered mfid14p1 ONLINE 0 0 0 27.0M resilvered mfid34p1 ONLINE 0 0 0 4.14G resilvered raidz1 ONLINE 0 0 0 mfid15p1 ONLINE 0 0 0 mfid16p1 ONLINE 0 0 0 gpt/disk-e1:s20 ONLINE 0 0 0 mfid18p1 ONLINE 0 0 0 raidz1 ONLINE 0 0 0 mfid19p1 ONLINE 0 0 0 mfid20p1 ONLINE 0 0 0 gpt/disk-e2:s0 ONLINE 0 0 0 mfid22p1 ONLINE 0 0 0 raidz1 ONLINE 0 0 0 mfid23p1 ONLINE 0 0 0 mfid24p1 ONLINE 0 0 0 mfid25p1 ONLINE 0 0 0 mfid26p1 ONLINE 0 0 0 raidz1 ONLINE 0 0 0 mfid27p1 ONLINE 0 0 0 mfid28p1 ONLINE 0 0 0 mfid29p1 ONLINE 0 0 0 mfid30p1 ONLINE 0 0 0 spares mfid32p1 AVAIL mfid31p1 AVAIL Notice that 3 of the devices are actually still used as gpt/ labels. Also device numbers changed. If I zpool export tank my pool all the /dev/gpt/disk-XX devices come back. *While* having the pool imported I can still see that all the devices with gpart show: # gpart show -l => 34 3904294845 mfid1 GPT (1.8T) 34 3903897600 1 disk-e1:s3 (1.8T) 3903897634 397245 - free - (194M) => 34 3904294845 mfid2 GPT (1.8T) 34 3903897600 1 disk-e1:s4 (1.8T) 3903897634 397245 - free - (194M) => 34 3904294845 mfid3 GPT (1.8T) 34 3903897600 1 disk-e1:s5 (1.8T) 3903897634 397245 - free - (194M) => 34 3904294845 mfid4 GPT (1.8T) 34 3903897600 1 disk-e1:s6 (1.8T) 3903897634 397245 - free - (194M) ... snip snip all the disks are here snip snip ... => 34 3904294845 mfid33 GPT (1.8T) 34 3903897600 1 newdisk-e1:s2 (1.8T) 3903897634 397245 - free - (194M) => 34 3904294845 mfid34 GPT (1.8T) 34 3903897600 1 newdisk-e1:s17 (1.8T) 3903897634 397245 - free - (194M) but only those 3 devices in /dev/gpt and absolutely nothing in /dev/gptid/ So is there a way to bring all the gpt labeled partitions back into the pool instead of using the mfidXX devices? Any help is highly appreciated. I can provide any additional information you may need. Thank you, Rumen Telbizov On Wed, Oct 27, 2010 at 6:05 PM, Rumen Telbizov <telbizov@gmail.com> wrote: > Thanks Artem, > > I am mainly concerned about fixing this immediate problem first and then if > I > can provide more information for the developers so that they look into this > problem > I'd be happy to. > > I'll try OpenSolaris live CD and see how it goes. Either way I'll report > back here. > > Cheers, > Rumen Telbizov > > > On Wed, Oct 27, 2010 at 5:22 PM, Artem Belevich <fbsdlist@src.cx> wrote: > >> Are you interested in what's wrong or in how to fix it? >> >> If fixing is the priority, I'd boot from OpenSolaris live CD and would >> try importing the array there. Just make sure you don't upgrade ZFS to >> a version that is newer than the one FreeBSD supports. >> >> Opensolaris may be able to fix the array. Once it's done, export it, >> boot back to FreeBSD and re-import it. >> >> --Artem >> >> >> >> On Wed, Oct 27, 2010 at 4:22 PM, Rumen Telbizov <telbizov@gmail.com> >> wrote: >> > No ideas whatsoever? >> > >> > On Tue, Oct 26, 2010 at 1:04 PM, Rumen Telbizov <telbizov@gmail.com> >> wrote: >> > >> >> Hello everyone, >> >> >> >> After a few days of struggle with my degraded zpool on a backup server >> I >> >> decided to ask for >> >> help here or at least get some clues as to what might be wrong with it. >> >> Here's the current state of the zpool: >> >> >> >> # zpool status >> >> >> >> pool: tank >> >> state: DEGRADED >> >> status: One or more devices has experienced an error resulting in data >> >> corruption. Applications may be affected. >> >> action: Restore the file in question if possible. Otherwise restore >> the >> >> entire pool from backup. >> >> see: http://www.sun.com/msg/ZFS-8000-8A >> >> scrub: none requested >> >> config: >> >> >> >> NAME STATE READ WRITE CKSUM >> >> tank DEGRADED 0 0 0 >> >> raidz1 DEGRADED 0 0 0 >> >> spare DEGRADED 0 0 0 >> >> replacing DEGRADED 0 0 0 >> >> 17307041822177798519 UNAVAIL 0 299 0 was >> >> /dev/gpt/disk-e1:s2 >> >> gpt/newdisk-e1:s2 ONLINE 0 0 0 >> >> gpt/disk-e2:s10 ONLINE 0 0 0 >> >> gpt/disk-e1:s3 ONLINE 30 0 0 >> >> gpt/disk-e1:s4 ONLINE 0 0 0 >> >> gpt/disk-e1:s5 ONLINE 0 0 0 >> >> raidz1 ONLINE 0 0 0 >> >> gpt/disk-e1:s6 ONLINE 0 0 0 >> >> gpt/disk-e1:s7 ONLINE 0 0 0 >> >> gpt/disk-e1:s8 ONLINE 0 0 0 >> >> gpt/disk-e1:s9 ONLINE 0 0 0 >> >> raidz1 ONLINE 0 0 0 >> >> gpt/disk-e1:s10 ONLINE 0 0 0 >> >> gpt/disk-e1:s11 ONLINE 0 0 0 >> >> gpt/disk-e1:s12 ONLINE 0 0 0 >> >> gpt/disk-e1:s13 ONLINE 0 0 0 >> >> raidz1 DEGRADED 0 0 0 >> >> gpt/disk-e1:s14 ONLINE 0 0 0 >> >> gpt/disk-e1:s15 ONLINE 0 0 0 >> >> gpt/disk-e1:s16 ONLINE 0 0 0 >> >> spare DEGRADED 0 0 0 >> >> replacing DEGRADED 0 0 0 >> >> 15258738282880603331 UNAVAIL 0 48 0 was >> >> /dev/gpt/disk-e1:s17 >> >> gpt/newdisk-e1:s17 ONLINE 0 0 0 >> >> gpt/disk-e2:s11 ONLINE 0 0 0 >> >> raidz1 ONLINE 0 0 0 >> >> gpt/disk-e1:s18 ONLINE 0 0 0 >> >> gpt/disk-e1:s19 ONLINE 0 0 0 >> >> gpt/disk-e1:s20 ONLINE 0 0 0 >> >> gpt/disk-e1:s21 ONLINE 0 0 0 >> >> raidz1 ONLINE 0 0 0 >> >> gpt/disk-e1:s22 ONLINE 0 0 0 >> >> gpt/disk-e1:s23 ONLINE 0 0 0 >> >> gpt/disk-e2:s0 ONLINE 0 0 0 >> >> gpt/disk-e2:s1 ONLINE 0 0 0 >> >> raidz1 ONLINE 0 0 0 >> >> gpt/disk-e2:s2 ONLINE 0 0 0 >> >> gpt/disk-e2:s3 ONLINE 0 0 0 >> >> gpt/disk-e2:s4 ONLINE 0 0 0 >> >> gpt/disk-e2:s5 ONLINE 0 0 0 >> >> raidz1 ONLINE 0 0 0 >> >> gpt/disk-e2:s6 ONLINE 0 0 0 >> >> gpt/disk-e2:s7 ONLINE 0 0 0 >> >> gpt/disk-e2:s8 ONLINE 0 0 0 >> >> gpt/disk-e2:s9 ONLINE 0 0 0 >> >> spares >> >> gpt/disk-e2:s10 INUSE currently in use >> >> gpt/disk-e2:s11 INUSE currently in use >> >> gpt/disk-e1:s2 UNAVAIL cannot open >> >> gpt/newdisk-e1:s17 INUSE currently in use >> >> >> >> errors: 4 data errors, use '-v' for a list >> >> >> >> >> >> The problem is: after replacing the bad drives and resilvering the >> old/bad >> >> drives cannot be detached. >> >> The replace command didn't remove it automatically and manual detach >> fails. >> >> Here are some examples: >> >> >> >> # zpool detach tank 15258738282880603331 >> >> cannot detach 15258738282880603331: no valid replicas >> >> # zpool detach tank gpt/disk-e2:s11 >> >> cannot detach gpt/disk-e2:s11: no valid replicas >> >> # zpool detach tank gpt/newdisk-e1:s17 >> >> cannot detach gpt/newdisk-e1:s17: no valid replicas >> >> # zpool detach tank gpt/disk-e1:s17 >> >> cannot detach gpt/disk-e1:s17: no valid replicas >> >> >> >> >> >> Here's more information and history of events. >> >> This is a 36 disk SuperMicro 847 machine with 2T WD RE4 disks organized >> in >> >> raidz1 groups as >> >> depicted above. zpool deals only with partitions like those: >> >> >> >> => 34 3904294845 mfid30 GPT (1.8T) >> >> 34 3903897600 1 disk-e2:s9 (1.8T) >> >> 3903897634 397245 - free - (194M) >> >> >> >> mfidXX devices are disks connected to a SuperMicro/LSI controller and >> >> presented as jbods. JBODs in this adapter >> >> are actually constructed as raid0 array of 1 disk but this should be >> >> irrelevant in this case. >> >> >> >> This machine was working fine since September 6th but two of the disks >> (in >> >> different raidz1 vdevs) were going >> >> pretty bad and accumulated quite a bit of errors until eventually they >> >> died. This is how they looked like: >> >> >> >> raidz1 DEGRADED 0 0 0 >> >> gpt/disk-e1:s2 UNAVAIL 44 59.5K 0 experienced >> I/O >> >> failures >> >> gpt/disk-e1:s3 ONLINE 0 0 0 >> >> gpt/disk-e1:s4 ONLINE 0 0 0 >> >> gpt/disk-e1:s5 ONLINE 0 0 0 >> >> >> >> raidz1 DEGRADED 0 0 0 >> >> gpt/disk-e1:s14 ONLINE 0 0 0 >> >> gpt/disk-e1:s15 ONLINE 0 0 0 >> >> gpt/disk-e1:s16 ONLINE 0 0 0 >> >> gpt/disk-e1:s17 UNAVAIL 1.56K 49.0K 0 experienced >> I/O >> >> failures >> >> >> >> >> >> I did have two spare disks ready to replace them. So after they died >> here's >> >> what I executed: >> >> >> >> # zpool replace tank gpt/disk-e1:s2 gpt/disk-e2:s10 >> >> # zpool replace tank gpt/disk-e1:s17 gpt/disk-e2:s11 >> >> >> >> Resilvering started. While in the middle of it though the kernel >> paniced >> >> and I had to reboot the machine. >> >> After reboot I waited until the resilvering is complete. Now that it >> was >> >> complete I expected to see the old/bad >> >> device removed from the vdev but it was still there. Trying detach was >> >> complaining with no valid replicas. >> >> I sent colo technician to replace both those defective drives with >> brand >> >> new ones. Once I had them inserted >> >> I recreated them exactly the same way as the ones that I had before - >> jbod >> >> and gpart labeled partition with the >> >> same name! Then I added them as spares: >> >> >> >> # zpool add tank spare gpt/disk-e1:s2 >> >> # zpool add tank spare gpt/disk-e1:s17 >> >> >> >> That actually made it worse I think since now I had the same device >> name >> >> both as a 'previous' failed device >> >> inside the raidz1 group and as a hot spare spare device. I couldn't do >> >> anything with it. >> >> What I did was to export the pool fail the disk on the controller, >> import >> >> the pool and check that zfs could open >> >> it anymore (as a part of the hot spares). Then I recreated that >> >> disk/partition with a new label 'newdisk-XXX' >> >> and tried to replace the device that originally failed (and was only >> >> presented with a number). So I did this: >> >> >> >> # zpool replace tank gpt/disk-e1:s17 gpt/newdisk-e1:s17 >> >> # zpool replace tank gpt/disk-e1:s2 gpt/newdisk-e1:s2 >> >> >> >> Resilvering completed after 17 hours or so and I expected for the >> >> 'replacing' operation to disappear and the >> >> replaced device to go away. But it didn't! Instead I have the state of >> the >> >> pool as shown in the beginning of >> >> the email. >> >> As for the 'errors: 4 data errors, use '-v' for a list' I suspect that >> >> it's due another failing >> >> device (gpt/disk-e1:s3) inside the first (currently degraded) raidz1 >> vdev. >> >> Those 4 corrupted files actually >> >> could be read sometimes so that tells me that the disk has trouble >> reading >> >> *sometimes* those bad blocks. >> >> >> >> Here's the output of zdb -l tank >> >> >> >> version=14 >> >> name='tank' >> >> state=0 >> >> txg=200225 >> >> pool_guid=13504509992978610301 >> >> hostid=409325918 >> >> hostname='XXXX' >> >> vdev_tree >> >> type='root' >> >> id=0 >> >> guid=13504509992978610301 >> >> children[0] >> >> type='raidz' >> >> id=0 >> >> guid=3740854890192825394 >> >> nparity=1 >> >> metaslab_array=33 >> >> metaslab_shift=36 >> >> ashift=9 >> >> asize=7995163410432 >> >> is_log=0 >> >> children[0] >> >> type='spare' >> >> id=0 >> >> guid=16171901098004278313 >> >> whole_disk=0 >> >> children[0] >> >> type='replacing' >> >> id=0 >> >> guid=2754550310390861576 >> >> whole_disk=0 >> >> children[0] >> >> type='disk' >> >> id=0 >> >> guid=17307041822177798519 >> >> path='/dev/gpt/disk-e1:s2' >> >> whole_disk=0 >> >> not_present=1 >> >> DTL=246 >> >> children[1] >> >> type='disk' >> >> id=1 >> >> guid=1641394056824955485 >> >> path='/dev/gpt/newdisk-e1:s2' >> >> whole_disk=0 >> >> DTL=55 >> >> children[1] >> >> type='disk' >> >> id=1 >> >> guid=13150356781300468512 >> >> path='/dev/gpt/disk-e2:s10' >> >> whole_disk=0 >> >> is_spare=1 >> >> DTL=1289 >> >> children[1] >> >> type='disk' >> >> id=1 >> >> guid=6047192237176807561 >> >> path='/dev/gpt/disk-e1:s3' >> >> whole_disk=0 >> >> DTL=250 >> >> children[2] >> >> type='disk' >> >> id=2 >> >> guid=9178318500891071208 >> >> path='/dev/gpt/disk-e1:s4' >> >> whole_disk=0 >> >> DTL=249 >> >> children[3] >> >> type='disk' >> >> id=3 >> >> guid=2567999855746767831 >> >> path='/dev/gpt/disk-e1:s5' >> >> whole_disk=0 >> >> DTL=248 >> >> children[1] >> >> type='raidz' >> >> id=1 >> >> guid=17097047310177793733 >> >> nparity=1 >> >> metaslab_array=31 >> >> metaslab_shift=36 >> >> ashift=9 >> >> asize=7995163410432 >> >> is_log=0 >> >> children[0] >> >> type='disk' >> >> id=0 >> >> guid=14513380297393196654 >> >> path='/dev/gpt/disk-e1:s6' >> >> whole_disk=0 >> >> DTL=266 >> >> children[1] >> >> type='disk' >> >> id=1 >> >> guid=7673391645329839273 >> >> path='/dev/gpt/disk-e1:s7' >> >> whole_disk=0 >> >> DTL=265 >> >> children[2] >> >> type='disk' >> >> id=2 >> >> guid=15189132305590412134 >> >> path='/dev/gpt/disk-e1:s8' >> >> whole_disk=0 >> >> DTL=264 >> >> children[3] >> >> type='disk' >> >> id=3 >> >> guid=17171875527714022076 >> >> path='/dev/gpt/disk-e1:s9' >> >> whole_disk=0 >> >> DTL=263 >> >> children[2] >> >> type='raidz' >> >> id=2 >> >> guid=4551002265962803186 >> >> nparity=1 >> >> metaslab_array=30 >> >> metaslab_shift=36 >> >> ashift=9 >> >> asize=7995163410432 >> >> is_log=0 >> >> children[0] >> >> type='disk' >> >> id=0 >> >> guid=12104241519484712161 >> >> path='/dev/gpt/disk-e1:s10' >> >> whole_disk=0 >> >> DTL=262 >> >> children[1] >> >> type='disk' >> >> id=1 >> >> guid=3950210349623142325 >> >> path='/dev/gpt/disk-e1:s11' >> >> whole_disk=0 >> >> DTL=261 >> >> children[2] >> >> type='disk' >> >> id=2 >> >> guid=14559903955698640085 >> >> path='/dev/gpt/disk-e1:s12' >> >> whole_disk=0 >> >> DTL=260 >> >> children[3] >> >> type='disk' >> >> id=3 >> >> guid=12364155114844220066 >> >> path='/dev/gpt/disk-e1:s13' >> >> whole_disk=0 >> >> DTL=259 >> >> children[3] >> >> type='raidz' >> >> id=3 >> >> guid=12517231224568010294 >> >> nparity=1 >> >> metaslab_array=29 >> >> metaslab_shift=36 >> >> ashift=9 >> >> asize=7995163410432 >> >> is_log=0 >> >> children[0] >> >> type='disk' >> >> id=0 >> >> guid=7655789038925330983 >> >> path='/dev/gpt/disk-e1:s14' >> >> whole_disk=0 >> >> DTL=258 >> >> children[1] >> >> type='disk' >> >> id=1 >> >> guid=17815755378968233141 >> >> path='/dev/gpt/disk-e1:s15' >> >> whole_disk=0 >> >> DTL=257 >> >> children[2] >> >> type='disk' >> >> id=2 >> >> guid=9590421681925673767 >> >> path='/dev/gpt/disk-e1:s16' >> >> whole_disk=0 >> >> DTL=256 >> >> children[3] >> >> type='spare' >> >> id=3 >> >> guid=4015417100051235398 >> >> whole_disk=0 >> >> children[0] >> >> type='replacing' >> >> id=0 >> >> guid=11653429697330193176 >> >> whole_disk=0 >> >> children[0] >> >> type='disk' >> >> id=0 >> >> guid=15258738282880603331 >> >> path='/dev/gpt/disk-e1:s17' >> >> whole_disk=0 >> >> not_present=1 >> >> DTL=255 >> >> children[1] >> >> type='disk' >> >> id=1 >> >> guid=908651380690954833 >> >> path='/dev/gpt/newdisk-e1:s17' >> >> whole_disk=0 >> >> is_spare=1 >> >> DTL=52 >> >> children[1] >> >> type='disk' >> >> id=1 >> >> guid=7250934196571906160 >> >> path='/dev/gpt/disk-e2:s11' >> >> whole_disk=0 >> >> is_spare=1 >> >> DTL=1292 >> >> children[4] >> >> type='raidz' >> >> id=4 >> >> guid=7622366288306613136 >> >> nparity=1 >> >> metaslab_array=28 >> >> metaslab_shift=36 >> >> ashift=9 >> >> asize=7995163410432 >> >> is_log=0 >> >> children[0] >> >> type='disk' >> >> id=0 >> >> guid=11283483106921343963 >> >> path='/dev/gpt/disk-e1:s18' >> >> whole_disk=0 >> >> DTL=254 >> >> children[1] >> >> type='disk' >> >> id=1 >> >> guid=14900597968455968576 >> >> path='/dev/gpt/disk-e1:s19' >> >> whole_disk=0 >> >> DTL=253 >> >> children[2] >> >> type='disk' >> >> id=2 >> >> guid=4140592611852504513 >> >> path='/dev/gpt/disk-e1:s20' >> >> whole_disk=0 >> >> DTL=252 >> >> children[3] >> >> type='disk' >> >> id=3 >> >> guid=2794215380207576975 >> >> path='/dev/gpt/disk-e1:s21' >> >> whole_disk=0 >> >> DTL=251 >> >> children[5] >> >> type='raidz' >> >> id=5 >> >> guid=17655293908271300889 >> >> nparity=1 >> >> metaslab_array=27 >> >> metaslab_shift=36 >> >> ashift=9 >> >> asize=7995163410432 >> >> is_log=0 >> >> children[0] >> >> type='disk' >> >> id=0 >> >> guid=5274146379037055039 >> >> path='/dev/gpt/disk-e1:s22' >> >> whole_disk=0 >> >> DTL=278 >> >> children[1] >> >> type='disk' >> >> id=1 >> >> guid=8651755019404873686 >> >> path='/dev/gpt/disk-e1:s23' >> >> whole_disk=0 >> >> DTL=277 >> >> children[2] >> >> type='disk' >> >> id=2 >> >> guid=16827379661759988976 >> >> path='/dev/gpt/disk-e2:s0' >> >> whole_disk=0 >> >> DTL=276 >> >> children[3] >> >> type='disk' >> >> id=3 >> >> guid=2524967151333933972 >> >> path='/dev/gpt/disk-e2:s1' >> >> whole_disk=0 >> >> DTL=275 >> >> children[6] >> >> type='raidz' >> >> id=6 >> >> guid=2413519694016115220 >> >> nparity=1 >> >> metaslab_array=26 >> >> metaslab_shift=36 >> >> ashift=9 >> >> asize=7995163410432 >> >> is_log=0 >> >> children[0] >> >> type='disk' >> >> id=0 >> >> guid=16361968944335143412 >> >> path='/dev/gpt/disk-e2:s2' >> >> whole_disk=0 >> >> DTL=274 >> >> children[1] >> >> type='disk' >> >> id=1 >> >> guid=10054650477559530937 >> >> path='/dev/gpt/disk-e2:s3' >> >> whole_disk=0 >> >> DTL=273 >> >> children[2] >> >> type='disk' >> >> id=2 >> >> guid=17105959045159531558 >> >> path='/dev/gpt/disk-e2:s4' >> >> whole_disk=0 >> >> DTL=272 >> >> children[3] >> >> type='disk' >> >> id=3 >> >> guid=17370453969371497663 >> >> path='/dev/gpt/disk-e2:s5' >> >> whole_disk=0 >> >> DTL=271 >> >> children[7] >> >> type='raidz' >> >> id=7 >> >> guid=4614010953103453823 >> >> nparity=1 >> >> metaslab_array=24 >> >> metaslab_shift=36 >> >> ashift=9 >> >> asize=7995163410432 >> >> is_log=0 >> >> children[0] >> >> type='disk' >> >> id=0 >> >> guid=10090128057592036175 >> >> path='/dev/gpt/disk-e2:s6' >> >> whole_disk=0 >> >> DTL=270 >> >> children[1] >> >> type='disk' >> >> id=1 >> >> guid=16676544025008223925 >> >> path='/dev/gpt/disk-e2:s7' >> >> whole_disk=0 >> >> DTL=269 >> >> children[2] >> >> type='disk' >> >> id=2 >> >> guid=11777789246954957292 >> >> path='/dev/gpt/disk-e2:s8' >> >> whole_disk=0 >> >> DTL=268 >> >> children[3] >> >> type='disk' >> >> id=3 >> >> guid=3406600121427522915 >> >> path='/dev/gpt/disk-e2:s9' >> >> whole_disk=0 >> >> DTL=267 >> >> >> >> OS: >> >> 8.1-STABLE FreeBSD 8.1-STABLE #0: Sun Sep 5 00:22:45 PDT 2010 amd64 >> >> >> >> Hardware: >> >> Chassis: SuperMicro 847E1 (two backplanes 24 disks front and 12 >> >> disks in the back) >> >> Motherboard: X8SIL >> >> CPU: 1 x X3430 @ 2.40GHz >> >> RAM: 16G >> >> HDD Controller: SuperMicro / LSI 9260 (pciconf -lv SAS1078 PCI-X >> >> Fusion-MPT SAS) : 2 ports >> >> Disks: 36 x 2T Western Digital RE4 >> >> >> >> >> >> >> >> Any help would be appreciated. Let me know what additional information >> I >> >> should provide. >> >> Thank you in advance, >> >> -- >> >> Rumen Telbizov >> >> >> >> >> > >> > >> > -- >> > Rumen Telbizov >> > http://telbizov.com >> > _______________________________________________ >> > freebsd-stable@freebsd.org mailing list >> > http://lists.freebsd.org/mailman/listinfo/freebsd-stable >> > To unsubscribe, send any mail to " >> freebsd-stable-unsubscribe@freebsd.org" >> > >> > > > > -- > Rumen Telbizov > http://telbizov.com > -- Rumen Telbizov http://telbizov.com
Want to link to this message? Use this URL: <https://mail-archive.FreeBSD.org/cgi/mid.cgi?AANLkTimMM82=rqMQQfZZYTcaM_CU%2B01xPeZUGAik8H3v>