Date: Wed, 23 Apr 2014 14:03:07 +0200 From: Johan Hendriks <joh.hendriks@gmail.com> To: Hugo Lombard <hal@elizium.za.net>, freebsd-fs@freebsd.org Subject: Re: ZFS unable to import pool Message-ID: <5357ABFB.9060702@gmail.com> In-Reply-To: <20140423120042.GK2830@sludge.elizium.za.net> References: <B493BD80-CDC2-4BA6-AC14-AE16B373A051@icloud.com> <20140423064203.GD2830@sludge.elizium.za.net> <B1024D84-EBBE-4A9B-82C4-5C19B5A66B60@icloud.com> <20140423080056.GE2830@sludge.elizium.za.net> <20140423091852.GH2830@sludge.elizium.za.net> <20140423100126.GJ2830@sludge.elizium.za.net> <5357937D.4080302@gmail.com> <20140423120042.GK2830@sludge.elizium.za.net>
next in thread | previous in thread | raw e-mail | index | archive | help
op 23-04-14 14:00, Hugo Lombard schreef: > On Wed, Apr 23, 2014 at 12:18:37PM +0200, Johan Hendriks wrote: >> Did you in the past add an extra disk to the pool? >> This could explain the whole issue as the pool is missing a whole vdev. >> > I agree that there's a vdev missing... > > I was able to "simulate" the current problematic import state (sans > failed "disk7", since that doesn't seem to be the stumbling block) by > adding 5 disks [1] to get to here: > > # zpool status test > pool: test > state: ONLINE > scan: none requested > config: > > NAME STATE READ WRITE CKSUM > test ONLINE 0 0 0 > raidz1-0 ONLINE 0 0 0 > md3 ONLINE 0 0 0 > md4 ONLINE 0 0 0 > md5 ONLINE 0 0 0 > md6 ONLINE 0 0 0 > md7 ONLINE 0 0 0 > raidz1-2 ONLINE 0 0 0 > md8 ONLINE 0 0 0 > md9 ONLINE 0 0 0 > md10 ONLINE 0 0 0 > md11 ONLINE 0 0 0 > md12 ONLINE 0 0 0 > logs > md1s1 ONLINE 0 0 0 > cache > md1s2 ONLINE 0 0 0 > > errors: No known data errors > # > > Then exporting it, and removing md8-md12, which results in: > > # zpool import > pool: test > id: 8932371712846778254 > state: UNAVAIL > status: One or more devices are missing from the system. > action: The pool cannot be imported. Attach the missing > devices and try again. > see: http://illumos.org/msg/ZFS-8000-6X > config: > > test UNAVAIL missing device > raidz1-0 ONLINE > md3 ONLINE > md4 ONLINE > md5 ONLINE > md6 ONLINE > md7 ONLINE > cache > md1s2 > logs > md1s1 ONLINE > > Additional devices are known to be part of this pool, though their > exact configuration cannot be determined. > # > > One more data point: In the 'zdb -l' output on the log device it shows > > vdev_children: 2 > > for the pool consisting of raidz1 + log + cache, but it shows > > vdev_children: 3 > > for the pool with raidz1 + raidz1 + log + cache. The pool in the > problem report also shows 'vdev_children: 3' [2] > > > > [1] Trying to add a single device resulted in zpool add complaining > with: > > mismatched replication level: pool uses raidz and new vdev is disk > > and trying it with three disks said: > > mismatched replication level: pool uses 5-way raidz and new vdev uses 3-way raidz > > > [2] http://lists.freebsd.org/pipermail/freebsd-fs/2014-April/019340.html > But you can force it.... If you force it, it will add a vdev not the same as the current vdev. So you will have a raidz1 and a single no parity vdev in the pool. If you destroy the single disk vdev then you will get a pool which can not be repaired as far as I know. regards Johan
Want to link to this message? Use this URL: <https://mail-archive.FreeBSD.org/cgi/mid.cgi?5357ABFB.9060702>