Date: Wed, 23 Apr 2014 14:00:42 +0200 From: Hugo Lombard <hal@elizium.za.net> To: Johan Hendriks <joh.hendriks@gmail.com> Cc: freebsd-fs@freebsd.org Subject: Re: ZFS unable to import pool Message-ID: <20140423120042.GK2830@sludge.elizium.za.net> In-Reply-To: <5357937D.4080302@gmail.com> References: <B493BD80-CDC2-4BA6-AC14-AE16B373A051@icloud.com> <20140423064203.GD2830@sludge.elizium.za.net> <B1024D84-EBBE-4A9B-82C4-5C19B5A66B60@icloud.com> <20140423080056.GE2830@sludge.elizium.za.net> <20140423091852.GH2830@sludge.elizium.za.net> <20140423100126.GJ2830@sludge.elizium.za.net> <5357937D.4080302@gmail.com>
next in thread | previous in thread | raw e-mail | index | archive | help
On Wed, Apr 23, 2014 at 12:18:37PM +0200, Johan Hendriks wrote: > > Did you in the past add an extra disk to the pool? > This could explain the whole issue as the pool is missing a whole vdev. > I agree that there's a vdev missing... I was able to "simulate" the current problematic import state (sans failed "disk7", since that doesn't seem to be the stumbling block) by adding 5 disks [1] to get to here: # zpool status test pool: test state: ONLINE scan: none requested config: NAME STATE READ WRITE CKSUM test ONLINE 0 0 0 raidz1-0 ONLINE 0 0 0 md3 ONLINE 0 0 0 md4 ONLINE 0 0 0 md5 ONLINE 0 0 0 md6 ONLINE 0 0 0 md7 ONLINE 0 0 0 raidz1-2 ONLINE 0 0 0 md8 ONLINE 0 0 0 md9 ONLINE 0 0 0 md10 ONLINE 0 0 0 md11 ONLINE 0 0 0 md12 ONLINE 0 0 0 logs md1s1 ONLINE 0 0 0 cache md1s2 ONLINE 0 0 0 errors: No known data errors # Then exporting it, and removing md8-md12, which results in: # zpool import pool: test id: 8932371712846778254 state: UNAVAIL status: One or more devices are missing from the system. action: The pool cannot be imported. Attach the missing devices and try again. see: http://illumos.org/msg/ZFS-8000-6X config: test UNAVAIL missing device raidz1-0 ONLINE md3 ONLINE md4 ONLINE md5 ONLINE md6 ONLINE md7 ONLINE cache md1s2 logs md1s1 ONLINE Additional devices are known to be part of this pool, though their exact configuration cannot be determined. # One more data point: In the 'zdb -l' output on the log device it shows vdev_children: 2 for the pool consisting of raidz1 + log + cache, but it shows vdev_children: 3 for the pool with raidz1 + raidz1 + log + cache. The pool in the problem report also shows 'vdev_children: 3' [2] [1] Trying to add a single device resulted in zpool add complaining with: mismatched replication level: pool uses raidz and new vdev is disk and trying it with three disks said: mismatched replication level: pool uses 5-way raidz and new vdev uses 3-way raidz [2] http://lists.freebsd.org/pipermail/freebsd-fs/2014-April/019340.html -- Hugo Lombard .___. (o,o) /) ) ---"-"---
Want to link to this message? Use this URL: <https://mail-archive.FreeBSD.org/cgi/mid.cgi?20140423120042.GK2830>