Date: Sun, 21 Mar 2010 19:12:25 -0500 (CDT) From: Wes Morgan <morganw@chemikals.org> To: Baldur Gislason <baldur@foo.is> Cc: freebsd-fs@freebsd.org Subject: Re: Frustration: replace not doing what I expected. Message-ID: <alpine.BSF.2.00.1003211909510.43670@ibyngvyr> In-Reply-To: <20100321161051.GM63370@gremlin.foo.is> References: <20100317214234.GF63370@gremlin.foo.is> <alpine.BSF.2.00.1003211030160.43670@ibyngvyr> <20100321161051.GM63370@gremlin.foo.is>
next in thread | previous in thread | raw e-mail | index | archive | help
On Sun, 21 Mar 2010, Baldur Gislason wrote: > I got it working, what I had to do was to delete a file that the > resilver process reported as being corrupted. Then run a scrub again > and it would upgrade the pool status to healthy. > Good, I'm glad it worked out. Those errors can be extremely frustrating. One thing that I am curious about, though. A single device failure on a raidz1 shouldn't have resulted in corruption. Were you running degraded for a long period of time? You might want to check your other disks if it was a read failure. > Baldur > > On Sun, Mar 21, 2010 at 10:34:44AM -0500, Wes Morgan wrote: > > On Wed, 17 Mar 2010, Baldur Gislason wrote: > > > > > A drive failed in a pool and I had to replace it. > > > I did zpool replace ad18 ad18, the pool resilvered for 5 hours > > > and finished but did not return from degraded mode. > > > I tried removing the cache file and reimporting the pool, no change, it > > > hasn't gotten rid of the old drive which does not exist anymore. > > > > Hmmm. I've successfully replaced a drive that way before, and I'm sure > > many other people have. Did you offline ad18 before doing both the > > physical drive replacement and the zpool replace? I can't recall if that > > is necessary or not. Can you send the relevant output from zpool history? > > > > The "old" device is part of the metadata on the drive labels, so there is > > no way to remove it like you're wanting without either zfs deciding to > > remove it or rewriting the labels by hand. > > > > > > > pool: zirconium > > > state: DEGRADED > > > status: One or more devices has experienced an error resulting in data > > > corruption. Applications may be affected. > > > action: Restore the file in question if possible. Otherwise restore the > > > entire pool from backup. > > > see: http://www.sun.com/msg/ZFS-8000-8A > > > scrub: none requested > > > config: > > > > > > NAME STATE READ WRITE CKSUM > > > zirconium DEGRADED 0 0 0 > > > raidz1 DEGRADED 0 0 0 > > > ad4 ONLINE 0 0 0 > > > ad6 ONLINE 0 0 0 > > > replacing DEGRADED 0 0 0 > > > 2614810928866691230 UNAVAIL 0 962 0 was /dev/ad18/old > > > ad18 ONLINE 0 0 0 > > > ad20 ONLINE 0 0 0 > _______________________________________________ > freebsd-fs@freebsd.org mailing list > http://lists.freebsd.org/mailman/listinfo/freebsd-fs > To unsubscribe, send any mail to "freebsd-fs-unsubscribe@freebsd.org" >
Want to link to this message? Use this URL: <https://mail-archive.FreeBSD.org/cgi/mid.cgi?alpine.BSF.2.00.1003211909510.43670>