From owner-freebsd-stable@FreeBSD.ORG Tue Jul 7 22:26:32 2009 Return-Path: Delivered-To: freebsd-stable@freebsd.org Received: from mx1.freebsd.org (mx1.freebsd.org [IPv6:2001:4f8:fff6::34]) by hub.freebsd.org (Postfix) with ESMTP id 6F4351065673 for ; Tue, 7 Jul 2009 22:26:32 +0000 (UTC) (envelope-from mahlon@martini.nu) Received: from martini.nu (martini.nu [198.145.180.83]) by mx1.freebsd.org (Postfix) with SMTP id 46CF48FC18 for ; Tue, 7 Jul 2009 22:26:32 +0000 (UTC) (envelope-from mahlon@martini.nu) Received: (qmail 86030 invoked by uid 1000); 7 Jul 2009 22:26:31 -0000 Date: Tue, 7 Jul 2009 15:26:31 -0700 From: "Mahlon E. Smith" To: Freddie Cash Message-ID: <20090707222631.GA70750@martini.nu> Mail-Followup-To: "Mahlon E. Smith" , Freddie Cash , freebsd-stable@freebsd.org References: <20090707195614.GA24326@martini.nu> MIME-Version: 1.0 Content-Type: multipart/signed; micalg=pgp-sha1; protocol="application/pgp-signature"; boundary="gBBFr7Ir9EOA20Yy" Content-Disposition: inline In-Reply-To: X-GPG-Fingerprint: 19B8 DDB3 0156 3A03 FA80 8278 C0BE 6BFB 3606 B267 X-Sysinfo: FreeBSD 7.0-RELEASE-p1 i386 User-Agent: Mutt/1.5.17 (2007-11-01) Cc: freebsd-stable@freebsd.org Subject: Re: ZFS: drive replacement performance X-BeenThere: freebsd-stable@freebsd.org X-Mailman-Version: 2.1.5 Precedence: list List-Id: Production branch of FreeBSD source code List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-List-Received-Date: Tue, 07 Jul 2009 22:26:32 -0000 --gBBFr7Ir9EOA20Yy Content-Type: text/plain; charset=us-ascii Content-Disposition: inline Content-Transfer-Encoding: quoted-printable On Tue, Jul 07, 2009, Freddie Cash wrote: >=20 > This is why we've started using glabel(8) to label our drives, and then a= dd > the labels to the pool: > # zpool create store raidz1 label/disk01 label/disk02 label/disk03 >=20 > That way, it does matter where the kernel detects the drives or what the > physical device node is called, GEOM picks up the label, and ZFS uses the > label. Ah, slick. I'll definitely be doing that moving forward. Wonder if I could do it piecemeal now via a shell game, labeling and replacing each individual drive? Will put that on my "try it" list. > > Once I swapped drives, I issued a 'zpool replace'. > > >=20 > See comment at the end: what's the replace command that you used? After the reboot that shuffled device order, the 'da2' changed to that ID number. To have it accept the replace command, I had to use the number itself -- I couldn't use 'da2' since that was now elsewhere, in use, on the raidz1. Surprisingly, it worked. Or at least, it appeared to. % zpool replace store 2025342973333799752 da8 > There's something wrong here. It definitely should be incrementing. Even > when we did the foolish thing of creating a 24-drive raidz2 vdev and had = to > replace a drive, the progress bar did change. Never got above 39% as it > kept restarting, but it did increment. Strangely, the ETA is jumping all over the place, from 50 hours to 2000+ hours. Never seen the percent complete over 0.01% done, but then it goes back to 0.00%. > I'd redo the replace command, and check the output of "zpool status" > to make sure it's showing the proper device node and not some random > string of numbers like it is. Hmm, I'm hunting for it but I don't see it -- know of any way to stop a replace in progress? Thanks for the quick reply, Freddie! -- Mahlon E. Smith =20 http://www.martini.nu/contact.html --gBBFr7Ir9EOA20Yy Content-Type: application/pgp-signature Content-Disposition: inline -----BEGIN PGP SIGNATURE----- iD8DBQFKU8uX1bsjBDapbeMRAoX+AKCfikB4n4yDPt5Dv9H1GsiY+iVqbQCePobm +SxZArzPWuj+CpLicMXhsl4= =n8gd -----END PGP SIGNATURE----- --gBBFr7Ir9EOA20Yy--