Date: Tue, 7 Jul 2009 11:11:34 -0700 From: Freddie Cash <fjwcash@gmail.com> To: freebsd-isp@freebsd.org Subject: Re: Re[2]: ZFS in productions 64 bit Message-ID: <b269bc570907071111o3b850b84q203bead00ad72597@mail.gmail.com> In-Reply-To: <225769512.20090707203135@homelink.ru> References: <4A5209CA.4030304@interazioni.it> <b269bc570907061352u1389d231k8ba35cc5de2d83cb@mail.gmail.com> <4A5343B6.2090608@ibctech.ca> <b269bc570907070851m720e0be6ief726027d4d20994@mail.gmail.com> <225769512.20090707203135@homelink.ru>
next in thread | previous in thread | raw e-mail | index | archive | help
On Tue, Jul 7, 2009 at 9:31 AM, Dennis Yusupoff <dyr@homelink.ru> wrote: > > If there's anything missing from there that you would like to know, just > > ask. :) > > At first, I would like to say thanks for your detailed "success-story" > report. It was great! > So, now a questions. ;) > Have you got any HDD failure, and if yes, how do you repair filesystem > and so on? > We've had one drive fail so far, which is how we discovered that our intial pool setup was horribly, horribly, horribly misconfigured. We originally used a single raidz2 vdev using all 24 harddrives. NOT RECOMMENDED!!! Our throughput was horrible (taking almost 8 hours to complete a backup run of less than 80 servers). Spent over a week trying to get that new drive to resilver, but it just thrashed the drives. Then I found a bunch of articles online that describe how the raidz implementation works (limited to the IOps of a single drive), and that one should not use more than 8 or 9 drives in a raidz vdev. We built the secondary server using the 3-raidz vdev layout, and copied over as much data as we could (lost 3 months of daily backups, saved 2 months). Then we rebuilt the primary servers using the 3-raidz vdev layout, and copied the data back. Since then, we haven't had any other harddrive issues. And, we now run a "zpool scrub" every weekend to check for filesystem inconsistencies, bad checksums, bad data, and so on. So far, no issues found. > Why are you use software RAID, not hardware? > For the flexibility, and all the integrity features of ZFS. The pooled storage concept is just so much nicer/easier to work with than hardware RAID arrays, separate LUNs, separate volume managers, separate partitions, etc. Need more storage? Just add another raidz vdev to the pool. Instantly have more storage space, and performance increases as well (the pool stripes across all the vdevs by default). Don't have any more drive bays? Then just replace the drives in the raidz vdev with larger ones. All the space becomes available to the pool. And *all* the filesystems use that pool, so they all get access to the extra space (no reformatting, no repartitioning, no offline expansion required). Add in the snapshots feature, that actually works without slowing down the system (UFS) or requiring "wasted"/used space (LVM), and it's hard to use hardware RAID anymore. :) Or course, we do still use hardware RAID controllers, for the disk management and alerting features, the onboard cache, the fast buses (PCI-X/PCIe), multi-lane cabling, hot-plug support, etc; we just don't use the actual RAID features. All of our Linux servers still use hardware RAID (5 and 10), with LVM on top, and XFS on top of that. But it's just not as nice of a storage stack to work with. :) -- Freddie Cash fjwcash@gmail.com
Want to link to this message? Use this URL: <https://mail-archive.FreeBSD.org/cgi/mid.cgi?b269bc570907071111o3b850b84q203bead00ad72597>
