Date: Thu, 26 Mar 2020 12:57:32 -0600 From: Bob Proulx <bob@proulx.com> To: freebsd-questions@freebsd.org Subject: Re: drive selection for disk arrays Message-ID: <20200326124648725158537@bob.proulx.com> In-Reply-To: <713db821-8f69-b41a-75b7-a412a0824c43@holgerdanske.com> References: <20200325081814.GK35528@mithril.foucry.net> <713db821-8f69-b41a-75b7-a412a0824c43@holgerdanske.com>
next in thread | previous in thread | raw e-mail | index | archive | help
David Christensen wrote: > Have anyone seen a failure involving multiple similar drives all failing in > the same mode at the same time? For a corporate setting with an SLA and so forth the usual solution is enough drives as hot spares and a fast enough SLA response time to replace drives quickly before too many fail. But even so I have twice seen large corporate arrays with multiple drives failed. They weren't running ZFS though and so didn't detect it until too late. Then they had to restore from backup. Twice now I have had two sibling disks that started out brand new together with relatively close serial numbers in a RAID1 configuration fail within a week of each other. Both times they were in a rack environment in a controlled access room. One was mine and I caught it soon enough with a replacement. One was a client site where they did not and it became a restore from backup task for me. For myself I always buy dissimilar drives to decouple failure modes. If that is not possible then I remix older drives into a set to decouple failure modes. For myself I would rather have one brand new drive with one older drive than two brand new drives. Regardless I always ensure that backup is operating properly. Bob
Want to link to this message? Use this URL: <https://mail-archive.FreeBSD.org/cgi/mid.cgi?20200326124648725158537>