Skip site navigation (1)Skip section navigation (2)
Date:      Tue, 14 Jul 2009 09:23:45 +0100
From:      Matthew Seaman <m.seaman@infracaninophile.co.uk>
To:        mahlerrd@yahoo.com
Cc:        Free BSD Questions list <freebsd-questions@freebsd.org>
Subject:   Re: ZFS or UFS for 4TB hardware RAID6?
Message-ID:  <4A5C4091.3030208@infracaninophile.co.uk>
In-Reply-To: <42310.1585.qm@web51008.mail.re2.yahoo.com>
References:  <42310.1585.qm@web51008.mail.re2.yahoo.com>

next in thread | previous in thread | raw e-mail | index | archive | help
This is an OpenPGP/MIME signed message (RFC 2440 and 3156)
--------------enig8D44B299A69B95A1567A379A
Content-Type: text/plain; charset=ISO-8859-1; format=flowed
Content-Transfer-Encoding: quoted-printable

Richard Mahlerwein wrote:
=20
> With 4 drives, you could get much, much higher performance out of
> RAID10 (which is alternatively called RAID0+1 or RAID1+0 depending on
> the manufacturer

Uh -- no.  RAID10 and RAID0+1 are superficially similar but quite differe=
nt
things.  The main differentiator is resilience to disk failure. RAID10 ta=
kes
the raw disks in pairs, creates a mirror across each pair, and then strip=
es
across all the sets of mirrors.  RAID0+1 divides the raw disks into two e=
qual
sets, constructs stripes across each set of disks, and then mirrors the
two stripes.

Read/Write performance is similar in either case: both perform well for=20
the sort of small randomly distributed IO operations you'ld get when eg.
running a RDBMS.  However, consider what happens if you get a disk failur=
e.
In the RAID10 case *one* of your N/2 mirrors is degraded but the other N-=
1
drives in the array operate as normal.  In the RAID0+1 case, one of the
2 stripes is immediately out of action and the whole IO load is carried b=
y
the N/2 drives in the other stripe.

Now consider what happens if a second drive should fail.  In the RAID10
case, you're still up and running so long as the failed drive is one of
the N-2 disks that aren't the mirror pair of the 1st failed drive.
In the RAID0+1 case, you're out of action if the 2nd disk to fail is one
of the N/2 drives from the working stripe.  Or in other words, if two
random disks fail in a RAID10, chances are the RAID will still work.  If
two arbitrarily selected disks fail in a RAID0+1 chances are basically
even that the whole RAID is out of action[*].

I don't think I've ever seen a manufacturer say RAID1+0 instead of RAID10=
,
but I suppose all things are possible.  My impression was that the 0+1=20
terminology was specifically invented to make it more visually distinctiv=
e
-- ie to prevent confusion between '01' and '10'.

	Cheers,

	Matthew

[*] Astute students of probability will point out that this really only
makes a difference for N > 4, and for N=3D4 chances are evens either way =

that failure of two drives would take out the RAID.

--=20
Dr Matthew J Seaman MA, D.Phil.                   7 Priory Courtyard
                                                  Flat 3
PGP: http://www.infracaninophile.co.uk/pgpkey     Ramsgate
                                                  Kent, CT11 9PW


--------------enig8D44B299A69B95A1567A379A
Content-Type: application/pgp-signature; name="signature.asc"
Content-Description: OpenPGP digital signature
Content-Disposition: attachment; filename="signature.asc"

-----BEGIN PGP SIGNATURE-----
Version: GnuPG v2.0.12 (FreeBSD)
Comment: Using GnuPG with Mozilla - http://enigmail.mozdev.org

iEYEAREIAAYFAkpcQJcACgkQ8Mjk52CukIz1JQCeNeREHTenaloe/RskSFVGLMRf
srwAoImZpbdpWoU2QXiC7scW7lJmfyYM
=J4t9
-----END PGP SIGNATURE-----

--------------enig8D44B299A69B95A1567A379A--



Want to link to this message? Use this URL: <https://mail-archive.FreeBSD.org/cgi/mid.cgi?4A5C4091.3030208>