Date: Fri, 20 Jan 2012 14:13:16 +0200 From: Nikolay Denev <ndenev@gmail.com> To: Alexander Motin <mav@FreeBSD.org> Cc: Gary Palmer <gpalmer@freebsd.org>, FreeBSD-Current <freebsd-current@freebsd.org>, Dennis K?gel <dk@neveragain.de>, "freebsd-geom@freebsd.org" <freebsd-geom@freebsd.org> Subject: Re: RFC: GEOM MULTIPATH rewrite Message-ID: <25C45DA0-4B52-42E4-A1A3-DD5168451423@gmail.com> In-Reply-To: <4F19503B.2090200@FreeBSD.org> References: <4EAF00A6.5060903@FreeBSD.org> <05E0E64F-5EC4-425A-81E4-B6C35320608B@neveragain.de> <4EB05566.3060700@FreeBSD.org> <20111114210957.GA68559@in-addr.com> <059C17DB-3A7B-41AA-BF91-2F8EBAF17D01@gmail.com> <4F19474A.9020600@FreeBSD.org> <-2439788735531654851@unknownmsgid> <4F19503B.2090200@FreeBSD.org>
next in thread | previous in thread | raw e-mail | index | archive | help
On Jan 20, 2012, at 1:30 PM, Alexander Motin wrote: > On 01/20/12 13:08, Nikolay Denev wrote: >> On 20.01.2012, at 12:51, Alexander Motin<mav@freebsd.org> wrote: >>=20 >>> On 01/20/12 10:09, Nikolay Denev wrote: >>>> Another thing I've observed is that active/active probably only = makes sense if you are accessing single LUN. >>>> In my tests where I have 24 LUNS that form 4 vdevs in a single = zpool, the highest performance was achieved >>>> when I split the active paths among the controllers installed in = the server importing the pool. (basically "gmultipath rotate $LUN" in = rc.local for half of the paths) >>>> Using active/active in this situation resulted in fluctuating = performance. >>>=20 >>> How big was fluctuation? Between speed of one and all paths? >>>=20 >>> Several active/active devices without knowledge about each other = with some probability will send part of requests via the same links, = while ZFS itself already does some balancing between vdevs. >>>=20 >>> -- >>> Alexander Motin >>=20 >> I will test in a bit and post results. >>=20 >> P.S.: Is there a way to enable/disable active-active on the fly? I'm >> currently re-labeling to achieve that. >=20 > No, there is not now. But for experiments you may achieve the same = results by manually marking as failed all paths except one. It is not = dangerous, as if that link fail, all other will resurrect automatically. >=20 > --=20 > Alexander Motin I had to destroy and relabel anyways, since I was not using = active-active currently. Here's what I did (maybe a little too verbose): gmultipath label -A -v LD_0 /dev/da0 /dev/da24=20 gmultipath label -A -v LD_1 /dev/da1 /dev/da25=20 gmultipath label -A -v LD_2 /dev/da2 /dev/da26=20 gmultipath label -A -v LD_3 /dev/da3 /dev/da27=20 gmultipath label -A -v LD_4 /dev/da4 /dev/da28=20 gmultipath label -A -v LD_5 /dev/da5 /dev/da29=20 gmultipath label -A -v LD_6 /dev/da6 /dev/da30=20 gmultipath label -A -v LD_7 /dev/da7 /dev/da31=20 gmultipath label -A -v LD_8 /dev/da8 /dev/da32=20 gmultipath label -A -v LD_9 /dev/da9 /dev/da33=20 gmultipath label -A -v LD_10 /dev/da10 /dev/da34=20 gmultipath label -A -v LD_11 /dev/da11 /dev/da35=20 gmultipath label -A -v LD_12 /dev/da12 /dev/da36=20 gmultipath label -A -v LD_13 /dev/da13 /dev/da37=20 gmultipath label -A -v LD_14 /dev/da14 /dev/da38=20 gmultipath label -A -v LD_15 /dev/da15 /dev/da39=20 gmultipath label -A -v LD_16 /dev/da16 /dev/da40=20 gmultipath label -A -v LD_17 /dev/da17 /dev/da41=20 gmultipath label -A -v LD_18 /dev/da18 /dev/da42=20 gmultipath label -A -v LD_19 /dev/da19 /dev/da43=20 gmultipath label -A -v LD_20 /dev/da20 /dev/da44=20 gmultipath label -A -v LD_21 /dev/da21 /dev/da45=20 gmultipath label -A -v LD_22 /dev/da22 /dev/da46=20 gmultipath label -A -v LD_23 /dev/da23 /dev/da47=20 :~# gmultipath status Name Status Components multipath/LD_0 OPTIMAL da0 (ACTIVE) da24 (ACTIVE) multipath/LD_1 OPTIMAL da1 (ACTIVE) da25 (ACTIVE) multipath/LD_2 OPTIMAL da2 (ACTIVE) da26 (ACTIVE) multipath/LD_3 OPTIMAL da3 (ACTIVE) da27 (ACTIVE) multipath/LD_4 OPTIMAL da4 (ACTIVE) da28 (ACTIVE) multipath/LD_5 OPTIMAL da5 (ACTIVE) da29 (ACTIVE) multipath/LD_6 OPTIMAL da6 (ACTIVE) da30 (ACTIVE) multipath/LD_7 OPTIMAL da7 (ACTIVE) da31 (ACTIVE) multipath/LD_8 OPTIMAL da8 (ACTIVE) da32 (ACTIVE) multipath/LD_9 OPTIMAL da9 (ACTIVE) da33 (ACTIVE) multipath/LD_10 OPTIMAL da10 (ACTIVE) da34 (ACTIVE) multipath/LD_11 OPTIMAL da11 (ACTIVE) da35 (ACTIVE) multipath/LD_12 OPTIMAL da12 (ACTIVE) da36 (ACTIVE) multipath/LD_13 OPTIMAL da13 (ACTIVE) da37 (ACTIVE) multipath/LD_14 OPTIMAL da14 (ACTIVE) da38 (ACTIVE) multipath/LD_15 OPTIMAL da15 (ACTIVE) da39 (ACTIVE) multipath/LD_16 OPTIMAL da16 (ACTIVE) da40 (ACTIVE) multipath/LD_17 OPTIMAL da17 (ACTIVE) da41 (ACTIVE) multipath/LD_18 OPTIMAL da18 (ACTIVE) da42 (ACTIVE) multipath/LD_19 OPTIMAL da19 (ACTIVE) da43 (ACTIVE) multipath/LD_20 OPTIMAL da20 (ACTIVE) da44 (ACTIVE) multipath/LD_21 OPTIMAL da21 (ACTIVE) da45 (ACTIVE) multipath/LD_22 OPTIMAL da22 (ACTIVE) da46 (ACTIVE) multipath/LD_23 OPTIMAL da23 (ACTIVE) da47 (ACTIVE) :~# zpool import tank :~# zpool status pool: tank state: ONLINE scan: none requested config: NAME STATE READ WRITE CKSUM tank ONLINE 0 0 0 raidz2-0 ONLINE 0 0 0 multipath/LD_0 ONLINE 0 0 0 multipath/LD_1 ONLINE 0 0 0 multipath/LD_2 ONLINE 0 0 0 multipath/LD_3 ONLINE 0 0 0 multipath/LD_4 ONLINE 0 0 0 multipath/LD_5 ONLINE 0 0 0 raidz2-1 ONLINE 0 0 0 multipath/LD_6 ONLINE 0 0 0 multipath/LD_7 ONLINE 0 0 0 multipath/LD_8 ONLINE 0 0 0 multipath/LD_9 ONLINE 0 0 0 multipath/LD_10 ONLINE 0 0 0 multipath/LD_11 ONLINE 0 0 0 raidz2-2 ONLINE 0 0 0 multipath/LD_12 ONLINE 0 0 0 multipath/LD_13 ONLINE 0 0 0 multipath/LD_14 ONLINE 0 0 0 multipath/LD_15 ONLINE 0 0 0 multipath/LD_16 ONLINE 0 0 0 multipath/LD_17 ONLINE 0 0 0 raidz2-3 ONLINE 0 0 0 multipath/LD_18 ONLINE 0 0 0 multipath/LD_19 ONLINE 0 0 0 multipath/LD_20 ONLINE 0 0 0 multipath/LD_21 ONLINE 0 0 0 multipath/LD_22 ONLINE 0 0 0 multipath/LD_23 ONLINE 0 0 0 errors: No known data errors And now a very naive benchmark : :~# dd if=3D/dev/zero of=3D/tank/TEST bs=3D1M count=3D512 =20 512+0 records in 512+0 records out 536870912 bytes transferred in 7.282780 secs (73717855 bytes/sec) :~# dd if=3D/dev/zero of=3D/tank/TEST bs=3D1M count=3D512 512+0 records in 512+0 records out 536870912 bytes transferred in 38.422724 secs (13972745 bytes/sec) :~# dd if=3D/dev/zero of=3D/tank/TEST bs=3D1M count=3D512 512+0 records in 512+0 records out 536870912 bytes transferred in 10.810989 secs (49659740 bytes/sec) Now deactivate the alternative paths : /sbin/gmultipath fail LD_0 da24 /sbin/gmultipath fail LD_1 da25 /sbin/gmultipath fail LD_2 da26 /sbin/gmultipath fail LD_3 da27 /sbin/gmultipath fail LD_4 da28 /sbin/gmultipath fail LD_5 da29 /sbin/gmultipath fail LD_6 da6 /sbin/gmultipath fail LD_7 da7 /sbin/gmultipath fail LD_8 da8 /sbin/gmultipath fail LD_9 da9 /sbin/gmultipath fail LD_10 da10 /sbin/gmultipath fail LD_11 da11 /sbin/gmultipath fail LD_12 da36 /sbin/gmultipath fail LD_13 da37 /sbin/gmultipath fail LD_14 da38 /sbin/gmultipath fail LD_15 da39 /sbin/gmultipath fail LD_16 da40 /sbin/gmultipath fail LD_17 da41 /sbin/gmultipath fail LD_18 da18 /sbin/gmultipath fail LD_19 da19 /sbin/gmultipath fail LD_20 da20 /sbin/gmultipath fail LD_21 da21 /sbin/gmultipath fail LD_22 da22 /sbin/gmultipath fail LD_23 da23 :~# gmultipath status Name Status Components multipath/LD_0 DEGRADED da0 (ACTIVE) da24 (FAIL) multipath/LD_1 DEGRADED da1 (ACTIVE) da25 (FAIL) multipath/LD_2 DEGRADED da2 (ACTIVE) da26 (FAIL) multipath/LD_3 DEGRADED da3 (ACTIVE) da27 (FAIL) multipath/LD_4 DEGRADED da4 (ACTIVE) da28 (FAIL) multipath/LD_5 DEGRADED da5 (ACTIVE) da29 (FAIL) multipath/LD_6 DEGRADED da6 (FAIL) da30 (ACTIVE) multipath/LD_7 DEGRADED da7 (FAIL) da31 (ACTIVE) multipath/LD_8 DEGRADED da8 (FAIL) da32 (ACTIVE) multipath/LD_9 DEGRADED da9 (FAIL) da33 (ACTIVE) multipath/LD_10 DEGRADED da10 (FAIL) da34 (ACTIVE) multipath/LD_11 DEGRADED da11 (FAIL) da35 (ACTIVE) multipath/LD_12 DEGRADED da12 (ACTIVE) da36 (FAIL) multipath/LD_13 DEGRADED da13 (ACTIVE) da37 (FAIL) multipath/LD_14 DEGRADED da14 (ACTIVE) da38 (FAIL) multipath/LD_15 DEGRADED da15 (ACTIVE) da39 (FAIL) multipath/LD_16 DEGRADED da16 (ACTIVE) da40 (FAIL) multipath/LD_17 DEGRADED da17 (ACTIVE) da41 (FAIL) multipath/LD_18 DEGRADED da18 (FAIL) da42 (ACTIVE) multipath/LD_19 DEGRADED da19 (FAIL) da43 (ACTIVE) multipath/LD_20 DEGRADED da20 (FAIL) da44 (ACTIVE) multipath/LD_21 DEGRADED da21 (FAIL) da45 (ACTIVE) multipath/LD_22 DEGRADED da22 (FAIL) da46 (ACTIVE) multipath/LD_23 DEGRADED da23 (FAIL) da47 (ACTIVE) And the benchmark again: :~# dd if=3D/dev/zero of=3D/tank/TEST bs=3D1M count=3D512 512+0 records in 512+0 records out 536870912 bytes transferred in 1.083226 secs (495622270 bytes/sec) :~# dd if=3D/dev/zero of=3D/tank/TEST bs=3D1M count=3D512 512+0 records in 512+0 records out 536870912 bytes transferred in 1.409975 secs (380766249 bytes/sec) :~# dd if=3D/dev/zero of=3D/tank/TEST bs=3D1M count=3D512 512+0 records in 512+0 records out 536870912 bytes transferred in 1.136110 secs (472551848 bytes/sec) P.S.: The server is running 8.2-STABLE, dual port isp(4) card, and is = directly connected to a 4Gbps Xyratex dual-controller (active-active) = storage array. All the 24 SAS drives are setup as single disk RAID0 LUNs.=
Want to link to this message? Use this URL: <https://mail-archive.FreeBSD.org/cgi/mid.cgi?25C45DA0-4B52-42E4-A1A3-DD5168451423>