Date: Thu, 25 Jun 2015 12:56:36 -0700 From: Scott Larson <stl@wiredrive.com> To: =?UTF-8?B?R2Vycml0IEvDvGhu?= <gerrit.kuehn@aei.mpg.de> Cc: freebsd-net@freebsd.org, "carsten.aulbert@aei.mpg.de" <carsten.aulbert@aei.mpg.de> Subject: Re: NFS on 10G interface terribly slow Message-ID: <CAFt8naF7xmZW8bgVrhrL=CaPXiVURqDLsNN5-NHDg=hiv-Qmtw@mail.gmail.com> In-Reply-To: <20150625145238.12cf9da3b368ef0b9a30f193@aei.mpg.de> References: <20150625145238.12cf9da3b368ef0b9a30f193@aei.mpg.de>
next in thread | previous in thread | raw e-mail | index | archive | help
We've got 10.0 and 10.1 servers accessing Isilon and Nexenta via NFS with Intel 10G gear and bursting to near wire speed with the stock MTU/rsize/wsize works as expected. TSO definitely needs to be enabled for that performance. The fact iperf gives you the expected throughput but NFS does not would have me looking at tuning for the NFS platform. Other things to look at: Are all the servers involved negotiating the correct speed and duplex, with TSO? Does it need to have the network stack tuned with whatever it's equivalent of maxsockbuf and send/recvbuf are? Do the switch ports and NIC counters show any drops or errors? On the FBSD servers you could also run 'netstat -i -w 1' under load to see if drops are occurring locally, or 'systat -vmstat' for resource contention problems. But again, a similar setup here and no such issues have appeared. *[image: userimage]Scott Larson[image: los angeles] <https://www.google.com/maps/place/4216+Glencoe+Ave,+Marina+Del+Rey,+CA+902= 92/@33.9892151,-118.4421334,17z/data=3D!3m1!4b1!4m2!3m1!1s0x80c2ba88ffae914= d:0x14e1d00084d4d09c>Lead Systems Administrator[image: wdlogo] <https://www.wiredrive.com/> [image: linkedin] <https://www.linkedin.com/company/wiredrive> [image: facebook] <https://www.twitter.com/wiredrive> [image: twitter] <https://www.facebook.com/wiredrive> [image: instagram] <https://www.instagram.com/wiredrive>T 310 823 8238 x1106 <310%20823%208238%20x1106> | M 310 904 8818 <310%20904%208818>* On Thu, Jun 25, 2015 at 5:52 AM, Gerrit K=C3=BChn <gerrit.kuehn@aei.mpg.de> wrote: > Hi all, > > We have a recent FreeBSD 10.1 installation here that is supposed to act a= s > nfs (v3) client to an Oracle x4-2l server running Soalris 11.2. > We have Intel 10-Gigabit X540-AT2 NICs on both ends, iperf is showing > plenty of bandwidth (9.xGB/s) in both directions. > However, nfs appears to be terribly slow, especially for writing: > > root@crest:~ # dd if=3D/dev/zero of=3D/net/hellpool/Z bs=3D1024k count=3D= 1000 > 1000+0 records in > 1000+0 records out > 1048576000 bytes transferred in 20.263190 secs (51747824 bytes/sec) > > > Reading appears to be faster, but still far away from full bandwidth: > > root@crest:~ # dd of=3D/dev/null if=3D/net/hellpool/Z bs=3D1024k > 1000+0 records in > 1000+0 records out > 1048576000 bytes transferred in 5.129869 secs (204406000 bytes/sec) > > > We have already tried to tune rsize/wsize parameters, but they appear to > have little (if any) impact on these results. Also, neither stripping dow= n > rxsum, txsum, tso etc. from the interface nor increasing MTU to 9000 for > jumbo frames did improve anything. > It is quite embarrassing to achieve way less than 1GBE performance with > 10GBE equipment. Are there any hints what else might be causing this (and > how to fix it)? > > > cu > Gerrit > _______________________________________________ > freebsd-net@freebsd.org mailing list > http://lists.freebsd.org/mailman/listinfo/freebsd-net > To unsubscribe, send any mail to "freebsd-net-unsubscribe@freebsd.org" >
Want to link to this message? Use this URL: <https://mail-archive.FreeBSD.org/cgi/mid.cgi?CAFt8naF7xmZW8bgVrhrL=CaPXiVURqDLsNN5-NHDg=hiv-Qmtw>