Date: Thu, 22 Apr 2010 11:41:28 -0400 From: Stephen Sanders <ssanders@softhammer.net> To: Jack Vogel <jfvogel@gmail.com> Cc: Brandon Gooch <jamesbrandongooch@gmail.com>, freebsd-performance@freebsd.org Subject: Re: FreeBSD 8.0 ixgbe Poor Performance Message-ID: <4BD06E28.3060609@softhammer.net> In-Reply-To: <v2t2a41acea1004211353nbfc4e68cy6dfaae6f47f86034@mail.gmail.com> References: <4BCF0C9A.10005@softhammer.net> <y2y179b97fb1004210804s6ca12944qf194f3a6d8c33cfe@mail.gmail.com> <x2j2a41acea1004211113kf8e4de95s9ff5c1669156b82c@mail.gmail.com> <4BCF5783.9050007@softhammer.net> <v2t2a41acea1004211353nbfc4e68cy6dfaae6f47f86034@mail.gmail.com>
next in thread | previous in thread | raw e-mail | index | archive | help
I believe that "pciconf -lvc" showed that the cards were in the correct slot. I'm not sure as to what all of the output means but I'm guessing that " cap 10[a0] = PCI-Express 2 endpoint max data 128(256) link x8(x8)" means that the card is an 8 lane card and is using all 8 lanes. Setting kern.ipc.maxsockbuf to16777216 got a better result with ipref TCP testing. The rate when from ~2.5Gpbs to ~5.5Gbps. Running iperf in UDP test mode is still yielding ~2.5Gbps. Running tcpreplay tests is also yielding ~2.5Gbps as well. Command lines for iperf testing are: ipref -t 10 -w 2.5m -l 2.5m -c 169.1.0.2 iperf -s -w 2.5m -B 169.1.0.2 iperf -t 10 -w 2.5m -c 169.1.0.2 -u iperf -s -w 2.5m -B 169.1.0.2 -u For the tcpdump test, I'm sending output to /dev/null and using the cache flag on tcpreplay in order to avoid limiting my network interface throughput to the disk speed. Commands lines for this test are: tcpdump -i ix1 -w /dev/null tcpreplay -i ix1 -t -l 0 -K ./rate.pcap Please forgive my lack of kernel building prowess but I'm guessing that the latest driver needs to be built in a FreeBSD STABLE tree. I ran into an undefined symbol "drbr_needs_enqueue" in the ixgbe code I downloaded. Thanks for all the help. On 4/21/2010 4:53 PM, Jack Vogel wrote: > Use my new driver and it will tell you when it comes up with the slot > speed is, > and if its substandard it will SQUAWK loudly at you :) > > I think the S5000PAL only has Gen1 PCIE slots which is going to limit you > somewhat. Would recommend a current generation (x58 or 5520 chipset) > system if you want the full benefit of 10G. > > BTW, you dont way what adapter, 82598 or 82599, you are using? > > Jack > > > On Wed, Apr 21, 2010 at 12:52 PM, Stephen Sanders > <ssanders@softhammer.net <mailto:ssanders@softhammer.net>> wrote: > > I'd be most pleased to get near 9k. > > I'm running FreeBSD 8.0 amd64 on both of the the test hosts. I've > reset > the configurations to system default as I was getting no where with > sysctl and loader.conf settings. > > The motherboards have been configured to do MSI interrupts. The > S5000PAL has a MSI to old style interrupt BIOS setting that > confuses the > driver interrupt setup. > > The 10Gbps cards should be plugged into the 8x PCI-E slots on both > hosts. I'm double checking that claim right now and will get back > later. > > Thanks > > > On 4/21/2010 2:13 PM, Jack Vogel wrote: > > When you get into the 10G world your performance will only be > as good > > as your weakest link, what I mean is if you connect to something > that has > > less than stellar bus and/or memory performance it is going to > throttle > > everything. > > > > Running back to back with two good systems you should be able to get > > near line rate (9K range). Things that can effect that: 64 bit > kernel, > > TSO, LRO, how many queues come to mind. The default driver config > > should get you there, so tell me more about your hardware/os > config?? > > > > Jack > > > > > > > > On Wed, Apr 21, 2010 at 8:04 AM, Brandon Gooch > > <jamesbrandongooch@gmail.com > <mailto:jamesbrandongooch@gmail.com>>wrote: > > > > > >> On Wed, Apr 21, 2010 at 9:32 AM, Stephen Sanders > >> <ssanders@softhammer.net <mailto:ssanders@softhammer.net>> wrote: > >> > >>> I am running speed tests on a pair of systems equipped with > Intel 10Gbps > >>> cards and am getting poor performance. > >>> > >>> iperf and tcpdump testing indicates that the card is running > at roughly > >>> 2.5Gbps max transmit/receive. > >>> > >>> My attempts at turning fiddling with netisr, polling, and > varying the > >>> buffer sizes has been fruitless. I'm sure there is something > that I'm > >>> missing so I'm hoping for suggestions. > >>> > >>> There are two systems that are connected head to head via > cross over > >>> cable. The two systems have the same hardware configuration. The > >>> hardware is as follows: > >>> > >>> 2 Intel E5430 (Quad core) @ 2.66 Ghz > >>> Intel S5000PAL Motherboard > >>> 16GB Memory > >>> > >>> My iperf command line for the client is: > >>> > >>> iperf -t 10 -c 169.0.0.1 -w 2.5M -l 2.5M > >>> > >>> My TCP dump test command lines are: > >>> > >>> tcpdump -i ix0 -w/dev/null > >>> tcpreplay -i ix0 -t -l 0 -K ./test.pcap > >>> > >> If you're running 8.0-RELEASE, you might try updating to 8-STABLE. > >> Jack Vogel recently committed updated Intel NIC driver code: > >> > >> http://svn.freebsd.org/viewvc/base/stable/8/sys/dev/ixgbe/ > >> > >> -Brandon > >> _______________________________________________ > >> freebsd-performance@freebsd.org > <mailto:freebsd-performance@freebsd.org> mailing list > >> http://lists.freebsd.org/mailman/listinfo/freebsd-performance > >> To unsubscribe, send any mail to " > >> freebsd-performance-unsubscribe@freebsd.org > <mailto:freebsd-performance-unsubscribe@freebsd.org>" > >> > >> > > _______________________________________________ > > freebsd-performance@freebsd.org > <mailto:freebsd-performance@freebsd.org> mailing list > > http://lists.freebsd.org/mailman/listinfo/freebsd-performance > > To unsubscribe, send any mail to > "freebsd-performance-unsubscribe@freebsd.org > <mailto:freebsd-performance-unsubscribe@freebsd.org>" > > > > > >
Want to link to this message? Use this URL: <https://mail-archive.FreeBSD.org/cgi/mid.cgi?4BD06E28.3060609>