Date: Wed, 19 Feb 2020 13:52:51 -0800 From: Eric Joyner <erj@freebsd.org> To: KIRIYAMA Kazuhiko <kiri@truefc.org> Cc: "freebsd-net@freebsd.org" <freebsd-net@freebsd.org> Subject: Re: How to work with in 1GbE network ? Message-ID: <CA%2Bb0zg_rWnyiOa3pfQqFoQyNZg1rkAqD5ZYrQ=YM6SrqUfiEZQ@mail.gmail.com> In-Reply-To: <202002190603.01J63naa005208@kx.truefc.org> References: <202002190603.01J63naa005208@kx.truefc.org>
next in thread | previous in thread | raw e-mail | index | archive | help
Have you tried turning off jumbo frames? - Eric On Tue, Feb 18, 2020 at 10:04 PM KIRIYAMA Kazuhiko <kiri@truefc.org> wrote: > Hi, all > > I wonder how to work ixgbe in 1GbE network. I tried to test > in below: > > internet > | > +-------+--------+ > | Netgear JGS516 | > +---+-----+------+ +----------------------+ > | +---------+ 13.0-CURRENT(r356739)| src_host > | +----------------------+ > | +----------------------+ > +----+ 13.0-CURRENT(r353025)| dest_host > +----------------------+ > > And try to NFS mount dest_host in src_host, but mount does > not work smoothly. It takes about 9 second !!! : > > # /usr/bin/time > time* timeout* > # /usr/bin/time -h mount -t nfs dest_host:/.dake /.dake > 9.15s real 0.04s user 0.02s sys > # nfsstat -m > dest_host:/.dake on /.dake > > nfsv3,tcp,resvport,hard,cto,lockd,sec=sys,acdirmin=3,acdirmax=60,acregmin=5,acregmax=60,nametimeo=60,negnametimeo=60,rsize=65536,wsize=65536,readdirsize=65536,readahead=1,wcommitsize=16777216,timeout=120,retrans=2 > # /usr/bin/time -h umount /.dake > 27.26s real 0.04s user 0.02s sys > > src_host to dest_host was set to mtu 9000: > > # route get dest_host > route to: xxx.xxx.xxx.xxx.foo > destination: xxx.xxx.xxx.xxx.foo > mask: xxx.xxx.xxx.xxx > fib: 0 > interface: ix0 > flags: <UP,DONE,PINNED> > recvpipe sendpipe ssthresh rtt,msec mtu weight expire > 0 0 0 0 9000 1 0 > # > > What's wrong ? src_host environments are as follows: > > # uname -a > FreeBSD src_host 13.0-CURRENT FreeBSD 13.0-CURRENT #0 r356739M: Tue Jan 28 > 21:49:59 JST 2020 root@msrvkx:/usr/obj/usr/src/amd64.amd64/sys/XIJ > amd64 > # ifconfig ix0 > ix0: flags=8843<UP,BROADCAST,RUNNING,SIMPLEX,MULTICAST> metric 0 mtu 9000 > > options=4e538bb<RXCSUM,TXCSUM,VLAN_MTU,VLAN_HWTAGGING,JUMBO_MTU,VLAN_HWCSUM,WOL_UCAST,WOL_MCAST,WOL_MAGIC,VLAN_HWFILTER,VLAN_HWTSO,RXCSUM_IPV6,TXCSUM_IPV6,NOMAP> > ether 3c:ec:ef:01:a4:e0 > inet xxx.xxx.xxx.xxx netmask 0xfffffff8 broadcast xxx.xxx.xxx.xxx > media: Ethernet autoselect (1000baseT > <full-duplex,rxpause,txpause>) > status: active > nd6 options=29<PERFORMNUD,IFDISABLED,AUTO_LINKLOCAL> > # sysctl -a|grep jumbo > kern.ipc.nmbjumbo16: 680520 > kern.ipc.nmbjumbo9: 1209814 > kern.ipc.nmbjumbop: 4083125 > vm.uma.mbuf_jumbo_16k.stats.xdomain: 0 > vm.uma.mbuf_jumbo_16k.stats.fails: 0 > vm.uma.mbuf_jumbo_16k.stats.frees: 0 > vm.uma.mbuf_jumbo_16k.stats.allocs: 0 > vm.uma.mbuf_jumbo_16k.stats.current: 0 > vm.uma.mbuf_jumbo_16k.domain.0.wss: 0 > vm.uma.mbuf_jumbo_16k.domain.0.imin: 0 > vm.uma.mbuf_jumbo_16k.domain.0.imax: 0 > vm.uma.mbuf_jumbo_16k.domain.0.nitems: 0 > vm.uma.mbuf_jumbo_16k.limit.bucket_cnt: 0 > vm.uma.mbuf_jumbo_16k.limit.bucket_max: 18446744073709551615 > vm.uma.mbuf_jumbo_16k.limit.sleeps: 0 > vm.uma.mbuf_jumbo_16k.limit.sleepers: 0 > vm.uma.mbuf_jumbo_16k.limit.max_items: 680520 > vm.uma.mbuf_jumbo_16k.limit.items: 0 > vm.uma.mbuf_jumbo_16k.keg.domain.0.free: 0 > vm.uma.mbuf_jumbo_16k.keg.domain.0.pages: 0 > vm.uma.mbuf_jumbo_16k.keg.efficiency: 99 > vm.uma.mbuf_jumbo_16k.keg.align: 7 > vm.uma.mbuf_jumbo_16k.keg.ipers: 1 > vm.uma.mbuf_jumbo_16k.keg.ppera: 4 > vm.uma.mbuf_jumbo_16k.keg.rsize: 16384 > vm.uma.mbuf_jumbo_16k.keg.name: mbuf_jumbo_16k > vm.uma.mbuf_jumbo_16k.bucket_size_max: 253 > vm.uma.mbuf_jumbo_16k.bucket_size: 253 > vm.uma.mbuf_jumbo_16k.flags: > 0x43a10000<TRASH,LIMIT,CTORDTOR,VTOSLAB,OFFPAGE,FIRSTTOUCH> > vm.uma.mbuf_jumbo_16k.size: 16384 > vm.uma.mbuf_jumbo_9k.stats.xdomain: 0 > vm.uma.mbuf_jumbo_9k.stats.fails: 0 > vm.uma.mbuf_jumbo_9k.stats.frees: 0 > vm.uma.mbuf_jumbo_9k.stats.allocs: 0 > vm.uma.mbuf_jumbo_9k.stats.current: 0 > vm.uma.mbuf_jumbo_9k.domain.0.wss: 0 > vm.uma.mbuf_jumbo_9k.domain.0.imin: 0 > vm.uma.mbuf_jumbo_9k.domain.0.imax: 0 > vm.uma.mbuf_jumbo_9k.domain.0.nitems: 0 > vm.uma.mbuf_jumbo_9k.limit.bucket_cnt: 0 > vm.uma.mbuf_jumbo_9k.limit.bucket_max: 18446744073709551615 > vm.uma.mbuf_jumbo_9k.limit.sleeps: 0 > vm.uma.mbuf_jumbo_9k.limit.sleepers: 0 > vm.uma.mbuf_jumbo_9k.limit.max_items: 1209814 > vm.uma.mbuf_jumbo_9k.limit.items: 0 > vm.uma.mbuf_jumbo_9k.keg.domain.0.free: 0 > vm.uma.mbuf_jumbo_9k.keg.domain.0.pages: 0 > vm.uma.mbuf_jumbo_9k.keg.efficiency: 75 > vm.uma.mbuf_jumbo_9k.keg.align: 7 > vm.uma.mbuf_jumbo_9k.keg.ipers: 1 > vm.uma.mbuf_jumbo_9k.keg.ppera: 3 > vm.uma.mbuf_jumbo_9k.keg.rsize: 9216 > vm.uma.mbuf_jumbo_9k.keg.name: mbuf_jumbo_9k > vm.uma.mbuf_jumbo_9k.bucket_size_max: 253 > vm.uma.mbuf_jumbo_9k.bucket_size: 253 > vm.uma.mbuf_jumbo_9k.flags: 0x43010000<TRASH,LIMIT,CTORDTOR,FIRSTTOUCH> > vm.uma.mbuf_jumbo_9k.size: 9216 > vm.uma.mbuf_jumbo_page.stats.xdomain: 0 > vm.uma.mbuf_jumbo_page.stats.fails: 0 > vm.uma.mbuf_jumbo_page.stats.frees: 2199 > vm.uma.mbuf_jumbo_page.stats.allocs: 67734 > vm.uma.mbuf_jumbo_page.stats.current: 65535 > vm.uma.mbuf_jumbo_page.domain.0.wss: 0 > vm.uma.mbuf_jumbo_page.domain.0.imin: 0 > vm.uma.mbuf_jumbo_page.domain.0.imax: 0 > vm.uma.mbuf_jumbo_page.domain.0.nitems: 0 > vm.uma.mbuf_jumbo_page.limit.bucket_cnt: 0 > vm.uma.mbuf_jumbo_page.limit.bucket_max: 18446744073709551615 > vm.uma.mbuf_jumbo_page.limit.sleeps: 0 > vm.uma.mbuf_jumbo_page.limit.sleepers: 0 > vm.uma.mbuf_jumbo_page.limit.max_items: 4083125 > vm.uma.mbuf_jumbo_page.limit.items: 67298 > vm.uma.mbuf_jumbo_page.keg.domain.0.free: 0 > vm.uma.mbuf_jumbo_page.keg.domain.0.pages: 67298 > vm.uma.mbuf_jumbo_page.keg.efficiency: 97 > vm.uma.mbuf_jumbo_page.keg.align: 7 > vm.uma.mbuf_jumbo_page.keg.ipers: 1 > vm.uma.mbuf_jumbo_page.keg.ppera: 1 > vm.uma.mbuf_jumbo_page.keg.rsize: 4096 > vm.uma.mbuf_jumbo_page.keg.name: mbuf_jumbo_page > vm.uma.mbuf_jumbo_page.bucket_size_max: 253 > vm.uma.mbuf_jumbo_page.bucket_size: 253 > vm.uma.mbuf_jumbo_page.flags: > 0x43a10000<TRASH,LIMIT,CTORDTOR,VTOSLAB,OFFPAGE,FIRSTTOUCH> > vm.uma.mbuf_jumbo_page.size: 4096 > # sysctl -a | grep nmbclusters > kern.ipc.nmbclusters: 8166250 > # sysctl -a | grep intr_storm_threshold > hw.intr_storm_threshold: 0 > # > > and dest_host environments are as follows: > > # uname -a > FreeBSD dest_host 13.0-CURRENT FreeBSD 13.0-CURRENT #0 r353025: Thu Oct 3 > 19:38:47 JST 2019 admin@dest_host:/ds/obj/current/13.0/r353025/ds/src/current/13.0/r353025/amd64.amd64/sys/GENERIC > amd64 > # ifconfig igb0 > igb0: flags=8943<UP,BROADCAST,RUNNING,PROMISC,SIMPLEX,MULTICAST> metric 0 > mtu 9000 > > options=4a520b9<RXCSUM,VLAN_MTU,VLAN_HWTAGGING,JUMBO_MTU,VLAN_HWCSUM,WOL_MAGIC,VLAN_HWFILTER,VLAN_HWTSO,RXCSUM_IPV6,NOMAP> > ether 0c:c4:7a:b3:cf:d4 > inet xxx.xxx.xxx.xxx netmask 0xfffffff8 broadcast xxx.xxx.xxx.xxx > media: Ethernet autoselect (1000baseT <full-duplex>) > status: active > nd6 options=29<PERFORMNUD,IFDISABLED,AUTO_LINKLOCAL> > # sysctl -a|grep jumbo > kern.ipc.nmbjumbo16: 339123 > kern.ipc.nmbjumbo9: 602886 > kern.ipc.nmbjumbop: 2034741 > # sysctl -a | grep nmbclusters > kern.ipc.nmbclusters: 4069482 > # sysctl -a | grep intr_storm_threshold > hw.intr_storm_threshold: 0 > # > > Best regards > --- > Kazuhiko Kiriyama > _______________________________________________ > freebsd-net@freebsd.org mailing list > https://lists.freebsd.org/mailman/listinfo/freebsd-net > To unsubscribe, send any mail to "freebsd-net-unsubscribe@freebsd.org" >
Want to link to this message? Use this URL: <https://mail-archive.FreeBSD.org/cgi/mid.cgi?CA%2Bb0zg_rWnyiOa3pfQqFoQyNZg1rkAqD5ZYrQ=YM6SrqUfiEZQ>