Skip site navigation (1)Skip section navigation (2)
Date:      Thu, 20 Feb 2020 10:00:35 +0900
From:      KIRIYAMA Kazuhiko <kiri@truefc.org>
To:        alan somers <asomers@gmail.com>
Cc:        Eric Joyner <erj@freebsd.org>, FreeBSD Net <freebsd-net@freebsd.org>, KIRIYAMA Kazuhiko <kiri@truefc.org>
Subject:   Re: How to work with in 1GbE network ?
Message-ID:  <202002200100.01K10ZYc024946@kx.truefc.org>
In-Reply-To: <CAOtMX2gkD63ZM48=MtpqA3iL7-cPR09hmu-AiiKoNpjqbvt0Pw@mail.gmail.com>
References:  <202002190603.01J63naa005208@kx.truefc.org> <CA%2Bb0zg_rWnyiOa3pfQqFoQyNZg1rkAqD5ZYrQ=YM6SrqUfiEZQ@mail.gmail.com> <CAOtMX2gkD63ZM48=MtpqA3iL7-cPR09hmu-AiiKoNpjqbvt0Pw@mail.gmail.com>

next in thread | previous in thread | raw e-mail | index | archive | help
On Thu, 20 Feb 2020 07:17:35 +0900,
alan somers wrote:
> 
> Make sure that dns resolution is working, forward and reverse.

That's it !  I've mistaken to reset /etc/resolv.conf in which
included local IP DNS.

Thanks for pointed out my rudimentary mistakes.

> 
> On Wed, Feb 19, 2020, 2:53 PM Eric Joyner <erj@freebsd.org> wrote:
> 
> > Have you tried turning off jumbo frames?
> >
> > - Eric
> >
> > On Tue, Feb 18, 2020 at 10:04 PM KIRIYAMA Kazuhiko <kiri@truefc.org>
> > wrote:
> >
> > > Hi, all
> > >
> > > I wonder how to work ixgbe in 1GbE network. I tried to test
> > > in below:
> > >
> > >     internet
> > >         |
> > > +-------+--------+
> > > | Netgear JGS516 |
> > > +---+-----+------+  +----------------------+
> > >     |     +---------+ 13.0-CURRENT(r356739)| src_host
> > >     |               +----------------------+
> > >     |    +----------------------+
> > >     +----+ 13.0-CURRENT(r353025)| dest_host
> > >          +----------------------+
> > >
> > > And try to NFS mount dest_host in src_host, but mount does
> > > not work smoothly. It takes about 9 second !!! :
> > >
> > > # /usr/bin/time
> > > time*    timeout*
> > > # /usr/bin/time -h mount -t nfs dest_host:/.dake /.dake
> > >         9.15s real              0.04s user              0.02s sys
> > > # nfsstat -m
> > > dest_host:/.dake on /.dake
> > >
> > >
> > nfsv3,tcp,resvport,hard,cto,lockd,sec=sys,acdirmin=3,acdirmax=60,acregmin=5,acregmax=60,nametimeo=60,negnametimeo=60,rsize=65536,wsize=65536,readdirsize=65536,readahead=1,wcommitsize=16777216,timeout=120,retrans=2
> > > # /usr/bin/time -h umount /.dake
> > >         27.26s real             0.04s user              0.02s sys
> > >
> > > src_host to dest_host was set to mtu 9000:
> > >
> > > # route get dest_host
> > >    route to: xxx.xxx.xxx.xxx.foo
> > > destination: xxx.xxx.xxx.xxx.foo
> > >        mask: xxx.xxx.xxx.xxx
> > >         fib: 0
> > >   interface: ix0
> > >       flags: <UP,DONE,PINNED>
> > >  recvpipe  sendpipe  ssthresh  rtt,msec    mtu        weight    expire
> > >        0         0         0         0      9000         1         0
> > > #
> > >
> > > What's wrong ?  src_host environments are as follows:
> > >
> > > # uname -a
> > > FreeBSD src_host 13.0-CURRENT FreeBSD 13.0-CURRENT #0 r356739M: Tue Jan
> > 28
> > > 21:49:59 JST 2020     root@msrvkx:/usr/obj/usr/src/amd64.amd64/sys/XIJ
> > > amd64
> > > # ifconfig ix0
> > > ix0: flags=8843<UP,BROADCAST,RUNNING,SIMPLEX,MULTICAST> metric 0 mtu 9000
> > >
> > >
> > options=4e538bb<RXCSUM,TXCSUM,VLAN_MTU,VLAN_HWTAGGING,JUMBO_MTU,VLAN_HWCSUM,WOL_UCAST,WOL_MCAST,WOL_MAGIC,VLAN_HWFILTER,VLAN_HWTSO,RXCSUM_IPV6,TXCSUM_IPV6,NOMAP>
> > >         ether 3c:ec:ef:01:a4:e0
> > >         inet xxx.xxx.xxx.xxx netmask 0xfffffff8 broadcast xxx.xxx.xxx.xxx
> > >         media: Ethernet autoselect (1000baseT
> > > <full-duplex,rxpause,txpause>)
> > >         status: active
> > >         nd6 options=29<PERFORMNUD,IFDISABLED,AUTO_LINKLOCAL>
> > > # sysctl -a|grep jumbo
> > > kern.ipc.nmbjumbo16: 680520
> > > kern.ipc.nmbjumbo9: 1209814
> > > kern.ipc.nmbjumbop: 4083125
> > > vm.uma.mbuf_jumbo_16k.stats.xdomain: 0
> > > vm.uma.mbuf_jumbo_16k.stats.fails: 0
> > > vm.uma.mbuf_jumbo_16k.stats.frees: 0
> > > vm.uma.mbuf_jumbo_16k.stats.allocs: 0
> > > vm.uma.mbuf_jumbo_16k.stats.current: 0
> > > vm.uma.mbuf_jumbo_16k.domain.0.wss: 0
> > > vm.uma.mbuf_jumbo_16k.domain.0.imin: 0
> > > vm.uma.mbuf_jumbo_16k.domain.0.imax: 0
> > > vm.uma.mbuf_jumbo_16k.domain.0.nitems: 0
> > > vm.uma.mbuf_jumbo_16k.limit.bucket_cnt: 0
> > > vm.uma.mbuf_jumbo_16k.limit.bucket_max: 18446744073709551615
> > > vm.uma.mbuf_jumbo_16k.limit.sleeps: 0
> > > vm.uma.mbuf_jumbo_16k.limit.sleepers: 0
> > > vm.uma.mbuf_jumbo_16k.limit.max_items: 680520
> > > vm.uma.mbuf_jumbo_16k.limit.items: 0
> > > vm.uma.mbuf_jumbo_16k.keg.domain.0.free: 0
> > > vm.uma.mbuf_jumbo_16k.keg.domain.0.pages: 0
> > > vm.uma.mbuf_jumbo_16k.keg.efficiency: 99
> > > vm.uma.mbuf_jumbo_16k.keg.align: 7
> > > vm.uma.mbuf_jumbo_16k.keg.ipers: 1
> > > vm.uma.mbuf_jumbo_16k.keg.ppera: 4
> > > vm.uma.mbuf_jumbo_16k.keg.rsize: 16384
> > > vm.uma.mbuf_jumbo_16k.keg.name: mbuf_jumbo_16k
> > > vm.uma.mbuf_jumbo_16k.bucket_size_max: 253
> > > vm.uma.mbuf_jumbo_16k.bucket_size: 253
> > > vm.uma.mbuf_jumbo_16k.flags:
> > > 0x43a10000<TRASH,LIMIT,CTORDTOR,VTOSLAB,OFFPAGE,FIRSTTOUCH>
> > > vm.uma.mbuf_jumbo_16k.size: 16384
> > > vm.uma.mbuf_jumbo_9k.stats.xdomain: 0
> > > vm.uma.mbuf_jumbo_9k.stats.fails: 0
> > > vm.uma.mbuf_jumbo_9k.stats.frees: 0
> > > vm.uma.mbuf_jumbo_9k.stats.allocs: 0
> > > vm.uma.mbuf_jumbo_9k.stats.current: 0
> > > vm.uma.mbuf_jumbo_9k.domain.0.wss: 0
> > > vm.uma.mbuf_jumbo_9k.domain.0.imin: 0
> > > vm.uma.mbuf_jumbo_9k.domain.0.imax: 0
> > > vm.uma.mbuf_jumbo_9k.domain.0.nitems: 0
> > > vm.uma.mbuf_jumbo_9k.limit.bucket_cnt: 0
> > > vm.uma.mbuf_jumbo_9k.limit.bucket_max: 18446744073709551615
> > > vm.uma.mbuf_jumbo_9k.limit.sleeps: 0
> > > vm.uma.mbuf_jumbo_9k.limit.sleepers: 0
> > > vm.uma.mbuf_jumbo_9k.limit.max_items: 1209814
> > > vm.uma.mbuf_jumbo_9k.limit.items: 0
> > > vm.uma.mbuf_jumbo_9k.keg.domain.0.free: 0
> > > vm.uma.mbuf_jumbo_9k.keg.domain.0.pages: 0
> > > vm.uma.mbuf_jumbo_9k.keg.efficiency: 75
> > > vm.uma.mbuf_jumbo_9k.keg.align: 7
> > > vm.uma.mbuf_jumbo_9k.keg.ipers: 1
> > > vm.uma.mbuf_jumbo_9k.keg.ppera: 3
> > > vm.uma.mbuf_jumbo_9k.keg.rsize: 9216
> > > vm.uma.mbuf_jumbo_9k.keg.name: mbuf_jumbo_9k
> > > vm.uma.mbuf_jumbo_9k.bucket_size_max: 253
> > > vm.uma.mbuf_jumbo_9k.bucket_size: 253
> > > vm.uma.mbuf_jumbo_9k.flags: 0x43010000<TRASH,LIMIT,CTORDTOR,FIRSTTOUCH>
> > > vm.uma.mbuf_jumbo_9k.size: 9216
> > > vm.uma.mbuf_jumbo_page.stats.xdomain: 0
> > > vm.uma.mbuf_jumbo_page.stats.fails: 0
> > > vm.uma.mbuf_jumbo_page.stats.frees: 2199
> > > vm.uma.mbuf_jumbo_page.stats.allocs: 67734
> > > vm.uma.mbuf_jumbo_page.stats.current: 65535
> > > vm.uma.mbuf_jumbo_page.domain.0.wss: 0
> > > vm.uma.mbuf_jumbo_page.domain.0.imin: 0
> > > vm.uma.mbuf_jumbo_page.domain.0.imax: 0
> > > vm.uma.mbuf_jumbo_page.domain.0.nitems: 0
> > > vm.uma.mbuf_jumbo_page.limit.bucket_cnt: 0
> > > vm.uma.mbuf_jumbo_page.limit.bucket_max: 18446744073709551615
> > > vm.uma.mbuf_jumbo_page.limit.sleeps: 0
> > > vm.uma.mbuf_jumbo_page.limit.sleepers: 0
> > > vm.uma.mbuf_jumbo_page.limit.max_items: 4083125
> > > vm.uma.mbuf_jumbo_page.limit.items: 67298
> > > vm.uma.mbuf_jumbo_page.keg.domain.0.free: 0
> > > vm.uma.mbuf_jumbo_page.keg.domain.0.pages: 67298
> > > vm.uma.mbuf_jumbo_page.keg.efficiency: 97
> > > vm.uma.mbuf_jumbo_page.keg.align: 7
> > > vm.uma.mbuf_jumbo_page.keg.ipers: 1
> > > vm.uma.mbuf_jumbo_page.keg.ppera: 1
> > > vm.uma.mbuf_jumbo_page.keg.rsize: 4096
> > > vm.uma.mbuf_jumbo_page.keg.name: mbuf_jumbo_page
> > > vm.uma.mbuf_jumbo_page.bucket_size_max: 253
> > > vm.uma.mbuf_jumbo_page.bucket_size: 253
> > > vm.uma.mbuf_jumbo_page.flags:
> > > 0x43a10000<TRASH,LIMIT,CTORDTOR,VTOSLAB,OFFPAGE,FIRSTTOUCH>
> > > vm.uma.mbuf_jumbo_page.size: 4096
> > > # sysctl -a | grep nmbclusters
> > > kern.ipc.nmbclusters: 8166250
> > > # sysctl -a | grep intr_storm_threshold
> > > hw.intr_storm_threshold: 0
> > > #
> > >
> > > and dest_host environments are as follows:
> > >
> > > # uname -a
> > > FreeBSD dest_host 13.0-CURRENT FreeBSD 13.0-CURRENT #0 r353025: Thu Oct
> > 3
> > > 19:38:47 JST 2019     admin@dest_host
> > :/ds/obj/current/13.0/r353025/ds/src/current/13.0/r353025/amd64.amd64/sys/GENERIC
> > > amd64
> > > # ifconfig igb0
> > > igb0: flags=8943<UP,BROADCAST,RUNNING,PROMISC,SIMPLEX,MULTICAST> metric 0
> > > mtu 9000
> > >
> > >
> > options=4a520b9<RXCSUM,VLAN_MTU,VLAN_HWTAGGING,JUMBO_MTU,VLAN_HWCSUM,WOL_MAGIC,VLAN_HWFILTER,VLAN_HWTSO,RXCSUM_IPV6,NOMAP>
> > >         ether 0c:c4:7a:b3:cf:d4
> > >         inet xxx.xxx.xxx.xxx netmask 0xfffffff8 broadcast xxx.xxx.xxx.xxx
> > >         media: Ethernet autoselect (1000baseT <full-duplex>)
> > >         status: active
> > >         nd6 options=29<PERFORMNUD,IFDISABLED,AUTO_LINKLOCAL>
> > > # sysctl -a|grep jumbo
> > > kern.ipc.nmbjumbo16: 339123
> > > kern.ipc.nmbjumbo9: 602886
> > > kern.ipc.nmbjumbop: 2034741
> > > # sysctl -a | grep nmbclusters
> > > kern.ipc.nmbclusters: 4069482
> > > # sysctl -a | grep intr_storm_threshold
> > > hw.intr_storm_threshold: 0
> > > #
> > >
> > > Best regards
> > > ---
> > > Kazuhiko Kiriyama
> > > _______________________________________________
> > > freebsd-net@freebsd.org mailing list
> > > https://lists.freebsd.org/mailman/listinfo/freebsd-net
> > > To unsubscribe, send any mail to "freebsd-net-unsubscribe@freebsd.org"
> > >
> > _______________________________________________
> > freebsd-net@freebsd.org mailing list
> > https://lists.freebsd.org/mailman/listinfo/freebsd-net
> > To unsubscribe, send any mail to "freebsd-net-unsubscribe@freebsd.org"
> >
> _______________________________________________
> freebsd-net@freebsd.org mailing list
> https://lists.freebsd.org/mailman/listinfo/freebsd-net
> To unsubscribe, send any mail to "freebsd-net-unsubscribe@freebsd.org"
> 
---
Kazuhiko Kiriyama



Want to link to this message? Use this URL: <https://mail-archive.FreeBSD.org/cgi/mid.cgi?202002200100.01K10ZYc024946>