Re: [CFT] VMware vmxnet3 ethernet driver

From: Bryan Venteicher <bryanv_at_daemoninthecloset.org>
Date: Mon, 26 Aug 2013 23:18:37 -0500 (CDT)
----- Original Message -----
> Bezüglich Bryan Venteicher's Nachricht vom 05.08.2013 02:12 (localtime):
> > Hi,
> >
> > I've ported the OpenBSD vmxnet3 ethernet driver to FreeBSD. I did a
> > lot of cleanup, bug fixes, new features, etc (+2000 new lines) along
> > the way so there is not much of a resemblance left.
> >
> > The driver is in good enough shape I'd like additional testers. A patch
> > against -CURRENT is at [1]. Alternatively, the driver and a Makefile is
> > at [2]; this should compile at least as far back as 9.1. I can look at
> > 8-STABLE if there is interest.
> >
> > Obviously, besides reports of 'it works', I'm interested performance vs
> > the emulated e1000, and (for those using it) the VMware tools vmxnet3
> > driver. Hopefully it is no worse :)
> 
> Hello Bryan,
> 
> thanks a lot for your hard work!
> 
> It seems if_vmx doesn't support jumbo frames. If I set mtu 9000, I get
> »vmx0: cannot populate Rx queue 0«, I have no problems using jumbo
> frames with vmxnet3.
> 

This could fail for two reasons - could not allocate an mbuf cluster,
or the call to bus_dmamap_load_mbuf_sg() failed. For the former, you
should check vmstat -z. For the later, the behavior of bus_dmamap_load_mbuf_sg()
changed between 9.1 and 9.2, and I know it was broken for awhile. I don't
recall exactly when I fixed it (I think shortly after I made the original
announcement). Could you retry with the files from HEAD _at_ [1]? Also, there
are new sysctl oids (dev.vmx.X.mbuf_load_failed & dev.vmx.X.mgetcl_failed)
for these errors.

I just compiled the driver on 9.2-RC2 with the sources from HEAD and was
able to change the MTU to 9000.

[1]- http://svnweb.freebsd.org/base/head/sys/dev/vmware/vmxnet3/

> I took a oldish host (4x2,8GHz Core2[LGA775]) with recent software: ESXi
> 5.1U1 and FreeBSD-9.2-RC2
> Two guests are connected to one MTU9000 "VMware Software Switch".
> 

I've got a few performance things to still look at. What's the sysctl 
dev.vmx.X output for the if_vmx<->if_vmx tests?


> Simple iperf (standard TCP) results:
> 
> vmxnet3jumbo <-> vmxnet3jumbo
>     5.3Gbits/sec, load: 40-60%Sys 0.5-2%Intr
> 
> vmxnet3 <-> vmxnet3
>     1.85 GBits/sec, load: 60-80%Sys 0-0.8%Intr
> 
> 
> if_vmx <-> if_vmx
>     1.51 GBits/sec, load: 10-45%Sys 40-48%Intr
>                                                                 !!!
> if_vmxjumbo <-> if_vmxjumbo not possible
> 
> 
> if_em(e1000) <-> if_em(e1000)
>     1.23 GBits/sec, load: 80-60%Sys 0.5-8%Intr
> 
> if_em(e1000)jumbo <-> if_em(e1000)jumbo
>     2.27Gbits/sec, load: 40-30%Sys 0.5-5%Intr
> 
> 
> if_igb(e1000e)junmbo <-> if_igb(e1000e)jumbo
>     5.03 Gbits/s, load: 70-60%Sys 0.5%Intr
> 
> if_igb(e1000e) <-> if_igb(e1000e)
>     1.39 Gbits/s, load: 60-80%Sys 0.5%Intr
> 
> 
> f_igb(e1000e) <-> if_igb(e1000e), both hw.em.[rt]xd=4096
>     1.66 Gbits/s, load: 65-90%Sys 0.5%Intr
> 
> if_igb(e1000e)junmbo <-> if_igb(e1000e)jumbo, both hw.em.[rt]xd=4096
>     4.81 Gbits/s, load: 65%Sys 0.5%Intr
> 
> Conclusion:
> if_vmx performs well compared to the regular emulated nics and standard
> MTU, but it's behind tuned e1000e nic emulation and can't reach vmxnet3
> performance with regular mtu. If one needs throughput, the missing jumbo
> frame support in if_vmx  is a show stopper.
> 
> e1000e is preferable over e1000, even if not officially choosable with
> "FreeBSD"-selection as guest (edit .vmx and alter ethernet0.virtualDev =
> "e1000e", and dont forget to set hw.em.enable_msix=0 in loader.conf,
> although the driver e1000e attaches is if_igb!)
> 
> Thanks,
> 
> -Harry
> 
>
Received on Tue Aug 27 2013 - 02:26:11 UTC

This archive was generated by hypermail 2.4.0 : Wed May 19 2021 - 11:40:40 UTC