On Mon, Jan 11, 2010 at 08:29:27PM -0800, David Ehrmann wrote: > Pyun YongHyeon wrote: > >It seems iperf on FreeBSD was broken. It incorrectly generates > >huge-packet with IP length 0 so other host disconnected the > >TCP connection. Not sure it could be related with threading though. > >Use netperf instead, it would be more reliable than iperf. > > > I saw a lot of warnings when I opened the cap file in Wireshark about > the length in the IP header being wrong. I'll start looking into netperf > Yeah, there must be a bug in iperf thread handling. Maybe default configuration of iperf should be changed until iperf bug is fixed. > >It's normal see some dropped frames under high network load. And > >you can't compare gigabit controller to fast ethernet controller. > > > Very true, and that's why I tried a lower load. I was a little > surprised to see it choking at just 1 Mb/s (that's bits, not bytes), though. Even though vge(4) is not one of the best controller you can still get more than 650Mbps TX, 920Mbps RX for bulk TCP transfers. For smaller sized TCP segments the number would be much lower than that but that's normal for virtually all controllers. I have a local patch which pushes TX performance numbers up to 800Mbps for vge(4) but it requires fast CPU to do that so I'm not sure whether I want to put it to tree or not. Since I didn't ever see this low TX performance numbers on PCIe based controllers there could be incorrectly programmed registers but datasheet said nothing about this issue. > >I have exact the same revision of the hardware and I don't have > >encountered your issue here. Instead of measuring performance > >number with broken iperf, check whether you still get > >"Connection reset by peer" message with csup(1) when you use vge(4) > >interface. If you still see the message, please send me tcpdump > >capture in private. > > > csup is still working. > > I actually think I *might* have the problem solved. Switching the mount > from UDP (why it was the default for NFS in this Linux distro, I don't > know) to TCP seems to have fixed it. My guess is that some sort of race > condition occurred or there's a bug in someone's NFS flow control > mechanism. A 10x CPU and network performance difference must be more > than is usually tested. I hope. > This could be related with NFS. If NFS over TCP works without problems on vge(4) it's more likely NFS issue. > I'll keep testing NFS over TCP and see if it fixed my problem.Received on Tue Jan 12 2010 - 16:29:44 UTC
This archive was generated by hypermail 2.4.0 : Wed May 19 2021 - 11:40:00 UTC