iperf3 udp flood behavior at higher rates
Roman Yeryomin
leroi.lists at gmail.com
Wed May 4 01:02:14 PDT 2016
On 3 May 2016 at 02:18, Dave Taht <dave.taht at gmail.com> wrote:
> to fork the fq_codel_drop discussion a bit...
>
> I have up and running two new boxes[1] that are my hope to be able to
> test ath10k/ath9k hardware with, for this test, using one in the
> middle as a router and a nuc i3 box as the server, all ports pure
> ethernet... there's a switch in the way, too.
>
> On tcp via netperf I get expected ~940 mbits.
>
> On udp via iperf3 (again, all pure ethernet) - in neither case below
> am I seeing any drops in the qdisc itself anywhere on the path, yet am
> only achieving 500mbit.
That's interesting, I have no problems with UDP over ethernet.
What about TCP with iperf3?
> ?
>
> 1) Using the
>
> iperf3 -c 172.26.16.130 -u -b900M -R -l1472 -t600
>
> udp flood version, I get some loss on the initial burst, but none
> *reported* after that, and peak at about ~500Mbits.
>
> [ ID] Interval Transfer Bandwidth Jitter
> Lost/Total Datagrams
> [ 4] 0.00-1.00 sec 52.1 MBytes 437 Mbits/sec 0.037 ms
> 1276/38379 (3.3%)
> [ 4] 1.00-2.00 sec 54.3 MBytes 456 Mbits/sec 0.042 ms 0/38699 (0%)
> [ 4] 2.00-3.00 sec 56.1 MBytes 470 Mbits/sec 0.030 ms 0/39933 (0%)
>
> 2) Flipping the sense of the test by getting rid of -R (from the nuc)
>
> iperf3 -c 172.26.16.130 -u -b900M -l1472 -t600
>
> I get on the other side a steady state throughput of a little over
> 520mbits (with 41% loss reported consistently)
>
> [ 5] 37.00-38.00 sec 64.2 MBytes 539 Mbits/sec 0.026 ms
> 31613/77355 (41%)
> [ 5] 38.00-39.00 sec 62.8 MBytes 527 Mbits/sec 0.023 ms
> 31517/76255 (41%)
> [ 5] 39.00-40.00 sec 62.0 MBytes 520 Mbits/sec 0.033 ms
> 31052/75201 (41%)
>
> On the other:
>
> [ 4] 77.00-78.00 sec 111 MBytes 929 Mbits/sec 78915
> [ 4] 78.00-79.00 sec 103 MBytes 864 Mbits/sec 73371
> [ 4] 79.00-80.00 sec 108 MBytes 907 Mbits/sec 77034
> [ 4] 80.00-81.00 sec 107 MBytes 900 Mbits/sec 76423
> [ 4] 81.00-82.00 sec 104 MBytes 875 Mbits/sec 74277
> [ 4] 82.00-83.00 sec 113 MBytes 950 Mbits/sec 80666
>
>
> Thinking that perhaps I was seeing loss in the rx ring, I used ethtool
> to increase that from the default 256 to 4096...
>
> only to hang things thoroughly... :( and I'm watching things reboot now.
>
> Netperf does not have a multi-hop capable udp flood test (rick jones
> can explain why... )
>
> As I recall on this thread iperf3 was being run on a mac box as a
> client, and I'll dig one up - but was it also osx on the other side of
> the test?
>
> And what other params would I tweak on linux to see a udp flood go faster?
I would try making packets smaller (-l), maybe they are fragmented somewhere.
> Topology looks like this:
>
> apu1 <-> apu2 <-> switch <-> nuc.
>
> I could put another switch in the way, I am always nervous about
> invoking hw flow control...
>
> [1] http://www.pcengines.ch/apu2c4.htm
More information about the ath10k
mailing list