Poor TCP performance with ath10k in 4.0 kernel, again.
Ben Greear
greearb at candelatech.com
Tue May 19 16:22:58 PDT 2015
Additional info & pkt capture at bottom...
On 05/18/2015 02:27 PM, Ben Greear wrote:
> Disclosure: I am working with a patched 4.0 kernel, patched ath10k driver, and
> patched (CT) ath10k firmware. Traffic generator is of our own making.
>
> First, this general problem has been reported before, but the
> work-arounds previously suggested do not fully resolve my problems.
>
> The basic issue is that when the sending socket is directly on top
> of a wifi interface (ath10k driver), then TCP throughput sucks.
>
> For instance, if AP interface sends to station, with 10 concurrent
> TCP streams, I see about 426Mbps. With 100 streams, I see total throughput
> of 750Mbps. These were maybe 10-30 second tests that I did.
>
> Interestingly, a single stream connection performs very poorly at first,
> but at least in one test, it eventually ran quite fast. It is too
> complicated to describe in words, but the graph is here:
>
> http://www.candelatech.com/downloads/single-tcp-4.0.pdf
>
> The 10-stream test did not go above about 450Mbps even after running for more than
> 1 minute, and it was fairly stable around the 450Mbps range after the first few seconds.
>
> 100-stream test shows nice stable aggregate throughput:
>
> http://www.candelatech.com/downloads/100-tcp-4.0.pdf
>
> I have tweaked the kernel tcp_limit_output_bytes setting
> (tested at 1024k too, did not make any significant difference).
>
> # cat /proc/sys/net/ipv4/tcp_limit_output_bytes
> 2048000
>
> I have tried forcing TCP send/rcv buffers to be 1MB and 2MB, but that
> did not make obvious difference except that it started at the maximum
> rate very quickly instead of taking a few seconds to train up to full speed.
>
> If I run a single-stream TCP test, sending on eth1 (Intel 1G NIC)
> through the AP machine, then single stream download is about 540 Mbps, and ramps up
> quickly. So, the AP can definitely send the needed amount of TCP packets.
>
> UDP throughput in download direction, single stream, is about 770Mbps, regardless
> of whether I originate the socket on the AP or if I pass it through the AP.
> send/recv bufs are set to 1MB for UDP sockets.
>
> The 3.17 kernel shows similar behaviour, and the 3.14 kernel is a lot better
> for TCP traffic.
>
> Are there tweaks other than tcp_limit_output_bytes that might
> improve this behaviour?
>
> I will be happy to grab captures or provide any other debugging info
> that someone thinks will be helpful.
>
> Thanks,
> Ben
Here is a capture for the single stream vap -> station test case.
http://www.candelatech.com/downloads/vap-to-sta-1-stream.pcap.bz2
It starts fairly slow, and manages up to around 440Mbps before it plateaus.
The qdisc is pfifo_fast (this is Fedora 19 system). The interface being used for
this test is 'vap1'. This is the sender system.
[root at ct523-1ac-lr201408006507 tmp]# tc qdisc
qdisc pfifo_fast 0: dev eth0 root refcnt 2 bands 3 priomap 1 2 2 2 1 2 0 0 1 1 1 1 1 1 1 1
qdisc pfifo_fast 0: dev eth1 root refcnt 2 bands 3 priomap 1 2 2 2 1 2 0 0 1 1 1 1 1 1 1 1
qdisc mq 0: dev wlan0 root
qdisc pfifo_fast 0: dev wlan0 parent :1 bands 3 priomap 1 2 2 2 1 2 0 0 1 1 1 1 1 1 1 1
qdisc pfifo_fast 0: dev wlan0 parent :2 bands 3 priomap 1 2 2 2 1 2 0 0 1 1 1 1 1 1 1 1
qdisc pfifo_fast 0: dev wlan0 parent :3 bands 3 priomap 1 2 2 2 1 2 0 0 1 1 1 1 1 1 1 1
qdisc pfifo_fast 0: dev wlan0 parent :4 bands 3 priomap 1 2 2 2 1 2 0 0 1 1 1 1 1 1 1 1
qdisc mq 0: dev vap1 root
qdisc pfifo_fast 0: dev vap1 parent :1 bands 3 priomap 1 2 2 2 1 2 0 0 1 1 1 1 1 1 1 1
qdisc pfifo_fast 0: dev vap1 parent :2 bands 3 priomap 1 2 2 2 1 2 0 0 1 1 1 1 1 1 1 1
qdisc pfifo_fast 0: dev vap1 parent :3 bands 3 priomap 1 2 2 2 1 2 0 0 1 1 1 1 1 1 1 1
qdisc pfifo_fast 0: dev vap1 parent :4 bands 3 priomap 1 2 2 2 1 2 0 0 1 1 1 1 1 1 1 1
Thanks,
Ben
--
Ben Greear <greearb at candelatech.com>
Candela Technologies Inc http://www.candelatech.com
More information about the ath10k
mailing list