[LEDE-DEV] [PATCH] mvebu: Add BQL patch for mvneta driver.
Toke Høiland-Jørgensen
toke at toke.dk
Mon Nov 21 02:12:09 PST 2016
This adds the patch submitted to upstream that adds BQL to the mvneta
driver: https://patchwork.kernel.org/patch/9328413/. Helps latency under
load when the physical link is saturated.
Signed-off-by: Toke Høiland-Jørgensen <toke at toke.dk>
---
.../147-net-mvneta-add-BQL-support.patch | 83 ++++++++++++++++++++++
1 file changed, 83 insertions(+)
create mode 100644 target/linux/mvebu/patches-4.4/147-net-mvneta-add-BQL-support.patch
diff --git a/target/linux/mvebu/patches-4.4/147-net-mvneta-add-BQL-support.patch b/target/linux/mvebu/patches-4.4/147-net-mvneta-add-BQL-support.patch
new file mode 100644
index 0000000..7bd2593
--- /dev/null
+++ b/target/linux/mvebu/patches-4.4/147-net-mvneta-add-BQL-support.patch
@@ -0,0 +1,83 @@
+--- a/drivers/net/ethernet/marvell/mvneta.c
++++ b/drivers/net/ethernet/marvell/mvneta.c
+@@ -1695,8 +1695,10 @@ static struct mvneta_tx_queue *mvneta_tx
+
+ /* Free tx queue skbuffs */
+ static void mvneta_txq_bufs_free(struct mvneta_port *pp,
+- struct mvneta_tx_queue *txq, int num)
++ struct mvneta_tx_queue *txq, int num,
++ struct netdev_queue *nq)
+ {
++ unsigned int bytes_compl = 0, pkts_compl = 0;
+ int i;
+
+ for (i = 0; i < num; i++) {
+@@ -1704,6 +1706,11 @@ static void mvneta_txq_bufs_free(struct
+ txq->txq_get_index;
+ struct sk_buff *skb = txq->tx_skb[txq->txq_get_index];
+
++ if (skb) {
++ bytes_compl += skb->len;
++ pkts_compl++;
++ }
++
+ mvneta_txq_inc_get(txq);
+
+ if (!IS_TSO_HEADER(txq, tx_desc->buf_phys_addr))
+@@ -1714,6 +1721,8 @@ static void mvneta_txq_bufs_free(struct
+ continue;
+ dev_kfree_skb_any(skb);
+ }
++
++ netdev_tx_completed_queue(nq, pkts_compl, bytes_compl);
+ }
+
+ /* Handle end of transmission */
+@@ -1727,7 +1736,7 @@ static void mvneta_txq_done(struct mvnet
+ if (!tx_done)
+ return;
+
+- mvneta_txq_bufs_free(pp, txq, tx_done);
++ mvneta_txq_bufs_free(pp, txq, tx_done, nq);
+
+ txq->count -= tx_done;
+
+@@ -2334,6 +2343,8 @@ out:
+ struct mvneta_pcpu_stats *stats = this_cpu_ptr(pp->stats);
+ struct netdev_queue *nq = netdev_get_tx_queue(dev, txq_id);
+
++ netdev_tx_sent_queue(nq, len);
++
+ txq->count += frags;
+ mvneta_txq_pend_desc_add(pp, txq, frags);
+
+@@ -2358,9 +2369,10 @@ static void mvneta_txq_done_force(struct
+ struct mvneta_tx_queue *txq)
+
+ {
++ struct netdev_queue *nq = netdev_get_tx_queue(pp->dev, txq->id);
+ int tx_done = txq->count;
+
+- mvneta_txq_bufs_free(pp, txq, tx_done);
++ mvneta_txq_bufs_free(pp, txq, tx_done, nq);
+
+ /* reset txq */
+ txq->count = 0;
+@@ -2841,6 +2853,8 @@ static int mvneta_txq_init(struct mvneta
+ static void mvneta_txq_deinit(struct mvneta_port *pp,
+ struct mvneta_tx_queue *txq)
+ {
++ struct netdev_queue *nq = netdev_get_tx_queue(pp->dev, txq->id);
++
+ kfree(txq->tx_skb);
+
+ if (txq->tso_hdrs)
+@@ -2852,6 +2866,8 @@ static void mvneta_txq_deinit(struct mvn
+ txq->size * MVNETA_DESC_ALIGNED_SIZE,
+ txq->descs, txq->descs_phys);
+
++ netdev_tx_reset_queue(nq);
++
+ txq->descs = NULL;
+ txq->last_desc = 0;
+ txq->next_desc_to_proc = 0;
--
2.10.2
More information about the Lede-dev
mailing list