[openwrt/openwrt] kernel: bump 5.15 to 5.15.82
LEDE Commits
lede-commits at lists.infradead.org
Sat Dec 10 17:43:00 PST 2022
hauke pushed a commit to openwrt/openwrt.git, branch master:
https://git.openwrt.org/68426e54eda4f00595390f0c61610b9850bb31db
commit 68426e54eda4f00595390f0c61610b9850bb31db
Author: John Audia <therealgraysky at proton.me>
AuthorDate: Thu Dec 8 08:32:44 2022 -0500
kernel: bump 5.15 to 5.15.82
All patches automatically rebased.
Build system: x86_64
Build-tested: bcm2711/RPi4B, filogic/xiaomi_redmi-router-ax6000-stock
Run-tested: bcm2711/RPi4B, filogic/xiaomi_redmi-router-ax6000-stock
Signed-off-by: John Audia <therealgraysky at proton.me>
---
include/kernel-5.15 | 4 +-
...950-0070-MMC-added-alternative-MMC-driver.patch | 6 +-
...-v6.1-03-mm-vmscan.c-refactor-shrink_node.patch | 6 +-
....1-04-mm-multigenerational-lru-groundwork.patch | 2 +-
...5-mm-multigenerational-lru-mm_struct-list.patch | 8 +-
...20-v6.1-06-mm-multigenerational-lru-aging.patch | 10 +-
...v6.1-07-mm-multigenerational-lru-eviction.patch | 24 +--
...8-mm-multigenerational-lru-user-interface.patch | 4 +-
...glru-don-t-sync-disk-for-each-aging-cycle.patch | 2 +-
...t-stmicro-stmmac-move-queue-reset-to-dedi.patch | 16 +-
...t-stmicro-stmmac-first-disable-all-queues.patch | 4 +-
...t-stmicro-stmmac-move-dma-conf-to-dedicat.patch | 194 ++++++++++-----------
...t-stmicro-stmmac-generate-stmmac-dma-conf.patch | 114 ++++++------
...t-stmicro-stmmac-permit-MTU-change-with-i.patch | 4 +-
.../pending-5.15/630-packet_socket_type.patch | 6 +-
...-add-detach-callback-to-struct-phy_driver.patch | 2 +-
...et-phy-simplify-phy_link_change-arguments.patch | 2 +-
17 files changed, 204 insertions(+), 204 deletions(-)
diff --git a/include/kernel-5.15 b/include/kernel-5.15
index 1e811b42ff..18444aaa84 100644
--- a/include/kernel-5.15
+++ b/include/kernel-5.15
@@ -1,2 +1,2 @@
-LINUX_VERSION-5.15 = .81
-LINUX_KERNEL_HASH-5.15.81 = 8f885cdebd754d6e63b920cf6c3e5713e91bbf5f52e9d99eb0054ef7e8f096ab
+LINUX_VERSION-5.15 = .82
+LINUX_KERNEL_HASH-5.15.82 = fceef6bb79bac494663ccde34453521fc616cd94272fd30564752b3742381b65
diff --git a/target/linux/bcm27xx/patches-5.15/950-0070-MMC-added-alternative-MMC-driver.patch b/target/linux/bcm27xx/patches-5.15/950-0070-MMC-added-alternative-MMC-driver.patch
index 3e2d3135c5..799b5c8eab 100644
--- a/target/linux/bcm27xx/patches-5.15/950-0070-MMC-added-alternative-MMC-driver.patch
+++ b/target/linux/bcm27xx/patches-5.15/950-0070-MMC-added-alternative-MMC-driver.patch
@@ -291,7 +291,7 @@ bcm2835-mmc: uninitialized_var is no more
goto out;
--- a/drivers/mmc/core/core.c
+++ b/drivers/mmc/core/core.c
-@@ -1807,7 +1807,8 @@ EXPORT_SYMBOL(mmc_erase);
+@@ -1812,7 +1812,8 @@ EXPORT_SYMBOL(mmc_erase);
int mmc_can_erase(struct mmc_card *card)
{
@@ -1984,7 +1984,7 @@ bcm2835-mmc: uninitialized_var is no more
#define MAX_TUNING_LOOP 40
-@@ -3143,7 +3143,7 @@ static void sdhci_timeout_timer(struct t
+@@ -3188,7 +3188,7 @@ static void sdhci_timeout_timer(struct t
spin_lock_irqsave(&host->lock, flags);
if (host->cmd && !sdhci_data_line_cmd(host->cmd)) {
@@ -1993,7 +1993,7 @@ bcm2835-mmc: uninitialized_var is no more
mmc_hostname(host->mmc));
sdhci_dumpregs(host);
-@@ -3165,7 +3165,7 @@ static void sdhci_timeout_data_timer(str
+@@ -3210,7 +3210,7 @@ static void sdhci_timeout_data_timer(str
if (host->data || host->data_cmd ||
(host->cmd && sdhci_data_line_cmd(host->cmd))) {
diff --git a/target/linux/generic/backport-5.15/020-v6.1-03-mm-vmscan.c-refactor-shrink_node.patch b/target/linux/generic/backport-5.15/020-v6.1-03-mm-vmscan.c-refactor-shrink_node.patch
index 35b6894a64..31161e2a0f 100644
--- a/target/linux/generic/backport-5.15/020-v6.1-03-mm-vmscan.c-refactor-shrink_node.patch
+++ b/target/linux/generic/backport-5.15/020-v6.1-03-mm-vmscan.c-refactor-shrink_node.patch
@@ -15,7 +15,7 @@ Change-Id: Iae734b5b4030205b7db6e8c841f747b6f6ae1a04
--- a/mm/vmscan.c
+++ b/mm/vmscan.c
-@@ -2562,6 +2562,103 @@ enum scan_balance {
+@@ -2497,6 +2497,103 @@ enum scan_balance {
SCAN_FILE,
};
@@ -119,7 +119,7 @@ Change-Id: Iae734b5b4030205b7db6e8c841f747b6f6ae1a04
/*
* Determine how aggressively the anon and file LRU lists should be
* scanned. The relative value of each set of LRU lists is determined
-@@ -3030,7 +3127,6 @@ static void shrink_node(pg_data_t *pgdat
+@@ -2965,7 +3062,6 @@ static void shrink_node(pg_data_t *pgdat
unsigned long nr_reclaimed, nr_scanned;
struct lruvec *target_lruvec;
bool reclaimable = false;
@@ -127,7 +127,7 @@ Change-Id: Iae734b5b4030205b7db6e8c841f747b6f6ae1a04
target_lruvec = mem_cgroup_lruvec(sc->target_mem_cgroup, pgdat);
-@@ -3046,93 +3142,7 @@ again:
+@@ -2981,93 +3077,7 @@ again:
nr_reclaimed = sc->nr_reclaimed;
nr_scanned = sc->nr_scanned;
diff --git a/target/linux/generic/backport-5.15/020-v6.1-04-mm-multigenerational-lru-groundwork.patch b/target/linux/generic/backport-5.15/020-v6.1-04-mm-multigenerational-lru-groundwork.patch
index fbc15d8c91..533ea2aa20 100644
--- a/target/linux/generic/backport-5.15/020-v6.1-04-mm-multigenerational-lru-groundwork.patch
+++ b/target/linux/generic/backport-5.15/020-v6.1-04-mm-multigenerational-lru-groundwork.patch
@@ -720,7 +720,7 @@ Change-Id: I71de7cd15b8dfa6f9fdd838023474693c4fee0a7
#include <asm/tlbflush.h>
#include <asm/div64.h>
-@@ -2880,6 +2881,273 @@ static bool can_age_anon_pages(struct pg
+@@ -2815,6 +2816,273 @@ static bool can_age_anon_pages(struct pg
return can_demote(pgdat->node_id, sc);
}
diff --git a/target/linux/generic/backport-5.15/020-v6.1-05-mm-multigenerational-lru-mm_struct-list.patch b/target/linux/generic/backport-5.15/020-v6.1-05-mm-multigenerational-lru-mm_struct-list.patch
index 2592f18e06..736b4f5152 100644
--- a/target/linux/generic/backport-5.15/020-v6.1-05-mm-multigenerational-lru-mm_struct-list.patch
+++ b/target/linux/generic/backport-5.15/020-v6.1-05-mm-multigenerational-lru-mm_struct-list.patch
@@ -387,7 +387,7 @@ Change-Id: I25d9eda8c6bdc7c3653b9f210a159d6c247c81e8
.dfl_cftypes = memory_files,
--- a/mm/vmscan.c
+++ b/mm/vmscan.c
-@@ -2929,6 +2929,306 @@ static bool __maybe_unused seq_is_valid(
+@@ -2864,6 +2864,306 @@ static bool __maybe_unused seq_is_valid(
}
/******************************************************************************
@@ -694,7 +694,7 @@ Change-Id: I25d9eda8c6bdc7c3653b9f210a159d6c247c81e8
* state change
******************************************************************************/
-@@ -3112,6 +3412,7 @@ void lru_gen_init_state(struct mem_cgrou
+@@ -3047,6 +3347,7 @@ void lru_gen_init_state(struct mem_cgrou
int i;
int gen, type, zone;
struct lrugen *lrugen = &lruvec->evictable;
@@ -702,7 +702,7 @@ Change-Id: I25d9eda8c6bdc7c3653b9f210a159d6c247c81e8
lrugen->max_seq = MIN_NR_GENS + 1;
lrugen->enabled[0] = lru_gen_enabled() && lru_gen_nr_swapfiles;
-@@ -3122,6 +3423,17 @@ void lru_gen_init_state(struct mem_cgrou
+@@ -3057,6 +3358,17 @@ void lru_gen_init_state(struct mem_cgrou
for_each_gen_type_zone(gen, type, zone)
INIT_LIST_HEAD(&lrugen->lists[gen][type][zone]);
@@ -720,7 +720,7 @@ Change-Id: I25d9eda8c6bdc7c3653b9f210a159d6c247c81e8
}
#ifdef CONFIG_MEMCG
-@@ -3129,18 +3441,37 @@ void lru_gen_init_memcg(struct mem_cgrou
+@@ -3064,18 +3376,37 @@ void lru_gen_init_memcg(struct mem_cgrou
{
int nid;
diff --git a/target/linux/generic/backport-5.15/020-v6.1-06-mm-multigenerational-lru-aging.patch b/target/linux/generic/backport-5.15/020-v6.1-06-mm-multigenerational-lru-aging.patch
index efe3f2e00a..2ff49681c7 100644
--- a/target/linux/generic/backport-5.15/020-v6.1-06-mm-multigenerational-lru-aging.patch
+++ b/target/linux/generic/backport-5.15/020-v6.1-06-mm-multigenerational-lru-aging.patch
@@ -193,7 +193,7 @@ Change-Id: I3ae8abc3100d023cecb3a699d86020ae6fc10a45
#include <asm/tlbflush.h>
#include <asm/div64.h>
-@@ -2887,6 +2889,15 @@ static bool can_age_anon_pages(struct pg
+@@ -2822,6 +2824,15 @@ static bool can_age_anon_pages(struct pg
* shorthand helpers
******************************************************************************/
@@ -209,7 +209,7 @@ Change-Id: I3ae8abc3100d023cecb3a699d86020ae6fc10a45
#define for_each_gen_type_zone(gen, type, zone) \
for ((gen) = 0; (gen) < MAX_NR_GENS; (gen)++) \
for ((type) = 0; (type) < ANON_AND_FILE; (type)++) \
-@@ -2899,6 +2910,12 @@ static int page_lru_gen(struct page *pag
+@@ -2834,6 +2845,12 @@ static int page_lru_gen(struct page *pag
return ((flags & LRU_GEN_MASK) >> LRU_GEN_PGOFF) - 1;
}
@@ -222,7 +222,7 @@ Change-Id: I3ae8abc3100d023cecb3a699d86020ae6fc10a45
static struct lruvec *get_lruvec(int nid, struct mem_cgroup *memcg)
{
struct pglist_data *pgdat = NODE_DATA(nid);
-@@ -3229,6 +3246,926 @@ done:
+@@ -3164,6 +3181,926 @@ done:
}
/******************************************************************************
@@ -1149,7 +1149,7 @@ Change-Id: I3ae8abc3100d023cecb3a699d86020ae6fc10a45
* state change
******************************************************************************/
-@@ -3477,6 +4414,12 @@ static int __init init_lru_gen(void)
+@@ -3412,6 +4349,12 @@ static int __init init_lru_gen(void)
};
late_initcall(init_lru_gen);
@@ -1162,7 +1162,7 @@ Change-Id: I3ae8abc3100d023cecb3a699d86020ae6fc10a45
#endif /* CONFIG_LRU_GEN */
static void shrink_lruvec(struct lruvec *lruvec, struct scan_control *sc)
-@@ -4331,6 +5274,11 @@ static void age_active_anon(struct pglis
+@@ -4266,6 +5209,11 @@ static void age_active_anon(struct pglis
struct mem_cgroup *memcg;
struct lruvec *lruvec;
diff --git a/target/linux/generic/backport-5.15/020-v6.1-07-mm-multigenerational-lru-eviction.patch b/target/linux/generic/backport-5.15/020-v6.1-07-mm-multigenerational-lru-eviction.patch
index 6edb8f9fa7..a75fedecaa 100644
--- a/target/linux/generic/backport-5.15/020-v6.1-07-mm-multigenerational-lru-eviction.patch
+++ b/target/linux/generic/backport-5.15/020-v6.1-07-mm-multigenerational-lru-eviction.patch
@@ -212,7 +212,7 @@ Change-Id: I64c06d8f2cdb83ac7d56c7e1d07f043483956cac
may_enter_fs = (sc->gfp_mask & __GFP_FS) ||
(PageSwapCache(page) && (sc->gfp_mask & __GFP_IO));
-@@ -2570,6 +2577,9 @@ static void prepare_scan_count(pg_data_t
+@@ -2505,6 +2512,9 @@ static void prepare_scan_count(pg_data_t
unsigned long file;
struct lruvec *target_lruvec;
@@ -222,7 +222,7 @@ Change-Id: I64c06d8f2cdb83ac7d56c7e1d07f043483956cac
target_lruvec = mem_cgroup_lruvec(sc->target_mem_cgroup, pgdat);
/*
-@@ -2910,6 +2920,17 @@ static int page_lru_gen(struct page *pag
+@@ -2845,6 +2855,17 @@ static int page_lru_gen(struct page *pag
return ((flags & LRU_GEN_MASK) >> LRU_GEN_PGOFF) - 1;
}
@@ -240,7 +240,7 @@ Change-Id: I64c06d8f2cdb83ac7d56c7e1d07f043483956cac
static int get_swappiness(struct mem_cgroup *memcg)
{
return mem_cgroup_get_nr_swap_pages(memcg) >= MIN_BATCH_SIZE ?
-@@ -3246,6 +3267,91 @@ done:
+@@ -3181,6 +3202,91 @@ done:
}
/******************************************************************************
@@ -332,7 +332,7 @@ Change-Id: I64c06d8f2cdb83ac7d56c7e1d07f043483956cac
* the aging
******************************************************************************/
-@@ -3265,6 +3371,7 @@ static int page_update_gen(struct page *
+@@ -3200,6 +3306,7 @@ static int page_update_gen(struct page *
new_flags &= ~LRU_GEN_MASK;
new_flags |= (gen + 1UL) << LRU_GEN_PGOFF;
@@ -340,7 +340,7 @@ Change-Id: I64c06d8f2cdb83ac7d56c7e1d07f043483956cac
} while (new_flags != old_flags &&
cmpxchg(&page->flags, old_flags, new_flags) != old_flags);
-@@ -3296,6 +3403,7 @@ static void page_inc_gen(struct page *pa
+@@ -3231,6 +3338,7 @@ static void page_inc_gen(struct page *pa
new_flags &= ~LRU_GEN_MASK;
new_flags |= (new_gen + 1UL) << LRU_GEN_PGOFF;
@@ -348,7 +348,7 @@ Change-Id: I64c06d8f2cdb83ac7d56c7e1d07f043483956cac
/* for end_page_writeback() */
if (reclaiming)
new_flags |= BIT(PG_reclaim);
-@@ -3787,6 +3895,7 @@ static bool inc_min_seq(struct lruvec *l
+@@ -3722,6 +3830,7 @@ static bool inc_min_seq(struct lruvec *l
}
}
@@ -356,7 +356,7 @@ Change-Id: I64c06d8f2cdb83ac7d56c7e1d07f043483956cac
WRITE_ONCE(lrugen->min_seq[type], lrugen->min_seq[type] + 1);
return true;
-@@ -3824,6 +3933,8 @@ next:
+@@ -3759,6 +3868,8 @@ next:
if (min_seq[type] == lrugen->min_seq[type])
continue;
@@ -365,7 +365,7 @@ Change-Id: I64c06d8f2cdb83ac7d56c7e1d07f043483956cac
WRITE_ONCE(lrugen->min_seq[type], min_seq[type]);
success = true;
}
-@@ -3885,6 +3996,9 @@ static void inc_max_seq(struct lruvec *l
+@@ -3820,6 +3931,9 @@ static void inc_max_seq(struct lruvec *l
}
}
@@ -375,7 +375,7 @@ Change-Id: I64c06d8f2cdb83ac7d56c7e1d07f043483956cac
WRITE_ONCE(lrugen->timestamps[gen], jiffies);
/* make sure all preceding modifications appear first */
smp_store_release(&lrugen->max_seq, lrugen->max_seq + 1);
-@@ -4166,6 +4280,433 @@ void lru_gen_look_around(struct page_vma
+@@ -4101,6 +4215,433 @@ void lru_gen_look_around(struct page_vma
}
/******************************************************************************
@@ -809,7 +809,7 @@ Change-Id: I64c06d8f2cdb83ac7d56c7e1d07f043483956cac
* state change
******************************************************************************/
-@@ -4420,6 +4961,10 @@ static void lru_gen_age_node(struct pgli
+@@ -4355,6 +4896,10 @@ static void lru_gen_age_node(struct pgli
{
}
@@ -820,7 +820,7 @@ Change-Id: I64c06d8f2cdb83ac7d56c7e1d07f043483956cac
#endif /* CONFIG_LRU_GEN */
static void shrink_lruvec(struct lruvec *lruvec, struct scan_control *sc)
-@@ -4433,6 +4978,11 @@ static void shrink_lruvec(struct lruvec
+@@ -4368,6 +4913,11 @@ static void shrink_lruvec(struct lruvec
bool proportional_reclaim;
struct blk_plug plug;
@@ -832,7 +832,7 @@ Change-Id: I64c06d8f2cdb83ac7d56c7e1d07f043483956cac
get_scan_count(lruvec, sc, nr);
/* Record the original scan target for proportional adjustments later */
-@@ -4904,6 +5454,9 @@ static void snapshot_refaults(struct mem
+@@ -4839,6 +5389,9 @@ static void snapshot_refaults(struct mem
struct lruvec *target_lruvec;
unsigned long refaults;
diff --git a/target/linux/generic/backport-5.15/020-v6.1-08-mm-multigenerational-lru-user-interface.patch b/target/linux/generic/backport-5.15/020-v6.1-08-mm-multigenerational-lru-user-interface.patch
index a1a749fc38..f0753ea802 100644
--- a/target/linux/generic/backport-5.15/020-v6.1-08-mm-multigenerational-lru-user-interface.patch
+++ b/target/linux/generic/backport-5.15/020-v6.1-08-mm-multigenerational-lru-user-interface.patch
@@ -67,7 +67,7 @@ Change-Id: I4448e60029badbe347aa3b624f429b280cc3a3d3
#include <asm/tlbflush.h>
#include <asm/div64.h>
-@@ -4882,6 +4884,413 @@ unlock:
+@@ -4817,6 +4819,413 @@ unlock:
}
/******************************************************************************
@@ -481,7 +481,7 @@ Change-Id: I4448e60029badbe347aa3b624f429b280cc3a3d3
* initialization
******************************************************************************/
-@@ -4951,6 +5360,12 @@ static int __init init_lru_gen(void)
+@@ -4886,6 +5295,12 @@ static int __init init_lru_gen(void)
BUILD_BUG_ON(BIT(LRU_GEN_WIDTH) <= MAX_NR_GENS);
BUILD_BUG_ON(sizeof(MM_STAT_CODES) != NR_MM_STATS + 1);
diff --git a/target/linux/generic/backport-5.15/021-v6.1-mm-mglru-don-t-sync-disk-for-each-aging-cycle.patch b/target/linux/generic/backport-5.15/021-v6.1-mm-mglru-don-t-sync-disk-for-each-aging-cycle.patch
index 269e5eb204..6cc4b3368f 100644
--- a/target/linux/generic/backport-5.15/021-v6.1-mm-mglru-don-t-sync-disk-for-each-aging-cycle.patch
+++ b/target/linux/generic/backport-5.15/021-v6.1-mm-mglru-don-t-sync-disk-for-each-aging-cycle.patch
@@ -21,7 +21,7 @@ Signed-off-by: Andrew Morton <akpm at linux-foundation.org>
--- a/mm/vmscan.c
+++ b/mm/vmscan.c
-@@ -4072,8 +4072,6 @@ static bool try_to_inc_max_seq(struct lr
+@@ -4007,8 +4007,6 @@ static bool try_to_inc_max_seq(struct lr
if (wq_has_sleeper(&lruvec->mm_walk.wait))
wake_up_all(&lruvec->mm_walk.wait);
diff --git a/target/linux/generic/backport-5.15/775-v6.0-01-net-ethernet-stmicro-stmmac-move-queue-reset-to-dedi.patch b/target/linux/generic/backport-5.15/775-v6.0-01-net-ethernet-stmicro-stmmac-move-queue-reset-to-dedi.patch
index 78a5494bf2..7890c99551 100644
--- a/target/linux/generic/backport-5.15/775-v6.0-01-net-ethernet-stmicro-stmmac-move-queue-reset-to-dedi.patch
+++ b/target/linux/generic/backport-5.15/775-v6.0-01-net-ethernet-stmicro-stmmac-move-queue-reset-to-dedi.patch
@@ -27,7 +27,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
static void stmmac_tx_timer_arm(struct stmmac_priv *priv, u32 queue);
static void stmmac_flush_tx_descriptors(struct stmmac_priv *priv, int queue);
-@@ -1697,9 +1700,6 @@ static int __init_dma_rx_desc_rings(stru
+@@ -1705,9 +1708,6 @@ static int __init_dma_rx_desc_rings(stru
return -ENOMEM;
}
@@ -37,7 +37,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
/* Setup the chained descriptor addresses */
if (priv->mode == STMMAC_CHAIN_MODE) {
if (priv->extend_desc)
-@@ -1805,12 +1805,6 @@ static int __init_dma_tx_desc_rings(stru
+@@ -1813,12 +1813,6 @@ static int __init_dma_tx_desc_rings(stru
tx_q->tx_skbuff[i] = NULL;
}
@@ -50,7 +50,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
return 0;
}
-@@ -2679,10 +2673,7 @@ static void stmmac_tx_err(struct stmmac_
+@@ -2687,10 +2681,7 @@ static void stmmac_tx_err(struct stmmac_
stmmac_stop_tx_dma(priv, chan);
dma_free_tx_skbufs(priv, chan);
stmmac_clear_tx_descriptors(priv, chan);
@@ -62,7 +62,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
stmmac_init_tx_chan(priv, priv->ioaddr, priv->plat->dma_cfg,
tx_q->dma_tx_phy, chan);
stmmac_start_tx_dma(priv, chan);
-@@ -3766,6 +3757,8 @@ static int stmmac_open(struct net_device
+@@ -3774,6 +3765,8 @@ static int stmmac_open(struct net_device
}
}
@@ -71,7 +71,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
ret = stmmac_hw_setup(dev, true);
if (ret < 0) {
netdev_err(priv->dev, "%s: Hw setup failed\n", __func__);
-@@ -6409,6 +6402,7 @@ void stmmac_enable_rx_queue(struct stmma
+@@ -6417,6 +6410,7 @@ void stmmac_enable_rx_queue(struct stmma
return;
}
@@ -79,7 +79,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
stmmac_clear_rx_descriptors(priv, queue);
stmmac_init_rx_chan(priv, priv->ioaddr, priv->plat->dma_cfg,
-@@ -6470,6 +6464,7 @@ void stmmac_enable_tx_queue(struct stmma
+@@ -6478,6 +6472,7 @@ void stmmac_enable_tx_queue(struct stmma
return;
}
@@ -87,7 +87,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
stmmac_clear_tx_descriptors(priv, queue);
stmmac_init_tx_chan(priv, priv->ioaddr, priv->plat->dma_cfg,
-@@ -7390,6 +7385,25 @@ int stmmac_suspend(struct device *dev)
+@@ -7398,6 +7393,25 @@ int stmmac_suspend(struct device *dev)
}
EXPORT_SYMBOL_GPL(stmmac_suspend);
@@ -113,7 +113,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
/**
* stmmac_reset_queues_param - reset queue parameters
* @priv: device pointer
-@@ -7400,22 +7414,11 @@ static void stmmac_reset_queues_param(st
+@@ -7408,22 +7422,11 @@ static void stmmac_reset_queues_param(st
u32 tx_cnt = priv->plat->tx_queues_to_use;
u32 queue;
diff --git a/target/linux/generic/backport-5.15/775-v6.0-02-net-ethernet-stmicro-stmmac-first-disable-all-queues.patch b/target/linux/generic/backport-5.15/775-v6.0-02-net-ethernet-stmicro-stmmac-first-disable-all-queues.patch
index 0bb28c79e7..ee3b92f06b 100644
--- a/target/linux/generic/backport-5.15/775-v6.0-02-net-ethernet-stmicro-stmmac-first-disable-all-queues.patch
+++ b/target/linux/generic/backport-5.15/775-v6.0-02-net-ethernet-stmicro-stmmac-first-disable-all-queues.patch
@@ -17,7 +17,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
--- a/drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
+++ b/drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
-@@ -3818,8 +3818,6 @@ static int stmmac_release(struct net_dev
+@@ -3826,8 +3826,6 @@ static int stmmac_release(struct net_dev
struct stmmac_priv *priv = netdev_priv(dev);
u32 chan;
@@ -26,7 +26,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
if (device_may_wakeup(priv->device))
phylink_speed_down(priv->phylink, false);
/* Stop and disconnect the PHY */
-@@ -3831,6 +3829,8 @@ static int stmmac_release(struct net_dev
+@@ -3839,6 +3837,8 @@ static int stmmac_release(struct net_dev
for (chan = 0; chan < priv->plat->tx_queues_to_use; chan++)
hrtimer_cancel(&priv->tx_queue[chan].txtimer);
diff --git a/target/linux/generic/backport-5.15/775-v6.0-03-net-ethernet-stmicro-stmmac-move-dma-conf-to-dedicat.patch b/target/linux/generic/backport-5.15/775-v6.0-03-net-ethernet-stmicro-stmmac-move-dma-conf-to-dedicat.patch
index 5cfb14aaab..46375a6fa9 100644
--- a/target/linux/generic/backport-5.15/775-v6.0-03-net-ethernet-stmicro-stmmac-move-dma-conf-to-dedicat.patch
+++ b/target/linux/generic/backport-5.15/775-v6.0-03-net-ethernet-stmicro-stmmac-move-dma-conf-to-dedicat.patch
@@ -189,7 +189,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
if (tx_q->dirty_tx != tx_q->cur_tx)
return -EBUSY; /* still unfinished work */
-@@ -1294,7 +1294,7 @@ static void stmmac_display_rx_rings(stru
+@@ -1302,7 +1302,7 @@ static void stmmac_display_rx_rings(stru
/* Display RX rings */
for (queue = 0; queue < rx_cnt; queue++) {
@@ -198,7 +198,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
pr_info("\tRX Queue %u rings\n", queue);
-@@ -1307,7 +1307,7 @@ static void stmmac_display_rx_rings(stru
+@@ -1315,7 +1315,7 @@ static void stmmac_display_rx_rings(stru
}
/* Display RX ring */
@@ -207,7 +207,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
rx_q->dma_rx_phy, desc_size);
}
}
-@@ -1321,7 +1321,7 @@ static void stmmac_display_tx_rings(stru
+@@ -1329,7 +1329,7 @@ static void stmmac_display_tx_rings(stru
/* Display TX rings */
for (queue = 0; queue < tx_cnt; queue++) {
@@ -216,7 +216,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
pr_info("\tTX Queue %d rings\n", queue);
-@@ -1336,7 +1336,7 @@ static void stmmac_display_tx_rings(stru
+@@ -1344,7 +1344,7 @@ static void stmmac_display_tx_rings(stru
desc_size = sizeof(struct dma_desc);
}
@@ -225,7 +225,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
tx_q->dma_tx_phy, desc_size);
}
}
-@@ -1377,21 +1377,21 @@ static int stmmac_set_bfsize(int mtu, in
+@@ -1385,21 +1385,21 @@ static int stmmac_set_bfsize(int mtu, in
*/
static void stmmac_clear_rx_descriptors(struct stmmac_priv *priv, u32 queue)
{
@@ -253,7 +253,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
}
/**
-@@ -1403,12 +1403,12 @@ static void stmmac_clear_rx_descriptors(
+@@ -1411,12 +1411,12 @@ static void stmmac_clear_rx_descriptors(
*/
static void stmmac_clear_tx_descriptors(struct stmmac_priv *priv, u32 queue)
{
@@ -269,7 +269,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
struct dma_desc *p;
if (priv->extend_desc)
-@@ -1456,7 +1456,7 @@ static void stmmac_clear_descriptors(str
+@@ -1464,7 +1464,7 @@ static void stmmac_clear_descriptors(str
static int stmmac_init_rx_buffers(struct stmmac_priv *priv, struct dma_desc *p,
int i, gfp_t flags, u32 queue)
{
@@ -278,7 +278,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
struct stmmac_rx_buffer *buf = &rx_q->buf_pool[i];
if (!buf->page) {
-@@ -1481,7 +1481,7 @@ static int stmmac_init_rx_buffers(struct
+@@ -1489,7 +1489,7 @@ static int stmmac_init_rx_buffers(struct
buf->addr = page_pool_get_dma_addr(buf->page) + buf->page_offset;
stmmac_set_desc_addr(priv, p, buf->addr);
@@ -287,7 +287,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
stmmac_init_desc3(priv, p);
return 0;
-@@ -1495,7 +1495,7 @@ static int stmmac_init_rx_buffers(struct
+@@ -1503,7 +1503,7 @@ static int stmmac_init_rx_buffers(struct
*/
static void stmmac_free_rx_buffer(struct stmmac_priv *priv, u32 queue, int i)
{
@@ -296,7 +296,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
struct stmmac_rx_buffer *buf = &rx_q->buf_pool[i];
if (buf->page)
-@@ -1515,7 +1515,7 @@ static void stmmac_free_rx_buffer(struct
+@@ -1523,7 +1523,7 @@ static void stmmac_free_rx_buffer(struct
*/
static void stmmac_free_tx_buffer(struct stmmac_priv *priv, u32 queue, int i)
{
@@ -305,7 +305,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
if (tx_q->tx_skbuff_dma[i].buf &&
tx_q->tx_skbuff_dma[i].buf_type != STMMAC_TXBUF_T_XDP_TX) {
-@@ -1560,17 +1560,17 @@ static void dma_free_rx_skbufs(struct st
+@@ -1568,17 +1568,17 @@ static void dma_free_rx_skbufs(struct st
{
int i;
@@ -326,7 +326,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
struct dma_desc *p;
int ret;
-@@ -1597,10 +1597,10 @@ static int stmmac_alloc_rx_buffers(struc
+@@ -1605,10 +1605,10 @@ static int stmmac_alloc_rx_buffers(struc
*/
static void dma_free_rx_xskbufs(struct stmmac_priv *priv, u32 queue)
{
@@ -339,7 +339,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
struct stmmac_rx_buffer *buf = &rx_q->buf_pool[i];
if (!buf->xdp)
-@@ -1613,10 +1613,10 @@ static void dma_free_rx_xskbufs(struct s
+@@ -1621,10 +1621,10 @@ static void dma_free_rx_xskbufs(struct s
static int stmmac_alloc_rx_buffers_zc(struct stmmac_priv *priv, u32 queue)
{
@@ -352,7 +352,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
struct stmmac_rx_buffer *buf;
dma_addr_t dma_addr;
struct dma_desc *p;
-@@ -1659,7 +1659,7 @@ static struct xsk_buff_pool *stmmac_get_
+@@ -1667,7 +1667,7 @@ static struct xsk_buff_pool *stmmac_get_
*/
static int __init_dma_rx_desc_rings(struct stmmac_priv *priv, u32 queue, gfp_t flags)
{
@@ -361,7 +361,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
int ret;
netif_dbg(priv, probe, priv->dev,
-@@ -1705,11 +1705,11 @@ static int __init_dma_rx_desc_rings(stru
+@@ -1713,11 +1713,11 @@ static int __init_dma_rx_desc_rings(stru
if (priv->extend_desc)
stmmac_mode_init(priv, rx_q->dma_erx,
rx_q->dma_rx_phy,
@@ -375,7 +375,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
}
return 0;
-@@ -1736,7 +1736,7 @@ static int init_dma_rx_desc_rings(struct
+@@ -1744,7 +1744,7 @@ static int init_dma_rx_desc_rings(struct
err_init_rx_buffers:
while (queue >= 0) {
@@ -384,7 +384,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
if (rx_q->xsk_pool)
dma_free_rx_xskbufs(priv, queue);
-@@ -1765,7 +1765,7 @@ err_init_rx_buffers:
+@@ -1773,7 +1773,7 @@ err_init_rx_buffers:
*/
static int __init_dma_tx_desc_rings(struct stmmac_priv *priv, u32 queue)
{
@@ -393,7 +393,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
int i;
netif_dbg(priv, probe, priv->dev,
-@@ -1777,16 +1777,16 @@ static int __init_dma_tx_desc_rings(stru
+@@ -1785,16 +1785,16 @@ static int __init_dma_tx_desc_rings(stru
if (priv->extend_desc)
stmmac_mode_init(priv, tx_q->dma_etx,
tx_q->dma_tx_phy,
@@ -413,7 +413,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
struct dma_desc *p;
if (priv->extend_desc)
-@@ -1856,12 +1856,12 @@ static int init_dma_desc_rings(struct ne
+@@ -1864,12 +1864,12 @@ static int init_dma_desc_rings(struct ne
*/
static void dma_free_tx_skbufs(struct stmmac_priv *priv, u32 queue)
{
@@ -428,7 +428,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
stmmac_free_tx_buffer(priv, queue, i);
if (tx_q->xsk_pool && tx_q->xsk_frames_done) {
-@@ -1891,7 +1891,7 @@ static void stmmac_free_tx_skbufs(struct
+@@ -1899,7 +1899,7 @@ static void stmmac_free_tx_skbufs(struct
*/
static void __free_dma_rx_desc_resources(struct stmmac_priv *priv, u32 queue)
{
@@ -437,7 +437,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
/* Release the DMA RX socket buffers */
if (rx_q->xsk_pool)
-@@ -1904,11 +1904,11 @@ static void __free_dma_rx_desc_resources
+@@ -1912,11 +1912,11 @@ static void __free_dma_rx_desc_resources
/* Free DMA regions of consistent memory previously allocated */
if (!priv->extend_desc)
@@ -451,7 +451,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
sizeof(struct dma_extended_desc),
rx_q->dma_erx, rx_q->dma_rx_phy);
-@@ -1937,7 +1937,7 @@ static void free_dma_rx_desc_resources(s
+@@ -1945,7 +1945,7 @@ static void free_dma_rx_desc_resources(s
*/
static void __free_dma_tx_desc_resources(struct stmmac_priv *priv, u32 queue)
{
@@ -460,7 +460,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
size_t size;
void *addr;
-@@ -1955,7 +1955,7 @@ static void __free_dma_tx_desc_resources
+@@ -1963,7 +1963,7 @@ static void __free_dma_tx_desc_resources
addr = tx_q->dma_tx;
}
@@ -469,7 +469,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
dma_free_coherent(priv->device, size, addr, tx_q->dma_tx_phy);
-@@ -1984,7 +1984,7 @@ static void free_dma_tx_desc_resources(s
+@@ -1992,7 +1992,7 @@ static void free_dma_tx_desc_resources(s
*/
static int __alloc_dma_rx_desc_resources(struct stmmac_priv *priv, u32 queue)
{
@@ -478,7 +478,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
struct stmmac_channel *ch = &priv->channel[queue];
bool xdp_prog = stmmac_xdp_is_enabled(priv);
struct page_pool_params pp_params = { 0 };
-@@ -1996,8 +1996,8 @@ static int __alloc_dma_rx_desc_resources
+@@ -2004,8 +2004,8 @@ static int __alloc_dma_rx_desc_resources
rx_q->priv_data = priv;
pp_params.flags = PP_FLAG_DMA_MAP | PP_FLAG_DMA_SYNC_DEV;
@@ -489,7 +489,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
pp_params.order = ilog2(num_pages);
pp_params.nid = dev_to_node(priv->device);
pp_params.dev = priv->device;
-@@ -2012,7 +2012,7 @@ static int __alloc_dma_rx_desc_resources
+@@ -2020,7 +2020,7 @@ static int __alloc_dma_rx_desc_resources
return ret;
}
@@ -498,7 +498,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
sizeof(*rx_q->buf_pool),
GFP_KERNEL);
if (!rx_q->buf_pool)
-@@ -2020,7 +2020,7 @@ static int __alloc_dma_rx_desc_resources
+@@ -2028,7 +2028,7 @@ static int __alloc_dma_rx_desc_resources
if (priv->extend_desc) {
rx_q->dma_erx = dma_alloc_coherent(priv->device,
@@ -507,7 +507,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
sizeof(struct dma_extended_desc),
&rx_q->dma_rx_phy,
GFP_KERNEL);
-@@ -2029,7 +2029,7 @@ static int __alloc_dma_rx_desc_resources
+@@ -2037,7 +2037,7 @@ static int __alloc_dma_rx_desc_resources
} else {
rx_q->dma_rx = dma_alloc_coherent(priv->device,
@@ -516,7 +516,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
sizeof(struct dma_desc),
&rx_q->dma_rx_phy,
GFP_KERNEL);
-@@ -2086,20 +2086,20 @@ err_dma:
+@@ -2094,20 +2094,20 @@ err_dma:
*/
static int __alloc_dma_tx_desc_resources(struct stmmac_priv *priv, u32 queue)
{
@@ -540,7 +540,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
sizeof(struct sk_buff *),
GFP_KERNEL);
if (!tx_q->tx_skbuff)
-@@ -2112,7 +2112,7 @@ static int __alloc_dma_tx_desc_resources
+@@ -2120,7 +2120,7 @@ static int __alloc_dma_tx_desc_resources
else
size = sizeof(struct dma_desc);
@@ -549,7 +549,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
addr = dma_alloc_coherent(priv->device, size,
&tx_q->dma_tx_phy, GFP_KERNEL);
-@@ -2356,7 +2356,7 @@ static void stmmac_dma_operation_mode(st
+@@ -2364,7 +2364,7 @@ static void stmmac_dma_operation_mode(st
/* configure all channels */
for (chan = 0; chan < rx_channels_count; chan++) {
@@ -558,7 +558,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
u32 buf_size;
qmode = priv->plat->rx_queues_cfg[chan].mode_to_use;
-@@ -2371,7 +2371,7 @@ static void stmmac_dma_operation_mode(st
+@@ -2379,7 +2379,7 @@ static void stmmac_dma_operation_mode(st
chan);
} else {
stmmac_set_dma_bfsize(priv, priv->ioaddr,
@@ -567,7 +567,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
chan);
}
}
-@@ -2387,7 +2387,7 @@ static void stmmac_dma_operation_mode(st
+@@ -2395,7 +2395,7 @@ static void stmmac_dma_operation_mode(st
static bool stmmac_xdp_xmit_zc(struct stmmac_priv *priv, u32 queue, u32 budget)
{
struct netdev_queue *nq = netdev_get_tx_queue(priv->dev, queue);
@@ -576,7 +576,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
struct xsk_buff_pool *pool = tx_q->xsk_pool;
unsigned int entry = tx_q->cur_tx;
struct dma_desc *tx_desc = NULL;
-@@ -2462,7 +2462,7 @@ static bool stmmac_xdp_xmit_zc(struct st
+@@ -2470,7 +2470,7 @@ static bool stmmac_xdp_xmit_zc(struct st
stmmac_enable_dma_transmission(priv, priv->ioaddr);
@@ -585,7 +585,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
entry = tx_q->cur_tx;
}
-@@ -2488,7 +2488,7 @@ static bool stmmac_xdp_xmit_zc(struct st
+@@ -2496,7 +2496,7 @@ static bool stmmac_xdp_xmit_zc(struct st
*/
static int stmmac_tx_clean(struct stmmac_priv *priv, int budget, u32 queue)
{
@@ -594,7 +594,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
unsigned int bytes_compl = 0, pkts_compl = 0;
unsigned int entry, xmits = 0, count = 0;
-@@ -2501,7 +2501,7 @@ static int stmmac_tx_clean(struct stmmac
+@@ -2509,7 +2509,7 @@ static int stmmac_tx_clean(struct stmmac
entry = tx_q->dirty_tx;
/* Try to clean all TX complete frame in 1 shot */
@@ -603,7 +603,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
struct xdp_frame *xdpf;
struct sk_buff *skb;
struct dma_desc *p;
-@@ -2601,7 +2601,7 @@ static int stmmac_tx_clean(struct stmmac
+@@ -2609,7 +2609,7 @@ static int stmmac_tx_clean(struct stmmac
stmmac_release_tx_desc(priv, p, priv->mode);
@@ -612,7 +612,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
}
tx_q->dirty_tx = entry;
-@@ -2666,7 +2666,7 @@ static int stmmac_tx_clean(struct stmmac
+@@ -2674,7 +2674,7 @@ static int stmmac_tx_clean(struct stmmac
*/
static void stmmac_tx_err(struct stmmac_priv *priv, u32 chan)
{
@@ -621,7 +621,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
netif_tx_stop_queue(netdev_get_tx_queue(priv->dev, chan));
-@@ -2733,8 +2733,8 @@ static int stmmac_napi_check(struct stmm
+@@ -2741,8 +2741,8 @@ static int stmmac_napi_check(struct stmm
{
int status = stmmac_dma_interrupt_status(priv, priv->ioaddr,
&priv->xstats, chan, dir);
@@ -632,7 +632,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
struct stmmac_channel *ch = &priv->channel[chan];
struct napi_struct *rx_napi;
struct napi_struct *tx_napi;
-@@ -2910,7 +2910,7 @@ static int stmmac_init_dma_engine(struct
+@@ -2918,7 +2918,7 @@ static int stmmac_init_dma_engine(struct
/* DMA RX Channel Configuration */
for (chan = 0; chan < rx_channels_count; chan++) {
@@ -641,7 +641,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
stmmac_init_rx_chan(priv, priv->ioaddr, priv->plat->dma_cfg,
rx_q->dma_rx_phy, chan);
-@@ -2924,7 +2924,7 @@ static int stmmac_init_dma_engine(struct
+@@ -2932,7 +2932,7 @@ static int stmmac_init_dma_engine(struct
/* DMA TX Channel Configuration */
for (chan = 0; chan < tx_channels_count; chan++) {
@@ -650,7 +650,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
stmmac_init_tx_chan(priv, priv->ioaddr, priv->plat->dma_cfg,
tx_q->dma_tx_phy, chan);
-@@ -2939,7 +2939,7 @@ static int stmmac_init_dma_engine(struct
+@@ -2947,7 +2947,7 @@ static int stmmac_init_dma_engine(struct
static void stmmac_tx_timer_arm(struct stmmac_priv *priv, u32 queue)
{
@@ -659,7 +659,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
hrtimer_start(&tx_q->txtimer,
STMMAC_COAL_TIMER(priv->tx_coal_timer[queue]),
-@@ -2989,7 +2989,7 @@ static void stmmac_init_coalesce(struct
+@@ -2997,7 +2997,7 @@ static void stmmac_init_coalesce(struct
u32 chan;
for (chan = 0; chan < tx_channel_count; chan++) {
@@ -668,7 +668,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
priv->tx_coal_frames[chan] = STMMAC_TX_FRAMES;
priv->tx_coal_timer[chan] = STMMAC_COAL_TX_TIMER;
-@@ -3011,12 +3011,12 @@ static void stmmac_set_rings_length(stru
+@@ -3019,12 +3019,12 @@ static void stmmac_set_rings_length(stru
/* set TX ring length */
for (chan = 0; chan < tx_channels_count; chan++)
stmmac_set_tx_ring_len(priv, priv->ioaddr,
@@ -683,7 +683,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
}
/**
-@@ -3351,7 +3351,7 @@ static int stmmac_hw_setup(struct net_de
+@@ -3359,7 +3359,7 @@ static int stmmac_hw_setup(struct net_de
/* Enable TSO */
if (priv->tso) {
for (chan = 0; chan < tx_cnt; chan++) {
@@ -692,7 +692,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
/* TSO and TBS cannot co-exist */
if (tx_q->tbs & STMMAC_TBS_AVAIL)
-@@ -3373,7 +3373,7 @@ static int stmmac_hw_setup(struct net_de
+@@ -3381,7 +3381,7 @@ static int stmmac_hw_setup(struct net_de
/* TBS */
for (chan = 0; chan < tx_cnt; chan++) {
@@ -701,7 +701,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
int enable = tx_q->tbs & STMMAC_TBS_AVAIL;
stmmac_enable_tbs(priv, priv->ioaddr, enable, chan);
-@@ -3417,7 +3417,7 @@ static void stmmac_free_irq(struct net_d
+@@ -3425,7 +3425,7 @@ static void stmmac_free_irq(struct net_d
for (j = irq_idx - 1; j >= 0; j--) {
if (priv->tx_irq[j] > 0) {
irq_set_affinity_hint(priv->tx_irq[j], NULL);
@@ -710,7 +710,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
}
}
irq_idx = priv->plat->rx_queues_to_use;
-@@ -3426,7 +3426,7 @@ static void stmmac_free_irq(struct net_d
+@@ -3434,7 +3434,7 @@ static void stmmac_free_irq(struct net_d
for (j = irq_idx - 1; j >= 0; j--) {
if (priv->rx_irq[j] > 0) {
irq_set_affinity_hint(priv->rx_irq[j], NULL);
@@ -719,7 +719,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
}
}
-@@ -3559,7 +3559,7 @@ static int stmmac_request_irq_multi_msi(
+@@ -3567,7 +3567,7 @@ static int stmmac_request_irq_multi_msi(
sprintf(int_name, "%s:%s-%d", dev->name, "rx", i);
ret = request_irq(priv->rx_irq[i],
stmmac_msi_intr_rx,
@@ -728,7 +728,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
if (unlikely(ret < 0)) {
netdev_err(priv->dev,
"%s: alloc rx-%d MSI %d (error: %d)\n",
-@@ -3582,7 +3582,7 @@ static int stmmac_request_irq_multi_msi(
+@@ -3590,7 +3590,7 @@ static int stmmac_request_irq_multi_msi(
sprintf(int_name, "%s:%s-%d", dev->name, "tx", i);
ret = request_irq(priv->tx_irq[i],
stmmac_msi_intr_tx,
@@ -737,7 +737,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
if (unlikely(ret < 0)) {
netdev_err(priv->dev,
"%s: alloc tx-%d MSI %d (error: %d)\n",
-@@ -3713,21 +3713,21 @@ static int stmmac_open(struct net_device
+@@ -3721,21 +3721,21 @@ static int stmmac_open(struct net_device
bfsize = 0;
if (bfsize < BUF_SIZE_16KiB)
@@ -766,7 +766,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
int tbs_en = priv->plat->tx_queues_cfg[chan].tbs_en;
/* Setup per-TXQ tbs flag before TX descriptor alloc */
-@@ -3785,7 +3785,7 @@ irq_error:
+@@ -3793,7 +3793,7 @@ irq_error:
phylink_stop(priv->phylink);
for (chan = 0; chan < priv->plat->tx_queues_to_use; chan++)
@@ -775,7 +775,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
stmmac_hw_teardown(dev);
init_error:
-@@ -3827,7 +3827,7 @@ static int stmmac_release(struct net_dev
+@@ -3835,7 +3835,7 @@ static int stmmac_release(struct net_dev
stmmac_disable_all_queues(priv);
for (chan = 0; chan < priv->plat->tx_queues_to_use; chan++)
@@ -784,7 +784,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
netif_tx_disable(dev);
-@@ -3891,7 +3891,7 @@ static bool stmmac_vlan_insert(struct st
+@@ -3899,7 +3899,7 @@ static bool stmmac_vlan_insert(struct st
return false;
stmmac_set_tx_owner(priv, p);
@@ -793,7 +793,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
return true;
}
-@@ -3909,7 +3909,7 @@ static bool stmmac_vlan_insert(struct st
+@@ -3917,7 +3917,7 @@ static bool stmmac_vlan_insert(struct st
static void stmmac_tso_allocator(struct stmmac_priv *priv, dma_addr_t des,
int total_len, bool last_segment, u32 queue)
{
@@ -802,7 +802,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
struct dma_desc *desc;
u32 buff_size;
int tmp_len;
-@@ -3920,7 +3920,7 @@ static void stmmac_tso_allocator(struct
+@@ -3928,7 +3928,7 @@ static void stmmac_tso_allocator(struct
dma_addr_t curr_addr;
tx_q->cur_tx = STMMAC_GET_ENTRY(tx_q->cur_tx,
@@ -811,7 +811,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
WARN_ON(tx_q->tx_skbuff[tx_q->cur_tx]);
if (tx_q->tbs & STMMAC_TBS_AVAIL)
-@@ -3948,7 +3948,7 @@ static void stmmac_tso_allocator(struct
+@@ -3956,7 +3956,7 @@ static void stmmac_tso_allocator(struct
static void stmmac_flush_tx_descriptors(struct stmmac_priv *priv, int queue)
{
@@ -820,7 +820,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
int desc_size;
if (likely(priv->extend_desc))
-@@ -4010,7 +4010,7 @@ static netdev_tx_t stmmac_tso_xmit(struc
+@@ -4018,7 +4018,7 @@ static netdev_tx_t stmmac_tso_xmit(struc
dma_addr_t des;
int i;
@@ -829,7 +829,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
first_tx = tx_q->cur_tx;
/* Compute header lengths */
-@@ -4050,7 +4050,7 @@ static netdev_tx_t stmmac_tso_xmit(struc
+@@ -4058,7 +4058,7 @@ static netdev_tx_t stmmac_tso_xmit(struc
stmmac_set_mss(priv, mss_desc, mss);
tx_q->mss = mss;
tx_q->cur_tx = STMMAC_GET_ENTRY(tx_q->cur_tx,
@@ -838,7 +838,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
WARN_ON(tx_q->tx_skbuff[tx_q->cur_tx]);
}
-@@ -4162,7 +4162,7 @@ static netdev_tx_t stmmac_tso_xmit(struc
+@@ -4170,7 +4170,7 @@ static netdev_tx_t stmmac_tso_xmit(struc
* ndo_start_xmit will fill this descriptor the next time it's
* called and stmmac_tx_clean may clean up to this descriptor.
*/
@@ -847,7 +847,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
if (unlikely(stmmac_tx_avail(priv, queue) <= (MAX_SKB_FRAGS + 1))) {
netif_dbg(priv, hw, priv->dev, "%s: stop transmitted packets\n",
-@@ -4250,7 +4250,7 @@ static netdev_tx_t stmmac_xmit(struct sk
+@@ -4258,7 +4258,7 @@ static netdev_tx_t stmmac_xmit(struct sk
int entry, first_tx;
dma_addr_t des;
@@ -856,7 +856,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
first_tx = tx_q->cur_tx;
if (priv->tx_path_in_lpi_mode && priv->eee_sw_timer_en)
-@@ -4313,7 +4313,7 @@ static netdev_tx_t stmmac_xmit(struct sk
+@@ -4321,7 +4321,7 @@ static netdev_tx_t stmmac_xmit(struct sk
int len = skb_frag_size(frag);
bool last_segment = (i == (nfrags - 1));
@@ -865,7 +865,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
WARN_ON(tx_q->tx_skbuff[entry]);
if (likely(priv->extend_desc))
-@@ -4384,7 +4384,7 @@ static netdev_tx_t stmmac_xmit(struct sk
+@@ -4392,7 +4392,7 @@ static netdev_tx_t stmmac_xmit(struct sk
* ndo_start_xmit will fill this descriptor the next time it's
* called and stmmac_tx_clean may clean up to this descriptor.
*/
@@ -874,7 +874,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
tx_q->cur_tx = entry;
if (netif_msg_pktdata(priv)) {
-@@ -4499,7 +4499,7 @@ static void stmmac_rx_vlan(struct net_de
+@@ -4507,7 +4507,7 @@ static void stmmac_rx_vlan(struct net_de
*/
static inline void stmmac_rx_refill(struct stmmac_priv *priv, u32 queue)
{
@@ -883,7 +883,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
int dirty = stmmac_rx_dirty(priv, queue);
unsigned int entry = rx_q->dirty_rx;
-@@ -4549,7 +4549,7 @@ static inline void stmmac_rx_refill(stru
+@@ -4557,7 +4557,7 @@ static inline void stmmac_rx_refill(stru
dma_wmb();
stmmac_set_rx_owner(priv, p, use_rx_wd);
@@ -892,7 +892,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
}
rx_q->dirty_rx = entry;
rx_q->rx_tail_addr = rx_q->dma_rx_phy +
-@@ -4577,12 +4577,12 @@ static unsigned int stmmac_rx_buf1_len(s
+@@ -4585,12 +4585,12 @@ static unsigned int stmmac_rx_buf1_len(s
/* First descriptor, not last descriptor and not split header */
if (status & rx_not_ls)
@@ -907,7 +907,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
}
static unsigned int stmmac_rx_buf2_len(struct stmmac_priv *priv,
-@@ -4598,7 +4598,7 @@ static unsigned int stmmac_rx_buf2_len(s
+@@ -4606,7 +4606,7 @@ static unsigned int stmmac_rx_buf2_len(s
/* Not last descriptor */
if (status & rx_not_ls)
@@ -916,7 +916,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
plen = stmmac_get_rx_frame_len(priv, p, coe);
-@@ -4609,7 +4609,7 @@ static unsigned int stmmac_rx_buf2_len(s
+@@ -4617,7 +4617,7 @@ static unsigned int stmmac_rx_buf2_len(s
static int stmmac_xdp_xmit_xdpf(struct stmmac_priv *priv, int queue,
struct xdp_frame *xdpf, bool dma_map)
{
@@ -925,7 +925,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
unsigned int entry = tx_q->cur_tx;
struct dma_desc *tx_desc;
dma_addr_t dma_addr;
-@@ -4672,7 +4672,7 @@ static int stmmac_xdp_xmit_xdpf(struct s
+@@ -4680,7 +4680,7 @@ static int stmmac_xdp_xmit_xdpf(struct s
stmmac_enable_dma_transmission(priv, priv->ioaddr);
@@ -934,7 +934,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
tx_q->cur_tx = entry;
return STMMAC_XDP_TX;
-@@ -4846,7 +4846,7 @@ static void stmmac_dispatch_skb_zc(struc
+@@ -4854,7 +4854,7 @@ static void stmmac_dispatch_skb_zc(struc
static bool stmmac_rx_refill_zc(struct stmmac_priv *priv, u32 queue, u32 budget)
{
@@ -943,7 +943,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
unsigned int entry = rx_q->dirty_rx;
struct dma_desc *rx_desc = NULL;
bool ret = true;
-@@ -4889,7 +4889,7 @@ static bool stmmac_rx_refill_zc(struct s
+@@ -4897,7 +4897,7 @@ static bool stmmac_rx_refill_zc(struct s
dma_wmb();
stmmac_set_rx_owner(priv, rx_desc, use_rx_wd);
@@ -952,7 +952,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
}
if (rx_desc) {
-@@ -4904,7 +4904,7 @@ static bool stmmac_rx_refill_zc(struct s
+@@ -4912,7 +4912,7 @@ static bool stmmac_rx_refill_zc(struct s
static int stmmac_rx_zc(struct stmmac_priv *priv, int limit, u32 queue)
{
@@ -961,7 +961,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
unsigned int count = 0, error = 0, len = 0;
int dirty = stmmac_rx_dirty(priv, queue);
unsigned int next_entry = rx_q->cur_rx;
-@@ -4926,7 +4926,7 @@ static int stmmac_rx_zc(struct stmmac_pr
+@@ -4934,7 +4934,7 @@ static int stmmac_rx_zc(struct stmmac_pr
desc_size = sizeof(struct dma_desc);
}
@@ -970,7 +970,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
rx_q->dma_rx_phy, desc_size);
}
while (count < limit) {
-@@ -4973,7 +4973,7 @@ read_again:
+@@ -4981,7 +4981,7 @@ read_again:
/* Prefetch the next RX descriptor */
rx_q->cur_rx = STMMAC_GET_ENTRY(rx_q->cur_rx,
@@ -979,7 +979,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
next_entry = rx_q->cur_rx;
if (priv->extend_desc)
-@@ -5094,7 +5094,7 @@ read_again:
+@@ -5102,7 +5102,7 @@ read_again:
*/
static int stmmac_rx(struct stmmac_priv *priv, int limit, u32 queue)
{
@@ -988,7 +988,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
struct stmmac_channel *ch = &priv->channel[queue];
unsigned int count = 0, error = 0, len = 0;
int status = 0, coe = priv->hw->rx_csum;
-@@ -5107,7 +5107,7 @@ static int stmmac_rx(struct stmmac_priv
+@@ -5115,7 +5115,7 @@ static int stmmac_rx(struct stmmac_priv
int buf_sz;
dma_dir = page_pool_get_dma_dir(rx_q->page_pool);
@@ -997,7 +997,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
if (netif_msg_rx_status(priv)) {
void *rx_head;
-@@ -5121,7 +5121,7 @@ static int stmmac_rx(struct stmmac_priv
+@@ -5129,7 +5129,7 @@ static int stmmac_rx(struct stmmac_priv
desc_size = sizeof(struct dma_desc);
}
@@ -1006,7 +1006,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
rx_q->dma_rx_phy, desc_size);
}
while (count < limit) {
-@@ -5165,7 +5165,7 @@ read_again:
+@@ -5173,7 +5173,7 @@ read_again:
break;
rx_q->cur_rx = STMMAC_GET_ENTRY(rx_q->cur_rx,
@@ -1015,7 +1015,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
next_entry = rx_q->cur_rx;
if (priv->extend_desc)
-@@ -5299,7 +5299,7 @@ read_again:
+@@ -5307,7 +5307,7 @@ read_again:
buf1_len, dma_dir);
skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags,
buf->page, buf->page_offset, buf1_len,
@@ -1024,7 +1024,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
/* Data payload appended into SKB */
page_pool_release_page(rx_q->page_pool, buf->page);
-@@ -5311,7 +5311,7 @@ read_again:
+@@ -5319,7 +5319,7 @@ read_again:
buf2_len, dma_dir);
skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags,
buf->sec_page, 0, buf2_len,
@@ -1033,7 +1033,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
/* Data payload appended into SKB */
page_pool_release_page(rx_q->page_pool, buf->sec_page);
-@@ -5753,11 +5753,13 @@ static irqreturn_t stmmac_safety_interru
+@@ -5761,11 +5761,13 @@ static irqreturn_t stmmac_safety_interru
static irqreturn_t stmmac_msi_intr_tx(int irq, void *data)
{
struct stmmac_tx_queue *tx_q = (struct stmmac_tx_queue *)data;
@@ -1048,7 +1048,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
if (unlikely(!data)) {
netdev_err(priv->dev, "%s: invalid dev pointer\n", __func__);
-@@ -5797,10 +5799,12 @@ static irqreturn_t stmmac_msi_intr_tx(in
+@@ -5805,10 +5807,12 @@ static irqreturn_t stmmac_msi_intr_tx(in
static irqreturn_t stmmac_msi_intr_rx(int irq, void *data)
{
struct stmmac_rx_queue *rx_q = (struct stmmac_rx_queue *)data;
@@ -1062,7 +1062,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
if (unlikely(!data)) {
netdev_err(priv->dev, "%s: invalid dev pointer\n", __func__);
-@@ -5831,10 +5835,10 @@ static void stmmac_poll_controller(struc
+@@ -5839,10 +5843,10 @@ static void stmmac_poll_controller(struc
if (priv->plat->multi_msi_en) {
for (i = 0; i < priv->plat->rx_queues_to_use; i++)
@@ -1075,7 +1075,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
} else {
disable_irq(dev->irq);
stmmac_interrupt(dev->irq, dev);
-@@ -6015,34 +6019,34 @@ static int stmmac_rings_status_show(stru
+@@ -6023,34 +6027,34 @@ static int stmmac_rings_status_show(stru
return 0;
for (queue = 0; queue < rx_count; queue++) {
@@ -1116,7 +1116,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
}
}
-@@ -6383,7 +6387,7 @@ void stmmac_disable_rx_queue(struct stmm
+@@ -6391,7 +6395,7 @@ void stmmac_disable_rx_queue(struct stmm
void stmmac_enable_rx_queue(struct stmmac_priv *priv, u32 queue)
{
@@ -1125,7 +1125,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
struct stmmac_channel *ch = &priv->channel[queue];
unsigned long flags;
u32 buf_size;
-@@ -6420,7 +6424,7 @@ void stmmac_enable_rx_queue(struct stmma
+@@ -6428,7 +6432,7 @@ void stmmac_enable_rx_queue(struct stmma
rx_q->queue_index);
} else {
stmmac_set_dma_bfsize(priv, priv->ioaddr,
@@ -1134,7 +1134,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
rx_q->queue_index);
}
-@@ -6446,7 +6450,7 @@ void stmmac_disable_tx_queue(struct stmm
+@@ -6454,7 +6458,7 @@ void stmmac_disable_tx_queue(struct stmm
void stmmac_enable_tx_queue(struct stmmac_priv *priv, u32 queue)
{
@@ -1143,7 +1143,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
struct stmmac_channel *ch = &priv->channel[queue];
unsigned long flags;
int ret;
-@@ -6496,7 +6500,7 @@ void stmmac_xdp_release(struct net_devic
+@@ -6504,7 +6508,7 @@ void stmmac_xdp_release(struct net_devic
stmmac_disable_all_queues(priv);
for (chan = 0; chan < priv->plat->tx_queues_to_use; chan++)
@@ -1152,7 +1152,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
/* Free the IRQ lines */
stmmac_free_irq(dev, REQ_IRQ_ERR_ALL, 0);
-@@ -6555,7 +6559,7 @@ int stmmac_xdp_open(struct net_device *d
+@@ -6563,7 +6567,7 @@ int stmmac_xdp_open(struct net_device *d
/* DMA RX Channel Configuration */
for (chan = 0; chan < rx_cnt; chan++) {
@@ -1161,7 +1161,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
stmmac_init_rx_chan(priv, priv->ioaddr, priv->plat->dma_cfg,
rx_q->dma_rx_phy, chan);
-@@ -6573,7 +6577,7 @@ int stmmac_xdp_open(struct net_device *d
+@@ -6581,7 +6585,7 @@ int stmmac_xdp_open(struct net_device *d
rx_q->queue_index);
} else {
stmmac_set_dma_bfsize(priv, priv->ioaddr,
@@ -1170,7 +1170,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
rx_q->queue_index);
}
-@@ -6582,7 +6586,7 @@ int stmmac_xdp_open(struct net_device *d
+@@ -6590,7 +6594,7 @@ int stmmac_xdp_open(struct net_device *d
/* DMA TX Channel Configuration */
for (chan = 0; chan < tx_cnt; chan++) {
@@ -1179,7 +1179,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
stmmac_init_tx_chan(priv, priv->ioaddr, priv->plat->dma_cfg,
tx_q->dma_tx_phy, chan);
-@@ -6615,7 +6619,7 @@ int stmmac_xdp_open(struct net_device *d
+@@ -6623,7 +6627,7 @@ int stmmac_xdp_open(struct net_device *d
irq_error:
for (chan = 0; chan < priv->plat->tx_queues_to_use; chan++)
@@ -1188,7 +1188,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
stmmac_hw_teardown(dev);
init_error:
-@@ -6642,8 +6646,8 @@ int stmmac_xsk_wakeup(struct net_device
+@@ -6650,8 +6654,8 @@ int stmmac_xsk_wakeup(struct net_device
queue >= priv->plat->tx_queues_to_use)
return -EINVAL;
@@ -1199,7 +1199,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
ch = &priv->channel[queue];
if (!rx_q->xsk_pool && !tx_q->xsk_pool)
-@@ -6899,8 +6903,8 @@ int stmmac_reinit_ringparam(struct net_d
+@@ -6907,8 +6911,8 @@ int stmmac_reinit_ringparam(struct net_d
if (netif_running(dev))
stmmac_release(dev);
@@ -1210,7 +1210,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
if (netif_running(dev))
ret = stmmac_open(dev);
-@@ -7336,7 +7340,7 @@ int stmmac_suspend(struct device *dev)
+@@ -7344,7 +7348,7 @@ int stmmac_suspend(struct device *dev)
stmmac_disable_all_queues(priv);
for (chan = 0; chan < priv->plat->tx_queues_to_use; chan++)
@@ -1219,7 +1219,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
if (priv->eee_enabled) {
priv->tx_path_in_lpi_mode = false;
-@@ -7387,7 +7391,7 @@ EXPORT_SYMBOL_GPL(stmmac_suspend);
+@@ -7395,7 +7399,7 @@ EXPORT_SYMBOL_GPL(stmmac_suspend);
static void stmmac_reset_rx_queue(struct stmmac_priv *priv, u32 queue)
{
@@ -1228,7 +1228,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
rx_q->cur_rx = 0;
rx_q->dirty_rx = 0;
-@@ -7395,7 +7399,7 @@ static void stmmac_reset_rx_queue(struct
+@@ -7403,7 +7407,7 @@ static void stmmac_reset_rx_queue(struct
static void stmmac_reset_tx_queue(struct stmmac_priv *priv, u32 queue)
{
diff --git a/target/linux/generic/backport-5.15/775-v6.0-04-net-ethernet-stmicro-stmmac-generate-stmmac-dma-conf.patch b/target/linux/generic/backport-5.15/775-v6.0-04-net-ethernet-stmicro-stmmac-generate-stmmac-dma-conf.patch
index 4c41d3c743..12149edbe9 100644
--- a/target/linux/generic/backport-5.15/775-v6.0-04-net-ethernet-stmicro-stmmac-generate-stmmac-dma-conf.patch
+++ b/target/linux/generic/backport-5.15/775-v6.0-04-net-ethernet-stmicro-stmmac-generate-stmmac-dma-conf.patch
@@ -17,7 +17,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
--- a/drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
+++ b/drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
-@@ -1285,7 +1285,8 @@ static int stmmac_phy_setup(struct stmma
+@@ -1293,7 +1293,8 @@ static int stmmac_phy_setup(struct stmma
return 0;
}
@@ -27,7 +27,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
{
u32 rx_cnt = priv->plat->rx_queues_to_use;
unsigned int desc_size;
-@@ -1294,7 +1295,7 @@ static void stmmac_display_rx_rings(stru
+@@ -1302,7 +1303,7 @@ static void stmmac_display_rx_rings(stru
/* Display RX rings */
for (queue = 0; queue < rx_cnt; queue++) {
@@ -36,7 +36,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
pr_info("\tRX Queue %u rings\n", queue);
-@@ -1307,12 +1308,13 @@ static void stmmac_display_rx_rings(stru
+@@ -1315,12 +1316,13 @@ static void stmmac_display_rx_rings(stru
}
/* Display RX ring */
@@ -52,7 +52,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
{
u32 tx_cnt = priv->plat->tx_queues_to_use;
unsigned int desc_size;
-@@ -1321,7 +1323,7 @@ static void stmmac_display_tx_rings(stru
+@@ -1329,7 +1331,7 @@ static void stmmac_display_tx_rings(stru
/* Display TX rings */
for (queue = 0; queue < tx_cnt; queue++) {
@@ -61,7 +61,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
pr_info("\tTX Queue %d rings\n", queue);
-@@ -1336,18 +1338,19 @@ static void stmmac_display_tx_rings(stru
+@@ -1344,18 +1346,19 @@ static void stmmac_display_tx_rings(stru
desc_size = sizeof(struct dma_desc);
}
@@ -85,7 +85,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
}
static int stmmac_set_bfsize(int mtu, int bufsize)
-@@ -1371,44 +1374,50 @@ static int stmmac_set_bfsize(int mtu, in
+@@ -1379,44 +1382,50 @@ static int stmmac_set_bfsize(int mtu, in
/**
* stmmac_clear_rx_descriptors - clear RX descriptors
* @priv: driver private structure
@@ -147,7 +147,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
struct dma_desc *p;
if (priv->extend_desc)
-@@ -1425,10 +1434,12 @@ static void stmmac_clear_tx_descriptors(
+@@ -1433,10 +1442,12 @@ static void stmmac_clear_tx_descriptors(
/**
* stmmac_clear_descriptors - clear descriptors
* @priv: driver private structure
@@ -161,7 +161,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
{
u32 rx_queue_cnt = priv->plat->rx_queues_to_use;
u32 tx_queue_cnt = priv->plat->tx_queues_to_use;
-@@ -1436,16 +1447,17 @@ static void stmmac_clear_descriptors(str
+@@ -1444,16 +1455,17 @@ static void stmmac_clear_descriptors(str
/* Clear the RX descriptors */
for (queue = 0; queue < rx_queue_cnt; queue++)
@@ -181,7 +181,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
* @p: descriptor pointer
* @i: descriptor index
* @flags: gfp flag
-@@ -1453,10 +1465,12 @@ static void stmmac_clear_descriptors(str
+@@ -1461,10 +1473,12 @@ static void stmmac_clear_descriptors(str
* Description: this function is called to allocate a receive buffer, perform
* the DMA mapping and init the descriptor.
*/
@@ -196,7 +196,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
struct stmmac_rx_buffer *buf = &rx_q->buf_pool[i];
if (!buf->page) {
-@@ -1481,7 +1495,7 @@ static int stmmac_init_rx_buffers(struct
+@@ -1489,7 +1503,7 @@ static int stmmac_init_rx_buffers(struct
buf->addr = page_pool_get_dma_addr(buf->page) + buf->page_offset;
stmmac_set_desc_addr(priv, p, buf->addr);
@@ -205,7 +205,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
stmmac_init_desc3(priv, p);
return 0;
-@@ -1490,12 +1504,13 @@ static int stmmac_init_rx_buffers(struct
+@@ -1498,12 +1512,13 @@ static int stmmac_init_rx_buffers(struct
/**
* stmmac_free_rx_buffer - free RX dma buffers
* @priv: private structure
@@ -222,7 +222,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
struct stmmac_rx_buffer *buf = &rx_q->buf_pool[i];
if (buf->page)
-@@ -1510,12 +1525,15 @@ static void stmmac_free_rx_buffer(struct
+@@ -1518,12 +1533,15 @@ static void stmmac_free_rx_buffer(struct
/**
* stmmac_free_tx_buffer - free RX dma buffers
* @priv: private structure
@@ -240,7 +240,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
if (tx_q->tx_skbuff_dma[i].buf &&
tx_q->tx_skbuff_dma[i].buf_type != STMMAC_TXBUF_T_XDP_TX) {
-@@ -1554,23 +1572,28 @@ static void stmmac_free_tx_buffer(struct
+@@ -1562,23 +1580,28 @@ static void stmmac_free_tx_buffer(struct
/**
* dma_free_rx_skbufs - free RX dma buffers
* @priv: private structure
@@ -276,7 +276,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
struct dma_desc *p;
int ret;
-@@ -1579,7 +1602,7 @@ static int stmmac_alloc_rx_buffers(struc
+@@ -1587,7 +1610,7 @@ static int stmmac_alloc_rx_buffers(struc
else
p = rx_q->dma_rx + i;
@@ -285,7 +285,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
queue);
if (ret)
return ret;
-@@ -1593,14 +1616,17 @@ static int stmmac_alloc_rx_buffers(struc
+@@ -1601,14 +1624,17 @@ static int stmmac_alloc_rx_buffers(struc
/**
* dma_free_rx_xskbufs - free RX dma buffers from XSK pool
* @priv: private structure
@@ -306,7 +306,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
struct stmmac_rx_buffer *buf = &rx_q->buf_pool[i];
if (!buf->xdp)
-@@ -1611,12 +1637,14 @@ static void dma_free_rx_xskbufs(struct s
+@@ -1619,12 +1645,14 @@ static void dma_free_rx_xskbufs(struct s
}
}
@@ -324,7 +324,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
struct stmmac_rx_buffer *buf;
dma_addr_t dma_addr;
struct dma_desc *p;
-@@ -1651,22 +1679,25 @@ static struct xsk_buff_pool *stmmac_get_
+@@ -1659,22 +1687,25 @@ static struct xsk_buff_pool *stmmac_get_
/**
* __init_dma_rx_desc_rings - init the RX descriptor ring (per queue)
* @priv: driver private structure
@@ -353,7 +353,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
xdp_rxq_info_unreg_mem_model(&rx_q->xdp_rxq);
-@@ -1693,9 +1724,9 @@ static int __init_dma_rx_desc_rings(stru
+@@ -1701,9 +1732,9 @@ static int __init_dma_rx_desc_rings(stru
/* RX XDP ZC buffer pool may not be populated, e.g.
* xdpsock TX-only.
*/
@@ -365,7 +365,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
if (ret < 0)
return -ENOMEM;
}
-@@ -1705,17 +1736,19 @@ static int __init_dma_rx_desc_rings(stru
+@@ -1713,17 +1744,19 @@ static int __init_dma_rx_desc_rings(stru
if (priv->extend_desc)
stmmac_mode_init(priv, rx_q->dma_erx,
rx_q->dma_rx_phy,
@@ -388,7 +388,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
{
struct stmmac_priv *priv = netdev_priv(dev);
u32 rx_count = priv->plat->rx_queues_to_use;
-@@ -1727,7 +1760,7 @@ static int init_dma_rx_desc_rings(struct
+@@ -1735,7 +1768,7 @@ static int init_dma_rx_desc_rings(struct
"SKB addresses:\nskb\t\tskb data\tdma data\n");
for (queue = 0; queue < rx_count; queue++) {
@@ -397,7 +397,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
if (ret)
goto err_init_rx_buffers;
}
-@@ -1736,12 +1769,12 @@ static int init_dma_rx_desc_rings(struct
+@@ -1744,12 +1777,12 @@ static int init_dma_rx_desc_rings(struct
err_init_rx_buffers:
while (queue >= 0) {
@@ -413,7 +413,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
rx_q->buf_alloc_num = 0;
rx_q->xsk_pool = NULL;
-@@ -1758,14 +1791,17 @@ err_init_rx_buffers:
+@@ -1766,14 +1799,17 @@ err_init_rx_buffers:
/**
* __init_dma_tx_desc_rings - init the TX descriptor ring (per queue)
* @priv: driver private structure
@@ -434,7 +434,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
int i;
netif_dbg(priv, probe, priv->dev,
-@@ -1777,16 +1813,16 @@ static int __init_dma_tx_desc_rings(stru
+@@ -1785,16 +1821,16 @@ static int __init_dma_tx_desc_rings(stru
if (priv->extend_desc)
stmmac_mode_init(priv, tx_q->dma_etx,
tx_q->dma_tx_phy,
@@ -454,7 +454,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
struct dma_desc *p;
if (priv->extend_desc)
-@@ -1808,7 +1844,8 @@ static int __init_dma_tx_desc_rings(stru
+@@ -1816,7 +1852,8 @@ static int __init_dma_tx_desc_rings(stru
return 0;
}
@@ -464,7 +464,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
{
struct stmmac_priv *priv = netdev_priv(dev);
u32 tx_queue_cnt;
-@@ -1817,7 +1854,7 @@ static int init_dma_tx_desc_rings(struct
+@@ -1825,7 +1862,7 @@ static int init_dma_tx_desc_rings(struct
tx_queue_cnt = priv->plat->tx_queues_to_use;
for (queue = 0; queue < tx_queue_cnt; queue++)
@@ -473,7 +473,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
return 0;
}
-@@ -1825,26 +1862,29 @@ static int init_dma_tx_desc_rings(struct
+@@ -1833,26 +1870,29 @@ static int init_dma_tx_desc_rings(struct
/**
* init_dma_desc_rings - init the RX/TX descriptor rings
* @dev: net device structure
@@ -508,7 +508,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
return ret;
}
-@@ -1852,17 +1892,20 @@ static int init_dma_desc_rings(struct ne
+@@ -1860,17 +1900,20 @@ static int init_dma_desc_rings(struct ne
/**
* dma_free_tx_skbufs - free TX dma buffers
* @priv: private structure
@@ -533,7 +533,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
if (tx_q->xsk_pool && tx_q->xsk_frames_done) {
xsk_tx_completed(tx_q->xsk_pool, tx_q->xsk_frames_done);
-@@ -1881,34 +1924,37 @@ static void stmmac_free_tx_skbufs(struct
+@@ -1889,34 +1932,37 @@ static void stmmac_free_tx_skbufs(struct
u32 queue;
for (queue = 0; queue < tx_queue_cnt; queue++)
@@ -578,7 +578,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
sizeof(struct dma_extended_desc),
rx_q->dma_erx, rx_q->dma_rx_phy);
-@@ -1920,29 +1966,33 @@ static void __free_dma_rx_desc_resources
+@@ -1928,29 +1974,33 @@ static void __free_dma_rx_desc_resources
page_pool_destroy(rx_q->page_pool);
}
@@ -617,7 +617,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
if (priv->extend_desc) {
size = sizeof(struct dma_extended_desc);
-@@ -1955,7 +2005,7 @@ static void __free_dma_tx_desc_resources
+@@ -1963,7 +2013,7 @@ static void __free_dma_tx_desc_resources
addr = tx_q->dma_tx;
}
@@ -626,7 +626,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
dma_free_coherent(priv->device, size, addr, tx_q->dma_tx_phy);
-@@ -1963,28 +2013,32 @@ static void __free_dma_tx_desc_resources
+@@ -1971,28 +2021,32 @@ static void __free_dma_tx_desc_resources
kfree(tx_q->tx_skbuff);
}
@@ -663,7 +663,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
struct stmmac_channel *ch = &priv->channel[queue];
bool xdp_prog = stmmac_xdp_is_enabled(priv);
struct page_pool_params pp_params = { 0 };
-@@ -1996,8 +2050,8 @@ static int __alloc_dma_rx_desc_resources
+@@ -2004,8 +2058,8 @@ static int __alloc_dma_rx_desc_resources
rx_q->priv_data = priv;
pp_params.flags = PP_FLAG_DMA_MAP | PP_FLAG_DMA_SYNC_DEV;
@@ -674,7 +674,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
pp_params.order = ilog2(num_pages);
pp_params.nid = dev_to_node(priv->device);
pp_params.dev = priv->device;
-@@ -2012,7 +2066,7 @@ static int __alloc_dma_rx_desc_resources
+@@ -2020,7 +2074,7 @@ static int __alloc_dma_rx_desc_resources
return ret;
}
@@ -683,7 +683,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
sizeof(*rx_q->buf_pool),
GFP_KERNEL);
if (!rx_q->buf_pool)
-@@ -2020,7 +2074,7 @@ static int __alloc_dma_rx_desc_resources
+@@ -2028,7 +2082,7 @@ static int __alloc_dma_rx_desc_resources
if (priv->extend_desc) {
rx_q->dma_erx = dma_alloc_coherent(priv->device,
@@ -692,7 +692,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
sizeof(struct dma_extended_desc),
&rx_q->dma_rx_phy,
GFP_KERNEL);
-@@ -2029,7 +2083,7 @@ static int __alloc_dma_rx_desc_resources
+@@ -2037,7 +2091,7 @@ static int __alloc_dma_rx_desc_resources
} else {
rx_q->dma_rx = dma_alloc_coherent(priv->device,
@@ -701,7 +701,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
sizeof(struct dma_desc),
&rx_q->dma_rx_phy,
GFP_KERNEL);
-@@ -2054,7 +2108,8 @@ static int __alloc_dma_rx_desc_resources
+@@ -2062,7 +2116,8 @@ static int __alloc_dma_rx_desc_resources
return 0;
}
@@ -711,7 +711,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
{
u32 rx_count = priv->plat->rx_queues_to_use;
u32 queue;
-@@ -2062,7 +2117,7 @@ static int alloc_dma_rx_desc_resources(s
+@@ -2070,7 +2125,7 @@ static int alloc_dma_rx_desc_resources(s
/* RX queues buffers and DMA */
for (queue = 0; queue < rx_count; queue++) {
@@ -720,7 +720,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
if (ret)
goto err_dma;
}
-@@ -2070,7 +2125,7 @@ static int alloc_dma_rx_desc_resources(s
+@@ -2078,7 +2133,7 @@ static int alloc_dma_rx_desc_resources(s
return 0;
err_dma:
@@ -729,7 +729,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
return ret;
}
-@@ -2078,28 +2133,31 @@ err_dma:
+@@ -2086,28 +2141,31 @@ err_dma:
/**
* __alloc_dma_tx_desc_resources - alloc TX resources (per queue).
* @priv: private structure
@@ -765,7 +765,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
sizeof(struct sk_buff *),
GFP_KERNEL);
if (!tx_q->tx_skbuff)
-@@ -2112,7 +2170,7 @@ static int __alloc_dma_tx_desc_resources
+@@ -2120,7 +2178,7 @@ static int __alloc_dma_tx_desc_resources
else
size = sizeof(struct dma_desc);
@@ -774,7 +774,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
addr = dma_alloc_coherent(priv->device, size,
&tx_q->dma_tx_phy, GFP_KERNEL);
-@@ -2129,7 +2187,8 @@ static int __alloc_dma_tx_desc_resources
+@@ -2137,7 +2195,8 @@ static int __alloc_dma_tx_desc_resources
return 0;
}
@@ -784,7 +784,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
{
u32 tx_count = priv->plat->tx_queues_to_use;
u32 queue;
-@@ -2137,7 +2196,7 @@ static int alloc_dma_tx_desc_resources(s
+@@ -2145,7 +2204,7 @@ static int alloc_dma_tx_desc_resources(s
/* TX queues buffers and DMA */
for (queue = 0; queue < tx_count; queue++) {
@@ -793,7 +793,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
if (ret)
goto err_dma;
}
-@@ -2145,27 +2204,29 @@ static int alloc_dma_tx_desc_resources(s
+@@ -2153,27 +2212,29 @@ static int alloc_dma_tx_desc_resources(s
return 0;
err_dma:
@@ -827,7 +827,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
return ret;
}
-@@ -2173,16 +2234,18 @@ static int alloc_dma_desc_resources(stru
+@@ -2181,16 +2242,18 @@ static int alloc_dma_desc_resources(stru
/**
* free_dma_desc_resources - free dma desc resources
* @priv: private structure
@@ -849,7 +849,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
}
/**
-@@ -2671,8 +2734,8 @@ static void stmmac_tx_err(struct stmmac_
+@@ -2679,8 +2742,8 @@ static void stmmac_tx_err(struct stmmac_
netif_tx_stop_queue(netdev_get_tx_queue(priv->dev, chan));
stmmac_stop_tx_dma(priv, chan);
@@ -860,7 +860,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
stmmac_reset_tx_queue(priv, chan);
stmmac_init_tx_chan(priv, priv->ioaddr, priv->plat->dma_cfg,
tx_q->dma_tx_phy, chan);
-@@ -3669,19 +3732,93 @@ static int stmmac_request_irq(struct net
+@@ -3677,19 +3740,93 @@ static int stmmac_request_irq(struct net
}
/**
@@ -957,7 +957,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
u32 chan;
int ret;
-@@ -3708,45 +3845,10 @@ static int stmmac_open(struct net_device
+@@ -3716,45 +3853,10 @@ static int stmmac_open(struct net_device
memset(&priv->xstats, 0, sizeof(struct stmmac_extra_stats));
priv->xstats.threshold = tc;
@@ -1005,7 +1005,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
if (priv->plat->serdes_powerup) {
ret = priv->plat->serdes_powerup(dev, priv->plat->bsp_priv);
-@@ -3789,14 +3891,28 @@ irq_error:
+@@ -3797,14 +3899,28 @@ irq_error:
stmmac_hw_teardown(dev);
init_error:
@@ -1036,7 +1036,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
static void stmmac_fpe_stop_wq(struct stmmac_priv *priv)
{
set_bit(__FPE_REMOVING, &priv->fpe_task_state);
-@@ -3843,7 +3959,7 @@ static int stmmac_release(struct net_dev
+@@ -3851,7 +3967,7 @@ static int stmmac_release(struct net_dev
stmmac_stop_all_dma(priv);
/* Release and free the Rx/Tx resources */
@@ -1045,7 +1045,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
/* Disable the MAC Rx/Tx */
stmmac_mac_set(priv, priv->ioaddr, false);
-@@ -6382,7 +6498,7 @@ void stmmac_disable_rx_queue(struct stmm
+@@ -6390,7 +6506,7 @@ void stmmac_disable_rx_queue(struct stmm
spin_unlock_irqrestore(&ch->lock, flags);
stmmac_stop_rx_dma(priv, queue);
@@ -1054,7 +1054,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
}
void stmmac_enable_rx_queue(struct stmmac_priv *priv, u32 queue)
-@@ -6393,21 +6509,21 @@ void stmmac_enable_rx_queue(struct stmma
+@@ -6401,21 +6517,21 @@ void stmmac_enable_rx_queue(struct stmma
u32 buf_size;
int ret;
@@ -1080,7 +1080,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
stmmac_init_rx_chan(priv, priv->ioaddr, priv->plat->dma_cfg,
rx_q->dma_rx_phy, rx_q->queue_index);
-@@ -6445,7 +6561,7 @@ void stmmac_disable_tx_queue(struct stmm
+@@ -6453,7 +6569,7 @@ void stmmac_disable_tx_queue(struct stmm
spin_unlock_irqrestore(&ch->lock, flags);
stmmac_stop_tx_dma(priv, queue);
@@ -1089,7 +1089,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
}
void stmmac_enable_tx_queue(struct stmmac_priv *priv, u32 queue)
-@@ -6455,21 +6571,21 @@ void stmmac_enable_tx_queue(struct stmma
+@@ -6463,21 +6579,21 @@ void stmmac_enable_tx_queue(struct stmma
unsigned long flags;
int ret;
@@ -1115,7 +1115,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
stmmac_init_tx_chan(priv, priv->ioaddr, priv->plat->dma_cfg,
tx_q->dma_tx_phy, tx_q->queue_index);
-@@ -6509,7 +6625,7 @@ void stmmac_xdp_release(struct net_devic
+@@ -6517,7 +6633,7 @@ void stmmac_xdp_release(struct net_devic
stmmac_stop_all_dma(priv);
/* Release and free the Rx/Tx resources */
@@ -1124,7 +1124,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
/* Disable the MAC Rx/Tx */
stmmac_mac_set(priv, priv->ioaddr, false);
-@@ -6534,14 +6650,14 @@ int stmmac_xdp_open(struct net_device *d
+@@ -6542,14 +6658,14 @@ int stmmac_xdp_open(struct net_device *d
u32 chan;
int ret;
@@ -1141,7 +1141,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
if (ret < 0) {
netdev_err(dev, "%s: DMA descriptors initialization failed\n",
__func__);
-@@ -6623,7 +6739,7 @@ irq_error:
+@@ -6631,7 +6747,7 @@ irq_error:
stmmac_hw_teardown(dev);
init_error:
@@ -1150,7 +1150,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
dma_desc_error:
return ret;
}
-@@ -7482,7 +7598,7 @@ int stmmac_resume(struct device *dev)
+@@ -7490,7 +7606,7 @@ int stmmac_resume(struct device *dev)
stmmac_reset_queues_param(priv);
stmmac_free_tx_skbufs(priv);
diff --git a/target/linux/generic/backport-5.15/775-v6.0-05-net-ethernet-stmicro-stmmac-permit-MTU-change-with-i.patch b/target/linux/generic/backport-5.15/775-v6.0-05-net-ethernet-stmicro-stmmac-permit-MTU-change-with-i.patch
index 996c28119e..d1d8c622fc 100644
--- a/target/linux/generic/backport-5.15/775-v6.0-05-net-ethernet-stmicro-stmmac-permit-MTU-change-with-i.patch
+++ b/target/linux/generic/backport-5.15/775-v6.0-05-net-ethernet-stmicro-stmmac-permit-MTU-change-with-i.patch
@@ -19,7 +19,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
--- a/drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
+++ b/drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
-@@ -5611,18 +5611,15 @@ static int stmmac_change_mtu(struct net_
+@@ -5619,18 +5619,15 @@ static int stmmac_change_mtu(struct net_
{
struct stmmac_priv *priv = netdev_priv(dev);
int txfifosz = priv->plat->tx_fifo_size;
@@ -40,7 +40,7 @@ Signed-off-by: Jakub Kicinski <kuba at kernel.org>
if (stmmac_xdp_is_enabled(priv) && new_mtu > ETH_DATA_LEN) {
netdev_dbg(priv->dev, "Jumbo frames not supported for XDP\n");
return -EINVAL;
-@@ -5634,8 +5631,29 @@ static int stmmac_change_mtu(struct net_
+@@ -5642,8 +5639,29 @@ static int stmmac_change_mtu(struct net_
if ((txfifosz < new_mtu) || (new_mtu > BUF_SIZE_16KiB))
return -EINVAL;
diff --git a/target/linux/generic/pending-5.15/630-packet_socket_type.patch b/target/linux/generic/pending-5.15/630-packet_socket_type.patch
index c61935ffb0..c305aca6e0 100644
--- a/target/linux/generic/pending-5.15/630-packet_socket_type.patch
+++ b/target/linux/generic/pending-5.15/630-packet_socket_type.patch
@@ -87,7 +87,7 @@ Signed-off-by: Felix Fietkau <nbd at nbd.name>
if (!net_eq(dev_net(dev), sock_net(sk)))
goto drop;
-@@ -3330,6 +3332,7 @@ static int packet_create(struct net *net
+@@ -3329,6 +3331,7 @@ static int packet_create(struct net *net
mutex_init(&po->pg_vec_lock);
po->rollover = NULL;
po->prot_hook.func = packet_rcv;
@@ -95,7 +95,7 @@ Signed-off-by: Felix Fietkau <nbd at nbd.name>
if (sock->type == SOCK_PACKET)
po->prot_hook.func = packet_rcv_spkt;
-@@ -3971,6 +3974,16 @@ packet_setsockopt(struct socket *sock, i
+@@ -3969,6 +3972,16 @@ packet_setsockopt(struct socket *sock, i
po->xmit = val ? packet_direct_xmit : dev_queue_xmit;
return 0;
}
@@ -112,7 +112,7 @@ Signed-off-by: Felix Fietkau <nbd at nbd.name>
default:
return -ENOPROTOOPT;
}
-@@ -4027,6 +4040,13 @@ static int packet_getsockopt(struct sock
+@@ -4025,6 +4038,13 @@ static int packet_getsockopt(struct sock
case PACKET_VNET_HDR:
val = po->has_vnet_hdr;
break;
diff --git a/target/linux/generic/pending-5.15/703-phy-add-detach-callback-to-struct-phy_driver.patch b/target/linux/generic/pending-5.15/703-phy-add-detach-callback-to-struct-phy_driver.patch
index 976239c128..5baccf73cb 100644
--- a/target/linux/generic/pending-5.15/703-phy-add-detach-callback-to-struct-phy_driver.patch
+++ b/target/linux/generic/pending-5.15/703-phy-add-detach-callback-to-struct-phy_driver.patch
@@ -11,7 +11,7 @@ Signed-off-by: Gabor Juhos <juhosg at openwrt.org>
--- a/drivers/net/phy/phy_device.c
+++ b/drivers/net/phy/phy_device.c
-@@ -1746,6 +1746,9 @@ void phy_detach(struct phy_device *phyde
+@@ -1748,6 +1748,9 @@ void phy_detach(struct phy_device *phyde
struct module *ndev_owner = NULL;
struct mii_bus *bus;
diff --git a/target/linux/ramips/patches-5.15/720-Revert-net-phy-simplify-phy_link_change-arguments.patch b/target/linux/ramips/patches-5.15/720-Revert-net-phy-simplify-phy_link_change-arguments.patch
index 192f5b48da..d53a8c2d79 100644
--- a/target/linux/ramips/patches-5.15/720-Revert-net-phy-simplify-phy_link_change-arguments.patch
+++ b/target/linux/ramips/patches-5.15/720-Revert-net-phy-simplify-phy_link_change-arguments.patch
@@ -71,7 +71,7 @@ still required by target/linux/ramips/files/drivers/net/ethernet/ralink/mdio.c
break;
--- a/drivers/net/phy/phy_device.c
+++ b/drivers/net/phy/phy_device.c
-@@ -1031,14 +1031,16 @@ struct phy_device *phy_find_first(struct
+@@ -1032,14 +1032,16 @@ struct phy_device *phy_find_first(struct
}
EXPORT_SYMBOL(phy_find_first);
More information about the lede-commits
mailing list