diff options
Diffstat (limited to 'target/linux/generic/backport-5.15/775-v6.0-03-net-ethernet-stmicro-stmmac-move-dma-conf-to-dedicat.patch')
-rw-r--r-- | target/linux/generic/backport-5.15/775-v6.0-03-net-ethernet-stmicro-stmmac-move-dma-conf-to-dedicat.patch | 92 |
1 files changed, 46 insertions, 46 deletions
diff --git a/target/linux/generic/backport-5.15/775-v6.0-03-net-ethernet-stmicro-stmmac-move-dma-conf-to-dedicat.patch b/target/linux/generic/backport-5.15/775-v6.0-03-net-ethernet-stmicro-stmmac-move-dma-conf-to-dedicat.patch index 980d821cf5..ec3607feeb 100644 --- a/target/linux/generic/backport-5.15/775-v6.0-03-net-ethernet-stmicro-stmmac-move-dma-conf-to-dedicat.patch +++ b/target/linux/generic/backport-5.15/775-v6.0-03-net-ethernet-stmicro-stmmac-move-dma-conf-to-dedicat.patch @@ -766,7 +766,7 @@ Signed-off-by: Jakub Kicinski <kuba@kernel.org> int tbs_en = priv->plat->tx_queues_cfg[chan].tbs_en; /* Setup per-TXQ tbs flag before TX descriptor alloc */ -@@ -3776,7 +3776,7 @@ irq_error: +@@ -3785,7 +3785,7 @@ irq_error: phylink_stop(priv->phylink); for (chan = 0; chan < priv->plat->tx_queues_to_use; chan++) @@ -775,7 +775,7 @@ Signed-off-by: Jakub Kicinski <kuba@kernel.org> stmmac_hw_teardown(dev); init_error: -@@ -3818,7 +3818,7 @@ static int stmmac_release(struct net_dev +@@ -3827,7 +3827,7 @@ static int stmmac_release(struct net_dev stmmac_disable_all_queues(priv); for (chan = 0; chan < priv->plat->tx_queues_to_use; chan++) @@ -784,7 +784,7 @@ Signed-off-by: Jakub Kicinski <kuba@kernel.org> netif_tx_disable(dev); -@@ -3878,7 +3878,7 @@ static bool stmmac_vlan_insert(struct st +@@ -3891,7 +3891,7 @@ static bool stmmac_vlan_insert(struct st return false; stmmac_set_tx_owner(priv, p); @@ -793,7 +793,7 @@ Signed-off-by: Jakub Kicinski <kuba@kernel.org> return true; } -@@ -3896,7 +3896,7 @@ static bool stmmac_vlan_insert(struct st +@@ -3909,7 +3909,7 @@ static bool stmmac_vlan_insert(struct st static void stmmac_tso_allocator(struct stmmac_priv *priv, dma_addr_t des, int total_len, bool last_segment, u32 queue) { @@ -802,7 +802,7 @@ Signed-off-by: Jakub Kicinski <kuba@kernel.org> struct dma_desc *desc; u32 buff_size; int tmp_len; -@@ -3907,7 +3907,7 @@ static void stmmac_tso_allocator(struct +@@ -3920,7 +3920,7 @@ static void stmmac_tso_allocator(struct dma_addr_t curr_addr; tx_q->cur_tx = STMMAC_GET_ENTRY(tx_q->cur_tx, @@ -811,7 +811,7 @@ Signed-off-by: Jakub Kicinski <kuba@kernel.org> WARN_ON(tx_q->tx_skbuff[tx_q->cur_tx]); if (tx_q->tbs & STMMAC_TBS_AVAIL) -@@ -3935,7 +3935,7 @@ static void stmmac_tso_allocator(struct +@@ -3948,7 +3948,7 @@ static void stmmac_tso_allocator(struct static void stmmac_flush_tx_descriptors(struct stmmac_priv *priv, int queue) { @@ -820,7 +820,7 @@ Signed-off-by: Jakub Kicinski <kuba@kernel.org> int desc_size; if (likely(priv->extend_desc)) -@@ -3997,7 +3997,7 @@ static netdev_tx_t stmmac_tso_xmit(struc +@@ -4010,7 +4010,7 @@ static netdev_tx_t stmmac_tso_xmit(struc dma_addr_t des; int i; @@ -829,7 +829,7 @@ Signed-off-by: Jakub Kicinski <kuba@kernel.org> first_tx = tx_q->cur_tx; /* Compute header lengths */ -@@ -4037,7 +4037,7 @@ static netdev_tx_t stmmac_tso_xmit(struc +@@ -4050,7 +4050,7 @@ static netdev_tx_t stmmac_tso_xmit(struc stmmac_set_mss(priv, mss_desc, mss); tx_q->mss = mss; tx_q->cur_tx = STMMAC_GET_ENTRY(tx_q->cur_tx, @@ -838,7 +838,7 @@ Signed-off-by: Jakub Kicinski <kuba@kernel.org> WARN_ON(tx_q->tx_skbuff[tx_q->cur_tx]); } -@@ -4149,7 +4149,7 @@ static netdev_tx_t stmmac_tso_xmit(struc +@@ -4162,7 +4162,7 @@ static netdev_tx_t stmmac_tso_xmit(struc * ndo_start_xmit will fill this descriptor the next time it's * called and stmmac_tx_clean may clean up to this descriptor. */ @@ -847,7 +847,7 @@ Signed-off-by: Jakub Kicinski <kuba@kernel.org> if (unlikely(stmmac_tx_avail(priv, queue) <= (MAX_SKB_FRAGS + 1))) { netif_dbg(priv, hw, priv->dev, "%s: stop transmitted packets\n", -@@ -4237,7 +4237,7 @@ static netdev_tx_t stmmac_xmit(struct sk +@@ -4250,7 +4250,7 @@ static netdev_tx_t stmmac_xmit(struct sk int entry, first_tx; dma_addr_t des; @@ -856,7 +856,7 @@ Signed-off-by: Jakub Kicinski <kuba@kernel.org> first_tx = tx_q->cur_tx; if (priv->tx_path_in_lpi_mode && priv->eee_sw_timer_en) -@@ -4300,7 +4300,7 @@ static netdev_tx_t stmmac_xmit(struct sk +@@ -4313,7 +4313,7 @@ static netdev_tx_t stmmac_xmit(struct sk int len = skb_frag_size(frag); bool last_segment = (i == (nfrags - 1)); @@ -865,7 +865,7 @@ Signed-off-by: Jakub Kicinski <kuba@kernel.org> WARN_ON(tx_q->tx_skbuff[entry]); if (likely(priv->extend_desc)) -@@ -4371,7 +4371,7 @@ static netdev_tx_t stmmac_xmit(struct sk +@@ -4384,7 +4384,7 @@ static netdev_tx_t stmmac_xmit(struct sk * ndo_start_xmit will fill this descriptor the next time it's * called and stmmac_tx_clean may clean up to this descriptor. */ @@ -874,7 +874,7 @@ Signed-off-by: Jakub Kicinski <kuba@kernel.org> tx_q->cur_tx = entry; if (netif_msg_pktdata(priv)) { -@@ -4486,7 +4486,7 @@ static void stmmac_rx_vlan(struct net_de +@@ -4499,7 +4499,7 @@ static void stmmac_rx_vlan(struct net_de */ static inline void stmmac_rx_refill(struct stmmac_priv *priv, u32 queue) { @@ -883,7 +883,7 @@ Signed-off-by: Jakub Kicinski <kuba@kernel.org> int dirty = stmmac_rx_dirty(priv, queue); unsigned int entry = rx_q->dirty_rx; -@@ -4536,7 +4536,7 @@ static inline void stmmac_rx_refill(stru +@@ -4549,7 +4549,7 @@ static inline void stmmac_rx_refill(stru dma_wmb(); stmmac_set_rx_owner(priv, p, use_rx_wd); @@ -892,7 +892,7 @@ Signed-off-by: Jakub Kicinski <kuba@kernel.org> } rx_q->dirty_rx = entry; rx_q->rx_tail_addr = rx_q->dma_rx_phy + -@@ -4564,12 +4564,12 @@ static unsigned int stmmac_rx_buf1_len(s +@@ -4577,12 +4577,12 @@ static unsigned int stmmac_rx_buf1_len(s /* First descriptor, not last descriptor and not split header */ if (status & rx_not_ls) @@ -907,7 +907,7 @@ Signed-off-by: Jakub Kicinski <kuba@kernel.org> } static unsigned int stmmac_rx_buf2_len(struct stmmac_priv *priv, -@@ -4585,7 +4585,7 @@ static unsigned int stmmac_rx_buf2_len(s +@@ -4598,7 +4598,7 @@ static unsigned int stmmac_rx_buf2_len(s /* Not last descriptor */ if (status & rx_not_ls) @@ -916,7 +916,7 @@ Signed-off-by: Jakub Kicinski <kuba@kernel.org> plen = stmmac_get_rx_frame_len(priv, p, coe); -@@ -4596,7 +4596,7 @@ static unsigned int stmmac_rx_buf2_len(s +@@ -4609,7 +4609,7 @@ static unsigned int stmmac_rx_buf2_len(s static int stmmac_xdp_xmit_xdpf(struct stmmac_priv *priv, int queue, struct xdp_frame *xdpf, bool dma_map) { @@ -925,7 +925,7 @@ Signed-off-by: Jakub Kicinski <kuba@kernel.org> unsigned int entry = tx_q->cur_tx; struct dma_desc *tx_desc; dma_addr_t dma_addr; -@@ -4659,7 +4659,7 @@ static int stmmac_xdp_xmit_xdpf(struct s +@@ -4672,7 +4672,7 @@ static int stmmac_xdp_xmit_xdpf(struct s stmmac_enable_dma_transmission(priv, priv->ioaddr); @@ -934,7 +934,7 @@ Signed-off-by: Jakub Kicinski <kuba@kernel.org> tx_q->cur_tx = entry; return STMMAC_XDP_TX; -@@ -4833,7 +4833,7 @@ static void stmmac_dispatch_skb_zc(struc +@@ -4846,7 +4846,7 @@ static void stmmac_dispatch_skb_zc(struc static bool stmmac_rx_refill_zc(struct stmmac_priv *priv, u32 queue, u32 budget) { @@ -943,7 +943,7 @@ Signed-off-by: Jakub Kicinski <kuba@kernel.org> unsigned int entry = rx_q->dirty_rx; struct dma_desc *rx_desc = NULL; bool ret = true; -@@ -4876,7 +4876,7 @@ static bool stmmac_rx_refill_zc(struct s +@@ -4889,7 +4889,7 @@ static bool stmmac_rx_refill_zc(struct s dma_wmb(); stmmac_set_rx_owner(priv, rx_desc, use_rx_wd); @@ -952,7 +952,7 @@ Signed-off-by: Jakub Kicinski <kuba@kernel.org> } if (rx_desc) { -@@ -4891,7 +4891,7 @@ static bool stmmac_rx_refill_zc(struct s +@@ -4904,7 +4904,7 @@ static bool stmmac_rx_refill_zc(struct s static int stmmac_rx_zc(struct stmmac_priv *priv, int limit, u32 queue) { @@ -961,7 +961,7 @@ Signed-off-by: Jakub Kicinski <kuba@kernel.org> unsigned int count = 0, error = 0, len = 0; int dirty = stmmac_rx_dirty(priv, queue); unsigned int next_entry = rx_q->cur_rx; -@@ -4913,7 +4913,7 @@ static int stmmac_rx_zc(struct stmmac_pr +@@ -4926,7 +4926,7 @@ static int stmmac_rx_zc(struct stmmac_pr desc_size = sizeof(struct dma_desc); } @@ -970,7 +970,7 @@ Signed-off-by: Jakub Kicinski <kuba@kernel.org> rx_q->dma_rx_phy, desc_size); } while (count < limit) { -@@ -4960,7 +4960,7 @@ read_again: +@@ -4973,7 +4973,7 @@ read_again: /* Prefetch the next RX descriptor */ rx_q->cur_rx = STMMAC_GET_ENTRY(rx_q->cur_rx, @@ -979,7 +979,7 @@ Signed-off-by: Jakub Kicinski <kuba@kernel.org> next_entry = rx_q->cur_rx; if (priv->extend_desc) -@@ -5081,7 +5081,7 @@ read_again: +@@ -5094,7 +5094,7 @@ read_again: */ static int stmmac_rx(struct stmmac_priv *priv, int limit, u32 queue) { @@ -988,7 +988,7 @@ Signed-off-by: Jakub Kicinski <kuba@kernel.org> struct stmmac_channel *ch = &priv->channel[queue]; unsigned int count = 0, error = 0, len = 0; int status = 0, coe = priv->hw->rx_csum; -@@ -5094,7 +5094,7 @@ static int stmmac_rx(struct stmmac_priv +@@ -5107,7 +5107,7 @@ static int stmmac_rx(struct stmmac_priv int buf_sz; dma_dir = page_pool_get_dma_dir(rx_q->page_pool); @@ -997,7 +997,7 @@ Signed-off-by: Jakub Kicinski <kuba@kernel.org> if (netif_msg_rx_status(priv)) { void *rx_head; -@@ -5108,7 +5108,7 @@ static int stmmac_rx(struct stmmac_priv +@@ -5121,7 +5121,7 @@ static int stmmac_rx(struct stmmac_priv desc_size = sizeof(struct dma_desc); } @@ -1006,7 +1006,7 @@ Signed-off-by: Jakub Kicinski <kuba@kernel.org> rx_q->dma_rx_phy, desc_size); } while (count < limit) { -@@ -5152,7 +5152,7 @@ read_again: +@@ -5165,7 +5165,7 @@ read_again: break; rx_q->cur_rx = STMMAC_GET_ENTRY(rx_q->cur_rx, @@ -1015,7 +1015,7 @@ Signed-off-by: Jakub Kicinski <kuba@kernel.org> next_entry = rx_q->cur_rx; if (priv->extend_desc) -@@ -5286,7 +5286,7 @@ read_again: +@@ -5299,7 +5299,7 @@ read_again: buf1_len, dma_dir); skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags, buf->page, buf->page_offset, buf1_len, @@ -1024,7 +1024,7 @@ Signed-off-by: Jakub Kicinski <kuba@kernel.org> /* Data payload appended into SKB */ page_pool_release_page(rx_q->page_pool, buf->page); -@@ -5298,7 +5298,7 @@ read_again: +@@ -5311,7 +5311,7 @@ read_again: buf2_len, dma_dir); skb_add_rx_frag(skb, skb_shinfo(skb)->nr_frags, buf->sec_page, 0, buf2_len, @@ -1033,7 +1033,7 @@ Signed-off-by: Jakub Kicinski <kuba@kernel.org> /* Data payload appended into SKB */ page_pool_release_page(rx_q->page_pool, buf->sec_page); -@@ -5740,11 +5740,13 @@ static irqreturn_t stmmac_safety_interru +@@ -5753,11 +5753,13 @@ static irqreturn_t stmmac_safety_interru static irqreturn_t stmmac_msi_intr_tx(int irq, void *data) { struct stmmac_tx_queue *tx_q = (struct stmmac_tx_queue *)data; @@ -1048,7 +1048,7 @@ Signed-off-by: Jakub Kicinski <kuba@kernel.org> if (unlikely(!data)) { netdev_err(priv->dev, "%s: invalid dev pointer\n", __func__); -@@ -5784,10 +5786,12 @@ static irqreturn_t stmmac_msi_intr_tx(in +@@ -5797,10 +5799,12 @@ static irqreturn_t stmmac_msi_intr_tx(in static irqreturn_t stmmac_msi_intr_rx(int irq, void *data) { struct stmmac_rx_queue *rx_q = (struct stmmac_rx_queue *)data; @@ -1062,7 +1062,7 @@ Signed-off-by: Jakub Kicinski <kuba@kernel.org> if (unlikely(!data)) { netdev_err(priv->dev, "%s: invalid dev pointer\n", __func__); -@@ -5818,10 +5822,10 @@ static void stmmac_poll_controller(struc +@@ -5831,10 +5835,10 @@ static void stmmac_poll_controller(struc if (priv->plat->multi_msi_en) { for (i = 0; i < priv->plat->rx_queues_to_use; i++) @@ -1075,7 +1075,7 @@ Signed-off-by: Jakub Kicinski <kuba@kernel.org> } else { disable_irq(dev->irq); stmmac_interrupt(dev->irq, dev); -@@ -6002,34 +6006,34 @@ static int stmmac_rings_status_show(stru +@@ -6015,34 +6019,34 @@ static int stmmac_rings_status_show(stru return 0; for (queue = 0; queue < rx_count; queue++) { @@ -1116,7 +1116,7 @@ Signed-off-by: Jakub Kicinski <kuba@kernel.org> } } -@@ -6370,7 +6374,7 @@ void stmmac_disable_rx_queue(struct stmm +@@ -6383,7 +6387,7 @@ void stmmac_disable_rx_queue(struct stmm void stmmac_enable_rx_queue(struct stmmac_priv *priv, u32 queue) { @@ -1125,7 +1125,7 @@ Signed-off-by: Jakub Kicinski <kuba@kernel.org> struct stmmac_channel *ch = &priv->channel[queue]; unsigned long flags; u32 buf_size; -@@ -6407,7 +6411,7 @@ void stmmac_enable_rx_queue(struct stmma +@@ -6420,7 +6424,7 @@ void stmmac_enable_rx_queue(struct stmma rx_q->queue_index); } else { stmmac_set_dma_bfsize(priv, priv->ioaddr, @@ -1134,7 +1134,7 @@ Signed-off-by: Jakub Kicinski <kuba@kernel.org> rx_q->queue_index); } -@@ -6433,7 +6437,7 @@ void stmmac_disable_tx_queue(struct stmm +@@ -6446,7 +6450,7 @@ void stmmac_disable_tx_queue(struct stmm void stmmac_enable_tx_queue(struct stmmac_priv *priv, u32 queue) { @@ -1143,7 +1143,7 @@ Signed-off-by: Jakub Kicinski <kuba@kernel.org> struct stmmac_channel *ch = &priv->channel[queue]; unsigned long flags; int ret; -@@ -6480,7 +6484,7 @@ void stmmac_xdp_release(struct net_devic +@@ -6493,7 +6497,7 @@ void stmmac_xdp_release(struct net_devic stmmac_disable_all_queues(priv); for (chan = 0; chan < priv->plat->tx_queues_to_use; chan++) @@ -1152,7 +1152,7 @@ Signed-off-by: Jakub Kicinski <kuba@kernel.org> /* Free the IRQ lines */ stmmac_free_irq(dev, REQ_IRQ_ERR_ALL, 0); -@@ -6539,7 +6543,7 @@ int stmmac_xdp_open(struct net_device *d +@@ -6552,7 +6556,7 @@ int stmmac_xdp_open(struct net_device *d /* DMA RX Channel Configuration */ for (chan = 0; chan < rx_cnt; chan++) { @@ -1161,7 +1161,7 @@ Signed-off-by: Jakub Kicinski <kuba@kernel.org> stmmac_init_rx_chan(priv, priv->ioaddr, priv->plat->dma_cfg, rx_q->dma_rx_phy, chan); -@@ -6557,7 +6561,7 @@ int stmmac_xdp_open(struct net_device *d +@@ -6570,7 +6574,7 @@ int stmmac_xdp_open(struct net_device *d rx_q->queue_index); } else { stmmac_set_dma_bfsize(priv, priv->ioaddr, @@ -1170,7 +1170,7 @@ Signed-off-by: Jakub Kicinski <kuba@kernel.org> rx_q->queue_index); } -@@ -6566,7 +6570,7 @@ int stmmac_xdp_open(struct net_device *d +@@ -6579,7 +6583,7 @@ int stmmac_xdp_open(struct net_device *d /* DMA TX Channel Configuration */ for (chan = 0; chan < tx_cnt; chan++) { @@ -1179,7 +1179,7 @@ Signed-off-by: Jakub Kicinski <kuba@kernel.org> stmmac_init_tx_chan(priv, priv->ioaddr, priv->plat->dma_cfg, tx_q->dma_tx_phy, chan); -@@ -6599,7 +6603,7 @@ int stmmac_xdp_open(struct net_device *d +@@ -6612,7 +6616,7 @@ int stmmac_xdp_open(struct net_device *d irq_error: for (chan = 0; chan < priv->plat->tx_queues_to_use; chan++) @@ -1188,7 +1188,7 @@ Signed-off-by: Jakub Kicinski <kuba@kernel.org> stmmac_hw_teardown(dev); init_error: -@@ -6626,8 +6630,8 @@ int stmmac_xsk_wakeup(struct net_device +@@ -6639,8 +6643,8 @@ int stmmac_xsk_wakeup(struct net_device queue >= priv->plat->tx_queues_to_use) return -EINVAL; @@ -1199,7 +1199,7 @@ Signed-off-by: Jakub Kicinski <kuba@kernel.org> ch = &priv->channel[queue]; if (!rx_q->xsk_pool && !tx_q->xsk_pool) -@@ -6883,8 +6887,8 @@ int stmmac_reinit_ringparam(struct net_d +@@ -6896,8 +6900,8 @@ int stmmac_reinit_ringparam(struct net_d if (netif_running(dev)) stmmac_release(dev); @@ -1210,7 +1210,7 @@ Signed-off-by: Jakub Kicinski <kuba@kernel.org> if (netif_running(dev)) ret = stmmac_open(dev); -@@ -7330,7 +7334,7 @@ int stmmac_suspend(struct device *dev) +@@ -7333,7 +7337,7 @@ int stmmac_suspend(struct device *dev) stmmac_disable_all_queues(priv); for (chan = 0; chan < priv->plat->tx_queues_to_use; chan++) @@ -1219,7 +1219,7 @@ Signed-off-by: Jakub Kicinski <kuba@kernel.org> if (priv->eee_enabled) { priv->tx_path_in_lpi_mode = false; -@@ -7381,7 +7385,7 @@ EXPORT_SYMBOL_GPL(stmmac_suspend); +@@ -7384,7 +7388,7 @@ EXPORT_SYMBOL_GPL(stmmac_suspend); static void stmmac_reset_rx_queue(struct stmmac_priv *priv, u32 queue) { @@ -1228,7 +1228,7 @@ Signed-off-by: Jakub Kicinski <kuba@kernel.org> rx_q->cur_rx = 0; rx_q->dirty_rx = 0; -@@ -7389,7 +7393,7 @@ static void stmmac_reset_rx_queue(struct +@@ -7392,7 +7396,7 @@ static void stmmac_reset_rx_queue(struct static void stmmac_reset_tx_queue(struct stmmac_priv *priv, u32 queue) { |