aboutsummaryrefslogtreecommitdiffstats
path: root/package/boot/uboot-mediatek/patches/002-0007-net-mediatek-stop-using-bitfileds-for-DMA-descriptor.patch
diff options
context:
space:
mode:
Diffstat (limited to 'package/boot/uboot-mediatek/patches/002-0007-net-mediatek-stop-using-bitfileds-for-DMA-descriptor.patch')
-rw-r--r--package/boot/uboot-mediatek/patches/002-0007-net-mediatek-stop-using-bitfileds-for-DMA-descriptor.patch317
1 files changed, 317 insertions, 0 deletions
diff --git a/package/boot/uboot-mediatek/patches/002-0007-net-mediatek-stop-using-bitfileds-for-DMA-descriptor.patch b/package/boot/uboot-mediatek/patches/002-0007-net-mediatek-stop-using-bitfileds-for-DMA-descriptor.patch
new file mode 100644
index 0000000000..3cf45b7f3e
--- /dev/null
+++ b/package/boot/uboot-mediatek/patches/002-0007-net-mediatek-stop-using-bitfileds-for-DMA-descriptor.patch
@@ -0,0 +1,317 @@
+From b978c067075fddbac341bf551ebef29e78767b75 Mon Sep 17 00:00:00 2001
+From: Weijie Gao <weijie.gao@mediatek.com>
+Date: Wed, 27 Jul 2022 09:32:29 +0800
+Subject: [PATCH 07/31] net: mediatek: stop using bitfileds for DMA descriptors
+
+This patch is a preparation for adding a new version of PDMA of which the
+DMA descriptor fields has changed. Using bitfields will result in a complex
+modification. Convert bitfields to u32 units can solve this problem easily.
+
+Reviewed-by: Simon Glass <sjg@chromium.org>
+Signed-off-by: Weijie Gao <weijie.gao@mediatek.com>
+---
+ drivers/net/mtk_eth.c | 144 ++++++++++++++----------------------------
+ drivers/net/mtk_eth.h | 32 ++++++++++
+ 2 files changed, 80 insertions(+), 96 deletions(-)
+
+--- a/drivers/net/mtk_eth.c
++++ b/drivers/net/mtk_eth.c
+@@ -65,77 +65,6 @@
+ (DP_DISCARD << MC_DP_S) | \
+ (DP_DISCARD << UN_DP_S))
+
+-struct pdma_rxd_info1 {
+- u32 PDP0;
+-};
+-
+-struct pdma_rxd_info2 {
+- u32 PLEN1 : 14;
+- u32 LS1 : 1;
+- u32 UN_USED : 1;
+- u32 PLEN0 : 14;
+- u32 LS0 : 1;
+- u32 DDONE : 1;
+-};
+-
+-struct pdma_rxd_info3 {
+- u32 PDP1;
+-};
+-
+-struct pdma_rxd_info4 {
+- u32 FOE_ENTRY : 14;
+- u32 CRSN : 5;
+- u32 SP : 3;
+- u32 L4F : 1;
+- u32 L4VLD : 1;
+- u32 TACK : 1;
+- u32 IP4F : 1;
+- u32 IP4 : 1;
+- u32 IP6 : 1;
+- u32 UN_USED : 4;
+-};
+-
+-struct pdma_rxdesc {
+- struct pdma_rxd_info1 rxd_info1;
+- struct pdma_rxd_info2 rxd_info2;
+- struct pdma_rxd_info3 rxd_info3;
+- struct pdma_rxd_info4 rxd_info4;
+-};
+-
+-struct pdma_txd_info1 {
+- u32 SDP0;
+-};
+-
+-struct pdma_txd_info2 {
+- u32 SDL1 : 14;
+- u32 LS1 : 1;
+- u32 BURST : 1;
+- u32 SDL0 : 14;
+- u32 LS0 : 1;
+- u32 DDONE : 1;
+-};
+-
+-struct pdma_txd_info3 {
+- u32 SDP1;
+-};
+-
+-struct pdma_txd_info4 {
+- u32 VLAN_TAG : 16;
+- u32 INS : 1;
+- u32 RESV : 2;
+- u32 UDF : 6;
+- u32 FPORT : 3;
+- u32 TSO : 1;
+- u32 TUI_CO : 3;
+-};
+-
+-struct pdma_txdesc {
+- struct pdma_txd_info1 txd_info1;
+- struct pdma_txd_info2 txd_info2;
+- struct pdma_txd_info3 txd_info3;
+- struct pdma_txd_info4 txd_info4;
+-};
+-
+ enum mtk_switch {
+ SW_NONE,
+ SW_MT7530,
+@@ -151,13 +80,15 @@ enum mtk_switch {
+ struct mtk_soc_data {
+ u32 caps;
+ u32 ana_rgc3;
++ u32 txd_size;
++ u32 rxd_size;
+ };
+
+ struct mtk_eth_priv {
+ char pkt_pool[TOTAL_PKT_BUF_SIZE] __aligned(ARCH_DMA_MINALIGN);
+
+- struct pdma_txdesc *tx_ring_noc;
+- struct pdma_rxdesc *rx_ring_noc;
++ void *tx_ring_noc;
++ void *rx_ring_noc;
+
+ int rx_dma_owner_idx0;
+ int tx_cpu_owner_idx0;
+@@ -1202,14 +1133,16 @@ static void mtk_mac_init(struct mtk_eth_
+ static void mtk_eth_fifo_init(struct mtk_eth_priv *priv)
+ {
+ char *pkt_base = priv->pkt_pool;
++ struct mtk_tx_dma *txd;
++ struct mtk_rx_dma *rxd;
+ int i;
+
+ mtk_pdma_rmw(priv, PDMA_GLO_CFG_REG, 0xffff0000, 0);
+ udelay(500);
+
+- memset(priv->tx_ring_noc, 0, NUM_TX_DESC * sizeof(struct pdma_txdesc));
+- memset(priv->rx_ring_noc, 0, NUM_RX_DESC * sizeof(struct pdma_rxdesc));
+- memset(priv->pkt_pool, 0, TOTAL_PKT_BUF_SIZE);
++ memset(priv->tx_ring_noc, 0, NUM_TX_DESC * priv->soc->txd_size);
++ memset(priv->rx_ring_noc, 0, NUM_RX_DESC * priv->soc->rxd_size);
++ memset(priv->pkt_pool, 0xff, TOTAL_PKT_BUF_SIZE);
+
+ flush_dcache_range((ulong)pkt_base,
+ (ulong)(pkt_base + TOTAL_PKT_BUF_SIZE));
+@@ -1218,17 +1151,21 @@ static void mtk_eth_fifo_init(struct mtk
+ priv->tx_cpu_owner_idx0 = 0;
+
+ for (i = 0; i < NUM_TX_DESC; i++) {
+- priv->tx_ring_noc[i].txd_info2.LS0 = 1;
+- priv->tx_ring_noc[i].txd_info2.DDONE = 1;
+- priv->tx_ring_noc[i].txd_info4.FPORT = priv->gmac_id + 1;
++ txd = priv->tx_ring_noc + i * priv->soc->txd_size;
++
++ txd->txd1 = virt_to_phys(pkt_base);
++ txd->txd2 = PDMA_TXD2_DDONE | PDMA_TXD2_LS0;
++ txd->txd4 = PDMA_TXD4_FPORT_SET(priv->gmac_id + 1);
+
+- priv->tx_ring_noc[i].txd_info1.SDP0 = virt_to_phys(pkt_base);
+ pkt_base += PKTSIZE_ALIGN;
+ }
+
+ for (i = 0; i < NUM_RX_DESC; i++) {
+- priv->rx_ring_noc[i].rxd_info2.PLEN0 = PKTSIZE_ALIGN;
+- priv->rx_ring_noc[i].rxd_info1.PDP0 = virt_to_phys(pkt_base);
++ rxd = priv->rx_ring_noc + i * priv->soc->rxd_size;
++
++ rxd->rxd1 = virt_to_phys(pkt_base);
++ rxd->rxd2 = PDMA_RXD2_PLEN0_SET(PKTSIZE_ALIGN);
++
+ pkt_base += PKTSIZE_ALIGN;
+ }
+
+@@ -1315,20 +1252,22 @@ static int mtk_eth_send(struct udevice *
+ {
+ struct mtk_eth_priv *priv = dev_get_priv(dev);
+ u32 idx = priv->tx_cpu_owner_idx0;
++ struct mtk_tx_dma *txd;
+ void *pkt_base;
+
+- if (!priv->tx_ring_noc[idx].txd_info2.DDONE) {
++ txd = priv->tx_ring_noc + idx * priv->soc->txd_size;
++
++ if (!(txd->txd2 & PDMA_TXD2_DDONE)) {
+ debug("mtk-eth: TX DMA descriptor ring is full\n");
+ return -EPERM;
+ }
+
+- pkt_base = (void *)phys_to_virt(priv->tx_ring_noc[idx].txd_info1.SDP0);
++ pkt_base = (void *)phys_to_virt(txd->txd1);
+ memcpy(pkt_base, packet, length);
+ flush_dcache_range((ulong)pkt_base, (ulong)pkt_base +
+ roundup(length, ARCH_DMA_MINALIGN));
+
+- priv->tx_ring_noc[idx].txd_info2.SDL0 = length;
+- priv->tx_ring_noc[idx].txd_info2.DDONE = 0;
++ txd->txd2 = PDMA_TXD2_LS0 | PDMA_TXD2_SDL0_SET(length);
+
+ priv->tx_cpu_owner_idx0 = (priv->tx_cpu_owner_idx0 + 1) % NUM_TX_DESC;
+ mtk_pdma_write(priv, TX_CTX_IDX_REG(0), priv->tx_cpu_owner_idx0);
+@@ -1340,16 +1279,20 @@ static int mtk_eth_recv(struct udevice *
+ {
+ struct mtk_eth_priv *priv = dev_get_priv(dev);
+ u32 idx = priv->rx_dma_owner_idx0;
++ struct mtk_rx_dma *rxd;
+ uchar *pkt_base;
+ u32 length;
+
+- if (!priv->rx_ring_noc[idx].rxd_info2.DDONE) {
++ rxd = priv->rx_ring_noc + idx * priv->soc->rxd_size;
++
++ if (!(rxd->rxd2 & PDMA_RXD2_DDONE)) {
+ debug("mtk-eth: RX DMA descriptor ring is empty\n");
+ return -EAGAIN;
+ }
+
+- length = priv->rx_ring_noc[idx].rxd_info2.PLEN0;
+- pkt_base = (void *)phys_to_virt(priv->rx_ring_noc[idx].rxd_info1.PDP0);
++ length = PDMA_RXD2_PLEN0_GET(rxd->rxd2);
++
++ pkt_base = (void *)phys_to_virt(rxd->rxd1);
+ invalidate_dcache_range((ulong)pkt_base, (ulong)pkt_base +
+ roundup(length, ARCH_DMA_MINALIGN));
+
+@@ -1363,10 +1306,11 @@ static int mtk_eth_free_pkt(struct udevi
+ {
+ struct mtk_eth_priv *priv = dev_get_priv(dev);
+ u32 idx = priv->rx_dma_owner_idx0;
++ struct mtk_rx_dma *rxd;
++
++ rxd = priv->rx_ring_noc + idx * priv->soc->rxd_size;
+
+- priv->rx_ring_noc[idx].rxd_info2.DDONE = 0;
+- priv->rx_ring_noc[idx].rxd_info2.LS0 = 0;
+- priv->rx_ring_noc[idx].rxd_info2.PLEN0 = PKTSIZE_ALIGN;
++ rxd->rxd2 = PDMA_RXD2_PLEN0_SET(PKTSIZE_ALIGN);
+
+ mtk_pdma_write(priv, RX_CRX_IDX_REG(0), idx);
+ priv->rx_dma_owner_idx0 = (priv->rx_dma_owner_idx0 + 1) % NUM_RX_DESC;
+@@ -1393,11 +1337,11 @@ static int mtk_eth_probe(struct udevice
+ return ret;
+
+ /* Prepare for tx/rx rings */
+- priv->tx_ring_noc = (struct pdma_txdesc *)
+- noncached_alloc(sizeof(struct pdma_txdesc) * NUM_TX_DESC,
++ priv->tx_ring_noc = (void *)
++ noncached_alloc(priv->soc->txd_size * NUM_TX_DESC,
+ ARCH_DMA_MINALIGN);
+- priv->rx_ring_noc = (struct pdma_rxdesc *)
+- noncached_alloc(sizeof(struct pdma_rxdesc) * NUM_RX_DESC,
++ priv->rx_ring_noc = (void *)
++ noncached_alloc(priv->soc->rxd_size * NUM_RX_DESC,
+ ARCH_DMA_MINALIGN);
+
+ /* Set MAC mode */
+@@ -1554,18 +1498,26 @@ static int mtk_eth_of_to_plat(struct ude
+
+ static const struct mtk_soc_data mt7629_data = {
+ .ana_rgc3 = 0x128,
++ .txd_size = sizeof(struct mtk_tx_dma),
++ .rxd_size = sizeof(struct mtk_rx_dma),
+ };
+
+ static const struct mtk_soc_data mt7623_data = {
+ .caps = MT7623_CAPS,
++ .txd_size = sizeof(struct mtk_tx_dma),
++ .rxd_size = sizeof(struct mtk_rx_dma),
+ };
+
+ static const struct mtk_soc_data mt7622_data = {
+ .ana_rgc3 = 0x2028,
++ .txd_size = sizeof(struct mtk_tx_dma),
++ .rxd_size = sizeof(struct mtk_rx_dma),
+ };
+
+ static const struct mtk_soc_data mt7621_data = {
+ .caps = MT7621_CAPS,
++ .txd_size = sizeof(struct mtk_tx_dma),
++ .rxd_size = sizeof(struct mtk_rx_dma),
+ };
+
+ static const struct udevice_id mtk_eth_ids[] = {
+--- a/drivers/net/mtk_eth.h
++++ b/drivers/net/mtk_eth.h
+@@ -10,6 +10,7 @@
+ #define _MTK_ETH_H_
+
+ #include <linux/bitops.h>
++#include <linux/bitfield.h>
+
+ enum mkt_eth_capabilities {
+ MTK_TRGMII_BIT,
+@@ -435,4 +436,35 @@ enum mkt_eth_capabilities {
+ #define PHY_POWER_SAVING_M 0x300
+ #define PHY_POWER_SAVING_TX 0x0
+
++/* PDMA descriptors */
++struct mtk_rx_dma {
++ unsigned int rxd1;
++ unsigned int rxd2;
++ unsigned int rxd3;
++ unsigned int rxd4;
++} __packed __aligned(4);
++
++struct mtk_tx_dma {
++ unsigned int txd1;
++ unsigned int txd2;
++ unsigned int txd3;
++ unsigned int txd4;
++} __packed __aligned(4);
++
++/* PDMA TXD fields */
++#define PDMA_TXD2_DDONE BIT(31)
++#define PDMA_TXD2_LS0 BIT(30)
++#define PDMA_TXD2_SDL0_M GENMASK(29, 16)
++#define PDMA_TXD2_SDL0_SET(_v) FIELD_PREP(PDMA_TXD2_SDL0_M, (_v))
++
++#define PDMA_TXD4_FPORT_M GENMASK(27, 25)
++#define PDMA_TXD4_FPORT_SET(_v) FIELD_PREP(PDMA_TXD4_FPORT_M, (_v))
++
++/* PDMA RXD fields */
++#define PDMA_RXD2_DDONE BIT(31)
++#define PDMA_RXD2_LS0 BIT(30)
++#define PDMA_RXD2_PLEN0_M GENMASK(29, 16)
++#define PDMA_RXD2_PLEN0_GET(_v) FIELD_GET(PDMA_RXD2_PLEN0_M, (_v))
++#define PDMA_RXD2_PLEN0_SET(_v) FIELD_PREP(PDMA_RXD2_PLEN0_M, (_v))
++
+ #endif /* _MTK_ETH_H_ */