aboutsummaryrefslogtreecommitdiffstats
path: root/target/linux/mvebu/patches-5.4/009-net-mvneta-add-XDP_TX-support.patch
diff options
context:
space:
mode:
Diffstat (limited to 'target/linux/mvebu/patches-5.4/009-net-mvneta-add-XDP_TX-support.patch')
-rw-r--r--target/linux/mvebu/patches-5.4/009-net-mvneta-add-XDP_TX-support.patch175
1 files changed, 0 insertions, 175 deletions
diff --git a/target/linux/mvebu/patches-5.4/009-net-mvneta-add-XDP_TX-support.patch b/target/linux/mvebu/patches-5.4/009-net-mvneta-add-XDP_TX-support.patch
deleted file mode 100644
index 8d9b775f7b..0000000000
--- a/target/linux/mvebu/patches-5.4/009-net-mvneta-add-XDP_TX-support.patch
+++ /dev/null
@@ -1,175 +0,0 @@
-From b0a43db9087a21d96e1a0b716b8d9963064b2d58 Mon Sep 17 00:00:00 2001
-From: Lorenzo Bianconi <lorenzo@kernel.org>
-Date: Sat, 19 Oct 2019 10:13:27 +0200
-Subject: [PATCH 7/7] net: mvneta: add XDP_TX support
-
-Implement XDP_TX verdict and ndo_xdp_xmit net_device_ops function
-pointer
-
-Signed-off-by: Lorenzo Bianconi <lorenzo@kernel.org>
-Signed-off-by: David S. Miller <davem@davemloft.net>
----
- drivers/net/ethernet/marvell/mvneta.c | 128 ++++++++++++++++++++++++--
- 1 file changed, 121 insertions(+), 7 deletions(-)
-
---- a/drivers/net/ethernet/marvell/mvneta.c
-+++ b/drivers/net/ethernet/marvell/mvneta.c
-@@ -1813,16 +1813,19 @@ static void mvneta_txq_bufs_free(struct
-
- mvneta_txq_inc_get(txq);
-
-- if (!IS_TSO_HEADER(txq, tx_desc->buf_phys_addr))
-+ if (!IS_TSO_HEADER(txq, tx_desc->buf_phys_addr) &&
-+ buf->type != MVNETA_TYPE_XDP_TX)
- dma_unmap_single(pp->dev->dev.parent,
- tx_desc->buf_phys_addr,
- tx_desc->data_size, DMA_TO_DEVICE);
-- if (!buf->skb)
-- continue;
--
-- bytes_compl += buf->skb->len;
-- pkts_compl++;
-- dev_kfree_skb_any(buf->skb);
-+ if (buf->type == MVNETA_TYPE_SKB && buf->skb) {
-+ bytes_compl += buf->skb->len;
-+ pkts_compl++;
-+ dev_kfree_skb_any(buf->skb);
-+ } else if (buf->type == MVNETA_TYPE_XDP_TX ||
-+ buf->type == MVNETA_TYPE_XDP_NDO) {
-+ xdp_return_frame(buf->xdpf);
-+ }
- }
-
- netdev_tx_completed_queue(nq, pkts_compl, bytes_compl);
-@@ -1987,6 +1990,111 @@ int mvneta_rx_refill_queue(struct mvneta
- }
-
- static int
-+mvneta_xdp_submit_frame(struct mvneta_port *pp, struct mvneta_tx_queue *txq,
-+ struct xdp_frame *xdpf, bool dma_map)
-+{
-+ struct mvneta_tx_desc *tx_desc;
-+ struct mvneta_tx_buf *buf;
-+ dma_addr_t dma_addr;
-+
-+ if (txq->count >= txq->tx_stop_threshold)
-+ return MVNETA_XDP_DROPPED;
-+
-+ tx_desc = mvneta_txq_next_desc_get(txq);
-+
-+ buf = &txq->buf[txq->txq_put_index];
-+ if (dma_map) {
-+ /* ndo_xdp_xmit */
-+ dma_addr = dma_map_single(pp->dev->dev.parent, xdpf->data,
-+ xdpf->len, DMA_TO_DEVICE);
-+ if (dma_mapping_error(pp->dev->dev.parent, dma_addr)) {
-+ mvneta_txq_desc_put(txq);
-+ return MVNETA_XDP_DROPPED;
-+ }
-+ buf->type = MVNETA_TYPE_XDP_NDO;
-+ } else {
-+ struct page *page = virt_to_page(xdpf->data);
-+
-+ dma_addr = page_pool_get_dma_addr(page) +
-+ sizeof(*xdpf) + xdpf->headroom;
-+ dma_sync_single_for_device(pp->dev->dev.parent, dma_addr,
-+ xdpf->len, DMA_BIDIRECTIONAL);
-+ buf->type = MVNETA_TYPE_XDP_TX;
-+ }
-+ buf->xdpf = xdpf;
-+
-+ tx_desc->command = MVNETA_TXD_FLZ_DESC;
-+ tx_desc->buf_phys_addr = dma_addr;
-+ tx_desc->data_size = xdpf->len;
-+
-+ mvneta_update_stats(pp, 1, xdpf->len, true);
-+ mvneta_txq_inc_put(txq);
-+ txq->pending++;
-+ txq->count++;
-+
-+ return MVNETA_XDP_TX;
-+}
-+
-+static int
-+mvneta_xdp_xmit_back(struct mvneta_port *pp, struct xdp_buff *xdp)
-+{
-+ struct mvneta_tx_queue *txq;
-+ struct netdev_queue *nq;
-+ struct xdp_frame *xdpf;
-+ int cpu;
-+ u32 ret;
-+
-+ xdpf = convert_to_xdp_frame(xdp);
-+ if (unlikely(!xdpf))
-+ return MVNETA_XDP_DROPPED;
-+
-+ cpu = smp_processor_id();
-+ txq = &pp->txqs[cpu % txq_number];
-+ nq = netdev_get_tx_queue(pp->dev, txq->id);
-+
-+ __netif_tx_lock(nq, cpu);
-+ ret = mvneta_xdp_submit_frame(pp, txq, xdpf, false);
-+ if (ret == MVNETA_XDP_TX)
-+ mvneta_txq_pend_desc_add(pp, txq, 0);
-+ __netif_tx_unlock(nq);
-+
-+ return ret;
-+}
-+
-+static int
-+mvneta_xdp_xmit(struct net_device *dev, int num_frame,
-+ struct xdp_frame **frames, u32 flags)
-+{
-+ struct mvneta_port *pp = netdev_priv(dev);
-+ int cpu = smp_processor_id();
-+ struct mvneta_tx_queue *txq;
-+ struct netdev_queue *nq;
-+ int i, drops = 0;
-+ u32 ret;
-+
-+ if (unlikely(flags & ~XDP_XMIT_FLAGS_MASK))
-+ return -EINVAL;
-+
-+ txq = &pp->txqs[cpu % txq_number];
-+ nq = netdev_get_tx_queue(pp->dev, txq->id);
-+
-+ __netif_tx_lock(nq, cpu);
-+ for (i = 0; i < num_frame; i++) {
-+ ret = mvneta_xdp_submit_frame(pp, txq, frames[i], true);
-+ if (ret != MVNETA_XDP_TX) {
-+ xdp_return_frame_rx_napi(frames[i]);
-+ drops++;
-+ }
-+ }
-+
-+ if (unlikely(flags & XDP_XMIT_FLUSH))
-+ mvneta_txq_pend_desc_add(pp, txq, 0);
-+ __netif_tx_unlock(nq);
-+
-+ return num_frame - drops;
-+}
-+
-+static int
- mvneta_run_xdp(struct mvneta_port *pp, struct mvneta_rx_queue *rxq,
- struct bpf_prog *prog, struct xdp_buff *xdp)
- {
-@@ -2008,6 +2116,11 @@ mvneta_run_xdp(struct mvneta_port *pp, s
- }
- break;
- }
-+ case XDP_TX:
-+ ret = mvneta_xdp_xmit_back(pp, xdp);
-+ if (ret != MVNETA_XDP_TX)
-+ xdp_return_buff(xdp);
-+ break;
- default:
- bpf_warn_invalid_xdp_action(act);
- /* fall through */
-@@ -4581,6 +4694,7 @@ static const struct net_device_ops mvnet
- .ndo_get_stats64 = mvneta_get_stats64,
- .ndo_do_ioctl = mvneta_ioctl,
- .ndo_bpf = mvneta_xdp,
-+ .ndo_xdp_xmit = mvneta_xdp_xmit,
- };
-
- static const struct ethtool_ops mvneta_eth_tool_ops = {