aboutsummaryrefslogtreecommitdiffstats
path: root/target/linux/layerscape/patches-4.14/711-dpaa-bqman-support-layerscape.patch
diff options
context:
space:
mode:
Diffstat (limited to 'target/linux/layerscape/patches-4.14/711-dpaa-bqman-support-layerscape.patch')
-rw-r--r--target/linux/layerscape/patches-4.14/711-dpaa-bqman-support-layerscape.patch550
1 files changed, 550 insertions, 0 deletions
diff --git a/target/linux/layerscape/patches-4.14/711-dpaa-bqman-support-layerscape.patch b/target/linux/layerscape/patches-4.14/711-dpaa-bqman-support-layerscape.patch
new file mode 100644
index 0000000000..89ace283b9
--- /dev/null
+++ b/target/linux/layerscape/patches-4.14/711-dpaa-bqman-support-layerscape.patch
@@ -0,0 +1,550 @@
+From 48dbe4b3a31795b8efdfff82f69eccd086052eed Mon Sep 17 00:00:00 2001
+From: Biwen Li <biwen.li@nxp.com>
+Date: Fri, 16 Nov 2018 10:27:30 +0800
+Subject: [PATCH 16/39] dpaa-bqman: support layerscape
+This is an integrated patch of dpaa-bqman for layerscape
+
+Signed-off-by: Claudiu Manoil <claudiu.manoil@nxp.com>
+Signed-off-by: Madalin Bucur <madalin.bucur@nxp.com>
+Signed-off-by: Roy Pledge <roy.pledge@nxp.com>
+Signed-off-by: Stuart Yoder <stuart.yoder@nxp.com>
+Signed-off-by: Valentin Rothberg <valentinrothberg@gmail.com>
+Signed-off-by: Biwen Li <biwen.li@nxp.com>
+---
+ drivers/soc/fsl/qbman/Kconfig | 2 +-
+ drivers/soc/fsl/qbman/bman.c | 24 ++++-
+ drivers/soc/fsl/qbman/bman_ccsr.c | 35 ++++++-
+ drivers/soc/fsl/qbman/bman_portal.c | 12 ++-
+ drivers/soc/fsl/qbman/bman_priv.h | 3 +
+ drivers/soc/fsl/qbman/dpaa_sys.h | 8 +-
+ drivers/soc/fsl/qbman/qman.c | 46 ++++++++-
+ drivers/soc/fsl/qbman/qman_ccsr.c | 140 ++++++++++++++++++++++------
+ drivers/soc/fsl/qbman/qman_portal.c | 12 ++-
+ drivers/soc/fsl/qbman/qman_priv.h | 5 +-
+ drivers/soc/fsl/qbman/qman_test.h | 2 -
+ 11 files changed, 236 insertions(+), 53 deletions(-)
+
+--- a/drivers/soc/fsl/qbman/Kconfig
++++ b/drivers/soc/fsl/qbman/Kconfig
+@@ -1,6 +1,6 @@
+ menuconfig FSL_DPAA
+ bool "Freescale DPAA 1.x support"
+- depends on FSL_SOC_BOOKE
++ depends on (FSL_SOC_BOOKE || ARCH_LAYERSCAPE)
+ select GENERIC_ALLOCATOR
+ help
+ The Freescale Data Path Acceleration Architecture (DPAA) is a set of
+--- a/drivers/soc/fsl/qbman/bman.c
++++ b/drivers/soc/fsl/qbman/bman.c
+@@ -35,6 +35,27 @@
+
+ /* Portal register assists */
+
++#if defined(CONFIG_ARM) || defined(CONFIG_ARM64)
++/* Cache-inhibited register offsets */
++#define BM_REG_RCR_PI_CINH 0x3000
++#define BM_REG_RCR_CI_CINH 0x3100
++#define BM_REG_RCR_ITR 0x3200
++#define BM_REG_CFG 0x3300
++#define BM_REG_SCN(n) (0x3400 + ((n) << 6))
++#define BM_REG_ISR 0x3e00
++#define BM_REG_IER 0x3e40
++#define BM_REG_ISDR 0x3e80
++#define BM_REG_IIR 0x3ec0
++
++/* Cache-enabled register offsets */
++#define BM_CL_CR 0x0000
++#define BM_CL_RR0 0x0100
++#define BM_CL_RR1 0x0140
++#define BM_CL_RCR 0x1000
++#define BM_CL_RCR_PI_CENA 0x3000
++#define BM_CL_RCR_CI_CENA 0x3100
++
++#else
+ /* Cache-inhibited register offsets */
+ #define BM_REG_RCR_PI_CINH 0x0000
+ #define BM_REG_RCR_CI_CINH 0x0004
+@@ -53,6 +74,7 @@
+ #define BM_CL_RCR 0x1000
+ #define BM_CL_RCR_PI_CENA 0x3000
+ #define BM_CL_RCR_CI_CENA 0x3100
++#endif
+
+ /*
+ * Portal modes.
+@@ -607,7 +629,7 @@ int bman_p_irqsource_add(struct bman_por
+ unsigned long irqflags;
+
+ local_irq_save(irqflags);
+- set_bits(bits & BM_PIRQ_VISIBLE, &p->irq_sources);
++ p->irq_sources |= bits & BM_PIRQ_VISIBLE;
+ bm_out(&p->p, BM_REG_IER, p->irq_sources);
+ local_irq_restore(irqflags);
+ return 0;
+--- a/drivers/soc/fsl/qbman/bman_ccsr.c
++++ b/drivers/soc/fsl/qbman/bman_ccsr.c
+@@ -170,10 +170,11 @@ static int fsl_bman_probe(struct platfor
+ {
+ int ret, err_irq;
+ struct device *dev = &pdev->dev;
+- struct device_node *node = dev->of_node;
++ struct device_node *mem_node, *node = dev->of_node;
+ struct resource *res;
+ u16 id, bm_pool_cnt;
+ u8 major, minor;
++ u64 size;
+
+ res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
+ if (!res) {
+@@ -201,6 +202,38 @@ static int fsl_bman_probe(struct platfor
+ return -ENODEV;
+ }
+
++ /*
++ * If FBPR memory wasn't defined using the qbman compatiable string
++ * try using the of_reserved_mem_device method
++ */
++ if (!fbpr_a) {
++ ret = of_reserved_mem_device_init(dev);
++ if (ret) {
++ dev_err(dev, "of_reserved_mem_device_init() failed 0x%x\n",
++ ret);
++ return -ENODEV;
++ }
++ mem_node = of_parse_phandle(dev->of_node, "memory-region", 0);
++ if (mem_node) {
++ ret = of_property_read_u64(mem_node, "size", &size);
++ if (ret) {
++ dev_err(dev, "FBPR: of_address_to_resource fails 0x%x\n",
++ ret);
++ return -ENODEV;
++ }
++ fbpr_sz = size;
++ } else {
++ dev_err(dev, "No memory-region found for FBPR\n");
++ return -ENODEV;
++ }
++ if (!dma_zalloc_coherent(dev, fbpr_sz, &fbpr_a, 0)) {
++ dev_err(dev, "Alloc FBPR memory failed\n");
++ return -ENODEV;
++ }
++ }
++
++ dev_dbg(dev, "Allocated FBPR 0x%llx 0x%zx\n", fbpr_a, fbpr_sz);
++
+ bm_set_memory(fbpr_a, fbpr_sz);
+
+ err_irq = platform_get_irq(pdev, 0);
+--- a/drivers/soc/fsl/qbman/bman_portal.c
++++ b/drivers/soc/fsl/qbman/bman_portal.c
+@@ -123,7 +123,14 @@ static int bman_portal_probe(struct plat
+ }
+ pcfg->irq = irq;
+
+- va = ioremap_prot(addr_phys[0]->start, resource_size(addr_phys[0]), 0);
++#ifdef CONFIG_PPC
++ /* PPC requires a cacheable/non-coherent mapping of the portal */
++ va = ioremap_prot(addr_phys[0]->start, resource_size(addr_phys[0]),
++ (pgprot_val(PAGE_KERNEL) & ~_PAGE_COHERENT));
++#else
++ /* For ARM we can use write combine mapping. */
++ va = ioremap_wc(addr_phys[0]->start, resource_size(addr_phys[0]));
++#endif
+ if (!va) {
+ dev_err(dev, "ioremap::CE failed\n");
+ goto err_ioremap1;
+@@ -131,8 +138,7 @@ static int bman_portal_probe(struct plat
+
+ pcfg->addr_virt[DPAA_PORTAL_CE] = va;
+
+- va = ioremap_prot(addr_phys[1]->start, resource_size(addr_phys[1]),
+- _PAGE_GUARDED | _PAGE_NO_CACHE);
++ va = ioremap(addr_phys[1]->start, resource_size(addr_phys[1]));
+ if (!va) {
+ dev_err(dev, "ioremap::CI failed\n");
+ goto err_ioremap2;
+--- a/drivers/soc/fsl/qbman/bman_priv.h
++++ b/drivers/soc/fsl/qbman/bman_priv.h
+@@ -33,6 +33,9 @@
+ #include "dpaa_sys.h"
+
+ #include <soc/fsl/bman.h>
++#include <linux/dma-contiguous.h>
++#include <linux/of_address.h>
++#include <linux/dma-mapping.h>
+
+ /* Portal processing (interrupt) sources */
+ #define BM_PIRQ_RCRI 0x00000002 /* RCR Ring (below threshold) */
+--- a/drivers/soc/fsl/qbman/dpaa_sys.h
++++ b/drivers/soc/fsl/qbman/dpaa_sys.h
+@@ -44,20 +44,18 @@
+ #include <linux/prefetch.h>
+ #include <linux/genalloc.h>
+ #include <asm/cacheflush.h>
++#include <linux/io.h>
++#include <linux/delay.h>
+
+ /* For 2-element tables related to cache-inhibited and cache-enabled mappings */
+ #define DPAA_PORTAL_CE 0
+ #define DPAA_PORTAL_CI 1
+
+-#if (L1_CACHE_BYTES != 32) && (L1_CACHE_BYTES != 64)
+-#error "Unsupported Cacheline Size"
+-#endif
+-
+ static inline void dpaa_flush(void *p)
+ {
+ #ifdef CONFIG_PPC
+ flush_dcache_range((unsigned long)p, (unsigned long)p+64);
+-#elif defined(CONFIG_ARM32)
++#elif defined(CONFIG_ARM)
+ __cpuc_flush_dcache_area(p, 64);
+ #elif defined(CONFIG_ARM64)
+ __flush_dcache_area(p, 64);
+--- a/drivers/soc/fsl/qbman/qman.c
++++ b/drivers/soc/fsl/qbman/qman.c
+@@ -41,6 +41,43 @@
+
+ /* Portal register assists */
+
++#if defined(CONFIG_ARM) || defined(CONFIG_ARM64)
++/* Cache-inhibited register offsets */
++#define QM_REG_EQCR_PI_CINH 0x3000
++#define QM_REG_EQCR_CI_CINH 0x3040
++#define QM_REG_EQCR_ITR 0x3080
++#define QM_REG_DQRR_PI_CINH 0x3100
++#define QM_REG_DQRR_CI_CINH 0x3140
++#define QM_REG_DQRR_ITR 0x3180
++#define QM_REG_DQRR_DCAP 0x31C0
++#define QM_REG_DQRR_SDQCR 0x3200
++#define QM_REG_DQRR_VDQCR 0x3240
++#define QM_REG_DQRR_PDQCR 0x3280
++#define QM_REG_MR_PI_CINH 0x3300
++#define QM_REG_MR_CI_CINH 0x3340
++#define QM_REG_MR_ITR 0x3380
++#define QM_REG_CFG 0x3500
++#define QM_REG_ISR 0x3600
++#define QM_REG_IER 0x3640
++#define QM_REG_ISDR 0x3680
++#define QM_REG_IIR 0x36C0
++#define QM_REG_ITPR 0x3740
++
++/* Cache-enabled register offsets */
++#define QM_CL_EQCR 0x0000
++#define QM_CL_DQRR 0x1000
++#define QM_CL_MR 0x2000
++#define QM_CL_EQCR_PI_CENA 0x3000
++#define QM_CL_EQCR_CI_CENA 0x3040
++#define QM_CL_DQRR_PI_CENA 0x3100
++#define QM_CL_DQRR_CI_CENA 0x3140
++#define QM_CL_MR_PI_CENA 0x3300
++#define QM_CL_MR_CI_CENA 0x3340
++#define QM_CL_CR 0x3800
++#define QM_CL_RR0 0x3900
++#define QM_CL_RR1 0x3940
++
++#else
+ /* Cache-inhibited register offsets */
+ #define QM_REG_EQCR_PI_CINH 0x0000
+ #define QM_REG_EQCR_CI_CINH 0x0004
+@@ -75,6 +112,7 @@
+ #define QM_CL_CR 0x3800
+ #define QM_CL_RR0 0x3900
+ #define QM_CL_RR1 0x3940
++#endif
+
+ /*
+ * BTW, the drivers (and h/w programming model) already obtain the required
+@@ -909,12 +947,12 @@ static inline int qm_mc_result_timeout(s
+
+ static inline void fq_set(struct qman_fq *fq, u32 mask)
+ {
+- set_bits(mask, &fq->flags);
++ fq->flags |= mask;
+ }
+
+ static inline void fq_clear(struct qman_fq *fq, u32 mask)
+ {
+- clear_bits(mask, &fq->flags);
++ fq->flags &= ~mask;
+ }
+
+ static inline int fq_isset(struct qman_fq *fq, u32 mask)
+@@ -1566,7 +1604,7 @@ void qman_p_irqsource_add(struct qman_po
+ unsigned long irqflags;
+
+ local_irq_save(irqflags);
+- set_bits(bits & QM_PIRQ_VISIBLE, &p->irq_sources);
++ p->irq_sources |= bits & QM_PIRQ_VISIBLE;
+ qm_out(&p->p, QM_REG_IER, p->irq_sources);
+ local_irq_restore(irqflags);
+ }
+@@ -1589,7 +1627,7 @@ void qman_p_irqsource_remove(struct qman
+ */
+ local_irq_save(irqflags);
+ bits &= QM_PIRQ_VISIBLE;
+- clear_bits(bits, &p->irq_sources);
++ p->irq_sources &= ~bits;
+ qm_out(&p->p, QM_REG_IER, p->irq_sources);
+ ier = qm_in(&p->p, QM_REG_IER);
+ /*
+--- a/drivers/soc/fsl/qbman/qman_ccsr.c
++++ b/drivers/soc/fsl/qbman/qman_ccsr.c
+@@ -401,21 +401,42 @@ static int qm_init_pfdr(struct device *d
+ }
+
+ /*
+- * Ideally we would use the DMA API to turn rmem->base into a DMA address
+- * (especially if iommu translations ever get involved). Unfortunately, the
+- * DMA API currently does not allow mapping anything that is not backed with
+- * a struct page.
++ * QMan needs two global memory areas initialized at boot time:
++ * 1) FQD: Frame Queue Descriptors used to manage frame queues
++ * 2) PFDR: Packed Frame Queue Descriptor Records used to store frames
++ * Both areas are reserved using the device tree reserved memory framework
++ * and the addresses and sizes are initialized when the QMan device is probed
+ */
+ static dma_addr_t fqd_a, pfdr_a;
+ static size_t fqd_sz, pfdr_sz;
+
++#ifdef CONFIG_PPC
++/*
++ * Support for PPC Device Tree backward compatibility when compatiable
++ * string is set to fsl-qman-fqd and fsl-qman-pfdr
++ */
++static int zero_priv_mem(phys_addr_t addr, size_t sz)
++{
++ /* map as cacheable, non-guarded */
++ void __iomem *tmpp = ioremap_prot(addr, sz, 0);
++
++ if (!tmpp)
++ return -ENOMEM;
++
++ memset_io(tmpp, 0, sz);
++ flush_dcache_range((unsigned long)tmpp,
++ (unsigned long)tmpp + sz);
++ iounmap(tmpp);
++
++ return 0;
++}
++
+ static int qman_fqd(struct reserved_mem *rmem)
+ {
+ fqd_a = rmem->base;
+ fqd_sz = rmem->size;
+
+ WARN_ON(!(fqd_a && fqd_sz));
+-
+ return 0;
+ }
+ RESERVEDMEM_OF_DECLARE(qman_fqd, "fsl,qman-fqd", qman_fqd);
+@@ -431,32 +452,13 @@ static int qman_pfdr(struct reserved_mem
+ }
+ RESERVEDMEM_OF_DECLARE(qman_pfdr, "fsl,qman-pfdr", qman_pfdr);
+
++#endif
++
+ static unsigned int qm_get_fqid_maxcnt(void)
+ {
+ return fqd_sz / 64;
+ }
+
+-/*
+- * Flush this memory range from data cache so that QMAN originated
+- * transactions for this memory region could be marked non-coherent.
+- */
+-static int zero_priv_mem(struct device *dev, struct device_node *node,
+- phys_addr_t addr, size_t sz)
+-{
+- /* map as cacheable, non-guarded */
+- void __iomem *tmpp = ioremap_prot(addr, sz, 0);
+-
+- if (!tmpp)
+- return -ENOMEM;
+-
+- memset_io(tmpp, 0, sz);
+- flush_dcache_range((unsigned long)tmpp,
+- (unsigned long)tmpp + sz);
+- iounmap(tmpp);
+-
+- return 0;
+-}
+-
+ static void log_edata_bits(struct device *dev, u32 bit_count)
+ {
+ u32 i, j, mask = 0xffffffff;
+@@ -687,11 +689,12 @@ static int qman_resource_init(struct dev
+ static int fsl_qman_probe(struct platform_device *pdev)
+ {
+ struct device *dev = &pdev->dev;
+- struct device_node *node = dev->of_node;
++ struct device_node *mem_node, *node = dev->of_node;
+ struct resource *res;
+ int ret, err_irq;
+ u16 id;
+ u8 major, minor;
++ u64 size;
+
+ res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
+ if (!res) {
+@@ -717,6 +720,8 @@ static int fsl_qman_probe(struct platfor
+ qman_ip_rev = QMAN_REV30;
+ else if (major == 3 && minor == 1)
+ qman_ip_rev = QMAN_REV31;
++ else if (major == 3 && minor == 2)
++ qman_ip_rev = QMAN_REV32;
+ else {
+ dev_err(dev, "Unknown QMan version\n");
+ return -ENODEV;
+@@ -727,10 +732,83 @@ static int fsl_qman_probe(struct platfor
+ qm_channel_caam = QMAN_CHANNEL_CAAM_REV3;
+ }
+
+- ret = zero_priv_mem(dev, node, fqd_a, fqd_sz);
+- WARN_ON(ret);
+- if (ret)
+- return -ENODEV;
++ if (fqd_a) {
++#ifdef CONFIG_PPC
++ /*
++ * For PPC backward DT compatibility
++ * FQD memory MUST be zero'd by software
++ */
++ zero_priv_mem(fqd_a, fqd_sz);
++#else
++ WARN(1, "Unexpected archiceture using non shared-dma-mem reservations");
++#endif
++ } else {
++ /*
++ * Order of memory regions is assumed as FQD followed by PFDR
++ * in order to ensure allocations from the correct regions the
++ * driver initializes then allocates each piece in order
++ */
++ ret = of_reserved_mem_device_init_by_idx(dev, dev->of_node, 0);
++ if (ret) {
++ dev_err(dev, "of_reserved_mem_device_init_by_idx(0) failed 0x%x\n",
++ ret);
++ return -ENODEV;
++ }
++ mem_node = of_parse_phandle(dev->of_node, "memory-region", 0);
++ if (mem_node) {
++ ret = of_property_read_u64(mem_node, "size", &size);
++ if (ret) {
++ dev_err(dev, "FQD: of_address_to_resource fails 0x%x\n",
++ ret);
++ return -ENODEV;
++ }
++ fqd_sz = size;
++ } else {
++ dev_err(dev, "No memory-region found for FQD\n");
++ return -ENODEV;
++ }
++ if (!dma_zalloc_coherent(dev, fqd_sz, &fqd_a, 0)) {
++ dev_err(dev, "Alloc FQD memory failed\n");
++ return -ENODEV;
++ }
++
++ /*
++ * Disassociate the FQD reserved memory area from the device
++ * because a device can only have one DMA memory area. This
++ * should be fine since the memory is allocated and initialized
++ * and only ever accessed by the QMan device from now on
++ */
++ of_reserved_mem_device_release(dev);
++ }
++ dev_dbg(dev, "Allocated FQD 0x%llx 0x%zx\n", fqd_a, fqd_sz);
++
++ if (!pfdr_a) {
++ /* Setup PFDR memory */
++ ret = of_reserved_mem_device_init_by_idx(dev, dev->of_node, 1);
++ if (ret) {
++ dev_err(dev, "of_reserved_mem_device_init(1) failed 0x%x\n",
++ ret);
++ return -ENODEV;
++ }
++ mem_node = of_parse_phandle(dev->of_node, "memory-region", 1);
++ if (mem_node) {
++ ret = of_property_read_u64(mem_node, "size", &size);
++ if (ret) {
++ dev_err(dev, "PFDR: of_address_to_resource fails 0x%x\n",
++ ret);
++ return -ENODEV;
++ }
++ pfdr_sz = size;
++ } else {
++ dev_err(dev, "No memory-region found for PFDR\n");
++ return -ENODEV;
++ }
++ if (!dma_zalloc_coherent(dev, pfdr_sz, &pfdr_a, 0)) {
++ dev_err(dev, "Alloc PFDR Failed size 0x%zx\n", pfdr_sz);
++ return -ENODEV;
++ }
++ }
++ dev_info(dev, "Allocated PFDR 0x%llx 0x%zx\n", pfdr_a, pfdr_sz);
+
+ ret = qman_init_ccsr(dev);
+ if (ret) {
+--- a/drivers/soc/fsl/qbman/qman_portal.c
++++ b/drivers/soc/fsl/qbman/qman_portal.c
+@@ -262,7 +262,14 @@ static int qman_portal_probe(struct plat
+ }
+ pcfg->irq = irq;
+
+- va = ioremap_prot(addr_phys[0]->start, resource_size(addr_phys[0]), 0);
++#ifdef CONFIG_PPC
++ /* PPC requires a cacheable/non-coherent mapping of the portal */
++ va = ioremap_prot(addr_phys[0]->start, resource_size(addr_phys[0]),
++ (pgprot_val(PAGE_KERNEL) & ~_PAGE_COHERENT));
++#else
++ /* For ARM we can use write combine mapping. */
++ va = ioremap_wc(addr_phys[0]->start, resource_size(addr_phys[0]));
++#endif
+ if (!va) {
+ dev_err(dev, "ioremap::CE failed\n");
+ goto err_ioremap1;
+@@ -270,8 +277,7 @@ static int qman_portal_probe(struct plat
+
+ pcfg->addr_virt[DPAA_PORTAL_CE] = va;
+
+- va = ioremap_prot(addr_phys[1]->start, resource_size(addr_phys[1]),
+- _PAGE_GUARDED | _PAGE_NO_CACHE);
++ va = ioremap(addr_phys[1]->start, resource_size(addr_phys[1]));
+ if (!va) {
+ dev_err(dev, "ioremap::CI failed\n");
+ goto err_ioremap2;
+--- a/drivers/soc/fsl/qbman/qman_priv.h
++++ b/drivers/soc/fsl/qbman/qman_priv.h
+@@ -28,13 +28,13 @@
+ * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
+ */
+
+-#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
+-
+ #include "dpaa_sys.h"
+
+ #include <soc/fsl/qman.h>
+ #include <linux/dma-mapping.h>
+ #include <linux/iommu.h>
++#include <linux/dma-contiguous.h>
++#include <linux/of_address.h>
+
+ #if defined(CONFIG_FSL_PAMU)
+ #include <asm/fsl_pamu_stash.h>
+@@ -187,6 +187,7 @@ struct qm_portal_config {
+ #define QMAN_REV20 0x0200
+ #define QMAN_REV30 0x0300
+ #define QMAN_REV31 0x0301
++#define QMAN_REV32 0x0302
+ extern u16 qman_ip_rev; /* 0 if uninitialised, otherwise QMAN_REVx */
+
+ #define QM_FQID_RANGE_START 1 /* FQID 0 reserved for internal use */
+--- a/drivers/soc/fsl/qbman/qman_test.h
++++ b/drivers/soc/fsl/qbman/qman_test.h
+@@ -30,7 +30,5 @@
+
+ #include "qman_priv.h"
+
+-#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
+-
+ int qman_test_stash(void);
+ int qman_test_api(void);